diff options
339 files changed, 6882 insertions, 4754 deletions
diff --git a/Documentation/ABI/testing/sysfs-driver-xen-blkback b/Documentation/ABI/testing/sysfs-driver-xen-blkback new file mode 100644 index 000000000000..8bb43b66eb55 --- /dev/null +++ b/Documentation/ABI/testing/sysfs-driver-xen-blkback | |||
@@ -0,0 +1,17 @@ | |||
1 | What: /sys/module/xen_blkback/parameters/max_buffer_pages | ||
2 | Date: March 2013 | ||
3 | KernelVersion: 3.11 | ||
4 | Contact: Roger Pau Monné <roger.pau@citrix.com> | ||
5 | Description: | ||
6 | Maximum number of free pages to keep in each block | ||
7 | backend buffer. | ||
8 | |||
9 | What: /sys/module/xen_blkback/parameters/max_persistent_grants | ||
10 | Date: March 2013 | ||
11 | KernelVersion: 3.11 | ||
12 | Contact: Roger Pau Monné <roger.pau@citrix.com> | ||
13 | Description: | ||
14 | Maximum number of grants to map persistently in | ||
15 | blkback. If the frontend tries to use more than | ||
16 | max_persistent_grants, the LRU kicks in and starts | ||
17 | removing 5% of max_persistent_grants every 100ms. | ||
diff --git a/Documentation/ABI/testing/sysfs-driver-xen-blkfront b/Documentation/ABI/testing/sysfs-driver-xen-blkfront new file mode 100644 index 000000000000..c0a6cb7eb314 --- /dev/null +++ b/Documentation/ABI/testing/sysfs-driver-xen-blkfront | |||
@@ -0,0 +1,10 @@ | |||
1 | What: /sys/module/xen_blkfront/parameters/max | ||
2 | Date: June 2013 | ||
3 | KernelVersion: 3.11 | ||
4 | Contact: Konrad Rzeszutek Wilk <konrad.wilk@oracle.com> | ||
5 | Description: | ||
6 | Maximum number of segments that the frontend will negotiate | ||
7 | with the backend for indirect descriptors. The default value | ||
8 | is 32 - higher value means more potential throughput but more | ||
9 | memory usage. The backend picks the minimum of the frontend | ||
10 | and its default backend value. | ||
diff --git a/Documentation/bcache.txt b/Documentation/bcache.txt index c3365f26b2d9..32b6c3189d98 100644 --- a/Documentation/bcache.txt +++ b/Documentation/bcache.txt | |||
@@ -46,29 +46,33 @@ you format your backing devices and cache device at the same time, you won't | |||
46 | have to manually attach: | 46 | have to manually attach: |
47 | make-bcache -B /dev/sda /dev/sdb -C /dev/sdc | 47 | make-bcache -B /dev/sda /dev/sdb -C /dev/sdc |
48 | 48 | ||
49 | To make bcache devices known to the kernel, echo them to /sys/fs/bcache/register: | 49 | bcache-tools now ships udev rules, and bcache devices are known to the kernel |
50 | immediately. Without udev, you can manually register devices like this: | ||
50 | 51 | ||
51 | echo /dev/sdb > /sys/fs/bcache/register | 52 | echo /dev/sdb > /sys/fs/bcache/register |
52 | echo /dev/sdc > /sys/fs/bcache/register | 53 | echo /dev/sdc > /sys/fs/bcache/register |
53 | 54 | ||
54 | To register your bcache devices automatically, you could add something like | 55 | Registering the backing device makes the bcache device show up in /dev; you can |
55 | this to an init script: | 56 | now format it and use it as normal. But the first time using a new bcache |
57 | device, it'll be running in passthrough mode until you attach it to a cache. | ||
58 | See the section on attaching. | ||
56 | 59 | ||
57 | echo /dev/sd* > /sys/fs/bcache/register_quiet | 60 | The devices show up as: |
58 | 61 | ||
59 | It'll look for bcache superblocks and ignore everything that doesn't have one. | 62 | /dev/bcache<N> |
60 | 63 | ||
61 | Registering the backing device makes the bcache show up in /dev; you can now | 64 | As well as (with udev): |
62 | format it and use it as normal. But the first time using a new bcache device, | ||
63 | it'll be running in passthrough mode until you attach it to a cache. See the | ||
64 | section on attaching. | ||
65 | 65 | ||
66 | The devices show up at /dev/bcacheN, and can be controlled via sysfs from | 66 | /dev/bcache/by-uuid/<uuid> |
67 | /sys/block/bcacheN/bcache: | 67 | /dev/bcache/by-label/<label> |
68 | |||
69 | To get started: | ||
68 | 70 | ||
69 | mkfs.ext4 /dev/bcache0 | 71 | mkfs.ext4 /dev/bcache0 |
70 | mount /dev/bcache0 /mnt | 72 | mount /dev/bcache0 /mnt |
71 | 73 | ||
74 | You can control bcache devices through sysfs at /sys/block/bcache<N>/bcache . | ||
75 | |||
72 | Cache devices are managed as sets; multiple caches per set isn't supported yet | 76 | Cache devices are managed as sets; multiple caches per set isn't supported yet |
73 | but will allow for mirroring of metadata and dirty data in the future. Your new | 77 | but will allow for mirroring of metadata and dirty data in the future. Your new |
74 | cache set shows up as /sys/fs/bcache/<UUID> | 78 | cache set shows up as /sys/fs/bcache/<UUID> |
@@ -80,11 +84,11 @@ must be attached to your cache set to enable caching. Attaching a backing | |||
80 | device to a cache set is done thusly, with the UUID of the cache set in | 84 | device to a cache set is done thusly, with the UUID of the cache set in |
81 | /sys/fs/bcache: | 85 | /sys/fs/bcache: |
82 | 86 | ||
83 | echo <UUID> > /sys/block/bcache0/bcache/attach | 87 | echo <CSET-UUID> > /sys/block/bcache0/bcache/attach |
84 | 88 | ||
85 | This only has to be done once. The next time you reboot, just reregister all | 89 | This only has to be done once. The next time you reboot, just reregister all |
86 | your bcache devices. If a backing device has data in a cache somewhere, the | 90 | your bcache devices. If a backing device has data in a cache somewhere, the |
87 | /dev/bcache# device won't be created until the cache shows up - particularly | 91 | /dev/bcache<N> device won't be created until the cache shows up - particularly |
88 | important if you have writeback caching turned on. | 92 | important if you have writeback caching turned on. |
89 | 93 | ||
90 | If you're booting up and your cache device is gone and never coming back, you | 94 | If you're booting up and your cache device is gone and never coming back, you |
@@ -191,6 +195,9 @@ want for getting the best possible numbers when benchmarking. | |||
191 | 195 | ||
192 | SYSFS - BACKING DEVICE: | 196 | SYSFS - BACKING DEVICE: |
193 | 197 | ||
198 | Available at /sys/block/<bdev>/bcache, /sys/block/bcache*/bcache and | ||
199 | (if attached) /sys/fs/bcache/<cset-uuid>/bdev* | ||
200 | |||
194 | attach | 201 | attach |
195 | Echo the UUID of a cache set to this file to enable caching. | 202 | Echo the UUID of a cache set to this file to enable caching. |
196 | 203 | ||
@@ -300,6 +307,8 @@ cache_readaheads | |||
300 | 307 | ||
301 | SYSFS - CACHE SET: | 308 | SYSFS - CACHE SET: |
302 | 309 | ||
310 | Available at /sys/fs/bcache/<cset-uuid> | ||
311 | |||
303 | average_key_size | 312 | average_key_size |
304 | Average data per key in the btree. | 313 | Average data per key in the btree. |
305 | 314 | ||
@@ -390,6 +399,8 @@ trigger_gc | |||
390 | 399 | ||
391 | SYSFS - CACHE DEVICE: | 400 | SYSFS - CACHE DEVICE: |
392 | 401 | ||
402 | Available at /sys/block/<cdev>/bcache | ||
403 | |||
393 | block_size | 404 | block_size |
394 | Minimum granularity of writes - should match hardware sector size. | 405 | Minimum granularity of writes - should match hardware sector size. |
395 | 406 | ||
diff --git a/Documentation/devicetree/bindings/clock/imx27-clock.txt b/Documentation/devicetree/bindings/clock/imx27-clock.txt index ab1a56e9de9d..7a2070393732 100644 --- a/Documentation/devicetree/bindings/clock/imx27-clock.txt +++ b/Documentation/devicetree/bindings/clock/imx27-clock.txt | |||
@@ -98,6 +98,7 @@ clocks and IDs. | |||
98 | fpm 83 | 98 | fpm 83 |
99 | mpll_osc_sel 84 | 99 | mpll_osc_sel 84 |
100 | mpll_sel 85 | 100 | mpll_sel 85 |
101 | spll_gate 86 | ||
101 | 102 | ||
102 | Examples: | 103 | Examples: |
103 | 104 | ||
diff --git a/Documentation/devicetree/bindings/vendor-prefixes.txt b/Documentation/devicetree/bindings/vendor-prefixes.txt index d5a79caec147..366ce9b87240 100644 --- a/Documentation/devicetree/bindings/vendor-prefixes.txt +++ b/Documentation/devicetree/bindings/vendor-prefixes.txt | |||
@@ -26,6 +26,7 @@ est ESTeem Wireless Modems | |||
26 | fsl Freescale Semiconductor | 26 | fsl Freescale Semiconductor |
27 | GEFanuc GE Fanuc Intelligent Platforms Embedded Systems, Inc. | 27 | GEFanuc GE Fanuc Intelligent Platforms Embedded Systems, Inc. |
28 | gef GE Fanuc Intelligent Platforms Embedded Systems, Inc. | 28 | gef GE Fanuc Intelligent Platforms Embedded Systems, Inc. |
29 | hisilicon Hisilicon Limited. | ||
29 | hp Hewlett Packard | 30 | hp Hewlett Packard |
30 | ibm International Business Machines (IBM) | 31 | ibm International Business Machines (IBM) |
31 | idt Integrated Device Technologies, Inc. | 32 | idt Integrated Device Technologies, Inc. |
@@ -43,6 +44,7 @@ nxp NXP Semiconductors | |||
43 | onnn ON Semiconductor Corp. | 44 | onnn ON Semiconductor Corp. |
44 | picochip Picochip Ltd | 45 | picochip Picochip Ltd |
45 | powervr PowerVR (deprecated, use img) | 46 | powervr PowerVR (deprecated, use img) |
47 | qca Qualcomm Atheros, Inc. | ||
46 | qcom Qualcomm, Inc. | 48 | qcom Qualcomm, Inc. |
47 | ralink Mediatek/Ralink Technology Corp. | 49 | ralink Mediatek/Ralink Technology Corp. |
48 | ramtron Ramtron International | 50 | ramtron Ramtron International |
diff --git a/Documentation/ja_JP/HOWTO b/Documentation/ja_JP/HOWTO index 050d37fe6d40..8148a47fc70e 100644 --- a/Documentation/ja_JP/HOWTO +++ b/Documentation/ja_JP/HOWTO | |||
@@ -11,14 +11,14 @@ for non English (read: Japanese) speakers and is not intended as a | |||
11 | fork. So if you have any comments or updates for this file, please try | 11 | fork. So if you have any comments or updates for this file, please try |
12 | to update the original English file first. | 12 | to update the original English file first. |
13 | 13 | ||
14 | Last Updated: 2011/03/31 | 14 | Last Updated: 2013/07/19 |
15 | ================================== | 15 | ================================== |
16 | ã“ã‚Œã¯ã€ | 16 | ã“ã‚Œã¯ã€ |
17 | linux-2.6.38/Documentation/HOWTO | 17 | linux-3.10/Documentation/HOWTO |
18 | ã®å’Œè¨³ã§ã™ã€‚ | 18 | ã®å’Œè¨³ã§ã™ã€‚ |
19 | 19 | ||
20 | 翻訳団体: JF プãƒã‚¸ã‚§ã‚¯ãƒˆ < http://www.linux.or.jp/JF/ > | 20 | 翻訳団体: JF プãƒã‚¸ã‚§ã‚¯ãƒˆ < http://linuxjf.sourceforge.jp/ > |
21 | 翻訳日: 2011/3/28 | 21 | 翻訳日: 2013/7/19 |
22 | 翻訳者: Tsugikazu Shibata <tshibata at ab dot jp dot nec dot com> | 22 | 翻訳者: Tsugikazu Shibata <tshibata at ab dot jp dot nec dot com> |
23 | æ ¡æ£è€…: æ¾å€‰ã•ã‚“ <nbh--mats at nifty dot com> | 23 | æ ¡æ£è€…: æ¾å€‰ã•ã‚“ <nbh--mats at nifty dot com> |
24 | å°æž— é›…å…¸ã•ã‚“ (Masanori Kobayasi) <zap03216 at nifty dot ne dot jp> | 24 | å°æž— é›…å…¸ã•ã‚“ (Masanori Kobayasi) <zap03216 at nifty dot ne dot jp> |
@@ -245,7 +245,7 @@ Linux カーãƒãƒ«ã‚½ãƒ¼ã‚¹ãƒ„リーã®ä¸ã«å«ã¾ã‚Œã‚‹ã€ãã‚Œã„ã«ã—ã€ä¿ | |||
245 | 自己å‚照方å¼ã§ã€ç´¢å¼•ãŒã¤ã„㟠web å½¢å¼ã§ã€ã‚½ãƒ¼ã‚¹ã‚³ãƒ¼ãƒ‰ã‚’å‚ç…§ã™ã‚‹ã“ã¨ãŒ | 245 | 自己å‚照方å¼ã§ã€ç´¢å¼•ãŒã¤ã„㟠web å½¢å¼ã§ã€ã‚½ãƒ¼ã‚¹ã‚³ãƒ¼ãƒ‰ã‚’å‚ç…§ã™ã‚‹ã“ã¨ãŒ |
246 | ã§ãã¾ã™ã€‚ã“ã®æœ€æ–°ã®ç´ æ™´ã—ã„カーãƒãƒ«ã‚³ãƒ¼ãƒ‰ã®ãƒªãƒã‚¸ãƒˆãƒªã¯ä»¥ä¸‹ã§è¦‹ã¤ã‹ã‚Š | 246 | ã§ãã¾ã™ã€‚ã“ã®æœ€æ–°ã®ç´ æ™´ã—ã„カーãƒãƒ«ã‚³ãƒ¼ãƒ‰ã®ãƒªãƒã‚¸ãƒˆãƒªã¯ä»¥ä¸‹ã§è¦‹ã¤ã‹ã‚Š |
247 | ã¾ã™- | 247 | ã¾ã™- |
248 | http://sosdg.org/~qiyong/lxr/ | 248 | http://lxr.linux.no/+trees |
249 | 249 | ||
250 | 開発プãƒã‚»ã‚¹ | 250 | 開発プãƒã‚»ã‚¹ |
251 | ----------------------- | 251 | ----------------------- |
@@ -253,24 +253,24 @@ Linux カーãƒãƒ«ã‚½ãƒ¼ã‚¹ãƒ„リーã®ä¸ã«å«ã¾ã‚Œã‚‹ã€ãã‚Œã„ã«ã—ã€ä¿ | |||
253 | Linux カーãƒãƒ«ã®é–‹ç™ºãƒ—ãƒã‚»ã‚¹ã¯ç¾åœ¨å¹¾ã¤ã‹ã®ç•°ãªã‚‹ãƒ¡ã‚¤ãƒ³ã‚«ãƒ¼ãƒãƒ«ã€Œãƒ–ラン | 253 | Linux カーãƒãƒ«ã®é–‹ç™ºãƒ—ãƒã‚»ã‚¹ã¯ç¾åœ¨å¹¾ã¤ã‹ã®ç•°ãªã‚‹ãƒ¡ã‚¤ãƒ³ã‚«ãƒ¼ãƒãƒ«ã€Œãƒ–ラン |
254 | ãƒã€ã¨å¤šæ•°ã®ã‚µãƒ–システム毎ã®ã‚«ãƒ¼ãƒãƒ«ãƒ–ランãƒã‹ã‚‰æ§‹æˆã•ã‚Œã¾ã™ã€‚ | 254 | ãƒã€ã¨å¤šæ•°ã®ã‚µãƒ–システム毎ã®ã‚«ãƒ¼ãƒãƒ«ãƒ–ランãƒã‹ã‚‰æ§‹æˆã•ã‚Œã¾ã™ã€‚ |
255 | ã“れらã®ãƒ–ランãƒã¨ã¯- | 255 | ã“れらã®ãƒ–ランãƒã¨ã¯- |
256 | - メイン㮠2.6.x カーãƒãƒ«ãƒ„リー | 256 | - メイン㮠3.x カーãƒãƒ«ãƒ„リー |
257 | - 2.6.x.y -stable カーãƒãƒ«ãƒ„リー | 257 | - 3.x.y -stable カーãƒãƒ«ãƒ„リー |
258 | - 2.6.x -git カーãƒãƒ«ãƒ‘ッム| 258 | - 3.x -git カーãƒãƒ«ãƒ‘ッム|
259 | - サブシステム毎ã®ã‚«ãƒ¼ãƒãƒ«ãƒ„リーã¨ãƒ‘ッム| 259 | - サブシステム毎ã®ã‚«ãƒ¼ãƒãƒ«ãƒ„リーã¨ãƒ‘ッム|
260 | - çµ±åˆãƒ†ã‚¹ãƒˆã®ãŸã‚ã® 2.6.x -next カーãƒãƒ«ãƒ„リー | 260 | - çµ±åˆãƒ†ã‚¹ãƒˆã®ãŸã‚ã® 3.x -next カーãƒãƒ«ãƒ„リー |
261 | 261 | ||
262 | 2.6.x カーãƒãƒ«ãƒ„リー | 262 | 3.x カーãƒãƒ«ãƒ„リー |
263 | ----------------- | 263 | ----------------- |
264 | 264 | ||
265 | 2.6.x カーãƒãƒ«ã¯ Linus Torvalds ã«ã‚ˆã£ã¦ãƒ¡ãƒ³ãƒ†ãƒŠãƒ³ã‚¹ã•ã‚Œã€kernel.org | 265 | 3.x カーãƒãƒ«ã¯ Linus Torvalds ã«ã‚ˆã£ã¦ãƒ¡ãƒ³ãƒ†ãƒŠãƒ³ã‚¹ã•ã‚Œã€kernel.org |
266 | ã® pub/linux/kernel/v2.6/ ディレクトリã«å˜åœ¨ã—ã¾ã™ã€‚ã“ã®é–‹ç™ºãƒ—ãƒã‚»ã‚¹ã¯ | 266 | ã® pub/linux/kernel/v3.x/ ディレクトリã«å˜åœ¨ã—ã¾ã™ã€‚ã“ã®é–‹ç™ºãƒ—ãƒã‚»ã‚¹ã¯ |
267 | 以下ã®ã¨ãŠã‚Š- | 267 | 以下ã®ã¨ãŠã‚Š- |
268 | 268 | ||
269 | - æ–°ã—ã„カーãƒãƒ«ãŒãƒªãƒªãƒ¼ã‚¹ã•ã‚ŒãŸç›´å¾Œã«ã€2週間ã®ç‰¹åˆ¥æœŸé–“ãŒè¨ã‘られ〠| 269 | - æ–°ã—ã„カーãƒãƒ«ãŒãƒªãƒªãƒ¼ã‚¹ã•ã‚ŒãŸç›´å¾Œã«ã€2週間ã®ç‰¹åˆ¥æœŸé–“ãŒè¨ã‘られ〠|
270 | ã“ã®æœŸé–“ä¸ã«ã€ãƒ¡ãƒ³ãƒ†ãƒŠé”㯠Linus ã«å¤§ããªå·®åˆ†ã‚’é€ã‚‹ã“ã¨ãŒã§ãã¾ã™ã€‚ | 270 | ã“ã®æœŸé–“ä¸ã«ã€ãƒ¡ãƒ³ãƒ†ãƒŠé”㯠Linus ã«å¤§ããªå·®åˆ†ã‚’é€ã‚‹ã“ã¨ãŒã§ãã¾ã™ã€‚ |
271 | ã“ã®ã‚ˆã†ãªå·®åˆ†ã¯é€šå¸¸ -next カーãƒãƒ«ã«æ•°é€±é–“å«ã¾ã‚Œã¦ããŸãƒ‘ッãƒã§ã™ã€‚ | 271 | ã“ã®ã‚ˆã†ãªå·®åˆ†ã¯é€šå¸¸ -next カーãƒãƒ«ã«æ•°é€±é–“å«ã¾ã‚Œã¦ããŸãƒ‘ッãƒã§ã™ã€‚ |
272 | 大ããªå¤‰æ›´ã¯ git(カーãƒãƒ«ã®ã‚½ãƒ¼ã‚¹ç®¡ç†ãƒ„ールã€è©³ç´°ã¯ | 272 | 大ããªå¤‰æ›´ã¯ git(カーãƒãƒ«ã®ã‚½ãƒ¼ã‚¹ç®¡ç†ãƒ„ールã€è©³ç´°ã¯ |
273 | http://git-scm.com/ å‚ç…§) を使ã£ã¦é€ã‚‹ã®ãŒå¥½ã¾ã—ã„ã‚„ã‚Šæ–¹ã§ã™ãŒã€ãƒ‘ッ | 273 | http://git-scm.com/ å‚ç…§) を使ã£ã¦é€ã‚‹ã®ãŒå¥½ã¾ã—ã„ã‚„ã‚Šæ–¹ã§ã™ãŒã€ãƒ‘ッ |
274 | ãƒãƒ•ã‚¡ã‚¤ãƒ«ã®å½¢å¼ã®ã¾ã¾é€ã‚‹ã®ã§ã‚‚å分ã§ã™ã€‚ | 274 | ãƒãƒ•ã‚¡ã‚¤ãƒ«ã®å½¢å¼ã®ã¾ã¾é€ã‚‹ã®ã§ã‚‚å分ã§ã™ã€‚ |
275 | 275 | ||
276 | - 2週間後ã€-rc1 カーãƒãƒ«ãŒãƒªãƒªãƒ¼ã‚¹ã•ã‚Œã€ã“ã®å¾Œã«ã¯ã‚«ãƒ¼ãƒãƒ«å…¨ä½“ã®å®‰å®š | 276 | - 2週間後ã€-rc1 カーãƒãƒ«ãŒãƒªãƒªãƒ¼ã‚¹ã•ã‚Œã€ã“ã®å¾Œã«ã¯ã‚«ãƒ¼ãƒãƒ«å…¨ä½“ã®å®‰å®š |
@@ -302,20 +302,20 @@ Andrew Morton ㌠Linux-kernel メーリングリストã«ã‚«ãƒ¼ãƒãƒ«ãƒªãƒªãƒ¼ã | |||
302 | 実ã«èªè˜ã•ã‚ŒãŸãƒã‚°ã®çŠ¶æ³ã«ã‚ˆã‚Šãƒªãƒªãƒ¼ã‚¹ã•ã‚Œã‚‹ã®ã§ã‚ã‚Šã€å‰ã‚‚ã£ã¦æ±ºã‚ら | 302 | 実ã«èªè˜ã•ã‚ŒãŸãƒã‚°ã®çŠ¶æ³ã«ã‚ˆã‚Šãƒªãƒªãƒ¼ã‚¹ã•ã‚Œã‚‹ã®ã§ã‚ã‚Šã€å‰ã‚‚ã£ã¦æ±ºã‚ら |
303 | ã‚ŒãŸè¨ˆç”»ã«ã‚ˆã£ã¦ãƒªãƒªãƒ¼ã‚¹ã•ã‚Œã‚‹ã‚‚ã®ã§ã¯ãªã„ã‹ã‚‰ã§ã™ã€‚〠| 303 | ã‚ŒãŸè¨ˆç”»ã«ã‚ˆã£ã¦ãƒªãƒªãƒ¼ã‚¹ã•ã‚Œã‚‹ã‚‚ã®ã§ã¯ãªã„ã‹ã‚‰ã§ã™ã€‚〠|
304 | 304 | ||
305 | 2.6.x.y -stable カーãƒãƒ«ãƒ„リー | 305 | 3.x.y -stable カーãƒãƒ«ãƒ„リー |
306 | --------------------------- | 306 | --------------------------- |
307 | 307 | ||
308 | ãƒãƒ¼ã‚¸ãƒ§ãƒ³ç•ªå·ãŒ4ã¤ã®æ•°å—ã«åˆ†ã‹ã‚Œã¦ã„るカーãƒãƒ«ã¯ -stable カーãƒãƒ«ã§ã™ã€‚ | 308 | ãƒãƒ¼ã‚¸ãƒ§ãƒ³ç•ªå·ãŒ3ã¤ã®æ•°å—ã«åˆ†ã‹ã‚Œã¦ã„るカーãƒãƒ«ã¯ -stable カーãƒãƒ«ã§ã™ã€‚ |
309 | ã“ã‚Œã«ã¯ã€2.6.x カーãƒãƒ«ã§è¦‹ã¤ã‹ã£ãŸã‚»ã‚ュリティå•é¡Œã‚„é‡å¤§ãªå¾Œæˆ»ã‚Šã«å¯¾ | 309 | ã“ã‚Œã«ã¯ã€3.x カーãƒãƒ«ã§è¦‹ã¤ã‹ã£ãŸã‚»ã‚ュリティå•é¡Œã‚„é‡å¤§ãªå¾Œæˆ»ã‚Šã«å¯¾ |
310 | ã™ã‚‹æ¯”較的å°ã•ã„é‡è¦ãªä¿®æ£ãŒå«ã¾ã‚Œã¾ã™ã€‚ | 310 | ã™ã‚‹æ¯”較的å°ã•ã„é‡è¦ãªä¿®æ£ãŒå«ã¾ã‚Œã¾ã™ã€‚ |
311 | 311 | ||
312 | ã“ã‚Œã¯ã€é–‹ç™º/実験的ãƒãƒ¼ã‚¸ãƒ§ãƒ³ã®ãƒ†ã‚¹ãƒˆã«å”力ã™ã‚‹ã“ã¨ã«èˆˆå‘³ãŒç„¡ã〠| 312 | ã“ã‚Œã¯ã€é–‹ç™º/実験的ãƒãƒ¼ã‚¸ãƒ§ãƒ³ã®ãƒ†ã‚¹ãƒˆã«å”力ã™ã‚‹ã“ã¨ã«èˆˆå‘³ãŒç„¡ã〠|
313 | 最新ã®å®‰å®šã—ãŸã‚«ãƒ¼ãƒãƒ«ã‚’使ã„ãŸã„ユーザã«æŽ¨å¥¨ã™ã‚‹ãƒ–ランãƒã§ã™ã€‚ | 313 | 最新ã®å®‰å®šã—ãŸã‚«ãƒ¼ãƒãƒ«ã‚’使ã„ãŸã„ユーザã«æŽ¨å¥¨ã™ã‚‹ãƒ–ランãƒã§ã™ã€‚ |
314 | 314 | ||
315 | ã‚‚ã—ã€2.6.x.y カーãƒãƒ«ãŒå˜åœ¨ã—ãªã„å ´åˆã«ã¯ã€ç•ªå·ãŒä¸€ç•ªå¤§ãã„ 2.6.x ㌠| 315 | ã‚‚ã—ã€3.x.y カーãƒãƒ«ãŒå˜åœ¨ã—ãªã„å ´åˆã«ã¯ã€ç•ªå·ãŒä¸€ç•ªå¤§ãã„ 3.x ㌠|
316 | 最新ã®å®‰å®šç‰ˆã‚«ãƒ¼ãƒãƒ«ã§ã™ã€‚ | 316 | 最新ã®å®‰å®šç‰ˆã‚«ãƒ¼ãƒãƒ«ã§ã™ã€‚ |
317 | 317 | ||
318 | 2.6.x.y 㯠"stable" ãƒãƒ¼ãƒ <stable@kernel.org> ã§ãƒ¡ãƒ³ãƒ†ã•ã‚Œã¦ãŠã‚Šã€å¿… | 318 | 3.x.y 㯠"stable" ãƒãƒ¼ãƒ <stable@kernel.org> ã§ãƒ¡ãƒ³ãƒ†ã•ã‚Œã¦ãŠã‚Šã€å¿… |
319 | è¦ã«å¿œã˜ã¦ãƒªãƒªãƒ¼ã‚¹ã•ã‚Œã¾ã™ã€‚通常ã®ãƒªãƒªãƒ¼ã‚¹æœŸé–“㯠2週間毎ã§ã™ãŒã€å·®ã—迫㣠| 319 | è¦ã«å¿œã˜ã¦ãƒªãƒªãƒ¼ã‚¹ã•ã‚Œã¾ã™ã€‚通常ã®ãƒªãƒªãƒ¼ã‚¹æœŸé–“㯠2週間毎ã§ã™ãŒã€å·®ã—迫㣠|
320 | ãŸå•é¡ŒãŒãªã‘ã‚Œã°ã‚‚ã†å°‘ã—é•·ããªã‚‹ã“ã¨ã‚‚ã‚ã‚Šã¾ã™ã€‚ã‚»ã‚ュリティ関連ã®å•é¡Œ | 320 | ãŸå•é¡ŒãŒãªã‘ã‚Œã°ã‚‚ã†å°‘ã—é•·ããªã‚‹ã“ã¨ã‚‚ã‚ã‚Šã¾ã™ã€‚ã‚»ã‚ュリティ関連ã®å•é¡Œ |
321 | ã®å ´åˆã¯ã“ã‚Œã«å¯¾ã—ã¦ã ã„ãŸã„ã®å ´åˆã€ã™ãã«ãƒªãƒªãƒ¼ã‚¹ãŒã•ã‚Œã¾ã™ã€‚ | 321 | ã®å ´åˆã¯ã“ã‚Œã«å¯¾ã—ã¦ã ã„ãŸã„ã®å ´åˆã€ã™ãã«ãƒªãƒªãƒ¼ã‚¹ãŒã•ã‚Œã¾ã™ã€‚ |
@@ -324,7 +324,7 @@ Andrew Morton ㌠Linux-kernel メーリングリストã«ã‚«ãƒ¼ãƒãƒ«ãƒªãƒªãƒ¼ã | |||
324 | イルã«ã¯ã©ã®ã‚ˆã†ãªç¨®é¡žã®å¤‰æ›´ãŒ -stable ツリーã«å—ã‘入れå¯èƒ½ã‹ã€ã¾ãŸãƒª | 324 | イルã«ã¯ã©ã®ã‚ˆã†ãªç¨®é¡žã®å¤‰æ›´ãŒ -stable ツリーã«å—ã‘入れå¯èƒ½ã‹ã€ã¾ãŸãƒª |
325 | リースプãƒã‚»ã‚¹ãŒã©ã†å‹•ãã‹ãŒè¨˜è¿°ã•ã‚Œã¦ã„ã¾ã™ã€‚ | 325 | リースプãƒã‚»ã‚¹ãŒã©ã†å‹•ãã‹ãŒè¨˜è¿°ã•ã‚Œã¦ã„ã¾ã™ã€‚ |
326 | 326 | ||
327 | 2.6.x -git パッム| 327 | 3.x -git パッム|
328 | ------------------ | 328 | ------------------ |
329 | 329 | ||
330 | git リãƒã‚¸ãƒˆãƒªã§ç®¡ç†ã•ã‚Œã¦ã„ã‚‹Linus ã®ã‚«ãƒ¼ãƒãƒ«ãƒ„リーã®æ¯Žæ—¥ã®ã‚¹ãƒŠãƒƒãƒ— | 330 | git リãƒã‚¸ãƒˆãƒªã§ç®¡ç†ã•ã‚Œã¦ã„ã‚‹Linus ã®ã‚«ãƒ¼ãƒãƒ«ãƒ„リーã®æ¯Žæ—¥ã®ã‚¹ãƒŠãƒƒãƒ— |
@@ -358,14 +358,14 @@ quilt シリーズã¨ã—ã¦å…¬é–‹ã•ã‚Œã¦ã„るパッãƒã‚ューも使ã‚ã‚Œã | |||
358 | ã‚’ã¤ã‘ã‚‹ã“ã¨ãŒã§ãã¾ã™ã€‚大部分ã®ã“れら㮠patchwork ã®ã‚µã‚¤ãƒˆã¯ | 358 | ã‚’ã¤ã‘ã‚‹ã“ã¨ãŒã§ãã¾ã™ã€‚大部分ã®ã“れら㮠patchwork ã®ã‚µã‚¤ãƒˆã¯ |
359 | http://patchwork.kernel.org/ ã§ãƒªã‚¹ãƒˆã•ã‚Œã¦ã„ã¾ã™ã€‚ | 359 | http://patchwork.kernel.org/ ã§ãƒªã‚¹ãƒˆã•ã‚Œã¦ã„ã¾ã™ã€‚ |
360 | 360 | ||
361 | çµ±åˆãƒ†ã‚¹ãƒˆã®ãŸã‚ã® 2.6.x -next カーãƒãƒ«ãƒ„リー | 361 | çµ±åˆãƒ†ã‚¹ãƒˆã®ãŸã‚ã® 3.x -next カーãƒãƒ«ãƒ„リー |
362 | --------------------------------------------- | 362 | --------------------------------------------- |
363 | 363 | ||
364 | サブシステムツリーã®æ›´æ–°å†…容ãŒãƒ¡ã‚¤ãƒ³ãƒ©ã‚¤ãƒ³ã® 2.6.x ツリーã«ãƒžãƒ¼ã‚¸ã•ã‚Œ | 364 | サブシステムツリーã®æ›´æ–°å†…容ãŒãƒ¡ã‚¤ãƒ³ãƒ©ã‚¤ãƒ³ã® 3.x ツリーã«ãƒžãƒ¼ã‚¸ã•ã‚Œ |
365 | ã‚‹å‰ã«ã€ãれらã¯çµ±åˆãƒ†ã‚¹ãƒˆã•ã‚Œã‚‹å¿…è¦ãŒã‚ã‚Šã¾ã™ã€‚ã“ã®ç›®çš„ã®ãŸã‚ã€å®Ÿè³ªçš„ | 365 | ã‚‹å‰ã«ã€ãれらã¯çµ±åˆãƒ†ã‚¹ãƒˆã•ã‚Œã‚‹å¿…è¦ãŒã‚ã‚Šã¾ã™ã€‚ã“ã®ç›®çš„ã®ãŸã‚ã€å®Ÿè³ªçš„ |
366 | ã«å…¨ã‚µãƒ–システムツリーã‹ã‚‰ã»ã¼æ¯Žæ—¥ãƒ—ルã•ã‚Œã¦ã§ãる特別ãªãƒ†ã‚¹ãƒˆç”¨ã®ãƒª | 366 | ã«å…¨ã‚µãƒ–システムツリーã‹ã‚‰ã»ã¼æ¯Žæ—¥ãƒ—ルã•ã‚Œã¦ã§ãる特別ãªãƒ†ã‚¹ãƒˆç”¨ã®ãƒª |
367 | ãƒã‚¸ãƒˆãƒªãŒå˜åœ¨ã—ã¾ã™- | 367 | ãƒã‚¸ãƒˆãƒªãŒå˜åœ¨ã—ã¾ã™- |
368 | http://git.kernel.org/?p=linux/kernel/git/sfr/linux-next.git | 368 | http://git.kernel.org/?p=linux/kernel/git/next/linux-next.git |
369 | http://linux.f-seidel.de/linux-next/pmwiki/ | 369 | http://linux.f-seidel.de/linux-next/pmwiki/ |
370 | 370 | ||
371 | ã“ã®ã‚„ã‚Šæ–¹ã«ã‚ˆã£ã¦ã€-next カーãƒãƒ«ã¯æ¬¡ã®ãƒžãƒ¼ã‚¸æ©Ÿä¼šã§ã©ã‚“ãªã‚‚ã®ãŒãƒ¡ã‚¤ãƒ³ | 371 | ã“ã®ã‚„ã‚Šæ–¹ã«ã‚ˆã£ã¦ã€-next カーãƒãƒ«ã¯æ¬¡ã®ãƒžãƒ¼ã‚¸æ©Ÿä¼šã§ã©ã‚“ãªã‚‚ã®ãŒãƒ¡ã‚¤ãƒ³ |
diff --git a/MAINTAINERS b/MAINTAINERS index bf61e04291ab..a26b10e52aea 100644 --- a/MAINTAINERS +++ b/MAINTAINERS | |||
@@ -1642,7 +1642,7 @@ S: Maintained | |||
1642 | F: drivers/net/hamradio/baycom* | 1642 | F: drivers/net/hamradio/baycom* |
1643 | 1643 | ||
1644 | BCACHE (BLOCK LAYER CACHE) | 1644 | BCACHE (BLOCK LAYER CACHE) |
1645 | M: Kent Overstreet <koverstreet@google.com> | 1645 | M: Kent Overstreet <kmo@daterainc.com> |
1646 | L: linux-bcache@vger.kernel.org | 1646 | L: linux-bcache@vger.kernel.org |
1647 | W: http://bcache.evilpiepirate.org | 1647 | W: http://bcache.evilpiepirate.org |
1648 | S: Maintained: | 1648 | S: Maintained: |
@@ -3346,7 +3346,7 @@ F: Documentation/firmware_class/ | |||
3346 | F: drivers/base/firmware*.c | 3346 | F: drivers/base/firmware*.c |
3347 | F: include/linux/firmware.h | 3347 | F: include/linux/firmware.h |
3348 | 3348 | ||
3349 | FLASHSYSTEM DRIVER (IBM FlashSystem 70/80 PCI SSD Flash Card) | 3349 | FLASH ADAPTER DRIVER (IBM Flash Adapter 900GB Full Height PCI Flash Card) |
3350 | M: Joshua Morris <josh.h.morris@us.ibm.com> | 3350 | M: Joshua Morris <josh.h.morris@us.ibm.com> |
3351 | M: Philip Kelleher <pjk1939@linux.vnet.ibm.com> | 3351 | M: Philip Kelleher <pjk1939@linux.vnet.ibm.com> |
3352 | S: Maintained | 3352 | S: Maintained |
@@ -3622,11 +3622,9 @@ F: drivers/isdn/gigaset/ | |||
3622 | F: include/uapi/linux/gigaset_dev.h | 3622 | F: include/uapi/linux/gigaset_dev.h |
3623 | 3623 | ||
3624 | GPIO SUBSYSTEM | 3624 | GPIO SUBSYSTEM |
3625 | M: Grant Likely <grant.likely@linaro.org> | ||
3626 | M: Linus Walleij <linus.walleij@linaro.org> | 3625 | M: Linus Walleij <linus.walleij@linaro.org> |
3627 | S: Maintained | 3626 | S: Maintained |
3628 | L: linux-gpio@vger.kernel.org | 3627 | L: linux-gpio@vger.kernel.org |
3629 | T: git git://git.secretlab.ca/git/linux-2.6.git | ||
3630 | F: Documentation/gpio.txt | 3628 | F: Documentation/gpio.txt |
3631 | F: drivers/gpio/ | 3629 | F: drivers/gpio/ |
3632 | F: include/linux/gpio* | 3630 | F: include/linux/gpio* |
@@ -4472,8 +4470,6 @@ F: drivers/irqchip/ | |||
4472 | 4470 | ||
4473 | IRQ DOMAINS (IRQ NUMBER MAPPING LIBRARY) | 4471 | IRQ DOMAINS (IRQ NUMBER MAPPING LIBRARY) |
4474 | M: Benjamin Herrenschmidt <benh@kernel.crashing.org> | 4472 | M: Benjamin Herrenschmidt <benh@kernel.crashing.org> |
4475 | M: Grant Likely <grant.likely@linaro.org> | ||
4476 | T: git git://git.secretlab.ca/git/linux-2.6.git irqdomain/next | ||
4477 | S: Maintained | 4473 | S: Maintained |
4478 | F: Documentation/IRQ-domain.txt | 4474 | F: Documentation/IRQ-domain.txt |
4479 | F: include/linux/irqdomain.h | 4475 | F: include/linux/irqdomain.h |
@@ -4990,7 +4986,7 @@ F: arch/powerpc/platforms/44x/ | |||
4990 | 4986 | ||
4991 | LINUX FOR POWERPC EMBEDDED XILINX VIRTEX | 4987 | LINUX FOR POWERPC EMBEDDED XILINX VIRTEX |
4992 | L: linuxppc-dev@lists.ozlabs.org | 4988 | L: linuxppc-dev@lists.ozlabs.org |
4993 | S: Unmaintained | 4989 | S: Orphan |
4994 | F: arch/powerpc/*/*virtex* | 4990 | F: arch/powerpc/*/*virtex* |
4995 | F: arch/powerpc/*/*/*virtex* | 4991 | F: arch/powerpc/*/*/*virtex* |
4996 | 4992 | ||
@@ -5886,7 +5882,7 @@ OMAP DEVICE TREE SUPPORT | |||
5886 | M: Benoît Cousson <b-cousson@ti.com> | 5882 | M: Benoît Cousson <b-cousson@ti.com> |
5887 | M: Tony Lindgren <tony@atomide.com> | 5883 | M: Tony Lindgren <tony@atomide.com> |
5888 | L: linux-omap@vger.kernel.org | 5884 | L: linux-omap@vger.kernel.org |
5889 | L: devicetree-discuss@lists.ozlabs.org (moderated for non-subscribers) | 5885 | L: devicetree@vger.kernel.org |
5890 | S: Maintained | 5886 | S: Maintained |
5891 | F: arch/arm/boot/dts/*omap* | 5887 | F: arch/arm/boot/dts/*omap* |
5892 | F: arch/arm/boot/dts/*am3* | 5888 | F: arch/arm/boot/dts/*am3* |
@@ -6050,17 +6046,28 @@ F: drivers/i2c/busses/i2c-ocores.c | |||
6050 | OPEN FIRMWARE AND FLATTENED DEVICE TREE | 6046 | OPEN FIRMWARE AND FLATTENED DEVICE TREE |
6051 | M: Grant Likely <grant.likely@linaro.org> | 6047 | M: Grant Likely <grant.likely@linaro.org> |
6052 | M: Rob Herring <rob.herring@calxeda.com> | 6048 | M: Rob Herring <rob.herring@calxeda.com> |
6053 | L: devicetree-discuss@lists.ozlabs.org (moderated for non-subscribers) | 6049 | L: devicetree@vger.kernel.org |
6054 | W: http://fdt.secretlab.ca | 6050 | W: http://fdt.secretlab.ca |
6055 | T: git git://git.secretlab.ca/git/linux-2.6.git | 6051 | T: git git://git.secretlab.ca/git/linux-2.6.git |
6056 | S: Maintained | 6052 | S: Maintained |
6057 | F: Documentation/devicetree | 6053 | F: drivers/of/ |
6058 | F: drivers/of | ||
6059 | F: include/linux/of*.h | 6054 | F: include/linux/of*.h |
6060 | F: scripts/dtc | 6055 | F: scripts/dtc/ |
6061 | K: of_get_property | 6056 | K: of_get_property |
6062 | K: of_match_table | 6057 | K: of_match_table |
6063 | 6058 | ||
6059 | OPEN FIRMWARE AND FLATTENED DEVICE TREE BINDINGS | ||
6060 | M: Rob Herring <rob.herring@calxeda.com> | ||
6061 | M: Pawel Moll <pawel.moll@arm.com> | ||
6062 | M: Mark Rutland <mark.rutland@arm.com> | ||
6063 | M: Stephen Warren <swarren@wwwdotorg.org> | ||
6064 | M: Ian Campbell <ian.campbell@citrix.com> | ||
6065 | L: devicetree@vger.kernel.org | ||
6066 | S: Maintained | ||
6067 | F: Documentation/devicetree/ | ||
6068 | F: arch/*/boot/dts/ | ||
6069 | F: include/dt-bindings/ | ||
6070 | |||
6064 | OPENRISC ARCHITECTURE | 6071 | OPENRISC ARCHITECTURE |
6065 | M: Jonas Bonn <jonas@southpole.se> | 6072 | M: Jonas Bonn <jonas@southpole.se> |
6066 | W: http://openrisc.net | 6073 | W: http://openrisc.net |
@@ -7746,7 +7753,6 @@ F: drivers/clk/spear/ | |||
7746 | 7753 | ||
7747 | SPI SUBSYSTEM | 7754 | SPI SUBSYSTEM |
7748 | M: Mark Brown <broonie@kernel.org> | 7755 | M: Mark Brown <broonie@kernel.org> |
7749 | M: Grant Likely <grant.likely@linaro.org> | ||
7750 | L: linux-spi@vger.kernel.org | 7756 | L: linux-spi@vger.kernel.org |
7751 | T: git git://git.kernel.org/pub/scm/linux/kernel/git/broonie/spi.git | 7757 | T: git git://git.kernel.org/pub/scm/linux/kernel/git/broonie/spi.git |
7752 | Q: http://patchwork.kernel.org/project/spi-devel-general/list/ | 7758 | Q: http://patchwork.kernel.org/project/spi-devel-general/list/ |
@@ -7812,7 +7818,7 @@ F: drivers/staging/asus_oled/ | |||
7812 | 7818 | ||
7813 | STAGING - COMEDI | 7819 | STAGING - COMEDI |
7814 | M: Ian Abbott <abbotti@mev.co.uk> | 7820 | M: Ian Abbott <abbotti@mev.co.uk> |
7815 | M: Mori Hess <fmhess@users.sourceforge.net> | 7821 | M: H Hartley Sweeten <hsweeten@visionengravers.com> |
7816 | S: Odd Fixes | 7822 | S: Odd Fixes |
7817 | F: drivers/staging/comedi/ | 7823 | F: drivers/staging/comedi/ |
7818 | 7824 | ||
@@ -9288,7 +9294,7 @@ S: Maintained | |||
9288 | F: drivers/net/ethernet/xilinx/xilinx_axienet* | 9294 | F: drivers/net/ethernet/xilinx/xilinx_axienet* |
9289 | 9295 | ||
9290 | XILINX SYSTEMACE DRIVER | 9296 | XILINX SYSTEMACE DRIVER |
9291 | S: Unmaintained | 9297 | S: Orphan |
9292 | F: drivers/block/xsysace.c | 9298 | F: drivers/block/xsysace.c |
9293 | 9299 | ||
9294 | XILINX UARTLITE SERIAL DRIVER | 9300 | XILINX UARTLITE SERIAL DRIVER |
@@ -1,7 +1,7 @@ | |||
1 | VERSION = 3 | 1 | VERSION = 3 |
2 | PATCHLEVEL = 11 | 2 | PATCHLEVEL = 11 |
3 | SUBLEVEL = 0 | 3 | SUBLEVEL = 0 |
4 | EXTRAVERSION = -rc2 | 4 | EXTRAVERSION = -rc3 |
5 | NAME = Linux for Workgroups | 5 | NAME = Linux for Workgroups |
6 | 6 | ||
7 | # *DOCUMENTATION* | 7 | # *DOCUMENTATION* |
diff --git a/arch/alpha/Kconfig b/arch/alpha/Kconfig index 837a1f2d8b96..082d9b4b5472 100644 --- a/arch/alpha/Kconfig +++ b/arch/alpha/Kconfig | |||
@@ -15,6 +15,7 @@ config ALPHA | |||
15 | select ARCH_WANT_OPTIONAL_GPIOLIB | 15 | select ARCH_WANT_OPTIONAL_GPIOLIB |
16 | select ARCH_WANT_IPC_PARSE_VERSION | 16 | select ARCH_WANT_IPC_PARSE_VERSION |
17 | select ARCH_HAVE_NMI_SAFE_CMPXCHG | 17 | select ARCH_HAVE_NMI_SAFE_CMPXCHG |
18 | select ARCH_HAS_ATOMIC64_DEC_IF_POSITIVE | ||
18 | select GENERIC_SMP_IDLE_THREAD | 19 | select GENERIC_SMP_IDLE_THREAD |
19 | select GENERIC_CMOS_UPDATE | 20 | select GENERIC_CMOS_UPDATE |
20 | select GENERIC_STRNCPY_FROM_USER | 21 | select GENERIC_STRNCPY_FROM_USER |
diff --git a/arch/alpha/include/asm/atomic.h b/arch/alpha/include/asm/atomic.h index c2cbe4fc391c..78b03ef39f6f 100644 --- a/arch/alpha/include/asm/atomic.h +++ b/arch/alpha/include/asm/atomic.h | |||
@@ -186,17 +186,24 @@ static __inline__ long atomic64_sub_return(long i, atomic64_t * v) | |||
186 | */ | 186 | */ |
187 | static __inline__ int __atomic_add_unless(atomic_t *v, int a, int u) | 187 | static __inline__ int __atomic_add_unless(atomic_t *v, int a, int u) |
188 | { | 188 | { |
189 | int c, old; | 189 | int c, new, old; |
190 | c = atomic_read(v); | 190 | smp_mb(); |
191 | for (;;) { | 191 | __asm__ __volatile__( |
192 | if (unlikely(c == (u))) | 192 | "1: ldl_l %[old],%[mem]\n" |
193 | break; | 193 | " cmpeq %[old],%[u],%[c]\n" |
194 | old = atomic_cmpxchg((v), c, c + (a)); | 194 | " addl %[old],%[a],%[new]\n" |
195 | if (likely(old == c)) | 195 | " bne %[c],2f\n" |
196 | break; | 196 | " stl_c %[new],%[mem]\n" |
197 | c = old; | 197 | " beq %[new],3f\n" |
198 | } | 198 | "2:\n" |
199 | return c; | 199 | ".subsection 2\n" |
200 | "3: br 1b\n" | ||
201 | ".previous" | ||
202 | : [old] "=&r"(old), [new] "=&r"(new), [c] "=&r"(c) | ||
203 | : [mem] "m"(*v), [a] "rI"(a), [u] "rI"((long)u) | ||
204 | : "memory"); | ||
205 | smp_mb(); | ||
206 | return old; | ||
200 | } | 207 | } |
201 | 208 | ||
202 | 209 | ||
@@ -207,21 +214,56 @@ static __inline__ int __atomic_add_unless(atomic_t *v, int a, int u) | |||
207 | * @u: ...unless v is equal to u. | 214 | * @u: ...unless v is equal to u. |
208 | * | 215 | * |
209 | * Atomically adds @a to @v, so long as it was not @u. | 216 | * Atomically adds @a to @v, so long as it was not @u. |
210 | * Returns the old value of @v. | 217 | * Returns true iff @v was not @u. |
211 | */ | 218 | */ |
212 | static __inline__ int atomic64_add_unless(atomic64_t *v, long a, long u) | 219 | static __inline__ int atomic64_add_unless(atomic64_t *v, long a, long u) |
213 | { | 220 | { |
214 | long c, old; | 221 | long c, tmp; |
215 | c = atomic64_read(v); | 222 | smp_mb(); |
216 | for (;;) { | 223 | __asm__ __volatile__( |
217 | if (unlikely(c == (u))) | 224 | "1: ldq_l %[tmp],%[mem]\n" |
218 | break; | 225 | " cmpeq %[tmp],%[u],%[c]\n" |
219 | old = atomic64_cmpxchg((v), c, c + (a)); | 226 | " addq %[tmp],%[a],%[tmp]\n" |
220 | if (likely(old == c)) | 227 | " bne %[c],2f\n" |
221 | break; | 228 | " stq_c %[tmp],%[mem]\n" |
222 | c = old; | 229 | " beq %[tmp],3f\n" |
223 | } | 230 | "2:\n" |
224 | return c != (u); | 231 | ".subsection 2\n" |
232 | "3: br 1b\n" | ||
233 | ".previous" | ||
234 | : [tmp] "=&r"(tmp), [c] "=&r"(c) | ||
235 | : [mem] "m"(*v), [a] "rI"(a), [u] "rI"(u) | ||
236 | : "memory"); | ||
237 | smp_mb(); | ||
238 | return !c; | ||
239 | } | ||
240 | |||
241 | /* | ||
242 | * atomic64_dec_if_positive - decrement by 1 if old value positive | ||
243 | * @v: pointer of type atomic_t | ||
244 | * | ||
245 | * The function returns the old value of *v minus 1, even if | ||
246 | * the atomic variable, v, was not decremented. | ||
247 | */ | ||
248 | static inline long atomic64_dec_if_positive(atomic64_t *v) | ||
249 | { | ||
250 | long old, tmp; | ||
251 | smp_mb(); | ||
252 | __asm__ __volatile__( | ||
253 | "1: ldq_l %[old],%[mem]\n" | ||
254 | " subq %[old],1,%[tmp]\n" | ||
255 | " ble %[old],2f\n" | ||
256 | " stq_c %[tmp],%[mem]\n" | ||
257 | " beq %[tmp],3f\n" | ||
258 | "2:\n" | ||
259 | ".subsection 2\n" | ||
260 | "3: br 1b\n" | ||
261 | ".previous" | ||
262 | : [old] "=&r"(old), [tmp] "=&r"(tmp) | ||
263 | : [mem] "m"(*v) | ||
264 | : "memory"); | ||
265 | smp_mb(); | ||
266 | return old - 1; | ||
225 | } | 267 | } |
226 | 268 | ||
227 | #define atomic64_inc_not_zero(v) atomic64_add_unless((v), 1, 0) | 269 | #define atomic64_inc_not_zero(v) atomic64_add_unless((v), 1, 0) |
diff --git a/arch/alpha/include/asm/param.h b/arch/alpha/include/asm/param.h index bf46af51941b..a5b68b268bcf 100644 --- a/arch/alpha/include/asm/param.h +++ b/arch/alpha/include/asm/param.h | |||
@@ -3,7 +3,9 @@ | |||
3 | 3 | ||
4 | #include <uapi/asm/param.h> | 4 | #include <uapi/asm/param.h> |
5 | 5 | ||
6 | #define HZ CONFIG_HZ | 6 | # undef HZ |
7 | #define USER_HZ HZ | 7 | # define HZ CONFIG_HZ |
8 | # define CLOCKS_PER_SEC HZ /* frequency at which times() counts */ | 8 | # define USER_HZ 1024 |
9 | # define CLOCKS_PER_SEC USER_HZ /* frequency at which times() counts */ | ||
10 | |||
9 | #endif /* _ASM_ALPHA_PARAM_H */ | 11 | #endif /* _ASM_ALPHA_PARAM_H */ |
diff --git a/arch/alpha/include/asm/spinlock.h b/arch/alpha/include/asm/spinlock.h index 3bba21e41b81..37b570d01202 100644 --- a/arch/alpha/include/asm/spinlock.h +++ b/arch/alpha/include/asm/spinlock.h | |||
@@ -168,8 +168,4 @@ static inline void arch_write_unlock(arch_rwlock_t * lock) | |||
168 | #define arch_read_lock_flags(lock, flags) arch_read_lock(lock) | 168 | #define arch_read_lock_flags(lock, flags) arch_read_lock(lock) |
169 | #define arch_write_lock_flags(lock, flags) arch_write_lock(lock) | 169 | #define arch_write_lock_flags(lock, flags) arch_write_lock(lock) |
170 | 170 | ||
171 | #define arch_spin_relax(lock) cpu_relax() | ||
172 | #define arch_read_relax(lock) cpu_relax() | ||
173 | #define arch_write_relax(lock) cpu_relax() | ||
174 | |||
175 | #endif /* _ALPHA_SPINLOCK_H */ | 171 | #endif /* _ALPHA_SPINLOCK_H */ |
diff --git a/arch/alpha/include/asm/unistd.h b/arch/alpha/include/asm/unistd.h index 43baee17acdf..f2c94402e2c8 100644 --- a/arch/alpha/include/asm/unistd.h +++ b/arch/alpha/include/asm/unistd.h | |||
@@ -3,8 +3,7 @@ | |||
3 | 3 | ||
4 | #include <uapi/asm/unistd.h> | 4 | #include <uapi/asm/unistd.h> |
5 | 5 | ||
6 | 6 | #define NR_SYSCALLS 508 | |
7 | #define NR_SYSCALLS 506 | ||
8 | 7 | ||
9 | #define __ARCH_WANT_OLD_READDIR | 8 | #define __ARCH_WANT_OLD_READDIR |
10 | #define __ARCH_WANT_STAT64 | 9 | #define __ARCH_WANT_STAT64 |
diff --git a/arch/alpha/include/uapi/asm/param.h b/arch/alpha/include/uapi/asm/param.h index 29daed819ebd..dbcd9834af6d 100644 --- a/arch/alpha/include/uapi/asm/param.h +++ b/arch/alpha/include/uapi/asm/param.h | |||
@@ -1,13 +1,7 @@ | |||
1 | #ifndef _UAPI_ASM_ALPHA_PARAM_H | 1 | #ifndef _UAPI_ASM_ALPHA_PARAM_H |
2 | #define _UAPI_ASM_ALPHA_PARAM_H | 2 | #define _UAPI_ASM_ALPHA_PARAM_H |
3 | 3 | ||
4 | /* ??? Gross. I don't want to parameterize this, and supposedly the | ||
5 | hardware ignores reprogramming. We also need userland buy-in to the | ||
6 | change in HZ, since this is visible in the wait4 resources etc. */ | ||
7 | |||
8 | #ifndef __KERNEL__ | ||
9 | #define HZ 1024 | 4 | #define HZ 1024 |
10 | #endif | ||
11 | 5 | ||
12 | #define EXEC_PAGESIZE 8192 | 6 | #define EXEC_PAGESIZE 8192 |
13 | 7 | ||
@@ -17,5 +11,4 @@ | |||
17 | 11 | ||
18 | #define MAXHOSTNAMELEN 64 /* max length of hostname */ | 12 | #define MAXHOSTNAMELEN 64 /* max length of hostname */ |
19 | 13 | ||
20 | |||
21 | #endif /* _UAPI_ASM_ALPHA_PARAM_H */ | 14 | #endif /* _UAPI_ASM_ALPHA_PARAM_H */ |
diff --git a/arch/alpha/include/uapi/asm/unistd.h b/arch/alpha/include/uapi/asm/unistd.h index 801d28bcea51..53ae7bb1bfd1 100644 --- a/arch/alpha/include/uapi/asm/unistd.h +++ b/arch/alpha/include/uapi/asm/unistd.h | |||
@@ -467,5 +467,7 @@ | |||
467 | #define __NR_sendmmsg 503 | 467 | #define __NR_sendmmsg 503 |
468 | #define __NR_process_vm_readv 504 | 468 | #define __NR_process_vm_readv 504 |
469 | #define __NR_process_vm_writev 505 | 469 | #define __NR_process_vm_writev 505 |
470 | #define __NR_kcmp 506 | ||
471 | #define __NR_finit_module 507 | ||
470 | 472 | ||
471 | #endif /* _UAPI_ALPHA_UNISTD_H */ | 473 | #endif /* _UAPI_ALPHA_UNISTD_H */ |
diff --git a/arch/alpha/kernel/entry.S b/arch/alpha/kernel/entry.S index f62a994ef126..a969b95ee5ac 100644 --- a/arch/alpha/kernel/entry.S +++ b/arch/alpha/kernel/entry.S | |||
@@ -12,11 +12,32 @@ | |||
12 | 12 | ||
13 | .text | 13 | .text |
14 | .set noat | 14 | .set noat |
15 | .cfi_sections .debug_frame | ||
15 | 16 | ||
16 | /* Stack offsets. */ | 17 | /* Stack offsets. */ |
17 | #define SP_OFF 184 | 18 | #define SP_OFF 184 |
18 | #define SWITCH_STACK_SIZE 320 | 19 | #define SWITCH_STACK_SIZE 320 |
19 | 20 | ||
21 | .macro CFI_START_OSF_FRAME func | ||
22 | .align 4 | ||
23 | .globl \func | ||
24 | .type \func,@function | ||
25 | \func: | ||
26 | .cfi_startproc simple | ||
27 | .cfi_return_column 64 | ||
28 | .cfi_def_cfa $sp, 48 | ||
29 | .cfi_rel_offset 64, 8 | ||
30 | .cfi_rel_offset $gp, 16 | ||
31 | .cfi_rel_offset $16, 24 | ||
32 | .cfi_rel_offset $17, 32 | ||
33 | .cfi_rel_offset $18, 40 | ||
34 | .endm | ||
35 | |||
36 | .macro CFI_END_OSF_FRAME func | ||
37 | .cfi_endproc | ||
38 | .size \func, . - \func | ||
39 | .endm | ||
40 | |||
20 | /* | 41 | /* |
21 | * This defines the normal kernel pt-regs layout. | 42 | * This defines the normal kernel pt-regs layout. |
22 | * | 43 | * |
@@ -27,100 +48,158 @@ | |||
27 | * the palcode-provided values are available to the signal handler. | 48 | * the palcode-provided values are available to the signal handler. |
28 | */ | 49 | */ |
29 | 50 | ||
30 | #define SAVE_ALL \ | 51 | .macro SAVE_ALL |
31 | subq $sp, SP_OFF, $sp; \ | 52 | subq $sp, SP_OFF, $sp |
32 | stq $0, 0($sp); \ | 53 | .cfi_adjust_cfa_offset SP_OFF |
33 | stq $1, 8($sp); \ | 54 | stq $0, 0($sp) |
34 | stq $2, 16($sp); \ | 55 | stq $1, 8($sp) |
35 | stq $3, 24($sp); \ | 56 | stq $2, 16($sp) |
36 | stq $4, 32($sp); \ | 57 | stq $3, 24($sp) |
37 | stq $28, 144($sp); \ | 58 | stq $4, 32($sp) |
38 | lda $2, alpha_mv; \ | 59 | stq $28, 144($sp) |
39 | stq $5, 40($sp); \ | 60 | .cfi_rel_offset $0, 0 |
40 | stq $6, 48($sp); \ | 61 | .cfi_rel_offset $1, 8 |
41 | stq $7, 56($sp); \ | 62 | .cfi_rel_offset $2, 16 |
42 | stq $8, 64($sp); \ | 63 | .cfi_rel_offset $3, 24 |
43 | stq $19, 72($sp); \ | 64 | .cfi_rel_offset $4, 32 |
44 | stq $20, 80($sp); \ | 65 | .cfi_rel_offset $28, 144 |
45 | stq $21, 88($sp); \ | 66 | lda $2, alpha_mv |
46 | ldq $2, HAE_CACHE($2); \ | 67 | stq $5, 40($sp) |
47 | stq $22, 96($sp); \ | 68 | stq $6, 48($sp) |
48 | stq $23, 104($sp); \ | 69 | stq $7, 56($sp) |
49 | stq $24, 112($sp); \ | 70 | stq $8, 64($sp) |
50 | stq $25, 120($sp); \ | 71 | stq $19, 72($sp) |
51 | stq $26, 128($sp); \ | 72 | stq $20, 80($sp) |
52 | stq $27, 136($sp); \ | 73 | stq $21, 88($sp) |
53 | stq $2, 152($sp); \ | 74 | ldq $2, HAE_CACHE($2) |
54 | stq $16, 160($sp); \ | 75 | stq $22, 96($sp) |
55 | stq $17, 168($sp); \ | 76 | stq $23, 104($sp) |
77 | stq $24, 112($sp) | ||
78 | stq $25, 120($sp) | ||
79 | stq $26, 128($sp) | ||
80 | stq $27, 136($sp) | ||
81 | stq $2, 152($sp) | ||
82 | stq $16, 160($sp) | ||
83 | stq $17, 168($sp) | ||
56 | stq $18, 176($sp) | 84 | stq $18, 176($sp) |
85 | .cfi_rel_offset $5, 40 | ||
86 | .cfi_rel_offset $6, 48 | ||
87 | .cfi_rel_offset $7, 56 | ||
88 | .cfi_rel_offset $8, 64 | ||
89 | .cfi_rel_offset $19, 72 | ||
90 | .cfi_rel_offset $20, 80 | ||
91 | .cfi_rel_offset $21, 88 | ||
92 | .cfi_rel_offset $22, 96 | ||
93 | .cfi_rel_offset $23, 104 | ||
94 | .cfi_rel_offset $24, 112 | ||
95 | .cfi_rel_offset $25, 120 | ||
96 | .cfi_rel_offset $26, 128 | ||
97 | .cfi_rel_offset $27, 136 | ||
98 | .endm | ||
57 | 99 | ||
58 | #define RESTORE_ALL \ | 100 | .macro RESTORE_ALL |
59 | lda $19, alpha_mv; \ | 101 | lda $19, alpha_mv |
60 | ldq $0, 0($sp); \ | 102 | ldq $0, 0($sp) |
61 | ldq $1, 8($sp); \ | 103 | ldq $1, 8($sp) |
62 | ldq $2, 16($sp); \ | 104 | ldq $2, 16($sp) |
63 | ldq $3, 24($sp); \ | 105 | ldq $3, 24($sp) |
64 | ldq $21, 152($sp); \ | 106 | ldq $21, 152($sp) |
65 | ldq $20, HAE_CACHE($19); \ | 107 | ldq $20, HAE_CACHE($19) |
66 | ldq $4, 32($sp); \ | 108 | ldq $4, 32($sp) |
67 | ldq $5, 40($sp); \ | 109 | ldq $5, 40($sp) |
68 | ldq $6, 48($sp); \ | 110 | ldq $6, 48($sp) |
69 | ldq $7, 56($sp); \ | 111 | ldq $7, 56($sp) |
70 | subq $20, $21, $20; \ | 112 | subq $20, $21, $20 |
71 | ldq $8, 64($sp); \ | 113 | ldq $8, 64($sp) |
72 | beq $20, 99f; \ | 114 | beq $20, 99f |
73 | ldq $20, HAE_REG($19); \ | 115 | ldq $20, HAE_REG($19) |
74 | stq $21, HAE_CACHE($19); \ | 116 | stq $21, HAE_CACHE($19) |
75 | stq $21, 0($20); \ | 117 | stq $21, 0($20) |
76 | 99:; \ | 118 | 99: ldq $19, 72($sp) |
77 | ldq $19, 72($sp); \ | 119 | ldq $20, 80($sp) |
78 | ldq $20, 80($sp); \ | 120 | ldq $21, 88($sp) |
79 | ldq $21, 88($sp); \ | 121 | ldq $22, 96($sp) |
80 | ldq $22, 96($sp); \ | 122 | ldq $23, 104($sp) |
81 | ldq $23, 104($sp); \ | 123 | ldq $24, 112($sp) |
82 | ldq $24, 112($sp); \ | 124 | ldq $25, 120($sp) |
83 | ldq $25, 120($sp); \ | 125 | ldq $26, 128($sp) |
84 | ldq $26, 128($sp); \ | 126 | ldq $27, 136($sp) |
85 | ldq $27, 136($sp); \ | 127 | ldq $28, 144($sp) |
86 | ldq $28, 144($sp); \ | ||
87 | addq $sp, SP_OFF, $sp | 128 | addq $sp, SP_OFF, $sp |
129 | .cfi_restore $0 | ||
130 | .cfi_restore $1 | ||
131 | .cfi_restore $2 | ||
132 | .cfi_restore $3 | ||
133 | .cfi_restore $4 | ||
134 | .cfi_restore $5 | ||
135 | .cfi_restore $6 | ||
136 | .cfi_restore $7 | ||
137 | .cfi_restore $8 | ||
138 | .cfi_restore $19 | ||
139 | .cfi_restore $20 | ||
140 | .cfi_restore $21 | ||
141 | .cfi_restore $22 | ||
142 | .cfi_restore $23 | ||
143 | .cfi_restore $24 | ||
144 | .cfi_restore $25 | ||
145 | .cfi_restore $26 | ||
146 | .cfi_restore $27 | ||
147 | .cfi_restore $28 | ||
148 | .cfi_adjust_cfa_offset -SP_OFF | ||
149 | .endm | ||
150 | |||
151 | .macro DO_SWITCH_STACK | ||
152 | bsr $1, do_switch_stack | ||
153 | .cfi_adjust_cfa_offset SWITCH_STACK_SIZE | ||
154 | .cfi_rel_offset $9, 0 | ||
155 | .cfi_rel_offset $10, 8 | ||
156 | .cfi_rel_offset $11, 16 | ||
157 | .cfi_rel_offset $12, 24 | ||
158 | .cfi_rel_offset $13, 32 | ||
159 | .cfi_rel_offset $14, 40 | ||
160 | .cfi_rel_offset $15, 48 | ||
161 | /* We don't really care about the FP registers for debugging. */ | ||
162 | .endm | ||
163 | |||
164 | .macro UNDO_SWITCH_STACK | ||
165 | bsr $1, undo_switch_stack | ||
166 | .cfi_restore $9 | ||
167 | .cfi_restore $10 | ||
168 | .cfi_restore $11 | ||
169 | .cfi_restore $12 | ||
170 | .cfi_restore $13 | ||
171 | .cfi_restore $14 | ||
172 | .cfi_restore $15 | ||
173 | .cfi_adjust_cfa_offset -SWITCH_STACK_SIZE | ||
174 | .endm | ||
88 | 175 | ||
89 | /* | 176 | /* |
90 | * Non-syscall kernel entry points. | 177 | * Non-syscall kernel entry points. |
91 | */ | 178 | */ |
92 | 179 | ||
93 | .align 4 | 180 | CFI_START_OSF_FRAME entInt |
94 | .globl entInt | ||
95 | .ent entInt | ||
96 | entInt: | ||
97 | SAVE_ALL | 181 | SAVE_ALL |
98 | lda $8, 0x3fff | 182 | lda $8, 0x3fff |
99 | lda $26, ret_from_sys_call | 183 | lda $26, ret_from_sys_call |
100 | bic $sp, $8, $8 | 184 | bic $sp, $8, $8 |
101 | mov $sp, $19 | 185 | mov $sp, $19 |
102 | jsr $31, do_entInt | 186 | jsr $31, do_entInt |
103 | .end entInt | 187 | CFI_END_OSF_FRAME entInt |
104 | 188 | ||
105 | .align 4 | 189 | CFI_START_OSF_FRAME entArith |
106 | .globl entArith | ||
107 | .ent entArith | ||
108 | entArith: | ||
109 | SAVE_ALL | 190 | SAVE_ALL |
110 | lda $8, 0x3fff | 191 | lda $8, 0x3fff |
111 | lda $26, ret_from_sys_call | 192 | lda $26, ret_from_sys_call |
112 | bic $sp, $8, $8 | 193 | bic $sp, $8, $8 |
113 | mov $sp, $18 | 194 | mov $sp, $18 |
114 | jsr $31, do_entArith | 195 | jsr $31, do_entArith |
115 | .end entArith | 196 | CFI_END_OSF_FRAME entArith |
116 | 197 | ||
117 | .align 4 | 198 | CFI_START_OSF_FRAME entMM |
118 | .globl entMM | ||
119 | .ent entMM | ||
120 | entMM: | ||
121 | SAVE_ALL | 199 | SAVE_ALL |
122 | /* save $9 - $15 so the inline exception code can manipulate them. */ | 200 | /* save $9 - $15 so the inline exception code can manipulate them. */ |
123 | subq $sp, 56, $sp | 201 | subq $sp, 56, $sp |
202 | .cfi_adjust_cfa_offset 56 | ||
124 | stq $9, 0($sp) | 203 | stq $9, 0($sp) |
125 | stq $10, 8($sp) | 204 | stq $10, 8($sp) |
126 | stq $11, 16($sp) | 205 | stq $11, 16($sp) |
@@ -128,6 +207,13 @@ entMM: | |||
128 | stq $13, 32($sp) | 207 | stq $13, 32($sp) |
129 | stq $14, 40($sp) | 208 | stq $14, 40($sp) |
130 | stq $15, 48($sp) | 209 | stq $15, 48($sp) |
210 | .cfi_rel_offset $9, 0 | ||
211 | .cfi_rel_offset $10, 8 | ||
212 | .cfi_rel_offset $11, 16 | ||
213 | .cfi_rel_offset $12, 24 | ||
214 | .cfi_rel_offset $13, 32 | ||
215 | .cfi_rel_offset $14, 40 | ||
216 | .cfi_rel_offset $15, 48 | ||
131 | addq $sp, 56, $19 | 217 | addq $sp, 56, $19 |
132 | /* handle the fault */ | 218 | /* handle the fault */ |
133 | lda $8, 0x3fff | 219 | lda $8, 0x3fff |
@@ -142,28 +228,33 @@ entMM: | |||
142 | ldq $14, 40($sp) | 228 | ldq $14, 40($sp) |
143 | ldq $15, 48($sp) | 229 | ldq $15, 48($sp) |
144 | addq $sp, 56, $sp | 230 | addq $sp, 56, $sp |
231 | .cfi_restore $9 | ||
232 | .cfi_restore $10 | ||
233 | .cfi_restore $11 | ||
234 | .cfi_restore $12 | ||
235 | .cfi_restore $13 | ||
236 | .cfi_restore $14 | ||
237 | .cfi_restore $15 | ||
238 | .cfi_adjust_cfa_offset -56 | ||
145 | /* finish up the syscall as normal. */ | 239 | /* finish up the syscall as normal. */ |
146 | br ret_from_sys_call | 240 | br ret_from_sys_call |
147 | .end entMM | 241 | CFI_END_OSF_FRAME entMM |
148 | 242 | ||
149 | .align 4 | 243 | CFI_START_OSF_FRAME entIF |
150 | .globl entIF | ||
151 | .ent entIF | ||
152 | entIF: | ||
153 | SAVE_ALL | 244 | SAVE_ALL |
154 | lda $8, 0x3fff | 245 | lda $8, 0x3fff |
155 | lda $26, ret_from_sys_call | 246 | lda $26, ret_from_sys_call |
156 | bic $sp, $8, $8 | 247 | bic $sp, $8, $8 |
157 | mov $sp, $17 | 248 | mov $sp, $17 |
158 | jsr $31, do_entIF | 249 | jsr $31, do_entIF |
159 | .end entIF | 250 | CFI_END_OSF_FRAME entIF |
160 | 251 | ||
161 | .align 4 | 252 | CFI_START_OSF_FRAME entUna |
162 | .globl entUna | ||
163 | .ent entUna | ||
164 | entUna: | ||
165 | lda $sp, -256($sp) | 253 | lda $sp, -256($sp) |
254 | .cfi_adjust_cfa_offset 256 | ||
166 | stq $0, 0($sp) | 255 | stq $0, 0($sp) |
256 | .cfi_rel_offset $0, 0 | ||
257 | .cfi_remember_state | ||
167 | ldq $0, 256($sp) /* get PS */ | 258 | ldq $0, 256($sp) /* get PS */ |
168 | stq $1, 8($sp) | 259 | stq $1, 8($sp) |
169 | stq $2, 16($sp) | 260 | stq $2, 16($sp) |
@@ -195,6 +286,32 @@ entUna: | |||
195 | stq $28, 224($sp) | 286 | stq $28, 224($sp) |
196 | mov $sp, $19 | 287 | mov $sp, $19 |
197 | stq $gp, 232($sp) | 288 | stq $gp, 232($sp) |
289 | .cfi_rel_offset $1, 1*8 | ||
290 | .cfi_rel_offset $2, 2*8 | ||
291 | .cfi_rel_offset $3, 3*8 | ||
292 | .cfi_rel_offset $4, 4*8 | ||
293 | .cfi_rel_offset $5, 5*8 | ||
294 | .cfi_rel_offset $6, 6*8 | ||
295 | .cfi_rel_offset $7, 7*8 | ||
296 | .cfi_rel_offset $8, 8*8 | ||
297 | .cfi_rel_offset $9, 9*8 | ||
298 | .cfi_rel_offset $10, 10*8 | ||
299 | .cfi_rel_offset $11, 11*8 | ||
300 | .cfi_rel_offset $12, 12*8 | ||
301 | .cfi_rel_offset $13, 13*8 | ||
302 | .cfi_rel_offset $14, 14*8 | ||
303 | .cfi_rel_offset $15, 15*8 | ||
304 | .cfi_rel_offset $19, 19*8 | ||
305 | .cfi_rel_offset $20, 20*8 | ||
306 | .cfi_rel_offset $21, 21*8 | ||
307 | .cfi_rel_offset $22, 22*8 | ||
308 | .cfi_rel_offset $23, 23*8 | ||
309 | .cfi_rel_offset $24, 24*8 | ||
310 | .cfi_rel_offset $25, 25*8 | ||
311 | .cfi_rel_offset $26, 26*8 | ||
312 | .cfi_rel_offset $27, 27*8 | ||
313 | .cfi_rel_offset $28, 28*8 | ||
314 | .cfi_rel_offset $29, 29*8 | ||
198 | lda $8, 0x3fff | 315 | lda $8, 0x3fff |
199 | stq $31, 248($sp) | 316 | stq $31, 248($sp) |
200 | bic $sp, $8, $8 | 317 | bic $sp, $8, $8 |
@@ -228,16 +345,45 @@ entUna: | |||
228 | ldq $28, 224($sp) | 345 | ldq $28, 224($sp) |
229 | ldq $gp, 232($sp) | 346 | ldq $gp, 232($sp) |
230 | lda $sp, 256($sp) | 347 | lda $sp, 256($sp) |
348 | .cfi_restore $1 | ||
349 | .cfi_restore $2 | ||
350 | .cfi_restore $3 | ||
351 | .cfi_restore $4 | ||
352 | .cfi_restore $5 | ||
353 | .cfi_restore $6 | ||
354 | .cfi_restore $7 | ||
355 | .cfi_restore $8 | ||
356 | .cfi_restore $9 | ||
357 | .cfi_restore $10 | ||
358 | .cfi_restore $11 | ||
359 | .cfi_restore $12 | ||
360 | .cfi_restore $13 | ||
361 | .cfi_restore $14 | ||
362 | .cfi_restore $15 | ||
363 | .cfi_restore $19 | ||
364 | .cfi_restore $20 | ||
365 | .cfi_restore $21 | ||
366 | .cfi_restore $22 | ||
367 | .cfi_restore $23 | ||
368 | .cfi_restore $24 | ||
369 | .cfi_restore $25 | ||
370 | .cfi_restore $26 | ||
371 | .cfi_restore $27 | ||
372 | .cfi_restore $28 | ||
373 | .cfi_restore $29 | ||
374 | .cfi_adjust_cfa_offset -256 | ||
231 | call_pal PAL_rti | 375 | call_pal PAL_rti |
232 | .end entUna | ||
233 | 376 | ||
234 | .align 4 | 377 | .align 4 |
235 | .ent entUnaUser | ||
236 | entUnaUser: | 378 | entUnaUser: |
379 | .cfi_restore_state | ||
237 | ldq $0, 0($sp) /* restore original $0 */ | 380 | ldq $0, 0($sp) /* restore original $0 */ |
238 | lda $sp, 256($sp) /* pop entUna's stack frame */ | 381 | lda $sp, 256($sp) /* pop entUna's stack frame */ |
382 | .cfi_restore $0 | ||
383 | .cfi_adjust_cfa_offset -256 | ||
239 | SAVE_ALL /* setup normal kernel stack */ | 384 | SAVE_ALL /* setup normal kernel stack */ |
240 | lda $sp, -56($sp) | 385 | lda $sp, -56($sp) |
386 | .cfi_adjust_cfa_offset 56 | ||
241 | stq $9, 0($sp) | 387 | stq $9, 0($sp) |
242 | stq $10, 8($sp) | 388 | stq $10, 8($sp) |
243 | stq $11, 16($sp) | 389 | stq $11, 16($sp) |
@@ -245,6 +391,13 @@ entUnaUser: | |||
245 | stq $13, 32($sp) | 391 | stq $13, 32($sp) |
246 | stq $14, 40($sp) | 392 | stq $14, 40($sp) |
247 | stq $15, 48($sp) | 393 | stq $15, 48($sp) |
394 | .cfi_rel_offset $9, 0 | ||
395 | .cfi_rel_offset $10, 8 | ||
396 | .cfi_rel_offset $11, 16 | ||
397 | .cfi_rel_offset $12, 24 | ||
398 | .cfi_rel_offset $13, 32 | ||
399 | .cfi_rel_offset $14, 40 | ||
400 | .cfi_rel_offset $15, 48 | ||
248 | lda $8, 0x3fff | 401 | lda $8, 0x3fff |
249 | addq $sp, 56, $19 | 402 | addq $sp, 56, $19 |
250 | bic $sp, $8, $8 | 403 | bic $sp, $8, $8 |
@@ -257,20 +410,25 @@ entUnaUser: | |||
257 | ldq $14, 40($sp) | 410 | ldq $14, 40($sp) |
258 | ldq $15, 48($sp) | 411 | ldq $15, 48($sp) |
259 | lda $sp, 56($sp) | 412 | lda $sp, 56($sp) |
413 | .cfi_restore $9 | ||
414 | .cfi_restore $10 | ||
415 | .cfi_restore $11 | ||
416 | .cfi_restore $12 | ||
417 | .cfi_restore $13 | ||
418 | .cfi_restore $14 | ||
419 | .cfi_restore $15 | ||
420 | .cfi_adjust_cfa_offset -56 | ||
260 | br ret_from_sys_call | 421 | br ret_from_sys_call |
261 | .end entUnaUser | 422 | CFI_END_OSF_FRAME entUna |
262 | 423 | ||
263 | .align 4 | 424 | CFI_START_OSF_FRAME entDbg |
264 | .globl entDbg | ||
265 | .ent entDbg | ||
266 | entDbg: | ||
267 | SAVE_ALL | 425 | SAVE_ALL |
268 | lda $8, 0x3fff | 426 | lda $8, 0x3fff |
269 | lda $26, ret_from_sys_call | 427 | lda $26, ret_from_sys_call |
270 | bic $sp, $8, $8 | 428 | bic $sp, $8, $8 |
271 | mov $sp, $16 | 429 | mov $sp, $16 |
272 | jsr $31, do_entDbg | 430 | jsr $31, do_entDbg |
273 | .end entDbg | 431 | CFI_END_OSF_FRAME entDbg |
274 | 432 | ||
275 | /* | 433 | /* |
276 | * The system call entry point is special. Most importantly, it looks | 434 | * The system call entry point is special. Most importantly, it looks |
@@ -285,8 +443,12 @@ entDbg: | |||
285 | 443 | ||
286 | .align 4 | 444 | .align 4 |
287 | .globl entSys | 445 | .globl entSys |
288 | .globl ret_from_sys_call | 446 | .type entSys, @function |
289 | .ent entSys | 447 | .cfi_startproc simple |
448 | .cfi_return_column 64 | ||
449 | .cfi_def_cfa $sp, 48 | ||
450 | .cfi_rel_offset 64, 8 | ||
451 | .cfi_rel_offset $gp, 16 | ||
290 | entSys: | 452 | entSys: |
291 | SAVE_ALL | 453 | SAVE_ALL |
292 | lda $8, 0x3fff | 454 | lda $8, 0x3fff |
@@ -300,6 +462,9 @@ entSys: | |||
300 | stq $17, SP_OFF+32($sp) | 462 | stq $17, SP_OFF+32($sp) |
301 | s8addq $0, $5, $5 | 463 | s8addq $0, $5, $5 |
302 | stq $18, SP_OFF+40($sp) | 464 | stq $18, SP_OFF+40($sp) |
465 | .cfi_rel_offset $16, SP_OFF+24 | ||
466 | .cfi_rel_offset $17, SP_OFF+32 | ||
467 | .cfi_rel_offset $18, SP_OFF+40 | ||
303 | blbs $3, strace | 468 | blbs $3, strace |
304 | beq $4, 1f | 469 | beq $4, 1f |
305 | ldq $27, 0($5) | 470 | ldq $27, 0($5) |
@@ -310,6 +475,7 @@ entSys: | |||
310 | stq $31, 72($sp) /* a3=0 => no error */ | 475 | stq $31, 72($sp) /* a3=0 => no error */ |
311 | 476 | ||
312 | .align 4 | 477 | .align 4 |
478 | .globl ret_from_sys_call | ||
313 | ret_from_sys_call: | 479 | ret_from_sys_call: |
314 | cmovne $26, 0, $18 /* $18 = 0 => non-restartable */ | 480 | cmovne $26, 0, $18 /* $18 = 0 => non-restartable */ |
315 | ldq $0, SP_OFF($sp) | 481 | ldq $0, SP_OFF($sp) |
@@ -324,10 +490,12 @@ ret_to_user: | |||
324 | and $17, _TIF_WORK_MASK, $2 | 490 | and $17, _TIF_WORK_MASK, $2 |
325 | bne $2, work_pending | 491 | bne $2, work_pending |
326 | restore_all: | 492 | restore_all: |
493 | .cfi_remember_state | ||
327 | RESTORE_ALL | 494 | RESTORE_ALL |
328 | call_pal PAL_rti | 495 | call_pal PAL_rti |
329 | 496 | ||
330 | ret_to_kernel: | 497 | ret_to_kernel: |
498 | .cfi_restore_state | ||
331 | lda $16, 7 | 499 | lda $16, 7 |
332 | call_pal PAL_swpipl | 500 | call_pal PAL_swpipl |
333 | br restore_all | 501 | br restore_all |
@@ -356,7 +524,6 @@ $ret_success: | |||
356 | stq $0, 0($sp) | 524 | stq $0, 0($sp) |
357 | stq $31, 72($sp) /* a3=0 => no error */ | 525 | stq $31, 72($sp) /* a3=0 => no error */ |
358 | br ret_from_sys_call | 526 | br ret_from_sys_call |
359 | .end entSys | ||
360 | 527 | ||
361 | /* | 528 | /* |
362 | * Do all cleanup when returning from all interrupts and system calls. | 529 | * Do all cleanup when returning from all interrupts and system calls. |
@@ -370,7 +537,7 @@ $ret_success: | |||
370 | */ | 537 | */ |
371 | 538 | ||
372 | .align 4 | 539 | .align 4 |
373 | .ent work_pending | 540 | .type work_pending, @function |
374 | work_pending: | 541 | work_pending: |
375 | and $17, _TIF_NOTIFY_RESUME | _TIF_SIGPENDING, $2 | 542 | and $17, _TIF_NOTIFY_RESUME | _TIF_SIGPENDING, $2 |
376 | bne $2, $work_notifysig | 543 | bne $2, $work_notifysig |
@@ -387,23 +554,22 @@ $work_resched: | |||
387 | 554 | ||
388 | $work_notifysig: | 555 | $work_notifysig: |
389 | mov $sp, $16 | 556 | mov $sp, $16 |
390 | bsr $1, do_switch_stack | 557 | DO_SWITCH_STACK |
391 | jsr $26, do_work_pending | 558 | jsr $26, do_work_pending |
392 | bsr $1, undo_switch_stack | 559 | UNDO_SWITCH_STACK |
393 | br restore_all | 560 | br restore_all |
394 | .end work_pending | ||
395 | 561 | ||
396 | /* | 562 | /* |
397 | * PTRACE syscall handler | 563 | * PTRACE syscall handler |
398 | */ | 564 | */ |
399 | 565 | ||
400 | .align 4 | 566 | .align 4 |
401 | .ent strace | 567 | .type strace, @function |
402 | strace: | 568 | strace: |
403 | /* set up signal stack, call syscall_trace */ | 569 | /* set up signal stack, call syscall_trace */ |
404 | bsr $1, do_switch_stack | 570 | DO_SWITCH_STACK |
405 | jsr $26, syscall_trace_enter /* returns the syscall number */ | 571 | jsr $26, syscall_trace_enter /* returns the syscall number */ |
406 | bsr $1, undo_switch_stack | 572 | UNDO_SWITCH_STACK |
407 | 573 | ||
408 | /* get the arguments back.. */ | 574 | /* get the arguments back.. */ |
409 | ldq $16, SP_OFF+24($sp) | 575 | ldq $16, SP_OFF+24($sp) |
@@ -431,9 +597,9 @@ ret_from_straced: | |||
431 | $strace_success: | 597 | $strace_success: |
432 | stq $0, 0($sp) /* save return value */ | 598 | stq $0, 0($sp) /* save return value */ |
433 | 599 | ||
434 | bsr $1, do_switch_stack | 600 | DO_SWITCH_STACK |
435 | jsr $26, syscall_trace_leave | 601 | jsr $26, syscall_trace_leave |
436 | bsr $1, undo_switch_stack | 602 | UNDO_SWITCH_STACK |
437 | br $31, ret_from_sys_call | 603 | br $31, ret_from_sys_call |
438 | 604 | ||
439 | .align 3 | 605 | .align 3 |
@@ -447,26 +613,31 @@ $strace_error: | |||
447 | stq $0, 0($sp) | 613 | stq $0, 0($sp) |
448 | stq $1, 72($sp) /* a3 for return */ | 614 | stq $1, 72($sp) /* a3 for return */ |
449 | 615 | ||
450 | bsr $1, do_switch_stack | 616 | DO_SWITCH_STACK |
451 | mov $18, $9 /* save old syscall number */ | 617 | mov $18, $9 /* save old syscall number */ |
452 | mov $19, $10 /* save old a3 */ | 618 | mov $19, $10 /* save old a3 */ |
453 | jsr $26, syscall_trace_leave | 619 | jsr $26, syscall_trace_leave |
454 | mov $9, $18 | 620 | mov $9, $18 |
455 | mov $10, $19 | 621 | mov $10, $19 |
456 | bsr $1, undo_switch_stack | 622 | UNDO_SWITCH_STACK |
457 | 623 | ||
458 | mov $31, $26 /* tell "ret_from_sys_call" we can restart */ | 624 | mov $31, $26 /* tell "ret_from_sys_call" we can restart */ |
459 | br ret_from_sys_call | 625 | br ret_from_sys_call |
460 | .end strace | 626 | CFI_END_OSF_FRAME entSys |
461 | 627 | ||
462 | /* | 628 | /* |
463 | * Save and restore the switch stack -- aka the balance of the user context. | 629 | * Save and restore the switch stack -- aka the balance of the user context. |
464 | */ | 630 | */ |
465 | 631 | ||
466 | .align 4 | 632 | .align 4 |
467 | .ent do_switch_stack | 633 | .type do_switch_stack, @function |
634 | .cfi_startproc simple | ||
635 | .cfi_return_column 64 | ||
636 | .cfi_def_cfa $sp, 0 | ||
637 | .cfi_register 64, $1 | ||
468 | do_switch_stack: | 638 | do_switch_stack: |
469 | lda $sp, -SWITCH_STACK_SIZE($sp) | 639 | lda $sp, -SWITCH_STACK_SIZE($sp) |
640 | .cfi_adjust_cfa_offset SWITCH_STACK_SIZE | ||
470 | stq $9, 0($sp) | 641 | stq $9, 0($sp) |
471 | stq $10, 8($sp) | 642 | stq $10, 8($sp) |
472 | stq $11, 16($sp) | 643 | stq $11, 16($sp) |
@@ -510,10 +681,14 @@ do_switch_stack: | |||
510 | stt $f0, 312($sp) # save fpcr in slot of $f31 | 681 | stt $f0, 312($sp) # save fpcr in slot of $f31 |
511 | ldt $f0, 64($sp) # dont let "do_switch_stack" change fp state. | 682 | ldt $f0, 64($sp) # dont let "do_switch_stack" change fp state. |
512 | ret $31, ($1), 1 | 683 | ret $31, ($1), 1 |
513 | .end do_switch_stack | 684 | .cfi_endproc |
685 | .size do_switch_stack, .-do_switch_stack | ||
514 | 686 | ||
515 | .align 4 | 687 | .align 4 |
516 | .ent undo_switch_stack | 688 | .type undo_switch_stack, @function |
689 | .cfi_startproc simple | ||
690 | .cfi_def_cfa $sp, 0 | ||
691 | .cfi_register 64, $1 | ||
517 | undo_switch_stack: | 692 | undo_switch_stack: |
518 | ldq $9, 0($sp) | 693 | ldq $9, 0($sp) |
519 | ldq $10, 8($sp) | 694 | ldq $10, 8($sp) |
@@ -558,7 +733,8 @@ undo_switch_stack: | |||
558 | ldt $f30, 304($sp) | 733 | ldt $f30, 304($sp) |
559 | lda $sp, SWITCH_STACK_SIZE($sp) | 734 | lda $sp, SWITCH_STACK_SIZE($sp) |
560 | ret $31, ($1), 1 | 735 | ret $31, ($1), 1 |
561 | .end undo_switch_stack | 736 | .cfi_endproc |
737 | .size undo_switch_stack, .-undo_switch_stack | ||
562 | 738 | ||
563 | /* | 739 | /* |
564 | * The meat of the context switch code. | 740 | * The meat of the context switch code. |
@@ -566,17 +742,18 @@ undo_switch_stack: | |||
566 | 742 | ||
567 | .align 4 | 743 | .align 4 |
568 | .globl alpha_switch_to | 744 | .globl alpha_switch_to |
569 | .ent alpha_switch_to | 745 | .type alpha_switch_to, @function |
746 | .cfi_startproc | ||
570 | alpha_switch_to: | 747 | alpha_switch_to: |
571 | .prologue 0 | 748 | DO_SWITCH_STACK |
572 | bsr $1, do_switch_stack | ||
573 | call_pal PAL_swpctx | 749 | call_pal PAL_swpctx |
574 | lda $8, 0x3fff | 750 | lda $8, 0x3fff |
575 | bsr $1, undo_switch_stack | 751 | UNDO_SWITCH_STACK |
576 | bic $sp, $8, $8 | 752 | bic $sp, $8, $8 |
577 | mov $17, $0 | 753 | mov $17, $0 |
578 | ret | 754 | ret |
579 | .end alpha_switch_to | 755 | .cfi_endproc |
756 | .size alpha_switch_to, .-alpha_switch_to | ||
580 | 757 | ||
581 | /* | 758 | /* |
582 | * New processes begin life here. | 759 | * New processes begin life here. |
diff --git a/arch/alpha/kernel/irq_alpha.c b/arch/alpha/kernel/irq_alpha.c index f433fc11877a..28e4429596f3 100644 --- a/arch/alpha/kernel/irq_alpha.c +++ b/arch/alpha/kernel/irq_alpha.c | |||
@@ -236,7 +236,7 @@ void __init | |||
236 | init_rtc_irq(void) | 236 | init_rtc_irq(void) |
237 | { | 237 | { |
238 | irq_set_chip_and_handler_name(RTC_IRQ, &dummy_irq_chip, | 238 | irq_set_chip_and_handler_name(RTC_IRQ, &dummy_irq_chip, |
239 | handle_simple_irq, "RTC"); | 239 | handle_percpu_irq, "RTC"); |
240 | setup_irq(RTC_IRQ, &timer_irqaction); | 240 | setup_irq(RTC_IRQ, &timer_irqaction); |
241 | } | 241 | } |
242 | 242 | ||
diff --git a/arch/alpha/kernel/smp.c b/arch/alpha/kernel/smp.c index 53b18a620e1c..9dbbcb3b9146 100644 --- a/arch/alpha/kernel/smp.c +++ b/arch/alpha/kernel/smp.c | |||
@@ -264,9 +264,10 @@ recv_secondary_console_msg(void) | |||
264 | if (cnt <= 0 || cnt >= 80) | 264 | if (cnt <= 0 || cnt >= 80) |
265 | strcpy(buf, "<<< BOGUS MSG >>>"); | 265 | strcpy(buf, "<<< BOGUS MSG >>>"); |
266 | else { | 266 | else { |
267 | cp1 = (char *) &cpu->ipc_buffer[11]; | 267 | cp1 = (char *) &cpu->ipc_buffer[1]; |
268 | cp2 = buf; | 268 | cp2 = buf; |
269 | strcpy(cp2, cp1); | 269 | memcpy(cp2, cp1, cnt); |
270 | cp2[cnt] = '\0'; | ||
270 | 271 | ||
271 | while ((cp2 = strchr(cp2, '\r')) != 0) { | 272 | while ((cp2 = strchr(cp2, '\r')) != 0) { |
272 | *cp2 = ' '; | 273 | *cp2 = ' '; |
diff --git a/arch/alpha/kernel/sys_dp264.c b/arch/alpha/kernel/sys_dp264.c index 5bf401f7ea97..6c35159bc00e 100644 --- a/arch/alpha/kernel/sys_dp264.c +++ b/arch/alpha/kernel/sys_dp264.c | |||
@@ -190,9 +190,6 @@ static struct irq_chip clipper_irq_type = { | |||
190 | static void | 190 | static void |
191 | dp264_device_interrupt(unsigned long vector) | 191 | dp264_device_interrupt(unsigned long vector) |
192 | { | 192 | { |
193 | #if 1 | ||
194 | printk("dp264_device_interrupt: NOT IMPLEMENTED YET!!\n"); | ||
195 | #else | ||
196 | unsigned long pld; | 193 | unsigned long pld; |
197 | unsigned int i; | 194 | unsigned int i; |
198 | 195 | ||
@@ -210,12 +207,7 @@ dp264_device_interrupt(unsigned long vector) | |||
210 | isa_device_interrupt(vector); | 207 | isa_device_interrupt(vector); |
211 | else | 208 | else |
212 | handle_irq(16 + i); | 209 | handle_irq(16 + i); |
213 | #if 0 | ||
214 | TSUNAMI_cchip->dir0.csr = 1UL << i; mb(); | ||
215 | tmp = TSUNAMI_cchip->dir0.csr; | ||
216 | #endif | ||
217 | } | 210 | } |
218 | #endif | ||
219 | } | 211 | } |
220 | 212 | ||
221 | static void | 213 | static void |
diff --git a/arch/alpha/kernel/sys_marvel.c b/arch/alpha/kernel/sys_marvel.c index 407accc80877..c92e389ff219 100644 --- a/arch/alpha/kernel/sys_marvel.c +++ b/arch/alpha/kernel/sys_marvel.c | |||
@@ -317,8 +317,9 @@ marvel_init_irq(void) | |||
317 | } | 317 | } |
318 | 318 | ||
319 | static int | 319 | static int |
320 | marvel_map_irq(struct pci_dev *dev, u8 slot, u8 pin) | 320 | marvel_map_irq(const struct pci_dev *cdev, u8 slot, u8 pin) |
321 | { | 321 | { |
322 | struct pci_dev *dev = (struct pci_dev *)cdev; | ||
322 | struct pci_controller *hose = dev->sysdata; | 323 | struct pci_controller *hose = dev->sysdata; |
323 | struct io7_port *io7_port = hose->sysdata; | 324 | struct io7_port *io7_port = hose->sysdata; |
324 | struct io7 *io7 = io7_port->io7; | 325 | struct io7 *io7 = io7_port->io7; |
diff --git a/arch/alpha/kernel/systbls.S b/arch/alpha/kernel/systbls.S index 4284ec798ec9..dca9b3fb0071 100644 --- a/arch/alpha/kernel/systbls.S +++ b/arch/alpha/kernel/systbls.S | |||
@@ -524,6 +524,8 @@ sys_call_table: | |||
524 | .quad sys_sendmmsg | 524 | .quad sys_sendmmsg |
525 | .quad sys_process_vm_readv | 525 | .quad sys_process_vm_readv |
526 | .quad sys_process_vm_writev /* 505 */ | 526 | .quad sys_process_vm_writev /* 505 */ |
527 | .quad sys_kcmp | ||
528 | .quad sys_finit_module | ||
527 | 529 | ||
528 | .size sys_call_table, . - sys_call_table | 530 | .size sys_call_table, . - sys_call_table |
529 | .type sys_call_table, @object | 531 | .type sys_call_table, @object |
diff --git a/arch/alpha/kernel/time.c b/arch/alpha/kernel/time.c index e336694ca042..ea3395036556 100644 --- a/arch/alpha/kernel/time.c +++ b/arch/alpha/kernel/time.c | |||
@@ -105,9 +105,7 @@ void arch_irq_work_raise(void) | |||
105 | 105 | ||
106 | static inline __u32 rpcc(void) | 106 | static inline __u32 rpcc(void) |
107 | { | 107 | { |
108 | __u32 result; | 108 | return __builtin_alpha_rpcc(); |
109 | asm volatile ("rpcc %0" : "=r"(result)); | ||
110 | return result; | ||
111 | } | 109 | } |
112 | 110 | ||
113 | int update_persistent_clock(struct timespec now) | 111 | int update_persistent_clock(struct timespec now) |
diff --git a/arch/alpha/kernel/traps.c b/arch/alpha/kernel/traps.c index be1fba334bd0..bd0665cdc840 100644 --- a/arch/alpha/kernel/traps.c +++ b/arch/alpha/kernel/traps.c | |||
@@ -66,8 +66,8 @@ dik_show_regs(struct pt_regs *regs, unsigned long *r9_15) | |||
66 | { | 66 | { |
67 | printk("pc = [<%016lx>] ra = [<%016lx>] ps = %04lx %s\n", | 67 | printk("pc = [<%016lx>] ra = [<%016lx>] ps = %04lx %s\n", |
68 | regs->pc, regs->r26, regs->ps, print_tainted()); | 68 | regs->pc, regs->r26, regs->ps, print_tainted()); |
69 | print_symbol("pc is at %s\n", regs->pc); | 69 | printk("pc is at %pSR\n", (void *)regs->pc); |
70 | print_symbol("ra is at %s\n", regs->r26 ); | 70 | printk("ra is at %pSR\n", (void *)regs->r26); |
71 | printk("v0 = %016lx t0 = %016lx t1 = %016lx\n", | 71 | printk("v0 = %016lx t0 = %016lx t1 = %016lx\n", |
72 | regs->r0, regs->r1, regs->r2); | 72 | regs->r0, regs->r1, regs->r2); |
73 | printk("t2 = %016lx t3 = %016lx t4 = %016lx\n", | 73 | printk("t2 = %016lx t3 = %016lx t4 = %016lx\n", |
@@ -132,9 +132,7 @@ dik_show_trace(unsigned long *sp) | |||
132 | continue; | 132 | continue; |
133 | if (tmp >= (unsigned long) &_etext) | 133 | if (tmp >= (unsigned long) &_etext) |
134 | continue; | 134 | continue; |
135 | printk("[<%lx>]", tmp); | 135 | printk("[<%lx>] %pSR\n", tmp, (void *)tmp); |
136 | print_symbol(" %s", tmp); | ||
137 | printk("\n"); | ||
138 | if (i > 40) { | 136 | if (i > 40) { |
139 | printk(" ..."); | 137 | printk(" ..."); |
140 | break; | 138 | break; |
diff --git a/arch/arm/Kconfig b/arch/arm/Kconfig index ba412e02ec0c..37c0f4e978d4 100644 --- a/arch/arm/Kconfig +++ b/arch/arm/Kconfig | |||
@@ -1600,8 +1600,7 @@ config LOCAL_TIMERS | |||
1600 | config ARCH_NR_GPIO | 1600 | config ARCH_NR_GPIO |
1601 | int | 1601 | int |
1602 | default 1024 if ARCH_SHMOBILE || ARCH_TEGRA | 1602 | default 1024 if ARCH_SHMOBILE || ARCH_TEGRA |
1603 | default 512 if SOC_OMAP5 | 1603 | default 512 if ARCH_EXYNOS || ARCH_KEYSTONE || SOC_OMAP5 |
1604 | default 512 if ARCH_KEYSTONE | ||
1605 | default 392 if ARCH_U8500 | 1604 | default 392 if ARCH_U8500 |
1606 | default 352 if ARCH_VT8500 | 1605 | default 352 if ARCH_VT8500 |
1607 | default 288 if ARCH_SUNXI | 1606 | default 288 if ARCH_SUNXI |
diff --git a/arch/arm/boot/dts/atlas6.dtsi b/arch/arm/boot/dts/atlas6.dtsi index 9866cd736dee..a0f2721ea583 100644 --- a/arch/arm/boot/dts/atlas6.dtsi +++ b/arch/arm/boot/dts/atlas6.dtsi | |||
@@ -485,6 +485,12 @@ | |||
485 | sirf,function = "usp0"; | 485 | sirf,function = "usp0"; |
486 | }; | 486 | }; |
487 | }; | 487 | }; |
488 | usp0_uart_nostreamctrl_pins_a: usp0@1 { | ||
489 | usp0 { | ||
490 | sirf,pins = "usp0_uart_nostreamctrl_grp"; | ||
491 | sirf,function = "usp0_uart_nostreamctrl"; | ||
492 | }; | ||
493 | }; | ||
488 | usp1_pins_a: usp1@0 { | 494 | usp1_pins_a: usp1@0 { |
489 | usp1 { | 495 | usp1 { |
490 | sirf,pins = "usp1grp"; | 496 | sirf,pins = "usp1grp"; |
@@ -515,16 +521,16 @@ | |||
515 | sirf,function = "pulse_count"; | 521 | sirf,function = "pulse_count"; |
516 | }; | 522 | }; |
517 | }; | 523 | }; |
518 | cko0_rst_pins_a: cko0_rst@0 { | 524 | cko0_pins_a: cko0@0 { |
519 | cko0_rst { | 525 | cko0 { |
520 | sirf,pins = "cko0_rstgrp"; | 526 | sirf,pins = "cko0grp"; |
521 | sirf,function = "cko0_rst"; | 527 | sirf,function = "cko0"; |
522 | }; | 528 | }; |
523 | }; | 529 | }; |
524 | cko1_rst_pins_a: cko1_rst@0 { | 530 | cko1_pins_a: cko1@0 { |
525 | cko1_rst { | 531 | cko1 { |
526 | sirf,pins = "cko1_rstgrp"; | 532 | sirf,pins = "cko1grp"; |
527 | sirf,function = "cko1_rst"; | 533 | sirf,function = "cko1"; |
528 | }; | 534 | }; |
529 | }; | 535 | }; |
530 | }; | 536 | }; |
diff --git a/arch/arm/boot/dts/imx28-apx4devkit.dts b/arch/arm/boot/dts/imx28-apx4devkit.dts index 43bf3c796cba..0e7fed47bd8d 100644 --- a/arch/arm/boot/dts/imx28-apx4devkit.dts +++ b/arch/arm/boot/dts/imx28-apx4devkit.dts | |||
@@ -147,7 +147,7 @@ | |||
147 | reg = <0x0a>; | 147 | reg = <0x0a>; |
148 | VDDA-supply = <®_3p3v>; | 148 | VDDA-supply = <®_3p3v>; |
149 | VDDIO-supply = <®_3p3v>; | 149 | VDDIO-supply = <®_3p3v>; |
150 | 150 | clocks = <&saif0>; | |
151 | }; | 151 | }; |
152 | 152 | ||
153 | pcf8563: rtc@51 { | 153 | pcf8563: rtc@51 { |
diff --git a/arch/arm/boot/dts/imx28-evk.dts b/arch/arm/boot/dts/imx28-evk.dts index 1f0d38d7b16f..e035f4664b97 100644 --- a/arch/arm/boot/dts/imx28-evk.dts +++ b/arch/arm/boot/dts/imx28-evk.dts | |||
@@ -195,7 +195,7 @@ | |||
195 | reg = <0x0a>; | 195 | reg = <0x0a>; |
196 | VDDA-supply = <®_3p3v>; | 196 | VDDA-supply = <®_3p3v>; |
197 | VDDIO-supply = <®_3p3v>; | 197 | VDDIO-supply = <®_3p3v>; |
198 | 198 | clocks = <&saif0>; | |
199 | }; | 199 | }; |
200 | 200 | ||
201 | at24@51 { | 201 | at24@51 { |
diff --git a/arch/arm/boot/dts/imx28-m28evk.dts b/arch/arm/boot/dts/imx28-m28evk.dts index 880df2f13be8..44d9da57736e 100644 --- a/arch/arm/boot/dts/imx28-m28evk.dts +++ b/arch/arm/boot/dts/imx28-m28evk.dts | |||
@@ -184,7 +184,7 @@ | |||
184 | reg = <0x0a>; | 184 | reg = <0x0a>; |
185 | VDDA-supply = <®_3p3v>; | 185 | VDDA-supply = <®_3p3v>; |
186 | VDDIO-supply = <®_3p3v>; | 186 | VDDIO-supply = <®_3p3v>; |
187 | 187 | clocks = <&saif0>; | |
188 | }; | 188 | }; |
189 | 189 | ||
190 | eeprom: eeprom@51 { | 190 | eeprom: eeprom@51 { |
diff --git a/arch/arm/boot/dts/imx28.dtsi b/arch/arm/boot/dts/imx28.dtsi index 6a8acb01b1d3..9524a0571281 100644 --- a/arch/arm/boot/dts/imx28.dtsi +++ b/arch/arm/boot/dts/imx28.dtsi | |||
@@ -837,6 +837,7 @@ | |||
837 | compatible = "fsl,imx28-saif"; | 837 | compatible = "fsl,imx28-saif"; |
838 | reg = <0x80042000 0x2000>; | 838 | reg = <0x80042000 0x2000>; |
839 | interrupts = <59 80>; | 839 | interrupts = <59 80>; |
840 | #clock-cells = <0>; | ||
840 | clocks = <&clks 53>; | 841 | clocks = <&clks 53>; |
841 | dmas = <&dma_apbx 4>; | 842 | dmas = <&dma_apbx 4>; |
842 | dma-names = "rx-tx"; | 843 | dma-names = "rx-tx"; |
diff --git a/arch/arm/boot/dts/imx51-babbage.dts b/arch/arm/boot/dts/imx51-babbage.dts index 6dd9486c755b..ad3471ca17c7 100644 --- a/arch/arm/boot/dts/imx51-babbage.dts +++ b/arch/arm/boot/dts/imx51-babbage.dts | |||
@@ -61,6 +61,16 @@ | |||
61 | mux-int-port = <2>; | 61 | mux-int-port = <2>; |
62 | mux-ext-port = <3>; | 62 | mux-ext-port = <3>; |
63 | }; | 63 | }; |
64 | |||
65 | clocks { | ||
66 | clk_26M: codec_clock { | ||
67 | compatible = "fixed-clock"; | ||
68 | reg=<0>; | ||
69 | #clock-cells = <0>; | ||
70 | clock-frequency = <26000000>; | ||
71 | gpios = <&gpio4 26 1>; | ||
72 | }; | ||
73 | }; | ||
64 | }; | 74 | }; |
65 | 75 | ||
66 | &esdhc1 { | 76 | &esdhc1 { |
@@ -229,6 +239,7 @@ | |||
229 | MX51_PAD_EIM_A27__GPIO2_21 0x5 | 239 | MX51_PAD_EIM_A27__GPIO2_21 0x5 |
230 | MX51_PAD_CSPI1_SS0__GPIO4_24 0x85 | 240 | MX51_PAD_CSPI1_SS0__GPIO4_24 0x85 |
231 | MX51_PAD_CSPI1_SS1__GPIO4_25 0x85 | 241 | MX51_PAD_CSPI1_SS1__GPIO4_25 0x85 |
242 | MX51_PAD_CSPI1_RDY__GPIO4_26 0x80000000 | ||
232 | >; | 243 | >; |
233 | }; | 244 | }; |
234 | }; | 245 | }; |
@@ -255,7 +266,7 @@ | |||
255 | sgtl5000: codec@0a { | 266 | sgtl5000: codec@0a { |
256 | compatible = "fsl,sgtl5000"; | 267 | compatible = "fsl,sgtl5000"; |
257 | reg = <0x0a>; | 268 | reg = <0x0a>; |
258 | clock-frequency = <26000000>; | 269 | clocks = <&clk_26M>; |
259 | VDDA-supply = <&vdig_reg>; | 270 | VDDA-supply = <&vdig_reg>; |
260 | VDDIO-supply = <&vvideo_reg>; | 271 | VDDIO-supply = <&vvideo_reg>; |
261 | }; | 272 | }; |
diff --git a/arch/arm/boot/dts/imx53-mba53.dts b/arch/arm/boot/dts/imx53-mba53.dts index aaa33bc99f78..a63090267941 100644 --- a/arch/arm/boot/dts/imx53-mba53.dts +++ b/arch/arm/boot/dts/imx53-mba53.dts | |||
@@ -27,7 +27,7 @@ | |||
27 | 27 | ||
28 | backlight { | 28 | backlight { |
29 | compatible = "pwm-backlight"; | 29 | compatible = "pwm-backlight"; |
30 | pwms = <&pwm2 0 50000 0 0>; | 30 | pwms = <&pwm2 0 50000>; |
31 | brightness-levels = <0 24 28 32 36 40 44 48 52 56 60 64 68 72 76 80 84 88 92 96 100>; | 31 | brightness-levels = <0 24 28 32 36 40 44 48 52 56 60 64 68 72 76 80 84 88 92 96 100>; |
32 | default-brightness-level = <10>; | 32 | default-brightness-level = <10>; |
33 | enable-gpios = <&gpio7 7 0>; | 33 | enable-gpios = <&gpio7 7 0>; |
diff --git a/arch/arm/boot/dts/imx53.dtsi b/arch/arm/boot/dts/imx53.dtsi index 3895fbba8fce..569aa9f2c4ed 100644 --- a/arch/arm/boot/dts/imx53.dtsi +++ b/arch/arm/boot/dts/imx53.dtsi | |||
@@ -725,15 +725,15 @@ | |||
725 | uart1 { | 725 | uart1 { |
726 | pinctrl_uart1_1: uart1grp-1 { | 726 | pinctrl_uart1_1: uart1grp-1 { |
727 | fsl,pins = < | 727 | fsl,pins = < |
728 | MX53_PAD_CSI0_DAT10__UART1_TXD_MUX 0x1c5 | 728 | MX53_PAD_CSI0_DAT10__UART1_TXD_MUX 0x1e4 |
729 | MX53_PAD_CSI0_DAT11__UART1_RXD_MUX 0x1c5 | 729 | MX53_PAD_CSI0_DAT11__UART1_RXD_MUX 0x1e4 |
730 | >; | 730 | >; |
731 | }; | 731 | }; |
732 | 732 | ||
733 | pinctrl_uart1_2: uart1grp-2 { | 733 | pinctrl_uart1_2: uart1grp-2 { |
734 | fsl,pins = < | 734 | fsl,pins = < |
735 | MX53_PAD_PATA_DIOW__UART1_TXD_MUX 0x1c5 | 735 | MX53_PAD_PATA_DIOW__UART1_TXD_MUX 0x1e4 |
736 | MX53_PAD_PATA_DMACK__UART1_RXD_MUX 0x1c5 | 736 | MX53_PAD_PATA_DMACK__UART1_RXD_MUX 0x1e4 |
737 | >; | 737 | >; |
738 | }; | 738 | }; |
739 | 739 | ||
@@ -748,8 +748,8 @@ | |||
748 | uart2 { | 748 | uart2 { |
749 | pinctrl_uart2_1: uart2grp-1 { | 749 | pinctrl_uart2_1: uart2grp-1 { |
750 | fsl,pins = < | 750 | fsl,pins = < |
751 | MX53_PAD_PATA_BUFFER_EN__UART2_RXD_MUX 0x1c5 | 751 | MX53_PAD_PATA_BUFFER_EN__UART2_RXD_MUX 0x1e4 |
752 | MX53_PAD_PATA_DMARQ__UART2_TXD_MUX 0x1c5 | 752 | MX53_PAD_PATA_DMARQ__UART2_TXD_MUX 0x1e4 |
753 | >; | 753 | >; |
754 | }; | 754 | }; |
755 | 755 | ||
@@ -766,17 +766,17 @@ | |||
766 | uart3 { | 766 | uart3 { |
767 | pinctrl_uart3_1: uart3grp-1 { | 767 | pinctrl_uart3_1: uart3grp-1 { |
768 | fsl,pins = < | 768 | fsl,pins = < |
769 | MX53_PAD_PATA_CS_0__UART3_TXD_MUX 0x1c5 | 769 | MX53_PAD_PATA_CS_0__UART3_TXD_MUX 0x1e4 |
770 | MX53_PAD_PATA_CS_1__UART3_RXD_MUX 0x1c5 | 770 | MX53_PAD_PATA_CS_1__UART3_RXD_MUX 0x1e4 |
771 | MX53_PAD_PATA_DA_1__UART3_CTS 0x1c5 | 771 | MX53_PAD_PATA_DA_1__UART3_CTS 0x1e4 |
772 | MX53_PAD_PATA_DA_2__UART3_RTS 0x1c5 | 772 | MX53_PAD_PATA_DA_2__UART3_RTS 0x1e4 |
773 | >; | 773 | >; |
774 | }; | 774 | }; |
775 | 775 | ||
776 | pinctrl_uart3_2: uart3grp-2 { | 776 | pinctrl_uart3_2: uart3grp-2 { |
777 | fsl,pins = < | 777 | fsl,pins = < |
778 | MX53_PAD_PATA_CS_0__UART3_TXD_MUX 0x1c5 | 778 | MX53_PAD_PATA_CS_0__UART3_TXD_MUX 0x1e4 |
779 | MX53_PAD_PATA_CS_1__UART3_RXD_MUX 0x1c5 | 779 | MX53_PAD_PATA_CS_1__UART3_RXD_MUX 0x1e4 |
780 | >; | 780 | >; |
781 | }; | 781 | }; |
782 | 782 | ||
@@ -785,8 +785,8 @@ | |||
785 | uart4 { | 785 | uart4 { |
786 | pinctrl_uart4_1: uart4grp-1 { | 786 | pinctrl_uart4_1: uart4grp-1 { |
787 | fsl,pins = < | 787 | fsl,pins = < |
788 | MX53_PAD_KEY_COL0__UART4_TXD_MUX 0x1c5 | 788 | MX53_PAD_KEY_COL0__UART4_TXD_MUX 0x1e4 |
789 | MX53_PAD_KEY_ROW0__UART4_RXD_MUX 0x1c5 | 789 | MX53_PAD_KEY_ROW0__UART4_RXD_MUX 0x1e4 |
790 | >; | 790 | >; |
791 | }; | 791 | }; |
792 | }; | 792 | }; |
@@ -794,8 +794,8 @@ | |||
794 | uart5 { | 794 | uart5 { |
795 | pinctrl_uart5_1: uart5grp-1 { | 795 | pinctrl_uart5_1: uart5grp-1 { |
796 | fsl,pins = < | 796 | fsl,pins = < |
797 | MX53_PAD_KEY_COL1__UART5_TXD_MUX 0x1c5 | 797 | MX53_PAD_KEY_COL1__UART5_TXD_MUX 0x1e4 |
798 | MX53_PAD_KEY_ROW1__UART5_RXD_MUX 0x1c5 | 798 | MX53_PAD_KEY_ROW1__UART5_RXD_MUX 0x1e4 |
799 | >; | 799 | >; |
800 | }; | 800 | }; |
801 | }; | 801 | }; |
diff --git a/arch/arm/boot/dts/prima2.dtsi b/arch/arm/boot/dts/prima2.dtsi index 05e9489cf95c..bbeb623fc2c6 100644 --- a/arch/arm/boot/dts/prima2.dtsi +++ b/arch/arm/boot/dts/prima2.dtsi | |||
@@ -515,16 +515,16 @@ | |||
515 | sirf,function = "pulse_count"; | 515 | sirf,function = "pulse_count"; |
516 | }; | 516 | }; |
517 | }; | 517 | }; |
518 | cko0_rst_pins_a: cko0_rst@0 { | 518 | cko0_pins_a: cko0@0 { |
519 | cko0_rst { | 519 | cko0 { |
520 | sirf,pins = "cko0_rstgrp"; | 520 | sirf,pins = "cko0grp"; |
521 | sirf,function = "cko0_rst"; | 521 | sirf,function = "cko0"; |
522 | }; | 522 | }; |
523 | }; | 523 | }; |
524 | cko1_rst_pins_a: cko1_rst@0 { | 524 | cko1_pins_a: cko1@0 { |
525 | cko1_rst { | 525 | cko1 { |
526 | sirf,pins = "cko1_rstgrp"; | 526 | sirf,pins = "cko1grp"; |
527 | sirf,function = "cko1_rst"; | 527 | sirf,function = "cko1"; |
528 | }; | 528 | }; |
529 | }; | 529 | }; |
530 | }; | 530 | }; |
diff --git a/arch/arm/boot/dts/stih416-pinctrl.dtsi b/arch/arm/boot/dts/stih416-pinctrl.dtsi index 957b21a71b4b..0f246c979262 100644 --- a/arch/arm/boot/dts/stih416-pinctrl.dtsi +++ b/arch/arm/boot/dts/stih416-pinctrl.dtsi | |||
@@ -166,6 +166,15 @@ | |||
166 | reg = <0x9000 0x100>; | 166 | reg = <0x9000 0x100>; |
167 | st,bank-name = "PIO31"; | 167 | st,bank-name = "PIO31"; |
168 | }; | 168 | }; |
169 | |||
170 | serial2-oe { | ||
171 | pinctrl_serial2_oe: serial2-1 { | ||
172 | st,pins { | ||
173 | output-enable = <&PIO11 3 ALT2 OUT>; | ||
174 | }; | ||
175 | }; | ||
176 | }; | ||
177 | |||
169 | }; | 178 | }; |
170 | 179 | ||
171 | pin-controller-rear { | 180 | pin-controller-rear { |
@@ -218,7 +227,6 @@ | |||
218 | st,pins { | 227 | st,pins { |
219 | tx = <&PIO17 4 ALT2 OUT>; | 228 | tx = <&PIO17 4 ALT2 OUT>; |
220 | rx = <&PIO17 5 ALT2 IN>; | 229 | rx = <&PIO17 5 ALT2 IN>; |
221 | output-enable = <&PIO11 3 ALT2 OUT>; | ||
222 | }; | 230 | }; |
223 | }; | 231 | }; |
224 | }; | 232 | }; |
diff --git a/arch/arm/boot/dts/stih416.dtsi b/arch/arm/boot/dts/stih416.dtsi index 3cecd9689a49..1a0326ea7d07 100644 --- a/arch/arm/boot/dts/stih416.dtsi +++ b/arch/arm/boot/dts/stih416.dtsi | |||
@@ -79,7 +79,7 @@ | |||
79 | interrupts = <0 197 0>; | 79 | interrupts = <0 197 0>; |
80 | clocks = <&CLK_S_ICN_REG_0>; | 80 | clocks = <&CLK_S_ICN_REG_0>; |
81 | pinctrl-names = "default"; | 81 | pinctrl-names = "default"; |
82 | pinctrl-0 = <&pinctrl_serial2>; | 82 | pinctrl-0 = <&pinctrl_serial2 &pinctrl_serial2_oe>; |
83 | }; | 83 | }; |
84 | 84 | ||
85 | /* SBC_UART1 */ | 85 | /* SBC_UART1 */ |
diff --git a/arch/arm/boot/dts/twl4030.dtsi b/arch/arm/boot/dts/twl4030.dtsi index b3034da00a37..ae6a17aed9ee 100644 --- a/arch/arm/boot/dts/twl4030.dtsi +++ b/arch/arm/boot/dts/twl4030.dtsi | |||
@@ -47,6 +47,12 @@ | |||
47 | regulator-max-microvolt = <3150000>; | 47 | regulator-max-microvolt = <3150000>; |
48 | }; | 48 | }; |
49 | 49 | ||
50 | vmmc2: regulator-vmmc2 { | ||
51 | compatible = "ti,twl4030-vmmc2"; | ||
52 | regulator-min-microvolt = <1850000>; | ||
53 | regulator-max-microvolt = <3150000>; | ||
54 | }; | ||
55 | |||
50 | vusb1v5: regulator-vusb1v5 { | 56 | vusb1v5: regulator-vusb1v5 { |
51 | compatible = "ti,twl4030-vusb1v5"; | 57 | compatible = "ti,twl4030-vusb1v5"; |
52 | }; | 58 | }; |
diff --git a/arch/arm/boot/dts/vf610.dtsi b/arch/arm/boot/dts/vf610.dtsi index e1eb7dadda80..67d929cf9804 100644 --- a/arch/arm/boot/dts/vf610.dtsi +++ b/arch/arm/boot/dts/vf610.dtsi | |||
@@ -442,8 +442,8 @@ | |||
442 | compatible = "fsl,mvf600-fec"; | 442 | compatible = "fsl,mvf600-fec"; |
443 | reg = <0x400d0000 0x1000>; | 443 | reg = <0x400d0000 0x1000>; |
444 | interrupts = <0 78 0x04>; | 444 | interrupts = <0 78 0x04>; |
445 | clocks = <&clks VF610_CLK_ENET>, | 445 | clocks = <&clks VF610_CLK_ENET0>, |
446 | <&clks VF610_CLK_ENET>, | 446 | <&clks VF610_CLK_ENET0>, |
447 | <&clks VF610_CLK_ENET>; | 447 | <&clks VF610_CLK_ENET>; |
448 | clock-names = "ipg", "ahb", "ptp"; | 448 | clock-names = "ipg", "ahb", "ptp"; |
449 | status = "disabled"; | 449 | status = "disabled"; |
@@ -453,8 +453,8 @@ | |||
453 | compatible = "fsl,mvf600-fec"; | 453 | compatible = "fsl,mvf600-fec"; |
454 | reg = <0x400d1000 0x1000>; | 454 | reg = <0x400d1000 0x1000>; |
455 | interrupts = <0 79 0x04>; | 455 | interrupts = <0 79 0x04>; |
456 | clocks = <&clks VF610_CLK_ENET>, | 456 | clocks = <&clks VF610_CLK_ENET1>, |
457 | <&clks VF610_CLK_ENET>, | 457 | <&clks VF610_CLK_ENET1>, |
458 | <&clks VF610_CLK_ENET>; | 458 | <&clks VF610_CLK_ENET>; |
459 | clock-names = "ipg", "ahb", "ptp"; | 459 | clock-names = "ipg", "ahb", "ptp"; |
460 | status = "disabled"; | 460 | status = "disabled"; |
diff --git a/arch/arm/common/edma.c b/arch/arm/common/edma.c index a432e6c1dac1..39ad030ac0c7 100644 --- a/arch/arm/common/edma.c +++ b/arch/arm/common/edma.c | |||
@@ -26,7 +26,6 @@ | |||
26 | #include <linux/io.h> | 26 | #include <linux/io.h> |
27 | #include <linux/slab.h> | 27 | #include <linux/slab.h> |
28 | #include <linux/edma.h> | 28 | #include <linux/edma.h> |
29 | #include <linux/err.h> | ||
30 | #include <linux/of_address.h> | 29 | #include <linux/of_address.h> |
31 | #include <linux/of_device.h> | 30 | #include <linux/of_device.h> |
32 | #include <linux/of_dma.h> | 31 | #include <linux/of_dma.h> |
diff --git a/arch/arm/configs/da8xx_omapl_defconfig b/arch/arm/configs/da8xx_omapl_defconfig index 7c868139bdb0..1571bea48bed 100644 --- a/arch/arm/configs/da8xx_omapl_defconfig +++ b/arch/arm/configs/da8xx_omapl_defconfig | |||
@@ -102,6 +102,8 @@ CONFIG_SND_SOC=m | |||
102 | CONFIG_SND_DAVINCI_SOC=m | 102 | CONFIG_SND_DAVINCI_SOC=m |
103 | # CONFIG_HID_SUPPORT is not set | 103 | # CONFIG_HID_SUPPORT is not set |
104 | # CONFIG_USB_SUPPORT is not set | 104 | # CONFIG_USB_SUPPORT is not set |
105 | CONFIG_DMADEVICES=y | ||
106 | CONFIG_TI_EDMA=y | ||
105 | CONFIG_EXT2_FS=y | 107 | CONFIG_EXT2_FS=y |
106 | CONFIG_EXT3_FS=y | 108 | CONFIG_EXT3_FS=y |
107 | CONFIG_XFS_FS=m | 109 | CONFIG_XFS_FS=m |
diff --git a/arch/arm/configs/davinci_all_defconfig b/arch/arm/configs/davinci_all_defconfig index c86fd75e181a..ab2f7378352c 100644 --- a/arch/arm/configs/davinci_all_defconfig +++ b/arch/arm/configs/davinci_all_defconfig | |||
@@ -162,6 +162,8 @@ CONFIG_LEDS_TRIGGERS=y | |||
162 | CONFIG_LEDS_TRIGGER_TIMER=m | 162 | CONFIG_LEDS_TRIGGER_TIMER=m |
163 | CONFIG_LEDS_TRIGGER_HEARTBEAT=m | 163 | CONFIG_LEDS_TRIGGER_HEARTBEAT=m |
164 | CONFIG_RTC_CLASS=y | 164 | CONFIG_RTC_CLASS=y |
165 | CONFIG_DMADEVICES=y | ||
166 | CONFIG_TI_EDMA=y | ||
165 | CONFIG_EXT2_FS=y | 167 | CONFIG_EXT2_FS=y |
166 | CONFIG_EXT3_FS=y | 168 | CONFIG_EXT3_FS=y |
167 | CONFIG_XFS_FS=m | 169 | CONFIG_XFS_FS=m |
diff --git a/arch/arm/configs/multi_v7_defconfig b/arch/arm/configs/multi_v7_defconfig index fe0bdc361d2c..6e572c64cf5a 100644 --- a/arch/arm/configs/multi_v7_defconfig +++ b/arch/arm/configs/multi_v7_defconfig | |||
@@ -53,6 +53,7 @@ CONFIG_IP_PNP=y | |||
53 | CONFIG_IP_PNP_DHCP=y | 53 | CONFIG_IP_PNP_DHCP=y |
54 | CONFIG_DEVTMPFS=y | 54 | CONFIG_DEVTMPFS=y |
55 | CONFIG_DEVTMPFS_MOUNT=y | 55 | CONFIG_DEVTMPFS_MOUNT=y |
56 | CONFIG_OMAP_OCP2SCP=y | ||
56 | CONFIG_BLK_DEV_SD=y | 57 | CONFIG_BLK_DEV_SD=y |
57 | CONFIG_ATA=y | 58 | CONFIG_ATA=y |
58 | CONFIG_SATA_AHCI_PLATFORM=y | 59 | CONFIG_SATA_AHCI_PLATFORM=y |
@@ -61,6 +62,7 @@ CONFIG_SATA_MV=y | |||
61 | CONFIG_NETDEVICES=y | 62 | CONFIG_NETDEVICES=y |
62 | CONFIG_SUN4I_EMAC=y | 63 | CONFIG_SUN4I_EMAC=y |
63 | CONFIG_NET_CALXEDA_XGMAC=y | 64 | CONFIG_NET_CALXEDA_XGMAC=y |
65 | CONFIG_KS8851=y | ||
64 | CONFIG_SMSC911X=y | 66 | CONFIG_SMSC911X=y |
65 | CONFIG_STMMAC_ETH=y | 67 | CONFIG_STMMAC_ETH=y |
66 | CONFIG_MDIO_SUN4I=y | 68 | CONFIG_MDIO_SUN4I=y |
@@ -89,6 +91,7 @@ CONFIG_I2C_DESIGNWARE_PLATFORM=y | |||
89 | CONFIG_I2C_SIRF=y | 91 | CONFIG_I2C_SIRF=y |
90 | CONFIG_I2C_TEGRA=y | 92 | CONFIG_I2C_TEGRA=y |
91 | CONFIG_SPI=y | 93 | CONFIG_SPI=y |
94 | CONFIG_SPI_OMAP24XX=y | ||
92 | CONFIG_SPI_PL022=y | 95 | CONFIG_SPI_PL022=y |
93 | CONFIG_SPI_SIRF=y | 96 | CONFIG_SPI_SIRF=y |
94 | CONFIG_SPI_TEGRA114=y | 97 | CONFIG_SPI_TEGRA114=y |
@@ -111,11 +114,12 @@ CONFIG_FB_SIMPLE=y | |||
111 | CONFIG_USB=y | 114 | CONFIG_USB=y |
112 | CONFIG_USB_XHCI_HCD=y | 115 | CONFIG_USB_XHCI_HCD=y |
113 | CONFIG_USB_EHCI_HCD=y | 116 | CONFIG_USB_EHCI_HCD=y |
114 | CONFIG_USB_EHCI_MXC=y | ||
115 | CONFIG_USB_EHCI_TEGRA=y | 117 | CONFIG_USB_EHCI_TEGRA=y |
116 | CONFIG_USB_EHCI_HCD_PLATFORM=y | 118 | CONFIG_USB_EHCI_HCD_PLATFORM=y |
117 | CONFIG_USB_ISP1760_HCD=y | 119 | CONFIG_USB_ISP1760_HCD=y |
118 | CONFIG_USB_STORAGE=y | 120 | CONFIG_USB_STORAGE=y |
121 | CONFIG_USB_CHIPIDEA=y | ||
122 | CONFIG_USB_CHIPIDEA_HOST=y | ||
119 | CONFIG_AB8500_USB=y | 123 | CONFIG_AB8500_USB=y |
120 | CONFIG_NOP_USB_XCEIV=y | 124 | CONFIG_NOP_USB_XCEIV=y |
121 | CONFIG_OMAP_USB2=y | 125 | CONFIG_OMAP_USB2=y |
diff --git a/arch/arm/configs/nhk8815_defconfig b/arch/arm/configs/nhk8815_defconfig index 35f8cf299fa2..263ae3869e32 100644 --- a/arch/arm/configs/nhk8815_defconfig +++ b/arch/arm/configs/nhk8815_defconfig | |||
@@ -1,6 +1,8 @@ | |||
1 | # CONFIG_LOCALVERSION_AUTO is not set | 1 | # CONFIG_LOCALVERSION_AUTO is not set |
2 | # CONFIG_SWAP is not set | 2 | # CONFIG_SWAP is not set |
3 | CONFIG_SYSVIPC=y | 3 | CONFIG_SYSVIPC=y |
4 | CONFIG_NO_HZ_IDLE=y | ||
5 | CONFIG_HIGH_RES_TIMERS=y | ||
4 | CONFIG_IKCONFIG=y | 6 | CONFIG_IKCONFIG=y |
5 | CONFIG_IKCONFIG_PROC=y | 7 | CONFIG_IKCONFIG_PROC=y |
6 | CONFIG_LOG_BUF_SHIFT=14 | 8 | CONFIG_LOG_BUF_SHIFT=14 |
@@ -48,7 +50,6 @@ CONFIG_UEVENT_HELPER_PATH="/sbin/hotplug" | |||
48 | CONFIG_MTD=y | 50 | CONFIG_MTD=y |
49 | CONFIG_MTD_TESTS=m | 51 | CONFIG_MTD_TESTS=m |
50 | CONFIG_MTD_CMDLINE_PARTS=y | 52 | CONFIG_MTD_CMDLINE_PARTS=y |
51 | CONFIG_MTD_CHAR=y | ||
52 | CONFIG_MTD_BLOCK=y | 53 | CONFIG_MTD_BLOCK=y |
53 | CONFIG_MTD_NAND_ECC_SMC=y | 54 | CONFIG_MTD_NAND_ECC_SMC=y |
54 | CONFIG_MTD_NAND=y | 55 | CONFIG_MTD_NAND=y |
@@ -94,8 +95,10 @@ CONFIG_I2C_GPIO=y | |||
94 | CONFIG_I2C_NOMADIK=y | 95 | CONFIG_I2C_NOMADIK=y |
95 | CONFIG_DEBUG_GPIO=y | 96 | CONFIG_DEBUG_GPIO=y |
96 | # CONFIG_HWMON is not set | 97 | # CONFIG_HWMON is not set |
98 | CONFIG_REGULATOR=y | ||
97 | CONFIG_MMC=y | 99 | CONFIG_MMC=y |
98 | CONFIG_MMC_CLKGATE=y | 100 | CONFIG_MMC_UNSAFE_RESUME=y |
101 | # CONFIG_MMC_BLOCK_BOUNCE is not set | ||
99 | CONFIG_MMC_ARMMMCI=y | 102 | CONFIG_MMC_ARMMMCI=y |
100 | CONFIG_NEW_LEDS=y | 103 | CONFIG_NEW_LEDS=y |
101 | CONFIG_LEDS_CLASS=y | 104 | CONFIG_LEDS_CLASS=y |
diff --git a/arch/arm/mach-davinci/board-dm365-evm.c b/arch/arm/mach-davinci/board-dm365-evm.c index afbc439f11d4..4cdb61c54459 100644 --- a/arch/arm/mach-davinci/board-dm365-evm.c +++ b/arch/arm/mach-davinci/board-dm365-evm.c | |||
@@ -505,7 +505,7 @@ static struct vpbe_output dm365evm_vpbe_outputs[] = { | |||
505 | /* | 505 | /* |
506 | * Amplifiers on the board | 506 | * Amplifiers on the board |
507 | */ | 507 | */ |
508 | struct ths7303_platform_data ths7303_pdata = { | 508 | static struct ths7303_platform_data ths7303_pdata = { |
509 | .ch_1 = 3, | 509 | .ch_1 = 3, |
510 | .ch_2 = 3, | 510 | .ch_2 = 3, |
511 | .ch_3 = 3, | 511 | .ch_3 = 3, |
diff --git a/arch/arm/mach-davinci/dm355.c b/arch/arm/mach-davinci/dm355.c index 42ef53f62c6c..86100d179694 100644 --- a/arch/arm/mach-davinci/dm355.c +++ b/arch/arm/mach-davinci/dm355.c | |||
@@ -860,7 +860,7 @@ static struct platform_device dm355_vpbe_display = { | |||
860 | }, | 860 | }, |
861 | }; | 861 | }; |
862 | 862 | ||
863 | struct venc_platform_data dm355_venc_pdata = { | 863 | static struct venc_platform_data dm355_venc_pdata = { |
864 | .setup_pinmux = dm355_vpbe_setup_pinmux, | 864 | .setup_pinmux = dm355_vpbe_setup_pinmux, |
865 | .setup_clock = dm355_venc_setup_clock, | 865 | .setup_clock = dm355_venc_setup_clock, |
866 | }; | 866 | }; |
diff --git a/arch/arm/mach-davinci/dm365.c b/arch/arm/mach-davinci/dm365.c index fa7af5eda52d..dad28029ba9b 100644 --- a/arch/arm/mach-davinci/dm365.c +++ b/arch/arm/mach-davinci/dm365.c | |||
@@ -1349,7 +1349,7 @@ static struct platform_device dm365_vpbe_display = { | |||
1349 | }, | 1349 | }, |
1350 | }; | 1350 | }; |
1351 | 1351 | ||
1352 | struct venc_platform_data dm365_venc_pdata = { | 1352 | static struct venc_platform_data dm365_venc_pdata = { |
1353 | .setup_pinmux = dm365_vpbe_setup_pinmux, | 1353 | .setup_pinmux = dm365_vpbe_setup_pinmux, |
1354 | .setup_clock = dm365_venc_setup_clock, | 1354 | .setup_clock = dm365_venc_setup_clock, |
1355 | }; | 1355 | }; |
diff --git a/arch/arm/mach-exynos/Kconfig b/arch/arm/mach-exynos/Kconfig index 855d4a7b462d..5952e68c76c4 100644 --- a/arch/arm/mach-exynos/Kconfig +++ b/arch/arm/mach-exynos/Kconfig | |||
@@ -92,6 +92,7 @@ config SOC_EXYNOS5440 | |||
92 | bool "SAMSUNG EXYNOS5440" | 92 | bool "SAMSUNG EXYNOS5440" |
93 | default y | 93 | default y |
94 | depends on ARCH_EXYNOS5 | 94 | depends on ARCH_EXYNOS5 |
95 | select ARCH_DMA_ADDR_T_64BIT if ARM_LPAE | ||
95 | select ARCH_HAS_OPP | 96 | select ARCH_HAS_OPP |
96 | select HAVE_ARM_ARCH_TIMER | 97 | select HAVE_ARM_ARCH_TIMER |
97 | select AUTO_ZRELADDR | 98 | select AUTO_ZRELADDR |
diff --git a/arch/arm/mach-exynos/Makefile b/arch/arm/mach-exynos/Makefile index e970a7a4e278..53696154aead 100644 --- a/arch/arm/mach-exynos/Makefile +++ b/arch/arm/mach-exynos/Makefile | |||
@@ -14,7 +14,7 @@ obj- := | |||
14 | 14 | ||
15 | obj-$(CONFIG_ARCH_EXYNOS) += common.o | 15 | obj-$(CONFIG_ARCH_EXYNOS) += common.o |
16 | 16 | ||
17 | obj-$(CONFIG_PM) += pm.o | 17 | obj-$(CONFIG_S5P_PM) += pm.o |
18 | obj-$(CONFIG_PM_GENERIC_DOMAINS) += pm_domains.o | 18 | obj-$(CONFIG_PM_GENERIC_DOMAINS) += pm_domains.o |
19 | obj-$(CONFIG_CPU_IDLE) += cpuidle.o | 19 | obj-$(CONFIG_CPU_IDLE) += cpuidle.o |
20 | 20 | ||
diff --git a/arch/arm/mach-exynos/common.c b/arch/arm/mach-exynos/common.c index 164685bd25c8..ba95e5db2501 100644 --- a/arch/arm/mach-exynos/common.c +++ b/arch/arm/mach-exynos/common.c | |||
@@ -58,7 +58,6 @@ static const char name_exynos5440[] = "EXYNOS5440"; | |||
58 | 58 | ||
59 | static void exynos4_map_io(void); | 59 | static void exynos4_map_io(void); |
60 | static void exynos5_map_io(void); | 60 | static void exynos5_map_io(void); |
61 | static void exynos5440_map_io(void); | ||
62 | static int exynos_init(void); | 61 | static int exynos_init(void); |
63 | 62 | ||
64 | static struct cpu_table cpu_ids[] __initdata = { | 63 | static struct cpu_table cpu_ids[] __initdata = { |
@@ -95,7 +94,6 @@ static struct cpu_table cpu_ids[] __initdata = { | |||
95 | }, { | 94 | }, { |
96 | .idcode = EXYNOS5440_SOC_ID, | 95 | .idcode = EXYNOS5440_SOC_ID, |
97 | .idmask = EXYNOS5_SOC_MASK, | 96 | .idmask = EXYNOS5_SOC_MASK, |
98 | .map_io = exynos5440_map_io, | ||
99 | .init = exynos_init, | 97 | .init = exynos_init, |
100 | .name = name_exynos5440, | 98 | .name = name_exynos5440, |
101 | }, | 99 | }, |
@@ -150,11 +148,6 @@ static struct map_desc exynos4_iodesc[] __initdata = { | |||
150 | .length = SZ_64K, | 148 | .length = SZ_64K, |
151 | .type = MT_DEVICE, | 149 | .type = MT_DEVICE, |
152 | }, { | 150 | }, { |
153 | .virtual = (unsigned long)S3C_VA_UART, | ||
154 | .pfn = __phys_to_pfn(EXYNOS4_PA_UART), | ||
155 | .length = SZ_512K, | ||
156 | .type = MT_DEVICE, | ||
157 | }, { | ||
158 | .virtual = (unsigned long)S5P_VA_CMU, | 151 | .virtual = (unsigned long)S5P_VA_CMU, |
159 | .pfn = __phys_to_pfn(EXYNOS4_PA_CMU), | 152 | .pfn = __phys_to_pfn(EXYNOS4_PA_CMU), |
160 | .length = SZ_128K, | 153 | .length = SZ_128K, |
@@ -268,20 +261,6 @@ static struct map_desc exynos5_iodesc[] __initdata = { | |||
268 | .pfn = __phys_to_pfn(EXYNOS5_PA_PMU), | 261 | .pfn = __phys_to_pfn(EXYNOS5_PA_PMU), |
269 | .length = SZ_64K, | 262 | .length = SZ_64K, |
270 | .type = MT_DEVICE, | 263 | .type = MT_DEVICE, |
271 | }, { | ||
272 | .virtual = (unsigned long)S3C_VA_UART, | ||
273 | .pfn = __phys_to_pfn(EXYNOS5_PA_UART), | ||
274 | .length = SZ_512K, | ||
275 | .type = MT_DEVICE, | ||
276 | }, | ||
277 | }; | ||
278 | |||
279 | static struct map_desc exynos5440_iodesc0[] __initdata = { | ||
280 | { | ||
281 | .virtual = (unsigned long)S3C_VA_UART, | ||
282 | .pfn = __phys_to_pfn(EXYNOS5440_PA_UART0), | ||
283 | .length = SZ_512K, | ||
284 | .type = MT_DEVICE, | ||
285 | }, | 264 | }, |
286 | }; | 265 | }; |
287 | 266 | ||
@@ -388,11 +367,6 @@ static void __init exynos5_map_io(void) | |||
388 | iotable_init(exynos5250_iodesc, ARRAY_SIZE(exynos5250_iodesc)); | 367 | iotable_init(exynos5250_iodesc, ARRAY_SIZE(exynos5250_iodesc)); |
389 | } | 368 | } |
390 | 369 | ||
391 | static void __init exynos5440_map_io(void) | ||
392 | { | ||
393 | iotable_init(exynos5440_iodesc0, ARRAY_SIZE(exynos5440_iodesc0)); | ||
394 | } | ||
395 | |||
396 | void __init exynos_init_time(void) | 370 | void __init exynos_init_time(void) |
397 | { | 371 | { |
398 | of_clk_init(NULL); | 372 | of_clk_init(NULL); |
diff --git a/arch/arm/mach-exynos/common.h b/arch/arm/mach-exynos/common.h index 3e156bcddcb4..972490fc09d6 100644 --- a/arch/arm/mach-exynos/common.h +++ b/arch/arm/mach-exynos/common.h | |||
@@ -97,6 +97,5 @@ struct exynos_pmu_conf { | |||
97 | }; | 97 | }; |
98 | 98 | ||
99 | extern void exynos_sys_powerdown_conf(enum sys_powerdown mode); | 99 | extern void exynos_sys_powerdown_conf(enum sys_powerdown mode); |
100 | extern void s3c_cpu_resume(void); | ||
101 | 100 | ||
102 | #endif /* __ARCH_ARM_MACH_EXYNOS_COMMON_H */ | 101 | #endif /* __ARCH_ARM_MACH_EXYNOS_COMMON_H */ |
diff --git a/arch/arm/mach-exynos/cpuidle.c b/arch/arm/mach-exynos/cpuidle.c index 17a18ff3d71e..225ee8431c72 100644 --- a/arch/arm/mach-exynos/cpuidle.c +++ b/arch/arm/mach-exynos/cpuidle.c | |||
@@ -25,6 +25,7 @@ | |||
25 | #include <mach/regs-pmu.h> | 25 | #include <mach/regs-pmu.h> |
26 | 26 | ||
27 | #include <plat/cpu.h> | 27 | #include <plat/cpu.h> |
28 | #include <plat/pm.h> | ||
28 | 29 | ||
29 | #include "common.h" | 30 | #include "common.h" |
30 | 31 | ||
diff --git a/arch/arm/mach-exynos/include/mach/memory.h b/arch/arm/mach-exynos/include/mach/memory.h index 374ef2cf7152..2a4cdb7cb326 100644 --- a/arch/arm/mach-exynos/include/mach/memory.h +++ b/arch/arm/mach-exynos/include/mach/memory.h | |||
@@ -15,8 +15,13 @@ | |||
15 | 15 | ||
16 | #define PLAT_PHYS_OFFSET UL(0x40000000) | 16 | #define PLAT_PHYS_OFFSET UL(0x40000000) |
17 | 17 | ||
18 | #ifndef CONFIG_ARM_LPAE | ||
18 | /* Maximum of 256MiB in one bank */ | 19 | /* Maximum of 256MiB in one bank */ |
19 | #define MAX_PHYSMEM_BITS 32 | 20 | #define MAX_PHYSMEM_BITS 32 |
20 | #define SECTION_SIZE_BITS 28 | 21 | #define SECTION_SIZE_BITS 28 |
22 | #else | ||
23 | #define MAX_PHYSMEM_BITS 36 | ||
24 | #define SECTION_SIZE_BITS 31 | ||
25 | #endif | ||
21 | 26 | ||
22 | #endif /* __ASM_ARCH_MEMORY_H */ | 27 | #endif /* __ASM_ARCH_MEMORY_H */ |
diff --git a/arch/arm/mach-exynos/pm.c b/arch/arm/mach-exynos/pm.c index 41c20692a13f..c679db577269 100644 --- a/arch/arm/mach-exynos/pm.c +++ b/arch/arm/mach-exynos/pm.c | |||
@@ -217,6 +217,9 @@ static __init int exynos_pm_drvinit(void) | |||
217 | struct clk *pll_base; | 217 | struct clk *pll_base; |
218 | unsigned int tmp; | 218 | unsigned int tmp; |
219 | 219 | ||
220 | if (soc_is_exynos5440()) | ||
221 | return 0; | ||
222 | |||
220 | s3c_pm_init(); | 223 | s3c_pm_init(); |
221 | 224 | ||
222 | /* All wakeup disable */ | 225 | /* All wakeup disable */ |
@@ -340,6 +343,9 @@ static struct syscore_ops exynos_pm_syscore_ops = { | |||
340 | 343 | ||
341 | static __init int exynos_pm_syscore_init(void) | 344 | static __init int exynos_pm_syscore_init(void) |
342 | { | 345 | { |
346 | if (soc_is_exynos5440()) | ||
347 | return 0; | ||
348 | |||
343 | register_syscore_ops(&exynos_pm_syscore_ops); | 349 | register_syscore_ops(&exynos_pm_syscore_ops); |
344 | return 0; | 350 | return 0; |
345 | } | 351 | } |
diff --git a/arch/arm/mach-footbridge/dc21285.c b/arch/arm/mach-footbridge/dc21285.c index a7cd2cf5e08d..3490a24f969e 100644 --- a/arch/arm/mach-footbridge/dc21285.c +++ b/arch/arm/mach-footbridge/dc21285.c | |||
@@ -276,8 +276,6 @@ int __init dc21285_setup(int nr, struct pci_sys_data *sys) | |||
276 | 276 | ||
277 | sys->mem_offset = DC21285_PCI_MEM; | 277 | sys->mem_offset = DC21285_PCI_MEM; |
278 | 278 | ||
279 | pci_ioremap_io(0, DC21285_PCI_IO); | ||
280 | |||
281 | pci_add_resource_offset(&sys->resources, &res[0], sys->mem_offset); | 279 | pci_add_resource_offset(&sys->resources, &res[0], sys->mem_offset); |
282 | pci_add_resource_offset(&sys->resources, &res[1], sys->mem_offset); | 280 | pci_add_resource_offset(&sys->resources, &res[1], sys->mem_offset); |
283 | 281 | ||
diff --git a/arch/arm/mach-highbank/highbank.c b/arch/arm/mach-highbank/highbank.c index dc5d6becd8c7..88815795fe26 100644 --- a/arch/arm/mach-highbank/highbank.c +++ b/arch/arm/mach-highbank/highbank.c | |||
@@ -115,6 +115,7 @@ static int highbank_platform_notifier(struct notifier_block *nb, | |||
115 | { | 115 | { |
116 | struct resource *res; | 116 | struct resource *res; |
117 | int reg = -1; | 117 | int reg = -1; |
118 | u32 val; | ||
118 | struct device *dev = __dev; | 119 | struct device *dev = __dev; |
119 | 120 | ||
120 | if (event != BUS_NOTIFY_ADD_DEVICE) | 121 | if (event != BUS_NOTIFY_ADD_DEVICE) |
@@ -141,10 +142,10 @@ static int highbank_platform_notifier(struct notifier_block *nb, | |||
141 | return NOTIFY_DONE; | 142 | return NOTIFY_DONE; |
142 | 143 | ||
143 | if (of_property_read_bool(dev->of_node, "dma-coherent")) { | 144 | if (of_property_read_bool(dev->of_node, "dma-coherent")) { |
144 | writel(0xff31, sregs_base + reg); | 145 | val = readl(sregs_base + reg); |
146 | writel(val | 0xff01, sregs_base + reg); | ||
145 | set_dma_ops(dev, &arm_coherent_dma_ops); | 147 | set_dma_ops(dev, &arm_coherent_dma_ops); |
146 | } else | 148 | } |
147 | writel(0, sregs_base + reg); | ||
148 | 149 | ||
149 | return NOTIFY_OK; | 150 | return NOTIFY_OK; |
150 | } | 151 | } |
diff --git a/arch/arm/mach-imx/clk-imx6q.c b/arch/arm/mach-imx/clk-imx6q.c index 4282e99f5ca1..86567d980b07 100644 --- a/arch/arm/mach-imx/clk-imx6q.c +++ b/arch/arm/mach-imx/clk-imx6q.c | |||
@@ -199,7 +199,8 @@ static const char *pcie_axi_sels[] = { "axi", "ahb", }; | |||
199 | static const char *ssi_sels[] = { "pll3_pfd2_508m", "pll3_pfd3_454m", "pll4_post_div", }; | 199 | static const char *ssi_sels[] = { "pll3_pfd2_508m", "pll3_pfd3_454m", "pll4_post_div", }; |
200 | static const char *usdhc_sels[] = { "pll2_pfd2_396m", "pll2_pfd0_352m", }; | 200 | static const char *usdhc_sels[] = { "pll2_pfd2_396m", "pll2_pfd0_352m", }; |
201 | static const char *enfc_sels[] = { "pll2_pfd0_352m", "pll2_bus", "pll3_usb_otg", "pll2_pfd2_396m", }; | 201 | static const char *enfc_sels[] = { "pll2_pfd0_352m", "pll2_bus", "pll3_usb_otg", "pll2_pfd2_396m", }; |
202 | static const char *emi_sels[] = { "axi", "pll3_usb_otg", "pll2_pfd2_396m", "pll2_pfd0_352m", }; | 202 | static const char *emi_sels[] = { "pll2_pfd2_396m", "pll3_usb_otg", "axi", "pll2_pfd0_352m", }; |
203 | static const char *emi_slow_sels[] = { "axi", "pll3_usb_otg", "pll2_pfd2_396m", "pll2_pfd0_352m", }; | ||
203 | static const char *vdo_axi_sels[] = { "axi", "ahb", }; | 204 | static const char *vdo_axi_sels[] = { "axi", "ahb", }; |
204 | static const char *vpu_axi_sels[] = { "axi", "pll2_pfd2_396m", "pll2_pfd0_352m", }; | 205 | static const char *vpu_axi_sels[] = { "axi", "pll2_pfd2_396m", "pll2_pfd0_352m", }; |
205 | static const char *cko1_sels[] = { "pll3_usb_otg", "pll2_bus", "pll1_sys", "pll5_video_div", | 206 | static const char *cko1_sels[] = { "pll3_usb_otg", "pll2_bus", "pll1_sys", "pll5_video_div", |
@@ -392,7 +393,7 @@ static void __init imx6q_clocks_init(struct device_node *ccm_node) | |||
392 | clk[usdhc4_sel] = imx_clk_mux("usdhc4_sel", base + 0x1c, 19, 1, usdhc_sels, ARRAY_SIZE(usdhc_sels)); | 393 | clk[usdhc4_sel] = imx_clk_mux("usdhc4_sel", base + 0x1c, 19, 1, usdhc_sels, ARRAY_SIZE(usdhc_sels)); |
393 | clk[enfc_sel] = imx_clk_mux("enfc_sel", base + 0x2c, 16, 2, enfc_sels, ARRAY_SIZE(enfc_sels)); | 394 | clk[enfc_sel] = imx_clk_mux("enfc_sel", base + 0x2c, 16, 2, enfc_sels, ARRAY_SIZE(enfc_sels)); |
394 | clk[emi_sel] = imx_clk_mux("emi_sel", base + 0x1c, 27, 2, emi_sels, ARRAY_SIZE(emi_sels)); | 395 | clk[emi_sel] = imx_clk_mux("emi_sel", base + 0x1c, 27, 2, emi_sels, ARRAY_SIZE(emi_sels)); |
395 | clk[emi_slow_sel] = imx_clk_mux("emi_slow_sel", base + 0x1c, 29, 2, emi_sels, ARRAY_SIZE(emi_sels)); | 396 | clk[emi_slow_sel] = imx_clk_mux("emi_slow_sel", base + 0x1c, 29, 2, emi_slow_sels, ARRAY_SIZE(emi_slow_sels)); |
396 | clk[vdo_axi_sel] = imx_clk_mux("vdo_axi_sel", base + 0x18, 11, 1, vdo_axi_sels, ARRAY_SIZE(vdo_axi_sels)); | 397 | clk[vdo_axi_sel] = imx_clk_mux("vdo_axi_sel", base + 0x18, 11, 1, vdo_axi_sels, ARRAY_SIZE(vdo_axi_sels)); |
397 | clk[vpu_axi_sel] = imx_clk_mux("vpu_axi_sel", base + 0x18, 14, 2, vpu_axi_sels, ARRAY_SIZE(vpu_axi_sels)); | 398 | clk[vpu_axi_sel] = imx_clk_mux("vpu_axi_sel", base + 0x18, 14, 2, vpu_axi_sels, ARRAY_SIZE(vpu_axi_sels)); |
398 | clk[cko1_sel] = imx_clk_mux("cko1_sel", base + 0x60, 0, 4, cko1_sels, ARRAY_SIZE(cko1_sels)); | 399 | clk[cko1_sel] = imx_clk_mux("cko1_sel", base + 0x60, 0, 4, cko1_sels, ARRAY_SIZE(cko1_sels)); |
diff --git a/arch/arm/mach-imx/clk-vf610.c b/arch/arm/mach-imx/clk-vf610.c index d617c0b7c809..b169a396d93b 100644 --- a/arch/arm/mach-imx/clk-vf610.c +++ b/arch/arm/mach-imx/clk-vf610.c | |||
@@ -183,6 +183,8 @@ static void __init vf610_clocks_init(struct device_node *ccm_node) | |||
183 | clk[VF610_CLK_ENET_TS_SEL] = imx_clk_mux("enet_ts_sel", CCM_CSCMR2, 0, 3, enet_ts_sels, 7); | 183 | clk[VF610_CLK_ENET_TS_SEL] = imx_clk_mux("enet_ts_sel", CCM_CSCMR2, 0, 3, enet_ts_sels, 7); |
184 | clk[VF610_CLK_ENET] = imx_clk_gate("enet", "enet_sel", CCM_CSCDR1, 24); | 184 | clk[VF610_CLK_ENET] = imx_clk_gate("enet", "enet_sel", CCM_CSCDR1, 24); |
185 | clk[VF610_CLK_ENET_TS] = imx_clk_gate("enet_ts", "enet_ts_sel", CCM_CSCDR1, 23); | 185 | clk[VF610_CLK_ENET_TS] = imx_clk_gate("enet_ts", "enet_ts_sel", CCM_CSCDR1, 23); |
186 | clk[VF610_CLK_ENET0] = imx_clk_gate2("enet0", "ipg_bus", CCM_CCGR9, CCM_CCGRx_CGn(0)); | ||
187 | clk[VF610_CLK_ENET1] = imx_clk_gate2("enet1", "ipg_bus", CCM_CCGR9, CCM_CCGRx_CGn(1)); | ||
186 | 188 | ||
187 | clk[VF610_CLK_PIT] = imx_clk_gate2("pit", "ipg_bus", CCM_CCGR1, CCM_CCGRx_CGn(7)); | 189 | clk[VF610_CLK_PIT] = imx_clk_gate2("pit", "ipg_bus", CCM_CCGR1, CCM_CCGRx_CGn(7)); |
188 | 190 | ||
diff --git a/arch/arm/mach-imx/mx27.h b/arch/arm/mach-imx/mx27.h index e074616d54ca..8a65f192e7f3 100644 --- a/arch/arm/mach-imx/mx27.h +++ b/arch/arm/mach-imx/mx27.h | |||
@@ -135,7 +135,7 @@ | |||
135 | #define MX27_INT_GPT4 (NR_IRQS_LEGACY + 4) | 135 | #define MX27_INT_GPT4 (NR_IRQS_LEGACY + 4) |
136 | #define MX27_INT_RTIC (NR_IRQS_LEGACY + 5) | 136 | #define MX27_INT_RTIC (NR_IRQS_LEGACY + 5) |
137 | #define MX27_INT_CSPI3 (NR_IRQS_LEGACY + 6) | 137 | #define MX27_INT_CSPI3 (NR_IRQS_LEGACY + 6) |
138 | #define MX27_INT_SDHC (NR_IRQS_LEGACY + 7) | 138 | #define MX27_INT_MSHC (NR_IRQS_LEGACY + 7) |
139 | #define MX27_INT_GPIO (NR_IRQS_LEGACY + 8) | 139 | #define MX27_INT_GPIO (NR_IRQS_LEGACY + 8) |
140 | #define MX27_INT_SDHC3 (NR_IRQS_LEGACY + 9) | 140 | #define MX27_INT_SDHC3 (NR_IRQS_LEGACY + 9) |
141 | #define MX27_INT_SDHC2 (NR_IRQS_LEGACY + 10) | 141 | #define MX27_INT_SDHC2 (NR_IRQS_LEGACY + 10) |
diff --git a/arch/arm/mach-keystone/keystone.c b/arch/arm/mach-keystone/keystone.c index fe4d9ff93a7e..b661c5c2870a 100644 --- a/arch/arm/mach-keystone/keystone.c +++ b/arch/arm/mach-keystone/keystone.c | |||
@@ -49,7 +49,7 @@ static const char *keystone_match[] __initconst = { | |||
49 | NULL, | 49 | NULL, |
50 | }; | 50 | }; |
51 | 51 | ||
52 | void keystone_restart(char mode, const char *cmd) | 52 | void keystone_restart(enum reboot_mode mode, const char *cmd) |
53 | { | 53 | { |
54 | u32 val; | 54 | u32 val; |
55 | 55 | ||
diff --git a/arch/arm/mach-omap2/Kconfig b/arch/arm/mach-omap2/Kconfig index 627fa7e41fba..3eed0006d189 100644 --- a/arch/arm/mach-omap2/Kconfig +++ b/arch/arm/mach-omap2/Kconfig | |||
@@ -62,7 +62,7 @@ config SOC_OMAP5 | |||
62 | select HAVE_SMP | 62 | select HAVE_SMP |
63 | select COMMON_CLK | 63 | select COMMON_CLK |
64 | select HAVE_ARM_ARCH_TIMER | 64 | select HAVE_ARM_ARCH_TIMER |
65 | select ARM_ERRATA_798181 | 65 | select ARM_ERRATA_798181 if SMP |
66 | 66 | ||
67 | config SOC_AM33XX | 67 | config SOC_AM33XX |
68 | bool "AM33XX support" | 68 | bool "AM33XX support" |
diff --git a/arch/arm/mach-omap2/board-generic.c b/arch/arm/mach-omap2/board-generic.c index e5fbfed69aa2..be5d005ebad2 100644 --- a/arch/arm/mach-omap2/board-generic.c +++ b/arch/arm/mach-omap2/board-generic.c | |||
@@ -15,6 +15,7 @@ | |||
15 | #include <linux/of_irq.h> | 15 | #include <linux/of_irq.h> |
16 | #include <linux/of_platform.h> | 16 | #include <linux/of_platform.h> |
17 | #include <linux/irqdomain.h> | 17 | #include <linux/irqdomain.h> |
18 | #include <linux/clk.h> | ||
18 | 19 | ||
19 | #include <asm/mach/arch.h> | 20 | #include <asm/mach/arch.h> |
20 | 21 | ||
@@ -35,6 +36,21 @@ static struct of_device_id omap_dt_match_table[] __initdata = { | |||
35 | { } | 36 | { } |
36 | }; | 37 | }; |
37 | 38 | ||
39 | /* | ||
40 | * Create alias for USB host PHY clock. | ||
41 | * Remove this when clock phandle can be provided via DT | ||
42 | */ | ||
43 | static void __init legacy_init_ehci_clk(char *clkname) | ||
44 | { | ||
45 | int ret; | ||
46 | |||
47 | ret = clk_add_alias("main_clk", NULL, clkname, NULL); | ||
48 | if (ret) { | ||
49 | pr_err("%s:Failed to add main_clk alias to %s :%d\n", | ||
50 | __func__, clkname, ret); | ||
51 | } | ||
52 | } | ||
53 | |||
38 | static void __init omap_generic_init(void) | 54 | static void __init omap_generic_init(void) |
39 | { | 55 | { |
40 | omap_sdrc_init(NULL, NULL); | 56 | omap_sdrc_init(NULL, NULL); |
@@ -45,10 +61,15 @@ static void __init omap_generic_init(void) | |||
45 | * HACK: call display setup code for selected boards to enable omapdss. | 61 | * HACK: call display setup code for selected boards to enable omapdss. |
46 | * This will be removed when omapdss supports DT. | 62 | * This will be removed when omapdss supports DT. |
47 | */ | 63 | */ |
48 | if (of_machine_is_compatible("ti,omap4-panda")) | 64 | if (of_machine_is_compatible("ti,omap4-panda")) { |
49 | omap4_panda_display_init_of(); | 65 | omap4_panda_display_init_of(); |
66 | legacy_init_ehci_clk("auxclk3_ck"); | ||
67 | |||
68 | } | ||
50 | else if (of_machine_is_compatible("ti,omap4-sdp")) | 69 | else if (of_machine_is_compatible("ti,omap4-sdp")) |
51 | omap_4430sdp_display_init_of(); | 70 | omap_4430sdp_display_init_of(); |
71 | else if (of_machine_is_compatible("ti,omap5-uevm")) | ||
72 | legacy_init_ehci_clk("auxclk1_ck"); | ||
52 | } | 73 | } |
53 | 74 | ||
54 | #ifdef CONFIG_SOC_OMAP2420 | 75 | #ifdef CONFIG_SOC_OMAP2420 |
diff --git a/arch/arm/mach-pxa/em-x270.c b/arch/arm/mach-pxa/em-x270.c index f6726bb4eb95..3a3362fa793e 100644 --- a/arch/arm/mach-pxa/em-x270.c +++ b/arch/arm/mach-pxa/em-x270.c | |||
@@ -477,16 +477,24 @@ static int em_x270_usb_hub_init(void) | |||
477 | /* USB Hub power-on and reset */ | 477 | /* USB Hub power-on and reset */ |
478 | gpio_direction_output(usb_hub_reset, 1); | 478 | gpio_direction_output(usb_hub_reset, 1); |
479 | gpio_direction_output(GPIO9_USB_VBUS_EN, 0); | 479 | gpio_direction_output(GPIO9_USB_VBUS_EN, 0); |
480 | regulator_enable(em_x270_usb_ldo); | 480 | err = regulator_enable(em_x270_usb_ldo); |
481 | if (err) | ||
482 | goto err_free_rst_gpio; | ||
483 | |||
481 | gpio_set_value(usb_hub_reset, 0); | 484 | gpio_set_value(usb_hub_reset, 0); |
482 | gpio_set_value(usb_hub_reset, 1); | 485 | gpio_set_value(usb_hub_reset, 1); |
483 | regulator_disable(em_x270_usb_ldo); | 486 | regulator_disable(em_x270_usb_ldo); |
484 | regulator_enable(em_x270_usb_ldo); | 487 | err = regulator_enable(em_x270_usb_ldo); |
488 | if (err) | ||
489 | goto err_free_rst_gpio; | ||
490 | |||
485 | gpio_set_value(usb_hub_reset, 0); | 491 | gpio_set_value(usb_hub_reset, 0); |
486 | gpio_set_value(GPIO9_USB_VBUS_EN, 1); | 492 | gpio_set_value(GPIO9_USB_VBUS_EN, 1); |
487 | 493 | ||
488 | return 0; | 494 | return 0; |
489 | 495 | ||
496 | err_free_rst_gpio: | ||
497 | gpio_free(usb_hub_reset); | ||
490 | err_free_vbus_gpio: | 498 | err_free_vbus_gpio: |
491 | gpio_free(GPIO9_USB_VBUS_EN); | 499 | gpio_free(GPIO9_USB_VBUS_EN); |
492 | err_free_usb_ldo: | 500 | err_free_usb_ldo: |
@@ -592,7 +600,7 @@ err_irq: | |||
592 | return err; | 600 | return err; |
593 | } | 601 | } |
594 | 602 | ||
595 | static void em_x270_mci_setpower(struct device *dev, unsigned int vdd) | 603 | static int em_x270_mci_setpower(struct device *dev, unsigned int vdd) |
596 | { | 604 | { |
597 | struct pxamci_platform_data* p_d = dev->platform_data; | 605 | struct pxamci_platform_data* p_d = dev->platform_data; |
598 | 606 | ||
@@ -600,10 +608,11 @@ static void em_x270_mci_setpower(struct device *dev, unsigned int vdd) | |||
600 | int vdd_uV = (2000 + (vdd - __ffs(MMC_VDD_20_21)) * 100) * 1000; | 608 | int vdd_uV = (2000 + (vdd - __ffs(MMC_VDD_20_21)) * 100) * 1000; |
601 | 609 | ||
602 | regulator_set_voltage(em_x270_sdio_ldo, vdd_uV, vdd_uV); | 610 | regulator_set_voltage(em_x270_sdio_ldo, vdd_uV, vdd_uV); |
603 | regulator_enable(em_x270_sdio_ldo); | 611 | return regulator_enable(em_x270_sdio_ldo); |
604 | } else { | 612 | } else { |
605 | regulator_disable(em_x270_sdio_ldo); | 613 | regulator_disable(em_x270_sdio_ldo); |
606 | } | 614 | } |
615 | return 0; | ||
607 | } | 616 | } |
608 | 617 | ||
609 | static void em_x270_mci_exit(struct device *dev, void *data) | 618 | static void em_x270_mci_exit(struct device *dev, void *data) |
diff --git a/arch/arm/mach-pxa/mainstone.c b/arch/arm/mach-pxa/mainstone.c index d2c652318376..dd70343c8708 100644 --- a/arch/arm/mach-pxa/mainstone.c +++ b/arch/arm/mach-pxa/mainstone.c | |||
@@ -408,7 +408,7 @@ static int mainstone_mci_init(struct device *dev, irq_handler_t mstone_detect_in | |||
408 | return err; | 408 | return err; |
409 | } | 409 | } |
410 | 410 | ||
411 | static void mainstone_mci_setpower(struct device *dev, unsigned int vdd) | 411 | static int mainstone_mci_setpower(struct device *dev, unsigned int vdd) |
412 | { | 412 | { |
413 | struct pxamci_platform_data* p_d = dev->platform_data; | 413 | struct pxamci_platform_data* p_d = dev->platform_data; |
414 | 414 | ||
@@ -420,6 +420,7 @@ static void mainstone_mci_setpower(struct device *dev, unsigned int vdd) | |||
420 | printk(KERN_DEBUG "%s: off\n", __func__); | 420 | printk(KERN_DEBUG "%s: off\n", __func__); |
421 | MST_MSCWR1 &= ~MST_MSCWR1_MMC_ON; | 421 | MST_MSCWR1 &= ~MST_MSCWR1_MMC_ON; |
422 | } | 422 | } |
423 | return 0; | ||
423 | } | 424 | } |
424 | 425 | ||
425 | static void mainstone_mci_exit(struct device *dev, void *data) | 426 | static void mainstone_mci_exit(struct device *dev, void *data) |
diff --git a/arch/arm/mach-pxa/pcm990-baseboard.c b/arch/arm/mach-pxa/pcm990-baseboard.c index fb7f1d1627dc..13e5b00eae90 100644 --- a/arch/arm/mach-pxa/pcm990-baseboard.c +++ b/arch/arm/mach-pxa/pcm990-baseboard.c | |||
@@ -335,7 +335,7 @@ static int pcm990_mci_init(struct device *dev, irq_handler_t mci_detect_int, | |||
335 | return err; | 335 | return err; |
336 | } | 336 | } |
337 | 337 | ||
338 | static void pcm990_mci_setpower(struct device *dev, unsigned int vdd) | 338 | static int pcm990_mci_setpower(struct device *dev, unsigned int vdd) |
339 | { | 339 | { |
340 | struct pxamci_platform_data *p_d = dev->platform_data; | 340 | struct pxamci_platform_data *p_d = dev->platform_data; |
341 | u8 val; | 341 | u8 val; |
@@ -348,6 +348,7 @@ static void pcm990_mci_setpower(struct device *dev, unsigned int vdd) | |||
348 | val &= ~PCM990_CTRL_MMC2PWR; | 348 | val &= ~PCM990_CTRL_MMC2PWR; |
349 | 349 | ||
350 | pcm990_cpld_writeb(PCM990_CTRL_MMC2PWR, PCM990_CTRL_REG5); | 350 | pcm990_cpld_writeb(PCM990_CTRL_MMC2PWR, PCM990_CTRL_REG5); |
351 | return 0; | ||
351 | } | 352 | } |
352 | 353 | ||
353 | static void pcm990_mci_exit(struct device *dev, void *data) | 354 | static void pcm990_mci_exit(struct device *dev, void *data) |
diff --git a/arch/arm/mach-pxa/poodle.c b/arch/arm/mach-pxa/poodle.c index 711d37e26bd8..aedf053a1de5 100644 --- a/arch/arm/mach-pxa/poodle.c +++ b/arch/arm/mach-pxa/poodle.c | |||
@@ -258,7 +258,7 @@ err_free_2: | |||
258 | return err; | 258 | return err; |
259 | } | 259 | } |
260 | 260 | ||
261 | static void poodle_mci_setpower(struct device *dev, unsigned int vdd) | 261 | static int poodle_mci_setpower(struct device *dev, unsigned int vdd) |
262 | { | 262 | { |
263 | struct pxamci_platform_data* p_d = dev->platform_data; | 263 | struct pxamci_platform_data* p_d = dev->platform_data; |
264 | 264 | ||
@@ -270,6 +270,8 @@ static void poodle_mci_setpower(struct device *dev, unsigned int vdd) | |||
270 | gpio_set_value(POODLE_GPIO_SD_PWR1, 0); | 270 | gpio_set_value(POODLE_GPIO_SD_PWR1, 0); |
271 | gpio_set_value(POODLE_GPIO_SD_PWR, 0); | 271 | gpio_set_value(POODLE_GPIO_SD_PWR, 0); |
272 | } | 272 | } |
273 | |||
274 | return 0; | ||
273 | } | 275 | } |
274 | 276 | ||
275 | static void poodle_mci_exit(struct device *dev, void *data) | 277 | static void poodle_mci_exit(struct device *dev, void *data) |
diff --git a/arch/arm/mach-pxa/spitz.c b/arch/arm/mach-pxa/spitz.c index 2125df0444e7..4c29173026e8 100644 --- a/arch/arm/mach-pxa/spitz.c +++ b/arch/arm/mach-pxa/spitz.c | |||
@@ -598,7 +598,7 @@ static inline void spitz_spi_init(void) {} | |||
598 | * NOTE: The card detect interrupt isn't debounced so we delay it by 250ms to | 598 | * NOTE: The card detect interrupt isn't debounced so we delay it by 250ms to |
599 | * give the card a chance to fully insert/eject. | 599 | * give the card a chance to fully insert/eject. |
600 | */ | 600 | */ |
601 | static void spitz_mci_setpower(struct device *dev, unsigned int vdd) | 601 | static int spitz_mci_setpower(struct device *dev, unsigned int vdd) |
602 | { | 602 | { |
603 | struct pxamci_platform_data* p_d = dev->platform_data; | 603 | struct pxamci_platform_data* p_d = dev->platform_data; |
604 | 604 | ||
@@ -606,6 +606,8 @@ static void spitz_mci_setpower(struct device *dev, unsigned int vdd) | |||
606 | spitz_card_pwr_ctrl(SCOOP_CPR_SD_3V, SCOOP_CPR_SD_3V); | 606 | spitz_card_pwr_ctrl(SCOOP_CPR_SD_3V, SCOOP_CPR_SD_3V); |
607 | else | 607 | else |
608 | spitz_card_pwr_ctrl(SCOOP_CPR_SD_3V, 0x0); | 608 | spitz_card_pwr_ctrl(SCOOP_CPR_SD_3V, 0x0); |
609 | |||
610 | return 0; | ||
609 | } | 611 | } |
610 | 612 | ||
611 | static struct pxamci_platform_data spitz_mci_platform_data = { | 613 | static struct pxamci_platform_data spitz_mci_platform_data = { |
diff --git a/arch/arm/mach-pxa/stargate2.c b/arch/arm/mach-pxa/stargate2.c index 88fde43c948c..62aea3e835f3 100644 --- a/arch/arm/mach-pxa/stargate2.c +++ b/arch/arm/mach-pxa/stargate2.c | |||
@@ -734,9 +734,10 @@ static int stargate2_mci_init(struct device *dev, | |||
734 | * | 734 | * |
735 | * Very simple control. Either it is on or off and is controlled by | 735 | * Very simple control. Either it is on or off and is controlled by |
736 | * a gpio pin */ | 736 | * a gpio pin */ |
737 | static void stargate2_mci_setpower(struct device *dev, unsigned int vdd) | 737 | static int stargate2_mci_setpower(struct device *dev, unsigned int vdd) |
738 | { | 738 | { |
739 | gpio_set_value(SG2_SD_POWER_ENABLE, !!vdd); | 739 | gpio_set_value(SG2_SD_POWER_ENABLE, !!vdd); |
740 | return 0; | ||
740 | } | 741 | } |
741 | 742 | ||
742 | static void stargate2_mci_exit(struct device *dev, void *data) | 743 | static void stargate2_mci_exit(struct device *dev, void *data) |
diff --git a/arch/arm/mach-s3c24xx/clock-s3c2410.c b/arch/arm/mach-s3c24xx/clock-s3c2410.c index 34fffdf6fc1d..564553694b54 100644 --- a/arch/arm/mach-s3c24xx/clock-s3c2410.c +++ b/arch/arm/mach-s3c24xx/clock-s3c2410.c | |||
@@ -119,66 +119,101 @@ static struct clk init_clocks_off[] = { | |||
119 | } | 119 | } |
120 | }; | 120 | }; |
121 | 121 | ||
122 | static struct clk init_clocks[] = { | 122 | static struct clk clk_lcd = { |
123 | { | 123 | .name = "lcd", |
124 | .name = "lcd", | 124 | .parent = &clk_h, |
125 | .parent = &clk_h, | 125 | .enable = s3c2410_clkcon_enable, |
126 | .enable = s3c2410_clkcon_enable, | 126 | .ctrlbit = S3C2410_CLKCON_LCDC, |
127 | .ctrlbit = S3C2410_CLKCON_LCDC, | 127 | }; |
128 | }, { | 128 | |
129 | .name = "gpio", | 129 | static struct clk clk_gpio = { |
130 | .parent = &clk_p, | 130 | .name = "gpio", |
131 | .enable = s3c2410_clkcon_enable, | 131 | .parent = &clk_p, |
132 | .ctrlbit = S3C2410_CLKCON_GPIO, | 132 | .enable = s3c2410_clkcon_enable, |
133 | }, { | 133 | .ctrlbit = S3C2410_CLKCON_GPIO, |
134 | .name = "usb-host", | 134 | }; |
135 | .parent = &clk_h, | 135 | |
136 | .enable = s3c2410_clkcon_enable, | 136 | static struct clk clk_usb_host = { |
137 | .ctrlbit = S3C2410_CLKCON_USBH, | 137 | .name = "usb-host", |
138 | }, { | 138 | .parent = &clk_h, |
139 | .name = "usb-device", | 139 | .enable = s3c2410_clkcon_enable, |
140 | .parent = &clk_h, | 140 | .ctrlbit = S3C2410_CLKCON_USBH, |
141 | .enable = s3c2410_clkcon_enable, | 141 | }; |
142 | .ctrlbit = S3C2410_CLKCON_USBD, | 142 | |
143 | }, { | 143 | static struct clk clk_usb_device = { |
144 | .name = "timers", | 144 | .name = "usb-device", |
145 | .parent = &clk_p, | 145 | .parent = &clk_h, |
146 | .enable = s3c2410_clkcon_enable, | 146 | .enable = s3c2410_clkcon_enable, |
147 | .ctrlbit = S3C2410_CLKCON_PWMT, | 147 | .ctrlbit = S3C2410_CLKCON_USBD, |
148 | }, { | 148 | }; |
149 | .name = "uart", | 149 | |
150 | .devname = "s3c2410-uart.0", | 150 | static struct clk clk_timers = { |
151 | .parent = &clk_p, | 151 | .name = "timers", |
152 | .enable = s3c2410_clkcon_enable, | 152 | .parent = &clk_p, |
153 | .ctrlbit = S3C2410_CLKCON_UART0, | 153 | .enable = s3c2410_clkcon_enable, |
154 | }, { | 154 | .ctrlbit = S3C2410_CLKCON_PWMT, |
155 | .name = "uart", | 155 | }; |
156 | .devname = "s3c2410-uart.1", | 156 | |
157 | .parent = &clk_p, | 157 | struct clk s3c24xx_clk_uart0 = { |
158 | .enable = s3c2410_clkcon_enable, | 158 | .name = "uart", |
159 | .ctrlbit = S3C2410_CLKCON_UART1, | 159 | .devname = "s3c2410-uart.0", |
160 | }, { | 160 | .parent = &clk_p, |
161 | .name = "uart", | 161 | .enable = s3c2410_clkcon_enable, |
162 | .devname = "s3c2410-uart.2", | 162 | .ctrlbit = S3C2410_CLKCON_UART0, |
163 | .parent = &clk_p, | 163 | }; |
164 | .enable = s3c2410_clkcon_enable, | 164 | |
165 | .ctrlbit = S3C2410_CLKCON_UART2, | 165 | struct clk s3c24xx_clk_uart1 = { |
166 | }, { | 166 | .name = "uart", |
167 | .name = "rtc", | 167 | .devname = "s3c2410-uart.1", |
168 | .parent = &clk_p, | 168 | .parent = &clk_p, |
169 | .enable = s3c2410_clkcon_enable, | 169 | .enable = s3c2410_clkcon_enable, |
170 | .ctrlbit = S3C2410_CLKCON_RTC, | 170 | .ctrlbit = S3C2410_CLKCON_UART1, |
171 | }, { | 171 | }; |
172 | .name = "watchdog", | 172 | |
173 | .parent = &clk_p, | 173 | struct clk s3c24xx_clk_uart2 = { |
174 | .ctrlbit = 0, | 174 | .name = "uart", |
175 | }, { | 175 | .devname = "s3c2410-uart.2", |
176 | .name = "usb-bus-host", | 176 | .parent = &clk_p, |
177 | .parent = &clk_usb_bus, | 177 | .enable = s3c2410_clkcon_enable, |
178 | }, { | 178 | .ctrlbit = S3C2410_CLKCON_UART2, |
179 | .name = "usb-bus-gadget", | 179 | }; |
180 | .parent = &clk_usb_bus, | 180 | |
181 | }, | 181 | static struct clk clk_rtc = { |
182 | .name = "rtc", | ||
183 | .parent = &clk_p, | ||
184 | .enable = s3c2410_clkcon_enable, | ||
185 | .ctrlbit = S3C2410_CLKCON_RTC, | ||
186 | }; | ||
187 | |||
188 | static struct clk clk_watchdog = { | ||
189 | .name = "watchdog", | ||
190 | .parent = &clk_p, | ||
191 | .ctrlbit = 0, | ||
192 | }; | ||
193 | |||
194 | static struct clk clk_usb_bus_host = { | ||
195 | .name = "usb-bus-host", | ||
196 | .parent = &clk_usb_bus, | ||
197 | }; | ||
198 | |||
199 | static struct clk clk_usb_bus_gadget = { | ||
200 | .name = "usb-bus-gadget", | ||
201 | .parent = &clk_usb_bus, | ||
202 | }; | ||
203 | |||
204 | static struct clk *init_clocks[] = { | ||
205 | &clk_lcd, | ||
206 | &clk_gpio, | ||
207 | &clk_usb_host, | ||
208 | &clk_usb_device, | ||
209 | &clk_timers, | ||
210 | &s3c24xx_clk_uart0, | ||
211 | &s3c24xx_clk_uart1, | ||
212 | &s3c24xx_clk_uart2, | ||
213 | &clk_rtc, | ||
214 | &clk_watchdog, | ||
215 | &clk_usb_bus_host, | ||
216 | &clk_usb_bus_gadget, | ||
182 | }; | 217 | }; |
183 | 218 | ||
184 | /* s3c2410_baseclk_add() | 219 | /* s3c2410_baseclk_add() |
@@ -195,7 +230,6 @@ int __init s3c2410_baseclk_add(void) | |||
195 | { | 230 | { |
196 | unsigned long clkslow = __raw_readl(S3C2410_CLKSLOW); | 231 | unsigned long clkslow = __raw_readl(S3C2410_CLKSLOW); |
197 | unsigned long clkcon = __raw_readl(S3C2410_CLKCON); | 232 | unsigned long clkcon = __raw_readl(S3C2410_CLKCON); |
198 | struct clk *clkp; | ||
199 | struct clk *xtal; | 233 | struct clk *xtal; |
200 | int ret; | 234 | int ret; |
201 | int ptr; | 235 | int ptr; |
@@ -207,8 +241,9 @@ int __init s3c2410_baseclk_add(void) | |||
207 | 241 | ||
208 | /* register clocks from clock array */ | 242 | /* register clocks from clock array */ |
209 | 243 | ||
210 | clkp = init_clocks; | 244 | for (ptr = 0; ptr < ARRAY_SIZE(init_clocks); ptr++) { |
211 | for (ptr = 0; ptr < ARRAY_SIZE(init_clocks); ptr++, clkp++) { | 245 | struct clk *clkp = init_clocks[ptr]; |
246 | |||
212 | /* ensure that we note the clock state */ | 247 | /* ensure that we note the clock state */ |
213 | 248 | ||
214 | clkp->usage = clkcon & clkp->ctrlbit ? 1 : 0; | 249 | clkp->usage = clkcon & clkp->ctrlbit ? 1 : 0; |
diff --git a/arch/arm/mach-s3c24xx/clock-s3c2440.c b/arch/arm/mach-s3c24xx/clock-s3c2440.c index 1069b5680826..aaf006d1d6dc 100644 --- a/arch/arm/mach-s3c24xx/clock-s3c2440.c +++ b/arch/arm/mach-s3c24xx/clock-s3c2440.c | |||
@@ -166,6 +166,9 @@ static struct clk_lookup s3c2440_clk_lookup[] = { | |||
166 | CLKDEV_INIT(NULL, "clk_uart_baud1", &s3c24xx_uclk), | 166 | CLKDEV_INIT(NULL, "clk_uart_baud1", &s3c24xx_uclk), |
167 | CLKDEV_INIT(NULL, "clk_uart_baud2", &clk_p), | 167 | CLKDEV_INIT(NULL, "clk_uart_baud2", &clk_p), |
168 | CLKDEV_INIT(NULL, "clk_uart_baud3", &s3c2440_clk_fclk_n), | 168 | CLKDEV_INIT(NULL, "clk_uart_baud3", &s3c2440_clk_fclk_n), |
169 | CLKDEV_INIT("s3c2440-uart.0", "uart", &s3c24xx_clk_uart0), | ||
170 | CLKDEV_INIT("s3c2440-uart.1", "uart", &s3c24xx_clk_uart1), | ||
171 | CLKDEV_INIT("s3c2440-uart.2", "uart", &s3c24xx_clk_uart2), | ||
169 | CLKDEV_INIT("s3c2440-camif", "camera", &s3c2440_clk_cam_upll), | 172 | CLKDEV_INIT("s3c2440-camif", "camera", &s3c2440_clk_cam_upll), |
170 | }; | 173 | }; |
171 | 174 | ||
diff --git a/arch/arm/mach-sti/Kconfig b/arch/arm/mach-sti/Kconfig index d04e3bfe1918..835833e3c4f8 100644 --- a/arch/arm/mach-sti/Kconfig +++ b/arch/arm/mach-sti/Kconfig | |||
@@ -11,8 +11,9 @@ menuconfig ARCH_STI | |||
11 | select HAVE_SMP | 11 | select HAVE_SMP |
12 | select HAVE_ARM_SCU if SMP | 12 | select HAVE_ARM_SCU if SMP |
13 | select ARCH_REQUIRE_GPIOLIB | 13 | select ARCH_REQUIRE_GPIOLIB |
14 | select ARM_ERRATA_720789 | ||
15 | select ARM_ERRATA_754322 | 14 | select ARM_ERRATA_754322 |
15 | select ARM_ERRATA_764369 | ||
16 | select ARM_ERRATA_775420 | ||
16 | select PL310_ERRATA_753970 if CACHE_PL310 | 17 | select PL310_ERRATA_753970 if CACHE_PL310 |
17 | select PL310_ERRATA_769419 if CACHE_PL310 | 18 | select PL310_ERRATA_769419 if CACHE_PL310 |
18 | help | 19 | help |
diff --git a/arch/arm/mach-zynq/common.c b/arch/arm/mach-zynq/common.c index 5b799c29886e..5f252569c689 100644 --- a/arch/arm/mach-zynq/common.c +++ b/arch/arm/mach-zynq/common.c | |||
@@ -91,7 +91,7 @@ static void __init zynq_map_io(void) | |||
91 | zynq_scu_map_io(); | 91 | zynq_scu_map_io(); |
92 | } | 92 | } |
93 | 93 | ||
94 | static void zynq_system_reset(char mode, const char *cmd) | 94 | static void zynq_system_reset(enum reboot_mode mode, const char *cmd) |
95 | { | 95 | { |
96 | zynq_slcr_system_reset(); | 96 | zynq_slcr_system_reset(); |
97 | } | 97 | } |
diff --git a/arch/arm/plat-samsung/Kconfig b/arch/arm/plat-samsung/Kconfig index 3dc5cbea86cc..a5b5ff6e68d2 100644 --- a/arch/arm/plat-samsung/Kconfig +++ b/arch/arm/plat-samsung/Kconfig | |||
@@ -29,6 +29,13 @@ config PLAT_S5P | |||
29 | help | 29 | help |
30 | Base platform code for Samsung's S5P series SoC. | 30 | Base platform code for Samsung's S5P series SoC. |
31 | 31 | ||
32 | config SAMSUNG_PM | ||
33 | bool | ||
34 | depends on PM && (PLAT_S3C24XX || ARCH_S3C64XX || ARCH_S5P64X0 || S5P_PM) | ||
35 | default y | ||
36 | help | ||
37 | Base platform power management code for samsung code | ||
38 | |||
32 | if PLAT_SAMSUNG | 39 | if PLAT_SAMSUNG |
33 | 40 | ||
34 | # boot configurations | 41 | # boot configurations |
diff --git a/arch/arm/plat-samsung/Makefile b/arch/arm/plat-samsung/Makefile index 98d07d8fc7a7..199bbe304d02 100644 --- a/arch/arm/plat-samsung/Makefile +++ b/arch/arm/plat-samsung/Makefile | |||
@@ -51,7 +51,7 @@ obj-$(CONFIG_SAMSUNG_DMADEV) += dma-ops.o | |||
51 | 51 | ||
52 | # PM support | 52 | # PM support |
53 | 53 | ||
54 | obj-$(CONFIG_PM) += pm.o | 54 | obj-$(CONFIG_SAMSUNG_PM) += pm.o |
55 | obj-$(CONFIG_SAMSUNG_PM_GPIO) += pm-gpio.o | 55 | obj-$(CONFIG_SAMSUNG_PM_GPIO) += pm-gpio.o |
56 | obj-$(CONFIG_SAMSUNG_PM_CHECK) += pm-check.o | 56 | obj-$(CONFIG_SAMSUNG_PM_CHECK) += pm-check.o |
57 | 57 | ||
diff --git a/arch/arm/plat-samsung/include/plat/clock.h b/arch/arm/plat-samsung/include/plat/clock.h index a62753dc15ba..df45d6edc98d 100644 --- a/arch/arm/plat-samsung/include/plat/clock.h +++ b/arch/arm/plat-samsung/include/plat/clock.h | |||
@@ -83,6 +83,11 @@ extern struct clk clk_ext; | |||
83 | extern struct clksrc_clk clk_epllref; | 83 | extern struct clksrc_clk clk_epllref; |
84 | extern struct clksrc_clk clk_esysclk; | 84 | extern struct clksrc_clk clk_esysclk; |
85 | 85 | ||
86 | /* S3C24XX UART clocks */ | ||
87 | extern struct clk s3c24xx_clk_uart0; | ||
88 | extern struct clk s3c24xx_clk_uart1; | ||
89 | extern struct clk s3c24xx_clk_uart2; | ||
90 | |||
86 | /* S3C64XX specific clocks */ | 91 | /* S3C64XX specific clocks */ |
87 | extern struct clk clk_h2; | 92 | extern struct clk clk_h2; |
88 | extern struct clk clk_27m; | 93 | extern struct clk clk_27m; |
diff --git a/arch/arm/plat-samsung/include/plat/pm.h b/arch/arm/plat-samsung/include/plat/pm.h index 5d47ca35cabd..6bc1a8f471e3 100644 --- a/arch/arm/plat-samsung/include/plat/pm.h +++ b/arch/arm/plat-samsung/include/plat/pm.h | |||
@@ -19,7 +19,7 @@ | |||
19 | 19 | ||
20 | struct device; | 20 | struct device; |
21 | 21 | ||
22 | #ifdef CONFIG_PM | 22 | #ifdef CONFIG_SAMSUNG_PM |
23 | 23 | ||
24 | extern __init int s3c_pm_init(void); | 24 | extern __init int s3c_pm_init(void); |
25 | extern __init int s3c64xx_pm_init(void); | 25 | extern __init int s3c64xx_pm_init(void); |
@@ -58,8 +58,6 @@ extern unsigned char pm_uart_udivslot; /* true to save UART UDIVSLOT */ | |||
58 | 58 | ||
59 | /* from sleep.S */ | 59 | /* from sleep.S */ |
60 | 60 | ||
61 | extern void s3c_cpu_resume(void); | ||
62 | |||
63 | extern int s3c2410_cpu_suspend(unsigned long); | 61 | extern int s3c2410_cpu_suspend(unsigned long); |
64 | 62 | ||
65 | /* sleep save info */ | 63 | /* sleep save info */ |
@@ -106,12 +104,14 @@ extern void s3c_pm_do_save(struct sleep_save *ptr, int count); | |||
106 | extern void s3c_pm_do_restore(struct sleep_save *ptr, int count); | 104 | extern void s3c_pm_do_restore(struct sleep_save *ptr, int count); |
107 | extern void s3c_pm_do_restore_core(struct sleep_save *ptr, int count); | 105 | extern void s3c_pm_do_restore_core(struct sleep_save *ptr, int count); |
108 | 106 | ||
109 | #ifdef CONFIG_PM | 107 | #ifdef CONFIG_SAMSUNG_PM |
110 | extern int s3c_irq_wake(struct irq_data *data, unsigned int state); | 108 | extern int s3c_irq_wake(struct irq_data *data, unsigned int state); |
111 | extern int s3c_irqext_wake(struct irq_data *data, unsigned int state); | 109 | extern int s3c_irqext_wake(struct irq_data *data, unsigned int state); |
110 | extern void s3c_cpu_resume(void); | ||
112 | #else | 111 | #else |
113 | #define s3c_irq_wake NULL | 112 | #define s3c_irq_wake NULL |
114 | #define s3c_irqext_wake NULL | 113 | #define s3c_irqext_wake NULL |
114 | #define s3c_cpu_resume NULL | ||
115 | #endif | 115 | #endif |
116 | 116 | ||
117 | /* PM debug functions */ | 117 | /* PM debug functions */ |
diff --git a/arch/arm/plat-samsung/pm.c b/arch/arm/plat-samsung/pm.c index ea3613642451..d0c23010b693 100644 --- a/arch/arm/plat-samsung/pm.c +++ b/arch/arm/plat-samsung/pm.c | |||
@@ -80,7 +80,7 @@ unsigned char pm_uart_udivslot; | |||
80 | 80 | ||
81 | #ifdef CONFIG_SAMSUNG_PM_DEBUG | 81 | #ifdef CONFIG_SAMSUNG_PM_DEBUG |
82 | 82 | ||
83 | static struct pm_uart_save uart_save[CONFIG_SERIAL_SAMSUNG_UARTS]; | 83 | static struct pm_uart_save uart_save; |
84 | 84 | ||
85 | static void s3c_pm_save_uart(unsigned int uart, struct pm_uart_save *save) | 85 | static void s3c_pm_save_uart(unsigned int uart, struct pm_uart_save *save) |
86 | { | 86 | { |
@@ -101,11 +101,7 @@ static void s3c_pm_save_uart(unsigned int uart, struct pm_uart_save *save) | |||
101 | 101 | ||
102 | static void s3c_pm_save_uarts(void) | 102 | static void s3c_pm_save_uarts(void) |
103 | { | 103 | { |
104 | struct pm_uart_save *save = uart_save; | 104 | s3c_pm_save_uart(CONFIG_DEBUG_S3C_UART, &uart_save); |
105 | unsigned int uart; | ||
106 | |||
107 | for (uart = 0; uart < CONFIG_SERIAL_SAMSUNG_UARTS; uart++, save++) | ||
108 | s3c_pm_save_uart(uart, save); | ||
109 | } | 105 | } |
110 | 106 | ||
111 | static void s3c_pm_restore_uart(unsigned int uart, struct pm_uart_save *save) | 107 | static void s3c_pm_restore_uart(unsigned int uart, struct pm_uart_save *save) |
@@ -126,11 +122,7 @@ static void s3c_pm_restore_uart(unsigned int uart, struct pm_uart_save *save) | |||
126 | 122 | ||
127 | static void s3c_pm_restore_uarts(void) | 123 | static void s3c_pm_restore_uarts(void) |
128 | { | 124 | { |
129 | struct pm_uart_save *save = uart_save; | 125 | s3c_pm_restore_uart(CONFIG_DEBUG_S3C_UART, &uart_save); |
130 | unsigned int uart; | ||
131 | |||
132 | for (uart = 0; uart < CONFIG_SERIAL_SAMSUNG_UARTS; uart++, save++) | ||
133 | s3c_pm_restore_uart(uart, save); | ||
134 | } | 126 | } |
135 | #else | 127 | #else |
136 | static void s3c_pm_save_uarts(void) { } | 128 | static void s3c_pm_save_uarts(void) { } |
diff --git a/arch/arm64/include/asm/thread_info.h b/arch/arm64/include/asm/thread_info.h index 3659e460071d..23a3c4791d86 100644 --- a/arch/arm64/include/asm/thread_info.h +++ b/arch/arm64/include/asm/thread_info.h | |||
@@ -24,10 +24,10 @@ | |||
24 | #include <linux/compiler.h> | 24 | #include <linux/compiler.h> |
25 | 25 | ||
26 | #ifndef CONFIG_ARM64_64K_PAGES | 26 | #ifndef CONFIG_ARM64_64K_PAGES |
27 | #define THREAD_SIZE_ORDER 1 | 27 | #define THREAD_SIZE_ORDER 2 |
28 | #endif | 28 | #endif |
29 | 29 | ||
30 | #define THREAD_SIZE 8192 | 30 | #define THREAD_SIZE 16384 |
31 | #define THREAD_START_SP (THREAD_SIZE - 16) | 31 | #define THREAD_START_SP (THREAD_SIZE - 16) |
32 | 32 | ||
33 | #ifndef __ASSEMBLY__ | 33 | #ifndef __ASSEMBLY__ |
diff --git a/arch/arm64/include/asm/virt.h b/arch/arm64/include/asm/virt.h index 439827271e3d..26e310c54344 100644 --- a/arch/arm64/include/asm/virt.h +++ b/arch/arm64/include/asm/virt.h | |||
@@ -21,6 +21,7 @@ | |||
21 | #define BOOT_CPU_MODE_EL2 (0x0e12b007) | 21 | #define BOOT_CPU_MODE_EL2 (0x0e12b007) |
22 | 22 | ||
23 | #ifndef __ASSEMBLY__ | 23 | #ifndef __ASSEMBLY__ |
24 | #include <asm/cacheflush.h> | ||
24 | 25 | ||
25 | /* | 26 | /* |
26 | * __boot_cpu_mode records what mode CPUs were booted in. | 27 | * __boot_cpu_mode records what mode CPUs were booted in. |
@@ -36,9 +37,20 @@ extern u32 __boot_cpu_mode[2]; | |||
36 | void __hyp_set_vectors(phys_addr_t phys_vector_base); | 37 | void __hyp_set_vectors(phys_addr_t phys_vector_base); |
37 | phys_addr_t __hyp_get_vectors(void); | 38 | phys_addr_t __hyp_get_vectors(void); |
38 | 39 | ||
40 | static inline void sync_boot_mode(void) | ||
41 | { | ||
42 | /* | ||
43 | * As secondaries write to __boot_cpu_mode with caches disabled, we | ||
44 | * must flush the corresponding cache entries to ensure the visibility | ||
45 | * of their writes. | ||
46 | */ | ||
47 | __flush_dcache_area(__boot_cpu_mode, sizeof(__boot_cpu_mode)); | ||
48 | } | ||
49 | |||
39 | /* Reports the availability of HYP mode */ | 50 | /* Reports the availability of HYP mode */ |
40 | static inline bool is_hyp_mode_available(void) | 51 | static inline bool is_hyp_mode_available(void) |
41 | { | 52 | { |
53 | sync_boot_mode(); | ||
42 | return (__boot_cpu_mode[0] == BOOT_CPU_MODE_EL2 && | 54 | return (__boot_cpu_mode[0] == BOOT_CPU_MODE_EL2 && |
43 | __boot_cpu_mode[1] == BOOT_CPU_MODE_EL2); | 55 | __boot_cpu_mode[1] == BOOT_CPU_MODE_EL2); |
44 | } | 56 | } |
@@ -46,6 +58,7 @@ static inline bool is_hyp_mode_available(void) | |||
46 | /* Check if the bootloader has booted CPUs in different modes */ | 58 | /* Check if the bootloader has booted CPUs in different modes */ |
47 | static inline bool is_hyp_mode_mismatched(void) | 59 | static inline bool is_hyp_mode_mismatched(void) |
48 | { | 60 | { |
61 | sync_boot_mode(); | ||
49 | return __boot_cpu_mode[0] != __boot_cpu_mode[1]; | 62 | return __boot_cpu_mode[0] != __boot_cpu_mode[1]; |
50 | } | 63 | } |
51 | 64 | ||
diff --git a/arch/arm64/kernel/entry.S b/arch/arm64/kernel/entry.S index 1d1314280a03..6ad781b21c08 100644 --- a/arch/arm64/kernel/entry.S +++ b/arch/arm64/kernel/entry.S | |||
@@ -121,7 +121,7 @@ | |||
121 | 121 | ||
122 | .macro get_thread_info, rd | 122 | .macro get_thread_info, rd |
123 | mov \rd, sp | 123 | mov \rd, sp |
124 | and \rd, \rd, #~((1 << 13) - 1) // top of 8K stack | 124 | and \rd, \rd, #~(THREAD_SIZE - 1) // top of stack |
125 | .endm | 125 | .endm |
126 | 126 | ||
127 | /* | 127 | /* |
diff --git a/arch/arm64/kernel/process.c b/arch/arm64/kernel/process.c index 1788bf6b471f..57fb55c44c90 100644 --- a/arch/arm64/kernel/process.c +++ b/arch/arm64/kernel/process.c | |||
@@ -81,7 +81,7 @@ void soft_restart(unsigned long addr) | |||
81 | void (*pm_power_off)(void); | 81 | void (*pm_power_off)(void); |
82 | EXPORT_SYMBOL_GPL(pm_power_off); | 82 | EXPORT_SYMBOL_GPL(pm_power_off); |
83 | 83 | ||
84 | void (*arm_pm_restart)(char str, const char *cmd); | 84 | void (*arm_pm_restart)(enum reboot_mode reboot_mode, const char *cmd); |
85 | EXPORT_SYMBOL_GPL(arm_pm_restart); | 85 | EXPORT_SYMBOL_GPL(arm_pm_restart); |
86 | 86 | ||
87 | void arch_cpu_idle_prepare(void) | 87 | void arch_cpu_idle_prepare(void) |
diff --git a/arch/powerpc/include/asm/eeh.h b/arch/powerpc/include/asm/eeh.h index 09a8743143f3..d3e5e9bc8f94 100644 --- a/arch/powerpc/include/asm/eeh.h +++ b/arch/powerpc/include/asm/eeh.h | |||
@@ -55,6 +55,8 @@ struct device_node; | |||
55 | #define EEH_PE_RECOVERING (1 << 1) /* Recovering PE */ | 55 | #define EEH_PE_RECOVERING (1 << 1) /* Recovering PE */ |
56 | #define EEH_PE_PHB_DEAD (1 << 2) /* Dead PHB */ | 56 | #define EEH_PE_PHB_DEAD (1 << 2) /* Dead PHB */ |
57 | 57 | ||
58 | #define EEH_PE_KEEP (1 << 8) /* Keep PE on hotplug */ | ||
59 | |||
58 | struct eeh_pe { | 60 | struct eeh_pe { |
59 | int type; /* PE type: PHB/Bus/Device */ | 61 | int type; /* PE type: PHB/Bus/Device */ |
60 | int state; /* PE EEH dependent mode */ | 62 | int state; /* PE EEH dependent mode */ |
@@ -72,8 +74,8 @@ struct eeh_pe { | |||
72 | struct list_head child; /* Child PEs */ | 74 | struct list_head child; /* Child PEs */ |
73 | }; | 75 | }; |
74 | 76 | ||
75 | #define eeh_pe_for_each_dev(pe, edev) \ | 77 | #define eeh_pe_for_each_dev(pe, edev, tmp) \ |
76 | list_for_each_entry(edev, &pe->edevs, list) | 78 | list_for_each_entry_safe(edev, tmp, &pe->edevs, list) |
77 | 79 | ||
78 | /* | 80 | /* |
79 | * The struct is used to trace EEH state for the associated | 81 | * The struct is used to trace EEH state for the associated |
@@ -82,7 +84,13 @@ struct eeh_pe { | |||
82 | * another tree except the currently existing tree of PCI | 84 | * another tree except the currently existing tree of PCI |
83 | * buses and PCI devices | 85 | * buses and PCI devices |
84 | */ | 86 | */ |
85 | #define EEH_DEV_IRQ_DISABLED (1<<0) /* Interrupt disabled */ | 87 | #define EEH_DEV_BRIDGE (1 << 0) /* PCI bridge */ |
88 | #define EEH_DEV_ROOT_PORT (1 << 1) /* PCIe root port */ | ||
89 | #define EEH_DEV_DS_PORT (1 << 2) /* Downstream port */ | ||
90 | #define EEH_DEV_IRQ_DISABLED (1 << 3) /* Interrupt disabled */ | ||
91 | #define EEH_DEV_DISCONNECTED (1 << 4) /* Removing from PE */ | ||
92 | |||
93 | #define EEH_DEV_SYSFS (1 << 8) /* Sysfs created */ | ||
86 | 94 | ||
87 | struct eeh_dev { | 95 | struct eeh_dev { |
88 | int mode; /* EEH mode */ | 96 | int mode; /* EEH mode */ |
@@ -90,11 +98,13 @@ struct eeh_dev { | |||
90 | int config_addr; /* Config address */ | 98 | int config_addr; /* Config address */ |
91 | int pe_config_addr; /* PE config address */ | 99 | int pe_config_addr; /* PE config address */ |
92 | u32 config_space[16]; /* Saved PCI config space */ | 100 | u32 config_space[16]; /* Saved PCI config space */ |
101 | u8 pcie_cap; /* Saved PCIe capability */ | ||
93 | struct eeh_pe *pe; /* Associated PE */ | 102 | struct eeh_pe *pe; /* Associated PE */ |
94 | struct list_head list; /* Form link list in the PE */ | 103 | struct list_head list; /* Form link list in the PE */ |
95 | struct pci_controller *phb; /* Associated PHB */ | 104 | struct pci_controller *phb; /* Associated PHB */ |
96 | struct device_node *dn; /* Associated device node */ | 105 | struct device_node *dn; /* Associated device node */ |
97 | struct pci_dev *pdev; /* Associated PCI device */ | 106 | struct pci_dev *pdev; /* Associated PCI device */ |
107 | struct pci_bus *bus; /* PCI bus for partial hotplug */ | ||
98 | }; | 108 | }; |
99 | 109 | ||
100 | static inline struct device_node *eeh_dev_to_of_node(struct eeh_dev *edev) | 110 | static inline struct device_node *eeh_dev_to_of_node(struct eeh_dev *edev) |
@@ -193,8 +203,10 @@ int eeh_phb_pe_create(struct pci_controller *phb); | |||
193 | struct eeh_pe *eeh_phb_pe_get(struct pci_controller *phb); | 203 | struct eeh_pe *eeh_phb_pe_get(struct pci_controller *phb); |
194 | struct eeh_pe *eeh_pe_get(struct eeh_dev *edev); | 204 | struct eeh_pe *eeh_pe_get(struct eeh_dev *edev); |
195 | int eeh_add_to_parent_pe(struct eeh_dev *edev); | 205 | int eeh_add_to_parent_pe(struct eeh_dev *edev); |
196 | int eeh_rmv_from_parent_pe(struct eeh_dev *edev, int purge_pe); | 206 | int eeh_rmv_from_parent_pe(struct eeh_dev *edev); |
197 | void eeh_pe_update_time_stamp(struct eeh_pe *pe); | 207 | void eeh_pe_update_time_stamp(struct eeh_pe *pe); |
208 | void *eeh_pe_traverse(struct eeh_pe *root, | ||
209 | eeh_traverse_func fn, void *flag); | ||
198 | void *eeh_pe_dev_traverse(struct eeh_pe *root, | 210 | void *eeh_pe_dev_traverse(struct eeh_pe *root, |
199 | eeh_traverse_func fn, void *flag); | 211 | eeh_traverse_func fn, void *flag); |
200 | void eeh_pe_restore_bars(struct eeh_pe *pe); | 212 | void eeh_pe_restore_bars(struct eeh_pe *pe); |
@@ -209,10 +221,12 @@ unsigned long eeh_check_failure(const volatile void __iomem *token, | |||
209 | unsigned long val); | 221 | unsigned long val); |
210 | int eeh_dev_check_failure(struct eeh_dev *edev); | 222 | int eeh_dev_check_failure(struct eeh_dev *edev); |
211 | void eeh_addr_cache_build(void); | 223 | void eeh_addr_cache_build(void); |
224 | void eeh_add_device_early(struct device_node *); | ||
212 | void eeh_add_device_tree_early(struct device_node *); | 225 | void eeh_add_device_tree_early(struct device_node *); |
226 | void eeh_add_device_late(struct pci_dev *); | ||
213 | void eeh_add_device_tree_late(struct pci_bus *); | 227 | void eeh_add_device_tree_late(struct pci_bus *); |
214 | void eeh_add_sysfs_files(struct pci_bus *); | 228 | void eeh_add_sysfs_files(struct pci_bus *); |
215 | void eeh_remove_bus_device(struct pci_dev *, int); | 229 | void eeh_remove_device(struct pci_dev *); |
216 | 230 | ||
217 | /** | 231 | /** |
218 | * EEH_POSSIBLE_ERROR() -- test for possible MMIO failure. | 232 | * EEH_POSSIBLE_ERROR() -- test for possible MMIO failure. |
@@ -252,13 +266,17 @@ static inline unsigned long eeh_check_failure(const volatile void __iomem *token | |||
252 | 266 | ||
253 | static inline void eeh_addr_cache_build(void) { } | 267 | static inline void eeh_addr_cache_build(void) { } |
254 | 268 | ||
269 | static inline void eeh_add_device_early(struct device_node *dn) { } | ||
270 | |||
255 | static inline void eeh_add_device_tree_early(struct device_node *dn) { } | 271 | static inline void eeh_add_device_tree_early(struct device_node *dn) { } |
256 | 272 | ||
273 | static inline void eeh_add_device_late(struct pci_dev *dev) { } | ||
274 | |||
257 | static inline void eeh_add_device_tree_late(struct pci_bus *bus) { } | 275 | static inline void eeh_add_device_tree_late(struct pci_bus *bus) { } |
258 | 276 | ||
259 | static inline void eeh_add_sysfs_files(struct pci_bus *bus) { } | 277 | static inline void eeh_add_sysfs_files(struct pci_bus *bus) { } |
260 | 278 | ||
261 | static inline void eeh_remove_bus_device(struct pci_dev *dev, int purge_pe) { } | 279 | static inline void eeh_remove_device(struct pci_dev *dev) { } |
262 | 280 | ||
263 | #define EEH_POSSIBLE_ERROR(val, type) (0) | 281 | #define EEH_POSSIBLE_ERROR(val, type) (0) |
264 | #define EEH_IO_ERROR_VALUE(size) (-1UL) | 282 | #define EEH_IO_ERROR_VALUE(size) (-1UL) |
diff --git a/arch/powerpc/include/asm/hw_irq.h b/arch/powerpc/include/asm/hw_irq.h index ba713f166fa5..10be1dd01c6b 100644 --- a/arch/powerpc/include/asm/hw_irq.h +++ b/arch/powerpc/include/asm/hw_irq.h | |||
@@ -96,10 +96,11 @@ static inline bool arch_irqs_disabled(void) | |||
96 | #endif | 96 | #endif |
97 | 97 | ||
98 | #define hard_irq_disable() do { \ | 98 | #define hard_irq_disable() do { \ |
99 | u8 _was_enabled = get_paca()->soft_enabled; \ | 99 | u8 _was_enabled; \ |
100 | __hard_irq_disable(); \ | 100 | __hard_irq_disable(); \ |
101 | get_paca()->soft_enabled = 0; \ | 101 | _was_enabled = local_paca->soft_enabled; \ |
102 | get_paca()->irq_happened |= PACA_IRQ_HARD_DIS; \ | 102 | local_paca->soft_enabled = 0; \ |
103 | local_paca->irq_happened |= PACA_IRQ_HARD_DIS; \ | ||
103 | if (_was_enabled) \ | 104 | if (_was_enabled) \ |
104 | trace_hardirqs_off(); \ | 105 | trace_hardirqs_off(); \ |
105 | } while(0) | 106 | } while(0) |
diff --git a/arch/powerpc/include/asm/module.h b/arch/powerpc/include/asm/module.h index c1df590ec444..49fa55bfbac4 100644 --- a/arch/powerpc/include/asm/module.h +++ b/arch/powerpc/include/asm/module.h | |||
@@ -82,10 +82,9 @@ struct exception_table_entry; | |||
82 | void sort_ex_table(struct exception_table_entry *start, | 82 | void sort_ex_table(struct exception_table_entry *start, |
83 | struct exception_table_entry *finish); | 83 | struct exception_table_entry *finish); |
84 | 84 | ||
85 | #ifdef CONFIG_MODVERSIONS | 85 | #if defined(CONFIG_MODVERSIONS) && defined(CONFIG_PPC64) |
86 | #define ARCH_RELOCATES_KCRCTAB | 86 | #define ARCH_RELOCATES_KCRCTAB |
87 | 87 | #define reloc_start PHYSICAL_START | |
88 | extern const unsigned long reloc_start[]; | ||
89 | #endif | 88 | #endif |
90 | #endif /* __KERNEL__ */ | 89 | #endif /* __KERNEL__ */ |
91 | #endif /* _ASM_POWERPC_MODULE_H */ | 90 | #endif /* _ASM_POWERPC_MODULE_H */ |
diff --git a/arch/powerpc/include/asm/pci-bridge.h b/arch/powerpc/include/asm/pci-bridge.h index 2c1d8cb9b265..32d0d2018faf 100644 --- a/arch/powerpc/include/asm/pci-bridge.h +++ b/arch/powerpc/include/asm/pci-bridge.h | |||
@@ -209,7 +209,6 @@ static inline struct eeh_dev *of_node_to_eeh_dev(struct device_node *dn) | |||
209 | extern struct pci_bus *pcibios_find_pci_bus(struct device_node *dn); | 209 | extern struct pci_bus *pcibios_find_pci_bus(struct device_node *dn); |
210 | 210 | ||
211 | /** Remove all of the PCI devices under this bus */ | 211 | /** Remove all of the PCI devices under this bus */ |
212 | extern void __pcibios_remove_pci_devices(struct pci_bus *bus, int purge_pe); | ||
213 | extern void pcibios_remove_pci_devices(struct pci_bus *bus); | 212 | extern void pcibios_remove_pci_devices(struct pci_bus *bus); |
214 | 213 | ||
215 | /** Discover new pci devices under this bus, and add them */ | 214 | /** Discover new pci devices under this bus, and add them */ |
diff --git a/arch/powerpc/include/asm/reg.h b/arch/powerpc/include/asm/reg.h index 5d7d9c2a5473..a6840e4e24f7 100644 --- a/arch/powerpc/include/asm/reg.h +++ b/arch/powerpc/include/asm/reg.h | |||
@@ -1088,7 +1088,8 @@ | |||
1088 | #define PVR_970MP 0x0044 | 1088 | #define PVR_970MP 0x0044 |
1089 | #define PVR_970GX 0x0045 | 1089 | #define PVR_970GX 0x0045 |
1090 | #define PVR_POWER7p 0x004A | 1090 | #define PVR_POWER7p 0x004A |
1091 | #define PVR_POWER8 0x004B | 1091 | #define PVR_POWER8E 0x004B |
1092 | #define PVR_POWER8 0x004D | ||
1092 | #define PVR_BE 0x0070 | 1093 | #define PVR_BE 0x0070 |
1093 | #define PVR_PA6T 0x0090 | 1094 | #define PVR_PA6T 0x0090 |
1094 | 1095 | ||
diff --git a/arch/powerpc/kernel/cputable.c b/arch/powerpc/kernel/cputable.c index 2a45d0f04385..22973a74df73 100644 --- a/arch/powerpc/kernel/cputable.c +++ b/arch/powerpc/kernel/cputable.c | |||
@@ -494,9 +494,27 @@ static struct cpu_spec __initdata cpu_specs[] = { | |||
494 | .cpu_restore = __restore_cpu_power7, | 494 | .cpu_restore = __restore_cpu_power7, |
495 | .platform = "power7+", | 495 | .platform = "power7+", |
496 | }, | 496 | }, |
497 | { /* Power8 */ | 497 | { /* Power8E */ |
498 | .pvr_mask = 0xffff0000, | 498 | .pvr_mask = 0xffff0000, |
499 | .pvr_value = 0x004b0000, | 499 | .pvr_value = 0x004b0000, |
500 | .cpu_name = "POWER8E (raw)", | ||
501 | .cpu_features = CPU_FTRS_POWER8, | ||
502 | .cpu_user_features = COMMON_USER_POWER8, | ||
503 | .cpu_user_features2 = COMMON_USER2_POWER8, | ||
504 | .mmu_features = MMU_FTRS_POWER8, | ||
505 | .icache_bsize = 128, | ||
506 | .dcache_bsize = 128, | ||
507 | .num_pmcs = 6, | ||
508 | .pmc_type = PPC_PMC_IBM, | ||
509 | .oprofile_cpu_type = "ppc64/power8", | ||
510 | .oprofile_type = PPC_OPROFILE_INVALID, | ||
511 | .cpu_setup = __setup_cpu_power8, | ||
512 | .cpu_restore = __restore_cpu_power8, | ||
513 | .platform = "power8", | ||
514 | }, | ||
515 | { /* Power8 */ | ||
516 | .pvr_mask = 0xffff0000, | ||
517 | .pvr_value = 0x004d0000, | ||
500 | .cpu_name = "POWER8 (raw)", | 518 | .cpu_name = "POWER8 (raw)", |
501 | .cpu_features = CPU_FTRS_POWER8, | 519 | .cpu_features = CPU_FTRS_POWER8, |
502 | .cpu_user_features = COMMON_USER_POWER8, | 520 | .cpu_user_features = COMMON_USER_POWER8, |
diff --git a/arch/powerpc/kernel/eeh.c b/arch/powerpc/kernel/eeh.c index 39954fe941b8..ea9414c8088d 100644 --- a/arch/powerpc/kernel/eeh.c +++ b/arch/powerpc/kernel/eeh.c | |||
@@ -231,7 +231,7 @@ static size_t eeh_gather_pci_data(struct eeh_dev *edev, char * buf, size_t len) | |||
231 | void eeh_slot_error_detail(struct eeh_pe *pe, int severity) | 231 | void eeh_slot_error_detail(struct eeh_pe *pe, int severity) |
232 | { | 232 | { |
233 | size_t loglen = 0; | 233 | size_t loglen = 0; |
234 | struct eeh_dev *edev; | 234 | struct eeh_dev *edev, *tmp; |
235 | bool valid_cfg_log = true; | 235 | bool valid_cfg_log = true; |
236 | 236 | ||
237 | /* | 237 | /* |
@@ -251,7 +251,7 @@ void eeh_slot_error_detail(struct eeh_pe *pe, int severity) | |||
251 | eeh_pe_restore_bars(pe); | 251 | eeh_pe_restore_bars(pe); |
252 | 252 | ||
253 | pci_regs_buf[0] = 0; | 253 | pci_regs_buf[0] = 0; |
254 | eeh_pe_for_each_dev(pe, edev) { | 254 | eeh_pe_for_each_dev(pe, edev, tmp) { |
255 | loglen += eeh_gather_pci_data(edev, pci_regs_buf + loglen, | 255 | loglen += eeh_gather_pci_data(edev, pci_regs_buf + loglen, |
256 | EEH_PCI_REGS_LOG_LEN - loglen); | 256 | EEH_PCI_REGS_LOG_LEN - loglen); |
257 | } | 257 | } |
@@ -499,8 +499,6 @@ unsigned long eeh_check_failure(const volatile void __iomem *token, unsigned lon | |||
499 | } | 499 | } |
500 | 500 | ||
501 | eeh_dev_check_failure(edev); | 501 | eeh_dev_check_failure(edev); |
502 | |||
503 | pci_dev_put(eeh_dev_to_pci_dev(edev)); | ||
504 | return val; | 502 | return val; |
505 | } | 503 | } |
506 | 504 | ||
@@ -838,7 +836,7 @@ core_initcall_sync(eeh_init); | |||
838 | * on the CEC architecture, type of the device, on earlier boot | 836 | * on the CEC architecture, type of the device, on earlier boot |
839 | * command-line arguments & etc. | 837 | * command-line arguments & etc. |
840 | */ | 838 | */ |
841 | static void eeh_add_device_early(struct device_node *dn) | 839 | void eeh_add_device_early(struct device_node *dn) |
842 | { | 840 | { |
843 | struct pci_controller *phb; | 841 | struct pci_controller *phb; |
844 | 842 | ||
@@ -886,7 +884,7 @@ EXPORT_SYMBOL_GPL(eeh_add_device_tree_early); | |||
886 | * This routine must be used to complete EEH initialization for PCI | 884 | * This routine must be used to complete EEH initialization for PCI |
887 | * devices that were added after system boot (e.g. hotplug, dlpar). | 885 | * devices that were added after system boot (e.g. hotplug, dlpar). |
888 | */ | 886 | */ |
889 | static void eeh_add_device_late(struct pci_dev *dev) | 887 | void eeh_add_device_late(struct pci_dev *dev) |
890 | { | 888 | { |
891 | struct device_node *dn; | 889 | struct device_node *dn; |
892 | struct eeh_dev *edev; | 890 | struct eeh_dev *edev; |
@@ -902,9 +900,23 @@ static void eeh_add_device_late(struct pci_dev *dev) | |||
902 | pr_debug("EEH: Already referenced !\n"); | 900 | pr_debug("EEH: Already referenced !\n"); |
903 | return; | 901 | return; |
904 | } | 902 | } |
905 | WARN_ON(edev->pdev); | ||
906 | 903 | ||
907 | pci_dev_get(dev); | 904 | /* |
905 | * The EEH cache might not be removed correctly because of | ||
906 | * unbalanced kref to the device during unplug time, which | ||
907 | * relies on pcibios_release_device(). So we have to remove | ||
908 | * that here explicitly. | ||
909 | */ | ||
910 | if (edev->pdev) { | ||
911 | eeh_rmv_from_parent_pe(edev); | ||
912 | eeh_addr_cache_rmv_dev(edev->pdev); | ||
913 | eeh_sysfs_remove_device(edev->pdev); | ||
914 | edev->mode &= ~EEH_DEV_SYSFS; | ||
915 | |||
916 | edev->pdev = NULL; | ||
917 | dev->dev.archdata.edev = NULL; | ||
918 | } | ||
919 | |||
908 | edev->pdev = dev; | 920 | edev->pdev = dev; |
909 | dev->dev.archdata.edev = edev; | 921 | dev->dev.archdata.edev = edev; |
910 | 922 | ||
@@ -967,7 +979,6 @@ EXPORT_SYMBOL_GPL(eeh_add_sysfs_files); | |||
967 | /** | 979 | /** |
968 | * eeh_remove_device - Undo EEH setup for the indicated pci device | 980 | * eeh_remove_device - Undo EEH setup for the indicated pci device |
969 | * @dev: pci device to be removed | 981 | * @dev: pci device to be removed |
970 | * @purge_pe: remove the PE or not | ||
971 | * | 982 | * |
972 | * This routine should be called when a device is removed from | 983 | * This routine should be called when a device is removed from |
973 | * a running system (e.g. by hotplug or dlpar). It unregisters | 984 | * a running system (e.g. by hotplug or dlpar). It unregisters |
@@ -975,7 +986,7 @@ EXPORT_SYMBOL_GPL(eeh_add_sysfs_files); | |||
975 | * this device will no longer be detected after this call; thus, | 986 | * this device will no longer be detected after this call; thus, |
976 | * i/o errors affecting this slot may leave this device unusable. | 987 | * i/o errors affecting this slot may leave this device unusable. |
977 | */ | 988 | */ |
978 | static void eeh_remove_device(struct pci_dev *dev, int purge_pe) | 989 | void eeh_remove_device(struct pci_dev *dev) |
979 | { | 990 | { |
980 | struct eeh_dev *edev; | 991 | struct eeh_dev *edev; |
981 | 992 | ||
@@ -986,42 +997,29 @@ static void eeh_remove_device(struct pci_dev *dev, int purge_pe) | |||
986 | /* Unregister the device with the EEH/PCI address search system */ | 997 | /* Unregister the device with the EEH/PCI address search system */ |
987 | pr_debug("EEH: Removing device %s\n", pci_name(dev)); | 998 | pr_debug("EEH: Removing device %s\n", pci_name(dev)); |
988 | 999 | ||
989 | if (!edev || !edev->pdev) { | 1000 | if (!edev || !edev->pdev || !edev->pe) { |
990 | pr_debug("EEH: Not referenced !\n"); | 1001 | pr_debug("EEH: Not referenced !\n"); |
991 | return; | 1002 | return; |
992 | } | 1003 | } |
1004 | |||
1005 | /* | ||
1006 | * During the hotplug for EEH error recovery, we need the EEH | ||
1007 | * device attached to the parent PE in order for BAR restore | ||
1008 | * a bit later. So we keep it for BAR restore and remove it | ||
1009 | * from the parent PE during the BAR resotre. | ||
1010 | */ | ||
993 | edev->pdev = NULL; | 1011 | edev->pdev = NULL; |
994 | dev->dev.archdata.edev = NULL; | 1012 | dev->dev.archdata.edev = NULL; |
995 | pci_dev_put(dev); | 1013 | if (!(edev->pe->state & EEH_PE_KEEP)) |
1014 | eeh_rmv_from_parent_pe(edev); | ||
1015 | else | ||
1016 | edev->mode |= EEH_DEV_DISCONNECTED; | ||
996 | 1017 | ||
997 | eeh_rmv_from_parent_pe(edev, purge_pe); | ||
998 | eeh_addr_cache_rmv_dev(dev); | 1018 | eeh_addr_cache_rmv_dev(dev); |
999 | eeh_sysfs_remove_device(dev); | 1019 | eeh_sysfs_remove_device(dev); |
1020 | edev->mode &= ~EEH_DEV_SYSFS; | ||
1000 | } | 1021 | } |
1001 | 1022 | ||
1002 | /** | ||
1003 | * eeh_remove_bus_device - Undo EEH setup for the indicated PCI device | ||
1004 | * @dev: PCI device | ||
1005 | * @purge_pe: remove the corresponding PE or not | ||
1006 | * | ||
1007 | * This routine must be called when a device is removed from the | ||
1008 | * running system through hotplug or dlpar. The corresponding | ||
1009 | * PCI address cache will be removed. | ||
1010 | */ | ||
1011 | void eeh_remove_bus_device(struct pci_dev *dev, int purge_pe) | ||
1012 | { | ||
1013 | struct pci_bus *bus = dev->subordinate; | ||
1014 | struct pci_dev *child, *tmp; | ||
1015 | |||
1016 | eeh_remove_device(dev, purge_pe); | ||
1017 | |||
1018 | if (bus && dev->hdr_type == PCI_HEADER_TYPE_BRIDGE) { | ||
1019 | list_for_each_entry_safe(child, tmp, &bus->devices, bus_list) | ||
1020 | eeh_remove_bus_device(child, purge_pe); | ||
1021 | } | ||
1022 | } | ||
1023 | EXPORT_SYMBOL_GPL(eeh_remove_bus_device); | ||
1024 | |||
1025 | static int proc_eeh_show(struct seq_file *m, void *v) | 1023 | static int proc_eeh_show(struct seq_file *m, void *v) |
1026 | { | 1024 | { |
1027 | if (0 == eeh_subsystem_enabled) { | 1025 | if (0 == eeh_subsystem_enabled) { |
diff --git a/arch/powerpc/kernel/eeh_cache.c b/arch/powerpc/kernel/eeh_cache.c index f9ac1232a746..e8c9fd546a5c 100644 --- a/arch/powerpc/kernel/eeh_cache.c +++ b/arch/powerpc/kernel/eeh_cache.c | |||
@@ -68,16 +68,12 @@ static inline struct eeh_dev *__eeh_addr_cache_get_device(unsigned long addr) | |||
68 | struct pci_io_addr_range *piar; | 68 | struct pci_io_addr_range *piar; |
69 | piar = rb_entry(n, struct pci_io_addr_range, rb_node); | 69 | piar = rb_entry(n, struct pci_io_addr_range, rb_node); |
70 | 70 | ||
71 | if (addr < piar->addr_lo) { | 71 | if (addr < piar->addr_lo) |
72 | n = n->rb_left; | 72 | n = n->rb_left; |
73 | } else { | 73 | else if (addr > piar->addr_hi) |
74 | if (addr > piar->addr_hi) { | 74 | n = n->rb_right; |
75 | n = n->rb_right; | 75 | else |
76 | } else { | 76 | return piar->edev; |
77 | pci_dev_get(piar->pcidev); | ||
78 | return piar->edev; | ||
79 | } | ||
80 | } | ||
81 | } | 77 | } |
82 | 78 | ||
83 | return NULL; | 79 | return NULL; |
@@ -156,7 +152,6 @@ eeh_addr_cache_insert(struct pci_dev *dev, unsigned long alo, | |||
156 | if (!piar) | 152 | if (!piar) |
157 | return NULL; | 153 | return NULL; |
158 | 154 | ||
159 | pci_dev_get(dev); | ||
160 | piar->addr_lo = alo; | 155 | piar->addr_lo = alo; |
161 | piar->addr_hi = ahi; | 156 | piar->addr_hi = ahi; |
162 | piar->edev = pci_dev_to_eeh_dev(dev); | 157 | piar->edev = pci_dev_to_eeh_dev(dev); |
@@ -250,7 +245,6 @@ restart: | |||
250 | 245 | ||
251 | if (piar->pcidev == dev) { | 246 | if (piar->pcidev == dev) { |
252 | rb_erase(n, &pci_io_addr_cache_root.rb_root); | 247 | rb_erase(n, &pci_io_addr_cache_root.rb_root); |
253 | pci_dev_put(piar->pcidev); | ||
254 | kfree(piar); | 248 | kfree(piar); |
255 | goto restart; | 249 | goto restart; |
256 | } | 250 | } |
@@ -302,12 +296,10 @@ void eeh_addr_cache_build(void) | |||
302 | if (!edev) | 296 | if (!edev) |
303 | continue; | 297 | continue; |
304 | 298 | ||
305 | pci_dev_get(dev); /* matching put is in eeh_remove_device() */ | ||
306 | dev->dev.archdata.edev = edev; | 299 | dev->dev.archdata.edev = edev; |
307 | edev->pdev = dev; | 300 | edev->pdev = dev; |
308 | 301 | ||
309 | eeh_addr_cache_insert_dev(dev); | 302 | eeh_addr_cache_insert_dev(dev); |
310 | |||
311 | eeh_sysfs_add_device(dev); | 303 | eeh_sysfs_add_device(dev); |
312 | } | 304 | } |
313 | 305 | ||
diff --git a/arch/powerpc/kernel/eeh_driver.c b/arch/powerpc/kernel/eeh_driver.c index 2b1ce17cae50..36bed5a12750 100644 --- a/arch/powerpc/kernel/eeh_driver.c +++ b/arch/powerpc/kernel/eeh_driver.c | |||
@@ -143,10 +143,14 @@ static void eeh_disable_irq(struct pci_dev *dev) | |||
143 | static void eeh_enable_irq(struct pci_dev *dev) | 143 | static void eeh_enable_irq(struct pci_dev *dev) |
144 | { | 144 | { |
145 | struct eeh_dev *edev = pci_dev_to_eeh_dev(dev); | 145 | struct eeh_dev *edev = pci_dev_to_eeh_dev(dev); |
146 | struct irq_desc *desc; | ||
146 | 147 | ||
147 | if ((edev->mode) & EEH_DEV_IRQ_DISABLED) { | 148 | if ((edev->mode) & EEH_DEV_IRQ_DISABLED) { |
148 | edev->mode &= ~EEH_DEV_IRQ_DISABLED; | 149 | edev->mode &= ~EEH_DEV_IRQ_DISABLED; |
149 | enable_irq(dev->irq); | 150 | |
151 | desc = irq_to_desc(dev->irq); | ||
152 | if (desc && desc->depth > 0) | ||
153 | enable_irq(dev->irq); | ||
150 | } | 154 | } |
151 | } | 155 | } |
152 | 156 | ||
@@ -338,6 +342,54 @@ static void *eeh_report_failure(void *data, void *userdata) | |||
338 | return NULL; | 342 | return NULL; |
339 | } | 343 | } |
340 | 344 | ||
345 | static void *eeh_rmv_device(void *data, void *userdata) | ||
346 | { | ||
347 | struct pci_driver *driver; | ||
348 | struct eeh_dev *edev = (struct eeh_dev *)data; | ||
349 | struct pci_dev *dev = eeh_dev_to_pci_dev(edev); | ||
350 | int *removed = (int *)userdata; | ||
351 | |||
352 | /* | ||
353 | * Actually, we should remove the PCI bridges as well. | ||
354 | * However, that's lots of complexity to do that, | ||
355 | * particularly some of devices under the bridge might | ||
356 | * support EEH. So we just care about PCI devices for | ||
357 | * simplicity here. | ||
358 | */ | ||
359 | if (!dev || (dev->hdr_type & PCI_HEADER_TYPE_BRIDGE)) | ||
360 | return NULL; | ||
361 | driver = eeh_pcid_get(dev); | ||
362 | if (driver && driver->err_handler) | ||
363 | return NULL; | ||
364 | |||
365 | /* Remove it from PCI subsystem */ | ||
366 | pr_debug("EEH: Removing %s without EEH sensitive driver\n", | ||
367 | pci_name(dev)); | ||
368 | edev->bus = dev->bus; | ||
369 | edev->mode |= EEH_DEV_DISCONNECTED; | ||
370 | (*removed)++; | ||
371 | |||
372 | pci_stop_and_remove_bus_device(dev); | ||
373 | |||
374 | return NULL; | ||
375 | } | ||
376 | |||
377 | static void *eeh_pe_detach_dev(void *data, void *userdata) | ||
378 | { | ||
379 | struct eeh_pe *pe = (struct eeh_pe *)data; | ||
380 | struct eeh_dev *edev, *tmp; | ||
381 | |||
382 | eeh_pe_for_each_dev(pe, edev, tmp) { | ||
383 | if (!(edev->mode & EEH_DEV_DISCONNECTED)) | ||
384 | continue; | ||
385 | |||
386 | edev->mode &= ~(EEH_DEV_DISCONNECTED | EEH_DEV_IRQ_DISABLED); | ||
387 | eeh_rmv_from_parent_pe(edev); | ||
388 | } | ||
389 | |||
390 | return NULL; | ||
391 | } | ||
392 | |||
341 | /** | 393 | /** |
342 | * eeh_reset_device - Perform actual reset of a pci slot | 394 | * eeh_reset_device - Perform actual reset of a pci slot |
343 | * @pe: EEH PE | 395 | * @pe: EEH PE |
@@ -349,8 +401,9 @@ static void *eeh_report_failure(void *data, void *userdata) | |||
349 | */ | 401 | */ |
350 | static int eeh_reset_device(struct eeh_pe *pe, struct pci_bus *bus) | 402 | static int eeh_reset_device(struct eeh_pe *pe, struct pci_bus *bus) |
351 | { | 403 | { |
404 | struct pci_bus *frozen_bus = eeh_pe_bus_get(pe); | ||
352 | struct timeval tstamp; | 405 | struct timeval tstamp; |
353 | int cnt, rc; | 406 | int cnt, rc, removed = 0; |
354 | 407 | ||
355 | /* pcibios will clear the counter; save the value */ | 408 | /* pcibios will clear the counter; save the value */ |
356 | cnt = pe->freeze_count; | 409 | cnt = pe->freeze_count; |
@@ -362,8 +415,11 @@ static int eeh_reset_device(struct eeh_pe *pe, struct pci_bus *bus) | |||
362 | * devices are expected to be attached soon when calling | 415 | * devices are expected to be attached soon when calling |
363 | * into pcibios_add_pci_devices(). | 416 | * into pcibios_add_pci_devices(). |
364 | */ | 417 | */ |
418 | eeh_pe_state_mark(pe, EEH_PE_KEEP); | ||
365 | if (bus) | 419 | if (bus) |
366 | __pcibios_remove_pci_devices(bus, 0); | 420 | pcibios_remove_pci_devices(bus); |
421 | else if (frozen_bus) | ||
422 | eeh_pe_dev_traverse(pe, eeh_rmv_device, &removed); | ||
367 | 423 | ||
368 | /* Reset the pci controller. (Asserts RST#; resets config space). | 424 | /* Reset the pci controller. (Asserts RST#; resets config space). |
369 | * Reconfigure bridges and devices. Don't try to bring the system | 425 | * Reconfigure bridges and devices. Don't try to bring the system |
@@ -384,9 +440,24 @@ static int eeh_reset_device(struct eeh_pe *pe, struct pci_bus *bus) | |||
384 | * potentially weird things happen. | 440 | * potentially weird things happen. |
385 | */ | 441 | */ |
386 | if (bus) { | 442 | if (bus) { |
443 | pr_info("EEH: Sleep 5s ahead of complete hotplug\n"); | ||
387 | ssleep(5); | 444 | ssleep(5); |
445 | |||
446 | /* | ||
447 | * The EEH device is still connected with its parent | ||
448 | * PE. We should disconnect it so the binding can be | ||
449 | * rebuilt when adding PCI devices. | ||
450 | */ | ||
451 | eeh_pe_traverse(pe, eeh_pe_detach_dev, NULL); | ||
388 | pcibios_add_pci_devices(bus); | 452 | pcibios_add_pci_devices(bus); |
453 | } else if (frozen_bus && removed) { | ||
454 | pr_info("EEH: Sleep 5s ahead of partial hotplug\n"); | ||
455 | ssleep(5); | ||
456 | |||
457 | eeh_pe_traverse(pe, eeh_pe_detach_dev, NULL); | ||
458 | pcibios_add_pci_devices(frozen_bus); | ||
389 | } | 459 | } |
460 | eeh_pe_state_clear(pe, EEH_PE_KEEP); | ||
390 | 461 | ||
391 | pe->tstamp = tstamp; | 462 | pe->tstamp = tstamp; |
392 | pe->freeze_count = cnt; | 463 | pe->freeze_count = cnt; |
diff --git a/arch/powerpc/kernel/eeh_pe.c b/arch/powerpc/kernel/eeh_pe.c index 016588a6f5ed..f9450537e335 100644 --- a/arch/powerpc/kernel/eeh_pe.c +++ b/arch/powerpc/kernel/eeh_pe.c | |||
@@ -149,8 +149,8 @@ static struct eeh_pe *eeh_pe_next(struct eeh_pe *pe, | |||
149 | * callback returns something other than NULL, or no more PEs | 149 | * callback returns something other than NULL, or no more PEs |
150 | * to be traversed. | 150 | * to be traversed. |
151 | */ | 151 | */ |
152 | static void *eeh_pe_traverse(struct eeh_pe *root, | 152 | void *eeh_pe_traverse(struct eeh_pe *root, |
153 | eeh_traverse_func fn, void *flag) | 153 | eeh_traverse_func fn, void *flag) |
154 | { | 154 | { |
155 | struct eeh_pe *pe; | 155 | struct eeh_pe *pe; |
156 | void *ret; | 156 | void *ret; |
@@ -176,7 +176,7 @@ void *eeh_pe_dev_traverse(struct eeh_pe *root, | |||
176 | eeh_traverse_func fn, void *flag) | 176 | eeh_traverse_func fn, void *flag) |
177 | { | 177 | { |
178 | struct eeh_pe *pe; | 178 | struct eeh_pe *pe; |
179 | struct eeh_dev *edev; | 179 | struct eeh_dev *edev, *tmp; |
180 | void *ret; | 180 | void *ret; |
181 | 181 | ||
182 | if (!root) { | 182 | if (!root) { |
@@ -186,7 +186,7 @@ void *eeh_pe_dev_traverse(struct eeh_pe *root, | |||
186 | 186 | ||
187 | /* Traverse root PE */ | 187 | /* Traverse root PE */ |
188 | for (pe = root; pe; pe = eeh_pe_next(pe, root)) { | 188 | for (pe = root; pe; pe = eeh_pe_next(pe, root)) { |
189 | eeh_pe_for_each_dev(pe, edev) { | 189 | eeh_pe_for_each_dev(pe, edev, tmp) { |
190 | ret = fn(edev, flag); | 190 | ret = fn(edev, flag); |
191 | if (ret) | 191 | if (ret) |
192 | return ret; | 192 | return ret; |
@@ -333,7 +333,7 @@ int eeh_add_to_parent_pe(struct eeh_dev *edev) | |||
333 | while (parent) { | 333 | while (parent) { |
334 | if (!(parent->type & EEH_PE_INVALID)) | 334 | if (!(parent->type & EEH_PE_INVALID)) |
335 | break; | 335 | break; |
336 | parent->type &= ~EEH_PE_INVALID; | 336 | parent->type &= ~(EEH_PE_INVALID | EEH_PE_KEEP); |
337 | parent = parent->parent; | 337 | parent = parent->parent; |
338 | } | 338 | } |
339 | pr_debug("EEH: Add %s to Device PE#%x, Parent PE#%x\n", | 339 | pr_debug("EEH: Add %s to Device PE#%x, Parent PE#%x\n", |
@@ -397,21 +397,20 @@ int eeh_add_to_parent_pe(struct eeh_dev *edev) | |||
397 | /** | 397 | /** |
398 | * eeh_rmv_from_parent_pe - Remove one EEH device from the associated PE | 398 | * eeh_rmv_from_parent_pe - Remove one EEH device from the associated PE |
399 | * @edev: EEH device | 399 | * @edev: EEH device |
400 | * @purge_pe: remove PE or not | ||
401 | * | 400 | * |
402 | * The PE hierarchy tree might be changed when doing PCI hotplug. | 401 | * The PE hierarchy tree might be changed when doing PCI hotplug. |
403 | * Also, the PCI devices or buses could be removed from the system | 402 | * Also, the PCI devices or buses could be removed from the system |
404 | * during EEH recovery. So we have to call the function remove the | 403 | * during EEH recovery. So we have to call the function remove the |
405 | * corresponding PE accordingly if necessary. | 404 | * corresponding PE accordingly if necessary. |
406 | */ | 405 | */ |
407 | int eeh_rmv_from_parent_pe(struct eeh_dev *edev, int purge_pe) | 406 | int eeh_rmv_from_parent_pe(struct eeh_dev *edev) |
408 | { | 407 | { |
409 | struct eeh_pe *pe, *parent, *child; | 408 | struct eeh_pe *pe, *parent, *child; |
410 | int cnt; | 409 | int cnt; |
411 | 410 | ||
412 | if (!edev->pe) { | 411 | if (!edev->pe) { |
413 | pr_warning("%s: No PE found for EEH device %s\n", | 412 | pr_debug("%s: No PE found for EEH device %s\n", |
414 | __func__, edev->dn->full_name); | 413 | __func__, edev->dn->full_name); |
415 | return -EEXIST; | 414 | return -EEXIST; |
416 | } | 415 | } |
417 | 416 | ||
@@ -431,7 +430,7 @@ int eeh_rmv_from_parent_pe(struct eeh_dev *edev, int purge_pe) | |||
431 | if (pe->type & EEH_PE_PHB) | 430 | if (pe->type & EEH_PE_PHB) |
432 | break; | 431 | break; |
433 | 432 | ||
434 | if (purge_pe) { | 433 | if (!(pe->state & EEH_PE_KEEP)) { |
435 | if (list_empty(&pe->edevs) && | 434 | if (list_empty(&pe->edevs) && |
436 | list_empty(&pe->child_list)) { | 435 | list_empty(&pe->child_list)) { |
437 | list_del(&pe->child); | 436 | list_del(&pe->child); |
@@ -502,7 +501,7 @@ static void *__eeh_pe_state_mark(void *data, void *flag) | |||
502 | { | 501 | { |
503 | struct eeh_pe *pe = (struct eeh_pe *)data; | 502 | struct eeh_pe *pe = (struct eeh_pe *)data; |
504 | int state = *((int *)flag); | 503 | int state = *((int *)flag); |
505 | struct eeh_dev *tmp; | 504 | struct eeh_dev *edev, *tmp; |
506 | struct pci_dev *pdev; | 505 | struct pci_dev *pdev; |
507 | 506 | ||
508 | /* | 507 | /* |
@@ -512,8 +511,8 @@ static void *__eeh_pe_state_mark(void *data, void *flag) | |||
512 | * the PCI device driver. | 511 | * the PCI device driver. |
513 | */ | 512 | */ |
514 | pe->state |= state; | 513 | pe->state |= state; |
515 | eeh_pe_for_each_dev(pe, tmp) { | 514 | eeh_pe_for_each_dev(pe, edev, tmp) { |
516 | pdev = eeh_dev_to_pci_dev(tmp); | 515 | pdev = eeh_dev_to_pci_dev(edev); |
517 | if (pdev) | 516 | if (pdev) |
518 | pdev->error_state = pci_channel_io_frozen; | 517 | pdev->error_state = pci_channel_io_frozen; |
519 | } | 518 | } |
@@ -579,7 +578,7 @@ void eeh_pe_state_clear(struct eeh_pe *pe, int state) | |||
579 | * blocked on normal path during the stage. So we need utilize | 578 | * blocked on normal path during the stage. So we need utilize |
580 | * eeh operations, which is always permitted. | 579 | * eeh operations, which is always permitted. |
581 | */ | 580 | */ |
582 | static void eeh_bridge_check_link(struct pci_dev *pdev, | 581 | static void eeh_bridge_check_link(struct eeh_dev *edev, |
583 | struct device_node *dn) | 582 | struct device_node *dn) |
584 | { | 583 | { |
585 | int cap; | 584 | int cap; |
@@ -590,16 +589,17 @@ static void eeh_bridge_check_link(struct pci_dev *pdev, | |||
590 | * We only check root port and downstream ports of | 589 | * We only check root port and downstream ports of |
591 | * PCIe switches | 590 | * PCIe switches |
592 | */ | 591 | */ |
593 | if (!pci_is_pcie(pdev) || | 592 | if (!(edev->mode & (EEH_DEV_ROOT_PORT | EEH_DEV_DS_PORT))) |
594 | (pci_pcie_type(pdev) != PCI_EXP_TYPE_ROOT_PORT && | ||
595 | pci_pcie_type(pdev) != PCI_EXP_TYPE_DOWNSTREAM)) | ||
596 | return; | 593 | return; |
597 | 594 | ||
598 | pr_debug("%s: Check PCIe link for %s ...\n", | 595 | pr_debug("%s: Check PCIe link for %04x:%02x:%02x.%01x ...\n", |
599 | __func__, pci_name(pdev)); | 596 | __func__, edev->phb->global_number, |
597 | edev->config_addr >> 8, | ||
598 | PCI_SLOT(edev->config_addr & 0xFF), | ||
599 | PCI_FUNC(edev->config_addr & 0xFF)); | ||
600 | 600 | ||
601 | /* Check slot status */ | 601 | /* Check slot status */ |
602 | cap = pdev->pcie_cap; | 602 | cap = edev->pcie_cap; |
603 | eeh_ops->read_config(dn, cap + PCI_EXP_SLTSTA, 2, &val); | 603 | eeh_ops->read_config(dn, cap + PCI_EXP_SLTSTA, 2, &val); |
604 | if (!(val & PCI_EXP_SLTSTA_PDS)) { | 604 | if (!(val & PCI_EXP_SLTSTA_PDS)) { |
605 | pr_debug(" No card in the slot (0x%04x) !\n", val); | 605 | pr_debug(" No card in the slot (0x%04x) !\n", val); |
@@ -653,8 +653,7 @@ static void eeh_bridge_check_link(struct pci_dev *pdev, | |||
653 | #define BYTE_SWAP(OFF) (8*((OFF)/4)+3-(OFF)) | 653 | #define BYTE_SWAP(OFF) (8*((OFF)/4)+3-(OFF)) |
654 | #define SAVED_BYTE(OFF) (((u8 *)(edev->config_space))[BYTE_SWAP(OFF)]) | 654 | #define SAVED_BYTE(OFF) (((u8 *)(edev->config_space))[BYTE_SWAP(OFF)]) |
655 | 655 | ||
656 | static void eeh_restore_bridge_bars(struct pci_dev *pdev, | 656 | static void eeh_restore_bridge_bars(struct eeh_dev *edev, |
657 | struct eeh_dev *edev, | ||
658 | struct device_node *dn) | 657 | struct device_node *dn) |
659 | { | 658 | { |
660 | int i; | 659 | int i; |
@@ -680,7 +679,7 @@ static void eeh_restore_bridge_bars(struct pci_dev *pdev, | |||
680 | eeh_ops->write_config(dn, PCI_COMMAND, 4, edev->config_space[1]); | 679 | eeh_ops->write_config(dn, PCI_COMMAND, 4, edev->config_space[1]); |
681 | 680 | ||
682 | /* Check the PCIe link is ready */ | 681 | /* Check the PCIe link is ready */ |
683 | eeh_bridge_check_link(pdev, dn); | 682 | eeh_bridge_check_link(edev, dn); |
684 | } | 683 | } |
685 | 684 | ||
686 | static void eeh_restore_device_bars(struct eeh_dev *edev, | 685 | static void eeh_restore_device_bars(struct eeh_dev *edev, |
@@ -729,19 +728,12 @@ static void eeh_restore_device_bars(struct eeh_dev *edev, | |||
729 | */ | 728 | */ |
730 | static void *eeh_restore_one_device_bars(void *data, void *flag) | 729 | static void *eeh_restore_one_device_bars(void *data, void *flag) |
731 | { | 730 | { |
732 | struct pci_dev *pdev = NULL; | ||
733 | struct eeh_dev *edev = (struct eeh_dev *)data; | 731 | struct eeh_dev *edev = (struct eeh_dev *)data; |
734 | struct device_node *dn = eeh_dev_to_of_node(edev); | 732 | struct device_node *dn = eeh_dev_to_of_node(edev); |
735 | 733 | ||
736 | /* Trace the PCI bridge */ | 734 | /* Do special restore for bridges */ |
737 | if (eeh_probe_mode_dev()) { | 735 | if (edev->mode & EEH_DEV_BRIDGE) |
738 | pdev = eeh_dev_to_pci_dev(edev); | 736 | eeh_restore_bridge_bars(edev, dn); |
739 | if (pdev->hdr_type != PCI_HEADER_TYPE_BRIDGE) | ||
740 | pdev = NULL; | ||
741 | } | ||
742 | |||
743 | if (pdev) | ||
744 | eeh_restore_bridge_bars(pdev, edev, dn); | ||
745 | else | 737 | else |
746 | eeh_restore_device_bars(edev, dn); | 738 | eeh_restore_device_bars(edev, dn); |
747 | 739 | ||
diff --git a/arch/powerpc/kernel/eeh_sysfs.c b/arch/powerpc/kernel/eeh_sysfs.c index e7ae3484918c..5d753d4f2c75 100644 --- a/arch/powerpc/kernel/eeh_sysfs.c +++ b/arch/powerpc/kernel/eeh_sysfs.c | |||
@@ -56,19 +56,40 @@ EEH_SHOW_ATTR(eeh_pe_config_addr, pe_config_addr, "0x%x"); | |||
56 | 56 | ||
57 | void eeh_sysfs_add_device(struct pci_dev *pdev) | 57 | void eeh_sysfs_add_device(struct pci_dev *pdev) |
58 | { | 58 | { |
59 | struct eeh_dev *edev = pci_dev_to_eeh_dev(pdev); | ||
59 | int rc=0; | 60 | int rc=0; |
60 | 61 | ||
62 | if (edev && (edev->mode & EEH_DEV_SYSFS)) | ||
63 | return; | ||
64 | |||
61 | rc += device_create_file(&pdev->dev, &dev_attr_eeh_mode); | 65 | rc += device_create_file(&pdev->dev, &dev_attr_eeh_mode); |
62 | rc += device_create_file(&pdev->dev, &dev_attr_eeh_config_addr); | 66 | rc += device_create_file(&pdev->dev, &dev_attr_eeh_config_addr); |
63 | rc += device_create_file(&pdev->dev, &dev_attr_eeh_pe_config_addr); | 67 | rc += device_create_file(&pdev->dev, &dev_attr_eeh_pe_config_addr); |
64 | 68 | ||
65 | if (rc) | 69 | if (rc) |
66 | printk(KERN_WARNING "EEH: Unable to create sysfs entries\n"); | 70 | printk(KERN_WARNING "EEH: Unable to create sysfs entries\n"); |
71 | else if (edev) | ||
72 | edev->mode |= EEH_DEV_SYSFS; | ||
67 | } | 73 | } |
68 | 74 | ||
69 | void eeh_sysfs_remove_device(struct pci_dev *pdev) | 75 | void eeh_sysfs_remove_device(struct pci_dev *pdev) |
70 | { | 76 | { |
77 | struct eeh_dev *edev = pci_dev_to_eeh_dev(pdev); | ||
78 | |||
79 | /* | ||
80 | * The parent directory might have been removed. We needn't | ||
81 | * continue for that case. | ||
82 | */ | ||
83 | if (!pdev->dev.kobj.sd) { | ||
84 | if (edev) | ||
85 | edev->mode &= ~EEH_DEV_SYSFS; | ||
86 | return; | ||
87 | } | ||
88 | |||
71 | device_remove_file(&pdev->dev, &dev_attr_eeh_mode); | 89 | device_remove_file(&pdev->dev, &dev_attr_eeh_mode); |
72 | device_remove_file(&pdev->dev, &dev_attr_eeh_config_addr); | 90 | device_remove_file(&pdev->dev, &dev_attr_eeh_config_addr); |
73 | device_remove_file(&pdev->dev, &dev_attr_eeh_pe_config_addr); | 91 | device_remove_file(&pdev->dev, &dev_attr_eeh_pe_config_addr); |
92 | |||
93 | if (edev) | ||
94 | edev->mode &= ~EEH_DEV_SYSFS; | ||
74 | } | 95 | } |
diff --git a/arch/powerpc/kernel/pci-common.c b/arch/powerpc/kernel/pci-common.c index f46914a0f33e..7d22a675fe1a 100644 --- a/arch/powerpc/kernel/pci-common.c +++ b/arch/powerpc/kernel/pci-common.c | |||
@@ -1462,6 +1462,8 @@ void pcibios_finish_adding_to_bus(struct pci_bus *bus) | |||
1462 | /* Allocate bus and devices resources */ | 1462 | /* Allocate bus and devices resources */ |
1463 | pcibios_allocate_bus_resources(bus); | 1463 | pcibios_allocate_bus_resources(bus); |
1464 | pcibios_claim_one_bus(bus); | 1464 | pcibios_claim_one_bus(bus); |
1465 | if (!pci_has_flag(PCI_PROBE_ONLY)) | ||
1466 | pci_assign_unassigned_bus_resources(bus); | ||
1465 | 1467 | ||
1466 | /* Fixup EEH */ | 1468 | /* Fixup EEH */ |
1467 | eeh_add_device_tree_late(bus); | 1469 | eeh_add_device_tree_late(bus); |
diff --git a/arch/powerpc/kernel/pci-hotplug.c b/arch/powerpc/kernel/pci-hotplug.c index 3f608800c06b..c1e17ae68a08 100644 --- a/arch/powerpc/kernel/pci-hotplug.c +++ b/arch/powerpc/kernel/pci-hotplug.c | |||
@@ -22,45 +22,40 @@ | |||
22 | #include <asm/eeh.h> | 22 | #include <asm/eeh.h> |
23 | 23 | ||
24 | /** | 24 | /** |
25 | * __pcibios_remove_pci_devices - remove all devices under this bus | 25 | * pcibios_release_device - release PCI device |
26 | * @dev: PCI device | ||
27 | * | ||
28 | * The function is called before releasing the indicated PCI device. | ||
29 | */ | ||
30 | void pcibios_release_device(struct pci_dev *dev) | ||
31 | { | ||
32 | eeh_remove_device(dev); | ||
33 | } | ||
34 | |||
35 | /** | ||
36 | * pcibios_remove_pci_devices - remove all devices under this bus | ||
26 | * @bus: the indicated PCI bus | 37 | * @bus: the indicated PCI bus |
27 | * @purge_pe: destroy the PE on removal of PCI devices | ||
28 | * | 38 | * |
29 | * Remove all of the PCI devices under this bus both from the | 39 | * Remove all of the PCI devices under this bus both from the |
30 | * linux pci device tree, and from the powerpc EEH address cache. | 40 | * linux pci device tree, and from the powerpc EEH address cache. |
31 | * By default, the corresponding PE will be destroied during the | ||
32 | * normal PCI hotplug path. For PCI hotplug during EEH recovery, | ||
33 | * the corresponding PE won't be destroied and deallocated. | ||
34 | */ | 41 | */ |
35 | void __pcibios_remove_pci_devices(struct pci_bus *bus, int purge_pe) | 42 | void pcibios_remove_pci_devices(struct pci_bus *bus) |
36 | { | 43 | { |
37 | struct pci_dev *dev, *tmp; | 44 | struct pci_dev *dev, *tmp; |
38 | struct pci_bus *child_bus; | 45 | struct pci_bus *child_bus; |
39 | 46 | ||
40 | /* First go down child busses */ | 47 | /* First go down child busses */ |
41 | list_for_each_entry(child_bus, &bus->children, node) | 48 | list_for_each_entry(child_bus, &bus->children, node) |
42 | __pcibios_remove_pci_devices(child_bus, purge_pe); | 49 | pcibios_remove_pci_devices(child_bus); |
43 | 50 | ||
44 | pr_debug("PCI: Removing devices on bus %04x:%02x\n", | 51 | pr_debug("PCI: Removing devices on bus %04x:%02x\n", |
45 | pci_domain_nr(bus), bus->number); | 52 | pci_domain_nr(bus), bus->number); |
46 | list_for_each_entry_safe(dev, tmp, &bus->devices, bus_list) { | 53 | list_for_each_entry_safe(dev, tmp, &bus->devices, bus_list) { |
47 | pr_debug(" * Removing %s...\n", pci_name(dev)); | 54 | pr_debug(" Removing %s...\n", pci_name(dev)); |
48 | eeh_remove_bus_device(dev, purge_pe); | ||
49 | pci_stop_and_remove_bus_device(dev); | 55 | pci_stop_and_remove_bus_device(dev); |
50 | } | 56 | } |
51 | } | 57 | } |
52 | 58 | ||
53 | /** | ||
54 | * pcibios_remove_pci_devices - remove all devices under this bus | ||
55 | * @bus: the indicated PCI bus | ||
56 | * | ||
57 | * Remove all of the PCI devices under this bus both from the | ||
58 | * linux pci device tree, and from the powerpc EEH address cache. | ||
59 | */ | ||
60 | void pcibios_remove_pci_devices(struct pci_bus *bus) | ||
61 | { | ||
62 | __pcibios_remove_pci_devices(bus, 1); | ||
63 | } | ||
64 | EXPORT_SYMBOL_GPL(pcibios_remove_pci_devices); | 59 | EXPORT_SYMBOL_GPL(pcibios_remove_pci_devices); |
65 | 60 | ||
66 | /** | 61 | /** |
@@ -76,7 +71,7 @@ EXPORT_SYMBOL_GPL(pcibios_remove_pci_devices); | |||
76 | */ | 71 | */ |
77 | void pcibios_add_pci_devices(struct pci_bus * bus) | 72 | void pcibios_add_pci_devices(struct pci_bus * bus) |
78 | { | 73 | { |
79 | int slotno, num, mode, pass, max; | 74 | int slotno, mode, pass, max; |
80 | struct pci_dev *dev; | 75 | struct pci_dev *dev; |
81 | struct device_node *dn = pci_bus_to_OF_node(bus); | 76 | struct device_node *dn = pci_bus_to_OF_node(bus); |
82 | 77 | ||
@@ -90,11 +85,15 @@ void pcibios_add_pci_devices(struct pci_bus * bus) | |||
90 | /* use ofdt-based probe */ | 85 | /* use ofdt-based probe */ |
91 | of_rescan_bus(dn, bus); | 86 | of_rescan_bus(dn, bus); |
92 | } else if (mode == PCI_PROBE_NORMAL) { | 87 | } else if (mode == PCI_PROBE_NORMAL) { |
93 | /* use legacy probe */ | 88 | /* |
89 | * Use legacy probe. In the partial hotplug case, we | ||
90 | * probably have grandchildren devices unplugged. So | ||
91 | * we don't check the return value from pci_scan_slot() in | ||
92 | * order for fully rescan all the way down to pick them up. | ||
93 | * They can have been removed during partial hotplug. | ||
94 | */ | ||
94 | slotno = PCI_SLOT(PCI_DN(dn->child)->devfn); | 95 | slotno = PCI_SLOT(PCI_DN(dn->child)->devfn); |
95 | num = pci_scan_slot(bus, PCI_DEVFN(slotno, 0)); | 96 | pci_scan_slot(bus, PCI_DEVFN(slotno, 0)); |
96 | if (!num) | ||
97 | return; | ||
98 | pcibios_setup_bus_devices(bus); | 97 | pcibios_setup_bus_devices(bus); |
99 | max = bus->busn_res.start; | 98 | max = bus->busn_res.start; |
100 | for (pass = 0; pass < 2; pass++) { | 99 | for (pass = 0; pass < 2; pass++) { |
diff --git a/arch/powerpc/kernel/pci_of_scan.c b/arch/powerpc/kernel/pci_of_scan.c index 6b0ba5854d99..15d9105323bf 100644 --- a/arch/powerpc/kernel/pci_of_scan.c +++ b/arch/powerpc/kernel/pci_of_scan.c | |||
@@ -230,11 +230,14 @@ void of_scan_pci_bridge(struct pci_dev *dev) | |||
230 | return; | 230 | return; |
231 | } | 231 | } |
232 | 232 | ||
233 | bus = pci_add_new_bus(dev->bus, dev, busrange[0]); | 233 | bus = pci_find_bus(pci_domain_nr(dev->bus), busrange[0]); |
234 | if (!bus) { | 234 | if (!bus) { |
235 | printk(KERN_ERR "Failed to create pci bus for %s\n", | 235 | bus = pci_add_new_bus(dev->bus, dev, busrange[0]); |
236 | node->full_name); | 236 | if (!bus) { |
237 | return; | 237 | printk(KERN_ERR "Failed to create pci bus for %s\n", |
238 | node->full_name); | ||
239 | return; | ||
240 | } | ||
238 | } | 241 | } |
239 | 242 | ||
240 | bus->primary = dev->bus->number; | 243 | bus->primary = dev->bus->number; |
@@ -292,6 +295,38 @@ void of_scan_pci_bridge(struct pci_dev *dev) | |||
292 | } | 295 | } |
293 | EXPORT_SYMBOL(of_scan_pci_bridge); | 296 | EXPORT_SYMBOL(of_scan_pci_bridge); |
294 | 297 | ||
298 | static struct pci_dev *of_scan_pci_dev(struct pci_bus *bus, | ||
299 | struct device_node *dn) | ||
300 | { | ||
301 | struct pci_dev *dev = NULL; | ||
302 | const u32 *reg; | ||
303 | int reglen, devfn; | ||
304 | |||
305 | pr_debug(" * %s\n", dn->full_name); | ||
306 | if (!of_device_is_available(dn)) | ||
307 | return NULL; | ||
308 | |||
309 | reg = of_get_property(dn, "reg", ®len); | ||
310 | if (reg == NULL || reglen < 20) | ||
311 | return NULL; | ||
312 | devfn = (reg[0] >> 8) & 0xff; | ||
313 | |||
314 | /* Check if the PCI device is already there */ | ||
315 | dev = pci_get_slot(bus, devfn); | ||
316 | if (dev) { | ||
317 | pci_dev_put(dev); | ||
318 | return dev; | ||
319 | } | ||
320 | |||
321 | /* create a new pci_dev for this device */ | ||
322 | dev = of_create_pci_dev(dn, bus, devfn); | ||
323 | if (!dev) | ||
324 | return NULL; | ||
325 | |||
326 | pr_debug(" dev header type: %x\n", dev->hdr_type); | ||
327 | return dev; | ||
328 | } | ||
329 | |||
295 | /** | 330 | /** |
296 | * __of_scan_bus - given a PCI bus node, setup bus and scan for child devices | 331 | * __of_scan_bus - given a PCI bus node, setup bus and scan for child devices |
297 | * @node: device tree node for the PCI bus | 332 | * @node: device tree node for the PCI bus |
@@ -302,8 +337,6 @@ static void __of_scan_bus(struct device_node *node, struct pci_bus *bus, | |||
302 | int rescan_existing) | 337 | int rescan_existing) |
303 | { | 338 | { |
304 | struct device_node *child; | 339 | struct device_node *child; |
305 | const u32 *reg; | ||
306 | int reglen, devfn; | ||
307 | struct pci_dev *dev; | 340 | struct pci_dev *dev; |
308 | 341 | ||
309 | pr_debug("of_scan_bus(%s) bus no %d...\n", | 342 | pr_debug("of_scan_bus(%s) bus no %d...\n", |
@@ -311,16 +344,7 @@ static void __of_scan_bus(struct device_node *node, struct pci_bus *bus, | |||
311 | 344 | ||
312 | /* Scan direct children */ | 345 | /* Scan direct children */ |
313 | for_each_child_of_node(node, child) { | 346 | for_each_child_of_node(node, child) { |
314 | pr_debug(" * %s\n", child->full_name); | 347 | dev = of_scan_pci_dev(bus, child); |
315 | if (!of_device_is_available(child)) | ||
316 | continue; | ||
317 | reg = of_get_property(child, "reg", ®len); | ||
318 | if (reg == NULL || reglen < 20) | ||
319 | continue; | ||
320 | devfn = (reg[0] >> 8) & 0xff; | ||
321 | |||
322 | /* create a new pci_dev for this device */ | ||
323 | dev = of_create_pci_dev(child, bus, devfn); | ||
324 | if (!dev) | 348 | if (!dev) |
325 | continue; | 349 | continue; |
326 | pr_debug(" dev header type: %x\n", dev->hdr_type); | 350 | pr_debug(" dev header type: %x\n", dev->hdr_type); |
diff --git a/arch/powerpc/kernel/prom_init.c b/arch/powerpc/kernel/prom_init.c index 5eccda9fd33f..607902424e73 100644 --- a/arch/powerpc/kernel/prom_init.c +++ b/arch/powerpc/kernel/prom_init.c | |||
@@ -644,7 +644,8 @@ unsigned char ibm_architecture_vec[] = { | |||
644 | W(0xfffe0000), W(0x003a0000), /* POWER5/POWER5+ */ | 644 | W(0xfffe0000), W(0x003a0000), /* POWER5/POWER5+ */ |
645 | W(0xffff0000), W(0x003e0000), /* POWER6 */ | 645 | W(0xffff0000), W(0x003e0000), /* POWER6 */ |
646 | W(0xffff0000), W(0x003f0000), /* POWER7 */ | 646 | W(0xffff0000), W(0x003f0000), /* POWER7 */ |
647 | W(0xffff0000), W(0x004b0000), /* POWER8 */ | 647 | W(0xffff0000), W(0x004b0000), /* POWER8E */ |
648 | W(0xffff0000), W(0x004d0000), /* POWER8 */ | ||
648 | W(0xffffffff), W(0x0f000004), /* all 2.07-compliant */ | 649 | W(0xffffffff), W(0x0f000004), /* all 2.07-compliant */ |
649 | W(0xffffffff), W(0x0f000003), /* all 2.06-compliant */ | 650 | W(0xffffffff), W(0x0f000003), /* all 2.06-compliant */ |
650 | W(0xffffffff), W(0x0f000002), /* all 2.05-compliant */ | 651 | W(0xffffffff), W(0x0f000002), /* all 2.05-compliant */ |
@@ -706,7 +707,7 @@ unsigned char ibm_architecture_vec[] = { | |||
706 | * must match by the macro below. Update the definition if | 707 | * must match by the macro below. Update the definition if |
707 | * the structure layout changes. | 708 | * the structure layout changes. |
708 | */ | 709 | */ |
709 | #define IBM_ARCH_VEC_NRCORES_OFFSET 117 | 710 | #define IBM_ARCH_VEC_NRCORES_OFFSET 125 |
710 | W(NR_CPUS), /* number of cores supported */ | 711 | W(NR_CPUS), /* number of cores supported */ |
711 | 0, | 712 | 0, |
712 | 0, | 713 | 0, |
diff --git a/arch/powerpc/kernel/vmlinux.lds.S b/arch/powerpc/kernel/vmlinux.lds.S index 654e479802f2..f096e72262f4 100644 --- a/arch/powerpc/kernel/vmlinux.lds.S +++ b/arch/powerpc/kernel/vmlinux.lds.S | |||
@@ -38,9 +38,6 @@ jiffies = jiffies_64 + 4; | |||
38 | #endif | 38 | #endif |
39 | SECTIONS | 39 | SECTIONS |
40 | { | 40 | { |
41 | . = 0; | ||
42 | reloc_start = .; | ||
43 | |||
44 | . = KERNELBASE; | 41 | . = KERNELBASE; |
45 | 42 | ||
46 | /* | 43 | /* |
diff --git a/arch/powerpc/mm/hash_native_64.c b/arch/powerpc/mm/hash_native_64.c index 3f0c30ae4791..c33d939120c9 100644 --- a/arch/powerpc/mm/hash_native_64.c +++ b/arch/powerpc/mm/hash_native_64.c | |||
@@ -43,6 +43,7 @@ static inline void __tlbie(unsigned long vpn, int psize, int apsize, int ssize) | |||
43 | { | 43 | { |
44 | unsigned long va; | 44 | unsigned long va; |
45 | unsigned int penc; | 45 | unsigned int penc; |
46 | unsigned long sllp; | ||
46 | 47 | ||
47 | /* | 48 | /* |
48 | * We need 14 to 65 bits of va for a tlibe of 4K page | 49 | * We need 14 to 65 bits of va for a tlibe of 4K page |
@@ -64,7 +65,9 @@ static inline void __tlbie(unsigned long vpn, int psize, int apsize, int ssize) | |||
64 | /* clear out bits after (52) [0....52.....63] */ | 65 | /* clear out bits after (52) [0....52.....63] */ |
65 | va &= ~((1ul << (64 - 52)) - 1); | 66 | va &= ~((1ul << (64 - 52)) - 1); |
66 | va |= ssize << 8; | 67 | va |= ssize << 8; |
67 | va |= mmu_psize_defs[apsize].sllp << 6; | 68 | sllp = ((mmu_psize_defs[apsize].sllp & SLB_VSID_L) >> 6) | |
69 | ((mmu_psize_defs[apsize].sllp & SLB_VSID_LP) >> 4); | ||
70 | va |= sllp << 5; | ||
68 | asm volatile(ASM_FTR_IFCLR("tlbie %0,0", PPC_TLBIE(%1,%0), %2) | 71 | asm volatile(ASM_FTR_IFCLR("tlbie %0,0", PPC_TLBIE(%1,%0), %2) |
69 | : : "r" (va), "r"(0), "i" (CPU_FTR_ARCH_206) | 72 | : : "r" (va), "r"(0), "i" (CPU_FTR_ARCH_206) |
70 | : "memory"); | 73 | : "memory"); |
@@ -98,6 +101,7 @@ static inline void __tlbiel(unsigned long vpn, int psize, int apsize, int ssize) | |||
98 | { | 101 | { |
99 | unsigned long va; | 102 | unsigned long va; |
100 | unsigned int penc; | 103 | unsigned int penc; |
104 | unsigned long sllp; | ||
101 | 105 | ||
102 | /* VPN_SHIFT can be atmost 12 */ | 106 | /* VPN_SHIFT can be atmost 12 */ |
103 | va = vpn << VPN_SHIFT; | 107 | va = vpn << VPN_SHIFT; |
@@ -113,7 +117,9 @@ static inline void __tlbiel(unsigned long vpn, int psize, int apsize, int ssize) | |||
113 | /* clear out bits after(52) [0....52.....63] */ | 117 | /* clear out bits after(52) [0....52.....63] */ |
114 | va &= ~((1ul << (64 - 52)) - 1); | 118 | va &= ~((1ul << (64 - 52)) - 1); |
115 | va |= ssize << 8; | 119 | va |= ssize << 8; |
116 | va |= mmu_psize_defs[apsize].sllp << 6; | 120 | sllp = ((mmu_psize_defs[apsize].sllp & SLB_VSID_L) >> 6) | |
121 | ((mmu_psize_defs[apsize].sllp & SLB_VSID_LP) >> 4); | ||
122 | va |= sllp << 5; | ||
117 | asm volatile(".long 0x7c000224 | (%0 << 11) | (0 << 21)" | 123 | asm volatile(".long 0x7c000224 | (%0 << 11) | (0 << 21)" |
118 | : : "r"(va) : "memory"); | 124 | : : "r"(va) : "memory"); |
119 | break; | 125 | break; |
@@ -554,6 +560,7 @@ static void hpte_decode(struct hash_pte *hpte, unsigned long slot, | |||
554 | seg_off |= vpi << shift; | 560 | seg_off |= vpi << shift; |
555 | } | 561 | } |
556 | *vpn = vsid << (SID_SHIFT - VPN_SHIFT) | seg_off >> VPN_SHIFT; | 562 | *vpn = vsid << (SID_SHIFT - VPN_SHIFT) | seg_off >> VPN_SHIFT; |
563 | break; | ||
557 | case MMU_SEGSIZE_1T: | 564 | case MMU_SEGSIZE_1T: |
558 | /* We only have 40 - 23 bits of seg_off in avpn */ | 565 | /* We only have 40 - 23 bits of seg_off in avpn */ |
559 | seg_off = (avpn & 0x1ffff) << 23; | 566 | seg_off = (avpn & 0x1ffff) << 23; |
@@ -563,6 +570,7 @@ static void hpte_decode(struct hash_pte *hpte, unsigned long slot, | |||
563 | seg_off |= vpi << shift; | 570 | seg_off |= vpi << shift; |
564 | } | 571 | } |
565 | *vpn = vsid << (SID_SHIFT_1T - VPN_SHIFT) | seg_off >> VPN_SHIFT; | 572 | *vpn = vsid << (SID_SHIFT_1T - VPN_SHIFT) | seg_off >> VPN_SHIFT; |
573 | break; | ||
566 | default: | 574 | default: |
567 | *vpn = size = 0; | 575 | *vpn = size = 0; |
568 | } | 576 | } |
diff --git a/arch/powerpc/perf/core-book3s.c b/arch/powerpc/perf/core-book3s.c index a3985aee77fe..24a45f91c65f 100644 --- a/arch/powerpc/perf/core-book3s.c +++ b/arch/powerpc/perf/core-book3s.c | |||
@@ -1252,8 +1252,11 @@ nocheck: | |||
1252 | 1252 | ||
1253 | ret = 0; | 1253 | ret = 0; |
1254 | out: | 1254 | out: |
1255 | if (has_branch_stack(event)) | 1255 | if (has_branch_stack(event)) { |
1256 | power_pmu_bhrb_enable(event); | 1256 | power_pmu_bhrb_enable(event); |
1257 | cpuhw->bhrb_filter = ppmu->bhrb_filter_map( | ||
1258 | event->attr.branch_sample_type); | ||
1259 | } | ||
1257 | 1260 | ||
1258 | perf_pmu_enable(event->pmu); | 1261 | perf_pmu_enable(event->pmu); |
1259 | local_irq_restore(flags); | 1262 | local_irq_restore(flags); |
diff --git a/arch/powerpc/perf/power8-pmu.c b/arch/powerpc/perf/power8-pmu.c index 96a64d6a8bdf..7466374d2787 100644 --- a/arch/powerpc/perf/power8-pmu.c +++ b/arch/powerpc/perf/power8-pmu.c | |||
@@ -561,18 +561,13 @@ static int power8_generic_events[] = { | |||
561 | static u64 power8_bhrb_filter_map(u64 branch_sample_type) | 561 | static u64 power8_bhrb_filter_map(u64 branch_sample_type) |
562 | { | 562 | { |
563 | u64 pmu_bhrb_filter = 0; | 563 | u64 pmu_bhrb_filter = 0; |
564 | u64 br_privilege = branch_sample_type & ONLY_PLM; | ||
565 | 564 | ||
566 | /* BHRB and regular PMU events share the same prvillege state | 565 | /* BHRB and regular PMU events share the same privilege state |
567 | * filter configuration. BHRB is always recorded along with a | 566 | * filter configuration. BHRB is always recorded along with a |
568 | * regular PMU event. So privilege state filter criteria for BHRB | 567 | * regular PMU event. As the privilege state filter is handled |
569 | * and the companion PMU events has to be the same. As a default | 568 | * in the basic PMC configuration of the accompanying regular |
570 | * "perf record" tool sets all privillege bits ON when no filter | 569 | * PMU event, we ignore any separate BHRB specific request. |
571 | * criteria is provided in the command line. So as along as all | ||
572 | * privillege bits are ON or they are OFF, we are good to go. | ||
573 | */ | 570 | */ |
574 | if ((br_privilege != 7) && (br_privilege != 0)) | ||
575 | return -1; | ||
576 | 571 | ||
577 | /* No branch filter requested */ | 572 | /* No branch filter requested */ |
578 | if (branch_sample_type & PERF_SAMPLE_BRANCH_ANY) | 573 | if (branch_sample_type & PERF_SAMPLE_BRANCH_ANY) |
@@ -621,10 +616,19 @@ static struct power_pmu power8_pmu = { | |||
621 | 616 | ||
622 | static int __init init_power8_pmu(void) | 617 | static int __init init_power8_pmu(void) |
623 | { | 618 | { |
619 | int rc; | ||
620 | |||
624 | if (!cur_cpu_spec->oprofile_cpu_type || | 621 | if (!cur_cpu_spec->oprofile_cpu_type || |
625 | strcmp(cur_cpu_spec->oprofile_cpu_type, "ppc64/power8")) | 622 | strcmp(cur_cpu_spec->oprofile_cpu_type, "ppc64/power8")) |
626 | return -ENODEV; | 623 | return -ENODEV; |
627 | 624 | ||
628 | return register_power_pmu(&power8_pmu); | 625 | rc = register_power_pmu(&power8_pmu); |
626 | if (rc) | ||
627 | return rc; | ||
628 | |||
629 | /* Tell userspace that EBB is supported */ | ||
630 | cur_cpu_spec->cpu_user_features2 |= PPC_FEATURE2_EBB; | ||
631 | |||
632 | return 0; | ||
629 | } | 633 | } |
630 | early_initcall(init_power8_pmu); | 634 | early_initcall(init_power8_pmu); |
diff --git a/arch/powerpc/platforms/powernv/eeh-powernv.c b/arch/powerpc/platforms/powernv/eeh-powernv.c index 969cce73055a..79663d26e6ea 100644 --- a/arch/powerpc/platforms/powernv/eeh-powernv.c +++ b/arch/powerpc/platforms/powernv/eeh-powernv.c | |||
@@ -114,7 +114,7 @@ static int powernv_eeh_dev_probe(struct pci_dev *dev, void *flag) | |||
114 | * the root bridge. So it's not reasonable to continue | 114 | * the root bridge. So it's not reasonable to continue |
115 | * the probing. | 115 | * the probing. |
116 | */ | 116 | */ |
117 | if (!dn || !edev) | 117 | if (!dn || !edev || edev->pe) |
118 | return 0; | 118 | return 0; |
119 | 119 | ||
120 | /* Skip for PCI-ISA bridge */ | 120 | /* Skip for PCI-ISA bridge */ |
@@ -122,8 +122,19 @@ static int powernv_eeh_dev_probe(struct pci_dev *dev, void *flag) | |||
122 | return 0; | 122 | return 0; |
123 | 123 | ||
124 | /* Initialize eeh device */ | 124 | /* Initialize eeh device */ |
125 | edev->class_code = dev->class; | 125 | edev->class_code = dev->class; |
126 | edev->mode = 0; | 126 | edev->mode &= 0xFFFFFF00; |
127 | if (dev->hdr_type == PCI_HEADER_TYPE_BRIDGE) | ||
128 | edev->mode |= EEH_DEV_BRIDGE; | ||
129 | if (pci_is_pcie(dev)) { | ||
130 | edev->pcie_cap = pci_pcie_cap(dev); | ||
131 | |||
132 | if (pci_pcie_type(dev) == PCI_EXP_TYPE_ROOT_PORT) | ||
133 | edev->mode |= EEH_DEV_ROOT_PORT; | ||
134 | else if (pci_pcie_type(dev) == PCI_EXP_TYPE_DOWNSTREAM) | ||
135 | edev->mode |= EEH_DEV_DS_PORT; | ||
136 | } | ||
137 | |||
127 | edev->config_addr = ((dev->bus->number << 8) | dev->devfn); | 138 | edev->config_addr = ((dev->bus->number << 8) | dev->devfn); |
128 | edev->pe_config_addr = phb->bdfn_to_pe(phb, dev->bus, dev->devfn & 0xff); | 139 | edev->pe_config_addr = phb->bdfn_to_pe(phb, dev->bus, dev->devfn & 0xff); |
129 | 140 | ||
diff --git a/arch/powerpc/platforms/powernv/pci-ioda.c b/arch/powerpc/platforms/powernv/pci-ioda.c index 49b57b9f835d..d8140b125e62 100644 --- a/arch/powerpc/platforms/powernv/pci-ioda.c +++ b/arch/powerpc/platforms/powernv/pci-ioda.c | |||
@@ -1266,7 +1266,7 @@ void __init pnv_pci_init_ioda_phb(struct device_node *np, | |||
1266 | opal_pci_set_pe(phb_id, 0, 0, 7, 1, 1 , OPAL_MAP_PE); | 1266 | opal_pci_set_pe(phb_id, 0, 0, 7, 1, 1 , OPAL_MAP_PE); |
1267 | } | 1267 | } |
1268 | 1268 | ||
1269 | void pnv_pci_init_ioda2_phb(struct device_node *np) | 1269 | void __init pnv_pci_init_ioda2_phb(struct device_node *np) |
1270 | { | 1270 | { |
1271 | pnv_pci_init_ioda_phb(np, 0, PNV_PHB_IODA2); | 1271 | pnv_pci_init_ioda_phb(np, 0, PNV_PHB_IODA2); |
1272 | } | 1272 | } |
diff --git a/arch/powerpc/platforms/pseries/Kconfig b/arch/powerpc/platforms/pseries/Kconfig index 1bd3399146ed..62b4f8025de0 100644 --- a/arch/powerpc/platforms/pseries/Kconfig +++ b/arch/powerpc/platforms/pseries/Kconfig | |||
@@ -19,7 +19,6 @@ config PPC_PSERIES | |||
19 | select ZLIB_DEFLATE | 19 | select ZLIB_DEFLATE |
20 | select PPC_DOORBELL | 20 | select PPC_DOORBELL |
21 | select HAVE_CONTEXT_TRACKING | 21 | select HAVE_CONTEXT_TRACKING |
22 | select HOTPLUG if SMP | ||
23 | select HOTPLUG_CPU if SMP | 22 | select HOTPLUG_CPU if SMP |
24 | default y | 23 | default y |
25 | 24 | ||
diff --git a/arch/powerpc/platforms/pseries/eeh_pseries.c b/arch/powerpc/platforms/pseries/eeh_pseries.c index b456b157d33d..7fbc25b1813f 100644 --- a/arch/powerpc/platforms/pseries/eeh_pseries.c +++ b/arch/powerpc/platforms/pseries/eeh_pseries.c | |||
@@ -133,6 +133,48 @@ static int pseries_eeh_init(void) | |||
133 | return 0; | 133 | return 0; |
134 | } | 134 | } |
135 | 135 | ||
136 | static int pseries_eeh_cap_start(struct device_node *dn) | ||
137 | { | ||
138 | struct pci_dn *pdn = PCI_DN(dn); | ||
139 | u32 status; | ||
140 | |||
141 | if (!pdn) | ||
142 | return 0; | ||
143 | |||
144 | rtas_read_config(pdn, PCI_STATUS, 2, &status); | ||
145 | if (!(status & PCI_STATUS_CAP_LIST)) | ||
146 | return 0; | ||
147 | |||
148 | return PCI_CAPABILITY_LIST; | ||
149 | } | ||
150 | |||
151 | |||
152 | static int pseries_eeh_find_cap(struct device_node *dn, int cap) | ||
153 | { | ||
154 | struct pci_dn *pdn = PCI_DN(dn); | ||
155 | int pos = pseries_eeh_cap_start(dn); | ||
156 | int cnt = 48; /* Maximal number of capabilities */ | ||
157 | u32 id; | ||
158 | |||
159 | if (!pos) | ||
160 | return 0; | ||
161 | |||
162 | while (cnt--) { | ||
163 | rtas_read_config(pdn, pos, 1, &pos); | ||
164 | if (pos < 0x40) | ||
165 | break; | ||
166 | pos &= ~3; | ||
167 | rtas_read_config(pdn, pos + PCI_CAP_LIST_ID, 1, &id); | ||
168 | if (id == 0xff) | ||
169 | break; | ||
170 | if (id == cap) | ||
171 | return pos; | ||
172 | pos += PCI_CAP_LIST_NEXT; | ||
173 | } | ||
174 | |||
175 | return 0; | ||
176 | } | ||
177 | |||
136 | /** | 178 | /** |
137 | * pseries_eeh_of_probe - EEH probe on the given device | 179 | * pseries_eeh_of_probe - EEH probe on the given device |
138 | * @dn: OF node | 180 | * @dn: OF node |
@@ -146,14 +188,16 @@ static void *pseries_eeh_of_probe(struct device_node *dn, void *flag) | |||
146 | { | 188 | { |
147 | struct eeh_dev *edev; | 189 | struct eeh_dev *edev; |
148 | struct eeh_pe pe; | 190 | struct eeh_pe pe; |
191 | struct pci_dn *pdn = PCI_DN(dn); | ||
149 | const u32 *class_code, *vendor_id, *device_id; | 192 | const u32 *class_code, *vendor_id, *device_id; |
150 | const u32 *regs; | 193 | const u32 *regs; |
194 | u32 pcie_flags; | ||
151 | int enable = 0; | 195 | int enable = 0; |
152 | int ret; | 196 | int ret; |
153 | 197 | ||
154 | /* Retrieve OF node and eeh device */ | 198 | /* Retrieve OF node and eeh device */ |
155 | edev = of_node_to_eeh_dev(dn); | 199 | edev = of_node_to_eeh_dev(dn); |
156 | if (!of_device_is_available(dn)) | 200 | if (edev->pe || !of_device_is_available(dn)) |
157 | return NULL; | 201 | return NULL; |
158 | 202 | ||
159 | /* Retrieve class/vendor/device IDs */ | 203 | /* Retrieve class/vendor/device IDs */ |
@@ -167,9 +211,26 @@ static void *pseries_eeh_of_probe(struct device_node *dn, void *flag) | |||
167 | if (dn->type && !strcmp(dn->type, "isa")) | 211 | if (dn->type && !strcmp(dn->type, "isa")) |
168 | return NULL; | 212 | return NULL; |
169 | 213 | ||
170 | /* Update class code and mode of eeh device */ | 214 | /* |
215 | * Update class code and mode of eeh device. We need | ||
216 | * correctly reflects that current device is root port | ||
217 | * or PCIe switch downstream port. | ||
218 | */ | ||
171 | edev->class_code = *class_code; | 219 | edev->class_code = *class_code; |
172 | edev->mode = 0; | 220 | edev->pcie_cap = pseries_eeh_find_cap(dn, PCI_CAP_ID_EXP); |
221 | edev->mode &= 0xFFFFFF00; | ||
222 | if ((edev->class_code >> 8) == PCI_CLASS_BRIDGE_PCI) { | ||
223 | edev->mode |= EEH_DEV_BRIDGE; | ||
224 | if (edev->pcie_cap) { | ||
225 | rtas_read_config(pdn, edev->pcie_cap + PCI_EXP_FLAGS, | ||
226 | 2, &pcie_flags); | ||
227 | pcie_flags = (pcie_flags & PCI_EXP_FLAGS_TYPE) >> 4; | ||
228 | if (pcie_flags == PCI_EXP_TYPE_ROOT_PORT) | ||
229 | edev->mode |= EEH_DEV_ROOT_PORT; | ||
230 | else if (pcie_flags == PCI_EXP_TYPE_DOWNSTREAM) | ||
231 | edev->mode |= EEH_DEV_DS_PORT; | ||
232 | } | ||
233 | } | ||
173 | 234 | ||
174 | /* Retrieve the device address */ | 235 | /* Retrieve the device address */ |
175 | regs = of_get_property(dn, "reg", NULL); | 236 | regs = of_get_property(dn, "reg", NULL); |
diff --git a/arch/powerpc/platforms/pseries/lpar.c b/arch/powerpc/platforms/pseries/lpar.c index 02d6e21619bb..8bad880bd177 100644 --- a/arch/powerpc/platforms/pseries/lpar.c +++ b/arch/powerpc/platforms/pseries/lpar.c | |||
@@ -146,7 +146,7 @@ static long pSeries_lpar_hpte_insert(unsigned long hpte_group, | |||
146 | flags = 0; | 146 | flags = 0; |
147 | 147 | ||
148 | /* Make pHyp happy */ | 148 | /* Make pHyp happy */ |
149 | if ((rflags & _PAGE_NO_CACHE) & !(rflags & _PAGE_WRITETHRU)) | 149 | if ((rflags & _PAGE_NO_CACHE) && !(rflags & _PAGE_WRITETHRU)) |
150 | hpte_r &= ~_PAGE_COHERENT; | 150 | hpte_r &= ~_PAGE_COHERENT; |
151 | if (firmware_has_feature(FW_FEATURE_XCMO) && !(hpte_r & HPTE_R_N)) | 151 | if (firmware_has_feature(FW_FEATURE_XCMO) && !(hpte_r & HPTE_R_N)) |
152 | flags |= H_COALESCE_CAND; | 152 | flags |= H_COALESCE_CAND; |
diff --git a/arch/powerpc/platforms/pseries/ras.c b/arch/powerpc/platforms/pseries/ras.c index 7b3cbde8c783..721c0586b284 100644 --- a/arch/powerpc/platforms/pseries/ras.c +++ b/arch/powerpc/platforms/pseries/ras.c | |||
@@ -287,6 +287,9 @@ static struct rtas_error_log *fwnmi_get_errinfo(struct pt_regs *regs) | |||
287 | unsigned long *savep; | 287 | unsigned long *savep; |
288 | struct rtas_error_log *h, *errhdr = NULL; | 288 | struct rtas_error_log *h, *errhdr = NULL; |
289 | 289 | ||
290 | /* Mask top two bits */ | ||
291 | regs->gpr[3] &= ~(0x3UL << 62); | ||
292 | |||
290 | if (!VALID_FWNMI_BUFFER(regs->gpr[3])) { | 293 | if (!VALID_FWNMI_BUFFER(regs->gpr[3])) { |
291 | printk(KERN_ERR "FWNMI: corrupt r3 0x%016lx\n", regs->gpr[3]); | 294 | printk(KERN_ERR "FWNMI: corrupt r3 0x%016lx\n", regs->gpr[3]); |
292 | return NULL; | 295 | return NULL; |
diff --git a/arch/x86/crypto/Makefile b/arch/x86/crypto/Makefile index 7d6ba9db1be9..6c63c358a7e6 100644 --- a/arch/x86/crypto/Makefile +++ b/arch/x86/crypto/Makefile | |||
@@ -27,7 +27,6 @@ obj-$(CONFIG_CRYPTO_SHA1_SSSE3) += sha1-ssse3.o | |||
27 | obj-$(CONFIG_CRYPTO_CRC32_PCLMUL) += crc32-pclmul.o | 27 | obj-$(CONFIG_CRYPTO_CRC32_PCLMUL) += crc32-pclmul.o |
28 | obj-$(CONFIG_CRYPTO_SHA256_SSSE3) += sha256-ssse3.o | 28 | obj-$(CONFIG_CRYPTO_SHA256_SSSE3) += sha256-ssse3.o |
29 | obj-$(CONFIG_CRYPTO_SHA512_SSSE3) += sha512-ssse3.o | 29 | obj-$(CONFIG_CRYPTO_SHA512_SSSE3) += sha512-ssse3.o |
30 | obj-$(CONFIG_CRYPTO_CRCT10DIF_PCLMUL) += crct10dif-pclmul.o | ||
31 | 30 | ||
32 | # These modules require assembler to support AVX. | 31 | # These modules require assembler to support AVX. |
33 | ifeq ($(avx_supported),yes) | 32 | ifeq ($(avx_supported),yes) |
@@ -82,4 +81,3 @@ crc32c-intel-$(CONFIG_64BIT) += crc32c-pcl-intel-asm_64.o | |||
82 | crc32-pclmul-y := crc32-pclmul_asm.o crc32-pclmul_glue.o | 81 | crc32-pclmul-y := crc32-pclmul_asm.o crc32-pclmul_glue.o |
83 | sha256-ssse3-y := sha256-ssse3-asm.o sha256-avx-asm.o sha256-avx2-asm.o sha256_ssse3_glue.o | 82 | sha256-ssse3-y := sha256-ssse3-asm.o sha256-avx-asm.o sha256-avx2-asm.o sha256_ssse3_glue.o |
84 | sha512-ssse3-y := sha512-ssse3-asm.o sha512-avx-asm.o sha512-avx2-asm.o sha512_ssse3_glue.o | 83 | sha512-ssse3-y := sha512-ssse3-asm.o sha512-avx-asm.o sha512-avx2-asm.o sha512_ssse3_glue.o |
85 | crct10dif-pclmul-y := crct10dif-pcl-asm_64.o crct10dif-pclmul_glue.o | ||
diff --git a/arch/x86/crypto/crct10dif-pcl-asm_64.S b/arch/x86/crypto/crct10dif-pcl-asm_64.S deleted file mode 100644 index 35e97569d05f..000000000000 --- a/arch/x86/crypto/crct10dif-pcl-asm_64.S +++ /dev/null | |||
@@ -1,643 +0,0 @@ | |||
1 | ######################################################################## | ||
2 | # Implement fast CRC-T10DIF computation with SSE and PCLMULQDQ instructions | ||
3 | # | ||
4 | # Copyright (c) 2013, Intel Corporation | ||
5 | # | ||
6 | # Authors: | ||
7 | # Erdinc Ozturk <erdinc.ozturk@intel.com> | ||
8 | # Vinodh Gopal <vinodh.gopal@intel.com> | ||
9 | # James Guilford <james.guilford@intel.com> | ||
10 | # Tim Chen <tim.c.chen@linux.intel.com> | ||
11 | # | ||
12 | # This software is available to you under a choice of one of two | ||
13 | # licenses. You may choose to be licensed under the terms of the GNU | ||
14 | # General Public License (GPL) Version 2, available from the file | ||
15 | # COPYING in the main directory of this source tree, or the | ||
16 | # OpenIB.org BSD license below: | ||
17 | # | ||
18 | # Redistribution and use in source and binary forms, with or without | ||
19 | # modification, are permitted provided that the following conditions are | ||
20 | # met: | ||
21 | # | ||
22 | # * Redistributions of source code must retain the above copyright | ||
23 | # notice, this list of conditions and the following disclaimer. | ||
24 | # | ||
25 | # * Redistributions in binary form must reproduce the above copyright | ||
26 | # notice, this list of conditions and the following disclaimer in the | ||
27 | # documentation and/or other materials provided with the | ||
28 | # distribution. | ||
29 | # | ||
30 | # * Neither the name of the Intel Corporation nor the names of its | ||
31 | # contributors may be used to endorse or promote products derived from | ||
32 | # this software without specific prior written permission. | ||
33 | # | ||
34 | # | ||
35 | # THIS SOFTWARE IS PROVIDED BY INTEL CORPORATION ""AS IS"" AND ANY | ||
36 | # EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE | ||
37 | # IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR | ||
38 | # PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL INTEL CORPORATION OR | ||
39 | # CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, | ||
40 | # EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, | ||
41 | # PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR | ||
42 | # PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF | ||
43 | # LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING | ||
44 | # NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS | ||
45 | # SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | ||
46 | ######################################################################## | ||
47 | # Function API: | ||
48 | # UINT16 crc_t10dif_pcl( | ||
49 | # UINT16 init_crc, //initial CRC value, 16 bits | ||
50 | # const unsigned char *buf, //buffer pointer to calculate CRC on | ||
51 | # UINT64 len //buffer length in bytes (64-bit data) | ||
52 | # ); | ||
53 | # | ||
54 | # Reference paper titled "Fast CRC Computation for Generic | ||
55 | # Polynomials Using PCLMULQDQ Instruction" | ||
56 | # URL: http://www.intel.com/content/dam/www/public/us/en/documents | ||
57 | # /white-papers/fast-crc-computation-generic-polynomials-pclmulqdq-paper.pdf | ||
58 | # | ||
59 | # | ||
60 | |||
61 | #include <linux/linkage.h> | ||
62 | |||
63 | .text | ||
64 | |||
65 | #define arg1 %rdi | ||
66 | #define arg2 %rsi | ||
67 | #define arg3 %rdx | ||
68 | |||
69 | #define arg1_low32 %edi | ||
70 | |||
71 | ENTRY(crc_t10dif_pcl) | ||
72 | .align 16 | ||
73 | |||
74 | # adjust the 16-bit initial_crc value, scale it to 32 bits | ||
75 | shl $16, arg1_low32 | ||
76 | |||
77 | # Allocate Stack Space | ||
78 | mov %rsp, %rcx | ||
79 | sub $16*2, %rsp | ||
80 | # align stack to 16 byte boundary | ||
81 | and $~(0x10 - 1), %rsp | ||
82 | |||
83 | # check if smaller than 256 | ||
84 | cmp $256, arg3 | ||
85 | |||
86 | # for sizes less than 128, we can't fold 64B at a time... | ||
87 | jl _less_than_128 | ||
88 | |||
89 | |||
90 | # load the initial crc value | ||
91 | movd arg1_low32, %xmm10 # initial crc | ||
92 | |||
93 | # crc value does not need to be byte-reflected, but it needs | ||
94 | # to be moved to the high part of the register. | ||
95 | # because data will be byte-reflected and will align with | ||
96 | # initial crc at correct place. | ||
97 | pslldq $12, %xmm10 | ||
98 | |||
99 | movdqa SHUF_MASK(%rip), %xmm11 | ||
100 | # receive the initial 64B data, xor the initial crc value | ||
101 | movdqu 16*0(arg2), %xmm0 | ||
102 | movdqu 16*1(arg2), %xmm1 | ||
103 | movdqu 16*2(arg2), %xmm2 | ||
104 | movdqu 16*3(arg2), %xmm3 | ||
105 | movdqu 16*4(arg2), %xmm4 | ||
106 | movdqu 16*5(arg2), %xmm5 | ||
107 | movdqu 16*6(arg2), %xmm6 | ||
108 | movdqu 16*7(arg2), %xmm7 | ||
109 | |||
110 | pshufb %xmm11, %xmm0 | ||
111 | # XOR the initial_crc value | ||
112 | pxor %xmm10, %xmm0 | ||
113 | pshufb %xmm11, %xmm1 | ||
114 | pshufb %xmm11, %xmm2 | ||
115 | pshufb %xmm11, %xmm3 | ||
116 | pshufb %xmm11, %xmm4 | ||
117 | pshufb %xmm11, %xmm5 | ||
118 | pshufb %xmm11, %xmm6 | ||
119 | pshufb %xmm11, %xmm7 | ||
120 | |||
121 | movdqa rk3(%rip), %xmm10 #xmm10 has rk3 and rk4 | ||
122 | #imm value of pclmulqdq instruction | ||
123 | #will determine which constant to use | ||
124 | |||
125 | ################################################################# | ||
126 | # we subtract 256 instead of 128 to save one instruction from the loop | ||
127 | sub $256, arg3 | ||
128 | |||
129 | # at this section of the code, there is 64*x+y (0<=y<64) bytes of | ||
130 | # buffer. The _fold_64_B_loop will fold 64B at a time | ||
131 | # until we have 64+y Bytes of buffer | ||
132 | |||
133 | |||
134 | # fold 64B at a time. This section of the code folds 4 xmm | ||
135 | # registers in parallel | ||
136 | _fold_64_B_loop: | ||
137 | |||
138 | # update the buffer pointer | ||
139 | add $128, arg2 # buf += 64# | ||
140 | |||
141 | movdqu 16*0(arg2), %xmm9 | ||
142 | movdqu 16*1(arg2), %xmm12 | ||
143 | pshufb %xmm11, %xmm9 | ||
144 | pshufb %xmm11, %xmm12 | ||
145 | movdqa %xmm0, %xmm8 | ||
146 | movdqa %xmm1, %xmm13 | ||
147 | pclmulqdq $0x0 , %xmm10, %xmm0 | ||
148 | pclmulqdq $0x11, %xmm10, %xmm8 | ||
149 | pclmulqdq $0x0 , %xmm10, %xmm1 | ||
150 | pclmulqdq $0x11, %xmm10, %xmm13 | ||
151 | pxor %xmm9 , %xmm0 | ||
152 | xorps %xmm8 , %xmm0 | ||
153 | pxor %xmm12, %xmm1 | ||
154 | xorps %xmm13, %xmm1 | ||
155 | |||
156 | movdqu 16*2(arg2), %xmm9 | ||
157 | movdqu 16*3(arg2), %xmm12 | ||
158 | pshufb %xmm11, %xmm9 | ||
159 | pshufb %xmm11, %xmm12 | ||
160 | movdqa %xmm2, %xmm8 | ||
161 | movdqa %xmm3, %xmm13 | ||
162 | pclmulqdq $0x0, %xmm10, %xmm2 | ||
163 | pclmulqdq $0x11, %xmm10, %xmm8 | ||
164 | pclmulqdq $0x0, %xmm10, %xmm3 | ||
165 | pclmulqdq $0x11, %xmm10, %xmm13 | ||
166 | pxor %xmm9 , %xmm2 | ||
167 | xorps %xmm8 , %xmm2 | ||
168 | pxor %xmm12, %xmm3 | ||
169 | xorps %xmm13, %xmm3 | ||
170 | |||
171 | movdqu 16*4(arg2), %xmm9 | ||
172 | movdqu 16*5(arg2), %xmm12 | ||
173 | pshufb %xmm11, %xmm9 | ||
174 | pshufb %xmm11, %xmm12 | ||
175 | movdqa %xmm4, %xmm8 | ||
176 | movdqa %xmm5, %xmm13 | ||
177 | pclmulqdq $0x0, %xmm10, %xmm4 | ||
178 | pclmulqdq $0x11, %xmm10, %xmm8 | ||
179 | pclmulqdq $0x0, %xmm10, %xmm5 | ||
180 | pclmulqdq $0x11, %xmm10, %xmm13 | ||
181 | pxor %xmm9 , %xmm4 | ||
182 | xorps %xmm8 , %xmm4 | ||
183 | pxor %xmm12, %xmm5 | ||
184 | xorps %xmm13, %xmm5 | ||
185 | |||
186 | movdqu 16*6(arg2), %xmm9 | ||
187 | movdqu 16*7(arg2), %xmm12 | ||
188 | pshufb %xmm11, %xmm9 | ||
189 | pshufb %xmm11, %xmm12 | ||
190 | movdqa %xmm6 , %xmm8 | ||
191 | movdqa %xmm7 , %xmm13 | ||
192 | pclmulqdq $0x0 , %xmm10, %xmm6 | ||
193 | pclmulqdq $0x11, %xmm10, %xmm8 | ||
194 | pclmulqdq $0x0 , %xmm10, %xmm7 | ||
195 | pclmulqdq $0x11, %xmm10, %xmm13 | ||
196 | pxor %xmm9 , %xmm6 | ||
197 | xorps %xmm8 , %xmm6 | ||
198 | pxor %xmm12, %xmm7 | ||
199 | xorps %xmm13, %xmm7 | ||
200 | |||
201 | sub $128, arg3 | ||
202 | |||
203 | # check if there is another 64B in the buffer to be able to fold | ||
204 | jge _fold_64_B_loop | ||
205 | ################################################################## | ||
206 | |||
207 | |||
208 | add $128, arg2 | ||
209 | # at this point, the buffer pointer is pointing at the last y Bytes | ||
210 | # of the buffer the 64B of folded data is in 4 of the xmm | ||
211 | # registers: xmm0, xmm1, xmm2, xmm3 | ||
212 | |||
213 | |||
214 | # fold the 8 xmm registers to 1 xmm register with different constants | ||
215 | |||
216 | movdqa rk9(%rip), %xmm10 | ||
217 | movdqa %xmm0, %xmm8 | ||
218 | pclmulqdq $0x11, %xmm10, %xmm0 | ||
219 | pclmulqdq $0x0 , %xmm10, %xmm8 | ||
220 | pxor %xmm8, %xmm7 | ||
221 | xorps %xmm0, %xmm7 | ||
222 | |||
223 | movdqa rk11(%rip), %xmm10 | ||
224 | movdqa %xmm1, %xmm8 | ||
225 | pclmulqdq $0x11, %xmm10, %xmm1 | ||
226 | pclmulqdq $0x0 , %xmm10, %xmm8 | ||
227 | pxor %xmm8, %xmm7 | ||
228 | xorps %xmm1, %xmm7 | ||
229 | |||
230 | movdqa rk13(%rip), %xmm10 | ||
231 | movdqa %xmm2, %xmm8 | ||
232 | pclmulqdq $0x11, %xmm10, %xmm2 | ||
233 | pclmulqdq $0x0 , %xmm10, %xmm8 | ||
234 | pxor %xmm8, %xmm7 | ||
235 | pxor %xmm2, %xmm7 | ||
236 | |||
237 | movdqa rk15(%rip), %xmm10 | ||
238 | movdqa %xmm3, %xmm8 | ||
239 | pclmulqdq $0x11, %xmm10, %xmm3 | ||
240 | pclmulqdq $0x0 , %xmm10, %xmm8 | ||
241 | pxor %xmm8, %xmm7 | ||
242 | xorps %xmm3, %xmm7 | ||
243 | |||
244 | movdqa rk17(%rip), %xmm10 | ||
245 | movdqa %xmm4, %xmm8 | ||
246 | pclmulqdq $0x11, %xmm10, %xmm4 | ||
247 | pclmulqdq $0x0 , %xmm10, %xmm8 | ||
248 | pxor %xmm8, %xmm7 | ||
249 | pxor %xmm4, %xmm7 | ||
250 | |||
251 | movdqa rk19(%rip), %xmm10 | ||
252 | movdqa %xmm5, %xmm8 | ||
253 | pclmulqdq $0x11, %xmm10, %xmm5 | ||
254 | pclmulqdq $0x0 , %xmm10, %xmm8 | ||
255 | pxor %xmm8, %xmm7 | ||
256 | xorps %xmm5, %xmm7 | ||
257 | |||
258 | movdqa rk1(%rip), %xmm10 #xmm10 has rk1 and rk2 | ||
259 | #imm value of pclmulqdq instruction | ||
260 | #will determine which constant to use | ||
261 | movdqa %xmm6, %xmm8 | ||
262 | pclmulqdq $0x11, %xmm10, %xmm6 | ||
263 | pclmulqdq $0x0 , %xmm10, %xmm8 | ||
264 | pxor %xmm8, %xmm7 | ||
265 | pxor %xmm6, %xmm7 | ||
266 | |||
267 | |||
268 | # instead of 64, we add 48 to the loop counter to save 1 instruction | ||
269 | # from the loop instead of a cmp instruction, we use the negative | ||
270 | # flag with the jl instruction | ||
271 | add $128-16, arg3 | ||
272 | jl _final_reduction_for_128 | ||
273 | |||
274 | # now we have 16+y bytes left to reduce. 16 Bytes is in register xmm7 | ||
275 | # and the rest is in memory. We can fold 16 bytes at a time if y>=16 | ||
276 | # continue folding 16B at a time | ||
277 | |||
278 | _16B_reduction_loop: | ||
279 | movdqa %xmm7, %xmm8 | ||
280 | pclmulqdq $0x11, %xmm10, %xmm7 | ||
281 | pclmulqdq $0x0 , %xmm10, %xmm8 | ||
282 | pxor %xmm8, %xmm7 | ||
283 | movdqu (arg2), %xmm0 | ||
284 | pshufb %xmm11, %xmm0 | ||
285 | pxor %xmm0 , %xmm7 | ||
286 | add $16, arg2 | ||
287 | sub $16, arg3 | ||
288 | # instead of a cmp instruction, we utilize the flags with the | ||
289 | # jge instruction equivalent of: cmp arg3, 16-16 | ||
290 | # check if there is any more 16B in the buffer to be able to fold | ||
291 | jge _16B_reduction_loop | ||
292 | |||
293 | #now we have 16+z bytes left to reduce, where 0<= z < 16. | ||
294 | #first, we reduce the data in the xmm7 register | ||
295 | |||
296 | |||
297 | _final_reduction_for_128: | ||
298 | # check if any more data to fold. If not, compute the CRC of | ||
299 | # the final 128 bits | ||
300 | add $16, arg3 | ||
301 | je _128_done | ||
302 | |||
303 | # here we are getting data that is less than 16 bytes. | ||
304 | # since we know that there was data before the pointer, we can | ||
305 | # offset the input pointer before the actual point, to receive | ||
306 | # exactly 16 bytes. after that the registers need to be adjusted. | ||
307 | _get_last_two_xmms: | ||
308 | movdqa %xmm7, %xmm2 | ||
309 | |||
310 | movdqu -16(arg2, arg3), %xmm1 | ||
311 | pshufb %xmm11, %xmm1 | ||
312 | |||
313 | # get rid of the extra data that was loaded before | ||
314 | # load the shift constant | ||
315 | lea pshufb_shf_table+16(%rip), %rax | ||
316 | sub arg3, %rax | ||
317 | movdqu (%rax), %xmm0 | ||
318 | |||
319 | # shift xmm2 to the left by arg3 bytes | ||
320 | pshufb %xmm0, %xmm2 | ||
321 | |||
322 | # shift xmm7 to the right by 16-arg3 bytes | ||
323 | pxor mask1(%rip), %xmm0 | ||
324 | pshufb %xmm0, %xmm7 | ||
325 | pblendvb %xmm2, %xmm1 #xmm0 is implicit | ||
326 | |||
327 | # fold 16 Bytes | ||
328 | movdqa %xmm1, %xmm2 | ||
329 | movdqa %xmm7, %xmm8 | ||
330 | pclmulqdq $0x11, %xmm10, %xmm7 | ||
331 | pclmulqdq $0x0 , %xmm10, %xmm8 | ||
332 | pxor %xmm8, %xmm7 | ||
333 | pxor %xmm2, %xmm7 | ||
334 | |||
335 | _128_done: | ||
336 | # compute crc of a 128-bit value | ||
337 | movdqa rk5(%rip), %xmm10 # rk5 and rk6 in xmm10 | ||
338 | movdqa %xmm7, %xmm0 | ||
339 | |||
340 | #64b fold | ||
341 | pclmulqdq $0x1, %xmm10, %xmm7 | ||
342 | pslldq $8 , %xmm0 | ||
343 | pxor %xmm0, %xmm7 | ||
344 | |||
345 | #32b fold | ||
346 | movdqa %xmm7, %xmm0 | ||
347 | |||
348 | pand mask2(%rip), %xmm0 | ||
349 | |||
350 | psrldq $12, %xmm7 | ||
351 | pclmulqdq $0x10, %xmm10, %xmm7 | ||
352 | pxor %xmm0, %xmm7 | ||
353 | |||
354 | #barrett reduction | ||
355 | _barrett: | ||
356 | movdqa rk7(%rip), %xmm10 # rk7 and rk8 in xmm10 | ||
357 | movdqa %xmm7, %xmm0 | ||
358 | pclmulqdq $0x01, %xmm10, %xmm7 | ||
359 | pslldq $4, %xmm7 | ||
360 | pclmulqdq $0x11, %xmm10, %xmm7 | ||
361 | |||
362 | pslldq $4, %xmm7 | ||
363 | pxor %xmm0, %xmm7 | ||
364 | pextrd $1, %xmm7, %eax | ||
365 | |||
366 | _cleanup: | ||
367 | # scale the result back to 16 bits | ||
368 | shr $16, %eax | ||
369 | mov %rcx, %rsp | ||
370 | ret | ||
371 | |||
372 | ######################################################################## | ||
373 | |||
374 | .align 16 | ||
375 | _less_than_128: | ||
376 | |||
377 | # check if there is enough buffer to be able to fold 16B at a time | ||
378 | cmp $32, arg3 | ||
379 | jl _less_than_32 | ||
380 | movdqa SHUF_MASK(%rip), %xmm11 | ||
381 | |||
382 | # now if there is, load the constants | ||
383 | movdqa rk1(%rip), %xmm10 # rk1 and rk2 in xmm10 | ||
384 | |||
385 | movd arg1_low32, %xmm0 # get the initial crc value | ||
386 | pslldq $12, %xmm0 # align it to its correct place | ||
387 | movdqu (arg2), %xmm7 # load the plaintext | ||
388 | pshufb %xmm11, %xmm7 # byte-reflect the plaintext | ||
389 | pxor %xmm0, %xmm7 | ||
390 | |||
391 | |||
392 | # update the buffer pointer | ||
393 | add $16, arg2 | ||
394 | |||
395 | # update the counter. subtract 32 instead of 16 to save one | ||
396 | # instruction from the loop | ||
397 | sub $32, arg3 | ||
398 | |||
399 | jmp _16B_reduction_loop | ||
400 | |||
401 | |||
402 | .align 16 | ||
403 | _less_than_32: | ||
404 | # mov initial crc to the return value. this is necessary for | ||
405 | # zero-length buffers. | ||
406 | mov arg1_low32, %eax | ||
407 | test arg3, arg3 | ||
408 | je _cleanup | ||
409 | |||
410 | movdqa SHUF_MASK(%rip), %xmm11 | ||
411 | |||
412 | movd arg1_low32, %xmm0 # get the initial crc value | ||
413 | pslldq $12, %xmm0 # align it to its correct place | ||
414 | |||
415 | cmp $16, arg3 | ||
416 | je _exact_16_left | ||
417 | jl _less_than_16_left | ||
418 | |||
419 | movdqu (arg2), %xmm7 # load the plaintext | ||
420 | pshufb %xmm11, %xmm7 # byte-reflect the plaintext | ||
421 | pxor %xmm0 , %xmm7 # xor the initial crc value | ||
422 | add $16, arg2 | ||
423 | sub $16, arg3 | ||
424 | movdqa rk1(%rip), %xmm10 # rk1 and rk2 in xmm10 | ||
425 | jmp _get_last_two_xmms | ||
426 | |||
427 | |||
428 | .align 16 | ||
429 | _less_than_16_left: | ||
430 | # use stack space to load data less than 16 bytes, zero-out | ||
431 | # the 16B in memory first. | ||
432 | |||
433 | pxor %xmm1, %xmm1 | ||
434 | mov %rsp, %r11 | ||
435 | movdqa %xmm1, (%r11) | ||
436 | |||
437 | cmp $4, arg3 | ||
438 | jl _only_less_than_4 | ||
439 | |||
440 | # backup the counter value | ||
441 | mov arg3, %r9 | ||
442 | cmp $8, arg3 | ||
443 | jl _less_than_8_left | ||
444 | |||
445 | # load 8 Bytes | ||
446 | mov (arg2), %rax | ||
447 | mov %rax, (%r11) | ||
448 | add $8, %r11 | ||
449 | sub $8, arg3 | ||
450 | add $8, arg2 | ||
451 | _less_than_8_left: | ||
452 | |||
453 | cmp $4, arg3 | ||
454 | jl _less_than_4_left | ||
455 | |||
456 | # load 4 Bytes | ||
457 | mov (arg2), %eax | ||
458 | mov %eax, (%r11) | ||
459 | add $4, %r11 | ||
460 | sub $4, arg3 | ||
461 | add $4, arg2 | ||
462 | _less_than_4_left: | ||
463 | |||
464 | cmp $2, arg3 | ||
465 | jl _less_than_2_left | ||
466 | |||
467 | # load 2 Bytes | ||
468 | mov (arg2), %ax | ||
469 | mov %ax, (%r11) | ||
470 | add $2, %r11 | ||
471 | sub $2, arg3 | ||
472 | add $2, arg2 | ||
473 | _less_than_2_left: | ||
474 | cmp $1, arg3 | ||
475 | jl _zero_left | ||
476 | |||
477 | # load 1 Byte | ||
478 | mov (arg2), %al | ||
479 | mov %al, (%r11) | ||
480 | _zero_left: | ||
481 | movdqa (%rsp), %xmm7 | ||
482 | pshufb %xmm11, %xmm7 | ||
483 | pxor %xmm0 , %xmm7 # xor the initial crc value | ||
484 | |||
485 | # shl r9, 4 | ||
486 | lea pshufb_shf_table+16(%rip), %rax | ||
487 | sub %r9, %rax | ||
488 | movdqu (%rax), %xmm0 | ||
489 | pxor mask1(%rip), %xmm0 | ||
490 | |||
491 | pshufb %xmm0, %xmm7 | ||
492 | jmp _128_done | ||
493 | |||
494 | .align 16 | ||
495 | _exact_16_left: | ||
496 | movdqu (arg2), %xmm7 | ||
497 | pshufb %xmm11, %xmm7 | ||
498 | pxor %xmm0 , %xmm7 # xor the initial crc value | ||
499 | |||
500 | jmp _128_done | ||
501 | |||
502 | _only_less_than_4: | ||
503 | cmp $3, arg3 | ||
504 | jl _only_less_than_3 | ||
505 | |||
506 | # load 3 Bytes | ||
507 | mov (arg2), %al | ||
508 | mov %al, (%r11) | ||
509 | |||
510 | mov 1(arg2), %al | ||
511 | mov %al, 1(%r11) | ||
512 | |||
513 | mov 2(arg2), %al | ||
514 | mov %al, 2(%r11) | ||
515 | |||
516 | movdqa (%rsp), %xmm7 | ||
517 | pshufb %xmm11, %xmm7 | ||
518 | pxor %xmm0 , %xmm7 # xor the initial crc value | ||
519 | |||
520 | psrldq $5, %xmm7 | ||
521 | |||
522 | jmp _barrett | ||
523 | _only_less_than_3: | ||
524 | cmp $2, arg3 | ||
525 | jl _only_less_than_2 | ||
526 | |||
527 | # load 2 Bytes | ||
528 | mov (arg2), %al | ||
529 | mov %al, (%r11) | ||
530 | |||
531 | mov 1(arg2), %al | ||
532 | mov %al, 1(%r11) | ||
533 | |||
534 | movdqa (%rsp), %xmm7 | ||
535 | pshufb %xmm11, %xmm7 | ||
536 | pxor %xmm0 , %xmm7 # xor the initial crc value | ||
537 | |||
538 | psrldq $6, %xmm7 | ||
539 | |||
540 | jmp _barrett | ||
541 | _only_less_than_2: | ||
542 | |||
543 | # load 1 Byte | ||
544 | mov (arg2), %al | ||
545 | mov %al, (%r11) | ||
546 | |||
547 | movdqa (%rsp), %xmm7 | ||
548 | pshufb %xmm11, %xmm7 | ||
549 | pxor %xmm0 , %xmm7 # xor the initial crc value | ||
550 | |||
551 | psrldq $7, %xmm7 | ||
552 | |||
553 | jmp _barrett | ||
554 | |||
555 | ENDPROC(crc_t10dif_pcl) | ||
556 | |||
557 | .data | ||
558 | |||
559 | # precomputed constants | ||
560 | # these constants are precomputed from the poly: | ||
561 | # 0x8bb70000 (0x8bb7 scaled to 32 bits) | ||
562 | .align 16 | ||
563 | # Q = 0x18BB70000 | ||
564 | # rk1 = 2^(32*3) mod Q << 32 | ||
565 | # rk2 = 2^(32*5) mod Q << 32 | ||
566 | # rk3 = 2^(32*15) mod Q << 32 | ||
567 | # rk4 = 2^(32*17) mod Q << 32 | ||
568 | # rk5 = 2^(32*3) mod Q << 32 | ||
569 | # rk6 = 2^(32*2) mod Q << 32 | ||
570 | # rk7 = floor(2^64/Q) | ||
571 | # rk8 = Q | ||
572 | rk1: | ||
573 | .quad 0x2d56000000000000 | ||
574 | rk2: | ||
575 | .quad 0x06df000000000000 | ||
576 | rk3: | ||
577 | .quad 0x9d9d000000000000 | ||
578 | rk4: | ||
579 | .quad 0x7cf5000000000000 | ||
580 | rk5: | ||
581 | .quad 0x2d56000000000000 | ||
582 | rk6: | ||
583 | .quad 0x1368000000000000 | ||
584 | rk7: | ||
585 | .quad 0x00000001f65a57f8 | ||
586 | rk8: | ||
587 | .quad 0x000000018bb70000 | ||
588 | |||
589 | rk9: | ||
590 | .quad 0xceae000000000000 | ||
591 | rk10: | ||
592 | .quad 0xbfd6000000000000 | ||
593 | rk11: | ||
594 | .quad 0x1e16000000000000 | ||
595 | rk12: | ||
596 | .quad 0x713c000000000000 | ||
597 | rk13: | ||
598 | .quad 0xf7f9000000000000 | ||
599 | rk14: | ||
600 | .quad 0x80a6000000000000 | ||
601 | rk15: | ||
602 | .quad 0x044c000000000000 | ||
603 | rk16: | ||
604 | .quad 0xe658000000000000 | ||
605 | rk17: | ||
606 | .quad 0xad18000000000000 | ||
607 | rk18: | ||
608 | .quad 0xa497000000000000 | ||
609 | rk19: | ||
610 | .quad 0x6ee3000000000000 | ||
611 | rk20: | ||
612 | .quad 0xe7b5000000000000 | ||
613 | |||
614 | |||
615 | |||
616 | mask1: | ||
617 | .octa 0x80808080808080808080808080808080 | ||
618 | mask2: | ||
619 | .octa 0x00000000FFFFFFFFFFFFFFFFFFFFFFFF | ||
620 | |||
621 | SHUF_MASK: | ||
622 | .octa 0x000102030405060708090A0B0C0D0E0F | ||
623 | |||
624 | pshufb_shf_table: | ||
625 | # use these values for shift constants for the pshufb instruction | ||
626 | # different alignments result in values as shown: | ||
627 | # DDQ 0x008f8e8d8c8b8a898887868584838281 # shl 15 (16-1) / shr1 | ||
628 | # DDQ 0x01008f8e8d8c8b8a8988878685848382 # shl 14 (16-3) / shr2 | ||
629 | # DDQ 0x0201008f8e8d8c8b8a89888786858483 # shl 13 (16-4) / shr3 | ||
630 | # DDQ 0x030201008f8e8d8c8b8a898887868584 # shl 12 (16-4) / shr4 | ||
631 | # DDQ 0x04030201008f8e8d8c8b8a8988878685 # shl 11 (16-5) / shr5 | ||
632 | # DDQ 0x0504030201008f8e8d8c8b8a89888786 # shl 10 (16-6) / shr6 | ||
633 | # DDQ 0x060504030201008f8e8d8c8b8a898887 # shl 9 (16-7) / shr7 | ||
634 | # DDQ 0x07060504030201008f8e8d8c8b8a8988 # shl 8 (16-8) / shr8 | ||
635 | # DDQ 0x0807060504030201008f8e8d8c8b8a89 # shl 7 (16-9) / shr9 | ||
636 | # DDQ 0x090807060504030201008f8e8d8c8b8a # shl 6 (16-10) / shr10 | ||
637 | # DDQ 0x0a090807060504030201008f8e8d8c8b # shl 5 (16-11) / shr11 | ||
638 | # DDQ 0x0b0a090807060504030201008f8e8d8c # shl 4 (16-12) / shr12 | ||
639 | # DDQ 0x0c0b0a090807060504030201008f8e8d # shl 3 (16-13) / shr13 | ||
640 | # DDQ 0x0d0c0b0a090807060504030201008f8e # shl 2 (16-14) / shr14 | ||
641 | # DDQ 0x0e0d0c0b0a090807060504030201008f # shl 1 (16-15) / shr15 | ||
642 | .octa 0x8f8e8d8c8b8a89888786858483828100 | ||
643 | .octa 0x000e0d0c0b0a09080706050403020100 | ||
diff --git a/arch/x86/crypto/crct10dif-pclmul_glue.c b/arch/x86/crypto/crct10dif-pclmul_glue.c deleted file mode 100644 index 7845d7fd54c0..000000000000 --- a/arch/x86/crypto/crct10dif-pclmul_glue.c +++ /dev/null | |||
@@ -1,151 +0,0 @@ | |||
1 | /* | ||
2 | * Cryptographic API. | ||
3 | * | ||
4 | * T10 Data Integrity Field CRC16 Crypto Transform using PCLMULQDQ Instructions | ||
5 | * | ||
6 | * Copyright (C) 2013 Intel Corporation | ||
7 | * Author: Tim Chen <tim.c.chen@linux.intel.com> | ||
8 | * | ||
9 | * This program is free software; you can redistribute it and/or modify it | ||
10 | * under the terms of the GNU General Public License as published by the Free | ||
11 | * Software Foundation; either version 2 of the License, or (at your option) | ||
12 | * any later version. | ||
13 | * | ||
14 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, | ||
15 | * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF | ||
16 | * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND | ||
17 | * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS | ||
18 | * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN | ||
19 | * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN | ||
20 | * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE | ||
21 | * SOFTWARE. | ||
22 | * | ||
23 | */ | ||
24 | |||
25 | #include <linux/types.h> | ||
26 | #include <linux/module.h> | ||
27 | #include <linux/crc-t10dif.h> | ||
28 | #include <crypto/internal/hash.h> | ||
29 | #include <linux/init.h> | ||
30 | #include <linux/string.h> | ||
31 | #include <linux/kernel.h> | ||
32 | #include <asm/i387.h> | ||
33 | #include <asm/cpufeature.h> | ||
34 | #include <asm/cpu_device_id.h> | ||
35 | |||
36 | asmlinkage __u16 crc_t10dif_pcl(__u16 crc, const unsigned char *buf, | ||
37 | size_t len); | ||
38 | |||
39 | struct chksum_desc_ctx { | ||
40 | __u16 crc; | ||
41 | }; | ||
42 | |||
43 | /* | ||
44 | * Steps through buffer one byte at at time, calculates reflected | ||
45 | * crc using table. | ||
46 | */ | ||
47 | |||
48 | static int chksum_init(struct shash_desc *desc) | ||
49 | { | ||
50 | struct chksum_desc_ctx *ctx = shash_desc_ctx(desc); | ||
51 | |||
52 | ctx->crc = 0; | ||
53 | |||
54 | return 0; | ||
55 | } | ||
56 | |||
57 | static int chksum_update(struct shash_desc *desc, const u8 *data, | ||
58 | unsigned int length) | ||
59 | { | ||
60 | struct chksum_desc_ctx *ctx = shash_desc_ctx(desc); | ||
61 | |||
62 | if (irq_fpu_usable()) { | ||
63 | kernel_fpu_begin(); | ||
64 | ctx->crc = crc_t10dif_pcl(ctx->crc, data, length); | ||
65 | kernel_fpu_end(); | ||
66 | } else | ||
67 | ctx->crc = crc_t10dif_generic(ctx->crc, data, length); | ||
68 | return 0; | ||
69 | } | ||
70 | |||
71 | static int chksum_final(struct shash_desc *desc, u8 *out) | ||
72 | { | ||
73 | struct chksum_desc_ctx *ctx = shash_desc_ctx(desc); | ||
74 | |||
75 | *(__u16 *)out = ctx->crc; | ||
76 | return 0; | ||
77 | } | ||
78 | |||
79 | static int __chksum_finup(__u16 *crcp, const u8 *data, unsigned int len, | ||
80 | u8 *out) | ||
81 | { | ||
82 | if (irq_fpu_usable()) { | ||
83 | kernel_fpu_begin(); | ||
84 | *(__u16 *)out = crc_t10dif_pcl(*crcp, data, len); | ||
85 | kernel_fpu_end(); | ||
86 | } else | ||
87 | *(__u16 *)out = crc_t10dif_generic(*crcp, data, len); | ||
88 | return 0; | ||
89 | } | ||
90 | |||
91 | static int chksum_finup(struct shash_desc *desc, const u8 *data, | ||
92 | unsigned int len, u8 *out) | ||
93 | { | ||
94 | struct chksum_desc_ctx *ctx = shash_desc_ctx(desc); | ||
95 | |||
96 | return __chksum_finup(&ctx->crc, data, len, out); | ||
97 | } | ||
98 | |||
99 | static int chksum_digest(struct shash_desc *desc, const u8 *data, | ||
100 | unsigned int length, u8 *out) | ||
101 | { | ||
102 | struct chksum_desc_ctx *ctx = shash_desc_ctx(desc); | ||
103 | |||
104 | return __chksum_finup(&ctx->crc, data, length, out); | ||
105 | } | ||
106 | |||
107 | static struct shash_alg alg = { | ||
108 | .digestsize = CRC_T10DIF_DIGEST_SIZE, | ||
109 | .init = chksum_init, | ||
110 | .update = chksum_update, | ||
111 | .final = chksum_final, | ||
112 | .finup = chksum_finup, | ||
113 | .digest = chksum_digest, | ||
114 | .descsize = sizeof(struct chksum_desc_ctx), | ||
115 | .base = { | ||
116 | .cra_name = "crct10dif", | ||
117 | .cra_driver_name = "crct10dif-pclmul", | ||
118 | .cra_priority = 200, | ||
119 | .cra_blocksize = CRC_T10DIF_BLOCK_SIZE, | ||
120 | .cra_module = THIS_MODULE, | ||
121 | } | ||
122 | }; | ||
123 | |||
124 | static const struct x86_cpu_id crct10dif_cpu_id[] = { | ||
125 | X86_FEATURE_MATCH(X86_FEATURE_PCLMULQDQ), | ||
126 | {} | ||
127 | }; | ||
128 | MODULE_DEVICE_TABLE(x86cpu, crct10dif_cpu_id); | ||
129 | |||
130 | static int __init crct10dif_intel_mod_init(void) | ||
131 | { | ||
132 | if (!x86_match_cpu(crct10dif_cpu_id)) | ||
133 | return -ENODEV; | ||
134 | |||
135 | return crypto_register_shash(&alg); | ||
136 | } | ||
137 | |||
138 | static void __exit crct10dif_intel_mod_fini(void) | ||
139 | { | ||
140 | crypto_unregister_shash(&alg); | ||
141 | } | ||
142 | |||
143 | module_init(crct10dif_intel_mod_init); | ||
144 | module_exit(crct10dif_intel_mod_fini); | ||
145 | |||
146 | MODULE_AUTHOR("Tim Chen <tim.c.chen@linux.intel.com>"); | ||
147 | MODULE_DESCRIPTION("T10 DIF CRC calculation accelerated with PCLMULQDQ."); | ||
148 | MODULE_LICENSE("GPL"); | ||
149 | |||
150 | MODULE_ALIAS("crct10dif"); | ||
151 | MODULE_ALIAS("crct10dif-pclmul"); | ||
diff --git a/crypto/Kconfig b/crypto/Kconfig index 69ce573f1224..aca01164f002 100644 --- a/crypto/Kconfig +++ b/crypto/Kconfig | |||
@@ -376,25 +376,6 @@ config CRYPTO_CRC32_PCLMUL | |||
376 | which will enable any routine to use the CRC-32-IEEE 802.3 checksum | 376 | which will enable any routine to use the CRC-32-IEEE 802.3 checksum |
377 | and gain better performance as compared with the table implementation. | 377 | and gain better performance as compared with the table implementation. |
378 | 378 | ||
379 | config CRYPTO_CRCT10DIF | ||
380 | tristate "CRCT10DIF algorithm" | ||
381 | select CRYPTO_HASH | ||
382 | help | ||
383 | CRC T10 Data Integrity Field computation is being cast as | ||
384 | a crypto transform. This allows for faster crc t10 diff | ||
385 | transforms to be used if they are available. | ||
386 | |||
387 | config CRYPTO_CRCT10DIF_PCLMUL | ||
388 | tristate "CRCT10DIF PCLMULQDQ hardware acceleration" | ||
389 | depends on X86 && 64BIT && CRC_T10DIF | ||
390 | select CRYPTO_HASH | ||
391 | help | ||
392 | For x86_64 processors with SSE4.2 and PCLMULQDQ supported, | ||
393 | CRC T10 DIF PCLMULQDQ computation can be hardware | ||
394 | accelerated PCLMULQDQ instruction. This option will create | ||
395 | 'crct10dif-plcmul' module, which is faster when computing the | ||
396 | crct10dif checksum as compared with the generic table implementation. | ||
397 | |||
398 | config CRYPTO_GHASH | 379 | config CRYPTO_GHASH |
399 | tristate "GHASH digest algorithm" | 380 | tristate "GHASH digest algorithm" |
400 | select CRYPTO_GF128MUL | 381 | select CRYPTO_GF128MUL |
diff --git a/crypto/Makefile b/crypto/Makefile index 2d5ed08a239f..2ba0df2f908f 100644 --- a/crypto/Makefile +++ b/crypto/Makefile | |||
@@ -83,7 +83,6 @@ obj-$(CONFIG_CRYPTO_ZLIB) += zlib.o | |||
83 | obj-$(CONFIG_CRYPTO_MICHAEL_MIC) += michael_mic.o | 83 | obj-$(CONFIG_CRYPTO_MICHAEL_MIC) += michael_mic.o |
84 | obj-$(CONFIG_CRYPTO_CRC32C) += crc32c.o | 84 | obj-$(CONFIG_CRYPTO_CRC32C) += crc32c.o |
85 | obj-$(CONFIG_CRYPTO_CRC32) += crc32.o | 85 | obj-$(CONFIG_CRYPTO_CRC32) += crc32.o |
86 | obj-$(CONFIG_CRYPTO_CRCT10DIF) += crct10dif.o | ||
87 | obj-$(CONFIG_CRYPTO_AUTHENC) += authenc.o authencesn.o | 86 | obj-$(CONFIG_CRYPTO_AUTHENC) += authenc.o authencesn.o |
88 | obj-$(CONFIG_CRYPTO_LZO) += lzo.o | 87 | obj-$(CONFIG_CRYPTO_LZO) += lzo.o |
89 | obj-$(CONFIG_CRYPTO_LZ4) += lz4.o | 88 | obj-$(CONFIG_CRYPTO_LZ4) += lz4.o |
diff --git a/crypto/crct10dif.c b/crypto/crct10dif.c deleted file mode 100644 index 92aca96d6b98..000000000000 --- a/crypto/crct10dif.c +++ /dev/null | |||
@@ -1,178 +0,0 @@ | |||
1 | /* | ||
2 | * Cryptographic API. | ||
3 | * | ||
4 | * T10 Data Integrity Field CRC16 Crypto Transform | ||
5 | * | ||
6 | * Copyright (c) 2007 Oracle Corporation. All rights reserved. | ||
7 | * Written by Martin K. Petersen <martin.petersen@oracle.com> | ||
8 | * Copyright (C) 2013 Intel Corporation | ||
9 | * Author: Tim Chen <tim.c.chen@linux.intel.com> | ||
10 | * | ||
11 | * This program is free software; you can redistribute it and/or modify it | ||
12 | * under the terms of the GNU General Public License as published by the Free | ||
13 | * Software Foundation; either version 2 of the License, or (at your option) | ||
14 | * any later version. | ||
15 | * | ||
16 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, | ||
17 | * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF | ||
18 | * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND | ||
19 | * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS | ||
20 | * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN | ||
21 | * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN | ||
22 | * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE | ||
23 | * SOFTWARE. | ||
24 | * | ||
25 | */ | ||
26 | |||
27 | #include <linux/types.h> | ||
28 | #include <linux/module.h> | ||
29 | #include <linux/crc-t10dif.h> | ||
30 | #include <crypto/internal/hash.h> | ||
31 | #include <linux/init.h> | ||
32 | #include <linux/string.h> | ||
33 | #include <linux/kernel.h> | ||
34 | |||
35 | struct chksum_desc_ctx { | ||
36 | __u16 crc; | ||
37 | }; | ||
38 | |||
39 | /* Table generated using the following polynomium: | ||
40 | * x^16 + x^15 + x^11 + x^9 + x^8 + x^7 + x^5 + x^4 + x^2 + x + 1 | ||
41 | * gt: 0x8bb7 | ||
42 | */ | ||
43 | static const __u16 t10_dif_crc_table[256] = { | ||
44 | 0x0000, 0x8BB7, 0x9CD9, 0x176E, 0xB205, 0x39B2, 0x2EDC, 0xA56B, | ||
45 | 0xEFBD, 0x640A, 0x7364, 0xF8D3, 0x5DB8, 0xD60F, 0xC161, 0x4AD6, | ||
46 | 0x54CD, 0xDF7A, 0xC814, 0x43A3, 0xE6C8, 0x6D7F, 0x7A11, 0xF1A6, | ||
47 | 0xBB70, 0x30C7, 0x27A9, 0xAC1E, 0x0975, 0x82C2, 0x95AC, 0x1E1B, | ||
48 | 0xA99A, 0x222D, 0x3543, 0xBEF4, 0x1B9F, 0x9028, 0x8746, 0x0CF1, | ||
49 | 0x4627, 0xCD90, 0xDAFE, 0x5149, 0xF422, 0x7F95, 0x68FB, 0xE34C, | ||
50 | 0xFD57, 0x76E0, 0x618E, 0xEA39, 0x4F52, 0xC4E5, 0xD38B, 0x583C, | ||
51 | 0x12EA, 0x995D, 0x8E33, 0x0584, 0xA0EF, 0x2B58, 0x3C36, 0xB781, | ||
52 | 0xD883, 0x5334, 0x445A, 0xCFED, 0x6A86, 0xE131, 0xF65F, 0x7DE8, | ||
53 | 0x373E, 0xBC89, 0xABE7, 0x2050, 0x853B, 0x0E8C, 0x19E2, 0x9255, | ||
54 | 0x8C4E, 0x07F9, 0x1097, 0x9B20, 0x3E4B, 0xB5FC, 0xA292, 0x2925, | ||
55 | 0x63F3, 0xE844, 0xFF2A, 0x749D, 0xD1F6, 0x5A41, 0x4D2F, 0xC698, | ||
56 | 0x7119, 0xFAAE, 0xEDC0, 0x6677, 0xC31C, 0x48AB, 0x5FC5, 0xD472, | ||
57 | 0x9EA4, 0x1513, 0x027D, 0x89CA, 0x2CA1, 0xA716, 0xB078, 0x3BCF, | ||
58 | 0x25D4, 0xAE63, 0xB90D, 0x32BA, 0x97D1, 0x1C66, 0x0B08, 0x80BF, | ||
59 | 0xCA69, 0x41DE, 0x56B0, 0xDD07, 0x786C, 0xF3DB, 0xE4B5, 0x6F02, | ||
60 | 0x3AB1, 0xB106, 0xA668, 0x2DDF, 0x88B4, 0x0303, 0x146D, 0x9FDA, | ||
61 | 0xD50C, 0x5EBB, 0x49D5, 0xC262, 0x6709, 0xECBE, 0xFBD0, 0x7067, | ||
62 | 0x6E7C, 0xE5CB, 0xF2A5, 0x7912, 0xDC79, 0x57CE, 0x40A0, 0xCB17, | ||
63 | 0x81C1, 0x0A76, 0x1D18, 0x96AF, 0x33C4, 0xB873, 0xAF1D, 0x24AA, | ||
64 | 0x932B, 0x189C, 0x0FF2, 0x8445, 0x212E, 0xAA99, 0xBDF7, 0x3640, | ||
65 | 0x7C96, 0xF721, 0xE04F, 0x6BF8, 0xCE93, 0x4524, 0x524A, 0xD9FD, | ||
66 | 0xC7E6, 0x4C51, 0x5B3F, 0xD088, 0x75E3, 0xFE54, 0xE93A, 0x628D, | ||
67 | 0x285B, 0xA3EC, 0xB482, 0x3F35, 0x9A5E, 0x11E9, 0x0687, 0x8D30, | ||
68 | 0xE232, 0x6985, 0x7EEB, 0xF55C, 0x5037, 0xDB80, 0xCCEE, 0x4759, | ||
69 | 0x0D8F, 0x8638, 0x9156, 0x1AE1, 0xBF8A, 0x343D, 0x2353, 0xA8E4, | ||
70 | 0xB6FF, 0x3D48, 0x2A26, 0xA191, 0x04FA, 0x8F4D, 0x9823, 0x1394, | ||
71 | 0x5942, 0xD2F5, 0xC59B, 0x4E2C, 0xEB47, 0x60F0, 0x779E, 0xFC29, | ||
72 | 0x4BA8, 0xC01F, 0xD771, 0x5CC6, 0xF9AD, 0x721A, 0x6574, 0xEEC3, | ||
73 | 0xA415, 0x2FA2, 0x38CC, 0xB37B, 0x1610, 0x9DA7, 0x8AC9, 0x017E, | ||
74 | 0x1F65, 0x94D2, 0x83BC, 0x080B, 0xAD60, 0x26D7, 0x31B9, 0xBA0E, | ||
75 | 0xF0D8, 0x7B6F, 0x6C01, 0xE7B6, 0x42DD, 0xC96A, 0xDE04, 0x55B3 | ||
76 | }; | ||
77 | |||
78 | __u16 crc_t10dif_generic(__u16 crc, const unsigned char *buffer, size_t len) | ||
79 | { | ||
80 | unsigned int i; | ||
81 | |||
82 | for (i = 0 ; i < len ; i++) | ||
83 | crc = (crc << 8) ^ t10_dif_crc_table[((crc >> 8) ^ buffer[i]) & 0xff]; | ||
84 | |||
85 | return crc; | ||
86 | } | ||
87 | EXPORT_SYMBOL(crc_t10dif_generic); | ||
88 | |||
89 | /* | ||
90 | * Steps through buffer one byte at at time, calculates reflected | ||
91 | * crc using table. | ||
92 | */ | ||
93 | |||
94 | static int chksum_init(struct shash_desc *desc) | ||
95 | { | ||
96 | struct chksum_desc_ctx *ctx = shash_desc_ctx(desc); | ||
97 | |||
98 | ctx->crc = 0; | ||
99 | |||
100 | return 0; | ||
101 | } | ||
102 | |||
103 | static int chksum_update(struct shash_desc *desc, const u8 *data, | ||
104 | unsigned int length) | ||
105 | { | ||
106 | struct chksum_desc_ctx *ctx = shash_desc_ctx(desc); | ||
107 | |||
108 | ctx->crc = crc_t10dif_generic(ctx->crc, data, length); | ||
109 | return 0; | ||
110 | } | ||
111 | |||
112 | static int chksum_final(struct shash_desc *desc, u8 *out) | ||
113 | { | ||
114 | struct chksum_desc_ctx *ctx = shash_desc_ctx(desc); | ||
115 | |||
116 | *(__u16 *)out = ctx->crc; | ||
117 | return 0; | ||
118 | } | ||
119 | |||
120 | static int __chksum_finup(__u16 *crcp, const u8 *data, unsigned int len, | ||
121 | u8 *out) | ||
122 | { | ||
123 | *(__u16 *)out = crc_t10dif_generic(*crcp, data, len); | ||
124 | return 0; | ||
125 | } | ||
126 | |||
127 | static int chksum_finup(struct shash_desc *desc, const u8 *data, | ||
128 | unsigned int len, u8 *out) | ||
129 | { | ||
130 | struct chksum_desc_ctx *ctx = shash_desc_ctx(desc); | ||
131 | |||
132 | return __chksum_finup(&ctx->crc, data, len, out); | ||
133 | } | ||
134 | |||
135 | static int chksum_digest(struct shash_desc *desc, const u8 *data, | ||
136 | unsigned int length, u8 *out) | ||
137 | { | ||
138 | struct chksum_desc_ctx *ctx = shash_desc_ctx(desc); | ||
139 | |||
140 | return __chksum_finup(&ctx->crc, data, length, out); | ||
141 | } | ||
142 | |||
143 | static struct shash_alg alg = { | ||
144 | .digestsize = CRC_T10DIF_DIGEST_SIZE, | ||
145 | .init = chksum_init, | ||
146 | .update = chksum_update, | ||
147 | .final = chksum_final, | ||
148 | .finup = chksum_finup, | ||
149 | .digest = chksum_digest, | ||
150 | .descsize = sizeof(struct chksum_desc_ctx), | ||
151 | .base = { | ||
152 | .cra_name = "crct10dif", | ||
153 | .cra_driver_name = "crct10dif-generic", | ||
154 | .cra_priority = 100, | ||
155 | .cra_blocksize = CRC_T10DIF_BLOCK_SIZE, | ||
156 | .cra_module = THIS_MODULE, | ||
157 | } | ||
158 | }; | ||
159 | |||
160 | static int __init crct10dif_mod_init(void) | ||
161 | { | ||
162 | int ret; | ||
163 | |||
164 | ret = crypto_register_shash(&alg); | ||
165 | return ret; | ||
166 | } | ||
167 | |||
168 | static void __exit crct10dif_mod_fini(void) | ||
169 | { | ||
170 | crypto_unregister_shash(&alg); | ||
171 | } | ||
172 | |||
173 | module_init(crct10dif_mod_init); | ||
174 | module_exit(crct10dif_mod_fini); | ||
175 | |||
176 | MODULE_AUTHOR("Tim Chen <tim.c.chen@linux.intel.com>"); | ||
177 | MODULE_DESCRIPTION("T10 DIF CRC calculation."); | ||
178 | MODULE_LICENSE("GPL"); | ||
diff --git a/crypto/tcrypt.c b/crypto/tcrypt.c index 25a5934f0e50..66d254ce0d11 100644 --- a/crypto/tcrypt.c +++ b/crypto/tcrypt.c | |||
@@ -1174,10 +1174,6 @@ static int do_test(int m) | |||
1174 | ret += tcrypt_test("ghash"); | 1174 | ret += tcrypt_test("ghash"); |
1175 | break; | 1175 | break; |
1176 | 1176 | ||
1177 | case 47: | ||
1178 | ret += tcrypt_test("crct10dif"); | ||
1179 | break; | ||
1180 | |||
1181 | case 100: | 1177 | case 100: |
1182 | ret += tcrypt_test("hmac(md5)"); | 1178 | ret += tcrypt_test("hmac(md5)"); |
1183 | break; | 1179 | break; |
@@ -1502,10 +1498,6 @@ static int do_test(int m) | |||
1502 | test_hash_speed("crc32c", sec, generic_hash_speed_template); | 1498 | test_hash_speed("crc32c", sec, generic_hash_speed_template); |
1503 | if (mode > 300 && mode < 400) break; | 1499 | if (mode > 300 && mode < 400) break; |
1504 | 1500 | ||
1505 | case 320: | ||
1506 | test_hash_speed("crct10dif", sec, generic_hash_speed_template); | ||
1507 | if (mode > 300 && mode < 400) break; | ||
1508 | |||
1509 | case 399: | 1501 | case 399: |
1510 | break; | 1502 | break; |
1511 | 1503 | ||
diff --git a/crypto/testmgr.c b/crypto/testmgr.c index 2f00607039e2..ecddf921a9db 100644 --- a/crypto/testmgr.c +++ b/crypto/testmgr.c | |||
@@ -2046,16 +2046,6 @@ static const struct alg_test_desc alg_test_descs[] = { | |||
2046 | } | 2046 | } |
2047 | } | 2047 | } |
2048 | }, { | 2048 | }, { |
2049 | .alg = "crct10dif", | ||
2050 | .test = alg_test_hash, | ||
2051 | .fips_allowed = 1, | ||
2052 | .suite = { | ||
2053 | .hash = { | ||
2054 | .vecs = crct10dif_tv_template, | ||
2055 | .count = CRCT10DIF_TEST_VECTORS | ||
2056 | } | ||
2057 | } | ||
2058 | }, { | ||
2059 | .alg = "cryptd(__driver-cbc-aes-aesni)", | 2049 | .alg = "cryptd(__driver-cbc-aes-aesni)", |
2060 | .test = alg_test_null, | 2050 | .test = alg_test_null, |
2061 | .fips_allowed = 1, | 2051 | .fips_allowed = 1, |
diff --git a/crypto/testmgr.h b/crypto/testmgr.h index 7d44aa3d6b44..1e701bc075b9 100644 --- a/crypto/testmgr.h +++ b/crypto/testmgr.h | |||
@@ -450,39 +450,6 @@ static struct hash_testvec rmd320_tv_template[] = { | |||
450 | } | 450 | } |
451 | }; | 451 | }; |
452 | 452 | ||
453 | #define CRCT10DIF_TEST_VECTORS 3 | ||
454 | static struct hash_testvec crct10dif_tv_template[] = { | ||
455 | { | ||
456 | .plaintext = "abc", | ||
457 | .psize = 3, | ||
458 | #ifdef __LITTLE_ENDIAN | ||
459 | .digest = "\x3b\x44", | ||
460 | #else | ||
461 | .digest = "\x44\x3b", | ||
462 | #endif | ||
463 | }, { | ||
464 | .plaintext = "1234567890123456789012345678901234567890" | ||
465 | "123456789012345678901234567890123456789", | ||
466 | .psize = 79, | ||
467 | #ifdef __LITTLE_ENDIAN | ||
468 | .digest = "\x70\x4b", | ||
469 | #else | ||
470 | .digest = "\x4b\x70", | ||
471 | #endif | ||
472 | }, { | ||
473 | .plaintext = | ||
474 | "abcddddddddddddddddddddddddddddddddddddddddddddddddddddd", | ||
475 | .psize = 56, | ||
476 | #ifdef __LITTLE_ENDIAN | ||
477 | .digest = "\xe3\x9c", | ||
478 | #else | ||
479 | .digest = "\x9c\xe3", | ||
480 | #endif | ||
481 | .np = 2, | ||
482 | .tap = { 28, 28 } | ||
483 | } | ||
484 | }; | ||
485 | |||
486 | /* | 453 | /* |
487 | * SHA1 test vectors from from FIPS PUB 180-1 | 454 | * SHA1 test vectors from from FIPS PUB 180-1 |
488 | * Long vector from CAVS 5.0 | 455 | * Long vector from CAVS 5.0 |
diff --git a/drivers/acpi/internal.h b/drivers/acpi/internal.h index 227aca77ee1e..5da44e81dd4d 100644 --- a/drivers/acpi/internal.h +++ b/drivers/acpi/internal.h | |||
@@ -169,10 +169,8 @@ int acpi_create_platform_device(struct acpi_device *adev, | |||
169 | -------------------------------------------------------------------------- */ | 169 | -------------------------------------------------------------------------- */ |
170 | #if defined(CONFIG_ACPI_VIDEO) || defined(CONFIG_ACPI_VIDEO_MODULE) | 170 | #if defined(CONFIG_ACPI_VIDEO) || defined(CONFIG_ACPI_VIDEO_MODULE) |
171 | bool acpi_video_backlight_quirks(void); | 171 | bool acpi_video_backlight_quirks(void); |
172 | bool acpi_video_verify_backlight_support(void); | ||
173 | #else | 172 | #else |
174 | static inline bool acpi_video_backlight_quirks(void) { return false; } | 173 | static inline bool acpi_video_backlight_quirks(void) { return false; } |
175 | static inline bool acpi_video_verify_backlight_support(void) { return false; } | ||
176 | #endif | 174 | #endif |
177 | 175 | ||
178 | #endif /* _ACPI_INTERNAL_H_ */ | 176 | #endif /* _ACPI_INTERNAL_H_ */ |
diff --git a/drivers/acpi/video.c b/drivers/acpi/video.c index 6dd237e79b4f..0ec434d2586d 100644 --- a/drivers/acpi/video.c +++ b/drivers/acpi/video.c | |||
@@ -911,7 +911,7 @@ static void acpi_video_device_find_cap(struct acpi_video_device *device) | |||
911 | if (acpi_video_init_brightness(device)) | 911 | if (acpi_video_init_brightness(device)) |
912 | return; | 912 | return; |
913 | 913 | ||
914 | if (acpi_video_verify_backlight_support()) { | 914 | if (acpi_video_backlight_support()) { |
915 | struct backlight_properties props; | 915 | struct backlight_properties props; |
916 | struct pci_dev *pdev; | 916 | struct pci_dev *pdev; |
917 | acpi_handle acpi_parent; | 917 | acpi_handle acpi_parent; |
@@ -1366,8 +1366,8 @@ acpi_video_switch_brightness(struct acpi_video_device *device, int event) | |||
1366 | unsigned long long level_current, level_next; | 1366 | unsigned long long level_current, level_next; |
1367 | int result = -EINVAL; | 1367 | int result = -EINVAL; |
1368 | 1368 | ||
1369 | /* no warning message if acpi_backlight=vendor or a quirk is used */ | 1369 | /* no warning message if acpi_backlight=vendor is used */ |
1370 | if (!acpi_video_verify_backlight_support()) | 1370 | if (!acpi_video_backlight_support()) |
1371 | return 0; | 1371 | return 0; |
1372 | 1372 | ||
1373 | if (!device->brightness) | 1373 | if (!device->brightness) |
@@ -1875,46 +1875,6 @@ static int acpi_video_bus_remove(struct acpi_device *device) | |||
1875 | return 0; | 1875 | return 0; |
1876 | } | 1876 | } |
1877 | 1877 | ||
1878 | static acpi_status video_unregister_backlight(acpi_handle handle, u32 lvl, | ||
1879 | void *context, void **rv) | ||
1880 | { | ||
1881 | struct acpi_device *acpi_dev; | ||
1882 | struct acpi_video_bus *video; | ||
1883 | struct acpi_video_device *dev, *next; | ||
1884 | |||
1885 | if (acpi_bus_get_device(handle, &acpi_dev)) | ||
1886 | return AE_OK; | ||
1887 | |||
1888 | if (acpi_match_device_ids(acpi_dev, video_device_ids)) | ||
1889 | return AE_OK; | ||
1890 | |||
1891 | video = acpi_driver_data(acpi_dev); | ||
1892 | if (!video) | ||
1893 | return AE_OK; | ||
1894 | |||
1895 | acpi_video_bus_stop_devices(video); | ||
1896 | mutex_lock(&video->device_list_lock); | ||
1897 | list_for_each_entry_safe(dev, next, &video->video_device_list, entry) { | ||
1898 | if (dev->backlight) { | ||
1899 | backlight_device_unregister(dev->backlight); | ||
1900 | dev->backlight = NULL; | ||
1901 | kfree(dev->brightness->levels); | ||
1902 | kfree(dev->brightness); | ||
1903 | } | ||
1904 | if (dev->cooling_dev) { | ||
1905 | sysfs_remove_link(&dev->dev->dev.kobj, | ||
1906 | "thermal_cooling"); | ||
1907 | sysfs_remove_link(&dev->cooling_dev->device.kobj, | ||
1908 | "device"); | ||
1909 | thermal_cooling_device_unregister(dev->cooling_dev); | ||
1910 | dev->cooling_dev = NULL; | ||
1911 | } | ||
1912 | } | ||
1913 | mutex_unlock(&video->device_list_lock); | ||
1914 | acpi_video_bus_start_devices(video); | ||
1915 | return AE_OK; | ||
1916 | } | ||
1917 | |||
1918 | static int __init is_i740(struct pci_dev *dev) | 1878 | static int __init is_i740(struct pci_dev *dev) |
1919 | { | 1879 | { |
1920 | if (dev->device == 0x00D1) | 1880 | if (dev->device == 0x00D1) |
@@ -1946,25 +1906,14 @@ static int __init intel_opregion_present(void) | |||
1946 | return opregion; | 1906 | return opregion; |
1947 | } | 1907 | } |
1948 | 1908 | ||
1949 | int __acpi_video_register(bool backlight_quirks) | 1909 | int acpi_video_register(void) |
1950 | { | 1910 | { |
1951 | bool no_backlight; | 1911 | int result = 0; |
1952 | int result; | ||
1953 | |||
1954 | no_backlight = backlight_quirks ? acpi_video_backlight_quirks() : false; | ||
1955 | |||
1956 | if (register_count) { | 1912 | if (register_count) { |
1957 | /* | 1913 | /* |
1958 | * If acpi_video_register() has been called already, don't try | 1914 | * if the function of acpi_video_register is already called, |
1959 | * to register acpi_video_bus, but unregister backlight devices | 1915 | * don't register the acpi_vide_bus again and return no error. |
1960 | * if no backlight support is requested. | ||
1961 | */ | 1916 | */ |
1962 | if (no_backlight) | ||
1963 | acpi_walk_namespace(ACPI_TYPE_DEVICE, ACPI_ROOT_OBJECT, | ||
1964 | ACPI_UINT32_MAX, | ||
1965 | video_unregister_backlight, | ||
1966 | NULL, NULL, NULL); | ||
1967 | |||
1968 | return 0; | 1917 | return 0; |
1969 | } | 1918 | } |
1970 | 1919 | ||
@@ -1980,7 +1929,7 @@ int __acpi_video_register(bool backlight_quirks) | |||
1980 | 1929 | ||
1981 | return 0; | 1930 | return 0; |
1982 | } | 1931 | } |
1983 | EXPORT_SYMBOL(__acpi_video_register); | 1932 | EXPORT_SYMBOL(acpi_video_register); |
1984 | 1933 | ||
1985 | void acpi_video_unregister(void) | 1934 | void acpi_video_unregister(void) |
1986 | { | 1935 | { |
diff --git a/drivers/acpi/video_detect.c b/drivers/acpi/video_detect.c index 826e52def080..c3397748ba46 100644 --- a/drivers/acpi/video_detect.c +++ b/drivers/acpi/video_detect.c | |||
@@ -238,12 +238,7 @@ static void acpi_video_caps_check(void) | |||
238 | 238 | ||
239 | bool acpi_video_backlight_quirks(void) | 239 | bool acpi_video_backlight_quirks(void) |
240 | { | 240 | { |
241 | if (acpi_gbl_osi_data >= ACPI_OSI_WIN_8) { | 241 | return acpi_gbl_osi_data >= ACPI_OSI_WIN_8; |
242 | acpi_video_caps_check(); | ||
243 | acpi_video_support |= ACPI_VIDEO_SKIP_BACKLIGHT; | ||
244 | return true; | ||
245 | } | ||
246 | return false; | ||
247 | } | 242 | } |
248 | EXPORT_SYMBOL(acpi_video_backlight_quirks); | 243 | EXPORT_SYMBOL(acpi_video_backlight_quirks); |
249 | 244 | ||
@@ -291,14 +286,6 @@ int acpi_video_backlight_support(void) | |||
291 | } | 286 | } |
292 | EXPORT_SYMBOL(acpi_video_backlight_support); | 287 | EXPORT_SYMBOL(acpi_video_backlight_support); |
293 | 288 | ||
294 | /* For the ACPI video driver use only. */ | ||
295 | bool acpi_video_verify_backlight_support(void) | ||
296 | { | ||
297 | return (acpi_video_support & ACPI_VIDEO_SKIP_BACKLIGHT) ? | ||
298 | false : acpi_video_backlight_support(); | ||
299 | } | ||
300 | EXPORT_SYMBOL(acpi_video_verify_backlight_support); | ||
301 | |||
302 | /* | 289 | /* |
303 | * Use acpi_backlight=vendor/video to force that backlight switching | 290 | * Use acpi_backlight=vendor/video to force that backlight switching |
304 | * is processed by vendor specific acpi drivers or video.ko driver. | 291 | * is processed by vendor specific acpi drivers or video.ko driver. |
diff --git a/drivers/ata/Kconfig b/drivers/ata/Kconfig index 80dc988f01e4..4e737728aee2 100644 --- a/drivers/ata/Kconfig +++ b/drivers/ata/Kconfig | |||
@@ -97,6 +97,15 @@ config SATA_AHCI_PLATFORM | |||
97 | 97 | ||
98 | If unsure, say N. | 98 | If unsure, say N. |
99 | 99 | ||
100 | config AHCI_IMX | ||
101 | tristate "Freescale i.MX AHCI SATA support" | ||
102 | depends on SATA_AHCI_PLATFORM && MFD_SYSCON | ||
103 | help | ||
104 | This option enables support for the Freescale i.MX SoC's | ||
105 | onboard AHCI SATA. | ||
106 | |||
107 | If unsure, say N. | ||
108 | |||
100 | config SATA_FSL | 109 | config SATA_FSL |
101 | tristate "Freescale 3.0Gbps SATA support" | 110 | tristate "Freescale 3.0Gbps SATA support" |
102 | depends on FSL_SOC | 111 | depends on FSL_SOC |
@@ -107,7 +116,7 @@ config SATA_FSL | |||
107 | If unsure, say N. | 116 | If unsure, say N. |
108 | 117 | ||
109 | config SATA_INIC162X | 118 | config SATA_INIC162X |
110 | tristate "Initio 162x SATA support" | 119 | tristate "Initio 162x SATA support (Very Experimental)" |
111 | depends on PCI | 120 | depends on PCI |
112 | help | 121 | help |
113 | This option enables support for Initio 162x Serial ATA. | 122 | This option enables support for Initio 162x Serial ATA. |
diff --git a/drivers/ata/Makefile b/drivers/ata/Makefile index c04d0fd038a3..46518c622460 100644 --- a/drivers/ata/Makefile +++ b/drivers/ata/Makefile | |||
@@ -10,6 +10,7 @@ obj-$(CONFIG_SATA_INIC162X) += sata_inic162x.o | |||
10 | obj-$(CONFIG_SATA_SIL24) += sata_sil24.o | 10 | obj-$(CONFIG_SATA_SIL24) += sata_sil24.o |
11 | obj-$(CONFIG_SATA_DWC) += sata_dwc_460ex.o | 11 | obj-$(CONFIG_SATA_DWC) += sata_dwc_460ex.o |
12 | obj-$(CONFIG_SATA_HIGHBANK) += sata_highbank.o libahci.o | 12 | obj-$(CONFIG_SATA_HIGHBANK) += sata_highbank.o libahci.o |
13 | obj-$(CONFIG_AHCI_IMX) += ahci_imx.o | ||
13 | 14 | ||
14 | # SFF w/ custom DMA | 15 | # SFF w/ custom DMA |
15 | obj-$(CONFIG_PDC_ADMA) += pdc_adma.o | 16 | obj-$(CONFIG_PDC_ADMA) += pdc_adma.o |
diff --git a/drivers/ata/ahci.c b/drivers/ata/ahci.c index 5064f3ea20f1..db4380d70031 100644 --- a/drivers/ata/ahci.c +++ b/drivers/ata/ahci.c | |||
@@ -1146,11 +1146,18 @@ int ahci_host_activate(struct ata_host *host, int irq, unsigned int n_msis) | |||
1146 | return rc; | 1146 | return rc; |
1147 | 1147 | ||
1148 | for (i = 0; i < host->n_ports; i++) { | 1148 | for (i = 0; i < host->n_ports; i++) { |
1149 | const char* desc; | ||
1149 | struct ahci_port_priv *pp = host->ports[i]->private_data; | 1150 | struct ahci_port_priv *pp = host->ports[i]->private_data; |
1150 | 1151 | ||
1152 | /* pp is NULL for dummy ports */ | ||
1153 | if (pp) | ||
1154 | desc = pp->irq_desc; | ||
1155 | else | ||
1156 | desc = dev_driver_string(host->dev); | ||
1157 | |||
1151 | rc = devm_request_threaded_irq(host->dev, | 1158 | rc = devm_request_threaded_irq(host->dev, |
1152 | irq + i, ahci_hw_interrupt, ahci_thread_fn, IRQF_SHARED, | 1159 | irq + i, ahci_hw_interrupt, ahci_thread_fn, IRQF_SHARED, |
1153 | pp->irq_desc, host->ports[i]); | 1160 | desc, host->ports[i]); |
1154 | if (rc) | 1161 | if (rc) |
1155 | goto out_free_irqs; | 1162 | goto out_free_irqs; |
1156 | } | 1163 | } |
diff --git a/drivers/ata/ahci_imx.c b/drivers/ata/ahci_imx.c new file mode 100644 index 000000000000..58debb0acc3a --- /dev/null +++ b/drivers/ata/ahci_imx.c | |||
@@ -0,0 +1,236 @@ | |||
1 | /* | ||
2 | * Freescale IMX AHCI SATA platform driver | ||
3 | * Copyright 2013 Freescale Semiconductor, Inc. | ||
4 | * | ||
5 | * based on the AHCI SATA platform driver by Jeff Garzik and Anton Vorontsov | ||
6 | * | ||
7 | * This program is free software; you can redistribute it and/or modify it | ||
8 | * under the terms and conditions of the GNU General Public License, | ||
9 | * version 2, as published by the Free Software Foundation. | ||
10 | * | ||
11 | * This program is distributed in the hope it will be useful, but WITHOUT | ||
12 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or | ||
13 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for | ||
14 | * more details. | ||
15 | * | ||
16 | * You should have received a copy of the GNU General Public License along with | ||
17 | * this program. If not, see <http://www.gnu.org/licenses/>. | ||
18 | */ | ||
19 | |||
20 | #include <linux/kernel.h> | ||
21 | #include <linux/module.h> | ||
22 | #include <linux/platform_device.h> | ||
23 | #include <linux/regmap.h> | ||
24 | #include <linux/ahci_platform.h> | ||
25 | #include <linux/of_device.h> | ||
26 | #include <linux/mfd/syscon.h> | ||
27 | #include <linux/mfd/syscon/imx6q-iomuxc-gpr.h> | ||
28 | #include "ahci.h" | ||
29 | |||
30 | enum { | ||
31 | HOST_TIMER1MS = 0xe0, /* Timer 1-ms */ | ||
32 | }; | ||
33 | |||
34 | struct imx_ahci_priv { | ||
35 | struct platform_device *ahci_pdev; | ||
36 | struct clk *sata_ref_clk; | ||
37 | struct clk *ahb_clk; | ||
38 | struct regmap *gpr; | ||
39 | }; | ||
40 | |||
41 | static int imx6q_sata_init(struct device *dev, void __iomem *mmio) | ||
42 | { | ||
43 | int ret = 0; | ||
44 | unsigned int reg_val; | ||
45 | struct imx_ahci_priv *imxpriv = dev_get_drvdata(dev->parent); | ||
46 | |||
47 | imxpriv->gpr = | ||
48 | syscon_regmap_lookup_by_compatible("fsl,imx6q-iomuxc-gpr"); | ||
49 | if (IS_ERR(imxpriv->gpr)) { | ||
50 | dev_err(dev, "failed to find fsl,imx6q-iomux-gpr regmap\n"); | ||
51 | return PTR_ERR(imxpriv->gpr); | ||
52 | } | ||
53 | |||
54 | ret = clk_prepare_enable(imxpriv->sata_ref_clk); | ||
55 | if (ret < 0) { | ||
56 | dev_err(dev, "prepare-enable sata_ref clock err:%d\n", ret); | ||
57 | return ret; | ||
58 | } | ||
59 | |||
60 | /* | ||
61 | * set PHY Paremeters, two steps to configure the GPR13, | ||
62 | * one write for rest of parameters, mask of first write | ||
63 | * is 0x07fffffd, and the other one write for setting | ||
64 | * the mpll_clk_en. | ||
65 | */ | ||
66 | regmap_update_bits(imxpriv->gpr, 0x34, IMX6Q_GPR13_SATA_RX_EQ_VAL_MASK | ||
67 | | IMX6Q_GPR13_SATA_RX_LOS_LVL_MASK | ||
68 | | IMX6Q_GPR13_SATA_RX_DPLL_MODE_MASK | ||
69 | | IMX6Q_GPR13_SATA_SPD_MODE_MASK | ||
70 | | IMX6Q_GPR13_SATA_MPLL_SS_EN | ||
71 | | IMX6Q_GPR13_SATA_TX_ATTEN_MASK | ||
72 | | IMX6Q_GPR13_SATA_TX_BOOST_MASK | ||
73 | | IMX6Q_GPR13_SATA_TX_LVL_MASK | ||
74 | | IMX6Q_GPR13_SATA_TX_EDGE_RATE | ||
75 | , IMX6Q_GPR13_SATA_RX_EQ_VAL_3_0_DB | ||
76 | | IMX6Q_GPR13_SATA_RX_LOS_LVL_SATA2M | ||
77 | | IMX6Q_GPR13_SATA_RX_DPLL_MODE_2P_4F | ||
78 | | IMX6Q_GPR13_SATA_SPD_MODE_3P0G | ||
79 | | IMX6Q_GPR13_SATA_MPLL_SS_EN | ||
80 | | IMX6Q_GPR13_SATA_TX_ATTEN_9_16 | ||
81 | | IMX6Q_GPR13_SATA_TX_BOOST_3_33_DB | ||
82 | | IMX6Q_GPR13_SATA_TX_LVL_1_025_V); | ||
83 | regmap_update_bits(imxpriv->gpr, 0x34, IMX6Q_GPR13_SATA_MPLL_CLK_EN, | ||
84 | IMX6Q_GPR13_SATA_MPLL_CLK_EN); | ||
85 | usleep_range(100, 200); | ||
86 | |||
87 | /* | ||
88 | * Configure the HWINIT bits of the HOST_CAP and HOST_PORTS_IMPL, | ||
89 | * and IP vendor specific register HOST_TIMER1MS. | ||
90 | * Configure CAP_SSS (support stagered spin up). | ||
91 | * Implement the port0. | ||
92 | * Get the ahb clock rate, and configure the TIMER1MS register. | ||
93 | */ | ||
94 | reg_val = readl(mmio + HOST_CAP); | ||
95 | if (!(reg_val & HOST_CAP_SSS)) { | ||
96 | reg_val |= HOST_CAP_SSS; | ||
97 | writel(reg_val, mmio + HOST_CAP); | ||
98 | } | ||
99 | reg_val = readl(mmio + HOST_PORTS_IMPL); | ||
100 | if (!(reg_val & 0x1)) { | ||
101 | reg_val |= 0x1; | ||
102 | writel(reg_val, mmio + HOST_PORTS_IMPL); | ||
103 | } | ||
104 | |||
105 | reg_val = clk_get_rate(imxpriv->ahb_clk) / 1000; | ||
106 | writel(reg_val, mmio + HOST_TIMER1MS); | ||
107 | |||
108 | return 0; | ||
109 | } | ||
110 | |||
111 | static void imx6q_sata_exit(struct device *dev) | ||
112 | { | ||
113 | struct imx_ahci_priv *imxpriv = dev_get_drvdata(dev->parent); | ||
114 | |||
115 | regmap_update_bits(imxpriv->gpr, 0x34, IMX6Q_GPR13_SATA_MPLL_CLK_EN, | ||
116 | !IMX6Q_GPR13_SATA_MPLL_CLK_EN); | ||
117 | clk_disable_unprepare(imxpriv->sata_ref_clk); | ||
118 | } | ||
119 | |||
120 | static struct ahci_platform_data imx6q_sata_pdata = { | ||
121 | .init = imx6q_sata_init, | ||
122 | .exit = imx6q_sata_exit, | ||
123 | }; | ||
124 | |||
125 | static const struct of_device_id imx_ahci_of_match[] = { | ||
126 | { .compatible = "fsl,imx6q-ahci", .data = &imx6q_sata_pdata}, | ||
127 | {}, | ||
128 | }; | ||
129 | MODULE_DEVICE_TABLE(of, imx_ahci_of_match); | ||
130 | |||
131 | static int imx_ahci_probe(struct platform_device *pdev) | ||
132 | { | ||
133 | struct device *dev = &pdev->dev; | ||
134 | struct resource *mem, *irq, res[2]; | ||
135 | const struct of_device_id *of_id; | ||
136 | const struct ahci_platform_data *pdata = NULL; | ||
137 | struct imx_ahci_priv *imxpriv; | ||
138 | struct device *ahci_dev; | ||
139 | struct platform_device *ahci_pdev; | ||
140 | int ret; | ||
141 | |||
142 | imxpriv = devm_kzalloc(dev, sizeof(*imxpriv), GFP_KERNEL); | ||
143 | if (!imxpriv) { | ||
144 | dev_err(dev, "can't alloc ahci_host_priv\n"); | ||
145 | return -ENOMEM; | ||
146 | } | ||
147 | |||
148 | ahci_pdev = platform_device_alloc("ahci", -1); | ||
149 | if (!ahci_pdev) | ||
150 | return -ENODEV; | ||
151 | |||
152 | ahci_dev = &ahci_pdev->dev; | ||
153 | ahci_dev->parent = dev; | ||
154 | |||
155 | imxpriv->ahb_clk = devm_clk_get(dev, "ahb"); | ||
156 | if (IS_ERR(imxpriv->ahb_clk)) { | ||
157 | dev_err(dev, "can't get ahb clock.\n"); | ||
158 | ret = PTR_ERR(imxpriv->ahb_clk); | ||
159 | goto err_out; | ||
160 | } | ||
161 | |||
162 | imxpriv->sata_ref_clk = devm_clk_get(dev, "sata_ref"); | ||
163 | if (IS_ERR(imxpriv->sata_ref_clk)) { | ||
164 | dev_err(dev, "can't get sata_ref clock.\n"); | ||
165 | ret = PTR_ERR(imxpriv->sata_ref_clk); | ||
166 | goto err_out; | ||
167 | } | ||
168 | |||
169 | imxpriv->ahci_pdev = ahci_pdev; | ||
170 | platform_set_drvdata(pdev, imxpriv); | ||
171 | |||
172 | of_id = of_match_device(imx_ahci_of_match, dev); | ||
173 | if (of_id) { | ||
174 | pdata = of_id->data; | ||
175 | } else { | ||
176 | ret = -EINVAL; | ||
177 | goto err_out; | ||
178 | } | ||
179 | |||
180 | mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); | ||
181 | irq = platform_get_resource(pdev, IORESOURCE_IRQ, 0); | ||
182 | if (!mem || !irq) { | ||
183 | dev_err(dev, "no mmio/irq resource\n"); | ||
184 | ret = -ENOMEM; | ||
185 | goto err_out; | ||
186 | } | ||
187 | |||
188 | res[0] = *mem; | ||
189 | res[1] = *irq; | ||
190 | |||
191 | ahci_dev->coherent_dma_mask = DMA_BIT_MASK(32); | ||
192 | ahci_dev->dma_mask = &ahci_dev->coherent_dma_mask; | ||
193 | ahci_dev->of_node = dev->of_node; | ||
194 | |||
195 | ret = platform_device_add_resources(ahci_pdev, res, 2); | ||
196 | if (ret) | ||
197 | goto err_out; | ||
198 | |||
199 | ret = platform_device_add_data(ahci_pdev, pdata, sizeof(*pdata)); | ||
200 | if (ret) | ||
201 | goto err_out; | ||
202 | |||
203 | ret = platform_device_add(ahci_pdev); | ||
204 | if (ret) { | ||
205 | err_out: | ||
206 | platform_device_put(ahci_pdev); | ||
207 | return ret; | ||
208 | } | ||
209 | |||
210 | return 0; | ||
211 | } | ||
212 | |||
213 | static int imx_ahci_remove(struct platform_device *pdev) | ||
214 | { | ||
215 | struct imx_ahci_priv *imxpriv = platform_get_drvdata(pdev); | ||
216 | struct platform_device *ahci_pdev = imxpriv->ahci_pdev; | ||
217 | |||
218 | platform_device_unregister(ahci_pdev); | ||
219 | return 0; | ||
220 | } | ||
221 | |||
222 | static struct platform_driver imx_ahci_driver = { | ||
223 | .probe = imx_ahci_probe, | ||
224 | .remove = imx_ahci_remove, | ||
225 | .driver = { | ||
226 | .name = "ahci-imx", | ||
227 | .owner = THIS_MODULE, | ||
228 | .of_match_table = imx_ahci_of_match, | ||
229 | }, | ||
230 | }; | ||
231 | module_platform_driver(imx_ahci_driver); | ||
232 | |||
233 | MODULE_DESCRIPTION("Freescale i.MX AHCI SATA platform driver"); | ||
234 | MODULE_AUTHOR("Richard Zhu <Hong-Xing.Zhu@freescale.com>"); | ||
235 | MODULE_LICENSE("GPL"); | ||
236 | MODULE_ALIAS("ahci:imx"); | ||
diff --git a/drivers/ata/ata_piix.c b/drivers/ata/ata_piix.c index b52a10c8eeb9..513ad7ed0c99 100644 --- a/drivers/ata/ata_piix.c +++ b/drivers/ata/ata_piix.c | |||
@@ -330,7 +330,7 @@ static const struct pci_device_id piix_pci_tbl[] = { | |||
330 | /* SATA Controller IDE (Wellsburg) */ | 330 | /* SATA Controller IDE (Wellsburg) */ |
331 | { 0x8086, 0x8d00, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_sata_snb }, | 331 | { 0x8086, 0x8d00, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_sata_snb }, |
332 | /* SATA Controller IDE (Wellsburg) */ | 332 | /* SATA Controller IDE (Wellsburg) */ |
333 | { 0x8086, 0x8d08, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_2port_sata }, | 333 | { 0x8086, 0x8d08, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_2port_sata_snb }, |
334 | /* SATA Controller IDE (Wellsburg) */ | 334 | /* SATA Controller IDE (Wellsburg) */ |
335 | { 0x8086, 0x8d60, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_sata_snb }, | 335 | { 0x8086, 0x8d60, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_sata_snb }, |
336 | /* SATA Controller IDE (Wellsburg) */ | 336 | /* SATA Controller IDE (Wellsburg) */ |
diff --git a/drivers/ata/libata-scsi.c b/drivers/ata/libata-scsi.c index 83c08907e042..b1e880a3c3da 100644 --- a/drivers/ata/libata-scsi.c +++ b/drivers/ata/libata-scsi.c | |||
@@ -206,8 +206,10 @@ static ssize_t ata_scsi_park_store(struct device *device, | |||
206 | unsigned long flags; | 206 | unsigned long flags; |
207 | int rc; | 207 | int rc; |
208 | 208 | ||
209 | rc = strict_strtol(buf, 10, &input); | 209 | rc = kstrtol(buf, 10, &input); |
210 | if (rc || input < -2) | 210 | if (rc) |
211 | return rc; | ||
212 | if (input < -2) | ||
211 | return -EINVAL; | 213 | return -EINVAL; |
212 | if (input > ATA_TMOUT_MAX_PARK) { | 214 | if (input > ATA_TMOUT_MAX_PARK) { |
213 | rc = -EOVERFLOW; | 215 | rc = -EOVERFLOW; |
diff --git a/drivers/ata/sata_inic162x.c b/drivers/ata/sata_inic162x.c index e45131748248..5c54d957370a 100644 --- a/drivers/ata/sata_inic162x.c +++ b/drivers/ata/sata_inic162x.c | |||
@@ -6,6 +6,18 @@ | |||
6 | * | 6 | * |
7 | * This file is released under GPL v2. | 7 | * This file is released under GPL v2. |
8 | * | 8 | * |
9 | * **** WARNING **** | ||
10 | * | ||
11 | * This driver never worked properly and unfortunately data corruption is | ||
12 | * relatively common. There isn't anyone working on the driver and there's | ||
13 | * no support from the vendor. Do not use this driver in any production | ||
14 | * environment. | ||
15 | * | ||
16 | * http://thread.gmane.org/gmane.linux.debian.devel.bugs.rc/378525/focus=54491 | ||
17 | * https://bugzilla.kernel.org/show_bug.cgi?id=60565 | ||
18 | * | ||
19 | * ***************** | ||
20 | * | ||
9 | * This controller is eccentric and easily locks up if something isn't | 21 | * This controller is eccentric and easily locks up if something isn't |
10 | * right. Documentation is available at initio's website but it only | 22 | * right. Documentation is available at initio's website but it only |
11 | * documents registers (not programming model). | 23 | * documents registers (not programming model). |
@@ -807,6 +819,8 @@ static int inic_init_one(struct pci_dev *pdev, const struct pci_device_id *ent) | |||
807 | 819 | ||
808 | ata_print_version_once(&pdev->dev, DRV_VERSION); | 820 | ata_print_version_once(&pdev->dev, DRV_VERSION); |
809 | 821 | ||
822 | dev_alert(&pdev->dev, "inic162x support is broken with common data corruption issues and will be disabled by default, contact linux-ide@vger.kernel.org if in production use\n"); | ||
823 | |||
810 | /* alloc host */ | 824 | /* alloc host */ |
811 | host = ata_host_alloc_pinfo(&pdev->dev, ppi, NR_PORTS); | 825 | host = ata_host_alloc_pinfo(&pdev->dev, ppi, NR_PORTS); |
812 | hpriv = devm_kzalloc(&pdev->dev, sizeof(*hpriv), GFP_KERNEL); | 826 | hpriv = devm_kzalloc(&pdev->dev, sizeof(*hpriv), GFP_KERNEL); |
diff --git a/drivers/block/Kconfig b/drivers/block/Kconfig index b81ddfea1da0..e07a5fd58ad7 100644 --- a/drivers/block/Kconfig +++ b/drivers/block/Kconfig | |||
@@ -532,11 +532,11 @@ config BLK_DEV_RBD | |||
532 | If unsure, say N. | 532 | If unsure, say N. |
533 | 533 | ||
534 | config BLK_DEV_RSXX | 534 | config BLK_DEV_RSXX |
535 | tristate "IBM FlashSystem 70/80 PCIe SSD Device Driver" | 535 | tristate "IBM Flash Adapter 900GB Full Height PCIe Device Driver" |
536 | depends on PCI | 536 | depends on PCI |
537 | help | 537 | help |
538 | Device driver for IBM's high speed PCIe SSD | 538 | Device driver for IBM's high speed PCIe SSD |
539 | storage devices: FlashSystem-70 and FlashSystem-80. | 539 | storage device: Flash Adapter 900GB Full Height. |
540 | 540 | ||
541 | To compile this driver as a module, choose M here: the | 541 | To compile this driver as a module, choose M here: the |
542 | module will be called rsxx. | 542 | module will be called rsxx. |
diff --git a/drivers/block/drbd/drbd_actlog.c b/drivers/block/drbd/drbd_actlog.c index 6608076dc39e..28c73ca320a8 100644 --- a/drivers/block/drbd/drbd_actlog.c +++ b/drivers/block/drbd/drbd_actlog.c | |||
@@ -659,6 +659,27 @@ void drbd_al_shrink(struct drbd_conf *mdev) | |||
659 | wake_up(&mdev->al_wait); | 659 | wake_up(&mdev->al_wait); |
660 | } | 660 | } |
661 | 661 | ||
662 | int drbd_initialize_al(struct drbd_conf *mdev, void *buffer) | ||
663 | { | ||
664 | struct al_transaction_on_disk *al = buffer; | ||
665 | struct drbd_md *md = &mdev->ldev->md; | ||
666 | sector_t al_base = md->md_offset + md->al_offset; | ||
667 | int al_size_4k = md->al_stripes * md->al_stripe_size_4k; | ||
668 | int i; | ||
669 | |||
670 | memset(al, 0, 4096); | ||
671 | al->magic = cpu_to_be32(DRBD_AL_MAGIC); | ||
672 | al->transaction_type = cpu_to_be16(AL_TR_INITIALIZED); | ||
673 | al->crc32c = cpu_to_be32(crc32c(0, al, 4096)); | ||
674 | |||
675 | for (i = 0; i < al_size_4k; i++) { | ||
676 | int err = drbd_md_sync_page_io(mdev, mdev->ldev, al_base + i * 8, WRITE); | ||
677 | if (err) | ||
678 | return err; | ||
679 | } | ||
680 | return 0; | ||
681 | } | ||
682 | |||
662 | static int w_update_odbm(struct drbd_work *w, int unused) | 683 | static int w_update_odbm(struct drbd_work *w, int unused) |
663 | { | 684 | { |
664 | struct update_odbm_work *udw = container_of(w, struct update_odbm_work, w); | 685 | struct update_odbm_work *udw = container_of(w, struct update_odbm_work, w); |
diff --git a/drivers/block/drbd/drbd_int.h b/drivers/block/drbd/drbd_int.h index f943aacfdad8..2d7f608d181c 100644 --- a/drivers/block/drbd/drbd_int.h +++ b/drivers/block/drbd/drbd_int.h | |||
@@ -832,6 +832,7 @@ struct drbd_tconn { /* is a resource from the config file */ | |||
832 | unsigned susp_nod:1; /* IO suspended because no data */ | 832 | unsigned susp_nod:1; /* IO suspended because no data */ |
833 | unsigned susp_fen:1; /* IO suspended because fence peer handler runs */ | 833 | unsigned susp_fen:1; /* IO suspended because fence peer handler runs */ |
834 | struct mutex cstate_mutex; /* Protects graceful disconnects */ | 834 | struct mutex cstate_mutex; /* Protects graceful disconnects */ |
835 | unsigned int connect_cnt; /* Inc each time a connection is established */ | ||
835 | 836 | ||
836 | unsigned long flags; | 837 | unsigned long flags; |
837 | struct net_conf *net_conf; /* content protected by rcu */ | 838 | struct net_conf *net_conf; /* content protected by rcu */ |
@@ -1132,6 +1133,7 @@ extern void drbd_mdev_cleanup(struct drbd_conf *mdev); | |||
1132 | void drbd_print_uuids(struct drbd_conf *mdev, const char *text); | 1133 | void drbd_print_uuids(struct drbd_conf *mdev, const char *text); |
1133 | 1134 | ||
1134 | extern void conn_md_sync(struct drbd_tconn *tconn); | 1135 | extern void conn_md_sync(struct drbd_tconn *tconn); |
1136 | extern void drbd_md_write(struct drbd_conf *mdev, void *buffer); | ||
1135 | extern void drbd_md_sync(struct drbd_conf *mdev); | 1137 | extern void drbd_md_sync(struct drbd_conf *mdev); |
1136 | extern int drbd_md_read(struct drbd_conf *mdev, struct drbd_backing_dev *bdev); | 1138 | extern int drbd_md_read(struct drbd_conf *mdev, struct drbd_backing_dev *bdev); |
1137 | extern void drbd_uuid_set(struct drbd_conf *mdev, int idx, u64 val) __must_hold(local); | 1139 | extern void drbd_uuid_set(struct drbd_conf *mdev, int idx, u64 val) __must_hold(local); |
@@ -1466,8 +1468,16 @@ extern void drbd_suspend_io(struct drbd_conf *mdev); | |||
1466 | extern void drbd_resume_io(struct drbd_conf *mdev); | 1468 | extern void drbd_resume_io(struct drbd_conf *mdev); |
1467 | extern char *ppsize(char *buf, unsigned long long size); | 1469 | extern char *ppsize(char *buf, unsigned long long size); |
1468 | extern sector_t drbd_new_dev_size(struct drbd_conf *, struct drbd_backing_dev *, sector_t, int); | 1470 | extern sector_t drbd_new_dev_size(struct drbd_conf *, struct drbd_backing_dev *, sector_t, int); |
1469 | enum determine_dev_size { dev_size_error = -1, unchanged = 0, shrunk = 1, grew = 2 }; | 1471 | enum determine_dev_size { |
1470 | extern enum determine_dev_size drbd_determine_dev_size(struct drbd_conf *, enum dds_flags) __must_hold(local); | 1472 | DS_ERROR_SHRINK = -3, |
1473 | DS_ERROR_SPACE_MD = -2, | ||
1474 | DS_ERROR = -1, | ||
1475 | DS_UNCHANGED = 0, | ||
1476 | DS_SHRUNK = 1, | ||
1477 | DS_GREW = 2 | ||
1478 | }; | ||
1479 | extern enum determine_dev_size | ||
1480 | drbd_determine_dev_size(struct drbd_conf *, enum dds_flags, struct resize_parms *) __must_hold(local); | ||
1471 | extern void resync_after_online_grow(struct drbd_conf *); | 1481 | extern void resync_after_online_grow(struct drbd_conf *); |
1472 | extern void drbd_reconsider_max_bio_size(struct drbd_conf *mdev); | 1482 | extern void drbd_reconsider_max_bio_size(struct drbd_conf *mdev); |
1473 | extern enum drbd_state_rv drbd_set_role(struct drbd_conf *mdev, | 1483 | extern enum drbd_state_rv drbd_set_role(struct drbd_conf *mdev, |
@@ -1633,6 +1643,7 @@ extern int __drbd_set_out_of_sync(struct drbd_conf *mdev, sector_t sector, | |||
1633 | #define drbd_set_out_of_sync(mdev, sector, size) \ | 1643 | #define drbd_set_out_of_sync(mdev, sector, size) \ |
1634 | __drbd_set_out_of_sync(mdev, sector, size, __FILE__, __LINE__) | 1644 | __drbd_set_out_of_sync(mdev, sector, size, __FILE__, __LINE__) |
1635 | extern void drbd_al_shrink(struct drbd_conf *mdev); | 1645 | extern void drbd_al_shrink(struct drbd_conf *mdev); |
1646 | extern int drbd_initialize_al(struct drbd_conf *, void *); | ||
1636 | 1647 | ||
1637 | /* drbd_nl.c */ | 1648 | /* drbd_nl.c */ |
1638 | /* state info broadcast */ | 1649 | /* state info broadcast */ |
diff --git a/drivers/block/drbd/drbd_main.c b/drivers/block/drbd/drbd_main.c index a5dca6affcbb..55635edf563b 100644 --- a/drivers/block/drbd/drbd_main.c +++ b/drivers/block/drbd/drbd_main.c | |||
@@ -2762,8 +2762,6 @@ int __init drbd_init(void) | |||
2762 | /* | 2762 | /* |
2763 | * allocate all necessary structs | 2763 | * allocate all necessary structs |
2764 | */ | 2764 | */ |
2765 | err = -ENOMEM; | ||
2766 | |||
2767 | init_waitqueue_head(&drbd_pp_wait); | 2765 | init_waitqueue_head(&drbd_pp_wait); |
2768 | 2766 | ||
2769 | drbd_proc = NULL; /* play safe for drbd_cleanup */ | 2767 | drbd_proc = NULL; /* play safe for drbd_cleanup */ |
@@ -2773,6 +2771,7 @@ int __init drbd_init(void) | |||
2773 | if (err) | 2771 | if (err) |
2774 | goto fail; | 2772 | goto fail; |
2775 | 2773 | ||
2774 | err = -ENOMEM; | ||
2776 | drbd_proc = proc_create_data("drbd", S_IFREG | S_IRUGO , NULL, &drbd_proc_fops, NULL); | 2775 | drbd_proc = proc_create_data("drbd", S_IFREG | S_IRUGO , NULL, &drbd_proc_fops, NULL); |
2777 | if (!drbd_proc) { | 2776 | if (!drbd_proc) { |
2778 | printk(KERN_ERR "drbd: unable to register proc file\n"); | 2777 | printk(KERN_ERR "drbd: unable to register proc file\n"); |
@@ -2803,7 +2802,6 @@ int __init drbd_init(void) | |||
2803 | fail: | 2802 | fail: |
2804 | drbd_cleanup(); | 2803 | drbd_cleanup(); |
2805 | if (err == -ENOMEM) | 2804 | if (err == -ENOMEM) |
2806 | /* currently always the case */ | ||
2807 | printk(KERN_ERR "drbd: ran out of memory\n"); | 2805 | printk(KERN_ERR "drbd: ran out of memory\n"); |
2808 | else | 2806 | else |
2809 | printk(KERN_ERR "drbd: initialization failure\n"); | 2807 | printk(KERN_ERR "drbd: initialization failure\n"); |
@@ -2881,34 +2879,14 @@ struct meta_data_on_disk { | |||
2881 | u8 reserved_u8[4096 - (7*8 + 10*4)]; | 2879 | u8 reserved_u8[4096 - (7*8 + 10*4)]; |
2882 | } __packed; | 2880 | } __packed; |
2883 | 2881 | ||
2884 | /** | 2882 | |
2885 | * drbd_md_sync() - Writes the meta data super block if the MD_DIRTY flag bit is set | 2883 | |
2886 | * @mdev: DRBD device. | 2884 | void drbd_md_write(struct drbd_conf *mdev, void *b) |
2887 | */ | ||
2888 | void drbd_md_sync(struct drbd_conf *mdev) | ||
2889 | { | 2885 | { |
2890 | struct meta_data_on_disk *buffer; | 2886 | struct meta_data_on_disk *buffer = b; |
2891 | sector_t sector; | 2887 | sector_t sector; |
2892 | int i; | 2888 | int i; |
2893 | 2889 | ||
2894 | /* Don't accidentally change the DRBD meta data layout. */ | ||
2895 | BUILD_BUG_ON(UI_SIZE != 4); | ||
2896 | BUILD_BUG_ON(sizeof(struct meta_data_on_disk) != 4096); | ||
2897 | |||
2898 | del_timer(&mdev->md_sync_timer); | ||
2899 | /* timer may be rearmed by drbd_md_mark_dirty() now. */ | ||
2900 | if (!test_and_clear_bit(MD_DIRTY, &mdev->flags)) | ||
2901 | return; | ||
2902 | |||
2903 | /* We use here D_FAILED and not D_ATTACHING because we try to write | ||
2904 | * metadata even if we detach due to a disk failure! */ | ||
2905 | if (!get_ldev_if_state(mdev, D_FAILED)) | ||
2906 | return; | ||
2907 | |||
2908 | buffer = drbd_md_get_buffer(mdev); | ||
2909 | if (!buffer) | ||
2910 | goto out; | ||
2911 | |||
2912 | memset(buffer, 0, sizeof(*buffer)); | 2890 | memset(buffer, 0, sizeof(*buffer)); |
2913 | 2891 | ||
2914 | buffer->la_size_sect = cpu_to_be64(drbd_get_capacity(mdev->this_bdev)); | 2892 | buffer->la_size_sect = cpu_to_be64(drbd_get_capacity(mdev->this_bdev)); |
@@ -2937,6 +2915,35 @@ void drbd_md_sync(struct drbd_conf *mdev) | |||
2937 | dev_err(DEV, "meta data update failed!\n"); | 2915 | dev_err(DEV, "meta data update failed!\n"); |
2938 | drbd_chk_io_error(mdev, 1, DRBD_META_IO_ERROR); | 2916 | drbd_chk_io_error(mdev, 1, DRBD_META_IO_ERROR); |
2939 | } | 2917 | } |
2918 | } | ||
2919 | |||
2920 | /** | ||
2921 | * drbd_md_sync() - Writes the meta data super block if the MD_DIRTY flag bit is set | ||
2922 | * @mdev: DRBD device. | ||
2923 | */ | ||
2924 | void drbd_md_sync(struct drbd_conf *mdev) | ||
2925 | { | ||
2926 | struct meta_data_on_disk *buffer; | ||
2927 | |||
2928 | /* Don't accidentally change the DRBD meta data layout. */ | ||
2929 | BUILD_BUG_ON(UI_SIZE != 4); | ||
2930 | BUILD_BUG_ON(sizeof(struct meta_data_on_disk) != 4096); | ||
2931 | |||
2932 | del_timer(&mdev->md_sync_timer); | ||
2933 | /* timer may be rearmed by drbd_md_mark_dirty() now. */ | ||
2934 | if (!test_and_clear_bit(MD_DIRTY, &mdev->flags)) | ||
2935 | return; | ||
2936 | |||
2937 | /* We use here D_FAILED and not D_ATTACHING because we try to write | ||
2938 | * metadata even if we detach due to a disk failure! */ | ||
2939 | if (!get_ldev_if_state(mdev, D_FAILED)) | ||
2940 | return; | ||
2941 | |||
2942 | buffer = drbd_md_get_buffer(mdev); | ||
2943 | if (!buffer) | ||
2944 | goto out; | ||
2945 | |||
2946 | drbd_md_write(mdev, buffer); | ||
2940 | 2947 | ||
2941 | /* Update mdev->ldev->md.la_size_sect, | 2948 | /* Update mdev->ldev->md.la_size_sect, |
2942 | * since we updated it on metadata. */ | 2949 | * since we updated it on metadata. */ |
diff --git a/drivers/block/drbd/drbd_nl.c b/drivers/block/drbd/drbd_nl.c index 9e3f441e7e84..8cc1e640f485 100644 --- a/drivers/block/drbd/drbd_nl.c +++ b/drivers/block/drbd/drbd_nl.c | |||
@@ -417,6 +417,7 @@ static enum drbd_fencing_p highest_fencing_policy(struct drbd_tconn *tconn) | |||
417 | 417 | ||
418 | bool conn_try_outdate_peer(struct drbd_tconn *tconn) | 418 | bool conn_try_outdate_peer(struct drbd_tconn *tconn) |
419 | { | 419 | { |
420 | unsigned int connect_cnt; | ||
420 | union drbd_state mask = { }; | 421 | union drbd_state mask = { }; |
421 | union drbd_state val = { }; | 422 | union drbd_state val = { }; |
422 | enum drbd_fencing_p fp; | 423 | enum drbd_fencing_p fp; |
@@ -428,6 +429,10 @@ bool conn_try_outdate_peer(struct drbd_tconn *tconn) | |||
428 | return false; | 429 | return false; |
429 | } | 430 | } |
430 | 431 | ||
432 | spin_lock_irq(&tconn->req_lock); | ||
433 | connect_cnt = tconn->connect_cnt; | ||
434 | spin_unlock_irq(&tconn->req_lock); | ||
435 | |||
431 | fp = highest_fencing_policy(tconn); | 436 | fp = highest_fencing_policy(tconn); |
432 | switch (fp) { | 437 | switch (fp) { |
433 | case FP_NOT_AVAIL: | 438 | case FP_NOT_AVAIL: |
@@ -492,8 +497,14 @@ bool conn_try_outdate_peer(struct drbd_tconn *tconn) | |||
492 | here, because we might were able to re-establish the connection in the | 497 | here, because we might were able to re-establish the connection in the |
493 | meantime. */ | 498 | meantime. */ |
494 | spin_lock_irq(&tconn->req_lock); | 499 | spin_lock_irq(&tconn->req_lock); |
495 | if (tconn->cstate < C_WF_REPORT_PARAMS && !test_bit(STATE_SENT, &tconn->flags)) | 500 | if (tconn->cstate < C_WF_REPORT_PARAMS && !test_bit(STATE_SENT, &tconn->flags)) { |
496 | _conn_request_state(tconn, mask, val, CS_VERBOSE); | 501 | if (tconn->connect_cnt != connect_cnt) |
502 | /* In case the connection was established and droped | ||
503 | while the fence-peer handler was running, ignore it */ | ||
504 | conn_info(tconn, "Ignoring fence-peer exit code\n"); | ||
505 | else | ||
506 | _conn_request_state(tconn, mask, val, CS_VERBOSE); | ||
507 | } | ||
497 | spin_unlock_irq(&tconn->req_lock); | 508 | spin_unlock_irq(&tconn->req_lock); |
498 | 509 | ||
499 | return conn_highest_pdsk(tconn) <= D_OUTDATED; | 510 | return conn_highest_pdsk(tconn) <= D_OUTDATED; |
@@ -816,15 +827,20 @@ void drbd_resume_io(struct drbd_conf *mdev) | |||
816 | * Returns 0 on success, negative return values indicate errors. | 827 | * Returns 0 on success, negative return values indicate errors. |
817 | * You should call drbd_md_sync() after calling this function. | 828 | * You should call drbd_md_sync() after calling this function. |
818 | */ | 829 | */ |
819 | enum determine_dev_size drbd_determine_dev_size(struct drbd_conf *mdev, enum dds_flags flags) __must_hold(local) | 830 | enum determine_dev_size |
831 | drbd_determine_dev_size(struct drbd_conf *mdev, enum dds_flags flags, struct resize_parms *rs) __must_hold(local) | ||
820 | { | 832 | { |
821 | sector_t prev_first_sect, prev_size; /* previous meta location */ | 833 | sector_t prev_first_sect, prev_size; /* previous meta location */ |
822 | sector_t la_size_sect, u_size; | 834 | sector_t la_size_sect, u_size; |
835 | struct drbd_md *md = &mdev->ldev->md; | ||
836 | u32 prev_al_stripe_size_4k; | ||
837 | u32 prev_al_stripes; | ||
823 | sector_t size; | 838 | sector_t size; |
824 | char ppb[10]; | 839 | char ppb[10]; |
840 | void *buffer; | ||
825 | 841 | ||
826 | int md_moved, la_size_changed; | 842 | int md_moved, la_size_changed; |
827 | enum determine_dev_size rv = unchanged; | 843 | enum determine_dev_size rv = DS_UNCHANGED; |
828 | 844 | ||
829 | /* race: | 845 | /* race: |
830 | * application request passes inc_ap_bio, | 846 | * application request passes inc_ap_bio, |
@@ -836,6 +852,11 @@ enum determine_dev_size drbd_determine_dev_size(struct drbd_conf *mdev, enum dds | |||
836 | * still lock the act_log to not trigger ASSERTs there. | 852 | * still lock the act_log to not trigger ASSERTs there. |
837 | */ | 853 | */ |
838 | drbd_suspend_io(mdev); | 854 | drbd_suspend_io(mdev); |
855 | buffer = drbd_md_get_buffer(mdev); /* Lock meta-data IO */ | ||
856 | if (!buffer) { | ||
857 | drbd_resume_io(mdev); | ||
858 | return DS_ERROR; | ||
859 | } | ||
839 | 860 | ||
840 | /* no wait necessary anymore, actually we could assert that */ | 861 | /* no wait necessary anymore, actually we could assert that */ |
841 | wait_event(mdev->al_wait, lc_try_lock(mdev->act_log)); | 862 | wait_event(mdev->al_wait, lc_try_lock(mdev->act_log)); |
@@ -844,7 +865,17 @@ enum determine_dev_size drbd_determine_dev_size(struct drbd_conf *mdev, enum dds | |||
844 | prev_size = mdev->ldev->md.md_size_sect; | 865 | prev_size = mdev->ldev->md.md_size_sect; |
845 | la_size_sect = mdev->ldev->md.la_size_sect; | 866 | la_size_sect = mdev->ldev->md.la_size_sect; |
846 | 867 | ||
847 | /* TODO: should only be some assert here, not (re)init... */ | 868 | if (rs) { |
869 | /* rs is non NULL if we should change the AL layout only */ | ||
870 | |||
871 | prev_al_stripes = md->al_stripes; | ||
872 | prev_al_stripe_size_4k = md->al_stripe_size_4k; | ||
873 | |||
874 | md->al_stripes = rs->al_stripes; | ||
875 | md->al_stripe_size_4k = rs->al_stripe_size / 4; | ||
876 | md->al_size_4k = (u64)rs->al_stripes * rs->al_stripe_size / 4; | ||
877 | } | ||
878 | |||
848 | drbd_md_set_sector_offsets(mdev, mdev->ldev); | 879 | drbd_md_set_sector_offsets(mdev, mdev->ldev); |
849 | 880 | ||
850 | rcu_read_lock(); | 881 | rcu_read_lock(); |
@@ -852,6 +883,21 @@ enum determine_dev_size drbd_determine_dev_size(struct drbd_conf *mdev, enum dds | |||
852 | rcu_read_unlock(); | 883 | rcu_read_unlock(); |
853 | size = drbd_new_dev_size(mdev, mdev->ldev, u_size, flags & DDSF_FORCED); | 884 | size = drbd_new_dev_size(mdev, mdev->ldev, u_size, flags & DDSF_FORCED); |
854 | 885 | ||
886 | if (size < la_size_sect) { | ||
887 | if (rs && u_size == 0) { | ||
888 | /* Remove "rs &&" later. This check should always be active, but | ||
889 | right now the receiver expects the permissive behavior */ | ||
890 | dev_warn(DEV, "Implicit shrink not allowed. " | ||
891 | "Use --size=%llus for explicit shrink.\n", | ||
892 | (unsigned long long)size); | ||
893 | rv = DS_ERROR_SHRINK; | ||
894 | } | ||
895 | if (u_size > size) | ||
896 | rv = DS_ERROR_SPACE_MD; | ||
897 | if (rv != DS_UNCHANGED) | ||
898 | goto err_out; | ||
899 | } | ||
900 | |||
855 | if (drbd_get_capacity(mdev->this_bdev) != size || | 901 | if (drbd_get_capacity(mdev->this_bdev) != size || |
856 | drbd_bm_capacity(mdev) != size) { | 902 | drbd_bm_capacity(mdev) != size) { |
857 | int err; | 903 | int err; |
@@ -867,7 +913,7 @@ enum determine_dev_size drbd_determine_dev_size(struct drbd_conf *mdev, enum dds | |||
867 | "Leaving size unchanged at size = %lu KB\n", | 913 | "Leaving size unchanged at size = %lu KB\n", |
868 | (unsigned long)size); | 914 | (unsigned long)size); |
869 | } | 915 | } |
870 | rv = dev_size_error; | 916 | rv = DS_ERROR; |
871 | } | 917 | } |
872 | /* racy, see comments above. */ | 918 | /* racy, see comments above. */ |
873 | drbd_set_my_capacity(mdev, size); | 919 | drbd_set_my_capacity(mdev, size); |
@@ -875,38 +921,57 @@ enum determine_dev_size drbd_determine_dev_size(struct drbd_conf *mdev, enum dds | |||
875 | dev_info(DEV, "size = %s (%llu KB)\n", ppsize(ppb, size>>1), | 921 | dev_info(DEV, "size = %s (%llu KB)\n", ppsize(ppb, size>>1), |
876 | (unsigned long long)size>>1); | 922 | (unsigned long long)size>>1); |
877 | } | 923 | } |
878 | if (rv == dev_size_error) | 924 | if (rv <= DS_ERROR) |
879 | goto out; | 925 | goto err_out; |
880 | 926 | ||
881 | la_size_changed = (la_size_sect != mdev->ldev->md.la_size_sect); | 927 | la_size_changed = (la_size_sect != mdev->ldev->md.la_size_sect); |
882 | 928 | ||
883 | md_moved = prev_first_sect != drbd_md_first_sector(mdev->ldev) | 929 | md_moved = prev_first_sect != drbd_md_first_sector(mdev->ldev) |
884 | || prev_size != mdev->ldev->md.md_size_sect; | 930 | || prev_size != mdev->ldev->md.md_size_sect; |
885 | 931 | ||
886 | if (la_size_changed || md_moved) { | 932 | if (la_size_changed || md_moved || rs) { |
887 | int err; | 933 | u32 prev_flags; |
888 | 934 | ||
889 | drbd_al_shrink(mdev); /* All extents inactive. */ | 935 | drbd_al_shrink(mdev); /* All extents inactive. */ |
936 | |||
937 | prev_flags = md->flags; | ||
938 | md->flags &= ~MDF_PRIMARY_IND; | ||
939 | drbd_md_write(mdev, buffer); | ||
940 | |||
890 | dev_info(DEV, "Writing the whole bitmap, %s\n", | 941 | dev_info(DEV, "Writing the whole bitmap, %s\n", |
891 | la_size_changed && md_moved ? "size changed and md moved" : | 942 | la_size_changed && md_moved ? "size changed and md moved" : |
892 | la_size_changed ? "size changed" : "md moved"); | 943 | la_size_changed ? "size changed" : "md moved"); |
893 | /* next line implicitly does drbd_suspend_io()+drbd_resume_io() */ | 944 | /* next line implicitly does drbd_suspend_io()+drbd_resume_io() */ |
894 | err = drbd_bitmap_io(mdev, md_moved ? &drbd_bm_write_all : &drbd_bm_write, | 945 | drbd_bitmap_io(mdev, md_moved ? &drbd_bm_write_all : &drbd_bm_write, |
895 | "size changed", BM_LOCKED_MASK); | 946 | "size changed", BM_LOCKED_MASK); |
896 | if (err) { | 947 | drbd_initialize_al(mdev, buffer); |
897 | rv = dev_size_error; | 948 | |
898 | goto out; | 949 | md->flags = prev_flags; |
899 | } | 950 | drbd_md_write(mdev, buffer); |
900 | drbd_md_mark_dirty(mdev); | 951 | |
952 | if (rs) | ||
953 | dev_info(DEV, "Changed AL layout to al-stripes = %d, al-stripe-size-kB = %d\n", | ||
954 | md->al_stripes, md->al_stripe_size_4k * 4); | ||
901 | } | 955 | } |
902 | 956 | ||
903 | if (size > la_size_sect) | 957 | if (size > la_size_sect) |
904 | rv = grew; | 958 | rv = DS_GREW; |
905 | if (size < la_size_sect) | 959 | if (size < la_size_sect) |
906 | rv = shrunk; | 960 | rv = DS_SHRUNK; |
907 | out: | 961 | |
962 | if (0) { | ||
963 | err_out: | ||
964 | if (rs) { | ||
965 | md->al_stripes = prev_al_stripes; | ||
966 | md->al_stripe_size_4k = prev_al_stripe_size_4k; | ||
967 | md->al_size_4k = (u64)prev_al_stripes * prev_al_stripe_size_4k; | ||
968 | |||
969 | drbd_md_set_sector_offsets(mdev, mdev->ldev); | ||
970 | } | ||
971 | } | ||
908 | lc_unlock(mdev->act_log); | 972 | lc_unlock(mdev->act_log); |
909 | wake_up(&mdev->al_wait); | 973 | wake_up(&mdev->al_wait); |
974 | drbd_md_put_buffer(mdev); | ||
910 | drbd_resume_io(mdev); | 975 | drbd_resume_io(mdev); |
911 | 976 | ||
912 | return rv; | 977 | return rv; |
@@ -1607,11 +1672,11 @@ int drbd_adm_attach(struct sk_buff *skb, struct genl_info *info) | |||
1607 | !drbd_md_test_flag(mdev->ldev, MDF_CONNECTED_IND)) | 1672 | !drbd_md_test_flag(mdev->ldev, MDF_CONNECTED_IND)) |
1608 | set_bit(USE_DEGR_WFC_T, &mdev->flags); | 1673 | set_bit(USE_DEGR_WFC_T, &mdev->flags); |
1609 | 1674 | ||
1610 | dd = drbd_determine_dev_size(mdev, 0); | 1675 | dd = drbd_determine_dev_size(mdev, 0, NULL); |
1611 | if (dd == dev_size_error) { | 1676 | if (dd <= DS_ERROR) { |
1612 | retcode = ERR_NOMEM_BITMAP; | 1677 | retcode = ERR_NOMEM_BITMAP; |
1613 | goto force_diskless_dec; | 1678 | goto force_diskless_dec; |
1614 | } else if (dd == grew) | 1679 | } else if (dd == DS_GREW) |
1615 | set_bit(RESYNC_AFTER_NEG, &mdev->flags); | 1680 | set_bit(RESYNC_AFTER_NEG, &mdev->flags); |
1616 | 1681 | ||
1617 | if (drbd_md_test_flag(mdev->ldev, MDF_FULL_SYNC) || | 1682 | if (drbd_md_test_flag(mdev->ldev, MDF_FULL_SYNC) || |
@@ -2305,6 +2370,7 @@ int drbd_adm_resize(struct sk_buff *skb, struct genl_info *info) | |||
2305 | struct drbd_conf *mdev; | 2370 | struct drbd_conf *mdev; |
2306 | enum drbd_ret_code retcode; | 2371 | enum drbd_ret_code retcode; |
2307 | enum determine_dev_size dd; | 2372 | enum determine_dev_size dd; |
2373 | bool change_al_layout = false; | ||
2308 | enum dds_flags ddsf; | 2374 | enum dds_flags ddsf; |
2309 | sector_t u_size; | 2375 | sector_t u_size; |
2310 | int err; | 2376 | int err; |
@@ -2315,31 +2381,33 @@ int drbd_adm_resize(struct sk_buff *skb, struct genl_info *info) | |||
2315 | if (retcode != NO_ERROR) | 2381 | if (retcode != NO_ERROR) |
2316 | goto fail; | 2382 | goto fail; |
2317 | 2383 | ||
2384 | mdev = adm_ctx.mdev; | ||
2385 | if (!get_ldev(mdev)) { | ||
2386 | retcode = ERR_NO_DISK; | ||
2387 | goto fail; | ||
2388 | } | ||
2389 | |||
2318 | memset(&rs, 0, sizeof(struct resize_parms)); | 2390 | memset(&rs, 0, sizeof(struct resize_parms)); |
2391 | rs.al_stripes = mdev->ldev->md.al_stripes; | ||
2392 | rs.al_stripe_size = mdev->ldev->md.al_stripe_size_4k * 4; | ||
2319 | if (info->attrs[DRBD_NLA_RESIZE_PARMS]) { | 2393 | if (info->attrs[DRBD_NLA_RESIZE_PARMS]) { |
2320 | err = resize_parms_from_attrs(&rs, info); | 2394 | err = resize_parms_from_attrs(&rs, info); |
2321 | if (err) { | 2395 | if (err) { |
2322 | retcode = ERR_MANDATORY_TAG; | 2396 | retcode = ERR_MANDATORY_TAG; |
2323 | drbd_msg_put_info(from_attrs_err_to_txt(err)); | 2397 | drbd_msg_put_info(from_attrs_err_to_txt(err)); |
2324 | goto fail; | 2398 | goto fail_ldev; |
2325 | } | 2399 | } |
2326 | } | 2400 | } |
2327 | 2401 | ||
2328 | mdev = adm_ctx.mdev; | ||
2329 | if (mdev->state.conn > C_CONNECTED) { | 2402 | if (mdev->state.conn > C_CONNECTED) { |
2330 | retcode = ERR_RESIZE_RESYNC; | 2403 | retcode = ERR_RESIZE_RESYNC; |
2331 | goto fail; | 2404 | goto fail_ldev; |
2332 | } | 2405 | } |
2333 | 2406 | ||
2334 | if (mdev->state.role == R_SECONDARY && | 2407 | if (mdev->state.role == R_SECONDARY && |
2335 | mdev->state.peer == R_SECONDARY) { | 2408 | mdev->state.peer == R_SECONDARY) { |
2336 | retcode = ERR_NO_PRIMARY; | 2409 | retcode = ERR_NO_PRIMARY; |
2337 | goto fail; | 2410 | goto fail_ldev; |
2338 | } | ||
2339 | |||
2340 | if (!get_ldev(mdev)) { | ||
2341 | retcode = ERR_NO_DISK; | ||
2342 | goto fail; | ||
2343 | } | 2411 | } |
2344 | 2412 | ||
2345 | if (rs.no_resync && mdev->tconn->agreed_pro_version < 93) { | 2413 | if (rs.no_resync && mdev->tconn->agreed_pro_version < 93) { |
@@ -2358,6 +2426,28 @@ int drbd_adm_resize(struct sk_buff *skb, struct genl_info *info) | |||
2358 | } | 2426 | } |
2359 | } | 2427 | } |
2360 | 2428 | ||
2429 | if (mdev->ldev->md.al_stripes != rs.al_stripes || | ||
2430 | mdev->ldev->md.al_stripe_size_4k != rs.al_stripe_size / 4) { | ||
2431 | u32 al_size_k = rs.al_stripes * rs.al_stripe_size; | ||
2432 | |||
2433 | if (al_size_k > (16 * 1024 * 1024)) { | ||
2434 | retcode = ERR_MD_LAYOUT_TOO_BIG; | ||
2435 | goto fail_ldev; | ||
2436 | } | ||
2437 | |||
2438 | if (al_size_k < MD_32kB_SECT/2) { | ||
2439 | retcode = ERR_MD_LAYOUT_TOO_SMALL; | ||
2440 | goto fail_ldev; | ||
2441 | } | ||
2442 | |||
2443 | if (mdev->state.conn != C_CONNECTED) { | ||
2444 | retcode = ERR_MD_LAYOUT_CONNECTED; | ||
2445 | goto fail_ldev; | ||
2446 | } | ||
2447 | |||
2448 | change_al_layout = true; | ||
2449 | } | ||
2450 | |||
2361 | if (mdev->ldev->known_size != drbd_get_capacity(mdev->ldev->backing_bdev)) | 2451 | if (mdev->ldev->known_size != drbd_get_capacity(mdev->ldev->backing_bdev)) |
2362 | mdev->ldev->known_size = drbd_get_capacity(mdev->ldev->backing_bdev); | 2452 | mdev->ldev->known_size = drbd_get_capacity(mdev->ldev->backing_bdev); |
2363 | 2453 | ||
@@ -2373,16 +2463,22 @@ int drbd_adm_resize(struct sk_buff *skb, struct genl_info *info) | |||
2373 | } | 2463 | } |
2374 | 2464 | ||
2375 | ddsf = (rs.resize_force ? DDSF_FORCED : 0) | (rs.no_resync ? DDSF_NO_RESYNC : 0); | 2465 | ddsf = (rs.resize_force ? DDSF_FORCED : 0) | (rs.no_resync ? DDSF_NO_RESYNC : 0); |
2376 | dd = drbd_determine_dev_size(mdev, ddsf); | 2466 | dd = drbd_determine_dev_size(mdev, ddsf, change_al_layout ? &rs : NULL); |
2377 | drbd_md_sync(mdev); | 2467 | drbd_md_sync(mdev); |
2378 | put_ldev(mdev); | 2468 | put_ldev(mdev); |
2379 | if (dd == dev_size_error) { | 2469 | if (dd == DS_ERROR) { |
2380 | retcode = ERR_NOMEM_BITMAP; | 2470 | retcode = ERR_NOMEM_BITMAP; |
2381 | goto fail; | 2471 | goto fail; |
2472 | } else if (dd == DS_ERROR_SPACE_MD) { | ||
2473 | retcode = ERR_MD_LAYOUT_NO_FIT; | ||
2474 | goto fail; | ||
2475 | } else if (dd == DS_ERROR_SHRINK) { | ||
2476 | retcode = ERR_IMPLICIT_SHRINK; | ||
2477 | goto fail; | ||
2382 | } | 2478 | } |
2383 | 2479 | ||
2384 | if (mdev->state.conn == C_CONNECTED) { | 2480 | if (mdev->state.conn == C_CONNECTED) { |
2385 | if (dd == grew) | 2481 | if (dd == DS_GREW) |
2386 | set_bit(RESIZE_PENDING, &mdev->flags); | 2482 | set_bit(RESIZE_PENDING, &mdev->flags); |
2387 | 2483 | ||
2388 | drbd_send_uuids(mdev); | 2484 | drbd_send_uuids(mdev); |
@@ -2658,7 +2754,6 @@ int nla_put_status_info(struct sk_buff *skb, struct drbd_conf *mdev, | |||
2658 | const struct sib_info *sib) | 2754 | const struct sib_info *sib) |
2659 | { | 2755 | { |
2660 | struct state_info *si = NULL; /* for sizeof(si->member); */ | 2756 | struct state_info *si = NULL; /* for sizeof(si->member); */ |
2661 | struct net_conf *nc; | ||
2662 | struct nlattr *nla; | 2757 | struct nlattr *nla; |
2663 | int got_ldev; | 2758 | int got_ldev; |
2664 | int err = 0; | 2759 | int err = 0; |
@@ -2688,13 +2783,19 @@ int nla_put_status_info(struct sk_buff *skb, struct drbd_conf *mdev, | |||
2688 | goto nla_put_failure; | 2783 | goto nla_put_failure; |
2689 | 2784 | ||
2690 | rcu_read_lock(); | 2785 | rcu_read_lock(); |
2691 | if (got_ldev) | 2786 | if (got_ldev) { |
2692 | if (disk_conf_to_skb(skb, rcu_dereference(mdev->ldev->disk_conf), exclude_sensitive)) | 2787 | struct disk_conf *disk_conf; |
2693 | goto nla_put_failure; | ||
2694 | 2788 | ||
2695 | nc = rcu_dereference(mdev->tconn->net_conf); | 2789 | disk_conf = rcu_dereference(mdev->ldev->disk_conf); |
2696 | if (nc) | 2790 | err = disk_conf_to_skb(skb, disk_conf, exclude_sensitive); |
2697 | err = net_conf_to_skb(skb, nc, exclude_sensitive); | 2791 | } |
2792 | if (!err) { | ||
2793 | struct net_conf *nc; | ||
2794 | |||
2795 | nc = rcu_dereference(mdev->tconn->net_conf); | ||
2796 | if (nc) | ||
2797 | err = net_conf_to_skb(skb, nc, exclude_sensitive); | ||
2798 | } | ||
2698 | rcu_read_unlock(); | 2799 | rcu_read_unlock(); |
2699 | if (err) | 2800 | if (err) |
2700 | goto nla_put_failure; | 2801 | goto nla_put_failure; |
diff --git a/drivers/block/drbd/drbd_receiver.c b/drivers/block/drbd/drbd_receiver.c index 4222affff488..cc29cd3bf78b 100644 --- a/drivers/block/drbd/drbd_receiver.c +++ b/drivers/block/drbd/drbd_receiver.c | |||
@@ -1039,6 +1039,8 @@ randomize: | |||
1039 | rcu_read_lock(); | 1039 | rcu_read_lock(); |
1040 | idr_for_each_entry(&tconn->volumes, mdev, vnr) { | 1040 | idr_for_each_entry(&tconn->volumes, mdev, vnr) { |
1041 | kref_get(&mdev->kref); | 1041 | kref_get(&mdev->kref); |
1042 | rcu_read_unlock(); | ||
1043 | |||
1042 | /* Prevent a race between resync-handshake and | 1044 | /* Prevent a race between resync-handshake and |
1043 | * being promoted to Primary. | 1045 | * being promoted to Primary. |
1044 | * | 1046 | * |
@@ -1049,8 +1051,6 @@ randomize: | |||
1049 | mutex_lock(mdev->state_mutex); | 1051 | mutex_lock(mdev->state_mutex); |
1050 | mutex_unlock(mdev->state_mutex); | 1052 | mutex_unlock(mdev->state_mutex); |
1051 | 1053 | ||
1052 | rcu_read_unlock(); | ||
1053 | |||
1054 | if (discard_my_data) | 1054 | if (discard_my_data) |
1055 | set_bit(DISCARD_MY_DATA, &mdev->flags); | 1055 | set_bit(DISCARD_MY_DATA, &mdev->flags); |
1056 | else | 1056 | else |
@@ -3545,7 +3545,7 @@ static int receive_sizes(struct drbd_tconn *tconn, struct packet_info *pi) | |||
3545 | { | 3545 | { |
3546 | struct drbd_conf *mdev; | 3546 | struct drbd_conf *mdev; |
3547 | struct p_sizes *p = pi->data; | 3547 | struct p_sizes *p = pi->data; |
3548 | enum determine_dev_size dd = unchanged; | 3548 | enum determine_dev_size dd = DS_UNCHANGED; |
3549 | sector_t p_size, p_usize, my_usize; | 3549 | sector_t p_size, p_usize, my_usize; |
3550 | int ldsc = 0; /* local disk size changed */ | 3550 | int ldsc = 0; /* local disk size changed */ |
3551 | enum dds_flags ddsf; | 3551 | enum dds_flags ddsf; |
@@ -3617,9 +3617,9 @@ static int receive_sizes(struct drbd_tconn *tconn, struct packet_info *pi) | |||
3617 | 3617 | ||
3618 | ddsf = be16_to_cpu(p->dds_flags); | 3618 | ddsf = be16_to_cpu(p->dds_flags); |
3619 | if (get_ldev(mdev)) { | 3619 | if (get_ldev(mdev)) { |
3620 | dd = drbd_determine_dev_size(mdev, ddsf); | 3620 | dd = drbd_determine_dev_size(mdev, ddsf, NULL); |
3621 | put_ldev(mdev); | 3621 | put_ldev(mdev); |
3622 | if (dd == dev_size_error) | 3622 | if (dd == DS_ERROR) |
3623 | return -EIO; | 3623 | return -EIO; |
3624 | drbd_md_sync(mdev); | 3624 | drbd_md_sync(mdev); |
3625 | } else { | 3625 | } else { |
@@ -3647,7 +3647,7 @@ static int receive_sizes(struct drbd_tconn *tconn, struct packet_info *pi) | |||
3647 | drbd_send_sizes(mdev, 0, ddsf); | 3647 | drbd_send_sizes(mdev, 0, ddsf); |
3648 | } | 3648 | } |
3649 | if (test_and_clear_bit(RESIZE_PENDING, &mdev->flags) || | 3649 | if (test_and_clear_bit(RESIZE_PENDING, &mdev->flags) || |
3650 | (dd == grew && mdev->state.conn == C_CONNECTED)) { | 3650 | (dd == DS_GREW && mdev->state.conn == C_CONNECTED)) { |
3651 | if (mdev->state.pdsk >= D_INCONSISTENT && | 3651 | if (mdev->state.pdsk >= D_INCONSISTENT && |
3652 | mdev->state.disk >= D_INCONSISTENT) { | 3652 | mdev->state.disk >= D_INCONSISTENT) { |
3653 | if (ddsf & DDSF_NO_RESYNC) | 3653 | if (ddsf & DDSF_NO_RESYNC) |
diff --git a/drivers/block/drbd/drbd_state.c b/drivers/block/drbd/drbd_state.c index 90c5be2b1d30..216d47b7e88b 100644 --- a/drivers/block/drbd/drbd_state.c +++ b/drivers/block/drbd/drbd_state.c | |||
@@ -1115,8 +1115,10 @@ __drbd_set_state(struct drbd_conf *mdev, union drbd_state ns, | |||
1115 | drbd_thread_restart_nowait(&mdev->tconn->receiver); | 1115 | drbd_thread_restart_nowait(&mdev->tconn->receiver); |
1116 | 1116 | ||
1117 | /* Resume AL writing if we get a connection */ | 1117 | /* Resume AL writing if we get a connection */ |
1118 | if (os.conn < C_CONNECTED && ns.conn >= C_CONNECTED) | 1118 | if (os.conn < C_CONNECTED && ns.conn >= C_CONNECTED) { |
1119 | drbd_resume_al(mdev); | 1119 | drbd_resume_al(mdev); |
1120 | mdev->tconn->connect_cnt++; | ||
1121 | } | ||
1120 | 1122 | ||
1121 | /* remember last attach time so request_timer_fn() won't | 1123 | /* remember last attach time so request_timer_fn() won't |
1122 | * kill newly established sessions while we are still trying to thaw | 1124 | * kill newly established sessions while we are still trying to thaw |
diff --git a/drivers/block/rsxx/core.c b/drivers/block/rsxx/core.c index 5af21f2db29c..6e85e21445eb 100644 --- a/drivers/block/rsxx/core.c +++ b/drivers/block/rsxx/core.c | |||
@@ -31,6 +31,8 @@ | |||
31 | #include <linux/slab.h> | 31 | #include <linux/slab.h> |
32 | #include <linux/bitops.h> | 32 | #include <linux/bitops.h> |
33 | #include <linux/delay.h> | 33 | #include <linux/delay.h> |
34 | #include <linux/debugfs.h> | ||
35 | #include <linux/seq_file.h> | ||
34 | 36 | ||
35 | #include <linux/genhd.h> | 37 | #include <linux/genhd.h> |
36 | #include <linux/idr.h> | 38 | #include <linux/idr.h> |
@@ -39,8 +41,9 @@ | |||
39 | #include "rsxx_cfg.h" | 41 | #include "rsxx_cfg.h" |
40 | 42 | ||
41 | #define NO_LEGACY 0 | 43 | #define NO_LEGACY 0 |
44 | #define SYNC_START_TIMEOUT (10 * 60) /* 10 minutes */ | ||
42 | 45 | ||
43 | MODULE_DESCRIPTION("IBM FlashSystem 70/80 PCIe SSD Device Driver"); | 46 | MODULE_DESCRIPTION("IBM Flash Adapter 900GB Full Height Device Driver"); |
44 | MODULE_AUTHOR("Joshua Morris/Philip Kelleher, IBM"); | 47 | MODULE_AUTHOR("Joshua Morris/Philip Kelleher, IBM"); |
45 | MODULE_LICENSE("GPL"); | 48 | MODULE_LICENSE("GPL"); |
46 | MODULE_VERSION(DRIVER_VERSION); | 49 | MODULE_VERSION(DRIVER_VERSION); |
@@ -49,9 +52,282 @@ static unsigned int force_legacy = NO_LEGACY; | |||
49 | module_param(force_legacy, uint, 0444); | 52 | module_param(force_legacy, uint, 0444); |
50 | MODULE_PARM_DESC(force_legacy, "Force the use of legacy type PCI interrupts"); | 53 | MODULE_PARM_DESC(force_legacy, "Force the use of legacy type PCI interrupts"); |
51 | 54 | ||
55 | static unsigned int sync_start = 1; | ||
56 | module_param(sync_start, uint, 0444); | ||
57 | MODULE_PARM_DESC(sync_start, "On by Default: Driver load will not complete " | ||
58 | "until the card startup has completed."); | ||
59 | |||
52 | static DEFINE_IDA(rsxx_disk_ida); | 60 | static DEFINE_IDA(rsxx_disk_ida); |
53 | static DEFINE_SPINLOCK(rsxx_ida_lock); | 61 | static DEFINE_SPINLOCK(rsxx_ida_lock); |
54 | 62 | ||
63 | /* --------------------Debugfs Setup ------------------- */ | ||
64 | |||
65 | struct rsxx_cram { | ||
66 | u32 f_pos; | ||
67 | u32 offset; | ||
68 | void *i_private; | ||
69 | }; | ||
70 | |||
71 | static int rsxx_attr_pci_regs_show(struct seq_file *m, void *p) | ||
72 | { | ||
73 | struct rsxx_cardinfo *card = m->private; | ||
74 | |||
75 | seq_printf(m, "HWID 0x%08x\n", | ||
76 | ioread32(card->regmap + HWID)); | ||
77 | seq_printf(m, "SCRATCH 0x%08x\n", | ||
78 | ioread32(card->regmap + SCRATCH)); | ||
79 | seq_printf(m, "IER 0x%08x\n", | ||
80 | ioread32(card->regmap + IER)); | ||
81 | seq_printf(m, "IPR 0x%08x\n", | ||
82 | ioread32(card->regmap + IPR)); | ||
83 | seq_printf(m, "CREG_CMD 0x%08x\n", | ||
84 | ioread32(card->regmap + CREG_CMD)); | ||
85 | seq_printf(m, "CREG_ADD 0x%08x\n", | ||
86 | ioread32(card->regmap + CREG_ADD)); | ||
87 | seq_printf(m, "CREG_CNT 0x%08x\n", | ||
88 | ioread32(card->regmap + CREG_CNT)); | ||
89 | seq_printf(m, "CREG_STAT 0x%08x\n", | ||
90 | ioread32(card->regmap + CREG_STAT)); | ||
91 | seq_printf(m, "CREG_DATA0 0x%08x\n", | ||
92 | ioread32(card->regmap + CREG_DATA0)); | ||
93 | seq_printf(m, "CREG_DATA1 0x%08x\n", | ||
94 | ioread32(card->regmap + CREG_DATA1)); | ||
95 | seq_printf(m, "CREG_DATA2 0x%08x\n", | ||
96 | ioread32(card->regmap + CREG_DATA2)); | ||
97 | seq_printf(m, "CREG_DATA3 0x%08x\n", | ||
98 | ioread32(card->regmap + CREG_DATA3)); | ||
99 | seq_printf(m, "CREG_DATA4 0x%08x\n", | ||
100 | ioread32(card->regmap + CREG_DATA4)); | ||
101 | seq_printf(m, "CREG_DATA5 0x%08x\n", | ||
102 | ioread32(card->regmap + CREG_DATA5)); | ||
103 | seq_printf(m, "CREG_DATA6 0x%08x\n", | ||
104 | ioread32(card->regmap + CREG_DATA6)); | ||
105 | seq_printf(m, "CREG_DATA7 0x%08x\n", | ||
106 | ioread32(card->regmap + CREG_DATA7)); | ||
107 | seq_printf(m, "INTR_COAL 0x%08x\n", | ||
108 | ioread32(card->regmap + INTR_COAL)); | ||
109 | seq_printf(m, "HW_ERROR 0x%08x\n", | ||
110 | ioread32(card->regmap + HW_ERROR)); | ||
111 | seq_printf(m, "DEBUG0 0x%08x\n", | ||
112 | ioread32(card->regmap + PCI_DEBUG0)); | ||
113 | seq_printf(m, "DEBUG1 0x%08x\n", | ||
114 | ioread32(card->regmap + PCI_DEBUG1)); | ||
115 | seq_printf(m, "DEBUG2 0x%08x\n", | ||
116 | ioread32(card->regmap + PCI_DEBUG2)); | ||
117 | seq_printf(m, "DEBUG3 0x%08x\n", | ||
118 | ioread32(card->regmap + PCI_DEBUG3)); | ||
119 | seq_printf(m, "DEBUG4 0x%08x\n", | ||
120 | ioread32(card->regmap + PCI_DEBUG4)); | ||
121 | seq_printf(m, "DEBUG5 0x%08x\n", | ||
122 | ioread32(card->regmap + PCI_DEBUG5)); | ||
123 | seq_printf(m, "DEBUG6 0x%08x\n", | ||
124 | ioread32(card->regmap + PCI_DEBUG6)); | ||
125 | seq_printf(m, "DEBUG7 0x%08x\n", | ||
126 | ioread32(card->regmap + PCI_DEBUG7)); | ||
127 | seq_printf(m, "RECONFIG 0x%08x\n", | ||
128 | ioread32(card->regmap + PCI_RECONFIG)); | ||
129 | |||
130 | return 0; | ||
131 | } | ||
132 | |||
133 | static int rsxx_attr_stats_show(struct seq_file *m, void *p) | ||
134 | { | ||
135 | struct rsxx_cardinfo *card = m->private; | ||
136 | int i; | ||
137 | |||
138 | for (i = 0; i < card->n_targets; i++) { | ||
139 | seq_printf(m, "Ctrl %d CRC Errors = %d\n", | ||
140 | i, card->ctrl[i].stats.crc_errors); | ||
141 | seq_printf(m, "Ctrl %d Hard Errors = %d\n", | ||
142 | i, card->ctrl[i].stats.hard_errors); | ||
143 | seq_printf(m, "Ctrl %d Soft Errors = %d\n", | ||
144 | i, card->ctrl[i].stats.soft_errors); | ||
145 | seq_printf(m, "Ctrl %d Writes Issued = %d\n", | ||
146 | i, card->ctrl[i].stats.writes_issued); | ||
147 | seq_printf(m, "Ctrl %d Writes Failed = %d\n", | ||
148 | i, card->ctrl[i].stats.writes_failed); | ||
149 | seq_printf(m, "Ctrl %d Reads Issued = %d\n", | ||
150 | i, card->ctrl[i].stats.reads_issued); | ||
151 | seq_printf(m, "Ctrl %d Reads Failed = %d\n", | ||
152 | i, card->ctrl[i].stats.reads_failed); | ||
153 | seq_printf(m, "Ctrl %d Reads Retried = %d\n", | ||
154 | i, card->ctrl[i].stats.reads_retried); | ||
155 | seq_printf(m, "Ctrl %d Discards Issued = %d\n", | ||
156 | i, card->ctrl[i].stats.discards_issued); | ||
157 | seq_printf(m, "Ctrl %d Discards Failed = %d\n", | ||
158 | i, card->ctrl[i].stats.discards_failed); | ||
159 | seq_printf(m, "Ctrl %d DMA SW Errors = %d\n", | ||
160 | i, card->ctrl[i].stats.dma_sw_err); | ||
161 | seq_printf(m, "Ctrl %d DMA HW Faults = %d\n", | ||
162 | i, card->ctrl[i].stats.dma_hw_fault); | ||
163 | seq_printf(m, "Ctrl %d DMAs Cancelled = %d\n", | ||
164 | i, card->ctrl[i].stats.dma_cancelled); | ||
165 | seq_printf(m, "Ctrl %d SW Queue Depth = %d\n", | ||
166 | i, card->ctrl[i].stats.sw_q_depth); | ||
167 | seq_printf(m, "Ctrl %d HW Queue Depth = %d\n", | ||
168 | i, atomic_read(&card->ctrl[i].stats.hw_q_depth)); | ||
169 | } | ||
170 | |||
171 | return 0; | ||
172 | } | ||
173 | |||
174 | static int rsxx_attr_stats_open(struct inode *inode, struct file *file) | ||
175 | { | ||
176 | return single_open(file, rsxx_attr_stats_show, inode->i_private); | ||
177 | } | ||
178 | |||
179 | static int rsxx_attr_pci_regs_open(struct inode *inode, struct file *file) | ||
180 | { | ||
181 | return single_open(file, rsxx_attr_pci_regs_show, inode->i_private); | ||
182 | } | ||
183 | |||
184 | static ssize_t rsxx_cram_read(struct file *fp, char __user *ubuf, | ||
185 | size_t cnt, loff_t *ppos) | ||
186 | { | ||
187 | struct rsxx_cram *info = fp->private_data; | ||
188 | struct rsxx_cardinfo *card = info->i_private; | ||
189 | char *buf; | ||
190 | int st; | ||
191 | |||
192 | buf = kzalloc(sizeof(*buf) * cnt, GFP_KERNEL); | ||
193 | if (!buf) | ||
194 | return -ENOMEM; | ||
195 | |||
196 | info->f_pos = (u32)*ppos + info->offset; | ||
197 | |||
198 | st = rsxx_creg_read(card, CREG_ADD_CRAM + info->f_pos, cnt, buf, 1); | ||
199 | if (st) | ||
200 | return st; | ||
201 | |||
202 | st = copy_to_user(ubuf, buf, cnt); | ||
203 | if (st) | ||
204 | return st; | ||
205 | |||
206 | info->offset += cnt; | ||
207 | |||
208 | kfree(buf); | ||
209 | |||
210 | return cnt; | ||
211 | } | ||
212 | |||
213 | static ssize_t rsxx_cram_write(struct file *fp, const char __user *ubuf, | ||
214 | size_t cnt, loff_t *ppos) | ||
215 | { | ||
216 | struct rsxx_cram *info = fp->private_data; | ||
217 | struct rsxx_cardinfo *card = info->i_private; | ||
218 | char *buf; | ||
219 | int st; | ||
220 | |||
221 | buf = kzalloc(sizeof(*buf) * cnt, GFP_KERNEL); | ||
222 | if (!buf) | ||
223 | return -ENOMEM; | ||
224 | |||
225 | st = copy_from_user(buf, ubuf, cnt); | ||
226 | if (st) | ||
227 | return st; | ||
228 | |||
229 | info->f_pos = (u32)*ppos + info->offset; | ||
230 | |||
231 | st = rsxx_creg_write(card, CREG_ADD_CRAM + info->f_pos, cnt, buf, 1); | ||
232 | if (st) | ||
233 | return st; | ||
234 | |||
235 | info->offset += cnt; | ||
236 | |||
237 | kfree(buf); | ||
238 | |||
239 | return cnt; | ||
240 | } | ||
241 | |||
242 | static int rsxx_cram_open(struct inode *inode, struct file *file) | ||
243 | { | ||
244 | struct rsxx_cram *info = kzalloc(sizeof(*info), GFP_KERNEL); | ||
245 | if (!info) | ||
246 | return -ENOMEM; | ||
247 | |||
248 | info->i_private = inode->i_private; | ||
249 | info->f_pos = file->f_pos; | ||
250 | file->private_data = info; | ||
251 | |||
252 | return 0; | ||
253 | } | ||
254 | |||
255 | static int rsxx_cram_release(struct inode *inode, struct file *file) | ||
256 | { | ||
257 | struct rsxx_cram *info = file->private_data; | ||
258 | |||
259 | if (!info) | ||
260 | return 0; | ||
261 | |||
262 | kfree(info); | ||
263 | file->private_data = NULL; | ||
264 | |||
265 | return 0; | ||
266 | } | ||
267 | |||
268 | static const struct file_operations debugfs_cram_fops = { | ||
269 | .owner = THIS_MODULE, | ||
270 | .open = rsxx_cram_open, | ||
271 | .read = rsxx_cram_read, | ||
272 | .write = rsxx_cram_write, | ||
273 | .release = rsxx_cram_release, | ||
274 | }; | ||
275 | |||
276 | static const struct file_operations debugfs_stats_fops = { | ||
277 | .owner = THIS_MODULE, | ||
278 | .open = rsxx_attr_stats_open, | ||
279 | .read = seq_read, | ||
280 | .llseek = seq_lseek, | ||
281 | .release = single_release, | ||
282 | }; | ||
283 | |||
284 | static const struct file_operations debugfs_pci_regs_fops = { | ||
285 | .owner = THIS_MODULE, | ||
286 | .open = rsxx_attr_pci_regs_open, | ||
287 | .read = seq_read, | ||
288 | .llseek = seq_lseek, | ||
289 | .release = single_release, | ||
290 | }; | ||
291 | |||
292 | static void rsxx_debugfs_dev_new(struct rsxx_cardinfo *card) | ||
293 | { | ||
294 | struct dentry *debugfs_stats; | ||
295 | struct dentry *debugfs_pci_regs; | ||
296 | struct dentry *debugfs_cram; | ||
297 | |||
298 | card->debugfs_dir = debugfs_create_dir(card->gendisk->disk_name, NULL); | ||
299 | if (IS_ERR_OR_NULL(card->debugfs_dir)) | ||
300 | goto failed_debugfs_dir; | ||
301 | |||
302 | debugfs_stats = debugfs_create_file("stats", S_IRUGO, | ||
303 | card->debugfs_dir, card, | ||
304 | &debugfs_stats_fops); | ||
305 | if (IS_ERR_OR_NULL(debugfs_stats)) | ||
306 | goto failed_debugfs_stats; | ||
307 | |||
308 | debugfs_pci_regs = debugfs_create_file("pci_regs", S_IRUGO, | ||
309 | card->debugfs_dir, card, | ||
310 | &debugfs_pci_regs_fops); | ||
311 | if (IS_ERR_OR_NULL(debugfs_pci_regs)) | ||
312 | goto failed_debugfs_pci_regs; | ||
313 | |||
314 | debugfs_cram = debugfs_create_file("cram", S_IRUGO | S_IWUSR, | ||
315 | card->debugfs_dir, card, | ||
316 | &debugfs_cram_fops); | ||
317 | if (IS_ERR_OR_NULL(debugfs_cram)) | ||
318 | goto failed_debugfs_cram; | ||
319 | |||
320 | return; | ||
321 | failed_debugfs_cram: | ||
322 | debugfs_remove(debugfs_pci_regs); | ||
323 | failed_debugfs_pci_regs: | ||
324 | debugfs_remove(debugfs_stats); | ||
325 | failed_debugfs_stats: | ||
326 | debugfs_remove(card->debugfs_dir); | ||
327 | failed_debugfs_dir: | ||
328 | card->debugfs_dir = NULL; | ||
329 | } | ||
330 | |||
55 | /*----------------- Interrupt Control & Handling -------------------*/ | 331 | /*----------------- Interrupt Control & Handling -------------------*/ |
56 | 332 | ||
57 | static void rsxx_mask_interrupts(struct rsxx_cardinfo *card) | 333 | static void rsxx_mask_interrupts(struct rsxx_cardinfo *card) |
@@ -163,12 +439,13 @@ static irqreturn_t rsxx_isr(int irq, void *pdata) | |||
163 | } | 439 | } |
164 | 440 | ||
165 | if (isr & CR_INTR_CREG) { | 441 | if (isr & CR_INTR_CREG) { |
166 | schedule_work(&card->creg_ctrl.done_work); | 442 | queue_work(card->creg_ctrl.creg_wq, |
443 | &card->creg_ctrl.done_work); | ||
167 | handled++; | 444 | handled++; |
168 | } | 445 | } |
169 | 446 | ||
170 | if (isr & CR_INTR_EVENT) { | 447 | if (isr & CR_INTR_EVENT) { |
171 | schedule_work(&card->event_work); | 448 | queue_work(card->event_wq, &card->event_work); |
172 | rsxx_disable_ier_and_isr(card, CR_INTR_EVENT); | 449 | rsxx_disable_ier_and_isr(card, CR_INTR_EVENT); |
173 | handled++; | 450 | handled++; |
174 | } | 451 | } |
@@ -329,7 +606,7 @@ static int rsxx_eeh_frozen(struct pci_dev *dev) | |||
329 | int i; | 606 | int i; |
330 | int st; | 607 | int st; |
331 | 608 | ||
332 | dev_warn(&dev->dev, "IBM FlashSystem PCI: preparing for slot reset.\n"); | 609 | dev_warn(&dev->dev, "IBM Flash Adapter PCI: preparing for slot reset.\n"); |
333 | 610 | ||
334 | card->eeh_state = 1; | 611 | card->eeh_state = 1; |
335 | rsxx_mask_interrupts(card); | 612 | rsxx_mask_interrupts(card); |
@@ -367,15 +644,26 @@ static void rsxx_eeh_failure(struct pci_dev *dev) | |||
367 | { | 644 | { |
368 | struct rsxx_cardinfo *card = pci_get_drvdata(dev); | 645 | struct rsxx_cardinfo *card = pci_get_drvdata(dev); |
369 | int i; | 646 | int i; |
647 | int cnt = 0; | ||
370 | 648 | ||
371 | dev_err(&dev->dev, "IBM FlashSystem PCI: disabling failed card.\n"); | 649 | dev_err(&dev->dev, "IBM Flash Adapter PCI: disabling failed card.\n"); |
372 | 650 | ||
373 | card->eeh_state = 1; | 651 | card->eeh_state = 1; |
652 | card->halt = 1; | ||
374 | 653 | ||
375 | for (i = 0; i < card->n_targets; i++) | 654 | for (i = 0; i < card->n_targets; i++) { |
376 | del_timer_sync(&card->ctrl[i].activity_timer); | 655 | spin_lock_bh(&card->ctrl[i].queue_lock); |
656 | cnt = rsxx_cleanup_dma_queue(&card->ctrl[i], | ||
657 | &card->ctrl[i].queue); | ||
658 | spin_unlock_bh(&card->ctrl[i].queue_lock); | ||
659 | |||
660 | cnt += rsxx_dma_cancel(&card->ctrl[i]); | ||
377 | 661 | ||
378 | rsxx_eeh_cancel_dmas(card); | 662 | if (cnt) |
663 | dev_info(CARD_TO_DEV(card), | ||
664 | "Freed %d queued DMAs on channel %d\n", | ||
665 | cnt, card->ctrl[i].id); | ||
666 | } | ||
379 | } | 667 | } |
380 | 668 | ||
381 | static int rsxx_eeh_fifo_flush_poll(struct rsxx_cardinfo *card) | 669 | static int rsxx_eeh_fifo_flush_poll(struct rsxx_cardinfo *card) |
@@ -432,7 +720,7 @@ static pci_ers_result_t rsxx_slot_reset(struct pci_dev *dev) | |||
432 | int st; | 720 | int st; |
433 | 721 | ||
434 | dev_warn(&dev->dev, | 722 | dev_warn(&dev->dev, |
435 | "IBM FlashSystem PCI: recovering from slot reset.\n"); | 723 | "IBM Flash Adapter PCI: recovering from slot reset.\n"); |
436 | 724 | ||
437 | st = pci_enable_device(dev); | 725 | st = pci_enable_device(dev); |
438 | if (st) | 726 | if (st) |
@@ -485,7 +773,7 @@ static pci_ers_result_t rsxx_slot_reset(struct pci_dev *dev) | |||
485 | &card->ctrl[i].issue_dma_work); | 773 | &card->ctrl[i].issue_dma_work); |
486 | } | 774 | } |
487 | 775 | ||
488 | dev_info(&dev->dev, "IBM FlashSystem PCI: recovery complete.\n"); | 776 | dev_info(&dev->dev, "IBM Flash Adapter PCI: recovery complete.\n"); |
489 | 777 | ||
490 | return PCI_ERS_RESULT_RECOVERED; | 778 | return PCI_ERS_RESULT_RECOVERED; |
491 | 779 | ||
@@ -528,6 +816,7 @@ static int rsxx_pci_probe(struct pci_dev *dev, | |||
528 | { | 816 | { |
529 | struct rsxx_cardinfo *card; | 817 | struct rsxx_cardinfo *card; |
530 | int st; | 818 | int st; |
819 | unsigned int sync_timeout; | ||
531 | 820 | ||
532 | dev_info(&dev->dev, "PCI-Flash SSD discovered\n"); | 821 | dev_info(&dev->dev, "PCI-Flash SSD discovered\n"); |
533 | 822 | ||
@@ -610,7 +899,11 @@ static int rsxx_pci_probe(struct pci_dev *dev, | |||
610 | } | 899 | } |
611 | 900 | ||
612 | /************* Setup Processor Command Interface *************/ | 901 | /************* Setup Processor Command Interface *************/ |
613 | rsxx_creg_setup(card); | 902 | st = rsxx_creg_setup(card); |
903 | if (st) { | ||
904 | dev_err(CARD_TO_DEV(card), "Failed to setup creg interface.\n"); | ||
905 | goto failed_creg_setup; | ||
906 | } | ||
614 | 907 | ||
615 | spin_lock_irq(&card->irq_lock); | 908 | spin_lock_irq(&card->irq_lock); |
616 | rsxx_enable_ier_and_isr(card, CR_INTR_CREG); | 909 | rsxx_enable_ier_and_isr(card, CR_INTR_CREG); |
@@ -650,6 +943,12 @@ static int rsxx_pci_probe(struct pci_dev *dev, | |||
650 | } | 943 | } |
651 | 944 | ||
652 | /************* Setup Card Event Handler *************/ | 945 | /************* Setup Card Event Handler *************/ |
946 | card->event_wq = create_singlethread_workqueue(DRIVER_NAME"_event"); | ||
947 | if (!card->event_wq) { | ||
948 | dev_err(CARD_TO_DEV(card), "Failed card event setup.\n"); | ||
949 | goto failed_event_handler; | ||
950 | } | ||
951 | |||
653 | INIT_WORK(&card->event_work, card_event_handler); | 952 | INIT_WORK(&card->event_work, card_event_handler); |
654 | 953 | ||
655 | st = rsxx_setup_dev(card); | 954 | st = rsxx_setup_dev(card); |
@@ -676,6 +975,33 @@ static int rsxx_pci_probe(struct pci_dev *dev, | |||
676 | if (st) | 975 | if (st) |
677 | dev_crit(CARD_TO_DEV(card), | 976 | dev_crit(CARD_TO_DEV(card), |
678 | "Failed issuing card startup\n"); | 977 | "Failed issuing card startup\n"); |
978 | if (sync_start) { | ||
979 | sync_timeout = SYNC_START_TIMEOUT; | ||
980 | |||
981 | dev_info(CARD_TO_DEV(card), | ||
982 | "Waiting for card to startup\n"); | ||
983 | |||
984 | do { | ||
985 | ssleep(1); | ||
986 | sync_timeout--; | ||
987 | |||
988 | rsxx_get_card_state(card, &card->state); | ||
989 | } while (sync_timeout && | ||
990 | (card->state == CARD_STATE_STARTING)); | ||
991 | |||
992 | if (card->state == CARD_STATE_STARTING) { | ||
993 | dev_warn(CARD_TO_DEV(card), | ||
994 | "Card startup timed out\n"); | ||
995 | card->size8 = 0; | ||
996 | } else { | ||
997 | dev_info(CARD_TO_DEV(card), | ||
998 | "card state: %s\n", | ||
999 | rsxx_card_state_to_str(card->state)); | ||
1000 | st = rsxx_get_card_size8(card, &card->size8); | ||
1001 | if (st) | ||
1002 | card->size8 = 0; | ||
1003 | } | ||
1004 | } | ||
679 | } else if (card->state == CARD_STATE_GOOD || | 1005 | } else if (card->state == CARD_STATE_GOOD || |
680 | card->state == CARD_STATE_RD_ONLY_FAULT) { | 1006 | card->state == CARD_STATE_RD_ONLY_FAULT) { |
681 | st = rsxx_get_card_size8(card, &card->size8); | 1007 | st = rsxx_get_card_size8(card, &card->size8); |
@@ -685,12 +1011,21 @@ static int rsxx_pci_probe(struct pci_dev *dev, | |||
685 | 1011 | ||
686 | rsxx_attach_dev(card); | 1012 | rsxx_attach_dev(card); |
687 | 1013 | ||
1014 | /************* Setup Debugfs *************/ | ||
1015 | rsxx_debugfs_dev_new(card); | ||
1016 | |||
688 | return 0; | 1017 | return 0; |
689 | 1018 | ||
690 | failed_create_dev: | 1019 | failed_create_dev: |
1020 | destroy_workqueue(card->event_wq); | ||
1021 | card->event_wq = NULL; | ||
1022 | failed_event_handler: | ||
691 | rsxx_dma_destroy(card); | 1023 | rsxx_dma_destroy(card); |
692 | failed_dma_setup: | 1024 | failed_dma_setup: |
693 | failed_compatiblity_check: | 1025 | failed_compatiblity_check: |
1026 | destroy_workqueue(card->creg_ctrl.creg_wq); | ||
1027 | card->creg_ctrl.creg_wq = NULL; | ||
1028 | failed_creg_setup: | ||
694 | spin_lock_irq(&card->irq_lock); | 1029 | spin_lock_irq(&card->irq_lock); |
695 | rsxx_disable_ier_and_isr(card, CR_INTR_ALL); | 1030 | rsxx_disable_ier_and_isr(card, CR_INTR_ALL); |
696 | spin_unlock_irq(&card->irq_lock); | 1031 | spin_unlock_irq(&card->irq_lock); |
@@ -756,6 +1091,8 @@ static void rsxx_pci_remove(struct pci_dev *dev) | |||
756 | /* Prevent work_structs from re-queuing themselves. */ | 1091 | /* Prevent work_structs from re-queuing themselves. */ |
757 | card->halt = 1; | 1092 | card->halt = 1; |
758 | 1093 | ||
1094 | debugfs_remove_recursive(card->debugfs_dir); | ||
1095 | |||
759 | free_irq(dev->irq, card); | 1096 | free_irq(dev->irq, card); |
760 | 1097 | ||
761 | if (!force_legacy) | 1098 | if (!force_legacy) |
diff --git a/drivers/block/rsxx/cregs.c b/drivers/block/rsxx/cregs.c index 4b5c020a0a65..926dce9c452f 100644 --- a/drivers/block/rsxx/cregs.c +++ b/drivers/block/rsxx/cregs.c | |||
@@ -431,6 +431,15 @@ static int __issue_creg_rw(struct rsxx_cardinfo *card, | |||
431 | *hw_stat = completion.creg_status; | 431 | *hw_stat = completion.creg_status; |
432 | 432 | ||
433 | if (completion.st) { | 433 | if (completion.st) { |
434 | /* | ||
435 | * This read is needed to verify that there has not been any | ||
436 | * extreme errors that might have occurred, i.e. EEH. The | ||
437 | * function iowrite32 will not detect EEH errors, so it is | ||
438 | * necessary that we recover if such an error is the reason | ||
439 | * for the timeout. This is a dummy read. | ||
440 | */ | ||
441 | ioread32(card->regmap + SCRATCH); | ||
442 | |||
434 | dev_warn(CARD_TO_DEV(card), | 443 | dev_warn(CARD_TO_DEV(card), |
435 | "creg command failed(%d x%08x)\n", | 444 | "creg command failed(%d x%08x)\n", |
436 | completion.st, addr); | 445 | completion.st, addr); |
@@ -727,6 +736,11 @@ int rsxx_creg_setup(struct rsxx_cardinfo *card) | |||
727 | { | 736 | { |
728 | card->creg_ctrl.active_cmd = NULL; | 737 | card->creg_ctrl.active_cmd = NULL; |
729 | 738 | ||
739 | card->creg_ctrl.creg_wq = | ||
740 | create_singlethread_workqueue(DRIVER_NAME"_creg"); | ||
741 | if (!card->creg_ctrl.creg_wq) | ||
742 | return -ENOMEM; | ||
743 | |||
730 | INIT_WORK(&card->creg_ctrl.done_work, creg_cmd_done); | 744 | INIT_WORK(&card->creg_ctrl.done_work, creg_cmd_done); |
731 | mutex_init(&card->creg_ctrl.reset_lock); | 745 | mutex_init(&card->creg_ctrl.reset_lock); |
732 | INIT_LIST_HEAD(&card->creg_ctrl.queue); | 746 | INIT_LIST_HEAD(&card->creg_ctrl.queue); |
diff --git a/drivers/block/rsxx/dev.c b/drivers/block/rsxx/dev.c index 4346d17d2949..d7af441880be 100644 --- a/drivers/block/rsxx/dev.c +++ b/drivers/block/rsxx/dev.c | |||
@@ -155,7 +155,8 @@ static void bio_dma_done_cb(struct rsxx_cardinfo *card, | |||
155 | atomic_set(&meta->error, 1); | 155 | atomic_set(&meta->error, 1); |
156 | 156 | ||
157 | if (atomic_dec_and_test(&meta->pending_dmas)) { | 157 | if (atomic_dec_and_test(&meta->pending_dmas)) { |
158 | disk_stats_complete(card, meta->bio, meta->start_time); | 158 | if (!card->eeh_state && card->gendisk) |
159 | disk_stats_complete(card, meta->bio, meta->start_time); | ||
159 | 160 | ||
160 | bio_endio(meta->bio, atomic_read(&meta->error) ? -EIO : 0); | 161 | bio_endio(meta->bio, atomic_read(&meta->error) ? -EIO : 0); |
161 | kmem_cache_free(bio_meta_pool, meta); | 162 | kmem_cache_free(bio_meta_pool, meta); |
@@ -170,6 +171,12 @@ static void rsxx_make_request(struct request_queue *q, struct bio *bio) | |||
170 | 171 | ||
171 | might_sleep(); | 172 | might_sleep(); |
172 | 173 | ||
174 | if (!card) | ||
175 | goto req_err; | ||
176 | |||
177 | if (bio->bi_sector + (bio->bi_size >> 9) > get_capacity(card->gendisk)) | ||
178 | goto req_err; | ||
179 | |||
173 | if (unlikely(card->halt)) { | 180 | if (unlikely(card->halt)) { |
174 | st = -EFAULT; | 181 | st = -EFAULT; |
175 | goto req_err; | 182 | goto req_err; |
@@ -196,7 +203,8 @@ static void rsxx_make_request(struct request_queue *q, struct bio *bio) | |||
196 | atomic_set(&bio_meta->pending_dmas, 0); | 203 | atomic_set(&bio_meta->pending_dmas, 0); |
197 | bio_meta->start_time = jiffies; | 204 | bio_meta->start_time = jiffies; |
198 | 205 | ||
199 | disk_stats_start(card, bio); | 206 | if (!unlikely(card->halt)) |
207 | disk_stats_start(card, bio); | ||
200 | 208 | ||
201 | dev_dbg(CARD_TO_DEV(card), "BIO[%c]: meta: %p addr8: x%llx size: %d\n", | 209 | dev_dbg(CARD_TO_DEV(card), "BIO[%c]: meta: %p addr8: x%llx size: %d\n", |
202 | bio_data_dir(bio) ? 'W' : 'R', bio_meta, | 210 | bio_data_dir(bio) ? 'W' : 'R', bio_meta, |
@@ -225,24 +233,6 @@ static bool rsxx_discard_supported(struct rsxx_cardinfo *card) | |||
225 | return (pci_rev >= RSXX_DISCARD_SUPPORT); | 233 | return (pci_rev >= RSXX_DISCARD_SUPPORT); |
226 | } | 234 | } |
227 | 235 | ||
228 | static unsigned short rsxx_get_logical_block_size( | ||
229 | struct rsxx_cardinfo *card) | ||
230 | { | ||
231 | u32 capabilities = 0; | ||
232 | int st; | ||
233 | |||
234 | st = rsxx_get_card_capabilities(card, &capabilities); | ||
235 | if (st) | ||
236 | dev_warn(CARD_TO_DEV(card), | ||
237 | "Failed reading card capabilities register\n"); | ||
238 | |||
239 | /* Earlier firmware did not have support for 512 byte accesses */ | ||
240 | if (capabilities & CARD_CAP_SUBPAGE_WRITES) | ||
241 | return 512; | ||
242 | else | ||
243 | return RSXX_HW_BLK_SIZE; | ||
244 | } | ||
245 | |||
246 | int rsxx_attach_dev(struct rsxx_cardinfo *card) | 236 | int rsxx_attach_dev(struct rsxx_cardinfo *card) |
247 | { | 237 | { |
248 | mutex_lock(&card->dev_lock); | 238 | mutex_lock(&card->dev_lock); |
@@ -305,7 +295,7 @@ int rsxx_setup_dev(struct rsxx_cardinfo *card) | |||
305 | return -ENOMEM; | 295 | return -ENOMEM; |
306 | } | 296 | } |
307 | 297 | ||
308 | blk_size = rsxx_get_logical_block_size(card); | 298 | blk_size = card->config.data.block_size; |
309 | 299 | ||
310 | blk_queue_make_request(card->queue, rsxx_make_request); | 300 | blk_queue_make_request(card->queue, rsxx_make_request); |
311 | blk_queue_bounce_limit(card->queue, BLK_BOUNCE_ANY); | 301 | blk_queue_bounce_limit(card->queue, BLK_BOUNCE_ANY); |
@@ -347,6 +337,7 @@ void rsxx_destroy_dev(struct rsxx_cardinfo *card) | |||
347 | card->gendisk = NULL; | 337 | card->gendisk = NULL; |
348 | 338 | ||
349 | blk_cleanup_queue(card->queue); | 339 | blk_cleanup_queue(card->queue); |
340 | card->queue->queuedata = NULL; | ||
350 | unregister_blkdev(card->major, DRIVER_NAME); | 341 | unregister_blkdev(card->major, DRIVER_NAME); |
351 | } | 342 | } |
352 | 343 | ||
diff --git a/drivers/block/rsxx/dma.c b/drivers/block/rsxx/dma.c index 0607513cfb41..bed32f16b084 100644 --- a/drivers/block/rsxx/dma.c +++ b/drivers/block/rsxx/dma.c | |||
@@ -245,6 +245,22 @@ static void rsxx_complete_dma(struct rsxx_dma_ctrl *ctrl, | |||
245 | kmem_cache_free(rsxx_dma_pool, dma); | 245 | kmem_cache_free(rsxx_dma_pool, dma); |
246 | } | 246 | } |
247 | 247 | ||
248 | int rsxx_cleanup_dma_queue(struct rsxx_dma_ctrl *ctrl, | ||
249 | struct list_head *q) | ||
250 | { | ||
251 | struct rsxx_dma *dma; | ||
252 | struct rsxx_dma *tmp; | ||
253 | int cnt = 0; | ||
254 | |||
255 | list_for_each_entry_safe(dma, tmp, q, list) { | ||
256 | list_del(&dma->list); | ||
257 | rsxx_complete_dma(ctrl, dma, DMA_CANCELLED); | ||
258 | cnt++; | ||
259 | } | ||
260 | |||
261 | return cnt; | ||
262 | } | ||
263 | |||
248 | static void rsxx_requeue_dma(struct rsxx_dma_ctrl *ctrl, | 264 | static void rsxx_requeue_dma(struct rsxx_dma_ctrl *ctrl, |
249 | struct rsxx_dma *dma) | 265 | struct rsxx_dma *dma) |
250 | { | 266 | { |
@@ -252,9 +268,10 @@ static void rsxx_requeue_dma(struct rsxx_dma_ctrl *ctrl, | |||
252 | * Requeued DMAs go to the front of the queue so they are issued | 268 | * Requeued DMAs go to the front of the queue so they are issued |
253 | * first. | 269 | * first. |
254 | */ | 270 | */ |
255 | spin_lock(&ctrl->queue_lock); | 271 | spin_lock_bh(&ctrl->queue_lock); |
272 | ctrl->stats.sw_q_depth++; | ||
256 | list_add(&dma->list, &ctrl->queue); | 273 | list_add(&dma->list, &ctrl->queue); |
257 | spin_unlock(&ctrl->queue_lock); | 274 | spin_unlock_bh(&ctrl->queue_lock); |
258 | } | 275 | } |
259 | 276 | ||
260 | static void rsxx_handle_dma_error(struct rsxx_dma_ctrl *ctrl, | 277 | static void rsxx_handle_dma_error(struct rsxx_dma_ctrl *ctrl, |
@@ -329,6 +346,7 @@ static void rsxx_handle_dma_error(struct rsxx_dma_ctrl *ctrl, | |||
329 | static void dma_engine_stalled(unsigned long data) | 346 | static void dma_engine_stalled(unsigned long data) |
330 | { | 347 | { |
331 | struct rsxx_dma_ctrl *ctrl = (struct rsxx_dma_ctrl *)data; | 348 | struct rsxx_dma_ctrl *ctrl = (struct rsxx_dma_ctrl *)data; |
349 | int cnt; | ||
332 | 350 | ||
333 | if (atomic_read(&ctrl->stats.hw_q_depth) == 0 || | 351 | if (atomic_read(&ctrl->stats.hw_q_depth) == 0 || |
334 | unlikely(ctrl->card->eeh_state)) | 352 | unlikely(ctrl->card->eeh_state)) |
@@ -349,18 +367,28 @@ static void dma_engine_stalled(unsigned long data) | |||
349 | "DMA channel %d has stalled, faulting interface.\n", | 367 | "DMA channel %d has stalled, faulting interface.\n", |
350 | ctrl->id); | 368 | ctrl->id); |
351 | ctrl->card->dma_fault = 1; | 369 | ctrl->card->dma_fault = 1; |
370 | |||
371 | /* Clean up the DMA queue */ | ||
372 | spin_lock(&ctrl->queue_lock); | ||
373 | cnt = rsxx_cleanup_dma_queue(ctrl, &ctrl->queue); | ||
374 | spin_unlock(&ctrl->queue_lock); | ||
375 | |||
376 | cnt += rsxx_dma_cancel(ctrl); | ||
377 | |||
378 | if (cnt) | ||
379 | dev_info(CARD_TO_DEV(ctrl->card), | ||
380 | "Freed %d queued DMAs on channel %d\n", | ||
381 | cnt, ctrl->id); | ||
352 | } | 382 | } |
353 | } | 383 | } |
354 | 384 | ||
355 | static void rsxx_issue_dmas(struct work_struct *work) | 385 | static void rsxx_issue_dmas(struct rsxx_dma_ctrl *ctrl) |
356 | { | 386 | { |
357 | struct rsxx_dma_ctrl *ctrl; | ||
358 | struct rsxx_dma *dma; | 387 | struct rsxx_dma *dma; |
359 | int tag; | 388 | int tag; |
360 | int cmds_pending = 0; | 389 | int cmds_pending = 0; |
361 | struct hw_cmd *hw_cmd_buf; | 390 | struct hw_cmd *hw_cmd_buf; |
362 | 391 | ||
363 | ctrl = container_of(work, struct rsxx_dma_ctrl, issue_dma_work); | ||
364 | hw_cmd_buf = ctrl->cmd.buf; | 392 | hw_cmd_buf = ctrl->cmd.buf; |
365 | 393 | ||
366 | if (unlikely(ctrl->card->halt) || | 394 | if (unlikely(ctrl->card->halt) || |
@@ -368,22 +396,22 @@ static void rsxx_issue_dmas(struct work_struct *work) | |||
368 | return; | 396 | return; |
369 | 397 | ||
370 | while (1) { | 398 | while (1) { |
371 | spin_lock(&ctrl->queue_lock); | 399 | spin_lock_bh(&ctrl->queue_lock); |
372 | if (list_empty(&ctrl->queue)) { | 400 | if (list_empty(&ctrl->queue)) { |
373 | spin_unlock(&ctrl->queue_lock); | 401 | spin_unlock_bh(&ctrl->queue_lock); |
374 | break; | 402 | break; |
375 | } | 403 | } |
376 | spin_unlock(&ctrl->queue_lock); | 404 | spin_unlock_bh(&ctrl->queue_lock); |
377 | 405 | ||
378 | tag = pop_tracker(ctrl->trackers); | 406 | tag = pop_tracker(ctrl->trackers); |
379 | if (tag == -1) | 407 | if (tag == -1) |
380 | break; | 408 | break; |
381 | 409 | ||
382 | spin_lock(&ctrl->queue_lock); | 410 | spin_lock_bh(&ctrl->queue_lock); |
383 | dma = list_entry(ctrl->queue.next, struct rsxx_dma, list); | 411 | dma = list_entry(ctrl->queue.next, struct rsxx_dma, list); |
384 | list_del(&dma->list); | 412 | list_del(&dma->list); |
385 | ctrl->stats.sw_q_depth--; | 413 | ctrl->stats.sw_q_depth--; |
386 | spin_unlock(&ctrl->queue_lock); | 414 | spin_unlock_bh(&ctrl->queue_lock); |
387 | 415 | ||
388 | /* | 416 | /* |
389 | * This will catch any DMAs that slipped in right before the | 417 | * This will catch any DMAs that slipped in right before the |
@@ -440,9 +468,8 @@ static void rsxx_issue_dmas(struct work_struct *work) | |||
440 | } | 468 | } |
441 | } | 469 | } |
442 | 470 | ||
443 | static void rsxx_dma_done(struct work_struct *work) | 471 | static void rsxx_dma_done(struct rsxx_dma_ctrl *ctrl) |
444 | { | 472 | { |
445 | struct rsxx_dma_ctrl *ctrl; | ||
446 | struct rsxx_dma *dma; | 473 | struct rsxx_dma *dma; |
447 | unsigned long flags; | 474 | unsigned long flags; |
448 | u16 count; | 475 | u16 count; |
@@ -450,7 +477,6 @@ static void rsxx_dma_done(struct work_struct *work) | |||
450 | u8 tag; | 477 | u8 tag; |
451 | struct hw_status *hw_st_buf; | 478 | struct hw_status *hw_st_buf; |
452 | 479 | ||
453 | ctrl = container_of(work, struct rsxx_dma_ctrl, dma_done_work); | ||
454 | hw_st_buf = ctrl->status.buf; | 480 | hw_st_buf = ctrl->status.buf; |
455 | 481 | ||
456 | if (unlikely(ctrl->card->halt) || | 482 | if (unlikely(ctrl->card->halt) || |
@@ -520,33 +546,32 @@ static void rsxx_dma_done(struct work_struct *work) | |||
520 | rsxx_enable_ier(ctrl->card, CR_INTR_DMA(ctrl->id)); | 546 | rsxx_enable_ier(ctrl->card, CR_INTR_DMA(ctrl->id)); |
521 | spin_unlock_irqrestore(&ctrl->card->irq_lock, flags); | 547 | spin_unlock_irqrestore(&ctrl->card->irq_lock, flags); |
522 | 548 | ||
523 | spin_lock(&ctrl->queue_lock); | 549 | spin_lock_bh(&ctrl->queue_lock); |
524 | if (ctrl->stats.sw_q_depth) | 550 | if (ctrl->stats.sw_q_depth) |
525 | queue_work(ctrl->issue_wq, &ctrl->issue_dma_work); | 551 | queue_work(ctrl->issue_wq, &ctrl->issue_dma_work); |
526 | spin_unlock(&ctrl->queue_lock); | 552 | spin_unlock_bh(&ctrl->queue_lock); |
527 | } | 553 | } |
528 | 554 | ||
529 | static int rsxx_cleanup_dma_queue(struct rsxx_cardinfo *card, | 555 | static void rsxx_schedule_issue(struct work_struct *work) |
530 | struct list_head *q) | ||
531 | { | 556 | { |
532 | struct rsxx_dma *dma; | 557 | struct rsxx_dma_ctrl *ctrl; |
533 | struct rsxx_dma *tmp; | ||
534 | int cnt = 0; | ||
535 | 558 | ||
536 | list_for_each_entry_safe(dma, tmp, q, list) { | 559 | ctrl = container_of(work, struct rsxx_dma_ctrl, issue_dma_work); |
537 | list_del(&dma->list); | ||
538 | 560 | ||
539 | if (dma->dma_addr) | 561 | mutex_lock(&ctrl->work_lock); |
540 | pci_unmap_page(card->dev, dma->dma_addr, | 562 | rsxx_issue_dmas(ctrl); |
541 | get_dma_size(dma), | 563 | mutex_unlock(&ctrl->work_lock); |
542 | (dma->cmd == HW_CMD_BLK_WRITE) ? | 564 | } |
543 | PCI_DMA_TODEVICE : | ||
544 | PCI_DMA_FROMDEVICE); | ||
545 | kmem_cache_free(rsxx_dma_pool, dma); | ||
546 | cnt++; | ||
547 | } | ||
548 | 565 | ||
549 | return cnt; | 566 | static void rsxx_schedule_done(struct work_struct *work) |
567 | { | ||
568 | struct rsxx_dma_ctrl *ctrl; | ||
569 | |||
570 | ctrl = container_of(work, struct rsxx_dma_ctrl, dma_done_work); | ||
571 | |||
572 | mutex_lock(&ctrl->work_lock); | ||
573 | rsxx_dma_done(ctrl); | ||
574 | mutex_unlock(&ctrl->work_lock); | ||
550 | } | 575 | } |
551 | 576 | ||
552 | static int rsxx_queue_discard(struct rsxx_cardinfo *card, | 577 | static int rsxx_queue_discard(struct rsxx_cardinfo *card, |
@@ -698,10 +723,10 @@ int rsxx_dma_queue_bio(struct rsxx_cardinfo *card, | |||
698 | 723 | ||
699 | for (i = 0; i < card->n_targets; i++) { | 724 | for (i = 0; i < card->n_targets; i++) { |
700 | if (!list_empty(&dma_list[i])) { | 725 | if (!list_empty(&dma_list[i])) { |
701 | spin_lock(&card->ctrl[i].queue_lock); | 726 | spin_lock_bh(&card->ctrl[i].queue_lock); |
702 | card->ctrl[i].stats.sw_q_depth += dma_cnt[i]; | 727 | card->ctrl[i].stats.sw_q_depth += dma_cnt[i]; |
703 | list_splice_tail(&dma_list[i], &card->ctrl[i].queue); | 728 | list_splice_tail(&dma_list[i], &card->ctrl[i].queue); |
704 | spin_unlock(&card->ctrl[i].queue_lock); | 729 | spin_unlock_bh(&card->ctrl[i].queue_lock); |
705 | 730 | ||
706 | queue_work(card->ctrl[i].issue_wq, | 731 | queue_work(card->ctrl[i].issue_wq, |
707 | &card->ctrl[i].issue_dma_work); | 732 | &card->ctrl[i].issue_dma_work); |
@@ -711,8 +736,11 @@ int rsxx_dma_queue_bio(struct rsxx_cardinfo *card, | |||
711 | return 0; | 736 | return 0; |
712 | 737 | ||
713 | bvec_err: | 738 | bvec_err: |
714 | for (i = 0; i < card->n_targets; i++) | 739 | for (i = 0; i < card->n_targets; i++) { |
715 | rsxx_cleanup_dma_queue(card, &dma_list[i]); | 740 | spin_lock_bh(&card->ctrl[i].queue_lock); |
741 | rsxx_cleanup_dma_queue(&card->ctrl[i], &dma_list[i]); | ||
742 | spin_unlock_bh(&card->ctrl[i].queue_lock); | ||
743 | } | ||
716 | 744 | ||
717 | return st; | 745 | return st; |
718 | } | 746 | } |
@@ -780,6 +808,7 @@ static int rsxx_dma_ctrl_init(struct pci_dev *dev, | |||
780 | spin_lock_init(&ctrl->trackers->lock); | 808 | spin_lock_init(&ctrl->trackers->lock); |
781 | 809 | ||
782 | spin_lock_init(&ctrl->queue_lock); | 810 | spin_lock_init(&ctrl->queue_lock); |
811 | mutex_init(&ctrl->work_lock); | ||
783 | INIT_LIST_HEAD(&ctrl->queue); | 812 | INIT_LIST_HEAD(&ctrl->queue); |
784 | 813 | ||
785 | setup_timer(&ctrl->activity_timer, dma_engine_stalled, | 814 | setup_timer(&ctrl->activity_timer, dma_engine_stalled, |
@@ -793,8 +822,8 @@ static int rsxx_dma_ctrl_init(struct pci_dev *dev, | |||
793 | if (!ctrl->done_wq) | 822 | if (!ctrl->done_wq) |
794 | return -ENOMEM; | 823 | return -ENOMEM; |
795 | 824 | ||
796 | INIT_WORK(&ctrl->issue_dma_work, rsxx_issue_dmas); | 825 | INIT_WORK(&ctrl->issue_dma_work, rsxx_schedule_issue); |
797 | INIT_WORK(&ctrl->dma_done_work, rsxx_dma_done); | 826 | INIT_WORK(&ctrl->dma_done_work, rsxx_schedule_done); |
798 | 827 | ||
799 | st = rsxx_hw_buffers_init(dev, ctrl); | 828 | st = rsxx_hw_buffers_init(dev, ctrl); |
800 | if (st) | 829 | if (st) |
@@ -918,13 +947,30 @@ failed_dma_setup: | |||
918 | return st; | 947 | return st; |
919 | } | 948 | } |
920 | 949 | ||
950 | int rsxx_dma_cancel(struct rsxx_dma_ctrl *ctrl) | ||
951 | { | ||
952 | struct rsxx_dma *dma; | ||
953 | int i; | ||
954 | int cnt = 0; | ||
955 | |||
956 | /* Clean up issued DMAs */ | ||
957 | for (i = 0; i < RSXX_MAX_OUTSTANDING_CMDS; i++) { | ||
958 | dma = get_tracker_dma(ctrl->trackers, i); | ||
959 | if (dma) { | ||
960 | atomic_dec(&ctrl->stats.hw_q_depth); | ||
961 | rsxx_complete_dma(ctrl, dma, DMA_CANCELLED); | ||
962 | push_tracker(ctrl->trackers, i); | ||
963 | cnt++; | ||
964 | } | ||
965 | } | ||
966 | |||
967 | return cnt; | ||
968 | } | ||
921 | 969 | ||
922 | void rsxx_dma_destroy(struct rsxx_cardinfo *card) | 970 | void rsxx_dma_destroy(struct rsxx_cardinfo *card) |
923 | { | 971 | { |
924 | struct rsxx_dma_ctrl *ctrl; | 972 | struct rsxx_dma_ctrl *ctrl; |
925 | struct rsxx_dma *dma; | 973 | int i; |
926 | int i, j; | ||
927 | int cnt = 0; | ||
928 | 974 | ||
929 | for (i = 0; i < card->n_targets; i++) { | 975 | for (i = 0; i < card->n_targets; i++) { |
930 | ctrl = &card->ctrl[i]; | 976 | ctrl = &card->ctrl[i]; |
@@ -943,33 +989,11 @@ void rsxx_dma_destroy(struct rsxx_cardinfo *card) | |||
943 | del_timer_sync(&ctrl->activity_timer); | 989 | del_timer_sync(&ctrl->activity_timer); |
944 | 990 | ||
945 | /* Clean up the DMA queue */ | 991 | /* Clean up the DMA queue */ |
946 | spin_lock(&ctrl->queue_lock); | 992 | spin_lock_bh(&ctrl->queue_lock); |
947 | cnt = rsxx_cleanup_dma_queue(card, &ctrl->queue); | 993 | rsxx_cleanup_dma_queue(ctrl, &ctrl->queue); |
948 | spin_unlock(&ctrl->queue_lock); | 994 | spin_unlock_bh(&ctrl->queue_lock); |
949 | |||
950 | if (cnt) | ||
951 | dev_info(CARD_TO_DEV(card), | ||
952 | "Freed %d queued DMAs on channel %d\n", | ||
953 | cnt, i); | ||
954 | |||
955 | /* Clean up issued DMAs */ | ||
956 | for (j = 0; j < RSXX_MAX_OUTSTANDING_CMDS; j++) { | ||
957 | dma = get_tracker_dma(ctrl->trackers, j); | ||
958 | if (dma) { | ||
959 | pci_unmap_page(card->dev, dma->dma_addr, | ||
960 | get_dma_size(dma), | ||
961 | (dma->cmd == HW_CMD_BLK_WRITE) ? | ||
962 | PCI_DMA_TODEVICE : | ||
963 | PCI_DMA_FROMDEVICE); | ||
964 | kmem_cache_free(rsxx_dma_pool, dma); | ||
965 | cnt++; | ||
966 | } | ||
967 | } | ||
968 | 995 | ||
969 | if (cnt) | 996 | rsxx_dma_cancel(ctrl); |
970 | dev_info(CARD_TO_DEV(card), | ||
971 | "Freed %d pending DMAs on channel %d\n", | ||
972 | cnt, i); | ||
973 | 997 | ||
974 | vfree(ctrl->trackers); | 998 | vfree(ctrl->trackers); |
975 | 999 | ||
@@ -1013,7 +1037,7 @@ int rsxx_eeh_save_issued_dmas(struct rsxx_cardinfo *card) | |||
1013 | cnt++; | 1037 | cnt++; |
1014 | } | 1038 | } |
1015 | 1039 | ||
1016 | spin_lock(&card->ctrl[i].queue_lock); | 1040 | spin_lock_bh(&card->ctrl[i].queue_lock); |
1017 | list_splice(&issued_dmas[i], &card->ctrl[i].queue); | 1041 | list_splice(&issued_dmas[i], &card->ctrl[i].queue); |
1018 | 1042 | ||
1019 | atomic_sub(cnt, &card->ctrl[i].stats.hw_q_depth); | 1043 | atomic_sub(cnt, &card->ctrl[i].stats.hw_q_depth); |
@@ -1028,7 +1052,7 @@ int rsxx_eeh_save_issued_dmas(struct rsxx_cardinfo *card) | |||
1028 | PCI_DMA_TODEVICE : | 1052 | PCI_DMA_TODEVICE : |
1029 | PCI_DMA_FROMDEVICE); | 1053 | PCI_DMA_FROMDEVICE); |
1030 | } | 1054 | } |
1031 | spin_unlock(&card->ctrl[i].queue_lock); | 1055 | spin_unlock_bh(&card->ctrl[i].queue_lock); |
1032 | } | 1056 | } |
1033 | 1057 | ||
1034 | kfree(issued_dmas); | 1058 | kfree(issued_dmas); |
@@ -1036,30 +1060,13 @@ int rsxx_eeh_save_issued_dmas(struct rsxx_cardinfo *card) | |||
1036 | return 0; | 1060 | return 0; |
1037 | } | 1061 | } |
1038 | 1062 | ||
1039 | void rsxx_eeh_cancel_dmas(struct rsxx_cardinfo *card) | ||
1040 | { | ||
1041 | struct rsxx_dma *dma; | ||
1042 | struct rsxx_dma *tmp; | ||
1043 | int i; | ||
1044 | |||
1045 | for (i = 0; i < card->n_targets; i++) { | ||
1046 | spin_lock(&card->ctrl[i].queue_lock); | ||
1047 | list_for_each_entry_safe(dma, tmp, &card->ctrl[i].queue, list) { | ||
1048 | list_del(&dma->list); | ||
1049 | |||
1050 | rsxx_complete_dma(&card->ctrl[i], dma, DMA_CANCELLED); | ||
1051 | } | ||
1052 | spin_unlock(&card->ctrl[i].queue_lock); | ||
1053 | } | ||
1054 | } | ||
1055 | |||
1056 | int rsxx_eeh_remap_dmas(struct rsxx_cardinfo *card) | 1063 | int rsxx_eeh_remap_dmas(struct rsxx_cardinfo *card) |
1057 | { | 1064 | { |
1058 | struct rsxx_dma *dma; | 1065 | struct rsxx_dma *dma; |
1059 | int i; | 1066 | int i; |
1060 | 1067 | ||
1061 | for (i = 0; i < card->n_targets; i++) { | 1068 | for (i = 0; i < card->n_targets; i++) { |
1062 | spin_lock(&card->ctrl[i].queue_lock); | 1069 | spin_lock_bh(&card->ctrl[i].queue_lock); |
1063 | list_for_each_entry(dma, &card->ctrl[i].queue, list) { | 1070 | list_for_each_entry(dma, &card->ctrl[i].queue, list) { |
1064 | dma->dma_addr = pci_map_page(card->dev, dma->page, | 1071 | dma->dma_addr = pci_map_page(card->dev, dma->page, |
1065 | dma->pg_off, get_dma_size(dma), | 1072 | dma->pg_off, get_dma_size(dma), |
@@ -1067,12 +1074,12 @@ int rsxx_eeh_remap_dmas(struct rsxx_cardinfo *card) | |||
1067 | PCI_DMA_TODEVICE : | 1074 | PCI_DMA_TODEVICE : |
1068 | PCI_DMA_FROMDEVICE); | 1075 | PCI_DMA_FROMDEVICE); |
1069 | if (!dma->dma_addr) { | 1076 | if (!dma->dma_addr) { |
1070 | spin_unlock(&card->ctrl[i].queue_lock); | 1077 | spin_unlock_bh(&card->ctrl[i].queue_lock); |
1071 | kmem_cache_free(rsxx_dma_pool, dma); | 1078 | kmem_cache_free(rsxx_dma_pool, dma); |
1072 | return -ENOMEM; | 1079 | return -ENOMEM; |
1073 | } | 1080 | } |
1074 | } | 1081 | } |
1075 | spin_unlock(&card->ctrl[i].queue_lock); | 1082 | spin_unlock_bh(&card->ctrl[i].queue_lock); |
1076 | } | 1083 | } |
1077 | 1084 | ||
1078 | return 0; | 1085 | return 0; |
diff --git a/drivers/block/rsxx/rsxx_priv.h b/drivers/block/rsxx/rsxx_priv.h index 382e8bf5c03b..5ad5055a4104 100644 --- a/drivers/block/rsxx/rsxx_priv.h +++ b/drivers/block/rsxx/rsxx_priv.h | |||
@@ -39,6 +39,7 @@ | |||
39 | #include <linux/vmalloc.h> | 39 | #include <linux/vmalloc.h> |
40 | #include <linux/timer.h> | 40 | #include <linux/timer.h> |
41 | #include <linux/ioctl.h> | 41 | #include <linux/ioctl.h> |
42 | #include <linux/delay.h> | ||
42 | 43 | ||
43 | #include "rsxx.h" | 44 | #include "rsxx.h" |
44 | #include "rsxx_cfg.h" | 45 | #include "rsxx_cfg.h" |
@@ -114,6 +115,7 @@ struct rsxx_dma_ctrl { | |||
114 | struct timer_list activity_timer; | 115 | struct timer_list activity_timer; |
115 | struct dma_tracker_list *trackers; | 116 | struct dma_tracker_list *trackers; |
116 | struct rsxx_dma_stats stats; | 117 | struct rsxx_dma_stats stats; |
118 | struct mutex work_lock; | ||
117 | }; | 119 | }; |
118 | 120 | ||
119 | struct rsxx_cardinfo { | 121 | struct rsxx_cardinfo { |
@@ -134,6 +136,7 @@ struct rsxx_cardinfo { | |||
134 | spinlock_t lock; | 136 | spinlock_t lock; |
135 | bool active; | 137 | bool active; |
136 | struct creg_cmd *active_cmd; | 138 | struct creg_cmd *active_cmd; |
139 | struct workqueue_struct *creg_wq; | ||
137 | struct work_struct done_work; | 140 | struct work_struct done_work; |
138 | struct list_head queue; | 141 | struct list_head queue; |
139 | unsigned int q_depth; | 142 | unsigned int q_depth; |
@@ -154,6 +157,7 @@ struct rsxx_cardinfo { | |||
154 | int buf_len; | 157 | int buf_len; |
155 | } log; | 158 | } log; |
156 | 159 | ||
160 | struct workqueue_struct *event_wq; | ||
157 | struct work_struct event_work; | 161 | struct work_struct event_work; |
158 | unsigned int state; | 162 | unsigned int state; |
159 | u64 size8; | 163 | u64 size8; |
@@ -181,6 +185,8 @@ struct rsxx_cardinfo { | |||
181 | 185 | ||
182 | int n_targets; | 186 | int n_targets; |
183 | struct rsxx_dma_ctrl *ctrl; | 187 | struct rsxx_dma_ctrl *ctrl; |
188 | |||
189 | struct dentry *debugfs_dir; | ||
184 | }; | 190 | }; |
185 | 191 | ||
186 | enum rsxx_pci_regmap { | 192 | enum rsxx_pci_regmap { |
@@ -283,6 +289,7 @@ enum rsxx_creg_addr { | |||
283 | CREG_ADD_CAPABILITIES = 0x80001050, | 289 | CREG_ADD_CAPABILITIES = 0x80001050, |
284 | CREG_ADD_LOG = 0x80002000, | 290 | CREG_ADD_LOG = 0x80002000, |
285 | CREG_ADD_NUM_TARGETS = 0x80003000, | 291 | CREG_ADD_NUM_TARGETS = 0x80003000, |
292 | CREG_ADD_CRAM = 0xA0000000, | ||
286 | CREG_ADD_CONFIG = 0xB0000000, | 293 | CREG_ADD_CONFIG = 0xB0000000, |
287 | }; | 294 | }; |
288 | 295 | ||
@@ -372,6 +379,8 @@ typedef void (*rsxx_dma_cb)(struct rsxx_cardinfo *card, | |||
372 | int rsxx_dma_setup(struct rsxx_cardinfo *card); | 379 | int rsxx_dma_setup(struct rsxx_cardinfo *card); |
373 | void rsxx_dma_destroy(struct rsxx_cardinfo *card); | 380 | void rsxx_dma_destroy(struct rsxx_cardinfo *card); |
374 | int rsxx_dma_init(void); | 381 | int rsxx_dma_init(void); |
382 | int rsxx_cleanup_dma_queue(struct rsxx_dma_ctrl *ctrl, struct list_head *q); | ||
383 | int rsxx_dma_cancel(struct rsxx_dma_ctrl *ctrl); | ||
375 | void rsxx_dma_cleanup(void); | 384 | void rsxx_dma_cleanup(void); |
376 | void rsxx_dma_queue_reset(struct rsxx_cardinfo *card); | 385 | void rsxx_dma_queue_reset(struct rsxx_cardinfo *card); |
377 | int rsxx_dma_configure(struct rsxx_cardinfo *card); | 386 | int rsxx_dma_configure(struct rsxx_cardinfo *card); |
@@ -382,7 +391,6 @@ int rsxx_dma_queue_bio(struct rsxx_cardinfo *card, | |||
382 | void *cb_data); | 391 | void *cb_data); |
383 | int rsxx_hw_buffers_init(struct pci_dev *dev, struct rsxx_dma_ctrl *ctrl); | 392 | int rsxx_hw_buffers_init(struct pci_dev *dev, struct rsxx_dma_ctrl *ctrl); |
384 | int rsxx_eeh_save_issued_dmas(struct rsxx_cardinfo *card); | 393 | int rsxx_eeh_save_issued_dmas(struct rsxx_cardinfo *card); |
385 | void rsxx_eeh_cancel_dmas(struct rsxx_cardinfo *card); | ||
386 | int rsxx_eeh_remap_dmas(struct rsxx_cardinfo *card); | 394 | int rsxx_eeh_remap_dmas(struct rsxx_cardinfo *card); |
387 | 395 | ||
388 | /***** cregs.c *****/ | 396 | /***** cregs.c *****/ |
diff --git a/drivers/block/xen-blkback/blkback.c b/drivers/block/xen-blkback/blkback.c index dd5b2fed97e9..bf4b9d282c04 100644 --- a/drivers/block/xen-blkback/blkback.c +++ b/drivers/block/xen-blkback/blkback.c | |||
@@ -50,110 +50,118 @@ | |||
50 | #include "common.h" | 50 | #include "common.h" |
51 | 51 | ||
52 | /* | 52 | /* |
53 | * These are rather arbitrary. They are fairly large because adjacent requests | 53 | * Maximum number of unused free pages to keep in the internal buffer. |
54 | * pulled from a communication ring are quite likely to end up being part of | 54 | * Setting this to a value too low will reduce memory used in each backend, |
55 | * the same scatter/gather request at the disc. | 55 | * but can have a performance penalty. |
56 | * | 56 | * |
57 | * ** TRY INCREASING 'xen_blkif_reqs' IF WRITE SPEEDS SEEM TOO LOW ** | 57 | * A sane value is xen_blkif_reqs * BLKIF_MAX_SEGMENTS_PER_REQUEST, but can |
58 | * | 58 | * be set to a lower value that might degrade performance on some intensive |
59 | * This will increase the chances of being able to write whole tracks. | 59 | * IO workloads. |
60 | * 64 should be enough to keep us competitive with Linux. | ||
61 | */ | 60 | */ |
62 | static int xen_blkif_reqs = 64; | ||
63 | module_param_named(reqs, xen_blkif_reqs, int, 0); | ||
64 | MODULE_PARM_DESC(reqs, "Number of blkback requests to allocate"); | ||
65 | 61 | ||
66 | /* Run-time switchable: /sys/module/blkback/parameters/ */ | 62 | static int xen_blkif_max_buffer_pages = 1024; |
67 | static unsigned int log_stats; | 63 | module_param_named(max_buffer_pages, xen_blkif_max_buffer_pages, int, 0644); |
68 | module_param(log_stats, int, 0644); | 64 | MODULE_PARM_DESC(max_buffer_pages, |
65 | "Maximum number of free pages to keep in each block backend buffer"); | ||
69 | 66 | ||
70 | /* | 67 | /* |
71 | * Each outstanding request that we've passed to the lower device layers has a | 68 | * Maximum number of grants to map persistently in blkback. For maximum |
72 | * 'pending_req' allocated to it. Each buffer_head that completes decrements | 69 | * performance this should be the total numbers of grants that can be used |
73 | * the pendcnt towards zero. When it hits zero, the specified domain has a | 70 | * to fill the ring, but since this might become too high, specially with |
74 | * response queued for it, with the saved 'id' passed back. | 71 | * the use of indirect descriptors, we set it to a value that provides good |
72 | * performance without using too much memory. | ||
73 | * | ||
74 | * When the list of persistent grants is full we clean it up using a LRU | ||
75 | * algorithm. | ||
75 | */ | 76 | */ |
76 | struct pending_req { | ||
77 | struct xen_blkif *blkif; | ||
78 | u64 id; | ||
79 | int nr_pages; | ||
80 | atomic_t pendcnt; | ||
81 | unsigned short operation; | ||
82 | int status; | ||
83 | struct list_head free_list; | ||
84 | DECLARE_BITMAP(unmap_seg, BLKIF_MAX_SEGMENTS_PER_REQUEST); | ||
85 | }; | ||
86 | 77 | ||
87 | #define BLKBACK_INVALID_HANDLE (~0) | 78 | static int xen_blkif_max_pgrants = 1056; |
79 | module_param_named(max_persistent_grants, xen_blkif_max_pgrants, int, 0644); | ||
80 | MODULE_PARM_DESC(max_persistent_grants, | ||
81 | "Maximum number of grants to map persistently"); | ||
88 | 82 | ||
89 | struct xen_blkbk { | 83 | /* |
90 | struct pending_req *pending_reqs; | 84 | * The LRU mechanism to clean the lists of persistent grants needs to |
91 | /* List of all 'pending_req' available */ | 85 | * be executed periodically. The time interval between consecutive executions |
92 | struct list_head pending_free; | 86 | * of the purge mechanism is set in ms. |
93 | /* And its spinlock. */ | 87 | */ |
94 | spinlock_t pending_free_lock; | 88 | #define LRU_INTERVAL 100 |
95 | wait_queue_head_t pending_free_wq; | ||
96 | /* The list of all pages that are available. */ | ||
97 | struct page **pending_pages; | ||
98 | /* And the grant handles that are available. */ | ||
99 | grant_handle_t *pending_grant_handles; | ||
100 | }; | ||
101 | |||
102 | static struct xen_blkbk *blkbk; | ||
103 | 89 | ||
104 | /* | 90 | /* |
105 | * Maximum number of grant pages that can be mapped in blkback. | 91 | * When the persistent grants list is full we will remove unused grants |
106 | * BLKIF_MAX_SEGMENTS_PER_REQUEST * RING_SIZE is the maximum number of | 92 | * from the list. The percent number of grants to be removed at each LRU |
107 | * pages that blkback will persistently map. | 93 | * execution. |
108 | * Currently, this is: | ||
109 | * RING_SIZE = 32 (for all known ring types) | ||
110 | * BLKIF_MAX_SEGMENTS_PER_REQUEST = 11 | ||
111 | * sizeof(struct persistent_gnt) = 48 | ||
112 | * So the maximum memory used to store the grants is: | ||
113 | * 32 * 11 * 48 = 16896 bytes | ||
114 | */ | 94 | */ |
115 | static inline unsigned int max_mapped_grant_pages(enum blkif_protocol protocol) | 95 | #define LRU_PERCENT_CLEAN 5 |
96 | |||
97 | /* Run-time switchable: /sys/module/blkback/parameters/ */ | ||
98 | static unsigned int log_stats; | ||
99 | module_param(log_stats, int, 0644); | ||
100 | |||
101 | #define BLKBACK_INVALID_HANDLE (~0) | ||
102 | |||
103 | /* Number of free pages to remove on each call to free_xenballooned_pages */ | ||
104 | #define NUM_BATCH_FREE_PAGES 10 | ||
105 | |||
106 | static inline int get_free_page(struct xen_blkif *blkif, struct page **page) | ||
116 | { | 107 | { |
117 | switch (protocol) { | 108 | unsigned long flags; |
118 | case BLKIF_PROTOCOL_NATIVE: | 109 | |
119 | return __CONST_RING_SIZE(blkif, PAGE_SIZE) * | 110 | spin_lock_irqsave(&blkif->free_pages_lock, flags); |
120 | BLKIF_MAX_SEGMENTS_PER_REQUEST; | 111 | if (list_empty(&blkif->free_pages)) { |
121 | case BLKIF_PROTOCOL_X86_32: | 112 | BUG_ON(blkif->free_pages_num != 0); |
122 | return __CONST_RING_SIZE(blkif_x86_32, PAGE_SIZE) * | 113 | spin_unlock_irqrestore(&blkif->free_pages_lock, flags); |
123 | BLKIF_MAX_SEGMENTS_PER_REQUEST; | 114 | return alloc_xenballooned_pages(1, page, false); |
124 | case BLKIF_PROTOCOL_X86_64: | ||
125 | return __CONST_RING_SIZE(blkif_x86_64, PAGE_SIZE) * | ||
126 | BLKIF_MAX_SEGMENTS_PER_REQUEST; | ||
127 | default: | ||
128 | BUG(); | ||
129 | } | 115 | } |
116 | BUG_ON(blkif->free_pages_num == 0); | ||
117 | page[0] = list_first_entry(&blkif->free_pages, struct page, lru); | ||
118 | list_del(&page[0]->lru); | ||
119 | blkif->free_pages_num--; | ||
120 | spin_unlock_irqrestore(&blkif->free_pages_lock, flags); | ||
121 | |||
130 | return 0; | 122 | return 0; |
131 | } | 123 | } |
132 | 124 | ||
133 | 125 | static inline void put_free_pages(struct xen_blkif *blkif, struct page **page, | |
134 | /* | 126 | int num) |
135 | * Little helpful macro to figure out the index and virtual address of the | ||
136 | * pending_pages[..]. For each 'pending_req' we have have up to | ||
137 | * BLKIF_MAX_SEGMENTS_PER_REQUEST (11) pages. The seg would be from 0 through | ||
138 | * 10 and would index in the pending_pages[..]. | ||
139 | */ | ||
140 | static inline int vaddr_pagenr(struct pending_req *req, int seg) | ||
141 | { | 127 | { |
142 | return (req - blkbk->pending_reqs) * | 128 | unsigned long flags; |
143 | BLKIF_MAX_SEGMENTS_PER_REQUEST + seg; | 129 | int i; |
144 | } | ||
145 | 130 | ||
146 | #define pending_page(req, seg) pending_pages[vaddr_pagenr(req, seg)] | 131 | spin_lock_irqsave(&blkif->free_pages_lock, flags); |
132 | for (i = 0; i < num; i++) | ||
133 | list_add(&page[i]->lru, &blkif->free_pages); | ||
134 | blkif->free_pages_num += num; | ||
135 | spin_unlock_irqrestore(&blkif->free_pages_lock, flags); | ||
136 | } | ||
147 | 137 | ||
148 | static inline unsigned long vaddr(struct pending_req *req, int seg) | 138 | static inline void shrink_free_pagepool(struct xen_blkif *blkif, int num) |
149 | { | 139 | { |
150 | unsigned long pfn = page_to_pfn(blkbk->pending_page(req, seg)); | 140 | /* Remove requested pages in batches of NUM_BATCH_FREE_PAGES */ |
151 | return (unsigned long)pfn_to_kaddr(pfn); | 141 | struct page *page[NUM_BATCH_FREE_PAGES]; |
152 | } | 142 | unsigned int num_pages = 0; |
143 | unsigned long flags; | ||
153 | 144 | ||
154 | #define pending_handle(_req, _seg) \ | 145 | spin_lock_irqsave(&blkif->free_pages_lock, flags); |
155 | (blkbk->pending_grant_handles[vaddr_pagenr(_req, _seg)]) | 146 | while (blkif->free_pages_num > num) { |
147 | BUG_ON(list_empty(&blkif->free_pages)); | ||
148 | page[num_pages] = list_first_entry(&blkif->free_pages, | ||
149 | struct page, lru); | ||
150 | list_del(&page[num_pages]->lru); | ||
151 | blkif->free_pages_num--; | ||
152 | if (++num_pages == NUM_BATCH_FREE_PAGES) { | ||
153 | spin_unlock_irqrestore(&blkif->free_pages_lock, flags); | ||
154 | free_xenballooned_pages(num_pages, page); | ||
155 | spin_lock_irqsave(&blkif->free_pages_lock, flags); | ||
156 | num_pages = 0; | ||
157 | } | ||
158 | } | ||
159 | spin_unlock_irqrestore(&blkif->free_pages_lock, flags); | ||
160 | if (num_pages != 0) | ||
161 | free_xenballooned_pages(num_pages, page); | ||
162 | } | ||
156 | 163 | ||
164 | #define vaddr(page) ((unsigned long)pfn_to_kaddr(page_to_pfn(page))) | ||
157 | 165 | ||
158 | static int do_block_io_op(struct xen_blkif *blkif); | 166 | static int do_block_io_op(struct xen_blkif *blkif); |
159 | static int dispatch_rw_block_io(struct xen_blkif *blkif, | 167 | static int dispatch_rw_block_io(struct xen_blkif *blkif, |
@@ -170,13 +178,29 @@ static void make_response(struct xen_blkif *blkif, u64 id, | |||
170 | (n) = (&(pos)->node != NULL) ? rb_next(&(pos)->node) : NULL) | 178 | (n) = (&(pos)->node != NULL) ? rb_next(&(pos)->node) : NULL) |
171 | 179 | ||
172 | 180 | ||
173 | static void add_persistent_gnt(struct rb_root *root, | 181 | /* |
182 | * We don't need locking around the persistent grant helpers | ||
183 | * because blkback uses a single-thread for each backed, so we | ||
184 | * can be sure that this functions will never be called recursively. | ||
185 | * | ||
186 | * The only exception to that is put_persistent_grant, that can be called | ||
187 | * from interrupt context (by xen_blkbk_unmap), so we have to use atomic | ||
188 | * bit operations to modify the flags of a persistent grant and to count | ||
189 | * the number of used grants. | ||
190 | */ | ||
191 | static int add_persistent_gnt(struct xen_blkif *blkif, | ||
174 | struct persistent_gnt *persistent_gnt) | 192 | struct persistent_gnt *persistent_gnt) |
175 | { | 193 | { |
176 | struct rb_node **new = &(root->rb_node), *parent = NULL; | 194 | struct rb_node **new = NULL, *parent = NULL; |
177 | struct persistent_gnt *this; | 195 | struct persistent_gnt *this; |
178 | 196 | ||
197 | if (blkif->persistent_gnt_c >= xen_blkif_max_pgrants) { | ||
198 | if (!blkif->vbd.overflow_max_grants) | ||
199 | blkif->vbd.overflow_max_grants = 1; | ||
200 | return -EBUSY; | ||
201 | } | ||
179 | /* Figure out where to put new node */ | 202 | /* Figure out where to put new node */ |
203 | new = &blkif->persistent_gnts.rb_node; | ||
180 | while (*new) { | 204 | while (*new) { |
181 | this = container_of(*new, struct persistent_gnt, node); | 205 | this = container_of(*new, struct persistent_gnt, node); |
182 | 206 | ||
@@ -186,22 +210,28 @@ static void add_persistent_gnt(struct rb_root *root, | |||
186 | else if (persistent_gnt->gnt > this->gnt) | 210 | else if (persistent_gnt->gnt > this->gnt) |
187 | new = &((*new)->rb_right); | 211 | new = &((*new)->rb_right); |
188 | else { | 212 | else { |
189 | pr_alert(DRV_PFX " trying to add a gref that's already in the tree\n"); | 213 | pr_alert_ratelimited(DRV_PFX " trying to add a gref that's already in the tree\n"); |
190 | BUG(); | 214 | return -EINVAL; |
191 | } | 215 | } |
192 | } | 216 | } |
193 | 217 | ||
218 | bitmap_zero(persistent_gnt->flags, PERSISTENT_GNT_FLAGS_SIZE); | ||
219 | set_bit(PERSISTENT_GNT_ACTIVE, persistent_gnt->flags); | ||
194 | /* Add new node and rebalance tree. */ | 220 | /* Add new node and rebalance tree. */ |
195 | rb_link_node(&(persistent_gnt->node), parent, new); | 221 | rb_link_node(&(persistent_gnt->node), parent, new); |
196 | rb_insert_color(&(persistent_gnt->node), root); | 222 | rb_insert_color(&(persistent_gnt->node), &blkif->persistent_gnts); |
223 | blkif->persistent_gnt_c++; | ||
224 | atomic_inc(&blkif->persistent_gnt_in_use); | ||
225 | return 0; | ||
197 | } | 226 | } |
198 | 227 | ||
199 | static struct persistent_gnt *get_persistent_gnt(struct rb_root *root, | 228 | static struct persistent_gnt *get_persistent_gnt(struct xen_blkif *blkif, |
200 | grant_ref_t gref) | 229 | grant_ref_t gref) |
201 | { | 230 | { |
202 | struct persistent_gnt *data; | 231 | struct persistent_gnt *data; |
203 | struct rb_node *node = root->rb_node; | 232 | struct rb_node *node = NULL; |
204 | 233 | ||
234 | node = blkif->persistent_gnts.rb_node; | ||
205 | while (node) { | 235 | while (node) { |
206 | data = container_of(node, struct persistent_gnt, node); | 236 | data = container_of(node, struct persistent_gnt, node); |
207 | 237 | ||
@@ -209,13 +239,31 @@ static struct persistent_gnt *get_persistent_gnt(struct rb_root *root, | |||
209 | node = node->rb_left; | 239 | node = node->rb_left; |
210 | else if (gref > data->gnt) | 240 | else if (gref > data->gnt) |
211 | node = node->rb_right; | 241 | node = node->rb_right; |
212 | else | 242 | else { |
243 | if(test_bit(PERSISTENT_GNT_ACTIVE, data->flags)) { | ||
244 | pr_alert_ratelimited(DRV_PFX " requesting a grant already in use\n"); | ||
245 | return NULL; | ||
246 | } | ||
247 | set_bit(PERSISTENT_GNT_ACTIVE, data->flags); | ||
248 | atomic_inc(&blkif->persistent_gnt_in_use); | ||
213 | return data; | 249 | return data; |
250 | } | ||
214 | } | 251 | } |
215 | return NULL; | 252 | return NULL; |
216 | } | 253 | } |
217 | 254 | ||
218 | static void free_persistent_gnts(struct rb_root *root, unsigned int num) | 255 | static void put_persistent_gnt(struct xen_blkif *blkif, |
256 | struct persistent_gnt *persistent_gnt) | ||
257 | { | ||
258 | if(!test_bit(PERSISTENT_GNT_ACTIVE, persistent_gnt->flags)) | ||
259 | pr_alert_ratelimited(DRV_PFX " freeing a grant already unused"); | ||
260 | set_bit(PERSISTENT_GNT_WAS_ACTIVE, persistent_gnt->flags); | ||
261 | clear_bit(PERSISTENT_GNT_ACTIVE, persistent_gnt->flags); | ||
262 | atomic_dec(&blkif->persistent_gnt_in_use); | ||
263 | } | ||
264 | |||
265 | static void free_persistent_gnts(struct xen_blkif *blkif, struct rb_root *root, | ||
266 | unsigned int num) | ||
219 | { | 267 | { |
220 | struct gnttab_unmap_grant_ref unmap[BLKIF_MAX_SEGMENTS_PER_REQUEST]; | 268 | struct gnttab_unmap_grant_ref unmap[BLKIF_MAX_SEGMENTS_PER_REQUEST]; |
221 | struct page *pages[BLKIF_MAX_SEGMENTS_PER_REQUEST]; | 269 | struct page *pages[BLKIF_MAX_SEGMENTS_PER_REQUEST]; |
@@ -240,7 +288,7 @@ static void free_persistent_gnts(struct rb_root *root, unsigned int num) | |||
240 | ret = gnttab_unmap_refs(unmap, NULL, pages, | 288 | ret = gnttab_unmap_refs(unmap, NULL, pages, |
241 | segs_to_unmap); | 289 | segs_to_unmap); |
242 | BUG_ON(ret); | 290 | BUG_ON(ret); |
243 | free_xenballooned_pages(segs_to_unmap, pages); | 291 | put_free_pages(blkif, pages, segs_to_unmap); |
244 | segs_to_unmap = 0; | 292 | segs_to_unmap = 0; |
245 | } | 293 | } |
246 | 294 | ||
@@ -251,21 +299,148 @@ static void free_persistent_gnts(struct rb_root *root, unsigned int num) | |||
251 | BUG_ON(num != 0); | 299 | BUG_ON(num != 0); |
252 | } | 300 | } |
253 | 301 | ||
302 | static void unmap_purged_grants(struct work_struct *work) | ||
303 | { | ||
304 | struct gnttab_unmap_grant_ref unmap[BLKIF_MAX_SEGMENTS_PER_REQUEST]; | ||
305 | struct page *pages[BLKIF_MAX_SEGMENTS_PER_REQUEST]; | ||
306 | struct persistent_gnt *persistent_gnt; | ||
307 | int ret, segs_to_unmap = 0; | ||
308 | struct xen_blkif *blkif = container_of(work, typeof(*blkif), persistent_purge_work); | ||
309 | |||
310 | while(!list_empty(&blkif->persistent_purge_list)) { | ||
311 | persistent_gnt = list_first_entry(&blkif->persistent_purge_list, | ||
312 | struct persistent_gnt, | ||
313 | remove_node); | ||
314 | list_del(&persistent_gnt->remove_node); | ||
315 | |||
316 | gnttab_set_unmap_op(&unmap[segs_to_unmap], | ||
317 | vaddr(persistent_gnt->page), | ||
318 | GNTMAP_host_map, | ||
319 | persistent_gnt->handle); | ||
320 | |||
321 | pages[segs_to_unmap] = persistent_gnt->page; | ||
322 | |||
323 | if (++segs_to_unmap == BLKIF_MAX_SEGMENTS_PER_REQUEST) { | ||
324 | ret = gnttab_unmap_refs(unmap, NULL, pages, | ||
325 | segs_to_unmap); | ||
326 | BUG_ON(ret); | ||
327 | put_free_pages(blkif, pages, segs_to_unmap); | ||
328 | segs_to_unmap = 0; | ||
329 | } | ||
330 | kfree(persistent_gnt); | ||
331 | } | ||
332 | if (segs_to_unmap > 0) { | ||
333 | ret = gnttab_unmap_refs(unmap, NULL, pages, segs_to_unmap); | ||
334 | BUG_ON(ret); | ||
335 | put_free_pages(blkif, pages, segs_to_unmap); | ||
336 | } | ||
337 | } | ||
338 | |||
339 | static void purge_persistent_gnt(struct xen_blkif *blkif) | ||
340 | { | ||
341 | struct persistent_gnt *persistent_gnt; | ||
342 | struct rb_node *n; | ||
343 | unsigned int num_clean, total; | ||
344 | bool scan_used = false, clean_used = false; | ||
345 | struct rb_root *root; | ||
346 | |||
347 | if (blkif->persistent_gnt_c < xen_blkif_max_pgrants || | ||
348 | (blkif->persistent_gnt_c == xen_blkif_max_pgrants && | ||
349 | !blkif->vbd.overflow_max_grants)) { | ||
350 | return; | ||
351 | } | ||
352 | |||
353 | if (work_pending(&blkif->persistent_purge_work)) { | ||
354 | pr_alert_ratelimited(DRV_PFX "Scheduled work from previous purge is still pending, cannot purge list\n"); | ||
355 | return; | ||
356 | } | ||
357 | |||
358 | num_clean = (xen_blkif_max_pgrants / 100) * LRU_PERCENT_CLEAN; | ||
359 | num_clean = blkif->persistent_gnt_c - xen_blkif_max_pgrants + num_clean; | ||
360 | num_clean = min(blkif->persistent_gnt_c, num_clean); | ||
361 | if ((num_clean == 0) || | ||
362 | (num_clean > (blkif->persistent_gnt_c - atomic_read(&blkif->persistent_gnt_in_use)))) | ||
363 | return; | ||
364 | |||
365 | /* | ||
366 | * At this point, we can assure that there will be no calls | ||
367 | * to get_persistent_grant (because we are executing this code from | ||
368 | * xen_blkif_schedule), there can only be calls to put_persistent_gnt, | ||
369 | * which means that the number of currently used grants will go down, | ||
370 | * but never up, so we will always be able to remove the requested | ||
371 | * number of grants. | ||
372 | */ | ||
373 | |||
374 | total = num_clean; | ||
375 | |||
376 | pr_debug(DRV_PFX "Going to purge %u persistent grants\n", num_clean); | ||
377 | |||
378 | INIT_LIST_HEAD(&blkif->persistent_purge_list); | ||
379 | root = &blkif->persistent_gnts; | ||
380 | purge_list: | ||
381 | foreach_grant_safe(persistent_gnt, n, root, node) { | ||
382 | BUG_ON(persistent_gnt->handle == | ||
383 | BLKBACK_INVALID_HANDLE); | ||
384 | |||
385 | if (clean_used) { | ||
386 | clear_bit(PERSISTENT_GNT_WAS_ACTIVE, persistent_gnt->flags); | ||
387 | continue; | ||
388 | } | ||
389 | |||
390 | if (test_bit(PERSISTENT_GNT_ACTIVE, persistent_gnt->flags)) | ||
391 | continue; | ||
392 | if (!scan_used && | ||
393 | (test_bit(PERSISTENT_GNT_WAS_ACTIVE, persistent_gnt->flags))) | ||
394 | continue; | ||
395 | |||
396 | rb_erase(&persistent_gnt->node, root); | ||
397 | list_add(&persistent_gnt->remove_node, | ||
398 | &blkif->persistent_purge_list); | ||
399 | if (--num_clean == 0) | ||
400 | goto finished; | ||
401 | } | ||
402 | /* | ||
403 | * If we get here it means we also need to start cleaning | ||
404 | * grants that were used since last purge in order to cope | ||
405 | * with the requested num | ||
406 | */ | ||
407 | if (!scan_used && !clean_used) { | ||
408 | pr_debug(DRV_PFX "Still missing %u purged frames\n", num_clean); | ||
409 | scan_used = true; | ||
410 | goto purge_list; | ||
411 | } | ||
412 | finished: | ||
413 | if (!clean_used) { | ||
414 | pr_debug(DRV_PFX "Finished scanning for grants to clean, removing used flag\n"); | ||
415 | clean_used = true; | ||
416 | goto purge_list; | ||
417 | } | ||
418 | |||
419 | blkif->persistent_gnt_c -= (total - num_clean); | ||
420 | blkif->vbd.overflow_max_grants = 0; | ||
421 | |||
422 | /* We can defer this work */ | ||
423 | INIT_WORK(&blkif->persistent_purge_work, unmap_purged_grants); | ||
424 | schedule_work(&blkif->persistent_purge_work); | ||
425 | pr_debug(DRV_PFX "Purged %u/%u\n", (total - num_clean), total); | ||
426 | return; | ||
427 | } | ||
428 | |||
254 | /* | 429 | /* |
255 | * Retrieve from the 'pending_reqs' a free pending_req structure to be used. | 430 | * Retrieve from the 'pending_reqs' a free pending_req structure to be used. |
256 | */ | 431 | */ |
257 | static struct pending_req *alloc_req(void) | 432 | static struct pending_req *alloc_req(struct xen_blkif *blkif) |
258 | { | 433 | { |
259 | struct pending_req *req = NULL; | 434 | struct pending_req *req = NULL; |
260 | unsigned long flags; | 435 | unsigned long flags; |
261 | 436 | ||
262 | spin_lock_irqsave(&blkbk->pending_free_lock, flags); | 437 | spin_lock_irqsave(&blkif->pending_free_lock, flags); |
263 | if (!list_empty(&blkbk->pending_free)) { | 438 | if (!list_empty(&blkif->pending_free)) { |
264 | req = list_entry(blkbk->pending_free.next, struct pending_req, | 439 | req = list_entry(blkif->pending_free.next, struct pending_req, |
265 | free_list); | 440 | free_list); |
266 | list_del(&req->free_list); | 441 | list_del(&req->free_list); |
267 | } | 442 | } |
268 | spin_unlock_irqrestore(&blkbk->pending_free_lock, flags); | 443 | spin_unlock_irqrestore(&blkif->pending_free_lock, flags); |
269 | return req; | 444 | return req; |
270 | } | 445 | } |
271 | 446 | ||
@@ -273,17 +448,17 @@ static struct pending_req *alloc_req(void) | |||
273 | * Return the 'pending_req' structure back to the freepool. We also | 448 | * Return the 'pending_req' structure back to the freepool. We also |
274 | * wake up the thread if it was waiting for a free page. | 449 | * wake up the thread if it was waiting for a free page. |
275 | */ | 450 | */ |
276 | static void free_req(struct pending_req *req) | 451 | static void free_req(struct xen_blkif *blkif, struct pending_req *req) |
277 | { | 452 | { |
278 | unsigned long flags; | 453 | unsigned long flags; |
279 | int was_empty; | 454 | int was_empty; |
280 | 455 | ||
281 | spin_lock_irqsave(&blkbk->pending_free_lock, flags); | 456 | spin_lock_irqsave(&blkif->pending_free_lock, flags); |
282 | was_empty = list_empty(&blkbk->pending_free); | 457 | was_empty = list_empty(&blkif->pending_free); |
283 | list_add(&req->free_list, &blkbk->pending_free); | 458 | list_add(&req->free_list, &blkif->pending_free); |
284 | spin_unlock_irqrestore(&blkbk->pending_free_lock, flags); | 459 | spin_unlock_irqrestore(&blkif->pending_free_lock, flags); |
285 | if (was_empty) | 460 | if (was_empty) |
286 | wake_up(&blkbk->pending_free_wq); | 461 | wake_up(&blkif->pending_free_wq); |
287 | } | 462 | } |
288 | 463 | ||
289 | /* | 464 | /* |
@@ -382,10 +557,12 @@ irqreturn_t xen_blkif_be_int(int irq, void *dev_id) | |||
382 | static void print_stats(struct xen_blkif *blkif) | 557 | static void print_stats(struct xen_blkif *blkif) |
383 | { | 558 | { |
384 | pr_info("xen-blkback (%s): oo %3llu | rd %4llu | wr %4llu | f %4llu" | 559 | pr_info("xen-blkback (%s): oo %3llu | rd %4llu | wr %4llu | f %4llu" |
385 | " | ds %4llu\n", | 560 | " | ds %4llu | pg: %4u/%4d\n", |
386 | current->comm, blkif->st_oo_req, | 561 | current->comm, blkif->st_oo_req, |
387 | blkif->st_rd_req, blkif->st_wr_req, | 562 | blkif->st_rd_req, blkif->st_wr_req, |
388 | blkif->st_f_req, blkif->st_ds_req); | 563 | blkif->st_f_req, blkif->st_ds_req, |
564 | blkif->persistent_gnt_c, | ||
565 | xen_blkif_max_pgrants); | ||
389 | blkif->st_print = jiffies + msecs_to_jiffies(10 * 1000); | 566 | blkif->st_print = jiffies + msecs_to_jiffies(10 * 1000); |
390 | blkif->st_rd_req = 0; | 567 | blkif->st_rd_req = 0; |
391 | blkif->st_wr_req = 0; | 568 | blkif->st_wr_req = 0; |
@@ -397,6 +574,8 @@ int xen_blkif_schedule(void *arg) | |||
397 | { | 574 | { |
398 | struct xen_blkif *blkif = arg; | 575 | struct xen_blkif *blkif = arg; |
399 | struct xen_vbd *vbd = &blkif->vbd; | 576 | struct xen_vbd *vbd = &blkif->vbd; |
577 | unsigned long timeout; | ||
578 | int ret; | ||
400 | 579 | ||
401 | xen_blkif_get(blkif); | 580 | xen_blkif_get(blkif); |
402 | 581 | ||
@@ -406,27 +585,52 @@ int xen_blkif_schedule(void *arg) | |||
406 | if (unlikely(vbd->size != vbd_sz(vbd))) | 585 | if (unlikely(vbd->size != vbd_sz(vbd))) |
407 | xen_vbd_resize(blkif); | 586 | xen_vbd_resize(blkif); |
408 | 587 | ||
409 | wait_event_interruptible( | 588 | timeout = msecs_to_jiffies(LRU_INTERVAL); |
589 | |||
590 | timeout = wait_event_interruptible_timeout( | ||
410 | blkif->wq, | 591 | blkif->wq, |
411 | blkif->waiting_reqs || kthread_should_stop()); | 592 | blkif->waiting_reqs || kthread_should_stop(), |
412 | wait_event_interruptible( | 593 | timeout); |
413 | blkbk->pending_free_wq, | 594 | if (timeout == 0) |
414 | !list_empty(&blkbk->pending_free) || | 595 | goto purge_gnt_list; |
415 | kthread_should_stop()); | 596 | timeout = wait_event_interruptible_timeout( |
597 | blkif->pending_free_wq, | ||
598 | !list_empty(&blkif->pending_free) || | ||
599 | kthread_should_stop(), | ||
600 | timeout); | ||
601 | if (timeout == 0) | ||
602 | goto purge_gnt_list; | ||
416 | 603 | ||
417 | blkif->waiting_reqs = 0; | 604 | blkif->waiting_reqs = 0; |
418 | smp_mb(); /* clear flag *before* checking for work */ | 605 | smp_mb(); /* clear flag *before* checking for work */ |
419 | 606 | ||
420 | if (do_block_io_op(blkif)) | 607 | ret = do_block_io_op(blkif); |
608 | if (ret > 0) | ||
421 | blkif->waiting_reqs = 1; | 609 | blkif->waiting_reqs = 1; |
610 | if (ret == -EACCES) | ||
611 | wait_event_interruptible(blkif->shutdown_wq, | ||
612 | kthread_should_stop()); | ||
613 | |||
614 | purge_gnt_list: | ||
615 | if (blkif->vbd.feature_gnt_persistent && | ||
616 | time_after(jiffies, blkif->next_lru)) { | ||
617 | purge_persistent_gnt(blkif); | ||
618 | blkif->next_lru = jiffies + msecs_to_jiffies(LRU_INTERVAL); | ||
619 | } | ||
620 | |||
621 | /* Shrink if we have more than xen_blkif_max_buffer_pages */ | ||
622 | shrink_free_pagepool(blkif, xen_blkif_max_buffer_pages); | ||
422 | 623 | ||
423 | if (log_stats && time_after(jiffies, blkif->st_print)) | 624 | if (log_stats && time_after(jiffies, blkif->st_print)) |
424 | print_stats(blkif); | 625 | print_stats(blkif); |
425 | } | 626 | } |
426 | 627 | ||
628 | /* Since we are shutting down remove all pages from the buffer */ | ||
629 | shrink_free_pagepool(blkif, 0 /* All */); | ||
630 | |||
427 | /* Free all persistent grant pages */ | 631 | /* Free all persistent grant pages */ |
428 | if (!RB_EMPTY_ROOT(&blkif->persistent_gnts)) | 632 | if (!RB_EMPTY_ROOT(&blkif->persistent_gnts)) |
429 | free_persistent_gnts(&blkif->persistent_gnts, | 633 | free_persistent_gnts(blkif, &blkif->persistent_gnts, |
430 | blkif->persistent_gnt_c); | 634 | blkif->persistent_gnt_c); |
431 | 635 | ||
432 | BUG_ON(!RB_EMPTY_ROOT(&blkif->persistent_gnts)); | 636 | BUG_ON(!RB_EMPTY_ROOT(&blkif->persistent_gnts)); |
@@ -441,148 +645,98 @@ int xen_blkif_schedule(void *arg) | |||
441 | return 0; | 645 | return 0; |
442 | } | 646 | } |
443 | 647 | ||
444 | struct seg_buf { | ||
445 | unsigned int offset; | ||
446 | unsigned int nsec; | ||
447 | }; | ||
448 | /* | 648 | /* |
449 | * Unmap the grant references, and also remove the M2P over-rides | 649 | * Unmap the grant references, and also remove the M2P over-rides |
450 | * used in the 'pending_req'. | 650 | * used in the 'pending_req'. |
451 | */ | 651 | */ |
452 | static void xen_blkbk_unmap(struct pending_req *req) | 652 | static void xen_blkbk_unmap(struct xen_blkif *blkif, |
653 | struct grant_page *pages[], | ||
654 | int num) | ||
453 | { | 655 | { |
454 | struct gnttab_unmap_grant_ref unmap[BLKIF_MAX_SEGMENTS_PER_REQUEST]; | 656 | struct gnttab_unmap_grant_ref unmap[BLKIF_MAX_SEGMENTS_PER_REQUEST]; |
455 | struct page *pages[BLKIF_MAX_SEGMENTS_PER_REQUEST]; | 657 | struct page *unmap_pages[BLKIF_MAX_SEGMENTS_PER_REQUEST]; |
456 | unsigned int i, invcount = 0; | 658 | unsigned int i, invcount = 0; |
457 | grant_handle_t handle; | ||
458 | int ret; | 659 | int ret; |
459 | 660 | ||
460 | for (i = 0; i < req->nr_pages; i++) { | 661 | for (i = 0; i < num; i++) { |
461 | if (!test_bit(i, req->unmap_seg)) | 662 | if (pages[i]->persistent_gnt != NULL) { |
663 | put_persistent_gnt(blkif, pages[i]->persistent_gnt); | ||
462 | continue; | 664 | continue; |
463 | handle = pending_handle(req, i); | 665 | } |
464 | if (handle == BLKBACK_INVALID_HANDLE) | 666 | if (pages[i]->handle == BLKBACK_INVALID_HANDLE) |
465 | continue; | 667 | continue; |
466 | gnttab_set_unmap_op(&unmap[invcount], vaddr(req, i), | 668 | unmap_pages[invcount] = pages[i]->page; |
467 | GNTMAP_host_map, handle); | 669 | gnttab_set_unmap_op(&unmap[invcount], vaddr(pages[i]->page), |
468 | pending_handle(req, i) = BLKBACK_INVALID_HANDLE; | 670 | GNTMAP_host_map, pages[i]->handle); |
469 | pages[invcount] = virt_to_page(vaddr(req, i)); | 671 | pages[i]->handle = BLKBACK_INVALID_HANDLE; |
470 | invcount++; | 672 | if (++invcount == BLKIF_MAX_SEGMENTS_PER_REQUEST) { |
673 | ret = gnttab_unmap_refs(unmap, NULL, unmap_pages, | ||
674 | invcount); | ||
675 | BUG_ON(ret); | ||
676 | put_free_pages(blkif, unmap_pages, invcount); | ||
677 | invcount = 0; | ||
678 | } | ||
679 | } | ||
680 | if (invcount) { | ||
681 | ret = gnttab_unmap_refs(unmap, NULL, unmap_pages, invcount); | ||
682 | BUG_ON(ret); | ||
683 | put_free_pages(blkif, unmap_pages, invcount); | ||
471 | } | 684 | } |
472 | |||
473 | ret = gnttab_unmap_refs(unmap, NULL, pages, invcount); | ||
474 | BUG_ON(ret); | ||
475 | } | 685 | } |
476 | 686 | ||
477 | static int xen_blkbk_map(struct blkif_request *req, | 687 | static int xen_blkbk_map(struct xen_blkif *blkif, |
478 | struct pending_req *pending_req, | 688 | struct grant_page *pages[], |
479 | struct seg_buf seg[], | 689 | int num, bool ro) |
480 | struct page *pages[]) | ||
481 | { | 690 | { |
482 | struct gnttab_map_grant_ref map[BLKIF_MAX_SEGMENTS_PER_REQUEST]; | 691 | struct gnttab_map_grant_ref map[BLKIF_MAX_SEGMENTS_PER_REQUEST]; |
483 | struct persistent_gnt *persistent_gnts[BLKIF_MAX_SEGMENTS_PER_REQUEST]; | ||
484 | struct page *pages_to_gnt[BLKIF_MAX_SEGMENTS_PER_REQUEST]; | 692 | struct page *pages_to_gnt[BLKIF_MAX_SEGMENTS_PER_REQUEST]; |
485 | struct persistent_gnt *persistent_gnt = NULL; | 693 | struct persistent_gnt *persistent_gnt = NULL; |
486 | struct xen_blkif *blkif = pending_req->blkif; | ||
487 | phys_addr_t addr = 0; | 694 | phys_addr_t addr = 0; |
488 | int i, j; | 695 | int i, seg_idx, new_map_idx; |
489 | bool new_map; | ||
490 | int nseg = req->u.rw.nr_segments; | ||
491 | int segs_to_map = 0; | 696 | int segs_to_map = 0; |
492 | int ret = 0; | 697 | int ret = 0; |
698 | int last_map = 0, map_until = 0; | ||
493 | int use_persistent_gnts; | 699 | int use_persistent_gnts; |
494 | 700 | ||
495 | use_persistent_gnts = (blkif->vbd.feature_gnt_persistent); | 701 | use_persistent_gnts = (blkif->vbd.feature_gnt_persistent); |
496 | 702 | ||
497 | BUG_ON(blkif->persistent_gnt_c > | ||
498 | max_mapped_grant_pages(pending_req->blkif->blk_protocol)); | ||
499 | |||
500 | /* | 703 | /* |
501 | * Fill out preq.nr_sects with proper amount of sectors, and setup | 704 | * Fill out preq.nr_sects with proper amount of sectors, and setup |
502 | * assign map[..] with the PFN of the page in our domain with the | 705 | * assign map[..] with the PFN of the page in our domain with the |
503 | * corresponding grant reference for each page. | 706 | * corresponding grant reference for each page. |
504 | */ | 707 | */ |
505 | for (i = 0; i < nseg; i++) { | 708 | again: |
709 | for (i = map_until; i < num; i++) { | ||
506 | uint32_t flags; | 710 | uint32_t flags; |
507 | 711 | ||
508 | if (use_persistent_gnts) | 712 | if (use_persistent_gnts) |
509 | persistent_gnt = get_persistent_gnt( | 713 | persistent_gnt = get_persistent_gnt( |
510 | &blkif->persistent_gnts, | 714 | blkif, |
511 | req->u.rw.seg[i].gref); | 715 | pages[i]->gref); |
512 | 716 | ||
513 | if (persistent_gnt) { | 717 | if (persistent_gnt) { |
514 | /* | 718 | /* |
515 | * We are using persistent grants and | 719 | * We are using persistent grants and |
516 | * the grant is already mapped | 720 | * the grant is already mapped |
517 | */ | 721 | */ |
518 | new_map = false; | 722 | pages[i]->page = persistent_gnt->page; |
519 | } else if (use_persistent_gnts && | 723 | pages[i]->persistent_gnt = persistent_gnt; |
520 | blkif->persistent_gnt_c < | ||
521 | max_mapped_grant_pages(blkif->blk_protocol)) { | ||
522 | /* | ||
523 | * We are using persistent grants, the grant is | ||
524 | * not mapped but we have room for it | ||
525 | */ | ||
526 | new_map = true; | ||
527 | persistent_gnt = kmalloc( | ||
528 | sizeof(struct persistent_gnt), | ||
529 | GFP_KERNEL); | ||
530 | if (!persistent_gnt) | ||
531 | return -ENOMEM; | ||
532 | if (alloc_xenballooned_pages(1, &persistent_gnt->page, | ||
533 | false)) { | ||
534 | kfree(persistent_gnt); | ||
535 | return -ENOMEM; | ||
536 | } | ||
537 | persistent_gnt->gnt = req->u.rw.seg[i].gref; | ||
538 | persistent_gnt->handle = BLKBACK_INVALID_HANDLE; | ||
539 | |||
540 | pages_to_gnt[segs_to_map] = | ||
541 | persistent_gnt->page; | ||
542 | addr = (unsigned long) pfn_to_kaddr( | ||
543 | page_to_pfn(persistent_gnt->page)); | ||
544 | |||
545 | add_persistent_gnt(&blkif->persistent_gnts, | ||
546 | persistent_gnt); | ||
547 | blkif->persistent_gnt_c++; | ||
548 | pr_debug(DRV_PFX " grant %u added to the tree of persistent grants, using %u/%u\n", | ||
549 | persistent_gnt->gnt, blkif->persistent_gnt_c, | ||
550 | max_mapped_grant_pages(blkif->blk_protocol)); | ||
551 | } else { | 724 | } else { |
552 | /* | 725 | if (get_free_page(blkif, &pages[i]->page)) |
553 | * We are either using persistent grants and | 726 | goto out_of_memory; |
554 | * hit the maximum limit of grants mapped, | 727 | addr = vaddr(pages[i]->page); |
555 | * or we are not using persistent grants. | 728 | pages_to_gnt[segs_to_map] = pages[i]->page; |
556 | */ | 729 | pages[i]->persistent_gnt = NULL; |
557 | if (use_persistent_gnts && | ||
558 | !blkif->vbd.overflow_max_grants) { | ||
559 | blkif->vbd.overflow_max_grants = 1; | ||
560 | pr_alert(DRV_PFX " domain %u, device %#x is using maximum number of persistent grants\n", | ||
561 | blkif->domid, blkif->vbd.handle); | ||
562 | } | ||
563 | new_map = true; | ||
564 | pages[i] = blkbk->pending_page(pending_req, i); | ||
565 | addr = vaddr(pending_req, i); | ||
566 | pages_to_gnt[segs_to_map] = | ||
567 | blkbk->pending_page(pending_req, i); | ||
568 | } | ||
569 | |||
570 | if (persistent_gnt) { | ||
571 | pages[i] = persistent_gnt->page; | ||
572 | persistent_gnts[i] = persistent_gnt; | ||
573 | } else { | ||
574 | persistent_gnts[i] = NULL; | ||
575 | } | ||
576 | |||
577 | if (new_map) { | ||
578 | flags = GNTMAP_host_map; | 730 | flags = GNTMAP_host_map; |
579 | if (!persistent_gnt && | 731 | if (!use_persistent_gnts && ro) |
580 | (pending_req->operation != BLKIF_OP_READ)) | ||
581 | flags |= GNTMAP_readonly; | 732 | flags |= GNTMAP_readonly; |
582 | gnttab_set_map_op(&map[segs_to_map++], addr, | 733 | gnttab_set_map_op(&map[segs_to_map++], addr, |
583 | flags, req->u.rw.seg[i].gref, | 734 | flags, pages[i]->gref, |
584 | blkif->domid); | 735 | blkif->domid); |
585 | } | 736 | } |
737 | map_until = i + 1; | ||
738 | if (segs_to_map == BLKIF_MAX_SEGMENTS_PER_REQUEST) | ||
739 | break; | ||
586 | } | 740 | } |
587 | 741 | ||
588 | if (segs_to_map) { | 742 | if (segs_to_map) { |
@@ -595,49 +749,133 @@ static int xen_blkbk_map(struct blkif_request *req, | |||
595 | * so that when we access vaddr(pending_req,i) it has the contents of | 749 | * so that when we access vaddr(pending_req,i) it has the contents of |
596 | * the page from the other domain. | 750 | * the page from the other domain. |
597 | */ | 751 | */ |
598 | bitmap_zero(pending_req->unmap_seg, BLKIF_MAX_SEGMENTS_PER_REQUEST); | 752 | for (seg_idx = last_map, new_map_idx = 0; seg_idx < map_until; seg_idx++) { |
599 | for (i = 0, j = 0; i < nseg; i++) { | 753 | if (!pages[seg_idx]->persistent_gnt) { |
600 | if (!persistent_gnts[i] || | ||
601 | persistent_gnts[i]->handle == BLKBACK_INVALID_HANDLE) { | ||
602 | /* This is a newly mapped grant */ | 754 | /* This is a newly mapped grant */ |
603 | BUG_ON(j >= segs_to_map); | 755 | BUG_ON(new_map_idx >= segs_to_map); |
604 | if (unlikely(map[j].status != 0)) { | 756 | if (unlikely(map[new_map_idx].status != 0)) { |
605 | pr_debug(DRV_PFX "invalid buffer -- could not remap it\n"); | 757 | pr_debug(DRV_PFX "invalid buffer -- could not remap it\n"); |
606 | map[j].handle = BLKBACK_INVALID_HANDLE; | 758 | pages[seg_idx]->handle = BLKBACK_INVALID_HANDLE; |
607 | ret |= 1; | 759 | ret |= 1; |
608 | if (persistent_gnts[i]) { | 760 | goto next; |
609 | rb_erase(&persistent_gnts[i]->node, | ||
610 | &blkif->persistent_gnts); | ||
611 | blkif->persistent_gnt_c--; | ||
612 | kfree(persistent_gnts[i]); | ||
613 | persistent_gnts[i] = NULL; | ||
614 | } | ||
615 | } | 761 | } |
762 | pages[seg_idx]->handle = map[new_map_idx].handle; | ||
763 | } else { | ||
764 | continue; | ||
616 | } | 765 | } |
617 | if (persistent_gnts[i]) { | 766 | if (use_persistent_gnts && |
618 | if (persistent_gnts[i]->handle == | 767 | blkif->persistent_gnt_c < xen_blkif_max_pgrants) { |
619 | BLKBACK_INVALID_HANDLE) { | 768 | /* |
769 | * We are using persistent grants, the grant is | ||
770 | * not mapped but we might have room for it. | ||
771 | */ | ||
772 | persistent_gnt = kmalloc(sizeof(struct persistent_gnt), | ||
773 | GFP_KERNEL); | ||
774 | if (!persistent_gnt) { | ||
620 | /* | 775 | /* |
621 | * If this is a new persistent grant | 776 | * If we don't have enough memory to |
622 | * save the handler | 777 | * allocate the persistent_gnt struct |
778 | * map this grant non-persistenly | ||
623 | */ | 779 | */ |
624 | persistent_gnts[i]->handle = map[j++].handle; | 780 | goto next; |
625 | } | 781 | } |
626 | pending_handle(pending_req, i) = | 782 | persistent_gnt->gnt = map[new_map_idx].ref; |
627 | persistent_gnts[i]->handle; | 783 | persistent_gnt->handle = map[new_map_idx].handle; |
784 | persistent_gnt->page = pages[seg_idx]->page; | ||
785 | if (add_persistent_gnt(blkif, | ||
786 | persistent_gnt)) { | ||
787 | kfree(persistent_gnt); | ||
788 | persistent_gnt = NULL; | ||
789 | goto next; | ||
790 | } | ||
791 | pages[seg_idx]->persistent_gnt = persistent_gnt; | ||
792 | pr_debug(DRV_PFX " grant %u added to the tree of persistent grants, using %u/%u\n", | ||
793 | persistent_gnt->gnt, blkif->persistent_gnt_c, | ||
794 | xen_blkif_max_pgrants); | ||
795 | goto next; | ||
796 | } | ||
797 | if (use_persistent_gnts && !blkif->vbd.overflow_max_grants) { | ||
798 | blkif->vbd.overflow_max_grants = 1; | ||
799 | pr_debug(DRV_PFX " domain %u, device %#x is using maximum number of persistent grants\n", | ||
800 | blkif->domid, blkif->vbd.handle); | ||
801 | } | ||
802 | /* | ||
803 | * We could not map this grant persistently, so use it as | ||
804 | * a non-persistent grant. | ||
805 | */ | ||
806 | next: | ||
807 | new_map_idx++; | ||
808 | } | ||
809 | segs_to_map = 0; | ||
810 | last_map = map_until; | ||
811 | if (map_until != num) | ||
812 | goto again; | ||
628 | 813 | ||
629 | if (ret) | 814 | return ret; |
630 | continue; | 815 | |
631 | } else { | 816 | out_of_memory: |
632 | pending_handle(pending_req, i) = map[j++].handle; | 817 | pr_alert(DRV_PFX "%s: out of memory\n", __func__); |
633 | bitmap_set(pending_req->unmap_seg, i, 1); | 818 | put_free_pages(blkif, pages_to_gnt, segs_to_map); |
819 | return -ENOMEM; | ||
820 | } | ||
821 | |||
822 | static int xen_blkbk_map_seg(struct pending_req *pending_req) | ||
823 | { | ||
824 | int rc; | ||
825 | |||
826 | rc = xen_blkbk_map(pending_req->blkif, pending_req->segments, | ||
827 | pending_req->nr_pages, | ||
828 | (pending_req->operation != BLKIF_OP_READ)); | ||
829 | |||
830 | return rc; | ||
831 | } | ||
634 | 832 | ||
635 | if (ret) | 833 | static int xen_blkbk_parse_indirect(struct blkif_request *req, |
636 | continue; | 834 | struct pending_req *pending_req, |
835 | struct seg_buf seg[], | ||
836 | struct phys_req *preq) | ||
837 | { | ||
838 | struct grant_page **pages = pending_req->indirect_pages; | ||
839 | struct xen_blkif *blkif = pending_req->blkif; | ||
840 | int indirect_grefs, rc, n, nseg, i; | ||
841 | struct blkif_request_segment_aligned *segments = NULL; | ||
842 | |||
843 | nseg = pending_req->nr_pages; | ||
844 | indirect_grefs = INDIRECT_PAGES(nseg); | ||
845 | BUG_ON(indirect_grefs > BLKIF_MAX_INDIRECT_PAGES_PER_REQUEST); | ||
846 | |||
847 | for (i = 0; i < indirect_grefs; i++) | ||
848 | pages[i]->gref = req->u.indirect.indirect_grefs[i]; | ||
849 | |||
850 | rc = xen_blkbk_map(blkif, pages, indirect_grefs, true); | ||
851 | if (rc) | ||
852 | goto unmap; | ||
853 | |||
854 | for (n = 0, i = 0; n < nseg; n++) { | ||
855 | if ((n % SEGS_PER_INDIRECT_FRAME) == 0) { | ||
856 | /* Map indirect segments */ | ||
857 | if (segments) | ||
858 | kunmap_atomic(segments); | ||
859 | segments = kmap_atomic(pages[n/SEGS_PER_INDIRECT_FRAME]->page); | ||
860 | } | ||
861 | i = n % SEGS_PER_INDIRECT_FRAME; | ||
862 | pending_req->segments[n]->gref = segments[i].gref; | ||
863 | seg[n].nsec = segments[i].last_sect - | ||
864 | segments[i].first_sect + 1; | ||
865 | seg[n].offset = (segments[i].first_sect << 9); | ||
866 | if ((segments[i].last_sect >= (PAGE_SIZE >> 9)) || | ||
867 | (segments[i].last_sect < segments[i].first_sect)) { | ||
868 | rc = -EINVAL; | ||
869 | goto unmap; | ||
637 | } | 870 | } |
638 | seg[i].offset = (req->u.rw.seg[i].first_sect << 9); | 871 | preq->nr_sects += seg[n].nsec; |
639 | } | 872 | } |
640 | return ret; | 873 | |
874 | unmap: | ||
875 | if (segments) | ||
876 | kunmap_atomic(segments); | ||
877 | xen_blkbk_unmap(blkif, pages, indirect_grefs); | ||
878 | return rc; | ||
641 | } | 879 | } |
642 | 880 | ||
643 | static int dispatch_discard_io(struct xen_blkif *blkif, | 881 | static int dispatch_discard_io(struct xen_blkif *blkif, |
@@ -647,7 +885,18 @@ static int dispatch_discard_io(struct xen_blkif *blkif, | |||
647 | int status = BLKIF_RSP_OKAY; | 885 | int status = BLKIF_RSP_OKAY; |
648 | struct block_device *bdev = blkif->vbd.bdev; | 886 | struct block_device *bdev = blkif->vbd.bdev; |
649 | unsigned long secure; | 887 | unsigned long secure; |
888 | struct phys_req preq; | ||
889 | |||
890 | preq.sector_number = req->u.discard.sector_number; | ||
891 | preq.nr_sects = req->u.discard.nr_sectors; | ||
650 | 892 | ||
893 | err = xen_vbd_translate(&preq, blkif, WRITE); | ||
894 | if (err) { | ||
895 | pr_warn(DRV_PFX "access denied: DISCARD [%llu->%llu] on dev=%04x\n", | ||
896 | preq.sector_number, | ||
897 | preq.sector_number + preq.nr_sects, blkif->vbd.pdevice); | ||
898 | goto fail_response; | ||
899 | } | ||
651 | blkif->st_ds_req++; | 900 | blkif->st_ds_req++; |
652 | 901 | ||
653 | xen_blkif_get(blkif); | 902 | xen_blkif_get(blkif); |
@@ -658,7 +907,7 @@ static int dispatch_discard_io(struct xen_blkif *blkif, | |||
658 | err = blkdev_issue_discard(bdev, req->u.discard.sector_number, | 907 | err = blkdev_issue_discard(bdev, req->u.discard.sector_number, |
659 | req->u.discard.nr_sectors, | 908 | req->u.discard.nr_sectors, |
660 | GFP_KERNEL, secure); | 909 | GFP_KERNEL, secure); |
661 | 910 | fail_response: | |
662 | if (err == -EOPNOTSUPP) { | 911 | if (err == -EOPNOTSUPP) { |
663 | pr_debug(DRV_PFX "discard op failed, not supported\n"); | 912 | pr_debug(DRV_PFX "discard op failed, not supported\n"); |
664 | status = BLKIF_RSP_EOPNOTSUPP; | 913 | status = BLKIF_RSP_EOPNOTSUPP; |
@@ -674,7 +923,7 @@ static int dispatch_other_io(struct xen_blkif *blkif, | |||
674 | struct blkif_request *req, | 923 | struct blkif_request *req, |
675 | struct pending_req *pending_req) | 924 | struct pending_req *pending_req) |
676 | { | 925 | { |
677 | free_req(pending_req); | 926 | free_req(blkif, pending_req); |
678 | make_response(blkif, req->u.other.id, req->operation, | 927 | make_response(blkif, req->u.other.id, req->operation, |
679 | BLKIF_RSP_EOPNOTSUPP); | 928 | BLKIF_RSP_EOPNOTSUPP); |
680 | return -EIO; | 929 | return -EIO; |
@@ -726,7 +975,9 @@ static void __end_block_io_op(struct pending_req *pending_req, int error) | |||
726 | * the proper response on the ring. | 975 | * the proper response on the ring. |
727 | */ | 976 | */ |
728 | if (atomic_dec_and_test(&pending_req->pendcnt)) { | 977 | if (atomic_dec_and_test(&pending_req->pendcnt)) { |
729 | xen_blkbk_unmap(pending_req); | 978 | xen_blkbk_unmap(pending_req->blkif, |
979 | pending_req->segments, | ||
980 | pending_req->nr_pages); | ||
730 | make_response(pending_req->blkif, pending_req->id, | 981 | make_response(pending_req->blkif, pending_req->id, |
731 | pending_req->operation, pending_req->status); | 982 | pending_req->operation, pending_req->status); |
732 | xen_blkif_put(pending_req->blkif); | 983 | xen_blkif_put(pending_req->blkif); |
@@ -734,7 +985,7 @@ static void __end_block_io_op(struct pending_req *pending_req, int error) | |||
734 | if (atomic_read(&pending_req->blkif->drain)) | 985 | if (atomic_read(&pending_req->blkif->drain)) |
735 | complete(&pending_req->blkif->drain_complete); | 986 | complete(&pending_req->blkif->drain_complete); |
736 | } | 987 | } |
737 | free_req(pending_req); | 988 | free_req(pending_req->blkif, pending_req); |
738 | } | 989 | } |
739 | } | 990 | } |
740 | 991 | ||
@@ -767,6 +1018,12 @@ __do_block_io_op(struct xen_blkif *blkif) | |||
767 | rp = blk_rings->common.sring->req_prod; | 1018 | rp = blk_rings->common.sring->req_prod; |
768 | rmb(); /* Ensure we see queued requests up to 'rp'. */ | 1019 | rmb(); /* Ensure we see queued requests up to 'rp'. */ |
769 | 1020 | ||
1021 | if (RING_REQUEST_PROD_OVERFLOW(&blk_rings->common, rp)) { | ||
1022 | rc = blk_rings->common.rsp_prod_pvt; | ||
1023 | pr_warn(DRV_PFX "Frontend provided bogus ring requests (%d - %d = %d). Halting ring processing on dev=%04x\n", | ||
1024 | rp, rc, rp - rc, blkif->vbd.pdevice); | ||
1025 | return -EACCES; | ||
1026 | } | ||
770 | while (rc != rp) { | 1027 | while (rc != rp) { |
771 | 1028 | ||
772 | if (RING_REQUEST_CONS_OVERFLOW(&blk_rings->common, rc)) | 1029 | if (RING_REQUEST_CONS_OVERFLOW(&blk_rings->common, rc)) |
@@ -777,7 +1034,7 @@ __do_block_io_op(struct xen_blkif *blkif) | |||
777 | break; | 1034 | break; |
778 | } | 1035 | } |
779 | 1036 | ||
780 | pending_req = alloc_req(); | 1037 | pending_req = alloc_req(blkif); |
781 | if (NULL == pending_req) { | 1038 | if (NULL == pending_req) { |
782 | blkif->st_oo_req++; | 1039 | blkif->st_oo_req++; |
783 | more_to_do = 1; | 1040 | more_to_do = 1; |
@@ -807,11 +1064,12 @@ __do_block_io_op(struct xen_blkif *blkif) | |||
807 | case BLKIF_OP_WRITE: | 1064 | case BLKIF_OP_WRITE: |
808 | case BLKIF_OP_WRITE_BARRIER: | 1065 | case BLKIF_OP_WRITE_BARRIER: |
809 | case BLKIF_OP_FLUSH_DISKCACHE: | 1066 | case BLKIF_OP_FLUSH_DISKCACHE: |
1067 | case BLKIF_OP_INDIRECT: | ||
810 | if (dispatch_rw_block_io(blkif, &req, pending_req)) | 1068 | if (dispatch_rw_block_io(blkif, &req, pending_req)) |
811 | goto done; | 1069 | goto done; |
812 | break; | 1070 | break; |
813 | case BLKIF_OP_DISCARD: | 1071 | case BLKIF_OP_DISCARD: |
814 | free_req(pending_req); | 1072 | free_req(blkif, pending_req); |
815 | if (dispatch_discard_io(blkif, &req)) | 1073 | if (dispatch_discard_io(blkif, &req)) |
816 | goto done; | 1074 | goto done; |
817 | break; | 1075 | break; |
@@ -853,17 +1111,28 @@ static int dispatch_rw_block_io(struct xen_blkif *blkif, | |||
853 | struct pending_req *pending_req) | 1111 | struct pending_req *pending_req) |
854 | { | 1112 | { |
855 | struct phys_req preq; | 1113 | struct phys_req preq; |
856 | struct seg_buf seg[BLKIF_MAX_SEGMENTS_PER_REQUEST]; | 1114 | struct seg_buf *seg = pending_req->seg; |
857 | unsigned int nseg; | 1115 | unsigned int nseg; |
858 | struct bio *bio = NULL; | 1116 | struct bio *bio = NULL; |
859 | struct bio *biolist[BLKIF_MAX_SEGMENTS_PER_REQUEST]; | 1117 | struct bio **biolist = pending_req->biolist; |
860 | int i, nbio = 0; | 1118 | int i, nbio = 0; |
861 | int operation; | 1119 | int operation; |
862 | struct blk_plug plug; | 1120 | struct blk_plug plug; |
863 | bool drain = false; | 1121 | bool drain = false; |
864 | struct page *pages[BLKIF_MAX_SEGMENTS_PER_REQUEST]; | 1122 | struct grant_page **pages = pending_req->segments; |
1123 | unsigned short req_operation; | ||
1124 | |||
1125 | req_operation = req->operation == BLKIF_OP_INDIRECT ? | ||
1126 | req->u.indirect.indirect_op : req->operation; | ||
1127 | if ((req->operation == BLKIF_OP_INDIRECT) && | ||
1128 | (req_operation != BLKIF_OP_READ) && | ||
1129 | (req_operation != BLKIF_OP_WRITE)) { | ||
1130 | pr_debug(DRV_PFX "Invalid indirect operation (%u)\n", | ||
1131 | req_operation); | ||
1132 | goto fail_response; | ||
1133 | } | ||
865 | 1134 | ||
866 | switch (req->operation) { | 1135 | switch (req_operation) { |
867 | case BLKIF_OP_READ: | 1136 | case BLKIF_OP_READ: |
868 | blkif->st_rd_req++; | 1137 | blkif->st_rd_req++; |
869 | operation = READ; | 1138 | operation = READ; |
@@ -885,33 +1154,47 @@ static int dispatch_rw_block_io(struct xen_blkif *blkif, | |||
885 | } | 1154 | } |
886 | 1155 | ||
887 | /* Check that the number of segments is sane. */ | 1156 | /* Check that the number of segments is sane. */ |
888 | nseg = req->u.rw.nr_segments; | 1157 | nseg = req->operation == BLKIF_OP_INDIRECT ? |
1158 | req->u.indirect.nr_segments : req->u.rw.nr_segments; | ||
889 | 1159 | ||
890 | if (unlikely(nseg == 0 && operation != WRITE_FLUSH) || | 1160 | if (unlikely(nseg == 0 && operation != WRITE_FLUSH) || |
891 | unlikely(nseg > BLKIF_MAX_SEGMENTS_PER_REQUEST)) { | 1161 | unlikely((req->operation != BLKIF_OP_INDIRECT) && |
1162 | (nseg > BLKIF_MAX_SEGMENTS_PER_REQUEST)) || | ||
1163 | unlikely((req->operation == BLKIF_OP_INDIRECT) && | ||
1164 | (nseg > MAX_INDIRECT_SEGMENTS))) { | ||
892 | pr_debug(DRV_PFX "Bad number of segments in request (%d)\n", | 1165 | pr_debug(DRV_PFX "Bad number of segments in request (%d)\n", |
893 | nseg); | 1166 | nseg); |
894 | /* Haven't submitted any bio's yet. */ | 1167 | /* Haven't submitted any bio's yet. */ |
895 | goto fail_response; | 1168 | goto fail_response; |
896 | } | 1169 | } |
897 | 1170 | ||
898 | preq.sector_number = req->u.rw.sector_number; | ||
899 | preq.nr_sects = 0; | 1171 | preq.nr_sects = 0; |
900 | 1172 | ||
901 | pending_req->blkif = blkif; | 1173 | pending_req->blkif = blkif; |
902 | pending_req->id = req->u.rw.id; | 1174 | pending_req->id = req->u.rw.id; |
903 | pending_req->operation = req->operation; | 1175 | pending_req->operation = req_operation; |
904 | pending_req->status = BLKIF_RSP_OKAY; | 1176 | pending_req->status = BLKIF_RSP_OKAY; |
905 | pending_req->nr_pages = nseg; | 1177 | pending_req->nr_pages = nseg; |
906 | 1178 | ||
907 | for (i = 0; i < nseg; i++) { | 1179 | if (req->operation != BLKIF_OP_INDIRECT) { |
908 | seg[i].nsec = req->u.rw.seg[i].last_sect - | 1180 | preq.dev = req->u.rw.handle; |
909 | req->u.rw.seg[i].first_sect + 1; | 1181 | preq.sector_number = req->u.rw.sector_number; |
910 | if ((req->u.rw.seg[i].last_sect >= (PAGE_SIZE >> 9)) || | 1182 | for (i = 0; i < nseg; i++) { |
911 | (req->u.rw.seg[i].last_sect < req->u.rw.seg[i].first_sect)) | 1183 | pages[i]->gref = req->u.rw.seg[i].gref; |
1184 | seg[i].nsec = req->u.rw.seg[i].last_sect - | ||
1185 | req->u.rw.seg[i].first_sect + 1; | ||
1186 | seg[i].offset = (req->u.rw.seg[i].first_sect << 9); | ||
1187 | if ((req->u.rw.seg[i].last_sect >= (PAGE_SIZE >> 9)) || | ||
1188 | (req->u.rw.seg[i].last_sect < | ||
1189 | req->u.rw.seg[i].first_sect)) | ||
1190 | goto fail_response; | ||
1191 | preq.nr_sects += seg[i].nsec; | ||
1192 | } | ||
1193 | } else { | ||
1194 | preq.dev = req->u.indirect.handle; | ||
1195 | preq.sector_number = req->u.indirect.sector_number; | ||
1196 | if (xen_blkbk_parse_indirect(req, pending_req, seg, &preq)) | ||
912 | goto fail_response; | 1197 | goto fail_response; |
913 | preq.nr_sects += seg[i].nsec; | ||
914 | |||
915 | } | 1198 | } |
916 | 1199 | ||
917 | if (xen_vbd_translate(&preq, blkif, operation) != 0) { | 1200 | if (xen_vbd_translate(&preq, blkif, operation) != 0) { |
@@ -948,7 +1231,7 @@ static int dispatch_rw_block_io(struct xen_blkif *blkif, | |||
948 | * the hypercall to unmap the grants - that is all done in | 1231 | * the hypercall to unmap the grants - that is all done in |
949 | * xen_blkbk_unmap. | 1232 | * xen_blkbk_unmap. |
950 | */ | 1233 | */ |
951 | if (xen_blkbk_map(req, pending_req, seg, pages)) | 1234 | if (xen_blkbk_map_seg(pending_req)) |
952 | goto fail_flush; | 1235 | goto fail_flush; |
953 | 1236 | ||
954 | /* | 1237 | /* |
@@ -960,11 +1243,12 @@ static int dispatch_rw_block_io(struct xen_blkif *blkif, | |||
960 | for (i = 0; i < nseg; i++) { | 1243 | for (i = 0; i < nseg; i++) { |
961 | while ((bio == NULL) || | 1244 | while ((bio == NULL) || |
962 | (bio_add_page(bio, | 1245 | (bio_add_page(bio, |
963 | pages[i], | 1246 | pages[i]->page, |
964 | seg[i].nsec << 9, | 1247 | seg[i].nsec << 9, |
965 | seg[i].offset) == 0)) { | 1248 | seg[i].offset) == 0)) { |
966 | 1249 | ||
967 | bio = bio_alloc(GFP_KERNEL, nseg-i); | 1250 | int nr_iovecs = min_t(int, (nseg-i), BIO_MAX_PAGES); |
1251 | bio = bio_alloc(GFP_KERNEL, nr_iovecs); | ||
968 | if (unlikely(bio == NULL)) | 1252 | if (unlikely(bio == NULL)) |
969 | goto fail_put_bio; | 1253 | goto fail_put_bio; |
970 | 1254 | ||
@@ -1009,11 +1293,12 @@ static int dispatch_rw_block_io(struct xen_blkif *blkif, | |||
1009 | return 0; | 1293 | return 0; |
1010 | 1294 | ||
1011 | fail_flush: | 1295 | fail_flush: |
1012 | xen_blkbk_unmap(pending_req); | 1296 | xen_blkbk_unmap(blkif, pending_req->segments, |
1297 | pending_req->nr_pages); | ||
1013 | fail_response: | 1298 | fail_response: |
1014 | /* Haven't submitted any bio's yet. */ | 1299 | /* Haven't submitted any bio's yet. */ |
1015 | make_response(blkif, req->u.rw.id, req->operation, BLKIF_RSP_ERROR); | 1300 | make_response(blkif, req->u.rw.id, req_operation, BLKIF_RSP_ERROR); |
1016 | free_req(pending_req); | 1301 | free_req(blkif, pending_req); |
1017 | msleep(1); /* back off a bit */ | 1302 | msleep(1); /* back off a bit */ |
1018 | return -EIO; | 1303 | return -EIO; |
1019 | 1304 | ||
@@ -1070,73 +1355,20 @@ static void make_response(struct xen_blkif *blkif, u64 id, | |||
1070 | 1355 | ||
1071 | static int __init xen_blkif_init(void) | 1356 | static int __init xen_blkif_init(void) |
1072 | { | 1357 | { |
1073 | int i, mmap_pages; | ||
1074 | int rc = 0; | 1358 | int rc = 0; |
1075 | 1359 | ||
1076 | if (!xen_domain()) | 1360 | if (!xen_domain()) |
1077 | return -ENODEV; | 1361 | return -ENODEV; |
1078 | 1362 | ||
1079 | blkbk = kzalloc(sizeof(struct xen_blkbk), GFP_KERNEL); | ||
1080 | if (!blkbk) { | ||
1081 | pr_alert(DRV_PFX "%s: out of memory!\n", __func__); | ||
1082 | return -ENOMEM; | ||
1083 | } | ||
1084 | |||
1085 | mmap_pages = xen_blkif_reqs * BLKIF_MAX_SEGMENTS_PER_REQUEST; | ||
1086 | |||
1087 | blkbk->pending_reqs = kzalloc(sizeof(blkbk->pending_reqs[0]) * | ||
1088 | xen_blkif_reqs, GFP_KERNEL); | ||
1089 | blkbk->pending_grant_handles = kmalloc(sizeof(blkbk->pending_grant_handles[0]) * | ||
1090 | mmap_pages, GFP_KERNEL); | ||
1091 | blkbk->pending_pages = kzalloc(sizeof(blkbk->pending_pages[0]) * | ||
1092 | mmap_pages, GFP_KERNEL); | ||
1093 | |||
1094 | if (!blkbk->pending_reqs || !blkbk->pending_grant_handles || | ||
1095 | !blkbk->pending_pages) { | ||
1096 | rc = -ENOMEM; | ||
1097 | goto out_of_memory; | ||
1098 | } | ||
1099 | |||
1100 | for (i = 0; i < mmap_pages; i++) { | ||
1101 | blkbk->pending_grant_handles[i] = BLKBACK_INVALID_HANDLE; | ||
1102 | blkbk->pending_pages[i] = alloc_page(GFP_KERNEL); | ||
1103 | if (blkbk->pending_pages[i] == NULL) { | ||
1104 | rc = -ENOMEM; | ||
1105 | goto out_of_memory; | ||
1106 | } | ||
1107 | } | ||
1108 | rc = xen_blkif_interface_init(); | 1363 | rc = xen_blkif_interface_init(); |
1109 | if (rc) | 1364 | if (rc) |
1110 | goto failed_init; | 1365 | goto failed_init; |
1111 | 1366 | ||
1112 | INIT_LIST_HEAD(&blkbk->pending_free); | ||
1113 | spin_lock_init(&blkbk->pending_free_lock); | ||
1114 | init_waitqueue_head(&blkbk->pending_free_wq); | ||
1115 | |||
1116 | for (i = 0; i < xen_blkif_reqs; i++) | ||
1117 | list_add_tail(&blkbk->pending_reqs[i].free_list, | ||
1118 | &blkbk->pending_free); | ||
1119 | |||
1120 | rc = xen_blkif_xenbus_init(); | 1367 | rc = xen_blkif_xenbus_init(); |
1121 | if (rc) | 1368 | if (rc) |
1122 | goto failed_init; | 1369 | goto failed_init; |
1123 | 1370 | ||
1124 | return 0; | ||
1125 | |||
1126 | out_of_memory: | ||
1127 | pr_alert(DRV_PFX "%s: out of memory\n", __func__); | ||
1128 | failed_init: | 1371 | failed_init: |
1129 | kfree(blkbk->pending_reqs); | ||
1130 | kfree(blkbk->pending_grant_handles); | ||
1131 | if (blkbk->pending_pages) { | ||
1132 | for (i = 0; i < mmap_pages; i++) { | ||
1133 | if (blkbk->pending_pages[i]) | ||
1134 | __free_page(blkbk->pending_pages[i]); | ||
1135 | } | ||
1136 | kfree(blkbk->pending_pages); | ||
1137 | } | ||
1138 | kfree(blkbk); | ||
1139 | blkbk = NULL; | ||
1140 | return rc; | 1372 | return rc; |
1141 | } | 1373 | } |
1142 | 1374 | ||
diff --git a/drivers/block/xen-blkback/common.h b/drivers/block/xen-blkback/common.h index 60103e2517ba..8d8807563d99 100644 --- a/drivers/block/xen-blkback/common.h +++ b/drivers/block/xen-blkback/common.h | |||
@@ -50,6 +50,19 @@ | |||
50 | __func__, __LINE__, ##args) | 50 | __func__, __LINE__, ##args) |
51 | 51 | ||
52 | 52 | ||
53 | /* | ||
54 | * This is the maximum number of segments that would be allowed in indirect | ||
55 | * requests. This value will also be passed to the frontend. | ||
56 | */ | ||
57 | #define MAX_INDIRECT_SEGMENTS 256 | ||
58 | |||
59 | #define SEGS_PER_INDIRECT_FRAME \ | ||
60 | (PAGE_SIZE/sizeof(struct blkif_request_segment_aligned)) | ||
61 | #define MAX_INDIRECT_PAGES \ | ||
62 | ((MAX_INDIRECT_SEGMENTS + SEGS_PER_INDIRECT_FRAME - 1)/SEGS_PER_INDIRECT_FRAME) | ||
63 | #define INDIRECT_PAGES(_segs) \ | ||
64 | ((_segs + SEGS_PER_INDIRECT_FRAME - 1)/SEGS_PER_INDIRECT_FRAME) | ||
65 | |||
53 | /* Not a real protocol. Used to generate ring structs which contain | 66 | /* Not a real protocol. Used to generate ring structs which contain |
54 | * the elements common to all protocols only. This way we get a | 67 | * the elements common to all protocols only. This way we get a |
55 | * compiler-checkable way to use common struct elements, so we can | 68 | * compiler-checkable way to use common struct elements, so we can |
@@ -83,12 +96,31 @@ struct blkif_x86_32_request_other { | |||
83 | uint64_t id; /* private guest value, echoed in resp */ | 96 | uint64_t id; /* private guest value, echoed in resp */ |
84 | } __attribute__((__packed__)); | 97 | } __attribute__((__packed__)); |
85 | 98 | ||
99 | struct blkif_x86_32_request_indirect { | ||
100 | uint8_t indirect_op; | ||
101 | uint16_t nr_segments; | ||
102 | uint64_t id; | ||
103 | blkif_sector_t sector_number; | ||
104 | blkif_vdev_t handle; | ||
105 | uint16_t _pad1; | ||
106 | grant_ref_t indirect_grefs[BLKIF_MAX_INDIRECT_PAGES_PER_REQUEST]; | ||
107 | /* | ||
108 | * The maximum number of indirect segments (and pages) that will | ||
109 | * be used is determined by MAX_INDIRECT_SEGMENTS, this value | ||
110 | * is also exported to the guest (via xenstore | ||
111 | * feature-max-indirect-segments entry), so the frontend knows how | ||
112 | * many indirect segments the backend supports. | ||
113 | */ | ||
114 | uint64_t _pad2; /* make it 64 byte aligned */ | ||
115 | } __attribute__((__packed__)); | ||
116 | |||
86 | struct blkif_x86_32_request { | 117 | struct blkif_x86_32_request { |
87 | uint8_t operation; /* BLKIF_OP_??? */ | 118 | uint8_t operation; /* BLKIF_OP_??? */ |
88 | union { | 119 | union { |
89 | struct blkif_x86_32_request_rw rw; | 120 | struct blkif_x86_32_request_rw rw; |
90 | struct blkif_x86_32_request_discard discard; | 121 | struct blkif_x86_32_request_discard discard; |
91 | struct blkif_x86_32_request_other other; | 122 | struct blkif_x86_32_request_other other; |
123 | struct blkif_x86_32_request_indirect indirect; | ||
92 | } u; | 124 | } u; |
93 | } __attribute__((__packed__)); | 125 | } __attribute__((__packed__)); |
94 | 126 | ||
@@ -127,12 +159,32 @@ struct blkif_x86_64_request_other { | |||
127 | uint64_t id; /* private guest value, echoed in resp */ | 159 | uint64_t id; /* private guest value, echoed in resp */ |
128 | } __attribute__((__packed__)); | 160 | } __attribute__((__packed__)); |
129 | 161 | ||
162 | struct blkif_x86_64_request_indirect { | ||
163 | uint8_t indirect_op; | ||
164 | uint16_t nr_segments; | ||
165 | uint32_t _pad1; /* offsetof(blkif_..,u.indirect.id)==8 */ | ||
166 | uint64_t id; | ||
167 | blkif_sector_t sector_number; | ||
168 | blkif_vdev_t handle; | ||
169 | uint16_t _pad2; | ||
170 | grant_ref_t indirect_grefs[BLKIF_MAX_INDIRECT_PAGES_PER_REQUEST]; | ||
171 | /* | ||
172 | * The maximum number of indirect segments (and pages) that will | ||
173 | * be used is determined by MAX_INDIRECT_SEGMENTS, this value | ||
174 | * is also exported to the guest (via xenstore | ||
175 | * feature-max-indirect-segments entry), so the frontend knows how | ||
176 | * many indirect segments the backend supports. | ||
177 | */ | ||
178 | uint32_t _pad3; /* make it 64 byte aligned */ | ||
179 | } __attribute__((__packed__)); | ||
180 | |||
130 | struct blkif_x86_64_request { | 181 | struct blkif_x86_64_request { |
131 | uint8_t operation; /* BLKIF_OP_??? */ | 182 | uint8_t operation; /* BLKIF_OP_??? */ |
132 | union { | 183 | union { |
133 | struct blkif_x86_64_request_rw rw; | 184 | struct blkif_x86_64_request_rw rw; |
134 | struct blkif_x86_64_request_discard discard; | 185 | struct blkif_x86_64_request_discard discard; |
135 | struct blkif_x86_64_request_other other; | 186 | struct blkif_x86_64_request_other other; |
187 | struct blkif_x86_64_request_indirect indirect; | ||
136 | } u; | 188 | } u; |
137 | } __attribute__((__packed__)); | 189 | } __attribute__((__packed__)); |
138 | 190 | ||
@@ -182,12 +234,26 @@ struct xen_vbd { | |||
182 | 234 | ||
183 | struct backend_info; | 235 | struct backend_info; |
184 | 236 | ||
237 | /* Number of available flags */ | ||
238 | #define PERSISTENT_GNT_FLAGS_SIZE 2 | ||
239 | /* This persistent grant is currently in use */ | ||
240 | #define PERSISTENT_GNT_ACTIVE 0 | ||
241 | /* | ||
242 | * This persistent grant has been used, this flag is set when we remove the | ||
243 | * PERSISTENT_GNT_ACTIVE, to know that this grant has been used recently. | ||
244 | */ | ||
245 | #define PERSISTENT_GNT_WAS_ACTIVE 1 | ||
246 | |||
247 | /* Number of requests that we can fit in a ring */ | ||
248 | #define XEN_BLKIF_REQS 32 | ||
185 | 249 | ||
186 | struct persistent_gnt { | 250 | struct persistent_gnt { |
187 | struct page *page; | 251 | struct page *page; |
188 | grant_ref_t gnt; | 252 | grant_ref_t gnt; |
189 | grant_handle_t handle; | 253 | grant_handle_t handle; |
254 | DECLARE_BITMAP(flags, PERSISTENT_GNT_FLAGS_SIZE); | ||
190 | struct rb_node node; | 255 | struct rb_node node; |
256 | struct list_head remove_node; | ||
191 | }; | 257 | }; |
192 | 258 | ||
193 | struct xen_blkif { | 259 | struct xen_blkif { |
@@ -219,6 +285,23 @@ struct xen_blkif { | |||
219 | /* tree to store persistent grants */ | 285 | /* tree to store persistent grants */ |
220 | struct rb_root persistent_gnts; | 286 | struct rb_root persistent_gnts; |
221 | unsigned int persistent_gnt_c; | 287 | unsigned int persistent_gnt_c; |
288 | atomic_t persistent_gnt_in_use; | ||
289 | unsigned long next_lru; | ||
290 | |||
291 | /* used by the kworker that offload work from the persistent purge */ | ||
292 | struct list_head persistent_purge_list; | ||
293 | struct work_struct persistent_purge_work; | ||
294 | |||
295 | /* buffer of free pages to map grant refs */ | ||
296 | spinlock_t free_pages_lock; | ||
297 | int free_pages_num; | ||
298 | struct list_head free_pages; | ||
299 | |||
300 | /* List of all 'pending_req' available */ | ||
301 | struct list_head pending_free; | ||
302 | /* And its spinlock. */ | ||
303 | spinlock_t pending_free_lock; | ||
304 | wait_queue_head_t pending_free_wq; | ||
222 | 305 | ||
223 | /* statistics */ | 306 | /* statistics */ |
224 | unsigned long st_print; | 307 | unsigned long st_print; |
@@ -231,6 +314,41 @@ struct xen_blkif { | |||
231 | unsigned long long st_wr_sect; | 314 | unsigned long long st_wr_sect; |
232 | 315 | ||
233 | wait_queue_head_t waiting_to_free; | 316 | wait_queue_head_t waiting_to_free; |
317 | /* Thread shutdown wait queue. */ | ||
318 | wait_queue_head_t shutdown_wq; | ||
319 | }; | ||
320 | |||
321 | struct seg_buf { | ||
322 | unsigned long offset; | ||
323 | unsigned int nsec; | ||
324 | }; | ||
325 | |||
326 | struct grant_page { | ||
327 | struct page *page; | ||
328 | struct persistent_gnt *persistent_gnt; | ||
329 | grant_handle_t handle; | ||
330 | grant_ref_t gref; | ||
331 | }; | ||
332 | |||
333 | /* | ||
334 | * Each outstanding request that we've passed to the lower device layers has a | ||
335 | * 'pending_req' allocated to it. Each buffer_head that completes decrements | ||
336 | * the pendcnt towards zero. When it hits zero, the specified domain has a | ||
337 | * response queued for it, with the saved 'id' passed back. | ||
338 | */ | ||
339 | struct pending_req { | ||
340 | struct xen_blkif *blkif; | ||
341 | u64 id; | ||
342 | int nr_pages; | ||
343 | atomic_t pendcnt; | ||
344 | unsigned short operation; | ||
345 | int status; | ||
346 | struct list_head free_list; | ||
347 | struct grant_page *segments[MAX_INDIRECT_SEGMENTS]; | ||
348 | /* Indirect descriptors */ | ||
349 | struct grant_page *indirect_pages[MAX_INDIRECT_PAGES]; | ||
350 | struct seg_buf seg[MAX_INDIRECT_SEGMENTS]; | ||
351 | struct bio *biolist[MAX_INDIRECT_SEGMENTS]; | ||
234 | }; | 352 | }; |
235 | 353 | ||
236 | 354 | ||
@@ -257,6 +375,7 @@ int xen_blkif_xenbus_init(void); | |||
257 | 375 | ||
258 | irqreturn_t xen_blkif_be_int(int irq, void *dev_id); | 376 | irqreturn_t xen_blkif_be_int(int irq, void *dev_id); |
259 | int xen_blkif_schedule(void *arg); | 377 | int xen_blkif_schedule(void *arg); |
378 | int xen_blkif_purge_persistent(void *arg); | ||
260 | 379 | ||
261 | int xen_blkbk_flush_diskcache(struct xenbus_transaction xbt, | 380 | int xen_blkbk_flush_diskcache(struct xenbus_transaction xbt, |
262 | struct backend_info *be, int state); | 381 | struct backend_info *be, int state); |
@@ -268,7 +387,7 @@ struct xenbus_device *xen_blkbk_xenbus(struct backend_info *be); | |||
268 | static inline void blkif_get_x86_32_req(struct blkif_request *dst, | 387 | static inline void blkif_get_x86_32_req(struct blkif_request *dst, |
269 | struct blkif_x86_32_request *src) | 388 | struct blkif_x86_32_request *src) |
270 | { | 389 | { |
271 | int i, n = BLKIF_MAX_SEGMENTS_PER_REQUEST; | 390 | int i, n = BLKIF_MAX_SEGMENTS_PER_REQUEST, j; |
272 | dst->operation = src->operation; | 391 | dst->operation = src->operation; |
273 | switch (src->operation) { | 392 | switch (src->operation) { |
274 | case BLKIF_OP_READ: | 393 | case BLKIF_OP_READ: |
@@ -291,6 +410,18 @@ static inline void blkif_get_x86_32_req(struct blkif_request *dst, | |||
291 | dst->u.discard.sector_number = src->u.discard.sector_number; | 410 | dst->u.discard.sector_number = src->u.discard.sector_number; |
292 | dst->u.discard.nr_sectors = src->u.discard.nr_sectors; | 411 | dst->u.discard.nr_sectors = src->u.discard.nr_sectors; |
293 | break; | 412 | break; |
413 | case BLKIF_OP_INDIRECT: | ||
414 | dst->u.indirect.indirect_op = src->u.indirect.indirect_op; | ||
415 | dst->u.indirect.nr_segments = src->u.indirect.nr_segments; | ||
416 | dst->u.indirect.handle = src->u.indirect.handle; | ||
417 | dst->u.indirect.id = src->u.indirect.id; | ||
418 | dst->u.indirect.sector_number = src->u.indirect.sector_number; | ||
419 | barrier(); | ||
420 | j = min(MAX_INDIRECT_PAGES, INDIRECT_PAGES(dst->u.indirect.nr_segments)); | ||
421 | for (i = 0; i < j; i++) | ||
422 | dst->u.indirect.indirect_grefs[i] = | ||
423 | src->u.indirect.indirect_grefs[i]; | ||
424 | break; | ||
294 | default: | 425 | default: |
295 | /* | 426 | /* |
296 | * Don't know how to translate this op. Only get the | 427 | * Don't know how to translate this op. Only get the |
@@ -304,7 +435,7 @@ static inline void blkif_get_x86_32_req(struct blkif_request *dst, | |||
304 | static inline void blkif_get_x86_64_req(struct blkif_request *dst, | 435 | static inline void blkif_get_x86_64_req(struct blkif_request *dst, |
305 | struct blkif_x86_64_request *src) | 436 | struct blkif_x86_64_request *src) |
306 | { | 437 | { |
307 | int i, n = BLKIF_MAX_SEGMENTS_PER_REQUEST; | 438 | int i, n = BLKIF_MAX_SEGMENTS_PER_REQUEST, j; |
308 | dst->operation = src->operation; | 439 | dst->operation = src->operation; |
309 | switch (src->operation) { | 440 | switch (src->operation) { |
310 | case BLKIF_OP_READ: | 441 | case BLKIF_OP_READ: |
@@ -327,6 +458,18 @@ static inline void blkif_get_x86_64_req(struct blkif_request *dst, | |||
327 | dst->u.discard.sector_number = src->u.discard.sector_number; | 458 | dst->u.discard.sector_number = src->u.discard.sector_number; |
328 | dst->u.discard.nr_sectors = src->u.discard.nr_sectors; | 459 | dst->u.discard.nr_sectors = src->u.discard.nr_sectors; |
329 | break; | 460 | break; |
461 | case BLKIF_OP_INDIRECT: | ||
462 | dst->u.indirect.indirect_op = src->u.indirect.indirect_op; | ||
463 | dst->u.indirect.nr_segments = src->u.indirect.nr_segments; | ||
464 | dst->u.indirect.handle = src->u.indirect.handle; | ||
465 | dst->u.indirect.id = src->u.indirect.id; | ||
466 | dst->u.indirect.sector_number = src->u.indirect.sector_number; | ||
467 | barrier(); | ||
468 | j = min(MAX_INDIRECT_PAGES, INDIRECT_PAGES(dst->u.indirect.nr_segments)); | ||
469 | for (i = 0; i < j; i++) | ||
470 | dst->u.indirect.indirect_grefs[i] = | ||
471 | src->u.indirect.indirect_grefs[i]; | ||
472 | break; | ||
330 | default: | 473 | default: |
331 | /* | 474 | /* |
332 | * Don't know how to translate this op. Only get the | 475 | * Don't know how to translate this op. Only get the |
diff --git a/drivers/block/xen-blkback/xenbus.c b/drivers/block/xen-blkback/xenbus.c index 04608a6502d7..fe5c3cd10c34 100644 --- a/drivers/block/xen-blkback/xenbus.c +++ b/drivers/block/xen-blkback/xenbus.c | |||
@@ -98,12 +98,17 @@ static void xen_update_blkif_status(struct xen_blkif *blkif) | |||
98 | err = PTR_ERR(blkif->xenblkd); | 98 | err = PTR_ERR(blkif->xenblkd); |
99 | blkif->xenblkd = NULL; | 99 | blkif->xenblkd = NULL; |
100 | xenbus_dev_error(blkif->be->dev, err, "start xenblkd"); | 100 | xenbus_dev_error(blkif->be->dev, err, "start xenblkd"); |
101 | return; | ||
101 | } | 102 | } |
102 | } | 103 | } |
103 | 104 | ||
104 | static struct xen_blkif *xen_blkif_alloc(domid_t domid) | 105 | static struct xen_blkif *xen_blkif_alloc(domid_t domid) |
105 | { | 106 | { |
106 | struct xen_blkif *blkif; | 107 | struct xen_blkif *blkif; |
108 | struct pending_req *req, *n; | ||
109 | int i, j; | ||
110 | |||
111 | BUILD_BUG_ON(MAX_INDIRECT_PAGES > BLKIF_MAX_INDIRECT_PAGES_PER_REQUEST); | ||
107 | 112 | ||
108 | blkif = kmem_cache_zalloc(xen_blkif_cachep, GFP_KERNEL); | 113 | blkif = kmem_cache_zalloc(xen_blkif_cachep, GFP_KERNEL); |
109 | if (!blkif) | 114 | if (!blkif) |
@@ -118,8 +123,57 @@ static struct xen_blkif *xen_blkif_alloc(domid_t domid) | |||
118 | blkif->st_print = jiffies; | 123 | blkif->st_print = jiffies; |
119 | init_waitqueue_head(&blkif->waiting_to_free); | 124 | init_waitqueue_head(&blkif->waiting_to_free); |
120 | blkif->persistent_gnts.rb_node = NULL; | 125 | blkif->persistent_gnts.rb_node = NULL; |
126 | spin_lock_init(&blkif->free_pages_lock); | ||
127 | INIT_LIST_HEAD(&blkif->free_pages); | ||
128 | blkif->free_pages_num = 0; | ||
129 | atomic_set(&blkif->persistent_gnt_in_use, 0); | ||
130 | |||
131 | INIT_LIST_HEAD(&blkif->pending_free); | ||
132 | |||
133 | for (i = 0; i < XEN_BLKIF_REQS; i++) { | ||
134 | req = kzalloc(sizeof(*req), GFP_KERNEL); | ||
135 | if (!req) | ||
136 | goto fail; | ||
137 | list_add_tail(&req->free_list, | ||
138 | &blkif->pending_free); | ||
139 | for (j = 0; j < MAX_INDIRECT_SEGMENTS; j++) { | ||
140 | req->segments[j] = kzalloc(sizeof(*req->segments[0]), | ||
141 | GFP_KERNEL); | ||
142 | if (!req->segments[j]) | ||
143 | goto fail; | ||
144 | } | ||
145 | for (j = 0; j < MAX_INDIRECT_PAGES; j++) { | ||
146 | req->indirect_pages[j] = kzalloc(sizeof(*req->indirect_pages[0]), | ||
147 | GFP_KERNEL); | ||
148 | if (!req->indirect_pages[j]) | ||
149 | goto fail; | ||
150 | } | ||
151 | } | ||
152 | spin_lock_init(&blkif->pending_free_lock); | ||
153 | init_waitqueue_head(&blkif->pending_free_wq); | ||
154 | init_waitqueue_head(&blkif->shutdown_wq); | ||
121 | 155 | ||
122 | return blkif; | 156 | return blkif; |
157 | |||
158 | fail: | ||
159 | list_for_each_entry_safe(req, n, &blkif->pending_free, free_list) { | ||
160 | list_del(&req->free_list); | ||
161 | for (j = 0; j < MAX_INDIRECT_SEGMENTS; j++) { | ||
162 | if (!req->segments[j]) | ||
163 | break; | ||
164 | kfree(req->segments[j]); | ||
165 | } | ||
166 | for (j = 0; j < MAX_INDIRECT_PAGES; j++) { | ||
167 | if (!req->indirect_pages[j]) | ||
168 | break; | ||
169 | kfree(req->indirect_pages[j]); | ||
170 | } | ||
171 | kfree(req); | ||
172 | } | ||
173 | |||
174 | kmem_cache_free(xen_blkif_cachep, blkif); | ||
175 | |||
176 | return ERR_PTR(-ENOMEM); | ||
123 | } | 177 | } |
124 | 178 | ||
125 | static int xen_blkif_map(struct xen_blkif *blkif, unsigned long shared_page, | 179 | static int xen_blkif_map(struct xen_blkif *blkif, unsigned long shared_page, |
@@ -178,6 +232,7 @@ static void xen_blkif_disconnect(struct xen_blkif *blkif) | |||
178 | { | 232 | { |
179 | if (blkif->xenblkd) { | 233 | if (blkif->xenblkd) { |
180 | kthread_stop(blkif->xenblkd); | 234 | kthread_stop(blkif->xenblkd); |
235 | wake_up(&blkif->shutdown_wq); | ||
181 | blkif->xenblkd = NULL; | 236 | blkif->xenblkd = NULL; |
182 | } | 237 | } |
183 | 238 | ||
@@ -198,8 +253,28 @@ static void xen_blkif_disconnect(struct xen_blkif *blkif) | |||
198 | 253 | ||
199 | static void xen_blkif_free(struct xen_blkif *blkif) | 254 | static void xen_blkif_free(struct xen_blkif *blkif) |
200 | { | 255 | { |
256 | struct pending_req *req, *n; | ||
257 | int i = 0, j; | ||
258 | |||
201 | if (!atomic_dec_and_test(&blkif->refcnt)) | 259 | if (!atomic_dec_and_test(&blkif->refcnt)) |
202 | BUG(); | 260 | BUG(); |
261 | |||
262 | /* Check that there is no request in use */ | ||
263 | list_for_each_entry_safe(req, n, &blkif->pending_free, free_list) { | ||
264 | list_del(&req->free_list); | ||
265 | |||
266 | for (j = 0; j < MAX_INDIRECT_SEGMENTS; j++) | ||
267 | kfree(req->segments[j]); | ||
268 | |||
269 | for (j = 0; j < MAX_INDIRECT_PAGES; j++) | ||
270 | kfree(req->indirect_pages[j]); | ||
271 | |||
272 | kfree(req); | ||
273 | i++; | ||
274 | } | ||
275 | |||
276 | WARN_ON(i != XEN_BLKIF_REQS); | ||
277 | |||
203 | kmem_cache_free(xen_blkif_cachep, blkif); | 278 | kmem_cache_free(xen_blkif_cachep, blkif); |
204 | } | 279 | } |
205 | 280 | ||
@@ -678,6 +753,11 @@ again: | |||
678 | dev->nodename); | 753 | dev->nodename); |
679 | goto abort; | 754 | goto abort; |
680 | } | 755 | } |
756 | err = xenbus_printf(xbt, dev->nodename, "feature-max-indirect-segments", "%u", | ||
757 | MAX_INDIRECT_SEGMENTS); | ||
758 | if (err) | ||
759 | dev_warn(&dev->dev, "writing %s/feature-max-indirect-segments (%d)", | ||
760 | dev->nodename, err); | ||
681 | 761 | ||
682 | err = xenbus_printf(xbt, dev->nodename, "sectors", "%llu", | 762 | err = xenbus_printf(xbt, dev->nodename, "sectors", "%llu", |
683 | (unsigned long long)vbd_sz(&be->blkif->vbd)); | 763 | (unsigned long long)vbd_sz(&be->blkif->vbd)); |
@@ -704,6 +784,11 @@ again: | |||
704 | dev->nodename); | 784 | dev->nodename); |
705 | goto abort; | 785 | goto abort; |
706 | } | 786 | } |
787 | err = xenbus_printf(xbt, dev->nodename, "physical-sector-size", "%u", | ||
788 | bdev_physical_block_size(be->blkif->vbd.bdev)); | ||
789 | if (err) | ||
790 | xenbus_dev_error(dev, err, "writing %s/physical-sector-size", | ||
791 | dev->nodename); | ||
707 | 792 | ||
708 | err = xenbus_transaction_end(xbt, 0); | 793 | err = xenbus_transaction_end(xbt, 0); |
709 | if (err == -EAGAIN) | 794 | if (err == -EAGAIN) |
diff --git a/drivers/block/xen-blkfront.c b/drivers/block/xen-blkfront.c index d89ef86220f4..a4660bbee8a6 100644 --- a/drivers/block/xen-blkfront.c +++ b/drivers/block/xen-blkfront.c | |||
@@ -74,12 +74,30 @@ struct grant { | |||
74 | struct blk_shadow { | 74 | struct blk_shadow { |
75 | struct blkif_request req; | 75 | struct blkif_request req; |
76 | struct request *request; | 76 | struct request *request; |
77 | struct grant *grants_used[BLKIF_MAX_SEGMENTS_PER_REQUEST]; | 77 | struct grant **grants_used; |
78 | struct grant **indirect_grants; | ||
79 | struct scatterlist *sg; | ||
80 | }; | ||
81 | |||
82 | struct split_bio { | ||
83 | struct bio *bio; | ||
84 | atomic_t pending; | ||
85 | int err; | ||
78 | }; | 86 | }; |
79 | 87 | ||
80 | static DEFINE_MUTEX(blkfront_mutex); | 88 | static DEFINE_MUTEX(blkfront_mutex); |
81 | static const struct block_device_operations xlvbd_block_fops; | 89 | static const struct block_device_operations xlvbd_block_fops; |
82 | 90 | ||
91 | /* | ||
92 | * Maximum number of segments in indirect requests, the actual value used by | ||
93 | * the frontend driver is the minimum of this value and the value provided | ||
94 | * by the backend driver. | ||
95 | */ | ||
96 | |||
97 | static unsigned int xen_blkif_max_segments = 32; | ||
98 | module_param_named(max, xen_blkif_max_segments, int, S_IRUGO); | ||
99 | MODULE_PARM_DESC(max, "Maximum amount of segments in indirect requests (default is 32)"); | ||
100 | |||
83 | #define BLK_RING_SIZE __CONST_RING_SIZE(blkif, PAGE_SIZE) | 101 | #define BLK_RING_SIZE __CONST_RING_SIZE(blkif, PAGE_SIZE) |
84 | 102 | ||
85 | /* | 103 | /* |
@@ -98,7 +116,6 @@ struct blkfront_info | |||
98 | enum blkif_state connected; | 116 | enum blkif_state connected; |
99 | int ring_ref; | 117 | int ring_ref; |
100 | struct blkif_front_ring ring; | 118 | struct blkif_front_ring ring; |
101 | struct scatterlist sg[BLKIF_MAX_SEGMENTS_PER_REQUEST]; | ||
102 | unsigned int evtchn, irq; | 119 | unsigned int evtchn, irq; |
103 | struct request_queue *rq; | 120 | struct request_queue *rq; |
104 | struct work_struct work; | 121 | struct work_struct work; |
@@ -114,6 +131,7 @@ struct blkfront_info | |||
114 | unsigned int discard_granularity; | 131 | unsigned int discard_granularity; |
115 | unsigned int discard_alignment; | 132 | unsigned int discard_alignment; |
116 | unsigned int feature_persistent:1; | 133 | unsigned int feature_persistent:1; |
134 | unsigned int max_indirect_segments; | ||
117 | int is_ready; | 135 | int is_ready; |
118 | }; | 136 | }; |
119 | 137 | ||
@@ -142,6 +160,13 @@ static DEFINE_SPINLOCK(minor_lock); | |||
142 | 160 | ||
143 | #define DEV_NAME "xvd" /* name in /dev */ | 161 | #define DEV_NAME "xvd" /* name in /dev */ |
144 | 162 | ||
163 | #define SEGS_PER_INDIRECT_FRAME \ | ||
164 | (PAGE_SIZE/sizeof(struct blkif_request_segment_aligned)) | ||
165 | #define INDIRECT_GREFS(_segs) \ | ||
166 | ((_segs + SEGS_PER_INDIRECT_FRAME - 1)/SEGS_PER_INDIRECT_FRAME) | ||
167 | |||
168 | static int blkfront_setup_indirect(struct blkfront_info *info); | ||
169 | |||
145 | static int get_id_from_freelist(struct blkfront_info *info) | 170 | static int get_id_from_freelist(struct blkfront_info *info) |
146 | { | 171 | { |
147 | unsigned long free = info->shadow_free; | 172 | unsigned long free = info->shadow_free; |
@@ -358,7 +383,8 @@ static int blkif_queue_request(struct request *req) | |||
358 | struct blkif_request *ring_req; | 383 | struct blkif_request *ring_req; |
359 | unsigned long id; | 384 | unsigned long id; |
360 | unsigned int fsect, lsect; | 385 | unsigned int fsect, lsect; |
361 | int i, ref; | 386 | int i, ref, n; |
387 | struct blkif_request_segment_aligned *segments = NULL; | ||
362 | 388 | ||
363 | /* | 389 | /* |
364 | * Used to store if we are able to queue the request by just using | 390 | * Used to store if we are able to queue the request by just using |
@@ -369,21 +395,27 @@ static int blkif_queue_request(struct request *req) | |||
369 | grant_ref_t gref_head; | 395 | grant_ref_t gref_head; |
370 | struct grant *gnt_list_entry = NULL; | 396 | struct grant *gnt_list_entry = NULL; |
371 | struct scatterlist *sg; | 397 | struct scatterlist *sg; |
398 | int nseg, max_grefs; | ||
372 | 399 | ||
373 | if (unlikely(info->connected != BLKIF_STATE_CONNECTED)) | 400 | if (unlikely(info->connected != BLKIF_STATE_CONNECTED)) |
374 | return 1; | 401 | return 1; |
375 | 402 | ||
376 | /* Check if we have enought grants to allocate a requests */ | 403 | max_grefs = info->max_indirect_segments ? |
377 | if (info->persistent_gnts_c < BLKIF_MAX_SEGMENTS_PER_REQUEST) { | 404 | info->max_indirect_segments + |
405 | INDIRECT_GREFS(info->max_indirect_segments) : | ||
406 | BLKIF_MAX_SEGMENTS_PER_REQUEST; | ||
407 | |||
408 | /* Check if we have enough grants to allocate a requests */ | ||
409 | if (info->persistent_gnts_c < max_grefs) { | ||
378 | new_persistent_gnts = 1; | 410 | new_persistent_gnts = 1; |
379 | if (gnttab_alloc_grant_references( | 411 | if (gnttab_alloc_grant_references( |
380 | BLKIF_MAX_SEGMENTS_PER_REQUEST - info->persistent_gnts_c, | 412 | max_grefs - info->persistent_gnts_c, |
381 | &gref_head) < 0) { | 413 | &gref_head) < 0) { |
382 | gnttab_request_free_callback( | 414 | gnttab_request_free_callback( |
383 | &info->callback, | 415 | &info->callback, |
384 | blkif_restart_queue_callback, | 416 | blkif_restart_queue_callback, |
385 | info, | 417 | info, |
386 | BLKIF_MAX_SEGMENTS_PER_REQUEST); | 418 | max_grefs); |
387 | return 1; | 419 | return 1; |
388 | } | 420 | } |
389 | } else | 421 | } else |
@@ -394,42 +426,67 @@ static int blkif_queue_request(struct request *req) | |||
394 | id = get_id_from_freelist(info); | 426 | id = get_id_from_freelist(info); |
395 | info->shadow[id].request = req; | 427 | info->shadow[id].request = req; |
396 | 428 | ||
397 | ring_req->u.rw.id = id; | ||
398 | ring_req->u.rw.sector_number = (blkif_sector_t)blk_rq_pos(req); | ||
399 | ring_req->u.rw.handle = info->handle; | ||
400 | |||
401 | ring_req->operation = rq_data_dir(req) ? | ||
402 | BLKIF_OP_WRITE : BLKIF_OP_READ; | ||
403 | |||
404 | if (req->cmd_flags & (REQ_FLUSH | REQ_FUA)) { | ||
405 | /* | ||
406 | * Ideally we can do an unordered flush-to-disk. In case the | ||
407 | * backend onlysupports barriers, use that. A barrier request | ||
408 | * a superset of FUA, so we can implement it the same | ||
409 | * way. (It's also a FLUSH+FUA, since it is | ||
410 | * guaranteed ordered WRT previous writes.) | ||
411 | */ | ||
412 | ring_req->operation = info->flush_op; | ||
413 | } | ||
414 | |||
415 | if (unlikely(req->cmd_flags & (REQ_DISCARD | REQ_SECURE))) { | 429 | if (unlikely(req->cmd_flags & (REQ_DISCARD | REQ_SECURE))) { |
416 | /* id, sector_number and handle are set above. */ | ||
417 | ring_req->operation = BLKIF_OP_DISCARD; | 430 | ring_req->operation = BLKIF_OP_DISCARD; |
418 | ring_req->u.discard.nr_sectors = blk_rq_sectors(req); | 431 | ring_req->u.discard.nr_sectors = blk_rq_sectors(req); |
432 | ring_req->u.discard.id = id; | ||
433 | ring_req->u.discard.sector_number = (blkif_sector_t)blk_rq_pos(req); | ||
419 | if ((req->cmd_flags & REQ_SECURE) && info->feature_secdiscard) | 434 | if ((req->cmd_flags & REQ_SECURE) && info->feature_secdiscard) |
420 | ring_req->u.discard.flag = BLKIF_DISCARD_SECURE; | 435 | ring_req->u.discard.flag = BLKIF_DISCARD_SECURE; |
421 | else | 436 | else |
422 | ring_req->u.discard.flag = 0; | 437 | ring_req->u.discard.flag = 0; |
423 | } else { | 438 | } else { |
424 | ring_req->u.rw.nr_segments = blk_rq_map_sg(req->q, req, | 439 | BUG_ON(info->max_indirect_segments == 0 && |
425 | info->sg); | 440 | req->nr_phys_segments > BLKIF_MAX_SEGMENTS_PER_REQUEST); |
426 | BUG_ON(ring_req->u.rw.nr_segments > | 441 | BUG_ON(info->max_indirect_segments && |
427 | BLKIF_MAX_SEGMENTS_PER_REQUEST); | 442 | req->nr_phys_segments > info->max_indirect_segments); |
428 | 443 | nseg = blk_rq_map_sg(req->q, req, info->shadow[id].sg); | |
429 | for_each_sg(info->sg, sg, ring_req->u.rw.nr_segments, i) { | 444 | ring_req->u.rw.id = id; |
445 | if (nseg > BLKIF_MAX_SEGMENTS_PER_REQUEST) { | ||
446 | /* | ||
447 | * The indirect operation can only be a BLKIF_OP_READ or | ||
448 | * BLKIF_OP_WRITE | ||
449 | */ | ||
450 | BUG_ON(req->cmd_flags & (REQ_FLUSH | REQ_FUA)); | ||
451 | ring_req->operation = BLKIF_OP_INDIRECT; | ||
452 | ring_req->u.indirect.indirect_op = rq_data_dir(req) ? | ||
453 | BLKIF_OP_WRITE : BLKIF_OP_READ; | ||
454 | ring_req->u.indirect.sector_number = (blkif_sector_t)blk_rq_pos(req); | ||
455 | ring_req->u.indirect.handle = info->handle; | ||
456 | ring_req->u.indirect.nr_segments = nseg; | ||
457 | } else { | ||
458 | ring_req->u.rw.sector_number = (blkif_sector_t)blk_rq_pos(req); | ||
459 | ring_req->u.rw.handle = info->handle; | ||
460 | ring_req->operation = rq_data_dir(req) ? | ||
461 | BLKIF_OP_WRITE : BLKIF_OP_READ; | ||
462 | if (req->cmd_flags & (REQ_FLUSH | REQ_FUA)) { | ||
463 | /* | ||
464 | * Ideally we can do an unordered flush-to-disk. In case the | ||
465 | * backend onlysupports barriers, use that. A barrier request | ||
466 | * a superset of FUA, so we can implement it the same | ||
467 | * way. (It's also a FLUSH+FUA, since it is | ||
468 | * guaranteed ordered WRT previous writes.) | ||
469 | */ | ||
470 | ring_req->operation = info->flush_op; | ||
471 | } | ||
472 | ring_req->u.rw.nr_segments = nseg; | ||
473 | } | ||
474 | for_each_sg(info->shadow[id].sg, sg, nseg, i) { | ||
430 | fsect = sg->offset >> 9; | 475 | fsect = sg->offset >> 9; |
431 | lsect = fsect + (sg->length >> 9) - 1; | 476 | lsect = fsect + (sg->length >> 9) - 1; |
432 | 477 | ||
478 | if ((ring_req->operation == BLKIF_OP_INDIRECT) && | ||
479 | (i % SEGS_PER_INDIRECT_FRAME == 0)) { | ||
480 | if (segments) | ||
481 | kunmap_atomic(segments); | ||
482 | |||
483 | n = i / SEGS_PER_INDIRECT_FRAME; | ||
484 | gnt_list_entry = get_grant(&gref_head, info); | ||
485 | info->shadow[id].indirect_grants[n] = gnt_list_entry; | ||
486 | segments = kmap_atomic(pfn_to_page(gnt_list_entry->pfn)); | ||
487 | ring_req->u.indirect.indirect_grefs[n] = gnt_list_entry->gref; | ||
488 | } | ||
489 | |||
433 | gnt_list_entry = get_grant(&gref_head, info); | 490 | gnt_list_entry = get_grant(&gref_head, info); |
434 | ref = gnt_list_entry->gref; | 491 | ref = gnt_list_entry->gref; |
435 | 492 | ||
@@ -441,8 +498,7 @@ static int blkif_queue_request(struct request *req) | |||
441 | 498 | ||
442 | BUG_ON(sg->offset + sg->length > PAGE_SIZE); | 499 | BUG_ON(sg->offset + sg->length > PAGE_SIZE); |
443 | 500 | ||
444 | shared_data = kmap_atomic( | 501 | shared_data = kmap_atomic(pfn_to_page(gnt_list_entry->pfn)); |
445 | pfn_to_page(gnt_list_entry->pfn)); | ||
446 | bvec_data = kmap_atomic(sg_page(sg)); | 502 | bvec_data = kmap_atomic(sg_page(sg)); |
447 | 503 | ||
448 | /* | 504 | /* |
@@ -461,13 +517,23 @@ static int blkif_queue_request(struct request *req) | |||
461 | kunmap_atomic(bvec_data); | 517 | kunmap_atomic(bvec_data); |
462 | kunmap_atomic(shared_data); | 518 | kunmap_atomic(shared_data); |
463 | } | 519 | } |
464 | 520 | if (ring_req->operation != BLKIF_OP_INDIRECT) { | |
465 | ring_req->u.rw.seg[i] = | 521 | ring_req->u.rw.seg[i] = |
466 | (struct blkif_request_segment) { | 522 | (struct blkif_request_segment) { |
467 | .gref = ref, | 523 | .gref = ref, |
468 | .first_sect = fsect, | 524 | .first_sect = fsect, |
469 | .last_sect = lsect }; | 525 | .last_sect = lsect }; |
526 | } else { | ||
527 | n = i % SEGS_PER_INDIRECT_FRAME; | ||
528 | segments[n] = | ||
529 | (struct blkif_request_segment_aligned) { | ||
530 | .gref = ref, | ||
531 | .first_sect = fsect, | ||
532 | .last_sect = lsect }; | ||
533 | } | ||
470 | } | 534 | } |
535 | if (segments) | ||
536 | kunmap_atomic(segments); | ||
471 | } | 537 | } |
472 | 538 | ||
473 | info->ring.req_prod_pvt++; | 539 | info->ring.req_prod_pvt++; |
@@ -542,7 +608,9 @@ wait: | |||
542 | flush_requests(info); | 608 | flush_requests(info); |
543 | } | 609 | } |
544 | 610 | ||
545 | static int xlvbd_init_blk_queue(struct gendisk *gd, u16 sector_size) | 611 | static int xlvbd_init_blk_queue(struct gendisk *gd, u16 sector_size, |
612 | unsigned int physical_sector_size, | ||
613 | unsigned int segments) | ||
546 | { | 614 | { |
547 | struct request_queue *rq; | 615 | struct request_queue *rq; |
548 | struct blkfront_info *info = gd->private_data; | 616 | struct blkfront_info *info = gd->private_data; |
@@ -564,14 +632,15 @@ static int xlvbd_init_blk_queue(struct gendisk *gd, u16 sector_size) | |||
564 | 632 | ||
565 | /* Hard sector size and max sectors impersonate the equiv. hardware. */ | 633 | /* Hard sector size and max sectors impersonate the equiv. hardware. */ |
566 | blk_queue_logical_block_size(rq, sector_size); | 634 | blk_queue_logical_block_size(rq, sector_size); |
567 | blk_queue_max_hw_sectors(rq, 512); | 635 | blk_queue_physical_block_size(rq, physical_sector_size); |
636 | blk_queue_max_hw_sectors(rq, (segments * PAGE_SIZE) / 512); | ||
568 | 637 | ||
569 | /* Each segment in a request is up to an aligned page in size. */ | 638 | /* Each segment in a request is up to an aligned page in size. */ |
570 | blk_queue_segment_boundary(rq, PAGE_SIZE - 1); | 639 | blk_queue_segment_boundary(rq, PAGE_SIZE - 1); |
571 | blk_queue_max_segment_size(rq, PAGE_SIZE); | 640 | blk_queue_max_segment_size(rq, PAGE_SIZE); |
572 | 641 | ||
573 | /* Ensure a merged request will fit in a single I/O ring slot. */ | 642 | /* Ensure a merged request will fit in a single I/O ring slot. */ |
574 | blk_queue_max_segments(rq, BLKIF_MAX_SEGMENTS_PER_REQUEST); | 643 | blk_queue_max_segments(rq, segments); |
575 | 644 | ||
576 | /* Make sure buffer addresses are sector-aligned. */ | 645 | /* Make sure buffer addresses are sector-aligned. */ |
577 | blk_queue_dma_alignment(rq, 511); | 646 | blk_queue_dma_alignment(rq, 511); |
@@ -588,13 +657,16 @@ static int xlvbd_init_blk_queue(struct gendisk *gd, u16 sector_size) | |||
588 | static void xlvbd_flush(struct blkfront_info *info) | 657 | static void xlvbd_flush(struct blkfront_info *info) |
589 | { | 658 | { |
590 | blk_queue_flush(info->rq, info->feature_flush); | 659 | blk_queue_flush(info->rq, info->feature_flush); |
591 | printk(KERN_INFO "blkfront: %s: %s: %s %s\n", | 660 | printk(KERN_INFO "blkfront: %s: %s: %s %s %s %s %s\n", |
592 | info->gd->disk_name, | 661 | info->gd->disk_name, |
593 | info->flush_op == BLKIF_OP_WRITE_BARRIER ? | 662 | info->flush_op == BLKIF_OP_WRITE_BARRIER ? |
594 | "barrier" : (info->flush_op == BLKIF_OP_FLUSH_DISKCACHE ? | 663 | "barrier" : (info->flush_op == BLKIF_OP_FLUSH_DISKCACHE ? |
595 | "flush diskcache" : "barrier or flush"), | 664 | "flush diskcache" : "barrier or flush"), |
596 | info->feature_flush ? "enabled" : "disabled", | 665 | info->feature_flush ? "enabled;" : "disabled;", |
597 | info->feature_persistent ? "using persistent grants" : ""); | 666 | "persistent grants:", |
667 | info->feature_persistent ? "enabled;" : "disabled;", | ||
668 | "indirect descriptors:", | ||
669 | info->max_indirect_segments ? "enabled;" : "disabled;"); | ||
598 | } | 670 | } |
599 | 671 | ||
600 | static int xen_translate_vdev(int vdevice, int *minor, unsigned int *offset) | 672 | static int xen_translate_vdev(int vdevice, int *minor, unsigned int *offset) |
@@ -667,7 +739,8 @@ static char *encode_disk_name(char *ptr, unsigned int n) | |||
667 | 739 | ||
668 | static int xlvbd_alloc_gendisk(blkif_sector_t capacity, | 740 | static int xlvbd_alloc_gendisk(blkif_sector_t capacity, |
669 | struct blkfront_info *info, | 741 | struct blkfront_info *info, |
670 | u16 vdisk_info, u16 sector_size) | 742 | u16 vdisk_info, u16 sector_size, |
743 | unsigned int physical_sector_size) | ||
671 | { | 744 | { |
672 | struct gendisk *gd; | 745 | struct gendisk *gd; |
673 | int nr_minors = 1; | 746 | int nr_minors = 1; |
@@ -734,7 +807,9 @@ static int xlvbd_alloc_gendisk(blkif_sector_t capacity, | |||
734 | gd->driverfs_dev = &(info->xbdev->dev); | 807 | gd->driverfs_dev = &(info->xbdev->dev); |
735 | set_capacity(gd, capacity); | 808 | set_capacity(gd, capacity); |
736 | 809 | ||
737 | if (xlvbd_init_blk_queue(gd, sector_size)) { | 810 | if (xlvbd_init_blk_queue(gd, sector_size, physical_sector_size, |
811 | info->max_indirect_segments ? : | ||
812 | BLKIF_MAX_SEGMENTS_PER_REQUEST)) { | ||
738 | del_gendisk(gd); | 813 | del_gendisk(gd); |
739 | goto release; | 814 | goto release; |
740 | } | 815 | } |
@@ -818,6 +893,7 @@ static void blkif_free(struct blkfront_info *info, int suspend) | |||
818 | { | 893 | { |
819 | struct grant *persistent_gnt; | 894 | struct grant *persistent_gnt; |
820 | struct grant *n; | 895 | struct grant *n; |
896 | int i, j, segs; | ||
821 | 897 | ||
822 | /* Prevent new requests being issued until we fix things up. */ | 898 | /* Prevent new requests being issued until we fix things up. */ |
823 | spin_lock_irq(&info->io_lock); | 899 | spin_lock_irq(&info->io_lock); |
@@ -843,6 +919,47 @@ static void blkif_free(struct blkfront_info *info, int suspend) | |||
843 | } | 919 | } |
844 | BUG_ON(info->persistent_gnts_c != 0); | 920 | BUG_ON(info->persistent_gnts_c != 0); |
845 | 921 | ||
922 | for (i = 0; i < BLK_RING_SIZE; i++) { | ||
923 | /* | ||
924 | * Clear persistent grants present in requests already | ||
925 | * on the shared ring | ||
926 | */ | ||
927 | if (!info->shadow[i].request) | ||
928 | goto free_shadow; | ||
929 | |||
930 | segs = info->shadow[i].req.operation == BLKIF_OP_INDIRECT ? | ||
931 | info->shadow[i].req.u.indirect.nr_segments : | ||
932 | info->shadow[i].req.u.rw.nr_segments; | ||
933 | for (j = 0; j < segs; j++) { | ||
934 | persistent_gnt = info->shadow[i].grants_used[j]; | ||
935 | gnttab_end_foreign_access(persistent_gnt->gref, 0, 0UL); | ||
936 | __free_page(pfn_to_page(persistent_gnt->pfn)); | ||
937 | kfree(persistent_gnt); | ||
938 | } | ||
939 | |||
940 | if (info->shadow[i].req.operation != BLKIF_OP_INDIRECT) | ||
941 | /* | ||
942 | * If this is not an indirect operation don't try to | ||
943 | * free indirect segments | ||
944 | */ | ||
945 | goto free_shadow; | ||
946 | |||
947 | for (j = 0; j < INDIRECT_GREFS(segs); j++) { | ||
948 | persistent_gnt = info->shadow[i].indirect_grants[j]; | ||
949 | gnttab_end_foreign_access(persistent_gnt->gref, 0, 0UL); | ||
950 | __free_page(pfn_to_page(persistent_gnt->pfn)); | ||
951 | kfree(persistent_gnt); | ||
952 | } | ||
953 | |||
954 | free_shadow: | ||
955 | kfree(info->shadow[i].grants_used); | ||
956 | info->shadow[i].grants_used = NULL; | ||
957 | kfree(info->shadow[i].indirect_grants); | ||
958 | info->shadow[i].indirect_grants = NULL; | ||
959 | kfree(info->shadow[i].sg); | ||
960 | info->shadow[i].sg = NULL; | ||
961 | } | ||
962 | |||
846 | /* No more gnttab callback work. */ | 963 | /* No more gnttab callback work. */ |
847 | gnttab_cancel_free_callback(&info->callback); | 964 | gnttab_cancel_free_callback(&info->callback); |
848 | spin_unlock_irq(&info->io_lock); | 965 | spin_unlock_irq(&info->io_lock); |
@@ -867,12 +984,13 @@ static void blkif_completion(struct blk_shadow *s, struct blkfront_info *info, | |||
867 | struct blkif_response *bret) | 984 | struct blkif_response *bret) |
868 | { | 985 | { |
869 | int i = 0; | 986 | int i = 0; |
870 | struct bio_vec *bvec; | 987 | struct scatterlist *sg; |
871 | struct req_iterator iter; | ||
872 | unsigned long flags; | ||
873 | char *bvec_data; | 988 | char *bvec_data; |
874 | void *shared_data; | 989 | void *shared_data; |
875 | unsigned int offset = 0; | 990 | int nseg; |
991 | |||
992 | nseg = s->req.operation == BLKIF_OP_INDIRECT ? | ||
993 | s->req.u.indirect.nr_segments : s->req.u.rw.nr_segments; | ||
876 | 994 | ||
877 | if (bret->operation == BLKIF_OP_READ) { | 995 | if (bret->operation == BLKIF_OP_READ) { |
878 | /* | 996 | /* |
@@ -881,26 +999,29 @@ static void blkif_completion(struct blk_shadow *s, struct blkfront_info *info, | |||
881 | * than PAGE_SIZE, we have to keep track of the current offset, | 999 | * than PAGE_SIZE, we have to keep track of the current offset, |
882 | * to be sure we are copying the data from the right shared page. | 1000 | * to be sure we are copying the data from the right shared page. |
883 | */ | 1001 | */ |
884 | rq_for_each_segment(bvec, s->request, iter) { | 1002 | for_each_sg(s->sg, sg, nseg, i) { |
885 | BUG_ON((bvec->bv_offset + bvec->bv_len) > PAGE_SIZE); | 1003 | BUG_ON(sg->offset + sg->length > PAGE_SIZE); |
886 | if (bvec->bv_offset < offset) | ||
887 | i++; | ||
888 | BUG_ON(i >= s->req.u.rw.nr_segments); | ||
889 | shared_data = kmap_atomic( | 1004 | shared_data = kmap_atomic( |
890 | pfn_to_page(s->grants_used[i]->pfn)); | 1005 | pfn_to_page(s->grants_used[i]->pfn)); |
891 | bvec_data = bvec_kmap_irq(bvec, &flags); | 1006 | bvec_data = kmap_atomic(sg_page(sg)); |
892 | memcpy(bvec_data, shared_data + bvec->bv_offset, | 1007 | memcpy(bvec_data + sg->offset, |
893 | bvec->bv_len); | 1008 | shared_data + sg->offset, |
894 | bvec_kunmap_irq(bvec_data, &flags); | 1009 | sg->length); |
1010 | kunmap_atomic(bvec_data); | ||
895 | kunmap_atomic(shared_data); | 1011 | kunmap_atomic(shared_data); |
896 | offset = bvec->bv_offset + bvec->bv_len; | ||
897 | } | 1012 | } |
898 | } | 1013 | } |
899 | /* Add the persistent grant into the list of free grants */ | 1014 | /* Add the persistent grant into the list of free grants */ |
900 | for (i = 0; i < s->req.u.rw.nr_segments; i++) { | 1015 | for (i = 0; i < nseg; i++) { |
901 | list_add(&s->grants_used[i]->node, &info->persistent_gnts); | 1016 | list_add(&s->grants_used[i]->node, &info->persistent_gnts); |
902 | info->persistent_gnts_c++; | 1017 | info->persistent_gnts_c++; |
903 | } | 1018 | } |
1019 | if (s->req.operation == BLKIF_OP_INDIRECT) { | ||
1020 | for (i = 0; i < INDIRECT_GREFS(nseg); i++) { | ||
1021 | list_add(&s->indirect_grants[i]->node, &info->persistent_gnts); | ||
1022 | info->persistent_gnts_c++; | ||
1023 | } | ||
1024 | } | ||
904 | } | 1025 | } |
905 | 1026 | ||
906 | static irqreturn_t blkif_interrupt(int irq, void *dev_id) | 1027 | static irqreturn_t blkif_interrupt(int irq, void *dev_id) |
@@ -1034,14 +1155,6 @@ static int setup_blkring(struct xenbus_device *dev, | |||
1034 | SHARED_RING_INIT(sring); | 1155 | SHARED_RING_INIT(sring); |
1035 | FRONT_RING_INIT(&info->ring, sring, PAGE_SIZE); | 1156 | FRONT_RING_INIT(&info->ring, sring, PAGE_SIZE); |
1036 | 1157 | ||
1037 | sg_init_table(info->sg, BLKIF_MAX_SEGMENTS_PER_REQUEST); | ||
1038 | |||
1039 | /* Allocate memory for grants */ | ||
1040 | err = fill_grant_buffer(info, BLK_RING_SIZE * | ||
1041 | BLKIF_MAX_SEGMENTS_PER_REQUEST); | ||
1042 | if (err) | ||
1043 | goto fail; | ||
1044 | |||
1045 | err = xenbus_grant_ring(dev, virt_to_mfn(info->ring.sring)); | 1158 | err = xenbus_grant_ring(dev, virt_to_mfn(info->ring.sring)); |
1046 | if (err < 0) { | 1159 | if (err < 0) { |
1047 | free_page((unsigned long)sring); | 1160 | free_page((unsigned long)sring); |
@@ -1223,13 +1336,84 @@ static int blkfront_probe(struct xenbus_device *dev, | |||
1223 | return 0; | 1336 | return 0; |
1224 | } | 1337 | } |
1225 | 1338 | ||
1339 | /* | ||
1340 | * This is a clone of md_trim_bio, used to split a bio into smaller ones | ||
1341 | */ | ||
1342 | static void trim_bio(struct bio *bio, int offset, int size) | ||
1343 | { | ||
1344 | /* 'bio' is a cloned bio which we need to trim to match | ||
1345 | * the given offset and size. | ||
1346 | * This requires adjusting bi_sector, bi_size, and bi_io_vec | ||
1347 | */ | ||
1348 | int i; | ||
1349 | struct bio_vec *bvec; | ||
1350 | int sofar = 0; | ||
1351 | |||
1352 | size <<= 9; | ||
1353 | if (offset == 0 && size == bio->bi_size) | ||
1354 | return; | ||
1355 | |||
1356 | bio->bi_sector += offset; | ||
1357 | bio->bi_size = size; | ||
1358 | offset <<= 9; | ||
1359 | clear_bit(BIO_SEG_VALID, &bio->bi_flags); | ||
1360 | |||
1361 | while (bio->bi_idx < bio->bi_vcnt && | ||
1362 | bio->bi_io_vec[bio->bi_idx].bv_len <= offset) { | ||
1363 | /* remove this whole bio_vec */ | ||
1364 | offset -= bio->bi_io_vec[bio->bi_idx].bv_len; | ||
1365 | bio->bi_idx++; | ||
1366 | } | ||
1367 | if (bio->bi_idx < bio->bi_vcnt) { | ||
1368 | bio->bi_io_vec[bio->bi_idx].bv_offset += offset; | ||
1369 | bio->bi_io_vec[bio->bi_idx].bv_len -= offset; | ||
1370 | } | ||
1371 | /* avoid any complications with bi_idx being non-zero*/ | ||
1372 | if (bio->bi_idx) { | ||
1373 | memmove(bio->bi_io_vec, bio->bi_io_vec+bio->bi_idx, | ||
1374 | (bio->bi_vcnt - bio->bi_idx) * sizeof(struct bio_vec)); | ||
1375 | bio->bi_vcnt -= bio->bi_idx; | ||
1376 | bio->bi_idx = 0; | ||
1377 | } | ||
1378 | /* Make sure vcnt and last bv are not too big */ | ||
1379 | bio_for_each_segment(bvec, bio, i) { | ||
1380 | if (sofar + bvec->bv_len > size) | ||
1381 | bvec->bv_len = size - sofar; | ||
1382 | if (bvec->bv_len == 0) { | ||
1383 | bio->bi_vcnt = i; | ||
1384 | break; | ||
1385 | } | ||
1386 | sofar += bvec->bv_len; | ||
1387 | } | ||
1388 | } | ||
1389 | |||
1390 | static void split_bio_end(struct bio *bio, int error) | ||
1391 | { | ||
1392 | struct split_bio *split_bio = bio->bi_private; | ||
1393 | |||
1394 | if (error) | ||
1395 | split_bio->err = error; | ||
1396 | |||
1397 | if (atomic_dec_and_test(&split_bio->pending)) { | ||
1398 | split_bio->bio->bi_phys_segments = 0; | ||
1399 | bio_endio(split_bio->bio, split_bio->err); | ||
1400 | kfree(split_bio); | ||
1401 | } | ||
1402 | bio_put(bio); | ||
1403 | } | ||
1226 | 1404 | ||
1227 | static int blkif_recover(struct blkfront_info *info) | 1405 | static int blkif_recover(struct blkfront_info *info) |
1228 | { | 1406 | { |
1229 | int i; | 1407 | int i; |
1230 | struct blkif_request *req; | 1408 | struct request *req, *n; |
1231 | struct blk_shadow *copy; | 1409 | struct blk_shadow *copy; |
1232 | int j; | 1410 | int rc; |
1411 | struct bio *bio, *cloned_bio; | ||
1412 | struct bio_list bio_list, merge_bio; | ||
1413 | unsigned int segs, offset; | ||
1414 | int pending, size; | ||
1415 | struct split_bio *split_bio; | ||
1416 | struct list_head requests; | ||
1233 | 1417 | ||
1234 | /* Stage 1: Make a safe copy of the shadow state. */ | 1418 | /* Stage 1: Make a safe copy of the shadow state. */ |
1235 | copy = kmemdup(info->shadow, sizeof(info->shadow), | 1419 | copy = kmemdup(info->shadow, sizeof(info->shadow), |
@@ -1244,36 +1428,64 @@ static int blkif_recover(struct blkfront_info *info) | |||
1244 | info->shadow_free = info->ring.req_prod_pvt; | 1428 | info->shadow_free = info->ring.req_prod_pvt; |
1245 | info->shadow[BLK_RING_SIZE-1].req.u.rw.id = 0x0fffffff; | 1429 | info->shadow[BLK_RING_SIZE-1].req.u.rw.id = 0x0fffffff; |
1246 | 1430 | ||
1247 | /* Stage 3: Find pending requests and requeue them. */ | 1431 | rc = blkfront_setup_indirect(info); |
1432 | if (rc) { | ||
1433 | kfree(copy); | ||
1434 | return rc; | ||
1435 | } | ||
1436 | |||
1437 | segs = info->max_indirect_segments ? : BLKIF_MAX_SEGMENTS_PER_REQUEST; | ||
1438 | blk_queue_max_segments(info->rq, segs); | ||
1439 | bio_list_init(&bio_list); | ||
1440 | INIT_LIST_HEAD(&requests); | ||
1248 | for (i = 0; i < BLK_RING_SIZE; i++) { | 1441 | for (i = 0; i < BLK_RING_SIZE; i++) { |
1249 | /* Not in use? */ | 1442 | /* Not in use? */ |
1250 | if (!copy[i].request) | 1443 | if (!copy[i].request) |
1251 | continue; | 1444 | continue; |
1252 | 1445 | ||
1253 | /* Grab a request slot and copy shadow state into it. */ | 1446 | /* |
1254 | req = RING_GET_REQUEST(&info->ring, info->ring.req_prod_pvt); | 1447 | * Get the bios in the request so we can re-queue them. |
1255 | *req = copy[i].req; | 1448 | */ |
1256 | 1449 | if (copy[i].request->cmd_flags & | |
1257 | /* We get a new request id, and must reset the shadow state. */ | 1450 | (REQ_FLUSH | REQ_FUA | REQ_DISCARD | REQ_SECURE)) { |
1258 | req->u.rw.id = get_id_from_freelist(info); | 1451 | /* |
1259 | memcpy(&info->shadow[req->u.rw.id], ©[i], sizeof(copy[i])); | 1452 | * Flush operations don't contain bios, so |
1260 | 1453 | * we need to requeue the whole request | |
1261 | if (req->operation != BLKIF_OP_DISCARD) { | 1454 | */ |
1262 | /* Rewrite any grant references invalidated by susp/resume. */ | 1455 | list_add(©[i].request->queuelist, &requests); |
1263 | for (j = 0; j < req->u.rw.nr_segments; j++) | 1456 | continue; |
1264 | gnttab_grant_foreign_access_ref( | ||
1265 | req->u.rw.seg[j].gref, | ||
1266 | info->xbdev->otherend_id, | ||
1267 | pfn_to_mfn(copy[i].grants_used[j]->pfn), | ||
1268 | 0); | ||
1269 | } | 1457 | } |
1270 | info->shadow[req->u.rw.id].req = *req; | 1458 | merge_bio.head = copy[i].request->bio; |
1271 | 1459 | merge_bio.tail = copy[i].request->biotail; | |
1272 | info->ring.req_prod_pvt++; | 1460 | bio_list_merge(&bio_list, &merge_bio); |
1461 | copy[i].request->bio = NULL; | ||
1462 | blk_put_request(copy[i].request); | ||
1273 | } | 1463 | } |
1274 | 1464 | ||
1275 | kfree(copy); | 1465 | kfree(copy); |
1276 | 1466 | ||
1467 | /* | ||
1468 | * Empty the queue, this is important because we might have | ||
1469 | * requests in the queue with more segments than what we | ||
1470 | * can handle now. | ||
1471 | */ | ||
1472 | spin_lock_irq(&info->io_lock); | ||
1473 | while ((req = blk_fetch_request(info->rq)) != NULL) { | ||
1474 | if (req->cmd_flags & | ||
1475 | (REQ_FLUSH | REQ_FUA | REQ_DISCARD | REQ_SECURE)) { | ||
1476 | list_add(&req->queuelist, &requests); | ||
1477 | continue; | ||
1478 | } | ||
1479 | merge_bio.head = req->bio; | ||
1480 | merge_bio.tail = req->biotail; | ||
1481 | bio_list_merge(&bio_list, &merge_bio); | ||
1482 | req->bio = NULL; | ||
1483 | if (req->cmd_flags & (REQ_FLUSH | REQ_FUA)) | ||
1484 | pr_alert("diskcache flush request found!\n"); | ||
1485 | __blk_put_request(info->rq, req); | ||
1486 | } | ||
1487 | spin_unlock_irq(&info->io_lock); | ||
1488 | |||
1277 | xenbus_switch_state(info->xbdev, XenbusStateConnected); | 1489 | xenbus_switch_state(info->xbdev, XenbusStateConnected); |
1278 | 1490 | ||
1279 | spin_lock_irq(&info->io_lock); | 1491 | spin_lock_irq(&info->io_lock); |
@@ -1281,14 +1493,50 @@ static int blkif_recover(struct blkfront_info *info) | |||
1281 | /* Now safe for us to use the shared ring */ | 1493 | /* Now safe for us to use the shared ring */ |
1282 | info->connected = BLKIF_STATE_CONNECTED; | 1494 | info->connected = BLKIF_STATE_CONNECTED; |
1283 | 1495 | ||
1284 | /* Send off requeued requests */ | ||
1285 | flush_requests(info); | ||
1286 | |||
1287 | /* Kick any other new requests queued since we resumed */ | 1496 | /* Kick any other new requests queued since we resumed */ |
1288 | kick_pending_request_queues(info); | 1497 | kick_pending_request_queues(info); |
1289 | 1498 | ||
1499 | list_for_each_entry_safe(req, n, &requests, queuelist) { | ||
1500 | /* Requeue pending requests (flush or discard) */ | ||
1501 | list_del_init(&req->queuelist); | ||
1502 | BUG_ON(req->nr_phys_segments > segs); | ||
1503 | blk_requeue_request(info->rq, req); | ||
1504 | } | ||
1290 | spin_unlock_irq(&info->io_lock); | 1505 | spin_unlock_irq(&info->io_lock); |
1291 | 1506 | ||
1507 | while ((bio = bio_list_pop(&bio_list)) != NULL) { | ||
1508 | /* Traverse the list of pending bios and re-queue them */ | ||
1509 | if (bio_segments(bio) > segs) { | ||
1510 | /* | ||
1511 | * This bio has more segments than what we can | ||
1512 | * handle, we have to split it. | ||
1513 | */ | ||
1514 | pending = (bio_segments(bio) + segs - 1) / segs; | ||
1515 | split_bio = kzalloc(sizeof(*split_bio), GFP_NOIO); | ||
1516 | BUG_ON(split_bio == NULL); | ||
1517 | atomic_set(&split_bio->pending, pending); | ||
1518 | split_bio->bio = bio; | ||
1519 | for (i = 0; i < pending; i++) { | ||
1520 | offset = (i * segs * PAGE_SIZE) >> 9; | ||
1521 | size = min((unsigned int)(segs * PAGE_SIZE) >> 9, | ||
1522 | (unsigned int)(bio->bi_size >> 9) - offset); | ||
1523 | cloned_bio = bio_clone(bio, GFP_NOIO); | ||
1524 | BUG_ON(cloned_bio == NULL); | ||
1525 | trim_bio(cloned_bio, offset, size); | ||
1526 | cloned_bio->bi_private = split_bio; | ||
1527 | cloned_bio->bi_end_io = split_bio_end; | ||
1528 | submit_bio(cloned_bio->bi_rw, cloned_bio); | ||
1529 | } | ||
1530 | /* | ||
1531 | * Now we have to wait for all those smaller bios to | ||
1532 | * end, so we can also end the "parent" bio. | ||
1533 | */ | ||
1534 | continue; | ||
1535 | } | ||
1536 | /* We don't need to split this bio */ | ||
1537 | submit_bio(bio->bi_rw, bio); | ||
1538 | } | ||
1539 | |||
1292 | return 0; | 1540 | return 0; |
1293 | } | 1541 | } |
1294 | 1542 | ||
@@ -1308,8 +1556,12 @@ static int blkfront_resume(struct xenbus_device *dev) | |||
1308 | blkif_free(info, info->connected == BLKIF_STATE_CONNECTED); | 1556 | blkif_free(info, info->connected == BLKIF_STATE_CONNECTED); |
1309 | 1557 | ||
1310 | err = talk_to_blkback(dev, info); | 1558 | err = talk_to_blkback(dev, info); |
1311 | if (info->connected == BLKIF_STATE_SUSPENDED && !err) | 1559 | |
1312 | err = blkif_recover(info); | 1560 | /* |
1561 | * We have to wait for the backend to switch to | ||
1562 | * connected state, since we want to read which | ||
1563 | * features it supports. | ||
1564 | */ | ||
1313 | 1565 | ||
1314 | return err; | 1566 | return err; |
1315 | } | 1567 | } |
@@ -1387,6 +1639,60 @@ static void blkfront_setup_discard(struct blkfront_info *info) | |||
1387 | kfree(type); | 1639 | kfree(type); |
1388 | } | 1640 | } |
1389 | 1641 | ||
1642 | static int blkfront_setup_indirect(struct blkfront_info *info) | ||
1643 | { | ||
1644 | unsigned int indirect_segments, segs; | ||
1645 | int err, i; | ||
1646 | |||
1647 | err = xenbus_gather(XBT_NIL, info->xbdev->otherend, | ||
1648 | "feature-max-indirect-segments", "%u", &indirect_segments, | ||
1649 | NULL); | ||
1650 | if (err) { | ||
1651 | info->max_indirect_segments = 0; | ||
1652 | segs = BLKIF_MAX_SEGMENTS_PER_REQUEST; | ||
1653 | } else { | ||
1654 | info->max_indirect_segments = min(indirect_segments, | ||
1655 | xen_blkif_max_segments); | ||
1656 | segs = info->max_indirect_segments; | ||
1657 | } | ||
1658 | |||
1659 | err = fill_grant_buffer(info, (segs + INDIRECT_GREFS(segs)) * BLK_RING_SIZE); | ||
1660 | if (err) | ||
1661 | goto out_of_memory; | ||
1662 | |||
1663 | for (i = 0; i < BLK_RING_SIZE; i++) { | ||
1664 | info->shadow[i].grants_used = kzalloc( | ||
1665 | sizeof(info->shadow[i].grants_used[0]) * segs, | ||
1666 | GFP_NOIO); | ||
1667 | info->shadow[i].sg = kzalloc(sizeof(info->shadow[i].sg[0]) * segs, GFP_NOIO); | ||
1668 | if (info->max_indirect_segments) | ||
1669 | info->shadow[i].indirect_grants = kzalloc( | ||
1670 | sizeof(info->shadow[i].indirect_grants[0]) * | ||
1671 | INDIRECT_GREFS(segs), | ||
1672 | GFP_NOIO); | ||
1673 | if ((info->shadow[i].grants_used == NULL) || | ||
1674 | (info->shadow[i].sg == NULL) || | ||
1675 | (info->max_indirect_segments && | ||
1676 | (info->shadow[i].indirect_grants == NULL))) | ||
1677 | goto out_of_memory; | ||
1678 | sg_init_table(info->shadow[i].sg, segs); | ||
1679 | } | ||
1680 | |||
1681 | |||
1682 | return 0; | ||
1683 | |||
1684 | out_of_memory: | ||
1685 | for (i = 0; i < BLK_RING_SIZE; i++) { | ||
1686 | kfree(info->shadow[i].grants_used); | ||
1687 | info->shadow[i].grants_used = NULL; | ||
1688 | kfree(info->shadow[i].sg); | ||
1689 | info->shadow[i].sg = NULL; | ||
1690 | kfree(info->shadow[i].indirect_grants); | ||
1691 | info->shadow[i].indirect_grants = NULL; | ||
1692 | } | ||
1693 | return -ENOMEM; | ||
1694 | } | ||
1695 | |||
1390 | /* | 1696 | /* |
1391 | * Invoked when the backend is finally 'ready' (and has told produced | 1697 | * Invoked when the backend is finally 'ready' (and has told produced |
1392 | * the details about the physical device - #sectors, size, etc). | 1698 | * the details about the physical device - #sectors, size, etc). |
@@ -1395,6 +1701,7 @@ static void blkfront_connect(struct blkfront_info *info) | |||
1395 | { | 1701 | { |
1396 | unsigned long long sectors; | 1702 | unsigned long long sectors; |
1397 | unsigned long sector_size; | 1703 | unsigned long sector_size; |
1704 | unsigned int physical_sector_size; | ||
1398 | unsigned int binfo; | 1705 | unsigned int binfo; |
1399 | int err; | 1706 | int err; |
1400 | int barrier, flush, discard, persistent; | 1707 | int barrier, flush, discard, persistent; |
@@ -1414,8 +1721,15 @@ static void blkfront_connect(struct blkfront_info *info) | |||
1414 | set_capacity(info->gd, sectors); | 1721 | set_capacity(info->gd, sectors); |
1415 | revalidate_disk(info->gd); | 1722 | revalidate_disk(info->gd); |
1416 | 1723 | ||
1417 | /* fall through */ | 1724 | return; |
1418 | case BLKIF_STATE_SUSPENDED: | 1725 | case BLKIF_STATE_SUSPENDED: |
1726 | /* | ||
1727 | * If we are recovering from suspension, we need to wait | ||
1728 | * for the backend to announce it's features before | ||
1729 | * reconnecting, at least we need to know if the backend | ||
1730 | * supports indirect descriptors, and how many. | ||
1731 | */ | ||
1732 | blkif_recover(info); | ||
1419 | return; | 1733 | return; |
1420 | 1734 | ||
1421 | default: | 1735 | default: |
@@ -1437,6 +1751,16 @@ static void blkfront_connect(struct blkfront_info *info) | |||
1437 | return; | 1751 | return; |
1438 | } | 1752 | } |
1439 | 1753 | ||
1754 | /* | ||
1755 | * physcial-sector-size is a newer field, so old backends may not | ||
1756 | * provide this. Assume physical sector size to be the same as | ||
1757 | * sector_size in that case. | ||
1758 | */ | ||
1759 | err = xenbus_scanf(XBT_NIL, info->xbdev->otherend, | ||
1760 | "physical-sector-size", "%u", &physical_sector_size); | ||
1761 | if (err != 1) | ||
1762 | physical_sector_size = sector_size; | ||
1763 | |||
1440 | info->feature_flush = 0; | 1764 | info->feature_flush = 0; |
1441 | info->flush_op = 0; | 1765 | info->flush_op = 0; |
1442 | 1766 | ||
@@ -1483,7 +1807,15 @@ static void blkfront_connect(struct blkfront_info *info) | |||
1483 | else | 1807 | else |
1484 | info->feature_persistent = persistent; | 1808 | info->feature_persistent = persistent; |
1485 | 1809 | ||
1486 | err = xlvbd_alloc_gendisk(sectors, info, binfo, sector_size); | 1810 | err = blkfront_setup_indirect(info); |
1811 | if (err) { | ||
1812 | xenbus_dev_fatal(info->xbdev, err, "setup_indirect at %s", | ||
1813 | info->xbdev->otherend); | ||
1814 | return; | ||
1815 | } | ||
1816 | |||
1817 | err = xlvbd_alloc_gendisk(sectors, info, binfo, sector_size, | ||
1818 | physical_sector_size); | ||
1487 | if (err) { | 1819 | if (err) { |
1488 | xenbus_dev_fatal(info->xbdev, err, "xlvbd_add at %s", | 1820 | xenbus_dev_fatal(info->xbdev, err, "xlvbd_add at %s", |
1489 | info->xbdev->otherend); | 1821 | info->xbdev->otherend); |
diff --git a/drivers/cpufreq/intel_pstate.c b/drivers/cpufreq/intel_pstate.c index b012d7600e1a..7cde885011ed 100644 --- a/drivers/cpufreq/intel_pstate.c +++ b/drivers/cpufreq/intel_pstate.c | |||
@@ -103,10 +103,10 @@ struct pstate_adjust_policy { | |||
103 | static struct pstate_adjust_policy default_policy = { | 103 | static struct pstate_adjust_policy default_policy = { |
104 | .sample_rate_ms = 10, | 104 | .sample_rate_ms = 10, |
105 | .deadband = 0, | 105 | .deadband = 0, |
106 | .setpoint = 109, | 106 | .setpoint = 97, |
107 | .p_gain_pct = 17, | 107 | .p_gain_pct = 20, |
108 | .d_gain_pct = 0, | 108 | .d_gain_pct = 0, |
109 | .i_gain_pct = 4, | 109 | .i_gain_pct = 0, |
110 | }; | 110 | }; |
111 | 111 | ||
112 | struct perf_limits { | 112 | struct perf_limits { |
@@ -468,12 +468,12 @@ static inline void intel_pstate_set_sample_time(struct cpudata *cpu) | |||
468 | static inline int intel_pstate_get_scaled_busy(struct cpudata *cpu) | 468 | static inline int intel_pstate_get_scaled_busy(struct cpudata *cpu) |
469 | { | 469 | { |
470 | int32_t busy_scaled; | 470 | int32_t busy_scaled; |
471 | int32_t core_busy, turbo_pstate, current_pstate; | 471 | int32_t core_busy, max_pstate, current_pstate; |
472 | 472 | ||
473 | core_busy = int_tofp(cpu->samples[cpu->sample_ptr].core_pct_busy); | 473 | core_busy = int_tofp(cpu->samples[cpu->sample_ptr].core_pct_busy); |
474 | turbo_pstate = int_tofp(cpu->pstate.turbo_pstate); | 474 | max_pstate = int_tofp(cpu->pstate.max_pstate); |
475 | current_pstate = int_tofp(cpu->pstate.current_pstate); | 475 | current_pstate = int_tofp(cpu->pstate.current_pstate); |
476 | busy_scaled = mul_fp(core_busy, div_fp(turbo_pstate, current_pstate)); | 476 | busy_scaled = mul_fp(core_busy, div_fp(max_pstate, current_pstate)); |
477 | 477 | ||
478 | return fp_toint(busy_scaled); | 478 | return fp_toint(busy_scaled); |
479 | } | 479 | } |
diff --git a/drivers/crypto/caam/caamhash.c b/drivers/crypto/caam/caamhash.c index 5996521a1caf..84573b4d6f92 100644 --- a/drivers/crypto/caam/caamhash.c +++ b/drivers/crypto/caam/caamhash.c | |||
@@ -429,7 +429,7 @@ static int hash_digest_key(struct caam_hash_ctx *ctx, const u8 *key_in, | |||
429 | dma_addr_t src_dma, dst_dma; | 429 | dma_addr_t src_dma, dst_dma; |
430 | int ret = 0; | 430 | int ret = 0; |
431 | 431 | ||
432 | desc = kmalloc(CAAM_CMD_SZ * 6 + CAAM_PTR_SZ * 2, GFP_KERNEL | GFP_DMA); | 432 | desc = kmalloc(CAAM_CMD_SZ * 8 + CAAM_PTR_SZ * 2, GFP_KERNEL | GFP_DMA); |
433 | if (!desc) { | 433 | if (!desc) { |
434 | dev_err(jrdev, "unable to allocate key input memory\n"); | 434 | dev_err(jrdev, "unable to allocate key input memory\n"); |
435 | return -ENOMEM; | 435 | return -ENOMEM; |
diff --git a/drivers/edac/edac_mc.c b/drivers/edac/edac_mc.c index 27e86d938262..89e109022d78 100644 --- a/drivers/edac/edac_mc.c +++ b/drivers/edac/edac_mc.c | |||
@@ -48,6 +48,8 @@ static LIST_HEAD(mc_devices); | |||
48 | */ | 48 | */ |
49 | static void const *edac_mc_owner; | 49 | static void const *edac_mc_owner; |
50 | 50 | ||
51 | static struct bus_type mc_bus[EDAC_MAX_MCS]; | ||
52 | |||
51 | unsigned edac_dimm_info_location(struct dimm_info *dimm, char *buf, | 53 | unsigned edac_dimm_info_location(struct dimm_info *dimm, char *buf, |
52 | unsigned len) | 54 | unsigned len) |
53 | { | 55 | { |
@@ -723,6 +725,11 @@ int edac_mc_add_mc(struct mem_ctl_info *mci) | |||
723 | int ret = -EINVAL; | 725 | int ret = -EINVAL; |
724 | edac_dbg(0, "\n"); | 726 | edac_dbg(0, "\n"); |
725 | 727 | ||
728 | if (mci->mc_idx >= EDAC_MAX_MCS) { | ||
729 | pr_warn_once("Too many memory controllers: %d\n", mci->mc_idx); | ||
730 | return -ENODEV; | ||
731 | } | ||
732 | |||
726 | #ifdef CONFIG_EDAC_DEBUG | 733 | #ifdef CONFIG_EDAC_DEBUG |
727 | if (edac_debug_level >= 3) | 734 | if (edac_debug_level >= 3) |
728 | edac_mc_dump_mci(mci); | 735 | edac_mc_dump_mci(mci); |
@@ -762,6 +769,8 @@ int edac_mc_add_mc(struct mem_ctl_info *mci) | |||
762 | /* set load time so that error rate can be tracked */ | 769 | /* set load time so that error rate can be tracked */ |
763 | mci->start_time = jiffies; | 770 | mci->start_time = jiffies; |
764 | 771 | ||
772 | mci->bus = &mc_bus[mci->mc_idx]; | ||
773 | |||
765 | if (edac_create_sysfs_mci_device(mci)) { | 774 | if (edac_create_sysfs_mci_device(mci)) { |
766 | edac_mc_printk(mci, KERN_WARNING, | 775 | edac_mc_printk(mci, KERN_WARNING, |
767 | "failed to create sysfs device\n"); | 776 | "failed to create sysfs device\n"); |
diff --git a/drivers/edac/edac_mc_sysfs.c b/drivers/edac/edac_mc_sysfs.c index ef15a7e613bc..e7c32c4f7837 100644 --- a/drivers/edac/edac_mc_sysfs.c +++ b/drivers/edac/edac_mc_sysfs.c | |||
@@ -370,7 +370,7 @@ static int edac_create_csrow_object(struct mem_ctl_info *mci, | |||
370 | return -ENODEV; | 370 | return -ENODEV; |
371 | 371 | ||
372 | csrow->dev.type = &csrow_attr_type; | 372 | csrow->dev.type = &csrow_attr_type; |
373 | csrow->dev.bus = &mci->bus; | 373 | csrow->dev.bus = mci->bus; |
374 | device_initialize(&csrow->dev); | 374 | device_initialize(&csrow->dev); |
375 | csrow->dev.parent = &mci->dev; | 375 | csrow->dev.parent = &mci->dev; |
376 | csrow->mci = mci; | 376 | csrow->mci = mci; |
@@ -605,7 +605,7 @@ static int edac_create_dimm_object(struct mem_ctl_info *mci, | |||
605 | dimm->mci = mci; | 605 | dimm->mci = mci; |
606 | 606 | ||
607 | dimm->dev.type = &dimm_attr_type; | 607 | dimm->dev.type = &dimm_attr_type; |
608 | dimm->dev.bus = &mci->bus; | 608 | dimm->dev.bus = mci->bus; |
609 | device_initialize(&dimm->dev); | 609 | device_initialize(&dimm->dev); |
610 | 610 | ||
611 | dimm->dev.parent = &mci->dev; | 611 | dimm->dev.parent = &mci->dev; |
@@ -975,11 +975,13 @@ int edac_create_sysfs_mci_device(struct mem_ctl_info *mci) | |||
975 | * The memory controller needs its own bus, in order to avoid | 975 | * The memory controller needs its own bus, in order to avoid |
976 | * namespace conflicts at /sys/bus/edac. | 976 | * namespace conflicts at /sys/bus/edac. |
977 | */ | 977 | */ |
978 | mci->bus.name = kasprintf(GFP_KERNEL, "mc%d", mci->mc_idx); | 978 | mci->bus->name = kasprintf(GFP_KERNEL, "mc%d", mci->mc_idx); |
979 | if (!mci->bus.name) | 979 | if (!mci->bus->name) |
980 | return -ENOMEM; | 980 | return -ENOMEM; |
981 | edac_dbg(0, "creating bus %s\n", mci->bus.name); | 981 | |
982 | err = bus_register(&mci->bus); | 982 | edac_dbg(0, "creating bus %s\n", mci->bus->name); |
983 | |||
984 | err = bus_register(mci->bus); | ||
983 | if (err < 0) | 985 | if (err < 0) |
984 | return err; | 986 | return err; |
985 | 987 | ||
@@ -988,7 +990,7 @@ int edac_create_sysfs_mci_device(struct mem_ctl_info *mci) | |||
988 | device_initialize(&mci->dev); | 990 | device_initialize(&mci->dev); |
989 | 991 | ||
990 | mci->dev.parent = mci_pdev; | 992 | mci->dev.parent = mci_pdev; |
991 | mci->dev.bus = &mci->bus; | 993 | mci->dev.bus = mci->bus; |
992 | dev_set_name(&mci->dev, "mc%d", mci->mc_idx); | 994 | dev_set_name(&mci->dev, "mc%d", mci->mc_idx); |
993 | dev_set_drvdata(&mci->dev, mci); | 995 | dev_set_drvdata(&mci->dev, mci); |
994 | pm_runtime_forbid(&mci->dev); | 996 | pm_runtime_forbid(&mci->dev); |
@@ -997,8 +999,8 @@ int edac_create_sysfs_mci_device(struct mem_ctl_info *mci) | |||
997 | err = device_add(&mci->dev); | 999 | err = device_add(&mci->dev); |
998 | if (err < 0) { | 1000 | if (err < 0) { |
999 | edac_dbg(1, "failure: create device %s\n", dev_name(&mci->dev)); | 1001 | edac_dbg(1, "failure: create device %s\n", dev_name(&mci->dev)); |
1000 | bus_unregister(&mci->bus); | 1002 | bus_unregister(mci->bus); |
1001 | kfree(mci->bus.name); | 1003 | kfree(mci->bus->name); |
1002 | return err; | 1004 | return err; |
1003 | } | 1005 | } |
1004 | 1006 | ||
@@ -1064,8 +1066,8 @@ fail: | |||
1064 | } | 1066 | } |
1065 | fail2: | 1067 | fail2: |
1066 | device_unregister(&mci->dev); | 1068 | device_unregister(&mci->dev); |
1067 | bus_unregister(&mci->bus); | 1069 | bus_unregister(mci->bus); |
1068 | kfree(mci->bus.name); | 1070 | kfree(mci->bus->name); |
1069 | return err; | 1071 | return err; |
1070 | } | 1072 | } |
1071 | 1073 | ||
@@ -1098,8 +1100,8 @@ void edac_unregister_sysfs(struct mem_ctl_info *mci) | |||
1098 | { | 1100 | { |
1099 | edac_dbg(1, "Unregistering device %s\n", dev_name(&mci->dev)); | 1101 | edac_dbg(1, "Unregistering device %s\n", dev_name(&mci->dev)); |
1100 | device_unregister(&mci->dev); | 1102 | device_unregister(&mci->dev); |
1101 | bus_unregister(&mci->bus); | 1103 | bus_unregister(mci->bus); |
1102 | kfree(mci->bus.name); | 1104 | kfree(mci->bus->name); |
1103 | } | 1105 | } |
1104 | 1106 | ||
1105 | static void mc_attr_release(struct device *dev) | 1107 | static void mc_attr_release(struct device *dev) |
diff --git a/drivers/edac/i5100_edac.c b/drivers/edac/i5100_edac.c index 1b635178cc44..157b934e8ce3 100644 --- a/drivers/edac/i5100_edac.c +++ b/drivers/edac/i5100_edac.c | |||
@@ -974,7 +974,7 @@ static int i5100_setup_debugfs(struct mem_ctl_info *mci) | |||
974 | if (!i5100_debugfs) | 974 | if (!i5100_debugfs) |
975 | return -ENODEV; | 975 | return -ENODEV; |
976 | 976 | ||
977 | priv->debugfs = debugfs_create_dir(mci->bus.name, i5100_debugfs); | 977 | priv->debugfs = debugfs_create_dir(mci->bus->name, i5100_debugfs); |
978 | 978 | ||
979 | if (!priv->debugfs) | 979 | if (!priv->debugfs) |
980 | return -ENOMEM; | 980 | return -ENOMEM; |
diff --git a/drivers/gpio/gpio-msm-v2.c b/drivers/gpio/gpio-msm-v2.c index f4491a497cc8..c2fa77086eb5 100644 --- a/drivers/gpio/gpio-msm-v2.c +++ b/drivers/gpio/gpio-msm-v2.c | |||
@@ -378,7 +378,7 @@ static int msm_gpio_probe(struct platform_device *pdev) | |||
378 | int ret, ngpio; | 378 | int ret, ngpio; |
379 | struct resource *res; | 379 | struct resource *res; |
380 | 380 | ||
381 | if (!of_property_read_u32(pdev->dev.of_node, "ngpio", &ngpio)) { | 381 | if (of_property_read_u32(pdev->dev.of_node, "ngpio", &ngpio)) { |
382 | dev_err(&pdev->dev, "%s: ngpio property missing\n", __func__); | 382 | dev_err(&pdev->dev, "%s: ngpio property missing\n", __func__); |
383 | return -EINVAL; | 383 | return -EINVAL; |
384 | } | 384 | } |
diff --git a/drivers/gpio/gpio-omap.c b/drivers/gpio/gpio-omap.c index dfeb3a3a8f20..c57244ef428b 100644 --- a/drivers/gpio/gpio-omap.c +++ b/drivers/gpio/gpio-omap.c | |||
@@ -1037,6 +1037,18 @@ omap_mpuio_alloc_gc(struct gpio_bank *bank, unsigned int irq_start, | |||
1037 | IRQ_NOREQUEST | IRQ_NOPROBE, 0); | 1037 | IRQ_NOREQUEST | IRQ_NOPROBE, 0); |
1038 | } | 1038 | } |
1039 | 1039 | ||
1040 | #if defined(CONFIG_OF_GPIO) | ||
1041 | static inline bool omap_gpio_chip_boot_dt(struct gpio_chip *chip) | ||
1042 | { | ||
1043 | return chip->of_node != NULL; | ||
1044 | } | ||
1045 | #else | ||
1046 | static inline bool omap_gpio_chip_boot_dt(struct gpio_chip *chip) | ||
1047 | { | ||
1048 | return false; | ||
1049 | } | ||
1050 | #endif | ||
1051 | |||
1040 | static void omap_gpio_chip_init(struct gpio_bank *bank) | 1052 | static void omap_gpio_chip_init(struct gpio_bank *bank) |
1041 | { | 1053 | { |
1042 | int j; | 1054 | int j; |
@@ -1068,24 +1080,68 @@ static void omap_gpio_chip_init(struct gpio_bank *bank) | |||
1068 | 1080 | ||
1069 | gpiochip_add(&bank->chip); | 1081 | gpiochip_add(&bank->chip); |
1070 | 1082 | ||
1071 | for (j = 0; j < bank->width; j++) { | 1083 | /* |
1072 | int irq = irq_create_mapping(bank->domain, j); | 1084 | * REVISIT these explicit calls to irq_create_mapping() |
1073 | irq_set_lockdep_class(irq, &gpio_lock_class); | 1085 | * to do the GPIO to IRQ domain mapping for each GPIO in |
1074 | irq_set_chip_data(irq, bank); | 1086 | * the bank can be removed once all OMAP platforms have |
1075 | if (bank->is_mpuio) { | 1087 | * been migrated to Device Tree boot only. |
1076 | omap_mpuio_alloc_gc(bank, irq, bank->width); | 1088 | * Since in DT boot irq_create_mapping() is called from |
1077 | } else { | 1089 | * irq_create_of_mapping() only for the GPIO lines that |
1078 | irq_set_chip_and_handler(irq, &gpio_irq_chip, | 1090 | * are used as interrupts. |
1079 | handle_simple_irq); | 1091 | */ |
1080 | set_irq_flags(irq, IRQF_VALID); | 1092 | if (!omap_gpio_chip_boot_dt(&bank->chip)) |
1081 | } | 1093 | for (j = 0; j < bank->width; j++) |
1082 | } | 1094 | irq_create_mapping(bank->domain, j); |
1083 | irq_set_chained_handler(bank->irq, gpio_irq_handler); | 1095 | irq_set_chained_handler(bank->irq, gpio_irq_handler); |
1084 | irq_set_handler_data(bank->irq, bank); | 1096 | irq_set_handler_data(bank->irq, bank); |
1085 | } | 1097 | } |
1086 | 1098 | ||
1087 | static const struct of_device_id omap_gpio_match[]; | 1099 | static const struct of_device_id omap_gpio_match[]; |
1088 | 1100 | ||
1101 | static int omap_gpio_irq_map(struct irq_domain *d, unsigned int virq, | ||
1102 | irq_hw_number_t hwirq) | ||
1103 | { | ||
1104 | struct gpio_bank *bank = d->host_data; | ||
1105 | int gpio; | ||
1106 | int ret; | ||
1107 | |||
1108 | if (!bank) | ||
1109 | return -EINVAL; | ||
1110 | |||
1111 | irq_set_lockdep_class(virq, &gpio_lock_class); | ||
1112 | irq_set_chip_data(virq, bank); | ||
1113 | if (bank->is_mpuio) { | ||
1114 | omap_mpuio_alloc_gc(bank, virq, bank->width); | ||
1115 | } else { | ||
1116 | irq_set_chip_and_handler(virq, &gpio_irq_chip, | ||
1117 | handle_simple_irq); | ||
1118 | set_irq_flags(virq, IRQF_VALID); | ||
1119 | } | ||
1120 | |||
1121 | /* | ||
1122 | * REVISIT most GPIO IRQ chip drivers need to call | ||
1123 | * gpio_request() before a GPIO line can be used as an | ||
1124 | * IRQ. Ideally this should be handled by the IRQ core | ||
1125 | * but until then this has to be done on a per driver | ||
1126 | * basis. Remove this once this is managed by the core. | ||
1127 | */ | ||
1128 | if (omap_gpio_chip_boot_dt(&bank->chip)) { | ||
1129 | gpio = irq_to_gpio(bank, hwirq); | ||
1130 | ret = gpio_request_one(gpio, GPIOF_IN, NULL); | ||
1131 | if (ret) { | ||
1132 | dev_err(bank->dev, "Could not request GPIO%d\n", gpio); | ||
1133 | return ret; | ||
1134 | } | ||
1135 | } | ||
1136 | |||
1137 | return 0; | ||
1138 | } | ||
1139 | |||
1140 | static struct irq_domain_ops omap_gpio_irq_ops = { | ||
1141 | .xlate = irq_domain_xlate_onetwocell, | ||
1142 | .map = omap_gpio_irq_map, | ||
1143 | }; | ||
1144 | |||
1089 | static int omap_gpio_probe(struct platform_device *pdev) | 1145 | static int omap_gpio_probe(struct platform_device *pdev) |
1090 | { | 1146 | { |
1091 | struct device *dev = &pdev->dev; | 1147 | struct device *dev = &pdev->dev; |
@@ -1151,10 +1207,10 @@ static int omap_gpio_probe(struct platform_device *pdev) | |||
1151 | } | 1207 | } |
1152 | 1208 | ||
1153 | bank->domain = irq_domain_add_legacy(node, bank->width, irq_base, | 1209 | bank->domain = irq_domain_add_legacy(node, bank->width, irq_base, |
1154 | 0, &irq_domain_simple_ops, NULL); | 1210 | 0, &omap_gpio_irq_ops, bank); |
1155 | #else | 1211 | #else |
1156 | bank->domain = irq_domain_add_linear(node, bank->width, | 1212 | bank->domain = irq_domain_add_linear(node, bank->width, |
1157 | &irq_domain_simple_ops, NULL); | 1213 | &omap_gpio_irq_ops, bank); |
1158 | #endif | 1214 | #endif |
1159 | if (!bank->domain) { | 1215 | if (!bank->domain) { |
1160 | dev_err(dev, "Couldn't register an IRQ domain\n"); | 1216 | dev_err(dev, "Couldn't register an IRQ domain\n"); |
diff --git a/drivers/gpu/drm/drm_crtc_helper.c b/drivers/gpu/drm/drm_crtc_helper.c index 738a4294d820..6a647493ca7f 100644 --- a/drivers/gpu/drm/drm_crtc_helper.c +++ b/drivers/gpu/drm/drm_crtc_helper.c | |||
@@ -677,6 +677,11 @@ int drm_crtc_helper_set_config(struct drm_mode_set *set) | |||
677 | /* don't break so fail path works correct */ | 677 | /* don't break so fail path works correct */ |
678 | fail = 1; | 678 | fail = 1; |
679 | break; | 679 | break; |
680 | |||
681 | if (connector->dpms != DRM_MODE_DPMS_ON) { | ||
682 | DRM_DEBUG_KMS("connector dpms not on, full mode switch\n"); | ||
683 | mode_changed = true; | ||
684 | } | ||
680 | } | 685 | } |
681 | } | 686 | } |
682 | 687 | ||
@@ -754,6 +759,12 @@ int drm_crtc_helper_set_config(struct drm_mode_set *set) | |||
754 | ret = -EINVAL; | 759 | ret = -EINVAL; |
755 | goto fail; | 760 | goto fail; |
756 | } | 761 | } |
762 | DRM_DEBUG_KMS("Setting connector DPMS state to on\n"); | ||
763 | for (i = 0; i < set->num_connectors; i++) { | ||
764 | DRM_DEBUG_KMS("\t[CONNECTOR:%d:%s] set DPMS on\n", set->connectors[i]->base.id, | ||
765 | drm_get_connector_name(set->connectors[i])); | ||
766 | set->connectors[i]->funcs->dpms(set->connectors[i], DRM_MODE_DPMS_ON); | ||
767 | } | ||
757 | } | 768 | } |
758 | drm_helper_disable_unused_functions(dev); | 769 | drm_helper_disable_unused_functions(dev); |
759 | } else if (fb_changed) { | 770 | } else if (fb_changed) { |
@@ -771,22 +782,6 @@ int drm_crtc_helper_set_config(struct drm_mode_set *set) | |||
771 | } | 782 | } |
772 | } | 783 | } |
773 | 784 | ||
774 | /* | ||
775 | * crtc set_config helpers implicit set the crtc and all connected | ||
776 | * encoders to DPMS on for a full mode set. But for just an fb update it | ||
777 | * doesn't do that. To not confuse userspace, do an explicit DPMS_ON | ||
778 | * unconditionally. This will also ensure driver internal dpms state is | ||
779 | * consistent again. | ||
780 | */ | ||
781 | if (set->crtc->enabled) { | ||
782 | DRM_DEBUG_KMS("Setting connector DPMS state to on\n"); | ||
783 | for (i = 0; i < set->num_connectors; i++) { | ||
784 | DRM_DEBUG_KMS("\t[CONNECTOR:%d:%s] set DPMS on\n", set->connectors[i]->base.id, | ||
785 | drm_get_connector_name(set->connectors[i])); | ||
786 | set->connectors[i]->funcs->dpms(set->connectors[i], DRM_MODE_DPMS_ON); | ||
787 | } | ||
788 | } | ||
789 | |||
790 | kfree(save_connectors); | 785 | kfree(save_connectors); |
791 | kfree(save_encoders); | 786 | kfree(save_encoders); |
792 | kfree(save_crtcs); | 787 | kfree(save_crtcs); |
diff --git a/drivers/gpu/drm/i915/i915_dma.c b/drivers/gpu/drm/i915/i915_dma.c index cf188ab7051a..66c63808fa35 100644 --- a/drivers/gpu/drm/i915/i915_dma.c +++ b/drivers/gpu/drm/i915/i915_dma.c | |||
@@ -1495,6 +1495,15 @@ int i915_driver_load(struct drm_device *dev, unsigned long flags) | |||
1495 | dev_priv->dev = dev; | 1495 | dev_priv->dev = dev; |
1496 | dev_priv->info = info; | 1496 | dev_priv->info = info; |
1497 | 1497 | ||
1498 | spin_lock_init(&dev_priv->irq_lock); | ||
1499 | spin_lock_init(&dev_priv->gpu_error.lock); | ||
1500 | spin_lock_init(&dev_priv->rps.lock); | ||
1501 | spin_lock_init(&dev_priv->gt_lock); | ||
1502 | spin_lock_init(&dev_priv->backlight.lock); | ||
1503 | mutex_init(&dev_priv->dpio_lock); | ||
1504 | mutex_init(&dev_priv->rps.hw_lock); | ||
1505 | mutex_init(&dev_priv->modeset_restore_lock); | ||
1506 | |||
1498 | i915_dump_device_info(dev_priv); | 1507 | i915_dump_device_info(dev_priv); |
1499 | 1508 | ||
1500 | if (i915_get_bridge_dev(dev)) { | 1509 | if (i915_get_bridge_dev(dev)) { |
@@ -1585,6 +1594,7 @@ int i915_driver_load(struct drm_device *dev, unsigned long flags) | |||
1585 | intel_detect_pch(dev); | 1594 | intel_detect_pch(dev); |
1586 | 1595 | ||
1587 | intel_irq_init(dev); | 1596 | intel_irq_init(dev); |
1597 | intel_gt_sanitize(dev); | ||
1588 | intel_gt_init(dev); | 1598 | intel_gt_init(dev); |
1589 | 1599 | ||
1590 | /* Try to make sure MCHBAR is enabled before poking at it */ | 1600 | /* Try to make sure MCHBAR is enabled before poking at it */ |
@@ -1610,15 +1620,6 @@ int i915_driver_load(struct drm_device *dev, unsigned long flags) | |||
1610 | if (!IS_I945G(dev) && !IS_I945GM(dev)) | 1620 | if (!IS_I945G(dev) && !IS_I945GM(dev)) |
1611 | pci_enable_msi(dev->pdev); | 1621 | pci_enable_msi(dev->pdev); |
1612 | 1622 | ||
1613 | spin_lock_init(&dev_priv->irq_lock); | ||
1614 | spin_lock_init(&dev_priv->gpu_error.lock); | ||
1615 | spin_lock_init(&dev_priv->rps.lock); | ||
1616 | spin_lock_init(&dev_priv->backlight.lock); | ||
1617 | mutex_init(&dev_priv->dpio_lock); | ||
1618 | |||
1619 | mutex_init(&dev_priv->rps.hw_lock); | ||
1620 | mutex_init(&dev_priv->modeset_restore_lock); | ||
1621 | |||
1622 | dev_priv->num_plane = 1; | 1623 | dev_priv->num_plane = 1; |
1623 | if (IS_VALLEYVIEW(dev)) | 1624 | if (IS_VALLEYVIEW(dev)) |
1624 | dev_priv->num_plane = 2; | 1625 | dev_priv->num_plane = 2; |
@@ -1648,7 +1649,7 @@ int i915_driver_load(struct drm_device *dev, unsigned long flags) | |||
1648 | if (INTEL_INFO(dev)->num_pipes) { | 1649 | if (INTEL_INFO(dev)->num_pipes) { |
1649 | /* Must be done after probing outputs */ | 1650 | /* Must be done after probing outputs */ |
1650 | intel_opregion_init(dev); | 1651 | intel_opregion_init(dev); |
1651 | acpi_video_register_with_quirks(); | 1652 | acpi_video_register(); |
1652 | } | 1653 | } |
1653 | 1654 | ||
1654 | if (IS_GEN5(dev)) | 1655 | if (IS_GEN5(dev)) |
diff --git a/drivers/gpu/drm/i915/i915_drv.c b/drivers/gpu/drm/i915/i915_drv.c index f4af1ca0fb62..45b3c030f483 100644 --- a/drivers/gpu/drm/i915/i915_drv.c +++ b/drivers/gpu/drm/i915/i915_drv.c | |||
@@ -706,7 +706,7 @@ static int i915_drm_thaw(struct drm_device *dev) | |||
706 | { | 706 | { |
707 | int error = 0; | 707 | int error = 0; |
708 | 708 | ||
709 | intel_gt_reset(dev); | 709 | intel_gt_sanitize(dev); |
710 | 710 | ||
711 | if (drm_core_check_feature(dev, DRIVER_MODESET)) { | 711 | if (drm_core_check_feature(dev, DRIVER_MODESET)) { |
712 | mutex_lock(&dev->struct_mutex); | 712 | mutex_lock(&dev->struct_mutex); |
@@ -732,7 +732,7 @@ int i915_resume(struct drm_device *dev) | |||
732 | 732 | ||
733 | pci_set_master(dev->pdev); | 733 | pci_set_master(dev->pdev); |
734 | 734 | ||
735 | intel_gt_reset(dev); | 735 | intel_gt_sanitize(dev); |
736 | 736 | ||
737 | /* | 737 | /* |
738 | * Platforms with opregion should have sane BIOS, older ones (gen3 and | 738 | * Platforms with opregion should have sane BIOS, older ones (gen3 and |
@@ -1253,21 +1253,21 @@ hsw_unclaimed_reg_check(struct drm_i915_private *dev_priv, u32 reg) | |||
1253 | 1253 | ||
1254 | #define __i915_read(x, y) \ | 1254 | #define __i915_read(x, y) \ |
1255 | u##x i915_read##x(struct drm_i915_private *dev_priv, u32 reg) { \ | 1255 | u##x i915_read##x(struct drm_i915_private *dev_priv, u32 reg) { \ |
1256 | unsigned long irqflags; \ | ||
1256 | u##x val = 0; \ | 1257 | u##x val = 0; \ |
1258 | spin_lock_irqsave(&dev_priv->gt_lock, irqflags); \ | ||
1257 | if (IS_GEN5(dev_priv->dev)) \ | 1259 | if (IS_GEN5(dev_priv->dev)) \ |
1258 | ilk_dummy_write(dev_priv); \ | 1260 | ilk_dummy_write(dev_priv); \ |
1259 | if (NEEDS_FORCE_WAKE((dev_priv), (reg))) { \ | 1261 | if (NEEDS_FORCE_WAKE((dev_priv), (reg))) { \ |
1260 | unsigned long irqflags; \ | ||
1261 | spin_lock_irqsave(&dev_priv->gt_lock, irqflags); \ | ||
1262 | if (dev_priv->forcewake_count == 0) \ | 1262 | if (dev_priv->forcewake_count == 0) \ |
1263 | dev_priv->gt.force_wake_get(dev_priv); \ | 1263 | dev_priv->gt.force_wake_get(dev_priv); \ |
1264 | val = read##y(dev_priv->regs + reg); \ | 1264 | val = read##y(dev_priv->regs + reg); \ |
1265 | if (dev_priv->forcewake_count == 0) \ | 1265 | if (dev_priv->forcewake_count == 0) \ |
1266 | dev_priv->gt.force_wake_put(dev_priv); \ | 1266 | dev_priv->gt.force_wake_put(dev_priv); \ |
1267 | spin_unlock_irqrestore(&dev_priv->gt_lock, irqflags); \ | ||
1268 | } else { \ | 1267 | } else { \ |
1269 | val = read##y(dev_priv->regs + reg); \ | 1268 | val = read##y(dev_priv->regs + reg); \ |
1270 | } \ | 1269 | } \ |
1270 | spin_unlock_irqrestore(&dev_priv->gt_lock, irqflags); \ | ||
1271 | trace_i915_reg_rw(false, reg, val, sizeof(val)); \ | 1271 | trace_i915_reg_rw(false, reg, val, sizeof(val)); \ |
1272 | return val; \ | 1272 | return val; \ |
1273 | } | 1273 | } |
@@ -1280,8 +1280,10 @@ __i915_read(64, q) | |||
1280 | 1280 | ||
1281 | #define __i915_write(x, y) \ | 1281 | #define __i915_write(x, y) \ |
1282 | void i915_write##x(struct drm_i915_private *dev_priv, u32 reg, u##x val) { \ | 1282 | void i915_write##x(struct drm_i915_private *dev_priv, u32 reg, u##x val) { \ |
1283 | unsigned long irqflags; \ | ||
1283 | u32 __fifo_ret = 0; \ | 1284 | u32 __fifo_ret = 0; \ |
1284 | trace_i915_reg_rw(true, reg, val, sizeof(val)); \ | 1285 | trace_i915_reg_rw(true, reg, val, sizeof(val)); \ |
1286 | spin_lock_irqsave(&dev_priv->gt_lock, irqflags); \ | ||
1285 | if (NEEDS_FORCE_WAKE((dev_priv), (reg))) { \ | 1287 | if (NEEDS_FORCE_WAKE((dev_priv), (reg))) { \ |
1286 | __fifo_ret = __gen6_gt_wait_for_fifo(dev_priv); \ | 1288 | __fifo_ret = __gen6_gt_wait_for_fifo(dev_priv); \ |
1287 | } \ | 1289 | } \ |
@@ -1293,6 +1295,7 @@ void i915_write##x(struct drm_i915_private *dev_priv, u32 reg, u##x val) { \ | |||
1293 | gen6_gt_check_fifodbg(dev_priv); \ | 1295 | gen6_gt_check_fifodbg(dev_priv); \ |
1294 | } \ | 1296 | } \ |
1295 | hsw_unclaimed_reg_check(dev_priv, reg); \ | 1297 | hsw_unclaimed_reg_check(dev_priv, reg); \ |
1298 | spin_unlock_irqrestore(&dev_priv->gt_lock, irqflags); \ | ||
1296 | } | 1299 | } |
1297 | __i915_write(8, b) | 1300 | __i915_write(8, b) |
1298 | __i915_write(16, w) | 1301 | __i915_write(16, w) |
diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h index a416645bcd23..d2ee3343c943 100644 --- a/drivers/gpu/drm/i915/i915_drv.h +++ b/drivers/gpu/drm/i915/i915_drv.h | |||
@@ -555,6 +555,7 @@ enum intel_sbi_destination { | |||
555 | #define QUIRK_PIPEA_FORCE (1<<0) | 555 | #define QUIRK_PIPEA_FORCE (1<<0) |
556 | #define QUIRK_LVDS_SSC_DISABLE (1<<1) | 556 | #define QUIRK_LVDS_SSC_DISABLE (1<<1) |
557 | #define QUIRK_INVERT_BRIGHTNESS (1<<2) | 557 | #define QUIRK_INVERT_BRIGHTNESS (1<<2) |
558 | #define QUIRK_NO_PCH_PWM_ENABLE (1<<3) | ||
558 | 559 | ||
559 | struct intel_fbdev; | 560 | struct intel_fbdev; |
560 | struct intel_fbc_work; | 561 | struct intel_fbc_work; |
@@ -1583,7 +1584,7 @@ void i915_handle_error(struct drm_device *dev, bool wedged); | |||
1583 | extern void intel_irq_init(struct drm_device *dev); | 1584 | extern void intel_irq_init(struct drm_device *dev); |
1584 | extern void intel_hpd_init(struct drm_device *dev); | 1585 | extern void intel_hpd_init(struct drm_device *dev); |
1585 | extern void intel_gt_init(struct drm_device *dev); | 1586 | extern void intel_gt_init(struct drm_device *dev); |
1586 | extern void intel_gt_reset(struct drm_device *dev); | 1587 | extern void intel_gt_sanitize(struct drm_device *dev); |
1587 | 1588 | ||
1588 | void i915_error_state_free(struct kref *error_ref); | 1589 | void i915_error_state_free(struct kref *error_ref); |
1589 | 1590 | ||
diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c index 97afd2639fb6..d9e2208cfe98 100644 --- a/drivers/gpu/drm/i915/i915_gem.c +++ b/drivers/gpu/drm/i915/i915_gem.c | |||
@@ -2258,7 +2258,17 @@ void i915_gem_restore_fences(struct drm_device *dev) | |||
2258 | 2258 | ||
2259 | for (i = 0; i < dev_priv->num_fence_regs; i++) { | 2259 | for (i = 0; i < dev_priv->num_fence_regs; i++) { |
2260 | struct drm_i915_fence_reg *reg = &dev_priv->fence_regs[i]; | 2260 | struct drm_i915_fence_reg *reg = &dev_priv->fence_regs[i]; |
2261 | i915_gem_write_fence(dev, i, reg->obj); | 2261 | |
2262 | /* | ||
2263 | * Commit delayed tiling changes if we have an object still | ||
2264 | * attached to the fence, otherwise just clear the fence. | ||
2265 | */ | ||
2266 | if (reg->obj) { | ||
2267 | i915_gem_object_update_fence(reg->obj, reg, | ||
2268 | reg->obj->tiling_mode); | ||
2269 | } else { | ||
2270 | i915_gem_write_fence(dev, i, NULL); | ||
2271 | } | ||
2262 | } | 2272 | } |
2263 | } | 2273 | } |
2264 | 2274 | ||
@@ -2795,6 +2805,10 @@ static void i915_gem_write_fence(struct drm_device *dev, int reg, | |||
2795 | if (i915_gem_object_needs_mb(dev_priv->fence_regs[reg].obj)) | 2805 | if (i915_gem_object_needs_mb(dev_priv->fence_regs[reg].obj)) |
2796 | mb(); | 2806 | mb(); |
2797 | 2807 | ||
2808 | WARN(obj && (!obj->stride || !obj->tiling_mode), | ||
2809 | "bogus fence setup with stride: 0x%x, tiling mode: %i\n", | ||
2810 | obj->stride, obj->tiling_mode); | ||
2811 | |||
2798 | switch (INTEL_INFO(dev)->gen) { | 2812 | switch (INTEL_INFO(dev)->gen) { |
2799 | case 7: | 2813 | case 7: |
2800 | case 6: | 2814 | case 6: |
@@ -2836,6 +2850,7 @@ static void i915_gem_object_update_fence(struct drm_i915_gem_object *obj, | |||
2836 | fence->obj = NULL; | 2850 | fence->obj = NULL; |
2837 | list_del_init(&fence->lru_list); | 2851 | list_del_init(&fence->lru_list); |
2838 | } | 2852 | } |
2853 | obj->fence_dirty = false; | ||
2839 | } | 2854 | } |
2840 | 2855 | ||
2841 | static int | 2856 | static int |
@@ -2965,7 +2980,6 @@ i915_gem_object_get_fence(struct drm_i915_gem_object *obj) | |||
2965 | return 0; | 2980 | return 0; |
2966 | 2981 | ||
2967 | i915_gem_object_update_fence(obj, reg, enable); | 2982 | i915_gem_object_update_fence(obj, reg, enable); |
2968 | obj->fence_dirty = false; | ||
2969 | 2983 | ||
2970 | return 0; | 2984 | return 0; |
2971 | } | 2985 | } |
diff --git a/drivers/gpu/drm/i915/intel_ddi.c b/drivers/gpu/drm/i915/intel_ddi.c index 324211ac9c55..b042ee5c4070 100644 --- a/drivers/gpu/drm/i915/intel_ddi.c +++ b/drivers/gpu/drm/i915/intel_ddi.c | |||
@@ -301,7 +301,7 @@ static void intel_ddi_mode_set(struct drm_encoder *encoder, | |||
301 | struct intel_digital_port *intel_dig_port = | 301 | struct intel_digital_port *intel_dig_port = |
302 | enc_to_dig_port(encoder); | 302 | enc_to_dig_port(encoder); |
303 | 303 | ||
304 | intel_dp->DP = intel_dig_port->port_reversal | | 304 | intel_dp->DP = intel_dig_port->saved_port_bits | |
305 | DDI_BUF_CTL_ENABLE | DDI_BUF_EMP_400MV_0DB_HSW; | 305 | DDI_BUF_CTL_ENABLE | DDI_BUF_EMP_400MV_0DB_HSW; |
306 | intel_dp->DP |= DDI_PORT_WIDTH(intel_dp->lane_count); | 306 | intel_dp->DP |= DDI_PORT_WIDTH(intel_dp->lane_count); |
307 | 307 | ||
@@ -1109,7 +1109,8 @@ static void intel_enable_ddi(struct intel_encoder *intel_encoder) | |||
1109 | * enabling the port. | 1109 | * enabling the port. |
1110 | */ | 1110 | */ |
1111 | I915_WRITE(DDI_BUF_CTL(port), | 1111 | I915_WRITE(DDI_BUF_CTL(port), |
1112 | intel_dig_port->port_reversal | DDI_BUF_CTL_ENABLE); | 1112 | intel_dig_port->saved_port_bits | |
1113 | DDI_BUF_CTL_ENABLE); | ||
1113 | } else if (type == INTEL_OUTPUT_EDP) { | 1114 | } else if (type == INTEL_OUTPUT_EDP) { |
1114 | struct intel_dp *intel_dp = enc_to_intel_dp(encoder); | 1115 | struct intel_dp *intel_dp = enc_to_intel_dp(encoder); |
1115 | 1116 | ||
@@ -1347,8 +1348,9 @@ void intel_ddi_init(struct drm_device *dev, enum port port) | |||
1347 | intel_encoder->get_config = intel_ddi_get_config; | 1348 | intel_encoder->get_config = intel_ddi_get_config; |
1348 | 1349 | ||
1349 | intel_dig_port->port = port; | 1350 | intel_dig_port->port = port; |
1350 | intel_dig_port->port_reversal = I915_READ(DDI_BUF_CTL(port)) & | 1351 | intel_dig_port->saved_port_bits = I915_READ(DDI_BUF_CTL(port)) & |
1351 | DDI_BUF_PORT_REVERSAL; | 1352 | (DDI_BUF_PORT_REVERSAL | |
1353 | DDI_A_4_LANES); | ||
1352 | intel_dig_port->dp.output_reg = DDI_BUF_CTL(port); | 1354 | intel_dig_port->dp.output_reg = DDI_BUF_CTL(port); |
1353 | 1355 | ||
1354 | intel_encoder->type = INTEL_OUTPUT_UNKNOWN; | 1356 | intel_encoder->type = INTEL_OUTPUT_UNKNOWN; |
diff --git a/drivers/gpu/drm/i915/intel_display.c b/drivers/gpu/drm/i915/intel_display.c index 85f3eb74d2b7..5fb305840db8 100644 --- a/drivers/gpu/drm/i915/intel_display.c +++ b/drivers/gpu/drm/i915/intel_display.c | |||
@@ -4913,22 +4913,19 @@ static void i9xx_get_pfit_config(struct intel_crtc *crtc, | |||
4913 | uint32_t tmp; | 4913 | uint32_t tmp; |
4914 | 4914 | ||
4915 | tmp = I915_READ(PFIT_CONTROL); | 4915 | tmp = I915_READ(PFIT_CONTROL); |
4916 | if (!(tmp & PFIT_ENABLE)) | ||
4917 | return; | ||
4916 | 4918 | ||
4919 | /* Check whether the pfit is attached to our pipe. */ | ||
4917 | if (INTEL_INFO(dev)->gen < 4) { | 4920 | if (INTEL_INFO(dev)->gen < 4) { |
4918 | if (crtc->pipe != PIPE_B) | 4921 | if (crtc->pipe != PIPE_B) |
4919 | return; | 4922 | return; |
4920 | |||
4921 | /* gen2/3 store dither state in pfit control, needs to match */ | ||
4922 | pipe_config->gmch_pfit.control = tmp & PANEL_8TO6_DITHER_ENABLE; | ||
4923 | } else { | 4923 | } else { |
4924 | if ((tmp & PFIT_PIPE_MASK) != (crtc->pipe << PFIT_PIPE_SHIFT)) | 4924 | if ((tmp & PFIT_PIPE_MASK) != (crtc->pipe << PFIT_PIPE_SHIFT)) |
4925 | return; | 4925 | return; |
4926 | } | 4926 | } |
4927 | 4927 | ||
4928 | if (!(tmp & PFIT_ENABLE)) | 4928 | pipe_config->gmch_pfit.control = tmp; |
4929 | return; | ||
4930 | |||
4931 | pipe_config->gmch_pfit.control = I915_READ(PFIT_CONTROL); | ||
4932 | pipe_config->gmch_pfit.pgm_ratios = I915_READ(PFIT_PGM_RATIOS); | 4929 | pipe_config->gmch_pfit.pgm_ratios = I915_READ(PFIT_PGM_RATIOS); |
4933 | if (INTEL_INFO(dev)->gen < 5) | 4930 | if (INTEL_INFO(dev)->gen < 5) |
4934 | pipe_config->gmch_pfit.lvds_border_bits = | 4931 | pipe_config->gmch_pfit.lvds_border_bits = |
@@ -8317,6 +8314,8 @@ check_shared_dpll_state(struct drm_device *dev) | |||
8317 | pll->active, pll->refcount); | 8314 | pll->active, pll->refcount); |
8318 | WARN(pll->active && !pll->on, | 8315 | WARN(pll->active && !pll->on, |
8319 | "pll in active use but not on in sw tracking\n"); | 8316 | "pll in active use but not on in sw tracking\n"); |
8317 | WARN(pll->on && !pll->active, | ||
8318 | "pll in on but not on in use in sw tracking\n"); | ||
8320 | WARN(pll->on != active, | 8319 | WARN(pll->on != active, |
8321 | "pll on state mismatch (expected %i, found %i)\n", | 8320 | "pll on state mismatch (expected %i, found %i)\n", |
8322 | pll->on, active); | 8321 | pll->on, active); |
@@ -8541,15 +8540,20 @@ static void intel_set_config_restore_state(struct drm_device *dev, | |||
8541 | } | 8540 | } |
8542 | 8541 | ||
8543 | static bool | 8542 | static bool |
8544 | is_crtc_connector_off(struct drm_crtc *crtc, struct drm_connector *connectors, | 8543 | is_crtc_connector_off(struct drm_mode_set *set) |
8545 | int num_connectors) | ||
8546 | { | 8544 | { |
8547 | int i; | 8545 | int i; |
8548 | 8546 | ||
8549 | for (i = 0; i < num_connectors; i++) | 8547 | if (set->num_connectors == 0) |
8550 | if (connectors[i].encoder && | 8548 | return false; |
8551 | connectors[i].encoder->crtc == crtc && | 8549 | |
8552 | connectors[i].dpms != DRM_MODE_DPMS_ON) | 8550 | if (WARN_ON(set->connectors == NULL)) |
8551 | return false; | ||
8552 | |||
8553 | for (i = 0; i < set->num_connectors; i++) | ||
8554 | if (set->connectors[i]->encoder && | ||
8555 | set->connectors[i]->encoder->crtc == set->crtc && | ||
8556 | set->connectors[i]->dpms != DRM_MODE_DPMS_ON) | ||
8553 | return true; | 8557 | return true; |
8554 | 8558 | ||
8555 | return false; | 8559 | return false; |
@@ -8562,10 +8566,8 @@ intel_set_config_compute_mode_changes(struct drm_mode_set *set, | |||
8562 | 8566 | ||
8563 | /* We should be able to check here if the fb has the same properties | 8567 | /* We should be able to check here if the fb has the same properties |
8564 | * and then just flip_or_move it */ | 8568 | * and then just flip_or_move it */ |
8565 | if (set->connectors != NULL && | 8569 | if (is_crtc_connector_off(set)) { |
8566 | is_crtc_connector_off(set->crtc, *set->connectors, | 8570 | config->mode_changed = true; |
8567 | set->num_connectors)) { | ||
8568 | config->mode_changed = true; | ||
8569 | } else if (set->crtc->fb != set->fb) { | 8571 | } else if (set->crtc->fb != set->fb) { |
8570 | /* If we have no fb then treat it as a full mode set */ | 8572 | /* If we have no fb then treat it as a full mode set */ |
8571 | if (set->crtc->fb == NULL) { | 8573 | if (set->crtc->fb == NULL) { |
@@ -9398,6 +9400,17 @@ static void quirk_invert_brightness(struct drm_device *dev) | |||
9398 | DRM_INFO("applying inverted panel brightness quirk\n"); | 9400 | DRM_INFO("applying inverted panel brightness quirk\n"); |
9399 | } | 9401 | } |
9400 | 9402 | ||
9403 | /* | ||
9404 | * Some machines (Dell XPS13) suffer broken backlight controls if | ||
9405 | * BLM_PCH_PWM_ENABLE is set. | ||
9406 | */ | ||
9407 | static void quirk_no_pcm_pwm_enable(struct drm_device *dev) | ||
9408 | { | ||
9409 | struct drm_i915_private *dev_priv = dev->dev_private; | ||
9410 | dev_priv->quirks |= QUIRK_NO_PCH_PWM_ENABLE; | ||
9411 | DRM_INFO("applying no-PCH_PWM_ENABLE quirk\n"); | ||
9412 | } | ||
9413 | |||
9401 | struct intel_quirk { | 9414 | struct intel_quirk { |
9402 | int device; | 9415 | int device; |
9403 | int subsystem_vendor; | 9416 | int subsystem_vendor; |
@@ -9467,6 +9480,11 @@ static struct intel_quirk intel_quirks[] = { | |||
9467 | 9480 | ||
9468 | /* Acer Aspire 4736Z */ | 9481 | /* Acer Aspire 4736Z */ |
9469 | { 0x2a42, 0x1025, 0x0260, quirk_invert_brightness }, | 9482 | { 0x2a42, 0x1025, 0x0260, quirk_invert_brightness }, |
9483 | |||
9484 | /* Dell XPS13 HD Sandy Bridge */ | ||
9485 | { 0x0116, 0x1028, 0x052e, quirk_no_pcm_pwm_enable }, | ||
9486 | /* Dell XPS13 HD and XPS13 FHD Ivy Bridge */ | ||
9487 | { 0x0166, 0x1028, 0x058b, quirk_no_pcm_pwm_enable }, | ||
9470 | }; | 9488 | }; |
9471 | 9489 | ||
9472 | static void intel_init_quirks(struct drm_device *dev) | 9490 | static void intel_init_quirks(struct drm_device *dev) |
@@ -9817,8 +9835,8 @@ static void intel_modeset_readout_hw_state(struct drm_device *dev) | |||
9817 | } | 9835 | } |
9818 | pll->refcount = pll->active; | 9836 | pll->refcount = pll->active; |
9819 | 9837 | ||
9820 | DRM_DEBUG_KMS("%s hw state readout: refcount %i\n", | 9838 | DRM_DEBUG_KMS("%s hw state readout: refcount %i, on %i\n", |
9821 | pll->name, pll->refcount); | 9839 | pll->name, pll->refcount, pll->on); |
9822 | } | 9840 | } |
9823 | 9841 | ||
9824 | list_for_each_entry(encoder, &dev->mode_config.encoder_list, | 9842 | list_for_each_entry(encoder, &dev->mode_config.encoder_list, |
@@ -9869,6 +9887,7 @@ void intel_modeset_setup_hw_state(struct drm_device *dev, | |||
9869 | struct drm_plane *plane; | 9887 | struct drm_plane *plane; |
9870 | struct intel_crtc *crtc; | 9888 | struct intel_crtc *crtc; |
9871 | struct intel_encoder *encoder; | 9889 | struct intel_encoder *encoder; |
9890 | int i; | ||
9872 | 9891 | ||
9873 | intel_modeset_readout_hw_state(dev); | 9892 | intel_modeset_readout_hw_state(dev); |
9874 | 9893 | ||
@@ -9884,6 +9903,18 @@ void intel_modeset_setup_hw_state(struct drm_device *dev, | |||
9884 | intel_dump_pipe_config(crtc, &crtc->config, "[setup_hw_state]"); | 9903 | intel_dump_pipe_config(crtc, &crtc->config, "[setup_hw_state]"); |
9885 | } | 9904 | } |
9886 | 9905 | ||
9906 | for (i = 0; i < dev_priv->num_shared_dpll; i++) { | ||
9907 | struct intel_shared_dpll *pll = &dev_priv->shared_dplls[i]; | ||
9908 | |||
9909 | if (!pll->on || pll->active) | ||
9910 | continue; | ||
9911 | |||
9912 | DRM_DEBUG_KMS("%s enabled but not in use, disabling\n", pll->name); | ||
9913 | |||
9914 | pll->disable(dev_priv, pll); | ||
9915 | pll->on = false; | ||
9916 | } | ||
9917 | |||
9887 | if (force_restore) { | 9918 | if (force_restore) { |
9888 | /* | 9919 | /* |
9889 | * We need to use raw interfaces for restoring state to avoid | 9920 | * We need to use raw interfaces for restoring state to avoid |
diff --git a/drivers/gpu/drm/i915/intel_drv.h b/drivers/gpu/drm/i915/intel_drv.h index c8c9b6f48230..b7d6e09456ce 100644 --- a/drivers/gpu/drm/i915/intel_drv.h +++ b/drivers/gpu/drm/i915/intel_drv.h | |||
@@ -504,7 +504,7 @@ struct intel_dp { | |||
504 | struct intel_digital_port { | 504 | struct intel_digital_port { |
505 | struct intel_encoder base; | 505 | struct intel_encoder base; |
506 | enum port port; | 506 | enum port port; |
507 | u32 port_reversal; | 507 | u32 saved_port_bits; |
508 | struct intel_dp dp; | 508 | struct intel_dp dp; |
509 | struct intel_hdmi hdmi; | 509 | struct intel_hdmi hdmi; |
510 | }; | 510 | }; |
diff --git a/drivers/gpu/drm/i915/intel_hdmi.c b/drivers/gpu/drm/i915/intel_hdmi.c index 98df2a0c85bd..2fd3fd5b943e 100644 --- a/drivers/gpu/drm/i915/intel_hdmi.c +++ b/drivers/gpu/drm/i915/intel_hdmi.c | |||
@@ -785,10 +785,22 @@ static void intel_disable_hdmi(struct intel_encoder *encoder) | |||
785 | } | 785 | } |
786 | } | 786 | } |
787 | 787 | ||
788 | static int hdmi_portclock_limit(struct intel_hdmi *hdmi) | ||
789 | { | ||
790 | struct drm_device *dev = intel_hdmi_to_dev(hdmi); | ||
791 | |||
792 | if (IS_G4X(dev)) | ||
793 | return 165000; | ||
794 | else if (IS_HASWELL(dev)) | ||
795 | return 300000; | ||
796 | else | ||
797 | return 225000; | ||
798 | } | ||
799 | |||
788 | static int intel_hdmi_mode_valid(struct drm_connector *connector, | 800 | static int intel_hdmi_mode_valid(struct drm_connector *connector, |
789 | struct drm_display_mode *mode) | 801 | struct drm_display_mode *mode) |
790 | { | 802 | { |
791 | if (mode->clock > 165000) | 803 | if (mode->clock > hdmi_portclock_limit(intel_attached_hdmi(connector))) |
792 | return MODE_CLOCK_HIGH; | 804 | return MODE_CLOCK_HIGH; |
793 | if (mode->clock < 20000) | 805 | if (mode->clock < 20000) |
794 | return MODE_CLOCK_LOW; | 806 | return MODE_CLOCK_LOW; |
@@ -806,6 +818,7 @@ bool intel_hdmi_compute_config(struct intel_encoder *encoder, | |||
806 | struct drm_device *dev = encoder->base.dev; | 818 | struct drm_device *dev = encoder->base.dev; |
807 | struct drm_display_mode *adjusted_mode = &pipe_config->adjusted_mode; | 819 | struct drm_display_mode *adjusted_mode = &pipe_config->adjusted_mode; |
808 | int clock_12bpc = pipe_config->requested_mode.clock * 3 / 2; | 820 | int clock_12bpc = pipe_config->requested_mode.clock * 3 / 2; |
821 | int portclock_limit = hdmi_portclock_limit(intel_hdmi); | ||
809 | int desired_bpp; | 822 | int desired_bpp; |
810 | 823 | ||
811 | if (intel_hdmi->color_range_auto) { | 824 | if (intel_hdmi->color_range_auto) { |
@@ -829,7 +842,7 @@ bool intel_hdmi_compute_config(struct intel_encoder *encoder, | |||
829 | * outputs. We also need to check that the higher clock still fits | 842 | * outputs. We also need to check that the higher clock still fits |
830 | * within limits. | 843 | * within limits. |
831 | */ | 844 | */ |
832 | if (pipe_config->pipe_bpp > 8*3 && clock_12bpc <= 225000 | 845 | if (pipe_config->pipe_bpp > 8*3 && clock_12bpc <= portclock_limit |
833 | && HAS_PCH_SPLIT(dev)) { | 846 | && HAS_PCH_SPLIT(dev)) { |
834 | DRM_DEBUG_KMS("picking bpc to 12 for HDMI output\n"); | 847 | DRM_DEBUG_KMS("picking bpc to 12 for HDMI output\n"); |
835 | desired_bpp = 12*3; | 848 | desired_bpp = 12*3; |
@@ -846,7 +859,7 @@ bool intel_hdmi_compute_config(struct intel_encoder *encoder, | |||
846 | pipe_config->pipe_bpp = desired_bpp; | 859 | pipe_config->pipe_bpp = desired_bpp; |
847 | } | 860 | } |
848 | 861 | ||
849 | if (adjusted_mode->clock > 225000) { | 862 | if (adjusted_mode->clock > portclock_limit) { |
850 | DRM_DEBUG_KMS("too high HDMI clock, rejecting mode\n"); | 863 | DRM_DEBUG_KMS("too high HDMI clock, rejecting mode\n"); |
851 | return false; | 864 | return false; |
852 | } | 865 | } |
diff --git a/drivers/gpu/drm/i915/intel_lvds.c b/drivers/gpu/drm/i915/intel_lvds.c index 021e8daa022d..61348eae2f04 100644 --- a/drivers/gpu/drm/i915/intel_lvds.c +++ b/drivers/gpu/drm/i915/intel_lvds.c | |||
@@ -109,6 +109,13 @@ static void intel_lvds_get_config(struct intel_encoder *encoder, | |||
109 | flags |= DRM_MODE_FLAG_PVSYNC; | 109 | flags |= DRM_MODE_FLAG_PVSYNC; |
110 | 110 | ||
111 | pipe_config->adjusted_mode.flags |= flags; | 111 | pipe_config->adjusted_mode.flags |= flags; |
112 | |||
113 | /* gen2/3 store dither state in pfit control, needs to match */ | ||
114 | if (INTEL_INFO(dev)->gen < 4) { | ||
115 | tmp = I915_READ(PFIT_CONTROL); | ||
116 | |||
117 | pipe_config->gmch_pfit.control |= tmp & PANEL_8TO6_DITHER_ENABLE; | ||
118 | } | ||
112 | } | 119 | } |
113 | 120 | ||
114 | /* The LVDS pin pair needs to be on before the DPLLs are enabled. | 121 | /* The LVDS pin pair needs to be on before the DPLLs are enabled. |
@@ -290,14 +297,11 @@ static bool intel_lvds_compute_config(struct intel_encoder *intel_encoder, | |||
290 | 297 | ||
291 | intel_pch_panel_fitting(intel_crtc, pipe_config, | 298 | intel_pch_panel_fitting(intel_crtc, pipe_config, |
292 | intel_connector->panel.fitting_mode); | 299 | intel_connector->panel.fitting_mode); |
293 | return true; | ||
294 | } else { | 300 | } else { |
295 | intel_gmch_panel_fitting(intel_crtc, pipe_config, | 301 | intel_gmch_panel_fitting(intel_crtc, pipe_config, |
296 | intel_connector->panel.fitting_mode); | 302 | intel_connector->panel.fitting_mode); |
297 | } | ||
298 | 303 | ||
299 | drm_mode_set_crtcinfo(adjusted_mode, 0); | 304 | } |
300 | pipe_config->timings_set = true; | ||
301 | 305 | ||
302 | /* | 306 | /* |
303 | * XXX: It would be nice to support lower refresh rates on the | 307 | * XXX: It would be nice to support lower refresh rates on the |
diff --git a/drivers/gpu/drm/i915/intel_panel.c b/drivers/gpu/drm/i915/intel_panel.c index 80bea1d3209f..67e2c1f1c9a8 100644 --- a/drivers/gpu/drm/i915/intel_panel.c +++ b/drivers/gpu/drm/i915/intel_panel.c | |||
@@ -194,6 +194,9 @@ void intel_gmch_panel_fitting(struct intel_crtc *intel_crtc, | |||
194 | adjusted_mode->vdisplay == mode->vdisplay) | 194 | adjusted_mode->vdisplay == mode->vdisplay) |
195 | goto out; | 195 | goto out; |
196 | 196 | ||
197 | drm_mode_set_crtcinfo(adjusted_mode, 0); | ||
198 | pipe_config->timings_set = true; | ||
199 | |||
197 | switch (fitting_mode) { | 200 | switch (fitting_mode) { |
198 | case DRM_MODE_SCALE_CENTER: | 201 | case DRM_MODE_SCALE_CENTER: |
199 | /* | 202 | /* |
@@ -580,7 +583,8 @@ void intel_panel_enable_backlight(struct drm_device *dev, | |||
580 | POSTING_READ(reg); | 583 | POSTING_READ(reg); |
581 | I915_WRITE(reg, tmp | BLM_PWM_ENABLE); | 584 | I915_WRITE(reg, tmp | BLM_PWM_ENABLE); |
582 | 585 | ||
583 | if (HAS_PCH_SPLIT(dev)) { | 586 | if (HAS_PCH_SPLIT(dev) && |
587 | !(dev_priv->quirks & QUIRK_NO_PCH_PWM_ENABLE)) { | ||
584 | tmp = I915_READ(BLC_PWM_PCH_CTL1); | 588 | tmp = I915_READ(BLC_PWM_PCH_CTL1); |
585 | tmp |= BLM_PCH_PWM_ENABLE; | 589 | tmp |= BLM_PCH_PWM_ENABLE; |
586 | tmp &= ~BLM_PCH_OVERRIDE_ENABLE; | 590 | tmp &= ~BLM_PCH_OVERRIDE_ENABLE; |
diff --git a/drivers/gpu/drm/i915/intel_pm.c b/drivers/gpu/drm/i915/intel_pm.c index d10e6735771f..51a2a60f5bfc 100644 --- a/drivers/gpu/drm/i915/intel_pm.c +++ b/drivers/gpu/drm/i915/intel_pm.c | |||
@@ -5476,7 +5476,7 @@ static void vlv_force_wake_put(struct drm_i915_private *dev_priv) | |||
5476 | gen6_gt_check_fifodbg(dev_priv); | 5476 | gen6_gt_check_fifodbg(dev_priv); |
5477 | } | 5477 | } |
5478 | 5478 | ||
5479 | void intel_gt_reset(struct drm_device *dev) | 5479 | void intel_gt_sanitize(struct drm_device *dev) |
5480 | { | 5480 | { |
5481 | struct drm_i915_private *dev_priv = dev->dev_private; | 5481 | struct drm_i915_private *dev_priv = dev->dev_private; |
5482 | 5482 | ||
@@ -5487,16 +5487,16 @@ void intel_gt_reset(struct drm_device *dev) | |||
5487 | if (IS_IVYBRIDGE(dev) || IS_HASWELL(dev)) | 5487 | if (IS_IVYBRIDGE(dev) || IS_HASWELL(dev)) |
5488 | __gen6_gt_force_wake_mt_reset(dev_priv); | 5488 | __gen6_gt_force_wake_mt_reset(dev_priv); |
5489 | } | 5489 | } |
5490 | |||
5491 | /* BIOS often leaves RC6 enabled, but disable it for hw init */ | ||
5492 | if (INTEL_INFO(dev)->gen >= 6) | ||
5493 | intel_disable_gt_powersave(dev); | ||
5490 | } | 5494 | } |
5491 | 5495 | ||
5492 | void intel_gt_init(struct drm_device *dev) | 5496 | void intel_gt_init(struct drm_device *dev) |
5493 | { | 5497 | { |
5494 | struct drm_i915_private *dev_priv = dev->dev_private; | 5498 | struct drm_i915_private *dev_priv = dev->dev_private; |
5495 | 5499 | ||
5496 | spin_lock_init(&dev_priv->gt_lock); | ||
5497 | |||
5498 | intel_gt_reset(dev); | ||
5499 | |||
5500 | if (IS_VALLEYVIEW(dev)) { | 5500 | if (IS_VALLEYVIEW(dev)) { |
5501 | dev_priv->gt.force_wake_get = vlv_force_wake_get; | 5501 | dev_priv->gt.force_wake_get = vlv_force_wake_get; |
5502 | dev_priv->gt.force_wake_put = vlv_force_wake_put; | 5502 | dev_priv->gt.force_wake_put = vlv_force_wake_put; |
diff --git a/drivers/gpu/drm/nouveau/core/engine/bsp/nvc0.c b/drivers/gpu/drm/nouveau/core/engine/bsp/nvc0.c index 262c9f5f5f60..ce860de43e61 100644 --- a/drivers/gpu/drm/nouveau/core/engine/bsp/nvc0.c +++ b/drivers/gpu/drm/nouveau/core/engine/bsp/nvc0.c | |||
@@ -90,6 +90,7 @@ nvc0_bsp_ctor(struct nouveau_object *parent, struct nouveau_object *engine, | |||
90 | return ret; | 90 | return ret; |
91 | 91 | ||
92 | nv_subdev(priv)->unit = 0x00008000; | 92 | nv_subdev(priv)->unit = 0x00008000; |
93 | nv_subdev(priv)->intr = nouveau_falcon_intr; | ||
93 | nv_engine(priv)->cclass = &nvc0_bsp_cclass; | 94 | nv_engine(priv)->cclass = &nvc0_bsp_cclass; |
94 | nv_engine(priv)->sclass = nvc0_bsp_sclass; | 95 | nv_engine(priv)->sclass = nvc0_bsp_sclass; |
95 | return 0; | 96 | return 0; |
diff --git a/drivers/gpu/drm/nouveau/core/engine/bsp/nve0.c b/drivers/gpu/drm/nouveau/core/engine/bsp/nve0.c index c46882c83982..ba6aeca0285e 100644 --- a/drivers/gpu/drm/nouveau/core/engine/bsp/nve0.c +++ b/drivers/gpu/drm/nouveau/core/engine/bsp/nve0.c | |||
@@ -90,6 +90,7 @@ nve0_bsp_ctor(struct nouveau_object *parent, struct nouveau_object *engine, | |||
90 | return ret; | 90 | return ret; |
91 | 91 | ||
92 | nv_subdev(priv)->unit = 0x00008000; | 92 | nv_subdev(priv)->unit = 0x00008000; |
93 | nv_subdev(priv)->intr = nouveau_falcon_intr; | ||
93 | nv_engine(priv)->cclass = &nve0_bsp_cclass; | 94 | nv_engine(priv)->cclass = &nve0_bsp_cclass; |
94 | nv_engine(priv)->sclass = nve0_bsp_sclass; | 95 | nv_engine(priv)->sclass = nve0_bsp_sclass; |
95 | return 0; | 96 | return 0; |
diff --git a/drivers/gpu/drm/nouveau/core/engine/falcon.c b/drivers/gpu/drm/nouveau/core/engine/falcon.c index 3c7a31f7590e..e03fc8e4dc1d 100644 --- a/drivers/gpu/drm/nouveau/core/engine/falcon.c +++ b/drivers/gpu/drm/nouveau/core/engine/falcon.c | |||
@@ -23,6 +23,25 @@ | |||
23 | #include <engine/falcon.h> | 23 | #include <engine/falcon.h> |
24 | #include <subdev/timer.h> | 24 | #include <subdev/timer.h> |
25 | 25 | ||
26 | void | ||
27 | nouveau_falcon_intr(struct nouveau_subdev *subdev) | ||
28 | { | ||
29 | struct nouveau_falcon *falcon = (void *)subdev; | ||
30 | u32 dispatch = nv_ro32(falcon, 0x01c); | ||
31 | u32 intr = nv_ro32(falcon, 0x008) & dispatch & ~(dispatch >> 16); | ||
32 | |||
33 | if (intr & 0x00000010) { | ||
34 | nv_debug(falcon, "ucode halted\n"); | ||
35 | nv_wo32(falcon, 0x004, 0x00000010); | ||
36 | intr &= ~0x00000010; | ||
37 | } | ||
38 | |||
39 | if (intr) { | ||
40 | nv_error(falcon, "unhandled intr 0x%08x\n", intr); | ||
41 | nv_wo32(falcon, 0x004, intr); | ||
42 | } | ||
43 | } | ||
44 | |||
26 | u32 | 45 | u32 |
27 | _nouveau_falcon_rd32(struct nouveau_object *object, u64 addr) | 46 | _nouveau_falcon_rd32(struct nouveau_object *object, u64 addr) |
28 | { | 47 | { |
diff --git a/drivers/gpu/drm/nouveau/core/engine/ppp/nvc0.c b/drivers/gpu/drm/nouveau/core/engine/ppp/nvc0.c index 98072c1ff360..73719aaa62d6 100644 --- a/drivers/gpu/drm/nouveau/core/engine/ppp/nvc0.c +++ b/drivers/gpu/drm/nouveau/core/engine/ppp/nvc0.c | |||
@@ -90,6 +90,7 @@ nvc0_ppp_ctor(struct nouveau_object *parent, struct nouveau_object *engine, | |||
90 | return ret; | 90 | return ret; |
91 | 91 | ||
92 | nv_subdev(priv)->unit = 0x00000002; | 92 | nv_subdev(priv)->unit = 0x00000002; |
93 | nv_subdev(priv)->intr = nouveau_falcon_intr; | ||
93 | nv_engine(priv)->cclass = &nvc0_ppp_cclass; | 94 | nv_engine(priv)->cclass = &nvc0_ppp_cclass; |
94 | nv_engine(priv)->sclass = nvc0_ppp_sclass; | 95 | nv_engine(priv)->sclass = nvc0_ppp_sclass; |
95 | return 0; | 96 | return 0; |
diff --git a/drivers/gpu/drm/nouveau/core/engine/vp/nvc0.c b/drivers/gpu/drm/nouveau/core/engine/vp/nvc0.c index 1879229b60eb..ac1f62aace72 100644 --- a/drivers/gpu/drm/nouveau/core/engine/vp/nvc0.c +++ b/drivers/gpu/drm/nouveau/core/engine/vp/nvc0.c | |||
@@ -90,6 +90,7 @@ nvc0_vp_ctor(struct nouveau_object *parent, struct nouveau_object *engine, | |||
90 | return ret; | 90 | return ret; |
91 | 91 | ||
92 | nv_subdev(priv)->unit = 0x00020000; | 92 | nv_subdev(priv)->unit = 0x00020000; |
93 | nv_subdev(priv)->intr = nouveau_falcon_intr; | ||
93 | nv_engine(priv)->cclass = &nvc0_vp_cclass; | 94 | nv_engine(priv)->cclass = &nvc0_vp_cclass; |
94 | nv_engine(priv)->sclass = nvc0_vp_sclass; | 95 | nv_engine(priv)->sclass = nvc0_vp_sclass; |
95 | return 0; | 96 | return 0; |
diff --git a/drivers/gpu/drm/nouveau/core/engine/vp/nve0.c b/drivers/gpu/drm/nouveau/core/engine/vp/nve0.c index d28ecbf7bc49..d4c3108479c9 100644 --- a/drivers/gpu/drm/nouveau/core/engine/vp/nve0.c +++ b/drivers/gpu/drm/nouveau/core/engine/vp/nve0.c | |||
@@ -90,6 +90,7 @@ nve0_vp_ctor(struct nouveau_object *parent, struct nouveau_object *engine, | |||
90 | return ret; | 90 | return ret; |
91 | 91 | ||
92 | nv_subdev(priv)->unit = 0x00020000; | 92 | nv_subdev(priv)->unit = 0x00020000; |
93 | nv_subdev(priv)->intr = nouveau_falcon_intr; | ||
93 | nv_engine(priv)->cclass = &nve0_vp_cclass; | 94 | nv_engine(priv)->cclass = &nve0_vp_cclass; |
94 | nv_engine(priv)->sclass = nve0_vp_sclass; | 95 | nv_engine(priv)->sclass = nve0_vp_sclass; |
95 | return 0; | 96 | return 0; |
diff --git a/drivers/gpu/drm/nouveau/core/include/engine/falcon.h b/drivers/gpu/drm/nouveau/core/include/engine/falcon.h index 1edec386ab36..181aa7da524d 100644 --- a/drivers/gpu/drm/nouveau/core/include/engine/falcon.h +++ b/drivers/gpu/drm/nouveau/core/include/engine/falcon.h | |||
@@ -72,6 +72,8 @@ int nouveau_falcon_create_(struct nouveau_object *, struct nouveau_object *, | |||
72 | struct nouveau_oclass *, u32, bool, const char *, | 72 | struct nouveau_oclass *, u32, bool, const char *, |
73 | const char *, int, void **); | 73 | const char *, int, void **); |
74 | 74 | ||
75 | void nouveau_falcon_intr(struct nouveau_subdev *subdev); | ||
76 | |||
75 | #define _nouveau_falcon_dtor _nouveau_engine_dtor | 77 | #define _nouveau_falcon_dtor _nouveau_engine_dtor |
76 | int _nouveau_falcon_init(struct nouveau_object *); | 78 | int _nouveau_falcon_init(struct nouveau_object *); |
77 | int _nouveau_falcon_fini(struct nouveau_object *, bool); | 79 | int _nouveau_falcon_fini(struct nouveau_object *, bool); |
diff --git a/drivers/gpu/drm/nouveau/nouveau_bo.c b/drivers/gpu/drm/nouveau/nouveau_bo.c index 4b1afb131380..4e7ee5f4155c 100644 --- a/drivers/gpu/drm/nouveau/nouveau_bo.c +++ b/drivers/gpu/drm/nouveau/nouveau_bo.c | |||
@@ -148,6 +148,7 @@ nouveau_bo_del_ttm(struct ttm_buffer_object *bo) | |||
148 | 148 | ||
149 | if (unlikely(nvbo->gem)) | 149 | if (unlikely(nvbo->gem)) |
150 | DRM_ERROR("bo %p still attached to GEM object\n", bo); | 150 | DRM_ERROR("bo %p still attached to GEM object\n", bo); |
151 | WARN_ON(nvbo->pin_refcnt > 0); | ||
151 | nv10_bo_put_tile_region(dev, nvbo->tile, NULL); | 152 | nv10_bo_put_tile_region(dev, nvbo->tile, NULL); |
152 | kfree(nvbo); | 153 | kfree(nvbo); |
153 | } | 154 | } |
@@ -197,6 +198,12 @@ nouveau_bo_new(struct drm_device *dev, int size, int align, | |||
197 | size_t acc_size; | 198 | size_t acc_size; |
198 | int ret; | 199 | int ret; |
199 | int type = ttm_bo_type_device; | 200 | int type = ttm_bo_type_device; |
201 | int max_size = INT_MAX & ~((1 << drm->client.base.vm->vmm->lpg_shift) - 1); | ||
202 | |||
203 | if (size <= 0 || size > max_size) { | ||
204 | nv_warn(drm, "skipped size %x\n", (u32)size); | ||
205 | return -EINVAL; | ||
206 | } | ||
200 | 207 | ||
201 | if (sg) | 208 | if (sg) |
202 | type = ttm_bo_type_sg; | 209 | type = ttm_bo_type_sg; |
@@ -340,13 +347,15 @@ nouveau_bo_unpin(struct nouveau_bo *nvbo) | |||
340 | { | 347 | { |
341 | struct nouveau_drm *drm = nouveau_bdev(nvbo->bo.bdev); | 348 | struct nouveau_drm *drm = nouveau_bdev(nvbo->bo.bdev); |
342 | struct ttm_buffer_object *bo = &nvbo->bo; | 349 | struct ttm_buffer_object *bo = &nvbo->bo; |
343 | int ret; | 350 | int ret, ref; |
344 | 351 | ||
345 | ret = ttm_bo_reserve(bo, false, false, false, 0); | 352 | ret = ttm_bo_reserve(bo, false, false, false, 0); |
346 | if (ret) | 353 | if (ret) |
347 | return ret; | 354 | return ret; |
348 | 355 | ||
349 | if (--nvbo->pin_refcnt) | 356 | ref = --nvbo->pin_refcnt; |
357 | WARN_ON_ONCE(ref < 0); | ||
358 | if (ref) | ||
350 | goto out; | 359 | goto out; |
351 | 360 | ||
352 | nouveau_bo_placement_set(nvbo, bo->mem.placement, 0); | 361 | nouveau_bo_placement_set(nvbo, bo->mem.placement, 0); |
@@ -578,7 +587,7 @@ nve0_bo_move_init(struct nouveau_channel *chan, u32 handle) | |||
578 | int ret = RING_SPACE(chan, 2); | 587 | int ret = RING_SPACE(chan, 2); |
579 | if (ret == 0) { | 588 | if (ret == 0) { |
580 | BEGIN_NVC0(chan, NvSubCopy, 0x0000, 1); | 589 | BEGIN_NVC0(chan, NvSubCopy, 0x0000, 1); |
581 | OUT_RING (chan, handle); | 590 | OUT_RING (chan, handle & 0x0000ffff); |
582 | FIRE_RING (chan); | 591 | FIRE_RING (chan); |
583 | } | 592 | } |
584 | return ret; | 593 | return ret; |
@@ -973,7 +982,7 @@ nouveau_bo_move_m2mf(struct ttm_buffer_object *bo, int evict, bool intr, | |||
973 | struct ttm_mem_reg *old_mem = &bo->mem; | 982 | struct ttm_mem_reg *old_mem = &bo->mem; |
974 | int ret; | 983 | int ret; |
975 | 984 | ||
976 | mutex_lock(&chan->cli->mutex); | 985 | mutex_lock_nested(&chan->cli->mutex, SINGLE_DEPTH_NESTING); |
977 | 986 | ||
978 | /* create temporary vmas for the transfer and attach them to the | 987 | /* create temporary vmas for the transfer and attach them to the |
979 | * old nouveau_mem node, these will get cleaned up after ttm has | 988 | * old nouveau_mem node, these will get cleaned up after ttm has |
@@ -1014,7 +1023,7 @@ nouveau_bo_move_init(struct nouveau_drm *drm) | |||
1014 | struct ttm_mem_reg *, struct ttm_mem_reg *); | 1023 | struct ttm_mem_reg *, struct ttm_mem_reg *); |
1015 | int (*init)(struct nouveau_channel *, u32 handle); | 1024 | int (*init)(struct nouveau_channel *, u32 handle); |
1016 | } _methods[] = { | 1025 | } _methods[] = { |
1017 | { "COPY", 0, 0xa0b5, nve0_bo_move_copy, nve0_bo_move_init }, | 1026 | { "COPY", 4, 0xa0b5, nve0_bo_move_copy, nve0_bo_move_init }, |
1018 | { "GRCE", 0, 0xa0b5, nve0_bo_move_copy, nvc0_bo_move_init }, | 1027 | { "GRCE", 0, 0xa0b5, nve0_bo_move_copy, nvc0_bo_move_init }, |
1019 | { "COPY1", 5, 0x90b8, nvc0_bo_move_copy, nvc0_bo_move_init }, | 1028 | { "COPY1", 5, 0x90b8, nvc0_bo_move_copy, nvc0_bo_move_init }, |
1020 | { "COPY0", 4, 0x90b5, nvc0_bo_move_copy, nvc0_bo_move_init }, | 1029 | { "COPY0", 4, 0x90b5, nvc0_bo_move_copy, nvc0_bo_move_init }, |
@@ -1034,7 +1043,7 @@ nouveau_bo_move_init(struct nouveau_drm *drm) | |||
1034 | struct nouveau_channel *chan; | 1043 | struct nouveau_channel *chan; |
1035 | u32 handle = (mthd->engine << 16) | mthd->oclass; | 1044 | u32 handle = (mthd->engine << 16) | mthd->oclass; |
1036 | 1045 | ||
1037 | if (mthd->init == nve0_bo_move_init) | 1046 | if (mthd->engine) |
1038 | chan = drm->cechan; | 1047 | chan = drm->cechan; |
1039 | else | 1048 | else |
1040 | chan = drm->channel; | 1049 | chan = drm->channel; |
diff --git a/drivers/gpu/drm/nouveau/nouveau_display.c b/drivers/gpu/drm/nouveau/nouveau_display.c index 708b2d1c0037..907d20ef6d4d 100644 --- a/drivers/gpu/drm/nouveau/nouveau_display.c +++ b/drivers/gpu/drm/nouveau/nouveau_display.c | |||
@@ -138,7 +138,7 @@ nouveau_user_framebuffer_create(struct drm_device *dev, | |||
138 | { | 138 | { |
139 | struct nouveau_framebuffer *nouveau_fb; | 139 | struct nouveau_framebuffer *nouveau_fb; |
140 | struct drm_gem_object *gem; | 140 | struct drm_gem_object *gem; |
141 | int ret; | 141 | int ret = -ENOMEM; |
142 | 142 | ||
143 | gem = drm_gem_object_lookup(dev, file_priv, mode_cmd->handles[0]); | 143 | gem = drm_gem_object_lookup(dev, file_priv, mode_cmd->handles[0]); |
144 | if (!gem) | 144 | if (!gem) |
@@ -146,15 +146,19 @@ nouveau_user_framebuffer_create(struct drm_device *dev, | |||
146 | 146 | ||
147 | nouveau_fb = kzalloc(sizeof(struct nouveau_framebuffer), GFP_KERNEL); | 147 | nouveau_fb = kzalloc(sizeof(struct nouveau_framebuffer), GFP_KERNEL); |
148 | if (!nouveau_fb) | 148 | if (!nouveau_fb) |
149 | return ERR_PTR(-ENOMEM); | 149 | goto err_unref; |
150 | 150 | ||
151 | ret = nouveau_framebuffer_init(dev, nouveau_fb, mode_cmd, nouveau_gem_object(gem)); | 151 | ret = nouveau_framebuffer_init(dev, nouveau_fb, mode_cmd, nouveau_gem_object(gem)); |
152 | if (ret) { | 152 | if (ret) |
153 | drm_gem_object_unreference(gem); | 153 | goto err; |
154 | return ERR_PTR(ret); | ||
155 | } | ||
156 | 154 | ||
157 | return &nouveau_fb->base; | 155 | return &nouveau_fb->base; |
156 | |||
157 | err: | ||
158 | kfree(nouveau_fb); | ||
159 | err_unref: | ||
160 | drm_gem_object_unreference(gem); | ||
161 | return ERR_PTR(ret); | ||
158 | } | 162 | } |
159 | 163 | ||
160 | static const struct drm_mode_config_funcs nouveau_mode_config_funcs = { | 164 | static const struct drm_mode_config_funcs nouveau_mode_config_funcs = { |
@@ -524,9 +528,12 @@ nouveau_crtc_page_flip(struct drm_crtc *crtc, struct drm_framebuffer *fb, | |||
524 | struct nouveau_page_flip_state *s; | 528 | struct nouveau_page_flip_state *s; |
525 | struct nouveau_channel *chan = NULL; | 529 | struct nouveau_channel *chan = NULL; |
526 | struct nouveau_fence *fence; | 530 | struct nouveau_fence *fence; |
527 | struct list_head res; | 531 | struct ttm_validate_buffer resv[2] = { |
528 | struct ttm_validate_buffer res_val[2]; | 532 | { .bo = &old_bo->bo }, |
533 | { .bo = &new_bo->bo }, | ||
534 | }; | ||
529 | struct ww_acquire_ctx ticket; | 535 | struct ww_acquire_ctx ticket; |
536 | LIST_HEAD(res); | ||
530 | int ret; | 537 | int ret; |
531 | 538 | ||
532 | if (!drm->channel) | 539 | if (!drm->channel) |
@@ -545,27 +552,19 @@ nouveau_crtc_page_flip(struct drm_crtc *crtc, struct drm_framebuffer *fb, | |||
545 | chan = drm->channel; | 552 | chan = drm->channel; |
546 | spin_unlock(&old_bo->bo.bdev->fence_lock); | 553 | spin_unlock(&old_bo->bo.bdev->fence_lock); |
547 | 554 | ||
548 | mutex_lock(&chan->cli->mutex); | ||
549 | |||
550 | if (new_bo != old_bo) { | 555 | if (new_bo != old_bo) { |
551 | ret = nouveau_bo_pin(new_bo, TTM_PL_FLAG_VRAM); | 556 | ret = nouveau_bo_pin(new_bo, TTM_PL_FLAG_VRAM); |
552 | if (likely(!ret)) { | 557 | if (ret) |
553 | res_val[0].bo = &old_bo->bo; | 558 | goto fail_free; |
554 | res_val[1].bo = &new_bo->bo; | ||
555 | INIT_LIST_HEAD(&res); | ||
556 | list_add_tail(&res_val[0].head, &res); | ||
557 | list_add_tail(&res_val[1].head, &res); | ||
558 | ret = ttm_eu_reserve_buffers(&ticket, &res); | ||
559 | if (ret) | ||
560 | nouveau_bo_unpin(new_bo); | ||
561 | } | ||
562 | } else | ||
563 | ret = ttm_bo_reserve(&new_bo->bo, false, false, false, 0); | ||
564 | 559 | ||
565 | if (ret) { | 560 | list_add(&resv[1].head, &res); |
566 | mutex_unlock(&chan->cli->mutex); | ||
567 | goto fail_free; | ||
568 | } | 561 | } |
562 | list_add(&resv[0].head, &res); | ||
563 | |||
564 | mutex_lock(&chan->cli->mutex); | ||
565 | ret = ttm_eu_reserve_buffers(&ticket, &res); | ||
566 | if (ret) | ||
567 | goto fail_unpin; | ||
569 | 568 | ||
570 | /* Initialize a page flip struct */ | 569 | /* Initialize a page flip struct */ |
571 | *s = (struct nouveau_page_flip_state) | 570 | *s = (struct nouveau_page_flip_state) |
@@ -576,10 +575,8 @@ nouveau_crtc_page_flip(struct drm_crtc *crtc, struct drm_framebuffer *fb, | |||
576 | /* Emit a page flip */ | 575 | /* Emit a page flip */ |
577 | if (nv_device(drm->device)->card_type >= NV_50) { | 576 | if (nv_device(drm->device)->card_type >= NV_50) { |
578 | ret = nv50_display_flip_next(crtc, fb, chan, 0); | 577 | ret = nv50_display_flip_next(crtc, fb, chan, 0); |
579 | if (ret) { | 578 | if (ret) |
580 | mutex_unlock(&chan->cli->mutex); | ||
581 | goto fail_unreserve; | 579 | goto fail_unreserve; |
582 | } | ||
583 | } | 580 | } |
584 | 581 | ||
585 | ret = nouveau_page_flip_emit(chan, old_bo, new_bo, s, &fence); | 582 | ret = nouveau_page_flip_emit(chan, old_bo, new_bo, s, &fence); |
@@ -590,22 +587,18 @@ nouveau_crtc_page_flip(struct drm_crtc *crtc, struct drm_framebuffer *fb, | |||
590 | /* Update the crtc struct and cleanup */ | 587 | /* Update the crtc struct and cleanup */ |
591 | crtc->fb = fb; | 588 | crtc->fb = fb; |
592 | 589 | ||
593 | if (old_bo != new_bo) { | 590 | ttm_eu_fence_buffer_objects(&ticket, &res, fence); |
594 | ttm_eu_fence_buffer_objects(&ticket, &res, fence); | 591 | if (old_bo != new_bo) |
595 | nouveau_bo_unpin(old_bo); | 592 | nouveau_bo_unpin(old_bo); |
596 | } else { | ||
597 | nouveau_bo_fence(new_bo, fence); | ||
598 | ttm_bo_unreserve(&new_bo->bo); | ||
599 | } | ||
600 | nouveau_fence_unref(&fence); | 593 | nouveau_fence_unref(&fence); |
601 | return 0; | 594 | return 0; |
602 | 595 | ||
603 | fail_unreserve: | 596 | fail_unreserve: |
604 | if (old_bo != new_bo) { | 597 | ttm_eu_backoff_reservation(&ticket, &res); |
605 | ttm_eu_backoff_reservation(&ticket, &res); | 598 | fail_unpin: |
599 | mutex_unlock(&chan->cli->mutex); | ||
600 | if (old_bo != new_bo) | ||
606 | nouveau_bo_unpin(new_bo); | 601 | nouveau_bo_unpin(new_bo); |
607 | } else | ||
608 | ttm_bo_unreserve(&new_bo->bo); | ||
609 | fail_free: | 602 | fail_free: |
610 | kfree(s); | 603 | kfree(s); |
611 | return ret; | 604 | return ret; |
diff --git a/drivers/gpu/drm/nouveau/nouveau_drm.c b/drivers/gpu/drm/nouveau/nouveau_drm.c index 218a4b522fe5..61972668fd05 100644 --- a/drivers/gpu/drm/nouveau/nouveau_drm.c +++ b/drivers/gpu/drm/nouveau/nouveau_drm.c | |||
@@ -192,6 +192,18 @@ nouveau_accel_init(struct nouveau_drm *drm) | |||
192 | 192 | ||
193 | arg0 = NVE0_CHANNEL_IND_ENGINE_GR; | 193 | arg0 = NVE0_CHANNEL_IND_ENGINE_GR; |
194 | arg1 = 1; | 194 | arg1 = 1; |
195 | } else | ||
196 | if (device->chipset >= 0xa3 && | ||
197 | device->chipset != 0xaa && | ||
198 | device->chipset != 0xac) { | ||
199 | ret = nouveau_channel_new(drm, &drm->client, NVDRM_DEVICE, | ||
200 | NVDRM_CHAN + 1, NvDmaFB, NvDmaTT, | ||
201 | &drm->cechan); | ||
202 | if (ret) | ||
203 | NV_ERROR(drm, "failed to create ce channel, %d\n", ret); | ||
204 | |||
205 | arg0 = NvDmaFB; | ||
206 | arg1 = NvDmaTT; | ||
195 | } else { | 207 | } else { |
196 | arg0 = NvDmaFB; | 208 | arg0 = NvDmaFB; |
197 | arg1 = NvDmaTT; | 209 | arg1 = NvDmaTT; |
@@ -284,8 +296,6 @@ static int nouveau_drm_probe(struct pci_dev *pdev, | |||
284 | return 0; | 296 | return 0; |
285 | } | 297 | } |
286 | 298 | ||
287 | static struct lock_class_key drm_client_lock_class_key; | ||
288 | |||
289 | static int | 299 | static int |
290 | nouveau_drm_load(struct drm_device *dev, unsigned long flags) | 300 | nouveau_drm_load(struct drm_device *dev, unsigned long flags) |
291 | { | 301 | { |
@@ -297,7 +307,6 @@ nouveau_drm_load(struct drm_device *dev, unsigned long flags) | |||
297 | ret = nouveau_cli_create(pdev, "DRM", sizeof(*drm), (void**)&drm); | 307 | ret = nouveau_cli_create(pdev, "DRM", sizeof(*drm), (void**)&drm); |
298 | if (ret) | 308 | if (ret) |
299 | return ret; | 309 | return ret; |
300 | lockdep_set_class(&drm->client.mutex, &drm_client_lock_class_key); | ||
301 | 310 | ||
302 | dev->dev_private = drm; | 311 | dev->dev_private = drm; |
303 | drm->dev = dev; | 312 | drm->dev = dev; |
diff --git a/drivers/gpu/drm/nouveau/nouveau_fbcon.c b/drivers/gpu/drm/nouveau/nouveau_fbcon.c index 9352010030e9..4c1bc061fae2 100644 --- a/drivers/gpu/drm/nouveau/nouveau_fbcon.c +++ b/drivers/gpu/drm/nouveau/nouveau_fbcon.c | |||
@@ -385,6 +385,7 @@ out_unlock: | |||
385 | mutex_unlock(&dev->struct_mutex); | 385 | mutex_unlock(&dev->struct_mutex); |
386 | if (chan) | 386 | if (chan) |
387 | nouveau_bo_vma_del(nvbo, &fbcon->nouveau_fb.vma); | 387 | nouveau_bo_vma_del(nvbo, &fbcon->nouveau_fb.vma); |
388 | nouveau_bo_unmap(nvbo); | ||
388 | out_unpin: | 389 | out_unpin: |
389 | nouveau_bo_unpin(nvbo); | 390 | nouveau_bo_unpin(nvbo); |
390 | out_unref: | 391 | out_unref: |
diff --git a/drivers/gpu/drm/nouveau/nouveau_fence.c b/drivers/gpu/drm/nouveau/nouveau_fence.c index 1680d9187bab..be3149932c2d 100644 --- a/drivers/gpu/drm/nouveau/nouveau_fence.c +++ b/drivers/gpu/drm/nouveau/nouveau_fence.c | |||
@@ -143,7 +143,7 @@ nouveau_fence_emit(struct nouveau_fence *fence, struct nouveau_channel *chan) | |||
143 | int ret; | 143 | int ret; |
144 | 144 | ||
145 | fence->channel = chan; | 145 | fence->channel = chan; |
146 | fence->timeout = jiffies + (3 * DRM_HZ); | 146 | fence->timeout = jiffies + (15 * DRM_HZ); |
147 | fence->sequence = ++fctx->sequence; | 147 | fence->sequence = ++fctx->sequence; |
148 | 148 | ||
149 | ret = fctx->emit(fence); | 149 | ret = fctx->emit(fence); |
diff --git a/drivers/gpu/drm/nouveau/nouveau_gem.c b/drivers/gpu/drm/nouveau/nouveau_gem.c index e72d09c068a8..830cb7bad922 100644 --- a/drivers/gpu/drm/nouveau/nouveau_gem.c +++ b/drivers/gpu/drm/nouveau/nouveau_gem.c | |||
@@ -50,12 +50,6 @@ nouveau_gem_object_del(struct drm_gem_object *gem) | |||
50 | return; | 50 | return; |
51 | nvbo->gem = NULL; | 51 | nvbo->gem = NULL; |
52 | 52 | ||
53 | /* Lockdep hates you for doing reserve with gem object lock held */ | ||
54 | if (WARN_ON_ONCE(nvbo->pin_refcnt)) { | ||
55 | nvbo->pin_refcnt = 1; | ||
56 | nouveau_bo_unpin(nvbo); | ||
57 | } | ||
58 | |||
59 | if (gem->import_attach) | 53 | if (gem->import_attach) |
60 | drm_prime_gem_destroy(gem, nvbo->bo.sg); | 54 | drm_prime_gem_destroy(gem, nvbo->bo.sg); |
61 | 55 | ||
diff --git a/drivers/gpu/drm/nouveau/nv50_display.c b/drivers/gpu/drm/nouveau/nv50_display.c index 54dc6355b0c2..8b40a36c1b57 100644 --- a/drivers/gpu/drm/nouveau/nv50_display.c +++ b/drivers/gpu/drm/nouveau/nv50_display.c | |||
@@ -355,6 +355,7 @@ struct nv50_oimm { | |||
355 | 355 | ||
356 | struct nv50_head { | 356 | struct nv50_head { |
357 | struct nouveau_crtc base; | 357 | struct nouveau_crtc base; |
358 | struct nouveau_bo *image; | ||
358 | struct nv50_curs curs; | 359 | struct nv50_curs curs; |
359 | struct nv50_sync sync; | 360 | struct nv50_sync sync; |
360 | struct nv50_ovly ovly; | 361 | struct nv50_ovly ovly; |
@@ -517,9 +518,10 @@ nv50_display_flip_next(struct drm_crtc *crtc, struct drm_framebuffer *fb, | |||
517 | { | 518 | { |
518 | struct nouveau_framebuffer *nv_fb = nouveau_framebuffer(fb); | 519 | struct nouveau_framebuffer *nv_fb = nouveau_framebuffer(fb); |
519 | struct nouveau_crtc *nv_crtc = nouveau_crtc(crtc); | 520 | struct nouveau_crtc *nv_crtc = nouveau_crtc(crtc); |
521 | struct nv50_head *head = nv50_head(crtc); | ||
520 | struct nv50_sync *sync = nv50_sync(crtc); | 522 | struct nv50_sync *sync = nv50_sync(crtc); |
521 | int head = nv_crtc->index, ret; | ||
522 | u32 *push; | 523 | u32 *push; |
524 | int ret; | ||
523 | 525 | ||
524 | swap_interval <<= 4; | 526 | swap_interval <<= 4; |
525 | if (swap_interval == 0) | 527 | if (swap_interval == 0) |
@@ -537,7 +539,7 @@ nv50_display_flip_next(struct drm_crtc *crtc, struct drm_framebuffer *fb, | |||
537 | return ret; | 539 | return ret; |
538 | 540 | ||
539 | BEGIN_NV04(chan, 0, NV11_SUBCHAN_DMA_SEMAPHORE, 2); | 541 | BEGIN_NV04(chan, 0, NV11_SUBCHAN_DMA_SEMAPHORE, 2); |
540 | OUT_RING (chan, NvEvoSema0 + head); | 542 | OUT_RING (chan, NvEvoSema0 + nv_crtc->index); |
541 | OUT_RING (chan, sync->addr ^ 0x10); | 543 | OUT_RING (chan, sync->addr ^ 0x10); |
542 | BEGIN_NV04(chan, 0, NV11_SUBCHAN_SEMAPHORE_RELEASE, 1); | 544 | BEGIN_NV04(chan, 0, NV11_SUBCHAN_SEMAPHORE_RELEASE, 1); |
543 | OUT_RING (chan, sync->data + 1); | 545 | OUT_RING (chan, sync->data + 1); |
@@ -546,7 +548,7 @@ nv50_display_flip_next(struct drm_crtc *crtc, struct drm_framebuffer *fb, | |||
546 | OUT_RING (chan, sync->data); | 548 | OUT_RING (chan, sync->data); |
547 | } else | 549 | } else |
548 | if (chan && nv_mclass(chan->object) < NVC0_CHANNEL_IND_CLASS) { | 550 | if (chan && nv_mclass(chan->object) < NVC0_CHANNEL_IND_CLASS) { |
549 | u64 addr = nv84_fence_crtc(chan, head) + sync->addr; | 551 | u64 addr = nv84_fence_crtc(chan, nv_crtc->index) + sync->addr; |
550 | ret = RING_SPACE(chan, 12); | 552 | ret = RING_SPACE(chan, 12); |
551 | if (ret) | 553 | if (ret) |
552 | return ret; | 554 | return ret; |
@@ -565,7 +567,7 @@ nv50_display_flip_next(struct drm_crtc *crtc, struct drm_framebuffer *fb, | |||
565 | OUT_RING (chan, NV84_SUBCHAN_SEMAPHORE_TRIGGER_ACQUIRE_EQUAL); | 567 | OUT_RING (chan, NV84_SUBCHAN_SEMAPHORE_TRIGGER_ACQUIRE_EQUAL); |
566 | } else | 568 | } else |
567 | if (chan) { | 569 | if (chan) { |
568 | u64 addr = nv84_fence_crtc(chan, head) + sync->addr; | 570 | u64 addr = nv84_fence_crtc(chan, nv_crtc->index) + sync->addr; |
569 | ret = RING_SPACE(chan, 10); | 571 | ret = RING_SPACE(chan, 10); |
570 | if (ret) | 572 | if (ret) |
571 | return ret; | 573 | return ret; |
@@ -630,6 +632,8 @@ nv50_display_flip_next(struct drm_crtc *crtc, struct drm_framebuffer *fb, | |||
630 | evo_mthd(push, 0x0080, 1); | 632 | evo_mthd(push, 0x0080, 1); |
631 | evo_data(push, 0x00000000); | 633 | evo_data(push, 0x00000000); |
632 | evo_kick(push, sync); | 634 | evo_kick(push, sync); |
635 | |||
636 | nouveau_bo_ref(nv_fb->nvbo, &head->image); | ||
633 | return 0; | 637 | return 0; |
634 | } | 638 | } |
635 | 639 | ||
@@ -1038,18 +1042,17 @@ static int | |||
1038 | nv50_crtc_swap_fbs(struct drm_crtc *crtc, struct drm_framebuffer *old_fb) | 1042 | nv50_crtc_swap_fbs(struct drm_crtc *crtc, struct drm_framebuffer *old_fb) |
1039 | { | 1043 | { |
1040 | struct nouveau_framebuffer *nvfb = nouveau_framebuffer(crtc->fb); | 1044 | struct nouveau_framebuffer *nvfb = nouveau_framebuffer(crtc->fb); |
1045 | struct nv50_head *head = nv50_head(crtc); | ||
1041 | int ret; | 1046 | int ret; |
1042 | 1047 | ||
1043 | ret = nouveau_bo_pin(nvfb->nvbo, TTM_PL_FLAG_VRAM); | 1048 | ret = nouveau_bo_pin(nvfb->nvbo, TTM_PL_FLAG_VRAM); |
1044 | if (ret) | 1049 | if (ret == 0) { |
1045 | return ret; | 1050 | if (head->image) |
1046 | 1051 | nouveau_bo_unpin(head->image); | |
1047 | if (old_fb) { | 1052 | nouveau_bo_ref(nvfb->nvbo, &head->image); |
1048 | nvfb = nouveau_framebuffer(old_fb); | ||
1049 | nouveau_bo_unpin(nvfb->nvbo); | ||
1050 | } | 1053 | } |
1051 | 1054 | ||
1052 | return 0; | 1055 | return ret; |
1053 | } | 1056 | } |
1054 | 1057 | ||
1055 | static int | 1058 | static int |
@@ -1198,6 +1201,15 @@ nv50_crtc_lut_load(struct drm_crtc *crtc) | |||
1198 | } | 1201 | } |
1199 | } | 1202 | } |
1200 | 1203 | ||
1204 | static void | ||
1205 | nv50_crtc_disable(struct drm_crtc *crtc) | ||
1206 | { | ||
1207 | struct nv50_head *head = nv50_head(crtc); | ||
1208 | if (head->image) | ||
1209 | nouveau_bo_unpin(head->image); | ||
1210 | nouveau_bo_ref(NULL, &head->image); | ||
1211 | } | ||
1212 | |||
1201 | static int | 1213 | static int |
1202 | nv50_crtc_cursor_set(struct drm_crtc *crtc, struct drm_file *file_priv, | 1214 | nv50_crtc_cursor_set(struct drm_crtc *crtc, struct drm_file *file_priv, |
1203 | uint32_t handle, uint32_t width, uint32_t height) | 1215 | uint32_t handle, uint32_t width, uint32_t height) |
@@ -1271,18 +1283,29 @@ nv50_crtc_destroy(struct drm_crtc *crtc) | |||
1271 | struct nouveau_crtc *nv_crtc = nouveau_crtc(crtc); | 1283 | struct nouveau_crtc *nv_crtc = nouveau_crtc(crtc); |
1272 | struct nv50_disp *disp = nv50_disp(crtc->dev); | 1284 | struct nv50_disp *disp = nv50_disp(crtc->dev); |
1273 | struct nv50_head *head = nv50_head(crtc); | 1285 | struct nv50_head *head = nv50_head(crtc); |
1286 | |||
1274 | nv50_dmac_destroy(disp->core, &head->ovly.base); | 1287 | nv50_dmac_destroy(disp->core, &head->ovly.base); |
1275 | nv50_pioc_destroy(disp->core, &head->oimm.base); | 1288 | nv50_pioc_destroy(disp->core, &head->oimm.base); |
1276 | nv50_dmac_destroy(disp->core, &head->sync.base); | 1289 | nv50_dmac_destroy(disp->core, &head->sync.base); |
1277 | nv50_pioc_destroy(disp->core, &head->curs.base); | 1290 | nv50_pioc_destroy(disp->core, &head->curs.base); |
1291 | |||
1292 | /*XXX: this shouldn't be necessary, but the core doesn't call | ||
1293 | * disconnect() during the cleanup paths | ||
1294 | */ | ||
1295 | if (head->image) | ||
1296 | nouveau_bo_unpin(head->image); | ||
1297 | nouveau_bo_ref(NULL, &head->image); | ||
1298 | |||
1278 | nouveau_bo_unmap(nv_crtc->cursor.nvbo); | 1299 | nouveau_bo_unmap(nv_crtc->cursor.nvbo); |
1279 | if (nv_crtc->cursor.nvbo) | 1300 | if (nv_crtc->cursor.nvbo) |
1280 | nouveau_bo_unpin(nv_crtc->cursor.nvbo); | 1301 | nouveau_bo_unpin(nv_crtc->cursor.nvbo); |
1281 | nouveau_bo_ref(NULL, &nv_crtc->cursor.nvbo); | 1302 | nouveau_bo_ref(NULL, &nv_crtc->cursor.nvbo); |
1303 | |||
1282 | nouveau_bo_unmap(nv_crtc->lut.nvbo); | 1304 | nouveau_bo_unmap(nv_crtc->lut.nvbo); |
1283 | if (nv_crtc->lut.nvbo) | 1305 | if (nv_crtc->lut.nvbo) |
1284 | nouveau_bo_unpin(nv_crtc->lut.nvbo); | 1306 | nouveau_bo_unpin(nv_crtc->lut.nvbo); |
1285 | nouveau_bo_ref(NULL, &nv_crtc->lut.nvbo); | 1307 | nouveau_bo_ref(NULL, &nv_crtc->lut.nvbo); |
1308 | |||
1286 | drm_crtc_cleanup(crtc); | 1309 | drm_crtc_cleanup(crtc); |
1287 | kfree(crtc); | 1310 | kfree(crtc); |
1288 | } | 1311 | } |
@@ -1296,6 +1319,7 @@ static const struct drm_crtc_helper_funcs nv50_crtc_hfunc = { | |||
1296 | .mode_set_base = nv50_crtc_mode_set_base, | 1319 | .mode_set_base = nv50_crtc_mode_set_base, |
1297 | .mode_set_base_atomic = nv50_crtc_mode_set_base_atomic, | 1320 | .mode_set_base_atomic = nv50_crtc_mode_set_base_atomic, |
1298 | .load_lut = nv50_crtc_lut_load, | 1321 | .load_lut = nv50_crtc_lut_load, |
1322 | .disable = nv50_crtc_disable, | ||
1299 | }; | 1323 | }; |
1300 | 1324 | ||
1301 | static const struct drm_crtc_funcs nv50_crtc_func = { | 1325 | static const struct drm_crtc_funcs nv50_crtc_func = { |
diff --git a/drivers/gpu/drm/qxl/qxl_cmd.c b/drivers/gpu/drm/qxl/qxl_cmd.c index 93c2f2cceb51..eb89653a7a17 100644 --- a/drivers/gpu/drm/qxl/qxl_cmd.c +++ b/drivers/gpu/drm/qxl/qxl_cmd.c | |||
@@ -179,9 +179,10 @@ qxl_push_command_ring_release(struct qxl_device *qdev, struct qxl_release *relea | |||
179 | uint32_t type, bool interruptible) | 179 | uint32_t type, bool interruptible) |
180 | { | 180 | { |
181 | struct qxl_command cmd; | 181 | struct qxl_command cmd; |
182 | struct qxl_bo_list *entry = list_first_entry(&release->bos, struct qxl_bo_list, tv.head); | ||
182 | 183 | ||
183 | cmd.type = type; | 184 | cmd.type = type; |
184 | cmd.data = qxl_bo_physical_address(qdev, release->bos[0], release->release_offset); | 185 | cmd.data = qxl_bo_physical_address(qdev, to_qxl_bo(entry->tv.bo), release->release_offset); |
185 | 186 | ||
186 | return qxl_ring_push(qdev->command_ring, &cmd, interruptible); | 187 | return qxl_ring_push(qdev->command_ring, &cmd, interruptible); |
187 | } | 188 | } |
@@ -191,9 +192,10 @@ qxl_push_cursor_ring_release(struct qxl_device *qdev, struct qxl_release *releas | |||
191 | uint32_t type, bool interruptible) | 192 | uint32_t type, bool interruptible) |
192 | { | 193 | { |
193 | struct qxl_command cmd; | 194 | struct qxl_command cmd; |
195 | struct qxl_bo_list *entry = list_first_entry(&release->bos, struct qxl_bo_list, tv.head); | ||
194 | 196 | ||
195 | cmd.type = type; | 197 | cmd.type = type; |
196 | cmd.data = qxl_bo_physical_address(qdev, release->bos[0], release->release_offset); | 198 | cmd.data = qxl_bo_physical_address(qdev, to_qxl_bo(entry->tv.bo), release->release_offset); |
197 | 199 | ||
198 | return qxl_ring_push(qdev->cursor_ring, &cmd, interruptible); | 200 | return qxl_ring_push(qdev->cursor_ring, &cmd, interruptible); |
199 | } | 201 | } |
@@ -214,7 +216,6 @@ int qxl_garbage_collect(struct qxl_device *qdev) | |||
214 | struct qxl_release *release; | 216 | struct qxl_release *release; |
215 | uint64_t id, next_id; | 217 | uint64_t id, next_id; |
216 | int i = 0; | 218 | int i = 0; |
217 | int ret; | ||
218 | union qxl_release_info *info; | 219 | union qxl_release_info *info; |
219 | 220 | ||
220 | while (qxl_ring_pop(qdev->release_ring, &id)) { | 221 | while (qxl_ring_pop(qdev->release_ring, &id)) { |
@@ -224,17 +225,10 @@ int qxl_garbage_collect(struct qxl_device *qdev) | |||
224 | if (release == NULL) | 225 | if (release == NULL) |
225 | break; | 226 | break; |
226 | 227 | ||
227 | ret = qxl_release_reserve(qdev, release, false); | ||
228 | if (ret) { | ||
229 | qxl_io_log(qdev, "failed to reserve release on garbage collect %lld\n", id); | ||
230 | DRM_ERROR("failed to reserve release %lld\n", id); | ||
231 | } | ||
232 | |||
233 | info = qxl_release_map(qdev, release); | 228 | info = qxl_release_map(qdev, release); |
234 | next_id = info->next; | 229 | next_id = info->next; |
235 | qxl_release_unmap(qdev, release, info); | 230 | qxl_release_unmap(qdev, release, info); |
236 | 231 | ||
237 | qxl_release_unreserve(qdev, release); | ||
238 | QXL_INFO(qdev, "popped %lld, next %lld\n", id, | 232 | QXL_INFO(qdev, "popped %lld, next %lld\n", id, |
239 | next_id); | 233 | next_id); |
240 | 234 | ||
@@ -259,27 +253,29 @@ int qxl_garbage_collect(struct qxl_device *qdev) | |||
259 | return i; | 253 | return i; |
260 | } | 254 | } |
261 | 255 | ||
262 | int qxl_alloc_bo_reserved(struct qxl_device *qdev, unsigned long size, | 256 | int qxl_alloc_bo_reserved(struct qxl_device *qdev, |
257 | struct qxl_release *release, | ||
258 | unsigned long size, | ||
263 | struct qxl_bo **_bo) | 259 | struct qxl_bo **_bo) |
264 | { | 260 | { |
265 | struct qxl_bo *bo; | 261 | struct qxl_bo *bo; |
266 | int ret; | 262 | int ret; |
267 | 263 | ||
268 | ret = qxl_bo_create(qdev, size, false /* not kernel - device */, | 264 | ret = qxl_bo_create(qdev, size, false /* not kernel - device */, |
269 | QXL_GEM_DOMAIN_VRAM, NULL, &bo); | 265 | false, QXL_GEM_DOMAIN_VRAM, NULL, &bo); |
270 | if (ret) { | 266 | if (ret) { |
271 | DRM_ERROR("failed to allocate VRAM BO\n"); | 267 | DRM_ERROR("failed to allocate VRAM BO\n"); |
272 | return ret; | 268 | return ret; |
273 | } | 269 | } |
274 | ret = qxl_bo_reserve(bo, false); | 270 | ret = qxl_release_list_add(release, bo); |
275 | if (unlikely(ret != 0)) | 271 | if (ret) |
276 | goto out_unref; | 272 | goto out_unref; |
277 | 273 | ||
278 | *_bo = bo; | 274 | *_bo = bo; |
279 | return 0; | 275 | return 0; |
280 | out_unref: | 276 | out_unref: |
281 | qxl_bo_unref(&bo); | 277 | qxl_bo_unref(&bo); |
282 | return 0; | 278 | return ret; |
283 | } | 279 | } |
284 | 280 | ||
285 | static int wait_for_io_cmd_user(struct qxl_device *qdev, uint8_t val, long port, bool intr) | 281 | static int wait_for_io_cmd_user(struct qxl_device *qdev, uint8_t val, long port, bool intr) |
@@ -503,6 +499,10 @@ int qxl_hw_surface_alloc(struct qxl_device *qdev, | |||
503 | if (ret) | 499 | if (ret) |
504 | return ret; | 500 | return ret; |
505 | 501 | ||
502 | ret = qxl_release_reserve_list(release, true); | ||
503 | if (ret) | ||
504 | return ret; | ||
505 | |||
506 | cmd = (struct qxl_surface_cmd *)qxl_release_map(qdev, release); | 506 | cmd = (struct qxl_surface_cmd *)qxl_release_map(qdev, release); |
507 | cmd->type = QXL_SURFACE_CMD_CREATE; | 507 | cmd->type = QXL_SURFACE_CMD_CREATE; |
508 | cmd->u.surface_create.format = surf->surf.format; | 508 | cmd->u.surface_create.format = surf->surf.format; |
@@ -524,14 +524,11 @@ int qxl_hw_surface_alloc(struct qxl_device *qdev, | |||
524 | 524 | ||
525 | surf->surf_create = release; | 525 | surf->surf_create = release; |
526 | 526 | ||
527 | /* no need to add a release to the fence for this bo, | 527 | /* no need to add a release to the fence for this surface bo, |
528 | since it is only released when we ask to destroy the surface | 528 | since it is only released when we ask to destroy the surface |
529 | and it would never signal otherwise */ | 529 | and it would never signal otherwise */ |
530 | qxl_fence_releaseable(qdev, release); | ||
531 | |||
532 | qxl_push_command_ring_release(qdev, release, QXL_CMD_SURFACE, false); | 530 | qxl_push_command_ring_release(qdev, release, QXL_CMD_SURFACE, false); |
533 | 531 | qxl_release_fence_buffer_objects(release); | |
534 | qxl_release_unreserve(qdev, release); | ||
535 | 532 | ||
536 | surf->hw_surf_alloc = true; | 533 | surf->hw_surf_alloc = true; |
537 | spin_lock(&qdev->surf_id_idr_lock); | 534 | spin_lock(&qdev->surf_id_idr_lock); |
@@ -573,12 +570,9 @@ int qxl_hw_surface_dealloc(struct qxl_device *qdev, | |||
573 | cmd->surface_id = id; | 570 | cmd->surface_id = id; |
574 | qxl_release_unmap(qdev, release, &cmd->release_info); | 571 | qxl_release_unmap(qdev, release, &cmd->release_info); |
575 | 572 | ||
576 | qxl_fence_releaseable(qdev, release); | ||
577 | |||
578 | qxl_push_command_ring_release(qdev, release, QXL_CMD_SURFACE, false); | 573 | qxl_push_command_ring_release(qdev, release, QXL_CMD_SURFACE, false); |
579 | 574 | ||
580 | qxl_release_unreserve(qdev, release); | 575 | qxl_release_fence_buffer_objects(release); |
581 | |||
582 | 576 | ||
583 | return 0; | 577 | return 0; |
584 | } | 578 | } |
diff --git a/drivers/gpu/drm/qxl/qxl_display.c b/drivers/gpu/drm/qxl/qxl_display.c index f76f5dd7bfc4..835caba026d3 100644 --- a/drivers/gpu/drm/qxl/qxl_display.c +++ b/drivers/gpu/drm/qxl/qxl_display.c | |||
@@ -179,7 +179,7 @@ static void qxl_crtc_destroy(struct drm_crtc *crtc) | |||
179 | kfree(qxl_crtc); | 179 | kfree(qxl_crtc); |
180 | } | 180 | } |
181 | 181 | ||
182 | static void | 182 | static int |
183 | qxl_hide_cursor(struct qxl_device *qdev) | 183 | qxl_hide_cursor(struct qxl_device *qdev) |
184 | { | 184 | { |
185 | struct qxl_release *release; | 185 | struct qxl_release *release; |
@@ -188,14 +188,22 @@ qxl_hide_cursor(struct qxl_device *qdev) | |||
188 | 188 | ||
189 | ret = qxl_alloc_release_reserved(qdev, sizeof(*cmd), QXL_RELEASE_CURSOR_CMD, | 189 | ret = qxl_alloc_release_reserved(qdev, sizeof(*cmd), QXL_RELEASE_CURSOR_CMD, |
190 | &release, NULL); | 190 | &release, NULL); |
191 | if (ret) | ||
192 | return ret; | ||
193 | |||
194 | ret = qxl_release_reserve_list(release, true); | ||
195 | if (ret) { | ||
196 | qxl_release_free(qdev, release); | ||
197 | return ret; | ||
198 | } | ||
191 | 199 | ||
192 | cmd = (struct qxl_cursor_cmd *)qxl_release_map(qdev, release); | 200 | cmd = (struct qxl_cursor_cmd *)qxl_release_map(qdev, release); |
193 | cmd->type = QXL_CURSOR_HIDE; | 201 | cmd->type = QXL_CURSOR_HIDE; |
194 | qxl_release_unmap(qdev, release, &cmd->release_info); | 202 | qxl_release_unmap(qdev, release, &cmd->release_info); |
195 | 203 | ||
196 | qxl_fence_releaseable(qdev, release); | ||
197 | qxl_push_cursor_ring_release(qdev, release, QXL_CMD_CURSOR, false); | 204 | qxl_push_cursor_ring_release(qdev, release, QXL_CMD_CURSOR, false); |
198 | qxl_release_unreserve(qdev, release); | 205 | qxl_release_fence_buffer_objects(release); |
206 | return 0; | ||
199 | } | 207 | } |
200 | 208 | ||
201 | static int qxl_crtc_cursor_set2(struct drm_crtc *crtc, | 209 | static int qxl_crtc_cursor_set2(struct drm_crtc *crtc, |
@@ -216,10 +224,8 @@ static int qxl_crtc_cursor_set2(struct drm_crtc *crtc, | |||
216 | 224 | ||
217 | int size = 64*64*4; | 225 | int size = 64*64*4; |
218 | int ret = 0; | 226 | int ret = 0; |
219 | if (!handle) { | 227 | if (!handle) |
220 | qxl_hide_cursor(qdev); | 228 | return qxl_hide_cursor(qdev); |
221 | return 0; | ||
222 | } | ||
223 | 229 | ||
224 | obj = drm_gem_object_lookup(crtc->dev, file_priv, handle); | 230 | obj = drm_gem_object_lookup(crtc->dev, file_priv, handle); |
225 | if (!obj) { | 231 | if (!obj) { |
@@ -234,8 +240,9 @@ static int qxl_crtc_cursor_set2(struct drm_crtc *crtc, | |||
234 | goto out_unref; | 240 | goto out_unref; |
235 | 241 | ||
236 | ret = qxl_bo_pin(user_bo, QXL_GEM_DOMAIN_CPU, NULL); | 242 | ret = qxl_bo_pin(user_bo, QXL_GEM_DOMAIN_CPU, NULL); |
243 | qxl_bo_unreserve(user_bo); | ||
237 | if (ret) | 244 | if (ret) |
238 | goto out_unreserve; | 245 | goto out_unref; |
239 | 246 | ||
240 | ret = qxl_bo_kmap(user_bo, &user_ptr); | 247 | ret = qxl_bo_kmap(user_bo, &user_ptr); |
241 | if (ret) | 248 | if (ret) |
@@ -246,14 +253,20 @@ static int qxl_crtc_cursor_set2(struct drm_crtc *crtc, | |||
246 | &release, NULL); | 253 | &release, NULL); |
247 | if (ret) | 254 | if (ret) |
248 | goto out_kunmap; | 255 | goto out_kunmap; |
249 | ret = qxl_alloc_bo_reserved(qdev, sizeof(struct qxl_cursor) + size, | 256 | |
250 | &cursor_bo); | 257 | ret = qxl_alloc_bo_reserved(qdev, release, sizeof(struct qxl_cursor) + size, |
258 | &cursor_bo); | ||
251 | if (ret) | 259 | if (ret) |
252 | goto out_free_release; | 260 | goto out_free_release; |
253 | ret = qxl_bo_kmap(cursor_bo, (void **)&cursor); | 261 | |
262 | ret = qxl_release_reserve_list(release, false); | ||
254 | if (ret) | 263 | if (ret) |
255 | goto out_free_bo; | 264 | goto out_free_bo; |
256 | 265 | ||
266 | ret = qxl_bo_kmap(cursor_bo, (void **)&cursor); | ||
267 | if (ret) | ||
268 | goto out_backoff; | ||
269 | |||
257 | cursor->header.unique = 0; | 270 | cursor->header.unique = 0; |
258 | cursor->header.type = SPICE_CURSOR_TYPE_ALPHA; | 271 | cursor->header.type = SPICE_CURSOR_TYPE_ALPHA; |
259 | cursor->header.width = 64; | 272 | cursor->header.width = 64; |
@@ -269,11 +282,7 @@ static int qxl_crtc_cursor_set2(struct drm_crtc *crtc, | |||
269 | 282 | ||
270 | qxl_bo_kunmap(cursor_bo); | 283 | qxl_bo_kunmap(cursor_bo); |
271 | 284 | ||
272 | /* finish with the userspace bo */ | ||
273 | qxl_bo_kunmap(user_bo); | 285 | qxl_bo_kunmap(user_bo); |
274 | qxl_bo_unpin(user_bo); | ||
275 | qxl_bo_unreserve(user_bo); | ||
276 | drm_gem_object_unreference_unlocked(obj); | ||
277 | 286 | ||
278 | cmd = (struct qxl_cursor_cmd *)qxl_release_map(qdev, release); | 287 | cmd = (struct qxl_cursor_cmd *)qxl_release_map(qdev, release); |
279 | cmd->type = QXL_CURSOR_SET; | 288 | cmd->type = QXL_CURSOR_SET; |
@@ -281,30 +290,35 @@ static int qxl_crtc_cursor_set2(struct drm_crtc *crtc, | |||
281 | cmd->u.set.position.y = qcrtc->cur_y; | 290 | cmd->u.set.position.y = qcrtc->cur_y; |
282 | 291 | ||
283 | cmd->u.set.shape = qxl_bo_physical_address(qdev, cursor_bo, 0); | 292 | cmd->u.set.shape = qxl_bo_physical_address(qdev, cursor_bo, 0); |
284 | qxl_release_add_res(qdev, release, cursor_bo); | ||
285 | 293 | ||
286 | cmd->u.set.visible = 1; | 294 | cmd->u.set.visible = 1; |
287 | qxl_release_unmap(qdev, release, &cmd->release_info); | 295 | qxl_release_unmap(qdev, release, &cmd->release_info); |
288 | 296 | ||
289 | qxl_fence_releaseable(qdev, release); | ||
290 | qxl_push_cursor_ring_release(qdev, release, QXL_CMD_CURSOR, false); | 297 | qxl_push_cursor_ring_release(qdev, release, QXL_CMD_CURSOR, false); |
291 | qxl_release_unreserve(qdev, release); | 298 | qxl_release_fence_buffer_objects(release); |
299 | |||
300 | /* finish with the userspace bo */ | ||
301 | ret = qxl_bo_reserve(user_bo, false); | ||
302 | if (!ret) { | ||
303 | qxl_bo_unpin(user_bo); | ||
304 | qxl_bo_unreserve(user_bo); | ||
305 | } | ||
306 | drm_gem_object_unreference_unlocked(obj); | ||
292 | 307 | ||
293 | qxl_bo_unreserve(cursor_bo); | ||
294 | qxl_bo_unref(&cursor_bo); | 308 | qxl_bo_unref(&cursor_bo); |
295 | 309 | ||
296 | return ret; | 310 | return ret; |
311 | |||
312 | out_backoff: | ||
313 | qxl_release_backoff_reserve_list(release); | ||
297 | out_free_bo: | 314 | out_free_bo: |
298 | qxl_bo_unref(&cursor_bo); | 315 | qxl_bo_unref(&cursor_bo); |
299 | out_free_release: | 316 | out_free_release: |
300 | qxl_release_unreserve(qdev, release); | ||
301 | qxl_release_free(qdev, release); | 317 | qxl_release_free(qdev, release); |
302 | out_kunmap: | 318 | out_kunmap: |
303 | qxl_bo_kunmap(user_bo); | 319 | qxl_bo_kunmap(user_bo); |
304 | out_unpin: | 320 | out_unpin: |
305 | qxl_bo_unpin(user_bo); | 321 | qxl_bo_unpin(user_bo); |
306 | out_unreserve: | ||
307 | qxl_bo_unreserve(user_bo); | ||
308 | out_unref: | 322 | out_unref: |
309 | drm_gem_object_unreference_unlocked(obj); | 323 | drm_gem_object_unreference_unlocked(obj); |
310 | return ret; | 324 | return ret; |
@@ -322,6 +336,14 @@ static int qxl_crtc_cursor_move(struct drm_crtc *crtc, | |||
322 | 336 | ||
323 | ret = qxl_alloc_release_reserved(qdev, sizeof(*cmd), QXL_RELEASE_CURSOR_CMD, | 337 | ret = qxl_alloc_release_reserved(qdev, sizeof(*cmd), QXL_RELEASE_CURSOR_CMD, |
324 | &release, NULL); | 338 | &release, NULL); |
339 | if (ret) | ||
340 | return ret; | ||
341 | |||
342 | ret = qxl_release_reserve_list(release, true); | ||
343 | if (ret) { | ||
344 | qxl_release_free(qdev, release); | ||
345 | return ret; | ||
346 | } | ||
325 | 347 | ||
326 | qcrtc->cur_x = x; | 348 | qcrtc->cur_x = x; |
327 | qcrtc->cur_y = y; | 349 | qcrtc->cur_y = y; |
@@ -332,9 +354,9 @@ static int qxl_crtc_cursor_move(struct drm_crtc *crtc, | |||
332 | cmd->u.position.y = qcrtc->cur_y; | 354 | cmd->u.position.y = qcrtc->cur_y; |
333 | qxl_release_unmap(qdev, release, &cmd->release_info); | 355 | qxl_release_unmap(qdev, release, &cmd->release_info); |
334 | 356 | ||
335 | qxl_fence_releaseable(qdev, release); | ||
336 | qxl_push_cursor_ring_release(qdev, release, QXL_CMD_CURSOR, false); | 357 | qxl_push_cursor_ring_release(qdev, release, QXL_CMD_CURSOR, false); |
337 | qxl_release_unreserve(qdev, release); | 358 | qxl_release_fence_buffer_objects(release); |
359 | |||
338 | return 0; | 360 | return 0; |
339 | } | 361 | } |
340 | 362 | ||
diff --git a/drivers/gpu/drm/qxl/qxl_draw.c b/drivers/gpu/drm/qxl/qxl_draw.c index 3c8c3dbf9378..56e1d633875e 100644 --- a/drivers/gpu/drm/qxl/qxl_draw.c +++ b/drivers/gpu/drm/qxl/qxl_draw.c | |||
@@ -23,25 +23,29 @@ | |||
23 | #include "qxl_drv.h" | 23 | #include "qxl_drv.h" |
24 | #include "qxl_object.h" | 24 | #include "qxl_object.h" |
25 | 25 | ||
26 | static int alloc_clips(struct qxl_device *qdev, | ||
27 | struct qxl_release *release, | ||
28 | unsigned num_clips, | ||
29 | struct qxl_bo **clips_bo) | ||
30 | { | ||
31 | int size = sizeof(struct qxl_clip_rects) + sizeof(struct qxl_rect) * num_clips; | ||
32 | |||
33 | return qxl_alloc_bo_reserved(qdev, release, size, clips_bo); | ||
34 | } | ||
35 | |||
26 | /* returns a pointer to the already allocated qxl_rect array inside | 36 | /* returns a pointer to the already allocated qxl_rect array inside |
27 | * the qxl_clip_rects. This is *not* the same as the memory allocated | 37 | * the qxl_clip_rects. This is *not* the same as the memory allocated |
28 | * on the device, it is offset to qxl_clip_rects.chunk.data */ | 38 | * on the device, it is offset to qxl_clip_rects.chunk.data */ |
29 | static struct qxl_rect *drawable_set_clipping(struct qxl_device *qdev, | 39 | static struct qxl_rect *drawable_set_clipping(struct qxl_device *qdev, |
30 | struct qxl_drawable *drawable, | 40 | struct qxl_drawable *drawable, |
31 | unsigned num_clips, | 41 | unsigned num_clips, |
32 | struct qxl_bo **clips_bo, | 42 | struct qxl_bo *clips_bo) |
33 | struct qxl_release *release) | ||
34 | { | 43 | { |
35 | struct qxl_clip_rects *dev_clips; | 44 | struct qxl_clip_rects *dev_clips; |
36 | int ret; | 45 | int ret; |
37 | int size = sizeof(*dev_clips) + sizeof(struct qxl_rect) * num_clips; | ||
38 | ret = qxl_alloc_bo_reserved(qdev, size, clips_bo); | ||
39 | if (ret) | ||
40 | return NULL; | ||
41 | 46 | ||
42 | ret = qxl_bo_kmap(*clips_bo, (void **)&dev_clips); | 47 | ret = qxl_bo_kmap(clips_bo, (void **)&dev_clips); |
43 | if (ret) { | 48 | if (ret) { |
44 | qxl_bo_unref(clips_bo); | ||
45 | return NULL; | 49 | return NULL; |
46 | } | 50 | } |
47 | dev_clips->num_rects = num_clips; | 51 | dev_clips->num_rects = num_clips; |
@@ -52,20 +56,34 @@ static struct qxl_rect *drawable_set_clipping(struct qxl_device *qdev, | |||
52 | } | 56 | } |
53 | 57 | ||
54 | static int | 58 | static int |
59 | alloc_drawable(struct qxl_device *qdev, struct qxl_release **release) | ||
60 | { | ||
61 | int ret; | ||
62 | ret = qxl_alloc_release_reserved(qdev, sizeof(struct qxl_drawable), | ||
63 | QXL_RELEASE_DRAWABLE, release, | ||
64 | NULL); | ||
65 | return ret; | ||
66 | } | ||
67 | |||
68 | static void | ||
69 | free_drawable(struct qxl_device *qdev, struct qxl_release *release) | ||
70 | { | ||
71 | qxl_release_free(qdev, release); | ||
72 | } | ||
73 | |||
74 | /* release needs to be reserved at this point */ | ||
75 | static int | ||
55 | make_drawable(struct qxl_device *qdev, int surface, uint8_t type, | 76 | make_drawable(struct qxl_device *qdev, int surface, uint8_t type, |
56 | const struct qxl_rect *rect, | 77 | const struct qxl_rect *rect, |
57 | struct qxl_release **release) | 78 | struct qxl_release *release) |
58 | { | 79 | { |
59 | struct qxl_drawable *drawable; | 80 | struct qxl_drawable *drawable; |
60 | int i, ret; | 81 | int i; |
61 | 82 | ||
62 | ret = qxl_alloc_release_reserved(qdev, sizeof(*drawable), | 83 | drawable = (struct qxl_drawable *)qxl_release_map(qdev, release); |
63 | QXL_RELEASE_DRAWABLE, release, | 84 | if (!drawable) |
64 | NULL); | 85 | return -ENOMEM; |
65 | if (ret) | ||
66 | return ret; | ||
67 | 86 | ||
68 | drawable = (struct qxl_drawable *)qxl_release_map(qdev, *release); | ||
69 | drawable->type = type; | 87 | drawable->type = type; |
70 | 88 | ||
71 | drawable->surface_id = surface; /* Only primary for now */ | 89 | drawable->surface_id = surface; /* Only primary for now */ |
@@ -91,14 +109,23 @@ make_drawable(struct qxl_device *qdev, int surface, uint8_t type, | |||
91 | drawable->bbox = *rect; | 109 | drawable->bbox = *rect; |
92 | 110 | ||
93 | drawable->mm_time = qdev->rom->mm_clock; | 111 | drawable->mm_time = qdev->rom->mm_clock; |
94 | qxl_release_unmap(qdev, *release, &drawable->release_info); | 112 | qxl_release_unmap(qdev, release, &drawable->release_info); |
95 | return 0; | 113 | return 0; |
96 | } | 114 | } |
97 | 115 | ||
98 | static int qxl_palette_create_1bit(struct qxl_bo **palette_bo, | 116 | static int alloc_palette_object(struct qxl_device *qdev, |
117 | struct qxl_release *release, | ||
118 | struct qxl_bo **palette_bo) | ||
119 | { | ||
120 | return qxl_alloc_bo_reserved(qdev, release, | ||
121 | sizeof(struct qxl_palette) + sizeof(uint32_t) * 2, | ||
122 | palette_bo); | ||
123 | } | ||
124 | |||
125 | static int qxl_palette_create_1bit(struct qxl_bo *palette_bo, | ||
126 | struct qxl_release *release, | ||
99 | const struct qxl_fb_image *qxl_fb_image) | 127 | const struct qxl_fb_image *qxl_fb_image) |
100 | { | 128 | { |
101 | struct qxl_device *qdev = qxl_fb_image->qdev; | ||
102 | const struct fb_image *fb_image = &qxl_fb_image->fb_image; | 129 | const struct fb_image *fb_image = &qxl_fb_image->fb_image; |
103 | uint32_t visual = qxl_fb_image->visual; | 130 | uint32_t visual = qxl_fb_image->visual; |
104 | const uint32_t *pseudo_palette = qxl_fb_image->pseudo_palette; | 131 | const uint32_t *pseudo_palette = qxl_fb_image->pseudo_palette; |
@@ -108,12 +135,7 @@ static int qxl_palette_create_1bit(struct qxl_bo **palette_bo, | |||
108 | static uint64_t unique; /* we make no attempt to actually set this | 135 | static uint64_t unique; /* we make no attempt to actually set this |
109 | * correctly globaly, since that would require | 136 | * correctly globaly, since that would require |
110 | * tracking all of our palettes. */ | 137 | * tracking all of our palettes. */ |
111 | 138 | ret = qxl_bo_kmap(palette_bo, (void **)&pal); | |
112 | ret = qxl_alloc_bo_reserved(qdev, | ||
113 | sizeof(struct qxl_palette) + sizeof(uint32_t) * 2, | ||
114 | palette_bo); | ||
115 | |||
116 | ret = qxl_bo_kmap(*palette_bo, (void **)&pal); | ||
117 | pal->num_ents = 2; | 139 | pal->num_ents = 2; |
118 | pal->unique = unique++; | 140 | pal->unique = unique++; |
119 | if (visual == FB_VISUAL_TRUECOLOR || visual == FB_VISUAL_DIRECTCOLOR) { | 141 | if (visual == FB_VISUAL_TRUECOLOR || visual == FB_VISUAL_DIRECTCOLOR) { |
@@ -126,7 +148,7 @@ static int qxl_palette_create_1bit(struct qxl_bo **palette_bo, | |||
126 | } | 148 | } |
127 | pal->ents[0] = bgcolor; | 149 | pal->ents[0] = bgcolor; |
128 | pal->ents[1] = fgcolor; | 150 | pal->ents[1] = fgcolor; |
129 | qxl_bo_kunmap(*palette_bo); | 151 | qxl_bo_kunmap(palette_bo); |
130 | return 0; | 152 | return 0; |
131 | } | 153 | } |
132 | 154 | ||
@@ -144,44 +166,63 @@ void qxl_draw_opaque_fb(const struct qxl_fb_image *qxl_fb_image, | |||
144 | const char *src = fb_image->data; | 166 | const char *src = fb_image->data; |
145 | int depth = fb_image->depth; | 167 | int depth = fb_image->depth; |
146 | struct qxl_release *release; | 168 | struct qxl_release *release; |
147 | struct qxl_bo *image_bo; | ||
148 | struct qxl_image *image; | 169 | struct qxl_image *image; |
149 | int ret; | 170 | int ret; |
150 | 171 | struct qxl_drm_image *dimage; | |
172 | struct qxl_bo *palette_bo = NULL; | ||
151 | if (stride == 0) | 173 | if (stride == 0) |
152 | stride = depth * width / 8; | 174 | stride = depth * width / 8; |
153 | 175 | ||
176 | ret = alloc_drawable(qdev, &release); | ||
177 | if (ret) | ||
178 | return; | ||
179 | |||
180 | ret = qxl_image_alloc_objects(qdev, release, | ||
181 | &dimage, | ||
182 | height, stride); | ||
183 | if (ret) | ||
184 | goto out_free_drawable; | ||
185 | |||
186 | if (depth == 1) { | ||
187 | ret = alloc_palette_object(qdev, release, &palette_bo); | ||
188 | if (ret) | ||
189 | goto out_free_image; | ||
190 | } | ||
191 | |||
192 | /* do a reservation run over all the objects we just allocated */ | ||
193 | ret = qxl_release_reserve_list(release, true); | ||
194 | if (ret) | ||
195 | goto out_free_palette; | ||
196 | |||
154 | rect.left = x; | 197 | rect.left = x; |
155 | rect.right = x + width; | 198 | rect.right = x + width; |
156 | rect.top = y; | 199 | rect.top = y; |
157 | rect.bottom = y + height; | 200 | rect.bottom = y + height; |
158 | 201 | ||
159 | ret = make_drawable(qdev, 0, QXL_DRAW_COPY, &rect, &release); | 202 | ret = make_drawable(qdev, 0, QXL_DRAW_COPY, &rect, release); |
160 | if (ret) | 203 | if (ret) { |
161 | return; | 204 | qxl_release_backoff_reserve_list(release); |
205 | goto out_free_palette; | ||
206 | } | ||
162 | 207 | ||
163 | ret = qxl_image_create(qdev, release, &image_bo, | 208 | ret = qxl_image_init(qdev, release, dimage, |
164 | (const uint8_t *)src, 0, 0, | 209 | (const uint8_t *)src, 0, 0, |
165 | width, height, depth, stride); | 210 | width, height, depth, stride); |
166 | if (ret) { | 211 | if (ret) { |
167 | qxl_release_unreserve(qdev, release); | 212 | qxl_release_backoff_reserve_list(release); |
168 | qxl_release_free(qdev, release); | 213 | qxl_release_free(qdev, release); |
169 | return; | 214 | return; |
170 | } | 215 | } |
171 | 216 | ||
172 | if (depth == 1) { | 217 | if (depth == 1) { |
173 | struct qxl_bo *palette_bo; | ||
174 | void *ptr; | 218 | void *ptr; |
175 | ret = qxl_palette_create_1bit(&palette_bo, qxl_fb_image); | 219 | ret = qxl_palette_create_1bit(palette_bo, release, qxl_fb_image); |
176 | qxl_release_add_res(qdev, release, palette_bo); | ||
177 | 220 | ||
178 | ptr = qxl_bo_kmap_atomic_page(qdev, image_bo, 0); | 221 | ptr = qxl_bo_kmap_atomic_page(qdev, dimage->bo, 0); |
179 | image = ptr; | 222 | image = ptr; |
180 | image->u.bitmap.palette = | 223 | image->u.bitmap.palette = |
181 | qxl_bo_physical_address(qdev, palette_bo, 0); | 224 | qxl_bo_physical_address(qdev, palette_bo, 0); |
182 | qxl_bo_kunmap_atomic_page(qdev, image_bo, ptr); | 225 | qxl_bo_kunmap_atomic_page(qdev, dimage->bo, ptr); |
183 | qxl_bo_unreserve(palette_bo); | ||
184 | qxl_bo_unref(&palette_bo); | ||
185 | } | 226 | } |
186 | 227 | ||
187 | drawable = (struct qxl_drawable *)qxl_release_map(qdev, release); | 228 | drawable = (struct qxl_drawable *)qxl_release_map(qdev, release); |
@@ -199,16 +240,20 @@ void qxl_draw_opaque_fb(const struct qxl_fb_image *qxl_fb_image, | |||
199 | drawable->u.copy.mask.bitmap = 0; | 240 | drawable->u.copy.mask.bitmap = 0; |
200 | 241 | ||
201 | drawable->u.copy.src_bitmap = | 242 | drawable->u.copy.src_bitmap = |
202 | qxl_bo_physical_address(qdev, image_bo, 0); | 243 | qxl_bo_physical_address(qdev, dimage->bo, 0); |
203 | qxl_release_unmap(qdev, release, &drawable->release_info); | 244 | qxl_release_unmap(qdev, release, &drawable->release_info); |
204 | 245 | ||
205 | qxl_release_add_res(qdev, release, image_bo); | ||
206 | qxl_bo_unreserve(image_bo); | ||
207 | qxl_bo_unref(&image_bo); | ||
208 | |||
209 | qxl_fence_releaseable(qdev, release); | ||
210 | qxl_push_command_ring_release(qdev, release, QXL_CMD_DRAW, false); | 246 | qxl_push_command_ring_release(qdev, release, QXL_CMD_DRAW, false); |
211 | qxl_release_unreserve(qdev, release); | 247 | qxl_release_fence_buffer_objects(release); |
248 | |||
249 | out_free_palette: | ||
250 | if (palette_bo) | ||
251 | qxl_bo_unref(&palette_bo); | ||
252 | out_free_image: | ||
253 | qxl_image_free_objects(qdev, dimage); | ||
254 | out_free_drawable: | ||
255 | if (ret) | ||
256 | free_drawable(qdev, release); | ||
212 | } | 257 | } |
213 | 258 | ||
214 | /* push a draw command using the given clipping rectangles as | 259 | /* push a draw command using the given clipping rectangles as |
@@ -243,10 +288,14 @@ void qxl_draw_dirty_fb(struct qxl_device *qdev, | |||
243 | int depth = qxl_fb->base.bits_per_pixel; | 288 | int depth = qxl_fb->base.bits_per_pixel; |
244 | uint8_t *surface_base; | 289 | uint8_t *surface_base; |
245 | struct qxl_release *release; | 290 | struct qxl_release *release; |
246 | struct qxl_bo *image_bo; | ||
247 | struct qxl_bo *clips_bo; | 291 | struct qxl_bo *clips_bo; |
292 | struct qxl_drm_image *dimage; | ||
248 | int ret; | 293 | int ret; |
249 | 294 | ||
295 | ret = alloc_drawable(qdev, &release); | ||
296 | if (ret) | ||
297 | return; | ||
298 | |||
250 | left = clips->x1; | 299 | left = clips->x1; |
251 | right = clips->x2; | 300 | right = clips->x2; |
252 | top = clips->y1; | 301 | top = clips->y1; |
@@ -263,36 +312,52 @@ void qxl_draw_dirty_fb(struct qxl_device *qdev, | |||
263 | 312 | ||
264 | width = right - left; | 313 | width = right - left; |
265 | height = bottom - top; | 314 | height = bottom - top; |
315 | |||
316 | ret = alloc_clips(qdev, release, num_clips, &clips_bo); | ||
317 | if (ret) | ||
318 | goto out_free_drawable; | ||
319 | |||
320 | ret = qxl_image_alloc_objects(qdev, release, | ||
321 | &dimage, | ||
322 | height, stride); | ||
323 | if (ret) | ||
324 | goto out_free_clips; | ||
325 | |||
326 | /* do a reservation run over all the objects we just allocated */ | ||
327 | ret = qxl_release_reserve_list(release, true); | ||
328 | if (ret) | ||
329 | goto out_free_image; | ||
330 | |||
266 | drawable_rect.left = left; | 331 | drawable_rect.left = left; |
267 | drawable_rect.right = right; | 332 | drawable_rect.right = right; |
268 | drawable_rect.top = top; | 333 | drawable_rect.top = top; |
269 | drawable_rect.bottom = bottom; | 334 | drawable_rect.bottom = bottom; |
335 | |||
270 | ret = make_drawable(qdev, 0, QXL_DRAW_COPY, &drawable_rect, | 336 | ret = make_drawable(qdev, 0, QXL_DRAW_COPY, &drawable_rect, |
271 | &release); | 337 | release); |
272 | if (ret) | 338 | if (ret) |
273 | return; | 339 | goto out_release_backoff; |
274 | 340 | ||
275 | ret = qxl_bo_kmap(bo, (void **)&surface_base); | 341 | ret = qxl_bo_kmap(bo, (void **)&surface_base); |
276 | if (ret) | 342 | if (ret) |
277 | goto out_unref; | 343 | goto out_release_backoff; |
278 | 344 | ||
279 | ret = qxl_image_create(qdev, release, &image_bo, surface_base, | 345 | |
280 | left, top, width, height, depth, stride); | 346 | ret = qxl_image_init(qdev, release, dimage, surface_base, |
347 | left, top, width, height, depth, stride); | ||
281 | qxl_bo_kunmap(bo); | 348 | qxl_bo_kunmap(bo); |
282 | if (ret) | 349 | if (ret) |
283 | goto out_unref; | 350 | goto out_release_backoff; |
351 | |||
352 | rects = drawable_set_clipping(qdev, drawable, num_clips, clips_bo); | ||
353 | if (!rects) | ||
354 | goto out_release_backoff; | ||
284 | 355 | ||
285 | rects = drawable_set_clipping(qdev, drawable, num_clips, &clips_bo, release); | ||
286 | if (!rects) { | ||
287 | qxl_bo_unref(&image_bo); | ||
288 | goto out_unref; | ||
289 | } | ||
290 | drawable = (struct qxl_drawable *)qxl_release_map(qdev, release); | 356 | drawable = (struct qxl_drawable *)qxl_release_map(qdev, release); |
291 | 357 | ||
292 | drawable->clip.type = SPICE_CLIP_TYPE_RECTS; | 358 | drawable->clip.type = SPICE_CLIP_TYPE_RECTS; |
293 | drawable->clip.data = qxl_bo_physical_address(qdev, | 359 | drawable->clip.data = qxl_bo_physical_address(qdev, |
294 | clips_bo, 0); | 360 | clips_bo, 0); |
295 | qxl_release_add_res(qdev, release, clips_bo); | ||
296 | 361 | ||
297 | drawable->u.copy.src_area.top = 0; | 362 | drawable->u.copy.src_area.top = 0; |
298 | drawable->u.copy.src_area.bottom = height; | 363 | drawable->u.copy.src_area.bottom = height; |
@@ -306,11 +371,9 @@ void qxl_draw_dirty_fb(struct qxl_device *qdev, | |||
306 | drawable->u.copy.mask.pos.y = 0; | 371 | drawable->u.copy.mask.pos.y = 0; |
307 | drawable->u.copy.mask.bitmap = 0; | 372 | drawable->u.copy.mask.bitmap = 0; |
308 | 373 | ||
309 | drawable->u.copy.src_bitmap = qxl_bo_physical_address(qdev, image_bo, 0); | 374 | drawable->u.copy.src_bitmap = qxl_bo_physical_address(qdev, dimage->bo, 0); |
310 | qxl_release_unmap(qdev, release, &drawable->release_info); | 375 | qxl_release_unmap(qdev, release, &drawable->release_info); |
311 | qxl_release_add_res(qdev, release, image_bo); | 376 | |
312 | qxl_bo_unreserve(image_bo); | ||
313 | qxl_bo_unref(&image_bo); | ||
314 | clips_ptr = clips; | 377 | clips_ptr = clips; |
315 | for (i = 0; i < num_clips; i++, clips_ptr += inc) { | 378 | for (i = 0; i < num_clips; i++, clips_ptr += inc) { |
316 | rects[i].left = clips_ptr->x1; | 379 | rects[i].left = clips_ptr->x1; |
@@ -319,17 +382,22 @@ void qxl_draw_dirty_fb(struct qxl_device *qdev, | |||
319 | rects[i].bottom = clips_ptr->y2; | 382 | rects[i].bottom = clips_ptr->y2; |
320 | } | 383 | } |
321 | qxl_bo_kunmap(clips_bo); | 384 | qxl_bo_kunmap(clips_bo); |
322 | qxl_bo_unreserve(clips_bo); | ||
323 | qxl_bo_unref(&clips_bo); | ||
324 | 385 | ||
325 | qxl_fence_releaseable(qdev, release); | ||
326 | qxl_push_command_ring_release(qdev, release, QXL_CMD_DRAW, false); | 386 | qxl_push_command_ring_release(qdev, release, QXL_CMD_DRAW, false); |
327 | qxl_release_unreserve(qdev, release); | 387 | qxl_release_fence_buffer_objects(release); |
328 | return; | 388 | |
389 | out_release_backoff: | ||
390 | if (ret) | ||
391 | qxl_release_backoff_reserve_list(release); | ||
392 | out_free_image: | ||
393 | qxl_image_free_objects(qdev, dimage); | ||
394 | out_free_clips: | ||
395 | qxl_bo_unref(&clips_bo); | ||
396 | out_free_drawable: | ||
397 | /* only free drawable on error */ | ||
398 | if (ret) | ||
399 | free_drawable(qdev, release); | ||
329 | 400 | ||
330 | out_unref: | ||
331 | qxl_release_unreserve(qdev, release); | ||
332 | qxl_release_free(qdev, release); | ||
333 | } | 401 | } |
334 | 402 | ||
335 | void qxl_draw_copyarea(struct qxl_device *qdev, | 403 | void qxl_draw_copyarea(struct qxl_device *qdev, |
@@ -342,22 +410,36 @@ void qxl_draw_copyarea(struct qxl_device *qdev, | |||
342 | struct qxl_release *release; | 410 | struct qxl_release *release; |
343 | int ret; | 411 | int ret; |
344 | 412 | ||
413 | ret = alloc_drawable(qdev, &release); | ||
414 | if (ret) | ||
415 | return; | ||
416 | |||
417 | /* do a reservation run over all the objects we just allocated */ | ||
418 | ret = qxl_release_reserve_list(release, true); | ||
419 | if (ret) | ||
420 | goto out_free_release; | ||
421 | |||
345 | rect.left = dx; | 422 | rect.left = dx; |
346 | rect.top = dy; | 423 | rect.top = dy; |
347 | rect.right = dx + width; | 424 | rect.right = dx + width; |
348 | rect.bottom = dy + height; | 425 | rect.bottom = dy + height; |
349 | ret = make_drawable(qdev, 0, QXL_COPY_BITS, &rect, &release); | 426 | ret = make_drawable(qdev, 0, QXL_COPY_BITS, &rect, release); |
350 | if (ret) | 427 | if (ret) { |
351 | return; | 428 | qxl_release_backoff_reserve_list(release); |
429 | goto out_free_release; | ||
430 | } | ||
352 | 431 | ||
353 | drawable = (struct qxl_drawable *)qxl_release_map(qdev, release); | 432 | drawable = (struct qxl_drawable *)qxl_release_map(qdev, release); |
354 | drawable->u.copy_bits.src_pos.x = sx; | 433 | drawable->u.copy_bits.src_pos.x = sx; |
355 | drawable->u.copy_bits.src_pos.y = sy; | 434 | drawable->u.copy_bits.src_pos.y = sy; |
356 | |||
357 | qxl_release_unmap(qdev, release, &drawable->release_info); | 435 | qxl_release_unmap(qdev, release, &drawable->release_info); |
358 | qxl_fence_releaseable(qdev, release); | 436 | |
359 | qxl_push_command_ring_release(qdev, release, QXL_CMD_DRAW, false); | 437 | qxl_push_command_ring_release(qdev, release, QXL_CMD_DRAW, false); |
360 | qxl_release_unreserve(qdev, release); | 438 | qxl_release_fence_buffer_objects(release); |
439 | |||
440 | out_free_release: | ||
441 | if (ret) | ||
442 | free_drawable(qdev, release); | ||
361 | } | 443 | } |
362 | 444 | ||
363 | void qxl_draw_fill(struct qxl_draw_fill *qxl_draw_fill_rec) | 445 | void qxl_draw_fill(struct qxl_draw_fill *qxl_draw_fill_rec) |
@@ -370,10 +452,21 @@ void qxl_draw_fill(struct qxl_draw_fill *qxl_draw_fill_rec) | |||
370 | struct qxl_release *release; | 452 | struct qxl_release *release; |
371 | int ret; | 453 | int ret; |
372 | 454 | ||
373 | ret = make_drawable(qdev, 0, QXL_DRAW_FILL, &rect, &release); | 455 | ret = alloc_drawable(qdev, &release); |
374 | if (ret) | 456 | if (ret) |
375 | return; | 457 | return; |
376 | 458 | ||
459 | /* do a reservation run over all the objects we just allocated */ | ||
460 | ret = qxl_release_reserve_list(release, true); | ||
461 | if (ret) | ||
462 | goto out_free_release; | ||
463 | |||
464 | ret = make_drawable(qdev, 0, QXL_DRAW_FILL, &rect, release); | ||
465 | if (ret) { | ||
466 | qxl_release_backoff_reserve_list(release); | ||
467 | goto out_free_release; | ||
468 | } | ||
469 | |||
377 | drawable = (struct qxl_drawable *)qxl_release_map(qdev, release); | 470 | drawable = (struct qxl_drawable *)qxl_release_map(qdev, release); |
378 | drawable->u.fill.brush.type = SPICE_BRUSH_TYPE_SOLID; | 471 | drawable->u.fill.brush.type = SPICE_BRUSH_TYPE_SOLID; |
379 | drawable->u.fill.brush.u.color = color; | 472 | drawable->u.fill.brush.u.color = color; |
@@ -384,7 +477,11 @@ void qxl_draw_fill(struct qxl_draw_fill *qxl_draw_fill_rec) | |||
384 | drawable->u.fill.mask.bitmap = 0; | 477 | drawable->u.fill.mask.bitmap = 0; |
385 | 478 | ||
386 | qxl_release_unmap(qdev, release, &drawable->release_info); | 479 | qxl_release_unmap(qdev, release, &drawable->release_info); |
387 | qxl_fence_releaseable(qdev, release); | 480 | |
388 | qxl_push_command_ring_release(qdev, release, QXL_CMD_DRAW, false); | 481 | qxl_push_command_ring_release(qdev, release, QXL_CMD_DRAW, false); |
389 | qxl_release_unreserve(qdev, release); | 482 | qxl_release_fence_buffer_objects(release); |
483 | |||
484 | out_free_release: | ||
485 | if (ret) | ||
486 | free_drawable(qdev, release); | ||
390 | } | 487 | } |
diff --git a/drivers/gpu/drm/qxl/qxl_drv.h b/drivers/gpu/drm/qxl/qxl_drv.h index aacb791464a3..7e96f4f11738 100644 --- a/drivers/gpu/drm/qxl/qxl_drv.h +++ b/drivers/gpu/drm/qxl/qxl_drv.h | |||
@@ -42,6 +42,9 @@ | |||
42 | #include <ttm/ttm_placement.h> | 42 | #include <ttm/ttm_placement.h> |
43 | #include <ttm/ttm_module.h> | 43 | #include <ttm/ttm_module.h> |
44 | 44 | ||
45 | /* just for ttm_validate_buffer */ | ||
46 | #include <ttm/ttm_execbuf_util.h> | ||
47 | |||
45 | #include <drm/qxl_drm.h> | 48 | #include <drm/qxl_drm.h> |
46 | #include "qxl_dev.h" | 49 | #include "qxl_dev.h" |
47 | 50 | ||
@@ -118,9 +121,9 @@ struct qxl_bo { | |||
118 | uint32_t surface_id; | 121 | uint32_t surface_id; |
119 | struct qxl_fence fence; /* per bo fence - list of releases */ | 122 | struct qxl_fence fence; /* per bo fence - list of releases */ |
120 | struct qxl_release *surf_create; | 123 | struct qxl_release *surf_create; |
121 | atomic_t reserve_count; | ||
122 | }; | 124 | }; |
123 | #define gem_to_qxl_bo(gobj) container_of((gobj), struct qxl_bo, gem_base) | 125 | #define gem_to_qxl_bo(gobj) container_of((gobj), struct qxl_bo, gem_base) |
126 | #define to_qxl_bo(tobj) container_of((tobj), struct qxl_bo, tbo) | ||
124 | 127 | ||
125 | struct qxl_gem { | 128 | struct qxl_gem { |
126 | struct mutex mutex; | 129 | struct mutex mutex; |
@@ -128,12 +131,7 @@ struct qxl_gem { | |||
128 | }; | 131 | }; |
129 | 132 | ||
130 | struct qxl_bo_list { | 133 | struct qxl_bo_list { |
131 | struct list_head lhead; | 134 | struct ttm_validate_buffer tv; |
132 | struct qxl_bo *bo; | ||
133 | }; | ||
134 | |||
135 | struct qxl_reloc_list { | ||
136 | struct list_head bos; | ||
137 | }; | 135 | }; |
138 | 136 | ||
139 | struct qxl_crtc { | 137 | struct qxl_crtc { |
@@ -195,10 +193,20 @@ enum { | |||
195 | struct qxl_release { | 193 | struct qxl_release { |
196 | int id; | 194 | int id; |
197 | int type; | 195 | int type; |
198 | int bo_count; | ||
199 | uint32_t release_offset; | 196 | uint32_t release_offset; |
200 | uint32_t surface_release_id; | 197 | uint32_t surface_release_id; |
201 | struct qxl_bo *bos[QXL_MAX_RES]; | 198 | struct ww_acquire_ctx ticket; |
199 | struct list_head bos; | ||
200 | }; | ||
201 | |||
202 | struct qxl_drm_chunk { | ||
203 | struct list_head head; | ||
204 | struct qxl_bo *bo; | ||
205 | }; | ||
206 | |||
207 | struct qxl_drm_image { | ||
208 | struct qxl_bo *bo; | ||
209 | struct list_head chunk_list; | ||
202 | }; | 210 | }; |
203 | 211 | ||
204 | struct qxl_fb_image { | 212 | struct qxl_fb_image { |
@@ -314,6 +322,7 @@ struct qxl_device { | |||
314 | struct workqueue_struct *gc_queue; | 322 | struct workqueue_struct *gc_queue; |
315 | struct work_struct gc_work; | 323 | struct work_struct gc_work; |
316 | 324 | ||
325 | struct work_struct fb_work; | ||
317 | }; | 326 | }; |
318 | 327 | ||
319 | /* forward declaration for QXL_INFO_IO */ | 328 | /* forward declaration for QXL_INFO_IO */ |
@@ -433,12 +442,19 @@ int qxl_mmap(struct file *filp, struct vm_area_struct *vma); | |||
433 | 442 | ||
434 | /* qxl image */ | 443 | /* qxl image */ |
435 | 444 | ||
436 | int qxl_image_create(struct qxl_device *qdev, | 445 | int qxl_image_init(struct qxl_device *qdev, |
437 | struct qxl_release *release, | 446 | struct qxl_release *release, |
438 | struct qxl_bo **image_bo, | 447 | struct qxl_drm_image *dimage, |
439 | const uint8_t *data, | 448 | const uint8_t *data, |
440 | int x, int y, int width, int height, | 449 | int x, int y, int width, int height, |
441 | int depth, int stride); | 450 | int depth, int stride); |
451 | int | ||
452 | qxl_image_alloc_objects(struct qxl_device *qdev, | ||
453 | struct qxl_release *release, | ||
454 | struct qxl_drm_image **image_ptr, | ||
455 | int height, int stride); | ||
456 | void qxl_image_free_objects(struct qxl_device *qdev, struct qxl_drm_image *dimage); | ||
457 | |||
442 | void qxl_update_screen(struct qxl_device *qxl); | 458 | void qxl_update_screen(struct qxl_device *qxl); |
443 | 459 | ||
444 | /* qxl io operations (qxl_cmd.c) */ | 460 | /* qxl io operations (qxl_cmd.c) */ |
@@ -459,20 +475,15 @@ int qxl_ring_push(struct qxl_ring *ring, const void *new_elt, bool interruptible | |||
459 | void qxl_io_flush_release(struct qxl_device *qdev); | 475 | void qxl_io_flush_release(struct qxl_device *qdev); |
460 | void qxl_io_flush_surfaces(struct qxl_device *qdev); | 476 | void qxl_io_flush_surfaces(struct qxl_device *qdev); |
461 | 477 | ||
462 | int qxl_release_reserve(struct qxl_device *qdev, | ||
463 | struct qxl_release *release, bool no_wait); | ||
464 | void qxl_release_unreserve(struct qxl_device *qdev, | ||
465 | struct qxl_release *release); | ||
466 | union qxl_release_info *qxl_release_map(struct qxl_device *qdev, | 478 | union qxl_release_info *qxl_release_map(struct qxl_device *qdev, |
467 | struct qxl_release *release); | 479 | struct qxl_release *release); |
468 | void qxl_release_unmap(struct qxl_device *qdev, | 480 | void qxl_release_unmap(struct qxl_device *qdev, |
469 | struct qxl_release *release, | 481 | struct qxl_release *release, |
470 | union qxl_release_info *info); | 482 | union qxl_release_info *info); |
471 | /* | 483 | int qxl_release_list_add(struct qxl_release *release, struct qxl_bo *bo); |
472 | * qxl_bo_add_resource. | 484 | int qxl_release_reserve_list(struct qxl_release *release, bool no_intr); |
473 | * | 485 | void qxl_release_backoff_reserve_list(struct qxl_release *release); |
474 | */ | 486 | void qxl_release_fence_buffer_objects(struct qxl_release *release); |
475 | void qxl_bo_add_resource(struct qxl_bo *main_bo, struct qxl_bo *resource); | ||
476 | 487 | ||
477 | int qxl_alloc_surface_release_reserved(struct qxl_device *qdev, | 488 | int qxl_alloc_surface_release_reserved(struct qxl_device *qdev, |
478 | enum qxl_surface_cmd_type surface_cmd_type, | 489 | enum qxl_surface_cmd_type surface_cmd_type, |
@@ -481,15 +492,16 @@ int qxl_alloc_surface_release_reserved(struct qxl_device *qdev, | |||
481 | int qxl_alloc_release_reserved(struct qxl_device *qdev, unsigned long size, | 492 | int qxl_alloc_release_reserved(struct qxl_device *qdev, unsigned long size, |
482 | int type, struct qxl_release **release, | 493 | int type, struct qxl_release **release, |
483 | struct qxl_bo **rbo); | 494 | struct qxl_bo **rbo); |
484 | int qxl_fence_releaseable(struct qxl_device *qdev, | 495 | |
485 | struct qxl_release *release); | ||
486 | int | 496 | int |
487 | qxl_push_command_ring_release(struct qxl_device *qdev, struct qxl_release *release, | 497 | qxl_push_command_ring_release(struct qxl_device *qdev, struct qxl_release *release, |
488 | uint32_t type, bool interruptible); | 498 | uint32_t type, bool interruptible); |
489 | int | 499 | int |
490 | qxl_push_cursor_ring_release(struct qxl_device *qdev, struct qxl_release *release, | 500 | qxl_push_cursor_ring_release(struct qxl_device *qdev, struct qxl_release *release, |
491 | uint32_t type, bool interruptible); | 501 | uint32_t type, bool interruptible); |
492 | int qxl_alloc_bo_reserved(struct qxl_device *qdev, unsigned long size, | 502 | int qxl_alloc_bo_reserved(struct qxl_device *qdev, |
503 | struct qxl_release *release, | ||
504 | unsigned long size, | ||
493 | struct qxl_bo **_bo); | 505 | struct qxl_bo **_bo); |
494 | /* qxl drawing commands */ | 506 | /* qxl drawing commands */ |
495 | 507 | ||
@@ -510,15 +522,9 @@ void qxl_draw_copyarea(struct qxl_device *qdev, | |||
510 | u32 sx, u32 sy, | 522 | u32 sx, u32 sy, |
511 | u32 dx, u32 dy); | 523 | u32 dx, u32 dy); |
512 | 524 | ||
513 | uint64_t | ||
514 | qxl_release_alloc(struct qxl_device *qdev, int type, | ||
515 | struct qxl_release **ret); | ||
516 | |||
517 | void qxl_release_free(struct qxl_device *qdev, | 525 | void qxl_release_free(struct qxl_device *qdev, |
518 | struct qxl_release *release); | 526 | struct qxl_release *release); |
519 | void qxl_release_add_res(struct qxl_device *qdev, | 527 | |
520 | struct qxl_release *release, | ||
521 | struct qxl_bo *bo); | ||
522 | /* used by qxl_debugfs_release */ | 528 | /* used by qxl_debugfs_release */ |
523 | struct qxl_release *qxl_release_from_id_locked(struct qxl_device *qdev, | 529 | struct qxl_release *qxl_release_from_id_locked(struct qxl_device *qdev, |
524 | uint64_t id); | 530 | uint64_t id); |
@@ -561,7 +567,7 @@ void qxl_surface_evict(struct qxl_device *qdev, struct qxl_bo *surf, bool freein | |||
561 | int qxl_update_surface(struct qxl_device *qdev, struct qxl_bo *surf); | 567 | int qxl_update_surface(struct qxl_device *qdev, struct qxl_bo *surf); |
562 | 568 | ||
563 | /* qxl_fence.c */ | 569 | /* qxl_fence.c */ |
564 | int qxl_fence_add_release(struct qxl_fence *qfence, uint32_t rel_id); | 570 | void qxl_fence_add_release_locked(struct qxl_fence *qfence, uint32_t rel_id); |
565 | int qxl_fence_remove_release(struct qxl_fence *qfence, uint32_t rel_id); | 571 | int qxl_fence_remove_release(struct qxl_fence *qfence, uint32_t rel_id); |
566 | int qxl_fence_init(struct qxl_device *qdev, struct qxl_fence *qfence); | 572 | int qxl_fence_init(struct qxl_device *qdev, struct qxl_fence *qfence); |
567 | void qxl_fence_fini(struct qxl_fence *qfence); | 573 | void qxl_fence_fini(struct qxl_fence *qfence); |
diff --git a/drivers/gpu/drm/qxl/qxl_fb.c b/drivers/gpu/drm/qxl/qxl_fb.c index 76f39d88d684..88722f233430 100644 --- a/drivers/gpu/drm/qxl/qxl_fb.c +++ b/drivers/gpu/drm/qxl/qxl_fb.c | |||
@@ -37,12 +37,29 @@ | |||
37 | 37 | ||
38 | #define QXL_DIRTY_DELAY (HZ / 30) | 38 | #define QXL_DIRTY_DELAY (HZ / 30) |
39 | 39 | ||
40 | #define QXL_FB_OP_FILLRECT 1 | ||
41 | #define QXL_FB_OP_COPYAREA 2 | ||
42 | #define QXL_FB_OP_IMAGEBLIT 3 | ||
43 | |||
44 | struct qxl_fb_op { | ||
45 | struct list_head head; | ||
46 | int op_type; | ||
47 | union { | ||
48 | struct fb_fillrect fr; | ||
49 | struct fb_copyarea ca; | ||
50 | struct fb_image ib; | ||
51 | } op; | ||
52 | void *img_data; | ||
53 | }; | ||
54 | |||
40 | struct qxl_fbdev { | 55 | struct qxl_fbdev { |
41 | struct drm_fb_helper helper; | 56 | struct drm_fb_helper helper; |
42 | struct qxl_framebuffer qfb; | 57 | struct qxl_framebuffer qfb; |
43 | struct list_head fbdev_list; | 58 | struct list_head fbdev_list; |
44 | struct qxl_device *qdev; | 59 | struct qxl_device *qdev; |
45 | 60 | ||
61 | spinlock_t delayed_ops_lock; | ||
62 | struct list_head delayed_ops; | ||
46 | void *shadow; | 63 | void *shadow; |
47 | int size; | 64 | int size; |
48 | 65 | ||
@@ -164,8 +181,69 @@ static struct fb_deferred_io qxl_defio = { | |||
164 | .deferred_io = qxl_deferred_io, | 181 | .deferred_io = qxl_deferred_io, |
165 | }; | 182 | }; |
166 | 183 | ||
167 | static void qxl_fb_fillrect(struct fb_info *info, | 184 | static void qxl_fb_delayed_fillrect(struct qxl_fbdev *qfbdev, |
168 | const struct fb_fillrect *fb_rect) | 185 | const struct fb_fillrect *fb_rect) |
186 | { | ||
187 | struct qxl_fb_op *op; | ||
188 | unsigned long flags; | ||
189 | |||
190 | op = kmalloc(sizeof(struct qxl_fb_op), GFP_ATOMIC | __GFP_NOWARN); | ||
191 | if (!op) | ||
192 | return; | ||
193 | |||
194 | op->op.fr = *fb_rect; | ||
195 | op->img_data = NULL; | ||
196 | op->op_type = QXL_FB_OP_FILLRECT; | ||
197 | |||
198 | spin_lock_irqsave(&qfbdev->delayed_ops_lock, flags); | ||
199 | list_add_tail(&op->head, &qfbdev->delayed_ops); | ||
200 | spin_unlock_irqrestore(&qfbdev->delayed_ops_lock, flags); | ||
201 | } | ||
202 | |||
203 | static void qxl_fb_delayed_copyarea(struct qxl_fbdev *qfbdev, | ||
204 | const struct fb_copyarea *fb_copy) | ||
205 | { | ||
206 | struct qxl_fb_op *op; | ||
207 | unsigned long flags; | ||
208 | |||
209 | op = kmalloc(sizeof(struct qxl_fb_op), GFP_ATOMIC | __GFP_NOWARN); | ||
210 | if (!op) | ||
211 | return; | ||
212 | |||
213 | op->op.ca = *fb_copy; | ||
214 | op->img_data = NULL; | ||
215 | op->op_type = QXL_FB_OP_COPYAREA; | ||
216 | |||
217 | spin_lock_irqsave(&qfbdev->delayed_ops_lock, flags); | ||
218 | list_add_tail(&op->head, &qfbdev->delayed_ops); | ||
219 | spin_unlock_irqrestore(&qfbdev->delayed_ops_lock, flags); | ||
220 | } | ||
221 | |||
222 | static void qxl_fb_delayed_imageblit(struct qxl_fbdev *qfbdev, | ||
223 | const struct fb_image *fb_image) | ||
224 | { | ||
225 | struct qxl_fb_op *op; | ||
226 | unsigned long flags; | ||
227 | uint32_t size = fb_image->width * fb_image->height * (fb_image->depth >= 8 ? fb_image->depth / 8 : 1); | ||
228 | |||
229 | op = kmalloc(sizeof(struct qxl_fb_op) + size, GFP_ATOMIC | __GFP_NOWARN); | ||
230 | if (!op) | ||
231 | return; | ||
232 | |||
233 | op->op.ib = *fb_image; | ||
234 | op->img_data = (void *)(op + 1); | ||
235 | op->op_type = QXL_FB_OP_IMAGEBLIT; | ||
236 | |||
237 | memcpy(op->img_data, fb_image->data, size); | ||
238 | |||
239 | op->op.ib.data = op->img_data; | ||
240 | spin_lock_irqsave(&qfbdev->delayed_ops_lock, flags); | ||
241 | list_add_tail(&op->head, &qfbdev->delayed_ops); | ||
242 | spin_unlock_irqrestore(&qfbdev->delayed_ops_lock, flags); | ||
243 | } | ||
244 | |||
245 | static void qxl_fb_fillrect_internal(struct fb_info *info, | ||
246 | const struct fb_fillrect *fb_rect) | ||
169 | { | 247 | { |
170 | struct qxl_fbdev *qfbdev = info->par; | 248 | struct qxl_fbdev *qfbdev = info->par; |
171 | struct qxl_device *qdev = qfbdev->qdev; | 249 | struct qxl_device *qdev = qfbdev->qdev; |
@@ -203,17 +281,28 @@ static void qxl_fb_fillrect(struct fb_info *info, | |||
203 | qxl_draw_fill_rec.rect = rect; | 281 | qxl_draw_fill_rec.rect = rect; |
204 | qxl_draw_fill_rec.color = color; | 282 | qxl_draw_fill_rec.color = color; |
205 | qxl_draw_fill_rec.rop = rop; | 283 | qxl_draw_fill_rec.rop = rop; |
284 | |||
285 | qxl_draw_fill(&qxl_draw_fill_rec); | ||
286 | } | ||
287 | |||
288 | static void qxl_fb_fillrect(struct fb_info *info, | ||
289 | const struct fb_fillrect *fb_rect) | ||
290 | { | ||
291 | struct qxl_fbdev *qfbdev = info->par; | ||
292 | struct qxl_device *qdev = qfbdev->qdev; | ||
293 | |||
206 | if (!drm_can_sleep()) { | 294 | if (!drm_can_sleep()) { |
207 | qxl_io_log(qdev, | 295 | qxl_fb_delayed_fillrect(qfbdev, fb_rect); |
208 | "%s: TODO use RCU, mysterious locks with spin_lock\n", | 296 | schedule_work(&qdev->fb_work); |
209 | __func__); | ||
210 | return; | 297 | return; |
211 | } | 298 | } |
212 | qxl_draw_fill(&qxl_draw_fill_rec); | 299 | /* make sure any previous work is done */ |
300 | flush_work(&qdev->fb_work); | ||
301 | qxl_fb_fillrect_internal(info, fb_rect); | ||
213 | } | 302 | } |
214 | 303 | ||
215 | static void qxl_fb_copyarea(struct fb_info *info, | 304 | static void qxl_fb_copyarea_internal(struct fb_info *info, |
216 | const struct fb_copyarea *region) | 305 | const struct fb_copyarea *region) |
217 | { | 306 | { |
218 | struct qxl_fbdev *qfbdev = info->par; | 307 | struct qxl_fbdev *qfbdev = info->par; |
219 | 308 | ||
@@ -223,37 +312,89 @@ static void qxl_fb_copyarea(struct fb_info *info, | |||
223 | region->dx, region->dy); | 312 | region->dx, region->dy); |
224 | } | 313 | } |
225 | 314 | ||
315 | static void qxl_fb_copyarea(struct fb_info *info, | ||
316 | const struct fb_copyarea *region) | ||
317 | { | ||
318 | struct qxl_fbdev *qfbdev = info->par; | ||
319 | struct qxl_device *qdev = qfbdev->qdev; | ||
320 | |||
321 | if (!drm_can_sleep()) { | ||
322 | qxl_fb_delayed_copyarea(qfbdev, region); | ||
323 | schedule_work(&qdev->fb_work); | ||
324 | return; | ||
325 | } | ||
326 | /* make sure any previous work is done */ | ||
327 | flush_work(&qdev->fb_work); | ||
328 | qxl_fb_copyarea_internal(info, region); | ||
329 | } | ||
330 | |||
226 | static void qxl_fb_imageblit_safe(struct qxl_fb_image *qxl_fb_image) | 331 | static void qxl_fb_imageblit_safe(struct qxl_fb_image *qxl_fb_image) |
227 | { | 332 | { |
228 | qxl_draw_opaque_fb(qxl_fb_image, 0); | 333 | qxl_draw_opaque_fb(qxl_fb_image, 0); |
229 | } | 334 | } |
230 | 335 | ||
336 | static void qxl_fb_imageblit_internal(struct fb_info *info, | ||
337 | const struct fb_image *image) | ||
338 | { | ||
339 | struct qxl_fbdev *qfbdev = info->par; | ||
340 | struct qxl_fb_image qxl_fb_image; | ||
341 | |||
342 | /* ensure proper order rendering operations - TODO: must do this | ||
343 | * for everything. */ | ||
344 | qxl_fb_image_init(&qxl_fb_image, qfbdev->qdev, info, image); | ||
345 | qxl_fb_imageblit_safe(&qxl_fb_image); | ||
346 | } | ||
347 | |||
231 | static void qxl_fb_imageblit(struct fb_info *info, | 348 | static void qxl_fb_imageblit(struct fb_info *info, |
232 | const struct fb_image *image) | 349 | const struct fb_image *image) |
233 | { | 350 | { |
234 | struct qxl_fbdev *qfbdev = info->par; | 351 | struct qxl_fbdev *qfbdev = info->par; |
235 | struct qxl_device *qdev = qfbdev->qdev; | 352 | struct qxl_device *qdev = qfbdev->qdev; |
236 | struct qxl_fb_image qxl_fb_image; | ||
237 | 353 | ||
238 | if (!drm_can_sleep()) { | 354 | if (!drm_can_sleep()) { |
239 | /* we cannot do any ttm_bo allocation since that will fail on | 355 | qxl_fb_delayed_imageblit(qfbdev, image); |
240 | * ioremap_wc..__get_vm_area_node, so queue the work item | 356 | schedule_work(&qdev->fb_work); |
241 | * instead This can happen from printk inside an interrupt | ||
242 | * context, i.e.: smp_apic_timer_interrupt..check_cpu_stall */ | ||
243 | qxl_io_log(qdev, | ||
244 | "%s: TODO use RCU, mysterious locks with spin_lock\n", | ||
245 | __func__); | ||
246 | return; | 357 | return; |
247 | } | 358 | } |
359 | /* make sure any previous work is done */ | ||
360 | flush_work(&qdev->fb_work); | ||
361 | qxl_fb_imageblit_internal(info, image); | ||
362 | } | ||
248 | 363 | ||
249 | /* ensure proper order of rendering operations - TODO: must do this | 364 | static void qxl_fb_work(struct work_struct *work) |
250 | * for everything. */ | 365 | { |
251 | qxl_fb_image_init(&qxl_fb_image, qfbdev->qdev, info, image); | 366 | struct qxl_device *qdev = container_of(work, struct qxl_device, fb_work); |
252 | qxl_fb_imageblit_safe(&qxl_fb_image); | 367 | unsigned long flags; |
368 | struct qxl_fb_op *entry, *tmp; | ||
369 | struct qxl_fbdev *qfbdev = qdev->mode_info.qfbdev; | ||
370 | |||
371 | /* since the irq context just adds entries to the end of the | ||
372 | list dropping the lock should be fine, as entry isn't modified | ||
373 | in the operation code */ | ||
374 | spin_lock_irqsave(&qfbdev->delayed_ops_lock, flags); | ||
375 | list_for_each_entry_safe(entry, tmp, &qfbdev->delayed_ops, head) { | ||
376 | spin_unlock_irqrestore(&qfbdev->delayed_ops_lock, flags); | ||
377 | switch (entry->op_type) { | ||
378 | case QXL_FB_OP_FILLRECT: | ||
379 | qxl_fb_fillrect_internal(qfbdev->helper.fbdev, &entry->op.fr); | ||
380 | break; | ||
381 | case QXL_FB_OP_COPYAREA: | ||
382 | qxl_fb_copyarea_internal(qfbdev->helper.fbdev, &entry->op.ca); | ||
383 | break; | ||
384 | case QXL_FB_OP_IMAGEBLIT: | ||
385 | qxl_fb_imageblit_internal(qfbdev->helper.fbdev, &entry->op.ib); | ||
386 | break; | ||
387 | } | ||
388 | spin_lock_irqsave(&qfbdev->delayed_ops_lock, flags); | ||
389 | list_del(&entry->head); | ||
390 | kfree(entry); | ||
391 | } | ||
392 | spin_unlock_irqrestore(&qfbdev->delayed_ops_lock, flags); | ||
253 | } | 393 | } |
254 | 394 | ||
255 | int qxl_fb_init(struct qxl_device *qdev) | 395 | int qxl_fb_init(struct qxl_device *qdev) |
256 | { | 396 | { |
397 | INIT_WORK(&qdev->fb_work, qxl_fb_work); | ||
257 | return 0; | 398 | return 0; |
258 | } | 399 | } |
259 | 400 | ||
@@ -536,7 +677,8 @@ int qxl_fbdev_init(struct qxl_device *qdev) | |||
536 | qfbdev->qdev = qdev; | 677 | qfbdev->qdev = qdev; |
537 | qdev->mode_info.qfbdev = qfbdev; | 678 | qdev->mode_info.qfbdev = qfbdev; |
538 | qfbdev->helper.funcs = &qxl_fb_helper_funcs; | 679 | qfbdev->helper.funcs = &qxl_fb_helper_funcs; |
539 | 680 | spin_lock_init(&qfbdev->delayed_ops_lock); | |
681 | INIT_LIST_HEAD(&qfbdev->delayed_ops); | ||
540 | ret = drm_fb_helper_init(qdev->ddev, &qfbdev->helper, | 682 | ret = drm_fb_helper_init(qdev->ddev, &qfbdev->helper, |
541 | qxl_num_crtc /* num_crtc - QXL supports just 1 */, | 683 | qxl_num_crtc /* num_crtc - QXL supports just 1 */, |
542 | QXLFB_CONN_LIMIT); | 684 | QXLFB_CONN_LIMIT); |
diff --git a/drivers/gpu/drm/qxl/qxl_fence.c b/drivers/gpu/drm/qxl/qxl_fence.c index 63c6715ad385..ae59e91cfb9a 100644 --- a/drivers/gpu/drm/qxl/qxl_fence.c +++ b/drivers/gpu/drm/qxl/qxl_fence.c | |||
@@ -49,17 +49,11 @@ | |||
49 | 49 | ||
50 | For some reason every so often qxl hw fails to release, things go wrong. | 50 | For some reason every so often qxl hw fails to release, things go wrong. |
51 | */ | 51 | */ |
52 | 52 | /* must be called with the fence lock held */ | |
53 | 53 | void qxl_fence_add_release_locked(struct qxl_fence *qfence, uint32_t rel_id) | |
54 | int qxl_fence_add_release(struct qxl_fence *qfence, uint32_t rel_id) | ||
55 | { | 54 | { |
56 | struct qxl_bo *bo = container_of(qfence, struct qxl_bo, fence); | ||
57 | |||
58 | spin_lock(&bo->tbo.bdev->fence_lock); | ||
59 | radix_tree_insert(&qfence->tree, rel_id, qfence); | 55 | radix_tree_insert(&qfence->tree, rel_id, qfence); |
60 | qfence->num_active_releases++; | 56 | qfence->num_active_releases++; |
61 | spin_unlock(&bo->tbo.bdev->fence_lock); | ||
62 | return 0; | ||
63 | } | 57 | } |
64 | 58 | ||
65 | int qxl_fence_remove_release(struct qxl_fence *qfence, uint32_t rel_id) | 59 | int qxl_fence_remove_release(struct qxl_fence *qfence, uint32_t rel_id) |
diff --git a/drivers/gpu/drm/qxl/qxl_gem.c b/drivers/gpu/drm/qxl/qxl_gem.c index a235693aabba..25e1777fb0a2 100644 --- a/drivers/gpu/drm/qxl/qxl_gem.c +++ b/drivers/gpu/drm/qxl/qxl_gem.c | |||
@@ -55,7 +55,7 @@ int qxl_gem_object_create(struct qxl_device *qdev, int size, | |||
55 | /* At least align on page size */ | 55 | /* At least align on page size */ |
56 | if (alignment < PAGE_SIZE) | 56 | if (alignment < PAGE_SIZE) |
57 | alignment = PAGE_SIZE; | 57 | alignment = PAGE_SIZE; |
58 | r = qxl_bo_create(qdev, size, kernel, initial_domain, surf, &qbo); | 58 | r = qxl_bo_create(qdev, size, kernel, false, initial_domain, surf, &qbo); |
59 | if (r) { | 59 | if (r) { |
60 | if (r != -ERESTARTSYS) | 60 | if (r != -ERESTARTSYS) |
61 | DRM_ERROR( | 61 | DRM_ERROR( |
diff --git a/drivers/gpu/drm/qxl/qxl_image.c b/drivers/gpu/drm/qxl/qxl_image.c index cf856206996b..7fbcc35e8ad3 100644 --- a/drivers/gpu/drm/qxl/qxl_image.c +++ b/drivers/gpu/drm/qxl/qxl_image.c | |||
@@ -30,31 +30,100 @@ | |||
30 | #include "qxl_object.h" | 30 | #include "qxl_object.h" |
31 | 31 | ||
32 | static int | 32 | static int |
33 | qxl_image_create_helper(struct qxl_device *qdev, | 33 | qxl_allocate_chunk(struct qxl_device *qdev, |
34 | struct qxl_release *release, | ||
35 | struct qxl_drm_image *image, | ||
36 | unsigned int chunk_size) | ||
37 | { | ||
38 | struct qxl_drm_chunk *chunk; | ||
39 | int ret; | ||
40 | |||
41 | chunk = kmalloc(sizeof(struct qxl_drm_chunk), GFP_KERNEL); | ||
42 | if (!chunk) | ||
43 | return -ENOMEM; | ||
44 | |||
45 | ret = qxl_alloc_bo_reserved(qdev, release, chunk_size, &chunk->bo); | ||
46 | if (ret) { | ||
47 | kfree(chunk); | ||
48 | return ret; | ||
49 | } | ||
50 | |||
51 | list_add_tail(&chunk->head, &image->chunk_list); | ||
52 | return 0; | ||
53 | } | ||
54 | |||
55 | int | ||
56 | qxl_image_alloc_objects(struct qxl_device *qdev, | ||
34 | struct qxl_release *release, | 57 | struct qxl_release *release, |
35 | struct qxl_bo **image_bo, | 58 | struct qxl_drm_image **image_ptr, |
36 | const uint8_t *data, | 59 | int height, int stride) |
37 | int width, int height, | 60 | { |
38 | int depth, unsigned int hash, | 61 | struct qxl_drm_image *image; |
39 | int stride) | 62 | int ret; |
63 | |||
64 | image = kmalloc(sizeof(struct qxl_drm_image), GFP_KERNEL); | ||
65 | if (!image) | ||
66 | return -ENOMEM; | ||
67 | |||
68 | INIT_LIST_HEAD(&image->chunk_list); | ||
69 | |||
70 | ret = qxl_alloc_bo_reserved(qdev, release, sizeof(struct qxl_image), &image->bo); | ||
71 | if (ret) { | ||
72 | kfree(image); | ||
73 | return ret; | ||
74 | } | ||
75 | |||
76 | ret = qxl_allocate_chunk(qdev, release, image, sizeof(struct qxl_data_chunk) + stride * height); | ||
77 | if (ret) { | ||
78 | qxl_bo_unref(&image->bo); | ||
79 | kfree(image); | ||
80 | return ret; | ||
81 | } | ||
82 | *image_ptr = image; | ||
83 | return 0; | ||
84 | } | ||
85 | |||
86 | void qxl_image_free_objects(struct qxl_device *qdev, struct qxl_drm_image *dimage) | ||
40 | { | 87 | { |
88 | struct qxl_drm_chunk *chunk, *tmp; | ||
89 | |||
90 | list_for_each_entry_safe(chunk, tmp, &dimage->chunk_list, head) { | ||
91 | qxl_bo_unref(&chunk->bo); | ||
92 | kfree(chunk); | ||
93 | } | ||
94 | |||
95 | qxl_bo_unref(&dimage->bo); | ||
96 | kfree(dimage); | ||
97 | } | ||
98 | |||
99 | static int | ||
100 | qxl_image_init_helper(struct qxl_device *qdev, | ||
101 | struct qxl_release *release, | ||
102 | struct qxl_drm_image *dimage, | ||
103 | const uint8_t *data, | ||
104 | int width, int height, | ||
105 | int depth, unsigned int hash, | ||
106 | int stride) | ||
107 | { | ||
108 | struct qxl_drm_chunk *drv_chunk; | ||
41 | struct qxl_image *image; | 109 | struct qxl_image *image; |
42 | struct qxl_data_chunk *chunk; | 110 | struct qxl_data_chunk *chunk; |
43 | int i; | 111 | int i; |
44 | int chunk_stride; | 112 | int chunk_stride; |
45 | int linesize = width * depth / 8; | 113 | int linesize = width * depth / 8; |
46 | struct qxl_bo *chunk_bo; | 114 | struct qxl_bo *chunk_bo, *image_bo; |
47 | int ret; | ||
48 | void *ptr; | 115 | void *ptr; |
49 | /* Chunk */ | 116 | /* Chunk */ |
50 | /* FIXME: Check integer overflow */ | 117 | /* FIXME: Check integer overflow */ |
51 | /* TODO: variable number of chunks */ | 118 | /* TODO: variable number of chunks */ |
119 | |||
120 | drv_chunk = list_first_entry(&dimage->chunk_list, struct qxl_drm_chunk, head); | ||
121 | |||
122 | chunk_bo = drv_chunk->bo; | ||
52 | chunk_stride = stride; /* TODO: should use linesize, but it renders | 123 | chunk_stride = stride; /* TODO: should use linesize, but it renders |
53 | wrong (check the bitmaps are sent correctly | 124 | wrong (check the bitmaps are sent correctly |
54 | first) */ | 125 | first) */ |
55 | ret = qxl_alloc_bo_reserved(qdev, sizeof(*chunk) + height * chunk_stride, | 126 | |
56 | &chunk_bo); | ||
57 | |||
58 | ptr = qxl_bo_kmap_atomic_page(qdev, chunk_bo, 0); | 127 | ptr = qxl_bo_kmap_atomic_page(qdev, chunk_bo, 0); |
59 | chunk = ptr; | 128 | chunk = ptr; |
60 | chunk->data_size = height * chunk_stride; | 129 | chunk->data_size = height * chunk_stride; |
@@ -102,7 +171,6 @@ qxl_image_create_helper(struct qxl_device *qdev, | |||
102 | while (remain > 0) { | 171 | while (remain > 0) { |
103 | page_base = out_offset & PAGE_MASK; | 172 | page_base = out_offset & PAGE_MASK; |
104 | page_offset = offset_in_page(out_offset); | 173 | page_offset = offset_in_page(out_offset); |
105 | |||
106 | size = min((int)(PAGE_SIZE - page_offset), remain); | 174 | size = min((int)(PAGE_SIZE - page_offset), remain); |
107 | 175 | ||
108 | ptr = qxl_bo_kmap_atomic_page(qdev, chunk_bo, page_base); | 176 | ptr = qxl_bo_kmap_atomic_page(qdev, chunk_bo, page_base); |
@@ -116,14 +184,10 @@ qxl_image_create_helper(struct qxl_device *qdev, | |||
116 | } | 184 | } |
117 | } | 185 | } |
118 | } | 186 | } |
119 | |||
120 | |||
121 | qxl_bo_kunmap(chunk_bo); | 187 | qxl_bo_kunmap(chunk_bo); |
122 | 188 | ||
123 | /* Image */ | 189 | image_bo = dimage->bo; |
124 | ret = qxl_alloc_bo_reserved(qdev, sizeof(*image), image_bo); | 190 | ptr = qxl_bo_kmap_atomic_page(qdev, image_bo, 0); |
125 | |||
126 | ptr = qxl_bo_kmap_atomic_page(qdev, *image_bo, 0); | ||
127 | image = ptr; | 191 | image = ptr; |
128 | 192 | ||
129 | image->descriptor.id = 0; | 193 | image->descriptor.id = 0; |
@@ -154,23 +218,20 @@ qxl_image_create_helper(struct qxl_device *qdev, | |||
154 | image->u.bitmap.stride = chunk_stride; | 218 | image->u.bitmap.stride = chunk_stride; |
155 | image->u.bitmap.palette = 0; | 219 | image->u.bitmap.palette = 0; |
156 | image->u.bitmap.data = qxl_bo_physical_address(qdev, chunk_bo, 0); | 220 | image->u.bitmap.data = qxl_bo_physical_address(qdev, chunk_bo, 0); |
157 | qxl_release_add_res(qdev, release, chunk_bo); | ||
158 | qxl_bo_unreserve(chunk_bo); | ||
159 | qxl_bo_unref(&chunk_bo); | ||
160 | 221 | ||
161 | qxl_bo_kunmap_atomic_page(qdev, *image_bo, ptr); | 222 | qxl_bo_kunmap_atomic_page(qdev, image_bo, ptr); |
162 | 223 | ||
163 | return 0; | 224 | return 0; |
164 | } | 225 | } |
165 | 226 | ||
166 | int qxl_image_create(struct qxl_device *qdev, | 227 | int qxl_image_init(struct qxl_device *qdev, |
167 | struct qxl_release *release, | 228 | struct qxl_release *release, |
168 | struct qxl_bo **image_bo, | 229 | struct qxl_drm_image *dimage, |
169 | const uint8_t *data, | 230 | const uint8_t *data, |
170 | int x, int y, int width, int height, | 231 | int x, int y, int width, int height, |
171 | int depth, int stride) | 232 | int depth, int stride) |
172 | { | 233 | { |
173 | data += y * stride + x * (depth / 8); | 234 | data += y * stride + x * (depth / 8); |
174 | return qxl_image_create_helper(qdev, release, image_bo, data, | 235 | return qxl_image_init_helper(qdev, release, dimage, data, |
175 | width, height, depth, 0, stride); | 236 | width, height, depth, 0, stride); |
176 | } | 237 | } |
diff --git a/drivers/gpu/drm/qxl/qxl_ioctl.c b/drivers/gpu/drm/qxl/qxl_ioctl.c index 27f45e49250d..6de33563d6f1 100644 --- a/drivers/gpu/drm/qxl/qxl_ioctl.c +++ b/drivers/gpu/drm/qxl/qxl_ioctl.c | |||
@@ -68,55 +68,60 @@ static int qxl_map_ioctl(struct drm_device *dev, void *data, | |||
68 | &qxl_map->offset); | 68 | &qxl_map->offset); |
69 | } | 69 | } |
70 | 70 | ||
71 | struct qxl_reloc_info { | ||
72 | int type; | ||
73 | struct qxl_bo *dst_bo; | ||
74 | uint32_t dst_offset; | ||
75 | struct qxl_bo *src_bo; | ||
76 | int src_offset; | ||
77 | }; | ||
78 | |||
71 | /* | 79 | /* |
72 | * dst must be validated, i.e. whole bo on vram/surfacesram (right now all bo's | 80 | * dst must be validated, i.e. whole bo on vram/surfacesram (right now all bo's |
73 | * are on vram). | 81 | * are on vram). |
74 | * *(dst + dst_off) = qxl_bo_physical_address(src, src_off) | 82 | * *(dst + dst_off) = qxl_bo_physical_address(src, src_off) |
75 | */ | 83 | */ |
76 | static void | 84 | static void |
77 | apply_reloc(struct qxl_device *qdev, struct qxl_bo *dst, uint64_t dst_off, | 85 | apply_reloc(struct qxl_device *qdev, struct qxl_reloc_info *info) |
78 | struct qxl_bo *src, uint64_t src_off) | ||
79 | { | 86 | { |
80 | void *reloc_page; | 87 | void *reloc_page; |
81 | 88 | reloc_page = qxl_bo_kmap_atomic_page(qdev, info->dst_bo, info->dst_offset & PAGE_MASK); | |
82 | reloc_page = qxl_bo_kmap_atomic_page(qdev, dst, dst_off & PAGE_MASK); | 89 | *(uint64_t *)(reloc_page + (info->dst_offset & ~PAGE_MASK)) = qxl_bo_physical_address(qdev, |
83 | *(uint64_t *)(reloc_page + (dst_off & ~PAGE_MASK)) = qxl_bo_physical_address(qdev, | 90 | info->src_bo, |
84 | src, src_off); | 91 | info->src_offset); |
85 | qxl_bo_kunmap_atomic_page(qdev, dst, reloc_page); | 92 | qxl_bo_kunmap_atomic_page(qdev, info->dst_bo, reloc_page); |
86 | } | 93 | } |
87 | 94 | ||
88 | static void | 95 | static void |
89 | apply_surf_reloc(struct qxl_device *qdev, struct qxl_bo *dst, uint64_t dst_off, | 96 | apply_surf_reloc(struct qxl_device *qdev, struct qxl_reloc_info *info) |
90 | struct qxl_bo *src) | ||
91 | { | 97 | { |
92 | uint32_t id = 0; | 98 | uint32_t id = 0; |
93 | void *reloc_page; | 99 | void *reloc_page; |
94 | 100 | ||
95 | if (src && !src->is_primary) | 101 | if (info->src_bo && !info->src_bo->is_primary) |
96 | id = src->surface_id; | 102 | id = info->src_bo->surface_id; |
97 | 103 | ||
98 | reloc_page = qxl_bo_kmap_atomic_page(qdev, dst, dst_off & PAGE_MASK); | 104 | reloc_page = qxl_bo_kmap_atomic_page(qdev, info->dst_bo, info->dst_offset & PAGE_MASK); |
99 | *(uint32_t *)(reloc_page + (dst_off & ~PAGE_MASK)) = id; | 105 | *(uint32_t *)(reloc_page + (info->dst_offset & ~PAGE_MASK)) = id; |
100 | qxl_bo_kunmap_atomic_page(qdev, dst, reloc_page); | 106 | qxl_bo_kunmap_atomic_page(qdev, info->dst_bo, reloc_page); |
101 | } | 107 | } |
102 | 108 | ||
103 | /* return holding the reference to this object */ | 109 | /* return holding the reference to this object */ |
104 | static struct qxl_bo *qxlhw_handle_to_bo(struct qxl_device *qdev, | 110 | static struct qxl_bo *qxlhw_handle_to_bo(struct qxl_device *qdev, |
105 | struct drm_file *file_priv, uint64_t handle, | 111 | struct drm_file *file_priv, uint64_t handle, |
106 | struct qxl_reloc_list *reloc_list) | 112 | struct qxl_release *release) |
107 | { | 113 | { |
108 | struct drm_gem_object *gobj; | 114 | struct drm_gem_object *gobj; |
109 | struct qxl_bo *qobj; | 115 | struct qxl_bo *qobj; |
110 | int ret; | 116 | int ret; |
111 | 117 | ||
112 | gobj = drm_gem_object_lookup(qdev->ddev, file_priv, handle); | 118 | gobj = drm_gem_object_lookup(qdev->ddev, file_priv, handle); |
113 | if (!gobj) { | 119 | if (!gobj) |
114 | DRM_ERROR("bad bo handle %lld\n", handle); | ||
115 | return NULL; | 120 | return NULL; |
116 | } | 121 | |
117 | qobj = gem_to_qxl_bo(gobj); | 122 | qobj = gem_to_qxl_bo(gobj); |
118 | 123 | ||
119 | ret = qxl_bo_list_add(reloc_list, qobj); | 124 | ret = qxl_release_list_add(release, qobj); |
120 | if (ret) | 125 | if (ret) |
121 | return NULL; | 126 | return NULL; |
122 | 127 | ||
@@ -129,151 +134,177 @@ static struct qxl_bo *qxlhw_handle_to_bo(struct qxl_device *qdev, | |||
129 | * However, the command as passed from user space must *not* contain the initial | 134 | * However, the command as passed from user space must *not* contain the initial |
130 | * QXLReleaseInfo struct (first XXX bytes) | 135 | * QXLReleaseInfo struct (first XXX bytes) |
131 | */ | 136 | */ |
132 | static int qxl_execbuffer_ioctl(struct drm_device *dev, void *data, | 137 | static int qxl_process_single_command(struct qxl_device *qdev, |
133 | struct drm_file *file_priv) | 138 | struct drm_qxl_command *cmd, |
139 | struct drm_file *file_priv) | ||
134 | { | 140 | { |
135 | struct qxl_device *qdev = dev->dev_private; | 141 | struct qxl_reloc_info *reloc_info; |
136 | struct drm_qxl_execbuffer *execbuffer = data; | 142 | int release_type; |
137 | struct drm_qxl_command user_cmd; | 143 | struct qxl_release *release; |
138 | int cmd_num; | 144 | struct qxl_bo *cmd_bo; |
139 | struct qxl_bo *reloc_src_bo; | ||
140 | struct qxl_bo *reloc_dst_bo; | ||
141 | struct drm_qxl_reloc reloc; | ||
142 | void *fb_cmd; | 145 | void *fb_cmd; |
143 | int i, ret; | 146 | int i, j, ret, num_relocs; |
144 | struct qxl_reloc_list reloc_list; | ||
145 | int unwritten; | 147 | int unwritten; |
146 | uint32_t reloc_dst_offset; | ||
147 | INIT_LIST_HEAD(&reloc_list.bos); | ||
148 | 148 | ||
149 | for (cmd_num = 0; cmd_num < execbuffer->commands_num; ++cmd_num) { | 149 | switch (cmd->type) { |
150 | struct qxl_release *release; | 150 | case QXL_CMD_DRAW: |
151 | struct qxl_bo *cmd_bo; | 151 | release_type = QXL_RELEASE_DRAWABLE; |
152 | int release_type; | 152 | break; |
153 | struct drm_qxl_command *commands = | 153 | case QXL_CMD_SURFACE: |
154 | (struct drm_qxl_command *)(uintptr_t)execbuffer->commands; | 154 | case QXL_CMD_CURSOR: |
155 | default: | ||
156 | DRM_DEBUG("Only draw commands in execbuffers\n"); | ||
157 | return -EINVAL; | ||
158 | break; | ||
159 | } | ||
155 | 160 | ||
156 | if (DRM_COPY_FROM_USER(&user_cmd, &commands[cmd_num], | 161 | if (cmd->command_size > PAGE_SIZE - sizeof(union qxl_release_info)) |
157 | sizeof(user_cmd))) | 162 | return -EINVAL; |
158 | return -EFAULT; | ||
159 | switch (user_cmd.type) { | ||
160 | case QXL_CMD_DRAW: | ||
161 | release_type = QXL_RELEASE_DRAWABLE; | ||
162 | break; | ||
163 | case QXL_CMD_SURFACE: | ||
164 | case QXL_CMD_CURSOR: | ||
165 | default: | ||
166 | DRM_DEBUG("Only draw commands in execbuffers\n"); | ||
167 | return -EINVAL; | ||
168 | break; | ||
169 | } | ||
170 | 163 | ||
171 | if (user_cmd.command_size > PAGE_SIZE - sizeof(union qxl_release_info)) | 164 | if (!access_ok(VERIFY_READ, |
172 | return -EINVAL; | 165 | (void *)(unsigned long)cmd->command, |
166 | cmd->command_size)) | ||
167 | return -EFAULT; | ||
173 | 168 | ||
174 | if (!access_ok(VERIFY_READ, | 169 | reloc_info = kmalloc(sizeof(struct qxl_reloc_info) * cmd->relocs_num, GFP_KERNEL); |
175 | (void *)(unsigned long)user_cmd.command, | 170 | if (!reloc_info) |
176 | user_cmd.command_size)) | 171 | return -ENOMEM; |
177 | return -EFAULT; | ||
178 | 172 | ||
179 | ret = qxl_alloc_release_reserved(qdev, | 173 | ret = qxl_alloc_release_reserved(qdev, |
180 | sizeof(union qxl_release_info) + | 174 | sizeof(union qxl_release_info) + |
181 | user_cmd.command_size, | 175 | cmd->command_size, |
182 | release_type, | 176 | release_type, |
183 | &release, | 177 | &release, |
184 | &cmd_bo); | 178 | &cmd_bo); |
185 | if (ret) | 179 | if (ret) |
186 | return ret; | 180 | goto out_free_reloc; |
187 | 181 | ||
188 | /* TODO copy slow path code from i915 */ | 182 | /* TODO copy slow path code from i915 */ |
189 | fb_cmd = qxl_bo_kmap_atomic_page(qdev, cmd_bo, (release->release_offset & PAGE_SIZE)); | 183 | fb_cmd = qxl_bo_kmap_atomic_page(qdev, cmd_bo, (release->release_offset & PAGE_SIZE)); |
190 | unwritten = __copy_from_user_inatomic_nocache(fb_cmd + sizeof(union qxl_release_info) + (release->release_offset & ~PAGE_SIZE), (void *)(unsigned long)user_cmd.command, user_cmd.command_size); | 184 | unwritten = __copy_from_user_inatomic_nocache(fb_cmd + sizeof(union qxl_release_info) + (release->release_offset & ~PAGE_SIZE), (void *)(unsigned long)cmd->command, cmd->command_size); |
191 | 185 | ||
192 | { | 186 | { |
193 | struct qxl_drawable *draw = fb_cmd; | 187 | struct qxl_drawable *draw = fb_cmd; |
188 | draw->mm_time = qdev->rom->mm_clock; | ||
189 | } | ||
194 | 190 | ||
195 | draw->mm_time = qdev->rom->mm_clock; | 191 | qxl_bo_kunmap_atomic_page(qdev, cmd_bo, fb_cmd); |
196 | } | 192 | if (unwritten) { |
197 | qxl_bo_kunmap_atomic_page(qdev, cmd_bo, fb_cmd); | 193 | DRM_ERROR("got unwritten %d\n", unwritten); |
198 | if (unwritten) { | 194 | ret = -EFAULT; |
199 | DRM_ERROR("got unwritten %d\n", unwritten); | 195 | goto out_free_release; |
200 | qxl_release_unreserve(qdev, release); | 196 | } |
201 | qxl_release_free(qdev, release); | 197 | |
202 | return -EFAULT; | 198 | /* fill out reloc info structs */ |
199 | num_relocs = 0; | ||
200 | for (i = 0; i < cmd->relocs_num; ++i) { | ||
201 | struct drm_qxl_reloc reloc; | ||
202 | |||
203 | if (DRM_COPY_FROM_USER(&reloc, | ||
204 | &((struct drm_qxl_reloc *)(uintptr_t)cmd->relocs)[i], | ||
205 | sizeof(reloc))) { | ||
206 | ret = -EFAULT; | ||
207 | goto out_free_bos; | ||
203 | } | 208 | } |
204 | 209 | ||
205 | for (i = 0 ; i < user_cmd.relocs_num; ++i) { | 210 | /* add the bos to the list of bos to validate - |
206 | if (DRM_COPY_FROM_USER(&reloc, | 211 | need to validate first then process relocs? */ |
207 | &((struct drm_qxl_reloc *)(uintptr_t)user_cmd.relocs)[i], | 212 | if (reloc.reloc_type != QXL_RELOC_TYPE_BO && reloc.reloc_type != QXL_RELOC_TYPE_SURF) { |
208 | sizeof(reloc))) { | 213 | DRM_DEBUG("unknown reloc type %d\n", reloc_info[i].type); |
209 | qxl_bo_list_unreserve(&reloc_list, true); | ||
210 | qxl_release_unreserve(qdev, release); | ||
211 | qxl_release_free(qdev, release); | ||
212 | return -EFAULT; | ||
213 | } | ||
214 | 214 | ||
215 | /* add the bos to the list of bos to validate - | 215 | ret = -EINVAL; |
216 | need to validate first then process relocs? */ | 216 | goto out_free_bos; |
217 | if (reloc.dst_handle) { | 217 | } |
218 | reloc_dst_bo = qxlhw_handle_to_bo(qdev, file_priv, | 218 | reloc_info[i].type = reloc.reloc_type; |
219 | reloc.dst_handle, &reloc_list); | 219 | |
220 | if (!reloc_dst_bo) { | 220 | if (reloc.dst_handle) { |
221 | qxl_bo_list_unreserve(&reloc_list, true); | 221 | reloc_info[i].dst_bo = qxlhw_handle_to_bo(qdev, file_priv, |
222 | qxl_release_unreserve(qdev, release); | 222 | reloc.dst_handle, release); |
223 | qxl_release_free(qdev, release); | 223 | if (!reloc_info[i].dst_bo) { |
224 | return -EINVAL; | 224 | ret = -EINVAL; |
225 | } | 225 | reloc_info[i].src_bo = NULL; |
226 | reloc_dst_offset = 0; | 226 | goto out_free_bos; |
227 | } else { | ||
228 | reloc_dst_bo = cmd_bo; | ||
229 | reloc_dst_offset = release->release_offset; | ||
230 | } | 227 | } |
231 | 228 | reloc_info[i].dst_offset = reloc.dst_offset; | |
232 | /* reserve and validate the reloc dst bo */ | 229 | } else { |
233 | if (reloc.reloc_type == QXL_RELOC_TYPE_BO || reloc.src_handle > 0) { | 230 | reloc_info[i].dst_bo = cmd_bo; |
234 | reloc_src_bo = | 231 | reloc_info[i].dst_offset = reloc.dst_offset + release->release_offset; |
235 | qxlhw_handle_to_bo(qdev, file_priv, | 232 | } |
236 | reloc.src_handle, &reloc_list); | 233 | num_relocs++; |
237 | if (!reloc_src_bo) { | 234 | |
238 | if (reloc_dst_bo != cmd_bo) | 235 | /* reserve and validate the reloc dst bo */ |
239 | drm_gem_object_unreference_unlocked(&reloc_dst_bo->gem_base); | 236 | if (reloc.reloc_type == QXL_RELOC_TYPE_BO || reloc.src_handle > 0) { |
240 | qxl_bo_list_unreserve(&reloc_list, true); | 237 | reloc_info[i].src_bo = |
241 | qxl_release_unreserve(qdev, release); | 238 | qxlhw_handle_to_bo(qdev, file_priv, |
242 | qxl_release_free(qdev, release); | 239 | reloc.src_handle, release); |
243 | return -EINVAL; | 240 | if (!reloc_info[i].src_bo) { |
244 | } | 241 | if (reloc_info[i].dst_bo != cmd_bo) |
245 | } else | 242 | drm_gem_object_unreference_unlocked(&reloc_info[i].dst_bo->gem_base); |
246 | reloc_src_bo = NULL; | 243 | ret = -EINVAL; |
247 | if (reloc.reloc_type == QXL_RELOC_TYPE_BO) { | 244 | goto out_free_bos; |
248 | apply_reloc(qdev, reloc_dst_bo, reloc_dst_offset + reloc.dst_offset, | ||
249 | reloc_src_bo, reloc.src_offset); | ||
250 | } else if (reloc.reloc_type == QXL_RELOC_TYPE_SURF) { | ||
251 | apply_surf_reloc(qdev, reloc_dst_bo, reloc_dst_offset + reloc.dst_offset, reloc_src_bo); | ||
252 | } else { | ||
253 | DRM_ERROR("unknown reloc type %d\n", reloc.reloc_type); | ||
254 | return -EINVAL; | ||
255 | } | 245 | } |
246 | reloc_info[i].src_offset = reloc.src_offset; | ||
247 | } else { | ||
248 | reloc_info[i].src_bo = NULL; | ||
249 | reloc_info[i].src_offset = 0; | ||
250 | } | ||
251 | } | ||
256 | 252 | ||
257 | if (reloc_src_bo && reloc_src_bo != cmd_bo) { | 253 | /* validate all buffers */ |
258 | qxl_release_add_res(qdev, release, reloc_src_bo); | 254 | ret = qxl_release_reserve_list(release, false); |
259 | drm_gem_object_unreference_unlocked(&reloc_src_bo->gem_base); | 255 | if (ret) |
260 | } | 256 | goto out_free_bos; |
261 | 257 | ||
262 | if (reloc_dst_bo != cmd_bo) | 258 | for (i = 0; i < cmd->relocs_num; ++i) { |
263 | drm_gem_object_unreference_unlocked(&reloc_dst_bo->gem_base); | 259 | if (reloc_info[i].type == QXL_RELOC_TYPE_BO) |
264 | } | 260 | apply_reloc(qdev, &reloc_info[i]); |
265 | qxl_fence_releaseable(qdev, release); | 261 | else if (reloc_info[i].type == QXL_RELOC_TYPE_SURF) |
262 | apply_surf_reloc(qdev, &reloc_info[i]); | ||
263 | } | ||
266 | 264 | ||
267 | ret = qxl_push_command_ring_release(qdev, release, user_cmd.type, true); | 265 | ret = qxl_push_command_ring_release(qdev, release, cmd->type, true); |
268 | if (ret == -ERESTARTSYS) { | 266 | if (ret) |
269 | qxl_release_unreserve(qdev, release); | 267 | qxl_release_backoff_reserve_list(release); |
270 | qxl_release_free(qdev, release); | 268 | else |
271 | qxl_bo_list_unreserve(&reloc_list, true); | 269 | qxl_release_fence_buffer_objects(release); |
270 | |||
271 | out_free_bos: | ||
272 | for (j = 0; j < num_relocs; j++) { | ||
273 | if (reloc_info[j].dst_bo != cmd_bo) | ||
274 | drm_gem_object_unreference_unlocked(&reloc_info[j].dst_bo->gem_base); | ||
275 | if (reloc_info[j].src_bo && reloc_info[j].src_bo != cmd_bo) | ||
276 | drm_gem_object_unreference_unlocked(&reloc_info[j].src_bo->gem_base); | ||
277 | } | ||
278 | out_free_release: | ||
279 | if (ret) | ||
280 | qxl_release_free(qdev, release); | ||
281 | out_free_reloc: | ||
282 | kfree(reloc_info); | ||
283 | return ret; | ||
284 | } | ||
285 | |||
286 | static int qxl_execbuffer_ioctl(struct drm_device *dev, void *data, | ||
287 | struct drm_file *file_priv) | ||
288 | { | ||
289 | struct qxl_device *qdev = dev->dev_private; | ||
290 | struct drm_qxl_execbuffer *execbuffer = data; | ||
291 | struct drm_qxl_command user_cmd; | ||
292 | int cmd_num; | ||
293 | int ret; | ||
294 | |||
295 | for (cmd_num = 0; cmd_num < execbuffer->commands_num; ++cmd_num) { | ||
296 | |||
297 | struct drm_qxl_command *commands = | ||
298 | (struct drm_qxl_command *)(uintptr_t)execbuffer->commands; | ||
299 | |||
300 | if (DRM_COPY_FROM_USER(&user_cmd, &commands[cmd_num], | ||
301 | sizeof(user_cmd))) | ||
302 | return -EFAULT; | ||
303 | |||
304 | ret = qxl_process_single_command(qdev, &user_cmd, file_priv); | ||
305 | if (ret) | ||
272 | return ret; | 306 | return ret; |
273 | } | ||
274 | qxl_release_unreserve(qdev, release); | ||
275 | } | 307 | } |
276 | qxl_bo_list_unreserve(&reloc_list, 0); | ||
277 | return 0; | 308 | return 0; |
278 | } | 309 | } |
279 | 310 | ||
@@ -305,7 +336,7 @@ static int qxl_update_area_ioctl(struct drm_device *dev, void *data, | |||
305 | goto out; | 336 | goto out; |
306 | 337 | ||
307 | if (!qobj->pin_count) { | 338 | if (!qobj->pin_count) { |
308 | qxl_ttm_placement_from_domain(qobj, qobj->type); | 339 | qxl_ttm_placement_from_domain(qobj, qobj->type, false); |
309 | ret = ttm_bo_validate(&qobj->tbo, &qobj->placement, | 340 | ret = ttm_bo_validate(&qobj->tbo, &qobj->placement, |
310 | true, false); | 341 | true, false); |
311 | if (unlikely(ret)) | 342 | if (unlikely(ret)) |
diff --git a/drivers/gpu/drm/qxl/qxl_object.c b/drivers/gpu/drm/qxl/qxl_object.c index 1191fe7788c9..aa161cddd87e 100644 --- a/drivers/gpu/drm/qxl/qxl_object.c +++ b/drivers/gpu/drm/qxl/qxl_object.c | |||
@@ -51,20 +51,21 @@ bool qxl_ttm_bo_is_qxl_bo(struct ttm_buffer_object *bo) | |||
51 | return false; | 51 | return false; |
52 | } | 52 | } |
53 | 53 | ||
54 | void qxl_ttm_placement_from_domain(struct qxl_bo *qbo, u32 domain) | 54 | void qxl_ttm_placement_from_domain(struct qxl_bo *qbo, u32 domain, bool pinned) |
55 | { | 55 | { |
56 | u32 c = 0; | 56 | u32 c = 0; |
57 | u32 pflag = pinned ? TTM_PL_FLAG_NO_EVICT : 0; | ||
57 | 58 | ||
58 | qbo->placement.fpfn = 0; | 59 | qbo->placement.fpfn = 0; |
59 | qbo->placement.lpfn = 0; | 60 | qbo->placement.lpfn = 0; |
60 | qbo->placement.placement = qbo->placements; | 61 | qbo->placement.placement = qbo->placements; |
61 | qbo->placement.busy_placement = qbo->placements; | 62 | qbo->placement.busy_placement = qbo->placements; |
62 | if (domain == QXL_GEM_DOMAIN_VRAM) | 63 | if (domain == QXL_GEM_DOMAIN_VRAM) |
63 | qbo->placements[c++] = TTM_PL_FLAG_CACHED | TTM_PL_FLAG_VRAM; | 64 | qbo->placements[c++] = TTM_PL_FLAG_CACHED | TTM_PL_FLAG_VRAM | pflag; |
64 | if (domain == QXL_GEM_DOMAIN_SURFACE) | 65 | if (domain == QXL_GEM_DOMAIN_SURFACE) |
65 | qbo->placements[c++] = TTM_PL_FLAG_CACHED | TTM_PL_FLAG_PRIV0; | 66 | qbo->placements[c++] = TTM_PL_FLAG_CACHED | TTM_PL_FLAG_PRIV0 | pflag; |
66 | if (domain == QXL_GEM_DOMAIN_CPU) | 67 | if (domain == QXL_GEM_DOMAIN_CPU) |
67 | qbo->placements[c++] = TTM_PL_MASK_CACHING | TTM_PL_FLAG_SYSTEM; | 68 | qbo->placements[c++] = TTM_PL_MASK_CACHING | TTM_PL_FLAG_SYSTEM | pflag; |
68 | if (!c) | 69 | if (!c) |
69 | qbo->placements[c++] = TTM_PL_MASK_CACHING | TTM_PL_FLAG_SYSTEM; | 70 | qbo->placements[c++] = TTM_PL_MASK_CACHING | TTM_PL_FLAG_SYSTEM; |
70 | qbo->placement.num_placement = c; | 71 | qbo->placement.num_placement = c; |
@@ -73,7 +74,7 @@ void qxl_ttm_placement_from_domain(struct qxl_bo *qbo, u32 domain) | |||
73 | 74 | ||
74 | 75 | ||
75 | int qxl_bo_create(struct qxl_device *qdev, | 76 | int qxl_bo_create(struct qxl_device *qdev, |
76 | unsigned long size, bool kernel, u32 domain, | 77 | unsigned long size, bool kernel, bool pinned, u32 domain, |
77 | struct qxl_surface *surf, | 78 | struct qxl_surface *surf, |
78 | struct qxl_bo **bo_ptr) | 79 | struct qxl_bo **bo_ptr) |
79 | { | 80 | { |
@@ -99,15 +100,15 @@ int qxl_bo_create(struct qxl_device *qdev, | |||
99 | } | 100 | } |
100 | bo->gem_base.driver_private = NULL; | 101 | bo->gem_base.driver_private = NULL; |
101 | bo->type = domain; | 102 | bo->type = domain; |
102 | bo->pin_count = 0; | 103 | bo->pin_count = pinned ? 1 : 0; |
103 | bo->surface_id = 0; | 104 | bo->surface_id = 0; |
104 | qxl_fence_init(qdev, &bo->fence); | 105 | qxl_fence_init(qdev, &bo->fence); |
105 | INIT_LIST_HEAD(&bo->list); | 106 | INIT_LIST_HEAD(&bo->list); |
106 | atomic_set(&bo->reserve_count, 0); | 107 | |
107 | if (surf) | 108 | if (surf) |
108 | bo->surf = *surf; | 109 | bo->surf = *surf; |
109 | 110 | ||
110 | qxl_ttm_placement_from_domain(bo, domain); | 111 | qxl_ttm_placement_from_domain(bo, domain, pinned); |
111 | 112 | ||
112 | r = ttm_bo_init(&qdev->mman.bdev, &bo->tbo, size, type, | 113 | r = ttm_bo_init(&qdev->mman.bdev, &bo->tbo, size, type, |
113 | &bo->placement, 0, !kernel, NULL, size, | 114 | &bo->placement, 0, !kernel, NULL, size, |
@@ -228,7 +229,7 @@ struct qxl_bo *qxl_bo_ref(struct qxl_bo *bo) | |||
228 | int qxl_bo_pin(struct qxl_bo *bo, u32 domain, u64 *gpu_addr) | 229 | int qxl_bo_pin(struct qxl_bo *bo, u32 domain, u64 *gpu_addr) |
229 | { | 230 | { |
230 | struct qxl_device *qdev = (struct qxl_device *)bo->gem_base.dev->dev_private; | 231 | struct qxl_device *qdev = (struct qxl_device *)bo->gem_base.dev->dev_private; |
231 | int r, i; | 232 | int r; |
232 | 233 | ||
233 | if (bo->pin_count) { | 234 | if (bo->pin_count) { |
234 | bo->pin_count++; | 235 | bo->pin_count++; |
@@ -236,9 +237,7 @@ int qxl_bo_pin(struct qxl_bo *bo, u32 domain, u64 *gpu_addr) | |||
236 | *gpu_addr = qxl_bo_gpu_offset(bo); | 237 | *gpu_addr = qxl_bo_gpu_offset(bo); |
237 | return 0; | 238 | return 0; |
238 | } | 239 | } |
239 | qxl_ttm_placement_from_domain(bo, domain); | 240 | qxl_ttm_placement_from_domain(bo, domain, true); |
240 | for (i = 0; i < bo->placement.num_placement; i++) | ||
241 | bo->placements[i] |= TTM_PL_FLAG_NO_EVICT; | ||
242 | r = ttm_bo_validate(&bo->tbo, &bo->placement, false, false); | 241 | r = ttm_bo_validate(&bo->tbo, &bo->placement, false, false); |
243 | if (likely(r == 0)) { | 242 | if (likely(r == 0)) { |
244 | bo->pin_count = 1; | 243 | bo->pin_count = 1; |
@@ -317,53 +316,6 @@ int qxl_bo_check_id(struct qxl_device *qdev, struct qxl_bo *bo) | |||
317 | return 0; | 316 | return 0; |
318 | } | 317 | } |
319 | 318 | ||
320 | void qxl_bo_list_unreserve(struct qxl_reloc_list *reloc_list, bool failed) | ||
321 | { | ||
322 | struct qxl_bo_list *entry, *sf; | ||
323 | |||
324 | list_for_each_entry_safe(entry, sf, &reloc_list->bos, lhead) { | ||
325 | qxl_bo_unreserve(entry->bo); | ||
326 | list_del(&entry->lhead); | ||
327 | kfree(entry); | ||
328 | } | ||
329 | } | ||
330 | |||
331 | int qxl_bo_list_add(struct qxl_reloc_list *reloc_list, struct qxl_bo *bo) | ||
332 | { | ||
333 | struct qxl_bo_list *entry; | ||
334 | int ret; | ||
335 | |||
336 | list_for_each_entry(entry, &reloc_list->bos, lhead) { | ||
337 | if (entry->bo == bo) | ||
338 | return 0; | ||
339 | } | ||
340 | |||
341 | entry = kmalloc(sizeof(struct qxl_bo_list), GFP_KERNEL); | ||
342 | if (!entry) | ||
343 | return -ENOMEM; | ||
344 | |||
345 | entry->bo = bo; | ||
346 | list_add(&entry->lhead, &reloc_list->bos); | ||
347 | |||
348 | ret = qxl_bo_reserve(bo, false); | ||
349 | if (ret) | ||
350 | return ret; | ||
351 | |||
352 | if (!bo->pin_count) { | ||
353 | qxl_ttm_placement_from_domain(bo, bo->type); | ||
354 | ret = ttm_bo_validate(&bo->tbo, &bo->placement, | ||
355 | true, false); | ||
356 | if (ret) | ||
357 | return ret; | ||
358 | } | ||
359 | |||
360 | /* allocate a surface for reserved + validated buffers */ | ||
361 | ret = qxl_bo_check_id(bo->gem_base.dev->dev_private, bo); | ||
362 | if (ret) | ||
363 | return ret; | ||
364 | return 0; | ||
365 | } | ||
366 | |||
367 | int qxl_surf_evict(struct qxl_device *qdev) | 319 | int qxl_surf_evict(struct qxl_device *qdev) |
368 | { | 320 | { |
369 | return ttm_bo_evict_mm(&qdev->mman.bdev, TTM_PL_PRIV0); | 321 | return ttm_bo_evict_mm(&qdev->mman.bdev, TTM_PL_PRIV0); |
diff --git a/drivers/gpu/drm/qxl/qxl_object.h b/drivers/gpu/drm/qxl/qxl_object.h index ee7ad79ce781..8cb6167038e5 100644 --- a/drivers/gpu/drm/qxl/qxl_object.h +++ b/drivers/gpu/drm/qxl/qxl_object.h | |||
@@ -88,7 +88,7 @@ static inline int qxl_bo_wait(struct qxl_bo *bo, u32 *mem_type, | |||
88 | 88 | ||
89 | extern int qxl_bo_create(struct qxl_device *qdev, | 89 | extern int qxl_bo_create(struct qxl_device *qdev, |
90 | unsigned long size, | 90 | unsigned long size, |
91 | bool kernel, u32 domain, | 91 | bool kernel, bool pinned, u32 domain, |
92 | struct qxl_surface *surf, | 92 | struct qxl_surface *surf, |
93 | struct qxl_bo **bo_ptr); | 93 | struct qxl_bo **bo_ptr); |
94 | extern int qxl_bo_kmap(struct qxl_bo *bo, void **ptr); | 94 | extern int qxl_bo_kmap(struct qxl_bo *bo, void **ptr); |
@@ -99,9 +99,7 @@ extern struct qxl_bo *qxl_bo_ref(struct qxl_bo *bo); | |||
99 | extern void qxl_bo_unref(struct qxl_bo **bo); | 99 | extern void qxl_bo_unref(struct qxl_bo **bo); |
100 | extern int qxl_bo_pin(struct qxl_bo *bo, u32 domain, u64 *gpu_addr); | 100 | extern int qxl_bo_pin(struct qxl_bo *bo, u32 domain, u64 *gpu_addr); |
101 | extern int qxl_bo_unpin(struct qxl_bo *bo); | 101 | extern int qxl_bo_unpin(struct qxl_bo *bo); |
102 | extern void qxl_ttm_placement_from_domain(struct qxl_bo *qbo, u32 domain); | 102 | extern void qxl_ttm_placement_from_domain(struct qxl_bo *qbo, u32 domain, bool pinned); |
103 | extern bool qxl_ttm_bo_is_qxl_bo(struct ttm_buffer_object *bo); | 103 | extern bool qxl_ttm_bo_is_qxl_bo(struct ttm_buffer_object *bo); |
104 | 104 | ||
105 | extern int qxl_bo_list_add(struct qxl_reloc_list *reloc_list, struct qxl_bo *bo); | ||
106 | extern void qxl_bo_list_unreserve(struct qxl_reloc_list *reloc_list, bool failed); | ||
107 | #endif | 105 | #endif |
diff --git a/drivers/gpu/drm/qxl/qxl_release.c b/drivers/gpu/drm/qxl/qxl_release.c index b443d6751d5f..b61449e52cd5 100644 --- a/drivers/gpu/drm/qxl/qxl_release.c +++ b/drivers/gpu/drm/qxl/qxl_release.c | |||
@@ -38,7 +38,8 @@ | |||
38 | 38 | ||
39 | static const int release_size_per_bo[] = { RELEASE_SIZE, SURFACE_RELEASE_SIZE, RELEASE_SIZE }; | 39 | static const int release_size_per_bo[] = { RELEASE_SIZE, SURFACE_RELEASE_SIZE, RELEASE_SIZE }; |
40 | static const int releases_per_bo[] = { RELEASES_PER_BO, SURFACE_RELEASES_PER_BO, RELEASES_PER_BO }; | 40 | static const int releases_per_bo[] = { RELEASES_PER_BO, SURFACE_RELEASES_PER_BO, RELEASES_PER_BO }; |
41 | uint64_t | 41 | |
42 | static uint64_t | ||
42 | qxl_release_alloc(struct qxl_device *qdev, int type, | 43 | qxl_release_alloc(struct qxl_device *qdev, int type, |
43 | struct qxl_release **ret) | 44 | struct qxl_release **ret) |
44 | { | 45 | { |
@@ -53,9 +54,9 @@ qxl_release_alloc(struct qxl_device *qdev, int type, | |||
53 | return 0; | 54 | return 0; |
54 | } | 55 | } |
55 | release->type = type; | 56 | release->type = type; |
56 | release->bo_count = 0; | ||
57 | release->release_offset = 0; | 57 | release->release_offset = 0; |
58 | release->surface_release_id = 0; | 58 | release->surface_release_id = 0; |
59 | INIT_LIST_HEAD(&release->bos); | ||
59 | 60 | ||
60 | idr_preload(GFP_KERNEL); | 61 | idr_preload(GFP_KERNEL); |
61 | spin_lock(&qdev->release_idr_lock); | 62 | spin_lock(&qdev->release_idr_lock); |
@@ -77,20 +78,20 @@ void | |||
77 | qxl_release_free(struct qxl_device *qdev, | 78 | qxl_release_free(struct qxl_device *qdev, |
78 | struct qxl_release *release) | 79 | struct qxl_release *release) |
79 | { | 80 | { |
80 | int i; | 81 | struct qxl_bo_list *entry, *tmp; |
81 | 82 | QXL_INFO(qdev, "release %d, type %d\n", release->id, | |
82 | QXL_INFO(qdev, "release %d, type %d, %d bos\n", release->id, | 83 | release->type); |
83 | release->type, release->bo_count); | ||
84 | 84 | ||
85 | if (release->surface_release_id) | 85 | if (release->surface_release_id) |
86 | qxl_surface_id_dealloc(qdev, release->surface_release_id); | 86 | qxl_surface_id_dealloc(qdev, release->surface_release_id); |
87 | 87 | ||
88 | for (i = 0 ; i < release->bo_count; ++i) { | 88 | list_for_each_entry_safe(entry, tmp, &release->bos, tv.head) { |
89 | struct qxl_bo *bo = to_qxl_bo(entry->tv.bo); | ||
89 | QXL_INFO(qdev, "release %llx\n", | 90 | QXL_INFO(qdev, "release %llx\n", |
90 | release->bos[i]->tbo.addr_space_offset | 91 | entry->tv.bo->addr_space_offset |
91 | - DRM_FILE_OFFSET); | 92 | - DRM_FILE_OFFSET); |
92 | qxl_fence_remove_release(&release->bos[i]->fence, release->id); | 93 | qxl_fence_remove_release(&bo->fence, release->id); |
93 | qxl_bo_unref(&release->bos[i]); | 94 | qxl_bo_unref(&bo); |
94 | } | 95 | } |
95 | spin_lock(&qdev->release_idr_lock); | 96 | spin_lock(&qdev->release_idr_lock); |
96 | idr_remove(&qdev->release_idr, release->id); | 97 | idr_remove(&qdev->release_idr, release->id); |
@@ -98,83 +99,117 @@ qxl_release_free(struct qxl_device *qdev, | |||
98 | kfree(release); | 99 | kfree(release); |
99 | } | 100 | } |
100 | 101 | ||
101 | void | ||
102 | qxl_release_add_res(struct qxl_device *qdev, struct qxl_release *release, | ||
103 | struct qxl_bo *bo) | ||
104 | { | ||
105 | int i; | ||
106 | for (i = 0; i < release->bo_count; i++) | ||
107 | if (release->bos[i] == bo) | ||
108 | return; | ||
109 | |||
110 | if (release->bo_count >= QXL_MAX_RES) { | ||
111 | DRM_ERROR("exceeded max resource on a qxl_release item\n"); | ||
112 | return; | ||
113 | } | ||
114 | release->bos[release->bo_count++] = qxl_bo_ref(bo); | ||
115 | } | ||
116 | |||
117 | static int qxl_release_bo_alloc(struct qxl_device *qdev, | 102 | static int qxl_release_bo_alloc(struct qxl_device *qdev, |
118 | struct qxl_bo **bo) | 103 | struct qxl_bo **bo) |
119 | { | 104 | { |
120 | int ret; | 105 | int ret; |
121 | ret = qxl_bo_create(qdev, PAGE_SIZE, false, QXL_GEM_DOMAIN_VRAM, NULL, | 106 | /* pin releases bo's they are too messy to evict */ |
107 | ret = qxl_bo_create(qdev, PAGE_SIZE, false, true, | ||
108 | QXL_GEM_DOMAIN_VRAM, NULL, | ||
122 | bo); | 109 | bo); |
123 | return ret; | 110 | return ret; |
124 | } | 111 | } |
125 | 112 | ||
126 | int qxl_release_reserve(struct qxl_device *qdev, | 113 | int qxl_release_list_add(struct qxl_release *release, struct qxl_bo *bo) |
127 | struct qxl_release *release, bool no_wait) | 114 | { |
115 | struct qxl_bo_list *entry; | ||
116 | |||
117 | list_for_each_entry(entry, &release->bos, tv.head) { | ||
118 | if (entry->tv.bo == &bo->tbo) | ||
119 | return 0; | ||
120 | } | ||
121 | |||
122 | entry = kmalloc(sizeof(struct qxl_bo_list), GFP_KERNEL); | ||
123 | if (!entry) | ||
124 | return -ENOMEM; | ||
125 | |||
126 | qxl_bo_ref(bo); | ||
127 | entry->tv.bo = &bo->tbo; | ||
128 | list_add_tail(&entry->tv.head, &release->bos); | ||
129 | return 0; | ||
130 | } | ||
131 | |||
132 | static int qxl_release_validate_bo(struct qxl_bo *bo) | ||
128 | { | 133 | { |
129 | int ret; | 134 | int ret; |
130 | if (atomic_inc_return(&release->bos[0]->reserve_count) == 1) { | 135 | |
131 | ret = qxl_bo_reserve(release->bos[0], no_wait); | 136 | if (!bo->pin_count) { |
137 | qxl_ttm_placement_from_domain(bo, bo->type, false); | ||
138 | ret = ttm_bo_validate(&bo->tbo, &bo->placement, | ||
139 | true, false); | ||
132 | if (ret) | 140 | if (ret) |
133 | return ret; | 141 | return ret; |
134 | } | 142 | } |
143 | |||
144 | /* allocate a surface for reserved + validated buffers */ | ||
145 | ret = qxl_bo_check_id(bo->gem_base.dev->dev_private, bo); | ||
146 | if (ret) | ||
147 | return ret; | ||
148 | return 0; | ||
149 | } | ||
150 | |||
151 | int qxl_release_reserve_list(struct qxl_release *release, bool no_intr) | ||
152 | { | ||
153 | int ret; | ||
154 | struct qxl_bo_list *entry; | ||
155 | |||
156 | /* if only one object on the release its the release itself | ||
157 | since these objects are pinned no need to reserve */ | ||
158 | if (list_is_singular(&release->bos)) | ||
159 | return 0; | ||
160 | |||
161 | ret = ttm_eu_reserve_buffers(&release->ticket, &release->bos); | ||
162 | if (ret) | ||
163 | return ret; | ||
164 | |||
165 | list_for_each_entry(entry, &release->bos, tv.head) { | ||
166 | struct qxl_bo *bo = to_qxl_bo(entry->tv.bo); | ||
167 | |||
168 | ret = qxl_release_validate_bo(bo); | ||
169 | if (ret) { | ||
170 | ttm_eu_backoff_reservation(&release->ticket, &release->bos); | ||
171 | return ret; | ||
172 | } | ||
173 | } | ||
135 | return 0; | 174 | return 0; |
136 | } | 175 | } |
137 | 176 | ||
138 | void qxl_release_unreserve(struct qxl_device *qdev, | 177 | void qxl_release_backoff_reserve_list(struct qxl_release *release) |
139 | struct qxl_release *release) | ||
140 | { | 178 | { |
141 | if (atomic_dec_and_test(&release->bos[0]->reserve_count)) | 179 | /* if only one object on the release its the release itself |
142 | qxl_bo_unreserve(release->bos[0]); | 180 | since these objects are pinned no need to reserve */ |
181 | if (list_is_singular(&release->bos)) | ||
182 | return; | ||
183 | |||
184 | ttm_eu_backoff_reservation(&release->ticket, &release->bos); | ||
143 | } | 185 | } |
144 | 186 | ||
187 | |||
145 | int qxl_alloc_surface_release_reserved(struct qxl_device *qdev, | 188 | int qxl_alloc_surface_release_reserved(struct qxl_device *qdev, |
146 | enum qxl_surface_cmd_type surface_cmd_type, | 189 | enum qxl_surface_cmd_type surface_cmd_type, |
147 | struct qxl_release *create_rel, | 190 | struct qxl_release *create_rel, |
148 | struct qxl_release **release) | 191 | struct qxl_release **release) |
149 | { | 192 | { |
150 | int ret; | ||
151 | |||
152 | if (surface_cmd_type == QXL_SURFACE_CMD_DESTROY && create_rel) { | 193 | if (surface_cmd_type == QXL_SURFACE_CMD_DESTROY && create_rel) { |
153 | int idr_ret; | 194 | int idr_ret; |
195 | struct qxl_bo_list *entry = list_first_entry(&create_rel->bos, struct qxl_bo_list, tv.head); | ||
154 | struct qxl_bo *bo; | 196 | struct qxl_bo *bo; |
155 | union qxl_release_info *info; | 197 | union qxl_release_info *info; |
156 | 198 | ||
157 | /* stash the release after the create command */ | 199 | /* stash the release after the create command */ |
158 | idr_ret = qxl_release_alloc(qdev, QXL_RELEASE_SURFACE_CMD, release); | 200 | idr_ret = qxl_release_alloc(qdev, QXL_RELEASE_SURFACE_CMD, release); |
159 | bo = qxl_bo_ref(create_rel->bos[0]); | 201 | bo = qxl_bo_ref(to_qxl_bo(entry->tv.bo)); |
160 | 202 | ||
161 | (*release)->release_offset = create_rel->release_offset + 64; | 203 | (*release)->release_offset = create_rel->release_offset + 64; |
162 | 204 | ||
163 | qxl_release_add_res(qdev, *release, bo); | 205 | qxl_release_list_add(*release, bo); |
164 | 206 | ||
165 | ret = qxl_release_reserve(qdev, *release, false); | ||
166 | if (ret) { | ||
167 | DRM_ERROR("release reserve failed\n"); | ||
168 | goto out_unref; | ||
169 | } | ||
170 | info = qxl_release_map(qdev, *release); | 207 | info = qxl_release_map(qdev, *release); |
171 | info->id = idr_ret; | 208 | info->id = idr_ret; |
172 | qxl_release_unmap(qdev, *release, info); | 209 | qxl_release_unmap(qdev, *release, info); |
173 | 210 | ||
174 | |||
175 | out_unref: | ||
176 | qxl_bo_unref(&bo); | 211 | qxl_bo_unref(&bo); |
177 | return ret; | 212 | return 0; |
178 | } | 213 | } |
179 | 214 | ||
180 | return qxl_alloc_release_reserved(qdev, sizeof(struct qxl_surface_cmd), | 215 | return qxl_alloc_release_reserved(qdev, sizeof(struct qxl_surface_cmd), |
@@ -187,7 +222,7 @@ int qxl_alloc_release_reserved(struct qxl_device *qdev, unsigned long size, | |||
187 | { | 222 | { |
188 | struct qxl_bo *bo; | 223 | struct qxl_bo *bo; |
189 | int idr_ret; | 224 | int idr_ret; |
190 | int ret; | 225 | int ret = 0; |
191 | union qxl_release_info *info; | 226 | union qxl_release_info *info; |
192 | int cur_idx; | 227 | int cur_idx; |
193 | 228 | ||
@@ -216,11 +251,6 @@ int qxl_alloc_release_reserved(struct qxl_device *qdev, unsigned long size, | |||
216 | mutex_unlock(&qdev->release_mutex); | 251 | mutex_unlock(&qdev->release_mutex); |
217 | return ret; | 252 | return ret; |
218 | } | 253 | } |
219 | |||
220 | /* pin releases bo's they are too messy to evict */ | ||
221 | ret = qxl_bo_reserve(qdev->current_release_bo[cur_idx], false); | ||
222 | qxl_bo_pin(qdev->current_release_bo[cur_idx], QXL_GEM_DOMAIN_VRAM, NULL); | ||
223 | qxl_bo_unreserve(qdev->current_release_bo[cur_idx]); | ||
224 | } | 254 | } |
225 | 255 | ||
226 | bo = qxl_bo_ref(qdev->current_release_bo[cur_idx]); | 256 | bo = qxl_bo_ref(qdev->current_release_bo[cur_idx]); |
@@ -231,36 +261,18 @@ int qxl_alloc_release_reserved(struct qxl_device *qdev, unsigned long size, | |||
231 | if (rbo) | 261 | if (rbo) |
232 | *rbo = bo; | 262 | *rbo = bo; |
233 | 263 | ||
234 | qxl_release_add_res(qdev, *release, bo); | ||
235 | |||
236 | ret = qxl_release_reserve(qdev, *release, false); | ||
237 | mutex_unlock(&qdev->release_mutex); | 264 | mutex_unlock(&qdev->release_mutex); |
238 | if (ret) | 265 | |
239 | goto out_unref; | 266 | qxl_release_list_add(*release, bo); |
240 | 267 | ||
241 | info = qxl_release_map(qdev, *release); | 268 | info = qxl_release_map(qdev, *release); |
242 | info->id = idr_ret; | 269 | info->id = idr_ret; |
243 | qxl_release_unmap(qdev, *release, info); | 270 | qxl_release_unmap(qdev, *release, info); |
244 | 271 | ||
245 | out_unref: | ||
246 | qxl_bo_unref(&bo); | 272 | qxl_bo_unref(&bo); |
247 | return ret; | 273 | return ret; |
248 | } | 274 | } |
249 | 275 | ||
250 | int qxl_fence_releaseable(struct qxl_device *qdev, | ||
251 | struct qxl_release *release) | ||
252 | { | ||
253 | int i, ret; | ||
254 | for (i = 0; i < release->bo_count; i++) { | ||
255 | if (!release->bos[i]->tbo.sync_obj) | ||
256 | release->bos[i]->tbo.sync_obj = &release->bos[i]->fence; | ||
257 | ret = qxl_fence_add_release(&release->bos[i]->fence, release->id); | ||
258 | if (ret) | ||
259 | return ret; | ||
260 | } | ||
261 | return 0; | ||
262 | } | ||
263 | |||
264 | struct qxl_release *qxl_release_from_id_locked(struct qxl_device *qdev, | 276 | struct qxl_release *qxl_release_from_id_locked(struct qxl_device *qdev, |
265 | uint64_t id) | 277 | uint64_t id) |
266 | { | 278 | { |
@@ -273,10 +285,7 @@ struct qxl_release *qxl_release_from_id_locked(struct qxl_device *qdev, | |||
273 | DRM_ERROR("failed to find id in release_idr\n"); | 285 | DRM_ERROR("failed to find id in release_idr\n"); |
274 | return NULL; | 286 | return NULL; |
275 | } | 287 | } |
276 | if (release->bo_count < 1) { | 288 | |
277 | DRM_ERROR("read a released resource with 0 bos\n"); | ||
278 | return NULL; | ||
279 | } | ||
280 | return release; | 289 | return release; |
281 | } | 290 | } |
282 | 291 | ||
@@ -285,9 +294,12 @@ union qxl_release_info *qxl_release_map(struct qxl_device *qdev, | |||
285 | { | 294 | { |
286 | void *ptr; | 295 | void *ptr; |
287 | union qxl_release_info *info; | 296 | union qxl_release_info *info; |
288 | struct qxl_bo *bo = release->bos[0]; | 297 | struct qxl_bo_list *entry = list_first_entry(&release->bos, struct qxl_bo_list, tv.head); |
298 | struct qxl_bo *bo = to_qxl_bo(entry->tv.bo); | ||
289 | 299 | ||
290 | ptr = qxl_bo_kmap_atomic_page(qdev, bo, release->release_offset & PAGE_SIZE); | 300 | ptr = qxl_bo_kmap_atomic_page(qdev, bo, release->release_offset & PAGE_SIZE); |
301 | if (!ptr) | ||
302 | return NULL; | ||
291 | info = ptr + (release->release_offset & ~PAGE_SIZE); | 303 | info = ptr + (release->release_offset & ~PAGE_SIZE); |
292 | return info; | 304 | return info; |
293 | } | 305 | } |
@@ -296,9 +308,51 @@ void qxl_release_unmap(struct qxl_device *qdev, | |||
296 | struct qxl_release *release, | 308 | struct qxl_release *release, |
297 | union qxl_release_info *info) | 309 | union qxl_release_info *info) |
298 | { | 310 | { |
299 | struct qxl_bo *bo = release->bos[0]; | 311 | struct qxl_bo_list *entry = list_first_entry(&release->bos, struct qxl_bo_list, tv.head); |
312 | struct qxl_bo *bo = to_qxl_bo(entry->tv.bo); | ||
300 | void *ptr; | 313 | void *ptr; |
301 | 314 | ||
302 | ptr = ((void *)info) - (release->release_offset & ~PAGE_SIZE); | 315 | ptr = ((void *)info) - (release->release_offset & ~PAGE_SIZE); |
303 | qxl_bo_kunmap_atomic_page(qdev, bo, ptr); | 316 | qxl_bo_kunmap_atomic_page(qdev, bo, ptr); |
304 | } | 317 | } |
318 | |||
319 | void qxl_release_fence_buffer_objects(struct qxl_release *release) | ||
320 | { | ||
321 | struct ttm_validate_buffer *entry; | ||
322 | struct ttm_buffer_object *bo; | ||
323 | struct ttm_bo_global *glob; | ||
324 | struct ttm_bo_device *bdev; | ||
325 | struct ttm_bo_driver *driver; | ||
326 | struct qxl_bo *qbo; | ||
327 | |||
328 | /* if only one object on the release its the release itself | ||
329 | since these objects are pinned no need to reserve */ | ||
330 | if (list_is_singular(&release->bos)) | ||
331 | return; | ||
332 | |||
333 | bo = list_first_entry(&release->bos, struct ttm_validate_buffer, head)->bo; | ||
334 | bdev = bo->bdev; | ||
335 | driver = bdev->driver; | ||
336 | glob = bo->glob; | ||
337 | |||
338 | spin_lock(&glob->lru_lock); | ||
339 | spin_lock(&bdev->fence_lock); | ||
340 | |||
341 | list_for_each_entry(entry, &release->bos, head) { | ||
342 | bo = entry->bo; | ||
343 | qbo = to_qxl_bo(bo); | ||
344 | |||
345 | if (!entry->bo->sync_obj) | ||
346 | entry->bo->sync_obj = &qbo->fence; | ||
347 | |||
348 | qxl_fence_add_release_locked(&qbo->fence, release->id); | ||
349 | |||
350 | ttm_bo_add_to_lru(bo); | ||
351 | ww_mutex_unlock(&bo->resv->lock); | ||
352 | entry->reserved = false; | ||
353 | } | ||
354 | spin_unlock(&bdev->fence_lock); | ||
355 | spin_unlock(&glob->lru_lock); | ||
356 | ww_acquire_fini(&release->ticket); | ||
357 | } | ||
358 | |||
diff --git a/drivers/gpu/drm/qxl/qxl_ttm.c b/drivers/gpu/drm/qxl/qxl_ttm.c index 489cb8cece4d..1dfd84cda2a1 100644 --- a/drivers/gpu/drm/qxl/qxl_ttm.c +++ b/drivers/gpu/drm/qxl/qxl_ttm.c | |||
@@ -206,7 +206,7 @@ static void qxl_evict_flags(struct ttm_buffer_object *bo, | |||
206 | return; | 206 | return; |
207 | } | 207 | } |
208 | qbo = container_of(bo, struct qxl_bo, tbo); | 208 | qbo = container_of(bo, struct qxl_bo, tbo); |
209 | qxl_ttm_placement_from_domain(qbo, QXL_GEM_DOMAIN_CPU); | 209 | qxl_ttm_placement_from_domain(qbo, QXL_GEM_DOMAIN_CPU, false); |
210 | *placement = qbo->placement; | 210 | *placement = qbo->placement; |
211 | } | 211 | } |
212 | 212 | ||
diff --git a/drivers/gpu/drm/radeon/atombios_dp.c b/drivers/gpu/drm/radeon/atombios_dp.c index 064023bed480..32501f6ec991 100644 --- a/drivers/gpu/drm/radeon/atombios_dp.c +++ b/drivers/gpu/drm/radeon/atombios_dp.c | |||
@@ -44,6 +44,41 @@ static char *pre_emph_names[] = { | |||
44 | }; | 44 | }; |
45 | 45 | ||
46 | /***** radeon AUX functions *****/ | 46 | /***** radeon AUX functions *****/ |
47 | |||
48 | /* Atom needs data in little endian format | ||
49 | * so swap as appropriate when copying data to | ||
50 | * or from atom. Note that atom operates on | ||
51 | * dw units. | ||
52 | */ | ||
53 | static void radeon_copy_swap(u8 *dst, u8 *src, u8 num_bytes, bool to_le) | ||
54 | { | ||
55 | #ifdef __BIG_ENDIAN | ||
56 | u8 src_tmp[20], dst_tmp[20]; /* used for byteswapping */ | ||
57 | u32 *dst32, *src32; | ||
58 | int i; | ||
59 | |||
60 | memcpy(src_tmp, src, num_bytes); | ||
61 | src32 = (u32 *)src_tmp; | ||
62 | dst32 = (u32 *)dst_tmp; | ||
63 | if (to_le) { | ||
64 | for (i = 0; i < ((num_bytes + 3) / 4); i++) | ||
65 | dst32[i] = cpu_to_le32(src32[i]); | ||
66 | memcpy(dst, dst_tmp, num_bytes); | ||
67 | } else { | ||
68 | u8 dws = num_bytes & ~3; | ||
69 | for (i = 0; i < ((num_bytes + 3) / 4); i++) | ||
70 | dst32[i] = le32_to_cpu(src32[i]); | ||
71 | memcpy(dst, dst_tmp, dws); | ||
72 | if (num_bytes % 4) { | ||
73 | for (i = 0; i < (num_bytes % 4); i++) | ||
74 | dst[dws+i] = dst_tmp[dws+i]; | ||
75 | } | ||
76 | } | ||
77 | #else | ||
78 | memcpy(dst, src, num_bytes); | ||
79 | #endif | ||
80 | } | ||
81 | |||
47 | union aux_channel_transaction { | 82 | union aux_channel_transaction { |
48 | PROCESS_AUX_CHANNEL_TRANSACTION_PS_ALLOCATION v1; | 83 | PROCESS_AUX_CHANNEL_TRANSACTION_PS_ALLOCATION v1; |
49 | PROCESS_AUX_CHANNEL_TRANSACTION_PARAMETERS_V2 v2; | 84 | PROCESS_AUX_CHANNEL_TRANSACTION_PARAMETERS_V2 v2; |
@@ -65,10 +100,10 @@ static int radeon_process_aux_ch(struct radeon_i2c_chan *chan, | |||
65 | 100 | ||
66 | base = (unsigned char *)(rdev->mode_info.atom_context->scratch + 1); | 101 | base = (unsigned char *)(rdev->mode_info.atom_context->scratch + 1); |
67 | 102 | ||
68 | memcpy(base, send, send_bytes); | 103 | radeon_copy_swap(base, send, send_bytes, true); |
69 | 104 | ||
70 | args.v1.lpAuxRequest = 0 + 4; | 105 | args.v1.lpAuxRequest = cpu_to_le16((u16)(0 + 4)); |
71 | args.v1.lpDataOut = 16 + 4; | 106 | args.v1.lpDataOut = cpu_to_le16((u16)(16 + 4)); |
72 | args.v1.ucDataOutLen = 0; | 107 | args.v1.ucDataOutLen = 0; |
73 | args.v1.ucChannelID = chan->rec.i2c_id; | 108 | args.v1.ucChannelID = chan->rec.i2c_id; |
74 | args.v1.ucDelay = delay / 10; | 109 | args.v1.ucDelay = delay / 10; |
@@ -102,7 +137,7 @@ static int radeon_process_aux_ch(struct radeon_i2c_chan *chan, | |||
102 | recv_bytes = recv_size; | 137 | recv_bytes = recv_size; |
103 | 138 | ||
104 | if (recv && recv_size) | 139 | if (recv && recv_size) |
105 | memcpy(recv, base + 16, recv_bytes); | 140 | radeon_copy_swap(recv, base + 16, recv_bytes, false); |
106 | 141 | ||
107 | return recv_bytes; | 142 | return recv_bytes; |
108 | } | 143 | } |
diff --git a/drivers/gpu/drm/radeon/r600.c b/drivers/gpu/drm/radeon/r600.c index 393880a09412..10f712e37003 100644 --- a/drivers/gpu/drm/radeon/r600.c +++ b/drivers/gpu/drm/radeon/r600.c | |||
@@ -3166,7 +3166,7 @@ int r600_copy_cpdma(struct radeon_device *rdev, | |||
3166 | 3166 | ||
3167 | size_in_bytes = (num_gpu_pages << RADEON_GPU_PAGE_SHIFT); | 3167 | size_in_bytes = (num_gpu_pages << RADEON_GPU_PAGE_SHIFT); |
3168 | num_loops = DIV_ROUND_UP(size_in_bytes, 0x1fffff); | 3168 | num_loops = DIV_ROUND_UP(size_in_bytes, 0x1fffff); |
3169 | r = radeon_ring_lock(rdev, ring, num_loops * 6 + 21); | 3169 | r = radeon_ring_lock(rdev, ring, num_loops * 6 + 24); |
3170 | if (r) { | 3170 | if (r) { |
3171 | DRM_ERROR("radeon: moving bo (%d).\n", r); | 3171 | DRM_ERROR("radeon: moving bo (%d).\n", r); |
3172 | radeon_semaphore_free(rdev, &sem, NULL); | 3172 | radeon_semaphore_free(rdev, &sem, NULL); |
@@ -3181,6 +3181,9 @@ int r600_copy_cpdma(struct radeon_device *rdev, | |||
3181 | radeon_semaphore_free(rdev, &sem, NULL); | 3181 | radeon_semaphore_free(rdev, &sem, NULL); |
3182 | } | 3182 | } |
3183 | 3183 | ||
3184 | radeon_ring_write(ring, PACKET3(PACKET3_SET_CONFIG_REG, 1)); | ||
3185 | radeon_ring_write(ring, (WAIT_UNTIL - PACKET3_SET_CONFIG_REG_OFFSET) >> 2); | ||
3186 | radeon_ring_write(ring, WAIT_3D_IDLE_bit); | ||
3184 | for (i = 0; i < num_loops; i++) { | 3187 | for (i = 0; i < num_loops; i++) { |
3185 | cur_size_in_bytes = size_in_bytes; | 3188 | cur_size_in_bytes = size_in_bytes; |
3186 | if (cur_size_in_bytes > 0x1fffff) | 3189 | if (cur_size_in_bytes > 0x1fffff) |
diff --git a/drivers/gpu/drm/radeon/r600_dpm.c b/drivers/gpu/drm/radeon/r600_dpm.c index b88f54b134ab..e5c860f4ccbe 100644 --- a/drivers/gpu/drm/radeon/r600_dpm.c +++ b/drivers/gpu/drm/radeon/r600_dpm.c | |||
@@ -278,9 +278,9 @@ bool r600_dynamicpm_enabled(struct radeon_device *rdev) | |||
278 | void r600_enable_sclk_control(struct radeon_device *rdev, bool enable) | 278 | void r600_enable_sclk_control(struct radeon_device *rdev, bool enable) |
279 | { | 279 | { |
280 | if (enable) | 280 | if (enable) |
281 | WREG32_P(GENERAL_PWRMGT, 0, ~SCLK_PWRMGT_OFF); | 281 | WREG32_P(SCLK_PWRMGT_CNTL, 0, ~SCLK_PWRMGT_OFF); |
282 | else | 282 | else |
283 | WREG32_P(GENERAL_PWRMGT, SCLK_PWRMGT_OFF, ~SCLK_PWRMGT_OFF); | 283 | WREG32_P(SCLK_PWRMGT_CNTL, SCLK_PWRMGT_OFF, ~SCLK_PWRMGT_OFF); |
284 | } | 284 | } |
285 | 285 | ||
286 | void r600_enable_mclk_control(struct radeon_device *rdev, bool enable) | 286 | void r600_enable_mclk_control(struct radeon_device *rdev, bool enable) |
diff --git a/drivers/gpu/drm/radeon/radeon_asic.c b/drivers/gpu/drm/radeon/radeon_asic.c index 78bec1a58ed1..f8f8b3113ddd 100644 --- a/drivers/gpu/drm/radeon/radeon_asic.c +++ b/drivers/gpu/drm/radeon/radeon_asic.c | |||
@@ -1161,6 +1161,7 @@ static struct radeon_asic rv6xx_asic = { | |||
1161 | .get_mclk = &rv6xx_dpm_get_mclk, | 1161 | .get_mclk = &rv6xx_dpm_get_mclk, |
1162 | .print_power_state = &rv6xx_dpm_print_power_state, | 1162 | .print_power_state = &rv6xx_dpm_print_power_state, |
1163 | .debugfs_print_current_performance_level = &rv6xx_dpm_debugfs_print_current_performance_level, | 1163 | .debugfs_print_current_performance_level = &rv6xx_dpm_debugfs_print_current_performance_level, |
1164 | .force_performance_level = &rv6xx_dpm_force_performance_level, | ||
1164 | }, | 1165 | }, |
1165 | .pflip = { | 1166 | .pflip = { |
1166 | .pre_page_flip = &rs600_pre_page_flip, | 1167 | .pre_page_flip = &rs600_pre_page_flip, |
diff --git a/drivers/gpu/drm/radeon/radeon_asic.h b/drivers/gpu/drm/radeon/radeon_asic.h index ca1895709908..902479fa737f 100644 --- a/drivers/gpu/drm/radeon/radeon_asic.h +++ b/drivers/gpu/drm/radeon/radeon_asic.h | |||
@@ -421,6 +421,8 @@ void rv6xx_dpm_print_power_state(struct radeon_device *rdev, | |||
421 | struct radeon_ps *ps); | 421 | struct radeon_ps *ps); |
422 | void rv6xx_dpm_debugfs_print_current_performance_level(struct radeon_device *rdev, | 422 | void rv6xx_dpm_debugfs_print_current_performance_level(struct radeon_device *rdev, |
423 | struct seq_file *m); | 423 | struct seq_file *m); |
424 | int rv6xx_dpm_force_performance_level(struct radeon_device *rdev, | ||
425 | enum radeon_dpm_forced_level level); | ||
424 | /* rs780 dpm */ | 426 | /* rs780 dpm */ |
425 | int rs780_dpm_init(struct radeon_device *rdev); | 427 | int rs780_dpm_init(struct radeon_device *rdev); |
426 | int rs780_dpm_enable(struct radeon_device *rdev); | 428 | int rs780_dpm_enable(struct radeon_device *rdev); |
diff --git a/drivers/gpu/drm/radeon/radeon_combios.c b/drivers/gpu/drm/radeon/radeon_combios.c index 78edadc9e86b..68ce36056019 100644 --- a/drivers/gpu/drm/radeon/radeon_combios.c +++ b/drivers/gpu/drm/radeon/radeon_combios.c | |||
@@ -147,7 +147,7 @@ static uint16_t combios_get_table_offset(struct drm_device *dev, | |||
147 | enum radeon_combios_table_offset table) | 147 | enum radeon_combios_table_offset table) |
148 | { | 148 | { |
149 | struct radeon_device *rdev = dev->dev_private; | 149 | struct radeon_device *rdev = dev->dev_private; |
150 | int rev; | 150 | int rev, size; |
151 | uint16_t offset = 0, check_offset; | 151 | uint16_t offset = 0, check_offset; |
152 | 152 | ||
153 | if (!rdev->bios) | 153 | if (!rdev->bios) |
@@ -156,174 +156,106 @@ static uint16_t combios_get_table_offset(struct drm_device *dev, | |||
156 | switch (table) { | 156 | switch (table) { |
157 | /* absolute offset tables */ | 157 | /* absolute offset tables */ |
158 | case COMBIOS_ASIC_INIT_1_TABLE: | 158 | case COMBIOS_ASIC_INIT_1_TABLE: |
159 | check_offset = RBIOS16(rdev->bios_header_start + 0xc); | 159 | check_offset = 0xc; |
160 | if (check_offset) | ||
161 | offset = check_offset; | ||
162 | break; | 160 | break; |
163 | case COMBIOS_BIOS_SUPPORT_TABLE: | 161 | case COMBIOS_BIOS_SUPPORT_TABLE: |
164 | check_offset = RBIOS16(rdev->bios_header_start + 0x14); | 162 | check_offset = 0x14; |
165 | if (check_offset) | ||
166 | offset = check_offset; | ||
167 | break; | 163 | break; |
168 | case COMBIOS_DAC_PROGRAMMING_TABLE: | 164 | case COMBIOS_DAC_PROGRAMMING_TABLE: |
169 | check_offset = RBIOS16(rdev->bios_header_start + 0x2a); | 165 | check_offset = 0x2a; |
170 | if (check_offset) | ||
171 | offset = check_offset; | ||
172 | break; | 166 | break; |
173 | case COMBIOS_MAX_COLOR_DEPTH_TABLE: | 167 | case COMBIOS_MAX_COLOR_DEPTH_TABLE: |
174 | check_offset = RBIOS16(rdev->bios_header_start + 0x2c); | 168 | check_offset = 0x2c; |
175 | if (check_offset) | ||
176 | offset = check_offset; | ||
177 | break; | 169 | break; |
178 | case COMBIOS_CRTC_INFO_TABLE: | 170 | case COMBIOS_CRTC_INFO_TABLE: |
179 | check_offset = RBIOS16(rdev->bios_header_start + 0x2e); | 171 | check_offset = 0x2e; |
180 | if (check_offset) | ||
181 | offset = check_offset; | ||
182 | break; | 172 | break; |
183 | case COMBIOS_PLL_INFO_TABLE: | 173 | case COMBIOS_PLL_INFO_TABLE: |
184 | check_offset = RBIOS16(rdev->bios_header_start + 0x30); | 174 | check_offset = 0x30; |
185 | if (check_offset) | ||
186 | offset = check_offset; | ||
187 | break; | 175 | break; |
188 | case COMBIOS_TV_INFO_TABLE: | 176 | case COMBIOS_TV_INFO_TABLE: |
189 | check_offset = RBIOS16(rdev->bios_header_start + 0x32); | 177 | check_offset = 0x32; |
190 | if (check_offset) | ||
191 | offset = check_offset; | ||
192 | break; | 178 | break; |
193 | case COMBIOS_DFP_INFO_TABLE: | 179 | case COMBIOS_DFP_INFO_TABLE: |
194 | check_offset = RBIOS16(rdev->bios_header_start + 0x34); | 180 | check_offset = 0x34; |
195 | if (check_offset) | ||
196 | offset = check_offset; | ||
197 | break; | 181 | break; |
198 | case COMBIOS_HW_CONFIG_INFO_TABLE: | 182 | case COMBIOS_HW_CONFIG_INFO_TABLE: |
199 | check_offset = RBIOS16(rdev->bios_header_start + 0x36); | 183 | check_offset = 0x36; |
200 | if (check_offset) | ||
201 | offset = check_offset; | ||
202 | break; | 184 | break; |
203 | case COMBIOS_MULTIMEDIA_INFO_TABLE: | 185 | case COMBIOS_MULTIMEDIA_INFO_TABLE: |
204 | check_offset = RBIOS16(rdev->bios_header_start + 0x38); | 186 | check_offset = 0x38; |
205 | if (check_offset) | ||
206 | offset = check_offset; | ||
207 | break; | 187 | break; |
208 | case COMBIOS_TV_STD_PATCH_TABLE: | 188 | case COMBIOS_TV_STD_PATCH_TABLE: |
209 | check_offset = RBIOS16(rdev->bios_header_start + 0x3e); | 189 | check_offset = 0x3e; |
210 | if (check_offset) | ||
211 | offset = check_offset; | ||
212 | break; | 190 | break; |
213 | case COMBIOS_LCD_INFO_TABLE: | 191 | case COMBIOS_LCD_INFO_TABLE: |
214 | check_offset = RBIOS16(rdev->bios_header_start + 0x40); | 192 | check_offset = 0x40; |
215 | if (check_offset) | ||
216 | offset = check_offset; | ||
217 | break; | 193 | break; |
218 | case COMBIOS_MOBILE_INFO_TABLE: | 194 | case COMBIOS_MOBILE_INFO_TABLE: |
219 | check_offset = RBIOS16(rdev->bios_header_start + 0x42); | 195 | check_offset = 0x42; |
220 | if (check_offset) | ||
221 | offset = check_offset; | ||
222 | break; | 196 | break; |
223 | case COMBIOS_PLL_INIT_TABLE: | 197 | case COMBIOS_PLL_INIT_TABLE: |
224 | check_offset = RBIOS16(rdev->bios_header_start + 0x46); | 198 | check_offset = 0x46; |
225 | if (check_offset) | ||
226 | offset = check_offset; | ||
227 | break; | 199 | break; |
228 | case COMBIOS_MEM_CONFIG_TABLE: | 200 | case COMBIOS_MEM_CONFIG_TABLE: |
229 | check_offset = RBIOS16(rdev->bios_header_start + 0x48); | 201 | check_offset = 0x48; |
230 | if (check_offset) | ||
231 | offset = check_offset; | ||
232 | break; | 202 | break; |
233 | case COMBIOS_SAVE_MASK_TABLE: | 203 | case COMBIOS_SAVE_MASK_TABLE: |
234 | check_offset = RBIOS16(rdev->bios_header_start + 0x4a); | 204 | check_offset = 0x4a; |
235 | if (check_offset) | ||
236 | offset = check_offset; | ||
237 | break; | 205 | break; |
238 | case COMBIOS_HARDCODED_EDID_TABLE: | 206 | case COMBIOS_HARDCODED_EDID_TABLE: |
239 | check_offset = RBIOS16(rdev->bios_header_start + 0x4c); | 207 | check_offset = 0x4c; |
240 | if (check_offset) | ||
241 | offset = check_offset; | ||
242 | break; | 208 | break; |
243 | case COMBIOS_ASIC_INIT_2_TABLE: | 209 | case COMBIOS_ASIC_INIT_2_TABLE: |
244 | check_offset = RBIOS16(rdev->bios_header_start + 0x4e); | 210 | check_offset = 0x4e; |
245 | if (check_offset) | ||
246 | offset = check_offset; | ||
247 | break; | 211 | break; |
248 | case COMBIOS_CONNECTOR_INFO_TABLE: | 212 | case COMBIOS_CONNECTOR_INFO_TABLE: |
249 | check_offset = RBIOS16(rdev->bios_header_start + 0x50); | 213 | check_offset = 0x50; |
250 | if (check_offset) | ||
251 | offset = check_offset; | ||
252 | break; | 214 | break; |
253 | case COMBIOS_DYN_CLK_1_TABLE: | 215 | case COMBIOS_DYN_CLK_1_TABLE: |
254 | check_offset = RBIOS16(rdev->bios_header_start + 0x52); | 216 | check_offset = 0x52; |
255 | if (check_offset) | ||
256 | offset = check_offset; | ||
257 | break; | 217 | break; |
258 | case COMBIOS_RESERVED_MEM_TABLE: | 218 | case COMBIOS_RESERVED_MEM_TABLE: |
259 | check_offset = RBIOS16(rdev->bios_header_start + 0x54); | 219 | check_offset = 0x54; |
260 | if (check_offset) | ||
261 | offset = check_offset; | ||
262 | break; | 220 | break; |
263 | case COMBIOS_EXT_TMDS_INFO_TABLE: | 221 | case COMBIOS_EXT_TMDS_INFO_TABLE: |
264 | check_offset = RBIOS16(rdev->bios_header_start + 0x58); | 222 | check_offset = 0x58; |
265 | if (check_offset) | ||
266 | offset = check_offset; | ||
267 | break; | 223 | break; |
268 | case COMBIOS_MEM_CLK_INFO_TABLE: | 224 | case COMBIOS_MEM_CLK_INFO_TABLE: |
269 | check_offset = RBIOS16(rdev->bios_header_start + 0x5a); | 225 | check_offset = 0x5a; |
270 | if (check_offset) | ||
271 | offset = check_offset; | ||
272 | break; | 226 | break; |
273 | case COMBIOS_EXT_DAC_INFO_TABLE: | 227 | case COMBIOS_EXT_DAC_INFO_TABLE: |
274 | check_offset = RBIOS16(rdev->bios_header_start + 0x5c); | 228 | check_offset = 0x5c; |
275 | if (check_offset) | ||
276 | offset = check_offset; | ||
277 | break; | 229 | break; |
278 | case COMBIOS_MISC_INFO_TABLE: | 230 | case COMBIOS_MISC_INFO_TABLE: |
279 | check_offset = RBIOS16(rdev->bios_header_start + 0x5e); | 231 | check_offset = 0x5e; |
280 | if (check_offset) | ||
281 | offset = check_offset; | ||
282 | break; | 232 | break; |
283 | case COMBIOS_CRT_INFO_TABLE: | 233 | case COMBIOS_CRT_INFO_TABLE: |
284 | check_offset = RBIOS16(rdev->bios_header_start + 0x60); | 234 | check_offset = 0x60; |
285 | if (check_offset) | ||
286 | offset = check_offset; | ||
287 | break; | 235 | break; |
288 | case COMBIOS_INTEGRATED_SYSTEM_INFO_TABLE: | 236 | case COMBIOS_INTEGRATED_SYSTEM_INFO_TABLE: |
289 | check_offset = RBIOS16(rdev->bios_header_start + 0x62); | 237 | check_offset = 0x62; |
290 | if (check_offset) | ||
291 | offset = check_offset; | ||
292 | break; | 238 | break; |
293 | case COMBIOS_COMPONENT_VIDEO_INFO_TABLE: | 239 | case COMBIOS_COMPONENT_VIDEO_INFO_TABLE: |
294 | check_offset = RBIOS16(rdev->bios_header_start + 0x64); | 240 | check_offset = 0x64; |
295 | if (check_offset) | ||
296 | offset = check_offset; | ||
297 | break; | 241 | break; |
298 | case COMBIOS_FAN_SPEED_INFO_TABLE: | 242 | case COMBIOS_FAN_SPEED_INFO_TABLE: |
299 | check_offset = RBIOS16(rdev->bios_header_start + 0x66); | 243 | check_offset = 0x66; |
300 | if (check_offset) | ||
301 | offset = check_offset; | ||
302 | break; | 244 | break; |
303 | case COMBIOS_OVERDRIVE_INFO_TABLE: | 245 | case COMBIOS_OVERDRIVE_INFO_TABLE: |
304 | check_offset = RBIOS16(rdev->bios_header_start + 0x68); | 246 | check_offset = 0x68; |
305 | if (check_offset) | ||
306 | offset = check_offset; | ||
307 | break; | 247 | break; |
308 | case COMBIOS_OEM_INFO_TABLE: | 248 | case COMBIOS_OEM_INFO_TABLE: |
309 | check_offset = RBIOS16(rdev->bios_header_start + 0x6a); | 249 | check_offset = 0x6a; |
310 | if (check_offset) | ||
311 | offset = check_offset; | ||
312 | break; | 250 | break; |
313 | case COMBIOS_DYN_CLK_2_TABLE: | 251 | case COMBIOS_DYN_CLK_2_TABLE: |
314 | check_offset = RBIOS16(rdev->bios_header_start + 0x6c); | 252 | check_offset = 0x6c; |
315 | if (check_offset) | ||
316 | offset = check_offset; | ||
317 | break; | 253 | break; |
318 | case COMBIOS_POWER_CONNECTOR_INFO_TABLE: | 254 | case COMBIOS_POWER_CONNECTOR_INFO_TABLE: |
319 | check_offset = RBIOS16(rdev->bios_header_start + 0x6e); | 255 | check_offset = 0x6e; |
320 | if (check_offset) | ||
321 | offset = check_offset; | ||
322 | break; | 256 | break; |
323 | case COMBIOS_I2C_INFO_TABLE: | 257 | case COMBIOS_I2C_INFO_TABLE: |
324 | check_offset = RBIOS16(rdev->bios_header_start + 0x70); | 258 | check_offset = 0x70; |
325 | if (check_offset) | ||
326 | offset = check_offset; | ||
327 | break; | 259 | break; |
328 | /* relative offset tables */ | 260 | /* relative offset tables */ |
329 | case COMBIOS_ASIC_INIT_3_TABLE: /* offset from misc info */ | 261 | case COMBIOS_ASIC_INIT_3_TABLE: /* offset from misc info */ |
@@ -439,11 +371,16 @@ static uint16_t combios_get_table_offset(struct drm_device *dev, | |||
439 | } | 371 | } |
440 | break; | 372 | break; |
441 | default: | 373 | default: |
374 | check_offset = 0; | ||
442 | break; | 375 | break; |
443 | } | 376 | } |
444 | 377 | ||
445 | return offset; | 378 | size = RBIOS8(rdev->bios_header_start + 0x6); |
379 | /* check absolute offset tables */ | ||
380 | if (table < COMBIOS_ASIC_INIT_3_TABLE && check_offset && check_offset < size) | ||
381 | offset = RBIOS16(rdev->bios_header_start + check_offset); | ||
446 | 382 | ||
383 | return offset; | ||
447 | } | 384 | } |
448 | 385 | ||
449 | bool radeon_combios_check_hardcoded_edid(struct radeon_device *rdev) | 386 | bool radeon_combios_check_hardcoded_edid(struct radeon_device *rdev) |
@@ -965,16 +902,22 @@ struct radeon_encoder_primary_dac *radeon_combios_get_primary_dac_info(struct | |||
965 | dac = RBIOS8(dac_info + 0x3) & 0xf; | 902 | dac = RBIOS8(dac_info + 0x3) & 0xf; |
966 | p_dac->ps2_pdac_adj = (bg << 8) | (dac); | 903 | p_dac->ps2_pdac_adj = (bg << 8) | (dac); |
967 | } | 904 | } |
968 | /* if the values are all zeros, use the table */ | 905 | /* if the values are zeros, use the table */ |
969 | if (p_dac->ps2_pdac_adj) | 906 | if ((dac == 0) || (bg == 0)) |
907 | found = 0; | ||
908 | else | ||
970 | found = 1; | 909 | found = 1; |
971 | } | 910 | } |
972 | 911 | ||
973 | /* quirks */ | 912 | /* quirks */ |
913 | /* Radeon 7000 (RV100) */ | ||
914 | if (((dev->pdev->device == 0x5159) && | ||
915 | (dev->pdev->subsystem_vendor == 0x174B) && | ||
916 | (dev->pdev->subsystem_device == 0x7c28)) || | ||
974 | /* Radeon 9100 (R200) */ | 917 | /* Radeon 9100 (R200) */ |
975 | if ((dev->pdev->device == 0x514D) && | 918 | ((dev->pdev->device == 0x514D) && |
976 | (dev->pdev->subsystem_vendor == 0x174B) && | 919 | (dev->pdev->subsystem_vendor == 0x174B) && |
977 | (dev->pdev->subsystem_device == 0x7149)) { | 920 | (dev->pdev->subsystem_device == 0x7149))) { |
978 | /* vbios value is bad, use the default */ | 921 | /* vbios value is bad, use the default */ |
979 | found = 0; | 922 | found = 0; |
980 | } | 923 | } |
diff --git a/drivers/gpu/drm/radeon/radeon_gart.c b/drivers/gpu/drm/radeon/radeon_gart.c index d9d31a383276..6a51d943ccf4 100644 --- a/drivers/gpu/drm/radeon/radeon_gart.c +++ b/drivers/gpu/drm/radeon/radeon_gart.c | |||
@@ -466,7 +466,7 @@ int radeon_vm_manager_init(struct radeon_device *rdev) | |||
466 | size += rdev->vm_manager.max_pfn * 8; | 466 | size += rdev->vm_manager.max_pfn * 8; |
467 | size *= 2; | 467 | size *= 2; |
468 | r = radeon_sa_bo_manager_init(rdev, &rdev->vm_manager.sa_manager, | 468 | r = radeon_sa_bo_manager_init(rdev, &rdev->vm_manager.sa_manager, |
469 | RADEON_VM_PTB_ALIGN(size), | 469 | RADEON_GPU_PAGE_ALIGN(size), |
470 | RADEON_VM_PTB_ALIGN_SIZE, | 470 | RADEON_VM_PTB_ALIGN_SIZE, |
471 | RADEON_GEM_DOMAIN_VRAM); | 471 | RADEON_GEM_DOMAIN_VRAM); |
472 | if (r) { | 472 | if (r) { |
@@ -621,7 +621,7 @@ int radeon_vm_alloc_pt(struct radeon_device *rdev, struct radeon_vm *vm) | |||
621 | } | 621 | } |
622 | 622 | ||
623 | retry: | 623 | retry: |
624 | pd_size = RADEON_VM_PTB_ALIGN(radeon_vm_directory_size(rdev)); | 624 | pd_size = radeon_vm_directory_size(rdev); |
625 | r = radeon_sa_bo_new(rdev, &rdev->vm_manager.sa_manager, | 625 | r = radeon_sa_bo_new(rdev, &rdev->vm_manager.sa_manager, |
626 | &vm->page_directory, pd_size, | 626 | &vm->page_directory, pd_size, |
627 | RADEON_VM_PTB_ALIGN_SIZE, false); | 627 | RADEON_VM_PTB_ALIGN_SIZE, false); |
@@ -953,8 +953,8 @@ static int radeon_vm_update_pdes(struct radeon_device *rdev, | |||
953 | retry: | 953 | retry: |
954 | r = radeon_sa_bo_new(rdev, &rdev->vm_manager.sa_manager, | 954 | r = radeon_sa_bo_new(rdev, &rdev->vm_manager.sa_manager, |
955 | &vm->page_tables[pt_idx], | 955 | &vm->page_tables[pt_idx], |
956 | RADEON_VM_PTB_ALIGN(RADEON_VM_PTE_COUNT * 8), | 956 | RADEON_VM_PTE_COUNT * 8, |
957 | RADEON_VM_PTB_ALIGN_SIZE, false); | 957 | RADEON_GPU_PAGE_SIZE, false); |
958 | 958 | ||
959 | if (r == -ENOMEM) { | 959 | if (r == -ENOMEM) { |
960 | r = radeon_vm_evict(rdev, vm); | 960 | r = radeon_vm_evict(rdev, vm); |
diff --git a/drivers/gpu/drm/radeon/rv6xx_dpm.c b/drivers/gpu/drm/radeon/rv6xx_dpm.c index 65e33f387341..363018c60412 100644 --- a/drivers/gpu/drm/radeon/rv6xx_dpm.c +++ b/drivers/gpu/drm/radeon/rv6xx_dpm.c | |||
@@ -819,7 +819,7 @@ static void rv6xx_program_memory_timing_parameters(struct radeon_device *rdev) | |||
819 | POWERMODE1(calculate_memory_refresh_rate(rdev, | 819 | POWERMODE1(calculate_memory_refresh_rate(rdev, |
820 | pi->hw.sclks[R600_POWER_LEVEL_MEDIUM])) | | 820 | pi->hw.sclks[R600_POWER_LEVEL_MEDIUM])) | |
821 | POWERMODE2(calculate_memory_refresh_rate(rdev, | 821 | POWERMODE2(calculate_memory_refresh_rate(rdev, |
822 | pi->hw.sclks[R600_POWER_LEVEL_MEDIUM])) | | 822 | pi->hw.sclks[R600_POWER_LEVEL_HIGH])) | |
823 | POWERMODE3(calculate_memory_refresh_rate(rdev, | 823 | POWERMODE3(calculate_memory_refresh_rate(rdev, |
824 | pi->hw.sclks[R600_POWER_LEVEL_HIGH]))); | 824 | pi->hw.sclks[R600_POWER_LEVEL_HIGH]))); |
825 | WREG32(ARB_RFSH_RATE, arb_refresh_rate); | 825 | WREG32(ARB_RFSH_RATE, arb_refresh_rate); |
@@ -1182,10 +1182,10 @@ static void rv6xx_program_display_gap(struct radeon_device *rdev) | |||
1182 | u32 tmp = RREG32(CG_DISPLAY_GAP_CNTL); | 1182 | u32 tmp = RREG32(CG_DISPLAY_GAP_CNTL); |
1183 | 1183 | ||
1184 | tmp &= ~(DISP1_GAP_MCHG_MASK | DISP2_GAP_MCHG_MASK); | 1184 | tmp &= ~(DISP1_GAP_MCHG_MASK | DISP2_GAP_MCHG_MASK); |
1185 | if (RREG32(AVIVO_D1CRTC_CONTROL) & AVIVO_CRTC_EN) { | 1185 | if (rdev->pm.dpm.new_active_crtcs & 1) { |
1186 | tmp |= DISP1_GAP_MCHG(R600_PM_DISPLAY_GAP_VBLANK); | 1186 | tmp |= DISP1_GAP_MCHG(R600_PM_DISPLAY_GAP_VBLANK); |
1187 | tmp |= DISP2_GAP_MCHG(R600_PM_DISPLAY_GAP_IGNORE); | 1187 | tmp |= DISP2_GAP_MCHG(R600_PM_DISPLAY_GAP_IGNORE); |
1188 | } else if (RREG32(AVIVO_D2CRTC_CONTROL) & AVIVO_CRTC_EN) { | 1188 | } else if (rdev->pm.dpm.new_active_crtcs & 2) { |
1189 | tmp |= DISP1_GAP_MCHG(R600_PM_DISPLAY_GAP_IGNORE); | 1189 | tmp |= DISP1_GAP_MCHG(R600_PM_DISPLAY_GAP_IGNORE); |
1190 | tmp |= DISP2_GAP_MCHG(R600_PM_DISPLAY_GAP_VBLANK); | 1190 | tmp |= DISP2_GAP_MCHG(R600_PM_DISPLAY_GAP_VBLANK); |
1191 | } else { | 1191 | } else { |
@@ -1670,6 +1670,8 @@ int rv6xx_dpm_set_power_state(struct radeon_device *rdev) | |||
1670 | struct radeon_ps *old_ps = rdev->pm.dpm.current_ps; | 1670 | struct radeon_ps *old_ps = rdev->pm.dpm.current_ps; |
1671 | int ret; | 1671 | int ret; |
1672 | 1672 | ||
1673 | pi->restricted_levels = 0; | ||
1674 | |||
1673 | rv6xx_set_uvd_clock_before_set_eng_clock(rdev, new_ps, old_ps); | 1675 | rv6xx_set_uvd_clock_before_set_eng_clock(rdev, new_ps, old_ps); |
1674 | 1676 | ||
1675 | rv6xx_clear_vc(rdev); | 1677 | rv6xx_clear_vc(rdev); |
@@ -1756,6 +1758,8 @@ int rv6xx_dpm_set_power_state(struct radeon_device *rdev) | |||
1756 | 1758 | ||
1757 | rv6xx_set_uvd_clock_after_set_eng_clock(rdev, new_ps, old_ps); | 1759 | rv6xx_set_uvd_clock_after_set_eng_clock(rdev, new_ps, old_ps); |
1758 | 1760 | ||
1761 | rdev->pm.dpm.forced_level = RADEON_DPM_FORCED_LEVEL_AUTO; | ||
1762 | |||
1759 | return 0; | 1763 | return 0; |
1760 | } | 1764 | } |
1761 | 1765 | ||
@@ -2085,3 +2089,34 @@ u32 rv6xx_dpm_get_mclk(struct radeon_device *rdev, bool low) | |||
2085 | else | 2089 | else |
2086 | return requested_state->high.mclk; | 2090 | return requested_state->high.mclk; |
2087 | } | 2091 | } |
2092 | |||
2093 | int rv6xx_dpm_force_performance_level(struct radeon_device *rdev, | ||
2094 | enum radeon_dpm_forced_level level) | ||
2095 | { | ||
2096 | struct rv6xx_power_info *pi = rv6xx_get_pi(rdev); | ||
2097 | |||
2098 | if (level == RADEON_DPM_FORCED_LEVEL_HIGH) { | ||
2099 | pi->restricted_levels = 3; | ||
2100 | } else if (level == RADEON_DPM_FORCED_LEVEL_LOW) { | ||
2101 | pi->restricted_levels = 2; | ||
2102 | } else { | ||
2103 | pi->restricted_levels = 0; | ||
2104 | } | ||
2105 | |||
2106 | rv6xx_clear_vc(rdev); | ||
2107 | r600_power_level_enable(rdev, R600_POWER_LEVEL_LOW, true); | ||
2108 | r600_set_at(rdev, 0xFFFF, 0xFFFF, 0xFFFF, 0xFFFF); | ||
2109 | r600_wait_for_power_level(rdev, R600_POWER_LEVEL_LOW); | ||
2110 | r600_power_level_enable(rdev, R600_POWER_LEVEL_HIGH, false); | ||
2111 | r600_power_level_enable(rdev, R600_POWER_LEVEL_MEDIUM, false); | ||
2112 | rv6xx_enable_medium(rdev); | ||
2113 | rv6xx_enable_high(rdev); | ||
2114 | if (pi->restricted_levels == 3) | ||
2115 | r600_power_level_enable(rdev, R600_POWER_LEVEL_LOW, false); | ||
2116 | rv6xx_program_vc(rdev); | ||
2117 | rv6xx_program_at(rdev); | ||
2118 | |||
2119 | rdev->pm.dpm.forced_level = level; | ||
2120 | |||
2121 | return 0; | ||
2122 | } | ||
diff --git a/drivers/hv/hv_balloon.c b/drivers/hv/hv_balloon.c index 4c605c70ebf9..deb5c25305af 100644 --- a/drivers/hv/hv_balloon.c +++ b/drivers/hv/hv_balloon.c | |||
@@ -562,7 +562,7 @@ static void hv_mem_hot_add(unsigned long start, unsigned long size, | |||
562 | struct hv_hotadd_state *has) | 562 | struct hv_hotadd_state *has) |
563 | { | 563 | { |
564 | int ret = 0; | 564 | int ret = 0; |
565 | int i, nid, t; | 565 | int i, nid; |
566 | unsigned long start_pfn; | 566 | unsigned long start_pfn; |
567 | unsigned long processed_pfn; | 567 | unsigned long processed_pfn; |
568 | unsigned long total_pfn = pfn_count; | 568 | unsigned long total_pfn = pfn_count; |
@@ -607,14 +607,11 @@ static void hv_mem_hot_add(unsigned long start, unsigned long size, | |||
607 | 607 | ||
608 | /* | 608 | /* |
609 | * Wait for the memory block to be onlined. | 609 | * Wait for the memory block to be onlined. |
610 | * Since the hot add has succeeded, it is ok to | ||
611 | * proceed even if the pages in the hot added region | ||
612 | * have not been "onlined" within the allowed time. | ||
610 | */ | 613 | */ |
611 | t = wait_for_completion_timeout(&dm_device.ol_waitevent, 5*HZ); | 614 | wait_for_completion_timeout(&dm_device.ol_waitevent, 5*HZ); |
612 | if (t == 0) { | ||
613 | pr_info("hot_add memory timedout\n"); | ||
614 | has->ha_end_pfn -= HA_CHUNK; | ||
615 | has->covered_end_pfn -= processed_pfn; | ||
616 | break; | ||
617 | } | ||
618 | 615 | ||
619 | } | 616 | } |
620 | 617 | ||
@@ -978,6 +975,14 @@ static void post_status(struct hv_dynmem_device *dm) | |||
978 | dm->num_pages_ballooned + | 975 | dm->num_pages_ballooned + |
979 | compute_balloon_floor(); | 976 | compute_balloon_floor(); |
980 | 977 | ||
978 | /* | ||
979 | * If our transaction ID is no longer current, just don't | ||
980 | * send the status. This can happen if we were interrupted | ||
981 | * after we picked our transaction ID. | ||
982 | */ | ||
983 | if (status.hdr.trans_id != atomic_read(&trans_id)) | ||
984 | return; | ||
985 | |||
981 | vmbus_sendpacket(dm->dev->channel, &status, | 986 | vmbus_sendpacket(dm->dev->channel, &status, |
982 | sizeof(struct dm_status), | 987 | sizeof(struct dm_status), |
983 | (unsigned long)NULL, | 988 | (unsigned long)NULL, |
diff --git a/drivers/hv/vmbus_drv.c b/drivers/hv/vmbus_drv.c index a2464bf07c49..e8e071fc1d6d 100644 --- a/drivers/hv/vmbus_drv.c +++ b/drivers/hv/vmbus_drv.c | |||
@@ -690,7 +690,7 @@ int vmbus_device_register(struct hv_device *child_device_obj) | |||
690 | if (ret) | 690 | if (ret) |
691 | pr_err("Unable to register child device\n"); | 691 | pr_err("Unable to register child device\n"); |
692 | else | 692 | else |
693 | pr_info("child device %s registered\n", | 693 | pr_debug("child device %s registered\n", |
694 | dev_name(&child_device_obj->device)); | 694 | dev_name(&child_device_obj->device)); |
695 | 695 | ||
696 | return ret; | 696 | return ret; |
@@ -702,14 +702,14 @@ int vmbus_device_register(struct hv_device *child_device_obj) | |||
702 | */ | 702 | */ |
703 | void vmbus_device_unregister(struct hv_device *device_obj) | 703 | void vmbus_device_unregister(struct hv_device *device_obj) |
704 | { | 704 | { |
705 | pr_debug("child device %s unregistered\n", | ||
706 | dev_name(&device_obj->device)); | ||
707 | |||
705 | /* | 708 | /* |
706 | * Kick off the process of unregistering the device. | 709 | * Kick off the process of unregistering the device. |
707 | * This will call vmbus_remove() and eventually vmbus_device_release() | 710 | * This will call vmbus_remove() and eventually vmbus_device_release() |
708 | */ | 711 | */ |
709 | device_unregister(&device_obj->device); | 712 | device_unregister(&device_obj->device); |
710 | |||
711 | pr_info("child device %s unregistered\n", | ||
712 | dev_name(&device_obj->device)); | ||
713 | } | 713 | } |
714 | 714 | ||
715 | 715 | ||
diff --git a/drivers/md/bcache/alloc.c b/drivers/md/bcache/alloc.c index 048f2947e08b..e45f5575fd4d 100644 --- a/drivers/md/bcache/alloc.c +++ b/drivers/md/bcache/alloc.c | |||
@@ -63,7 +63,10 @@ | |||
63 | #include "bcache.h" | 63 | #include "bcache.h" |
64 | #include "btree.h" | 64 | #include "btree.h" |
65 | 65 | ||
66 | #include <linux/freezer.h> | ||
67 | #include <linux/kthread.h> | ||
66 | #include <linux/random.h> | 68 | #include <linux/random.h> |
69 | #include <trace/events/bcache.h> | ||
67 | 70 | ||
68 | #define MAX_IN_FLIGHT_DISCARDS 8U | 71 | #define MAX_IN_FLIGHT_DISCARDS 8U |
69 | 72 | ||
@@ -151,7 +154,7 @@ static void discard_finish(struct work_struct *w) | |||
151 | mutex_unlock(&ca->set->bucket_lock); | 154 | mutex_unlock(&ca->set->bucket_lock); |
152 | 155 | ||
153 | closure_wake_up(&ca->set->bucket_wait); | 156 | closure_wake_up(&ca->set->bucket_wait); |
154 | wake_up(&ca->set->alloc_wait); | 157 | wake_up_process(ca->alloc_thread); |
155 | 158 | ||
156 | closure_put(&ca->set->cl); | 159 | closure_put(&ca->set->cl); |
157 | } | 160 | } |
@@ -350,38 +353,30 @@ static void invalidate_buckets(struct cache *ca) | |||
350 | break; | 353 | break; |
351 | } | 354 | } |
352 | 355 | ||
353 | pr_debug("free %zu/%zu free_inc %zu/%zu unused %zu/%zu", | 356 | trace_bcache_alloc_invalidate(ca); |
354 | fifo_used(&ca->free), ca->free.size, | ||
355 | fifo_used(&ca->free_inc), ca->free_inc.size, | ||
356 | fifo_used(&ca->unused), ca->unused.size); | ||
357 | } | 357 | } |
358 | 358 | ||
359 | #define allocator_wait(ca, cond) \ | 359 | #define allocator_wait(ca, cond) \ |
360 | do { \ | 360 | do { \ |
361 | DEFINE_WAIT(__wait); \ | ||
362 | \ | ||
363 | while (1) { \ | 361 | while (1) { \ |
364 | prepare_to_wait(&ca->set->alloc_wait, \ | 362 | set_current_state(TASK_INTERRUPTIBLE); \ |
365 | &__wait, TASK_INTERRUPTIBLE); \ | ||
366 | if (cond) \ | 363 | if (cond) \ |
367 | break; \ | 364 | break; \ |
368 | \ | 365 | \ |
369 | mutex_unlock(&(ca)->set->bucket_lock); \ | 366 | mutex_unlock(&(ca)->set->bucket_lock); \ |
370 | if (test_bit(CACHE_SET_STOPPING_2, &ca->set->flags)) { \ | 367 | if (kthread_should_stop()) \ |
371 | finish_wait(&ca->set->alloc_wait, &__wait); \ | 368 | return 0; \ |
372 | closure_return(cl); \ | ||
373 | } \ | ||
374 | \ | 369 | \ |
370 | try_to_freeze(); \ | ||
375 | schedule(); \ | 371 | schedule(); \ |
376 | mutex_lock(&(ca)->set->bucket_lock); \ | 372 | mutex_lock(&(ca)->set->bucket_lock); \ |
377 | } \ | 373 | } \ |
378 | \ | 374 | __set_current_state(TASK_RUNNING); \ |
379 | finish_wait(&ca->set->alloc_wait, &__wait); \ | ||
380 | } while (0) | 375 | } while (0) |
381 | 376 | ||
382 | void bch_allocator_thread(struct closure *cl) | 377 | static int bch_allocator_thread(void *arg) |
383 | { | 378 | { |
384 | struct cache *ca = container_of(cl, struct cache, alloc); | 379 | struct cache *ca = arg; |
385 | 380 | ||
386 | mutex_lock(&ca->set->bucket_lock); | 381 | mutex_lock(&ca->set->bucket_lock); |
387 | 382 | ||
@@ -442,7 +437,7 @@ long bch_bucket_alloc(struct cache *ca, unsigned watermark, struct closure *cl) | |||
442 | { | 437 | { |
443 | long r = -1; | 438 | long r = -1; |
444 | again: | 439 | again: |
445 | wake_up(&ca->set->alloc_wait); | 440 | wake_up_process(ca->alloc_thread); |
446 | 441 | ||
447 | if (fifo_used(&ca->free) > ca->watermark[watermark] && | 442 | if (fifo_used(&ca->free) > ca->watermark[watermark] && |
448 | fifo_pop(&ca->free, r)) { | 443 | fifo_pop(&ca->free, r)) { |
@@ -476,9 +471,7 @@ again: | |||
476 | return r; | 471 | return r; |
477 | } | 472 | } |
478 | 473 | ||
479 | pr_debug("alloc failure: blocked %i free %zu free_inc %zu unused %zu", | 474 | trace_bcache_alloc_fail(ca); |
480 | atomic_read(&ca->set->prio_blocked), fifo_used(&ca->free), | ||
481 | fifo_used(&ca->free_inc), fifo_used(&ca->unused)); | ||
482 | 475 | ||
483 | if (cl) { | 476 | if (cl) { |
484 | closure_wait(&ca->set->bucket_wait, cl); | 477 | closure_wait(&ca->set->bucket_wait, cl); |
@@ -552,6 +545,17 @@ int bch_bucket_alloc_set(struct cache_set *c, unsigned watermark, | |||
552 | 545 | ||
553 | /* Init */ | 546 | /* Init */ |
554 | 547 | ||
548 | int bch_cache_allocator_start(struct cache *ca) | ||
549 | { | ||
550 | struct task_struct *k = kthread_run(bch_allocator_thread, | ||
551 | ca, "bcache_allocator"); | ||
552 | if (IS_ERR(k)) | ||
553 | return PTR_ERR(k); | ||
554 | |||
555 | ca->alloc_thread = k; | ||
556 | return 0; | ||
557 | } | ||
558 | |||
555 | void bch_cache_allocator_exit(struct cache *ca) | 559 | void bch_cache_allocator_exit(struct cache *ca) |
556 | { | 560 | { |
557 | struct discard *d; | 561 | struct discard *d; |
diff --git a/drivers/md/bcache/bcache.h b/drivers/md/bcache/bcache.h index d3e15b42a4ab..b39f6f0b45f2 100644 --- a/drivers/md/bcache/bcache.h +++ b/drivers/md/bcache/bcache.h | |||
@@ -178,7 +178,6 @@ | |||
178 | #define pr_fmt(fmt) "bcache: %s() " fmt "\n", __func__ | 178 | #define pr_fmt(fmt) "bcache: %s() " fmt "\n", __func__ |
179 | 179 | ||
180 | #include <linux/bio.h> | 180 | #include <linux/bio.h> |
181 | #include <linux/blktrace_api.h> | ||
182 | #include <linux/kobject.h> | 181 | #include <linux/kobject.h> |
183 | #include <linux/list.h> | 182 | #include <linux/list.h> |
184 | #include <linux/mutex.h> | 183 | #include <linux/mutex.h> |
@@ -388,8 +387,6 @@ struct keybuf_key { | |||
388 | typedef bool (keybuf_pred_fn)(struct keybuf *, struct bkey *); | 387 | typedef bool (keybuf_pred_fn)(struct keybuf *, struct bkey *); |
389 | 388 | ||
390 | struct keybuf { | 389 | struct keybuf { |
391 | keybuf_pred_fn *key_predicate; | ||
392 | |||
393 | struct bkey last_scanned; | 390 | struct bkey last_scanned; |
394 | spinlock_t lock; | 391 | spinlock_t lock; |
395 | 392 | ||
@@ -437,9 +434,12 @@ struct bcache_device { | |||
437 | 434 | ||
438 | /* If nonzero, we're detaching/unregistering from cache set */ | 435 | /* If nonzero, we're detaching/unregistering from cache set */ |
439 | atomic_t detaching; | 436 | atomic_t detaching; |
437 | int flush_done; | ||
438 | |||
439 | uint64_t nr_stripes; | ||
440 | unsigned stripe_size_bits; | ||
441 | atomic_t *stripe_sectors_dirty; | ||
440 | 442 | ||
441 | atomic_long_t sectors_dirty; | ||
442 | unsigned long sectors_dirty_gc; | ||
443 | unsigned long sectors_dirty_last; | 443 | unsigned long sectors_dirty_last; |
444 | long sectors_dirty_derivative; | 444 | long sectors_dirty_derivative; |
445 | 445 | ||
@@ -531,6 +531,7 @@ struct cached_dev { | |||
531 | unsigned sequential_merge:1; | 531 | unsigned sequential_merge:1; |
532 | unsigned verify:1; | 532 | unsigned verify:1; |
533 | 533 | ||
534 | unsigned partial_stripes_expensive:1; | ||
534 | unsigned writeback_metadata:1; | 535 | unsigned writeback_metadata:1; |
535 | unsigned writeback_running:1; | 536 | unsigned writeback_running:1; |
536 | unsigned char writeback_percent; | 537 | unsigned char writeback_percent; |
@@ -565,8 +566,7 @@ struct cache { | |||
565 | 566 | ||
566 | unsigned watermark[WATERMARK_MAX]; | 567 | unsigned watermark[WATERMARK_MAX]; |
567 | 568 | ||
568 | struct closure alloc; | 569 | struct task_struct *alloc_thread; |
569 | struct workqueue_struct *alloc_workqueue; | ||
570 | 570 | ||
571 | struct closure prio; | 571 | struct closure prio; |
572 | struct prio_set *disk_buckets; | 572 | struct prio_set *disk_buckets; |
@@ -664,13 +664,9 @@ struct gc_stat { | |||
664 | * CACHE_SET_STOPPING always gets set first when we're closing down a cache set; | 664 | * CACHE_SET_STOPPING always gets set first when we're closing down a cache set; |
665 | * we'll continue to run normally for awhile with CACHE_SET_STOPPING set (i.e. | 665 | * we'll continue to run normally for awhile with CACHE_SET_STOPPING set (i.e. |
666 | * flushing dirty data). | 666 | * flushing dirty data). |
667 | * | ||
668 | * CACHE_SET_STOPPING_2 gets set at the last phase, when it's time to shut down | ||
669 | * the allocation thread. | ||
670 | */ | 667 | */ |
671 | #define CACHE_SET_UNREGISTERING 0 | 668 | #define CACHE_SET_UNREGISTERING 0 |
672 | #define CACHE_SET_STOPPING 1 | 669 | #define CACHE_SET_STOPPING 1 |
673 | #define CACHE_SET_STOPPING_2 2 | ||
674 | 670 | ||
675 | struct cache_set { | 671 | struct cache_set { |
676 | struct closure cl; | 672 | struct closure cl; |
@@ -703,9 +699,6 @@ struct cache_set { | |||
703 | /* For the btree cache */ | 699 | /* For the btree cache */ |
704 | struct shrinker shrink; | 700 | struct shrinker shrink; |
705 | 701 | ||
706 | /* For the allocator itself */ | ||
707 | wait_queue_head_t alloc_wait; | ||
708 | |||
709 | /* For the btree cache and anything allocation related */ | 702 | /* For the btree cache and anything allocation related */ |
710 | struct mutex bucket_lock; | 703 | struct mutex bucket_lock; |
711 | 704 | ||
@@ -823,10 +816,9 @@ struct cache_set { | |||
823 | 816 | ||
824 | /* | 817 | /* |
825 | * A btree node on disk could have too many bsets for an iterator to fit | 818 | * A btree node on disk could have too many bsets for an iterator to fit |
826 | * on the stack - this is a single element mempool for btree_read_work() | 819 | * on the stack - have to dynamically allocate them |
827 | */ | 820 | */ |
828 | struct mutex fill_lock; | 821 | mempool_t *fill_iter; |
829 | struct btree_iter *fill_iter; | ||
830 | 822 | ||
831 | /* | 823 | /* |
832 | * btree_sort() is a merge sort and requires temporary space - single | 824 | * btree_sort() is a merge sort and requires temporary space - single |
@@ -834,6 +826,7 @@ struct cache_set { | |||
834 | */ | 826 | */ |
835 | struct mutex sort_lock; | 827 | struct mutex sort_lock; |
836 | struct bset *sort; | 828 | struct bset *sort; |
829 | unsigned sort_crit_factor; | ||
837 | 830 | ||
838 | /* List of buckets we're currently writing data to */ | 831 | /* List of buckets we're currently writing data to */ |
839 | struct list_head data_buckets; | 832 | struct list_head data_buckets; |
@@ -906,8 +899,6 @@ static inline unsigned local_clock_us(void) | |||
906 | return local_clock() >> 10; | 899 | return local_clock() >> 10; |
907 | } | 900 | } |
908 | 901 | ||
909 | #define MAX_BSETS 4U | ||
910 | |||
911 | #define BTREE_PRIO USHRT_MAX | 902 | #define BTREE_PRIO USHRT_MAX |
912 | #define INITIAL_PRIO 32768 | 903 | #define INITIAL_PRIO 32768 |
913 | 904 | ||
@@ -1112,23 +1103,6 @@ static inline void __bkey_put(struct cache_set *c, struct bkey *k) | |||
1112 | atomic_dec_bug(&PTR_BUCKET(c, k, i)->pin); | 1103 | atomic_dec_bug(&PTR_BUCKET(c, k, i)->pin); |
1113 | } | 1104 | } |
1114 | 1105 | ||
1115 | /* Blktrace macros */ | ||
1116 | |||
1117 | #define blktrace_msg(c, fmt, ...) \ | ||
1118 | do { \ | ||
1119 | struct request_queue *q = bdev_get_queue(c->bdev); \ | ||
1120 | if (q) \ | ||
1121 | blk_add_trace_msg(q, fmt, ##__VA_ARGS__); \ | ||
1122 | } while (0) | ||
1123 | |||
1124 | #define blktrace_msg_all(s, fmt, ...) \ | ||
1125 | do { \ | ||
1126 | struct cache *_c; \ | ||
1127 | unsigned i; \ | ||
1128 | for_each_cache(_c, (s), i) \ | ||
1129 | blktrace_msg(_c, fmt, ##__VA_ARGS__); \ | ||
1130 | } while (0) | ||
1131 | |||
1132 | static inline void cached_dev_put(struct cached_dev *dc) | 1106 | static inline void cached_dev_put(struct cached_dev *dc) |
1133 | { | 1107 | { |
1134 | if (atomic_dec_and_test(&dc->count)) | 1108 | if (atomic_dec_and_test(&dc->count)) |
@@ -1173,10 +1147,16 @@ static inline uint8_t bucket_disk_gen(struct bucket *b) | |||
1173 | static struct kobj_attribute ksysfs_##n = \ | 1147 | static struct kobj_attribute ksysfs_##n = \ |
1174 | __ATTR(n, S_IWUSR|S_IRUSR, show, store) | 1148 | __ATTR(n, S_IWUSR|S_IRUSR, show, store) |
1175 | 1149 | ||
1176 | /* Forward declarations */ | 1150 | static inline void wake_up_allocators(struct cache_set *c) |
1151 | { | ||
1152 | struct cache *ca; | ||
1153 | unsigned i; | ||
1154 | |||
1155 | for_each_cache(ca, c, i) | ||
1156 | wake_up_process(ca->alloc_thread); | ||
1157 | } | ||
1177 | 1158 | ||
1178 | void bch_writeback_queue(struct cached_dev *); | 1159 | /* Forward declarations */ |
1179 | void bch_writeback_add(struct cached_dev *, unsigned); | ||
1180 | 1160 | ||
1181 | void bch_count_io_errors(struct cache *, int, const char *); | 1161 | void bch_count_io_errors(struct cache *, int, const char *); |
1182 | void bch_bbio_count_io_errors(struct cache_set *, struct bio *, | 1162 | void bch_bbio_count_io_errors(struct cache_set *, struct bio *, |
@@ -1193,7 +1173,6 @@ void bch_submit_bbio(struct bio *, struct cache_set *, struct bkey *, unsigned); | |||
1193 | uint8_t bch_inc_gen(struct cache *, struct bucket *); | 1173 | uint8_t bch_inc_gen(struct cache *, struct bucket *); |
1194 | void bch_rescale_priorities(struct cache_set *, int); | 1174 | void bch_rescale_priorities(struct cache_set *, int); |
1195 | bool bch_bucket_add_unused(struct cache *, struct bucket *); | 1175 | bool bch_bucket_add_unused(struct cache *, struct bucket *); |
1196 | void bch_allocator_thread(struct closure *); | ||
1197 | 1176 | ||
1198 | long bch_bucket_alloc(struct cache *, unsigned, struct closure *); | 1177 | long bch_bucket_alloc(struct cache *, unsigned, struct closure *); |
1199 | void bch_bucket_free(struct cache_set *, struct bkey *); | 1178 | void bch_bucket_free(struct cache_set *, struct bkey *); |
@@ -1241,9 +1220,9 @@ void bch_cache_set_stop(struct cache_set *); | |||
1241 | struct cache_set *bch_cache_set_alloc(struct cache_sb *); | 1220 | struct cache_set *bch_cache_set_alloc(struct cache_sb *); |
1242 | void bch_btree_cache_free(struct cache_set *); | 1221 | void bch_btree_cache_free(struct cache_set *); |
1243 | int bch_btree_cache_alloc(struct cache_set *); | 1222 | int bch_btree_cache_alloc(struct cache_set *); |
1244 | void bch_cached_dev_writeback_init(struct cached_dev *); | ||
1245 | void bch_moving_init_cache_set(struct cache_set *); | 1223 | void bch_moving_init_cache_set(struct cache_set *); |
1246 | 1224 | ||
1225 | int bch_cache_allocator_start(struct cache *ca); | ||
1247 | void bch_cache_allocator_exit(struct cache *ca); | 1226 | void bch_cache_allocator_exit(struct cache *ca); |
1248 | int bch_cache_allocator_init(struct cache *ca); | 1227 | int bch_cache_allocator_init(struct cache *ca); |
1249 | 1228 | ||
diff --git a/drivers/md/bcache/bset.c b/drivers/md/bcache/bset.c index 1d27d3af3251..8010eed06a51 100644 --- a/drivers/md/bcache/bset.c +++ b/drivers/md/bcache/bset.c | |||
@@ -78,6 +78,7 @@ struct bkey *bch_keylist_pop(struct keylist *l) | |||
78 | bool __bch_ptr_invalid(struct cache_set *c, int level, const struct bkey *k) | 78 | bool __bch_ptr_invalid(struct cache_set *c, int level, const struct bkey *k) |
79 | { | 79 | { |
80 | unsigned i; | 80 | unsigned i; |
81 | char buf[80]; | ||
81 | 82 | ||
82 | if (level && (!KEY_PTRS(k) || !KEY_SIZE(k) || KEY_DIRTY(k))) | 83 | if (level && (!KEY_PTRS(k) || !KEY_SIZE(k) || KEY_DIRTY(k))) |
83 | goto bad; | 84 | goto bad; |
@@ -102,7 +103,8 @@ bool __bch_ptr_invalid(struct cache_set *c, int level, const struct bkey *k) | |||
102 | 103 | ||
103 | return false; | 104 | return false; |
104 | bad: | 105 | bad: |
105 | cache_bug(c, "spotted bad key %s: %s", pkey(k), bch_ptr_status(c, k)); | 106 | bch_bkey_to_text(buf, sizeof(buf), k); |
107 | cache_bug(c, "spotted bad key %s: %s", buf, bch_ptr_status(c, k)); | ||
106 | return true; | 108 | return true; |
107 | } | 109 | } |
108 | 110 | ||
@@ -162,10 +164,16 @@ bool bch_ptr_bad(struct btree *b, const struct bkey *k) | |||
162 | #ifdef CONFIG_BCACHE_EDEBUG | 164 | #ifdef CONFIG_BCACHE_EDEBUG |
163 | bug: | 165 | bug: |
164 | mutex_unlock(&b->c->bucket_lock); | 166 | mutex_unlock(&b->c->bucket_lock); |
165 | btree_bug(b, | 167 | |
168 | { | ||
169 | char buf[80]; | ||
170 | |||
171 | bch_bkey_to_text(buf, sizeof(buf), k); | ||
172 | btree_bug(b, | ||
166 | "inconsistent pointer %s: bucket %zu pin %i prio %i gen %i last_gc %i mark %llu gc_gen %i", | 173 | "inconsistent pointer %s: bucket %zu pin %i prio %i gen %i last_gc %i mark %llu gc_gen %i", |
167 | pkey(k), PTR_BUCKET_NR(b->c, k, i), atomic_read(&g->pin), | 174 | buf, PTR_BUCKET_NR(b->c, k, i), atomic_read(&g->pin), |
168 | g->prio, g->gen, g->last_gc, GC_MARK(g), g->gc_gen); | 175 | g->prio, g->gen, g->last_gc, GC_MARK(g), g->gc_gen); |
176 | } | ||
169 | return true; | 177 | return true; |
170 | #endif | 178 | #endif |
171 | } | 179 | } |
@@ -1084,33 +1092,39 @@ void bch_btree_sort_into(struct btree *b, struct btree *new) | |||
1084 | new->sets->size = 0; | 1092 | new->sets->size = 0; |
1085 | } | 1093 | } |
1086 | 1094 | ||
1095 | #define SORT_CRIT (4096 / sizeof(uint64_t)) | ||
1096 | |||
1087 | void bch_btree_sort_lazy(struct btree *b) | 1097 | void bch_btree_sort_lazy(struct btree *b) |
1088 | { | 1098 | { |
1089 | if (b->nsets) { | 1099 | unsigned crit = SORT_CRIT; |
1090 | unsigned i, j, keys = 0, total; | 1100 | int i; |
1091 | 1101 | ||
1092 | for (i = 0; i <= b->nsets; i++) | 1102 | /* Don't sort if nothing to do */ |
1093 | keys += b->sets[i].data->keys; | 1103 | if (!b->nsets) |
1094 | 1104 | goto out; | |
1095 | total = keys; | ||
1096 | 1105 | ||
1097 | for (j = 0; j < b->nsets; j++) { | 1106 | /* If not a leaf node, always sort */ |
1098 | if (keys * 2 < total || | 1107 | if (b->level) { |
1099 | keys < 1000) { | 1108 | bch_btree_sort(b); |
1100 | bch_btree_sort_partial(b, j); | 1109 | return; |
1101 | return; | 1110 | } |
1102 | } | ||
1103 | 1111 | ||
1104 | keys -= b->sets[j].data->keys; | 1112 | for (i = b->nsets - 1; i >= 0; --i) { |
1105 | } | 1113 | crit *= b->c->sort_crit_factor; |
1106 | 1114 | ||
1107 | /* Must sort if b->nsets == 3 or we'll overflow */ | 1115 | if (b->sets[i].data->keys < crit) { |
1108 | if (b->nsets >= (MAX_BSETS - 1) - b->level) { | 1116 | bch_btree_sort_partial(b, i); |
1109 | bch_btree_sort(b); | ||
1110 | return; | 1117 | return; |
1111 | } | 1118 | } |
1112 | } | 1119 | } |
1113 | 1120 | ||
1121 | /* Sort if we'd overflow */ | ||
1122 | if (b->nsets + 1 == MAX_BSETS) { | ||
1123 | bch_btree_sort(b); | ||
1124 | return; | ||
1125 | } | ||
1126 | |||
1127 | out: | ||
1114 | bset_build_written_tree(b); | 1128 | bset_build_written_tree(b); |
1115 | } | 1129 | } |
1116 | 1130 | ||
diff --git a/drivers/md/bcache/bset.h b/drivers/md/bcache/bset.h index 57a9cff41546..ae115a253d73 100644 --- a/drivers/md/bcache/bset.h +++ b/drivers/md/bcache/bset.h | |||
@@ -1,6 +1,8 @@ | |||
1 | #ifndef _BCACHE_BSET_H | 1 | #ifndef _BCACHE_BSET_H |
2 | #define _BCACHE_BSET_H | 2 | #define _BCACHE_BSET_H |
3 | 3 | ||
4 | #include <linux/slab.h> | ||
5 | |||
4 | /* | 6 | /* |
5 | * BKEYS: | 7 | * BKEYS: |
6 | * | 8 | * |
@@ -142,6 +144,8 @@ | |||
142 | 144 | ||
143 | /* Btree key comparison/iteration */ | 145 | /* Btree key comparison/iteration */ |
144 | 146 | ||
147 | #define MAX_BSETS 4U | ||
148 | |||
145 | struct btree_iter { | 149 | struct btree_iter { |
146 | size_t size, used; | 150 | size_t size, used; |
147 | struct btree_iter_set { | 151 | struct btree_iter_set { |
diff --git a/drivers/md/bcache/btree.c b/drivers/md/bcache/btree.c index 7a5658f04e62..ee372884c405 100644 --- a/drivers/md/bcache/btree.c +++ b/drivers/md/bcache/btree.c | |||
@@ -24,6 +24,7 @@ | |||
24 | #include "btree.h" | 24 | #include "btree.h" |
25 | #include "debug.h" | 25 | #include "debug.h" |
26 | #include "request.h" | 26 | #include "request.h" |
27 | #include "writeback.h" | ||
27 | 28 | ||
28 | #include <linux/slab.h> | 29 | #include <linux/slab.h> |
29 | #include <linux/bitops.h> | 30 | #include <linux/bitops.h> |
@@ -134,44 +135,17 @@ static uint64_t btree_csum_set(struct btree *b, struct bset *i) | |||
134 | return crc ^ 0xffffffffffffffffULL; | 135 | return crc ^ 0xffffffffffffffffULL; |
135 | } | 136 | } |
136 | 137 | ||
137 | static void btree_bio_endio(struct bio *bio, int error) | 138 | static void bch_btree_node_read_done(struct btree *b) |
138 | { | 139 | { |
139 | struct closure *cl = bio->bi_private; | ||
140 | struct btree *b = container_of(cl, struct btree, io.cl); | ||
141 | |||
142 | if (error) | ||
143 | set_btree_node_io_error(b); | ||
144 | |||
145 | bch_bbio_count_io_errors(b->c, bio, error, (bio->bi_rw & WRITE) | ||
146 | ? "writing btree" : "reading btree"); | ||
147 | closure_put(cl); | ||
148 | } | ||
149 | |||
150 | static void btree_bio_init(struct btree *b) | ||
151 | { | ||
152 | BUG_ON(b->bio); | ||
153 | b->bio = bch_bbio_alloc(b->c); | ||
154 | |||
155 | b->bio->bi_end_io = btree_bio_endio; | ||
156 | b->bio->bi_private = &b->io.cl; | ||
157 | } | ||
158 | |||
159 | void bch_btree_read_done(struct closure *cl) | ||
160 | { | ||
161 | struct btree *b = container_of(cl, struct btree, io.cl); | ||
162 | struct bset *i = b->sets[0].data; | ||
163 | struct btree_iter *iter = b->c->fill_iter; | ||
164 | const char *err = "bad btree header"; | 140 | const char *err = "bad btree header"; |
165 | BUG_ON(b->nsets || b->written); | 141 | struct bset *i = b->sets[0].data; |
166 | 142 | struct btree_iter *iter; | |
167 | bch_bbio_free(b->bio, b->c); | ||
168 | b->bio = NULL; | ||
169 | 143 | ||
170 | mutex_lock(&b->c->fill_lock); | 144 | iter = mempool_alloc(b->c->fill_iter, GFP_NOWAIT); |
145 | iter->size = b->c->sb.bucket_size / b->c->sb.block_size; | ||
171 | iter->used = 0; | 146 | iter->used = 0; |
172 | 147 | ||
173 | if (btree_node_io_error(b) || | 148 | if (!i->seq) |
174 | !i->seq) | ||
175 | goto err; | 149 | goto err; |
176 | 150 | ||
177 | for (; | 151 | for (; |
@@ -228,17 +202,8 @@ void bch_btree_read_done(struct closure *cl) | |||
228 | if (b->written < btree_blocks(b)) | 202 | if (b->written < btree_blocks(b)) |
229 | bch_bset_init_next(b); | 203 | bch_bset_init_next(b); |
230 | out: | 204 | out: |
231 | 205 | mempool_free(iter, b->c->fill_iter); | |
232 | mutex_unlock(&b->c->fill_lock); | 206 | return; |
233 | |||
234 | spin_lock(&b->c->btree_read_time_lock); | ||
235 | bch_time_stats_update(&b->c->btree_read_time, b->io_start_time); | ||
236 | spin_unlock(&b->c->btree_read_time_lock); | ||
237 | |||
238 | smp_wmb(); /* read_done is our write lock */ | ||
239 | set_btree_node_read_done(b); | ||
240 | |||
241 | closure_return(cl); | ||
242 | err: | 207 | err: |
243 | set_btree_node_io_error(b); | 208 | set_btree_node_io_error(b); |
244 | bch_cache_set_error(b->c, "%s at bucket %zu, block %zu, %u keys", | 209 | bch_cache_set_error(b->c, "%s at bucket %zu, block %zu, %u keys", |
@@ -247,48 +212,69 @@ err: | |||
247 | goto out; | 212 | goto out; |
248 | } | 213 | } |
249 | 214 | ||
250 | void bch_btree_read(struct btree *b) | 215 | static void btree_node_read_endio(struct bio *bio, int error) |
216 | { | ||
217 | struct closure *cl = bio->bi_private; | ||
218 | closure_put(cl); | ||
219 | } | ||
220 | |||
221 | void bch_btree_node_read(struct btree *b) | ||
251 | { | 222 | { |
252 | BUG_ON(b->nsets || b->written); | 223 | uint64_t start_time = local_clock(); |
224 | struct closure cl; | ||
225 | struct bio *bio; | ||
226 | |||
227 | trace_bcache_btree_read(b); | ||
228 | |||
229 | closure_init_stack(&cl); | ||
230 | |||
231 | bio = bch_bbio_alloc(b->c); | ||
232 | bio->bi_rw = REQ_META|READ_SYNC; | ||
233 | bio->bi_size = KEY_SIZE(&b->key) << 9; | ||
234 | bio->bi_end_io = btree_node_read_endio; | ||
235 | bio->bi_private = &cl; | ||
236 | |||
237 | bch_bio_map(bio, b->sets[0].data); | ||
238 | |||
239 | bch_submit_bbio(bio, b->c, &b->key, 0); | ||
240 | closure_sync(&cl); | ||
253 | 241 | ||
254 | if (!closure_trylock(&b->io.cl, &b->c->cl)) | 242 | if (!test_bit(BIO_UPTODATE, &bio->bi_flags)) |
255 | BUG(); | 243 | set_btree_node_io_error(b); |
256 | 244 | ||
257 | b->io_start_time = local_clock(); | 245 | bch_bbio_free(bio, b->c); |
258 | 246 | ||
259 | btree_bio_init(b); | 247 | if (btree_node_io_error(b)) |
260 | b->bio->bi_rw = REQ_META|READ_SYNC; | 248 | goto err; |
261 | b->bio->bi_size = KEY_SIZE(&b->key) << 9; | ||
262 | 249 | ||
263 | bch_bio_map(b->bio, b->sets[0].data); | 250 | bch_btree_node_read_done(b); |
264 | 251 | ||
265 | pr_debug("%s", pbtree(b)); | 252 | spin_lock(&b->c->btree_read_time_lock); |
266 | trace_bcache_btree_read(b->bio); | 253 | bch_time_stats_update(&b->c->btree_read_time, start_time); |
267 | bch_submit_bbio(b->bio, b->c, &b->key, 0); | 254 | spin_unlock(&b->c->btree_read_time_lock); |
268 | 255 | ||
269 | continue_at(&b->io.cl, bch_btree_read_done, system_wq); | 256 | return; |
257 | err: | ||
258 | bch_cache_set_error(b->c, "io error reading bucket %lu", | ||
259 | PTR_BUCKET_NR(b->c, &b->key, 0)); | ||
270 | } | 260 | } |
271 | 261 | ||
272 | static void btree_complete_write(struct btree *b, struct btree_write *w) | 262 | static void btree_complete_write(struct btree *b, struct btree_write *w) |
273 | { | 263 | { |
274 | if (w->prio_blocked && | 264 | if (w->prio_blocked && |
275 | !atomic_sub_return(w->prio_blocked, &b->c->prio_blocked)) | 265 | !atomic_sub_return(w->prio_blocked, &b->c->prio_blocked)) |
276 | wake_up(&b->c->alloc_wait); | 266 | wake_up_allocators(b->c); |
277 | 267 | ||
278 | if (w->journal) { | 268 | if (w->journal) { |
279 | atomic_dec_bug(w->journal); | 269 | atomic_dec_bug(w->journal); |
280 | __closure_wake_up(&b->c->journal.wait); | 270 | __closure_wake_up(&b->c->journal.wait); |
281 | } | 271 | } |
282 | 272 | ||
283 | if (w->owner) | ||
284 | closure_put(w->owner); | ||
285 | |||
286 | w->prio_blocked = 0; | 273 | w->prio_blocked = 0; |
287 | w->journal = NULL; | 274 | w->journal = NULL; |
288 | w->owner = NULL; | ||
289 | } | 275 | } |
290 | 276 | ||
291 | static void __btree_write_done(struct closure *cl) | 277 | static void __btree_node_write_done(struct closure *cl) |
292 | { | 278 | { |
293 | struct btree *b = container_of(cl, struct btree, io.cl); | 279 | struct btree *b = container_of(cl, struct btree, io.cl); |
294 | struct btree_write *w = btree_prev_write(b); | 280 | struct btree_write *w = btree_prev_write(b); |
@@ -304,7 +290,7 @@ static void __btree_write_done(struct closure *cl) | |||
304 | closure_return(cl); | 290 | closure_return(cl); |
305 | } | 291 | } |
306 | 292 | ||
307 | static void btree_write_done(struct closure *cl) | 293 | static void btree_node_write_done(struct closure *cl) |
308 | { | 294 | { |
309 | struct btree *b = container_of(cl, struct btree, io.cl); | 295 | struct btree *b = container_of(cl, struct btree, io.cl); |
310 | struct bio_vec *bv; | 296 | struct bio_vec *bv; |
@@ -313,10 +299,22 @@ static void btree_write_done(struct closure *cl) | |||
313 | __bio_for_each_segment(bv, b->bio, n, 0) | 299 | __bio_for_each_segment(bv, b->bio, n, 0) |
314 | __free_page(bv->bv_page); | 300 | __free_page(bv->bv_page); |
315 | 301 | ||
316 | __btree_write_done(cl); | 302 | __btree_node_write_done(cl); |
317 | } | 303 | } |
318 | 304 | ||
319 | static void do_btree_write(struct btree *b) | 305 | static void btree_node_write_endio(struct bio *bio, int error) |
306 | { | ||
307 | struct closure *cl = bio->bi_private; | ||
308 | struct btree *b = container_of(cl, struct btree, io.cl); | ||
309 | |||
310 | if (error) | ||
311 | set_btree_node_io_error(b); | ||
312 | |||
313 | bch_bbio_count_io_errors(b->c, bio, error, "writing btree"); | ||
314 | closure_put(cl); | ||
315 | } | ||
316 | |||
317 | static void do_btree_node_write(struct btree *b) | ||
320 | { | 318 | { |
321 | struct closure *cl = &b->io.cl; | 319 | struct closure *cl = &b->io.cl; |
322 | struct bset *i = b->sets[b->nsets].data; | 320 | struct bset *i = b->sets[b->nsets].data; |
@@ -325,15 +323,34 @@ static void do_btree_write(struct btree *b) | |||
325 | i->version = BCACHE_BSET_VERSION; | 323 | i->version = BCACHE_BSET_VERSION; |
326 | i->csum = btree_csum_set(b, i); | 324 | i->csum = btree_csum_set(b, i); |
327 | 325 | ||
328 | btree_bio_init(b); | 326 | BUG_ON(b->bio); |
329 | b->bio->bi_rw = REQ_META|WRITE_SYNC; | 327 | b->bio = bch_bbio_alloc(b->c); |
330 | b->bio->bi_size = set_blocks(i, b->c) * block_bytes(b->c); | 328 | |
329 | b->bio->bi_end_io = btree_node_write_endio; | ||
330 | b->bio->bi_private = &b->io.cl; | ||
331 | b->bio->bi_rw = REQ_META|WRITE_SYNC|REQ_FUA; | ||
332 | b->bio->bi_size = set_blocks(i, b->c) * block_bytes(b->c); | ||
331 | bch_bio_map(b->bio, i); | 333 | bch_bio_map(b->bio, i); |
332 | 334 | ||
335 | /* | ||
336 | * If we're appending to a leaf node, we don't technically need FUA - | ||
337 | * this write just needs to be persisted before the next journal write, | ||
338 | * which will be marked FLUSH|FUA. | ||
339 | * | ||
340 | * Similarly if we're writing a new btree root - the pointer is going to | ||
341 | * be in the next journal entry. | ||
342 | * | ||
343 | * But if we're writing a new btree node (that isn't a root) or | ||
344 | * appending to a non leaf btree node, we need either FUA or a flush | ||
345 | * when we write the parent with the new pointer. FUA is cheaper than a | ||
346 | * flush, and writes appending to leaf nodes aren't blocking anything so | ||
347 | * just make all btree node writes FUA to keep things sane. | ||
348 | */ | ||
349 | |||
333 | bkey_copy(&k.key, &b->key); | 350 | bkey_copy(&k.key, &b->key); |
334 | SET_PTR_OFFSET(&k.key, 0, PTR_OFFSET(&k.key, 0) + bset_offset(b, i)); | 351 | SET_PTR_OFFSET(&k.key, 0, PTR_OFFSET(&k.key, 0) + bset_offset(b, i)); |
335 | 352 | ||
336 | if (!bch_bio_alloc_pages(b->bio, GFP_NOIO)) { | 353 | if (!bio_alloc_pages(b->bio, GFP_NOIO)) { |
337 | int j; | 354 | int j; |
338 | struct bio_vec *bv; | 355 | struct bio_vec *bv; |
339 | void *base = (void *) ((unsigned long) i & ~(PAGE_SIZE - 1)); | 356 | void *base = (void *) ((unsigned long) i & ~(PAGE_SIZE - 1)); |
@@ -342,40 +359,41 @@ static void do_btree_write(struct btree *b) | |||
342 | memcpy(page_address(bv->bv_page), | 359 | memcpy(page_address(bv->bv_page), |
343 | base + j * PAGE_SIZE, PAGE_SIZE); | 360 | base + j * PAGE_SIZE, PAGE_SIZE); |
344 | 361 | ||
345 | trace_bcache_btree_write(b->bio); | ||
346 | bch_submit_bbio(b->bio, b->c, &k.key, 0); | 362 | bch_submit_bbio(b->bio, b->c, &k.key, 0); |
347 | 363 | ||
348 | continue_at(cl, btree_write_done, NULL); | 364 | continue_at(cl, btree_node_write_done, NULL); |
349 | } else { | 365 | } else { |
350 | b->bio->bi_vcnt = 0; | 366 | b->bio->bi_vcnt = 0; |
351 | bch_bio_map(b->bio, i); | 367 | bch_bio_map(b->bio, i); |
352 | 368 | ||
353 | trace_bcache_btree_write(b->bio); | ||
354 | bch_submit_bbio(b->bio, b->c, &k.key, 0); | 369 | bch_submit_bbio(b->bio, b->c, &k.key, 0); |
355 | 370 | ||
356 | closure_sync(cl); | 371 | closure_sync(cl); |
357 | __btree_write_done(cl); | 372 | __btree_node_write_done(cl); |
358 | } | 373 | } |
359 | } | 374 | } |
360 | 375 | ||
361 | static void __btree_write(struct btree *b) | 376 | void bch_btree_node_write(struct btree *b, struct closure *parent) |
362 | { | 377 | { |
363 | struct bset *i = b->sets[b->nsets].data; | 378 | struct bset *i = b->sets[b->nsets].data; |
364 | 379 | ||
380 | trace_bcache_btree_write(b); | ||
381 | |||
365 | BUG_ON(current->bio_list); | 382 | BUG_ON(current->bio_list); |
383 | BUG_ON(b->written >= btree_blocks(b)); | ||
384 | BUG_ON(b->written && !i->keys); | ||
385 | BUG_ON(b->sets->data->seq != i->seq); | ||
386 | bch_check_key_order(b, i); | ||
366 | 387 | ||
367 | closure_lock(&b->io, &b->c->cl); | ||
368 | cancel_delayed_work(&b->work); | 388 | cancel_delayed_work(&b->work); |
369 | 389 | ||
390 | /* If caller isn't waiting for write, parent refcount is cache set */ | ||
391 | closure_lock(&b->io, parent ?: &b->c->cl); | ||
392 | |||
370 | clear_bit(BTREE_NODE_dirty, &b->flags); | 393 | clear_bit(BTREE_NODE_dirty, &b->flags); |
371 | change_bit(BTREE_NODE_write_idx, &b->flags); | 394 | change_bit(BTREE_NODE_write_idx, &b->flags); |
372 | 395 | ||
373 | bch_check_key_order(b, i); | 396 | do_btree_node_write(b); |
374 | BUG_ON(b->written && !i->keys); | ||
375 | |||
376 | do_btree_write(b); | ||
377 | |||
378 | pr_debug("%s block %i keys %i", pbtree(b), b->written, i->keys); | ||
379 | 397 | ||
380 | b->written += set_blocks(i, b->c); | 398 | b->written += set_blocks(i, b->c); |
381 | atomic_long_add(set_blocks(i, b->c) * b->c->sb.block_size, | 399 | atomic_long_add(set_blocks(i, b->c) * b->c->sb.block_size, |
@@ -387,37 +405,31 @@ static void __btree_write(struct btree *b) | |||
387 | bch_bset_init_next(b); | 405 | bch_bset_init_next(b); |
388 | } | 406 | } |
389 | 407 | ||
390 | static void btree_write_work(struct work_struct *w) | 408 | static void btree_node_write_work(struct work_struct *w) |
391 | { | 409 | { |
392 | struct btree *b = container_of(to_delayed_work(w), struct btree, work); | 410 | struct btree *b = container_of(to_delayed_work(w), struct btree, work); |
393 | 411 | ||
394 | down_write(&b->lock); | 412 | rw_lock(true, b, b->level); |
395 | 413 | ||
396 | if (btree_node_dirty(b)) | 414 | if (btree_node_dirty(b)) |
397 | __btree_write(b); | 415 | bch_btree_node_write(b, NULL); |
398 | up_write(&b->lock); | 416 | rw_unlock(true, b); |
399 | } | 417 | } |
400 | 418 | ||
401 | void bch_btree_write(struct btree *b, bool now, struct btree_op *op) | 419 | static void bch_btree_leaf_dirty(struct btree *b, struct btree_op *op) |
402 | { | 420 | { |
403 | struct bset *i = b->sets[b->nsets].data; | 421 | struct bset *i = b->sets[b->nsets].data; |
404 | struct btree_write *w = btree_current_write(b); | 422 | struct btree_write *w = btree_current_write(b); |
405 | 423 | ||
406 | BUG_ON(b->written && | 424 | BUG_ON(!b->written); |
407 | (b->written >= btree_blocks(b) || | 425 | BUG_ON(!i->keys); |
408 | i->seq != b->sets[0].data->seq || | ||
409 | !i->keys)); | ||
410 | 426 | ||
411 | if (!btree_node_dirty(b)) { | 427 | if (!btree_node_dirty(b)) |
412 | set_btree_node_dirty(b); | 428 | queue_delayed_work(btree_io_wq, &b->work, 30 * HZ); |
413 | queue_delayed_work(btree_io_wq, &b->work, | ||
414 | msecs_to_jiffies(30000)); | ||
415 | } | ||
416 | 429 | ||
417 | w->prio_blocked += b->prio_blocked; | 430 | set_btree_node_dirty(b); |
418 | b->prio_blocked = 0; | ||
419 | 431 | ||
420 | if (op && op->journal && !b->level) { | 432 | if (op && op->journal) { |
421 | if (w->journal && | 433 | if (w->journal && |
422 | journal_pin_cmp(b->c, w, op)) { | 434 | journal_pin_cmp(b->c, w, op)) { |
423 | atomic_dec_bug(w->journal); | 435 | atomic_dec_bug(w->journal); |
@@ -430,23 +442,10 @@ void bch_btree_write(struct btree *b, bool now, struct btree_op *op) | |||
430 | } | 442 | } |
431 | } | 443 | } |
432 | 444 | ||
433 | if (current->bio_list) | ||
434 | return; | ||
435 | |||
436 | /* Force write if set is too big */ | 445 | /* Force write if set is too big */ |
437 | if (now || | 446 | if (set_bytes(i) > PAGE_SIZE - 48 && |
438 | b->level || | 447 | !current->bio_list) |
439 | set_bytes(i) > PAGE_SIZE - 48) { | 448 | bch_btree_node_write(b, NULL); |
440 | if (op && now) { | ||
441 | /* Must wait on multiple writes */ | ||
442 | BUG_ON(w->owner); | ||
443 | w->owner = &op->cl; | ||
444 | closure_get(&op->cl); | ||
445 | } | ||
446 | |||
447 | __btree_write(b); | ||
448 | } | ||
449 | BUG_ON(!b->written); | ||
450 | } | 449 | } |
451 | 450 | ||
452 | /* | 451 | /* |
@@ -559,7 +558,7 @@ static struct btree *mca_bucket_alloc(struct cache_set *c, | |||
559 | init_rwsem(&b->lock); | 558 | init_rwsem(&b->lock); |
560 | lockdep_set_novalidate_class(&b->lock); | 559 | lockdep_set_novalidate_class(&b->lock); |
561 | INIT_LIST_HEAD(&b->list); | 560 | INIT_LIST_HEAD(&b->list); |
562 | INIT_DELAYED_WORK(&b->work, btree_write_work); | 561 | INIT_DELAYED_WORK(&b->work, btree_node_write_work); |
563 | b->c = c; | 562 | b->c = c; |
564 | closure_init_unlocked(&b->io); | 563 | closure_init_unlocked(&b->io); |
565 | 564 | ||
@@ -582,7 +581,7 @@ static int mca_reap(struct btree *b, struct closure *cl, unsigned min_order) | |||
582 | BUG_ON(btree_node_dirty(b) && !b->sets[0].data); | 581 | BUG_ON(btree_node_dirty(b) && !b->sets[0].data); |
583 | 582 | ||
584 | if (cl && btree_node_dirty(b)) | 583 | if (cl && btree_node_dirty(b)) |
585 | bch_btree_write(b, true, NULL); | 584 | bch_btree_node_write(b, NULL); |
586 | 585 | ||
587 | if (cl) | 586 | if (cl) |
588 | closure_wait_event_async(&b->io.wait, cl, | 587 | closure_wait_event_async(&b->io.wait, cl, |
@@ -623,6 +622,13 @@ static int bch_mca_shrink(struct shrinker *shrink, struct shrink_control *sc) | |||
623 | else if (!mutex_trylock(&c->bucket_lock)) | 622 | else if (!mutex_trylock(&c->bucket_lock)) |
624 | return -1; | 623 | return -1; |
625 | 624 | ||
625 | /* | ||
626 | * It's _really_ critical that we don't free too many btree nodes - we | ||
627 | * have to always leave ourselves a reserve. The reserve is how we | ||
628 | * guarantee that allocating memory for a new btree node can always | ||
629 | * succeed, so that inserting keys into the btree can always succeed and | ||
630 | * IO can always make forward progress: | ||
631 | */ | ||
626 | nr /= c->btree_pages; | 632 | nr /= c->btree_pages; |
627 | nr = min_t(unsigned long, nr, mca_can_free(c)); | 633 | nr = min_t(unsigned long, nr, mca_can_free(c)); |
628 | 634 | ||
@@ -766,6 +772,8 @@ static struct btree *mca_cannibalize(struct cache_set *c, struct bkey *k, | |||
766 | int ret = -ENOMEM; | 772 | int ret = -ENOMEM; |
767 | struct btree *i; | 773 | struct btree *i; |
768 | 774 | ||
775 | trace_bcache_btree_cache_cannibalize(c); | ||
776 | |||
769 | if (!cl) | 777 | if (!cl) |
770 | return ERR_PTR(-ENOMEM); | 778 | return ERR_PTR(-ENOMEM); |
771 | 779 | ||
@@ -784,7 +792,6 @@ static struct btree *mca_cannibalize(struct cache_set *c, struct bkey *k, | |||
784 | return ERR_PTR(-EAGAIN); | 792 | return ERR_PTR(-EAGAIN); |
785 | } | 793 | } |
786 | 794 | ||
787 | /* XXX: tracepoint */ | ||
788 | c->try_harder = cl; | 795 | c->try_harder = cl; |
789 | c->try_harder_start = local_clock(); | 796 | c->try_harder_start = local_clock(); |
790 | retry: | 797 | retry: |
@@ -905,6 +912,9 @@ retry: | |||
905 | b = mca_find(c, k); | 912 | b = mca_find(c, k); |
906 | 913 | ||
907 | if (!b) { | 914 | if (!b) { |
915 | if (current->bio_list) | ||
916 | return ERR_PTR(-EAGAIN); | ||
917 | |||
908 | mutex_lock(&c->bucket_lock); | 918 | mutex_lock(&c->bucket_lock); |
909 | b = mca_alloc(c, k, level, &op->cl); | 919 | b = mca_alloc(c, k, level, &op->cl); |
910 | mutex_unlock(&c->bucket_lock); | 920 | mutex_unlock(&c->bucket_lock); |
@@ -914,7 +924,7 @@ retry: | |||
914 | if (IS_ERR(b)) | 924 | if (IS_ERR(b)) |
915 | return b; | 925 | return b; |
916 | 926 | ||
917 | bch_btree_read(b); | 927 | bch_btree_node_read(b); |
918 | 928 | ||
919 | if (!write) | 929 | if (!write) |
920 | downgrade_write(&b->lock); | 930 | downgrade_write(&b->lock); |
@@ -937,15 +947,12 @@ retry: | |||
937 | for (; i <= b->nsets; i++) | 947 | for (; i <= b->nsets; i++) |
938 | prefetch(b->sets[i].data); | 948 | prefetch(b->sets[i].data); |
939 | 949 | ||
940 | if (!closure_wait_event(&b->io.wait, &op->cl, | 950 | if (btree_node_io_error(b)) { |
941 | btree_node_read_done(b))) { | ||
942 | rw_unlock(write, b); | ||
943 | b = ERR_PTR(-EAGAIN); | ||
944 | } else if (btree_node_io_error(b)) { | ||
945 | rw_unlock(write, b); | 951 | rw_unlock(write, b); |
946 | b = ERR_PTR(-EIO); | 952 | return ERR_PTR(-EIO); |
947 | } else | 953 | } |
948 | BUG_ON(!b->written); | 954 | |
955 | BUG_ON(!b->written); | ||
949 | 956 | ||
950 | return b; | 957 | return b; |
951 | } | 958 | } |
@@ -959,7 +966,7 @@ static void btree_node_prefetch(struct cache_set *c, struct bkey *k, int level) | |||
959 | mutex_unlock(&c->bucket_lock); | 966 | mutex_unlock(&c->bucket_lock); |
960 | 967 | ||
961 | if (!IS_ERR_OR_NULL(b)) { | 968 | if (!IS_ERR_OR_NULL(b)) { |
962 | bch_btree_read(b); | 969 | bch_btree_node_read(b); |
963 | rw_unlock(true, b); | 970 | rw_unlock(true, b); |
964 | } | 971 | } |
965 | } | 972 | } |
@@ -970,24 +977,19 @@ static void btree_node_free(struct btree *b, struct btree_op *op) | |||
970 | { | 977 | { |
971 | unsigned i; | 978 | unsigned i; |
972 | 979 | ||
980 | trace_bcache_btree_node_free(b); | ||
981 | |||
973 | /* | 982 | /* |
974 | * The BUG_ON() in btree_node_get() implies that we must have a write | 983 | * The BUG_ON() in btree_node_get() implies that we must have a write |
975 | * lock on parent to free or even invalidate a node | 984 | * lock on parent to free or even invalidate a node |
976 | */ | 985 | */ |
977 | BUG_ON(op->lock <= b->level); | 986 | BUG_ON(op->lock <= b->level); |
978 | BUG_ON(b == b->c->root); | 987 | BUG_ON(b == b->c->root); |
979 | pr_debug("bucket %s", pbtree(b)); | ||
980 | 988 | ||
981 | if (btree_node_dirty(b)) | 989 | if (btree_node_dirty(b)) |
982 | btree_complete_write(b, btree_current_write(b)); | 990 | btree_complete_write(b, btree_current_write(b)); |
983 | clear_bit(BTREE_NODE_dirty, &b->flags); | 991 | clear_bit(BTREE_NODE_dirty, &b->flags); |
984 | 992 | ||
985 | if (b->prio_blocked && | ||
986 | !atomic_sub_return(b->prio_blocked, &b->c->prio_blocked)) | ||
987 | wake_up(&b->c->alloc_wait); | ||
988 | |||
989 | b->prio_blocked = 0; | ||
990 | |||
991 | cancel_delayed_work(&b->work); | 993 | cancel_delayed_work(&b->work); |
992 | 994 | ||
993 | mutex_lock(&b->c->bucket_lock); | 995 | mutex_lock(&b->c->bucket_lock); |
@@ -1028,17 +1030,20 @@ retry: | |||
1028 | goto retry; | 1030 | goto retry; |
1029 | } | 1031 | } |
1030 | 1032 | ||
1031 | set_btree_node_read_done(b); | ||
1032 | b->accessed = 1; | 1033 | b->accessed = 1; |
1033 | bch_bset_init_next(b); | 1034 | bch_bset_init_next(b); |
1034 | 1035 | ||
1035 | mutex_unlock(&c->bucket_lock); | 1036 | mutex_unlock(&c->bucket_lock); |
1037 | |||
1038 | trace_bcache_btree_node_alloc(b); | ||
1036 | return b; | 1039 | return b; |
1037 | err_free: | 1040 | err_free: |
1038 | bch_bucket_free(c, &k.key); | 1041 | bch_bucket_free(c, &k.key); |
1039 | __bkey_put(c, &k.key); | 1042 | __bkey_put(c, &k.key); |
1040 | err: | 1043 | err: |
1041 | mutex_unlock(&c->bucket_lock); | 1044 | mutex_unlock(&c->bucket_lock); |
1045 | |||
1046 | trace_bcache_btree_node_alloc_fail(b); | ||
1042 | return b; | 1047 | return b; |
1043 | } | 1048 | } |
1044 | 1049 | ||
@@ -1137,11 +1142,8 @@ static int btree_gc_mark_node(struct btree *b, unsigned *keys, | |||
1137 | gc->nkeys++; | 1142 | gc->nkeys++; |
1138 | 1143 | ||
1139 | gc->data += KEY_SIZE(k); | 1144 | gc->data += KEY_SIZE(k); |
1140 | if (KEY_DIRTY(k)) { | 1145 | if (KEY_DIRTY(k)) |
1141 | gc->dirty += KEY_SIZE(k); | 1146 | gc->dirty += KEY_SIZE(k); |
1142 | if (d) | ||
1143 | d->sectors_dirty_gc += KEY_SIZE(k); | ||
1144 | } | ||
1145 | } | 1147 | } |
1146 | 1148 | ||
1147 | for (t = b->sets; t <= &b->sets[b->nsets]; t++) | 1149 | for (t = b->sets; t <= &b->sets[b->nsets]; t++) |
@@ -1166,14 +1168,11 @@ static struct btree *btree_gc_alloc(struct btree *b, struct bkey *k, | |||
1166 | 1168 | ||
1167 | if (!IS_ERR_OR_NULL(n)) { | 1169 | if (!IS_ERR_OR_NULL(n)) { |
1168 | swap(b, n); | 1170 | swap(b, n); |
1171 | __bkey_put(b->c, &b->key); | ||
1169 | 1172 | ||
1170 | memcpy(k->ptr, b->key.ptr, | 1173 | memcpy(k->ptr, b->key.ptr, |
1171 | sizeof(uint64_t) * KEY_PTRS(&b->key)); | 1174 | sizeof(uint64_t) * KEY_PTRS(&b->key)); |
1172 | 1175 | ||
1173 | __bkey_put(b->c, &b->key); | ||
1174 | atomic_inc(&b->c->prio_blocked); | ||
1175 | b->prio_blocked++; | ||
1176 | |||
1177 | btree_node_free(n, op); | 1176 | btree_node_free(n, op); |
1178 | up_write(&n->lock); | 1177 | up_write(&n->lock); |
1179 | } | 1178 | } |
@@ -1278,7 +1277,7 @@ static void btree_gc_coalesce(struct btree *b, struct btree_op *op, | |||
1278 | btree_node_free(r->b, op); | 1277 | btree_node_free(r->b, op); |
1279 | up_write(&r->b->lock); | 1278 | up_write(&r->b->lock); |
1280 | 1279 | ||
1281 | pr_debug("coalesced %u nodes", nodes); | 1280 | trace_bcache_btree_gc_coalesce(nodes); |
1282 | 1281 | ||
1283 | gc->nodes--; | 1282 | gc->nodes--; |
1284 | nodes--; | 1283 | nodes--; |
@@ -1293,14 +1292,9 @@ static int btree_gc_recurse(struct btree *b, struct btree_op *op, | |||
1293 | void write(struct btree *r) | 1292 | void write(struct btree *r) |
1294 | { | 1293 | { |
1295 | if (!r->written) | 1294 | if (!r->written) |
1296 | bch_btree_write(r, true, op); | 1295 | bch_btree_node_write(r, &op->cl); |
1297 | else if (btree_node_dirty(r)) { | 1296 | else if (btree_node_dirty(r)) |
1298 | BUG_ON(btree_current_write(r)->owner); | 1297 | bch_btree_node_write(r, writes); |
1299 | btree_current_write(r)->owner = writes; | ||
1300 | closure_get(writes); | ||
1301 | |||
1302 | bch_btree_write(r, true, NULL); | ||
1303 | } | ||
1304 | 1298 | ||
1305 | up_write(&r->lock); | 1299 | up_write(&r->lock); |
1306 | } | 1300 | } |
@@ -1386,9 +1380,7 @@ static int bch_btree_gc_root(struct btree *b, struct btree_op *op, | |||
1386 | ret = btree_gc_recurse(b, op, writes, gc); | 1380 | ret = btree_gc_recurse(b, op, writes, gc); |
1387 | 1381 | ||
1388 | if (!b->written || btree_node_dirty(b)) { | 1382 | if (!b->written || btree_node_dirty(b)) { |
1389 | atomic_inc(&b->c->prio_blocked); | 1383 | bch_btree_node_write(b, n ? &op->cl : NULL); |
1390 | b->prio_blocked++; | ||
1391 | bch_btree_write(b, true, n ? op : NULL); | ||
1392 | } | 1384 | } |
1393 | 1385 | ||
1394 | if (!IS_ERR_OR_NULL(n)) { | 1386 | if (!IS_ERR_OR_NULL(n)) { |
@@ -1405,7 +1397,6 @@ static void btree_gc_start(struct cache_set *c) | |||
1405 | { | 1397 | { |
1406 | struct cache *ca; | 1398 | struct cache *ca; |
1407 | struct bucket *b; | 1399 | struct bucket *b; |
1408 | struct bcache_device **d; | ||
1409 | unsigned i; | 1400 | unsigned i; |
1410 | 1401 | ||
1411 | if (!c->gc_mark_valid) | 1402 | if (!c->gc_mark_valid) |
@@ -1419,16 +1410,12 @@ static void btree_gc_start(struct cache_set *c) | |||
1419 | for_each_cache(ca, c, i) | 1410 | for_each_cache(ca, c, i) |
1420 | for_each_bucket(b, ca) { | 1411 | for_each_bucket(b, ca) { |
1421 | b->gc_gen = b->gen; | 1412 | b->gc_gen = b->gen; |
1422 | if (!atomic_read(&b->pin)) | 1413 | if (!atomic_read(&b->pin)) { |
1423 | SET_GC_MARK(b, GC_MARK_RECLAIMABLE); | 1414 | SET_GC_MARK(b, GC_MARK_RECLAIMABLE); |
1415 | SET_GC_SECTORS_USED(b, 0); | ||
1416 | } | ||
1424 | } | 1417 | } |
1425 | 1418 | ||
1426 | for (d = c->devices; | ||
1427 | d < c->devices + c->nr_uuids; | ||
1428 | d++) | ||
1429 | if (*d) | ||
1430 | (*d)->sectors_dirty_gc = 0; | ||
1431 | |||
1432 | mutex_unlock(&c->bucket_lock); | 1419 | mutex_unlock(&c->bucket_lock); |
1433 | } | 1420 | } |
1434 | 1421 | ||
@@ -1437,7 +1424,6 @@ size_t bch_btree_gc_finish(struct cache_set *c) | |||
1437 | size_t available = 0; | 1424 | size_t available = 0; |
1438 | struct bucket *b; | 1425 | struct bucket *b; |
1439 | struct cache *ca; | 1426 | struct cache *ca; |
1440 | struct bcache_device **d; | ||
1441 | unsigned i; | 1427 | unsigned i; |
1442 | 1428 | ||
1443 | mutex_lock(&c->bucket_lock); | 1429 | mutex_lock(&c->bucket_lock); |
@@ -1480,22 +1466,6 @@ size_t bch_btree_gc_finish(struct cache_set *c) | |||
1480 | } | 1466 | } |
1481 | } | 1467 | } |
1482 | 1468 | ||
1483 | for (d = c->devices; | ||
1484 | d < c->devices + c->nr_uuids; | ||
1485 | d++) | ||
1486 | if (*d) { | ||
1487 | unsigned long last = | ||
1488 | atomic_long_read(&((*d)->sectors_dirty)); | ||
1489 | long difference = (*d)->sectors_dirty_gc - last; | ||
1490 | |||
1491 | pr_debug("sectors dirty off by %li", difference); | ||
1492 | |||
1493 | (*d)->sectors_dirty_last += difference; | ||
1494 | |||
1495 | atomic_long_set(&((*d)->sectors_dirty), | ||
1496 | (*d)->sectors_dirty_gc); | ||
1497 | } | ||
1498 | |||
1499 | mutex_unlock(&c->bucket_lock); | 1469 | mutex_unlock(&c->bucket_lock); |
1500 | return available; | 1470 | return available; |
1501 | } | 1471 | } |
@@ -1508,10 +1478,9 @@ static void bch_btree_gc(struct closure *cl) | |||
1508 | struct gc_stat stats; | 1478 | struct gc_stat stats; |
1509 | struct closure writes; | 1479 | struct closure writes; |
1510 | struct btree_op op; | 1480 | struct btree_op op; |
1511 | |||
1512 | uint64_t start_time = local_clock(); | 1481 | uint64_t start_time = local_clock(); |
1513 | trace_bcache_gc_start(c->sb.set_uuid); | 1482 | |
1514 | blktrace_msg_all(c, "Starting gc"); | 1483 | trace_bcache_gc_start(c); |
1515 | 1484 | ||
1516 | memset(&stats, 0, sizeof(struct gc_stat)); | 1485 | memset(&stats, 0, sizeof(struct gc_stat)); |
1517 | closure_init_stack(&writes); | 1486 | closure_init_stack(&writes); |
@@ -1520,14 +1489,14 @@ static void bch_btree_gc(struct closure *cl) | |||
1520 | 1489 | ||
1521 | btree_gc_start(c); | 1490 | btree_gc_start(c); |
1522 | 1491 | ||
1492 | atomic_inc(&c->prio_blocked); | ||
1493 | |||
1523 | ret = btree_root(gc_root, c, &op, &writes, &stats); | 1494 | ret = btree_root(gc_root, c, &op, &writes, &stats); |
1524 | closure_sync(&op.cl); | 1495 | closure_sync(&op.cl); |
1525 | closure_sync(&writes); | 1496 | closure_sync(&writes); |
1526 | 1497 | ||
1527 | if (ret) { | 1498 | if (ret) { |
1528 | blktrace_msg_all(c, "Stopped gc"); | ||
1529 | pr_warn("gc failed!"); | 1499 | pr_warn("gc failed!"); |
1530 | |||
1531 | continue_at(cl, bch_btree_gc, bch_gc_wq); | 1500 | continue_at(cl, bch_btree_gc, bch_gc_wq); |
1532 | } | 1501 | } |
1533 | 1502 | ||
@@ -1537,6 +1506,9 @@ static void bch_btree_gc(struct closure *cl) | |||
1537 | 1506 | ||
1538 | available = bch_btree_gc_finish(c); | 1507 | available = bch_btree_gc_finish(c); |
1539 | 1508 | ||
1509 | atomic_dec(&c->prio_blocked); | ||
1510 | wake_up_allocators(c); | ||
1511 | |||
1540 | bch_time_stats_update(&c->btree_gc_time, start_time); | 1512 | bch_time_stats_update(&c->btree_gc_time, start_time); |
1541 | 1513 | ||
1542 | stats.key_bytes *= sizeof(uint64_t); | 1514 | stats.key_bytes *= sizeof(uint64_t); |
@@ -1544,10 +1516,8 @@ static void bch_btree_gc(struct closure *cl) | |||
1544 | stats.data <<= 9; | 1516 | stats.data <<= 9; |
1545 | stats.in_use = (c->nbuckets - available) * 100 / c->nbuckets; | 1517 | stats.in_use = (c->nbuckets - available) * 100 / c->nbuckets; |
1546 | memcpy(&c->gc_stats, &stats, sizeof(struct gc_stat)); | 1518 | memcpy(&c->gc_stats, &stats, sizeof(struct gc_stat)); |
1547 | blktrace_msg_all(c, "Finished gc"); | ||
1548 | 1519 | ||
1549 | trace_bcache_gc_end(c->sb.set_uuid); | 1520 | trace_bcache_gc_end(c); |
1550 | wake_up(&c->alloc_wait); | ||
1551 | 1521 | ||
1552 | continue_at(cl, bch_moving_gc, bch_gc_wq); | 1522 | continue_at(cl, bch_moving_gc, bch_gc_wq); |
1553 | } | 1523 | } |
@@ -1654,14 +1624,14 @@ static bool fix_overlapping_extents(struct btree *b, | |||
1654 | struct btree_iter *iter, | 1624 | struct btree_iter *iter, |
1655 | struct btree_op *op) | 1625 | struct btree_op *op) |
1656 | { | 1626 | { |
1657 | void subtract_dirty(struct bkey *k, int sectors) | 1627 | void subtract_dirty(struct bkey *k, uint64_t offset, int sectors) |
1658 | { | 1628 | { |
1659 | struct bcache_device *d = b->c->devices[KEY_INODE(k)]; | 1629 | if (KEY_DIRTY(k)) |
1660 | 1630 | bcache_dev_sectors_dirty_add(b->c, KEY_INODE(k), | |
1661 | if (KEY_DIRTY(k) && d) | 1631 | offset, -sectors); |
1662 | atomic_long_sub(sectors, &d->sectors_dirty); | ||
1663 | } | 1632 | } |
1664 | 1633 | ||
1634 | uint64_t old_offset; | ||
1665 | unsigned old_size, sectors_found = 0; | 1635 | unsigned old_size, sectors_found = 0; |
1666 | 1636 | ||
1667 | while (1) { | 1637 | while (1) { |
@@ -1673,6 +1643,7 @@ static bool fix_overlapping_extents(struct btree *b, | |||
1673 | if (bkey_cmp(k, &START_KEY(insert)) <= 0) | 1643 | if (bkey_cmp(k, &START_KEY(insert)) <= 0) |
1674 | continue; | 1644 | continue; |
1675 | 1645 | ||
1646 | old_offset = KEY_START(k); | ||
1676 | old_size = KEY_SIZE(k); | 1647 | old_size = KEY_SIZE(k); |
1677 | 1648 | ||
1678 | /* | 1649 | /* |
@@ -1728,7 +1699,7 @@ static bool fix_overlapping_extents(struct btree *b, | |||
1728 | 1699 | ||
1729 | struct bkey *top; | 1700 | struct bkey *top; |
1730 | 1701 | ||
1731 | subtract_dirty(k, KEY_SIZE(insert)); | 1702 | subtract_dirty(k, KEY_START(insert), KEY_SIZE(insert)); |
1732 | 1703 | ||
1733 | if (bkey_written(b, k)) { | 1704 | if (bkey_written(b, k)) { |
1734 | /* | 1705 | /* |
@@ -1775,7 +1746,7 @@ static bool fix_overlapping_extents(struct btree *b, | |||
1775 | } | 1746 | } |
1776 | } | 1747 | } |
1777 | 1748 | ||
1778 | subtract_dirty(k, old_size - KEY_SIZE(k)); | 1749 | subtract_dirty(k, old_offset, old_size - KEY_SIZE(k)); |
1779 | } | 1750 | } |
1780 | 1751 | ||
1781 | check_failed: | 1752 | check_failed: |
@@ -1798,7 +1769,7 @@ static bool btree_insert_key(struct btree *b, struct btree_op *op, | |||
1798 | { | 1769 | { |
1799 | struct bset *i = b->sets[b->nsets].data; | 1770 | struct bset *i = b->sets[b->nsets].data; |
1800 | struct bkey *m, *prev; | 1771 | struct bkey *m, *prev; |
1801 | const char *status = "insert"; | 1772 | unsigned status = BTREE_INSERT_STATUS_INSERT; |
1802 | 1773 | ||
1803 | BUG_ON(bkey_cmp(k, &b->key) > 0); | 1774 | BUG_ON(bkey_cmp(k, &b->key) > 0); |
1804 | BUG_ON(b->level && !KEY_PTRS(k)); | 1775 | BUG_ON(b->level && !KEY_PTRS(k)); |
@@ -1831,17 +1802,17 @@ static bool btree_insert_key(struct btree *b, struct btree_op *op, | |||
1831 | goto insert; | 1802 | goto insert; |
1832 | 1803 | ||
1833 | /* prev is in the tree, if we merge we're done */ | 1804 | /* prev is in the tree, if we merge we're done */ |
1834 | status = "back merging"; | 1805 | status = BTREE_INSERT_STATUS_BACK_MERGE; |
1835 | if (prev && | 1806 | if (prev && |
1836 | bch_bkey_try_merge(b, prev, k)) | 1807 | bch_bkey_try_merge(b, prev, k)) |
1837 | goto merged; | 1808 | goto merged; |
1838 | 1809 | ||
1839 | status = "overwrote front"; | 1810 | status = BTREE_INSERT_STATUS_OVERWROTE; |
1840 | if (m != end(i) && | 1811 | if (m != end(i) && |
1841 | KEY_PTRS(m) == KEY_PTRS(k) && !KEY_SIZE(m)) | 1812 | KEY_PTRS(m) == KEY_PTRS(k) && !KEY_SIZE(m)) |
1842 | goto copy; | 1813 | goto copy; |
1843 | 1814 | ||
1844 | status = "front merge"; | 1815 | status = BTREE_INSERT_STATUS_FRONT_MERGE; |
1845 | if (m != end(i) && | 1816 | if (m != end(i) && |
1846 | bch_bkey_try_merge(b, k, m)) | 1817 | bch_bkey_try_merge(b, k, m)) |
1847 | goto copy; | 1818 | goto copy; |
@@ -1851,21 +1822,21 @@ static bool btree_insert_key(struct btree *b, struct btree_op *op, | |||
1851 | insert: shift_keys(b, m, k); | 1822 | insert: shift_keys(b, m, k); |
1852 | copy: bkey_copy(m, k); | 1823 | copy: bkey_copy(m, k); |
1853 | merged: | 1824 | merged: |
1854 | bch_check_keys(b, "%s for %s at %s: %s", status, | 1825 | if (KEY_DIRTY(k)) |
1855 | op_type(op), pbtree(b), pkey(k)); | 1826 | bcache_dev_sectors_dirty_add(b->c, KEY_INODE(k), |
1856 | bch_check_key_order_msg(b, i, "%s for %s at %s: %s", status, | 1827 | KEY_START(k), KEY_SIZE(k)); |
1857 | op_type(op), pbtree(b), pkey(k)); | 1828 | |
1829 | bch_check_keys(b, "%u for %s", status, op_type(op)); | ||
1858 | 1830 | ||
1859 | if (b->level && !KEY_OFFSET(k)) | 1831 | if (b->level && !KEY_OFFSET(k)) |
1860 | b->prio_blocked++; | 1832 | btree_current_write(b)->prio_blocked++; |
1861 | 1833 | ||
1862 | pr_debug("%s for %s at %s: %s", status, | 1834 | trace_bcache_btree_insert_key(b, k, op->type, status); |
1863 | op_type(op), pbtree(b), pkey(k)); | ||
1864 | 1835 | ||
1865 | return true; | 1836 | return true; |
1866 | } | 1837 | } |
1867 | 1838 | ||
1868 | bool bch_btree_insert_keys(struct btree *b, struct btree_op *op) | 1839 | static bool bch_btree_insert_keys(struct btree *b, struct btree_op *op) |
1869 | { | 1840 | { |
1870 | bool ret = false; | 1841 | bool ret = false; |
1871 | struct bkey *k; | 1842 | struct bkey *k; |
@@ -1896,7 +1867,7 @@ bool bch_btree_insert_check_key(struct btree *b, struct btree_op *op, | |||
1896 | should_split(b)) | 1867 | should_split(b)) |
1897 | goto out; | 1868 | goto out; |
1898 | 1869 | ||
1899 | op->replace = KEY(op->inode, bio_end(bio), bio_sectors(bio)); | 1870 | op->replace = KEY(op->inode, bio_end_sector(bio), bio_sectors(bio)); |
1900 | 1871 | ||
1901 | SET_KEY_PTRS(&op->replace, 1); | 1872 | SET_KEY_PTRS(&op->replace, 1); |
1902 | get_random_bytes(&op->replace.ptr[0], sizeof(uint64_t)); | 1873 | get_random_bytes(&op->replace.ptr[0], sizeof(uint64_t)); |
@@ -1907,7 +1878,6 @@ bool bch_btree_insert_check_key(struct btree *b, struct btree_op *op, | |||
1907 | 1878 | ||
1908 | BUG_ON(op->type != BTREE_INSERT); | 1879 | BUG_ON(op->type != BTREE_INSERT); |
1909 | BUG_ON(!btree_insert_key(b, op, &tmp.k)); | 1880 | BUG_ON(!btree_insert_key(b, op, &tmp.k)); |
1910 | bch_btree_write(b, false, NULL); | ||
1911 | ret = true; | 1881 | ret = true; |
1912 | out: | 1882 | out: |
1913 | downgrade_write(&b->lock); | 1883 | downgrade_write(&b->lock); |
@@ -1929,12 +1899,11 @@ static int btree_split(struct btree *b, struct btree_op *op) | |||
1929 | 1899 | ||
1930 | split = set_blocks(n1->sets[0].data, n1->c) > (btree_blocks(b) * 4) / 5; | 1900 | split = set_blocks(n1->sets[0].data, n1->c) > (btree_blocks(b) * 4) / 5; |
1931 | 1901 | ||
1932 | pr_debug("%ssplitting at %s keys %i", split ? "" : "not ", | ||
1933 | pbtree(b), n1->sets[0].data->keys); | ||
1934 | |||
1935 | if (split) { | 1902 | if (split) { |
1936 | unsigned keys = 0; | 1903 | unsigned keys = 0; |
1937 | 1904 | ||
1905 | trace_bcache_btree_node_split(b, n1->sets[0].data->keys); | ||
1906 | |||
1938 | n2 = bch_btree_node_alloc(b->c, b->level, &op->cl); | 1907 | n2 = bch_btree_node_alloc(b->c, b->level, &op->cl); |
1939 | if (IS_ERR(n2)) | 1908 | if (IS_ERR(n2)) |
1940 | goto err_free1; | 1909 | goto err_free1; |
@@ -1967,18 +1936,21 @@ static int btree_split(struct btree *b, struct btree_op *op) | |||
1967 | bkey_copy_key(&n2->key, &b->key); | 1936 | bkey_copy_key(&n2->key, &b->key); |
1968 | 1937 | ||
1969 | bch_keylist_add(&op->keys, &n2->key); | 1938 | bch_keylist_add(&op->keys, &n2->key); |
1970 | bch_btree_write(n2, true, op); | 1939 | bch_btree_node_write(n2, &op->cl); |
1971 | rw_unlock(true, n2); | 1940 | rw_unlock(true, n2); |
1972 | } else | 1941 | } else { |
1942 | trace_bcache_btree_node_compact(b, n1->sets[0].data->keys); | ||
1943 | |||
1973 | bch_btree_insert_keys(n1, op); | 1944 | bch_btree_insert_keys(n1, op); |
1945 | } | ||
1974 | 1946 | ||
1975 | bch_keylist_add(&op->keys, &n1->key); | 1947 | bch_keylist_add(&op->keys, &n1->key); |
1976 | bch_btree_write(n1, true, op); | 1948 | bch_btree_node_write(n1, &op->cl); |
1977 | 1949 | ||
1978 | if (n3) { | 1950 | if (n3) { |
1979 | bkey_copy_key(&n3->key, &MAX_KEY); | 1951 | bkey_copy_key(&n3->key, &MAX_KEY); |
1980 | bch_btree_insert_keys(n3, op); | 1952 | bch_btree_insert_keys(n3, op); |
1981 | bch_btree_write(n3, true, op); | 1953 | bch_btree_node_write(n3, &op->cl); |
1982 | 1954 | ||
1983 | closure_sync(&op->cl); | 1955 | closure_sync(&op->cl); |
1984 | bch_btree_set_root(n3); | 1956 | bch_btree_set_root(n3); |
@@ -2082,8 +2054,12 @@ static int bch_btree_insert_recurse(struct btree *b, struct btree_op *op, | |||
2082 | 2054 | ||
2083 | BUG_ON(write_block(b) != b->sets[b->nsets].data); | 2055 | BUG_ON(write_block(b) != b->sets[b->nsets].data); |
2084 | 2056 | ||
2085 | if (bch_btree_insert_keys(b, op)) | 2057 | if (bch_btree_insert_keys(b, op)) { |
2086 | bch_btree_write(b, false, op); | 2058 | if (!b->level) |
2059 | bch_btree_leaf_dirty(b, op); | ||
2060 | else | ||
2061 | bch_btree_node_write(b, &op->cl); | ||
2062 | } | ||
2087 | } | 2063 | } |
2088 | 2064 | ||
2089 | return 0; | 2065 | return 0; |
@@ -2140,6 +2116,11 @@ int bch_btree_insert(struct btree_op *op, struct cache_set *c) | |||
2140 | void bch_btree_set_root(struct btree *b) | 2116 | void bch_btree_set_root(struct btree *b) |
2141 | { | 2117 | { |
2142 | unsigned i; | 2118 | unsigned i; |
2119 | struct closure cl; | ||
2120 | |||
2121 | closure_init_stack(&cl); | ||
2122 | |||
2123 | trace_bcache_btree_set_root(b); | ||
2143 | 2124 | ||
2144 | BUG_ON(!b->written); | 2125 | BUG_ON(!b->written); |
2145 | 2126 | ||
@@ -2153,8 +2134,8 @@ void bch_btree_set_root(struct btree *b) | |||
2153 | b->c->root = b; | 2134 | b->c->root = b; |
2154 | __bkey_put(b->c, &b->key); | 2135 | __bkey_put(b->c, &b->key); |
2155 | 2136 | ||
2156 | bch_journal_meta(b->c, NULL); | 2137 | bch_journal_meta(b->c, &cl); |
2157 | pr_debug("%s for %pf", pbtree(b), __builtin_return_address(0)); | 2138 | closure_sync(&cl); |
2158 | } | 2139 | } |
2159 | 2140 | ||
2160 | /* Cache lookup */ | 2141 | /* Cache lookup */ |
@@ -2215,9 +2196,6 @@ static int submit_partial_cache_hit(struct btree *b, struct btree_op *op, | |||
2215 | KEY_OFFSET(k) - bio->bi_sector); | 2196 | KEY_OFFSET(k) - bio->bi_sector); |
2216 | 2197 | ||
2217 | n = bch_bio_split(bio, sectors, GFP_NOIO, s->d->bio_split); | 2198 | n = bch_bio_split(bio, sectors, GFP_NOIO, s->d->bio_split); |
2218 | if (!n) | ||
2219 | return -EAGAIN; | ||
2220 | |||
2221 | if (n == bio) | 2199 | if (n == bio) |
2222 | op->lookup_done = true; | 2200 | op->lookup_done = true; |
2223 | 2201 | ||
@@ -2240,7 +2218,6 @@ static int submit_partial_cache_hit(struct btree *b, struct btree_op *op, | |||
2240 | n->bi_end_io = bch_cache_read_endio; | 2218 | n->bi_end_io = bch_cache_read_endio; |
2241 | n->bi_private = &s->cl; | 2219 | n->bi_private = &s->cl; |
2242 | 2220 | ||
2243 | trace_bcache_cache_hit(n); | ||
2244 | __bch_submit_bbio(n, b->c); | 2221 | __bch_submit_bbio(n, b->c); |
2245 | } | 2222 | } |
2246 | 2223 | ||
@@ -2257,9 +2234,6 @@ int bch_btree_search_recurse(struct btree *b, struct btree_op *op) | |||
2257 | struct btree_iter iter; | 2234 | struct btree_iter iter; |
2258 | bch_btree_iter_init(b, &iter, &KEY(op->inode, bio->bi_sector, 0)); | 2235 | bch_btree_iter_init(b, &iter, &KEY(op->inode, bio->bi_sector, 0)); |
2259 | 2236 | ||
2260 | pr_debug("at %s searching for %u:%llu", pbtree(b), op->inode, | ||
2261 | (uint64_t) bio->bi_sector); | ||
2262 | |||
2263 | do { | 2237 | do { |
2264 | k = bch_btree_iter_next_filter(&iter, b, bch_ptr_bad); | 2238 | k = bch_btree_iter_next_filter(&iter, b, bch_ptr_bad); |
2265 | if (!k) { | 2239 | if (!k) { |
@@ -2303,7 +2277,8 @@ static inline int keybuf_nonoverlapping_cmp(struct keybuf_key *l, | |||
2303 | } | 2277 | } |
2304 | 2278 | ||
2305 | static int bch_btree_refill_keybuf(struct btree *b, struct btree_op *op, | 2279 | static int bch_btree_refill_keybuf(struct btree *b, struct btree_op *op, |
2306 | struct keybuf *buf, struct bkey *end) | 2280 | struct keybuf *buf, struct bkey *end, |
2281 | keybuf_pred_fn *pred) | ||
2307 | { | 2282 | { |
2308 | struct btree_iter iter; | 2283 | struct btree_iter iter; |
2309 | bch_btree_iter_init(b, &iter, &buf->last_scanned); | 2284 | bch_btree_iter_init(b, &iter, &buf->last_scanned); |
@@ -2322,11 +2297,9 @@ static int bch_btree_refill_keybuf(struct btree *b, struct btree_op *op, | |||
2322 | if (bkey_cmp(&buf->last_scanned, end) >= 0) | 2297 | if (bkey_cmp(&buf->last_scanned, end) >= 0) |
2323 | break; | 2298 | break; |
2324 | 2299 | ||
2325 | if (buf->key_predicate(buf, k)) { | 2300 | if (pred(buf, k)) { |
2326 | struct keybuf_key *w; | 2301 | struct keybuf_key *w; |
2327 | 2302 | ||
2328 | pr_debug("%s", pkey(k)); | ||
2329 | |||
2330 | spin_lock(&buf->lock); | 2303 | spin_lock(&buf->lock); |
2331 | 2304 | ||
2332 | w = array_alloc(&buf->freelist); | 2305 | w = array_alloc(&buf->freelist); |
@@ -2343,7 +2316,7 @@ static int bch_btree_refill_keybuf(struct btree *b, struct btree_op *op, | |||
2343 | if (!k) | 2316 | if (!k) |
2344 | break; | 2317 | break; |
2345 | 2318 | ||
2346 | btree(refill_keybuf, k, b, op, buf, end); | 2319 | btree(refill_keybuf, k, b, op, buf, end, pred); |
2347 | /* | 2320 | /* |
2348 | * Might get an error here, but can't really do anything | 2321 | * Might get an error here, but can't really do anything |
2349 | * and it'll get logged elsewhere. Just read what we | 2322 | * and it'll get logged elsewhere. Just read what we |
@@ -2361,7 +2334,7 @@ static int bch_btree_refill_keybuf(struct btree *b, struct btree_op *op, | |||
2361 | } | 2334 | } |
2362 | 2335 | ||
2363 | void bch_refill_keybuf(struct cache_set *c, struct keybuf *buf, | 2336 | void bch_refill_keybuf(struct cache_set *c, struct keybuf *buf, |
2364 | struct bkey *end) | 2337 | struct bkey *end, keybuf_pred_fn *pred) |
2365 | { | 2338 | { |
2366 | struct bkey start = buf->last_scanned; | 2339 | struct bkey start = buf->last_scanned; |
2367 | struct btree_op op; | 2340 | struct btree_op op; |
@@ -2369,7 +2342,7 @@ void bch_refill_keybuf(struct cache_set *c, struct keybuf *buf, | |||
2369 | 2342 | ||
2370 | cond_resched(); | 2343 | cond_resched(); |
2371 | 2344 | ||
2372 | btree_root(refill_keybuf, c, &op, buf, end); | 2345 | btree_root(refill_keybuf, c, &op, buf, end, pred); |
2373 | closure_sync(&op.cl); | 2346 | closure_sync(&op.cl); |
2374 | 2347 | ||
2375 | pr_debug("found %s keys from %llu:%llu to %llu:%llu", | 2348 | pr_debug("found %s keys from %llu:%llu to %llu:%llu", |
@@ -2455,7 +2428,8 @@ struct keybuf_key *bch_keybuf_next(struct keybuf *buf) | |||
2455 | 2428 | ||
2456 | struct keybuf_key *bch_keybuf_next_rescan(struct cache_set *c, | 2429 | struct keybuf_key *bch_keybuf_next_rescan(struct cache_set *c, |
2457 | struct keybuf *buf, | 2430 | struct keybuf *buf, |
2458 | struct bkey *end) | 2431 | struct bkey *end, |
2432 | keybuf_pred_fn *pred) | ||
2459 | { | 2433 | { |
2460 | struct keybuf_key *ret; | 2434 | struct keybuf_key *ret; |
2461 | 2435 | ||
@@ -2469,15 +2443,14 @@ struct keybuf_key *bch_keybuf_next_rescan(struct cache_set *c, | |||
2469 | break; | 2443 | break; |
2470 | } | 2444 | } |
2471 | 2445 | ||
2472 | bch_refill_keybuf(c, buf, end); | 2446 | bch_refill_keybuf(c, buf, end, pred); |
2473 | } | 2447 | } |
2474 | 2448 | ||
2475 | return ret; | 2449 | return ret; |
2476 | } | 2450 | } |
2477 | 2451 | ||
2478 | void bch_keybuf_init(struct keybuf *buf, keybuf_pred_fn *fn) | 2452 | void bch_keybuf_init(struct keybuf *buf) |
2479 | { | 2453 | { |
2480 | buf->key_predicate = fn; | ||
2481 | buf->last_scanned = MAX_KEY; | 2454 | buf->last_scanned = MAX_KEY; |
2482 | buf->keys = RB_ROOT; | 2455 | buf->keys = RB_ROOT; |
2483 | 2456 | ||
diff --git a/drivers/md/bcache/btree.h b/drivers/md/bcache/btree.h index af4a7092a28c..3333d3723633 100644 --- a/drivers/md/bcache/btree.h +++ b/drivers/md/bcache/btree.h | |||
@@ -102,7 +102,6 @@ | |||
102 | #include "debug.h" | 102 | #include "debug.h" |
103 | 103 | ||
104 | struct btree_write { | 104 | struct btree_write { |
105 | struct closure *owner; | ||
106 | atomic_t *journal; | 105 | atomic_t *journal; |
107 | 106 | ||
108 | /* If btree_split() frees a btree node, it writes a new pointer to that | 107 | /* If btree_split() frees a btree node, it writes a new pointer to that |
@@ -142,16 +141,12 @@ struct btree { | |||
142 | */ | 141 | */ |
143 | struct bset_tree sets[MAX_BSETS]; | 142 | struct bset_tree sets[MAX_BSETS]; |
144 | 143 | ||
145 | /* Used to refcount bio splits, also protects b->bio */ | 144 | /* For outstanding btree writes, used as a lock - protects write_idx */ |
146 | struct closure_with_waitlist io; | 145 | struct closure_with_waitlist io; |
147 | 146 | ||
148 | /* Gets transferred to w->prio_blocked - see the comment there */ | ||
149 | int prio_blocked; | ||
150 | |||
151 | struct list_head list; | 147 | struct list_head list; |
152 | struct delayed_work work; | 148 | struct delayed_work work; |
153 | 149 | ||
154 | uint64_t io_start_time; | ||
155 | struct btree_write writes[2]; | 150 | struct btree_write writes[2]; |
156 | struct bio *bio; | 151 | struct bio *bio; |
157 | }; | 152 | }; |
@@ -164,13 +159,11 @@ static inline void set_btree_node_ ## flag(struct btree *b) \ | |||
164 | { set_bit(BTREE_NODE_ ## flag, &b->flags); } \ | 159 | { set_bit(BTREE_NODE_ ## flag, &b->flags); } \ |
165 | 160 | ||
166 | enum btree_flags { | 161 | enum btree_flags { |
167 | BTREE_NODE_read_done, | ||
168 | BTREE_NODE_io_error, | 162 | BTREE_NODE_io_error, |
169 | BTREE_NODE_dirty, | 163 | BTREE_NODE_dirty, |
170 | BTREE_NODE_write_idx, | 164 | BTREE_NODE_write_idx, |
171 | }; | 165 | }; |
172 | 166 | ||
173 | BTREE_FLAG(read_done); | ||
174 | BTREE_FLAG(io_error); | 167 | BTREE_FLAG(io_error); |
175 | BTREE_FLAG(dirty); | 168 | BTREE_FLAG(dirty); |
176 | BTREE_FLAG(write_idx); | 169 | BTREE_FLAG(write_idx); |
@@ -278,6 +271,13 @@ struct btree_op { | |||
278 | BKEY_PADDED(replace); | 271 | BKEY_PADDED(replace); |
279 | }; | 272 | }; |
280 | 273 | ||
274 | enum { | ||
275 | BTREE_INSERT_STATUS_INSERT, | ||
276 | BTREE_INSERT_STATUS_BACK_MERGE, | ||
277 | BTREE_INSERT_STATUS_OVERWROTE, | ||
278 | BTREE_INSERT_STATUS_FRONT_MERGE, | ||
279 | }; | ||
280 | |||
281 | void bch_btree_op_init_stack(struct btree_op *); | 281 | void bch_btree_op_init_stack(struct btree_op *); |
282 | 282 | ||
283 | static inline void rw_lock(bool w, struct btree *b, int level) | 283 | static inline void rw_lock(bool w, struct btree *b, int level) |
@@ -293,9 +293,7 @@ static inline void rw_unlock(bool w, struct btree *b) | |||
293 | #ifdef CONFIG_BCACHE_EDEBUG | 293 | #ifdef CONFIG_BCACHE_EDEBUG |
294 | unsigned i; | 294 | unsigned i; |
295 | 295 | ||
296 | if (w && | 296 | if (w && b->key.ptr[0]) |
297 | b->key.ptr[0] && | ||
298 | btree_node_read_done(b)) | ||
299 | for (i = 0; i <= b->nsets; i++) | 297 | for (i = 0; i <= b->nsets; i++) |
300 | bch_check_key_order(b, b->sets[i].data); | 298 | bch_check_key_order(b, b->sets[i].data); |
301 | #endif | 299 | #endif |
@@ -370,9 +368,8 @@ static inline bool should_split(struct btree *b) | |||
370 | > btree_blocks(b)); | 368 | > btree_blocks(b)); |
371 | } | 369 | } |
372 | 370 | ||
373 | void bch_btree_read_done(struct closure *); | 371 | void bch_btree_node_read(struct btree *); |
374 | void bch_btree_read(struct btree *); | 372 | void bch_btree_node_write(struct btree *, struct closure *); |
375 | void bch_btree_write(struct btree *b, bool now, struct btree_op *op); | ||
376 | 373 | ||
377 | void bch_cannibalize_unlock(struct cache_set *, struct closure *); | 374 | void bch_cannibalize_unlock(struct cache_set *, struct closure *); |
378 | void bch_btree_set_root(struct btree *); | 375 | void bch_btree_set_root(struct btree *); |
@@ -380,7 +377,6 @@ struct btree *bch_btree_node_alloc(struct cache_set *, int, struct closure *); | |||
380 | struct btree *bch_btree_node_get(struct cache_set *, struct bkey *, | 377 | struct btree *bch_btree_node_get(struct cache_set *, struct bkey *, |
381 | int, struct btree_op *); | 378 | int, struct btree_op *); |
382 | 379 | ||
383 | bool bch_btree_insert_keys(struct btree *, struct btree_op *); | ||
384 | bool bch_btree_insert_check_key(struct btree *, struct btree_op *, | 380 | bool bch_btree_insert_check_key(struct btree *, struct btree_op *, |
385 | struct bio *); | 381 | struct bio *); |
386 | int bch_btree_insert(struct btree_op *, struct cache_set *); | 382 | int bch_btree_insert(struct btree_op *, struct cache_set *); |
@@ -393,13 +389,14 @@ void bch_moving_gc(struct closure *); | |||
393 | int bch_btree_check(struct cache_set *, struct btree_op *); | 389 | int bch_btree_check(struct cache_set *, struct btree_op *); |
394 | uint8_t __bch_btree_mark_key(struct cache_set *, int, struct bkey *); | 390 | uint8_t __bch_btree_mark_key(struct cache_set *, int, struct bkey *); |
395 | 391 | ||
396 | void bch_keybuf_init(struct keybuf *, keybuf_pred_fn *); | 392 | void bch_keybuf_init(struct keybuf *); |
397 | void bch_refill_keybuf(struct cache_set *, struct keybuf *, struct bkey *); | 393 | void bch_refill_keybuf(struct cache_set *, struct keybuf *, struct bkey *, |
394 | keybuf_pred_fn *); | ||
398 | bool bch_keybuf_check_overlapping(struct keybuf *, struct bkey *, | 395 | bool bch_keybuf_check_overlapping(struct keybuf *, struct bkey *, |
399 | struct bkey *); | 396 | struct bkey *); |
400 | void bch_keybuf_del(struct keybuf *, struct keybuf_key *); | 397 | void bch_keybuf_del(struct keybuf *, struct keybuf_key *); |
401 | struct keybuf_key *bch_keybuf_next(struct keybuf *); | 398 | struct keybuf_key *bch_keybuf_next(struct keybuf *); |
402 | struct keybuf_key *bch_keybuf_next_rescan(struct cache_set *, | 399 | struct keybuf_key *bch_keybuf_next_rescan(struct cache_set *, struct keybuf *, |
403 | struct keybuf *, struct bkey *); | 400 | struct bkey *, keybuf_pred_fn *); |
404 | 401 | ||
405 | #endif | 402 | #endif |
diff --git a/drivers/md/bcache/closure.c b/drivers/md/bcache/closure.c index bd05a9a8c7cf..9aba2017f0d1 100644 --- a/drivers/md/bcache/closure.c +++ b/drivers/md/bcache/closure.c | |||
@@ -66,16 +66,18 @@ static inline void closure_put_after_sub(struct closure *cl, int flags) | |||
66 | } else { | 66 | } else { |
67 | struct closure *parent = cl->parent; | 67 | struct closure *parent = cl->parent; |
68 | struct closure_waitlist *wait = closure_waitlist(cl); | 68 | struct closure_waitlist *wait = closure_waitlist(cl); |
69 | closure_fn *destructor = cl->fn; | ||
69 | 70 | ||
70 | closure_debug_destroy(cl); | 71 | closure_debug_destroy(cl); |
71 | 72 | ||
73 | smp_mb(); | ||
72 | atomic_set(&cl->remaining, -1); | 74 | atomic_set(&cl->remaining, -1); |
73 | 75 | ||
74 | if (wait) | 76 | if (wait) |
75 | closure_wake_up(wait); | 77 | closure_wake_up(wait); |
76 | 78 | ||
77 | if (cl->fn) | 79 | if (destructor) |
78 | cl->fn(cl); | 80 | destructor(cl); |
79 | 81 | ||
80 | if (parent) | 82 | if (parent) |
81 | closure_put(parent); | 83 | closure_put(parent); |
diff --git a/drivers/md/bcache/debug.c b/drivers/md/bcache/debug.c index 89fd5204924e..88e6411eab4f 100644 --- a/drivers/md/bcache/debug.c +++ b/drivers/md/bcache/debug.c | |||
@@ -47,11 +47,10 @@ const char *bch_ptr_status(struct cache_set *c, const struct bkey *k) | |||
47 | return ""; | 47 | return ""; |
48 | } | 48 | } |
49 | 49 | ||
50 | struct keyprint_hack bch_pkey(const struct bkey *k) | 50 | int bch_bkey_to_text(char *buf, size_t size, const struct bkey *k) |
51 | { | 51 | { |
52 | unsigned i = 0; | 52 | unsigned i = 0; |
53 | struct keyprint_hack r; | 53 | char *out = buf, *end = buf + size; |
54 | char *out = r.s, *end = r.s + KEYHACK_SIZE; | ||
55 | 54 | ||
56 | #define p(...) (out += scnprintf(out, end - out, __VA_ARGS__)) | 55 | #define p(...) (out += scnprintf(out, end - out, __VA_ARGS__)) |
57 | 56 | ||
@@ -75,16 +74,14 @@ struct keyprint_hack bch_pkey(const struct bkey *k) | |||
75 | if (KEY_CSUM(k)) | 74 | if (KEY_CSUM(k)) |
76 | p(" cs%llu %llx", KEY_CSUM(k), k->ptr[1]); | 75 | p(" cs%llu %llx", KEY_CSUM(k), k->ptr[1]); |
77 | #undef p | 76 | #undef p |
78 | return r; | 77 | return out - buf; |
79 | } | 78 | } |
80 | 79 | ||
81 | struct keyprint_hack bch_pbtree(const struct btree *b) | 80 | int bch_btree_to_text(char *buf, size_t size, const struct btree *b) |
82 | { | 81 | { |
83 | struct keyprint_hack r; | 82 | return scnprintf(buf, size, "%zu level %i/%i", |
84 | 83 | PTR_BUCKET_NR(b->c, &b->key, 0), | |
85 | snprintf(r.s, 40, "%zu level %i/%i", PTR_BUCKET_NR(b->c, &b->key, 0), | 84 | b->level, b->c->root ? b->c->root->level : -1); |
86 | b->level, b->c->root ? b->c->root->level : -1); | ||
87 | return r; | ||
88 | } | 85 | } |
89 | 86 | ||
90 | #if defined(CONFIG_BCACHE_DEBUG) || defined(CONFIG_BCACHE_EDEBUG) | 87 | #if defined(CONFIG_BCACHE_DEBUG) || defined(CONFIG_BCACHE_EDEBUG) |
@@ -100,10 +97,12 @@ static void dump_bset(struct btree *b, struct bset *i) | |||
100 | { | 97 | { |
101 | struct bkey *k; | 98 | struct bkey *k; |
102 | unsigned j; | 99 | unsigned j; |
100 | char buf[80]; | ||
103 | 101 | ||
104 | for (k = i->start; k < end(i); k = bkey_next(k)) { | 102 | for (k = i->start; k < end(i); k = bkey_next(k)) { |
103 | bch_bkey_to_text(buf, sizeof(buf), k); | ||
105 | printk(KERN_ERR "block %zu key %zi/%u: %s", index(i, b), | 104 | printk(KERN_ERR "block %zu key %zi/%u: %s", index(i, b), |
106 | (uint64_t *) k - i->d, i->keys, pkey(k)); | 105 | (uint64_t *) k - i->d, i->keys, buf); |
107 | 106 | ||
108 | for (j = 0; j < KEY_PTRS(k); j++) { | 107 | for (j = 0; j < KEY_PTRS(k); j++) { |
109 | size_t n = PTR_BUCKET_NR(b->c, k, j); | 108 | size_t n = PTR_BUCKET_NR(b->c, k, j); |
@@ -144,7 +143,7 @@ void bch_btree_verify(struct btree *b, struct bset *new) | |||
144 | v->written = 0; | 143 | v->written = 0; |
145 | v->level = b->level; | 144 | v->level = b->level; |
146 | 145 | ||
147 | bch_btree_read(v); | 146 | bch_btree_node_read(v); |
148 | closure_wait_event(&v->io.wait, &cl, | 147 | closure_wait_event(&v->io.wait, &cl, |
149 | atomic_read(&b->io.cl.remaining) == -1); | 148 | atomic_read(&b->io.cl.remaining) == -1); |
150 | 149 | ||
@@ -200,7 +199,7 @@ void bch_data_verify(struct search *s) | |||
200 | if (!check) | 199 | if (!check) |
201 | return; | 200 | return; |
202 | 201 | ||
203 | if (bch_bio_alloc_pages(check, GFP_NOIO)) | 202 | if (bio_alloc_pages(check, GFP_NOIO)) |
204 | goto out_put; | 203 | goto out_put; |
205 | 204 | ||
206 | check->bi_rw = READ_SYNC; | 205 | check->bi_rw = READ_SYNC; |
@@ -252,6 +251,7 @@ static void vdump_bucket_and_panic(struct btree *b, const char *fmt, | |||
252 | va_list args) | 251 | va_list args) |
253 | { | 252 | { |
254 | unsigned i; | 253 | unsigned i; |
254 | char buf[80]; | ||
255 | 255 | ||
256 | console_lock(); | 256 | console_lock(); |
257 | 257 | ||
@@ -262,7 +262,8 @@ static void vdump_bucket_and_panic(struct btree *b, const char *fmt, | |||
262 | 262 | ||
263 | console_unlock(); | 263 | console_unlock(); |
264 | 264 | ||
265 | panic("at %s\n", pbtree(b)); | 265 | bch_btree_to_text(buf, sizeof(buf), b); |
266 | panic("at %s\n", buf); | ||
266 | } | 267 | } |
267 | 268 | ||
268 | void bch_check_key_order_msg(struct btree *b, struct bset *i, | 269 | void bch_check_key_order_msg(struct btree *b, struct bset *i, |
@@ -337,6 +338,7 @@ static ssize_t bch_dump_read(struct file *file, char __user *buf, | |||
337 | { | 338 | { |
338 | struct dump_iterator *i = file->private_data; | 339 | struct dump_iterator *i = file->private_data; |
339 | ssize_t ret = 0; | 340 | ssize_t ret = 0; |
341 | char kbuf[80]; | ||
340 | 342 | ||
341 | while (size) { | 343 | while (size) { |
342 | struct keybuf_key *w; | 344 | struct keybuf_key *w; |
@@ -355,11 +357,12 @@ static ssize_t bch_dump_read(struct file *file, char __user *buf, | |||
355 | if (i->bytes) | 357 | if (i->bytes) |
356 | break; | 358 | break; |
357 | 359 | ||
358 | w = bch_keybuf_next_rescan(i->c, &i->keys, &MAX_KEY); | 360 | w = bch_keybuf_next_rescan(i->c, &i->keys, &MAX_KEY, dump_pred); |
359 | if (!w) | 361 | if (!w) |
360 | break; | 362 | break; |
361 | 363 | ||
362 | i->bytes = snprintf(i->buf, PAGE_SIZE, "%s\n", pkey(&w->key)); | 364 | bch_bkey_to_text(kbuf, sizeof(kbuf), &w->key); |
365 | i->bytes = snprintf(i->buf, PAGE_SIZE, "%s\n", kbuf); | ||
363 | bch_keybuf_del(&i->keys, w); | 366 | bch_keybuf_del(&i->keys, w); |
364 | } | 367 | } |
365 | 368 | ||
@@ -377,7 +380,7 @@ static int bch_dump_open(struct inode *inode, struct file *file) | |||
377 | 380 | ||
378 | file->private_data = i; | 381 | file->private_data = i; |
379 | i->c = c; | 382 | i->c = c; |
380 | bch_keybuf_init(&i->keys, dump_pred); | 383 | bch_keybuf_init(&i->keys); |
381 | i->keys.last_scanned = KEY(0, 0, 0); | 384 | i->keys.last_scanned = KEY(0, 0, 0); |
382 | 385 | ||
383 | return 0; | 386 | return 0; |
@@ -409,142 +412,6 @@ void bch_debug_init_cache_set(struct cache_set *c) | |||
409 | 412 | ||
410 | #endif | 413 | #endif |
411 | 414 | ||
412 | /* Fuzz tester has rotted: */ | ||
413 | #if 0 | ||
414 | |||
415 | static ssize_t btree_fuzz(struct kobject *k, struct kobj_attribute *a, | ||
416 | const char *buffer, size_t size) | ||
417 | { | ||
418 | void dump(struct btree *b) | ||
419 | { | ||
420 | struct bset *i; | ||
421 | |||
422 | for (i = b->sets[0].data; | ||
423 | index(i, b) < btree_blocks(b) && | ||
424 | i->seq == b->sets[0].data->seq; | ||
425 | i = ((void *) i) + set_blocks(i, b->c) * block_bytes(b->c)) | ||
426 | dump_bset(b, i); | ||
427 | } | ||
428 | |||
429 | struct cache_sb *sb; | ||
430 | struct cache_set *c; | ||
431 | struct btree *all[3], *b, *fill, *orig; | ||
432 | int j; | ||
433 | |||
434 | struct btree_op op; | ||
435 | bch_btree_op_init_stack(&op); | ||
436 | |||
437 | sb = kzalloc(sizeof(struct cache_sb), GFP_KERNEL); | ||
438 | if (!sb) | ||
439 | return -ENOMEM; | ||
440 | |||
441 | sb->bucket_size = 128; | ||
442 | sb->block_size = 4; | ||
443 | |||
444 | c = bch_cache_set_alloc(sb); | ||
445 | if (!c) | ||
446 | return -ENOMEM; | ||
447 | |||
448 | for (j = 0; j < 3; j++) { | ||
449 | BUG_ON(list_empty(&c->btree_cache)); | ||
450 | all[j] = list_first_entry(&c->btree_cache, struct btree, list); | ||
451 | list_del_init(&all[j]->list); | ||
452 | |||
453 | all[j]->key = KEY(0, 0, c->sb.bucket_size); | ||
454 | bkey_copy_key(&all[j]->key, &MAX_KEY); | ||
455 | } | ||
456 | |||
457 | b = all[0]; | ||
458 | fill = all[1]; | ||
459 | orig = all[2]; | ||
460 | |||
461 | while (1) { | ||
462 | for (j = 0; j < 3; j++) | ||
463 | all[j]->written = all[j]->nsets = 0; | ||
464 | |||
465 | bch_bset_init_next(b); | ||
466 | |||
467 | while (1) { | ||
468 | struct bset *i = write_block(b); | ||
469 | struct bkey *k = op.keys.top; | ||
470 | unsigned rand; | ||
471 | |||
472 | bkey_init(k); | ||
473 | rand = get_random_int(); | ||
474 | |||
475 | op.type = rand & 1 | ||
476 | ? BTREE_INSERT | ||
477 | : BTREE_REPLACE; | ||
478 | rand >>= 1; | ||
479 | |||
480 | SET_KEY_SIZE(k, bucket_remainder(c, rand)); | ||
481 | rand >>= c->bucket_bits; | ||
482 | rand &= 1024 * 512 - 1; | ||
483 | rand += c->sb.bucket_size; | ||
484 | SET_KEY_OFFSET(k, rand); | ||
485 | #if 0 | ||
486 | SET_KEY_PTRS(k, 1); | ||
487 | #endif | ||
488 | bch_keylist_push(&op.keys); | ||
489 | bch_btree_insert_keys(b, &op); | ||
490 | |||
491 | if (should_split(b) || | ||
492 | set_blocks(i, b->c) != | ||
493 | __set_blocks(i, i->keys + 15, b->c)) { | ||
494 | i->csum = csum_set(i); | ||
495 | |||
496 | memcpy(write_block(fill), | ||
497 | i, set_bytes(i)); | ||
498 | |||
499 | b->written += set_blocks(i, b->c); | ||
500 | fill->written = b->written; | ||
501 | if (b->written == btree_blocks(b)) | ||
502 | break; | ||
503 | |||
504 | bch_btree_sort_lazy(b); | ||
505 | bch_bset_init_next(b); | ||
506 | } | ||
507 | } | ||
508 | |||
509 | memcpy(orig->sets[0].data, | ||
510 | fill->sets[0].data, | ||
511 | btree_bytes(c)); | ||
512 | |||
513 | bch_btree_sort(b); | ||
514 | fill->written = 0; | ||
515 | bch_btree_read_done(&fill->io.cl); | ||
516 | |||
517 | if (b->sets[0].data->keys != fill->sets[0].data->keys || | ||
518 | memcmp(b->sets[0].data->start, | ||
519 | fill->sets[0].data->start, | ||
520 | b->sets[0].data->keys * sizeof(uint64_t))) { | ||
521 | struct bset *i = b->sets[0].data; | ||
522 | struct bkey *k, *l; | ||
523 | |||
524 | for (k = i->start, | ||
525 | l = fill->sets[0].data->start; | ||
526 | k < end(i); | ||
527 | k = bkey_next(k), l = bkey_next(l)) | ||
528 | if (bkey_cmp(k, l) || | ||
529 | KEY_SIZE(k) != KEY_SIZE(l)) | ||
530 | pr_err("key %zi differs: %s != %s", | ||
531 | (uint64_t *) k - i->d, | ||
532 | pkey(k), pkey(l)); | ||
533 | |||
534 | for (j = 0; j < 3; j++) { | ||
535 | pr_err("**** Set %i ****", j); | ||
536 | dump(all[j]); | ||
537 | } | ||
538 | panic("\n"); | ||
539 | } | ||
540 | |||
541 | pr_info("fuzz complete: %i keys", b->sets[0].data->keys); | ||
542 | } | ||
543 | } | ||
544 | |||
545 | kobj_attribute_write(fuzz, btree_fuzz); | ||
546 | #endif | ||
547 | |||
548 | void bch_debug_exit(void) | 415 | void bch_debug_exit(void) |
549 | { | 416 | { |
550 | if (!IS_ERR_OR_NULL(debug)) | 417 | if (!IS_ERR_OR_NULL(debug)) |
@@ -554,11 +421,6 @@ void bch_debug_exit(void) | |||
554 | int __init bch_debug_init(struct kobject *kobj) | 421 | int __init bch_debug_init(struct kobject *kobj) |
555 | { | 422 | { |
556 | int ret = 0; | 423 | int ret = 0; |
557 | #if 0 | ||
558 | ret = sysfs_create_file(kobj, &ksysfs_fuzz.attr); | ||
559 | if (ret) | ||
560 | return ret; | ||
561 | #endif | ||
562 | 424 | ||
563 | debug = debugfs_create_dir("bcache", NULL); | 425 | debug = debugfs_create_dir("bcache", NULL); |
564 | return ret; | 426 | return ret; |
diff --git a/drivers/md/bcache/debug.h b/drivers/md/bcache/debug.h index f9378a218148..1c39b5a2489b 100644 --- a/drivers/md/bcache/debug.h +++ b/drivers/md/bcache/debug.h | |||
@@ -3,15 +3,8 @@ | |||
3 | 3 | ||
4 | /* Btree/bkey debug printing */ | 4 | /* Btree/bkey debug printing */ |
5 | 5 | ||
6 | #define KEYHACK_SIZE 80 | 6 | int bch_bkey_to_text(char *buf, size_t size, const struct bkey *k); |
7 | struct keyprint_hack { | 7 | int bch_btree_to_text(char *buf, size_t size, const struct btree *b); |
8 | char s[KEYHACK_SIZE]; | ||
9 | }; | ||
10 | |||
11 | struct keyprint_hack bch_pkey(const struct bkey *k); | ||
12 | struct keyprint_hack bch_pbtree(const struct btree *b); | ||
13 | #define pkey(k) (&bch_pkey(k).s[0]) | ||
14 | #define pbtree(b) (&bch_pbtree(b).s[0]) | ||
15 | 8 | ||
16 | #ifdef CONFIG_BCACHE_EDEBUG | 9 | #ifdef CONFIG_BCACHE_EDEBUG |
17 | 10 | ||
diff --git a/drivers/md/bcache/io.c b/drivers/md/bcache/io.c index 48efd4dea645..9056632995b1 100644 --- a/drivers/md/bcache/io.c +++ b/drivers/md/bcache/io.c | |||
@@ -9,6 +9,8 @@ | |||
9 | #include "bset.h" | 9 | #include "bset.h" |
10 | #include "debug.h" | 10 | #include "debug.h" |
11 | 11 | ||
12 | #include <linux/blkdev.h> | ||
13 | |||
12 | static void bch_bi_idx_hack_endio(struct bio *bio, int error) | 14 | static void bch_bi_idx_hack_endio(struct bio *bio, int error) |
13 | { | 15 | { |
14 | struct bio *p = bio->bi_private; | 16 | struct bio *p = bio->bi_private; |
@@ -66,13 +68,6 @@ static void bch_generic_make_request_hack(struct bio *bio) | |||
66 | * The newly allocated bio will point to @bio's bi_io_vec, if the split was on a | 68 | * The newly allocated bio will point to @bio's bi_io_vec, if the split was on a |
67 | * bvec boundry; it is the caller's responsibility to ensure that @bio is not | 69 | * bvec boundry; it is the caller's responsibility to ensure that @bio is not |
68 | * freed before the split. | 70 | * freed before the split. |
69 | * | ||
70 | * If bch_bio_split() is running under generic_make_request(), it's not safe to | ||
71 | * allocate more than one bio from the same bio set. Therefore, if it is running | ||
72 | * under generic_make_request() it masks out __GFP_WAIT when doing the | ||
73 | * allocation. The caller must check for failure if there's any possibility of | ||
74 | * it being called from under generic_make_request(); it is then the caller's | ||
75 | * responsibility to retry from a safe context (by e.g. punting to workqueue). | ||
76 | */ | 71 | */ |
77 | struct bio *bch_bio_split(struct bio *bio, int sectors, | 72 | struct bio *bch_bio_split(struct bio *bio, int sectors, |
78 | gfp_t gfp, struct bio_set *bs) | 73 | gfp_t gfp, struct bio_set *bs) |
@@ -83,20 +78,13 @@ struct bio *bch_bio_split(struct bio *bio, int sectors, | |||
83 | 78 | ||
84 | BUG_ON(sectors <= 0); | 79 | BUG_ON(sectors <= 0); |
85 | 80 | ||
86 | /* | ||
87 | * If we're being called from underneath generic_make_request() and we | ||
88 | * already allocated any bios from this bio set, we risk deadlock if we | ||
89 | * use the mempool. So instead, we possibly fail and let the caller punt | ||
90 | * to workqueue or somesuch and retry in a safe context. | ||
91 | */ | ||
92 | if (current->bio_list) | ||
93 | gfp &= ~__GFP_WAIT; | ||
94 | |||
95 | if (sectors >= bio_sectors(bio)) | 81 | if (sectors >= bio_sectors(bio)) |
96 | return bio; | 82 | return bio; |
97 | 83 | ||
98 | if (bio->bi_rw & REQ_DISCARD) { | 84 | if (bio->bi_rw & REQ_DISCARD) { |
99 | ret = bio_alloc_bioset(gfp, 1, bs); | 85 | ret = bio_alloc_bioset(gfp, 1, bs); |
86 | if (!ret) | ||
87 | return NULL; | ||
100 | idx = 0; | 88 | idx = 0; |
101 | goto out; | 89 | goto out; |
102 | } | 90 | } |
@@ -160,17 +148,18 @@ static unsigned bch_bio_max_sectors(struct bio *bio) | |||
160 | struct request_queue *q = bdev_get_queue(bio->bi_bdev); | 148 | struct request_queue *q = bdev_get_queue(bio->bi_bdev); |
161 | unsigned max_segments = min_t(unsigned, BIO_MAX_PAGES, | 149 | unsigned max_segments = min_t(unsigned, BIO_MAX_PAGES, |
162 | queue_max_segments(q)); | 150 | queue_max_segments(q)); |
163 | struct bio_vec *bv, *end = bio_iovec(bio) + | ||
164 | min_t(int, bio_segments(bio), max_segments); | ||
165 | 151 | ||
166 | if (bio->bi_rw & REQ_DISCARD) | 152 | if (bio->bi_rw & REQ_DISCARD) |
167 | return min(ret, q->limits.max_discard_sectors); | 153 | return min(ret, q->limits.max_discard_sectors); |
168 | 154 | ||
169 | if (bio_segments(bio) > max_segments || | 155 | if (bio_segments(bio) > max_segments || |
170 | q->merge_bvec_fn) { | 156 | q->merge_bvec_fn) { |
157 | struct bio_vec *bv; | ||
158 | int i, seg = 0; | ||
159 | |||
171 | ret = 0; | 160 | ret = 0; |
172 | 161 | ||
173 | for (bv = bio_iovec(bio); bv < end; bv++) { | 162 | bio_for_each_segment(bv, bio, i) { |
174 | struct bvec_merge_data bvm = { | 163 | struct bvec_merge_data bvm = { |
175 | .bi_bdev = bio->bi_bdev, | 164 | .bi_bdev = bio->bi_bdev, |
176 | .bi_sector = bio->bi_sector, | 165 | .bi_sector = bio->bi_sector, |
@@ -178,10 +167,14 @@ static unsigned bch_bio_max_sectors(struct bio *bio) | |||
178 | .bi_rw = bio->bi_rw, | 167 | .bi_rw = bio->bi_rw, |
179 | }; | 168 | }; |
180 | 169 | ||
170 | if (seg == max_segments) | ||
171 | break; | ||
172 | |||
181 | if (q->merge_bvec_fn && | 173 | if (q->merge_bvec_fn && |
182 | q->merge_bvec_fn(q, &bvm, bv) < (int) bv->bv_len) | 174 | q->merge_bvec_fn(q, &bvm, bv) < (int) bv->bv_len) |
183 | break; | 175 | break; |
184 | 176 | ||
177 | seg++; | ||
185 | ret += bv->bv_len >> 9; | 178 | ret += bv->bv_len >> 9; |
186 | } | 179 | } |
187 | } | 180 | } |
@@ -218,30 +211,10 @@ static void bch_bio_submit_split_endio(struct bio *bio, int error) | |||
218 | closure_put(cl); | 211 | closure_put(cl); |
219 | } | 212 | } |
220 | 213 | ||
221 | static void __bch_bio_submit_split(struct closure *cl) | ||
222 | { | ||
223 | struct bio_split_hook *s = container_of(cl, struct bio_split_hook, cl); | ||
224 | struct bio *bio = s->bio, *n; | ||
225 | |||
226 | do { | ||
227 | n = bch_bio_split(bio, bch_bio_max_sectors(bio), | ||
228 | GFP_NOIO, s->p->bio_split); | ||
229 | if (!n) | ||
230 | continue_at(cl, __bch_bio_submit_split, system_wq); | ||
231 | |||
232 | n->bi_end_io = bch_bio_submit_split_endio; | ||
233 | n->bi_private = cl; | ||
234 | |||
235 | closure_get(cl); | ||
236 | bch_generic_make_request_hack(n); | ||
237 | } while (n != bio); | ||
238 | |||
239 | continue_at(cl, bch_bio_submit_split_done, NULL); | ||
240 | } | ||
241 | |||
242 | void bch_generic_make_request(struct bio *bio, struct bio_split_pool *p) | 214 | void bch_generic_make_request(struct bio *bio, struct bio_split_pool *p) |
243 | { | 215 | { |
244 | struct bio_split_hook *s; | 216 | struct bio_split_hook *s; |
217 | struct bio *n; | ||
245 | 218 | ||
246 | if (!bio_has_data(bio) && !(bio->bi_rw & REQ_DISCARD)) | 219 | if (!bio_has_data(bio) && !(bio->bi_rw & REQ_DISCARD)) |
247 | goto submit; | 220 | goto submit; |
@@ -250,6 +223,7 @@ void bch_generic_make_request(struct bio *bio, struct bio_split_pool *p) | |||
250 | goto submit; | 223 | goto submit; |
251 | 224 | ||
252 | s = mempool_alloc(p->bio_split_hook, GFP_NOIO); | 225 | s = mempool_alloc(p->bio_split_hook, GFP_NOIO); |
226 | closure_init(&s->cl, NULL); | ||
253 | 227 | ||
254 | s->bio = bio; | 228 | s->bio = bio; |
255 | s->p = p; | 229 | s->p = p; |
@@ -257,8 +231,18 @@ void bch_generic_make_request(struct bio *bio, struct bio_split_pool *p) | |||
257 | s->bi_private = bio->bi_private; | 231 | s->bi_private = bio->bi_private; |
258 | bio_get(bio); | 232 | bio_get(bio); |
259 | 233 | ||
260 | closure_call(&s->cl, __bch_bio_submit_split, NULL, NULL); | 234 | do { |
261 | return; | 235 | n = bch_bio_split(bio, bch_bio_max_sectors(bio), |
236 | GFP_NOIO, s->p->bio_split); | ||
237 | |||
238 | n->bi_end_io = bch_bio_submit_split_endio; | ||
239 | n->bi_private = &s->cl; | ||
240 | |||
241 | closure_get(&s->cl); | ||
242 | bch_generic_make_request_hack(n); | ||
243 | } while (n != bio); | ||
244 | |||
245 | continue_at(&s->cl, bch_bio_submit_split_done, NULL); | ||
262 | submit: | 246 | submit: |
263 | bch_generic_make_request_hack(bio); | 247 | bch_generic_make_request_hack(bio); |
264 | } | 248 | } |
diff --git a/drivers/md/bcache/journal.c b/drivers/md/bcache/journal.c index 8c8dfdcd9d4c..ba95ab84b2be 100644 --- a/drivers/md/bcache/journal.c +++ b/drivers/md/bcache/journal.c | |||
@@ -9,6 +9,8 @@ | |||
9 | #include "debug.h" | 9 | #include "debug.h" |
10 | #include "request.h" | 10 | #include "request.h" |
11 | 11 | ||
12 | #include <trace/events/bcache.h> | ||
13 | |||
12 | /* | 14 | /* |
13 | * Journal replay/recovery: | 15 | * Journal replay/recovery: |
14 | * | 16 | * |
@@ -182,9 +184,14 @@ bsearch: | |||
182 | pr_debug("starting binary search, l %u r %u", l, r); | 184 | pr_debug("starting binary search, l %u r %u", l, r); |
183 | 185 | ||
184 | while (l + 1 < r) { | 186 | while (l + 1 < r) { |
187 | seq = list_entry(list->prev, struct journal_replay, | ||
188 | list)->j.seq; | ||
189 | |||
185 | m = (l + r) >> 1; | 190 | m = (l + r) >> 1; |
191 | read_bucket(m); | ||
186 | 192 | ||
187 | if (read_bucket(m)) | 193 | if (seq != list_entry(list->prev, struct journal_replay, |
194 | list)->j.seq) | ||
188 | l = m; | 195 | l = m; |
189 | else | 196 | else |
190 | r = m; | 197 | r = m; |
@@ -300,7 +307,8 @@ int bch_journal_replay(struct cache_set *s, struct list_head *list, | |||
300 | for (k = i->j.start; | 307 | for (k = i->j.start; |
301 | k < end(&i->j); | 308 | k < end(&i->j); |
302 | k = bkey_next(k)) { | 309 | k = bkey_next(k)) { |
303 | pr_debug("%s", pkey(k)); | 310 | trace_bcache_journal_replay_key(k); |
311 | |||
304 | bkey_copy(op->keys.top, k); | 312 | bkey_copy(op->keys.top, k); |
305 | bch_keylist_push(&op->keys); | 313 | bch_keylist_push(&op->keys); |
306 | 314 | ||
@@ -384,7 +392,7 @@ out: | |||
384 | return; | 392 | return; |
385 | found: | 393 | found: |
386 | if (btree_node_dirty(best)) | 394 | if (btree_node_dirty(best)) |
387 | bch_btree_write(best, true, NULL); | 395 | bch_btree_node_write(best, NULL); |
388 | rw_unlock(true, best); | 396 | rw_unlock(true, best); |
389 | } | 397 | } |
390 | 398 | ||
@@ -617,7 +625,7 @@ static void journal_write_unlocked(struct closure *cl) | |||
617 | bio_reset(bio); | 625 | bio_reset(bio); |
618 | bio->bi_sector = PTR_OFFSET(k, i); | 626 | bio->bi_sector = PTR_OFFSET(k, i); |
619 | bio->bi_bdev = ca->bdev; | 627 | bio->bi_bdev = ca->bdev; |
620 | bio->bi_rw = REQ_WRITE|REQ_SYNC|REQ_META|REQ_FLUSH; | 628 | bio->bi_rw = REQ_WRITE|REQ_SYNC|REQ_META|REQ_FLUSH|REQ_FUA; |
621 | bio->bi_size = sectors << 9; | 629 | bio->bi_size = sectors << 9; |
622 | 630 | ||
623 | bio->bi_end_io = journal_write_endio; | 631 | bio->bi_end_io = journal_write_endio; |
@@ -712,7 +720,8 @@ void bch_journal(struct closure *cl) | |||
712 | spin_lock(&c->journal.lock); | 720 | spin_lock(&c->journal.lock); |
713 | 721 | ||
714 | if (journal_full(&c->journal)) { | 722 | if (journal_full(&c->journal)) { |
715 | /* XXX: tracepoint */ | 723 | trace_bcache_journal_full(c); |
724 | |||
716 | closure_wait(&c->journal.wait, cl); | 725 | closure_wait(&c->journal.wait, cl); |
717 | 726 | ||
718 | journal_reclaim(c); | 727 | journal_reclaim(c); |
@@ -728,13 +737,15 @@ void bch_journal(struct closure *cl) | |||
728 | 737 | ||
729 | if (b * c->sb.block_size > PAGE_SECTORS << JSET_BITS || | 738 | if (b * c->sb.block_size > PAGE_SECTORS << JSET_BITS || |
730 | b > c->journal.blocks_free) { | 739 | b > c->journal.blocks_free) { |
731 | /* XXX: If we were inserting so many keys that they won't fit in | 740 | trace_bcache_journal_entry_full(c); |
741 | |||
742 | /* | ||
743 | * XXX: If we were inserting so many keys that they won't fit in | ||
732 | * an _empty_ journal write, we'll deadlock. For now, handle | 744 | * an _empty_ journal write, we'll deadlock. For now, handle |
733 | * this in bch_keylist_realloc() - but something to think about. | 745 | * this in bch_keylist_realloc() - but something to think about. |
734 | */ | 746 | */ |
735 | BUG_ON(!w->data->keys); | 747 | BUG_ON(!w->data->keys); |
736 | 748 | ||
737 | /* XXX: tracepoint */ | ||
738 | BUG_ON(!closure_wait(&w->wait, cl)); | 749 | BUG_ON(!closure_wait(&w->wait, cl)); |
739 | 750 | ||
740 | closure_flush(&c->journal.io); | 751 | closure_flush(&c->journal.io); |
diff --git a/drivers/md/bcache/movinggc.c b/drivers/md/bcache/movinggc.c index 8589512c972e..1a3b4f4786c3 100644 --- a/drivers/md/bcache/movinggc.c +++ b/drivers/md/bcache/movinggc.c | |||
@@ -9,6 +9,8 @@ | |||
9 | #include "debug.h" | 9 | #include "debug.h" |
10 | #include "request.h" | 10 | #include "request.h" |
11 | 11 | ||
12 | #include <trace/events/bcache.h> | ||
13 | |||
12 | struct moving_io { | 14 | struct moving_io { |
13 | struct keybuf_key *w; | 15 | struct keybuf_key *w; |
14 | struct search s; | 16 | struct search s; |
@@ -44,14 +46,14 @@ static void write_moving_finish(struct closure *cl) | |||
44 | { | 46 | { |
45 | struct moving_io *io = container_of(cl, struct moving_io, s.cl); | 47 | struct moving_io *io = container_of(cl, struct moving_io, s.cl); |
46 | struct bio *bio = &io->bio.bio; | 48 | struct bio *bio = &io->bio.bio; |
47 | struct bio_vec *bv = bio_iovec_idx(bio, bio->bi_vcnt); | 49 | struct bio_vec *bv; |
50 | int i; | ||
48 | 51 | ||
49 | while (bv-- != bio->bi_io_vec) | 52 | bio_for_each_segment_all(bv, bio, i) |
50 | __free_page(bv->bv_page); | 53 | __free_page(bv->bv_page); |
51 | 54 | ||
52 | pr_debug("%s %s", io->s.op.insert_collision | 55 | if (io->s.op.insert_collision) |
53 | ? "collision moving" : "moved", | 56 | trace_bcache_gc_copy_collision(&io->w->key); |
54 | pkey(&io->w->key)); | ||
55 | 57 | ||
56 | bch_keybuf_del(&io->s.op.c->moving_gc_keys, io->w); | 58 | bch_keybuf_del(&io->s.op.c->moving_gc_keys, io->w); |
57 | 59 | ||
@@ -94,8 +96,6 @@ static void write_moving(struct closure *cl) | |||
94 | struct moving_io *io = container_of(s, struct moving_io, s); | 96 | struct moving_io *io = container_of(s, struct moving_io, s); |
95 | 97 | ||
96 | if (!s->error) { | 98 | if (!s->error) { |
97 | trace_bcache_write_moving(&io->bio.bio); | ||
98 | |||
99 | moving_init(io); | 99 | moving_init(io); |
100 | 100 | ||
101 | io->bio.bio.bi_sector = KEY_START(&io->w->key); | 101 | io->bio.bio.bi_sector = KEY_START(&io->w->key); |
@@ -122,7 +122,6 @@ static void read_moving_submit(struct closure *cl) | |||
122 | struct moving_io *io = container_of(s, struct moving_io, s); | 122 | struct moving_io *io = container_of(s, struct moving_io, s); |
123 | struct bio *bio = &io->bio.bio; | 123 | struct bio *bio = &io->bio.bio; |
124 | 124 | ||
125 | trace_bcache_read_moving(bio); | ||
126 | bch_submit_bbio(bio, s->op.c, &io->w->key, 0); | 125 | bch_submit_bbio(bio, s->op.c, &io->w->key, 0); |
127 | 126 | ||
128 | continue_at(cl, write_moving, bch_gc_wq); | 127 | continue_at(cl, write_moving, bch_gc_wq); |
@@ -138,7 +137,8 @@ static void read_moving(struct closure *cl) | |||
138 | /* XXX: if we error, background writeback could stall indefinitely */ | 137 | /* XXX: if we error, background writeback could stall indefinitely */ |
139 | 138 | ||
140 | while (!test_bit(CACHE_SET_STOPPING, &c->flags)) { | 139 | while (!test_bit(CACHE_SET_STOPPING, &c->flags)) { |
141 | w = bch_keybuf_next_rescan(c, &c->moving_gc_keys, &MAX_KEY); | 140 | w = bch_keybuf_next_rescan(c, &c->moving_gc_keys, |
141 | &MAX_KEY, moving_pred); | ||
142 | if (!w) | 142 | if (!w) |
143 | break; | 143 | break; |
144 | 144 | ||
@@ -159,10 +159,10 @@ static void read_moving(struct closure *cl) | |||
159 | bio->bi_rw = READ; | 159 | bio->bi_rw = READ; |
160 | bio->bi_end_io = read_moving_endio; | 160 | bio->bi_end_io = read_moving_endio; |
161 | 161 | ||
162 | if (bch_bio_alloc_pages(bio, GFP_KERNEL)) | 162 | if (bio_alloc_pages(bio, GFP_KERNEL)) |
163 | goto err; | 163 | goto err; |
164 | 164 | ||
165 | pr_debug("%s", pkey(&w->key)); | 165 | trace_bcache_gc_copy(&w->key); |
166 | 166 | ||
167 | closure_call(&io->s.cl, read_moving_submit, NULL, &c->gc.cl); | 167 | closure_call(&io->s.cl, read_moving_submit, NULL, &c->gc.cl); |
168 | 168 | ||
@@ -250,5 +250,5 @@ void bch_moving_gc(struct closure *cl) | |||
250 | 250 | ||
251 | void bch_moving_init_cache_set(struct cache_set *c) | 251 | void bch_moving_init_cache_set(struct cache_set *c) |
252 | { | 252 | { |
253 | bch_keybuf_init(&c->moving_gc_keys, moving_pred); | 253 | bch_keybuf_init(&c->moving_gc_keys); |
254 | } | 254 | } |
diff --git a/drivers/md/bcache/request.c b/drivers/md/bcache/request.c index e5ff12e52d5b..786a1a4f74d8 100644 --- a/drivers/md/bcache/request.c +++ b/drivers/md/bcache/request.c | |||
@@ -10,6 +10,7 @@ | |||
10 | #include "btree.h" | 10 | #include "btree.h" |
11 | #include "debug.h" | 11 | #include "debug.h" |
12 | #include "request.h" | 12 | #include "request.h" |
13 | #include "writeback.h" | ||
13 | 14 | ||
14 | #include <linux/cgroup.h> | 15 | #include <linux/cgroup.h> |
15 | #include <linux/module.h> | 16 | #include <linux/module.h> |
@@ -21,8 +22,6 @@ | |||
21 | 22 | ||
22 | #define CUTOFF_CACHE_ADD 95 | 23 | #define CUTOFF_CACHE_ADD 95 |
23 | #define CUTOFF_CACHE_READA 90 | 24 | #define CUTOFF_CACHE_READA 90 |
24 | #define CUTOFF_WRITEBACK 50 | ||
25 | #define CUTOFF_WRITEBACK_SYNC 75 | ||
26 | 25 | ||
27 | struct kmem_cache *bch_search_cache; | 26 | struct kmem_cache *bch_search_cache; |
28 | 27 | ||
@@ -489,6 +488,12 @@ static void bch_insert_data_loop(struct closure *cl) | |||
489 | bch_queue_gc(op->c); | 488 | bch_queue_gc(op->c); |
490 | } | 489 | } |
491 | 490 | ||
491 | /* | ||
492 | * Journal writes are marked REQ_FLUSH; if the original write was a | ||
493 | * flush, it'll wait on the journal write. | ||
494 | */ | ||
495 | bio->bi_rw &= ~(REQ_FLUSH|REQ_FUA); | ||
496 | |||
492 | do { | 497 | do { |
493 | unsigned i; | 498 | unsigned i; |
494 | struct bkey *k; | 499 | struct bkey *k; |
@@ -510,10 +515,6 @@ static void bch_insert_data_loop(struct closure *cl) | |||
510 | goto err; | 515 | goto err; |
511 | 516 | ||
512 | n = bch_bio_split(bio, KEY_SIZE(k), GFP_NOIO, split); | 517 | n = bch_bio_split(bio, KEY_SIZE(k), GFP_NOIO, split); |
513 | if (!n) { | ||
514 | __bkey_put(op->c, k); | ||
515 | continue_at(cl, bch_insert_data_loop, bcache_wq); | ||
516 | } | ||
517 | 518 | ||
518 | n->bi_end_io = bch_insert_data_endio; | 519 | n->bi_end_io = bch_insert_data_endio; |
519 | n->bi_private = cl; | 520 | n->bi_private = cl; |
@@ -530,10 +531,9 @@ static void bch_insert_data_loop(struct closure *cl) | |||
530 | if (KEY_CSUM(k)) | 531 | if (KEY_CSUM(k)) |
531 | bio_csum(n, k); | 532 | bio_csum(n, k); |
532 | 533 | ||
533 | pr_debug("%s", pkey(k)); | 534 | trace_bcache_cache_insert(k); |
534 | bch_keylist_push(&op->keys); | 535 | bch_keylist_push(&op->keys); |
535 | 536 | ||
536 | trace_bcache_cache_insert(n, n->bi_sector, n->bi_bdev); | ||
537 | n->bi_rw |= REQ_WRITE; | 537 | n->bi_rw |= REQ_WRITE; |
538 | bch_submit_bbio(n, op->c, k, 0); | 538 | bch_submit_bbio(n, op->c, k, 0); |
539 | } while (n != bio); | 539 | } while (n != bio); |
@@ -716,7 +716,7 @@ static struct search *search_alloc(struct bio *bio, struct bcache_device *d) | |||
716 | s->task = current; | 716 | s->task = current; |
717 | s->orig_bio = bio; | 717 | s->orig_bio = bio; |
718 | s->write = (bio->bi_rw & REQ_WRITE) != 0; | 718 | s->write = (bio->bi_rw & REQ_WRITE) != 0; |
719 | s->op.flush_journal = (bio->bi_rw & REQ_FLUSH) != 0; | 719 | s->op.flush_journal = (bio->bi_rw & (REQ_FLUSH|REQ_FUA)) != 0; |
720 | s->op.skip = (bio->bi_rw & REQ_DISCARD) != 0; | 720 | s->op.skip = (bio->bi_rw & REQ_DISCARD) != 0; |
721 | s->recoverable = 1; | 721 | s->recoverable = 1; |
722 | s->start_time = jiffies; | 722 | s->start_time = jiffies; |
@@ -784,11 +784,8 @@ static void request_read_error(struct closure *cl) | |||
784 | int i; | 784 | int i; |
785 | 785 | ||
786 | if (s->recoverable) { | 786 | if (s->recoverable) { |
787 | /* The cache read failed, but we can retry from the backing | 787 | /* Retry from the backing device: */ |
788 | * device. | 788 | trace_bcache_read_retry(s->orig_bio); |
789 | */ | ||
790 | pr_debug("recovering at sector %llu", | ||
791 | (uint64_t) s->orig_bio->bi_sector); | ||
792 | 789 | ||
793 | s->error = 0; | 790 | s->error = 0; |
794 | bv = s->bio.bio.bi_io_vec; | 791 | bv = s->bio.bio.bi_io_vec; |
@@ -806,7 +803,6 @@ static void request_read_error(struct closure *cl) | |||
806 | 803 | ||
807 | /* XXX: invalidate cache */ | 804 | /* XXX: invalidate cache */ |
808 | 805 | ||
809 | trace_bcache_read_retry(&s->bio.bio); | ||
810 | closure_bio_submit(&s->bio.bio, &s->cl, s->d); | 806 | closure_bio_submit(&s->bio.bio, &s->cl, s->d); |
811 | } | 807 | } |
812 | 808 | ||
@@ -827,53 +823,13 @@ static void request_read_done(struct closure *cl) | |||
827 | */ | 823 | */ |
828 | 824 | ||
829 | if (s->op.cache_bio) { | 825 | if (s->op.cache_bio) { |
830 | struct bio_vec *src, *dst; | ||
831 | unsigned src_offset, dst_offset, bytes; | ||
832 | void *dst_ptr; | ||
833 | |||
834 | bio_reset(s->op.cache_bio); | 826 | bio_reset(s->op.cache_bio); |
835 | s->op.cache_bio->bi_sector = s->cache_miss->bi_sector; | 827 | s->op.cache_bio->bi_sector = s->cache_miss->bi_sector; |
836 | s->op.cache_bio->bi_bdev = s->cache_miss->bi_bdev; | 828 | s->op.cache_bio->bi_bdev = s->cache_miss->bi_bdev; |
837 | s->op.cache_bio->bi_size = s->cache_bio_sectors << 9; | 829 | s->op.cache_bio->bi_size = s->cache_bio_sectors << 9; |
838 | bch_bio_map(s->op.cache_bio, NULL); | 830 | bch_bio_map(s->op.cache_bio, NULL); |
839 | 831 | ||
840 | src = bio_iovec(s->op.cache_bio); | 832 | bio_copy_data(s->cache_miss, s->op.cache_bio); |
841 | dst = bio_iovec(s->cache_miss); | ||
842 | src_offset = src->bv_offset; | ||
843 | dst_offset = dst->bv_offset; | ||
844 | dst_ptr = kmap(dst->bv_page); | ||
845 | |||
846 | while (1) { | ||
847 | if (dst_offset == dst->bv_offset + dst->bv_len) { | ||
848 | kunmap(dst->bv_page); | ||
849 | dst++; | ||
850 | if (dst == bio_iovec_idx(s->cache_miss, | ||
851 | s->cache_miss->bi_vcnt)) | ||
852 | break; | ||
853 | |||
854 | dst_offset = dst->bv_offset; | ||
855 | dst_ptr = kmap(dst->bv_page); | ||
856 | } | ||
857 | |||
858 | if (src_offset == src->bv_offset + src->bv_len) { | ||
859 | src++; | ||
860 | if (src == bio_iovec_idx(s->op.cache_bio, | ||
861 | s->op.cache_bio->bi_vcnt)) | ||
862 | BUG(); | ||
863 | |||
864 | src_offset = src->bv_offset; | ||
865 | } | ||
866 | |||
867 | bytes = min(dst->bv_offset + dst->bv_len - dst_offset, | ||
868 | src->bv_offset + src->bv_len - src_offset); | ||
869 | |||
870 | memcpy(dst_ptr + dst_offset, | ||
871 | page_address(src->bv_page) + src_offset, | ||
872 | bytes); | ||
873 | |||
874 | src_offset += bytes; | ||
875 | dst_offset += bytes; | ||
876 | } | ||
877 | 833 | ||
878 | bio_put(s->cache_miss); | 834 | bio_put(s->cache_miss); |
879 | s->cache_miss = NULL; | 835 | s->cache_miss = NULL; |
@@ -899,6 +855,7 @@ static void request_read_done_bh(struct closure *cl) | |||
899 | struct cached_dev *dc = container_of(s->d, struct cached_dev, disk); | 855 | struct cached_dev *dc = container_of(s->d, struct cached_dev, disk); |
900 | 856 | ||
901 | bch_mark_cache_accounting(s, !s->cache_miss, s->op.skip); | 857 | bch_mark_cache_accounting(s, !s->cache_miss, s->op.skip); |
858 | trace_bcache_read(s->orig_bio, !s->cache_miss, s->op.skip); | ||
902 | 859 | ||
903 | if (s->error) | 860 | if (s->error) |
904 | continue_at_nobarrier(cl, request_read_error, bcache_wq); | 861 | continue_at_nobarrier(cl, request_read_error, bcache_wq); |
@@ -917,9 +874,6 @@ static int cached_dev_cache_miss(struct btree *b, struct search *s, | |||
917 | struct bio *miss; | 874 | struct bio *miss; |
918 | 875 | ||
919 | miss = bch_bio_split(bio, sectors, GFP_NOIO, s->d->bio_split); | 876 | miss = bch_bio_split(bio, sectors, GFP_NOIO, s->d->bio_split); |
920 | if (!miss) | ||
921 | return -EAGAIN; | ||
922 | |||
923 | if (miss == bio) | 877 | if (miss == bio) |
924 | s->op.lookup_done = true; | 878 | s->op.lookup_done = true; |
925 | 879 | ||
@@ -938,8 +892,9 @@ static int cached_dev_cache_miss(struct btree *b, struct search *s, | |||
938 | reada = min(dc->readahead >> 9, | 892 | reada = min(dc->readahead >> 9, |
939 | sectors - bio_sectors(miss)); | 893 | sectors - bio_sectors(miss)); |
940 | 894 | ||
941 | if (bio_end(miss) + reada > bdev_sectors(miss->bi_bdev)) | 895 | if (bio_end_sector(miss) + reada > bdev_sectors(miss->bi_bdev)) |
942 | reada = bdev_sectors(miss->bi_bdev) - bio_end(miss); | 896 | reada = bdev_sectors(miss->bi_bdev) - |
897 | bio_end_sector(miss); | ||
943 | } | 898 | } |
944 | 899 | ||
945 | s->cache_bio_sectors = bio_sectors(miss) + reada; | 900 | s->cache_bio_sectors = bio_sectors(miss) + reada; |
@@ -963,13 +918,12 @@ static int cached_dev_cache_miss(struct btree *b, struct search *s, | |||
963 | goto out_put; | 918 | goto out_put; |
964 | 919 | ||
965 | bch_bio_map(s->op.cache_bio, NULL); | 920 | bch_bio_map(s->op.cache_bio, NULL); |
966 | if (bch_bio_alloc_pages(s->op.cache_bio, __GFP_NOWARN|GFP_NOIO)) | 921 | if (bio_alloc_pages(s->op.cache_bio, __GFP_NOWARN|GFP_NOIO)) |
967 | goto out_put; | 922 | goto out_put; |
968 | 923 | ||
969 | s->cache_miss = miss; | 924 | s->cache_miss = miss; |
970 | bio_get(s->op.cache_bio); | 925 | bio_get(s->op.cache_bio); |
971 | 926 | ||
972 | trace_bcache_cache_miss(s->orig_bio); | ||
973 | closure_bio_submit(s->op.cache_bio, &s->cl, s->d); | 927 | closure_bio_submit(s->op.cache_bio, &s->cl, s->d); |
974 | 928 | ||
975 | return ret; | 929 | return ret; |
@@ -1002,24 +956,13 @@ static void cached_dev_write_complete(struct closure *cl) | |||
1002 | cached_dev_bio_complete(cl); | 956 | cached_dev_bio_complete(cl); |
1003 | } | 957 | } |
1004 | 958 | ||
1005 | static bool should_writeback(struct cached_dev *dc, struct bio *bio) | ||
1006 | { | ||
1007 | unsigned threshold = (bio->bi_rw & REQ_SYNC) | ||
1008 | ? CUTOFF_WRITEBACK_SYNC | ||
1009 | : CUTOFF_WRITEBACK; | ||
1010 | |||
1011 | return !atomic_read(&dc->disk.detaching) && | ||
1012 | cache_mode(dc, bio) == CACHE_MODE_WRITEBACK && | ||
1013 | dc->disk.c->gc_stats.in_use < threshold; | ||
1014 | } | ||
1015 | |||
1016 | static void request_write(struct cached_dev *dc, struct search *s) | 959 | static void request_write(struct cached_dev *dc, struct search *s) |
1017 | { | 960 | { |
1018 | struct closure *cl = &s->cl; | 961 | struct closure *cl = &s->cl; |
1019 | struct bio *bio = &s->bio.bio; | 962 | struct bio *bio = &s->bio.bio; |
1020 | struct bkey start, end; | 963 | struct bkey start, end; |
1021 | start = KEY(dc->disk.id, bio->bi_sector, 0); | 964 | start = KEY(dc->disk.id, bio->bi_sector, 0); |
1022 | end = KEY(dc->disk.id, bio_end(bio), 0); | 965 | end = KEY(dc->disk.id, bio_end_sector(bio), 0); |
1023 | 966 | ||
1024 | bch_keybuf_check_overlapping(&s->op.c->moving_gc_keys, &start, &end); | 967 | bch_keybuf_check_overlapping(&s->op.c->moving_gc_keys, &start, &end); |
1025 | 968 | ||
@@ -1034,22 +977,37 @@ static void request_write(struct cached_dev *dc, struct search *s) | |||
1034 | if (bio->bi_rw & REQ_DISCARD) | 977 | if (bio->bi_rw & REQ_DISCARD) |
1035 | goto skip; | 978 | goto skip; |
1036 | 979 | ||
980 | if (should_writeback(dc, s->orig_bio, | ||
981 | cache_mode(dc, bio), | ||
982 | s->op.skip)) { | ||
983 | s->op.skip = false; | ||
984 | s->writeback = true; | ||
985 | } | ||
986 | |||
1037 | if (s->op.skip) | 987 | if (s->op.skip) |
1038 | goto skip; | 988 | goto skip; |
1039 | 989 | ||
1040 | if (should_writeback(dc, s->orig_bio)) | 990 | trace_bcache_write(s->orig_bio, s->writeback, s->op.skip); |
1041 | s->writeback = true; | ||
1042 | 991 | ||
1043 | if (!s->writeback) { | 992 | if (!s->writeback) { |
1044 | s->op.cache_bio = bio_clone_bioset(bio, GFP_NOIO, | 993 | s->op.cache_bio = bio_clone_bioset(bio, GFP_NOIO, |
1045 | dc->disk.bio_split); | 994 | dc->disk.bio_split); |
1046 | 995 | ||
1047 | trace_bcache_writethrough(s->orig_bio); | ||
1048 | closure_bio_submit(bio, cl, s->d); | 996 | closure_bio_submit(bio, cl, s->d); |
1049 | } else { | 997 | } else { |
1050 | s->op.cache_bio = bio; | 998 | bch_writeback_add(dc); |
1051 | trace_bcache_writeback(s->orig_bio); | 999 | |
1052 | bch_writeback_add(dc, bio_sectors(bio)); | 1000 | if (s->op.flush_journal) { |
1001 | /* Also need to send a flush to the backing device */ | ||
1002 | s->op.cache_bio = bio_clone_bioset(bio, GFP_NOIO, | ||
1003 | dc->disk.bio_split); | ||
1004 | |||
1005 | bio->bi_size = 0; | ||
1006 | bio->bi_vcnt = 0; | ||
1007 | closure_bio_submit(bio, cl, s->d); | ||
1008 | } else { | ||
1009 | s->op.cache_bio = bio; | ||
1010 | } | ||
1053 | } | 1011 | } |
1054 | out: | 1012 | out: |
1055 | closure_call(&s->op.cl, bch_insert_data, NULL, cl); | 1013 | closure_call(&s->op.cl, bch_insert_data, NULL, cl); |
@@ -1058,7 +1016,6 @@ skip: | |||
1058 | s->op.skip = true; | 1016 | s->op.skip = true; |
1059 | s->op.cache_bio = s->orig_bio; | 1017 | s->op.cache_bio = s->orig_bio; |
1060 | bio_get(s->op.cache_bio); | 1018 | bio_get(s->op.cache_bio); |
1061 | trace_bcache_write_skip(s->orig_bio); | ||
1062 | 1019 | ||
1063 | if ((bio->bi_rw & REQ_DISCARD) && | 1020 | if ((bio->bi_rw & REQ_DISCARD) && |
1064 | !blk_queue_discard(bdev_get_queue(dc->bdev))) | 1021 | !blk_queue_discard(bdev_get_queue(dc->bdev))) |
@@ -1088,9 +1045,10 @@ static void request_nodata(struct cached_dev *dc, struct search *s) | |||
1088 | 1045 | ||
1089 | /* Cached devices - read & write stuff */ | 1046 | /* Cached devices - read & write stuff */ |
1090 | 1047 | ||
1091 | int bch_get_congested(struct cache_set *c) | 1048 | unsigned bch_get_congested(struct cache_set *c) |
1092 | { | 1049 | { |
1093 | int i; | 1050 | int i; |
1051 | long rand; | ||
1094 | 1052 | ||
1095 | if (!c->congested_read_threshold_us && | 1053 | if (!c->congested_read_threshold_us && |
1096 | !c->congested_write_threshold_us) | 1054 | !c->congested_write_threshold_us) |
@@ -1106,7 +1064,13 @@ int bch_get_congested(struct cache_set *c) | |||
1106 | 1064 | ||
1107 | i += CONGESTED_MAX; | 1065 | i += CONGESTED_MAX; |
1108 | 1066 | ||
1109 | return i <= 0 ? 1 : fract_exp_two(i, 6); | 1067 | if (i > 0) |
1068 | i = fract_exp_two(i, 6); | ||
1069 | |||
1070 | rand = get_random_int(); | ||
1071 | i -= bitmap_weight(&rand, BITS_PER_LONG); | ||
1072 | |||
1073 | return i > 0 ? i : 1; | ||
1110 | } | 1074 | } |
1111 | 1075 | ||
1112 | static void add_sequential(struct task_struct *t) | 1076 | static void add_sequential(struct task_struct *t) |
@@ -1126,10 +1090,8 @@ static void check_should_skip(struct cached_dev *dc, struct search *s) | |||
1126 | { | 1090 | { |
1127 | struct cache_set *c = s->op.c; | 1091 | struct cache_set *c = s->op.c; |
1128 | struct bio *bio = &s->bio.bio; | 1092 | struct bio *bio = &s->bio.bio; |
1129 | |||
1130 | long rand; | ||
1131 | int cutoff = bch_get_congested(c); | ||
1132 | unsigned mode = cache_mode(dc, bio); | 1093 | unsigned mode = cache_mode(dc, bio); |
1094 | unsigned sectors, congested = bch_get_congested(c); | ||
1133 | 1095 | ||
1134 | if (atomic_read(&dc->disk.detaching) || | 1096 | if (atomic_read(&dc->disk.detaching) || |
1135 | c->gc_stats.in_use > CUTOFF_CACHE_ADD || | 1097 | c->gc_stats.in_use > CUTOFF_CACHE_ADD || |
@@ -1147,17 +1109,14 @@ static void check_should_skip(struct cached_dev *dc, struct search *s) | |||
1147 | goto skip; | 1109 | goto skip; |
1148 | } | 1110 | } |
1149 | 1111 | ||
1150 | if (!cutoff) { | 1112 | if (!congested && !dc->sequential_cutoff) |
1151 | cutoff = dc->sequential_cutoff >> 9; | 1113 | goto rescale; |
1152 | 1114 | ||
1153 | if (!cutoff) | 1115 | if (!congested && |
1154 | goto rescale; | 1116 | mode == CACHE_MODE_WRITEBACK && |
1155 | 1117 | (bio->bi_rw & REQ_WRITE) && | |
1156 | if (mode == CACHE_MODE_WRITEBACK && | 1118 | (bio->bi_rw & REQ_SYNC)) |
1157 | (bio->bi_rw & REQ_WRITE) && | 1119 | goto rescale; |
1158 | (bio->bi_rw & REQ_SYNC)) | ||
1159 | goto rescale; | ||
1160 | } | ||
1161 | 1120 | ||
1162 | if (dc->sequential_merge) { | 1121 | if (dc->sequential_merge) { |
1163 | struct io *i; | 1122 | struct io *i; |
@@ -1177,7 +1136,7 @@ found: | |||
1177 | if (i->sequential + bio->bi_size > i->sequential) | 1136 | if (i->sequential + bio->bi_size > i->sequential) |
1178 | i->sequential += bio->bi_size; | 1137 | i->sequential += bio->bi_size; |
1179 | 1138 | ||
1180 | i->last = bio_end(bio); | 1139 | i->last = bio_end_sector(bio); |
1181 | i->jiffies = jiffies + msecs_to_jiffies(5000); | 1140 | i->jiffies = jiffies + msecs_to_jiffies(5000); |
1182 | s->task->sequential_io = i->sequential; | 1141 | s->task->sequential_io = i->sequential; |
1183 | 1142 | ||
@@ -1192,12 +1151,19 @@ found: | |||
1192 | add_sequential(s->task); | 1151 | add_sequential(s->task); |
1193 | } | 1152 | } |
1194 | 1153 | ||
1195 | rand = get_random_int(); | 1154 | sectors = max(s->task->sequential_io, |
1196 | cutoff -= bitmap_weight(&rand, BITS_PER_LONG); | 1155 | s->task->sequential_io_avg) >> 9; |
1197 | 1156 | ||
1198 | if (cutoff <= (int) (max(s->task->sequential_io, | 1157 | if (dc->sequential_cutoff && |
1199 | s->task->sequential_io_avg) >> 9)) | 1158 | sectors >= dc->sequential_cutoff >> 9) { |
1159 | trace_bcache_bypass_sequential(s->orig_bio); | ||
1200 | goto skip; | 1160 | goto skip; |
1161 | } | ||
1162 | |||
1163 | if (congested && sectors >= congested) { | ||
1164 | trace_bcache_bypass_congested(s->orig_bio); | ||
1165 | goto skip; | ||
1166 | } | ||
1201 | 1167 | ||
1202 | rescale: | 1168 | rescale: |
1203 | bch_rescale_priorities(c, bio_sectors(bio)); | 1169 | bch_rescale_priorities(c, bio_sectors(bio)); |
@@ -1288,30 +1254,25 @@ void bch_cached_dev_request_init(struct cached_dev *dc) | |||
1288 | static int flash_dev_cache_miss(struct btree *b, struct search *s, | 1254 | static int flash_dev_cache_miss(struct btree *b, struct search *s, |
1289 | struct bio *bio, unsigned sectors) | 1255 | struct bio *bio, unsigned sectors) |
1290 | { | 1256 | { |
1257 | struct bio_vec *bv; | ||
1258 | int i; | ||
1259 | |||
1291 | /* Zero fill bio */ | 1260 | /* Zero fill bio */ |
1292 | 1261 | ||
1293 | while (bio->bi_idx != bio->bi_vcnt) { | 1262 | bio_for_each_segment(bv, bio, i) { |
1294 | struct bio_vec *bv = bio_iovec(bio); | ||
1295 | unsigned j = min(bv->bv_len >> 9, sectors); | 1263 | unsigned j = min(bv->bv_len >> 9, sectors); |
1296 | 1264 | ||
1297 | void *p = kmap(bv->bv_page); | 1265 | void *p = kmap(bv->bv_page); |
1298 | memset(p + bv->bv_offset, 0, j << 9); | 1266 | memset(p + bv->bv_offset, 0, j << 9); |
1299 | kunmap(bv->bv_page); | 1267 | kunmap(bv->bv_page); |
1300 | 1268 | ||
1301 | bv->bv_len -= j << 9; | 1269 | sectors -= j; |
1302 | bv->bv_offset += j << 9; | ||
1303 | |||
1304 | if (bv->bv_len) | ||
1305 | return 0; | ||
1306 | |||
1307 | bio->bi_sector += j; | ||
1308 | bio->bi_size -= j << 9; | ||
1309 | |||
1310 | bio->bi_idx++; | ||
1311 | sectors -= j; | ||
1312 | } | 1270 | } |
1313 | 1271 | ||
1314 | s->op.lookup_done = true; | 1272 | bio_advance(bio, min(sectors << 9, bio->bi_size)); |
1273 | |||
1274 | if (!bio->bi_size) | ||
1275 | s->op.lookup_done = true; | ||
1315 | 1276 | ||
1316 | return 0; | 1277 | return 0; |
1317 | } | 1278 | } |
@@ -1338,8 +1299,8 @@ static void flash_dev_make_request(struct request_queue *q, struct bio *bio) | |||
1338 | closure_call(&s->op.cl, btree_read_async, NULL, cl); | 1299 | closure_call(&s->op.cl, btree_read_async, NULL, cl); |
1339 | } else if (bio_has_data(bio) || s->op.skip) { | 1300 | } else if (bio_has_data(bio) || s->op.skip) { |
1340 | bch_keybuf_check_overlapping(&s->op.c->moving_gc_keys, | 1301 | bch_keybuf_check_overlapping(&s->op.c->moving_gc_keys, |
1341 | &KEY(d->id, bio->bi_sector, 0), | 1302 | &KEY(d->id, bio->bi_sector, 0), |
1342 | &KEY(d->id, bio_end(bio), 0)); | 1303 | &KEY(d->id, bio_end_sector(bio), 0)); |
1343 | 1304 | ||
1344 | s->writeback = true; | 1305 | s->writeback = true; |
1345 | s->op.cache_bio = bio; | 1306 | s->op.cache_bio = bio; |
diff --git a/drivers/md/bcache/request.h b/drivers/md/bcache/request.h index 254d9ab5707c..57dc4784f4f4 100644 --- a/drivers/md/bcache/request.h +++ b/drivers/md/bcache/request.h | |||
@@ -30,7 +30,7 @@ struct search { | |||
30 | }; | 30 | }; |
31 | 31 | ||
32 | void bch_cache_read_endio(struct bio *, int); | 32 | void bch_cache_read_endio(struct bio *, int); |
33 | int bch_get_congested(struct cache_set *); | 33 | unsigned bch_get_congested(struct cache_set *); |
34 | void bch_insert_data(struct closure *cl); | 34 | void bch_insert_data(struct closure *cl); |
35 | void bch_btree_insert_async(struct closure *); | 35 | void bch_btree_insert_async(struct closure *); |
36 | void bch_cache_read_endio(struct bio *, int); | 36 | void bch_cache_read_endio(struct bio *, int); |
diff --git a/drivers/md/bcache/super.c b/drivers/md/bcache/super.c index f88e2b653a3f..547c4c57b052 100644 --- a/drivers/md/bcache/super.c +++ b/drivers/md/bcache/super.c | |||
@@ -10,10 +10,13 @@ | |||
10 | #include "btree.h" | 10 | #include "btree.h" |
11 | #include "debug.h" | 11 | #include "debug.h" |
12 | #include "request.h" | 12 | #include "request.h" |
13 | #include "writeback.h" | ||
13 | 14 | ||
15 | #include <linux/blkdev.h> | ||
14 | #include <linux/buffer_head.h> | 16 | #include <linux/buffer_head.h> |
15 | #include <linux/debugfs.h> | 17 | #include <linux/debugfs.h> |
16 | #include <linux/genhd.h> | 18 | #include <linux/genhd.h> |
19 | #include <linux/kthread.h> | ||
17 | #include <linux/module.h> | 20 | #include <linux/module.h> |
18 | #include <linux/random.h> | 21 | #include <linux/random.h> |
19 | #include <linux/reboot.h> | 22 | #include <linux/reboot.h> |
@@ -342,6 +345,7 @@ static void uuid_io(struct cache_set *c, unsigned long rw, | |||
342 | struct closure *cl = &c->uuid_write.cl; | 345 | struct closure *cl = &c->uuid_write.cl; |
343 | struct uuid_entry *u; | 346 | struct uuid_entry *u; |
344 | unsigned i; | 347 | unsigned i; |
348 | char buf[80]; | ||
345 | 349 | ||
346 | BUG_ON(!parent); | 350 | BUG_ON(!parent); |
347 | closure_lock(&c->uuid_write, parent); | 351 | closure_lock(&c->uuid_write, parent); |
@@ -362,8 +366,8 @@ static void uuid_io(struct cache_set *c, unsigned long rw, | |||
362 | break; | 366 | break; |
363 | } | 367 | } |
364 | 368 | ||
365 | pr_debug("%s UUIDs at %s", rw & REQ_WRITE ? "wrote" : "read", | 369 | bch_bkey_to_text(buf, sizeof(buf), k); |
366 | pkey(&c->uuid_bucket)); | 370 | pr_debug("%s UUIDs at %s", rw & REQ_WRITE ? "wrote" : "read", buf); |
367 | 371 | ||
368 | for (u = c->uuids; u < c->uuids + c->nr_uuids; u++) | 372 | for (u = c->uuids; u < c->uuids + c->nr_uuids; u++) |
369 | if (!bch_is_zero(u->uuid, 16)) | 373 | if (!bch_is_zero(u->uuid, 16)) |
@@ -543,7 +547,6 @@ void bch_prio_write(struct cache *ca) | |||
543 | 547 | ||
544 | pr_debug("free %zu, free_inc %zu, unused %zu", fifo_used(&ca->free), | 548 | pr_debug("free %zu, free_inc %zu, unused %zu", fifo_used(&ca->free), |
545 | fifo_used(&ca->free_inc), fifo_used(&ca->unused)); | 549 | fifo_used(&ca->free_inc), fifo_used(&ca->unused)); |
546 | blktrace_msg(ca, "Starting priorities: " buckets_free(ca)); | ||
547 | 550 | ||
548 | for (i = prio_buckets(ca) - 1; i >= 0; --i) { | 551 | for (i = prio_buckets(ca) - 1; i >= 0; --i) { |
549 | long bucket; | 552 | long bucket; |
@@ -704,7 +707,8 @@ static void bcache_device_detach(struct bcache_device *d) | |||
704 | atomic_set(&d->detaching, 0); | 707 | atomic_set(&d->detaching, 0); |
705 | } | 708 | } |
706 | 709 | ||
707 | bcache_device_unlink(d); | 710 | if (!d->flush_done) |
711 | bcache_device_unlink(d); | ||
708 | 712 | ||
709 | d->c->devices[d->id] = NULL; | 713 | d->c->devices[d->id] = NULL; |
710 | closure_put(&d->c->caching); | 714 | closure_put(&d->c->caching); |
@@ -743,13 +747,35 @@ static void bcache_device_free(struct bcache_device *d) | |||
743 | mempool_destroy(d->unaligned_bvec); | 747 | mempool_destroy(d->unaligned_bvec); |
744 | if (d->bio_split) | 748 | if (d->bio_split) |
745 | bioset_free(d->bio_split); | 749 | bioset_free(d->bio_split); |
750 | if (is_vmalloc_addr(d->stripe_sectors_dirty)) | ||
751 | vfree(d->stripe_sectors_dirty); | ||
752 | else | ||
753 | kfree(d->stripe_sectors_dirty); | ||
746 | 754 | ||
747 | closure_debug_destroy(&d->cl); | 755 | closure_debug_destroy(&d->cl); |
748 | } | 756 | } |
749 | 757 | ||
750 | static int bcache_device_init(struct bcache_device *d, unsigned block_size) | 758 | static int bcache_device_init(struct bcache_device *d, unsigned block_size, |
759 | sector_t sectors) | ||
751 | { | 760 | { |
752 | struct request_queue *q; | 761 | struct request_queue *q; |
762 | size_t n; | ||
763 | |||
764 | if (!d->stripe_size_bits) | ||
765 | d->stripe_size_bits = 31; | ||
766 | |||
767 | d->nr_stripes = round_up(sectors, 1 << d->stripe_size_bits) >> | ||
768 | d->stripe_size_bits; | ||
769 | |||
770 | if (!d->nr_stripes || d->nr_stripes > SIZE_MAX / sizeof(atomic_t)) | ||
771 | return -ENOMEM; | ||
772 | |||
773 | n = d->nr_stripes * sizeof(atomic_t); | ||
774 | d->stripe_sectors_dirty = n < PAGE_SIZE << 6 | ||
775 | ? kzalloc(n, GFP_KERNEL) | ||
776 | : vzalloc(n); | ||
777 | if (!d->stripe_sectors_dirty) | ||
778 | return -ENOMEM; | ||
753 | 779 | ||
754 | if (!(d->bio_split = bioset_create(4, offsetof(struct bbio, bio))) || | 780 | if (!(d->bio_split = bioset_create(4, offsetof(struct bbio, bio))) || |
755 | !(d->unaligned_bvec = mempool_create_kmalloc_pool(1, | 781 | !(d->unaligned_bvec = mempool_create_kmalloc_pool(1, |
@@ -759,6 +785,7 @@ static int bcache_device_init(struct bcache_device *d, unsigned block_size) | |||
759 | !(q = blk_alloc_queue(GFP_KERNEL))) | 785 | !(q = blk_alloc_queue(GFP_KERNEL))) |
760 | return -ENOMEM; | 786 | return -ENOMEM; |
761 | 787 | ||
788 | set_capacity(d->disk, sectors); | ||
762 | snprintf(d->disk->disk_name, DISK_NAME_LEN, "bcache%i", bcache_minor); | 789 | snprintf(d->disk->disk_name, DISK_NAME_LEN, "bcache%i", bcache_minor); |
763 | 790 | ||
764 | d->disk->major = bcache_major; | 791 | d->disk->major = bcache_major; |
@@ -781,6 +808,8 @@ static int bcache_device_init(struct bcache_device *d, unsigned block_size) | |||
781 | set_bit(QUEUE_FLAG_NONROT, &d->disk->queue->queue_flags); | 808 | set_bit(QUEUE_FLAG_NONROT, &d->disk->queue->queue_flags); |
782 | set_bit(QUEUE_FLAG_DISCARD, &d->disk->queue->queue_flags); | 809 | set_bit(QUEUE_FLAG_DISCARD, &d->disk->queue->queue_flags); |
783 | 810 | ||
811 | blk_queue_flush(q, REQ_FLUSH|REQ_FUA); | ||
812 | |||
784 | return 0; | 813 | return 0; |
785 | } | 814 | } |
786 | 815 | ||
@@ -800,6 +829,17 @@ static void calc_cached_dev_sectors(struct cache_set *c) | |||
800 | void bch_cached_dev_run(struct cached_dev *dc) | 829 | void bch_cached_dev_run(struct cached_dev *dc) |
801 | { | 830 | { |
802 | struct bcache_device *d = &dc->disk; | 831 | struct bcache_device *d = &dc->disk; |
832 | char buf[SB_LABEL_SIZE + 1]; | ||
833 | char *env[] = { | ||
834 | "DRIVER=bcache", | ||
835 | kasprintf(GFP_KERNEL, "CACHED_UUID=%pU", dc->sb.uuid), | ||
836 | NULL, | ||
837 | NULL, | ||
838 | }; | ||
839 | |||
840 | memcpy(buf, dc->sb.label, SB_LABEL_SIZE); | ||
841 | buf[SB_LABEL_SIZE] = '\0'; | ||
842 | env[2] = kasprintf(GFP_KERNEL, "CACHED_LABEL=%s", buf); | ||
803 | 843 | ||
804 | if (atomic_xchg(&dc->running, 1)) | 844 | if (atomic_xchg(&dc->running, 1)) |
805 | return; | 845 | return; |
@@ -816,10 +856,12 @@ void bch_cached_dev_run(struct cached_dev *dc) | |||
816 | 856 | ||
817 | add_disk(d->disk); | 857 | add_disk(d->disk); |
818 | bd_link_disk_holder(dc->bdev, dc->disk.disk); | 858 | bd_link_disk_holder(dc->bdev, dc->disk.disk); |
819 | #if 0 | 859 | /* won't show up in the uevent file, use udevadm monitor -e instead |
820 | char *env[] = { "SYMLINK=label" , NULL }; | 860 | * only class / kset properties are persistent */ |
821 | kobject_uevent_env(&disk_to_dev(d->disk)->kobj, KOBJ_CHANGE, env); | 861 | kobject_uevent_env(&disk_to_dev(d->disk)->kobj, KOBJ_CHANGE, env); |
822 | #endif | 862 | kfree(env[1]); |
863 | kfree(env[2]); | ||
864 | |||
823 | if (sysfs_create_link(&d->kobj, &disk_to_dev(d->disk)->kobj, "dev") || | 865 | if (sysfs_create_link(&d->kobj, &disk_to_dev(d->disk)->kobj, "dev") || |
824 | sysfs_create_link(&disk_to_dev(d->disk)->kobj, &d->kobj, "bcache")) | 866 | sysfs_create_link(&disk_to_dev(d->disk)->kobj, &d->kobj, "bcache")) |
825 | pr_debug("error creating sysfs link"); | 867 | pr_debug("error creating sysfs link"); |
@@ -960,6 +1002,7 @@ int bch_cached_dev_attach(struct cached_dev *dc, struct cache_set *c) | |||
960 | atomic_set(&dc->count, 1); | 1002 | atomic_set(&dc->count, 1); |
961 | 1003 | ||
962 | if (BDEV_STATE(&dc->sb) == BDEV_STATE_DIRTY) { | 1004 | if (BDEV_STATE(&dc->sb) == BDEV_STATE_DIRTY) { |
1005 | bch_sectors_dirty_init(dc); | ||
963 | atomic_set(&dc->has_dirty, 1); | 1006 | atomic_set(&dc->has_dirty, 1); |
964 | atomic_inc(&dc->count); | 1007 | atomic_inc(&dc->count); |
965 | bch_writeback_queue(dc); | 1008 | bch_writeback_queue(dc); |
@@ -1014,6 +1057,14 @@ static void cached_dev_flush(struct closure *cl) | |||
1014 | struct cached_dev *dc = container_of(cl, struct cached_dev, disk.cl); | 1057 | struct cached_dev *dc = container_of(cl, struct cached_dev, disk.cl); |
1015 | struct bcache_device *d = &dc->disk; | 1058 | struct bcache_device *d = &dc->disk; |
1016 | 1059 | ||
1060 | mutex_lock(&bch_register_lock); | ||
1061 | d->flush_done = 1; | ||
1062 | |||
1063 | if (d->c) | ||
1064 | bcache_device_unlink(d); | ||
1065 | |||
1066 | mutex_unlock(&bch_register_lock); | ||
1067 | |||
1017 | bch_cache_accounting_destroy(&dc->accounting); | 1068 | bch_cache_accounting_destroy(&dc->accounting); |
1018 | kobject_del(&d->kobj); | 1069 | kobject_del(&d->kobj); |
1019 | 1070 | ||
@@ -1045,7 +1096,8 @@ static int cached_dev_init(struct cached_dev *dc, unsigned block_size) | |||
1045 | hlist_add_head(&io->hash, dc->io_hash + RECENT_IO); | 1096 | hlist_add_head(&io->hash, dc->io_hash + RECENT_IO); |
1046 | } | 1097 | } |
1047 | 1098 | ||
1048 | ret = bcache_device_init(&dc->disk, block_size); | 1099 | ret = bcache_device_init(&dc->disk, block_size, |
1100 | dc->bdev->bd_part->nr_sects - dc->sb.data_offset); | ||
1049 | if (ret) | 1101 | if (ret) |
1050 | return ret; | 1102 | return ret; |
1051 | 1103 | ||
@@ -1144,11 +1196,10 @@ static int flash_dev_run(struct cache_set *c, struct uuid_entry *u) | |||
1144 | 1196 | ||
1145 | kobject_init(&d->kobj, &bch_flash_dev_ktype); | 1197 | kobject_init(&d->kobj, &bch_flash_dev_ktype); |
1146 | 1198 | ||
1147 | if (bcache_device_init(d, block_bytes(c))) | 1199 | if (bcache_device_init(d, block_bytes(c), u->sectors)) |
1148 | goto err; | 1200 | goto err; |
1149 | 1201 | ||
1150 | bcache_device_attach(d, c, u - c->uuids); | 1202 | bcache_device_attach(d, c, u - c->uuids); |
1151 | set_capacity(d->disk, u->sectors); | ||
1152 | bch_flash_dev_request_init(d); | 1203 | bch_flash_dev_request_init(d); |
1153 | add_disk(d->disk); | 1204 | add_disk(d->disk); |
1154 | 1205 | ||
@@ -1255,9 +1306,10 @@ static void cache_set_free(struct closure *cl) | |||
1255 | free_pages((unsigned long) c->uuids, ilog2(bucket_pages(c))); | 1306 | free_pages((unsigned long) c->uuids, ilog2(bucket_pages(c))); |
1256 | free_pages((unsigned long) c->sort, ilog2(bucket_pages(c))); | 1307 | free_pages((unsigned long) c->sort, ilog2(bucket_pages(c))); |
1257 | 1308 | ||
1258 | kfree(c->fill_iter); | ||
1259 | if (c->bio_split) | 1309 | if (c->bio_split) |
1260 | bioset_free(c->bio_split); | 1310 | bioset_free(c->bio_split); |
1311 | if (c->fill_iter) | ||
1312 | mempool_destroy(c->fill_iter); | ||
1261 | if (c->bio_meta) | 1313 | if (c->bio_meta) |
1262 | mempool_destroy(c->bio_meta); | 1314 | mempool_destroy(c->bio_meta); |
1263 | if (c->search) | 1315 | if (c->search) |
@@ -1278,11 +1330,9 @@ static void cache_set_free(struct closure *cl) | |||
1278 | static void cache_set_flush(struct closure *cl) | 1330 | static void cache_set_flush(struct closure *cl) |
1279 | { | 1331 | { |
1280 | struct cache_set *c = container_of(cl, struct cache_set, caching); | 1332 | struct cache_set *c = container_of(cl, struct cache_set, caching); |
1333 | struct cache *ca; | ||
1281 | struct btree *b; | 1334 | struct btree *b; |
1282 | 1335 | unsigned i; | |
1283 | /* Shut down allocator threads */ | ||
1284 | set_bit(CACHE_SET_STOPPING_2, &c->flags); | ||
1285 | wake_up(&c->alloc_wait); | ||
1286 | 1336 | ||
1287 | bch_cache_accounting_destroy(&c->accounting); | 1337 | bch_cache_accounting_destroy(&c->accounting); |
1288 | 1338 | ||
@@ -1295,7 +1345,11 @@ static void cache_set_flush(struct closure *cl) | |||
1295 | /* Should skip this if we're unregistering because of an error */ | 1345 | /* Should skip this if we're unregistering because of an error */ |
1296 | list_for_each_entry(b, &c->btree_cache, list) | 1346 | list_for_each_entry(b, &c->btree_cache, list) |
1297 | if (btree_node_dirty(b)) | 1347 | if (btree_node_dirty(b)) |
1298 | bch_btree_write(b, true, NULL); | 1348 | bch_btree_node_write(b, NULL); |
1349 | |||
1350 | for_each_cache(ca, c, i) | ||
1351 | if (ca->alloc_thread) | ||
1352 | kthread_stop(ca->alloc_thread); | ||
1299 | 1353 | ||
1300 | closure_return(cl); | 1354 | closure_return(cl); |
1301 | } | 1355 | } |
@@ -1303,18 +1357,22 @@ static void cache_set_flush(struct closure *cl) | |||
1303 | static void __cache_set_unregister(struct closure *cl) | 1357 | static void __cache_set_unregister(struct closure *cl) |
1304 | { | 1358 | { |
1305 | struct cache_set *c = container_of(cl, struct cache_set, caching); | 1359 | struct cache_set *c = container_of(cl, struct cache_set, caching); |
1306 | struct cached_dev *dc, *t; | 1360 | struct cached_dev *dc; |
1307 | size_t i; | 1361 | size_t i; |
1308 | 1362 | ||
1309 | mutex_lock(&bch_register_lock); | 1363 | mutex_lock(&bch_register_lock); |
1310 | 1364 | ||
1311 | if (test_bit(CACHE_SET_UNREGISTERING, &c->flags)) | ||
1312 | list_for_each_entry_safe(dc, t, &c->cached_devs, list) | ||
1313 | bch_cached_dev_detach(dc); | ||
1314 | |||
1315 | for (i = 0; i < c->nr_uuids; i++) | 1365 | for (i = 0; i < c->nr_uuids; i++) |
1316 | if (c->devices[i] && UUID_FLASH_ONLY(&c->uuids[i])) | 1366 | if (c->devices[i]) { |
1317 | bcache_device_stop(c->devices[i]); | 1367 | if (!UUID_FLASH_ONLY(&c->uuids[i]) && |
1368 | test_bit(CACHE_SET_UNREGISTERING, &c->flags)) { | ||
1369 | dc = container_of(c->devices[i], | ||
1370 | struct cached_dev, disk); | ||
1371 | bch_cached_dev_detach(dc); | ||
1372 | } else { | ||
1373 | bcache_device_stop(c->devices[i]); | ||
1374 | } | ||
1375 | } | ||
1318 | 1376 | ||
1319 | mutex_unlock(&bch_register_lock); | 1377 | mutex_unlock(&bch_register_lock); |
1320 | 1378 | ||
@@ -1373,9 +1431,9 @@ struct cache_set *bch_cache_set_alloc(struct cache_sb *sb) | |||
1373 | c->btree_pages = max_t(int, c->btree_pages / 4, | 1431 | c->btree_pages = max_t(int, c->btree_pages / 4, |
1374 | BTREE_MAX_PAGES); | 1432 | BTREE_MAX_PAGES); |
1375 | 1433 | ||
1376 | init_waitqueue_head(&c->alloc_wait); | 1434 | c->sort_crit_factor = int_sqrt(c->btree_pages); |
1435 | |||
1377 | mutex_init(&c->bucket_lock); | 1436 | mutex_init(&c->bucket_lock); |
1378 | mutex_init(&c->fill_lock); | ||
1379 | mutex_init(&c->sort_lock); | 1437 | mutex_init(&c->sort_lock); |
1380 | spin_lock_init(&c->sort_time_lock); | 1438 | spin_lock_init(&c->sort_time_lock); |
1381 | closure_init_unlocked(&c->sb_write); | 1439 | closure_init_unlocked(&c->sb_write); |
@@ -1401,8 +1459,8 @@ struct cache_set *bch_cache_set_alloc(struct cache_sb *sb) | |||
1401 | !(c->bio_meta = mempool_create_kmalloc_pool(2, | 1459 | !(c->bio_meta = mempool_create_kmalloc_pool(2, |
1402 | sizeof(struct bbio) + sizeof(struct bio_vec) * | 1460 | sizeof(struct bbio) + sizeof(struct bio_vec) * |
1403 | bucket_pages(c))) || | 1461 | bucket_pages(c))) || |
1462 | !(c->fill_iter = mempool_create_kmalloc_pool(1, iter_size)) || | ||
1404 | !(c->bio_split = bioset_create(4, offsetof(struct bbio, bio))) || | 1463 | !(c->bio_split = bioset_create(4, offsetof(struct bbio, bio))) || |
1405 | !(c->fill_iter = kmalloc(iter_size, GFP_KERNEL)) || | ||
1406 | !(c->sort = alloc_bucket_pages(GFP_KERNEL, c)) || | 1464 | !(c->sort = alloc_bucket_pages(GFP_KERNEL, c)) || |
1407 | !(c->uuids = alloc_bucket_pages(GFP_KERNEL, c)) || | 1465 | !(c->uuids = alloc_bucket_pages(GFP_KERNEL, c)) || |
1408 | bch_journal_alloc(c) || | 1466 | bch_journal_alloc(c) || |
@@ -1410,8 +1468,6 @@ struct cache_set *bch_cache_set_alloc(struct cache_sb *sb) | |||
1410 | bch_open_buckets_alloc(c)) | 1468 | bch_open_buckets_alloc(c)) |
1411 | goto err; | 1469 | goto err; |
1412 | 1470 | ||
1413 | c->fill_iter->size = sb->bucket_size / sb->block_size; | ||
1414 | |||
1415 | c->congested_read_threshold_us = 2000; | 1471 | c->congested_read_threshold_us = 2000; |
1416 | c->congested_write_threshold_us = 20000; | 1472 | c->congested_write_threshold_us = 20000; |
1417 | c->error_limit = 8 << IO_ERROR_SHIFT; | 1473 | c->error_limit = 8 << IO_ERROR_SHIFT; |
@@ -1496,9 +1552,10 @@ static void run_cache_set(struct cache_set *c) | |||
1496 | */ | 1552 | */ |
1497 | bch_journal_next(&c->journal); | 1553 | bch_journal_next(&c->journal); |
1498 | 1554 | ||
1555 | err = "error starting allocator thread"; | ||
1499 | for_each_cache(ca, c, i) | 1556 | for_each_cache(ca, c, i) |
1500 | closure_call(&ca->alloc, bch_allocator_thread, | 1557 | if (bch_cache_allocator_start(ca)) |
1501 | system_wq, &c->cl); | 1558 | goto err; |
1502 | 1559 | ||
1503 | /* | 1560 | /* |
1504 | * First place it's safe to allocate: btree_check() and | 1561 | * First place it's safe to allocate: btree_check() and |
@@ -1531,17 +1588,16 @@ static void run_cache_set(struct cache_set *c) | |||
1531 | 1588 | ||
1532 | bch_btree_gc_finish(c); | 1589 | bch_btree_gc_finish(c); |
1533 | 1590 | ||
1591 | err = "error starting allocator thread"; | ||
1534 | for_each_cache(ca, c, i) | 1592 | for_each_cache(ca, c, i) |
1535 | closure_call(&ca->alloc, bch_allocator_thread, | 1593 | if (bch_cache_allocator_start(ca)) |
1536 | ca->alloc_workqueue, &c->cl); | 1594 | goto err; |
1537 | 1595 | ||
1538 | mutex_lock(&c->bucket_lock); | 1596 | mutex_lock(&c->bucket_lock); |
1539 | for_each_cache(ca, c, i) | 1597 | for_each_cache(ca, c, i) |
1540 | bch_prio_write(ca); | 1598 | bch_prio_write(ca); |
1541 | mutex_unlock(&c->bucket_lock); | 1599 | mutex_unlock(&c->bucket_lock); |
1542 | 1600 | ||
1543 | wake_up(&c->alloc_wait); | ||
1544 | |||
1545 | err = "cannot allocate new UUID bucket"; | 1601 | err = "cannot allocate new UUID bucket"; |
1546 | if (__uuid_write(c)) | 1602 | if (__uuid_write(c)) |
1547 | goto err_unlock_gc; | 1603 | goto err_unlock_gc; |
@@ -1552,7 +1608,7 @@ static void run_cache_set(struct cache_set *c) | |||
1552 | goto err_unlock_gc; | 1608 | goto err_unlock_gc; |
1553 | 1609 | ||
1554 | bkey_copy_key(&c->root->key, &MAX_KEY); | 1610 | bkey_copy_key(&c->root->key, &MAX_KEY); |
1555 | bch_btree_write(c->root, true, &op); | 1611 | bch_btree_node_write(c->root, &op.cl); |
1556 | 1612 | ||
1557 | bch_btree_set_root(c->root); | 1613 | bch_btree_set_root(c->root); |
1558 | rw_unlock(true, c->root); | 1614 | rw_unlock(true, c->root); |
@@ -1673,9 +1729,6 @@ void bch_cache_release(struct kobject *kobj) | |||
1673 | 1729 | ||
1674 | bio_split_pool_free(&ca->bio_split_hook); | 1730 | bio_split_pool_free(&ca->bio_split_hook); |
1675 | 1731 | ||
1676 | if (ca->alloc_workqueue) | ||
1677 | destroy_workqueue(ca->alloc_workqueue); | ||
1678 | |||
1679 | free_pages((unsigned long) ca->disk_buckets, ilog2(bucket_pages(ca))); | 1732 | free_pages((unsigned long) ca->disk_buckets, ilog2(bucket_pages(ca))); |
1680 | kfree(ca->prio_buckets); | 1733 | kfree(ca->prio_buckets); |
1681 | vfree(ca->buckets); | 1734 | vfree(ca->buckets); |
@@ -1723,7 +1776,6 @@ static int cache_alloc(struct cache_sb *sb, struct cache *ca) | |||
1723 | !(ca->prio_buckets = kzalloc(sizeof(uint64_t) * prio_buckets(ca) * | 1776 | !(ca->prio_buckets = kzalloc(sizeof(uint64_t) * prio_buckets(ca) * |
1724 | 2, GFP_KERNEL)) || | 1777 | 2, GFP_KERNEL)) || |
1725 | !(ca->disk_buckets = alloc_bucket_pages(GFP_KERNEL, ca)) || | 1778 | !(ca->disk_buckets = alloc_bucket_pages(GFP_KERNEL, ca)) || |
1726 | !(ca->alloc_workqueue = alloc_workqueue("bch_allocator", 0, 1)) || | ||
1727 | bio_split_pool_init(&ca->bio_split_hook)) | 1779 | bio_split_pool_init(&ca->bio_split_hook)) |
1728 | return -ENOMEM; | 1780 | return -ENOMEM; |
1729 | 1781 | ||
@@ -1786,6 +1838,36 @@ static ssize_t register_bcache(struct kobject *, struct kobj_attribute *, | |||
1786 | kobj_attribute_write(register, register_bcache); | 1838 | kobj_attribute_write(register, register_bcache); |
1787 | kobj_attribute_write(register_quiet, register_bcache); | 1839 | kobj_attribute_write(register_quiet, register_bcache); |
1788 | 1840 | ||
1841 | static bool bch_is_open_backing(struct block_device *bdev) { | ||
1842 | struct cache_set *c, *tc; | ||
1843 | struct cached_dev *dc, *t; | ||
1844 | |||
1845 | list_for_each_entry_safe(c, tc, &bch_cache_sets, list) | ||
1846 | list_for_each_entry_safe(dc, t, &c->cached_devs, list) | ||
1847 | if (dc->bdev == bdev) | ||
1848 | return true; | ||
1849 | list_for_each_entry_safe(dc, t, &uncached_devices, list) | ||
1850 | if (dc->bdev == bdev) | ||
1851 | return true; | ||
1852 | return false; | ||
1853 | } | ||
1854 | |||
1855 | static bool bch_is_open_cache(struct block_device *bdev) { | ||
1856 | struct cache_set *c, *tc; | ||
1857 | struct cache *ca; | ||
1858 | unsigned i; | ||
1859 | |||
1860 | list_for_each_entry_safe(c, tc, &bch_cache_sets, list) | ||
1861 | for_each_cache(ca, c, i) | ||
1862 | if (ca->bdev == bdev) | ||
1863 | return true; | ||
1864 | return false; | ||
1865 | } | ||
1866 | |||
1867 | static bool bch_is_open(struct block_device *bdev) { | ||
1868 | return bch_is_open_cache(bdev) || bch_is_open_backing(bdev); | ||
1869 | } | ||
1870 | |||
1789 | static ssize_t register_bcache(struct kobject *k, struct kobj_attribute *attr, | 1871 | static ssize_t register_bcache(struct kobject *k, struct kobj_attribute *attr, |
1790 | const char *buffer, size_t size) | 1872 | const char *buffer, size_t size) |
1791 | { | 1873 | { |
@@ -1810,8 +1892,13 @@ static ssize_t register_bcache(struct kobject *k, struct kobj_attribute *attr, | |||
1810 | FMODE_READ|FMODE_WRITE|FMODE_EXCL, | 1892 | FMODE_READ|FMODE_WRITE|FMODE_EXCL, |
1811 | sb); | 1893 | sb); |
1812 | if (IS_ERR(bdev)) { | 1894 | if (IS_ERR(bdev)) { |
1813 | if (bdev == ERR_PTR(-EBUSY)) | 1895 | if (bdev == ERR_PTR(-EBUSY)) { |
1814 | err = "device busy"; | 1896 | bdev = lookup_bdev(strim(path)); |
1897 | if (!IS_ERR(bdev) && bch_is_open(bdev)) | ||
1898 | err = "device already registered"; | ||
1899 | else | ||
1900 | err = "device busy"; | ||
1901 | } | ||
1815 | goto err; | 1902 | goto err; |
1816 | } | 1903 | } |
1817 | 1904 | ||
diff --git a/drivers/md/bcache/sysfs.c b/drivers/md/bcache/sysfs.c index 4d9cca47e4c6..12a2c2846f99 100644 --- a/drivers/md/bcache/sysfs.c +++ b/drivers/md/bcache/sysfs.c | |||
@@ -9,7 +9,9 @@ | |||
9 | #include "sysfs.h" | 9 | #include "sysfs.h" |
10 | #include "btree.h" | 10 | #include "btree.h" |
11 | #include "request.h" | 11 | #include "request.h" |
12 | #include "writeback.h" | ||
12 | 13 | ||
14 | #include <linux/blkdev.h> | ||
13 | #include <linux/sort.h> | 15 | #include <linux/sort.h> |
14 | 16 | ||
15 | static const char * const cache_replacement_policies[] = { | 17 | static const char * const cache_replacement_policies[] = { |
@@ -79,6 +81,9 @@ rw_attribute(writeback_rate_p_term_inverse); | |||
79 | rw_attribute(writeback_rate_d_smooth); | 81 | rw_attribute(writeback_rate_d_smooth); |
80 | read_attribute(writeback_rate_debug); | 82 | read_attribute(writeback_rate_debug); |
81 | 83 | ||
84 | read_attribute(stripe_size); | ||
85 | read_attribute(partial_stripes_expensive); | ||
86 | |||
82 | rw_attribute(synchronous); | 87 | rw_attribute(synchronous); |
83 | rw_attribute(journal_delay_ms); | 88 | rw_attribute(journal_delay_ms); |
84 | rw_attribute(discard); | 89 | rw_attribute(discard); |
@@ -127,7 +132,7 @@ SHOW(__bch_cached_dev) | |||
127 | char derivative[20]; | 132 | char derivative[20]; |
128 | char target[20]; | 133 | char target[20]; |
129 | bch_hprint(dirty, | 134 | bch_hprint(dirty, |
130 | atomic_long_read(&dc->disk.sectors_dirty) << 9); | 135 | bcache_dev_sectors_dirty(&dc->disk) << 9); |
131 | bch_hprint(derivative, dc->writeback_rate_derivative << 9); | 136 | bch_hprint(derivative, dc->writeback_rate_derivative << 9); |
132 | bch_hprint(target, dc->writeback_rate_target << 9); | 137 | bch_hprint(target, dc->writeback_rate_target << 9); |
133 | 138 | ||
@@ -143,7 +148,10 @@ SHOW(__bch_cached_dev) | |||
143 | } | 148 | } |
144 | 149 | ||
145 | sysfs_hprint(dirty_data, | 150 | sysfs_hprint(dirty_data, |
146 | atomic_long_read(&dc->disk.sectors_dirty) << 9); | 151 | bcache_dev_sectors_dirty(&dc->disk) << 9); |
152 | |||
153 | sysfs_hprint(stripe_size, (1 << dc->disk.stripe_size_bits) << 9); | ||
154 | var_printf(partial_stripes_expensive, "%u"); | ||
147 | 155 | ||
148 | var_printf(sequential_merge, "%i"); | 156 | var_printf(sequential_merge, "%i"); |
149 | var_hprint(sequential_cutoff); | 157 | var_hprint(sequential_cutoff); |
@@ -170,6 +178,7 @@ STORE(__cached_dev) | |||
170 | disk.kobj); | 178 | disk.kobj); |
171 | unsigned v = size; | 179 | unsigned v = size; |
172 | struct cache_set *c; | 180 | struct cache_set *c; |
181 | struct kobj_uevent_env *env; | ||
173 | 182 | ||
174 | #define d_strtoul(var) sysfs_strtoul(var, dc->var) | 183 | #define d_strtoul(var) sysfs_strtoul(var, dc->var) |
175 | #define d_strtoi_h(var) sysfs_hatoi(var, dc->var) | 184 | #define d_strtoi_h(var) sysfs_hatoi(var, dc->var) |
@@ -214,6 +223,7 @@ STORE(__cached_dev) | |||
214 | } | 223 | } |
215 | 224 | ||
216 | if (attr == &sysfs_label) { | 225 | if (attr == &sysfs_label) { |
226 | /* note: endlines are preserved */ | ||
217 | memcpy(dc->sb.label, buf, SB_LABEL_SIZE); | 227 | memcpy(dc->sb.label, buf, SB_LABEL_SIZE); |
218 | bch_write_bdev_super(dc, NULL); | 228 | bch_write_bdev_super(dc, NULL); |
219 | if (dc->disk.c) { | 229 | if (dc->disk.c) { |
@@ -221,6 +231,15 @@ STORE(__cached_dev) | |||
221 | buf, SB_LABEL_SIZE); | 231 | buf, SB_LABEL_SIZE); |
222 | bch_uuid_write(dc->disk.c); | 232 | bch_uuid_write(dc->disk.c); |
223 | } | 233 | } |
234 | env = kzalloc(sizeof(struct kobj_uevent_env), GFP_KERNEL); | ||
235 | if (!env) | ||
236 | return -ENOMEM; | ||
237 | add_uevent_var(env, "DRIVER=bcache"); | ||
238 | add_uevent_var(env, "CACHED_UUID=%pU", dc->sb.uuid), | ||
239 | add_uevent_var(env, "CACHED_LABEL=%s", buf); | ||
240 | kobject_uevent_env( | ||
241 | &disk_to_dev(dc->disk.disk)->kobj, KOBJ_CHANGE, env->envp); | ||
242 | kfree(env); | ||
224 | } | 243 | } |
225 | 244 | ||
226 | if (attr == &sysfs_attach) { | 245 | if (attr == &sysfs_attach) { |
@@ -284,6 +303,8 @@ static struct attribute *bch_cached_dev_files[] = { | |||
284 | &sysfs_writeback_rate_d_smooth, | 303 | &sysfs_writeback_rate_d_smooth, |
285 | &sysfs_writeback_rate_debug, | 304 | &sysfs_writeback_rate_debug, |
286 | &sysfs_dirty_data, | 305 | &sysfs_dirty_data, |
306 | &sysfs_stripe_size, | ||
307 | &sysfs_partial_stripes_expensive, | ||
287 | &sysfs_sequential_cutoff, | 308 | &sysfs_sequential_cutoff, |
288 | &sysfs_sequential_merge, | 309 | &sysfs_sequential_merge, |
289 | &sysfs_clear_stats, | 310 | &sysfs_clear_stats, |
@@ -665,12 +686,10 @@ SHOW(__bch_cache) | |||
665 | int cmp(const void *l, const void *r) | 686 | int cmp(const void *l, const void *r) |
666 | { return *((uint16_t *) r) - *((uint16_t *) l); } | 687 | { return *((uint16_t *) r) - *((uint16_t *) l); } |
667 | 688 | ||
668 | /* Number of quantiles we compute */ | ||
669 | const unsigned nq = 31; | ||
670 | |||
671 | size_t n = ca->sb.nbuckets, i, unused, btree; | 689 | size_t n = ca->sb.nbuckets, i, unused, btree; |
672 | uint64_t sum = 0; | 690 | uint64_t sum = 0; |
673 | uint16_t q[nq], *p, *cached; | 691 | /* Compute 31 quantiles */ |
692 | uint16_t q[31], *p, *cached; | ||
674 | ssize_t ret; | 693 | ssize_t ret; |
675 | 694 | ||
676 | cached = p = vmalloc(ca->sb.nbuckets * sizeof(uint16_t)); | 695 | cached = p = vmalloc(ca->sb.nbuckets * sizeof(uint16_t)); |
@@ -703,26 +722,29 @@ SHOW(__bch_cache) | |||
703 | if (n) | 722 | if (n) |
704 | do_div(sum, n); | 723 | do_div(sum, n); |
705 | 724 | ||
706 | for (i = 0; i < nq; i++) | 725 | for (i = 0; i < ARRAY_SIZE(q); i++) |
707 | q[i] = INITIAL_PRIO - cached[n * (i + 1) / (nq + 1)]; | 726 | q[i] = INITIAL_PRIO - cached[n * (i + 1) / |
727 | (ARRAY_SIZE(q) + 1)]; | ||
708 | 728 | ||
709 | vfree(p); | 729 | vfree(p); |
710 | 730 | ||
711 | ret = snprintf(buf, PAGE_SIZE, | 731 | ret = scnprintf(buf, PAGE_SIZE, |
712 | "Unused: %zu%%\n" | 732 | "Unused: %zu%%\n" |
713 | "Metadata: %zu%%\n" | 733 | "Metadata: %zu%%\n" |
714 | "Average: %llu\n" | 734 | "Average: %llu\n" |
715 | "Sectors per Q: %zu\n" | 735 | "Sectors per Q: %zu\n" |
716 | "Quantiles: [", | 736 | "Quantiles: [", |
717 | unused * 100 / (size_t) ca->sb.nbuckets, | 737 | unused * 100 / (size_t) ca->sb.nbuckets, |
718 | btree * 100 / (size_t) ca->sb.nbuckets, sum, | 738 | btree * 100 / (size_t) ca->sb.nbuckets, sum, |
719 | n * ca->sb.bucket_size / (nq + 1)); | 739 | n * ca->sb.bucket_size / (ARRAY_SIZE(q) + 1)); |
720 | 740 | ||
721 | for (i = 0; i < nq && ret < (ssize_t) PAGE_SIZE; i++) | 741 | for (i = 0; i < ARRAY_SIZE(q); i++) |
722 | ret += snprintf(buf + ret, PAGE_SIZE - ret, | 742 | ret += scnprintf(buf + ret, PAGE_SIZE - ret, |
723 | i < nq - 1 ? "%u " : "%u]\n", q[i]); | 743 | "%u ", q[i]); |
724 | 744 | ret--; | |
725 | buf[PAGE_SIZE - 1] = '\0'; | 745 | |
746 | ret += scnprintf(buf + ret, PAGE_SIZE - ret, "]\n"); | ||
747 | |||
726 | return ret; | 748 | return ret; |
727 | } | 749 | } |
728 | 750 | ||
diff --git a/drivers/md/bcache/trace.c b/drivers/md/bcache/trace.c index 983f9bb411bc..f7b6c197f90f 100644 --- a/drivers/md/bcache/trace.c +++ b/drivers/md/bcache/trace.c | |||
@@ -2,6 +2,7 @@ | |||
2 | #include "btree.h" | 2 | #include "btree.h" |
3 | #include "request.h" | 3 | #include "request.h" |
4 | 4 | ||
5 | #include <linux/blktrace_api.h> | ||
5 | #include <linux/module.h> | 6 | #include <linux/module.h> |
6 | 7 | ||
7 | #define CREATE_TRACE_POINTS | 8 | #define CREATE_TRACE_POINTS |
@@ -9,18 +10,44 @@ | |||
9 | 10 | ||
10 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_request_start); | 11 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_request_start); |
11 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_request_end); | 12 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_request_end); |
12 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_passthrough); | 13 | |
13 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_cache_hit); | 14 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_bypass_sequential); |
14 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_cache_miss); | 15 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_bypass_congested); |
16 | |||
17 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_read); | ||
18 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_write); | ||
15 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_read_retry); | 19 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_read_retry); |
16 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_writethrough); | 20 | |
17 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_writeback); | 21 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_cache_insert); |
18 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_write_skip); | 22 | |
23 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_journal_replay_key); | ||
24 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_journal_write); | ||
25 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_journal_full); | ||
26 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_journal_entry_full); | ||
27 | |||
28 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_btree_cache_cannibalize); | ||
29 | |||
19 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_btree_read); | 30 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_btree_read); |
20 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_btree_write); | 31 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_btree_write); |
21 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_write_dirty); | 32 | |
22 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_read_dirty); | 33 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_btree_node_alloc); |
23 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_journal_write); | 34 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_btree_node_alloc_fail); |
24 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_cache_insert); | 35 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_btree_node_free); |
36 | |||
37 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_btree_gc_coalesce); | ||
25 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_gc_start); | 38 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_gc_start); |
26 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_gc_end); | 39 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_gc_end); |
40 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_gc_copy); | ||
41 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_gc_copy_collision); | ||
42 | |||
43 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_btree_insert_key); | ||
44 | |||
45 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_btree_node_split); | ||
46 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_btree_node_compact); | ||
47 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_btree_set_root); | ||
48 | |||
49 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_alloc_invalidate); | ||
50 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_alloc_fail); | ||
51 | |||
52 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_writeback); | ||
53 | EXPORT_TRACEPOINT_SYMBOL_GPL(bcache_writeback_collision); | ||
diff --git a/drivers/md/bcache/util.c b/drivers/md/bcache/util.c index da3a99e85b1e..98eb81159a22 100644 --- a/drivers/md/bcache/util.c +++ b/drivers/md/bcache/util.c | |||
@@ -228,23 +228,6 @@ start: bv->bv_len = min_t(size_t, PAGE_SIZE - bv->bv_offset, | |||
228 | } | 228 | } |
229 | } | 229 | } |
230 | 230 | ||
231 | int bch_bio_alloc_pages(struct bio *bio, gfp_t gfp) | ||
232 | { | ||
233 | int i; | ||
234 | struct bio_vec *bv; | ||
235 | |||
236 | bio_for_each_segment(bv, bio, i) { | ||
237 | bv->bv_page = alloc_page(gfp); | ||
238 | if (!bv->bv_page) { | ||
239 | while (bv-- != bio->bi_io_vec + bio->bi_idx) | ||
240 | __free_page(bv->bv_page); | ||
241 | return -ENOMEM; | ||
242 | } | ||
243 | } | ||
244 | |||
245 | return 0; | ||
246 | } | ||
247 | |||
248 | /* | 231 | /* |
249 | * Portions Copyright (c) 1996-2001, PostgreSQL Global Development Group (Any | 232 | * Portions Copyright (c) 1996-2001, PostgreSQL Global Development Group (Any |
250 | * use permitted, subject to terms of PostgreSQL license; see.) | 233 | * use permitted, subject to terms of PostgreSQL license; see.) |
diff --git a/drivers/md/bcache/util.h b/drivers/md/bcache/util.h index 577393e38c3a..1ae2a73ad85f 100644 --- a/drivers/md/bcache/util.h +++ b/drivers/md/bcache/util.h | |||
@@ -15,8 +15,6 @@ | |||
15 | 15 | ||
16 | struct closure; | 16 | struct closure; |
17 | 17 | ||
18 | #include <trace/events/bcache.h> | ||
19 | |||
20 | #ifdef CONFIG_BCACHE_EDEBUG | 18 | #ifdef CONFIG_BCACHE_EDEBUG |
21 | 19 | ||
22 | #define atomic_dec_bug(v) BUG_ON(atomic_dec_return(v) < 0) | 20 | #define atomic_dec_bug(v) BUG_ON(atomic_dec_return(v) < 0) |
@@ -566,12 +564,8 @@ static inline unsigned fract_exp_two(unsigned x, unsigned fract_bits) | |||
566 | return x; | 564 | return x; |
567 | } | 565 | } |
568 | 566 | ||
569 | #define bio_end(bio) ((bio)->bi_sector + bio_sectors(bio)) | ||
570 | |||
571 | void bch_bio_map(struct bio *bio, void *base); | 567 | void bch_bio_map(struct bio *bio, void *base); |
572 | 568 | ||
573 | int bch_bio_alloc_pages(struct bio *bio, gfp_t gfp); | ||
574 | |||
575 | static inline sector_t bdev_sectors(struct block_device *bdev) | 569 | static inline sector_t bdev_sectors(struct block_device *bdev) |
576 | { | 570 | { |
577 | return bdev->bd_inode->i_size >> 9; | 571 | return bdev->bd_inode->i_size >> 9; |
diff --git a/drivers/md/bcache/writeback.c b/drivers/md/bcache/writeback.c index 2714ed3991d1..22cbff551628 100644 --- a/drivers/md/bcache/writeback.c +++ b/drivers/md/bcache/writeback.c | |||
@@ -9,6 +9,9 @@ | |||
9 | #include "bcache.h" | 9 | #include "bcache.h" |
10 | #include "btree.h" | 10 | #include "btree.h" |
11 | #include "debug.h" | 11 | #include "debug.h" |
12 | #include "writeback.h" | ||
13 | |||
14 | #include <trace/events/bcache.h> | ||
12 | 15 | ||
13 | static struct workqueue_struct *dirty_wq; | 16 | static struct workqueue_struct *dirty_wq; |
14 | 17 | ||
@@ -36,7 +39,7 @@ static void __update_writeback_rate(struct cached_dev *dc) | |||
36 | 39 | ||
37 | int change = 0; | 40 | int change = 0; |
38 | int64_t error; | 41 | int64_t error; |
39 | int64_t dirty = atomic_long_read(&dc->disk.sectors_dirty); | 42 | int64_t dirty = bcache_dev_sectors_dirty(&dc->disk); |
40 | int64_t derivative = dirty - dc->disk.sectors_dirty_last; | 43 | int64_t derivative = dirty - dc->disk.sectors_dirty_last; |
41 | 44 | ||
42 | dc->disk.sectors_dirty_last = dirty; | 45 | dc->disk.sectors_dirty_last = dirty; |
@@ -105,6 +108,31 @@ static bool dirty_pred(struct keybuf *buf, struct bkey *k) | |||
105 | return KEY_DIRTY(k); | 108 | return KEY_DIRTY(k); |
106 | } | 109 | } |
107 | 110 | ||
111 | static bool dirty_full_stripe_pred(struct keybuf *buf, struct bkey *k) | ||
112 | { | ||
113 | uint64_t stripe; | ||
114 | unsigned nr_sectors = KEY_SIZE(k); | ||
115 | struct cached_dev *dc = container_of(buf, struct cached_dev, | ||
116 | writeback_keys); | ||
117 | unsigned stripe_size = 1 << dc->disk.stripe_size_bits; | ||
118 | |||
119 | if (!KEY_DIRTY(k)) | ||
120 | return false; | ||
121 | |||
122 | stripe = KEY_START(k) >> dc->disk.stripe_size_bits; | ||
123 | while (1) { | ||
124 | if (atomic_read(dc->disk.stripe_sectors_dirty + stripe) != | ||
125 | stripe_size) | ||
126 | return false; | ||
127 | |||
128 | if (nr_sectors <= stripe_size) | ||
129 | return true; | ||
130 | |||
131 | nr_sectors -= stripe_size; | ||
132 | stripe++; | ||
133 | } | ||
134 | } | ||
135 | |||
108 | static void dirty_init(struct keybuf_key *w) | 136 | static void dirty_init(struct keybuf_key *w) |
109 | { | 137 | { |
110 | struct dirty_io *io = w->private; | 138 | struct dirty_io *io = w->private; |
@@ -149,7 +177,22 @@ static void refill_dirty(struct closure *cl) | |||
149 | searched_from_start = true; | 177 | searched_from_start = true; |
150 | } | 178 | } |
151 | 179 | ||
152 | bch_refill_keybuf(dc->disk.c, buf, &end); | 180 | if (dc->partial_stripes_expensive) { |
181 | uint64_t i; | ||
182 | |||
183 | for (i = 0; i < dc->disk.nr_stripes; i++) | ||
184 | if (atomic_read(dc->disk.stripe_sectors_dirty + i) == | ||
185 | 1 << dc->disk.stripe_size_bits) | ||
186 | goto full_stripes; | ||
187 | |||
188 | goto normal_refill; | ||
189 | full_stripes: | ||
190 | bch_refill_keybuf(dc->disk.c, buf, &end, | ||
191 | dirty_full_stripe_pred); | ||
192 | } else { | ||
193 | normal_refill: | ||
194 | bch_refill_keybuf(dc->disk.c, buf, &end, dirty_pred); | ||
195 | } | ||
153 | 196 | ||
154 | if (bkey_cmp(&buf->last_scanned, &end) >= 0 && searched_from_start) { | 197 | if (bkey_cmp(&buf->last_scanned, &end) >= 0 && searched_from_start) { |
155 | /* Searched the entire btree - delay awhile */ | 198 | /* Searched the entire btree - delay awhile */ |
@@ -181,10 +224,8 @@ void bch_writeback_queue(struct cached_dev *dc) | |||
181 | } | 224 | } |
182 | } | 225 | } |
183 | 226 | ||
184 | void bch_writeback_add(struct cached_dev *dc, unsigned sectors) | 227 | void bch_writeback_add(struct cached_dev *dc) |
185 | { | 228 | { |
186 | atomic_long_add(sectors, &dc->disk.sectors_dirty); | ||
187 | |||
188 | if (!atomic_read(&dc->has_dirty) && | 229 | if (!atomic_read(&dc->has_dirty) && |
189 | !atomic_xchg(&dc->has_dirty, 1)) { | 230 | !atomic_xchg(&dc->has_dirty, 1)) { |
190 | atomic_inc(&dc->count); | 231 | atomic_inc(&dc->count); |
@@ -203,6 +244,34 @@ void bch_writeback_add(struct cached_dev *dc, unsigned sectors) | |||
203 | } | 244 | } |
204 | } | 245 | } |
205 | 246 | ||
247 | void bcache_dev_sectors_dirty_add(struct cache_set *c, unsigned inode, | ||
248 | uint64_t offset, int nr_sectors) | ||
249 | { | ||
250 | struct bcache_device *d = c->devices[inode]; | ||
251 | unsigned stripe_size, stripe_offset; | ||
252 | uint64_t stripe; | ||
253 | |||
254 | if (!d) | ||
255 | return; | ||
256 | |||
257 | stripe_size = 1 << d->stripe_size_bits; | ||
258 | stripe = offset >> d->stripe_size_bits; | ||
259 | stripe_offset = offset & (stripe_size - 1); | ||
260 | |||
261 | while (nr_sectors) { | ||
262 | int s = min_t(unsigned, abs(nr_sectors), | ||
263 | stripe_size - stripe_offset); | ||
264 | |||
265 | if (nr_sectors < 0) | ||
266 | s = -s; | ||
267 | |||
268 | atomic_add(s, d->stripe_sectors_dirty + stripe); | ||
269 | nr_sectors -= s; | ||
270 | stripe_offset = 0; | ||
271 | stripe++; | ||
272 | } | ||
273 | } | ||
274 | |||
206 | /* Background writeback - IO loop */ | 275 | /* Background writeback - IO loop */ |
207 | 276 | ||
208 | static void dirty_io_destructor(struct closure *cl) | 277 | static void dirty_io_destructor(struct closure *cl) |
@@ -216,9 +285,10 @@ static void write_dirty_finish(struct closure *cl) | |||
216 | struct dirty_io *io = container_of(cl, struct dirty_io, cl); | 285 | struct dirty_io *io = container_of(cl, struct dirty_io, cl); |
217 | struct keybuf_key *w = io->bio.bi_private; | 286 | struct keybuf_key *w = io->bio.bi_private; |
218 | struct cached_dev *dc = io->dc; | 287 | struct cached_dev *dc = io->dc; |
219 | struct bio_vec *bv = bio_iovec_idx(&io->bio, io->bio.bi_vcnt); | 288 | struct bio_vec *bv; |
289 | int i; | ||
220 | 290 | ||
221 | while (bv-- != io->bio.bi_io_vec) | 291 | bio_for_each_segment_all(bv, &io->bio, i) |
222 | __free_page(bv->bv_page); | 292 | __free_page(bv->bv_page); |
223 | 293 | ||
224 | /* This is kind of a dumb way of signalling errors. */ | 294 | /* This is kind of a dumb way of signalling errors. */ |
@@ -236,10 +306,12 @@ static void write_dirty_finish(struct closure *cl) | |||
236 | for (i = 0; i < KEY_PTRS(&w->key); i++) | 306 | for (i = 0; i < KEY_PTRS(&w->key); i++) |
237 | atomic_inc(&PTR_BUCKET(dc->disk.c, &w->key, i)->pin); | 307 | atomic_inc(&PTR_BUCKET(dc->disk.c, &w->key, i)->pin); |
238 | 308 | ||
239 | pr_debug("clearing %s", pkey(&w->key)); | ||
240 | bch_btree_insert(&op, dc->disk.c); | 309 | bch_btree_insert(&op, dc->disk.c); |
241 | closure_sync(&op.cl); | 310 | closure_sync(&op.cl); |
242 | 311 | ||
312 | if (op.insert_collision) | ||
313 | trace_bcache_writeback_collision(&w->key); | ||
314 | |||
243 | atomic_long_inc(op.insert_collision | 315 | atomic_long_inc(op.insert_collision |
244 | ? &dc->disk.c->writeback_keys_failed | 316 | ? &dc->disk.c->writeback_keys_failed |
245 | : &dc->disk.c->writeback_keys_done); | 317 | : &dc->disk.c->writeback_keys_done); |
@@ -275,7 +347,6 @@ static void write_dirty(struct closure *cl) | |||
275 | io->bio.bi_bdev = io->dc->bdev; | 347 | io->bio.bi_bdev = io->dc->bdev; |
276 | io->bio.bi_end_io = dirty_endio; | 348 | io->bio.bi_end_io = dirty_endio; |
277 | 349 | ||
278 | trace_bcache_write_dirty(&io->bio); | ||
279 | closure_bio_submit(&io->bio, cl, &io->dc->disk); | 350 | closure_bio_submit(&io->bio, cl, &io->dc->disk); |
280 | 351 | ||
281 | continue_at(cl, write_dirty_finish, dirty_wq); | 352 | continue_at(cl, write_dirty_finish, dirty_wq); |
@@ -296,7 +367,6 @@ static void read_dirty_submit(struct closure *cl) | |||
296 | { | 367 | { |
297 | struct dirty_io *io = container_of(cl, struct dirty_io, cl); | 368 | struct dirty_io *io = container_of(cl, struct dirty_io, cl); |
298 | 369 | ||
299 | trace_bcache_read_dirty(&io->bio); | ||
300 | closure_bio_submit(&io->bio, cl, &io->dc->disk); | 370 | closure_bio_submit(&io->bio, cl, &io->dc->disk); |
301 | 371 | ||
302 | continue_at(cl, write_dirty, dirty_wq); | 372 | continue_at(cl, write_dirty, dirty_wq); |
@@ -349,10 +419,10 @@ static void read_dirty(struct closure *cl) | |||
349 | io->bio.bi_rw = READ; | 419 | io->bio.bi_rw = READ; |
350 | io->bio.bi_end_io = read_dirty_endio; | 420 | io->bio.bi_end_io = read_dirty_endio; |
351 | 421 | ||
352 | if (bch_bio_alloc_pages(&io->bio, GFP_KERNEL)) | 422 | if (bio_alloc_pages(&io->bio, GFP_KERNEL)) |
353 | goto err_free; | 423 | goto err_free; |
354 | 424 | ||
355 | pr_debug("%s", pkey(&w->key)); | 425 | trace_bcache_writeback(&w->key); |
356 | 426 | ||
357 | closure_call(&io->cl, read_dirty_submit, NULL, &dc->disk.cl); | 427 | closure_call(&io->cl, read_dirty_submit, NULL, &dc->disk.cl); |
358 | 428 | ||
@@ -375,12 +445,49 @@ err: | |||
375 | refill_dirty(cl); | 445 | refill_dirty(cl); |
376 | } | 446 | } |
377 | 447 | ||
448 | /* Init */ | ||
449 | |||
450 | static int bch_btree_sectors_dirty_init(struct btree *b, struct btree_op *op, | ||
451 | struct cached_dev *dc) | ||
452 | { | ||
453 | struct bkey *k; | ||
454 | struct btree_iter iter; | ||
455 | |||
456 | bch_btree_iter_init(b, &iter, &KEY(dc->disk.id, 0, 0)); | ||
457 | while ((k = bch_btree_iter_next_filter(&iter, b, bch_ptr_bad))) | ||
458 | if (!b->level) { | ||
459 | if (KEY_INODE(k) > dc->disk.id) | ||
460 | break; | ||
461 | |||
462 | if (KEY_DIRTY(k)) | ||
463 | bcache_dev_sectors_dirty_add(b->c, dc->disk.id, | ||
464 | KEY_START(k), | ||
465 | KEY_SIZE(k)); | ||
466 | } else { | ||
467 | btree(sectors_dirty_init, k, b, op, dc); | ||
468 | if (KEY_INODE(k) > dc->disk.id) | ||
469 | break; | ||
470 | |||
471 | cond_resched(); | ||
472 | } | ||
473 | |||
474 | return 0; | ||
475 | } | ||
476 | |||
477 | void bch_sectors_dirty_init(struct cached_dev *dc) | ||
478 | { | ||
479 | struct btree_op op; | ||
480 | |||
481 | bch_btree_op_init_stack(&op); | ||
482 | btree_root(sectors_dirty_init, dc->disk.c, &op, dc); | ||
483 | } | ||
484 | |||
378 | void bch_cached_dev_writeback_init(struct cached_dev *dc) | 485 | void bch_cached_dev_writeback_init(struct cached_dev *dc) |
379 | { | 486 | { |
380 | closure_init_unlocked(&dc->writeback); | 487 | closure_init_unlocked(&dc->writeback); |
381 | init_rwsem(&dc->writeback_lock); | 488 | init_rwsem(&dc->writeback_lock); |
382 | 489 | ||
383 | bch_keybuf_init(&dc->writeback_keys, dirty_pred); | 490 | bch_keybuf_init(&dc->writeback_keys); |
384 | 491 | ||
385 | dc->writeback_metadata = true; | 492 | dc->writeback_metadata = true; |
386 | dc->writeback_running = true; | 493 | dc->writeback_running = true; |
diff --git a/drivers/md/bcache/writeback.h b/drivers/md/bcache/writeback.h new file mode 100644 index 000000000000..c91f61bb95b6 --- /dev/null +++ b/drivers/md/bcache/writeback.h | |||
@@ -0,0 +1,64 @@ | |||
1 | #ifndef _BCACHE_WRITEBACK_H | ||
2 | #define _BCACHE_WRITEBACK_H | ||
3 | |||
4 | #define CUTOFF_WRITEBACK 40 | ||
5 | #define CUTOFF_WRITEBACK_SYNC 70 | ||
6 | |||
7 | static inline uint64_t bcache_dev_sectors_dirty(struct bcache_device *d) | ||
8 | { | ||
9 | uint64_t i, ret = 0; | ||
10 | |||
11 | for (i = 0; i < d->nr_stripes; i++) | ||
12 | ret += atomic_read(d->stripe_sectors_dirty + i); | ||
13 | |||
14 | return ret; | ||
15 | } | ||
16 | |||
17 | static inline bool bcache_dev_stripe_dirty(struct bcache_device *d, | ||
18 | uint64_t offset, | ||
19 | unsigned nr_sectors) | ||
20 | { | ||
21 | uint64_t stripe = offset >> d->stripe_size_bits; | ||
22 | |||
23 | while (1) { | ||
24 | if (atomic_read(d->stripe_sectors_dirty + stripe)) | ||
25 | return true; | ||
26 | |||
27 | if (nr_sectors <= 1 << d->stripe_size_bits) | ||
28 | return false; | ||
29 | |||
30 | nr_sectors -= 1 << d->stripe_size_bits; | ||
31 | stripe++; | ||
32 | } | ||
33 | } | ||
34 | |||
35 | static inline bool should_writeback(struct cached_dev *dc, struct bio *bio, | ||
36 | unsigned cache_mode, bool would_skip) | ||
37 | { | ||
38 | unsigned in_use = dc->disk.c->gc_stats.in_use; | ||
39 | |||
40 | if (cache_mode != CACHE_MODE_WRITEBACK || | ||
41 | atomic_read(&dc->disk.detaching) || | ||
42 | in_use > CUTOFF_WRITEBACK_SYNC) | ||
43 | return false; | ||
44 | |||
45 | if (dc->partial_stripes_expensive && | ||
46 | bcache_dev_stripe_dirty(&dc->disk, bio->bi_sector, | ||
47 | bio_sectors(bio))) | ||
48 | return true; | ||
49 | |||
50 | if (would_skip) | ||
51 | return false; | ||
52 | |||
53 | return bio->bi_rw & REQ_SYNC || | ||
54 | in_use <= CUTOFF_WRITEBACK; | ||
55 | } | ||
56 | |||
57 | void bcache_dev_sectors_dirty_add(struct cache_set *, unsigned, uint64_t, int); | ||
58 | void bch_writeback_queue(struct cached_dev *); | ||
59 | void bch_writeback_add(struct cached_dev *); | ||
60 | |||
61 | void bch_sectors_dirty_init(struct cached_dev *dc); | ||
62 | void bch_cached_dev_writeback_init(struct cached_dev *); | ||
63 | |||
64 | #endif | ||
diff --git a/drivers/md/raid10.c b/drivers/md/raid10.c index 957a719e8c2f..df7b0a06b0ea 100644 --- a/drivers/md/raid10.c +++ b/drivers/md/raid10.c | |||
@@ -2290,12 +2290,18 @@ static void recovery_request_write(struct mddev *mddev, struct r10bio *r10_bio) | |||
2290 | d = r10_bio->devs[1].devnum; | 2290 | d = r10_bio->devs[1].devnum; |
2291 | wbio = r10_bio->devs[1].bio; | 2291 | wbio = r10_bio->devs[1].bio; |
2292 | wbio2 = r10_bio->devs[1].repl_bio; | 2292 | wbio2 = r10_bio->devs[1].repl_bio; |
2293 | /* Need to test wbio2->bi_end_io before we call | ||
2294 | * generic_make_request as if the former is NULL, | ||
2295 | * the latter is free to free wbio2. | ||
2296 | */ | ||
2297 | if (wbio2 && !wbio2->bi_end_io) | ||
2298 | wbio2 = NULL; | ||
2293 | if (wbio->bi_end_io) { | 2299 | if (wbio->bi_end_io) { |
2294 | atomic_inc(&conf->mirrors[d].rdev->nr_pending); | 2300 | atomic_inc(&conf->mirrors[d].rdev->nr_pending); |
2295 | md_sync_acct(conf->mirrors[d].rdev->bdev, bio_sectors(wbio)); | 2301 | md_sync_acct(conf->mirrors[d].rdev->bdev, bio_sectors(wbio)); |
2296 | generic_make_request(wbio); | 2302 | generic_make_request(wbio); |
2297 | } | 2303 | } |
2298 | if (wbio2 && wbio2->bi_end_io) { | 2304 | if (wbio2) { |
2299 | atomic_inc(&conf->mirrors[d].replacement->nr_pending); | 2305 | atomic_inc(&conf->mirrors[d].replacement->nr_pending); |
2300 | md_sync_acct(conf->mirrors[d].replacement->bdev, | 2306 | md_sync_acct(conf->mirrors[d].replacement->bdev, |
2301 | bio_sectors(wbio2)); | 2307 | bio_sectors(wbio2)); |
diff --git a/drivers/md/raid5.c b/drivers/md/raid5.c index 2bf094a587cb..78ea44336e75 100644 --- a/drivers/md/raid5.c +++ b/drivers/md/raid5.c | |||
@@ -3462,6 +3462,7 @@ static void handle_stripe(struct stripe_head *sh) | |||
3462 | test_and_clear_bit(STRIPE_SYNC_REQUESTED, &sh->state)) { | 3462 | test_and_clear_bit(STRIPE_SYNC_REQUESTED, &sh->state)) { |
3463 | set_bit(STRIPE_SYNCING, &sh->state); | 3463 | set_bit(STRIPE_SYNCING, &sh->state); |
3464 | clear_bit(STRIPE_INSYNC, &sh->state); | 3464 | clear_bit(STRIPE_INSYNC, &sh->state); |
3465 | clear_bit(STRIPE_REPLACED, &sh->state); | ||
3465 | } | 3466 | } |
3466 | spin_unlock(&sh->stripe_lock); | 3467 | spin_unlock(&sh->stripe_lock); |
3467 | } | 3468 | } |
@@ -3607,19 +3608,23 @@ static void handle_stripe(struct stripe_head *sh) | |||
3607 | handle_parity_checks5(conf, sh, &s, disks); | 3608 | handle_parity_checks5(conf, sh, &s, disks); |
3608 | } | 3609 | } |
3609 | 3610 | ||
3610 | if (s.replacing && s.locked == 0 | 3611 | if ((s.replacing || s.syncing) && s.locked == 0 |
3611 | && !test_bit(STRIPE_INSYNC, &sh->state)) { | 3612 | && !test_bit(STRIPE_COMPUTE_RUN, &sh->state) |
3613 | && !test_bit(STRIPE_REPLACED, &sh->state)) { | ||
3612 | /* Write out to replacement devices where possible */ | 3614 | /* Write out to replacement devices where possible */ |
3613 | for (i = 0; i < conf->raid_disks; i++) | 3615 | for (i = 0; i < conf->raid_disks; i++) |
3614 | if (test_bit(R5_UPTODATE, &sh->dev[i].flags) && | 3616 | if (test_bit(R5_NeedReplace, &sh->dev[i].flags)) { |
3615 | test_bit(R5_NeedReplace, &sh->dev[i].flags)) { | 3617 | WARN_ON(!test_bit(R5_UPTODATE, &sh->dev[i].flags)); |
3616 | set_bit(R5_WantReplace, &sh->dev[i].flags); | 3618 | set_bit(R5_WantReplace, &sh->dev[i].flags); |
3617 | set_bit(R5_LOCKED, &sh->dev[i].flags); | 3619 | set_bit(R5_LOCKED, &sh->dev[i].flags); |
3618 | s.locked++; | 3620 | s.locked++; |
3619 | } | 3621 | } |
3620 | set_bit(STRIPE_INSYNC, &sh->state); | 3622 | if (s.replacing) |
3623 | set_bit(STRIPE_INSYNC, &sh->state); | ||
3624 | set_bit(STRIPE_REPLACED, &sh->state); | ||
3621 | } | 3625 | } |
3622 | if ((s.syncing || s.replacing) && s.locked == 0 && | 3626 | if ((s.syncing || s.replacing) && s.locked == 0 && |
3627 | !test_bit(STRIPE_COMPUTE_RUN, &sh->state) && | ||
3623 | test_bit(STRIPE_INSYNC, &sh->state)) { | 3628 | test_bit(STRIPE_INSYNC, &sh->state)) { |
3624 | md_done_sync(conf->mddev, STRIPE_SECTORS, 1); | 3629 | md_done_sync(conf->mddev, STRIPE_SECTORS, 1); |
3625 | clear_bit(STRIPE_SYNCING, &sh->state); | 3630 | clear_bit(STRIPE_SYNCING, &sh->state); |
diff --git a/drivers/md/raid5.h b/drivers/md/raid5.h index b0b663b119a8..70c49329ca9a 100644 --- a/drivers/md/raid5.h +++ b/drivers/md/raid5.h | |||
@@ -306,6 +306,7 @@ enum { | |||
306 | STRIPE_SYNC_REQUESTED, | 306 | STRIPE_SYNC_REQUESTED, |
307 | STRIPE_SYNCING, | 307 | STRIPE_SYNCING, |
308 | STRIPE_INSYNC, | 308 | STRIPE_INSYNC, |
309 | STRIPE_REPLACED, | ||
309 | STRIPE_PREREAD_ACTIVE, | 310 | STRIPE_PREREAD_ACTIVE, |
310 | STRIPE_DELAYED, | 311 | STRIPE_DELAYED, |
311 | STRIPE_DEGRADED, | 312 | STRIPE_DEGRADED, |
diff --git a/drivers/misc/atmel-ssc.c b/drivers/misc/atmel-ssc.c index f7b90661e321..e068a76a5f6f 100644 --- a/drivers/misc/atmel-ssc.c +++ b/drivers/misc/atmel-ssc.c | |||
@@ -66,14 +66,19 @@ EXPORT_SYMBOL(ssc_request); | |||
66 | 66 | ||
67 | void ssc_free(struct ssc_device *ssc) | 67 | void ssc_free(struct ssc_device *ssc) |
68 | { | 68 | { |
69 | bool disable_clk = true; | ||
70 | |||
69 | spin_lock(&user_lock); | 71 | spin_lock(&user_lock); |
70 | if (ssc->user) { | 72 | if (ssc->user) |
71 | ssc->user--; | 73 | ssc->user--; |
72 | clk_disable_unprepare(ssc->clk); | 74 | else { |
73 | } else { | 75 | disable_clk = false; |
74 | dev_dbg(&ssc->pdev->dev, "device already free\n"); | 76 | dev_dbg(&ssc->pdev->dev, "device already free\n"); |
75 | } | 77 | } |
76 | spin_unlock(&user_lock); | 78 | spin_unlock(&user_lock); |
79 | |||
80 | if (disable_clk) | ||
81 | clk_disable_unprepare(ssc->clk); | ||
77 | } | 82 | } |
78 | EXPORT_SYMBOL(ssc_free); | 83 | EXPORT_SYMBOL(ssc_free); |
79 | 84 | ||
diff --git a/drivers/misc/mei/hbm.c b/drivers/misc/mei/hbm.c index f9296abcf02a..6127ab64bb39 100644 --- a/drivers/misc/mei/hbm.c +++ b/drivers/misc/mei/hbm.c | |||
@@ -167,7 +167,7 @@ int mei_hbm_start_req(struct mei_device *dev) | |||
167 | 167 | ||
168 | dev->hbm_state = MEI_HBM_IDLE; | 168 | dev->hbm_state = MEI_HBM_IDLE; |
169 | if (mei_write_message(dev, mei_hdr, dev->wr_msg.data)) { | 169 | if (mei_write_message(dev, mei_hdr, dev->wr_msg.data)) { |
170 | dev_err(&dev->pdev->dev, "version message writet failed\n"); | 170 | dev_err(&dev->pdev->dev, "version message write failed\n"); |
171 | dev->dev_state = MEI_DEV_RESETTING; | 171 | dev->dev_state = MEI_DEV_RESETTING; |
172 | mei_reset(dev, 1); | 172 | mei_reset(dev, 1); |
173 | return -ENODEV; | 173 | return -ENODEV; |
diff --git a/drivers/misc/mei/hw-me.c b/drivers/misc/mei/hw-me.c index e4f8dec4dc3c..b22c7e247225 100644 --- a/drivers/misc/mei/hw-me.c +++ b/drivers/misc/mei/hw-me.c | |||
@@ -239,14 +239,18 @@ static int mei_me_hw_ready_wait(struct mei_device *dev) | |||
239 | if (mei_me_hw_is_ready(dev)) | 239 | if (mei_me_hw_is_ready(dev)) |
240 | return 0; | 240 | return 0; |
241 | 241 | ||
242 | dev->recvd_hw_ready = false; | ||
242 | mutex_unlock(&dev->device_lock); | 243 | mutex_unlock(&dev->device_lock); |
243 | err = wait_event_interruptible_timeout(dev->wait_hw_ready, | 244 | err = wait_event_interruptible_timeout(dev->wait_hw_ready, |
244 | dev->recvd_hw_ready, MEI_INTEROP_TIMEOUT); | 245 | dev->recvd_hw_ready, |
246 | mei_secs_to_jiffies(MEI_INTEROP_TIMEOUT)); | ||
245 | mutex_lock(&dev->device_lock); | 247 | mutex_lock(&dev->device_lock); |
246 | if (!err && !dev->recvd_hw_ready) { | 248 | if (!err && !dev->recvd_hw_ready) { |
249 | if (!err) | ||
250 | err = -ETIMEDOUT; | ||
247 | dev_err(&dev->pdev->dev, | 251 | dev_err(&dev->pdev->dev, |
248 | "wait hw ready failed. status = 0x%x\n", err); | 252 | "wait hw ready failed. status = %d\n", err); |
249 | return -ETIMEDOUT; | 253 | return err; |
250 | } | 254 | } |
251 | 255 | ||
252 | dev->recvd_hw_ready = false; | 256 | dev->recvd_hw_ready = false; |
@@ -483,7 +487,9 @@ irqreturn_t mei_me_irq_thread_handler(int irq, void *dev_id) | |||
483 | /* check if ME wants a reset */ | 487 | /* check if ME wants a reset */ |
484 | if (!mei_hw_is_ready(dev) && | 488 | if (!mei_hw_is_ready(dev) && |
485 | dev->dev_state != MEI_DEV_RESETTING && | 489 | dev->dev_state != MEI_DEV_RESETTING && |
486 | dev->dev_state != MEI_DEV_INITIALIZING) { | 490 | dev->dev_state != MEI_DEV_INITIALIZING && |
491 | dev->dev_state != MEI_DEV_POWER_DOWN && | ||
492 | dev->dev_state != MEI_DEV_POWER_UP) { | ||
487 | dev_dbg(&dev->pdev->dev, "FW not ready.\n"); | 493 | dev_dbg(&dev->pdev->dev, "FW not ready.\n"); |
488 | mei_reset(dev, 1); | 494 | mei_reset(dev, 1); |
489 | mutex_unlock(&dev->device_lock); | 495 | mutex_unlock(&dev->device_lock); |
diff --git a/drivers/misc/mei/init.c b/drivers/misc/mei/init.c index ed1d75203af6..e6f16f83ecde 100644 --- a/drivers/misc/mei/init.c +++ b/drivers/misc/mei/init.c | |||
@@ -148,7 +148,8 @@ void mei_reset(struct mei_device *dev, int interrupts_enabled) | |||
148 | 148 | ||
149 | dev->hbm_state = MEI_HBM_IDLE; | 149 | dev->hbm_state = MEI_HBM_IDLE; |
150 | 150 | ||
151 | if (dev->dev_state != MEI_DEV_INITIALIZING) { | 151 | if (dev->dev_state != MEI_DEV_INITIALIZING && |
152 | dev->dev_state != MEI_DEV_POWER_UP) { | ||
152 | if (dev->dev_state != MEI_DEV_DISABLED && | 153 | if (dev->dev_state != MEI_DEV_DISABLED && |
153 | dev->dev_state != MEI_DEV_POWER_DOWN) | 154 | dev->dev_state != MEI_DEV_POWER_DOWN) |
154 | dev->dev_state = MEI_DEV_RESETTING; | 155 | dev->dev_state = MEI_DEV_RESETTING; |
diff --git a/drivers/mmc/host/pxamci.c b/drivers/mmc/host/pxamci.c index 847b1996ce8e..2c5a91bb8ec3 100644 --- a/drivers/mmc/host/pxamci.c +++ b/drivers/mmc/host/pxamci.c | |||
@@ -128,7 +128,7 @@ static inline int pxamci_set_power(struct pxamci_host *host, | |||
128 | !!on ^ host->pdata->gpio_power_invert); | 128 | !!on ^ host->pdata->gpio_power_invert); |
129 | } | 129 | } |
130 | if (!host->vcc && host->pdata && host->pdata->setpower) | 130 | if (!host->vcc && host->pdata && host->pdata->setpower) |
131 | host->pdata->setpower(mmc_dev(host->mmc), vdd); | 131 | return host->pdata->setpower(mmc_dev(host->mmc), vdd); |
132 | 132 | ||
133 | return 0; | 133 | return 0; |
134 | } | 134 | } |
diff --git a/drivers/of/irq.c b/drivers/of/irq.c index a3c1c5aae6a9..1264923ade0f 100644 --- a/drivers/of/irq.c +++ b/drivers/of/irq.c | |||
@@ -345,6 +345,7 @@ int of_irq_to_resource(struct device_node *dev, int index, struct resource *r) | |||
345 | if (r && irq) { | 345 | if (r && irq) { |
346 | const char *name = NULL; | 346 | const char *name = NULL; |
347 | 347 | ||
348 | memset(r, 0, sizeof(*r)); | ||
348 | /* | 349 | /* |
349 | * Get optional "interrupts-names" property to add a name | 350 | * Get optional "interrupts-names" property to add a name |
350 | * to the resource. | 351 | * to the resource. |
@@ -482,8 +483,9 @@ void __init of_irq_init(const struct of_device_id *matches) | |||
482 | } | 483 | } |
483 | 484 | ||
484 | /* Get the next pending parent that might have children */ | 485 | /* Get the next pending parent that might have children */ |
485 | desc = list_first_entry(&intc_parent_list, typeof(*desc), list); | 486 | desc = list_first_entry_or_null(&intc_parent_list, |
486 | if (list_empty(&intc_parent_list) || !desc) { | 487 | typeof(*desc), list); |
488 | if (!desc) { | ||
487 | pr_err("of_irq_init: children remain, but no parents\n"); | 489 | pr_err("of_irq_init: children remain, but no parents\n"); |
488 | break; | 490 | break; |
489 | } | 491 | } |
diff --git a/drivers/pci/hotplug/rpadlpar_core.c b/drivers/pci/hotplug/rpadlpar_core.c index b29e20b7862f..bb7af78e4eed 100644 --- a/drivers/pci/hotplug/rpadlpar_core.c +++ b/drivers/pci/hotplug/rpadlpar_core.c | |||
@@ -388,7 +388,6 @@ int dlpar_remove_pci_slot(char *drc_name, struct device_node *dn) | |||
388 | /* Remove the EADS bridge device itself */ | 388 | /* Remove the EADS bridge device itself */ |
389 | BUG_ON(!bus->self); | 389 | BUG_ON(!bus->self); |
390 | pr_debug("PCI: Now removing bridge device %s\n", pci_name(bus->self)); | 390 | pr_debug("PCI: Now removing bridge device %s\n", pci_name(bus->self)); |
391 | eeh_remove_bus_device(bus->self, true); | ||
392 | pci_stop_and_remove_bus_device(bus->self); | 391 | pci_stop_and_remove_bus_device(bus->self); |
393 | 392 | ||
394 | return 0; | 393 | return 0; |
diff --git a/drivers/pinctrl/core.c b/drivers/pinctrl/core.c index 5b272bfd261d..2a00239661b3 100644 --- a/drivers/pinctrl/core.c +++ b/drivers/pinctrl/core.c | |||
@@ -1193,6 +1193,7 @@ void pinctrl_unregister_map(struct pinctrl_map const *map) | |||
1193 | list_for_each_entry(maps_node, &pinctrl_maps, node) { | 1193 | list_for_each_entry(maps_node, &pinctrl_maps, node) { |
1194 | if (maps_node->maps == map) { | 1194 | if (maps_node->maps == map) { |
1195 | list_del(&maps_node->node); | 1195 | list_del(&maps_node->node); |
1196 | kfree(maps_node); | ||
1196 | mutex_unlock(&pinctrl_maps_mutex); | 1197 | mutex_unlock(&pinctrl_maps_mutex); |
1197 | return; | 1198 | return; |
1198 | } | 1199 | } |
diff --git a/drivers/pinctrl/pinctrl-single.c b/drivers/pinctrl/pinctrl-single.c index 6866548fab31..7323cca440b5 100644 --- a/drivers/pinctrl/pinctrl-single.c +++ b/drivers/pinctrl/pinctrl-single.c | |||
@@ -1483,6 +1483,7 @@ static int pcs_add_gpio_func(struct device_node *node, struct pcs_device *pcs) | |||
1483 | return ret; | 1483 | return ret; |
1484 | } | 1484 | } |
1485 | 1485 | ||
1486 | #ifdef CONFIG_PM | ||
1486 | static int pinctrl_single_suspend(struct platform_device *pdev, | 1487 | static int pinctrl_single_suspend(struct platform_device *pdev, |
1487 | pm_message_t state) | 1488 | pm_message_t state) |
1488 | { | 1489 | { |
@@ -1505,6 +1506,7 @@ static int pinctrl_single_resume(struct platform_device *pdev) | |||
1505 | 1506 | ||
1506 | return pinctrl_force_default(pcs->pctl); | 1507 | return pinctrl_force_default(pcs->pctl); |
1507 | } | 1508 | } |
1509 | #endif | ||
1508 | 1510 | ||
1509 | static int pcs_probe(struct platform_device *pdev) | 1511 | static int pcs_probe(struct platform_device *pdev) |
1510 | { | 1512 | { |
diff --git a/drivers/pinctrl/sh-pfc/pfc-sh73a0.c b/drivers/pinctrl/sh-pfc/pfc-sh73a0.c index 7956df58d751..31f7d0e04aaa 100644 --- a/drivers/pinctrl/sh-pfc/pfc-sh73a0.c +++ b/drivers/pinctrl/sh-pfc/pfc-sh73a0.c | |||
@@ -3785,6 +3785,7 @@ static const struct regulator_desc sh73a0_vccq_mc0_desc = { | |||
3785 | 3785 | ||
3786 | static struct regulator_consumer_supply sh73a0_vccq_mc0_consumers[] = { | 3786 | static struct regulator_consumer_supply sh73a0_vccq_mc0_consumers[] = { |
3787 | REGULATOR_SUPPLY("vqmmc", "sh_mobile_sdhi.0"), | 3787 | REGULATOR_SUPPLY("vqmmc", "sh_mobile_sdhi.0"), |
3788 | REGULATOR_SUPPLY("vqmmc", "ee100000.sdhi"), | ||
3788 | }; | 3789 | }; |
3789 | 3790 | ||
3790 | static const struct regulator_init_data sh73a0_vccq_mc0_init_data = { | 3791 | static const struct regulator_init_data sh73a0_vccq_mc0_init_data = { |
diff --git a/drivers/pinctrl/sirf/pinctrl-atlas6.c b/drivers/pinctrl/sirf/pinctrl-atlas6.c index 1fa39a444171..867c9681763c 100644 --- a/drivers/pinctrl/sirf/pinctrl-atlas6.c +++ b/drivers/pinctrl/sirf/pinctrl-atlas6.c | |||
@@ -496,7 +496,7 @@ static const unsigned sdmmc5_pins[] = { 24, 25, 26 }; | |||
496 | static const struct sirfsoc_muxmask usp0_muxmask[] = { | 496 | static const struct sirfsoc_muxmask usp0_muxmask[] = { |
497 | { | 497 | { |
498 | .group = 1, | 498 | .group = 1, |
499 | .mask = BIT(19) | BIT(20) | BIT(21) | BIT(22), | 499 | .mask = BIT(19) | BIT(20) | BIT(21) | BIT(22) | BIT(23), |
500 | }, | 500 | }, |
501 | }; | 501 | }; |
502 | 502 | ||
@@ -507,8 +507,21 @@ static const struct sirfsoc_padmux usp0_padmux = { | |||
507 | .funcval = 0, | 507 | .funcval = 0, |
508 | }; | 508 | }; |
509 | 509 | ||
510 | static const unsigned usp0_pins[] = { 51, 52, 53, 54 }; | 510 | static const unsigned usp0_pins[] = { 51, 52, 53, 54, 55 }; |
511 | 511 | ||
512 | static const struct sirfsoc_muxmask usp0_uart_nostreamctrl_muxmask[] = { | ||
513 | { | ||
514 | .group = 1, | ||
515 | .mask = BIT(20) | BIT(21), | ||
516 | }, | ||
517 | }; | ||
518 | |||
519 | static const struct sirfsoc_padmux usp0_uart_nostreamctrl_padmux = { | ||
520 | .muxmask_counts = ARRAY_SIZE(usp0_uart_nostreamctrl_muxmask), | ||
521 | .muxmask = usp0_uart_nostreamctrl_muxmask, | ||
522 | }; | ||
523 | |||
524 | static const unsigned usp0_uart_nostreamctrl_pins[] = { 52, 53 }; | ||
512 | static const struct sirfsoc_muxmask usp1_muxmask[] = { | 525 | static const struct sirfsoc_muxmask usp1_muxmask[] = { |
513 | { | 526 | { |
514 | .group = 0, | 527 | .group = 0, |
@@ -822,6 +835,8 @@ static const struct sirfsoc_pin_group sirfsoc_pin_groups[] = { | |||
822 | SIRFSOC_PIN_GROUP("uart2grp", uart2_pins), | 835 | SIRFSOC_PIN_GROUP("uart2grp", uart2_pins), |
823 | SIRFSOC_PIN_GROUP("uart2_nostreamctrlgrp", uart2_nostreamctrl_pins), | 836 | SIRFSOC_PIN_GROUP("uart2_nostreamctrlgrp", uart2_nostreamctrl_pins), |
824 | SIRFSOC_PIN_GROUP("usp0grp", usp0_pins), | 837 | SIRFSOC_PIN_GROUP("usp0grp", usp0_pins), |
838 | SIRFSOC_PIN_GROUP("usp0_uart_nostreamctrl_grp", | ||
839 | usp0_uart_nostreamctrl_pins), | ||
825 | SIRFSOC_PIN_GROUP("usp1grp", usp1_pins), | 840 | SIRFSOC_PIN_GROUP("usp1grp", usp1_pins), |
826 | SIRFSOC_PIN_GROUP("i2c0grp", i2c0_pins), | 841 | SIRFSOC_PIN_GROUP("i2c0grp", i2c0_pins), |
827 | SIRFSOC_PIN_GROUP("i2c1grp", i2c1_pins), | 842 | SIRFSOC_PIN_GROUP("i2c1grp", i2c1_pins), |
@@ -862,6 +877,8 @@ static const char * const uart0grp[] = { "uart0grp" }; | |||
862 | static const char * const uart1grp[] = { "uart1grp" }; | 877 | static const char * const uart1grp[] = { "uart1grp" }; |
863 | static const char * const uart2grp[] = { "uart2grp" }; | 878 | static const char * const uart2grp[] = { "uart2grp" }; |
864 | static const char * const uart2_nostreamctrlgrp[] = { "uart2_nostreamctrlgrp" }; | 879 | static const char * const uart2_nostreamctrlgrp[] = { "uart2_nostreamctrlgrp" }; |
880 | static const char * const usp0_uart_nostreamctrl_grp[] = { | ||
881 | "usp0_uart_nostreamctrl_grp" }; | ||
865 | static const char * const usp0grp[] = { "usp0grp" }; | 882 | static const char * const usp0grp[] = { "usp0grp" }; |
866 | static const char * const usp1grp[] = { "usp1grp" }; | 883 | static const char * const usp1grp[] = { "usp1grp" }; |
867 | static const char * const i2c0grp[] = { "i2c0grp" }; | 884 | static const char * const i2c0grp[] = { "i2c0grp" }; |
@@ -904,6 +921,9 @@ static const struct sirfsoc_pmx_func sirfsoc_pmx_functions[] = { | |||
904 | SIRFSOC_PMX_FUNCTION("uart2", uart2grp, uart2_padmux), | 921 | SIRFSOC_PMX_FUNCTION("uart2", uart2grp, uart2_padmux), |
905 | SIRFSOC_PMX_FUNCTION("uart2_nostreamctrl", uart2_nostreamctrlgrp, uart2_nostreamctrl_padmux), | 922 | SIRFSOC_PMX_FUNCTION("uart2_nostreamctrl", uart2_nostreamctrlgrp, uart2_nostreamctrl_padmux), |
906 | SIRFSOC_PMX_FUNCTION("usp0", usp0grp, usp0_padmux), | 923 | SIRFSOC_PMX_FUNCTION("usp0", usp0grp, usp0_padmux), |
924 | SIRFSOC_PMX_FUNCTION("usp0_uart_nostreamctrl", | ||
925 | usp0_uart_nostreamctrl_grp, | ||
926 | usp0_uart_nostreamctrl_padmux), | ||
907 | SIRFSOC_PMX_FUNCTION("usp1", usp1grp, usp1_padmux), | 927 | SIRFSOC_PMX_FUNCTION("usp1", usp1grp, usp1_padmux), |
908 | SIRFSOC_PMX_FUNCTION("i2c0", i2c0grp, i2c0_padmux), | 928 | SIRFSOC_PMX_FUNCTION("i2c0", i2c0grp, i2c0_padmux), |
909 | SIRFSOC_PMX_FUNCTION("i2c1", i2c1grp, i2c1_padmux), | 929 | SIRFSOC_PMX_FUNCTION("i2c1", i2c1grp, i2c1_padmux), |
diff --git a/drivers/scsi/isci/request.c b/drivers/scsi/isci/request.c index 7b082157eb79..99d2930b18c8 100644 --- a/drivers/scsi/isci/request.c +++ b/drivers/scsi/isci/request.c | |||
@@ -185,7 +185,7 @@ static void sci_io_request_build_ssp_command_iu(struct isci_request *ireq) | |||
185 | cmd_iu->_r_c = 0; | 185 | cmd_iu->_r_c = 0; |
186 | 186 | ||
187 | sci_swab32_cpy(&cmd_iu->cdb, task->ssp_task.cmd->cmnd, | 187 | sci_swab32_cpy(&cmd_iu->cdb, task->ssp_task.cmd->cmnd, |
188 | task->ssp_task.cmd->cmd_len / sizeof(u32)); | 188 | (task->ssp_task.cmd->cmd_len+3) / sizeof(u32)); |
189 | } | 189 | } |
190 | 190 | ||
191 | static void sci_task_request_build_ssp_task_iu(struct isci_request *ireq) | 191 | static void sci_task_request_build_ssp_task_iu(struct isci_request *ireq) |
diff --git a/drivers/scsi/isci/task.c b/drivers/scsi/isci/task.c index 9bb020ac089c..0d30ca849e8f 100644 --- a/drivers/scsi/isci/task.c +++ b/drivers/scsi/isci/task.c | |||
@@ -491,6 +491,7 @@ int isci_task_abort_task(struct sas_task *task) | |||
491 | struct isci_tmf tmf; | 491 | struct isci_tmf tmf; |
492 | int ret = TMF_RESP_FUNC_FAILED; | 492 | int ret = TMF_RESP_FUNC_FAILED; |
493 | unsigned long flags; | 493 | unsigned long flags; |
494 | int target_done_already = 0; | ||
494 | 495 | ||
495 | /* Get the isci_request reference from the task. Note that | 496 | /* Get the isci_request reference from the task. Note that |
496 | * this check does not depend on the pending request list | 497 | * this check does not depend on the pending request list |
@@ -505,9 +506,11 @@ int isci_task_abort_task(struct sas_task *task) | |||
505 | /* If task is already done, the request isn't valid */ | 506 | /* If task is already done, the request isn't valid */ |
506 | if (!(task->task_state_flags & SAS_TASK_STATE_DONE) && | 507 | if (!(task->task_state_flags & SAS_TASK_STATE_DONE) && |
507 | (task->task_state_flags & SAS_TASK_AT_INITIATOR) && | 508 | (task->task_state_flags & SAS_TASK_AT_INITIATOR) && |
508 | old_request) | 509 | old_request) { |
509 | idev = isci_get_device(task->dev->lldd_dev); | 510 | idev = isci_get_device(task->dev->lldd_dev); |
510 | 511 | target_done_already = test_bit(IREQ_COMPLETE_IN_TARGET, | |
512 | &old_request->flags); | ||
513 | } | ||
511 | spin_unlock(&task->task_state_lock); | 514 | spin_unlock(&task->task_state_lock); |
512 | spin_unlock_irqrestore(&ihost->scic_lock, flags); | 515 | spin_unlock_irqrestore(&ihost->scic_lock, flags); |
513 | 516 | ||
@@ -561,7 +564,7 @@ int isci_task_abort_task(struct sas_task *task) | |||
561 | 564 | ||
562 | if (task->task_proto == SAS_PROTOCOL_SMP || | 565 | if (task->task_proto == SAS_PROTOCOL_SMP || |
563 | sas_protocol_ata(task->task_proto) || | 566 | sas_protocol_ata(task->task_proto) || |
564 | test_bit(IREQ_COMPLETE_IN_TARGET, &old_request->flags) || | 567 | target_done_already || |
565 | test_bit(IDEV_GONE, &idev->flags)) { | 568 | test_bit(IDEV_GONE, &idev->flags)) { |
566 | 569 | ||
567 | spin_unlock_irqrestore(&ihost->scic_lock, flags); | 570 | spin_unlock_irqrestore(&ihost->scic_lock, flags); |
diff --git a/drivers/scsi/mvsas/mv_sas.c b/drivers/scsi/mvsas/mv_sas.c index f14665a6293d..6b1b4e91e53f 100644 --- a/drivers/scsi/mvsas/mv_sas.c +++ b/drivers/scsi/mvsas/mv_sas.c | |||
@@ -1857,11 +1857,16 @@ int mvs_slot_complete(struct mvs_info *mvi, u32 rx_desc, u32 flags) | |||
1857 | goto out; | 1857 | goto out; |
1858 | } | 1858 | } |
1859 | 1859 | ||
1860 | /* error info record present */ | 1860 | /* |
1861 | if (unlikely((rx_desc & RXQ_ERR) && (*(u64 *) slot->response))) { | 1861 | * error info record present; slot->response is 32 bit aligned but may |
1862 | * not be 64 bit aligned, so check for zero in two 32 bit reads | ||
1863 | */ | ||
1864 | if (unlikely((rx_desc & RXQ_ERR) | ||
1865 | && (*((u32 *)slot->response) | ||
1866 | || *(((u32 *)slot->response) + 1)))) { | ||
1862 | mv_dprintk("port %d slot %d rx_desc %X has error info" | 1867 | mv_dprintk("port %d slot %d rx_desc %X has error info" |
1863 | "%016llX.\n", slot->port->sas_port.id, slot_idx, | 1868 | "%016llX.\n", slot->port->sas_port.id, slot_idx, |
1864 | rx_desc, (u64)(*(u64 *)slot->response)); | 1869 | rx_desc, get_unaligned_le64(slot->response)); |
1865 | tstat->stat = mvs_slot_err(mvi, task, slot_idx); | 1870 | tstat->stat = mvs_slot_err(mvi, task, slot_idx); |
1866 | tstat->resp = SAS_TASK_COMPLETE; | 1871 | tstat->resp = SAS_TASK_COMPLETE; |
1867 | goto out; | 1872 | goto out; |
diff --git a/drivers/scsi/mvsas/mv_sas.h b/drivers/scsi/mvsas/mv_sas.h index 60e2fb7f2dca..d6b19dc80bee 100644 --- a/drivers/scsi/mvsas/mv_sas.h +++ b/drivers/scsi/mvsas/mv_sas.h | |||
@@ -39,6 +39,7 @@ | |||
39 | #include <linux/irq.h> | 39 | #include <linux/irq.h> |
40 | #include <linux/slab.h> | 40 | #include <linux/slab.h> |
41 | #include <linux/vmalloc.h> | 41 | #include <linux/vmalloc.h> |
42 | #include <asm/unaligned.h> | ||
42 | #include <scsi/libsas.h> | 43 | #include <scsi/libsas.h> |
43 | #include <scsi/scsi.h> | 44 | #include <scsi/scsi.h> |
44 | #include <scsi/scsi_tcq.h> | 45 | #include <scsi/scsi_tcq.h> |
diff --git a/drivers/scsi/qla2xxx/qla_iocb.c b/drivers/scsi/qla2xxx/qla_iocb.c index 42ef481db942..ef0a5481b9dd 100644 --- a/drivers/scsi/qla2xxx/qla_iocb.c +++ b/drivers/scsi/qla2xxx/qla_iocb.c | |||
@@ -419,6 +419,8 @@ qla2x00_start_scsi(srb_t *sp) | |||
419 | __constant_cpu_to_le16(CF_SIMPLE_TAG); | 419 | __constant_cpu_to_le16(CF_SIMPLE_TAG); |
420 | break; | 420 | break; |
421 | } | 421 | } |
422 | } else { | ||
423 | cmd_pkt->control_flags = __constant_cpu_to_le16(CF_SIMPLE_TAG); | ||
422 | } | 424 | } |
423 | 425 | ||
424 | /* Load SCSI command packet. */ | 426 | /* Load SCSI command packet. */ |
@@ -1307,11 +1309,11 @@ qla24xx_build_scsi_crc_2_iocbs(srb_t *sp, struct cmd_type_crc_2 *cmd_pkt, | |||
1307 | fcp_cmnd->task_attribute = TSK_ORDERED; | 1309 | fcp_cmnd->task_attribute = TSK_ORDERED; |
1308 | break; | 1310 | break; |
1309 | default: | 1311 | default: |
1310 | fcp_cmnd->task_attribute = 0; | 1312 | fcp_cmnd->task_attribute = TSK_SIMPLE; |
1311 | break; | 1313 | break; |
1312 | } | 1314 | } |
1313 | } else { | 1315 | } else { |
1314 | fcp_cmnd->task_attribute = 0; | 1316 | fcp_cmnd->task_attribute = TSK_SIMPLE; |
1315 | } | 1317 | } |
1316 | 1318 | ||
1317 | cmd_pkt->fcp_rsp_dseg_len = 0; /* Let response come in status iocb */ | 1319 | cmd_pkt->fcp_rsp_dseg_len = 0; /* Let response come in status iocb */ |
@@ -1525,7 +1527,12 @@ qla24xx_start_scsi(srb_t *sp) | |||
1525 | case ORDERED_QUEUE_TAG: | 1527 | case ORDERED_QUEUE_TAG: |
1526 | cmd_pkt->task = TSK_ORDERED; | 1528 | cmd_pkt->task = TSK_ORDERED; |
1527 | break; | 1529 | break; |
1530 | default: | ||
1531 | cmd_pkt->task = TSK_SIMPLE; | ||
1532 | break; | ||
1528 | } | 1533 | } |
1534 | } else { | ||
1535 | cmd_pkt->task = TSK_SIMPLE; | ||
1529 | } | 1536 | } |
1530 | 1537 | ||
1531 | /* Load SCSI command packet. */ | 1538 | /* Load SCSI command packet. */ |
diff --git a/drivers/scsi/sd.c b/drivers/scsi/sd.c index 80f39b8b0223..86fcf2c313ad 100644 --- a/drivers/scsi/sd.c +++ b/drivers/scsi/sd.c | |||
@@ -838,10 +838,17 @@ static int scsi_setup_flush_cmnd(struct scsi_device *sdp, struct request *rq) | |||
838 | 838 | ||
839 | static void sd_unprep_fn(struct request_queue *q, struct request *rq) | 839 | static void sd_unprep_fn(struct request_queue *q, struct request *rq) |
840 | { | 840 | { |
841 | struct scsi_cmnd *SCpnt = rq->special; | ||
842 | |||
841 | if (rq->cmd_flags & REQ_DISCARD) { | 843 | if (rq->cmd_flags & REQ_DISCARD) { |
842 | free_page((unsigned long)rq->buffer); | 844 | free_page((unsigned long)rq->buffer); |
843 | rq->buffer = NULL; | 845 | rq->buffer = NULL; |
844 | } | 846 | } |
847 | if (SCpnt->cmnd != rq->cmd) { | ||
848 | mempool_free(SCpnt->cmnd, sd_cdb_pool); | ||
849 | SCpnt->cmnd = NULL; | ||
850 | SCpnt->cmd_len = 0; | ||
851 | } | ||
845 | } | 852 | } |
846 | 853 | ||
847 | /** | 854 | /** |
@@ -1720,21 +1727,6 @@ static int sd_done(struct scsi_cmnd *SCpnt) | |||
1720 | if (rq_data_dir(SCpnt->request) == READ && scsi_prot_sg_count(SCpnt)) | 1727 | if (rq_data_dir(SCpnt->request) == READ && scsi_prot_sg_count(SCpnt)) |
1721 | sd_dif_complete(SCpnt, good_bytes); | 1728 | sd_dif_complete(SCpnt, good_bytes); |
1722 | 1729 | ||
1723 | if (scsi_host_dif_capable(sdkp->device->host, sdkp->protection_type) | ||
1724 | == SD_DIF_TYPE2_PROTECTION && SCpnt->cmnd != SCpnt->request->cmd) { | ||
1725 | |||
1726 | /* We have to print a failed command here as the | ||
1727 | * extended CDB gets freed before scsi_io_completion() | ||
1728 | * is called. | ||
1729 | */ | ||
1730 | if (result) | ||
1731 | scsi_print_command(SCpnt); | ||
1732 | |||
1733 | mempool_free(SCpnt->cmnd, sd_cdb_pool); | ||
1734 | SCpnt->cmnd = NULL; | ||
1735 | SCpnt->cmd_len = 0; | ||
1736 | } | ||
1737 | |||
1738 | return good_bytes; | 1730 | return good_bytes; |
1739 | } | 1731 | } |
1740 | 1732 | ||
diff --git a/drivers/staging/android/logger.c b/drivers/staging/android/logger.c index 080abf2faf97..a8c344422a77 100644 --- a/drivers/staging/android/logger.c +++ b/drivers/staging/android/logger.c | |||
@@ -469,7 +469,7 @@ static ssize_t logger_aio_write(struct kiocb *iocb, const struct iovec *iov, | |||
469 | unsigned long nr_segs, loff_t ppos) | 469 | unsigned long nr_segs, loff_t ppos) |
470 | { | 470 | { |
471 | struct logger_log *log = file_get_log(iocb->ki_filp); | 471 | struct logger_log *log = file_get_log(iocb->ki_filp); |
472 | size_t orig = log->w_off; | 472 | size_t orig; |
473 | struct logger_entry header; | 473 | struct logger_entry header; |
474 | struct timespec now; | 474 | struct timespec now; |
475 | ssize_t ret = 0; | 475 | ssize_t ret = 0; |
@@ -490,6 +490,8 @@ static ssize_t logger_aio_write(struct kiocb *iocb, const struct iovec *iov, | |||
490 | 490 | ||
491 | mutex_lock(&log->mutex); | 491 | mutex_lock(&log->mutex); |
492 | 492 | ||
493 | orig = log->w_off; | ||
494 | |||
493 | /* | 495 | /* |
494 | * Fix up any readers, pulling them forward to the first readable | 496 | * Fix up any readers, pulling them forward to the first readable |
495 | * entry after (what will be) the new write offset. We do this now | 497 | * entry after (what will be) the new write offset. We do this now |
diff --git a/drivers/staging/comedi/TODO b/drivers/staging/comedi/TODO index b10f739b7e3e..fa8da9aada30 100644 --- a/drivers/staging/comedi/TODO +++ b/drivers/staging/comedi/TODO | |||
@@ -9,4 +9,4 @@ TODO: | |||
9 | Please send patches to Greg Kroah-Hartman <greg@kroah.com> and | 9 | Please send patches to Greg Kroah-Hartman <greg@kroah.com> and |
10 | copy: | 10 | copy: |
11 | Ian Abbott <abbotti@mev.co.uk> | 11 | Ian Abbott <abbotti@mev.co.uk> |
12 | Frank Mori Hess <fmhess@users.sourceforge.net> | 12 | H Hartley Sweeten <hsweeten@visionengravers.com> |
diff --git a/drivers/staging/comedi/comedi_fops.c b/drivers/staging/comedi/comedi_fops.c index 2f2ff8f08783..6d50e3033228 100644 --- a/drivers/staging/comedi/comedi_fops.c +++ b/drivers/staging/comedi/comedi_fops.c | |||
@@ -1414,22 +1414,19 @@ static int do_cmd_ioctl(struct comedi_device *dev, | |||
1414 | DPRINTK("subdevice busy\n"); | 1414 | DPRINTK("subdevice busy\n"); |
1415 | return -EBUSY; | 1415 | return -EBUSY; |
1416 | } | 1416 | } |
1417 | s->busy = file; | ||
1418 | 1417 | ||
1419 | /* make sure channel/gain list isn't too long */ | 1418 | /* make sure channel/gain list isn't too long */ |
1420 | if (cmd.chanlist_len > s->len_chanlist) { | 1419 | if (cmd.chanlist_len > s->len_chanlist) { |
1421 | DPRINTK("channel/gain list too long %u > %d\n", | 1420 | DPRINTK("channel/gain list too long %u > %d\n", |
1422 | cmd.chanlist_len, s->len_chanlist); | 1421 | cmd.chanlist_len, s->len_chanlist); |
1423 | ret = -EINVAL; | 1422 | return -EINVAL; |
1424 | goto cleanup; | ||
1425 | } | 1423 | } |
1426 | 1424 | ||
1427 | /* make sure channel/gain list isn't too short */ | 1425 | /* make sure channel/gain list isn't too short */ |
1428 | if (cmd.chanlist_len < 1) { | 1426 | if (cmd.chanlist_len < 1) { |
1429 | DPRINTK("channel/gain list too short %u < 1\n", | 1427 | DPRINTK("channel/gain list too short %u < 1\n", |
1430 | cmd.chanlist_len); | 1428 | cmd.chanlist_len); |
1431 | ret = -EINVAL; | 1429 | return -EINVAL; |
1432 | goto cleanup; | ||
1433 | } | 1430 | } |
1434 | 1431 | ||
1435 | async->cmd = cmd; | 1432 | async->cmd = cmd; |
@@ -1439,8 +1436,7 @@ static int do_cmd_ioctl(struct comedi_device *dev, | |||
1439 | kmalloc(async->cmd.chanlist_len * sizeof(int), GFP_KERNEL); | 1436 | kmalloc(async->cmd.chanlist_len * sizeof(int), GFP_KERNEL); |
1440 | if (!async->cmd.chanlist) { | 1437 | if (!async->cmd.chanlist) { |
1441 | DPRINTK("allocation failed\n"); | 1438 | DPRINTK("allocation failed\n"); |
1442 | ret = -ENOMEM; | 1439 | return -ENOMEM; |
1443 | goto cleanup; | ||
1444 | } | 1440 | } |
1445 | 1441 | ||
1446 | if (copy_from_user(async->cmd.chanlist, user_chanlist, | 1442 | if (copy_from_user(async->cmd.chanlist, user_chanlist, |
@@ -1492,6 +1488,9 @@ static int do_cmd_ioctl(struct comedi_device *dev, | |||
1492 | 1488 | ||
1493 | comedi_set_subdevice_runflags(s, ~0, SRF_USER | SRF_RUNNING); | 1489 | comedi_set_subdevice_runflags(s, ~0, SRF_USER | SRF_RUNNING); |
1494 | 1490 | ||
1491 | /* set s->busy _after_ setting SRF_RUNNING flag to avoid race with | ||
1492 | * comedi_read() or comedi_write() */ | ||
1493 | s->busy = file; | ||
1495 | ret = s->do_cmd(dev, s); | 1494 | ret = s->do_cmd(dev, s); |
1496 | if (ret == 0) | 1495 | if (ret == 0) |
1497 | return 0; | 1496 | return 0; |
@@ -1706,6 +1705,7 @@ static int do_cancel_ioctl(struct comedi_device *dev, unsigned int arg, | |||
1706 | void *file) | 1705 | void *file) |
1707 | { | 1706 | { |
1708 | struct comedi_subdevice *s; | 1707 | struct comedi_subdevice *s; |
1708 | int ret; | ||
1709 | 1709 | ||
1710 | if (arg >= dev->n_subdevices) | 1710 | if (arg >= dev->n_subdevices) |
1711 | return -EINVAL; | 1711 | return -EINVAL; |
@@ -1722,7 +1722,11 @@ static int do_cancel_ioctl(struct comedi_device *dev, unsigned int arg, | |||
1722 | if (s->busy != file) | 1722 | if (s->busy != file) |
1723 | return -EBUSY; | 1723 | return -EBUSY; |
1724 | 1724 | ||
1725 | return do_cancel(dev, s); | 1725 | ret = do_cancel(dev, s); |
1726 | if (comedi_get_subdevice_runflags(s) & SRF_USER) | ||
1727 | wake_up_interruptible(&s->async->wait_head); | ||
1728 | |||
1729 | return ret; | ||
1726 | } | 1730 | } |
1727 | 1731 | ||
1728 | /* | 1732 | /* |
@@ -2054,11 +2058,13 @@ static ssize_t comedi_write(struct file *file, const char __user *buf, | |||
2054 | 2058 | ||
2055 | if (!comedi_is_subdevice_running(s)) { | 2059 | if (!comedi_is_subdevice_running(s)) { |
2056 | if (count == 0) { | 2060 | if (count == 0) { |
2061 | mutex_lock(&dev->mutex); | ||
2057 | if (comedi_is_subdevice_in_error(s)) | 2062 | if (comedi_is_subdevice_in_error(s)) |
2058 | retval = -EPIPE; | 2063 | retval = -EPIPE; |
2059 | else | 2064 | else |
2060 | retval = 0; | 2065 | retval = 0; |
2061 | do_become_nonbusy(dev, s); | 2066 | do_become_nonbusy(dev, s); |
2067 | mutex_unlock(&dev->mutex); | ||
2062 | } | 2068 | } |
2063 | break; | 2069 | break; |
2064 | } | 2070 | } |
@@ -2157,11 +2163,13 @@ static ssize_t comedi_read(struct file *file, char __user *buf, size_t nbytes, | |||
2157 | 2163 | ||
2158 | if (n == 0) { | 2164 | if (n == 0) { |
2159 | if (!comedi_is_subdevice_running(s)) { | 2165 | if (!comedi_is_subdevice_running(s)) { |
2166 | mutex_lock(&dev->mutex); | ||
2160 | do_become_nonbusy(dev, s); | 2167 | do_become_nonbusy(dev, s); |
2161 | if (comedi_is_subdevice_in_error(s)) | 2168 | if (comedi_is_subdevice_in_error(s)) |
2162 | retval = -EPIPE; | 2169 | retval = -EPIPE; |
2163 | else | 2170 | else |
2164 | retval = 0; | 2171 | retval = 0; |
2172 | mutex_unlock(&dev->mutex); | ||
2165 | break; | 2173 | break; |
2166 | } | 2174 | } |
2167 | if (file->f_flags & O_NONBLOCK) { | 2175 | if (file->f_flags & O_NONBLOCK) { |
@@ -2199,9 +2207,11 @@ static ssize_t comedi_read(struct file *file, char __user *buf, size_t nbytes, | |||
2199 | buf += n; | 2207 | buf += n; |
2200 | break; /* makes device work like a pipe */ | 2208 | break; /* makes device work like a pipe */ |
2201 | } | 2209 | } |
2202 | if (comedi_is_subdevice_idle(s) && | 2210 | if (comedi_is_subdevice_idle(s)) { |
2203 | async->buf_read_count - async->buf_write_count == 0) { | 2211 | mutex_lock(&dev->mutex); |
2204 | do_become_nonbusy(dev, s); | 2212 | if (async->buf_read_count - async->buf_write_count == 0) |
2213 | do_become_nonbusy(dev, s); | ||
2214 | mutex_unlock(&dev->mutex); | ||
2205 | } | 2215 | } |
2206 | set_current_state(TASK_RUNNING); | 2216 | set_current_state(TASK_RUNNING); |
2207 | remove_wait_queue(&async->wait_head, &wait); | 2217 | remove_wait_queue(&async->wait_head, &wait); |
diff --git a/drivers/staging/frontier/alphatrack.c b/drivers/staging/frontier/alphatrack.c index 5590ebf1da15..817f837b240d 100644 --- a/drivers/staging/frontier/alphatrack.c +++ b/drivers/staging/frontier/alphatrack.c | |||
@@ -827,11 +827,11 @@ static void usb_alphatrack_disconnect(struct usb_interface *intf) | |||
827 | mutex_unlock(&dev->mtx); | 827 | mutex_unlock(&dev->mtx); |
828 | usb_alphatrack_delete(dev); | 828 | usb_alphatrack_delete(dev); |
829 | } else { | 829 | } else { |
830 | atomic_set(&dev->writes_pending, 0); | ||
830 | dev->intf = NULL; | 831 | dev->intf = NULL; |
831 | mutex_unlock(&dev->mtx); | 832 | mutex_unlock(&dev->mtx); |
832 | } | 833 | } |
833 | 834 | ||
834 | atomic_set(&dev->writes_pending, 0); | ||
835 | mutex_unlock(&disconnect_mutex); | 835 | mutex_unlock(&disconnect_mutex); |
836 | 836 | ||
837 | dev_info(&intf->dev, "Alphatrack Surface #%d now disconnected\n", | 837 | dev_info(&intf->dev, "Alphatrack Surface #%d now disconnected\n", |
diff --git a/drivers/staging/gdm72xx/gdm_qos.c b/drivers/staging/gdm72xx/gdm_qos.c index b795353e8348..cc3692439a5c 100644 --- a/drivers/staging/gdm72xx/gdm_qos.c +++ b/drivers/staging/gdm72xx/gdm_qos.c | |||
@@ -250,8 +250,8 @@ static void send_qos_list(struct nic *nic, struct list_head *head) | |||
250 | 250 | ||
251 | list_for_each_entry_safe(entry, n, head, list) { | 251 | list_for_each_entry_safe(entry, n, head, list) { |
252 | list_del(&entry->list); | 252 | list_del(&entry->list); |
253 | free_qos_entry(entry); | ||
254 | gdm_wimax_send_tx(entry->skb, entry->dev); | 253 | gdm_wimax_send_tx(entry->skb, entry->dev); |
254 | free_qos_entry(entry); | ||
255 | } | 255 | } |
256 | } | 256 | } |
257 | 257 | ||
diff --git a/drivers/staging/imx-drm/Kconfig b/drivers/staging/imx-drm/Kconfig index 22339059837f..bd0f2fd01db4 100644 --- a/drivers/staging/imx-drm/Kconfig +++ b/drivers/staging/imx-drm/Kconfig | |||
@@ -33,7 +33,6 @@ config DRM_IMX_TVE | |||
33 | config DRM_IMX_LDB | 33 | config DRM_IMX_LDB |
34 | tristate "Support for LVDS displays" | 34 | tristate "Support for LVDS displays" |
35 | depends on DRM_IMX | 35 | depends on DRM_IMX |
36 | select OF_VIDEOMODE | ||
37 | help | 36 | help |
38 | Choose this to enable the internal LVDS Display Bridge (LDB) | 37 | Choose this to enable the internal LVDS Display Bridge (LDB) |
39 | found on i.MX53 and i.MX6 processors. | 38 | found on i.MX53 and i.MX6 processors. |
diff --git a/drivers/staging/tidspbridge/pmgr/dbll.c b/drivers/staging/tidspbridge/pmgr/dbll.c index c191ae203565..41e88abe47af 100644 --- a/drivers/staging/tidspbridge/pmgr/dbll.c +++ b/drivers/staging/tidspbridge/pmgr/dbll.c | |||
@@ -1120,8 +1120,11 @@ static int dbll_rmm_alloc(struct dynamic_loader_allocate *this, | |||
1120 | or DYN_EXTERNAL, then mem granularity information is present | 1120 | or DYN_EXTERNAL, then mem granularity information is present |
1121 | within the section name - only process if there are at least three | 1121 | within the section name - only process if there are at least three |
1122 | tokens within the section name (just a minor optimization) */ | 1122 | tokens within the section name (just a minor optimization) */ |
1123 | if (count >= 3) | 1123 | if (count >= 3) { |
1124 | strict_strtol(sz_last_token, 10, (long *)&req); | 1124 | status = kstrtos32(sz_last_token, 10, &req); |
1125 | if (status) | ||
1126 | goto func_cont; | ||
1127 | } | ||
1125 | 1128 | ||
1126 | if ((req == 0) || (req == 1)) { | 1129 | if ((req == 0) || (req == 1)) { |
1127 | if (strcmp(sz_sec_last_token, "DYN_DARAM") == 0) { | 1130 | if (strcmp(sz_sec_last_token, "DYN_DARAM") == 0) { |
diff --git a/drivers/staging/zram/zram_drv.c b/drivers/staging/zram/zram_drv.c index 82c7202fd5cc..e77fb6ea40c9 100644 --- a/drivers/staging/zram/zram_drv.c +++ b/drivers/staging/zram/zram_drv.c | |||
@@ -527,8 +527,11 @@ static void zram_reset_device(struct zram *zram) | |||
527 | size_t index; | 527 | size_t index; |
528 | struct zram_meta *meta; | 528 | struct zram_meta *meta; |
529 | 529 | ||
530 | if (!zram->init_done) | 530 | down_write(&zram->init_lock); |
531 | if (!zram->init_done) { | ||
532 | up_write(&zram->init_lock); | ||
531 | return; | 533 | return; |
534 | } | ||
532 | 535 | ||
533 | meta = zram->meta; | 536 | meta = zram->meta; |
534 | zram->init_done = 0; | 537 | zram->init_done = 0; |
@@ -549,6 +552,7 @@ static void zram_reset_device(struct zram *zram) | |||
549 | 552 | ||
550 | zram->disksize = 0; | 553 | zram->disksize = 0; |
551 | set_capacity(zram->disk, 0); | 554 | set_capacity(zram->disk, 0); |
555 | up_write(&zram->init_lock); | ||
552 | } | 556 | } |
553 | 557 | ||
554 | static void zram_init_device(struct zram *zram, struct zram_meta *meta) | 558 | static void zram_init_device(struct zram *zram, struct zram_meta *meta) |
diff --git a/drivers/thermal/x86_pkg_temp_thermal.c b/drivers/thermal/x86_pkg_temp_thermal.c index 5de56f671a9d..f36950e4134f 100644 --- a/drivers/thermal/x86_pkg_temp_thermal.c +++ b/drivers/thermal/x86_pkg_temp_thermal.c | |||
@@ -54,6 +54,8 @@ MODULE_PARM_DESC(notify_delay_ms, | |||
54 | * is some wrong values returned by cpuid for number of thresholds. | 54 | * is some wrong values returned by cpuid for number of thresholds. |
55 | */ | 55 | */ |
56 | #define MAX_NUMBER_OF_TRIPS 2 | 56 | #define MAX_NUMBER_OF_TRIPS 2 |
57 | /* Limit number of package temp zones */ | ||
58 | #define MAX_PKG_TEMP_ZONE_IDS 256 | ||
57 | 59 | ||
58 | struct phy_dev_entry { | 60 | struct phy_dev_entry { |
59 | struct list_head list; | 61 | struct list_head list; |
@@ -394,12 +396,16 @@ static int pkg_temp_thermal_device_add(unsigned int cpu) | |||
394 | char buffer[30]; | 396 | char buffer[30]; |
395 | int thres_count; | 397 | int thres_count; |
396 | u32 eax, ebx, ecx, edx; | 398 | u32 eax, ebx, ecx, edx; |
399 | u8 *temp; | ||
397 | 400 | ||
398 | cpuid(6, &eax, &ebx, &ecx, &edx); | 401 | cpuid(6, &eax, &ebx, &ecx, &edx); |
399 | thres_count = ebx & 0x07; | 402 | thres_count = ebx & 0x07; |
400 | if (!thres_count) | 403 | if (!thres_count) |
401 | return -ENODEV; | 404 | return -ENODEV; |
402 | 405 | ||
406 | if (topology_physical_package_id(cpu) > MAX_PKG_TEMP_ZONE_IDS) | ||
407 | return -ENODEV; | ||
408 | |||
403 | thres_count = clamp_val(thres_count, 0, MAX_NUMBER_OF_TRIPS); | 409 | thres_count = clamp_val(thres_count, 0, MAX_NUMBER_OF_TRIPS); |
404 | 410 | ||
405 | err = get_tj_max(cpu, &tj_max); | 411 | err = get_tj_max(cpu, &tj_max); |
@@ -417,13 +423,14 @@ static int pkg_temp_thermal_device_add(unsigned int cpu) | |||
417 | spin_lock(&pkg_work_lock); | 423 | spin_lock(&pkg_work_lock); |
418 | if (topology_physical_package_id(cpu) > max_phy_id) | 424 | if (topology_physical_package_id(cpu) > max_phy_id) |
419 | max_phy_id = topology_physical_package_id(cpu); | 425 | max_phy_id = topology_physical_package_id(cpu); |
420 | pkg_work_scheduled = krealloc(pkg_work_scheduled, | 426 | temp = krealloc(pkg_work_scheduled, |
421 | (max_phy_id+1) * sizeof(u8), GFP_ATOMIC); | 427 | (max_phy_id+1) * sizeof(u8), GFP_ATOMIC); |
422 | if (!pkg_work_scheduled) { | 428 | if (!temp) { |
423 | spin_unlock(&pkg_work_lock); | 429 | spin_unlock(&pkg_work_lock); |
424 | err = -ENOMEM; | 430 | err = -ENOMEM; |
425 | goto err_ret_free; | 431 | goto err_ret_free; |
426 | } | 432 | } |
433 | pkg_work_scheduled = temp; | ||
427 | pkg_work_scheduled[topology_physical_package_id(cpu)] = 0; | 434 | pkg_work_scheduled[topology_physical_package_id(cpu)] = 0; |
428 | spin_unlock(&pkg_work_lock); | 435 | spin_unlock(&pkg_work_lock); |
429 | 436 | ||
@@ -511,7 +518,7 @@ static int get_core_online(unsigned int cpu) | |||
511 | 518 | ||
512 | /* Check if there is already an instance for this package */ | 519 | /* Check if there is already an instance for this package */ |
513 | if (!phdev) { | 520 | if (!phdev) { |
514 | if (!cpu_has(c, X86_FEATURE_DTHERM) && | 521 | if (!cpu_has(c, X86_FEATURE_DTHERM) || |
515 | !cpu_has(c, X86_FEATURE_PTS)) | 522 | !cpu_has(c, X86_FEATURE_PTS)) |
516 | return -ENODEV; | 523 | return -ENODEV; |
517 | if (pkg_temp_thermal_device_add(cpu)) | 524 | if (pkg_temp_thermal_device_add(cpu)) |
@@ -562,7 +569,7 @@ static struct notifier_block pkg_temp_thermal_notifier __refdata = { | |||
562 | }; | 569 | }; |
563 | 570 | ||
564 | static const struct x86_cpu_id __initconst pkg_temp_thermal_ids[] = { | 571 | static const struct x86_cpu_id __initconst pkg_temp_thermal_ids[] = { |
565 | { X86_VENDOR_INTEL, X86_FAMILY_ANY, X86_MODEL_ANY, X86_FEATURE_DTHERM }, | 572 | { X86_VENDOR_INTEL, X86_FAMILY_ANY, X86_MODEL_ANY, X86_FEATURE_PTS }, |
566 | {} | 573 | {} |
567 | }; | 574 | }; |
568 | MODULE_DEVICE_TABLE(x86cpu, pkg_temp_thermal_ids); | 575 | MODULE_DEVICE_TABLE(x86cpu, pkg_temp_thermal_ids); |
@@ -592,7 +599,6 @@ static int __init pkg_temp_thermal_init(void) | |||
592 | return 0; | 599 | return 0; |
593 | 600 | ||
594 | err_ret: | 601 | err_ret: |
595 | get_online_cpus(); | ||
596 | for_each_online_cpu(i) | 602 | for_each_online_cpu(i) |
597 | put_core_offline(i); | 603 | put_core_offline(i); |
598 | put_online_cpus(); | 604 | put_online_cpus(); |
diff --git a/drivers/tty/serial/8250/8250_early.c b/drivers/tty/serial/8250/8250_early.c index 721904f8efa9..946ddd2b3a54 100644 --- a/drivers/tty/serial/8250/8250_early.c +++ b/drivers/tty/serial/8250/8250_early.c | |||
@@ -193,7 +193,8 @@ static int __init parse_options(struct early_serial8250_device *device, | |||
193 | if (options) { | 193 | if (options) { |
194 | options++; | 194 | options++; |
195 | device->baud = simple_strtoul(options, NULL, 0); | 195 | device->baud = simple_strtoul(options, NULL, 0); |
196 | length = min(strcspn(options, " "), sizeof(device->options)); | 196 | length = min(strcspn(options, " ") + 1, |
197 | sizeof(device->options)); | ||
197 | strlcpy(device->options, options, length); | 198 | strlcpy(device->options, options, length); |
198 | } else { | 199 | } else { |
199 | device->baud = probe_baud(port); | 200 | device->baud = probe_baud(port); |
diff --git a/drivers/tty/serial/Kconfig b/drivers/tty/serial/Kconfig index 5e3d68917ffe..1456673bcca0 100644 --- a/drivers/tty/serial/Kconfig +++ b/drivers/tty/serial/Kconfig | |||
@@ -277,7 +277,7 @@ config SERIAL_TEGRA | |||
277 | select SERIAL_CORE | 277 | select SERIAL_CORE |
278 | help | 278 | help |
279 | Support for the on-chip UARTs on the NVIDIA Tegra series SOCs | 279 | Support for the on-chip UARTs on the NVIDIA Tegra series SOCs |
280 | providing /dev/ttyHS0, 1, 2, 3 and 4 (note, some machines may not | 280 | providing /dev/ttyTHS0, 1, 2, 3 and 4 (note, some machines may not |
281 | provide all of these ports, depending on how the serial port | 281 | provide all of these ports, depending on how the serial port |
282 | are enabled). This driver uses the APB DMA to achieve higher baudrate | 282 | are enabled). This driver uses the APB DMA to achieve higher baudrate |
283 | and better performance. | 283 | and better performance. |
diff --git a/drivers/tty/synclinkmp.c b/drivers/tty/synclinkmp.c index ff171384ea52..dc6e96996ead 100644 --- a/drivers/tty/synclinkmp.c +++ b/drivers/tty/synclinkmp.c | |||
@@ -3478,7 +3478,7 @@ static int alloc_buf_list(SLMP_INFO *info) | |||
3478 | for ( i = 0; i < info->rx_buf_count; i++ ) { | 3478 | for ( i = 0; i < info->rx_buf_count; i++ ) { |
3479 | /* calculate and store physical address of this buffer entry */ | 3479 | /* calculate and store physical address of this buffer entry */ |
3480 | info->rx_buf_list_ex[i].phys_entry = | 3480 | info->rx_buf_list_ex[i].phys_entry = |
3481 | info->buffer_list_phys + (i * sizeof(SCABUFSIZE)); | 3481 | info->buffer_list_phys + (i * SCABUFSIZE); |
3482 | 3482 | ||
3483 | /* calculate and store physical address of */ | 3483 | /* calculate and store physical address of */ |
3484 | /* next entry in cirular list of entries */ | 3484 | /* next entry in cirular list of entries */ |
diff --git a/drivers/usb/core/hub.c b/drivers/usb/core/hub.c index 4191db32f12c..4a8a1d68002c 100644 --- a/drivers/usb/core/hub.c +++ b/drivers/usb/core/hub.c | |||
@@ -668,6 +668,15 @@ resubmit: | |||
668 | static inline int | 668 | static inline int |
669 | hub_clear_tt_buffer (struct usb_device *hdev, u16 devinfo, u16 tt) | 669 | hub_clear_tt_buffer (struct usb_device *hdev, u16 devinfo, u16 tt) |
670 | { | 670 | { |
671 | /* Need to clear both directions for control ep */ | ||
672 | if (((devinfo >> 11) & USB_ENDPOINT_XFERTYPE_MASK) == | ||
673 | USB_ENDPOINT_XFER_CONTROL) { | ||
674 | int status = usb_control_msg(hdev, usb_sndctrlpipe(hdev, 0), | ||
675 | HUB_CLEAR_TT_BUFFER, USB_RT_PORT, | ||
676 | devinfo ^ 0x8000, tt, NULL, 0, 1000); | ||
677 | if (status) | ||
678 | return status; | ||
679 | } | ||
671 | return usb_control_msg(hdev, usb_sndctrlpipe(hdev, 0), | 680 | return usb_control_msg(hdev, usb_sndctrlpipe(hdev, 0), |
672 | HUB_CLEAR_TT_BUFFER, USB_RT_PORT, devinfo, | 681 | HUB_CLEAR_TT_BUFFER, USB_RT_PORT, devinfo, |
673 | tt, NULL, 0, 1000); | 682 | tt, NULL, 0, 1000); |
@@ -2848,6 +2857,15 @@ static int usb_disable_function_remotewakeup(struct usb_device *udev) | |||
2848 | USB_CTRL_SET_TIMEOUT); | 2857 | USB_CTRL_SET_TIMEOUT); |
2849 | } | 2858 | } |
2850 | 2859 | ||
2860 | /* Count of wakeup-enabled devices at or below udev */ | ||
2861 | static unsigned wakeup_enabled_descendants(struct usb_device *udev) | ||
2862 | { | ||
2863 | struct usb_hub *hub = usb_hub_to_struct_hub(udev); | ||
2864 | |||
2865 | return udev->do_remote_wakeup + | ||
2866 | (hub ? hub->wakeup_enabled_descendants : 0); | ||
2867 | } | ||
2868 | |||
2851 | /* | 2869 | /* |
2852 | * usb_port_suspend - suspend a usb device's upstream port | 2870 | * usb_port_suspend - suspend a usb device's upstream port |
2853 | * @udev: device that's no longer in active use, not a root hub | 2871 | * @udev: device that's no longer in active use, not a root hub |
@@ -2888,8 +2906,8 @@ static int usb_disable_function_remotewakeup(struct usb_device *udev) | |||
2888 | * Linux (2.6) currently has NO mechanisms to initiate that: no khubd | 2906 | * Linux (2.6) currently has NO mechanisms to initiate that: no khubd |
2889 | * timer, no SRP, no requests through sysfs. | 2907 | * timer, no SRP, no requests through sysfs. |
2890 | * | 2908 | * |
2891 | * If Runtime PM isn't enabled or used, non-SuperSpeed devices really get | 2909 | * If Runtime PM isn't enabled or used, non-SuperSpeed devices may not get |
2892 | * suspended only when their bus goes into global suspend (i.e., the root | 2910 | * suspended until their bus goes into global suspend (i.e., the root |
2893 | * hub is suspended). Nevertheless, we change @udev->state to | 2911 | * hub is suspended). Nevertheless, we change @udev->state to |
2894 | * USB_STATE_SUSPENDED as this is the device's "logical" state. The actual | 2912 | * USB_STATE_SUSPENDED as this is the device's "logical" state. The actual |
2895 | * upstream port setting is stored in @udev->port_is_suspended. | 2913 | * upstream port setting is stored in @udev->port_is_suspended. |
@@ -2960,15 +2978,21 @@ int usb_port_suspend(struct usb_device *udev, pm_message_t msg) | |||
2960 | /* see 7.1.7.6 */ | 2978 | /* see 7.1.7.6 */ |
2961 | if (hub_is_superspeed(hub->hdev)) | 2979 | if (hub_is_superspeed(hub->hdev)) |
2962 | status = hub_set_port_link_state(hub, port1, USB_SS_PORT_LS_U3); | 2980 | status = hub_set_port_link_state(hub, port1, USB_SS_PORT_LS_U3); |
2963 | else if (PMSG_IS_AUTO(msg)) | 2981 | |
2964 | status = set_port_feature(hub->hdev, port1, | ||
2965 | USB_PORT_FEAT_SUSPEND); | ||
2966 | /* | 2982 | /* |
2967 | * For system suspend, we do not need to enable the suspend feature | 2983 | * For system suspend, we do not need to enable the suspend feature |
2968 | * on individual USB-2 ports. The devices will automatically go | 2984 | * on individual USB-2 ports. The devices will automatically go |
2969 | * into suspend a few ms after the root hub stops sending packets. | 2985 | * into suspend a few ms after the root hub stops sending packets. |
2970 | * The USB 2.0 spec calls this "global suspend". | 2986 | * The USB 2.0 spec calls this "global suspend". |
2987 | * | ||
2988 | * However, many USB hubs have a bug: They don't relay wakeup requests | ||
2989 | * from a downstream port if the port's suspend feature isn't on. | ||
2990 | * Therefore we will turn on the suspend feature if udev or any of its | ||
2991 | * descendants is enabled for remote wakeup. | ||
2971 | */ | 2992 | */ |
2993 | else if (PMSG_IS_AUTO(msg) || wakeup_enabled_descendants(udev) > 0) | ||
2994 | status = set_port_feature(hub->hdev, port1, | ||
2995 | USB_PORT_FEAT_SUSPEND); | ||
2972 | else { | 2996 | else { |
2973 | really_suspend = false; | 2997 | really_suspend = false; |
2974 | status = 0; | 2998 | status = 0; |
@@ -3003,15 +3027,16 @@ int usb_port_suspend(struct usb_device *udev, pm_message_t msg) | |||
3003 | if (!PMSG_IS_AUTO(msg)) | 3027 | if (!PMSG_IS_AUTO(msg)) |
3004 | status = 0; | 3028 | status = 0; |
3005 | } else { | 3029 | } else { |
3006 | /* device has up to 10 msec to fully suspend */ | ||
3007 | dev_dbg(&udev->dev, "usb %ssuspend, wakeup %d\n", | 3030 | dev_dbg(&udev->dev, "usb %ssuspend, wakeup %d\n", |
3008 | (PMSG_IS_AUTO(msg) ? "auto-" : ""), | 3031 | (PMSG_IS_AUTO(msg) ? "auto-" : ""), |
3009 | udev->do_remote_wakeup); | 3032 | udev->do_remote_wakeup); |
3010 | usb_set_device_state(udev, USB_STATE_SUSPENDED); | ||
3011 | if (really_suspend) { | 3033 | if (really_suspend) { |
3012 | udev->port_is_suspended = 1; | 3034 | udev->port_is_suspended = 1; |
3035 | |||
3036 | /* device has up to 10 msec to fully suspend */ | ||
3013 | msleep(10); | 3037 | msleep(10); |
3014 | } | 3038 | } |
3039 | usb_set_device_state(udev, USB_STATE_SUSPENDED); | ||
3015 | } | 3040 | } |
3016 | 3041 | ||
3017 | /* | 3042 | /* |
@@ -3293,7 +3318,11 @@ static int hub_suspend(struct usb_interface *intf, pm_message_t msg) | |||
3293 | unsigned port1; | 3318 | unsigned port1; |
3294 | int status; | 3319 | int status; |
3295 | 3320 | ||
3296 | /* Warn if children aren't already suspended */ | 3321 | /* |
3322 | * Warn if children aren't already suspended. | ||
3323 | * Also, add up the number of wakeup-enabled descendants. | ||
3324 | */ | ||
3325 | hub->wakeup_enabled_descendants = 0; | ||
3297 | for (port1 = 1; port1 <= hdev->maxchild; port1++) { | 3326 | for (port1 = 1; port1 <= hdev->maxchild; port1++) { |
3298 | struct usb_device *udev; | 3327 | struct usb_device *udev; |
3299 | 3328 | ||
@@ -3303,6 +3332,9 @@ static int hub_suspend(struct usb_interface *intf, pm_message_t msg) | |||
3303 | if (PMSG_IS_AUTO(msg)) | 3332 | if (PMSG_IS_AUTO(msg)) |
3304 | return -EBUSY; | 3333 | return -EBUSY; |
3305 | } | 3334 | } |
3335 | if (udev) | ||
3336 | hub->wakeup_enabled_descendants += | ||
3337 | wakeup_enabled_descendants(udev); | ||
3306 | } | 3338 | } |
3307 | 3339 | ||
3308 | if (hdev->do_remote_wakeup && hub->quirk_check_port_auto_suspend) { | 3340 | if (hdev->do_remote_wakeup && hub->quirk_check_port_auto_suspend) { |
diff --git a/drivers/usb/core/hub.h b/drivers/usb/core/hub.h index 6508e02b3dac..4e4790dea343 100644 --- a/drivers/usb/core/hub.h +++ b/drivers/usb/core/hub.h | |||
@@ -59,6 +59,9 @@ struct usb_hub { | |||
59 | struct usb_tt tt; /* Transaction Translator */ | 59 | struct usb_tt tt; /* Transaction Translator */ |
60 | 60 | ||
61 | unsigned mA_per_port; /* current for each child */ | 61 | unsigned mA_per_port; /* current for each child */ |
62 | #ifdef CONFIG_PM | ||
63 | unsigned wakeup_enabled_descendants; | ||
64 | #endif | ||
62 | 65 | ||
63 | unsigned limited_power:1; | 66 | unsigned limited_power:1; |
64 | unsigned quiescing:1; | 67 | unsigned quiescing:1; |
diff --git a/drivers/usb/dwc3/Kconfig b/drivers/usb/dwc3/Kconfig index 757aa18027d0..2378958ea63e 100644 --- a/drivers/usb/dwc3/Kconfig +++ b/drivers/usb/dwc3/Kconfig | |||
@@ -1,6 +1,6 @@ | |||
1 | config USB_DWC3 | 1 | config USB_DWC3 |
2 | tristate "DesignWare USB3 DRD Core Support" | 2 | tristate "DesignWare USB3 DRD Core Support" |
3 | depends on (USB || USB_GADGET) && GENERIC_HARDIRQS | 3 | depends on (USB || USB_GADGET) && GENERIC_HARDIRQS && HAS_DMA |
4 | select USB_XHCI_PLATFORM if USB_SUPPORT && USB_XHCI_HCD | 4 | select USB_XHCI_PLATFORM if USB_SUPPORT && USB_XHCI_HCD |
5 | help | 5 | help |
6 | Say Y or M here if your system has a Dual Role SuperSpeed | 6 | Say Y or M here if your system has a Dual Role SuperSpeed |
diff --git a/drivers/usb/dwc3/core.c b/drivers/usb/dwc3/core.c index c35d49d39b76..358375e0b291 100644 --- a/drivers/usb/dwc3/core.c +++ b/drivers/usb/dwc3/core.c | |||
@@ -450,7 +450,7 @@ static int dwc3_probe(struct platform_device *pdev) | |||
450 | } | 450 | } |
451 | 451 | ||
452 | if (IS_ERR(dwc->usb3_phy)) { | 452 | if (IS_ERR(dwc->usb3_phy)) { |
453 | ret = PTR_ERR(dwc->usb2_phy); | 453 | ret = PTR_ERR(dwc->usb3_phy); |
454 | 454 | ||
455 | /* | 455 | /* |
456 | * if -ENXIO is returned, it means PHY layer wasn't | 456 | * if -ENXIO is returned, it means PHY layer wasn't |
diff --git a/drivers/usb/dwc3/core.h b/drivers/usb/dwc3/core.h index b69d322e3cab..27dad993b007 100644 --- a/drivers/usb/dwc3/core.h +++ b/drivers/usb/dwc3/core.h | |||
@@ -759,8 +759,8 @@ struct dwc3 { | |||
759 | 759 | ||
760 | struct dwc3_event_type { | 760 | struct dwc3_event_type { |
761 | u32 is_devspec:1; | 761 | u32 is_devspec:1; |
762 | u32 type:6; | 762 | u32 type:7; |
763 | u32 reserved8_31:25; | 763 | u32 reserved8_31:24; |
764 | } __packed; | 764 | } __packed; |
765 | 765 | ||
766 | #define DWC3_DEPEVT_XFERCOMPLETE 0x01 | 766 | #define DWC3_DEPEVT_XFERCOMPLETE 0x01 |
diff --git a/drivers/usb/dwc3/gadget.c b/drivers/usb/dwc3/gadget.c index b5e5b35df49c..f77083fedc68 100644 --- a/drivers/usb/dwc3/gadget.c +++ b/drivers/usb/dwc3/gadget.c | |||
@@ -1584,6 +1584,7 @@ err1: | |||
1584 | __dwc3_gadget_ep_disable(dwc->eps[0]); | 1584 | __dwc3_gadget_ep_disable(dwc->eps[0]); |
1585 | 1585 | ||
1586 | err0: | 1586 | err0: |
1587 | dwc->gadget_driver = NULL; | ||
1587 | spin_unlock_irqrestore(&dwc->lock, flags); | 1588 | spin_unlock_irqrestore(&dwc->lock, flags); |
1588 | 1589 | ||
1589 | return ret; | 1590 | return ret; |
diff --git a/drivers/usb/gadget/Kconfig b/drivers/usb/gadget/Kconfig index 62f6802f6e0f..8e9368330b10 100644 --- a/drivers/usb/gadget/Kconfig +++ b/drivers/usb/gadget/Kconfig | |||
@@ -193,6 +193,7 @@ config USB_FUSB300 | |||
193 | Faraday usb device controller FUSB300 driver | 193 | Faraday usb device controller FUSB300 driver |
194 | 194 | ||
195 | config USB_FOTG210_UDC | 195 | config USB_FOTG210_UDC |
196 | depends on HAS_DMA | ||
196 | tristate "Faraday FOTG210 USB Peripheral Controller" | 197 | tristate "Faraday FOTG210 USB Peripheral Controller" |
197 | help | 198 | help |
198 | Faraday USB2.0 OTG controller which can be configured as | 199 | Faraday USB2.0 OTG controller which can be configured as |
@@ -328,13 +329,14 @@ config USB_S3C_HSUDC | |||
328 | 329 | ||
329 | config USB_MV_UDC | 330 | config USB_MV_UDC |
330 | tristate "Marvell USB2.0 Device Controller" | 331 | tristate "Marvell USB2.0 Device Controller" |
331 | depends on GENERIC_HARDIRQS | 332 | depends on GENERIC_HARDIRQS && HAS_DMA |
332 | help | 333 | help |
333 | Marvell Socs (including PXA and MMP series) include a high speed | 334 | Marvell Socs (including PXA and MMP series) include a high speed |
334 | USB2.0 OTG controller, which can be configured as high speed or | 335 | USB2.0 OTG controller, which can be configured as high speed or |
335 | full speed USB peripheral. | 336 | full speed USB peripheral. |
336 | 337 | ||
337 | config USB_MV_U3D | 338 | config USB_MV_U3D |
339 | depends on HAS_DMA | ||
338 | tristate "MARVELL PXA2128 USB 3.0 controller" | 340 | tristate "MARVELL PXA2128 USB 3.0 controller" |
339 | help | 341 | help |
340 | MARVELL PXA2128 Processor series include a super speed USB3.0 device | 342 | MARVELL PXA2128 Processor series include a super speed USB3.0 device |
@@ -639,6 +641,7 @@ config USB_CONFIGFS_RNDIS | |||
639 | depends on USB_CONFIGFS | 641 | depends on USB_CONFIGFS |
640 | depends on NET | 642 | depends on NET |
641 | select USB_U_ETHER | 643 | select USB_U_ETHER |
644 | select USB_U_RNDIS | ||
642 | select USB_F_RNDIS | 645 | select USB_F_RNDIS |
643 | help | 646 | help |
644 | Microsoft Windows XP bundles the "Remote NDIS" (RNDIS) protocol, | 647 | Microsoft Windows XP bundles the "Remote NDIS" (RNDIS) protocol, |
diff --git a/drivers/usb/gadget/at91_udc.c b/drivers/usb/gadget/at91_udc.c index 073b938f9135..d9a6add0c852 100644 --- a/drivers/usb/gadget/at91_udc.c +++ b/drivers/usb/gadget/at91_udc.c | |||
@@ -870,8 +870,8 @@ static void clk_on(struct at91_udc *udc) | |||
870 | if (udc->clocked) | 870 | if (udc->clocked) |
871 | return; | 871 | return; |
872 | udc->clocked = 1; | 872 | udc->clocked = 1; |
873 | clk_enable(udc->iclk); | 873 | clk_prepare_enable(udc->iclk); |
874 | clk_enable(udc->fclk); | 874 | clk_prepare_enable(udc->fclk); |
875 | } | 875 | } |
876 | 876 | ||
877 | static void clk_off(struct at91_udc *udc) | 877 | static void clk_off(struct at91_udc *udc) |
@@ -880,8 +880,8 @@ static void clk_off(struct at91_udc *udc) | |||
880 | return; | 880 | return; |
881 | udc->clocked = 0; | 881 | udc->clocked = 0; |
882 | udc->gadget.speed = USB_SPEED_UNKNOWN; | 882 | udc->gadget.speed = USB_SPEED_UNKNOWN; |
883 | clk_disable(udc->fclk); | 883 | clk_disable_unprepare(udc->fclk); |
884 | clk_disable(udc->iclk); | 884 | clk_disable_unprepare(udc->iclk); |
885 | } | 885 | } |
886 | 886 | ||
887 | /* | 887 | /* |
@@ -1725,7 +1725,7 @@ static int at91udc_probe(struct platform_device *pdev) | |||
1725 | /* init software state */ | 1725 | /* init software state */ |
1726 | udc = &controller; | 1726 | udc = &controller; |
1727 | udc->gadget.dev.parent = dev; | 1727 | udc->gadget.dev.parent = dev; |
1728 | if (pdev->dev.of_node) | 1728 | if (IS_ENABLED(CONFIG_OF) && pdev->dev.of_node) |
1729 | at91udc_of_init(udc, pdev->dev.of_node); | 1729 | at91udc_of_init(udc, pdev->dev.of_node); |
1730 | else | 1730 | else |
1731 | memcpy(&udc->board, dev->platform_data, | 1731 | memcpy(&udc->board, dev->platform_data, |
@@ -1782,12 +1782,14 @@ static int at91udc_probe(struct platform_device *pdev) | |||
1782 | } | 1782 | } |
1783 | 1783 | ||
1784 | /* don't do anything until we have both gadget driver and VBUS */ | 1784 | /* don't do anything until we have both gadget driver and VBUS */ |
1785 | clk_enable(udc->iclk); | 1785 | retval = clk_prepare_enable(udc->iclk); |
1786 | if (retval) | ||
1787 | goto fail1; | ||
1786 | at91_udp_write(udc, AT91_UDP_TXVC, AT91_UDP_TXVC_TXVDIS); | 1788 | at91_udp_write(udc, AT91_UDP_TXVC, AT91_UDP_TXVC_TXVDIS); |
1787 | at91_udp_write(udc, AT91_UDP_IDR, 0xffffffff); | 1789 | at91_udp_write(udc, AT91_UDP_IDR, 0xffffffff); |
1788 | /* Clear all pending interrupts - UDP may be used by bootloader. */ | 1790 | /* Clear all pending interrupts - UDP may be used by bootloader. */ |
1789 | at91_udp_write(udc, AT91_UDP_ICR, 0xffffffff); | 1791 | at91_udp_write(udc, AT91_UDP_ICR, 0xffffffff); |
1790 | clk_disable(udc->iclk); | 1792 | clk_disable_unprepare(udc->iclk); |
1791 | 1793 | ||
1792 | /* request UDC and maybe VBUS irqs */ | 1794 | /* request UDC and maybe VBUS irqs */ |
1793 | udc->udp_irq = platform_get_irq(pdev, 0); | 1795 | udc->udp_irq = platform_get_irq(pdev, 0); |
diff --git a/drivers/usb/gadget/f_ecm.c b/drivers/usb/gadget/f_ecm.c index 5d3561ea1c15..edab45da3741 100644 --- a/drivers/usb/gadget/f_ecm.c +++ b/drivers/usb/gadget/f_ecm.c | |||
@@ -959,8 +959,11 @@ static struct usb_function_instance *ecm_alloc_inst(void) | |||
959 | mutex_init(&opts->lock); | 959 | mutex_init(&opts->lock); |
960 | opts->func_inst.free_func_inst = ecm_free_inst; | 960 | opts->func_inst.free_func_inst = ecm_free_inst; |
961 | opts->net = gether_setup_default(); | 961 | opts->net = gether_setup_default(); |
962 | if (IS_ERR(opts->net)) | 962 | if (IS_ERR(opts->net)) { |
963 | return ERR_PTR(PTR_ERR(opts->net)); | 963 | struct net_device *net = opts->net; |
964 | kfree(opts); | ||
965 | return ERR_CAST(net); | ||
966 | } | ||
964 | 967 | ||
965 | config_group_init_type_name(&opts->func_inst.group, "", &ecm_func_type); | 968 | config_group_init_type_name(&opts->func_inst.group, "", &ecm_func_type); |
966 | 969 | ||
diff --git a/drivers/usb/gadget/f_eem.c b/drivers/usb/gadget/f_eem.c index 90ee8022e8d8..d00392d879db 100644 --- a/drivers/usb/gadget/f_eem.c +++ b/drivers/usb/gadget/f_eem.c | |||
@@ -593,8 +593,11 @@ static struct usb_function_instance *eem_alloc_inst(void) | |||
593 | mutex_init(&opts->lock); | 593 | mutex_init(&opts->lock); |
594 | opts->func_inst.free_func_inst = eem_free_inst; | 594 | opts->func_inst.free_func_inst = eem_free_inst; |
595 | opts->net = gether_setup_default(); | 595 | opts->net = gether_setup_default(); |
596 | if (IS_ERR(opts->net)) | 596 | if (IS_ERR(opts->net)) { |
597 | return ERR_CAST(opts->net); | 597 | struct net_device *net = opts->net; |
598 | kfree(opts); | ||
599 | return ERR_CAST(net); | ||
600 | } | ||
598 | 601 | ||
599 | config_group_init_type_name(&opts->func_inst.group, "", &eem_func_type); | 602 | config_group_init_type_name(&opts->func_inst.group, "", &eem_func_type); |
600 | 603 | ||
diff --git a/drivers/usb/gadget/f_ncm.c b/drivers/usb/gadget/f_ncm.c index 952177f7eb9b..1c28fe13328a 100644 --- a/drivers/usb/gadget/f_ncm.c +++ b/drivers/usb/gadget/f_ncm.c | |||
@@ -1350,8 +1350,11 @@ static struct usb_function_instance *ncm_alloc_inst(void) | |||
1350 | mutex_init(&opts->lock); | 1350 | mutex_init(&opts->lock); |
1351 | opts->func_inst.free_func_inst = ncm_free_inst; | 1351 | opts->func_inst.free_func_inst = ncm_free_inst; |
1352 | opts->net = gether_setup_default(); | 1352 | opts->net = gether_setup_default(); |
1353 | if (IS_ERR(opts->net)) | 1353 | if (IS_ERR(opts->net)) { |
1354 | return ERR_PTR(PTR_ERR(opts->net)); | 1354 | struct net_device *net = opts->net; |
1355 | kfree(opts); | ||
1356 | return ERR_CAST(net); | ||
1357 | } | ||
1355 | 1358 | ||
1356 | config_group_init_type_name(&opts->func_inst.group, "", &ncm_func_type); | 1359 | config_group_init_type_name(&opts->func_inst.group, "", &ncm_func_type); |
1357 | 1360 | ||
diff --git a/drivers/usb/gadget/f_phonet.c b/drivers/usb/gadget/f_phonet.c index 7944fb0efe3b..1bf26e9f38cd 100644 --- a/drivers/usb/gadget/f_phonet.c +++ b/drivers/usb/gadget/f_phonet.c | |||
@@ -656,8 +656,11 @@ static struct usb_function_instance *phonet_alloc_inst(void) | |||
656 | 656 | ||
657 | opts->func_inst.free_func_inst = phonet_free_inst; | 657 | opts->func_inst.free_func_inst = phonet_free_inst; |
658 | opts->net = gphonet_setup_default(); | 658 | opts->net = gphonet_setup_default(); |
659 | if (IS_ERR(opts->net)) | 659 | if (IS_ERR(opts->net)) { |
660 | return ERR_PTR(PTR_ERR(opts->net)); | 660 | struct net_device *net = opts->net; |
661 | kfree(opts); | ||
662 | return ERR_CAST(net); | ||
663 | } | ||
661 | 664 | ||
662 | config_group_init_type_name(&opts->func_inst.group, "", | 665 | config_group_init_type_name(&opts->func_inst.group, "", |
663 | &phonet_func_type); | 666 | &phonet_func_type); |
diff --git a/drivers/usb/gadget/f_rndis.c b/drivers/usb/gadget/f_rndis.c index 191df35ae69d..717ed7f95639 100644 --- a/drivers/usb/gadget/f_rndis.c +++ b/drivers/usb/gadget/f_rndis.c | |||
@@ -963,8 +963,11 @@ static struct usb_function_instance *rndis_alloc_inst(void) | |||
963 | mutex_init(&opts->lock); | 963 | mutex_init(&opts->lock); |
964 | opts->func_inst.free_func_inst = rndis_free_inst; | 964 | opts->func_inst.free_func_inst = rndis_free_inst; |
965 | opts->net = gether_setup_default(); | 965 | opts->net = gether_setup_default(); |
966 | if (IS_ERR(opts->net)) | 966 | if (IS_ERR(opts->net)) { |
967 | return ERR_CAST(opts->net); | 967 | struct net_device *net = opts->net; |
968 | kfree(opts); | ||
969 | return ERR_CAST(net); | ||
970 | } | ||
968 | 971 | ||
969 | config_group_init_type_name(&opts->func_inst.group, "", | 972 | config_group_init_type_name(&opts->func_inst.group, "", |
970 | &rndis_func_type); | 973 | &rndis_func_type); |
diff --git a/drivers/usb/gadget/f_subset.c b/drivers/usb/gadget/f_subset.c index 5601e1d96c4f..7c8674fa7e80 100644 --- a/drivers/usb/gadget/f_subset.c +++ b/drivers/usb/gadget/f_subset.c | |||
@@ -505,8 +505,11 @@ static struct usb_function_instance *geth_alloc_inst(void) | |||
505 | mutex_init(&opts->lock); | 505 | mutex_init(&opts->lock); |
506 | opts->func_inst.free_func_inst = geth_free_inst; | 506 | opts->func_inst.free_func_inst = geth_free_inst; |
507 | opts->net = gether_setup_default(); | 507 | opts->net = gether_setup_default(); |
508 | if (IS_ERR(opts->net)) | 508 | if (IS_ERR(opts->net)) { |
509 | return ERR_CAST(opts->net); | 509 | struct net_device *net = opts->net; |
510 | kfree(opts); | ||
511 | return ERR_CAST(net); | ||
512 | } | ||
510 | 513 | ||
511 | config_group_init_type_name(&opts->func_inst.group, "", | 514 | config_group_init_type_name(&opts->func_inst.group, "", |
512 | &gether_func_type); | 515 | &gether_func_type); |
diff --git a/drivers/usb/gadget/fotg210-udc.c b/drivers/usb/gadget/fotg210-udc.c index cce5535b1dc6..10cd18ddd0d4 100644 --- a/drivers/usb/gadget/fotg210-udc.c +++ b/drivers/usb/gadget/fotg210-udc.c | |||
@@ -1074,7 +1074,7 @@ static struct usb_gadget_ops fotg210_gadget_ops = { | |||
1074 | .udc_stop = fotg210_udc_stop, | 1074 | .udc_stop = fotg210_udc_stop, |
1075 | }; | 1075 | }; |
1076 | 1076 | ||
1077 | static int __exit fotg210_udc_remove(struct platform_device *pdev) | 1077 | static int fotg210_udc_remove(struct platform_device *pdev) |
1078 | { | 1078 | { |
1079 | struct fotg210_udc *fotg210 = dev_get_drvdata(&pdev->dev); | 1079 | struct fotg210_udc *fotg210 = dev_get_drvdata(&pdev->dev); |
1080 | 1080 | ||
@@ -1088,7 +1088,7 @@ static int __exit fotg210_udc_remove(struct platform_device *pdev) | |||
1088 | return 0; | 1088 | return 0; |
1089 | } | 1089 | } |
1090 | 1090 | ||
1091 | static int __init fotg210_udc_probe(struct platform_device *pdev) | 1091 | static int fotg210_udc_probe(struct platform_device *pdev) |
1092 | { | 1092 | { |
1093 | struct resource *res, *ires; | 1093 | struct resource *res, *ires; |
1094 | struct fotg210_udc *fotg210 = NULL; | 1094 | struct fotg210_udc *fotg210 = NULL; |
diff --git a/drivers/usb/gadget/mv_u3d_core.c b/drivers/usb/gadget/mv_u3d_core.c index 07fdb3eaf48a..ec6a2d290398 100644 --- a/drivers/usb/gadget/mv_u3d_core.c +++ b/drivers/usb/gadget/mv_u3d_core.c | |||
@@ -1776,7 +1776,7 @@ static int mv_u3d_remove(struct platform_device *dev) | |||
1776 | kfree(u3d->eps); | 1776 | kfree(u3d->eps); |
1777 | 1777 | ||
1778 | if (u3d->irq) | 1778 | if (u3d->irq) |
1779 | free_irq(u3d->irq, &dev->dev); | 1779 | free_irq(u3d->irq, u3d); |
1780 | 1780 | ||
1781 | if (u3d->cap_regs) | 1781 | if (u3d->cap_regs) |
1782 | iounmap(u3d->cap_regs); | 1782 | iounmap(u3d->cap_regs); |
@@ -1974,7 +1974,7 @@ static int mv_u3d_probe(struct platform_device *dev) | |||
1974 | return 0; | 1974 | return 0; |
1975 | 1975 | ||
1976 | err_unregister: | 1976 | err_unregister: |
1977 | free_irq(u3d->irq, &dev->dev); | 1977 | free_irq(u3d->irq, u3d); |
1978 | err_request_irq: | 1978 | err_request_irq: |
1979 | err_get_irq: | 1979 | err_get_irq: |
1980 | kfree(u3d->status_req); | 1980 | kfree(u3d->status_req); |
diff --git a/drivers/usb/gadget/udc-core.c b/drivers/usb/gadget/udc-core.c index ffd8fa541101..c28ac9872030 100644 --- a/drivers/usb/gadget/udc-core.c +++ b/drivers/usb/gadget/udc-core.c | |||
@@ -50,6 +50,8 @@ static DEFINE_MUTEX(udc_lock); | |||
50 | 50 | ||
51 | /* ------------------------------------------------------------------------- */ | 51 | /* ------------------------------------------------------------------------- */ |
52 | 52 | ||
53 | #ifdef CONFIG_HAS_DMA | ||
54 | |||
53 | int usb_gadget_map_request(struct usb_gadget *gadget, | 55 | int usb_gadget_map_request(struct usb_gadget *gadget, |
54 | struct usb_request *req, int is_in) | 56 | struct usb_request *req, int is_in) |
55 | { | 57 | { |
@@ -99,6 +101,8 @@ void usb_gadget_unmap_request(struct usb_gadget *gadget, | |||
99 | } | 101 | } |
100 | EXPORT_SYMBOL_GPL(usb_gadget_unmap_request); | 102 | EXPORT_SYMBOL_GPL(usb_gadget_unmap_request); |
101 | 103 | ||
104 | #endif /* CONFIG_HAS_DMA */ | ||
105 | |||
102 | /* ------------------------------------------------------------------------- */ | 106 | /* ------------------------------------------------------------------------- */ |
103 | 107 | ||
104 | void usb_gadget_set_state(struct usb_gadget *gadget, | 108 | void usb_gadget_set_state(struct usb_gadget *gadget, |
@@ -194,9 +198,11 @@ int usb_add_gadget_udc_release(struct device *parent, struct usb_gadget *gadget, | |||
194 | dev_set_name(&gadget->dev, "gadget"); | 198 | dev_set_name(&gadget->dev, "gadget"); |
195 | gadget->dev.parent = parent; | 199 | gadget->dev.parent = parent; |
196 | 200 | ||
201 | #ifdef CONFIG_HAS_DMA | ||
197 | dma_set_coherent_mask(&gadget->dev, parent->coherent_dma_mask); | 202 | dma_set_coherent_mask(&gadget->dev, parent->coherent_dma_mask); |
198 | gadget->dev.dma_parms = parent->dma_parms; | 203 | gadget->dev.dma_parms = parent->dma_parms; |
199 | gadget->dev.dma_mask = parent->dma_mask; | 204 | gadget->dev.dma_mask = parent->dma_mask; |
205 | #endif | ||
200 | 206 | ||
201 | if (release) | 207 | if (release) |
202 | gadget->dev.release = release; | 208 | gadget->dev.release = release; |
diff --git a/drivers/usb/host/ehci-hub.c b/drivers/usb/host/ehci-hub.c index 2b702772d04d..6dce37555c4f 100644 --- a/drivers/usb/host/ehci-hub.c +++ b/drivers/usb/host/ehci-hub.c | |||
@@ -874,6 +874,7 @@ static int ehci_hub_control ( | |||
874 | ehci->reset_done[wIndex] = jiffies | 874 | ehci->reset_done[wIndex] = jiffies |
875 | + msecs_to_jiffies(20); | 875 | + msecs_to_jiffies(20); |
876 | usb_hcd_start_port_resume(&hcd->self, wIndex); | 876 | usb_hcd_start_port_resume(&hcd->self, wIndex); |
877 | set_bit(wIndex, &ehci->resuming_ports); | ||
877 | /* check the port again */ | 878 | /* check the port again */ |
878 | mod_timer(&ehci_to_hcd(ehci)->rh_timer, | 879 | mod_timer(&ehci_to_hcd(ehci)->rh_timer, |
879 | ehci->reset_done[wIndex]); | 880 | ehci->reset_done[wIndex]); |
diff --git a/drivers/usb/host/pci-quirks.h b/drivers/usb/host/pci-quirks.h index 4b8a2092432f..978c849f9c9a 100644 --- a/drivers/usb/host/pci-quirks.h +++ b/drivers/usb/host/pci-quirks.h | |||
@@ -13,6 +13,7 @@ void usb_enable_xhci_ports(struct pci_dev *xhci_pdev); | |||
13 | void usb_disable_xhci_ports(struct pci_dev *xhci_pdev); | 13 | void usb_disable_xhci_ports(struct pci_dev *xhci_pdev); |
14 | void sb800_prefetch(struct device *dev, int on); | 14 | void sb800_prefetch(struct device *dev, int on); |
15 | #else | 15 | #else |
16 | struct pci_dev; | ||
16 | static inline void usb_amd_quirk_pll_disable(void) {} | 17 | static inline void usb_amd_quirk_pll_disable(void) {} |
17 | static inline void usb_amd_quirk_pll_enable(void) {} | 18 | static inline void usb_amd_quirk_pll_enable(void) {} |
18 | static inline void usb_amd_dev_put(void) {} | 19 | static inline void usb_amd_dev_put(void) {} |
diff --git a/drivers/usb/host/xhci-pci.c b/drivers/usb/host/xhci-pci.c index cc24e39b97d5..f00cb203faea 100644 --- a/drivers/usb/host/xhci-pci.c +++ b/drivers/usb/host/xhci-pci.c | |||
@@ -93,7 +93,6 @@ static void xhci_pci_quirks(struct device *dev, struct xhci_hcd *xhci) | |||
93 | } | 93 | } |
94 | if (pdev->vendor == PCI_VENDOR_ID_INTEL && | 94 | if (pdev->vendor == PCI_VENDOR_ID_INTEL && |
95 | pdev->device == PCI_DEVICE_ID_INTEL_PANTHERPOINT_XHCI) { | 95 | pdev->device == PCI_DEVICE_ID_INTEL_PANTHERPOINT_XHCI) { |
96 | xhci->quirks |= XHCI_SPURIOUS_SUCCESS; | ||
97 | xhci->quirks |= XHCI_EP_LIMIT_QUIRK; | 96 | xhci->quirks |= XHCI_EP_LIMIT_QUIRK; |
98 | xhci->limit_active_eps = 64; | 97 | xhci->limit_active_eps = 64; |
99 | xhci->quirks |= XHCI_SW_BW_CHECKING; | 98 | xhci->quirks |= XHCI_SW_BW_CHECKING; |
diff --git a/drivers/usb/host/xhci-ring.c b/drivers/usb/host/xhci-ring.c index 1e57eafa6910..5b08cd85f8e7 100644 --- a/drivers/usb/host/xhci-ring.c +++ b/drivers/usb/host/xhci-ring.c | |||
@@ -434,7 +434,7 @@ static void ring_doorbell_for_active_rings(struct xhci_hcd *xhci, | |||
434 | 434 | ||
435 | /* A ring has pending URBs if its TD list is not empty */ | 435 | /* A ring has pending URBs if its TD list is not empty */ |
436 | if (!(ep->ep_state & EP_HAS_STREAMS)) { | 436 | if (!(ep->ep_state & EP_HAS_STREAMS)) { |
437 | if (!(list_empty(&ep->ring->td_list))) | 437 | if (ep->ring && !(list_empty(&ep->ring->td_list))) |
438 | xhci_ring_ep_doorbell(xhci, slot_id, ep_index, 0); | 438 | xhci_ring_ep_doorbell(xhci, slot_id, ep_index, 0); |
439 | return; | 439 | return; |
440 | } | 440 | } |
diff --git a/drivers/usb/host/xhci.c b/drivers/usb/host/xhci.c index 2c49f00260ca..41eb4fc33453 100644 --- a/drivers/usb/host/xhci.c +++ b/drivers/usb/host/xhci.c | |||
@@ -329,7 +329,7 @@ static void xhci_cleanup_msix(struct xhci_hcd *xhci) | |||
329 | return; | 329 | return; |
330 | } | 330 | } |
331 | 331 | ||
332 | static void xhci_msix_sync_irqs(struct xhci_hcd *xhci) | 332 | static void __maybe_unused xhci_msix_sync_irqs(struct xhci_hcd *xhci) |
333 | { | 333 | { |
334 | int i; | 334 | int i; |
335 | 335 | ||
@@ -1181,9 +1181,6 @@ static int xhci_check_args(struct usb_hcd *hcd, struct usb_device *udev, | |||
1181 | } | 1181 | } |
1182 | 1182 | ||
1183 | xhci = hcd_to_xhci(hcd); | 1183 | xhci = hcd_to_xhci(hcd); |
1184 | if (xhci->xhc_state & XHCI_STATE_HALTED) | ||
1185 | return -ENODEV; | ||
1186 | |||
1187 | if (check_virt_dev) { | 1184 | if (check_virt_dev) { |
1188 | if (!udev->slot_id || !xhci->devs[udev->slot_id]) { | 1185 | if (!udev->slot_id || !xhci->devs[udev->slot_id]) { |
1189 | printk(KERN_DEBUG "xHCI %s called with unaddressed " | 1186 | printk(KERN_DEBUG "xHCI %s called with unaddressed " |
@@ -1199,6 +1196,9 @@ static int xhci_check_args(struct usb_hcd *hcd, struct usb_device *udev, | |||
1199 | } | 1196 | } |
1200 | } | 1197 | } |
1201 | 1198 | ||
1199 | if (xhci->xhc_state & XHCI_STATE_HALTED) | ||
1200 | return -ENODEV; | ||
1201 | |||
1202 | return 1; | 1202 | return 1; |
1203 | } | 1203 | } |
1204 | 1204 | ||
@@ -3898,7 +3898,7 @@ int xhci_find_raw_port_number(struct usb_hcd *hcd, int port1) | |||
3898 | * Issue an Evaluate Context command to change the Maximum Exit Latency in the | 3898 | * Issue an Evaluate Context command to change the Maximum Exit Latency in the |
3899 | * slot context. If that succeeds, store the new MEL in the xhci_virt_device. | 3899 | * slot context. If that succeeds, store the new MEL in the xhci_virt_device. |
3900 | */ | 3900 | */ |
3901 | static int xhci_change_max_exit_latency(struct xhci_hcd *xhci, | 3901 | static int __maybe_unused xhci_change_max_exit_latency(struct xhci_hcd *xhci, |
3902 | struct usb_device *udev, u16 max_exit_latency) | 3902 | struct usb_device *udev, u16 max_exit_latency) |
3903 | { | 3903 | { |
3904 | struct xhci_virt_device *virt_dev; | 3904 | struct xhci_virt_device *virt_dev; |
@@ -4892,6 +4892,13 @@ int xhci_gen_setup(struct usb_hcd *hcd, xhci_get_quirks_t get_quirks) | |||
4892 | 4892 | ||
4893 | get_quirks(dev, xhci); | 4893 | get_quirks(dev, xhci); |
4894 | 4894 | ||
4895 | /* In xhci controllers which follow xhci 1.0 spec gives a spurious | ||
4896 | * success event after a short transfer. This quirk will ignore such | ||
4897 | * spurious event. | ||
4898 | */ | ||
4899 | if (xhci->hci_version > 0x96) | ||
4900 | xhci->quirks |= XHCI_SPURIOUS_SUCCESS; | ||
4901 | |||
4895 | /* Make sure the HC is halted. */ | 4902 | /* Make sure the HC is halted. */ |
4896 | retval = xhci_halt(xhci); | 4903 | retval = xhci_halt(xhci); |
4897 | if (retval) | 4904 | if (retval) |
diff --git a/drivers/usb/misc/sisusbvga/sisusb.c b/drivers/usb/misc/sisusbvga/sisusb.c index c21386ec5d35..de98906f786d 100644 --- a/drivers/usb/misc/sisusbvga/sisusb.c +++ b/drivers/usb/misc/sisusbvga/sisusb.c | |||
@@ -3247,6 +3247,7 @@ static const struct usb_device_id sisusb_table[] = { | |||
3247 | { USB_DEVICE(0x0711, 0x0903) }, | 3247 | { USB_DEVICE(0x0711, 0x0903) }, |
3248 | { USB_DEVICE(0x0711, 0x0918) }, | 3248 | { USB_DEVICE(0x0711, 0x0918) }, |
3249 | { USB_DEVICE(0x0711, 0x0920) }, | 3249 | { USB_DEVICE(0x0711, 0x0920) }, |
3250 | { USB_DEVICE(0x0711, 0x0950) }, | ||
3250 | { USB_DEVICE(0x182d, 0x021c) }, | 3251 | { USB_DEVICE(0x182d, 0x021c) }, |
3251 | { USB_DEVICE(0x182d, 0x0269) }, | 3252 | { USB_DEVICE(0x182d, 0x0269) }, |
3252 | { } | 3253 | { } |
diff --git a/drivers/usb/phy/phy-omap-usb3.c b/drivers/usb/phy/phy-omap-usb3.c index efe6e1464f45..a2fb30bbb971 100644 --- a/drivers/usb/phy/phy-omap-usb3.c +++ b/drivers/usb/phy/phy-omap-usb3.c | |||
@@ -71,9 +71,9 @@ static struct usb_dpll_params omap_usb3_dpll_params[NUM_SYS_CLKS] = { | |||
71 | {1250, 5, 4, 20, 0}, /* 12 MHz */ | 71 | {1250, 5, 4, 20, 0}, /* 12 MHz */ |
72 | {3125, 20, 4, 20, 0}, /* 16.8 MHz */ | 72 | {3125, 20, 4, 20, 0}, /* 16.8 MHz */ |
73 | {1172, 8, 4, 20, 65537}, /* 19.2 MHz */ | 73 | {1172, 8, 4, 20, 65537}, /* 19.2 MHz */ |
74 | {1000, 7, 4, 10, 0}, /* 20 MHz */ | ||
74 | {1250, 12, 4, 20, 0}, /* 26 MHz */ | 75 | {1250, 12, 4, 20, 0}, /* 26 MHz */ |
75 | {3125, 47, 4, 20, 92843}, /* 38.4 MHz */ | 76 | {3125, 47, 4, 20, 92843}, /* 38.4 MHz */ |
76 | {1000, 7, 4, 10, 0}, /* 20 MHz */ | ||
77 | 77 | ||
78 | }; | 78 | }; |
79 | 79 | ||
diff --git a/drivers/usb/phy/phy-samsung-usb2.c b/drivers/usb/phy/phy-samsung-usb2.c index 1011c16ade7e..758b86d0fcb3 100644 --- a/drivers/usb/phy/phy-samsung-usb2.c +++ b/drivers/usb/phy/phy-samsung-usb2.c | |||
@@ -388,7 +388,7 @@ static int samsung_usb2phy_probe(struct platform_device *pdev) | |||
388 | clk = devm_clk_get(dev, "otg"); | 388 | clk = devm_clk_get(dev, "otg"); |
389 | 389 | ||
390 | if (IS_ERR(clk)) { | 390 | if (IS_ERR(clk)) { |
391 | dev_err(dev, "Failed to get otg clock\n"); | 391 | dev_err(dev, "Failed to get usbhost/otg clock\n"); |
392 | return PTR_ERR(clk); | 392 | return PTR_ERR(clk); |
393 | } | 393 | } |
394 | 394 | ||
diff --git a/drivers/usb/renesas_usbhs/mod_gadget.c b/drivers/usb/renesas_usbhs/mod_gadget.c index ed4949faa70d..805940c37353 100644 --- a/drivers/usb/renesas_usbhs/mod_gadget.c +++ b/drivers/usb/renesas_usbhs/mod_gadget.c | |||
@@ -855,10 +855,6 @@ static int usbhsg_gadget_stop(struct usb_gadget *gadget, | |||
855 | struct usbhsg_gpriv *gpriv = usbhsg_gadget_to_gpriv(gadget); | 855 | struct usbhsg_gpriv *gpriv = usbhsg_gadget_to_gpriv(gadget); |
856 | struct usbhs_priv *priv = usbhsg_gpriv_to_priv(gpriv); | 856 | struct usbhs_priv *priv = usbhsg_gpriv_to_priv(gpriv); |
857 | 857 | ||
858 | if (!driver || | ||
859 | !driver->unbind) | ||
860 | return -EINVAL; | ||
861 | |||
862 | usbhsg_try_stop(priv, USBHSG_STATUS_REGISTERD); | 858 | usbhsg_try_stop(priv, USBHSG_STATUS_REGISTERD); |
863 | gpriv->driver = NULL; | 859 | gpriv->driver = NULL; |
864 | 860 | ||
diff --git a/drivers/usb/serial/cp210x.c b/drivers/usb/serial/cp210x.c index d6ef2f8da37d..0eae4ba3760e 100644 --- a/drivers/usb/serial/cp210x.c +++ b/drivers/usb/serial/cp210x.c | |||
@@ -53,6 +53,7 @@ static const struct usb_device_id id_table[] = { | |||
53 | { USB_DEVICE(0x0489, 0xE000) }, /* Pirelli Broadband S.p.A, DP-L10 SIP/GSM Mobile */ | 53 | { USB_DEVICE(0x0489, 0xE000) }, /* Pirelli Broadband S.p.A, DP-L10 SIP/GSM Mobile */ |
54 | { USB_DEVICE(0x0489, 0xE003) }, /* Pirelli Broadband S.p.A, DP-L10 SIP/GSM Mobile */ | 54 | { USB_DEVICE(0x0489, 0xE003) }, /* Pirelli Broadband S.p.A, DP-L10 SIP/GSM Mobile */ |
55 | { USB_DEVICE(0x0745, 0x1000) }, /* CipherLab USB CCD Barcode Scanner 1000 */ | 55 | { USB_DEVICE(0x0745, 0x1000) }, /* CipherLab USB CCD Barcode Scanner 1000 */ |
56 | { USB_DEVICE(0x0846, 0x1100) }, /* NetGear Managed Switch M4100 series, M5300 series, M7100 series */ | ||
56 | { USB_DEVICE(0x08e6, 0x5501) }, /* Gemalto Prox-PU/CU contactless smartcard reader */ | 57 | { USB_DEVICE(0x08e6, 0x5501) }, /* Gemalto Prox-PU/CU contactless smartcard reader */ |
57 | { USB_DEVICE(0x08FD, 0x000A) }, /* Digianswer A/S , ZigBee/802.15.4 MAC Device */ | 58 | { USB_DEVICE(0x08FD, 0x000A) }, /* Digianswer A/S , ZigBee/802.15.4 MAC Device */ |
58 | { USB_DEVICE(0x0BED, 0x1100) }, /* MEI (TM) Cashflow-SC Bill/Voucher Acceptor */ | 59 | { USB_DEVICE(0x0BED, 0x1100) }, /* MEI (TM) Cashflow-SC Bill/Voucher Acceptor */ |
@@ -118,6 +119,8 @@ static const struct usb_device_id id_table[] = { | |||
118 | { USB_DEVICE(0x10C4, 0x85F8) }, /* Virtenio Preon32 */ | 119 | { USB_DEVICE(0x10C4, 0x85F8) }, /* Virtenio Preon32 */ |
119 | { USB_DEVICE(0x10C4, 0x8664) }, /* AC-Services CAN-IF */ | 120 | { USB_DEVICE(0x10C4, 0x8664) }, /* AC-Services CAN-IF */ |
120 | { USB_DEVICE(0x10C4, 0x8665) }, /* AC-Services OBD-IF */ | 121 | { USB_DEVICE(0x10C4, 0x8665) }, /* AC-Services OBD-IF */ |
122 | { USB_DEVICE(0x10C4, 0x88A4) }, /* MMB Networks ZigBee USB Device */ | ||
123 | { USB_DEVICE(0x10C4, 0x88A5) }, /* Planet Innovation Ingeni ZigBee USB Device */ | ||
121 | { USB_DEVICE(0x10C4, 0xEA60) }, /* Silicon Labs factory default */ | 124 | { USB_DEVICE(0x10C4, 0xEA60) }, /* Silicon Labs factory default */ |
122 | { USB_DEVICE(0x10C4, 0xEA61) }, /* Silicon Labs factory default */ | 125 | { USB_DEVICE(0x10C4, 0xEA61) }, /* Silicon Labs factory default */ |
123 | { USB_DEVICE(0x10C4, 0xEA70) }, /* Silicon Labs factory default */ | 126 | { USB_DEVICE(0x10C4, 0xEA70) }, /* Silicon Labs factory default */ |
@@ -148,6 +151,7 @@ static const struct usb_device_id id_table[] = { | |||
148 | { USB_DEVICE(0x17F4, 0xAAAA) }, /* Wavesense Jazz blood glucose meter */ | 151 | { USB_DEVICE(0x17F4, 0xAAAA) }, /* Wavesense Jazz blood glucose meter */ |
149 | { USB_DEVICE(0x1843, 0x0200) }, /* Vaisala USB Instrument Cable */ | 152 | { USB_DEVICE(0x1843, 0x0200) }, /* Vaisala USB Instrument Cable */ |
150 | { USB_DEVICE(0x18EF, 0xE00F) }, /* ELV USB-I2C-Interface */ | 153 | { USB_DEVICE(0x18EF, 0xE00F) }, /* ELV USB-I2C-Interface */ |
154 | { USB_DEVICE(0x1ADB, 0x0001) }, /* Schweitzer Engineering C662 Cable */ | ||
151 | { USB_DEVICE(0x1BE3, 0x07A6) }, /* WAGO 750-923 USB Service Cable */ | 155 | { USB_DEVICE(0x1BE3, 0x07A6) }, /* WAGO 750-923 USB Service Cable */ |
152 | { USB_DEVICE(0x1E29, 0x0102) }, /* Festo CPX-USB */ | 156 | { USB_DEVICE(0x1E29, 0x0102) }, /* Festo CPX-USB */ |
153 | { USB_DEVICE(0x1E29, 0x0501) }, /* Festo CMSP */ | 157 | { USB_DEVICE(0x1E29, 0x0501) }, /* Festo CMSP */ |
diff --git a/drivers/usb/serial/mos7840.c b/drivers/usb/serial/mos7840.c index 0a818b238508..603fb70dde80 100644 --- a/drivers/usb/serial/mos7840.c +++ b/drivers/usb/serial/mos7840.c | |||
@@ -905,20 +905,20 @@ static int mos7840_open(struct tty_struct *tty, struct usb_serial_port *port) | |||
905 | status = mos7840_get_reg_sync(port, mos7840_port->SpRegOffset, &Data); | 905 | status = mos7840_get_reg_sync(port, mos7840_port->SpRegOffset, &Data); |
906 | if (status < 0) { | 906 | if (status < 0) { |
907 | dev_dbg(&port->dev, "Reading Spreg failed\n"); | 907 | dev_dbg(&port->dev, "Reading Spreg failed\n"); |
908 | return -1; | 908 | goto err; |
909 | } | 909 | } |
910 | Data |= 0x80; | 910 | Data |= 0x80; |
911 | status = mos7840_set_reg_sync(port, mos7840_port->SpRegOffset, Data); | 911 | status = mos7840_set_reg_sync(port, mos7840_port->SpRegOffset, Data); |
912 | if (status < 0) { | 912 | if (status < 0) { |
913 | dev_dbg(&port->dev, "writing Spreg failed\n"); | 913 | dev_dbg(&port->dev, "writing Spreg failed\n"); |
914 | return -1; | 914 | goto err; |
915 | } | 915 | } |
916 | 916 | ||
917 | Data &= ~0x80; | 917 | Data &= ~0x80; |
918 | status = mos7840_set_reg_sync(port, mos7840_port->SpRegOffset, Data); | 918 | status = mos7840_set_reg_sync(port, mos7840_port->SpRegOffset, Data); |
919 | if (status < 0) { | 919 | if (status < 0) { |
920 | dev_dbg(&port->dev, "writing Spreg failed\n"); | 920 | dev_dbg(&port->dev, "writing Spreg failed\n"); |
921 | return -1; | 921 | goto err; |
922 | } | 922 | } |
923 | /* End of block to be checked */ | 923 | /* End of block to be checked */ |
924 | 924 | ||
@@ -927,7 +927,7 @@ static int mos7840_open(struct tty_struct *tty, struct usb_serial_port *port) | |||
927 | &Data); | 927 | &Data); |
928 | if (status < 0) { | 928 | if (status < 0) { |
929 | dev_dbg(&port->dev, "Reading Controlreg failed\n"); | 929 | dev_dbg(&port->dev, "Reading Controlreg failed\n"); |
930 | return -1; | 930 | goto err; |
931 | } | 931 | } |
932 | Data |= 0x08; /* Driver done bit */ | 932 | Data |= 0x08; /* Driver done bit */ |
933 | Data |= 0x20; /* rx_disable */ | 933 | Data |= 0x20; /* rx_disable */ |
@@ -935,7 +935,7 @@ static int mos7840_open(struct tty_struct *tty, struct usb_serial_port *port) | |||
935 | mos7840_port->ControlRegOffset, Data); | 935 | mos7840_port->ControlRegOffset, Data); |
936 | if (status < 0) { | 936 | if (status < 0) { |
937 | dev_dbg(&port->dev, "writing Controlreg failed\n"); | 937 | dev_dbg(&port->dev, "writing Controlreg failed\n"); |
938 | return -1; | 938 | goto err; |
939 | } | 939 | } |
940 | /* do register settings here */ | 940 | /* do register settings here */ |
941 | /* Set all regs to the device default values. */ | 941 | /* Set all regs to the device default values. */ |
@@ -946,21 +946,21 @@ static int mos7840_open(struct tty_struct *tty, struct usb_serial_port *port) | |||
946 | status = mos7840_set_uart_reg(port, INTERRUPT_ENABLE_REGISTER, Data); | 946 | status = mos7840_set_uart_reg(port, INTERRUPT_ENABLE_REGISTER, Data); |
947 | if (status < 0) { | 947 | if (status < 0) { |
948 | dev_dbg(&port->dev, "disabling interrupts failed\n"); | 948 | dev_dbg(&port->dev, "disabling interrupts failed\n"); |
949 | return -1; | 949 | goto err; |
950 | } | 950 | } |
951 | /* Set FIFO_CONTROL_REGISTER to the default value */ | 951 | /* Set FIFO_CONTROL_REGISTER to the default value */ |
952 | Data = 0x00; | 952 | Data = 0x00; |
953 | status = mos7840_set_uart_reg(port, FIFO_CONTROL_REGISTER, Data); | 953 | status = mos7840_set_uart_reg(port, FIFO_CONTROL_REGISTER, Data); |
954 | if (status < 0) { | 954 | if (status < 0) { |
955 | dev_dbg(&port->dev, "Writing FIFO_CONTROL_REGISTER failed\n"); | 955 | dev_dbg(&port->dev, "Writing FIFO_CONTROL_REGISTER failed\n"); |
956 | return -1; | 956 | goto err; |
957 | } | 957 | } |
958 | 958 | ||
959 | Data = 0xcf; | 959 | Data = 0xcf; |
960 | status = mos7840_set_uart_reg(port, FIFO_CONTROL_REGISTER, Data); | 960 | status = mos7840_set_uart_reg(port, FIFO_CONTROL_REGISTER, Data); |
961 | if (status < 0) { | 961 | if (status < 0) { |
962 | dev_dbg(&port->dev, "Writing FIFO_CONTROL_REGISTER failed\n"); | 962 | dev_dbg(&port->dev, "Writing FIFO_CONTROL_REGISTER failed\n"); |
963 | return -1; | 963 | goto err; |
964 | } | 964 | } |
965 | 965 | ||
966 | Data = 0x03; | 966 | Data = 0x03; |
@@ -1103,6 +1103,15 @@ static int mos7840_open(struct tty_struct *tty, struct usb_serial_port *port) | |||
1103 | /* mos7840_change_port_settings(mos7840_port,old_termios); */ | 1103 | /* mos7840_change_port_settings(mos7840_port,old_termios); */ |
1104 | 1104 | ||
1105 | return 0; | 1105 | return 0; |
1106 | err: | ||
1107 | for (j = 0; j < NUM_URBS; ++j) { | ||
1108 | urb = mos7840_port->write_urb_pool[j]; | ||
1109 | if (!urb) | ||
1110 | continue; | ||
1111 | kfree(urb->transfer_buffer); | ||
1112 | usb_free_urb(urb); | ||
1113 | } | ||
1114 | return status; | ||
1106 | } | 1115 | } |
1107 | 1116 | ||
1108 | /***************************************************************************** | 1117 | /***************************************************************************** |
diff --git a/drivers/usb/serial/option.c b/drivers/usb/serial/option.c index 5dd857de05b0..1cf6f125f5f0 100644 --- a/drivers/usb/serial/option.c +++ b/drivers/usb/serial/option.c | |||
@@ -341,17 +341,12 @@ static void option_instat_callback(struct urb *urb); | |||
341 | #define OLIVETTI_VENDOR_ID 0x0b3c | 341 | #define OLIVETTI_VENDOR_ID 0x0b3c |
342 | #define OLIVETTI_PRODUCT_OLICARD100 0xc000 | 342 | #define OLIVETTI_PRODUCT_OLICARD100 0xc000 |
343 | #define OLIVETTI_PRODUCT_OLICARD145 0xc003 | 343 | #define OLIVETTI_PRODUCT_OLICARD145 0xc003 |
344 | #define OLIVETTI_PRODUCT_OLICARD200 0xc005 | ||
344 | 345 | ||
345 | /* Celot products */ | 346 | /* Celot products */ |
346 | #define CELOT_VENDOR_ID 0x211f | 347 | #define CELOT_VENDOR_ID 0x211f |
347 | #define CELOT_PRODUCT_CT680M 0x6801 | 348 | #define CELOT_PRODUCT_CT680M 0x6801 |
348 | 349 | ||
349 | /* ONDA Communication vendor id */ | ||
350 | #define ONDA_VENDOR_ID 0x1ee8 | ||
351 | |||
352 | /* ONDA MT825UP HSDPA 14.2 modem */ | ||
353 | #define ONDA_MT825UP 0x000b | ||
354 | |||
355 | /* Samsung products */ | 350 | /* Samsung products */ |
356 | #define SAMSUNG_VENDOR_ID 0x04e8 | 351 | #define SAMSUNG_VENDOR_ID 0x04e8 |
357 | #define SAMSUNG_PRODUCT_GT_B3730 0x6889 | 352 | #define SAMSUNG_PRODUCT_GT_B3730 0x6889 |
@@ -444,7 +439,8 @@ static void option_instat_callback(struct urb *urb); | |||
444 | 439 | ||
445 | /* Hyundai Petatel Inc. products */ | 440 | /* Hyundai Petatel Inc. products */ |
446 | #define PETATEL_VENDOR_ID 0x1ff4 | 441 | #define PETATEL_VENDOR_ID 0x1ff4 |
447 | #define PETATEL_PRODUCT_NP10T 0x600e | 442 | #define PETATEL_PRODUCT_NP10T_600A 0x600a |
443 | #define PETATEL_PRODUCT_NP10T_600E 0x600e | ||
448 | 444 | ||
449 | /* TP-LINK Incorporated products */ | 445 | /* TP-LINK Incorporated products */ |
450 | #define TPLINK_VENDOR_ID 0x2357 | 446 | #define TPLINK_VENDOR_ID 0x2357 |
@@ -782,6 +778,7 @@ static const struct usb_device_id option_ids[] = { | |||
782 | { USB_DEVICE(KYOCERA_VENDOR_ID, KYOCERA_PRODUCT_KPC650) }, | 778 | { USB_DEVICE(KYOCERA_VENDOR_ID, KYOCERA_PRODUCT_KPC650) }, |
783 | { USB_DEVICE(KYOCERA_VENDOR_ID, KYOCERA_PRODUCT_KPC680) }, | 779 | { USB_DEVICE(KYOCERA_VENDOR_ID, KYOCERA_PRODUCT_KPC680) }, |
784 | { USB_DEVICE(QUALCOMM_VENDOR_ID, 0x6613)}, /* Onda H600/ZTE MF330 */ | 780 | { USB_DEVICE(QUALCOMM_VENDOR_ID, 0x6613)}, /* Onda H600/ZTE MF330 */ |
781 | { USB_DEVICE(QUALCOMM_VENDOR_ID, 0x0023)}, /* ONYX 3G device */ | ||
785 | { USB_DEVICE(QUALCOMM_VENDOR_ID, 0x9000)}, /* SIMCom SIM5218 */ | 782 | { USB_DEVICE(QUALCOMM_VENDOR_ID, 0x9000)}, /* SIMCom SIM5218 */ |
786 | { USB_DEVICE(CMOTECH_VENDOR_ID, CMOTECH_PRODUCT_6280) }, /* BP3-USB & BP3-EXT HSDPA */ | 783 | { USB_DEVICE(CMOTECH_VENDOR_ID, CMOTECH_PRODUCT_6280) }, /* BP3-USB & BP3-EXT HSDPA */ |
787 | { USB_DEVICE(CMOTECH_VENDOR_ID, CMOTECH_PRODUCT_6008) }, | 784 | { USB_DEVICE(CMOTECH_VENDOR_ID, CMOTECH_PRODUCT_6008) }, |
@@ -817,7 +814,8 @@ static const struct usb_device_id option_ids[] = { | |||
817 | { USB_DEVICE_AND_INTERFACE_INFO(ZTE_VENDOR_ID, 0x0017, 0xff, 0xff, 0xff), | 814 | { USB_DEVICE_AND_INTERFACE_INFO(ZTE_VENDOR_ID, 0x0017, 0xff, 0xff, 0xff), |
818 | .driver_info = (kernel_ulong_t)&net_intf3_blacklist }, | 815 | .driver_info = (kernel_ulong_t)&net_intf3_blacklist }, |
819 | { USB_DEVICE_AND_INTERFACE_INFO(ZTE_VENDOR_ID, 0x0018, 0xff, 0xff, 0xff) }, | 816 | { USB_DEVICE_AND_INTERFACE_INFO(ZTE_VENDOR_ID, 0x0018, 0xff, 0xff, 0xff) }, |
820 | { USB_DEVICE_AND_INTERFACE_INFO(ZTE_VENDOR_ID, 0x0019, 0xff, 0xff, 0xff) }, | 817 | { USB_DEVICE_AND_INTERFACE_INFO(ZTE_VENDOR_ID, 0x0019, 0xff, 0xff, 0xff), |
818 | .driver_info = (kernel_ulong_t)&net_intf3_blacklist }, | ||
821 | { USB_DEVICE_AND_INTERFACE_INFO(ZTE_VENDOR_ID, 0x0020, 0xff, 0xff, 0xff) }, | 819 | { USB_DEVICE_AND_INTERFACE_INFO(ZTE_VENDOR_ID, 0x0020, 0xff, 0xff, 0xff) }, |
822 | { USB_DEVICE_AND_INTERFACE_INFO(ZTE_VENDOR_ID, 0x0021, 0xff, 0xff, 0xff), | 820 | { USB_DEVICE_AND_INTERFACE_INFO(ZTE_VENDOR_ID, 0x0021, 0xff, 0xff, 0xff), |
823 | .driver_info = (kernel_ulong_t)&net_intf4_blacklist }, | 821 | .driver_info = (kernel_ulong_t)&net_intf4_blacklist }, |
@@ -1256,8 +1254,8 @@ static const struct usb_device_id option_ids[] = { | |||
1256 | 1254 | ||
1257 | { USB_DEVICE(OLIVETTI_VENDOR_ID, OLIVETTI_PRODUCT_OLICARD100) }, | 1255 | { USB_DEVICE(OLIVETTI_VENDOR_ID, OLIVETTI_PRODUCT_OLICARD100) }, |
1258 | { USB_DEVICE(OLIVETTI_VENDOR_ID, OLIVETTI_PRODUCT_OLICARD145) }, | 1256 | { USB_DEVICE(OLIVETTI_VENDOR_ID, OLIVETTI_PRODUCT_OLICARD145) }, |
1257 | { USB_DEVICE(OLIVETTI_VENDOR_ID, OLIVETTI_PRODUCT_OLICARD200) }, | ||
1259 | { USB_DEVICE(CELOT_VENDOR_ID, CELOT_PRODUCT_CT680M) }, /* CT-650 CDMA 450 1xEVDO modem */ | 1258 | { USB_DEVICE(CELOT_VENDOR_ID, CELOT_PRODUCT_CT680M) }, /* CT-650 CDMA 450 1xEVDO modem */ |
1260 | { USB_DEVICE(ONDA_VENDOR_ID, ONDA_MT825UP) }, /* ONDA MT825UP modem */ | ||
1261 | { USB_DEVICE_AND_INTERFACE_INFO(SAMSUNG_VENDOR_ID, SAMSUNG_PRODUCT_GT_B3730, USB_CLASS_CDC_DATA, 0x00, 0x00) }, /* Samsung GT-B3730 LTE USB modem.*/ | 1259 | { USB_DEVICE_AND_INTERFACE_INFO(SAMSUNG_VENDOR_ID, SAMSUNG_PRODUCT_GT_B3730, USB_CLASS_CDC_DATA, 0x00, 0x00) }, /* Samsung GT-B3730 LTE USB modem.*/ |
1262 | { USB_DEVICE(YUGA_VENDOR_ID, YUGA_PRODUCT_CEM600) }, | 1260 | { USB_DEVICE(YUGA_VENDOR_ID, YUGA_PRODUCT_CEM600) }, |
1263 | { USB_DEVICE(YUGA_VENDOR_ID, YUGA_PRODUCT_CEM610) }, | 1261 | { USB_DEVICE(YUGA_VENDOR_ID, YUGA_PRODUCT_CEM610) }, |
@@ -1329,9 +1327,12 @@ static const struct usb_device_id option_ids[] = { | |||
1329 | { USB_DEVICE_AND_INTERFACE_INFO(MEDIATEK_VENDOR_ID, MEDIATEK_PRODUCT_DC_4COM2, 0xff, 0x02, 0x01) }, | 1327 | { USB_DEVICE_AND_INTERFACE_INFO(MEDIATEK_VENDOR_ID, MEDIATEK_PRODUCT_DC_4COM2, 0xff, 0x02, 0x01) }, |
1330 | { USB_DEVICE_AND_INTERFACE_INFO(MEDIATEK_VENDOR_ID, MEDIATEK_PRODUCT_DC_4COM2, 0xff, 0x00, 0x00) }, | 1328 | { USB_DEVICE_AND_INTERFACE_INFO(MEDIATEK_VENDOR_ID, MEDIATEK_PRODUCT_DC_4COM2, 0xff, 0x00, 0x00) }, |
1331 | { USB_DEVICE(CELLIENT_VENDOR_ID, CELLIENT_PRODUCT_MEN200) }, | 1329 | { USB_DEVICE(CELLIENT_VENDOR_ID, CELLIENT_PRODUCT_MEN200) }, |
1332 | { USB_DEVICE(PETATEL_VENDOR_ID, PETATEL_PRODUCT_NP10T) }, | 1330 | { USB_DEVICE(PETATEL_VENDOR_ID, PETATEL_PRODUCT_NP10T_600A) }, |
1331 | { USB_DEVICE(PETATEL_VENDOR_ID, PETATEL_PRODUCT_NP10T_600E) }, | ||
1333 | { USB_DEVICE(TPLINK_VENDOR_ID, TPLINK_PRODUCT_MA180), | 1332 | { USB_DEVICE(TPLINK_VENDOR_ID, TPLINK_PRODUCT_MA180), |
1334 | .driver_info = (kernel_ulong_t)&net_intf4_blacklist }, | 1333 | .driver_info = (kernel_ulong_t)&net_intf4_blacklist }, |
1334 | { USB_DEVICE(TPLINK_VENDOR_ID, 0x9000), /* TP-Link MA260 */ | ||
1335 | .driver_info = (kernel_ulong_t)&net_intf4_blacklist }, | ||
1335 | { USB_DEVICE(CHANGHONG_VENDOR_ID, CHANGHONG_PRODUCT_CH690) }, | 1336 | { USB_DEVICE(CHANGHONG_VENDOR_ID, CHANGHONG_PRODUCT_CH690) }, |
1336 | { USB_DEVICE_AND_INTERFACE_INFO(0x2001, 0x7d01, 0xff, 0x02, 0x01) }, /* D-Link DWM-156 (variant) */ | 1337 | { USB_DEVICE_AND_INTERFACE_INFO(0x2001, 0x7d01, 0xff, 0x02, 0x01) }, /* D-Link DWM-156 (variant) */ |
1337 | { USB_DEVICE_AND_INTERFACE_INFO(0x2001, 0x7d01, 0xff, 0x00, 0x00) }, /* D-Link DWM-156 (variant) */ | 1338 | { USB_DEVICE_AND_INTERFACE_INFO(0x2001, 0x7d01, 0xff, 0x00, 0x00) }, /* D-Link DWM-156 (variant) */ |
@@ -1339,6 +1340,8 @@ static const struct usb_device_id option_ids[] = { | |||
1339 | { USB_DEVICE_AND_INTERFACE_INFO(0x2001, 0x7d02, 0xff, 0x00, 0x00) }, | 1340 | { USB_DEVICE_AND_INTERFACE_INFO(0x2001, 0x7d02, 0xff, 0x00, 0x00) }, |
1340 | { USB_DEVICE_AND_INTERFACE_INFO(0x2001, 0x7d03, 0xff, 0x02, 0x01) }, | 1341 | { USB_DEVICE_AND_INTERFACE_INFO(0x2001, 0x7d03, 0xff, 0x02, 0x01) }, |
1341 | { USB_DEVICE_AND_INTERFACE_INFO(0x2001, 0x7d03, 0xff, 0x00, 0x00) }, | 1342 | { USB_DEVICE_AND_INTERFACE_INFO(0x2001, 0x7d03, 0xff, 0x00, 0x00) }, |
1343 | { USB_DEVICE_AND_INTERFACE_INFO(0x07d1, 0x3e01, 0xff, 0xff, 0xff) }, /* D-Link DWM-152/C1 */ | ||
1344 | { USB_DEVICE_AND_INTERFACE_INFO(0x07d1, 0x3e02, 0xff, 0xff, 0xff) }, /* D-Link DWM-156/C1 */ | ||
1342 | { } /* Terminating entry */ | 1345 | { } /* Terminating entry */ |
1343 | }; | 1346 | }; |
1344 | MODULE_DEVICE_TABLE(usb, option_ids); | 1347 | MODULE_DEVICE_TABLE(usb, option_ids); |
diff --git a/drivers/usb/serial/ti_usb_3410_5052.c b/drivers/usb/serial/ti_usb_3410_5052.c index 7182bb774b79..375b5a400b6f 100644 --- a/drivers/usb/serial/ti_usb_3410_5052.c +++ b/drivers/usb/serial/ti_usb_3410_5052.c | |||
@@ -371,7 +371,7 @@ static int ti_startup(struct usb_serial *serial) | |||
371 | usb_set_serial_data(serial, tdev); | 371 | usb_set_serial_data(serial, tdev); |
372 | 372 | ||
373 | /* determine device type */ | 373 | /* determine device type */ |
374 | if (usb_match_id(serial->interface, ti_id_table_3410)) | 374 | if (serial->type == &ti_1port_device) |
375 | tdev->td_is_3410 = 1; | 375 | tdev->td_is_3410 = 1; |
376 | dev_dbg(&dev->dev, "%s - device type is %s\n", __func__, | 376 | dev_dbg(&dev->dev, "%s - device type is %s\n", __func__, |
377 | tdev->td_is_3410 ? "3410" : "5052"); | 377 | tdev->td_is_3410 ? "3410" : "5052"); |
diff --git a/drivers/usb/storage/unusual_devs.h b/drivers/usb/storage/unusual_devs.h index 1799335288bd..c015f2c16729 100644 --- a/drivers/usb/storage/unusual_devs.h +++ b/drivers/usb/storage/unusual_devs.h | |||
@@ -665,6 +665,13 @@ UNUSUAL_DEV( 0x054c, 0x016a, 0x0000, 0x9999, | |||
665 | USB_SC_DEVICE, USB_PR_DEVICE, NULL, | 665 | USB_SC_DEVICE, USB_PR_DEVICE, NULL, |
666 | US_FL_FIX_INQUIRY ), | 666 | US_FL_FIX_INQUIRY ), |
667 | 667 | ||
668 | /* Submitted by Ren Bigcren <bigcren.ren@sonymobile.com> */ | ||
669 | UNUSUAL_DEV( 0x054c, 0x02a5, 0x0100, 0x0100, | ||
670 | "Sony Corp.", | ||
671 | "MicroVault Flash Drive", | ||
672 | USB_SC_DEVICE, USB_PR_DEVICE, NULL, | ||
673 | US_FL_NO_READ_CAPACITY_16 ), | ||
674 | |||
668 | /* floppy reports multiple luns */ | 675 | /* floppy reports multiple luns */ |
669 | UNUSUAL_DEV( 0x055d, 0x2020, 0x0000, 0x0210, | 676 | UNUSUAL_DEV( 0x055d, 0x2020, 0x0000, 0x0210, |
670 | "SAMSUNG", | 677 | "SAMSUNG", |
diff --git a/drivers/vhost/net.c b/drivers/vhost/net.c index 027be91db139..969a85960e9f 100644 --- a/drivers/vhost/net.c +++ b/drivers/vhost/net.c | |||
@@ -15,7 +15,6 @@ | |||
15 | #include <linux/moduleparam.h> | 15 | #include <linux/moduleparam.h> |
16 | #include <linux/mutex.h> | 16 | #include <linux/mutex.h> |
17 | #include <linux/workqueue.h> | 17 | #include <linux/workqueue.h> |
18 | #include <linux/rcupdate.h> | ||
19 | #include <linux/file.h> | 18 | #include <linux/file.h> |
20 | #include <linux/slab.h> | 19 | #include <linux/slab.h> |
21 | 20 | ||
@@ -346,12 +345,11 @@ static void handle_tx(struct vhost_net *net) | |||
346 | struct vhost_net_ubuf_ref *uninitialized_var(ubufs); | 345 | struct vhost_net_ubuf_ref *uninitialized_var(ubufs); |
347 | bool zcopy, zcopy_used; | 346 | bool zcopy, zcopy_used; |
348 | 347 | ||
349 | /* TODO: check that we are running from vhost_worker? */ | 348 | mutex_lock(&vq->mutex); |
350 | sock = rcu_dereference_check(vq->private_data, 1); | 349 | sock = vq->private_data; |
351 | if (!sock) | 350 | if (!sock) |
352 | return; | 351 | goto out; |
353 | 352 | ||
354 | mutex_lock(&vq->mutex); | ||
355 | vhost_disable_notify(&net->dev, vq); | 353 | vhost_disable_notify(&net->dev, vq); |
356 | 354 | ||
357 | hdr_size = nvq->vhost_hlen; | 355 | hdr_size = nvq->vhost_hlen; |
@@ -461,7 +459,7 @@ static void handle_tx(struct vhost_net *net) | |||
461 | break; | 459 | break; |
462 | } | 460 | } |
463 | } | 461 | } |
464 | 462 | out: | |
465 | mutex_unlock(&vq->mutex); | 463 | mutex_unlock(&vq->mutex); |
466 | } | 464 | } |
467 | 465 | ||
@@ -570,14 +568,14 @@ static void handle_rx(struct vhost_net *net) | |||
570 | s16 headcount; | 568 | s16 headcount; |
571 | size_t vhost_hlen, sock_hlen; | 569 | size_t vhost_hlen, sock_hlen; |
572 | size_t vhost_len, sock_len; | 570 | size_t vhost_len, sock_len; |
573 | /* TODO: check that we are running from vhost_worker? */ | 571 | struct socket *sock; |
574 | struct socket *sock = rcu_dereference_check(vq->private_data, 1); | ||
575 | |||
576 | if (!sock) | ||
577 | return; | ||
578 | 572 | ||
579 | mutex_lock(&vq->mutex); | 573 | mutex_lock(&vq->mutex); |
574 | sock = vq->private_data; | ||
575 | if (!sock) | ||
576 | goto out; | ||
580 | vhost_disable_notify(&net->dev, vq); | 577 | vhost_disable_notify(&net->dev, vq); |
578 | |||
581 | vhost_hlen = nvq->vhost_hlen; | 579 | vhost_hlen = nvq->vhost_hlen; |
582 | sock_hlen = nvq->sock_hlen; | 580 | sock_hlen = nvq->sock_hlen; |
583 | 581 | ||
@@ -652,7 +650,7 @@ static void handle_rx(struct vhost_net *net) | |||
652 | break; | 650 | break; |
653 | } | 651 | } |
654 | } | 652 | } |
655 | 653 | out: | |
656 | mutex_unlock(&vq->mutex); | 654 | mutex_unlock(&vq->mutex); |
657 | } | 655 | } |
658 | 656 | ||
@@ -750,8 +748,7 @@ static int vhost_net_enable_vq(struct vhost_net *n, | |||
750 | struct vhost_poll *poll = n->poll + (nvq - n->vqs); | 748 | struct vhost_poll *poll = n->poll + (nvq - n->vqs); |
751 | struct socket *sock; | 749 | struct socket *sock; |
752 | 750 | ||
753 | sock = rcu_dereference_protected(vq->private_data, | 751 | sock = vq->private_data; |
754 | lockdep_is_held(&vq->mutex)); | ||
755 | if (!sock) | 752 | if (!sock) |
756 | return 0; | 753 | return 0; |
757 | 754 | ||
@@ -764,10 +761,9 @@ static struct socket *vhost_net_stop_vq(struct vhost_net *n, | |||
764 | struct socket *sock; | 761 | struct socket *sock; |
765 | 762 | ||
766 | mutex_lock(&vq->mutex); | 763 | mutex_lock(&vq->mutex); |
767 | sock = rcu_dereference_protected(vq->private_data, | 764 | sock = vq->private_data; |
768 | lockdep_is_held(&vq->mutex)); | ||
769 | vhost_net_disable_vq(n, vq); | 765 | vhost_net_disable_vq(n, vq); |
770 | rcu_assign_pointer(vq->private_data, NULL); | 766 | vq->private_data = NULL; |
771 | mutex_unlock(&vq->mutex); | 767 | mutex_unlock(&vq->mutex); |
772 | return sock; | 768 | return sock; |
773 | } | 769 | } |
@@ -923,8 +919,7 @@ static long vhost_net_set_backend(struct vhost_net *n, unsigned index, int fd) | |||
923 | } | 919 | } |
924 | 920 | ||
925 | /* start polling new socket */ | 921 | /* start polling new socket */ |
926 | oldsock = rcu_dereference_protected(vq->private_data, | 922 | oldsock = vq->private_data; |
927 | lockdep_is_held(&vq->mutex)); | ||
928 | if (sock != oldsock) { | 923 | if (sock != oldsock) { |
929 | ubufs = vhost_net_ubuf_alloc(vq, | 924 | ubufs = vhost_net_ubuf_alloc(vq, |
930 | sock && vhost_sock_zcopy(sock)); | 925 | sock && vhost_sock_zcopy(sock)); |
@@ -934,7 +929,7 @@ static long vhost_net_set_backend(struct vhost_net *n, unsigned index, int fd) | |||
934 | } | 929 | } |
935 | 930 | ||
936 | vhost_net_disable_vq(n, vq); | 931 | vhost_net_disable_vq(n, vq); |
937 | rcu_assign_pointer(vq->private_data, sock); | 932 | vq->private_data = sock; |
938 | r = vhost_init_used(vq); | 933 | r = vhost_init_used(vq); |
939 | if (r) | 934 | if (r) |
940 | goto err_used; | 935 | goto err_used; |
@@ -968,7 +963,7 @@ static long vhost_net_set_backend(struct vhost_net *n, unsigned index, int fd) | |||
968 | return 0; | 963 | return 0; |
969 | 964 | ||
970 | err_used: | 965 | err_used: |
971 | rcu_assign_pointer(vq->private_data, oldsock); | 966 | vq->private_data = oldsock; |
972 | vhost_net_enable_vq(n, vq); | 967 | vhost_net_enable_vq(n, vq); |
973 | if (ubufs) | 968 | if (ubufs) |
974 | vhost_net_ubuf_put_wait_and_free(ubufs); | 969 | vhost_net_ubuf_put_wait_and_free(ubufs); |
diff --git a/drivers/vhost/scsi.c b/drivers/vhost/scsi.c index 06adf31a9248..0c27c7df1b09 100644 --- a/drivers/vhost/scsi.c +++ b/drivers/vhost/scsi.c | |||
@@ -902,19 +902,15 @@ vhost_scsi_handle_vq(struct vhost_scsi *vs, struct vhost_virtqueue *vq) | |||
902 | int head, ret; | 902 | int head, ret; |
903 | u8 target; | 903 | u8 target; |
904 | 904 | ||
905 | mutex_lock(&vq->mutex); | ||
905 | /* | 906 | /* |
906 | * We can handle the vq only after the endpoint is setup by calling the | 907 | * We can handle the vq only after the endpoint is setup by calling the |
907 | * VHOST_SCSI_SET_ENDPOINT ioctl. | 908 | * VHOST_SCSI_SET_ENDPOINT ioctl. |
908 | * | ||
909 | * TODO: Check that we are running from vhost_worker which acts | ||
910 | * as read-side critical section for vhost kind of RCU. | ||
911 | * See the comments in struct vhost_virtqueue in drivers/vhost/vhost.h | ||
912 | */ | 909 | */ |
913 | vs_tpg = rcu_dereference_check(vq->private_data, 1); | 910 | vs_tpg = vq->private_data; |
914 | if (!vs_tpg) | 911 | if (!vs_tpg) |
915 | return; | 912 | goto out; |
916 | 913 | ||
917 | mutex_lock(&vq->mutex); | ||
918 | vhost_disable_notify(&vs->dev, vq); | 914 | vhost_disable_notify(&vs->dev, vq); |
919 | 915 | ||
920 | for (;;) { | 916 | for (;;) { |
@@ -1064,6 +1060,7 @@ err_free: | |||
1064 | vhost_scsi_free_cmd(cmd); | 1060 | vhost_scsi_free_cmd(cmd); |
1065 | err_cmd: | 1061 | err_cmd: |
1066 | vhost_scsi_send_bad_target(vs, vq, head, out); | 1062 | vhost_scsi_send_bad_target(vs, vq, head, out); |
1063 | out: | ||
1067 | mutex_unlock(&vq->mutex); | 1064 | mutex_unlock(&vq->mutex); |
1068 | } | 1065 | } |
1069 | 1066 | ||
@@ -1232,9 +1229,8 @@ vhost_scsi_set_endpoint(struct vhost_scsi *vs, | |||
1232 | sizeof(vs->vs_vhost_wwpn)); | 1229 | sizeof(vs->vs_vhost_wwpn)); |
1233 | for (i = 0; i < VHOST_SCSI_MAX_VQ; i++) { | 1230 | for (i = 0; i < VHOST_SCSI_MAX_VQ; i++) { |
1234 | vq = &vs->vqs[i].vq; | 1231 | vq = &vs->vqs[i].vq; |
1235 | /* Flushing the vhost_work acts as synchronize_rcu */ | ||
1236 | mutex_lock(&vq->mutex); | 1232 | mutex_lock(&vq->mutex); |
1237 | rcu_assign_pointer(vq->private_data, vs_tpg); | 1233 | vq->private_data = vs_tpg; |
1238 | vhost_init_used(vq); | 1234 | vhost_init_used(vq); |
1239 | mutex_unlock(&vq->mutex); | 1235 | mutex_unlock(&vq->mutex); |
1240 | } | 1236 | } |
@@ -1313,9 +1309,8 @@ vhost_scsi_clear_endpoint(struct vhost_scsi *vs, | |||
1313 | if (match) { | 1309 | if (match) { |
1314 | for (i = 0; i < VHOST_SCSI_MAX_VQ; i++) { | 1310 | for (i = 0; i < VHOST_SCSI_MAX_VQ; i++) { |
1315 | vq = &vs->vqs[i].vq; | 1311 | vq = &vs->vqs[i].vq; |
1316 | /* Flushing the vhost_work acts as synchronize_rcu */ | ||
1317 | mutex_lock(&vq->mutex); | 1312 | mutex_lock(&vq->mutex); |
1318 | rcu_assign_pointer(vq->private_data, NULL); | 1313 | vq->private_data = NULL; |
1319 | mutex_unlock(&vq->mutex); | 1314 | mutex_unlock(&vq->mutex); |
1320 | } | 1315 | } |
1321 | } | 1316 | } |
diff --git a/drivers/vhost/test.c b/drivers/vhost/test.c index a73ea217f24d..339eae85859a 100644 --- a/drivers/vhost/test.c +++ b/drivers/vhost/test.c | |||
@@ -13,7 +13,6 @@ | |||
13 | #include <linux/module.h> | 13 | #include <linux/module.h> |
14 | #include <linux/mutex.h> | 14 | #include <linux/mutex.h> |
15 | #include <linux/workqueue.h> | 15 | #include <linux/workqueue.h> |
16 | #include <linux/rcupdate.h> | ||
17 | #include <linux/file.h> | 16 | #include <linux/file.h> |
18 | #include <linux/slab.h> | 17 | #include <linux/slab.h> |
19 | 18 | ||
@@ -200,9 +199,8 @@ static long vhost_test_run(struct vhost_test *n, int test) | |||
200 | priv = test ? n : NULL; | 199 | priv = test ? n : NULL; |
201 | 200 | ||
202 | /* start polling new socket */ | 201 | /* start polling new socket */ |
203 | oldpriv = rcu_dereference_protected(vq->private_data, | 202 | oldpriv = vq->private_data; |
204 | lockdep_is_held(&vq->mutex)); | 203 | vq->private_data = priv; |
205 | rcu_assign_pointer(vq->private_data, priv); | ||
206 | 204 | ||
207 | r = vhost_init_used(&n->vqs[index]); | 205 | r = vhost_init_used(&n->vqs[index]); |
208 | 206 | ||
diff --git a/drivers/vhost/vhost.h b/drivers/vhost/vhost.h index 42298cd23c73..4465ed5f316d 100644 --- a/drivers/vhost/vhost.h +++ b/drivers/vhost/vhost.h | |||
@@ -103,14 +103,8 @@ struct vhost_virtqueue { | |||
103 | struct iovec iov[UIO_MAXIOV]; | 103 | struct iovec iov[UIO_MAXIOV]; |
104 | struct iovec *indirect; | 104 | struct iovec *indirect; |
105 | struct vring_used_elem *heads; | 105 | struct vring_used_elem *heads; |
106 | /* We use a kind of RCU to access private pointer. | 106 | /* Protected by virtqueue mutex. */ |
107 | * All readers access it from worker, which makes it possible to | 107 | void *private_data; |
108 | * flush the vhost_work instead of synchronize_rcu. Therefore readers do | ||
109 | * not need to call rcu_read_lock/rcu_read_unlock: the beginning of | ||
110 | * vhost_work execution acts instead of rcu_read_lock() and the end of | ||
111 | * vhost_work execution acts instead of rcu_read_unlock(). | ||
112 | * Writers use virtqueue mutex. */ | ||
113 | void __rcu *private_data; | ||
114 | /* Log write descriptors */ | 108 | /* Log write descriptors */ |
115 | void __user *log_base; | 109 | void __user *log_base; |
116 | struct vhost_log *log; | 110 | struct vhost_log *log; |
diff --git a/drivers/video/backlight/max8925_bl.c b/drivers/video/backlight/max8925_bl.c index 5ca11b066b7e..886e797f75f9 100644 --- a/drivers/video/backlight/max8925_bl.c +++ b/drivers/video/backlight/max8925_bl.c | |||
@@ -101,33 +101,37 @@ static const struct backlight_ops max8925_backlight_ops = { | |||
101 | .get_brightness = max8925_backlight_get_brightness, | 101 | .get_brightness = max8925_backlight_get_brightness, |
102 | }; | 102 | }; |
103 | 103 | ||
104 | #ifdef CONFIG_OF | 104 | static void max8925_backlight_dt_init(struct platform_device *pdev) |
105 | static int max8925_backlight_dt_init(struct platform_device *pdev, | ||
106 | struct max8925_backlight_pdata *pdata) | ||
107 | { | 105 | { |
108 | struct device_node *nproot = pdev->dev.parent->of_node, *np; | 106 | struct device_node *nproot = pdev->dev.parent->of_node, *np; |
109 | int dual_string; | 107 | struct max8925_backlight_pdata *pdata; |
108 | u32 val; | ||
109 | |||
110 | if (!nproot || !IS_ENABLED(CONFIG_OF)) | ||
111 | return; | ||
112 | |||
113 | pdata = devm_kzalloc(&pdev->dev, | ||
114 | sizeof(struct max8925_backlight_pdata), | ||
115 | GFP_KERNEL); | ||
116 | if (!pdata) | ||
117 | return; | ||
110 | 118 | ||
111 | if (!nproot) | ||
112 | return -ENODEV; | ||
113 | np = of_find_node_by_name(nproot, "backlight"); | 119 | np = of_find_node_by_name(nproot, "backlight"); |
114 | if (!np) { | 120 | if (!np) { |
115 | dev_err(&pdev->dev, "failed to find backlight node\n"); | 121 | dev_err(&pdev->dev, "failed to find backlight node\n"); |
116 | return -ENODEV; | 122 | return; |
117 | } | 123 | } |
118 | 124 | ||
119 | of_property_read_u32(np, "maxim,max8925-dual-string", &dual_string); | 125 | if (!of_property_read_u32(np, "maxim,max8925-dual-string", &val)) |
120 | pdata->dual_string = dual_string; | 126 | pdata->dual_string = val; |
121 | return 0; | 127 | |
128 | pdev->dev.platform_data = pdata; | ||
122 | } | 129 | } |
123 | #else | ||
124 | #define max8925_backlight_dt_init(x, y) (-1) | ||
125 | #endif | ||
126 | 130 | ||
127 | static int max8925_backlight_probe(struct platform_device *pdev) | 131 | static int max8925_backlight_probe(struct platform_device *pdev) |
128 | { | 132 | { |
129 | struct max8925_chip *chip = dev_get_drvdata(pdev->dev.parent); | 133 | struct max8925_chip *chip = dev_get_drvdata(pdev->dev.parent); |
130 | struct max8925_backlight_pdata *pdata = pdev->dev.platform_data; | 134 | struct max8925_backlight_pdata *pdata; |
131 | struct max8925_backlight_data *data; | 135 | struct max8925_backlight_data *data; |
132 | struct backlight_device *bl; | 136 | struct backlight_device *bl; |
133 | struct backlight_properties props; | 137 | struct backlight_properties props; |
@@ -170,13 +174,10 @@ static int max8925_backlight_probe(struct platform_device *pdev) | |||
170 | platform_set_drvdata(pdev, bl); | 174 | platform_set_drvdata(pdev, bl); |
171 | 175 | ||
172 | value = 0; | 176 | value = 0; |
173 | if (pdev->dev.parent->of_node && !pdata) { | 177 | if (!pdev->dev.platform_data) |
174 | pdata = devm_kzalloc(&pdev->dev, | 178 | max8925_backlight_dt_init(pdev); |
175 | sizeof(struct max8925_backlight_pdata), | ||
176 | GFP_KERNEL); | ||
177 | max8925_backlight_dt_init(pdev, pdata); | ||
178 | } | ||
179 | 179 | ||
180 | pdata = pdev->dev.platform_data; | ||
180 | if (pdata) { | 181 | if (pdata) { |
181 | if (pdata->lxw_scl) | 182 | if (pdata->lxw_scl) |
182 | value |= (1 << 7); | 183 | value |= (1 << 7); |
diff --git a/fs/fuse/dir.c b/fs/fuse/dir.c index 0eda52738ec4..72a5d5b04494 100644 --- a/fs/fuse/dir.c +++ b/fs/fuse/dir.c | |||
@@ -1223,30 +1223,46 @@ static int fuse_direntplus_link(struct file *file, | |||
1223 | if (name.name[1] == '.' && name.len == 2) | 1223 | if (name.name[1] == '.' && name.len == 2) |
1224 | return 0; | 1224 | return 0; |
1225 | } | 1225 | } |
1226 | |||
1227 | if (invalid_nodeid(o->nodeid)) | ||
1228 | return -EIO; | ||
1229 | if (!fuse_valid_type(o->attr.mode)) | ||
1230 | return -EIO; | ||
1231 | |||
1226 | fc = get_fuse_conn(dir); | 1232 | fc = get_fuse_conn(dir); |
1227 | 1233 | ||
1228 | name.hash = full_name_hash(name.name, name.len); | 1234 | name.hash = full_name_hash(name.name, name.len); |
1229 | dentry = d_lookup(parent, &name); | 1235 | dentry = d_lookup(parent, &name); |
1230 | if (dentry && dentry->d_inode) { | 1236 | if (dentry) { |
1231 | inode = dentry->d_inode; | 1237 | inode = dentry->d_inode; |
1232 | if (get_node_id(inode) == o->nodeid) { | 1238 | if (!inode) { |
1239 | d_drop(dentry); | ||
1240 | } else if (get_node_id(inode) != o->nodeid || | ||
1241 | ((o->attr.mode ^ inode->i_mode) & S_IFMT)) { | ||
1242 | err = d_invalidate(dentry); | ||
1243 | if (err) | ||
1244 | goto out; | ||
1245 | } else if (is_bad_inode(inode)) { | ||
1246 | err = -EIO; | ||
1247 | goto out; | ||
1248 | } else { | ||
1233 | struct fuse_inode *fi; | 1249 | struct fuse_inode *fi; |
1234 | fi = get_fuse_inode(inode); | 1250 | fi = get_fuse_inode(inode); |
1235 | spin_lock(&fc->lock); | 1251 | spin_lock(&fc->lock); |
1236 | fi->nlookup++; | 1252 | fi->nlookup++; |
1237 | spin_unlock(&fc->lock); | 1253 | spin_unlock(&fc->lock); |
1238 | 1254 | ||
1255 | fuse_change_attributes(inode, &o->attr, | ||
1256 | entry_attr_timeout(o), | ||
1257 | attr_version); | ||
1258 | |||
1239 | /* | 1259 | /* |
1240 | * The other branch to 'found' comes via fuse_iget() | 1260 | * The other branch to 'found' comes via fuse_iget() |
1241 | * which bumps nlookup inside | 1261 | * which bumps nlookup inside |
1242 | */ | 1262 | */ |
1243 | goto found; | 1263 | goto found; |
1244 | } | 1264 | } |
1245 | err = d_invalidate(dentry); | ||
1246 | if (err) | ||
1247 | goto out; | ||
1248 | dput(dentry); | 1265 | dput(dentry); |
1249 | dentry = NULL; | ||
1250 | } | 1266 | } |
1251 | 1267 | ||
1252 | dentry = d_alloc(parent, &name); | 1268 | dentry = d_alloc(parent, &name); |
@@ -1259,25 +1275,30 @@ static int fuse_direntplus_link(struct file *file, | |||
1259 | if (!inode) | 1275 | if (!inode) |
1260 | goto out; | 1276 | goto out; |
1261 | 1277 | ||
1262 | alias = d_materialise_unique(dentry, inode); | 1278 | if (S_ISDIR(inode->i_mode)) { |
1263 | err = PTR_ERR(alias); | 1279 | mutex_lock(&fc->inst_mutex); |
1264 | if (IS_ERR(alias)) | 1280 | alias = fuse_d_add_directory(dentry, inode); |
1265 | goto out; | 1281 | mutex_unlock(&fc->inst_mutex); |
1282 | err = PTR_ERR(alias); | ||
1283 | if (IS_ERR(alias)) { | ||
1284 | iput(inode); | ||
1285 | goto out; | ||
1286 | } | ||
1287 | } else { | ||
1288 | alias = d_splice_alias(inode, dentry); | ||
1289 | } | ||
1290 | |||
1266 | if (alias) { | 1291 | if (alias) { |
1267 | dput(dentry); | 1292 | dput(dentry); |
1268 | dentry = alias; | 1293 | dentry = alias; |
1269 | } | 1294 | } |
1270 | 1295 | ||
1271 | found: | 1296 | found: |
1272 | fuse_change_attributes(inode, &o->attr, entry_attr_timeout(o), | ||
1273 | attr_version); | ||
1274 | |||
1275 | fuse_change_entry_timeout(dentry, o); | 1297 | fuse_change_entry_timeout(dentry, o); |
1276 | 1298 | ||
1277 | err = 0; | 1299 | err = 0; |
1278 | out: | 1300 | out: |
1279 | if (dentry) | 1301 | dput(dentry); |
1280 | dput(dentry); | ||
1281 | return err; | 1302 | return err; |
1282 | } | 1303 | } |
1283 | 1304 | ||
diff --git a/fs/nfs/nfs4xdr.c b/fs/nfs/nfs4xdr.c index c74d6168db99..3850b018815f 100644 --- a/fs/nfs/nfs4xdr.c +++ b/fs/nfs/nfs4xdr.c | |||
@@ -1118,11 +1118,11 @@ static void encode_attrs(struct xdr_stream *xdr, const struct iattr *iap, | |||
1118 | len, ((char *)p - (char *)q) + 4); | 1118 | len, ((char *)p - (char *)q) + 4); |
1119 | BUG(); | 1119 | BUG(); |
1120 | } | 1120 | } |
1121 | len = (char *)p - (char *)q - (bmval_len << 2); | ||
1122 | *q++ = htonl(bmval0); | 1121 | *q++ = htonl(bmval0); |
1123 | *q++ = htonl(bmval1); | 1122 | *q++ = htonl(bmval1); |
1124 | if (bmval_len == 3) | 1123 | if (bmval_len == 3) |
1125 | *q++ = htonl(bmval2); | 1124 | *q++ = htonl(bmval2); |
1125 | len = (char *)p - (char *)(q + 1); | ||
1126 | *q = htonl(len); | 1126 | *q = htonl(len); |
1127 | 1127 | ||
1128 | /* out: */ | 1128 | /* out: */ |
diff --git a/fs/nfsd/vfs.c b/fs/nfsd/vfs.c index 8ff6a0019b0b..c827acb0e943 100644 --- a/fs/nfsd/vfs.c +++ b/fs/nfsd/vfs.c | |||
@@ -830,9 +830,10 @@ nfsd_open(struct svc_rqst *rqstp, struct svc_fh *fhp, umode_t type, | |||
830 | flags = O_WRONLY|O_LARGEFILE; | 830 | flags = O_WRONLY|O_LARGEFILE; |
831 | } | 831 | } |
832 | *filp = dentry_open(&path, flags, current_cred()); | 832 | *filp = dentry_open(&path, flags, current_cred()); |
833 | if (IS_ERR(*filp)) | 833 | if (IS_ERR(*filp)) { |
834 | host_err = PTR_ERR(*filp); | 834 | host_err = PTR_ERR(*filp); |
835 | else { | 835 | *filp = NULL; |
836 | } else { | ||
836 | host_err = ima_file_check(*filp, may_flags); | 837 | host_err = ima_file_check(*filp, may_flags); |
837 | 838 | ||
838 | if (may_flags & NFSD_MAY_64BIT_COOKIE) | 839 | if (may_flags & NFSD_MAY_64BIT_COOKIE) |
diff --git a/fs/xfs/xfs_dinode.h b/fs/xfs/xfs_dinode.h index 07d735a80a0f..e5869b50dc41 100644 --- a/fs/xfs/xfs_dinode.h +++ b/fs/xfs/xfs_dinode.h | |||
@@ -39,6 +39,9 @@ typedef struct xfs_timestamp { | |||
39 | * There is a very similar struct icdinode in xfs_inode which matches the | 39 | * There is a very similar struct icdinode in xfs_inode which matches the |
40 | * layout of the first 96 bytes of this structure, but is kept in native | 40 | * layout of the first 96 bytes of this structure, but is kept in native |
41 | * format instead of big endian. | 41 | * format instead of big endian. |
42 | * | ||
43 | * Note: di_flushiter is only used by v1/2 inodes - it's effectively a zeroed | ||
44 | * padding field for v3 inodes. | ||
42 | */ | 45 | */ |
43 | typedef struct xfs_dinode { | 46 | typedef struct xfs_dinode { |
44 | __be16 di_magic; /* inode magic # = XFS_DINODE_MAGIC */ | 47 | __be16 di_magic; /* inode magic # = XFS_DINODE_MAGIC */ |
diff --git a/fs/xfs/xfs_inode.c b/fs/xfs/xfs_inode.c index b78481f99d9d..bb262c25c8de 100644 --- a/fs/xfs/xfs_inode.c +++ b/fs/xfs/xfs_inode.c | |||
@@ -896,7 +896,6 @@ xfs_dinode_to_disk( | |||
896 | to->di_projid_lo = cpu_to_be16(from->di_projid_lo); | 896 | to->di_projid_lo = cpu_to_be16(from->di_projid_lo); |
897 | to->di_projid_hi = cpu_to_be16(from->di_projid_hi); | 897 | to->di_projid_hi = cpu_to_be16(from->di_projid_hi); |
898 | memcpy(to->di_pad, from->di_pad, sizeof(to->di_pad)); | 898 | memcpy(to->di_pad, from->di_pad, sizeof(to->di_pad)); |
899 | to->di_flushiter = cpu_to_be16(from->di_flushiter); | ||
900 | to->di_atime.t_sec = cpu_to_be32(from->di_atime.t_sec); | 899 | to->di_atime.t_sec = cpu_to_be32(from->di_atime.t_sec); |
901 | to->di_atime.t_nsec = cpu_to_be32(from->di_atime.t_nsec); | 900 | to->di_atime.t_nsec = cpu_to_be32(from->di_atime.t_nsec); |
902 | to->di_mtime.t_sec = cpu_to_be32(from->di_mtime.t_sec); | 901 | to->di_mtime.t_sec = cpu_to_be32(from->di_mtime.t_sec); |
@@ -924,6 +923,9 @@ xfs_dinode_to_disk( | |||
924 | to->di_lsn = cpu_to_be64(from->di_lsn); | 923 | to->di_lsn = cpu_to_be64(from->di_lsn); |
925 | memcpy(to->di_pad2, from->di_pad2, sizeof(to->di_pad2)); | 924 | memcpy(to->di_pad2, from->di_pad2, sizeof(to->di_pad2)); |
926 | uuid_copy(&to->di_uuid, &from->di_uuid); | 925 | uuid_copy(&to->di_uuid, &from->di_uuid); |
926 | to->di_flushiter = 0; | ||
927 | } else { | ||
928 | to->di_flushiter = cpu_to_be16(from->di_flushiter); | ||
927 | } | 929 | } |
928 | } | 930 | } |
929 | 931 | ||
@@ -1029,10 +1031,14 @@ xfs_dinode_calc_crc( | |||
1029 | /* | 1031 | /* |
1030 | * Read the disk inode attributes into the in-core inode structure. | 1032 | * Read the disk inode attributes into the in-core inode structure. |
1031 | * | 1033 | * |
1032 | * If we are initialising a new inode and we are not utilising the | 1034 | * For version 5 superblocks, if we are initialising a new inode and we are not |
1033 | * XFS_MOUNT_IKEEP inode cluster mode, we can simple build the new inode core | 1035 | * utilising the XFS_MOUNT_IKEEP inode cluster mode, we can simple build the new |
1034 | * with a random generation number. If we are keeping inodes around, we need to | 1036 | * inode core with a random generation number. If we are keeping inodes around, |
1035 | * read the inode cluster to get the existing generation number off disk. | 1037 | * we need to read the inode cluster to get the existing generation number off |
1038 | * disk. Further, if we are using version 4 superblocks (i.e. v1/v2 inode | ||
1039 | * format) then log recovery is dependent on the di_flushiter field being | ||
1040 | * initialised from the current on-disk value and hence we must also read the | ||
1041 | * inode off disk. | ||
1036 | */ | 1042 | */ |
1037 | int | 1043 | int |
1038 | xfs_iread( | 1044 | xfs_iread( |
@@ -1054,6 +1060,7 @@ xfs_iread( | |||
1054 | 1060 | ||
1055 | /* shortcut IO on inode allocation if possible */ | 1061 | /* shortcut IO on inode allocation if possible */ |
1056 | if ((iget_flags & XFS_IGET_CREATE) && | 1062 | if ((iget_flags & XFS_IGET_CREATE) && |
1063 | xfs_sb_version_hascrc(&mp->m_sb) && | ||
1057 | !(mp->m_flags & XFS_MOUNT_IKEEP)) { | 1064 | !(mp->m_flags & XFS_MOUNT_IKEEP)) { |
1058 | /* initialise the on-disk inode core */ | 1065 | /* initialise the on-disk inode core */ |
1059 | memset(&ip->i_d, 0, sizeof(ip->i_d)); | 1066 | memset(&ip->i_d, 0, sizeof(ip->i_d)); |
@@ -2882,12 +2889,18 @@ xfs_iflush_int( | |||
2882 | __func__, ip->i_ino, ip->i_d.di_forkoff, ip); | 2889 | __func__, ip->i_ino, ip->i_d.di_forkoff, ip); |
2883 | goto corrupt_out; | 2890 | goto corrupt_out; |
2884 | } | 2891 | } |
2892 | |||
2885 | /* | 2893 | /* |
2886 | * bump the flush iteration count, used to detect flushes which | 2894 | * Inode item log recovery for v1/v2 inodes are dependent on the |
2887 | * postdate a log record during recovery. This is redundant as we now | 2895 | * di_flushiter count for correct sequencing. We bump the flush |
2888 | * log every change and hence this can't happen. Still, it doesn't hurt. | 2896 | * iteration count so we can detect flushes which postdate a log record |
2897 | * during recovery. This is redundant as we now log every change and | ||
2898 | * hence this can't happen but we need to still do it to ensure | ||
2899 | * backwards compatibility with old kernels that predate logging all | ||
2900 | * inode changes. | ||
2889 | */ | 2901 | */ |
2890 | ip->i_d.di_flushiter++; | 2902 | if (ip->i_d.di_version < 3) |
2903 | ip->i_d.di_flushiter++; | ||
2891 | 2904 | ||
2892 | /* | 2905 | /* |
2893 | * Copy the dirty parts of the inode into the on-disk | 2906 | * Copy the dirty parts of the inode into the on-disk |
diff --git a/fs/xfs/xfs_log_recover.c b/fs/xfs/xfs_log_recover.c index 6fcc910a50b9..7681b19aa5dc 100644 --- a/fs/xfs/xfs_log_recover.c +++ b/fs/xfs/xfs_log_recover.c | |||
@@ -2592,8 +2592,16 @@ xlog_recover_inode_pass2( | |||
2592 | goto error; | 2592 | goto error; |
2593 | } | 2593 | } |
2594 | 2594 | ||
2595 | /* Skip replay when the on disk inode is newer than the log one */ | 2595 | /* |
2596 | if (dicp->di_flushiter < be16_to_cpu(dip->di_flushiter)) { | 2596 | * di_flushiter is only valid for v1/2 inodes. All changes for v3 inodes |
2597 | * are transactional and if ordering is necessary we can determine that | ||
2598 | * more accurately by the LSN field in the V3 inode core. Don't trust | ||
2599 | * the inode versions we might be changing them here - use the | ||
2600 | * superblock flag to determine whether we need to look at di_flushiter | ||
2601 | * to skip replay when the on disk inode is newer than the log one | ||
2602 | */ | ||
2603 | if (!xfs_sb_version_hascrc(&mp->m_sb) && | ||
2604 | dicp->di_flushiter < be16_to_cpu(dip->di_flushiter)) { | ||
2597 | /* | 2605 | /* |
2598 | * Deal with the wrap case, DI_MAX_FLUSH is less | 2606 | * Deal with the wrap case, DI_MAX_FLUSH is less |
2599 | * than smaller numbers | 2607 | * than smaller numbers |
@@ -2608,6 +2616,7 @@ xlog_recover_inode_pass2( | |||
2608 | goto error; | 2616 | goto error; |
2609 | } | 2617 | } |
2610 | } | 2618 | } |
2619 | |||
2611 | /* Take the opportunity to reset the flush iteration count */ | 2620 | /* Take the opportunity to reset the flush iteration count */ |
2612 | dicp->di_flushiter = 0; | 2621 | dicp->di_flushiter = 0; |
2613 | 2622 | ||
diff --git a/include/acpi/video.h b/include/acpi/video.h index b26dc4fb7ba8..61109f2609fc 100644 --- a/include/acpi/video.h +++ b/include/acpi/video.h | |||
@@ -17,21 +17,12 @@ struct acpi_device; | |||
17 | #define ACPI_VIDEO_DISPLAY_LEGACY_TV 0x0200 | 17 | #define ACPI_VIDEO_DISPLAY_LEGACY_TV 0x0200 |
18 | 18 | ||
19 | #if (defined CONFIG_ACPI_VIDEO || defined CONFIG_ACPI_VIDEO_MODULE) | 19 | #if (defined CONFIG_ACPI_VIDEO || defined CONFIG_ACPI_VIDEO_MODULE) |
20 | extern int __acpi_video_register(bool backlight_quirks); | 20 | extern int acpi_video_register(void); |
21 | static inline int acpi_video_register(void) | ||
22 | { | ||
23 | return __acpi_video_register(false); | ||
24 | } | ||
25 | static inline int acpi_video_register_with_quirks(void) | ||
26 | { | ||
27 | return __acpi_video_register(true); | ||
28 | } | ||
29 | extern void acpi_video_unregister(void); | 21 | extern void acpi_video_unregister(void); |
30 | extern int acpi_video_get_edid(struct acpi_device *device, int type, | 22 | extern int acpi_video_get_edid(struct acpi_device *device, int type, |
31 | int device_id, void **edid); | 23 | int device_id, void **edid); |
32 | #else | 24 | #else |
33 | static inline int acpi_video_register(void) { return 0; } | 25 | static inline int acpi_video_register(void) { return 0; } |
34 | static inline int acpi_video_register_with_quirks(void) { return 0; } | ||
35 | static inline void acpi_video_unregister(void) { return; } | 26 | static inline void acpi_video_unregister(void) { return; } |
36 | static inline int acpi_video_get_edid(struct acpi_device *device, int type, | 27 | static inline int acpi_video_get_edid(struct acpi_device *device, int type, |
37 | int device_id, void **edid) | 28 | int device_id, void **edid) |
diff --git a/include/dt-bindings/clock/vf610-clock.h b/include/dt-bindings/clock/vf610-clock.h index 15e997fa78f2..4aa2b48cd151 100644 --- a/include/dt-bindings/clock/vf610-clock.h +++ b/include/dt-bindings/clock/vf610-clock.h | |||
@@ -158,6 +158,8 @@ | |||
158 | #define VF610_CLK_GPU_SEL 145 | 158 | #define VF610_CLK_GPU_SEL 145 |
159 | #define VF610_CLK_GPU_EN 146 | 159 | #define VF610_CLK_GPU_EN 146 |
160 | #define VF610_CLK_GPU2D 147 | 160 | #define VF610_CLK_GPU2D 147 |
161 | #define VF610_CLK_END 148 | 161 | #define VF610_CLK_ENET0 148 |
162 | #define VF610_CLK_ENET1 149 | ||
163 | #define VF610_CLK_END 150 | ||
162 | 164 | ||
163 | #endif /* __DT_BINDINGS_CLOCK_VF610_H */ | 165 | #endif /* __DT_BINDINGS_CLOCK_VF610_H */ |
diff --git a/include/dt-bindings/pinctrl/am33xx.h b/include/dt-bindings/pinctrl/am33xx.h index 469e0325e6f4..2fbc804e1a45 100644 --- a/include/dt-bindings/pinctrl/am33xx.h +++ b/include/dt-bindings/pinctrl/am33xx.h | |||
@@ -5,7 +5,7 @@ | |||
5 | #ifndef _DT_BINDINGS_PINCTRL_AM33XX_H | 5 | #ifndef _DT_BINDINGS_PINCTRL_AM33XX_H |
6 | #define _DT_BINDINGS_PINCTRL_AM33XX_H | 6 | #define _DT_BINDINGS_PINCTRL_AM33XX_H |
7 | 7 | ||
8 | #include <include/dt-bindings/pinctrl/omap.h> | 8 | #include <dt-bindings/pinctrl/omap.h> |
9 | 9 | ||
10 | /* am33xx specific mux bit defines */ | 10 | /* am33xx specific mux bit defines */ |
11 | #undef PULL_ENA | 11 | #undef PULL_ENA |
diff --git a/include/linux/acpi.h b/include/linux/acpi.h index 6ad72f92469c..353ba256f368 100644 --- a/include/linux/acpi.h +++ b/include/linux/acpi.h | |||
@@ -191,7 +191,6 @@ extern bool wmi_has_guid(const char *guid); | |||
191 | #define ACPI_VIDEO_BACKLIGHT_DMI_VIDEO 0x0200 | 191 | #define ACPI_VIDEO_BACKLIGHT_DMI_VIDEO 0x0200 |
192 | #define ACPI_VIDEO_OUTPUT_SWITCHING_DMI_VENDOR 0x0400 | 192 | #define ACPI_VIDEO_OUTPUT_SWITCHING_DMI_VENDOR 0x0400 |
193 | #define ACPI_VIDEO_OUTPUT_SWITCHING_DMI_VIDEO 0x0800 | 193 | #define ACPI_VIDEO_OUTPUT_SWITCHING_DMI_VIDEO 0x0800 |
194 | #define ACPI_VIDEO_SKIP_BACKLIGHT 0x1000 | ||
195 | 194 | ||
196 | #if defined(CONFIG_ACPI_VIDEO) || defined(CONFIG_ACPI_VIDEO_MODULE) | 195 | #if defined(CONFIG_ACPI_VIDEO) || defined(CONFIG_ACPI_VIDEO_MODULE) |
197 | 196 | ||
diff --git a/include/linux/cgroup.h b/include/linux/cgroup.h index 297462b9f41a..e9ac882868c0 100644 --- a/include/linux/cgroup.h +++ b/include/linux/cgroup.h | |||
@@ -542,8 +542,7 @@ int cgroup_rm_cftypes(struct cgroup_subsys *ss, struct cftype *cfts); | |||
542 | bool cgroup_is_descendant(struct cgroup *cgrp, struct cgroup *ancestor); | 542 | bool cgroup_is_descendant(struct cgroup *cgrp, struct cgroup *ancestor); |
543 | 543 | ||
544 | int cgroup_path(const struct cgroup *cgrp, char *buf, int buflen); | 544 | int cgroup_path(const struct cgroup *cgrp, char *buf, int buflen); |
545 | int task_cgroup_path_from_hierarchy(struct task_struct *task, int hierarchy_id, | 545 | int task_cgroup_path(struct task_struct *task, char *buf, size_t buflen); |
546 | char *buf, size_t buflen); | ||
547 | 546 | ||
548 | int cgroup_task_count(const struct cgroup *cgrp); | 547 | int cgroup_task_count(const struct cgroup *cgrp); |
549 | 548 | ||
diff --git a/include/linux/cgroup_subsys.h b/include/linux/cgroup_subsys.h index 6e7ec64b69ab..b613ffd402d1 100644 --- a/include/linux/cgroup_subsys.h +++ b/include/linux/cgroup_subsys.h | |||
@@ -1,86 +1,55 @@ | |||
1 | /* Add subsystem definitions of the form SUBSYS(<name>) in this | 1 | /* |
2 | * file. Surround each one by a line of comment markers so that | 2 | * List of cgroup subsystems. |
3 | * patches don't collide | 3 | * |
4 | * DO NOT ADD ANY SUBSYSTEM WITHOUT EXPLICIT ACKS FROM CGROUP MAINTAINERS. | ||
4 | */ | 5 | */ |
5 | |||
6 | /* */ | ||
7 | |||
8 | /* */ | ||
9 | |||
10 | #if IS_SUBSYS_ENABLED(CONFIG_CPUSETS) | 6 | #if IS_SUBSYS_ENABLED(CONFIG_CPUSETS) |
11 | SUBSYS(cpuset) | 7 | SUBSYS(cpuset) |
12 | #endif | 8 | #endif |
13 | 9 | ||
14 | /* */ | ||
15 | |||
16 | #if IS_SUBSYS_ENABLED(CONFIG_CGROUP_DEBUG) | 10 | #if IS_SUBSYS_ENABLED(CONFIG_CGROUP_DEBUG) |
17 | SUBSYS(debug) | 11 | SUBSYS(debug) |
18 | #endif | 12 | #endif |
19 | 13 | ||
20 | /* */ | ||
21 | |||
22 | #if IS_SUBSYS_ENABLED(CONFIG_CGROUP_SCHED) | 14 | #if IS_SUBSYS_ENABLED(CONFIG_CGROUP_SCHED) |
23 | SUBSYS(cpu_cgroup) | 15 | SUBSYS(cpu_cgroup) |
24 | #endif | 16 | #endif |
25 | 17 | ||
26 | /* */ | ||
27 | |||
28 | #if IS_SUBSYS_ENABLED(CONFIG_CGROUP_CPUACCT) | 18 | #if IS_SUBSYS_ENABLED(CONFIG_CGROUP_CPUACCT) |
29 | SUBSYS(cpuacct) | 19 | SUBSYS(cpuacct) |
30 | #endif | 20 | #endif |
31 | 21 | ||
32 | /* */ | ||
33 | |||
34 | #if IS_SUBSYS_ENABLED(CONFIG_MEMCG) | 22 | #if IS_SUBSYS_ENABLED(CONFIG_MEMCG) |
35 | SUBSYS(mem_cgroup) | 23 | SUBSYS(mem_cgroup) |
36 | #endif | 24 | #endif |
37 | 25 | ||
38 | /* */ | ||
39 | |||
40 | #if IS_SUBSYS_ENABLED(CONFIG_CGROUP_DEVICE) | 26 | #if IS_SUBSYS_ENABLED(CONFIG_CGROUP_DEVICE) |
41 | SUBSYS(devices) | 27 | SUBSYS(devices) |
42 | #endif | 28 | #endif |
43 | 29 | ||
44 | /* */ | ||
45 | |||
46 | #if IS_SUBSYS_ENABLED(CONFIG_CGROUP_FREEZER) | 30 | #if IS_SUBSYS_ENABLED(CONFIG_CGROUP_FREEZER) |
47 | SUBSYS(freezer) | 31 | SUBSYS(freezer) |
48 | #endif | 32 | #endif |
49 | 33 | ||
50 | /* */ | ||
51 | |||
52 | #if IS_SUBSYS_ENABLED(CONFIG_NET_CLS_CGROUP) | 34 | #if IS_SUBSYS_ENABLED(CONFIG_NET_CLS_CGROUP) |
53 | SUBSYS(net_cls) | 35 | SUBSYS(net_cls) |
54 | #endif | 36 | #endif |
55 | 37 | ||
56 | /* */ | ||
57 | |||
58 | #if IS_SUBSYS_ENABLED(CONFIG_BLK_CGROUP) | 38 | #if IS_SUBSYS_ENABLED(CONFIG_BLK_CGROUP) |
59 | SUBSYS(blkio) | 39 | SUBSYS(blkio) |
60 | #endif | 40 | #endif |
61 | 41 | ||
62 | /* */ | ||
63 | |||
64 | #if IS_SUBSYS_ENABLED(CONFIG_CGROUP_PERF) | 42 | #if IS_SUBSYS_ENABLED(CONFIG_CGROUP_PERF) |
65 | SUBSYS(perf) | 43 | SUBSYS(perf) |
66 | #endif | 44 | #endif |
67 | 45 | ||
68 | /* */ | ||
69 | |||
70 | #if IS_SUBSYS_ENABLED(CONFIG_NETPRIO_CGROUP) | 46 | #if IS_SUBSYS_ENABLED(CONFIG_NETPRIO_CGROUP) |
71 | SUBSYS(net_prio) | 47 | SUBSYS(net_prio) |
72 | #endif | 48 | #endif |
73 | 49 | ||
74 | /* */ | ||
75 | |||
76 | #if IS_SUBSYS_ENABLED(CONFIG_CGROUP_HUGETLB) | 50 | #if IS_SUBSYS_ENABLED(CONFIG_CGROUP_HUGETLB) |
77 | SUBSYS(hugetlb) | 51 | SUBSYS(hugetlb) |
78 | #endif | 52 | #endif |
79 | 53 | /* | |
80 | /* */ | 54 | * DO NOT ADD ANY SUBSYSTEM WITHOUT EXPLICIT ACKS FROM CGROUP MAINTAINERS. |
81 | 55 | */ | |
82 | #ifdef CONFIG_CGROUP_BCACHE | ||
83 | SUBSYS(bcache) | ||
84 | #endif | ||
85 | |||
86 | /* */ | ||
diff --git a/include/linux/crc-t10dif.h b/include/linux/crc-t10dif.h index b3cb71f0d3b0..a9c96d865ee7 100644 --- a/include/linux/crc-t10dif.h +++ b/include/linux/crc-t10dif.h | |||
@@ -3,10 +3,6 @@ | |||
3 | 3 | ||
4 | #include <linux/types.h> | 4 | #include <linux/types.h> |
5 | 5 | ||
6 | #define CRC_T10DIF_DIGEST_SIZE 2 | ||
7 | #define CRC_T10DIF_BLOCK_SIZE 1 | ||
8 | |||
9 | __u16 crc_t10dif_generic(__u16 crc, const unsigned char *buffer, size_t len); | ||
10 | __u16 crc_t10dif(unsigned char const *, size_t); | 6 | __u16 crc_t10dif(unsigned char const *, size_t); |
11 | 7 | ||
12 | #endif | 8 | #endif |
diff --git a/include/linux/drbd.h b/include/linux/drbd.h index 1b4d4ee1168f..de7d74ab3de6 100644 --- a/include/linux/drbd.h +++ b/include/linux/drbd.h | |||
@@ -177,7 +177,11 @@ enum drbd_ret_code { | |||
177 | ERR_NEED_APV_100 = 163, | 177 | ERR_NEED_APV_100 = 163, |
178 | ERR_NEED_ALLOW_TWO_PRI = 164, | 178 | ERR_NEED_ALLOW_TWO_PRI = 164, |
179 | ERR_MD_UNCLEAN = 165, | 179 | ERR_MD_UNCLEAN = 165, |
180 | 180 | ERR_MD_LAYOUT_CONNECTED = 166, | |
181 | ERR_MD_LAYOUT_TOO_BIG = 167, | ||
182 | ERR_MD_LAYOUT_TOO_SMALL = 168, | ||
183 | ERR_MD_LAYOUT_NO_FIT = 169, | ||
184 | ERR_IMPLICIT_SHRINK = 170, | ||
181 | /* insert new ones above this line */ | 185 | /* insert new ones above this line */ |
182 | AFTER_LAST_ERR_CODE | 186 | AFTER_LAST_ERR_CODE |
183 | }; | 187 | }; |
diff --git a/include/linux/drbd_genl.h b/include/linux/drbd_genl.h index d0d8fac8a6e4..e8c44572b8cb 100644 --- a/include/linux/drbd_genl.h +++ b/include/linux/drbd_genl.h | |||
@@ -181,6 +181,8 @@ GENL_struct(DRBD_NLA_RESIZE_PARMS, 7, resize_parms, | |||
181 | __u64_field(1, DRBD_GENLA_F_MANDATORY, resize_size) | 181 | __u64_field(1, DRBD_GENLA_F_MANDATORY, resize_size) |
182 | __flg_field(2, DRBD_GENLA_F_MANDATORY, resize_force) | 182 | __flg_field(2, DRBD_GENLA_F_MANDATORY, resize_force) |
183 | __flg_field(3, DRBD_GENLA_F_MANDATORY, no_resync) | 183 | __flg_field(3, DRBD_GENLA_F_MANDATORY, no_resync) |
184 | __u32_field_def(4, 0 /* OPTIONAL */, al_stripes, DRBD_AL_STRIPES_DEF) | ||
185 | __u32_field_def(5, 0 /* OPTIONAL */, al_stripe_size, DRBD_AL_STRIPE_SIZE_DEF) | ||
184 | ) | 186 | ) |
185 | 187 | ||
186 | GENL_struct(DRBD_NLA_STATE_INFO, 8, state_info, | 188 | GENL_struct(DRBD_NLA_STATE_INFO, 8, state_info, |
diff --git a/include/linux/drbd_limits.h b/include/linux/drbd_limits.h index 1fedf2b17cc8..17e50bb00521 100644 --- a/include/linux/drbd_limits.h +++ b/include/linux/drbd_limits.h | |||
@@ -215,4 +215,13 @@ | |||
215 | #define DRBD_ALWAYS_ASBP_DEF 0 | 215 | #define DRBD_ALWAYS_ASBP_DEF 0 |
216 | #define DRBD_USE_RLE_DEF 1 | 216 | #define DRBD_USE_RLE_DEF 1 |
217 | 217 | ||
218 | #define DRBD_AL_STRIPES_MIN 1 | ||
219 | #define DRBD_AL_STRIPES_MAX 1024 | ||
220 | #define DRBD_AL_STRIPES_DEF 1 | ||
221 | #define DRBD_AL_STRIPES_SCALE '1' | ||
222 | |||
223 | #define DRBD_AL_STRIPE_SIZE_MIN 4 | ||
224 | #define DRBD_AL_STRIPE_SIZE_MAX 16777216 | ||
225 | #define DRBD_AL_STRIPE_SIZE_DEF 32 | ||
226 | #define DRBD_AL_STRIPE_SIZE_SCALE 'k' /* kilobytes */ | ||
218 | #endif | 227 | #endif |
diff --git a/include/linux/edac.h b/include/linux/edac.h index 0b763276f619..5c6d7fbaf89e 100644 --- a/include/linux/edac.h +++ b/include/linux/edac.h | |||
@@ -622,7 +622,7 @@ struct edac_raw_error_desc { | |||
622 | */ | 622 | */ |
623 | struct mem_ctl_info { | 623 | struct mem_ctl_info { |
624 | struct device dev; | 624 | struct device dev; |
625 | struct bus_type bus; | 625 | struct bus_type *bus; |
626 | 626 | ||
627 | struct list_head link; /* for global list of mem_ctl_info structs */ | 627 | struct list_head link; /* for global list of mem_ctl_info structs */ |
628 | 628 | ||
@@ -742,4 +742,9 @@ struct mem_ctl_info { | |||
742 | #endif | 742 | #endif |
743 | }; | 743 | }; |
744 | 744 | ||
745 | /* | ||
746 | * Maximum number of memory controllers in the coherent fabric. | ||
747 | */ | ||
748 | #define EDAC_MAX_MCS 16 | ||
749 | |||
745 | #endif | 750 | #endif |
diff --git a/include/linux/mfd/syscon/imx6q-iomuxc-gpr.h b/include/linux/mfd/syscon/imx6q-iomuxc-gpr.h index dab34a1deb2c..b6bdcd66c07d 100644 --- a/include/linux/mfd/syscon/imx6q-iomuxc-gpr.h +++ b/include/linux/mfd/syscon/imx6q-iomuxc-gpr.h | |||
@@ -103,15 +103,15 @@ | |||
103 | #define IMX6Q_GPR1_EXC_MON_MASK BIT(22) | 103 | #define IMX6Q_GPR1_EXC_MON_MASK BIT(22) |
104 | #define IMX6Q_GPR1_EXC_MON_OKAY 0x0 | 104 | #define IMX6Q_GPR1_EXC_MON_OKAY 0x0 |
105 | #define IMX6Q_GPR1_EXC_MON_SLVE BIT(22) | 105 | #define IMX6Q_GPR1_EXC_MON_SLVE BIT(22) |
106 | #define IMX6Q_GPR1_MIPI_IPU2_SEL_MASK BIT(21) | 106 | #define IMX6Q_GPR1_ENET_CLK_SEL_MASK BIT(21) |
107 | #define IMX6Q_GPR1_MIPI_IPU2_SEL_GASKET 0x0 | 107 | #define IMX6Q_GPR1_ENET_CLK_SEL_PAD 0 |
108 | #define IMX6Q_GPR1_MIPI_IPU2_SEL_IOMUX BIT(21) | 108 | #define IMX6Q_GPR1_ENET_CLK_SEL_ANATOP BIT(21) |
109 | #define IMX6Q_GPR1_MIPI_IPU1_MUX_MASK BIT(20) | 109 | #define IMX6Q_GPR1_MIPI_IPU2_MUX_MASK BIT(20) |
110 | #define IMX6Q_GPR1_MIPI_IPU1_MUX_GASKET 0x0 | ||
111 | #define IMX6Q_GPR1_MIPI_IPU1_MUX_IOMUX BIT(20) | ||
112 | #define IMX6Q_GPR1_MIPI_IPU2_MUX_MASK BIT(19) | ||
113 | #define IMX6Q_GPR1_MIPI_IPU2_MUX_GASKET 0x0 | 110 | #define IMX6Q_GPR1_MIPI_IPU2_MUX_GASKET 0x0 |
114 | #define IMX6Q_GPR1_MIPI_IPU2_MUX_IOMUX BIT(19) | 111 | #define IMX6Q_GPR1_MIPI_IPU2_MUX_IOMUX BIT(20) |
112 | #define IMX6Q_GPR1_MIPI_IPU1_MUX_MASK BIT(19) | ||
113 | #define IMX6Q_GPR1_MIPI_IPU1_MUX_GASKET 0x0 | ||
114 | #define IMX6Q_GPR1_MIPI_IPU1_MUX_IOMUX BIT(19) | ||
115 | #define IMX6Q_GPR1_PCIE_TEST_PD BIT(18) | 115 | #define IMX6Q_GPR1_PCIE_TEST_PD BIT(18) |
116 | #define IMX6Q_GPR1_IPU_VPU_MUX_MASK BIT(17) | 116 | #define IMX6Q_GPR1_IPU_VPU_MUX_MASK BIT(17) |
117 | #define IMX6Q_GPR1_IPU_VPU_MUX_IPU1 0x0 | 117 | #define IMX6Q_GPR1_IPU_VPU_MUX_IPU1 0x0 |
@@ -279,41 +279,88 @@ | |||
279 | #define IMX6Q_GPR13_CAN2_STOP_REQ BIT(29) | 279 | #define IMX6Q_GPR13_CAN2_STOP_REQ BIT(29) |
280 | #define IMX6Q_GPR13_CAN1_STOP_REQ BIT(28) | 280 | #define IMX6Q_GPR13_CAN1_STOP_REQ BIT(28) |
281 | #define IMX6Q_GPR13_ENET_STOP_REQ BIT(27) | 281 | #define IMX6Q_GPR13_ENET_STOP_REQ BIT(27) |
282 | #define IMX6Q_GPR13_SATA_PHY_8_MASK (0x7 << 24) | 282 | #define IMX6Q_GPR13_SATA_RX_EQ_VAL_MASK (0x7 << 24) |
283 | #define IMX6Q_GPR13_SATA_PHY_8_0_5_DB (0x0 << 24) | 283 | #define IMX6Q_GPR13_SATA_RX_EQ_VAL_0_5_DB (0x0 << 24) |
284 | #define IMX6Q_GPR13_SATA_PHY_8_1_0_DB (0x1 << 24) | 284 | #define IMX6Q_GPR13_SATA_RX_EQ_VAL_1_0_DB (0x1 << 24) |
285 | #define IMX6Q_GPR13_SATA_PHY_8_1_5_DB (0x2 << 24) | 285 | #define IMX6Q_GPR13_SATA_RX_EQ_VAL_1_5_DB (0x2 << 24) |
286 | #define IMX6Q_GPR13_SATA_PHY_8_2_0_DB (0x3 << 24) | 286 | #define IMX6Q_GPR13_SATA_RX_EQ_VAL_2_0_DB (0x3 << 24) |
287 | #define IMX6Q_GPR13_SATA_PHY_8_2_5_DB (0x4 << 24) | 287 | #define IMX6Q_GPR13_SATA_RX_EQ_VAL_2_5_DB (0x4 << 24) |
288 | #define IMX6Q_GPR13_SATA_PHY_8_3_0_DB (0x5 << 24) | 288 | #define IMX6Q_GPR13_SATA_RX_EQ_VAL_3_0_DB (0x5 << 24) |
289 | #define IMX6Q_GPR13_SATA_PHY_8_3_5_DB (0x6 << 24) | 289 | #define IMX6Q_GPR13_SATA_RX_EQ_VAL_3_5_DB (0x6 << 24) |
290 | #define IMX6Q_GPR13_SATA_PHY_8_4_0_DB (0x7 << 24) | 290 | #define IMX6Q_GPR13_SATA_RX_EQ_VAL_4_0_DB (0x7 << 24) |
291 | #define IMX6Q_GPR13_SATA_PHY_7_MASK (0x1f << 19) | 291 | #define IMX6Q_GPR13_SATA_RX_LOS_LVL_MASK (0x1f << 19) |
292 | #define IMX6Q_GPR13_SATA_PHY_7_SATA1I (0x10 << 19) | 292 | #define IMX6Q_GPR13_SATA_RX_LOS_LVL_SATA1I (0x10 << 19) |
293 | #define IMX6Q_GPR13_SATA_PHY_7_SATA1M (0x10 << 19) | 293 | #define IMX6Q_GPR13_SATA_RX_LOS_LVL_SATA1M (0x10 << 19) |
294 | #define IMX6Q_GPR13_SATA_PHY_7_SATA1X (0x1a << 19) | 294 | #define IMX6Q_GPR13_SATA_RX_LOS_LVL_SATA1X (0x1a << 19) |
295 | #define IMX6Q_GPR13_SATA_PHY_7_SATA2I (0x12 << 19) | 295 | #define IMX6Q_GPR13_SATA_RX_LOS_LVL_SATA2I (0x12 << 19) |
296 | #define IMX6Q_GPR13_SATA_PHY_7_SATA2M (0x12 << 19) | 296 | #define IMX6Q_GPR13_SATA_RX_LOS_LVL_SATA2M (0x12 << 19) |
297 | #define IMX6Q_GPR13_SATA_PHY_7_SATA2X (0x1a << 19) | 297 | #define IMX6Q_GPR13_SATA_RX_LOS_LVL_SATA2X (0x1a << 19) |
298 | #define IMX6Q_GPR13_SATA_PHY_6_MASK (0x7 << 16) | 298 | #define IMX6Q_GPR13_SATA_RX_DPLL_MODE_MASK (0x7 << 16) |
299 | #define IMX6Q_GPR13_SATA_SPEED_MASK BIT(15) | 299 | #define IMX6Q_GPR13_SATA_RX_DPLL_MODE_1P_1F (0x0 << 16) |
300 | #define IMX6Q_GPR13_SATA_SPEED_1P5G 0x0 | 300 | #define IMX6Q_GPR13_SATA_RX_DPLL_MODE_2P_2F (0x1 << 16) |
301 | #define IMX6Q_GPR13_SATA_SPEED_3P0G BIT(15) | 301 | #define IMX6Q_GPR13_SATA_RX_DPLL_MODE_1P_4F (0x2 << 16) |
302 | #define IMX6Q_GPR13_SATA_PHY_5 BIT(14) | 302 | #define IMX6Q_GPR13_SATA_RX_DPLL_MODE_2P_4F (0x3 << 16) |
303 | #define IMX6Q_GPR13_SATA_PHY_4_MASK (0x7 << 11) | 303 | #define IMX6Q_GPR13_SATA_SPD_MODE_MASK BIT(15) |
304 | #define IMX6Q_GPR13_SATA_PHY_4_16_16 (0x0 << 11) | 304 | #define IMX6Q_GPR13_SATA_SPD_MODE_1P5G 0x0 |
305 | #define IMX6Q_GPR13_SATA_PHY_4_14_16 (0x1 << 11) | 305 | #define IMX6Q_GPR13_SATA_SPD_MODE_3P0G BIT(15) |
306 | #define IMX6Q_GPR13_SATA_PHY_4_12_16 (0x2 << 11) | 306 | #define IMX6Q_GPR13_SATA_MPLL_SS_EN BIT(14) |
307 | #define IMX6Q_GPR13_SATA_PHY_4_10_16 (0x3 << 11) | 307 | #define IMX6Q_GPR13_SATA_TX_ATTEN_MASK (0x7 << 11) |
308 | #define IMX6Q_GPR13_SATA_PHY_4_9_16 (0x4 << 11) | 308 | #define IMX6Q_GPR13_SATA_TX_ATTEN_16_16 (0x0 << 11) |
309 | #define IMX6Q_GPR13_SATA_PHY_4_8_16 (0x5 << 11) | 309 | #define IMX6Q_GPR13_SATA_TX_ATTEN_14_16 (0x1 << 11) |
310 | #define IMX6Q_GPR13_SATA_PHY_3_MASK (0xf << 7) | 310 | #define IMX6Q_GPR13_SATA_TX_ATTEN_12_16 (0x2 << 11) |
311 | #define IMX6Q_GPR13_SATA_PHY_3_OFF 0x7 | 311 | #define IMX6Q_GPR13_SATA_TX_ATTEN_10_16 (0x3 << 11) |
312 | #define IMX6Q_GPR13_SATA_PHY_2_MASK (0x1f << 2) | 312 | #define IMX6Q_GPR13_SATA_TX_ATTEN_9_16 (0x4 << 11) |
313 | #define IMX6Q_GPR13_SATA_PHY_2_OFF 0x2 | 313 | #define IMX6Q_GPR13_SATA_TX_ATTEN_8_16 (0x5 << 11) |
314 | #define IMX6Q_GPR13_SATA_PHY_1_MASK (0x3 << 0) | 314 | #define IMX6Q_GPR13_SATA_TX_BOOST_MASK (0xf << 7) |
315 | #define IMX6Q_GPR13_SATA_PHY_1_FAST (0x0 << 0) | 315 | #define IMX6Q_GPR13_SATA_TX_BOOST_0_00_DB (0x0 << 7) |
316 | #define IMX6Q_GPR13_SATA_PHY_1_MED (0x1 << 0) | 316 | #define IMX6Q_GPR13_SATA_TX_BOOST_0_37_DB (0x1 << 7) |
317 | #define IMX6Q_GPR13_SATA_PHY_1_SLOW (0x2 << 0) | 317 | #define IMX6Q_GPR13_SATA_TX_BOOST_0_74_DB (0x2 << 7) |
318 | 318 | #define IMX6Q_GPR13_SATA_TX_BOOST_1_11_DB (0x3 << 7) | |
319 | #define IMX6Q_GPR13_SATA_TX_BOOST_1_48_DB (0x4 << 7) | ||
320 | #define IMX6Q_GPR13_SATA_TX_BOOST_1_85_DB (0x5 << 7) | ||
321 | #define IMX6Q_GPR13_SATA_TX_BOOST_2_22_DB (0x6 << 7) | ||
322 | #define IMX6Q_GPR13_SATA_TX_BOOST_2_59_DB (0x7 << 7) | ||
323 | #define IMX6Q_GPR13_SATA_TX_BOOST_2_96_DB (0x8 << 7) | ||
324 | #define IMX6Q_GPR13_SATA_TX_BOOST_3_33_DB (0x9 << 7) | ||
325 | #define IMX6Q_GPR13_SATA_TX_BOOST_3_70_DB (0xa << 7) | ||
326 | #define IMX6Q_GPR13_SATA_TX_BOOST_4_07_DB (0xb << 7) | ||
327 | #define IMX6Q_GPR13_SATA_TX_BOOST_4_44_DB (0xc << 7) | ||
328 | #define IMX6Q_GPR13_SATA_TX_BOOST_4_81_DB (0xd << 7) | ||
329 | #define IMX6Q_GPR13_SATA_TX_BOOST_5_28_DB (0xe << 7) | ||
330 | #define IMX6Q_GPR13_SATA_TX_BOOST_5_75_DB (0xf << 7) | ||
331 | #define IMX6Q_GPR13_SATA_TX_LVL_MASK (0x1f << 2) | ||
332 | #define IMX6Q_GPR13_SATA_TX_LVL_0_937_V (0x00 << 2) | ||
333 | #define IMX6Q_GPR13_SATA_TX_LVL_0_947_V (0x01 << 2) | ||
334 | #define IMX6Q_GPR13_SATA_TX_LVL_0_957_V (0x02 << 2) | ||
335 | #define IMX6Q_GPR13_SATA_TX_LVL_0_966_V (0x03 << 2) | ||
336 | #define IMX6Q_GPR13_SATA_TX_LVL_0_976_V (0x04 << 2) | ||
337 | #define IMX6Q_GPR13_SATA_TX_LVL_0_986_V (0x05 << 2) | ||
338 | #define IMX6Q_GPR13_SATA_TX_LVL_0_996_V (0x06 << 2) | ||
339 | #define IMX6Q_GPR13_SATA_TX_LVL_1_005_V (0x07 << 2) | ||
340 | #define IMX6Q_GPR13_SATA_TX_LVL_1_015_V (0x08 << 2) | ||
341 | #define IMX6Q_GPR13_SATA_TX_LVL_1_025_V (0x09 << 2) | ||
342 | #define IMX6Q_GPR13_SATA_TX_LVL_1_035_V (0x0a << 2) | ||
343 | #define IMX6Q_GPR13_SATA_TX_LVL_1_045_V (0x0b << 2) | ||
344 | #define IMX6Q_GPR13_SATA_TX_LVL_1_054_V (0x0c << 2) | ||
345 | #define IMX6Q_GPR13_SATA_TX_LVL_1_064_V (0x0d << 2) | ||
346 | #define IMX6Q_GPR13_SATA_TX_LVL_1_074_V (0x0e << 2) | ||
347 | #define IMX6Q_GPR13_SATA_TX_LVL_1_084_V (0x0f << 2) | ||
348 | #define IMX6Q_GPR13_SATA_TX_LVL_1_094_V (0x10 << 2) | ||
349 | #define IMX6Q_GPR13_SATA_TX_LVL_1_104_V (0x11 << 2) | ||
350 | #define IMX6Q_GPR13_SATA_TX_LVL_1_113_V (0x12 << 2) | ||
351 | #define IMX6Q_GPR13_SATA_TX_LVL_1_123_V (0x13 << 2) | ||
352 | #define IMX6Q_GPR13_SATA_TX_LVL_1_133_V (0x14 << 2) | ||
353 | #define IMX6Q_GPR13_SATA_TX_LVL_1_143_V (0x15 << 2) | ||
354 | #define IMX6Q_GPR13_SATA_TX_LVL_1_152_V (0x16 << 2) | ||
355 | #define IMX6Q_GPR13_SATA_TX_LVL_1_162_V (0x17 << 2) | ||
356 | #define IMX6Q_GPR13_SATA_TX_LVL_1_172_V (0x18 << 2) | ||
357 | #define IMX6Q_GPR13_SATA_TX_LVL_1_182_V (0x19 << 2) | ||
358 | #define IMX6Q_GPR13_SATA_TX_LVL_1_191_V (0x1a << 2) | ||
359 | #define IMX6Q_GPR13_SATA_TX_LVL_1_201_V (0x1b << 2) | ||
360 | #define IMX6Q_GPR13_SATA_TX_LVL_1_211_V (0x1c << 2) | ||
361 | #define IMX6Q_GPR13_SATA_TX_LVL_1_221_V (0x1d << 2) | ||
362 | #define IMX6Q_GPR13_SATA_TX_LVL_1_230_V (0x1e << 2) | ||
363 | #define IMX6Q_GPR13_SATA_TX_LVL_1_240_V (0x1f << 2) | ||
364 | #define IMX6Q_GPR13_SATA_MPLL_CLK_EN BIT(1) | ||
365 | #define IMX6Q_GPR13_SATA_TX_EDGE_RATE BIT(0) | ||
319 | #endif /* __LINUX_IMX6Q_IOMUXC_GPR_H */ | 366 | #endif /* __LINUX_IMX6Q_IOMUXC_GPR_H */ |
diff --git a/include/linux/platform_data/mmc-pxamci.h b/include/linux/platform_data/mmc-pxamci.h index 9eb515bb799d..1706b3597ce0 100644 --- a/include/linux/platform_data/mmc-pxamci.h +++ b/include/linux/platform_data/mmc-pxamci.h | |||
@@ -12,7 +12,7 @@ struct pxamci_platform_data { | |||
12 | unsigned long detect_delay_ms; /* delay in millisecond before detecting cards after interrupt */ | 12 | unsigned long detect_delay_ms; /* delay in millisecond before detecting cards after interrupt */ |
13 | int (*init)(struct device *, irq_handler_t , void *); | 13 | int (*init)(struct device *, irq_handler_t , void *); |
14 | int (*get_ro)(struct device *); | 14 | int (*get_ro)(struct device *); |
15 | void (*setpower)(struct device *, unsigned int); | 15 | int (*setpower)(struct device *, unsigned int); |
16 | void (*exit)(struct device *, void *); | 16 | void (*exit)(struct device *, void *); |
17 | int gpio_card_detect; /* gpio detecting card insertion */ | 17 | int gpio_card_detect; /* gpio detecting card insertion */ |
18 | int gpio_card_ro; /* gpio detecting read only toggle */ | 18 | int gpio_card_ro; /* gpio detecting read only toggle */ |
diff --git a/include/linux/shdma-base.h b/include/linux/shdma-base.h index 382cf710ca9a..5b1c9848124c 100644 --- a/include/linux/shdma-base.h +++ b/include/linux/shdma-base.h | |||
@@ -124,6 +124,10 @@ void shdma_chan_remove(struct shdma_chan *schan); | |||
124 | int shdma_init(struct device *dev, struct shdma_dev *sdev, | 124 | int shdma_init(struct device *dev, struct shdma_dev *sdev, |
125 | int chan_num); | 125 | int chan_num); |
126 | void shdma_cleanup(struct shdma_dev *sdev); | 126 | void shdma_cleanup(struct shdma_dev *sdev); |
127 | #if IS_ENABLED(CONFIG_SH_DMAE_BASE) | ||
127 | bool shdma_chan_filter(struct dma_chan *chan, void *arg); | 128 | bool shdma_chan_filter(struct dma_chan *chan, void *arg); |
129 | #else | ||
130 | #define shdma_chan_filter NULL | ||
131 | #endif | ||
128 | 132 | ||
129 | #endif | 133 | #endif |
diff --git a/include/linux/usb.h b/include/linux/usb.h index a232b7ece1f6..0eec2689b955 100644 --- a/include/linux/usb.h +++ b/include/linux/usb.h | |||
@@ -367,17 +367,6 @@ struct usb_bus { | |||
367 | 367 | ||
368 | /* ----------------------------------------------------------------------- */ | 368 | /* ----------------------------------------------------------------------- */ |
369 | 369 | ||
370 | /* This is arbitrary. | ||
371 | * From USB 2.0 spec Table 11-13, offset 7, a hub can | ||
372 | * have up to 255 ports. The most yet reported is 10. | ||
373 | * | ||
374 | * Current Wireless USB host hardware (Intel i1480 for example) allows | ||
375 | * up to 22 devices to connect. Upcoming hardware might raise that | ||
376 | * limit. Because the arrays need to add a bit for hub status data, we | ||
377 | * do 31, so plus one evens out to four bytes. | ||
378 | */ | ||
379 | #define USB_MAXCHILDREN (31) | ||
380 | |||
381 | struct usb_tt; | 370 | struct usb_tt; |
382 | 371 | ||
383 | enum usb_device_removable { | 372 | enum usb_device_removable { |
diff --git a/include/trace/events/bcache.h b/include/trace/events/bcache.h index 3cc5a0b278c3..5ebda976ea93 100644 --- a/include/trace/events/bcache.h +++ b/include/trace/events/bcache.h | |||
@@ -9,9 +9,7 @@ | |||
9 | struct search; | 9 | struct search; |
10 | 10 | ||
11 | DECLARE_EVENT_CLASS(bcache_request, | 11 | DECLARE_EVENT_CLASS(bcache_request, |
12 | |||
13 | TP_PROTO(struct search *s, struct bio *bio), | 12 | TP_PROTO(struct search *s, struct bio *bio), |
14 | |||
15 | TP_ARGS(s, bio), | 13 | TP_ARGS(s, bio), |
16 | 14 | ||
17 | TP_STRUCT__entry( | 15 | TP_STRUCT__entry( |
@@ -22,7 +20,6 @@ DECLARE_EVENT_CLASS(bcache_request, | |||
22 | __field(dev_t, orig_sector ) | 20 | __field(dev_t, orig_sector ) |
23 | __field(unsigned int, nr_sector ) | 21 | __field(unsigned int, nr_sector ) |
24 | __array(char, rwbs, 6 ) | 22 | __array(char, rwbs, 6 ) |
25 | __array(char, comm, TASK_COMM_LEN ) | ||
26 | ), | 23 | ), |
27 | 24 | ||
28 | TP_fast_assign( | 25 | TP_fast_assign( |
@@ -33,36 +30,66 @@ DECLARE_EVENT_CLASS(bcache_request, | |||
33 | __entry->orig_sector = bio->bi_sector - 16; | 30 | __entry->orig_sector = bio->bi_sector - 16; |
34 | __entry->nr_sector = bio->bi_size >> 9; | 31 | __entry->nr_sector = bio->bi_size >> 9; |
35 | blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_size); | 32 | blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_size); |
36 | memcpy(__entry->comm, current->comm, TASK_COMM_LEN); | ||
37 | ), | 33 | ), |
38 | 34 | ||
39 | TP_printk("%d,%d %s %llu + %u [%s] (from %d,%d @ %llu)", | 35 | TP_printk("%d,%d %s %llu + %u (from %d,%d @ %llu)", |
40 | MAJOR(__entry->dev), MINOR(__entry->dev), | 36 | MAJOR(__entry->dev), MINOR(__entry->dev), |
41 | __entry->rwbs, | 37 | __entry->rwbs, (unsigned long long)__entry->sector, |
42 | (unsigned long long)__entry->sector, | 38 | __entry->nr_sector, __entry->orig_major, __entry->orig_minor, |
43 | __entry->nr_sector, __entry->comm, | ||
44 | __entry->orig_major, __entry->orig_minor, | ||
45 | (unsigned long long)__entry->orig_sector) | 39 | (unsigned long long)__entry->orig_sector) |
46 | ); | 40 | ); |
47 | 41 | ||
48 | DEFINE_EVENT(bcache_request, bcache_request_start, | 42 | DECLARE_EVENT_CLASS(bkey, |
43 | TP_PROTO(struct bkey *k), | ||
44 | TP_ARGS(k), | ||
49 | 45 | ||
50 | TP_PROTO(struct search *s, struct bio *bio), | 46 | TP_STRUCT__entry( |
47 | __field(u32, size ) | ||
48 | __field(u32, inode ) | ||
49 | __field(u64, offset ) | ||
50 | __field(bool, dirty ) | ||
51 | ), | ||
51 | 52 | ||
52 | TP_ARGS(s, bio) | 53 | TP_fast_assign( |
54 | __entry->inode = KEY_INODE(k); | ||
55 | __entry->offset = KEY_OFFSET(k); | ||
56 | __entry->size = KEY_SIZE(k); | ||
57 | __entry->dirty = KEY_DIRTY(k); | ||
58 | ), | ||
59 | |||
60 | TP_printk("%u:%llu len %u dirty %u", __entry->inode, | ||
61 | __entry->offset, __entry->size, __entry->dirty) | ||
53 | ); | 62 | ); |
54 | 63 | ||
55 | DEFINE_EVENT(bcache_request, bcache_request_end, | 64 | DECLARE_EVENT_CLASS(btree_node, |
65 | TP_PROTO(struct btree *b), | ||
66 | TP_ARGS(b), | ||
67 | |||
68 | TP_STRUCT__entry( | ||
69 | __field(size_t, bucket ) | ||
70 | ), | ||
56 | 71 | ||
72 | TP_fast_assign( | ||
73 | __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0); | ||
74 | ), | ||
75 | |||
76 | TP_printk("bucket %zu", __entry->bucket) | ||
77 | ); | ||
78 | |||
79 | /* request.c */ | ||
80 | |||
81 | DEFINE_EVENT(bcache_request, bcache_request_start, | ||
57 | TP_PROTO(struct search *s, struct bio *bio), | 82 | TP_PROTO(struct search *s, struct bio *bio), |
83 | TP_ARGS(s, bio) | ||
84 | ); | ||
58 | 85 | ||
86 | DEFINE_EVENT(bcache_request, bcache_request_end, | ||
87 | TP_PROTO(struct search *s, struct bio *bio), | ||
59 | TP_ARGS(s, bio) | 88 | TP_ARGS(s, bio) |
60 | ); | 89 | ); |
61 | 90 | ||
62 | DECLARE_EVENT_CLASS(bcache_bio, | 91 | DECLARE_EVENT_CLASS(bcache_bio, |
63 | |||
64 | TP_PROTO(struct bio *bio), | 92 | TP_PROTO(struct bio *bio), |
65 | |||
66 | TP_ARGS(bio), | 93 | TP_ARGS(bio), |
67 | 94 | ||
68 | TP_STRUCT__entry( | 95 | TP_STRUCT__entry( |
@@ -70,7 +97,6 @@ DECLARE_EVENT_CLASS(bcache_bio, | |||
70 | __field(sector_t, sector ) | 97 | __field(sector_t, sector ) |
71 | __field(unsigned int, nr_sector ) | 98 | __field(unsigned int, nr_sector ) |
72 | __array(char, rwbs, 6 ) | 99 | __array(char, rwbs, 6 ) |
73 | __array(char, comm, TASK_COMM_LEN ) | ||
74 | ), | 100 | ), |
75 | 101 | ||
76 | TP_fast_assign( | 102 | TP_fast_assign( |
@@ -78,191 +104,328 @@ DECLARE_EVENT_CLASS(bcache_bio, | |||
78 | __entry->sector = bio->bi_sector; | 104 | __entry->sector = bio->bi_sector; |
79 | __entry->nr_sector = bio->bi_size >> 9; | 105 | __entry->nr_sector = bio->bi_size >> 9; |
80 | blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_size); | 106 | blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_size); |
81 | memcpy(__entry->comm, current->comm, TASK_COMM_LEN); | ||
82 | ), | 107 | ), |
83 | 108 | ||
84 | TP_printk("%d,%d %s %llu + %u [%s]", | 109 | TP_printk("%d,%d %s %llu + %u", |
85 | MAJOR(__entry->dev), MINOR(__entry->dev), | 110 | MAJOR(__entry->dev), MINOR(__entry->dev), __entry->rwbs, |
86 | __entry->rwbs, | 111 | (unsigned long long)__entry->sector, __entry->nr_sector) |
87 | (unsigned long long)__entry->sector, | ||
88 | __entry->nr_sector, __entry->comm) | ||
89 | ); | 112 | ); |
90 | 113 | ||
91 | 114 | DEFINE_EVENT(bcache_bio, bcache_bypass_sequential, | |
92 | DEFINE_EVENT(bcache_bio, bcache_passthrough, | ||
93 | |||
94 | TP_PROTO(struct bio *bio), | 115 | TP_PROTO(struct bio *bio), |
116 | TP_ARGS(bio) | ||
117 | ); | ||
95 | 118 | ||
119 | DEFINE_EVENT(bcache_bio, bcache_bypass_congested, | ||
120 | TP_PROTO(struct bio *bio), | ||
96 | TP_ARGS(bio) | 121 | TP_ARGS(bio) |
97 | ); | 122 | ); |
98 | 123 | ||
99 | DEFINE_EVENT(bcache_bio, bcache_cache_hit, | 124 | TRACE_EVENT(bcache_read, |
125 | TP_PROTO(struct bio *bio, bool hit, bool bypass), | ||
126 | TP_ARGS(bio, hit, bypass), | ||
100 | 127 | ||
101 | TP_PROTO(struct bio *bio), | 128 | TP_STRUCT__entry( |
129 | __field(dev_t, dev ) | ||
130 | __field(sector_t, sector ) | ||
131 | __field(unsigned int, nr_sector ) | ||
132 | __array(char, rwbs, 6 ) | ||
133 | __field(bool, cache_hit ) | ||
134 | __field(bool, bypass ) | ||
135 | ), | ||
102 | 136 | ||
103 | TP_ARGS(bio) | 137 | TP_fast_assign( |
138 | __entry->dev = bio->bi_bdev->bd_dev; | ||
139 | __entry->sector = bio->bi_sector; | ||
140 | __entry->nr_sector = bio->bi_size >> 9; | ||
141 | blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_size); | ||
142 | __entry->cache_hit = hit; | ||
143 | __entry->bypass = bypass; | ||
144 | ), | ||
145 | |||
146 | TP_printk("%d,%d %s %llu + %u hit %u bypass %u", | ||
147 | MAJOR(__entry->dev), MINOR(__entry->dev), | ||
148 | __entry->rwbs, (unsigned long long)__entry->sector, | ||
149 | __entry->nr_sector, __entry->cache_hit, __entry->bypass) | ||
104 | ); | 150 | ); |
105 | 151 | ||
106 | DEFINE_EVENT(bcache_bio, bcache_cache_miss, | 152 | TRACE_EVENT(bcache_write, |
153 | TP_PROTO(struct bio *bio, bool writeback, bool bypass), | ||
154 | TP_ARGS(bio, writeback, bypass), | ||
107 | 155 | ||
108 | TP_PROTO(struct bio *bio), | 156 | TP_STRUCT__entry( |
157 | __field(dev_t, dev ) | ||
158 | __field(sector_t, sector ) | ||
159 | __field(unsigned int, nr_sector ) | ||
160 | __array(char, rwbs, 6 ) | ||
161 | __field(bool, writeback ) | ||
162 | __field(bool, bypass ) | ||
163 | ), | ||
109 | 164 | ||
110 | TP_ARGS(bio) | 165 | TP_fast_assign( |
166 | __entry->dev = bio->bi_bdev->bd_dev; | ||
167 | __entry->sector = bio->bi_sector; | ||
168 | __entry->nr_sector = bio->bi_size >> 9; | ||
169 | blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_size); | ||
170 | __entry->writeback = writeback; | ||
171 | __entry->bypass = bypass; | ||
172 | ), | ||
173 | |||
174 | TP_printk("%d,%d %s %llu + %u hit %u bypass %u", | ||
175 | MAJOR(__entry->dev), MINOR(__entry->dev), | ||
176 | __entry->rwbs, (unsigned long long)__entry->sector, | ||
177 | __entry->nr_sector, __entry->writeback, __entry->bypass) | ||
111 | ); | 178 | ); |
112 | 179 | ||
113 | DEFINE_EVENT(bcache_bio, bcache_read_retry, | 180 | DEFINE_EVENT(bcache_bio, bcache_read_retry, |
114 | |||
115 | TP_PROTO(struct bio *bio), | 181 | TP_PROTO(struct bio *bio), |
116 | |||
117 | TP_ARGS(bio) | 182 | TP_ARGS(bio) |
118 | ); | 183 | ); |
119 | 184 | ||
120 | DEFINE_EVENT(bcache_bio, bcache_writethrough, | 185 | DEFINE_EVENT(bkey, bcache_cache_insert, |
186 | TP_PROTO(struct bkey *k), | ||
187 | TP_ARGS(k) | ||
188 | ); | ||
121 | 189 | ||
122 | TP_PROTO(struct bio *bio), | 190 | /* Journal */ |
123 | 191 | ||
124 | TP_ARGS(bio) | 192 | DECLARE_EVENT_CLASS(cache_set, |
125 | ); | 193 | TP_PROTO(struct cache_set *c), |
194 | TP_ARGS(c), | ||
126 | 195 | ||
127 | DEFINE_EVENT(bcache_bio, bcache_writeback, | 196 | TP_STRUCT__entry( |
197 | __array(char, uuid, 16 ) | ||
198 | ), | ||
128 | 199 | ||
129 | TP_PROTO(struct bio *bio), | 200 | TP_fast_assign( |
201 | memcpy(__entry->uuid, c->sb.set_uuid, 16); | ||
202 | ), | ||
130 | 203 | ||
131 | TP_ARGS(bio) | 204 | TP_printk("%pU", __entry->uuid) |
132 | ); | 205 | ); |
133 | 206 | ||
134 | DEFINE_EVENT(bcache_bio, bcache_write_skip, | 207 | DEFINE_EVENT(bkey, bcache_journal_replay_key, |
135 | 208 | TP_PROTO(struct bkey *k), | |
136 | TP_PROTO(struct bio *bio), | 209 | TP_ARGS(k) |
210 | ); | ||
137 | 211 | ||
138 | TP_ARGS(bio) | 212 | DEFINE_EVENT(cache_set, bcache_journal_full, |
213 | TP_PROTO(struct cache_set *c), | ||
214 | TP_ARGS(c) | ||
139 | ); | 215 | ); |
140 | 216 | ||
141 | DEFINE_EVENT(bcache_bio, bcache_btree_read, | 217 | DEFINE_EVENT(cache_set, bcache_journal_entry_full, |
218 | TP_PROTO(struct cache_set *c), | ||
219 | TP_ARGS(c) | ||
220 | ); | ||
142 | 221 | ||
222 | DEFINE_EVENT(bcache_bio, bcache_journal_write, | ||
143 | TP_PROTO(struct bio *bio), | 223 | TP_PROTO(struct bio *bio), |
144 | |||
145 | TP_ARGS(bio) | 224 | TP_ARGS(bio) |
146 | ); | 225 | ); |
147 | 226 | ||
148 | DEFINE_EVENT(bcache_bio, bcache_btree_write, | 227 | /* Btree */ |
149 | 228 | ||
150 | TP_PROTO(struct bio *bio), | 229 | DEFINE_EVENT(cache_set, bcache_btree_cache_cannibalize, |
230 | TP_PROTO(struct cache_set *c), | ||
231 | TP_ARGS(c) | ||
232 | ); | ||
151 | 233 | ||
152 | TP_ARGS(bio) | 234 | DEFINE_EVENT(btree_node, bcache_btree_read, |
235 | TP_PROTO(struct btree *b), | ||
236 | TP_ARGS(b) | ||
153 | ); | 237 | ); |
154 | 238 | ||
155 | DEFINE_EVENT(bcache_bio, bcache_write_dirty, | 239 | TRACE_EVENT(bcache_btree_write, |
240 | TP_PROTO(struct btree *b), | ||
241 | TP_ARGS(b), | ||
156 | 242 | ||
157 | TP_PROTO(struct bio *bio), | 243 | TP_STRUCT__entry( |
244 | __field(size_t, bucket ) | ||
245 | __field(unsigned, block ) | ||
246 | __field(unsigned, keys ) | ||
247 | ), | ||
158 | 248 | ||
159 | TP_ARGS(bio) | 249 | TP_fast_assign( |
250 | __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0); | ||
251 | __entry->block = b->written; | ||
252 | __entry->keys = b->sets[b->nsets].data->keys; | ||
253 | ), | ||
254 | |||
255 | TP_printk("bucket %zu", __entry->bucket) | ||
160 | ); | 256 | ); |
161 | 257 | ||
162 | DEFINE_EVENT(bcache_bio, bcache_read_dirty, | 258 | DEFINE_EVENT(btree_node, bcache_btree_node_alloc, |
259 | TP_PROTO(struct btree *b), | ||
260 | TP_ARGS(b) | ||
261 | ); | ||
163 | 262 | ||
164 | TP_PROTO(struct bio *bio), | 263 | DEFINE_EVENT(btree_node, bcache_btree_node_alloc_fail, |
264 | TP_PROTO(struct btree *b), | ||
265 | TP_ARGS(b) | ||
266 | ); | ||
165 | 267 | ||
166 | TP_ARGS(bio) | 268 | DEFINE_EVENT(btree_node, bcache_btree_node_free, |
269 | TP_PROTO(struct btree *b), | ||
270 | TP_ARGS(b) | ||
167 | ); | 271 | ); |
168 | 272 | ||
169 | DEFINE_EVENT(bcache_bio, bcache_write_moving, | 273 | TRACE_EVENT(bcache_btree_gc_coalesce, |
274 | TP_PROTO(unsigned nodes), | ||
275 | TP_ARGS(nodes), | ||
170 | 276 | ||
171 | TP_PROTO(struct bio *bio), | 277 | TP_STRUCT__entry( |
278 | __field(unsigned, nodes ) | ||
279 | ), | ||
172 | 280 | ||
173 | TP_ARGS(bio) | 281 | TP_fast_assign( |
282 | __entry->nodes = nodes; | ||
283 | ), | ||
284 | |||
285 | TP_printk("coalesced %u nodes", __entry->nodes) | ||
174 | ); | 286 | ); |
175 | 287 | ||
176 | DEFINE_EVENT(bcache_bio, bcache_read_moving, | 288 | DEFINE_EVENT(cache_set, bcache_gc_start, |
289 | TP_PROTO(struct cache_set *c), | ||
290 | TP_ARGS(c) | ||
291 | ); | ||
177 | 292 | ||
178 | TP_PROTO(struct bio *bio), | 293 | DEFINE_EVENT(cache_set, bcache_gc_end, |
294 | TP_PROTO(struct cache_set *c), | ||
295 | TP_ARGS(c) | ||
296 | ); | ||
179 | 297 | ||
180 | TP_ARGS(bio) | 298 | DEFINE_EVENT(bkey, bcache_gc_copy, |
299 | TP_PROTO(struct bkey *k), | ||
300 | TP_ARGS(k) | ||
181 | ); | 301 | ); |
182 | 302 | ||
183 | DEFINE_EVENT(bcache_bio, bcache_journal_write, | 303 | DEFINE_EVENT(bkey, bcache_gc_copy_collision, |
304 | TP_PROTO(struct bkey *k), | ||
305 | TP_ARGS(k) | ||
306 | ); | ||
184 | 307 | ||
185 | TP_PROTO(struct bio *bio), | 308 | TRACE_EVENT(bcache_btree_insert_key, |
309 | TP_PROTO(struct btree *b, struct bkey *k, unsigned op, unsigned status), | ||
310 | TP_ARGS(b, k, op, status), | ||
186 | 311 | ||
187 | TP_ARGS(bio) | 312 | TP_STRUCT__entry( |
188 | ); | 313 | __field(u64, btree_node ) |
314 | __field(u32, btree_level ) | ||
315 | __field(u32, inode ) | ||
316 | __field(u64, offset ) | ||
317 | __field(u32, size ) | ||
318 | __field(u8, dirty ) | ||
319 | __field(u8, op ) | ||
320 | __field(u8, status ) | ||
321 | ), | ||
189 | 322 | ||
190 | DECLARE_EVENT_CLASS(bcache_cache_bio, | 323 | TP_fast_assign( |
324 | __entry->btree_node = PTR_BUCKET_NR(b->c, &b->key, 0); | ||
325 | __entry->btree_level = b->level; | ||
326 | __entry->inode = KEY_INODE(k); | ||
327 | __entry->offset = KEY_OFFSET(k); | ||
328 | __entry->size = KEY_SIZE(k); | ||
329 | __entry->dirty = KEY_DIRTY(k); | ||
330 | __entry->op = op; | ||
331 | __entry->status = status; | ||
332 | ), | ||
191 | 333 | ||
192 | TP_PROTO(struct bio *bio, | 334 | TP_printk("%u for %u at %llu(%u): %u:%llu len %u dirty %u", |
193 | sector_t orig_sector, | 335 | __entry->status, __entry->op, |
194 | struct block_device* orig_bdev), | 336 | __entry->btree_node, __entry->btree_level, |
337 | __entry->inode, __entry->offset, | ||
338 | __entry->size, __entry->dirty) | ||
339 | ); | ||
195 | 340 | ||
196 | TP_ARGS(bio, orig_sector, orig_bdev), | 341 | DECLARE_EVENT_CLASS(btree_split, |
342 | TP_PROTO(struct btree *b, unsigned keys), | ||
343 | TP_ARGS(b, keys), | ||
197 | 344 | ||
198 | TP_STRUCT__entry( | 345 | TP_STRUCT__entry( |
199 | __field(dev_t, dev ) | 346 | __field(size_t, bucket ) |
200 | __field(dev_t, orig_dev ) | 347 | __field(unsigned, keys ) |
201 | __field(sector_t, sector ) | ||
202 | __field(sector_t, orig_sector ) | ||
203 | __field(unsigned int, nr_sector ) | ||
204 | __array(char, rwbs, 6 ) | ||
205 | __array(char, comm, TASK_COMM_LEN ) | ||
206 | ), | 348 | ), |
207 | 349 | ||
208 | TP_fast_assign( | 350 | TP_fast_assign( |
209 | __entry->dev = bio->bi_bdev->bd_dev; | 351 | __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0); |
210 | __entry->orig_dev = orig_bdev->bd_dev; | 352 | __entry->keys = keys; |
211 | __entry->sector = bio->bi_sector; | ||
212 | __entry->orig_sector = orig_sector; | ||
213 | __entry->nr_sector = bio->bi_size >> 9; | ||
214 | blk_fill_rwbs(__entry->rwbs, bio->bi_rw, bio->bi_size); | ||
215 | memcpy(__entry->comm, current->comm, TASK_COMM_LEN); | ||
216 | ), | 353 | ), |
217 | 354 | ||
218 | TP_printk("%d,%d %s %llu + %u [%s] (from %d,%d %llu)", | 355 | TP_printk("bucket %zu keys %u", __entry->bucket, __entry->keys) |
219 | MAJOR(__entry->dev), MINOR(__entry->dev), | ||
220 | __entry->rwbs, | ||
221 | (unsigned long long)__entry->sector, | ||
222 | __entry->nr_sector, __entry->comm, | ||
223 | MAJOR(__entry->orig_dev), MINOR(__entry->orig_dev), | ||
224 | (unsigned long long)__entry->orig_sector) | ||
225 | ); | 356 | ); |
226 | 357 | ||
227 | DEFINE_EVENT(bcache_cache_bio, bcache_cache_insert, | 358 | DEFINE_EVENT(btree_split, bcache_btree_node_split, |
228 | 359 | TP_PROTO(struct btree *b, unsigned keys), | |
229 | TP_PROTO(struct bio *bio, | 360 | TP_ARGS(b, keys) |
230 | sector_t orig_sector, | 361 | ); |
231 | struct block_device *orig_bdev), | ||
232 | 362 | ||
233 | TP_ARGS(bio, orig_sector, orig_bdev) | 363 | DEFINE_EVENT(btree_split, bcache_btree_node_compact, |
364 | TP_PROTO(struct btree *b, unsigned keys), | ||
365 | TP_ARGS(b, keys) | ||
234 | ); | 366 | ); |
235 | 367 | ||
236 | DECLARE_EVENT_CLASS(bcache_gc, | 368 | DEFINE_EVENT(btree_node, bcache_btree_set_root, |
369 | TP_PROTO(struct btree *b), | ||
370 | TP_ARGS(b) | ||
371 | ); | ||
237 | 372 | ||
238 | TP_PROTO(uint8_t *uuid), | 373 | /* Allocator */ |
239 | 374 | ||
240 | TP_ARGS(uuid), | 375 | TRACE_EVENT(bcache_alloc_invalidate, |
376 | TP_PROTO(struct cache *ca), | ||
377 | TP_ARGS(ca), | ||
241 | 378 | ||
242 | TP_STRUCT__entry( | 379 | TP_STRUCT__entry( |
243 | __field(uint8_t *, uuid) | 380 | __field(unsigned, free ) |
381 | __field(unsigned, free_inc ) | ||
382 | __field(unsigned, free_inc_size ) | ||
383 | __field(unsigned, unused ) | ||
244 | ), | 384 | ), |
245 | 385 | ||
246 | TP_fast_assign( | 386 | TP_fast_assign( |
247 | __entry->uuid = uuid; | 387 | __entry->free = fifo_used(&ca->free); |
388 | __entry->free_inc = fifo_used(&ca->free_inc); | ||
389 | __entry->free_inc_size = ca->free_inc.size; | ||
390 | __entry->unused = fifo_used(&ca->unused); | ||
248 | ), | 391 | ), |
249 | 392 | ||
250 | TP_printk("%pU", __entry->uuid) | 393 | TP_printk("free %u free_inc %u/%u unused %u", __entry->free, |
394 | __entry->free_inc, __entry->free_inc_size, __entry->unused) | ||
251 | ); | 395 | ); |
252 | 396 | ||
397 | TRACE_EVENT(bcache_alloc_fail, | ||
398 | TP_PROTO(struct cache *ca), | ||
399 | TP_ARGS(ca), | ||
253 | 400 | ||
254 | DEFINE_EVENT(bcache_gc, bcache_gc_start, | 401 | TP_STRUCT__entry( |
402 | __field(unsigned, free ) | ||
403 | __field(unsigned, free_inc ) | ||
404 | __field(unsigned, unused ) | ||
405 | __field(unsigned, blocked ) | ||
406 | ), | ||
255 | 407 | ||
256 | TP_PROTO(uint8_t *uuid), | 408 | TP_fast_assign( |
409 | __entry->free = fifo_used(&ca->free); | ||
410 | __entry->free_inc = fifo_used(&ca->free_inc); | ||
411 | __entry->unused = fifo_used(&ca->unused); | ||
412 | __entry->blocked = atomic_read(&ca->set->prio_blocked); | ||
413 | ), | ||
257 | 414 | ||
258 | TP_ARGS(uuid) | 415 | TP_printk("free %u free_inc %u unused %u blocked %u", __entry->free, |
416 | __entry->free_inc, __entry->unused, __entry->blocked) | ||
259 | ); | 417 | ); |
260 | 418 | ||
261 | DEFINE_EVENT(bcache_gc, bcache_gc_end, | 419 | /* Background writeback */ |
262 | 420 | ||
263 | TP_PROTO(uint8_t *uuid), | 421 | DEFINE_EVENT(bkey, bcache_writeback, |
422 | TP_PROTO(struct bkey *k), | ||
423 | TP_ARGS(k) | ||
424 | ); | ||
264 | 425 | ||
265 | TP_ARGS(uuid) | 426 | DEFINE_EVENT(bkey, bcache_writeback_collision, |
427 | TP_PROTO(struct bkey *k), | ||
428 | TP_ARGS(k) | ||
266 | ); | 429 | ); |
267 | 430 | ||
268 | #endif /* _TRACE_BCACHE_H */ | 431 | #endif /* _TRACE_BCACHE_H */ |
diff --git a/include/trace/ftrace.h b/include/trace/ftrace.h index d615f78cc6b6..41a6643e2136 100644 --- a/include/trace/ftrace.h +++ b/include/trace/ftrace.h | |||
@@ -670,10 +670,6 @@ perf_trace_##call(void *__data, proto) \ | |||
670 | sizeof(u64)); \ | 670 | sizeof(u64)); \ |
671 | __entry_size -= sizeof(u32); \ | 671 | __entry_size -= sizeof(u32); \ |
672 | \ | 672 | \ |
673 | if (WARN_ONCE(__entry_size > PERF_MAX_TRACE_SIZE, \ | ||
674 | "profile buffer not large enough")) \ | ||
675 | return; \ | ||
676 | \ | ||
677 | entry = (struct ftrace_raw_##call *)perf_trace_buf_prepare( \ | 673 | entry = (struct ftrace_raw_##call *)perf_trace_buf_prepare( \ |
678 | __entry_size, event_call->event.type, &__regs, &rctx); \ | 674 | __entry_size, event_call->event.type, &__regs, &rctx); \ |
679 | if (!entry) \ | 675 | if (!entry) \ |
diff --git a/include/uapi/linux/usb/ch11.h b/include/uapi/linux/usb/ch11.h index 7692dc69ccf7..331499d597fa 100644 --- a/include/uapi/linux/usb/ch11.h +++ b/include/uapi/linux/usb/ch11.h | |||
@@ -11,6 +11,17 @@ | |||
11 | 11 | ||
12 | #include <linux/types.h> /* __u8 etc */ | 12 | #include <linux/types.h> /* __u8 etc */ |
13 | 13 | ||
14 | /* This is arbitrary. | ||
15 | * From USB 2.0 spec Table 11-13, offset 7, a hub can | ||
16 | * have up to 255 ports. The most yet reported is 10. | ||
17 | * | ||
18 | * Current Wireless USB host hardware (Intel i1480 for example) allows | ||
19 | * up to 22 devices to connect. Upcoming hardware might raise that | ||
20 | * limit. Because the arrays need to add a bit for hub status data, we | ||
21 | * use 31, so plus one evens out to four bytes. | ||
22 | */ | ||
23 | #define USB_MAXCHILDREN 31 | ||
24 | |||
14 | /* | 25 | /* |
15 | * Hub request types | 26 | * Hub request types |
16 | */ | 27 | */ |
diff --git a/include/xen/interface/io/blkif.h b/include/xen/interface/io/blkif.h index ffd4652de91c..65e12099ef89 100644 --- a/include/xen/interface/io/blkif.h +++ b/include/xen/interface/io/blkif.h | |||
@@ -103,12 +103,46 @@ typedef uint64_t blkif_sector_t; | |||
103 | #define BLKIF_OP_DISCARD 5 | 103 | #define BLKIF_OP_DISCARD 5 |
104 | 104 | ||
105 | /* | 105 | /* |
106 | * Recognized if "feature-max-indirect-segments" in present in the backend | ||
107 | * xenbus info. The "feature-max-indirect-segments" node contains the maximum | ||
108 | * number of segments allowed by the backend per request. If the node is | ||
109 | * present, the frontend might use blkif_request_indirect structs in order to | ||
110 | * issue requests with more than BLKIF_MAX_SEGMENTS_PER_REQUEST (11). The | ||
111 | * maximum number of indirect segments is fixed by the backend, but the | ||
112 | * frontend can issue requests with any number of indirect segments as long as | ||
113 | * it's less than the number provided by the backend. The indirect_grefs field | ||
114 | * in blkif_request_indirect should be filled by the frontend with the | ||
115 | * grant references of the pages that are holding the indirect segments. | ||
116 | * This pages are filled with an array of blkif_request_segment_aligned | ||
117 | * that hold the information about the segments. The number of indirect | ||
118 | * pages to use is determined by the maximum number of segments | ||
119 | * a indirect request contains. Every indirect page can contain a maximum | ||
120 | * of 512 segments (PAGE_SIZE/sizeof(blkif_request_segment_aligned)), | ||
121 | * so to calculate the number of indirect pages to use we have to do | ||
122 | * ceil(indirect_segments/512). | ||
123 | * | ||
124 | * If a backend does not recognize BLKIF_OP_INDIRECT, it should *not* | ||
125 | * create the "feature-max-indirect-segments" node! | ||
126 | */ | ||
127 | #define BLKIF_OP_INDIRECT 6 | ||
128 | |||
129 | /* | ||
106 | * Maximum scatter/gather segments per request. | 130 | * Maximum scatter/gather segments per request. |
107 | * This is carefully chosen so that sizeof(struct blkif_ring) <= PAGE_SIZE. | 131 | * This is carefully chosen so that sizeof(struct blkif_ring) <= PAGE_SIZE. |
108 | * NB. This could be 12 if the ring indexes weren't stored in the same page. | 132 | * NB. This could be 12 if the ring indexes weren't stored in the same page. |
109 | */ | 133 | */ |
110 | #define BLKIF_MAX_SEGMENTS_PER_REQUEST 11 | 134 | #define BLKIF_MAX_SEGMENTS_PER_REQUEST 11 |
111 | 135 | ||
136 | #define BLKIF_MAX_INDIRECT_PAGES_PER_REQUEST 8 | ||
137 | |||
138 | struct blkif_request_segment_aligned { | ||
139 | grant_ref_t gref; /* reference to I/O buffer frame */ | ||
140 | /* @first_sect: first sector in frame to transfer (inclusive). */ | ||
141 | /* @last_sect: last sector in frame to transfer (inclusive). */ | ||
142 | uint8_t first_sect, last_sect; | ||
143 | uint16_t _pad; /* padding to make it 8 bytes, so it's cache-aligned */ | ||
144 | } __attribute__((__packed__)); | ||
145 | |||
112 | struct blkif_request_rw { | 146 | struct blkif_request_rw { |
113 | uint8_t nr_segments; /* number of segments */ | 147 | uint8_t nr_segments; /* number of segments */ |
114 | blkif_vdev_t handle; /* only for read/write requests */ | 148 | blkif_vdev_t handle; /* only for read/write requests */ |
@@ -147,12 +181,31 @@ struct blkif_request_other { | |||
147 | uint64_t id; /* private guest value, echoed in resp */ | 181 | uint64_t id; /* private guest value, echoed in resp */ |
148 | } __attribute__((__packed__)); | 182 | } __attribute__((__packed__)); |
149 | 183 | ||
184 | struct blkif_request_indirect { | ||
185 | uint8_t indirect_op; | ||
186 | uint16_t nr_segments; | ||
187 | #ifdef CONFIG_X86_64 | ||
188 | uint32_t _pad1; /* offsetof(blkif_...,u.indirect.id) == 8 */ | ||
189 | #endif | ||
190 | uint64_t id; | ||
191 | blkif_sector_t sector_number; | ||
192 | blkif_vdev_t handle; | ||
193 | uint16_t _pad2; | ||
194 | grant_ref_t indirect_grefs[BLKIF_MAX_INDIRECT_PAGES_PER_REQUEST]; | ||
195 | #ifdef CONFIG_X86_64 | ||
196 | uint32_t _pad3; /* make it 64 byte aligned */ | ||
197 | #else | ||
198 | uint64_t _pad3; /* make it 64 byte aligned */ | ||
199 | #endif | ||
200 | } __attribute__((__packed__)); | ||
201 | |||
150 | struct blkif_request { | 202 | struct blkif_request { |
151 | uint8_t operation; /* BLKIF_OP_??? */ | 203 | uint8_t operation; /* BLKIF_OP_??? */ |
152 | union { | 204 | union { |
153 | struct blkif_request_rw rw; | 205 | struct blkif_request_rw rw; |
154 | struct blkif_request_discard discard; | 206 | struct blkif_request_discard discard; |
155 | struct blkif_request_other other; | 207 | struct blkif_request_other other; |
208 | struct blkif_request_indirect indirect; | ||
156 | } u; | 209 | } u; |
157 | } __attribute__((__packed__)); | 210 | } __attribute__((__packed__)); |
158 | 211 | ||
diff --git a/include/xen/interface/io/ring.h b/include/xen/interface/io/ring.h index 75271b9a8f61..7d28aff605c7 100644 --- a/include/xen/interface/io/ring.h +++ b/include/xen/interface/io/ring.h | |||
@@ -188,6 +188,11 @@ struct __name##_back_ring { \ | |||
188 | #define RING_REQUEST_CONS_OVERFLOW(_r, _cons) \ | 188 | #define RING_REQUEST_CONS_OVERFLOW(_r, _cons) \ |
189 | (((_cons) - (_r)->rsp_prod_pvt) >= RING_SIZE(_r)) | 189 | (((_cons) - (_r)->rsp_prod_pvt) >= RING_SIZE(_r)) |
190 | 190 | ||
191 | /* Ill-behaved frontend determination: Can there be this many requests? */ | ||
192 | #define RING_REQUEST_PROD_OVERFLOW(_r, _prod) \ | ||
193 | (((_prod) - (_r)->rsp_prod_pvt) > RING_SIZE(_r)) | ||
194 | |||
195 | |||
191 | #define RING_PUSH_REQUESTS(_r) do { \ | 196 | #define RING_PUSH_REQUESTS(_r) do { \ |
192 | wmb(); /* back sees requests /before/ updated producer index */ \ | 197 | wmb(); /* back sees requests /before/ updated producer index */ \ |
193 | (_r)->sring->req_prod = (_r)->req_prod_pvt; \ | 198 | (_r)->sring->req_prod = (_r)->req_prod_pvt; \ |
diff --git a/kernel/cgroup.c b/kernel/cgroup.c index 0e0b20b8c5db..789ec4683db3 100644 --- a/kernel/cgroup.c +++ b/kernel/cgroup.c | |||
@@ -1845,36 +1845,43 @@ out: | |||
1845 | EXPORT_SYMBOL_GPL(cgroup_path); | 1845 | EXPORT_SYMBOL_GPL(cgroup_path); |
1846 | 1846 | ||
1847 | /** | 1847 | /** |
1848 | * task_cgroup_path_from_hierarchy - cgroup path of a task on a hierarchy | 1848 | * task_cgroup_path - cgroup path of a task in the first cgroup hierarchy |
1849 | * @task: target task | 1849 | * @task: target task |
1850 | * @hierarchy_id: the hierarchy to look up @task's cgroup from | ||
1851 | * @buf: the buffer to write the path into | 1850 | * @buf: the buffer to write the path into |
1852 | * @buflen: the length of the buffer | 1851 | * @buflen: the length of the buffer |
1853 | * | 1852 | * |
1854 | * Determine @task's cgroup on the hierarchy specified by @hierarchy_id and | 1853 | * Determine @task's cgroup on the first (the one with the lowest non-zero |
1855 | * copy its path into @buf. This function grabs cgroup_mutex and shouldn't | 1854 | * hierarchy_id) cgroup hierarchy and copy its path into @buf. This |
1856 | * be used inside locks used by cgroup controller callbacks. | 1855 | * function grabs cgroup_mutex and shouldn't be used inside locks used by |
1856 | * cgroup controller callbacks. | ||
1857 | * | ||
1858 | * Returns 0 on success, fails with -%ENAMETOOLONG if @buflen is too short. | ||
1857 | */ | 1859 | */ |
1858 | int task_cgroup_path_from_hierarchy(struct task_struct *task, int hierarchy_id, | 1860 | int task_cgroup_path(struct task_struct *task, char *buf, size_t buflen) |
1859 | char *buf, size_t buflen) | ||
1860 | { | 1861 | { |
1861 | struct cgroupfs_root *root; | 1862 | struct cgroupfs_root *root; |
1862 | struct cgroup *cgrp = NULL; | 1863 | struct cgroup *cgrp; |
1863 | int ret = -ENOENT; | 1864 | int hierarchy_id = 1, ret = 0; |
1865 | |||
1866 | if (buflen < 2) | ||
1867 | return -ENAMETOOLONG; | ||
1864 | 1868 | ||
1865 | mutex_lock(&cgroup_mutex); | 1869 | mutex_lock(&cgroup_mutex); |
1866 | 1870 | ||
1867 | root = idr_find(&cgroup_hierarchy_idr, hierarchy_id); | 1871 | root = idr_get_next(&cgroup_hierarchy_idr, &hierarchy_id); |
1872 | |||
1868 | if (root) { | 1873 | if (root) { |
1869 | cgrp = task_cgroup_from_root(task, root); | 1874 | cgrp = task_cgroup_from_root(task, root); |
1870 | ret = cgroup_path(cgrp, buf, buflen); | 1875 | ret = cgroup_path(cgrp, buf, buflen); |
1876 | } else { | ||
1877 | /* if no hierarchy exists, everyone is in "/" */ | ||
1878 | memcpy(buf, "/", 2); | ||
1871 | } | 1879 | } |
1872 | 1880 | ||
1873 | mutex_unlock(&cgroup_mutex); | 1881 | mutex_unlock(&cgroup_mutex); |
1874 | |||
1875 | return ret; | 1882 | return ret; |
1876 | } | 1883 | } |
1877 | EXPORT_SYMBOL_GPL(task_cgroup_path_from_hierarchy); | 1884 | EXPORT_SYMBOL_GPL(task_cgroup_path); |
1878 | 1885 | ||
1879 | /* | 1886 | /* |
1880 | * Control Group taskset | 1887 | * Control Group taskset |
diff --git a/kernel/trace/ftrace.c b/kernel/trace/ftrace.c index 67708f46baae..8ce9eefc5bb4 100644 --- a/kernel/trace/ftrace.c +++ b/kernel/trace/ftrace.c | |||
@@ -1441,12 +1441,22 @@ ftrace_hash_move(struct ftrace_ops *ops, int enable, | |||
1441 | * the hashes are freed with call_rcu_sched(). | 1441 | * the hashes are freed with call_rcu_sched(). |
1442 | */ | 1442 | */ |
1443 | static int | 1443 | static int |
1444 | ftrace_ops_test(struct ftrace_ops *ops, unsigned long ip) | 1444 | ftrace_ops_test(struct ftrace_ops *ops, unsigned long ip, void *regs) |
1445 | { | 1445 | { |
1446 | struct ftrace_hash *filter_hash; | 1446 | struct ftrace_hash *filter_hash; |
1447 | struct ftrace_hash *notrace_hash; | 1447 | struct ftrace_hash *notrace_hash; |
1448 | int ret; | 1448 | int ret; |
1449 | 1449 | ||
1450 | #ifdef CONFIG_DYNAMIC_FTRACE_WITH_REGS | ||
1451 | /* | ||
1452 | * There's a small race when adding ops that the ftrace handler | ||
1453 | * that wants regs, may be called without them. We can not | ||
1454 | * allow that handler to be called if regs is NULL. | ||
1455 | */ | ||
1456 | if (regs == NULL && (ops->flags & FTRACE_OPS_FL_SAVE_REGS)) | ||
1457 | return 0; | ||
1458 | #endif | ||
1459 | |||
1450 | filter_hash = rcu_dereference_raw_notrace(ops->filter_hash); | 1460 | filter_hash = rcu_dereference_raw_notrace(ops->filter_hash); |
1451 | notrace_hash = rcu_dereference_raw_notrace(ops->notrace_hash); | 1461 | notrace_hash = rcu_dereference_raw_notrace(ops->notrace_hash); |
1452 | 1462 | ||
@@ -4218,7 +4228,7 @@ static inline void ftrace_startup_enable(int command) { } | |||
4218 | # define ftrace_shutdown_sysctl() do { } while (0) | 4228 | # define ftrace_shutdown_sysctl() do { } while (0) |
4219 | 4229 | ||
4220 | static inline int | 4230 | static inline int |
4221 | ftrace_ops_test(struct ftrace_ops *ops, unsigned long ip) | 4231 | ftrace_ops_test(struct ftrace_ops *ops, unsigned long ip, void *regs) |
4222 | { | 4232 | { |
4223 | return 1; | 4233 | return 1; |
4224 | } | 4234 | } |
@@ -4241,7 +4251,7 @@ ftrace_ops_control_func(unsigned long ip, unsigned long parent_ip, | |||
4241 | do_for_each_ftrace_op(op, ftrace_control_list) { | 4251 | do_for_each_ftrace_op(op, ftrace_control_list) { |
4242 | if (!(op->flags & FTRACE_OPS_FL_STUB) && | 4252 | if (!(op->flags & FTRACE_OPS_FL_STUB) && |
4243 | !ftrace_function_local_disabled(op) && | 4253 | !ftrace_function_local_disabled(op) && |
4244 | ftrace_ops_test(op, ip)) | 4254 | ftrace_ops_test(op, ip, regs)) |
4245 | op->func(ip, parent_ip, op, regs); | 4255 | op->func(ip, parent_ip, op, regs); |
4246 | } while_for_each_ftrace_op(op); | 4256 | } while_for_each_ftrace_op(op); |
4247 | trace_recursion_clear(TRACE_CONTROL_BIT); | 4257 | trace_recursion_clear(TRACE_CONTROL_BIT); |
@@ -4274,7 +4284,7 @@ __ftrace_ops_list_func(unsigned long ip, unsigned long parent_ip, | |||
4274 | */ | 4284 | */ |
4275 | preempt_disable_notrace(); | 4285 | preempt_disable_notrace(); |
4276 | do_for_each_ftrace_op(op, ftrace_ops_list) { | 4286 | do_for_each_ftrace_op(op, ftrace_ops_list) { |
4277 | if (ftrace_ops_test(op, ip)) | 4287 | if (ftrace_ops_test(op, ip, regs)) |
4278 | op->func(ip, parent_ip, op, regs); | 4288 | op->func(ip, parent_ip, op, regs); |
4279 | } while_for_each_ftrace_op(op); | 4289 | } while_for_each_ftrace_op(op); |
4280 | preempt_enable_notrace(); | 4290 | preempt_enable_notrace(); |
diff --git a/kernel/trace/ring_buffer.c b/kernel/trace/ring_buffer.c index e444ff88f0a4..cc2f66f68dc5 100644 --- a/kernel/trace/ring_buffer.c +++ b/kernel/trace/ring_buffer.c | |||
@@ -36,11 +36,11 @@ int ring_buffer_print_entry_header(struct trace_seq *s) | |||
36 | { | 36 | { |
37 | int ret; | 37 | int ret; |
38 | 38 | ||
39 | ret = trace_seq_printf(s, "# compressed entry header\n"); | 39 | ret = trace_seq_puts(s, "# compressed entry header\n"); |
40 | ret = trace_seq_printf(s, "\ttype_len : 5 bits\n"); | 40 | ret = trace_seq_puts(s, "\ttype_len : 5 bits\n"); |
41 | ret = trace_seq_printf(s, "\ttime_delta : 27 bits\n"); | 41 | ret = trace_seq_puts(s, "\ttime_delta : 27 bits\n"); |
42 | ret = trace_seq_printf(s, "\tarray : 32 bits\n"); | 42 | ret = trace_seq_puts(s, "\tarray : 32 bits\n"); |
43 | ret = trace_seq_printf(s, "\n"); | 43 | ret = trace_seq_putc(s, '\n'); |
44 | ret = trace_seq_printf(s, "\tpadding : type == %d\n", | 44 | ret = trace_seq_printf(s, "\tpadding : type == %d\n", |
45 | RINGBUF_TYPE_PADDING); | 45 | RINGBUF_TYPE_PADDING); |
46 | ret = trace_seq_printf(s, "\ttime_extend : type == %d\n", | 46 | ret = trace_seq_printf(s, "\ttime_extend : type == %d\n", |
@@ -1066,7 +1066,7 @@ static int rb_check_list(struct ring_buffer_per_cpu *cpu_buffer, | |||
1066 | } | 1066 | } |
1067 | 1067 | ||
1068 | /** | 1068 | /** |
1069 | * check_pages - integrity check of buffer pages | 1069 | * rb_check_pages - integrity check of buffer pages |
1070 | * @cpu_buffer: CPU buffer with pages to test | 1070 | * @cpu_buffer: CPU buffer with pages to test |
1071 | * | 1071 | * |
1072 | * As a safety measure we check to make sure the data pages have not | 1072 | * As a safety measure we check to make sure the data pages have not |
@@ -1258,7 +1258,7 @@ static int rb_cpu_notify(struct notifier_block *self, | |||
1258 | #endif | 1258 | #endif |
1259 | 1259 | ||
1260 | /** | 1260 | /** |
1261 | * ring_buffer_alloc - allocate a new ring_buffer | 1261 | * __ring_buffer_alloc - allocate a new ring_buffer |
1262 | * @size: the size in bytes per cpu that is needed. | 1262 | * @size: the size in bytes per cpu that is needed. |
1263 | * @flags: attributes to set for the ring buffer. | 1263 | * @flags: attributes to set for the ring buffer. |
1264 | * | 1264 | * |
@@ -1607,6 +1607,7 @@ static void update_pages_handler(struct work_struct *work) | |||
1607 | * ring_buffer_resize - resize the ring buffer | 1607 | * ring_buffer_resize - resize the ring buffer |
1608 | * @buffer: the buffer to resize. | 1608 | * @buffer: the buffer to resize. |
1609 | * @size: the new size. | 1609 | * @size: the new size. |
1610 | * @cpu_id: the cpu buffer to resize | ||
1610 | * | 1611 | * |
1611 | * Minimum size is 2 * BUF_PAGE_SIZE. | 1612 | * Minimum size is 2 * BUF_PAGE_SIZE. |
1612 | * | 1613 | * |
@@ -3956,11 +3957,11 @@ EXPORT_SYMBOL_GPL(ring_buffer_consume); | |||
3956 | * expected. | 3957 | * expected. |
3957 | * | 3958 | * |
3958 | * After a sequence of ring_buffer_read_prepare calls, the user is | 3959 | * After a sequence of ring_buffer_read_prepare calls, the user is |
3959 | * expected to make at least one call to ring_buffer_prepare_sync. | 3960 | * expected to make at least one call to ring_buffer_read_prepare_sync. |
3960 | * Afterwards, ring_buffer_read_start is invoked to get things going | 3961 | * Afterwards, ring_buffer_read_start is invoked to get things going |
3961 | * for real. | 3962 | * for real. |
3962 | * | 3963 | * |
3963 | * This overall must be paired with ring_buffer_finish. | 3964 | * This overall must be paired with ring_buffer_read_finish. |
3964 | */ | 3965 | */ |
3965 | struct ring_buffer_iter * | 3966 | struct ring_buffer_iter * |
3966 | ring_buffer_read_prepare(struct ring_buffer *buffer, int cpu) | 3967 | ring_buffer_read_prepare(struct ring_buffer *buffer, int cpu) |
@@ -4009,7 +4010,7 @@ EXPORT_SYMBOL_GPL(ring_buffer_read_prepare_sync); | |||
4009 | * an intervening ring_buffer_read_prepare_sync must have been | 4010 | * an intervening ring_buffer_read_prepare_sync must have been |
4010 | * performed. | 4011 | * performed. |
4011 | * | 4012 | * |
4012 | * Must be paired with ring_buffer_finish. | 4013 | * Must be paired with ring_buffer_read_finish. |
4013 | */ | 4014 | */ |
4014 | void | 4015 | void |
4015 | ring_buffer_read_start(struct ring_buffer_iter *iter) | 4016 | ring_buffer_read_start(struct ring_buffer_iter *iter) |
@@ -4031,7 +4032,7 @@ ring_buffer_read_start(struct ring_buffer_iter *iter) | |||
4031 | EXPORT_SYMBOL_GPL(ring_buffer_read_start); | 4032 | EXPORT_SYMBOL_GPL(ring_buffer_read_start); |
4032 | 4033 | ||
4033 | /** | 4034 | /** |
4034 | * ring_buffer_finish - finish reading the iterator of the buffer | 4035 | * ring_buffer_read_finish - finish reading the iterator of the buffer |
4035 | * @iter: The iterator retrieved by ring_buffer_start | 4036 | * @iter: The iterator retrieved by ring_buffer_start |
4036 | * | 4037 | * |
4037 | * This re-enables the recording to the buffer, and frees the | 4038 | * This re-enables the recording to the buffer, and frees the |
@@ -4346,6 +4347,7 @@ EXPORT_SYMBOL_GPL(ring_buffer_swap_cpu); | |||
4346 | /** | 4347 | /** |
4347 | * ring_buffer_alloc_read_page - allocate a page to read from buffer | 4348 | * ring_buffer_alloc_read_page - allocate a page to read from buffer |
4348 | * @buffer: the buffer to allocate for. | 4349 | * @buffer: the buffer to allocate for. |
4350 | * @cpu: the cpu buffer to allocate. | ||
4349 | * | 4351 | * |
4350 | * This function is used in conjunction with ring_buffer_read_page. | 4352 | * This function is used in conjunction with ring_buffer_read_page. |
4351 | * When reading a full page from the ring buffer, these functions | 4353 | * When reading a full page from the ring buffer, these functions |
@@ -4403,7 +4405,7 @@ EXPORT_SYMBOL_GPL(ring_buffer_free_read_page); | |||
4403 | * to swap with a page in the ring buffer. | 4405 | * to swap with a page in the ring buffer. |
4404 | * | 4406 | * |
4405 | * for example: | 4407 | * for example: |
4406 | * rpage = ring_buffer_alloc_read_page(buffer); | 4408 | * rpage = ring_buffer_alloc_read_page(buffer, cpu); |
4407 | * if (!rpage) | 4409 | * if (!rpage) |
4408 | * return error; | 4410 | * return error; |
4409 | * ret = ring_buffer_read_page(buffer, &rpage, len, cpu, 0); | 4411 | * ret = ring_buffer_read_page(buffer, &rpage, len, cpu, 0); |
diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c index 0cd500bffd9b..882ec1dd1515 100644 --- a/kernel/trace/trace.c +++ b/kernel/trace/trace.c | |||
@@ -1224,18 +1224,17 @@ void tracing_reset_current(int cpu) | |||
1224 | tracing_reset(&global_trace.trace_buffer, cpu); | 1224 | tracing_reset(&global_trace.trace_buffer, cpu); |
1225 | } | 1225 | } |
1226 | 1226 | ||
1227 | /* Must have trace_types_lock held */ | ||
1227 | void tracing_reset_all_online_cpus(void) | 1228 | void tracing_reset_all_online_cpus(void) |
1228 | { | 1229 | { |
1229 | struct trace_array *tr; | 1230 | struct trace_array *tr; |
1230 | 1231 | ||
1231 | mutex_lock(&trace_types_lock); | ||
1232 | list_for_each_entry(tr, &ftrace_trace_arrays, list) { | 1232 | list_for_each_entry(tr, &ftrace_trace_arrays, list) { |
1233 | tracing_reset_online_cpus(&tr->trace_buffer); | 1233 | tracing_reset_online_cpus(&tr->trace_buffer); |
1234 | #ifdef CONFIG_TRACER_MAX_TRACE | 1234 | #ifdef CONFIG_TRACER_MAX_TRACE |
1235 | tracing_reset_online_cpus(&tr->max_buffer); | 1235 | tracing_reset_online_cpus(&tr->max_buffer); |
1236 | #endif | 1236 | #endif |
1237 | } | 1237 | } |
1238 | mutex_unlock(&trace_types_lock); | ||
1239 | } | 1238 | } |
1240 | 1239 | ||
1241 | #define SAVED_CMDLINES 128 | 1240 | #define SAVED_CMDLINES 128 |
@@ -2843,6 +2842,17 @@ static int s_show(struct seq_file *m, void *v) | |||
2843 | return 0; | 2842 | return 0; |
2844 | } | 2843 | } |
2845 | 2844 | ||
2845 | /* | ||
2846 | * Should be used after trace_array_get(), trace_types_lock | ||
2847 | * ensures that i_cdev was already initialized. | ||
2848 | */ | ||
2849 | static inline int tracing_get_cpu(struct inode *inode) | ||
2850 | { | ||
2851 | if (inode->i_cdev) /* See trace_create_cpu_file() */ | ||
2852 | return (long)inode->i_cdev - 1; | ||
2853 | return RING_BUFFER_ALL_CPUS; | ||
2854 | } | ||
2855 | |||
2846 | static const struct seq_operations tracer_seq_ops = { | 2856 | static const struct seq_operations tracer_seq_ops = { |
2847 | .start = s_start, | 2857 | .start = s_start, |
2848 | .next = s_next, | 2858 | .next = s_next, |
@@ -2851,9 +2861,9 @@ static const struct seq_operations tracer_seq_ops = { | |||
2851 | }; | 2861 | }; |
2852 | 2862 | ||
2853 | static struct trace_iterator * | 2863 | static struct trace_iterator * |
2854 | __tracing_open(struct trace_array *tr, struct trace_cpu *tc, | 2864 | __tracing_open(struct inode *inode, struct file *file, bool snapshot) |
2855 | struct inode *inode, struct file *file, bool snapshot) | ||
2856 | { | 2865 | { |
2866 | struct trace_array *tr = inode->i_private; | ||
2857 | struct trace_iterator *iter; | 2867 | struct trace_iterator *iter; |
2858 | int cpu; | 2868 | int cpu; |
2859 | 2869 | ||
@@ -2894,8 +2904,8 @@ __tracing_open(struct trace_array *tr, struct trace_cpu *tc, | |||
2894 | iter->trace_buffer = &tr->trace_buffer; | 2904 | iter->trace_buffer = &tr->trace_buffer; |
2895 | iter->snapshot = snapshot; | 2905 | iter->snapshot = snapshot; |
2896 | iter->pos = -1; | 2906 | iter->pos = -1; |
2907 | iter->cpu_file = tracing_get_cpu(inode); | ||
2897 | mutex_init(&iter->mutex); | 2908 | mutex_init(&iter->mutex); |
2898 | iter->cpu_file = tc->cpu; | ||
2899 | 2909 | ||
2900 | /* Notify the tracer early; before we stop tracing. */ | 2910 | /* Notify the tracer early; before we stop tracing. */ |
2901 | if (iter->trace && iter->trace->open) | 2911 | if (iter->trace && iter->trace->open) |
@@ -2971,45 +2981,22 @@ static int tracing_open_generic_tr(struct inode *inode, struct file *filp) | |||
2971 | filp->private_data = inode->i_private; | 2981 | filp->private_data = inode->i_private; |
2972 | 2982 | ||
2973 | return 0; | 2983 | return 0; |
2974 | |||
2975 | } | ||
2976 | |||
2977 | static int tracing_open_generic_tc(struct inode *inode, struct file *filp) | ||
2978 | { | ||
2979 | struct trace_cpu *tc = inode->i_private; | ||
2980 | struct trace_array *tr = tc->tr; | ||
2981 | |||
2982 | if (tracing_disabled) | ||
2983 | return -ENODEV; | ||
2984 | |||
2985 | if (trace_array_get(tr) < 0) | ||
2986 | return -ENODEV; | ||
2987 | |||
2988 | filp->private_data = inode->i_private; | ||
2989 | |||
2990 | return 0; | ||
2991 | |||
2992 | } | 2984 | } |
2993 | 2985 | ||
2994 | static int tracing_release(struct inode *inode, struct file *file) | 2986 | static int tracing_release(struct inode *inode, struct file *file) |
2995 | { | 2987 | { |
2988 | struct trace_array *tr = inode->i_private; | ||
2996 | struct seq_file *m = file->private_data; | 2989 | struct seq_file *m = file->private_data; |
2997 | struct trace_iterator *iter; | 2990 | struct trace_iterator *iter; |
2998 | struct trace_array *tr; | ||
2999 | int cpu; | 2991 | int cpu; |
3000 | 2992 | ||
3001 | /* Writes do not use seq_file, need to grab tr from inode */ | ||
3002 | if (!(file->f_mode & FMODE_READ)) { | 2993 | if (!(file->f_mode & FMODE_READ)) { |
3003 | struct trace_cpu *tc = inode->i_private; | 2994 | trace_array_put(tr); |
3004 | |||
3005 | trace_array_put(tc->tr); | ||
3006 | return 0; | 2995 | return 0; |
3007 | } | 2996 | } |
3008 | 2997 | ||
2998 | /* Writes do not use seq_file */ | ||
3009 | iter = m->private; | 2999 | iter = m->private; |
3010 | tr = iter->tr; | ||
3011 | trace_array_put(tr); | ||
3012 | |||
3013 | mutex_lock(&trace_types_lock); | 3000 | mutex_lock(&trace_types_lock); |
3014 | 3001 | ||
3015 | for_each_tracing_cpu(cpu) { | 3002 | for_each_tracing_cpu(cpu) { |
@@ -3023,6 +3010,9 @@ static int tracing_release(struct inode *inode, struct file *file) | |||
3023 | if (!iter->snapshot) | 3010 | if (!iter->snapshot) |
3024 | /* reenable tracing if it was previously enabled */ | 3011 | /* reenable tracing if it was previously enabled */ |
3025 | tracing_start_tr(tr); | 3012 | tracing_start_tr(tr); |
3013 | |||
3014 | __trace_array_put(tr); | ||
3015 | |||
3026 | mutex_unlock(&trace_types_lock); | 3016 | mutex_unlock(&trace_types_lock); |
3027 | 3017 | ||
3028 | mutex_destroy(&iter->mutex); | 3018 | mutex_destroy(&iter->mutex); |
@@ -3042,15 +3032,6 @@ static int tracing_release_generic_tr(struct inode *inode, struct file *file) | |||
3042 | return 0; | 3032 | return 0; |
3043 | } | 3033 | } |
3044 | 3034 | ||
3045 | static int tracing_release_generic_tc(struct inode *inode, struct file *file) | ||
3046 | { | ||
3047 | struct trace_cpu *tc = inode->i_private; | ||
3048 | struct trace_array *tr = tc->tr; | ||
3049 | |||
3050 | trace_array_put(tr); | ||
3051 | return 0; | ||
3052 | } | ||
3053 | |||
3054 | static int tracing_single_release_tr(struct inode *inode, struct file *file) | 3035 | static int tracing_single_release_tr(struct inode *inode, struct file *file) |
3055 | { | 3036 | { |
3056 | struct trace_array *tr = inode->i_private; | 3037 | struct trace_array *tr = inode->i_private; |
@@ -3062,8 +3043,7 @@ static int tracing_single_release_tr(struct inode *inode, struct file *file) | |||
3062 | 3043 | ||
3063 | static int tracing_open(struct inode *inode, struct file *file) | 3044 | static int tracing_open(struct inode *inode, struct file *file) |
3064 | { | 3045 | { |
3065 | struct trace_cpu *tc = inode->i_private; | 3046 | struct trace_array *tr = inode->i_private; |
3066 | struct trace_array *tr = tc->tr; | ||
3067 | struct trace_iterator *iter; | 3047 | struct trace_iterator *iter; |
3068 | int ret = 0; | 3048 | int ret = 0; |
3069 | 3049 | ||
@@ -3071,16 +3051,17 @@ static int tracing_open(struct inode *inode, struct file *file) | |||
3071 | return -ENODEV; | 3051 | return -ENODEV; |
3072 | 3052 | ||
3073 | /* If this file was open for write, then erase contents */ | 3053 | /* If this file was open for write, then erase contents */ |
3074 | if ((file->f_mode & FMODE_WRITE) && | 3054 | if ((file->f_mode & FMODE_WRITE) && (file->f_flags & O_TRUNC)) { |
3075 | (file->f_flags & O_TRUNC)) { | 3055 | int cpu = tracing_get_cpu(inode); |
3076 | if (tc->cpu == RING_BUFFER_ALL_CPUS) | 3056 | |
3057 | if (cpu == RING_BUFFER_ALL_CPUS) | ||
3077 | tracing_reset_online_cpus(&tr->trace_buffer); | 3058 | tracing_reset_online_cpus(&tr->trace_buffer); |
3078 | else | 3059 | else |
3079 | tracing_reset(&tr->trace_buffer, tc->cpu); | 3060 | tracing_reset(&tr->trace_buffer, cpu); |
3080 | } | 3061 | } |
3081 | 3062 | ||
3082 | if (file->f_mode & FMODE_READ) { | 3063 | if (file->f_mode & FMODE_READ) { |
3083 | iter = __tracing_open(tr, tc, inode, file, false); | 3064 | iter = __tracing_open(inode, file, false); |
3084 | if (IS_ERR(iter)) | 3065 | if (IS_ERR(iter)) |
3085 | ret = PTR_ERR(iter); | 3066 | ret = PTR_ERR(iter); |
3086 | else if (trace_flags & TRACE_ITER_LATENCY_FMT) | 3067 | else if (trace_flags & TRACE_ITER_LATENCY_FMT) |
@@ -3447,6 +3428,7 @@ tracing_trace_options_write(struct file *filp, const char __user *ubuf, | |||
3447 | static int tracing_trace_options_open(struct inode *inode, struct file *file) | 3428 | static int tracing_trace_options_open(struct inode *inode, struct file *file) |
3448 | { | 3429 | { |
3449 | struct trace_array *tr = inode->i_private; | 3430 | struct trace_array *tr = inode->i_private; |
3431 | int ret; | ||
3450 | 3432 | ||
3451 | if (tracing_disabled) | 3433 | if (tracing_disabled) |
3452 | return -ENODEV; | 3434 | return -ENODEV; |
@@ -3454,7 +3436,11 @@ static int tracing_trace_options_open(struct inode *inode, struct file *file) | |||
3454 | if (trace_array_get(tr) < 0) | 3436 | if (trace_array_get(tr) < 0) |
3455 | return -ENODEV; | 3437 | return -ENODEV; |
3456 | 3438 | ||
3457 | return single_open(file, tracing_trace_options_show, inode->i_private); | 3439 | ret = single_open(file, tracing_trace_options_show, inode->i_private); |
3440 | if (ret < 0) | ||
3441 | trace_array_put(tr); | ||
3442 | |||
3443 | return ret; | ||
3458 | } | 3444 | } |
3459 | 3445 | ||
3460 | static const struct file_operations tracing_iter_fops = { | 3446 | static const struct file_operations tracing_iter_fops = { |
@@ -3537,14 +3523,14 @@ static const char readme_msg[] = | |||
3537 | "\n snapshot\t\t- Like 'trace' but shows the content of the static snapshot buffer\n" | 3523 | "\n snapshot\t\t- Like 'trace' but shows the content of the static snapshot buffer\n" |
3538 | "\t\t\t Read the contents for more information\n" | 3524 | "\t\t\t Read the contents for more information\n" |
3539 | #endif | 3525 | #endif |
3540 | #ifdef CONFIG_STACKTRACE | 3526 | #ifdef CONFIG_STACK_TRACER |
3541 | " stack_trace\t\t- Shows the max stack trace when active\n" | 3527 | " stack_trace\t\t- Shows the max stack trace when active\n" |
3542 | " stack_max_size\t- Shows current max stack size that was traced\n" | 3528 | " stack_max_size\t- Shows current max stack size that was traced\n" |
3543 | "\t\t\t Write into this file to reset the max size (trigger a new trace)\n" | 3529 | "\t\t\t Write into this file to reset the max size (trigger a new trace)\n" |
3544 | #ifdef CONFIG_DYNAMIC_FTRACE | 3530 | #ifdef CONFIG_DYNAMIC_FTRACE |
3545 | " stack_trace_filter\t- Like set_ftrace_filter but limits what stack_trace traces\n" | 3531 | " stack_trace_filter\t- Like set_ftrace_filter but limits what stack_trace traces\n" |
3546 | #endif | 3532 | #endif |
3547 | #endif /* CONFIG_STACKTRACE */ | 3533 | #endif /* CONFIG_STACK_TRACER */ |
3548 | ; | 3534 | ; |
3549 | 3535 | ||
3550 | static ssize_t | 3536 | static ssize_t |
@@ -3941,8 +3927,7 @@ tracing_max_lat_write(struct file *filp, const char __user *ubuf, | |||
3941 | 3927 | ||
3942 | static int tracing_open_pipe(struct inode *inode, struct file *filp) | 3928 | static int tracing_open_pipe(struct inode *inode, struct file *filp) |
3943 | { | 3929 | { |
3944 | struct trace_cpu *tc = inode->i_private; | 3930 | struct trace_array *tr = inode->i_private; |
3945 | struct trace_array *tr = tc->tr; | ||
3946 | struct trace_iterator *iter; | 3931 | struct trace_iterator *iter; |
3947 | int ret = 0; | 3932 | int ret = 0; |
3948 | 3933 | ||
@@ -3958,6 +3943,7 @@ static int tracing_open_pipe(struct inode *inode, struct file *filp) | |||
3958 | iter = kzalloc(sizeof(*iter), GFP_KERNEL); | 3943 | iter = kzalloc(sizeof(*iter), GFP_KERNEL); |
3959 | if (!iter) { | 3944 | if (!iter) { |
3960 | ret = -ENOMEM; | 3945 | ret = -ENOMEM; |
3946 | __trace_array_put(tr); | ||
3961 | goto out; | 3947 | goto out; |
3962 | } | 3948 | } |
3963 | 3949 | ||
@@ -3987,9 +3973,9 @@ static int tracing_open_pipe(struct inode *inode, struct file *filp) | |||
3987 | if (trace_clocks[tr->clock_id].in_ns) | 3973 | if (trace_clocks[tr->clock_id].in_ns) |
3988 | iter->iter_flags |= TRACE_FILE_TIME_IN_NS; | 3974 | iter->iter_flags |= TRACE_FILE_TIME_IN_NS; |
3989 | 3975 | ||
3990 | iter->cpu_file = tc->cpu; | 3976 | iter->tr = tr; |
3991 | iter->tr = tc->tr; | 3977 | iter->trace_buffer = &tr->trace_buffer; |
3992 | iter->trace_buffer = &tc->tr->trace_buffer; | 3978 | iter->cpu_file = tracing_get_cpu(inode); |
3993 | mutex_init(&iter->mutex); | 3979 | mutex_init(&iter->mutex); |
3994 | filp->private_data = iter; | 3980 | filp->private_data = iter; |
3995 | 3981 | ||
@@ -4012,8 +3998,7 @@ fail: | |||
4012 | static int tracing_release_pipe(struct inode *inode, struct file *file) | 3998 | static int tracing_release_pipe(struct inode *inode, struct file *file) |
4013 | { | 3999 | { |
4014 | struct trace_iterator *iter = file->private_data; | 4000 | struct trace_iterator *iter = file->private_data; |
4015 | struct trace_cpu *tc = inode->i_private; | 4001 | struct trace_array *tr = inode->i_private; |
4016 | struct trace_array *tr = tc->tr; | ||
4017 | 4002 | ||
4018 | mutex_lock(&trace_types_lock); | 4003 | mutex_lock(&trace_types_lock); |
4019 | 4004 | ||
@@ -4366,15 +4351,16 @@ static ssize_t | |||
4366 | tracing_entries_read(struct file *filp, char __user *ubuf, | 4351 | tracing_entries_read(struct file *filp, char __user *ubuf, |
4367 | size_t cnt, loff_t *ppos) | 4352 | size_t cnt, loff_t *ppos) |
4368 | { | 4353 | { |
4369 | struct trace_cpu *tc = filp->private_data; | 4354 | struct inode *inode = file_inode(filp); |
4370 | struct trace_array *tr = tc->tr; | 4355 | struct trace_array *tr = inode->i_private; |
4356 | int cpu = tracing_get_cpu(inode); | ||
4371 | char buf[64]; | 4357 | char buf[64]; |
4372 | int r = 0; | 4358 | int r = 0; |
4373 | ssize_t ret; | 4359 | ssize_t ret; |
4374 | 4360 | ||
4375 | mutex_lock(&trace_types_lock); | 4361 | mutex_lock(&trace_types_lock); |
4376 | 4362 | ||
4377 | if (tc->cpu == RING_BUFFER_ALL_CPUS) { | 4363 | if (cpu == RING_BUFFER_ALL_CPUS) { |
4378 | int cpu, buf_size_same; | 4364 | int cpu, buf_size_same; |
4379 | unsigned long size; | 4365 | unsigned long size; |
4380 | 4366 | ||
@@ -4401,7 +4387,7 @@ tracing_entries_read(struct file *filp, char __user *ubuf, | |||
4401 | } else | 4387 | } else |
4402 | r = sprintf(buf, "X\n"); | 4388 | r = sprintf(buf, "X\n"); |
4403 | } else | 4389 | } else |
4404 | r = sprintf(buf, "%lu\n", per_cpu_ptr(tr->trace_buffer.data, tc->cpu)->entries >> 10); | 4390 | r = sprintf(buf, "%lu\n", per_cpu_ptr(tr->trace_buffer.data, cpu)->entries >> 10); |
4405 | 4391 | ||
4406 | mutex_unlock(&trace_types_lock); | 4392 | mutex_unlock(&trace_types_lock); |
4407 | 4393 | ||
@@ -4413,7 +4399,8 @@ static ssize_t | |||
4413 | tracing_entries_write(struct file *filp, const char __user *ubuf, | 4399 | tracing_entries_write(struct file *filp, const char __user *ubuf, |
4414 | size_t cnt, loff_t *ppos) | 4400 | size_t cnt, loff_t *ppos) |
4415 | { | 4401 | { |
4416 | struct trace_cpu *tc = filp->private_data; | 4402 | struct inode *inode = file_inode(filp); |
4403 | struct trace_array *tr = inode->i_private; | ||
4417 | unsigned long val; | 4404 | unsigned long val; |
4418 | int ret; | 4405 | int ret; |
4419 | 4406 | ||
@@ -4427,8 +4414,7 @@ tracing_entries_write(struct file *filp, const char __user *ubuf, | |||
4427 | 4414 | ||
4428 | /* value is in KB */ | 4415 | /* value is in KB */ |
4429 | val <<= 10; | 4416 | val <<= 10; |
4430 | 4417 | ret = tracing_resize_ring_buffer(tr, val, tracing_get_cpu(inode)); | |
4431 | ret = tracing_resize_ring_buffer(tc->tr, val, tc->cpu); | ||
4432 | if (ret < 0) | 4418 | if (ret < 0) |
4433 | return ret; | 4419 | return ret; |
4434 | 4420 | ||
@@ -4689,8 +4675,7 @@ struct ftrace_buffer_info { | |||
4689 | #ifdef CONFIG_TRACER_SNAPSHOT | 4675 | #ifdef CONFIG_TRACER_SNAPSHOT |
4690 | static int tracing_snapshot_open(struct inode *inode, struct file *file) | 4676 | static int tracing_snapshot_open(struct inode *inode, struct file *file) |
4691 | { | 4677 | { |
4692 | struct trace_cpu *tc = inode->i_private; | 4678 | struct trace_array *tr = inode->i_private; |
4693 | struct trace_array *tr = tc->tr; | ||
4694 | struct trace_iterator *iter; | 4679 | struct trace_iterator *iter; |
4695 | struct seq_file *m; | 4680 | struct seq_file *m; |
4696 | int ret = 0; | 4681 | int ret = 0; |
@@ -4699,26 +4684,29 @@ static int tracing_snapshot_open(struct inode *inode, struct file *file) | |||
4699 | return -ENODEV; | 4684 | return -ENODEV; |
4700 | 4685 | ||
4701 | if (file->f_mode & FMODE_READ) { | 4686 | if (file->f_mode & FMODE_READ) { |
4702 | iter = __tracing_open(tr, tc, inode, file, true); | 4687 | iter = __tracing_open(inode, file, true); |
4703 | if (IS_ERR(iter)) | 4688 | if (IS_ERR(iter)) |
4704 | ret = PTR_ERR(iter); | 4689 | ret = PTR_ERR(iter); |
4705 | } else { | 4690 | } else { |
4706 | /* Writes still need the seq_file to hold the private data */ | 4691 | /* Writes still need the seq_file to hold the private data */ |
4692 | ret = -ENOMEM; | ||
4707 | m = kzalloc(sizeof(*m), GFP_KERNEL); | 4693 | m = kzalloc(sizeof(*m), GFP_KERNEL); |
4708 | if (!m) | 4694 | if (!m) |
4709 | return -ENOMEM; | 4695 | goto out; |
4710 | iter = kzalloc(sizeof(*iter), GFP_KERNEL); | 4696 | iter = kzalloc(sizeof(*iter), GFP_KERNEL); |
4711 | if (!iter) { | 4697 | if (!iter) { |
4712 | kfree(m); | 4698 | kfree(m); |
4713 | return -ENOMEM; | 4699 | goto out; |
4714 | } | 4700 | } |
4701 | ret = 0; | ||
4702 | |||
4715 | iter->tr = tr; | 4703 | iter->tr = tr; |
4716 | iter->trace_buffer = &tc->tr->max_buffer; | 4704 | iter->trace_buffer = &tr->max_buffer; |
4717 | iter->cpu_file = tc->cpu; | 4705 | iter->cpu_file = tracing_get_cpu(inode); |
4718 | m->private = iter; | 4706 | m->private = iter; |
4719 | file->private_data = m; | 4707 | file->private_data = m; |
4720 | } | 4708 | } |
4721 | 4709 | out: | |
4722 | if (ret < 0) | 4710 | if (ret < 0) |
4723 | trace_array_put(tr); | 4711 | trace_array_put(tr); |
4724 | 4712 | ||
@@ -4873,11 +4861,11 @@ static const struct file_operations tracing_pipe_fops = { | |||
4873 | }; | 4861 | }; |
4874 | 4862 | ||
4875 | static const struct file_operations tracing_entries_fops = { | 4863 | static const struct file_operations tracing_entries_fops = { |
4876 | .open = tracing_open_generic_tc, | 4864 | .open = tracing_open_generic_tr, |
4877 | .read = tracing_entries_read, | 4865 | .read = tracing_entries_read, |
4878 | .write = tracing_entries_write, | 4866 | .write = tracing_entries_write, |
4879 | .llseek = generic_file_llseek, | 4867 | .llseek = generic_file_llseek, |
4880 | .release = tracing_release_generic_tc, | 4868 | .release = tracing_release_generic_tr, |
4881 | }; | 4869 | }; |
4882 | 4870 | ||
4883 | static const struct file_operations tracing_total_entries_fops = { | 4871 | static const struct file_operations tracing_total_entries_fops = { |
@@ -4929,8 +4917,7 @@ static const struct file_operations snapshot_raw_fops = { | |||
4929 | 4917 | ||
4930 | static int tracing_buffers_open(struct inode *inode, struct file *filp) | 4918 | static int tracing_buffers_open(struct inode *inode, struct file *filp) |
4931 | { | 4919 | { |
4932 | struct trace_cpu *tc = inode->i_private; | 4920 | struct trace_array *tr = inode->i_private; |
4933 | struct trace_array *tr = tc->tr; | ||
4934 | struct ftrace_buffer_info *info; | 4921 | struct ftrace_buffer_info *info; |
4935 | int ret; | 4922 | int ret; |
4936 | 4923 | ||
@@ -4948,10 +4935,8 @@ static int tracing_buffers_open(struct inode *inode, struct file *filp) | |||
4948 | 4935 | ||
4949 | mutex_lock(&trace_types_lock); | 4936 | mutex_lock(&trace_types_lock); |
4950 | 4937 | ||
4951 | tr->ref++; | ||
4952 | |||
4953 | info->iter.tr = tr; | 4938 | info->iter.tr = tr; |
4954 | info->iter.cpu_file = tc->cpu; | 4939 | info->iter.cpu_file = tracing_get_cpu(inode); |
4955 | info->iter.trace = tr->current_trace; | 4940 | info->iter.trace = tr->current_trace; |
4956 | info->iter.trace_buffer = &tr->trace_buffer; | 4941 | info->iter.trace_buffer = &tr->trace_buffer; |
4957 | info->spare = NULL; | 4942 | info->spare = NULL; |
@@ -5268,14 +5253,14 @@ static ssize_t | |||
5268 | tracing_stats_read(struct file *filp, char __user *ubuf, | 5253 | tracing_stats_read(struct file *filp, char __user *ubuf, |
5269 | size_t count, loff_t *ppos) | 5254 | size_t count, loff_t *ppos) |
5270 | { | 5255 | { |
5271 | struct trace_cpu *tc = filp->private_data; | 5256 | struct inode *inode = file_inode(filp); |
5272 | struct trace_array *tr = tc->tr; | 5257 | struct trace_array *tr = inode->i_private; |
5273 | struct trace_buffer *trace_buf = &tr->trace_buffer; | 5258 | struct trace_buffer *trace_buf = &tr->trace_buffer; |
5259 | int cpu = tracing_get_cpu(inode); | ||
5274 | struct trace_seq *s; | 5260 | struct trace_seq *s; |
5275 | unsigned long cnt; | 5261 | unsigned long cnt; |
5276 | unsigned long long t; | 5262 | unsigned long long t; |
5277 | unsigned long usec_rem; | 5263 | unsigned long usec_rem; |
5278 | int cpu = tc->cpu; | ||
5279 | 5264 | ||
5280 | s = kmalloc(sizeof(*s), GFP_KERNEL); | 5265 | s = kmalloc(sizeof(*s), GFP_KERNEL); |
5281 | if (!s) | 5266 | if (!s) |
@@ -5328,9 +5313,10 @@ tracing_stats_read(struct file *filp, char __user *ubuf, | |||
5328 | } | 5313 | } |
5329 | 5314 | ||
5330 | static const struct file_operations tracing_stats_fops = { | 5315 | static const struct file_operations tracing_stats_fops = { |
5331 | .open = tracing_open_generic, | 5316 | .open = tracing_open_generic_tr, |
5332 | .read = tracing_stats_read, | 5317 | .read = tracing_stats_read, |
5333 | .llseek = generic_file_llseek, | 5318 | .llseek = generic_file_llseek, |
5319 | .release = tracing_release_generic_tr, | ||
5334 | }; | 5320 | }; |
5335 | 5321 | ||
5336 | #ifdef CONFIG_DYNAMIC_FTRACE | 5322 | #ifdef CONFIG_DYNAMIC_FTRACE |
@@ -5519,10 +5505,20 @@ static struct dentry *tracing_dentry_percpu(struct trace_array *tr, int cpu) | |||
5519 | return tr->percpu_dir; | 5505 | return tr->percpu_dir; |
5520 | } | 5506 | } |
5521 | 5507 | ||
5508 | static struct dentry * | ||
5509 | trace_create_cpu_file(const char *name, umode_t mode, struct dentry *parent, | ||
5510 | void *data, long cpu, const struct file_operations *fops) | ||
5511 | { | ||
5512 | struct dentry *ret = trace_create_file(name, mode, parent, data, fops); | ||
5513 | |||
5514 | if (ret) /* See tracing_get_cpu() */ | ||
5515 | ret->d_inode->i_cdev = (void *)(cpu + 1); | ||
5516 | return ret; | ||
5517 | } | ||
5518 | |||
5522 | static void | 5519 | static void |
5523 | tracing_init_debugfs_percpu(struct trace_array *tr, long cpu) | 5520 | tracing_init_debugfs_percpu(struct trace_array *tr, long cpu) |
5524 | { | 5521 | { |
5525 | struct trace_array_cpu *data = per_cpu_ptr(tr->trace_buffer.data, cpu); | ||
5526 | struct dentry *d_percpu = tracing_dentry_percpu(tr, cpu); | 5522 | struct dentry *d_percpu = tracing_dentry_percpu(tr, cpu); |
5527 | struct dentry *d_cpu; | 5523 | struct dentry *d_cpu; |
5528 | char cpu_dir[30]; /* 30 characters should be more than enough */ | 5524 | char cpu_dir[30]; /* 30 characters should be more than enough */ |
@@ -5538,28 +5534,28 @@ tracing_init_debugfs_percpu(struct trace_array *tr, long cpu) | |||
5538 | } | 5534 | } |
5539 | 5535 | ||
5540 | /* per cpu trace_pipe */ | 5536 | /* per cpu trace_pipe */ |
5541 | trace_create_file("trace_pipe", 0444, d_cpu, | 5537 | trace_create_cpu_file("trace_pipe", 0444, d_cpu, |
5542 | (void *)&data->trace_cpu, &tracing_pipe_fops); | 5538 | tr, cpu, &tracing_pipe_fops); |
5543 | 5539 | ||
5544 | /* per cpu trace */ | 5540 | /* per cpu trace */ |
5545 | trace_create_file("trace", 0644, d_cpu, | 5541 | trace_create_cpu_file("trace", 0644, d_cpu, |
5546 | (void *)&data->trace_cpu, &tracing_fops); | 5542 | tr, cpu, &tracing_fops); |
5547 | 5543 | ||
5548 | trace_create_file("trace_pipe_raw", 0444, d_cpu, | 5544 | trace_create_cpu_file("trace_pipe_raw", 0444, d_cpu, |
5549 | (void *)&data->trace_cpu, &tracing_buffers_fops); | 5545 | tr, cpu, &tracing_buffers_fops); |
5550 | 5546 | ||
5551 | trace_create_file("stats", 0444, d_cpu, | 5547 | trace_create_cpu_file("stats", 0444, d_cpu, |
5552 | (void *)&data->trace_cpu, &tracing_stats_fops); | 5548 | tr, cpu, &tracing_stats_fops); |
5553 | 5549 | ||
5554 | trace_create_file("buffer_size_kb", 0444, d_cpu, | 5550 | trace_create_cpu_file("buffer_size_kb", 0444, d_cpu, |
5555 | (void *)&data->trace_cpu, &tracing_entries_fops); | 5551 | tr, cpu, &tracing_entries_fops); |
5556 | 5552 | ||
5557 | #ifdef CONFIG_TRACER_SNAPSHOT | 5553 | #ifdef CONFIG_TRACER_SNAPSHOT |
5558 | trace_create_file("snapshot", 0644, d_cpu, | 5554 | trace_create_cpu_file("snapshot", 0644, d_cpu, |
5559 | (void *)&data->trace_cpu, &snapshot_fops); | 5555 | tr, cpu, &snapshot_fops); |
5560 | 5556 | ||
5561 | trace_create_file("snapshot_raw", 0444, d_cpu, | 5557 | trace_create_cpu_file("snapshot_raw", 0444, d_cpu, |
5562 | (void *)&data->trace_cpu, &snapshot_raw_fops); | 5558 | tr, cpu, &snapshot_raw_fops); |
5563 | #endif | 5559 | #endif |
5564 | } | 5560 | } |
5565 | 5561 | ||
@@ -5868,17 +5864,6 @@ struct dentry *trace_instance_dir; | |||
5868 | static void | 5864 | static void |
5869 | init_tracer_debugfs(struct trace_array *tr, struct dentry *d_tracer); | 5865 | init_tracer_debugfs(struct trace_array *tr, struct dentry *d_tracer); |
5870 | 5866 | ||
5871 | static void init_trace_buffers(struct trace_array *tr, struct trace_buffer *buf) | ||
5872 | { | ||
5873 | int cpu; | ||
5874 | |||
5875 | for_each_tracing_cpu(cpu) { | ||
5876 | memset(per_cpu_ptr(buf->data, cpu), 0, sizeof(struct trace_array_cpu)); | ||
5877 | per_cpu_ptr(buf->data, cpu)->trace_cpu.cpu = cpu; | ||
5878 | per_cpu_ptr(buf->data, cpu)->trace_cpu.tr = tr; | ||
5879 | } | ||
5880 | } | ||
5881 | |||
5882 | static int | 5867 | static int |
5883 | allocate_trace_buffer(struct trace_array *tr, struct trace_buffer *buf, int size) | 5868 | allocate_trace_buffer(struct trace_array *tr, struct trace_buffer *buf, int size) |
5884 | { | 5869 | { |
@@ -5896,8 +5881,6 @@ allocate_trace_buffer(struct trace_array *tr, struct trace_buffer *buf, int size | |||
5896 | return -ENOMEM; | 5881 | return -ENOMEM; |
5897 | } | 5882 | } |
5898 | 5883 | ||
5899 | init_trace_buffers(tr, buf); | ||
5900 | |||
5901 | /* Allocate the first page for all buffers */ | 5884 | /* Allocate the first page for all buffers */ |
5902 | set_buffer_entries(&tr->trace_buffer, | 5885 | set_buffer_entries(&tr->trace_buffer, |
5903 | ring_buffer_size(tr->trace_buffer.buffer, 0)); | 5886 | ring_buffer_size(tr->trace_buffer.buffer, 0)); |
@@ -5964,17 +5947,15 @@ static int new_instance_create(const char *name) | |||
5964 | if (allocate_trace_buffers(tr, trace_buf_size) < 0) | 5947 | if (allocate_trace_buffers(tr, trace_buf_size) < 0) |
5965 | goto out_free_tr; | 5948 | goto out_free_tr; |
5966 | 5949 | ||
5967 | /* Holder for file callbacks */ | ||
5968 | tr->trace_cpu.cpu = RING_BUFFER_ALL_CPUS; | ||
5969 | tr->trace_cpu.tr = tr; | ||
5970 | |||
5971 | tr->dir = debugfs_create_dir(name, trace_instance_dir); | 5950 | tr->dir = debugfs_create_dir(name, trace_instance_dir); |
5972 | if (!tr->dir) | 5951 | if (!tr->dir) |
5973 | goto out_free_tr; | 5952 | goto out_free_tr; |
5974 | 5953 | ||
5975 | ret = event_trace_add_tracer(tr->dir, tr); | 5954 | ret = event_trace_add_tracer(tr->dir, tr); |
5976 | if (ret) | 5955 | if (ret) { |
5956 | debugfs_remove_recursive(tr->dir); | ||
5977 | goto out_free_tr; | 5957 | goto out_free_tr; |
5958 | } | ||
5978 | 5959 | ||
5979 | init_tracer_debugfs(tr, tr->dir); | 5960 | init_tracer_debugfs(tr, tr->dir); |
5980 | 5961 | ||
@@ -6120,13 +6101,13 @@ init_tracer_debugfs(struct trace_array *tr, struct dentry *d_tracer) | |||
6120 | tr, &tracing_iter_fops); | 6101 | tr, &tracing_iter_fops); |
6121 | 6102 | ||
6122 | trace_create_file("trace", 0644, d_tracer, | 6103 | trace_create_file("trace", 0644, d_tracer, |
6123 | (void *)&tr->trace_cpu, &tracing_fops); | 6104 | tr, &tracing_fops); |
6124 | 6105 | ||
6125 | trace_create_file("trace_pipe", 0444, d_tracer, | 6106 | trace_create_file("trace_pipe", 0444, d_tracer, |
6126 | (void *)&tr->trace_cpu, &tracing_pipe_fops); | 6107 | tr, &tracing_pipe_fops); |
6127 | 6108 | ||
6128 | trace_create_file("buffer_size_kb", 0644, d_tracer, | 6109 | trace_create_file("buffer_size_kb", 0644, d_tracer, |
6129 | (void *)&tr->trace_cpu, &tracing_entries_fops); | 6110 | tr, &tracing_entries_fops); |
6130 | 6111 | ||
6131 | trace_create_file("buffer_total_size_kb", 0444, d_tracer, | 6112 | trace_create_file("buffer_total_size_kb", 0444, d_tracer, |
6132 | tr, &tracing_total_entries_fops); | 6113 | tr, &tracing_total_entries_fops); |
@@ -6141,11 +6122,11 @@ init_tracer_debugfs(struct trace_array *tr, struct dentry *d_tracer) | |||
6141 | &trace_clock_fops); | 6122 | &trace_clock_fops); |
6142 | 6123 | ||
6143 | trace_create_file("tracing_on", 0644, d_tracer, | 6124 | trace_create_file("tracing_on", 0644, d_tracer, |
6144 | tr, &rb_simple_fops); | 6125 | tr, &rb_simple_fops); |
6145 | 6126 | ||
6146 | #ifdef CONFIG_TRACER_SNAPSHOT | 6127 | #ifdef CONFIG_TRACER_SNAPSHOT |
6147 | trace_create_file("snapshot", 0644, d_tracer, | 6128 | trace_create_file("snapshot", 0644, d_tracer, |
6148 | (void *)&tr->trace_cpu, &snapshot_fops); | 6129 | tr, &snapshot_fops); |
6149 | #endif | 6130 | #endif |
6150 | 6131 | ||
6151 | for_each_tracing_cpu(cpu) | 6132 | for_each_tracing_cpu(cpu) |
@@ -6439,10 +6420,6 @@ __init static int tracer_alloc_buffers(void) | |||
6439 | 6420 | ||
6440 | global_trace.flags = TRACE_ARRAY_FL_GLOBAL; | 6421 | global_trace.flags = TRACE_ARRAY_FL_GLOBAL; |
6441 | 6422 | ||
6442 | /* Holder for file callbacks */ | ||
6443 | global_trace.trace_cpu.cpu = RING_BUFFER_ALL_CPUS; | ||
6444 | global_trace.trace_cpu.tr = &global_trace; | ||
6445 | |||
6446 | INIT_LIST_HEAD(&global_trace.systems); | 6423 | INIT_LIST_HEAD(&global_trace.systems); |
6447 | INIT_LIST_HEAD(&global_trace.events); | 6424 | INIT_LIST_HEAD(&global_trace.events); |
6448 | list_add(&global_trace.list, &ftrace_trace_arrays); | 6425 | list_add(&global_trace.list, &ftrace_trace_arrays); |
diff --git a/kernel/trace/trace.h b/kernel/trace/trace.h index 4a4f6e1828b6..afaae41b0a02 100644 --- a/kernel/trace/trace.h +++ b/kernel/trace/trace.h | |||
@@ -130,19 +130,12 @@ enum trace_flag_type { | |||
130 | 130 | ||
131 | struct trace_array; | 131 | struct trace_array; |
132 | 132 | ||
133 | struct trace_cpu { | ||
134 | struct trace_array *tr; | ||
135 | struct dentry *dir; | ||
136 | int cpu; | ||
137 | }; | ||
138 | |||
139 | /* | 133 | /* |
140 | * The CPU trace array - it consists of thousands of trace entries | 134 | * The CPU trace array - it consists of thousands of trace entries |
141 | * plus some other descriptor data: (for example which task started | 135 | * plus some other descriptor data: (for example which task started |
142 | * the trace, etc.) | 136 | * the trace, etc.) |
143 | */ | 137 | */ |
144 | struct trace_array_cpu { | 138 | struct trace_array_cpu { |
145 | struct trace_cpu trace_cpu; | ||
146 | atomic_t disabled; | 139 | atomic_t disabled; |
147 | void *buffer_page; /* ring buffer spare */ | 140 | void *buffer_page; /* ring buffer spare */ |
148 | 141 | ||
@@ -196,7 +189,6 @@ struct trace_array { | |||
196 | bool allocated_snapshot; | 189 | bool allocated_snapshot; |
197 | #endif | 190 | #endif |
198 | int buffer_disabled; | 191 | int buffer_disabled; |
199 | struct trace_cpu trace_cpu; /* place holder */ | ||
200 | #ifdef CONFIG_FTRACE_SYSCALLS | 192 | #ifdef CONFIG_FTRACE_SYSCALLS |
201 | int sys_refcount_enter; | 193 | int sys_refcount_enter; |
202 | int sys_refcount_exit; | 194 | int sys_refcount_exit; |
@@ -214,7 +206,6 @@ struct trace_array { | |||
214 | struct dentry *event_dir; | 206 | struct dentry *event_dir; |
215 | struct list_head systems; | 207 | struct list_head systems; |
216 | struct list_head events; | 208 | struct list_head events; |
217 | struct task_struct *waiter; | ||
218 | int ref; | 209 | int ref; |
219 | }; | 210 | }; |
220 | 211 | ||
@@ -680,6 +671,15 @@ extern int trace_selftest_startup_sched_switch(struct tracer *trace, | |||
680 | struct trace_array *tr); | 671 | struct trace_array *tr); |
681 | extern int trace_selftest_startup_branch(struct tracer *trace, | 672 | extern int trace_selftest_startup_branch(struct tracer *trace, |
682 | struct trace_array *tr); | 673 | struct trace_array *tr); |
674 | /* | ||
675 | * Tracer data references selftest functions that only occur | ||
676 | * on boot up. These can be __init functions. Thus, when selftests | ||
677 | * are enabled, then the tracers need to reference __init functions. | ||
678 | */ | ||
679 | #define __tracer_data __refdata | ||
680 | #else | ||
681 | /* Tracers are seldom changed. Optimize when selftests are disabled. */ | ||
682 | #define __tracer_data __read_mostly | ||
683 | #endif /* CONFIG_FTRACE_STARTUP_TEST */ | 683 | #endif /* CONFIG_FTRACE_STARTUP_TEST */ |
684 | 684 | ||
685 | extern void *head_page(struct trace_array_cpu *data); | 685 | extern void *head_page(struct trace_array_cpu *data); |
diff --git a/kernel/trace/trace_event_perf.c b/kernel/trace/trace_event_perf.c index 84b1e045faba..80c36bcf66e8 100644 --- a/kernel/trace/trace_event_perf.c +++ b/kernel/trace/trace_event_perf.c | |||
@@ -236,6 +236,10 @@ __kprobes void *perf_trace_buf_prepare(int size, unsigned short type, | |||
236 | 236 | ||
237 | BUILD_BUG_ON(PERF_MAX_TRACE_SIZE % sizeof(unsigned long)); | 237 | BUILD_BUG_ON(PERF_MAX_TRACE_SIZE % sizeof(unsigned long)); |
238 | 238 | ||
239 | if (WARN_ONCE(size > PERF_MAX_TRACE_SIZE, | ||
240 | "perf buffer not large enough")) | ||
241 | return NULL; | ||
242 | |||
239 | pc = preempt_count(); | 243 | pc = preempt_count(); |
240 | 244 | ||
241 | *rctxp = perf_swevent_get_recursion_context(); | 245 | *rctxp = perf_swevent_get_recursion_context(); |
@@ -266,6 +270,10 @@ perf_ftrace_function_call(unsigned long ip, unsigned long parent_ip, | |||
266 | struct pt_regs regs; | 270 | struct pt_regs regs; |
267 | int rctx; | 271 | int rctx; |
268 | 272 | ||
273 | head = this_cpu_ptr(event_function.perf_events); | ||
274 | if (hlist_empty(head)) | ||
275 | return; | ||
276 | |||
269 | #define ENTRY_SIZE (ALIGN(sizeof(struct ftrace_entry) + sizeof(u32), \ | 277 | #define ENTRY_SIZE (ALIGN(sizeof(struct ftrace_entry) + sizeof(u32), \ |
270 | sizeof(u64)) - sizeof(u32)) | 278 | sizeof(u64)) - sizeof(u32)) |
271 | 279 | ||
@@ -279,8 +287,6 @@ perf_ftrace_function_call(unsigned long ip, unsigned long parent_ip, | |||
279 | 287 | ||
280 | entry->ip = ip; | 288 | entry->ip = ip; |
281 | entry->parent_ip = parent_ip; | 289 | entry->parent_ip = parent_ip; |
282 | |||
283 | head = this_cpu_ptr(event_function.perf_events); | ||
284 | perf_trace_buf_submit(entry, ENTRY_SIZE, rctx, 0, | 290 | perf_trace_buf_submit(entry, ENTRY_SIZE, rctx, 0, |
285 | 1, ®s, head, NULL); | 291 | 1, ®s, head, NULL); |
286 | 292 | ||
diff --git a/kernel/trace/trace_events.c b/kernel/trace/trace_events.c index 7d854290bf81..898f868833f2 100644 --- a/kernel/trace/trace_events.c +++ b/kernel/trace/trace_events.c | |||
@@ -826,59 +826,33 @@ enum { | |||
826 | static void *f_next(struct seq_file *m, void *v, loff_t *pos) | 826 | static void *f_next(struct seq_file *m, void *v, loff_t *pos) |
827 | { | 827 | { |
828 | struct ftrace_event_call *call = m->private; | 828 | struct ftrace_event_call *call = m->private; |
829 | struct ftrace_event_field *field; | ||
830 | struct list_head *common_head = &ftrace_common_fields; | 829 | struct list_head *common_head = &ftrace_common_fields; |
831 | struct list_head *head = trace_get_fields(call); | 830 | struct list_head *head = trace_get_fields(call); |
831 | struct list_head *node = v; | ||
832 | 832 | ||
833 | (*pos)++; | 833 | (*pos)++; |
834 | 834 | ||
835 | switch ((unsigned long)v) { | 835 | switch ((unsigned long)v) { |
836 | case FORMAT_HEADER: | 836 | case FORMAT_HEADER: |
837 | if (unlikely(list_empty(common_head))) | 837 | node = common_head; |
838 | return NULL; | 838 | break; |
839 | |||
840 | field = list_entry(common_head->prev, | ||
841 | struct ftrace_event_field, link); | ||
842 | return field; | ||
843 | 839 | ||
844 | case FORMAT_FIELD_SEPERATOR: | 840 | case FORMAT_FIELD_SEPERATOR: |
845 | if (unlikely(list_empty(head))) | 841 | node = head; |
846 | return NULL; | 842 | break; |
847 | |||
848 | field = list_entry(head->prev, struct ftrace_event_field, link); | ||
849 | return field; | ||
850 | 843 | ||
851 | case FORMAT_PRINTFMT: | 844 | case FORMAT_PRINTFMT: |
852 | /* all done */ | 845 | /* all done */ |
853 | return NULL; | 846 | return NULL; |
854 | } | 847 | } |
855 | 848 | ||
856 | field = v; | 849 | node = node->prev; |
857 | if (field->link.prev == common_head) | 850 | if (node == common_head) |
858 | return (void *)FORMAT_FIELD_SEPERATOR; | 851 | return (void *)FORMAT_FIELD_SEPERATOR; |
859 | else if (field->link.prev == head) | 852 | else if (node == head) |
860 | return (void *)FORMAT_PRINTFMT; | 853 | return (void *)FORMAT_PRINTFMT; |
861 | 854 | else | |
862 | field = list_entry(field->link.prev, struct ftrace_event_field, link); | 855 | return node; |
863 | |||
864 | return field; | ||
865 | } | ||
866 | |||
867 | static void *f_start(struct seq_file *m, loff_t *pos) | ||
868 | { | ||
869 | loff_t l = 0; | ||
870 | void *p; | ||
871 | |||
872 | /* Start by showing the header */ | ||
873 | if (!*pos) | ||
874 | return (void *)FORMAT_HEADER; | ||
875 | |||
876 | p = (void *)FORMAT_HEADER; | ||
877 | do { | ||
878 | p = f_next(m, p, &l); | ||
879 | } while (p && l < *pos); | ||
880 | |||
881 | return p; | ||
882 | } | 856 | } |
883 | 857 | ||
884 | static int f_show(struct seq_file *m, void *v) | 858 | static int f_show(struct seq_file *m, void *v) |
@@ -904,8 +878,7 @@ static int f_show(struct seq_file *m, void *v) | |||
904 | return 0; | 878 | return 0; |
905 | } | 879 | } |
906 | 880 | ||
907 | field = v; | 881 | field = list_entry(v, struct ftrace_event_field, link); |
908 | |||
909 | /* | 882 | /* |
910 | * Smartly shows the array type(except dynamic array). | 883 | * Smartly shows the array type(except dynamic array). |
911 | * Normal: | 884 | * Normal: |
@@ -932,6 +905,17 @@ static int f_show(struct seq_file *m, void *v) | |||
932 | return 0; | 905 | return 0; |
933 | } | 906 | } |
934 | 907 | ||
908 | static void *f_start(struct seq_file *m, loff_t *pos) | ||
909 | { | ||
910 | void *p = (void *)FORMAT_HEADER; | ||
911 | loff_t l = 0; | ||
912 | |||
913 | while (l < *pos && p) | ||
914 | p = f_next(m, p, &l); | ||
915 | |||
916 | return p; | ||
917 | } | ||
918 | |||
935 | static void f_stop(struct seq_file *m, void *p) | 919 | static void f_stop(struct seq_file *m, void *p) |
936 | { | 920 | { |
937 | } | 921 | } |
@@ -963,23 +947,14 @@ static ssize_t | |||
963 | event_id_read(struct file *filp, char __user *ubuf, size_t cnt, loff_t *ppos) | 947 | event_id_read(struct file *filp, char __user *ubuf, size_t cnt, loff_t *ppos) |
964 | { | 948 | { |
965 | struct ftrace_event_call *call = filp->private_data; | 949 | struct ftrace_event_call *call = filp->private_data; |
966 | struct trace_seq *s; | 950 | char buf[32]; |
967 | int r; | 951 | int len; |
968 | 952 | ||
969 | if (*ppos) | 953 | if (*ppos) |
970 | return 0; | 954 | return 0; |
971 | 955 | ||
972 | s = kmalloc(sizeof(*s), GFP_KERNEL); | 956 | len = sprintf(buf, "%d\n", call->event.type); |
973 | if (!s) | 957 | return simple_read_from_buffer(ubuf, cnt, ppos, buf, len); |
974 | return -ENOMEM; | ||
975 | |||
976 | trace_seq_init(s); | ||
977 | trace_seq_printf(s, "%d\n", call->event.type); | ||
978 | |||
979 | r = simple_read_from_buffer(ubuf, cnt, ppos, | ||
980 | s->buffer, s->len); | ||
981 | kfree(s); | ||
982 | return r; | ||
983 | } | 958 | } |
984 | 959 | ||
985 | static ssize_t | 960 | static ssize_t |
@@ -1218,6 +1193,7 @@ show_header(struct file *filp, char __user *ubuf, size_t cnt, loff_t *ppos) | |||
1218 | 1193 | ||
1219 | static int ftrace_event_avail_open(struct inode *inode, struct file *file); | 1194 | static int ftrace_event_avail_open(struct inode *inode, struct file *file); |
1220 | static int ftrace_event_set_open(struct inode *inode, struct file *file); | 1195 | static int ftrace_event_set_open(struct inode *inode, struct file *file); |
1196 | static int ftrace_event_release(struct inode *inode, struct file *file); | ||
1221 | 1197 | ||
1222 | static const struct seq_operations show_event_seq_ops = { | 1198 | static const struct seq_operations show_event_seq_ops = { |
1223 | .start = t_start, | 1199 | .start = t_start, |
@@ -1245,7 +1221,7 @@ static const struct file_operations ftrace_set_event_fops = { | |||
1245 | .read = seq_read, | 1221 | .read = seq_read, |
1246 | .write = ftrace_event_write, | 1222 | .write = ftrace_event_write, |
1247 | .llseek = seq_lseek, | 1223 | .llseek = seq_lseek, |
1248 | .release = seq_release, | 1224 | .release = ftrace_event_release, |
1249 | }; | 1225 | }; |
1250 | 1226 | ||
1251 | static const struct file_operations ftrace_enable_fops = { | 1227 | static const struct file_operations ftrace_enable_fops = { |
@@ -1323,6 +1299,15 @@ ftrace_event_open(struct inode *inode, struct file *file, | |||
1323 | return ret; | 1299 | return ret; |
1324 | } | 1300 | } |
1325 | 1301 | ||
1302 | static int ftrace_event_release(struct inode *inode, struct file *file) | ||
1303 | { | ||
1304 | struct trace_array *tr = inode->i_private; | ||
1305 | |||
1306 | trace_array_put(tr); | ||
1307 | |||
1308 | return seq_release(inode, file); | ||
1309 | } | ||
1310 | |||
1326 | static int | 1311 | static int |
1327 | ftrace_event_avail_open(struct inode *inode, struct file *file) | 1312 | ftrace_event_avail_open(struct inode *inode, struct file *file) |
1328 | { | 1313 | { |
@@ -1336,12 +1321,19 @@ ftrace_event_set_open(struct inode *inode, struct file *file) | |||
1336 | { | 1321 | { |
1337 | const struct seq_operations *seq_ops = &show_set_event_seq_ops; | 1322 | const struct seq_operations *seq_ops = &show_set_event_seq_ops; |
1338 | struct trace_array *tr = inode->i_private; | 1323 | struct trace_array *tr = inode->i_private; |
1324 | int ret; | ||
1325 | |||
1326 | if (trace_array_get(tr) < 0) | ||
1327 | return -ENODEV; | ||
1339 | 1328 | ||
1340 | if ((file->f_mode & FMODE_WRITE) && | 1329 | if ((file->f_mode & FMODE_WRITE) && |
1341 | (file->f_flags & O_TRUNC)) | 1330 | (file->f_flags & O_TRUNC)) |
1342 | ftrace_clear_events(tr); | 1331 | ftrace_clear_events(tr); |
1343 | 1332 | ||
1344 | return ftrace_event_open(inode, file, seq_ops); | 1333 | ret = ftrace_event_open(inode, file, seq_ops); |
1334 | if (ret < 0) | ||
1335 | trace_array_put(tr); | ||
1336 | return ret; | ||
1345 | } | 1337 | } |
1346 | 1338 | ||
1347 | static struct event_subsystem * | 1339 | static struct event_subsystem * |
diff --git a/kernel/trace/trace_events_filter.c b/kernel/trace/trace_events_filter.c index 0d883dc057d6..0c7b75a8acc8 100644 --- a/kernel/trace/trace_events_filter.c +++ b/kernel/trace/trace_events_filter.c | |||
@@ -646,7 +646,7 @@ void print_event_filter(struct ftrace_event_call *call, struct trace_seq *s) | |||
646 | if (filter && filter->filter_string) | 646 | if (filter && filter->filter_string) |
647 | trace_seq_printf(s, "%s\n", filter->filter_string); | 647 | trace_seq_printf(s, "%s\n", filter->filter_string); |
648 | else | 648 | else |
649 | trace_seq_printf(s, "none\n"); | 649 | trace_seq_puts(s, "none\n"); |
650 | mutex_unlock(&event_mutex); | 650 | mutex_unlock(&event_mutex); |
651 | } | 651 | } |
652 | 652 | ||
@@ -660,7 +660,7 @@ void print_subsystem_event_filter(struct event_subsystem *system, | |||
660 | if (filter && filter->filter_string) | 660 | if (filter && filter->filter_string) |
661 | trace_seq_printf(s, "%s\n", filter->filter_string); | 661 | trace_seq_printf(s, "%s\n", filter->filter_string); |
662 | else | 662 | else |
663 | trace_seq_printf(s, DEFAULT_SYS_FILTER_MESSAGE "\n"); | 663 | trace_seq_puts(s, DEFAULT_SYS_FILTER_MESSAGE "\n"); |
664 | mutex_unlock(&event_mutex); | 664 | mutex_unlock(&event_mutex); |
665 | } | 665 | } |
666 | 666 | ||
diff --git a/kernel/trace/trace_functions.c b/kernel/trace/trace_functions.c index b863f93b30f3..38fe1483c508 100644 --- a/kernel/trace/trace_functions.c +++ b/kernel/trace/trace_functions.c | |||
@@ -199,7 +199,7 @@ static int func_set_flag(u32 old_flags, u32 bit, int set) | |||
199 | return 0; | 199 | return 0; |
200 | } | 200 | } |
201 | 201 | ||
202 | static struct tracer function_trace __read_mostly = | 202 | static struct tracer function_trace __tracer_data = |
203 | { | 203 | { |
204 | .name = "function", | 204 | .name = "function", |
205 | .init = function_trace_init, | 205 | .init = function_trace_init, |
diff --git a/kernel/trace/trace_functions_graph.c b/kernel/trace/trace_functions_graph.c index 8388bc99f2ee..b5c09242683d 100644 --- a/kernel/trace/trace_functions_graph.c +++ b/kernel/trace/trace_functions_graph.c | |||
@@ -446,7 +446,7 @@ print_graph_proc(struct trace_seq *s, pid_t pid) | |||
446 | 446 | ||
447 | /* First spaces to align center */ | 447 | /* First spaces to align center */ |
448 | for (i = 0; i < spaces / 2; i++) { | 448 | for (i = 0; i < spaces / 2; i++) { |
449 | ret = trace_seq_printf(s, " "); | 449 | ret = trace_seq_putc(s, ' '); |
450 | if (!ret) | 450 | if (!ret) |
451 | return TRACE_TYPE_PARTIAL_LINE; | 451 | return TRACE_TYPE_PARTIAL_LINE; |
452 | } | 452 | } |
@@ -457,7 +457,7 @@ print_graph_proc(struct trace_seq *s, pid_t pid) | |||
457 | 457 | ||
458 | /* Last spaces to align center */ | 458 | /* Last spaces to align center */ |
459 | for (i = 0; i < spaces - (spaces / 2); i++) { | 459 | for (i = 0; i < spaces - (spaces / 2); i++) { |
460 | ret = trace_seq_printf(s, " "); | 460 | ret = trace_seq_putc(s, ' '); |
461 | if (!ret) | 461 | if (!ret) |
462 | return TRACE_TYPE_PARTIAL_LINE; | 462 | return TRACE_TYPE_PARTIAL_LINE; |
463 | } | 463 | } |
@@ -503,7 +503,7 @@ verif_pid(struct trace_seq *s, pid_t pid, int cpu, struct fgraph_data *data) | |||
503 | ------------------------------------------ | 503 | ------------------------------------------ |
504 | 504 | ||
505 | */ | 505 | */ |
506 | ret = trace_seq_printf(s, | 506 | ret = trace_seq_puts(s, |
507 | " ------------------------------------------\n"); | 507 | " ------------------------------------------\n"); |
508 | if (!ret) | 508 | if (!ret) |
509 | return TRACE_TYPE_PARTIAL_LINE; | 509 | return TRACE_TYPE_PARTIAL_LINE; |
@@ -516,7 +516,7 @@ verif_pid(struct trace_seq *s, pid_t pid, int cpu, struct fgraph_data *data) | |||
516 | if (ret == TRACE_TYPE_PARTIAL_LINE) | 516 | if (ret == TRACE_TYPE_PARTIAL_LINE) |
517 | return TRACE_TYPE_PARTIAL_LINE; | 517 | return TRACE_TYPE_PARTIAL_LINE; |
518 | 518 | ||
519 | ret = trace_seq_printf(s, " => "); | 519 | ret = trace_seq_puts(s, " => "); |
520 | if (!ret) | 520 | if (!ret) |
521 | return TRACE_TYPE_PARTIAL_LINE; | 521 | return TRACE_TYPE_PARTIAL_LINE; |
522 | 522 | ||
@@ -524,7 +524,7 @@ verif_pid(struct trace_seq *s, pid_t pid, int cpu, struct fgraph_data *data) | |||
524 | if (ret == TRACE_TYPE_PARTIAL_LINE) | 524 | if (ret == TRACE_TYPE_PARTIAL_LINE) |
525 | return TRACE_TYPE_PARTIAL_LINE; | 525 | return TRACE_TYPE_PARTIAL_LINE; |
526 | 526 | ||
527 | ret = trace_seq_printf(s, | 527 | ret = trace_seq_puts(s, |
528 | "\n ------------------------------------------\n\n"); | 528 | "\n ------------------------------------------\n\n"); |
529 | if (!ret) | 529 | if (!ret) |
530 | return TRACE_TYPE_PARTIAL_LINE; | 530 | return TRACE_TYPE_PARTIAL_LINE; |
@@ -645,7 +645,7 @@ print_graph_irq(struct trace_iterator *iter, unsigned long addr, | |||
645 | ret = print_graph_proc(s, pid); | 645 | ret = print_graph_proc(s, pid); |
646 | if (ret == TRACE_TYPE_PARTIAL_LINE) | 646 | if (ret == TRACE_TYPE_PARTIAL_LINE) |
647 | return TRACE_TYPE_PARTIAL_LINE; | 647 | return TRACE_TYPE_PARTIAL_LINE; |
648 | ret = trace_seq_printf(s, " | "); | 648 | ret = trace_seq_puts(s, " | "); |
649 | if (!ret) | 649 | if (!ret) |
650 | return TRACE_TYPE_PARTIAL_LINE; | 650 | return TRACE_TYPE_PARTIAL_LINE; |
651 | } | 651 | } |
@@ -657,9 +657,9 @@ print_graph_irq(struct trace_iterator *iter, unsigned long addr, | |||
657 | return ret; | 657 | return ret; |
658 | 658 | ||
659 | if (type == TRACE_GRAPH_ENT) | 659 | if (type == TRACE_GRAPH_ENT) |
660 | ret = trace_seq_printf(s, "==========>"); | 660 | ret = trace_seq_puts(s, "==========>"); |
661 | else | 661 | else |
662 | ret = trace_seq_printf(s, "<=========="); | 662 | ret = trace_seq_puts(s, "<=========="); |
663 | 663 | ||
664 | if (!ret) | 664 | if (!ret) |
665 | return TRACE_TYPE_PARTIAL_LINE; | 665 | return TRACE_TYPE_PARTIAL_LINE; |
@@ -668,7 +668,7 @@ print_graph_irq(struct trace_iterator *iter, unsigned long addr, | |||
668 | if (ret != TRACE_TYPE_HANDLED) | 668 | if (ret != TRACE_TYPE_HANDLED) |
669 | return ret; | 669 | return ret; |
670 | 670 | ||
671 | ret = trace_seq_printf(s, "\n"); | 671 | ret = trace_seq_putc(s, '\n'); |
672 | 672 | ||
673 | if (!ret) | 673 | if (!ret) |
674 | return TRACE_TYPE_PARTIAL_LINE; | 674 | return TRACE_TYPE_PARTIAL_LINE; |
@@ -705,13 +705,13 @@ trace_print_graph_duration(unsigned long long duration, struct trace_seq *s) | |||
705 | len += strlen(nsecs_str); | 705 | len += strlen(nsecs_str); |
706 | } | 706 | } |
707 | 707 | ||
708 | ret = trace_seq_printf(s, " us "); | 708 | ret = trace_seq_puts(s, " us "); |
709 | if (!ret) | 709 | if (!ret) |
710 | return TRACE_TYPE_PARTIAL_LINE; | 710 | return TRACE_TYPE_PARTIAL_LINE; |
711 | 711 | ||
712 | /* Print remaining spaces to fit the row's width */ | 712 | /* Print remaining spaces to fit the row's width */ |
713 | for (i = len; i < 7; i++) { | 713 | for (i = len; i < 7; i++) { |
714 | ret = trace_seq_printf(s, " "); | 714 | ret = trace_seq_putc(s, ' '); |
715 | if (!ret) | 715 | if (!ret) |
716 | return TRACE_TYPE_PARTIAL_LINE; | 716 | return TRACE_TYPE_PARTIAL_LINE; |
717 | } | 717 | } |
@@ -731,13 +731,13 @@ print_graph_duration(unsigned long long duration, struct trace_seq *s, | |||
731 | /* No real adata, just filling the column with spaces */ | 731 | /* No real adata, just filling the column with spaces */ |
732 | switch (duration) { | 732 | switch (duration) { |
733 | case DURATION_FILL_FULL: | 733 | case DURATION_FILL_FULL: |
734 | ret = trace_seq_printf(s, " | "); | 734 | ret = trace_seq_puts(s, " | "); |
735 | return ret ? TRACE_TYPE_HANDLED : TRACE_TYPE_PARTIAL_LINE; | 735 | return ret ? TRACE_TYPE_HANDLED : TRACE_TYPE_PARTIAL_LINE; |
736 | case DURATION_FILL_START: | 736 | case DURATION_FILL_START: |
737 | ret = trace_seq_printf(s, " "); | 737 | ret = trace_seq_puts(s, " "); |
738 | return ret ? TRACE_TYPE_HANDLED : TRACE_TYPE_PARTIAL_LINE; | 738 | return ret ? TRACE_TYPE_HANDLED : TRACE_TYPE_PARTIAL_LINE; |
739 | case DURATION_FILL_END: | 739 | case DURATION_FILL_END: |
740 | ret = trace_seq_printf(s, " |"); | 740 | ret = trace_seq_puts(s, " |"); |
741 | return ret ? TRACE_TYPE_HANDLED : TRACE_TYPE_PARTIAL_LINE; | 741 | return ret ? TRACE_TYPE_HANDLED : TRACE_TYPE_PARTIAL_LINE; |
742 | } | 742 | } |
743 | 743 | ||
@@ -745,10 +745,10 @@ print_graph_duration(unsigned long long duration, struct trace_seq *s, | |||
745 | if (flags & TRACE_GRAPH_PRINT_OVERHEAD) { | 745 | if (flags & TRACE_GRAPH_PRINT_OVERHEAD) { |
746 | /* Duration exceeded 100 msecs */ | 746 | /* Duration exceeded 100 msecs */ |
747 | if (duration > 100000ULL) | 747 | if (duration > 100000ULL) |
748 | ret = trace_seq_printf(s, "! "); | 748 | ret = trace_seq_puts(s, "! "); |
749 | /* Duration exceeded 10 msecs */ | 749 | /* Duration exceeded 10 msecs */ |
750 | else if (duration > 10000ULL) | 750 | else if (duration > 10000ULL) |
751 | ret = trace_seq_printf(s, "+ "); | 751 | ret = trace_seq_puts(s, "+ "); |
752 | } | 752 | } |
753 | 753 | ||
754 | /* | 754 | /* |
@@ -757,7 +757,7 @@ print_graph_duration(unsigned long long duration, struct trace_seq *s, | |||
757 | * to fill out the space. | 757 | * to fill out the space. |
758 | */ | 758 | */ |
759 | if (ret == -1) | 759 | if (ret == -1) |
760 | ret = trace_seq_printf(s, " "); | 760 | ret = trace_seq_puts(s, " "); |
761 | 761 | ||
762 | /* Catching here any failure happenned above */ | 762 | /* Catching here any failure happenned above */ |
763 | if (!ret) | 763 | if (!ret) |
@@ -767,7 +767,7 @@ print_graph_duration(unsigned long long duration, struct trace_seq *s, | |||
767 | if (ret != TRACE_TYPE_HANDLED) | 767 | if (ret != TRACE_TYPE_HANDLED) |
768 | return ret; | 768 | return ret; |
769 | 769 | ||
770 | ret = trace_seq_printf(s, "| "); | 770 | ret = trace_seq_puts(s, "| "); |
771 | if (!ret) | 771 | if (!ret) |
772 | return TRACE_TYPE_PARTIAL_LINE; | 772 | return TRACE_TYPE_PARTIAL_LINE; |
773 | 773 | ||
@@ -817,7 +817,7 @@ print_graph_entry_leaf(struct trace_iterator *iter, | |||
817 | 817 | ||
818 | /* Function */ | 818 | /* Function */ |
819 | for (i = 0; i < call->depth * TRACE_GRAPH_INDENT; i++) { | 819 | for (i = 0; i < call->depth * TRACE_GRAPH_INDENT; i++) { |
820 | ret = trace_seq_printf(s, " "); | 820 | ret = trace_seq_putc(s, ' '); |
821 | if (!ret) | 821 | if (!ret) |
822 | return TRACE_TYPE_PARTIAL_LINE; | 822 | return TRACE_TYPE_PARTIAL_LINE; |
823 | } | 823 | } |
@@ -858,7 +858,7 @@ print_graph_entry_nested(struct trace_iterator *iter, | |||
858 | 858 | ||
859 | /* Function */ | 859 | /* Function */ |
860 | for (i = 0; i < call->depth * TRACE_GRAPH_INDENT; i++) { | 860 | for (i = 0; i < call->depth * TRACE_GRAPH_INDENT; i++) { |
861 | ret = trace_seq_printf(s, " "); | 861 | ret = trace_seq_putc(s, ' '); |
862 | if (!ret) | 862 | if (!ret) |
863 | return TRACE_TYPE_PARTIAL_LINE; | 863 | return TRACE_TYPE_PARTIAL_LINE; |
864 | } | 864 | } |
@@ -917,7 +917,7 @@ print_graph_prologue(struct trace_iterator *iter, struct trace_seq *s, | |||
917 | if (ret == TRACE_TYPE_PARTIAL_LINE) | 917 | if (ret == TRACE_TYPE_PARTIAL_LINE) |
918 | return TRACE_TYPE_PARTIAL_LINE; | 918 | return TRACE_TYPE_PARTIAL_LINE; |
919 | 919 | ||
920 | ret = trace_seq_printf(s, " | "); | 920 | ret = trace_seq_puts(s, " | "); |
921 | if (!ret) | 921 | if (!ret) |
922 | return TRACE_TYPE_PARTIAL_LINE; | 922 | return TRACE_TYPE_PARTIAL_LINE; |
923 | } | 923 | } |
@@ -1117,7 +1117,7 @@ print_graph_return(struct ftrace_graph_ret *trace, struct trace_seq *s, | |||
1117 | 1117 | ||
1118 | /* Closing brace */ | 1118 | /* Closing brace */ |
1119 | for (i = 0; i < trace->depth * TRACE_GRAPH_INDENT; i++) { | 1119 | for (i = 0; i < trace->depth * TRACE_GRAPH_INDENT; i++) { |
1120 | ret = trace_seq_printf(s, " "); | 1120 | ret = trace_seq_putc(s, ' '); |
1121 | if (!ret) | 1121 | if (!ret) |
1122 | return TRACE_TYPE_PARTIAL_LINE; | 1122 | return TRACE_TYPE_PARTIAL_LINE; |
1123 | } | 1123 | } |
@@ -1129,7 +1129,7 @@ print_graph_return(struct ftrace_graph_ret *trace, struct trace_seq *s, | |||
1129 | * belongs to, write out the function name. | 1129 | * belongs to, write out the function name. |
1130 | */ | 1130 | */ |
1131 | if (func_match) { | 1131 | if (func_match) { |
1132 | ret = trace_seq_printf(s, "}\n"); | 1132 | ret = trace_seq_puts(s, "}\n"); |
1133 | if (!ret) | 1133 | if (!ret) |
1134 | return TRACE_TYPE_PARTIAL_LINE; | 1134 | return TRACE_TYPE_PARTIAL_LINE; |
1135 | } else { | 1135 | } else { |
@@ -1179,13 +1179,13 @@ print_graph_comment(struct trace_seq *s, struct trace_entry *ent, | |||
1179 | /* Indentation */ | 1179 | /* Indentation */ |
1180 | if (depth > 0) | 1180 | if (depth > 0) |
1181 | for (i = 0; i < (depth + 1) * TRACE_GRAPH_INDENT; i++) { | 1181 | for (i = 0; i < (depth + 1) * TRACE_GRAPH_INDENT; i++) { |
1182 | ret = trace_seq_printf(s, " "); | 1182 | ret = trace_seq_putc(s, ' '); |
1183 | if (!ret) | 1183 | if (!ret) |
1184 | return TRACE_TYPE_PARTIAL_LINE; | 1184 | return TRACE_TYPE_PARTIAL_LINE; |
1185 | } | 1185 | } |
1186 | 1186 | ||
1187 | /* The comment */ | 1187 | /* The comment */ |
1188 | ret = trace_seq_printf(s, "/* "); | 1188 | ret = trace_seq_puts(s, "/* "); |
1189 | if (!ret) | 1189 | if (!ret) |
1190 | return TRACE_TYPE_PARTIAL_LINE; | 1190 | return TRACE_TYPE_PARTIAL_LINE; |
1191 | 1191 | ||
@@ -1216,7 +1216,7 @@ print_graph_comment(struct trace_seq *s, struct trace_entry *ent, | |||
1216 | s->len--; | 1216 | s->len--; |
1217 | } | 1217 | } |
1218 | 1218 | ||
1219 | ret = trace_seq_printf(s, " */\n"); | 1219 | ret = trace_seq_puts(s, " */\n"); |
1220 | if (!ret) | 1220 | if (!ret) |
1221 | return TRACE_TYPE_PARTIAL_LINE; | 1221 | return TRACE_TYPE_PARTIAL_LINE; |
1222 | 1222 | ||
@@ -1448,7 +1448,7 @@ static struct trace_event graph_trace_ret_event = { | |||
1448 | .funcs = &graph_functions | 1448 | .funcs = &graph_functions |
1449 | }; | 1449 | }; |
1450 | 1450 | ||
1451 | static struct tracer graph_trace __read_mostly = { | 1451 | static struct tracer graph_trace __tracer_data = { |
1452 | .name = "function_graph", | 1452 | .name = "function_graph", |
1453 | .open = graph_trace_open, | 1453 | .open = graph_trace_open, |
1454 | .pipe_open = graph_trace_open, | 1454 | .pipe_open = graph_trace_open, |
diff --git a/kernel/trace/trace_kprobe.c b/kernel/trace/trace_kprobe.c index 7ed6976493c8..3811487e7a7a 100644 --- a/kernel/trace/trace_kprobe.c +++ b/kernel/trace/trace_kprobe.c | |||
@@ -243,11 +243,11 @@ find_event_file_link(struct trace_probe *tp, struct ftrace_event_file *file) | |||
243 | static int | 243 | static int |
244 | disable_trace_probe(struct trace_probe *tp, struct ftrace_event_file *file) | 244 | disable_trace_probe(struct trace_probe *tp, struct ftrace_event_file *file) |
245 | { | 245 | { |
246 | struct event_file_link *link = NULL; | ||
247 | int wait = 0; | ||
246 | int ret = 0; | 248 | int ret = 0; |
247 | 249 | ||
248 | if (file) { | 250 | if (file) { |
249 | struct event_file_link *link; | ||
250 | |||
251 | link = find_event_file_link(tp, file); | 251 | link = find_event_file_link(tp, file); |
252 | if (!link) { | 252 | if (!link) { |
253 | ret = -EINVAL; | 253 | ret = -EINVAL; |
@@ -255,10 +255,7 @@ disable_trace_probe(struct trace_probe *tp, struct ftrace_event_file *file) | |||
255 | } | 255 | } |
256 | 256 | ||
257 | list_del_rcu(&link->list); | 257 | list_del_rcu(&link->list); |
258 | /* synchronize with kprobe_trace_func/kretprobe_trace_func */ | 258 | wait = 1; |
259 | synchronize_sched(); | ||
260 | kfree(link); | ||
261 | |||
262 | if (!list_empty(&tp->files)) | 259 | if (!list_empty(&tp->files)) |
263 | goto out; | 260 | goto out; |
264 | 261 | ||
@@ -271,8 +268,22 @@ disable_trace_probe(struct trace_probe *tp, struct ftrace_event_file *file) | |||
271 | disable_kretprobe(&tp->rp); | 268 | disable_kretprobe(&tp->rp); |
272 | else | 269 | else |
273 | disable_kprobe(&tp->rp.kp); | 270 | disable_kprobe(&tp->rp.kp); |
271 | wait = 1; | ||
274 | } | 272 | } |
275 | out: | 273 | out: |
274 | if (wait) { | ||
275 | /* | ||
276 | * Synchronize with kprobe_trace_func/kretprobe_trace_func | ||
277 | * to ensure disabled (all running handlers are finished). | ||
278 | * This is not only for kfree(), but also the caller, | ||
279 | * trace_remove_event_call() supposes it for releasing | ||
280 | * event_call related objects, which will be accessed in | ||
281 | * the kprobe_trace_func/kretprobe_trace_func. | ||
282 | */ | ||
283 | synchronize_sched(); | ||
284 | kfree(link); /* Ignored if link == NULL */ | ||
285 | } | ||
286 | |||
276 | return ret; | 287 | return ret; |
277 | } | 288 | } |
278 | 289 | ||
@@ -1087,9 +1098,6 @@ kprobe_perf_func(struct trace_probe *tp, struct pt_regs *regs) | |||
1087 | __size = sizeof(*entry) + tp->size + dsize; | 1098 | __size = sizeof(*entry) + tp->size + dsize; |
1088 | size = ALIGN(__size + sizeof(u32), sizeof(u64)); | 1099 | size = ALIGN(__size + sizeof(u32), sizeof(u64)); |
1089 | size -= sizeof(u32); | 1100 | size -= sizeof(u32); |
1090 | if (WARN_ONCE(size > PERF_MAX_TRACE_SIZE, | ||
1091 | "profile buffer not large enough")) | ||
1092 | return; | ||
1093 | 1101 | ||
1094 | entry = perf_trace_buf_prepare(size, call->event.type, regs, &rctx); | 1102 | entry = perf_trace_buf_prepare(size, call->event.type, regs, &rctx); |
1095 | if (!entry) | 1103 | if (!entry) |
@@ -1120,9 +1128,6 @@ kretprobe_perf_func(struct trace_probe *tp, struct kretprobe_instance *ri, | |||
1120 | __size = sizeof(*entry) + tp->size + dsize; | 1128 | __size = sizeof(*entry) + tp->size + dsize; |
1121 | size = ALIGN(__size + sizeof(u32), sizeof(u64)); | 1129 | size = ALIGN(__size + sizeof(u32), sizeof(u64)); |
1122 | size -= sizeof(u32); | 1130 | size -= sizeof(u32); |
1123 | if (WARN_ONCE(size > PERF_MAX_TRACE_SIZE, | ||
1124 | "profile buffer not large enough")) | ||
1125 | return; | ||
1126 | 1131 | ||
1127 | entry = perf_trace_buf_prepare(size, call->event.type, regs, &rctx); | 1132 | entry = perf_trace_buf_prepare(size, call->event.type, regs, &rctx); |
1128 | if (!entry) | 1133 | if (!entry) |
diff --git a/kernel/trace/trace_mmiotrace.c b/kernel/trace/trace_mmiotrace.c index a5e8f4878bfa..b3dcfb2f0fef 100644 --- a/kernel/trace/trace_mmiotrace.c +++ b/kernel/trace/trace_mmiotrace.c | |||
@@ -90,7 +90,7 @@ static int mmio_print_pcidev(struct trace_seq *s, const struct pci_dev *dev) | |||
90 | if (drv) | 90 | if (drv) |
91 | ret += trace_seq_printf(s, " %s\n", drv->name); | 91 | ret += trace_seq_printf(s, " %s\n", drv->name); |
92 | else | 92 | else |
93 | ret += trace_seq_printf(s, " \n"); | 93 | ret += trace_seq_puts(s, " \n"); |
94 | return ret; | 94 | return ret; |
95 | } | 95 | } |
96 | 96 | ||
@@ -107,7 +107,7 @@ static void mmio_pipe_open(struct trace_iterator *iter) | |||
107 | struct header_iter *hiter; | 107 | struct header_iter *hiter; |
108 | struct trace_seq *s = &iter->seq; | 108 | struct trace_seq *s = &iter->seq; |
109 | 109 | ||
110 | trace_seq_printf(s, "VERSION 20070824\n"); | 110 | trace_seq_puts(s, "VERSION 20070824\n"); |
111 | 111 | ||
112 | hiter = kzalloc(sizeof(*hiter), GFP_KERNEL); | 112 | hiter = kzalloc(sizeof(*hiter), GFP_KERNEL); |
113 | if (!hiter) | 113 | if (!hiter) |
@@ -209,7 +209,7 @@ static enum print_line_t mmio_print_rw(struct trace_iterator *iter) | |||
209 | (rw->value >> 0) & 0xff, rw->pc, 0); | 209 | (rw->value >> 0) & 0xff, rw->pc, 0); |
210 | break; | 210 | break; |
211 | default: | 211 | default: |
212 | ret = trace_seq_printf(s, "rw what?\n"); | 212 | ret = trace_seq_puts(s, "rw what?\n"); |
213 | break; | 213 | break; |
214 | } | 214 | } |
215 | if (ret) | 215 | if (ret) |
@@ -245,7 +245,7 @@ static enum print_line_t mmio_print_map(struct trace_iterator *iter) | |||
245 | secs, usec_rem, m->map_id, 0UL, 0); | 245 | secs, usec_rem, m->map_id, 0UL, 0); |
246 | break; | 246 | break; |
247 | default: | 247 | default: |
248 | ret = trace_seq_printf(s, "map what?\n"); | 248 | ret = trace_seq_puts(s, "map what?\n"); |
249 | break; | 249 | break; |
250 | } | 250 | } |
251 | if (ret) | 251 | if (ret) |
diff --git a/kernel/trace/trace_output.c b/kernel/trace/trace_output.c index bb922d9ee51b..34e7cbac0c9c 100644 --- a/kernel/trace/trace_output.c +++ b/kernel/trace/trace_output.c | |||
@@ -78,7 +78,7 @@ enum print_line_t trace_print_printk_msg_only(struct trace_iterator *iter) | |||
78 | 78 | ||
79 | trace_assign_type(field, entry); | 79 | trace_assign_type(field, entry); |
80 | 80 | ||
81 | ret = trace_seq_printf(s, "%s", field->buf); | 81 | ret = trace_seq_puts(s, field->buf); |
82 | if (!ret) | 82 | if (!ret) |
83 | return TRACE_TYPE_PARTIAL_LINE; | 83 | return TRACE_TYPE_PARTIAL_LINE; |
84 | 84 | ||
@@ -558,14 +558,14 @@ seq_print_userip_objs(const struct userstack_entry *entry, struct trace_seq *s, | |||
558 | if (ret) | 558 | if (ret) |
559 | ret = trace_seq_puts(s, "??"); | 559 | ret = trace_seq_puts(s, "??"); |
560 | if (ret) | 560 | if (ret) |
561 | ret = trace_seq_puts(s, "\n"); | 561 | ret = trace_seq_putc(s, '\n'); |
562 | continue; | 562 | continue; |
563 | } | 563 | } |
564 | if (!ret) | 564 | if (!ret) |
565 | break; | 565 | break; |
566 | if (ret) | 566 | if (ret) |
567 | ret = seq_print_user_ip(s, mm, ip, sym_flags); | 567 | ret = seq_print_user_ip(s, mm, ip, sym_flags); |
568 | ret = trace_seq_puts(s, "\n"); | 568 | ret = trace_seq_putc(s, '\n'); |
569 | } | 569 | } |
570 | 570 | ||
571 | if (mm) | 571 | if (mm) |
@@ -579,7 +579,7 @@ seq_print_ip_sym(struct trace_seq *s, unsigned long ip, unsigned long sym_flags) | |||
579 | int ret; | 579 | int ret; |
580 | 580 | ||
581 | if (!ip) | 581 | if (!ip) |
582 | return trace_seq_printf(s, "0"); | 582 | return trace_seq_putc(s, '0'); |
583 | 583 | ||
584 | if (sym_flags & TRACE_ITER_SYM_OFFSET) | 584 | if (sym_flags & TRACE_ITER_SYM_OFFSET) |
585 | ret = seq_print_sym_offset(s, "%s", ip); | 585 | ret = seq_print_sym_offset(s, "%s", ip); |
@@ -964,14 +964,14 @@ static enum print_line_t trace_fn_trace(struct trace_iterator *iter, int flags, | |||
964 | goto partial; | 964 | goto partial; |
965 | 965 | ||
966 | if ((flags & TRACE_ITER_PRINT_PARENT) && field->parent_ip) { | 966 | if ((flags & TRACE_ITER_PRINT_PARENT) && field->parent_ip) { |
967 | if (!trace_seq_printf(s, " <-")) | 967 | if (!trace_seq_puts(s, " <-")) |
968 | goto partial; | 968 | goto partial; |
969 | if (!seq_print_ip_sym(s, | 969 | if (!seq_print_ip_sym(s, |
970 | field->parent_ip, | 970 | field->parent_ip, |
971 | flags)) | 971 | flags)) |
972 | goto partial; | 972 | goto partial; |
973 | } | 973 | } |
974 | if (!trace_seq_printf(s, "\n")) | 974 | if (!trace_seq_putc(s, '\n')) |
975 | goto partial; | 975 | goto partial; |
976 | 976 | ||
977 | return TRACE_TYPE_HANDLED; | 977 | return TRACE_TYPE_HANDLED; |
@@ -1210,7 +1210,7 @@ static enum print_line_t trace_stack_print(struct trace_iterator *iter, | |||
1210 | 1210 | ||
1211 | if (!seq_print_ip_sym(s, *p, flags)) | 1211 | if (!seq_print_ip_sym(s, *p, flags)) |
1212 | goto partial; | 1212 | goto partial; |
1213 | if (!trace_seq_puts(s, "\n")) | 1213 | if (!trace_seq_putc(s, '\n')) |
1214 | goto partial; | 1214 | goto partial; |
1215 | } | 1215 | } |
1216 | 1216 | ||
diff --git a/kernel/trace/trace_syscalls.c b/kernel/trace/trace_syscalls.c index 322e16461072..8fd03657bc7d 100644 --- a/kernel/trace/trace_syscalls.c +++ b/kernel/trace/trace_syscalls.c | |||
@@ -175,7 +175,7 @@ print_syscall_exit(struct trace_iterator *iter, int flags, | |||
175 | entry = syscall_nr_to_meta(syscall); | 175 | entry = syscall_nr_to_meta(syscall); |
176 | 176 | ||
177 | if (!entry) { | 177 | if (!entry) { |
178 | trace_seq_printf(s, "\n"); | 178 | trace_seq_putc(s, '\n'); |
179 | return TRACE_TYPE_HANDLED; | 179 | return TRACE_TYPE_HANDLED; |
180 | } | 180 | } |
181 | 181 | ||
@@ -566,15 +566,15 @@ static void perf_syscall_enter(void *ignore, struct pt_regs *regs, long id) | |||
566 | if (!sys_data) | 566 | if (!sys_data) |
567 | return; | 567 | return; |
568 | 568 | ||
569 | head = this_cpu_ptr(sys_data->enter_event->perf_events); | ||
570 | if (hlist_empty(head)) | ||
571 | return; | ||
572 | |||
569 | /* get the size after alignment with the u32 buffer size field */ | 573 | /* get the size after alignment with the u32 buffer size field */ |
570 | size = sizeof(unsigned long) * sys_data->nb_args + sizeof(*rec); | 574 | size = sizeof(unsigned long) * sys_data->nb_args + sizeof(*rec); |
571 | size = ALIGN(size + sizeof(u32), sizeof(u64)); | 575 | size = ALIGN(size + sizeof(u32), sizeof(u64)); |
572 | size -= sizeof(u32); | 576 | size -= sizeof(u32); |
573 | 577 | ||
574 | if (WARN_ONCE(size > PERF_MAX_TRACE_SIZE, | ||
575 | "perf buffer not large enough")) | ||
576 | return; | ||
577 | |||
578 | rec = (struct syscall_trace_enter *)perf_trace_buf_prepare(size, | 578 | rec = (struct syscall_trace_enter *)perf_trace_buf_prepare(size, |
579 | sys_data->enter_event->event.type, regs, &rctx); | 579 | sys_data->enter_event->event.type, regs, &rctx); |
580 | if (!rec) | 580 | if (!rec) |
@@ -583,8 +583,6 @@ static void perf_syscall_enter(void *ignore, struct pt_regs *regs, long id) | |||
583 | rec->nr = syscall_nr; | 583 | rec->nr = syscall_nr; |
584 | syscall_get_arguments(current, regs, 0, sys_data->nb_args, | 584 | syscall_get_arguments(current, regs, 0, sys_data->nb_args, |
585 | (unsigned long *)&rec->args); | 585 | (unsigned long *)&rec->args); |
586 | |||
587 | head = this_cpu_ptr(sys_data->enter_event->perf_events); | ||
588 | perf_trace_buf_submit(rec, size, rctx, 0, 1, regs, head, NULL); | 586 | perf_trace_buf_submit(rec, size, rctx, 0, 1, regs, head, NULL); |
589 | } | 587 | } |
590 | 588 | ||
@@ -642,18 +640,14 @@ static void perf_syscall_exit(void *ignore, struct pt_regs *regs, long ret) | |||
642 | if (!sys_data) | 640 | if (!sys_data) |
643 | return; | 641 | return; |
644 | 642 | ||
643 | head = this_cpu_ptr(sys_data->exit_event->perf_events); | ||
644 | if (hlist_empty(head)) | ||
645 | return; | ||
646 | |||
645 | /* We can probably do that at build time */ | 647 | /* We can probably do that at build time */ |
646 | size = ALIGN(sizeof(*rec) + sizeof(u32), sizeof(u64)); | 648 | size = ALIGN(sizeof(*rec) + sizeof(u32), sizeof(u64)); |
647 | size -= sizeof(u32); | 649 | size -= sizeof(u32); |
648 | 650 | ||
649 | /* | ||
650 | * Impossible, but be paranoid with the future | ||
651 | * How to put this check outside runtime? | ||
652 | */ | ||
653 | if (WARN_ONCE(size > PERF_MAX_TRACE_SIZE, | ||
654 | "exit event has grown above perf buffer size")) | ||
655 | return; | ||
656 | |||
657 | rec = (struct syscall_trace_exit *)perf_trace_buf_prepare(size, | 651 | rec = (struct syscall_trace_exit *)perf_trace_buf_prepare(size, |
658 | sys_data->exit_event->event.type, regs, &rctx); | 652 | sys_data->exit_event->event.type, regs, &rctx); |
659 | if (!rec) | 653 | if (!rec) |
@@ -661,8 +655,6 @@ static void perf_syscall_exit(void *ignore, struct pt_regs *regs, long ret) | |||
661 | 655 | ||
662 | rec->nr = syscall_nr; | 656 | rec->nr = syscall_nr; |
663 | rec->ret = syscall_get_return_value(current, regs); | 657 | rec->ret = syscall_get_return_value(current, regs); |
664 | |||
665 | head = this_cpu_ptr(sys_data->exit_event->perf_events); | ||
666 | perf_trace_buf_submit(rec, size, rctx, 0, 1, regs, head, NULL); | 658 | perf_trace_buf_submit(rec, size, rctx, 0, 1, regs, head, NULL); |
667 | } | 659 | } |
668 | 660 | ||
diff --git a/kernel/trace/trace_uprobe.c b/kernel/trace/trace_uprobe.c index d5d0cd368a56..a23d2d71188e 100644 --- a/kernel/trace/trace_uprobe.c +++ b/kernel/trace/trace_uprobe.c | |||
@@ -818,8 +818,6 @@ static void uprobe_perf_print(struct trace_uprobe *tu, | |||
818 | 818 | ||
819 | size = SIZEOF_TRACE_ENTRY(is_ret_probe(tu)); | 819 | size = SIZEOF_TRACE_ENTRY(is_ret_probe(tu)); |
820 | size = ALIGN(size + tu->size + sizeof(u32), sizeof(u64)) - sizeof(u32); | 820 | size = ALIGN(size + tu->size + sizeof(u32), sizeof(u64)) - sizeof(u32); |
821 | if (WARN_ONCE(size > PERF_MAX_TRACE_SIZE, "profile buffer not large enough")) | ||
822 | return; | ||
823 | 821 | ||
824 | preempt_disable(); | 822 | preempt_disable(); |
825 | head = this_cpu_ptr(call->perf_events); | 823 | head = this_cpu_ptr(call->perf_events); |
diff --git a/kernel/wait.c b/kernel/wait.c index ce0daa320a26..dec68bd4e9d8 100644 --- a/kernel/wait.c +++ b/kernel/wait.c | |||
@@ -333,7 +333,8 @@ int __wait_on_atomic_t(wait_queue_head_t *wq, struct wait_bit_queue *q, | |||
333 | prepare_to_wait(wq, &q->wait, mode); | 333 | prepare_to_wait(wq, &q->wait, mode); |
334 | val = q->key.flags; | 334 | val = q->key.flags; |
335 | if (atomic_read(val) == 0) | 335 | if (atomic_read(val) == 0) |
336 | ret = (*action)(val); | 336 | break; |
337 | ret = (*action)(val); | ||
337 | } while (!ret && atomic_read(val) != 0); | 338 | } while (!ret && atomic_read(val) != 0); |
338 | finish_wait(wq, &q->wait); | 339 | finish_wait(wq, &q->wait); |
339 | return ret; | 340 | return ret; |
diff --git a/lib/Kconfig b/lib/Kconfig index 35da51359d40..71d9f81f6eed 100644 --- a/lib/Kconfig +++ b/lib/Kconfig | |||
@@ -66,8 +66,6 @@ config CRC16 | |||
66 | 66 | ||
67 | config CRC_T10DIF | 67 | config CRC_T10DIF |
68 | tristate "CRC calculation for the T10 Data Integrity Field" | 68 | tristate "CRC calculation for the T10 Data Integrity Field" |
69 | select CRYPTO | ||
70 | select CRYPTO_CRCT10DIF | ||
71 | help | 69 | help |
72 | This option is only needed if a module that's not in the | 70 | This option is only needed if a module that's not in the |
73 | kernel tree needs to calculate CRC checks for use with the | 71 | kernel tree needs to calculate CRC checks for use with the |
diff --git a/lib/crc-t10dif.c b/lib/crc-t10dif.c index fe3428c07b47..fbbd66ed86cd 100644 --- a/lib/crc-t10dif.c +++ b/lib/crc-t10dif.c | |||
@@ -11,44 +11,57 @@ | |||
11 | #include <linux/types.h> | 11 | #include <linux/types.h> |
12 | #include <linux/module.h> | 12 | #include <linux/module.h> |
13 | #include <linux/crc-t10dif.h> | 13 | #include <linux/crc-t10dif.h> |
14 | #include <linux/err.h> | ||
15 | #include <linux/init.h> | ||
16 | #include <crypto/hash.h> | ||
17 | 14 | ||
18 | static struct crypto_shash *crct10dif_tfm; | 15 | /* Table generated using the following polynomium: |
16 | * x^16 + x^15 + x^11 + x^9 + x^8 + x^7 + x^5 + x^4 + x^2 + x + 1 | ||
17 | * gt: 0x8bb7 | ||
18 | */ | ||
19 | static const __u16 t10_dif_crc_table[256] = { | ||
20 | 0x0000, 0x8BB7, 0x9CD9, 0x176E, 0xB205, 0x39B2, 0x2EDC, 0xA56B, | ||
21 | 0xEFBD, 0x640A, 0x7364, 0xF8D3, 0x5DB8, 0xD60F, 0xC161, 0x4AD6, | ||
22 | 0x54CD, 0xDF7A, 0xC814, 0x43A3, 0xE6C8, 0x6D7F, 0x7A11, 0xF1A6, | ||
23 | 0xBB70, 0x30C7, 0x27A9, 0xAC1E, 0x0975, 0x82C2, 0x95AC, 0x1E1B, | ||
24 | 0xA99A, 0x222D, 0x3543, 0xBEF4, 0x1B9F, 0x9028, 0x8746, 0x0CF1, | ||
25 | 0x4627, 0xCD90, 0xDAFE, 0x5149, 0xF422, 0x7F95, 0x68FB, 0xE34C, | ||
26 | 0xFD57, 0x76E0, 0x618E, 0xEA39, 0x4F52, 0xC4E5, 0xD38B, 0x583C, | ||
27 | 0x12EA, 0x995D, 0x8E33, 0x0584, 0xA0EF, 0x2B58, 0x3C36, 0xB781, | ||
28 | 0xD883, 0x5334, 0x445A, 0xCFED, 0x6A86, 0xE131, 0xF65F, 0x7DE8, | ||
29 | 0x373E, 0xBC89, 0xABE7, 0x2050, 0x853B, 0x0E8C, 0x19E2, 0x9255, | ||
30 | 0x8C4E, 0x07F9, 0x1097, 0x9B20, 0x3E4B, 0xB5FC, 0xA292, 0x2925, | ||
31 | 0x63F3, 0xE844, 0xFF2A, 0x749D, 0xD1F6, 0x5A41, 0x4D2F, 0xC698, | ||
32 | 0x7119, 0xFAAE, 0xEDC0, 0x6677, 0xC31C, 0x48AB, 0x5FC5, 0xD472, | ||
33 | 0x9EA4, 0x1513, 0x027D, 0x89CA, 0x2CA1, 0xA716, 0xB078, 0x3BCF, | ||
34 | 0x25D4, 0xAE63, 0xB90D, 0x32BA, 0x97D1, 0x1C66, 0x0B08, 0x80BF, | ||
35 | 0xCA69, 0x41DE, 0x56B0, 0xDD07, 0x786C, 0xF3DB, 0xE4B5, 0x6F02, | ||
36 | 0x3AB1, 0xB106, 0xA668, 0x2DDF, 0x88B4, 0x0303, 0x146D, 0x9FDA, | ||
37 | 0xD50C, 0x5EBB, 0x49D5, 0xC262, 0x6709, 0xECBE, 0xFBD0, 0x7067, | ||
38 | 0x6E7C, 0xE5CB, 0xF2A5, 0x7912, 0xDC79, 0x57CE, 0x40A0, 0xCB17, | ||
39 | 0x81C1, 0x0A76, 0x1D18, 0x96AF, 0x33C4, 0xB873, 0xAF1D, 0x24AA, | ||
40 | 0x932B, 0x189C, 0x0FF2, 0x8445, 0x212E, 0xAA99, 0xBDF7, 0x3640, | ||
41 | 0x7C96, 0xF721, 0xE04F, 0x6BF8, 0xCE93, 0x4524, 0x524A, 0xD9FD, | ||
42 | 0xC7E6, 0x4C51, 0x5B3F, 0xD088, 0x75E3, 0xFE54, 0xE93A, 0x628D, | ||
43 | 0x285B, 0xA3EC, 0xB482, 0x3F35, 0x9A5E, 0x11E9, 0x0687, 0x8D30, | ||
44 | 0xE232, 0x6985, 0x7EEB, 0xF55C, 0x5037, 0xDB80, 0xCCEE, 0x4759, | ||
45 | 0x0D8F, 0x8638, 0x9156, 0x1AE1, 0xBF8A, 0x343D, 0x2353, 0xA8E4, | ||
46 | 0xB6FF, 0x3D48, 0x2A26, 0xA191, 0x04FA, 0x8F4D, 0x9823, 0x1394, | ||
47 | 0x5942, 0xD2F5, 0xC59B, 0x4E2C, 0xEB47, 0x60F0, 0x779E, 0xFC29, | ||
48 | 0x4BA8, 0xC01F, 0xD771, 0x5CC6, 0xF9AD, 0x721A, 0x6574, 0xEEC3, | ||
49 | 0xA415, 0x2FA2, 0x38CC, 0xB37B, 0x1610, 0x9DA7, 0x8AC9, 0x017E, | ||
50 | 0x1F65, 0x94D2, 0x83BC, 0x080B, 0xAD60, 0x26D7, 0x31B9, 0xBA0E, | ||
51 | 0xF0D8, 0x7B6F, 0x6C01, 0xE7B6, 0x42DD, 0xC96A, 0xDE04, 0x55B3 | ||
52 | }; | ||
19 | 53 | ||
20 | __u16 crc_t10dif(const unsigned char *buffer, size_t len) | 54 | __u16 crc_t10dif(const unsigned char *buffer, size_t len) |
21 | { | 55 | { |
22 | struct { | 56 | __u16 crc = 0; |
23 | struct shash_desc shash; | 57 | unsigned int i; |
24 | char ctx[2]; | ||
25 | } desc; | ||
26 | int err; | ||
27 | |||
28 | desc.shash.tfm = crct10dif_tfm; | ||
29 | desc.shash.flags = 0; | ||
30 | *(__u16 *)desc.ctx = 0; | ||
31 | 58 | ||
32 | err = crypto_shash_update(&desc.shash, buffer, len); | 59 | for (i = 0 ; i < len ; i++) |
33 | BUG_ON(err); | 60 | crc = (crc << 8) ^ t10_dif_crc_table[((crc >> 8) ^ buffer[i]) & 0xff]; |
34 | 61 | ||
35 | return *(__u16 *)desc.ctx; | 62 | return crc; |
36 | } | 63 | } |
37 | EXPORT_SYMBOL(crc_t10dif); | 64 | EXPORT_SYMBOL(crc_t10dif); |
38 | 65 | ||
39 | static int __init crc_t10dif_mod_init(void) | ||
40 | { | ||
41 | crct10dif_tfm = crypto_alloc_shash("crct10dif", 0, 0); | ||
42 | return PTR_RET(crct10dif_tfm); | ||
43 | } | ||
44 | |||
45 | static void __exit crc_t10dif_mod_fini(void) | ||
46 | { | ||
47 | crypto_free_shash(crct10dif_tfm); | ||
48 | } | ||
49 | |||
50 | module_init(crc_t10dif_mod_init); | ||
51 | module_exit(crc_t10dif_mod_fini); | ||
52 | |||
53 | MODULE_DESCRIPTION("T10 DIF CRC calculation"); | 66 | MODULE_DESCRIPTION("T10 DIF CRC calculation"); |
54 | MODULE_LICENSE("GPL"); | 67 | MODULE_LICENSE("GPL"); |
diff --git a/lib/mpi/longlong.h b/lib/mpi/longlong.h index d411355f238e..aac511417ad1 100644 --- a/lib/mpi/longlong.h +++ b/lib/mpi/longlong.h | |||
@@ -151,15 +151,12 @@ do { \ | |||
151 | #endif /* __a29k__ */ | 151 | #endif /* __a29k__ */ |
152 | 152 | ||
153 | #if defined(__alpha) && W_TYPE_SIZE == 64 | 153 | #if defined(__alpha) && W_TYPE_SIZE == 64 |
154 | #define umul_ppmm(ph, pl, m0, m1) \ | 154 | #define umul_ppmm(ph, pl, m0, m1) \ |
155 | do { \ | 155 | do { \ |
156 | UDItype __m0 = (m0), __m1 = (m1); \ | 156 | UDItype __m0 = (m0), __m1 = (m1); \ |
157 | __asm__ ("umulh %r1,%2,%0" \ | 157 | (ph) = __builtin_alpha_umulh(__m0, __m1); \ |
158 | : "=r" ((UDItype) ph) \ | 158 | (pl) = __m0 * __m1; \ |
159 | : "%rJ" (__m0), \ | 159 | } while (0) |
160 | "rI" (__m1)); \ | ||
161 | (pl) = __m0 * __m1; \ | ||
162 | } while (0) | ||
163 | #define UMUL_TIME 46 | 160 | #define UMUL_TIME 46 |
164 | #ifndef LONGLONG_STANDALONE | 161 | #ifndef LONGLONG_STANDALONE |
165 | #define udiv_qrnnd(q, r, n1, n0, d) \ | 162 | #define udiv_qrnnd(q, r, n1, n0, d) \ |
@@ -167,7 +164,7 @@ do { UDItype __r; \ | |||
167 | (q) = __udiv_qrnnd(&__r, (n1), (n0), (d)); \ | 164 | (q) = __udiv_qrnnd(&__r, (n1), (n0), (d)); \ |
168 | (r) = __r; \ | 165 | (r) = __r; \ |
169 | } while (0) | 166 | } while (0) |
170 | extern UDItype __udiv_qrnnd(); | 167 | extern UDItype __udiv_qrnnd(UDItype *, UDItype, UDItype, UDItype); |
171 | #define UDIV_TIME 220 | 168 | #define UDIV_TIME 220 |
172 | #endif /* LONGLONG_STANDALONE */ | 169 | #endif /* LONGLONG_STANDALONE */ |
173 | #endif /* __alpha */ | 170 | #endif /* __alpha */ |
diff --git a/sound/pci/hda/patch_sigmatel.c b/sound/pci/hda/patch_sigmatel.c index e2f83591161b..92b9b4324372 100644 --- a/sound/pci/hda/patch_sigmatel.c +++ b/sound/pci/hda/patch_sigmatel.c | |||
@@ -417,9 +417,11 @@ static void stac_update_outputs(struct hda_codec *codec) | |||
417 | val &= ~spec->eapd_mask; | 417 | val &= ~spec->eapd_mask; |
418 | else | 418 | else |
419 | val |= spec->eapd_mask; | 419 | val |= spec->eapd_mask; |
420 | if (spec->gpio_data != val) | 420 | if (spec->gpio_data != val) { |
421 | spec->gpio_data = val; | ||
421 | stac_gpio_set(codec, spec->gpio_mask, spec->gpio_dir, | 422 | stac_gpio_set(codec, spec->gpio_mask, spec->gpio_dir, |
422 | val); | 423 | val); |
424 | } | ||
423 | } | 425 | } |
424 | } | 426 | } |
425 | 427 | ||
@@ -3231,7 +3233,7 @@ static const struct hda_fixup stac927x_fixups[] = { | |||
3231 | /* configure the analog microphone on some laptops */ | 3233 | /* configure the analog microphone on some laptops */ |
3232 | { 0x0c, 0x90a79130 }, | 3234 | { 0x0c, 0x90a79130 }, |
3233 | /* correct the front output jack as a hp out */ | 3235 | /* correct the front output jack as a hp out */ |
3234 | { 0x0f, 0x0227011f }, | 3236 | { 0x0f, 0x0221101f }, |
3235 | /* correct the front input jack as a mic */ | 3237 | /* correct the front input jack as a mic */ |
3236 | { 0x0e, 0x02a79130 }, | 3238 | { 0x0e, 0x02a79130 }, |
3237 | {} | 3239 | {} |
@@ -3612,20 +3614,18 @@ static int stac_parse_auto_config(struct hda_codec *codec) | |||
3612 | static int stac_init(struct hda_codec *codec) | 3614 | static int stac_init(struct hda_codec *codec) |
3613 | { | 3615 | { |
3614 | struct sigmatel_spec *spec = codec->spec; | 3616 | struct sigmatel_spec *spec = codec->spec; |
3615 | unsigned int gpio; | ||
3616 | int i; | 3617 | int i; |
3617 | 3618 | ||
3618 | /* override some hints */ | 3619 | /* override some hints */ |
3619 | stac_store_hints(codec); | 3620 | stac_store_hints(codec); |
3620 | 3621 | ||
3621 | /* set up GPIO */ | 3622 | /* set up GPIO */ |
3622 | gpio = spec->gpio_data; | ||
3623 | /* turn on EAPD statically when spec->eapd_switch isn't set. | 3623 | /* turn on EAPD statically when spec->eapd_switch isn't set. |
3624 | * otherwise, unsol event will turn it on/off dynamically | 3624 | * otherwise, unsol event will turn it on/off dynamically |
3625 | */ | 3625 | */ |
3626 | if (!spec->eapd_switch) | 3626 | if (!spec->eapd_switch) |
3627 | gpio |= spec->eapd_mask; | 3627 | spec->gpio_data |= spec->eapd_mask; |
3628 | stac_gpio_set(codec, spec->gpio_mask, spec->gpio_dir, gpio); | 3628 | stac_gpio_set(codec, spec->gpio_mask, spec->gpio_dir, spec->gpio_data); |
3629 | 3629 | ||
3630 | snd_hda_gen_init(codec); | 3630 | snd_hda_gen_init(codec); |
3631 | 3631 | ||
@@ -3915,6 +3915,7 @@ static void stac_setup_gpio(struct hda_codec *codec) | |||
3915 | { | 3915 | { |
3916 | struct sigmatel_spec *spec = codec->spec; | 3916 | struct sigmatel_spec *spec = codec->spec; |
3917 | 3917 | ||
3918 | spec->gpio_mask |= spec->eapd_mask; | ||
3918 | if (spec->gpio_led) { | 3919 | if (spec->gpio_led) { |
3919 | if (!spec->vref_mute_led_nid) { | 3920 | if (!spec->vref_mute_led_nid) { |
3920 | spec->gpio_mask |= spec->gpio_led; | 3921 | spec->gpio_mask |= spec->gpio_led; |
diff --git a/sound/soc/cirrus/ep93xx-ac97.c b/sound/soc/cirrus/ep93xx-ac97.c index ac73c607410a..04491f0e8d1b 100644 --- a/sound/soc/cirrus/ep93xx-ac97.c +++ b/sound/soc/cirrus/ep93xx-ac97.c | |||
@@ -102,13 +102,13 @@ static struct ep93xx_ac97_info *ep93xx_ac97_info; | |||
102 | 102 | ||
103 | static struct ep93xx_dma_data ep93xx_ac97_pcm_out = { | 103 | static struct ep93xx_dma_data ep93xx_ac97_pcm_out = { |
104 | .name = "ac97-pcm-out", | 104 | .name = "ac97-pcm-out", |
105 | .dma_port = EP93XX_DMA_AAC1, | 105 | .port = EP93XX_DMA_AAC1, |
106 | .direction = DMA_MEM_TO_DEV, | 106 | .direction = DMA_MEM_TO_DEV, |
107 | }; | 107 | }; |
108 | 108 | ||
109 | static struct ep93xx_dma_data ep93xx_ac97_pcm_in = { | 109 | static struct ep93xx_dma_data ep93xx_ac97_pcm_in = { |
110 | .name = "ac97-pcm-in", | 110 | .name = "ac97-pcm-in", |
111 | .dma_port = EP93XX_DMA_AAC1, | 111 | .port = EP93XX_DMA_AAC1, |
112 | .direction = DMA_DEV_TO_MEM, | 112 | .direction = DMA_DEV_TO_MEM, |
113 | }; | 113 | }; |
114 | 114 | ||
diff --git a/sound/soc/codecs/max98088.c b/sound/soc/codecs/max98088.c index 3eeada57e87d..566a367c94fa 100644 --- a/sound/soc/codecs/max98088.c +++ b/sound/soc/codecs/max98088.c | |||
@@ -1612,7 +1612,7 @@ static int max98088_dai2_digital_mute(struct snd_soc_dai *codec_dai, int mute) | |||
1612 | 1612 | ||
1613 | static void max98088_sync_cache(struct snd_soc_codec *codec) | 1613 | static void max98088_sync_cache(struct snd_soc_codec *codec) |
1614 | { | 1614 | { |
1615 | u16 *reg_cache = codec->reg_cache; | 1615 | u8 *reg_cache = codec->reg_cache; |
1616 | int i; | 1616 | int i; |
1617 | 1617 | ||
1618 | if (!codec->cache_sync) | 1618 | if (!codec->cache_sync) |
diff --git a/sound/soc/codecs/sgtl5000.c b/sound/soc/codecs/sgtl5000.c index d659d3adcfb3..6c8a9e7bee25 100644 --- a/sound/soc/codecs/sgtl5000.c +++ b/sound/soc/codecs/sgtl5000.c | |||
@@ -1527,6 +1527,9 @@ static int sgtl5000_i2c_probe(struct i2c_client *client, | |||
1527 | if (IS_ERR(sgtl5000->mclk)) { | 1527 | if (IS_ERR(sgtl5000->mclk)) { |
1528 | ret = PTR_ERR(sgtl5000->mclk); | 1528 | ret = PTR_ERR(sgtl5000->mclk); |
1529 | dev_err(&client->dev, "Failed to get mclock: %d\n", ret); | 1529 | dev_err(&client->dev, "Failed to get mclock: %d\n", ret); |
1530 | /* Defer the probe to see if the clk will be provided later */ | ||
1531 | if (ret == -ENOENT) | ||
1532 | return -EPROBE_DEFER; | ||
1530 | return ret; | 1533 | return ret; |
1531 | } | 1534 | } |
1532 | 1535 | ||
diff --git a/sound/soc/soc-core.c b/sound/soc/soc-core.c index 0ec070cf7231..d82ee386eab5 100644 --- a/sound/soc/soc-core.c +++ b/sound/soc/soc-core.c | |||
@@ -3908,10 +3908,8 @@ int snd_soc_add_platform(struct device *dev, struct snd_soc_platform *platform, | |||
3908 | { | 3908 | { |
3909 | /* create platform component name */ | 3909 | /* create platform component name */ |
3910 | platform->name = fmt_single_name(dev, &platform->id); | 3910 | platform->name = fmt_single_name(dev, &platform->id); |
3911 | if (platform->name == NULL) { | 3911 | if (platform->name == NULL) |
3912 | kfree(platform); | ||
3913 | return -ENOMEM; | 3912 | return -ENOMEM; |
3914 | } | ||
3915 | 3913 | ||
3916 | platform->dev = dev; | 3914 | platform->dev = dev; |
3917 | platform->driver = platform_drv; | 3915 | platform->driver = platform_drv; |
diff --git a/sound/soc/tegra/tegra20_ac97.c b/sound/soc/tegra/tegra20_ac97.c index e58233f7df61..6c486625321b 100644 --- a/sound/soc/tegra/tegra20_ac97.c +++ b/sound/soc/tegra/tegra20_ac97.c | |||
@@ -389,9 +389,9 @@ static int tegra20_ac97_platform_probe(struct platform_device *pdev) | |||
389 | ac97->capture_dma_data.slave_id = of_dma[1]; | 389 | ac97->capture_dma_data.slave_id = of_dma[1]; |
390 | 390 | ||
391 | ac97->playback_dma_data.addr = mem->start + TEGRA20_AC97_FIFO_TX1; | 391 | ac97->playback_dma_data.addr = mem->start + TEGRA20_AC97_FIFO_TX1; |
392 | ac97->capture_dma_data.addr_width = DMA_SLAVE_BUSWIDTH_4_BYTES; | 392 | ac97->playback_dma_data.addr_width = DMA_SLAVE_BUSWIDTH_4_BYTES; |
393 | ac97->capture_dma_data.maxburst = 4; | 393 | ac97->playback_dma_data.maxburst = 4; |
394 | ac97->capture_dma_data.slave_id = of_dma[0]; | 394 | ac97->playback_dma_data.slave_id = of_dma[1]; |
395 | 395 | ||
396 | ret = tegra_asoc_utils_init(&ac97->util_data, &pdev->dev); | 396 | ret = tegra_asoc_utils_init(&ac97->util_data, &pdev->dev); |
397 | if (ret) | 397 | if (ret) |
diff --git a/sound/soc/tegra/tegra20_spdif.c b/sound/soc/tegra/tegra20_spdif.c index 5eaa12cdc6eb..551b3c93ce93 100644 --- a/sound/soc/tegra/tegra20_spdif.c +++ b/sound/soc/tegra/tegra20_spdif.c | |||
@@ -323,8 +323,8 @@ static int tegra20_spdif_platform_probe(struct platform_device *pdev) | |||
323 | } | 323 | } |
324 | 324 | ||
325 | spdif->playback_dma_data.addr = mem->start + TEGRA20_SPDIF_DATA_OUT; | 325 | spdif->playback_dma_data.addr = mem->start + TEGRA20_SPDIF_DATA_OUT; |
326 | spdif->capture_dma_data.addr_width = DMA_SLAVE_BUSWIDTH_4_BYTES; | 326 | spdif->playback_dma_data.addr_width = DMA_SLAVE_BUSWIDTH_4_BYTES; |
327 | spdif->capture_dma_data.maxburst = 4; | 327 | spdif->playback_dma_data.maxburst = 4; |
328 | spdif->playback_dma_data.slave_id = dmareq->start; | 328 | spdif->playback_dma_data.slave_id = dmareq->start; |
329 | 329 | ||
330 | pm_runtime_enable(&pdev->dev); | 330 | pm_runtime_enable(&pdev->dev); |
diff --git a/sound/usb/6fire/pcm.c b/sound/usb/6fire/pcm.c index 2aa4e13063a8..3d2551cc10f2 100644 --- a/sound/usb/6fire/pcm.c +++ b/sound/usb/6fire/pcm.c | |||
@@ -543,7 +543,7 @@ static snd_pcm_uframes_t usb6fire_pcm_pointer( | |||
543 | snd_pcm_uframes_t ret; | 543 | snd_pcm_uframes_t ret; |
544 | 544 | ||
545 | if (rt->panic || !sub) | 545 | if (rt->panic || !sub) |
546 | return SNDRV_PCM_STATE_XRUN; | 546 | return SNDRV_PCM_POS_XRUN; |
547 | 547 | ||
548 | spin_lock_irqsave(&sub->lock, flags); | 548 | spin_lock_irqsave(&sub->lock, flags); |
549 | ret = sub->dma_off; | 549 | ret = sub->dma_off; |
diff --git a/sound/usb/hiface/pcm.c b/sound/usb/hiface/pcm.c index 6430ed2a9f65..c21a3df9a0df 100644 --- a/sound/usb/hiface/pcm.c +++ b/sound/usb/hiface/pcm.c | |||
@@ -503,7 +503,7 @@ static snd_pcm_uframes_t hiface_pcm_pointer(struct snd_pcm_substream *alsa_sub) | |||
503 | snd_pcm_uframes_t dma_offset; | 503 | snd_pcm_uframes_t dma_offset; |
504 | 504 | ||
505 | if (rt->panic || !sub) | 505 | if (rt->panic || !sub) |
506 | return SNDRV_PCM_STATE_XRUN; | 506 | return SNDRV_PCM_POS_XRUN; |
507 | 507 | ||
508 | spin_lock_irqsave(&sub->lock, flags); | 508 | spin_lock_irqsave(&sub->lock, flags); |
509 | dma_offset = sub->dma_off; | 509 | dma_offset = sub->dma_off; |
diff --git a/tools/hv/hv_kvp_daemon.c b/tools/hv/hv_kvp_daemon.c index ca9fa4d32e07..07819bfa7dba 100644 --- a/tools/hv/hv_kvp_daemon.c +++ b/tools/hv/hv_kvp_daemon.c | |||
@@ -1026,9 +1026,10 @@ kvp_get_ip_info(int family, char *if_name, int op, | |||
1026 | 1026 | ||
1027 | if (sn_offset == 0) | 1027 | if (sn_offset == 0) |
1028 | strcpy(sn_str, cidr_mask); | 1028 | strcpy(sn_str, cidr_mask); |
1029 | else | 1029 | else { |
1030 | strcat((char *)ip_buffer->sub_net, ";"); | ||
1030 | strcat(sn_str, cidr_mask); | 1031 | strcat(sn_str, cidr_mask); |
1031 | strcat((char *)ip_buffer->sub_net, ";"); | 1032 | } |
1032 | sn_offset += strlen(sn_str) + 1; | 1033 | sn_offset += strlen(sn_str) + 1; |
1033 | } | 1034 | } |
1034 | 1035 | ||