diff options
author | Linus Torvalds <torvalds@linux-foundation.org> | 2008-04-18 11:38:06 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@linux-foundation.org> | 2008-04-18 11:38:06 -0400 |
commit | 8019aa946af5218bc4446c21e43cc19c9401ac68 (patch) | |
tree | 6681b73a560f5a4e4a4e8f97b3833a087a488439 | |
parent | 73e3e6481f56b3b5b618671a8d32b19a35f84316 (diff) | |
parent | 48feb3c419508487becfb9ea3afcc54c3eac6d80 (diff) |
Merge branch 'upstream-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/jgarzik/libata-dev
* 'upstream-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/jgarzik/libata-dev: (79 commits)
ata-acpi: don't call _GTF for disabled drive
sata_mv add temporary 3 second init delay for SiliconImage PMs
sata_mv remove redundant edma init code
sata_mv add basic port multiplier support
sata_mv fix SOC flags, enable NCQ on SOC
sata_mv disable hotplug for now
sata_mv cosmetics
sata_mv hardreset rework
[libata] improve Kconfig help text for new PMP, SFF options
libata: make EH fail gracefully if no reset method is available
libata: Be a bit more slack about early devices
libata: cable logic
libata: move link onlineness check out of softreset methods
libata: kill dead code paths in reset path
pata_scc: fix build breakage
libata: make PMP support optional
libata: implement PMP helpers
libata: separate PMP support code from core code
libata: make SFF support optional
libata: don't use ap->ioaddr in non-SFF drivers
...
83 files changed, 4787 insertions, 8801 deletions
diff --git a/drivers/ata/Kconfig b/drivers/ata/Kconfig index 25aba69b59b4..292aa9a0f02f 100644 --- a/drivers/ata/Kconfig +++ b/drivers/ata/Kconfig | |||
@@ -28,7 +28,7 @@ config ATA_NONSTANDARD | |||
28 | default n | 28 | default n |
29 | 29 | ||
30 | config ATA_ACPI | 30 | config ATA_ACPI |
31 | bool | 31 | bool "ATA ACPI Support" |
32 | depends on ACPI && PCI | 32 | depends on ACPI && PCI |
33 | select ACPI_DOCK | 33 | select ACPI_DOCK |
34 | default y | 34 | default y |
@@ -41,6 +41,13 @@ config ATA_ACPI | |||
41 | You can disable this at kernel boot time by using the | 41 | You can disable this at kernel boot time by using the |
42 | option libata.noacpi=1 | 42 | option libata.noacpi=1 |
43 | 43 | ||
44 | config SATA_PMP | ||
45 | bool "SATA Port Multiplier support" | ||
46 | default y | ||
47 | help | ||
48 | This option adds support for SATA Port Multipliers | ||
49 | (the SATA version of an ethernet hub, or SAS expander). | ||
50 | |||
44 | config SATA_AHCI | 51 | config SATA_AHCI |
45 | tristate "AHCI SATA support" | 52 | tristate "AHCI SATA support" |
46 | depends on PCI | 53 | depends on PCI |
@@ -49,6 +56,43 @@ config SATA_AHCI | |||
49 | 56 | ||
50 | If unsure, say N. | 57 | If unsure, say N. |
51 | 58 | ||
59 | config SATA_SIL24 | ||
60 | tristate "Silicon Image 3124/3132 SATA support" | ||
61 | depends on PCI | ||
62 | help | ||
63 | This option enables support for Silicon Image 3124/3132 Serial ATA. | ||
64 | |||
65 | If unsure, say N. | ||
66 | |||
67 | config SATA_FSL | ||
68 | tristate "Freescale 3.0Gbps SATA support" | ||
69 | depends on FSL_SOC | ||
70 | help | ||
71 | This option enables support for Freescale 3.0Gbps SATA controller. | ||
72 | It can be found on MPC837x and MPC8315. | ||
73 | |||
74 | If unsure, say N. | ||
75 | |||
76 | config ATA_SFF | ||
77 | bool "ATA SFF support" | ||
78 | default y | ||
79 | help | ||
80 | This option adds support for ATA controllers with SFF | ||
81 | compliant or similar programming interface. | ||
82 | |||
83 | SFF is the legacy IDE interface that has been around since | ||
84 | the dawn of time. Almost all PATA controllers have an | ||
85 | SFF interface. Many SATA controllers have an SFF interface | ||
86 | when configured into a legacy compatibility mode. | ||
87 | |||
88 | For users with exclusively modern controllers like AHCI, | ||
89 | Silicon Image 3124, or Marvell 6440, you may choose to | ||
90 | disable this uneeded SFF support. | ||
91 | |||
92 | If unsure, say Y. | ||
93 | |||
94 | if ATA_SFF | ||
95 | |||
52 | config SATA_SVW | 96 | config SATA_SVW |
53 | tristate "ServerWorks Frodo / Apple K2 SATA support" | 97 | tristate "ServerWorks Frodo / Apple K2 SATA support" |
54 | depends on PCI | 98 | depends on PCI |
@@ -125,14 +169,6 @@ config SATA_SIL | |||
125 | 169 | ||
126 | If unsure, say N. | 170 | If unsure, say N. |
127 | 171 | ||
128 | config SATA_SIL24 | ||
129 | tristate "Silicon Image 3124/3132 SATA support" | ||
130 | depends on PCI | ||
131 | help | ||
132 | This option enables support for Silicon Image 3124/3132 Serial ATA. | ||
133 | |||
134 | If unsure, say N. | ||
135 | |||
136 | config SATA_SIS | 172 | config SATA_SIS |
137 | tristate "SiS 964/965/966/180 SATA support" | 173 | tristate "SiS 964/965/966/180 SATA support" |
138 | depends on PCI | 174 | depends on PCI |
@@ -183,15 +219,6 @@ config PATA_ACPI | |||
183 | firmware in the BIOS. This driver can sometimes handle | 219 | firmware in the BIOS. This driver can sometimes handle |
184 | otherwise unsupported hardware. | 220 | otherwise unsupported hardware. |
185 | 221 | ||
186 | config SATA_FSL | ||
187 | tristate "Freescale 3.0Gbps SATA support" | ||
188 | depends on FSL_SOC | ||
189 | help | ||
190 | This option enables support for Freescale 3.0Gbps SATA controller. | ||
191 | It can be found on MPC837x and MPC8315. | ||
192 | |||
193 | If unsure, say N. | ||
194 | |||
195 | config PATA_ALI | 222 | config PATA_ALI |
196 | tristate "ALi PATA support (Experimental)" | 223 | tristate "ALi PATA support (Experimental)" |
197 | depends on PCI && EXPERIMENTAL | 224 | depends on PCI && EXPERIMENTAL |
@@ -679,4 +706,5 @@ config PATA_BF54X | |||
679 | 706 | ||
680 | If unsure, say N. | 707 | If unsure, say N. |
681 | 708 | ||
709 | endif # ATA_SFF | ||
682 | endif # ATA | 710 | endif # ATA |
diff --git a/drivers/ata/Makefile b/drivers/ata/Makefile index 0511e6f0bb58..1fbc2aa648b7 100644 --- a/drivers/ata/Makefile +++ b/drivers/ata/Makefile | |||
@@ -78,6 +78,7 @@ obj-$(CONFIG_ATA_GENERIC) += ata_generic.o | |||
78 | # Should be last libata driver | 78 | # Should be last libata driver |
79 | obj-$(CONFIG_PATA_LEGACY) += pata_legacy.o | 79 | obj-$(CONFIG_PATA_LEGACY) += pata_legacy.o |
80 | 80 | ||
81 | libata-objs := libata-core.o libata-scsi.o libata-sff.o libata-eh.o \ | 81 | libata-objs := libata-core.o libata-scsi.o libata-eh.o |
82 | libata-pmp.o | 82 | libata-$(CONFIG_ATA_SFF) += libata-sff.o |
83 | libata-$(CONFIG_SATA_PMP) += libata-pmp.o | ||
83 | libata-$(CONFIG_ATA_ACPI) += libata-acpi.o | 84 | libata-$(CONFIG_ATA_ACPI) += libata-acpi.o |
diff --git a/drivers/ata/ahci.c b/drivers/ata/ahci.c index b1eb4e24c86a..739ba3f222e8 100644 --- a/drivers/ata/ahci.c +++ b/drivers/ata/ahci.c | |||
@@ -62,7 +62,6 @@ enum { | |||
62 | AHCI_MAX_PORTS = 32, | 62 | AHCI_MAX_PORTS = 32, |
63 | AHCI_MAX_SG = 168, /* hardware max is 64K */ | 63 | AHCI_MAX_SG = 168, /* hardware max is 64K */ |
64 | AHCI_DMA_BOUNDARY = 0xffffffff, | 64 | AHCI_DMA_BOUNDARY = 0xffffffff, |
65 | AHCI_USE_CLUSTERING = 1, | ||
66 | AHCI_MAX_CMDS = 32, | 65 | AHCI_MAX_CMDS = 32, |
67 | AHCI_CMD_SZ = 32, | 66 | AHCI_CMD_SZ = 32, |
68 | AHCI_CMD_SLOT_SZ = AHCI_MAX_CMDS * AHCI_CMD_SZ, | 67 | AHCI_CMD_SLOT_SZ = AHCI_MAX_CMDS * AHCI_CMD_SZ, |
@@ -198,7 +197,6 @@ enum { | |||
198 | ATA_FLAG_MMIO | ATA_FLAG_PIO_DMA | | 197 | ATA_FLAG_MMIO | ATA_FLAG_PIO_DMA | |
199 | ATA_FLAG_ACPI_SATA | ATA_FLAG_AN | | 198 | ATA_FLAG_ACPI_SATA | ATA_FLAG_AN | |
200 | ATA_FLAG_IPM, | 199 | ATA_FLAG_IPM, |
201 | AHCI_LFLAG_COMMON = ATA_LFLAG_SKIP_D2H_BSY, | ||
202 | 200 | ||
203 | ICH_MAP = 0x90, /* ICH MAP register */ | 201 | ICH_MAP = 0x90, /* ICH MAP register */ |
204 | }; | 202 | }; |
@@ -245,19 +243,24 @@ static int ahci_scr_read(struct ata_port *ap, unsigned int sc_reg, u32 *val); | |||
245 | static int ahci_scr_write(struct ata_port *ap, unsigned int sc_reg, u32 val); | 243 | static int ahci_scr_write(struct ata_port *ap, unsigned int sc_reg, u32 val); |
246 | static int ahci_init_one(struct pci_dev *pdev, const struct pci_device_id *ent); | 244 | static int ahci_init_one(struct pci_dev *pdev, const struct pci_device_id *ent); |
247 | static unsigned int ahci_qc_issue(struct ata_queued_cmd *qc); | 245 | static unsigned int ahci_qc_issue(struct ata_queued_cmd *qc); |
248 | static void ahci_irq_clear(struct ata_port *ap); | 246 | static bool ahci_qc_fill_rtf(struct ata_queued_cmd *qc); |
249 | static int ahci_port_start(struct ata_port *ap); | 247 | static int ahci_port_start(struct ata_port *ap); |
250 | static void ahci_port_stop(struct ata_port *ap); | 248 | static void ahci_port_stop(struct ata_port *ap); |
251 | static void ahci_tf_read(struct ata_port *ap, struct ata_taskfile *tf); | ||
252 | static void ahci_qc_prep(struct ata_queued_cmd *qc); | 249 | static void ahci_qc_prep(struct ata_queued_cmd *qc); |
253 | static u8 ahci_check_status(struct ata_port *ap); | ||
254 | static void ahci_freeze(struct ata_port *ap); | 250 | static void ahci_freeze(struct ata_port *ap); |
255 | static void ahci_thaw(struct ata_port *ap); | 251 | static void ahci_thaw(struct ata_port *ap); |
256 | static void ahci_pmp_attach(struct ata_port *ap); | 252 | static void ahci_pmp_attach(struct ata_port *ap); |
257 | static void ahci_pmp_detach(struct ata_port *ap); | 253 | static void ahci_pmp_detach(struct ata_port *ap); |
254 | static int ahci_softreset(struct ata_link *link, unsigned int *class, | ||
255 | unsigned long deadline); | ||
256 | static int ahci_hardreset(struct ata_link *link, unsigned int *class, | ||
257 | unsigned long deadline); | ||
258 | static int ahci_vt8251_hardreset(struct ata_link *link, unsigned int *class, | ||
259 | unsigned long deadline); | ||
260 | static int ahci_p5wdh_hardreset(struct ata_link *link, unsigned int *class, | ||
261 | unsigned long deadline); | ||
262 | static void ahci_postreset(struct ata_link *link, unsigned int *class); | ||
258 | static void ahci_error_handler(struct ata_port *ap); | 263 | static void ahci_error_handler(struct ata_port *ap); |
259 | static void ahci_vt8251_error_handler(struct ata_port *ap); | ||
260 | static void ahci_p5wdh_error_handler(struct ata_port *ap); | ||
261 | static void ahci_post_internal_cmd(struct ata_queued_cmd *qc); | 264 | static void ahci_post_internal_cmd(struct ata_queued_cmd *qc); |
262 | static int ahci_port_resume(struct ata_port *ap); | 265 | static int ahci_port_resume(struct ata_port *ap); |
263 | static void ahci_dev_config(struct ata_device *dev); | 266 | static void ahci_dev_config(struct ata_device *dev); |
@@ -276,129 +279,54 @@ static struct class_device_attribute *ahci_shost_attrs[] = { | |||
276 | }; | 279 | }; |
277 | 280 | ||
278 | static struct scsi_host_template ahci_sht = { | 281 | static struct scsi_host_template ahci_sht = { |
279 | .module = THIS_MODULE, | 282 | ATA_NCQ_SHT(DRV_NAME), |
280 | .name = DRV_NAME, | ||
281 | .ioctl = ata_scsi_ioctl, | ||
282 | .queuecommand = ata_scsi_queuecmd, | ||
283 | .change_queue_depth = ata_scsi_change_queue_depth, | ||
284 | .can_queue = AHCI_MAX_CMDS - 1, | 283 | .can_queue = AHCI_MAX_CMDS - 1, |
285 | .this_id = ATA_SHT_THIS_ID, | ||
286 | .sg_tablesize = AHCI_MAX_SG, | 284 | .sg_tablesize = AHCI_MAX_SG, |
287 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
288 | .emulated = ATA_SHT_EMULATED, | ||
289 | .use_clustering = AHCI_USE_CLUSTERING, | ||
290 | .proc_name = DRV_NAME, | ||
291 | .dma_boundary = AHCI_DMA_BOUNDARY, | 285 | .dma_boundary = AHCI_DMA_BOUNDARY, |
292 | .slave_configure = ata_scsi_slave_config, | ||
293 | .slave_destroy = ata_scsi_slave_destroy, | ||
294 | .bios_param = ata_std_bios_param, | ||
295 | .shost_attrs = ahci_shost_attrs, | 286 | .shost_attrs = ahci_shost_attrs, |
296 | }; | 287 | }; |
297 | 288 | ||
298 | static const struct ata_port_operations ahci_ops = { | 289 | static struct ata_port_operations ahci_ops = { |
299 | .check_status = ahci_check_status, | 290 | .inherits = &sata_pmp_port_ops, |
300 | .check_altstatus = ahci_check_status, | ||
301 | .dev_select = ata_noop_dev_select, | ||
302 | |||
303 | .dev_config = ahci_dev_config, | ||
304 | |||
305 | .tf_read = ahci_tf_read, | ||
306 | 291 | ||
307 | .qc_defer = sata_pmp_qc_defer_cmd_switch, | 292 | .qc_defer = sata_pmp_qc_defer_cmd_switch, |
308 | .qc_prep = ahci_qc_prep, | 293 | .qc_prep = ahci_qc_prep, |
309 | .qc_issue = ahci_qc_issue, | 294 | .qc_issue = ahci_qc_issue, |
310 | 295 | .qc_fill_rtf = ahci_qc_fill_rtf, | |
311 | .irq_clear = ahci_irq_clear, | ||
312 | |||
313 | .scr_read = ahci_scr_read, | ||
314 | .scr_write = ahci_scr_write, | ||
315 | 296 | ||
316 | .freeze = ahci_freeze, | 297 | .freeze = ahci_freeze, |
317 | .thaw = ahci_thaw, | 298 | .thaw = ahci_thaw, |
318 | 299 | .softreset = ahci_softreset, | |
300 | .hardreset = ahci_hardreset, | ||
301 | .postreset = ahci_postreset, | ||
302 | .pmp_softreset = ahci_softreset, | ||
319 | .error_handler = ahci_error_handler, | 303 | .error_handler = ahci_error_handler, |
320 | .post_internal_cmd = ahci_post_internal_cmd, | 304 | .post_internal_cmd = ahci_post_internal_cmd, |
321 | 305 | .dev_config = ahci_dev_config, | |
322 | .pmp_attach = ahci_pmp_attach, | ||
323 | .pmp_detach = ahci_pmp_detach, | ||
324 | |||
325 | #ifdef CONFIG_PM | ||
326 | .port_suspend = ahci_port_suspend, | ||
327 | .port_resume = ahci_port_resume, | ||
328 | #endif | ||
329 | .enable_pm = ahci_enable_alpm, | ||
330 | .disable_pm = ahci_disable_alpm, | ||
331 | |||
332 | .port_start = ahci_port_start, | ||
333 | .port_stop = ahci_port_stop, | ||
334 | }; | ||
335 | |||
336 | static const struct ata_port_operations ahci_vt8251_ops = { | ||
337 | .check_status = ahci_check_status, | ||
338 | .check_altstatus = ahci_check_status, | ||
339 | .dev_select = ata_noop_dev_select, | ||
340 | |||
341 | .tf_read = ahci_tf_read, | ||
342 | |||
343 | .qc_defer = sata_pmp_qc_defer_cmd_switch, | ||
344 | .qc_prep = ahci_qc_prep, | ||
345 | .qc_issue = ahci_qc_issue, | ||
346 | |||
347 | .irq_clear = ahci_irq_clear, | ||
348 | 306 | ||
349 | .scr_read = ahci_scr_read, | 307 | .scr_read = ahci_scr_read, |
350 | .scr_write = ahci_scr_write, | 308 | .scr_write = ahci_scr_write, |
351 | |||
352 | .freeze = ahci_freeze, | ||
353 | .thaw = ahci_thaw, | ||
354 | |||
355 | .error_handler = ahci_vt8251_error_handler, | ||
356 | .post_internal_cmd = ahci_post_internal_cmd, | ||
357 | |||
358 | .pmp_attach = ahci_pmp_attach, | 309 | .pmp_attach = ahci_pmp_attach, |
359 | .pmp_detach = ahci_pmp_detach, | 310 | .pmp_detach = ahci_pmp_detach, |
360 | 311 | ||
312 | .enable_pm = ahci_enable_alpm, | ||
313 | .disable_pm = ahci_disable_alpm, | ||
361 | #ifdef CONFIG_PM | 314 | #ifdef CONFIG_PM |
362 | .port_suspend = ahci_port_suspend, | 315 | .port_suspend = ahci_port_suspend, |
363 | .port_resume = ahci_port_resume, | 316 | .port_resume = ahci_port_resume, |
364 | #endif | 317 | #endif |
365 | |||
366 | .port_start = ahci_port_start, | 318 | .port_start = ahci_port_start, |
367 | .port_stop = ahci_port_stop, | 319 | .port_stop = ahci_port_stop, |
368 | }; | 320 | }; |
369 | 321 | ||
370 | static const struct ata_port_operations ahci_p5wdh_ops = { | 322 | static struct ata_port_operations ahci_vt8251_ops = { |
371 | .check_status = ahci_check_status, | 323 | .inherits = &ahci_ops, |
372 | .check_altstatus = ahci_check_status, | 324 | .hardreset = ahci_vt8251_hardreset, |
373 | .dev_select = ata_noop_dev_select, | 325 | }; |
374 | |||
375 | .tf_read = ahci_tf_read, | ||
376 | |||
377 | .qc_defer = sata_pmp_qc_defer_cmd_switch, | ||
378 | .qc_prep = ahci_qc_prep, | ||
379 | .qc_issue = ahci_qc_issue, | ||
380 | |||
381 | .irq_clear = ahci_irq_clear, | ||
382 | |||
383 | .scr_read = ahci_scr_read, | ||
384 | .scr_write = ahci_scr_write, | ||
385 | |||
386 | .freeze = ahci_freeze, | ||
387 | .thaw = ahci_thaw, | ||
388 | |||
389 | .error_handler = ahci_p5wdh_error_handler, | ||
390 | .post_internal_cmd = ahci_post_internal_cmd, | ||
391 | |||
392 | .pmp_attach = ahci_pmp_attach, | ||
393 | .pmp_detach = ahci_pmp_detach, | ||
394 | |||
395 | #ifdef CONFIG_PM | ||
396 | .port_suspend = ahci_port_suspend, | ||
397 | .port_resume = ahci_port_resume, | ||
398 | #endif | ||
399 | 326 | ||
400 | .port_start = ahci_port_start, | 327 | static struct ata_port_operations ahci_p5wdh_ops = { |
401 | .port_stop = ahci_port_stop, | 328 | .inherits = &ahci_ops, |
329 | .hardreset = ahci_p5wdh_hardreset, | ||
402 | }; | 330 | }; |
403 | 331 | ||
404 | #define AHCI_HFLAGS(flags) .private_data = (void *)(flags) | 332 | #define AHCI_HFLAGS(flags) .private_data = (void *)(flags) |
@@ -407,7 +335,6 @@ static const struct ata_port_info ahci_port_info[] = { | |||
407 | /* board_ahci */ | 335 | /* board_ahci */ |
408 | { | 336 | { |
409 | .flags = AHCI_FLAG_COMMON, | 337 | .flags = AHCI_FLAG_COMMON, |
410 | .link_flags = AHCI_LFLAG_COMMON, | ||
411 | .pio_mask = 0x1f, /* pio0-4 */ | 338 | .pio_mask = 0x1f, /* pio0-4 */ |
412 | .udma_mask = ATA_UDMA6, | 339 | .udma_mask = ATA_UDMA6, |
413 | .port_ops = &ahci_ops, | 340 | .port_ops = &ahci_ops, |
@@ -416,7 +343,6 @@ static const struct ata_port_info ahci_port_info[] = { | |||
416 | { | 343 | { |
417 | AHCI_HFLAGS (AHCI_HFLAG_NO_NCQ | AHCI_HFLAG_NO_PMP), | 344 | AHCI_HFLAGS (AHCI_HFLAG_NO_NCQ | AHCI_HFLAG_NO_PMP), |
418 | .flags = AHCI_FLAG_COMMON, | 345 | .flags = AHCI_FLAG_COMMON, |
419 | .link_flags = AHCI_LFLAG_COMMON | ATA_LFLAG_HRST_TO_RESUME, | ||
420 | .pio_mask = 0x1f, /* pio0-4 */ | 346 | .pio_mask = 0x1f, /* pio0-4 */ |
421 | .udma_mask = ATA_UDMA6, | 347 | .udma_mask = ATA_UDMA6, |
422 | .port_ops = &ahci_vt8251_ops, | 348 | .port_ops = &ahci_vt8251_ops, |
@@ -425,7 +351,6 @@ static const struct ata_port_info ahci_port_info[] = { | |||
425 | { | 351 | { |
426 | AHCI_HFLAGS (AHCI_HFLAG_IGN_IRQ_IF_ERR), | 352 | AHCI_HFLAGS (AHCI_HFLAG_IGN_IRQ_IF_ERR), |
427 | .flags = AHCI_FLAG_COMMON, | 353 | .flags = AHCI_FLAG_COMMON, |
428 | .link_flags = AHCI_LFLAG_COMMON, | ||
429 | .pio_mask = 0x1f, /* pio0-4 */ | 354 | .pio_mask = 0x1f, /* pio0-4 */ |
430 | .udma_mask = ATA_UDMA6, | 355 | .udma_mask = ATA_UDMA6, |
431 | .port_ops = &ahci_ops, | 356 | .port_ops = &ahci_ops, |
@@ -436,7 +361,6 @@ static const struct ata_port_info ahci_port_info[] = { | |||
436 | AHCI_HFLAG_32BIT_ONLY | | 361 | AHCI_HFLAG_32BIT_ONLY | |
437 | AHCI_HFLAG_SECT255 | AHCI_HFLAG_NO_PMP), | 362 | AHCI_HFLAG_SECT255 | AHCI_HFLAG_NO_PMP), |
438 | .flags = AHCI_FLAG_COMMON, | 363 | .flags = AHCI_FLAG_COMMON, |
439 | .link_flags = AHCI_LFLAG_COMMON, | ||
440 | .pio_mask = 0x1f, /* pio0-4 */ | 364 | .pio_mask = 0x1f, /* pio0-4 */ |
441 | .udma_mask = ATA_UDMA6, | 365 | .udma_mask = ATA_UDMA6, |
442 | .port_ops = &ahci_ops, | 366 | .port_ops = &ahci_ops, |
@@ -447,7 +371,6 @@ static const struct ata_port_info ahci_port_info[] = { | |||
447 | AHCI_HFLAG_MV_PATA), | 371 | AHCI_HFLAG_MV_PATA), |
448 | .flags = ATA_FLAG_SATA | ATA_FLAG_NO_LEGACY | | 372 | .flags = ATA_FLAG_SATA | ATA_FLAG_NO_LEGACY | |
449 | ATA_FLAG_MMIO | ATA_FLAG_PIO_DMA, | 373 | ATA_FLAG_MMIO | ATA_FLAG_PIO_DMA, |
450 | .link_flags = AHCI_LFLAG_COMMON, | ||
451 | .pio_mask = 0x1f, /* pio0-4 */ | 374 | .pio_mask = 0x1f, /* pio0-4 */ |
452 | .udma_mask = ATA_UDMA6, | 375 | .udma_mask = ATA_UDMA6, |
453 | .port_ops = &ahci_ops, | 376 | .port_ops = &ahci_ops, |
@@ -457,7 +380,6 @@ static const struct ata_port_info ahci_port_info[] = { | |||
457 | AHCI_HFLAGS (AHCI_HFLAG_IGN_SERR_INTERNAL | | 380 | AHCI_HFLAGS (AHCI_HFLAG_IGN_SERR_INTERNAL | |
458 | AHCI_HFLAG_NO_PMP), | 381 | AHCI_HFLAG_NO_PMP), |
459 | .flags = AHCI_FLAG_COMMON, | 382 | .flags = AHCI_FLAG_COMMON, |
460 | .link_flags = AHCI_LFLAG_COMMON, | ||
461 | .pio_mask = 0x1f, /* pio0-4 */ | 383 | .pio_mask = 0x1f, /* pio0-4 */ |
462 | .udma_mask = ATA_UDMA6, | 384 | .udma_mask = ATA_UDMA6, |
463 | .port_ops = &ahci_ops, | 385 | .port_ops = &ahci_ops, |
@@ -1255,13 +1177,14 @@ static void ahci_fill_cmd_slot(struct ahci_port_priv *pp, unsigned int tag, | |||
1255 | 1177 | ||
1256 | static int ahci_kick_engine(struct ata_port *ap, int force_restart) | 1178 | static int ahci_kick_engine(struct ata_port *ap, int force_restart) |
1257 | { | 1179 | { |
1258 | void __iomem *port_mmio = ap->ioaddr.cmd_addr; | 1180 | void __iomem *port_mmio = ahci_port_base(ap); |
1259 | struct ahci_host_priv *hpriv = ap->host->private_data; | 1181 | struct ahci_host_priv *hpriv = ap->host->private_data; |
1182 | u8 status = readl(port_mmio + PORT_TFDATA) & 0xFF; | ||
1260 | u32 tmp; | 1183 | u32 tmp; |
1261 | int busy, rc; | 1184 | int busy, rc; |
1262 | 1185 | ||
1263 | /* do we need to kick the port? */ | 1186 | /* do we need to kick the port? */ |
1264 | busy = ahci_check_status(ap) & (ATA_BUSY | ATA_DRQ); | 1187 | busy = status & (ATA_BUSY | ATA_DRQ); |
1265 | if (!busy && !force_restart) | 1188 | if (!busy && !force_restart) |
1266 | return 0; | 1189 | return 0; |
1267 | 1190 | ||
@@ -1328,10 +1251,21 @@ static int ahci_exec_polled_cmd(struct ata_port *ap, int pmp, | |||
1328 | return 0; | 1251 | return 0; |
1329 | } | 1252 | } |
1330 | 1253 | ||
1331 | static int ahci_do_softreset(struct ata_link *link, unsigned int *class, | 1254 | static int ahci_check_ready(struct ata_link *link) |
1332 | int pmp, unsigned long deadline) | 1255 | { |
1256 | void __iomem *port_mmio = ahci_port_base(link->ap); | ||
1257 | u8 status = readl(port_mmio + PORT_TFDATA) & 0xFF; | ||
1258 | |||
1259 | if (!(status & ATA_BUSY)) | ||
1260 | return 1; | ||
1261 | return 0; | ||
1262 | } | ||
1263 | |||
1264 | static int ahci_softreset(struct ata_link *link, unsigned int *class, | ||
1265 | unsigned long deadline) | ||
1333 | { | 1266 | { |
1334 | struct ata_port *ap = link->ap; | 1267 | struct ata_port *ap = link->ap; |
1268 | int pmp = sata_srst_pmp(link); | ||
1335 | const char *reason = NULL; | 1269 | const char *reason = NULL; |
1336 | unsigned long now, msecs; | 1270 | unsigned long now, msecs; |
1337 | struct ata_taskfile tf; | 1271 | struct ata_taskfile tf; |
@@ -1339,12 +1273,6 @@ static int ahci_do_softreset(struct ata_link *link, unsigned int *class, | |||
1339 | 1273 | ||
1340 | DPRINTK("ENTER\n"); | 1274 | DPRINTK("ENTER\n"); |
1341 | 1275 | ||
1342 | if (ata_link_offline(link)) { | ||
1343 | DPRINTK("PHY reports no device\n"); | ||
1344 | *class = ATA_DEV_NONE; | ||
1345 | return 0; | ||
1346 | } | ||
1347 | |||
1348 | /* prepare for SRST (AHCI-1.1 10.4.1) */ | 1276 | /* prepare for SRST (AHCI-1.1 10.4.1) */ |
1349 | rc = ahci_kick_engine(ap, 1); | 1277 | rc = ahci_kick_engine(ap, 1); |
1350 | if (rc && rc != -EOPNOTSUPP) | 1278 | if (rc && rc != -EOPNOTSUPP) |
@@ -1374,10 +1302,8 @@ static int ahci_do_softreset(struct ata_link *link, unsigned int *class, | |||
1374 | tf.ctl &= ~ATA_SRST; | 1302 | tf.ctl &= ~ATA_SRST; |
1375 | ahci_exec_polled_cmd(ap, pmp, &tf, 0, 0, 0); | 1303 | ahci_exec_polled_cmd(ap, pmp, &tf, 0, 0, 0); |
1376 | 1304 | ||
1377 | /* wait a while before checking status */ | 1305 | /* wait for link to become ready */ |
1378 | ata_wait_after_reset(ap, deadline); | 1306 | rc = ata_wait_after_reset(link, deadline, ahci_check_ready); |
1379 | |||
1380 | rc = ata_wait_ready(ap, deadline); | ||
1381 | /* link occupied, -ENODEV too is an error */ | 1307 | /* link occupied, -ENODEV too is an error */ |
1382 | if (rc) { | 1308 | if (rc) { |
1383 | reason = "device not ready"; | 1309 | reason = "device not ready"; |
@@ -1393,24 +1319,15 @@ static int ahci_do_softreset(struct ata_link *link, unsigned int *class, | |||
1393 | return rc; | 1319 | return rc; |
1394 | } | 1320 | } |
1395 | 1321 | ||
1396 | static int ahci_softreset(struct ata_link *link, unsigned int *class, | ||
1397 | unsigned long deadline) | ||
1398 | { | ||
1399 | int pmp = 0; | ||
1400 | |||
1401 | if (link->ap->flags & ATA_FLAG_PMP) | ||
1402 | pmp = SATA_PMP_CTRL_PORT; | ||
1403 | |||
1404 | return ahci_do_softreset(link, class, pmp, deadline); | ||
1405 | } | ||
1406 | |||
1407 | static int ahci_hardreset(struct ata_link *link, unsigned int *class, | 1322 | static int ahci_hardreset(struct ata_link *link, unsigned int *class, |
1408 | unsigned long deadline) | 1323 | unsigned long deadline) |
1409 | { | 1324 | { |
1325 | const unsigned long *timing = sata_ehc_deb_timing(&link->eh_context); | ||
1410 | struct ata_port *ap = link->ap; | 1326 | struct ata_port *ap = link->ap; |
1411 | struct ahci_port_priv *pp = ap->private_data; | 1327 | struct ahci_port_priv *pp = ap->private_data; |
1412 | u8 *d2h_fis = pp->rx_fis + RX_FIS_D2H_REG; | 1328 | u8 *d2h_fis = pp->rx_fis + RX_FIS_D2H_REG; |
1413 | struct ata_taskfile tf; | 1329 | struct ata_taskfile tf; |
1330 | bool online; | ||
1414 | int rc; | 1331 | int rc; |
1415 | 1332 | ||
1416 | DPRINTK("ENTER\n"); | 1333 | DPRINTK("ENTER\n"); |
@@ -1422,14 +1339,13 @@ static int ahci_hardreset(struct ata_link *link, unsigned int *class, | |||
1422 | tf.command = 0x80; | 1339 | tf.command = 0x80; |
1423 | ata_tf_to_fis(&tf, 0, 0, d2h_fis); | 1340 | ata_tf_to_fis(&tf, 0, 0, d2h_fis); |
1424 | 1341 | ||
1425 | rc = sata_std_hardreset(link, class, deadline); | 1342 | rc = sata_link_hardreset(link, timing, deadline, &online, |
1343 | ahci_check_ready); | ||
1426 | 1344 | ||
1427 | ahci_start_engine(ap); | 1345 | ahci_start_engine(ap); |
1428 | 1346 | ||
1429 | if (rc == 0 && ata_link_online(link)) | 1347 | if (online) |
1430 | *class = ahci_dev_classify(ap); | 1348 | *class = ahci_dev_classify(ap); |
1431 | if (rc != -EAGAIN && *class == ATA_DEV_UNKNOWN) | ||
1432 | *class = ATA_DEV_NONE; | ||
1433 | 1349 | ||
1434 | DPRINTK("EXIT, rc=%d, class=%u\n", rc, *class); | 1350 | DPRINTK("EXIT, rc=%d, class=%u\n", rc, *class); |
1435 | return rc; | 1351 | return rc; |
@@ -1439,7 +1355,7 @@ static int ahci_vt8251_hardreset(struct ata_link *link, unsigned int *class, | |||
1439 | unsigned long deadline) | 1355 | unsigned long deadline) |
1440 | { | 1356 | { |
1441 | struct ata_port *ap = link->ap; | 1357 | struct ata_port *ap = link->ap; |
1442 | u32 serror; | 1358 | bool online; |
1443 | int rc; | 1359 | int rc; |
1444 | 1360 | ||
1445 | DPRINTK("ENTER\n"); | 1361 | DPRINTK("ENTER\n"); |
@@ -1447,11 +1363,7 @@ static int ahci_vt8251_hardreset(struct ata_link *link, unsigned int *class, | |||
1447 | ahci_stop_engine(ap); | 1363 | ahci_stop_engine(ap); |
1448 | 1364 | ||
1449 | rc = sata_link_hardreset(link, sata_ehc_deb_timing(&link->eh_context), | 1365 | rc = sata_link_hardreset(link, sata_ehc_deb_timing(&link->eh_context), |
1450 | deadline); | 1366 | deadline, &online, NULL); |
1451 | |||
1452 | /* vt8251 needs SError cleared for the port to operate */ | ||
1453 | ahci_scr_read(ap, SCR_ERROR, &serror); | ||
1454 | ahci_scr_write(ap, SCR_ERROR, serror); | ||
1455 | 1367 | ||
1456 | ahci_start_engine(ap); | 1368 | ahci_start_engine(ap); |
1457 | 1369 | ||
@@ -1460,7 +1372,7 @@ static int ahci_vt8251_hardreset(struct ata_link *link, unsigned int *class, | |||
1460 | /* vt8251 doesn't clear BSY on signature FIS reception, | 1372 | /* vt8251 doesn't clear BSY on signature FIS reception, |
1461 | * request follow-up softreset. | 1373 | * request follow-up softreset. |
1462 | */ | 1374 | */ |
1463 | return rc ?: -EAGAIN; | 1375 | return online ? -EAGAIN : rc; |
1464 | } | 1376 | } |
1465 | 1377 | ||
1466 | static int ahci_p5wdh_hardreset(struct ata_link *link, unsigned int *class, | 1378 | static int ahci_p5wdh_hardreset(struct ata_link *link, unsigned int *class, |
@@ -1470,6 +1382,7 @@ static int ahci_p5wdh_hardreset(struct ata_link *link, unsigned int *class, | |||
1470 | struct ahci_port_priv *pp = ap->private_data; | 1382 | struct ahci_port_priv *pp = ap->private_data; |
1471 | u8 *d2h_fis = pp->rx_fis + RX_FIS_D2H_REG; | 1383 | u8 *d2h_fis = pp->rx_fis + RX_FIS_D2H_REG; |
1472 | struct ata_taskfile tf; | 1384 | struct ata_taskfile tf; |
1385 | bool online; | ||
1473 | int rc; | 1386 | int rc; |
1474 | 1387 | ||
1475 | ahci_stop_engine(ap); | 1388 | ahci_stop_engine(ap); |
@@ -1480,16 +1393,10 @@ static int ahci_p5wdh_hardreset(struct ata_link *link, unsigned int *class, | |||
1480 | ata_tf_to_fis(&tf, 0, 0, d2h_fis); | 1393 | ata_tf_to_fis(&tf, 0, 0, d2h_fis); |
1481 | 1394 | ||
1482 | rc = sata_link_hardreset(link, sata_ehc_deb_timing(&link->eh_context), | 1395 | rc = sata_link_hardreset(link, sata_ehc_deb_timing(&link->eh_context), |
1483 | deadline); | 1396 | deadline, &online, NULL); |
1484 | 1397 | ||
1485 | ahci_start_engine(ap); | 1398 | ahci_start_engine(ap); |
1486 | 1399 | ||
1487 | if (rc || ata_link_offline(link)) | ||
1488 | return rc; | ||
1489 | |||
1490 | /* spec mandates ">= 2ms" before checking status */ | ||
1491 | msleep(150); | ||
1492 | |||
1493 | /* The pseudo configuration device on SIMG4726 attached to | 1400 | /* The pseudo configuration device on SIMG4726 attached to |
1494 | * ASUS P5W-DH Deluxe doesn't send signature FIS after | 1401 | * ASUS P5W-DH Deluxe doesn't send signature FIS after |
1495 | * hardreset if no device is attached to the first downstream | 1402 | * hardreset if no device is attached to the first downstream |
@@ -1503,11 +1410,13 @@ static int ahci_p5wdh_hardreset(struct ata_link *link, unsigned int *class, | |||
1503 | * have to be reset again. For most cases, this should | 1410 | * have to be reset again. For most cases, this should |
1504 | * suffice while making probing snappish enough. | 1411 | * suffice while making probing snappish enough. |
1505 | */ | 1412 | */ |
1506 | rc = ata_wait_ready(ap, jiffies + 2 * HZ); | 1413 | if (online) { |
1507 | if (rc) | 1414 | rc = ata_wait_after_reset(link, jiffies + 2 * HZ, |
1508 | ahci_kick_engine(ap, 0); | 1415 | ahci_check_ready); |
1509 | 1416 | if (rc) | |
1510 | return 0; | 1417 | ahci_kick_engine(ap, 0); |
1418 | } | ||
1419 | return rc; | ||
1511 | } | 1420 | } |
1512 | 1421 | ||
1513 | static void ahci_postreset(struct ata_link *link, unsigned int *class) | 1422 | static void ahci_postreset(struct ata_link *link, unsigned int *class) |
@@ -1530,27 +1439,6 @@ static void ahci_postreset(struct ata_link *link, unsigned int *class) | |||
1530 | } | 1439 | } |
1531 | } | 1440 | } |
1532 | 1441 | ||
1533 | static int ahci_pmp_softreset(struct ata_link *link, unsigned int *class, | ||
1534 | unsigned long deadline) | ||
1535 | { | ||
1536 | return ahci_do_softreset(link, class, link->pmp, deadline); | ||
1537 | } | ||
1538 | |||
1539 | static u8 ahci_check_status(struct ata_port *ap) | ||
1540 | { | ||
1541 | void __iomem *mmio = ap->ioaddr.cmd_addr; | ||
1542 | |||
1543 | return readl(mmio + PORT_TFDATA) & 0xFF; | ||
1544 | } | ||
1545 | |||
1546 | static void ahci_tf_read(struct ata_port *ap, struct ata_taskfile *tf) | ||
1547 | { | ||
1548 | struct ahci_port_priv *pp = ap->private_data; | ||
1549 | u8 *d2h_fis = pp->rx_fis + RX_FIS_D2H_REG; | ||
1550 | |||
1551 | ata_tf_from_fis(d2h_fis, tf); | ||
1552 | } | ||
1553 | |||
1554 | static unsigned int ahci_fill_sg(struct ata_queued_cmd *qc, void *cmd_tbl) | 1442 | static unsigned int ahci_fill_sg(struct ata_queued_cmd *qc, void *cmd_tbl) |
1555 | { | 1443 | { |
1556 | struct scatterlist *sg; | 1444 | struct scatterlist *sg; |
@@ -1663,27 +1551,27 @@ static void ahci_error_intr(struct ata_port *ap, u32 irq_stat) | |||
1663 | u32 *unk = (u32 *)(pp->rx_fis + RX_FIS_UNK); | 1551 | u32 *unk = (u32 *)(pp->rx_fis + RX_FIS_UNK); |
1664 | 1552 | ||
1665 | active_ehi->err_mask |= AC_ERR_HSM; | 1553 | active_ehi->err_mask |= AC_ERR_HSM; |
1666 | active_ehi->action |= ATA_EH_SOFTRESET; | 1554 | active_ehi->action |= ATA_EH_RESET; |
1667 | ata_ehi_push_desc(active_ehi, | 1555 | ata_ehi_push_desc(active_ehi, |
1668 | "unknown FIS %08x %08x %08x %08x" , | 1556 | "unknown FIS %08x %08x %08x %08x" , |
1669 | unk[0], unk[1], unk[2], unk[3]); | 1557 | unk[0], unk[1], unk[2], unk[3]); |
1670 | } | 1558 | } |
1671 | 1559 | ||
1672 | if (ap->nr_pmp_links && (irq_stat & PORT_IRQ_BAD_PMP)) { | 1560 | if (sata_pmp_attached(ap) && (irq_stat & PORT_IRQ_BAD_PMP)) { |
1673 | active_ehi->err_mask |= AC_ERR_HSM; | 1561 | active_ehi->err_mask |= AC_ERR_HSM; |
1674 | active_ehi->action |= ATA_EH_SOFTRESET; | 1562 | active_ehi->action |= ATA_EH_RESET; |
1675 | ata_ehi_push_desc(active_ehi, "incorrect PMP"); | 1563 | ata_ehi_push_desc(active_ehi, "incorrect PMP"); |
1676 | } | 1564 | } |
1677 | 1565 | ||
1678 | if (irq_stat & (PORT_IRQ_HBUS_ERR | PORT_IRQ_HBUS_DATA_ERR)) { | 1566 | if (irq_stat & (PORT_IRQ_HBUS_ERR | PORT_IRQ_HBUS_DATA_ERR)) { |
1679 | host_ehi->err_mask |= AC_ERR_HOST_BUS; | 1567 | host_ehi->err_mask |= AC_ERR_HOST_BUS; |
1680 | host_ehi->action |= ATA_EH_SOFTRESET; | 1568 | host_ehi->action |= ATA_EH_RESET; |
1681 | ata_ehi_push_desc(host_ehi, "host bus error"); | 1569 | ata_ehi_push_desc(host_ehi, "host bus error"); |
1682 | } | 1570 | } |
1683 | 1571 | ||
1684 | if (irq_stat & PORT_IRQ_IF_ERR) { | 1572 | if (irq_stat & PORT_IRQ_IF_ERR) { |
1685 | host_ehi->err_mask |= AC_ERR_ATA_BUS; | 1573 | host_ehi->err_mask |= AC_ERR_ATA_BUS; |
1686 | host_ehi->action |= ATA_EH_SOFTRESET; | 1574 | host_ehi->action |= ATA_EH_RESET; |
1687 | ata_ehi_push_desc(host_ehi, "interface fatal error"); | 1575 | ata_ehi_push_desc(host_ehi, "interface fatal error"); |
1688 | } | 1576 | } |
1689 | 1577 | ||
@@ -1704,7 +1592,7 @@ static void ahci_error_intr(struct ata_port *ap, u32 irq_stat) | |||
1704 | 1592 | ||
1705 | static void ahci_port_intr(struct ata_port *ap) | 1593 | static void ahci_port_intr(struct ata_port *ap) |
1706 | { | 1594 | { |
1707 | void __iomem *port_mmio = ap->ioaddr.cmd_addr; | 1595 | void __iomem *port_mmio = ahci_port_base(ap); |
1708 | struct ata_eh_info *ehi = &ap->link.eh_info; | 1596 | struct ata_eh_info *ehi = &ap->link.eh_info; |
1709 | struct ahci_port_priv *pp = ap->private_data; | 1597 | struct ahci_port_priv *pp = ap->private_data; |
1710 | struct ahci_host_priv *hpriv = ap->host->private_data; | 1598 | struct ahci_host_priv *hpriv = ap->host->private_data; |
@@ -1766,21 +1654,16 @@ static void ahci_port_intr(struct ata_port *ap) | |||
1766 | else | 1654 | else |
1767 | qc_active = readl(port_mmio + PORT_CMD_ISSUE); | 1655 | qc_active = readl(port_mmio + PORT_CMD_ISSUE); |
1768 | 1656 | ||
1769 | rc = ata_qc_complete_multiple(ap, qc_active, NULL); | 1657 | rc = ata_qc_complete_multiple(ap, qc_active); |
1770 | 1658 | ||
1771 | /* while resetting, invalid completions are expected */ | 1659 | /* while resetting, invalid completions are expected */ |
1772 | if (unlikely(rc < 0 && !resetting)) { | 1660 | if (unlikely(rc < 0 && !resetting)) { |
1773 | ehi->err_mask |= AC_ERR_HSM; | 1661 | ehi->err_mask |= AC_ERR_HSM; |
1774 | ehi->action |= ATA_EH_SOFTRESET; | 1662 | ehi->action |= ATA_EH_RESET; |
1775 | ata_port_freeze(ap); | 1663 | ata_port_freeze(ap); |
1776 | } | 1664 | } |
1777 | } | 1665 | } |
1778 | 1666 | ||
1779 | static void ahci_irq_clear(struct ata_port *ap) | ||
1780 | { | ||
1781 | /* TODO */ | ||
1782 | } | ||
1783 | |||
1784 | static irqreturn_t ahci_interrupt(int irq, void *dev_instance) | 1667 | static irqreturn_t ahci_interrupt(int irq, void *dev_instance) |
1785 | { | 1668 | { |
1786 | struct ata_host *host = dev_instance; | 1669 | struct ata_host *host = dev_instance; |
@@ -1854,6 +1737,15 @@ static unsigned int ahci_qc_issue(struct ata_queued_cmd *qc) | |||
1854 | return 0; | 1737 | return 0; |
1855 | } | 1738 | } |
1856 | 1739 | ||
1740 | static bool ahci_qc_fill_rtf(struct ata_queued_cmd *qc) | ||
1741 | { | ||
1742 | struct ahci_port_priv *pp = qc->ap->private_data; | ||
1743 | u8 *d2h_fis = pp->rx_fis + RX_FIS_D2H_REG; | ||
1744 | |||
1745 | ata_tf_from_fis(d2h_fis, &qc->result_tf); | ||
1746 | return true; | ||
1747 | } | ||
1748 | |||
1857 | static void ahci_freeze(struct ata_port *ap) | 1749 | static void ahci_freeze(struct ata_port *ap) |
1858 | { | 1750 | { |
1859 | void __iomem *port_mmio = ahci_port_base(ap); | 1751 | void __iomem *port_mmio = ahci_port_base(ap); |
@@ -1886,37 +1778,7 @@ static void ahci_error_handler(struct ata_port *ap) | |||
1886 | ahci_start_engine(ap); | 1778 | ahci_start_engine(ap); |
1887 | } | 1779 | } |
1888 | 1780 | ||
1889 | /* perform recovery */ | 1781 | sata_pmp_error_handler(ap); |
1890 | sata_pmp_do_eh(ap, ata_std_prereset, ahci_softreset, | ||
1891 | ahci_hardreset, ahci_postreset, | ||
1892 | sata_pmp_std_prereset, ahci_pmp_softreset, | ||
1893 | sata_pmp_std_hardreset, sata_pmp_std_postreset); | ||
1894 | } | ||
1895 | |||
1896 | static void ahci_vt8251_error_handler(struct ata_port *ap) | ||
1897 | { | ||
1898 | if (!(ap->pflags & ATA_PFLAG_FROZEN)) { | ||
1899 | /* restart engine */ | ||
1900 | ahci_stop_engine(ap); | ||
1901 | ahci_start_engine(ap); | ||
1902 | } | ||
1903 | |||
1904 | /* perform recovery */ | ||
1905 | ata_do_eh(ap, ata_std_prereset, ahci_softreset, ahci_vt8251_hardreset, | ||
1906 | ahci_postreset); | ||
1907 | } | ||
1908 | |||
1909 | static void ahci_p5wdh_error_handler(struct ata_port *ap) | ||
1910 | { | ||
1911 | if (!(ap->pflags & ATA_PFLAG_FROZEN)) { | ||
1912 | /* restart engine */ | ||
1913 | ahci_stop_engine(ap); | ||
1914 | ahci_start_engine(ap); | ||
1915 | } | ||
1916 | |||
1917 | /* perform recovery */ | ||
1918 | ata_do_eh(ap, ata_std_prereset, ahci_softreset, ahci_p5wdh_hardreset, | ||
1919 | ahci_postreset); | ||
1920 | } | 1782 | } |
1921 | 1783 | ||
1922 | static void ahci_post_internal_cmd(struct ata_queued_cmd *qc) | 1784 | static void ahci_post_internal_cmd(struct ata_queued_cmd *qc) |
@@ -1961,7 +1823,7 @@ static int ahci_port_resume(struct ata_port *ap) | |||
1961 | ahci_power_up(ap); | 1823 | ahci_power_up(ap); |
1962 | ahci_start_port(ap); | 1824 | ahci_start_port(ap); |
1963 | 1825 | ||
1964 | if (ap->nr_pmp_links) | 1826 | if (sata_pmp_attached(ap)) |
1965 | ahci_pmp_attach(ap); | 1827 | ahci_pmp_attach(ap); |
1966 | else | 1828 | else |
1967 | ahci_pmp_detach(ap); | 1829 | ahci_pmp_detach(ap); |
@@ -2324,7 +2186,6 @@ static int ahci_init_one(struct pci_dev *pdev, const struct pci_device_id *ent) | |||
2324 | 2186 | ||
2325 | for (i = 0; i < host->n_ports; i++) { | 2187 | for (i = 0; i < host->n_ports; i++) { |
2326 | struct ata_port *ap = host->ports[i]; | 2188 | struct ata_port *ap = host->ports[i]; |
2327 | void __iomem *port_mmio = ahci_port_base(ap); | ||
2328 | 2189 | ||
2329 | ata_port_pbar_desc(ap, AHCI_PCI_BAR, -1, "abar"); | 2190 | ata_port_pbar_desc(ap, AHCI_PCI_BAR, -1, "abar"); |
2330 | ata_port_pbar_desc(ap, AHCI_PCI_BAR, | 2191 | ata_port_pbar_desc(ap, AHCI_PCI_BAR, |
@@ -2333,12 +2194,8 @@ static int ahci_init_one(struct pci_dev *pdev, const struct pci_device_id *ent) | |||
2333 | /* set initial link pm policy */ | 2194 | /* set initial link pm policy */ |
2334 | ap->pm_policy = NOT_AVAILABLE; | 2195 | ap->pm_policy = NOT_AVAILABLE; |
2335 | 2196 | ||
2336 | /* standard SATA port setup */ | ||
2337 | if (hpriv->port_map & (1 << i)) | ||
2338 | ap->ioaddr.cmd_addr = port_mmio; | ||
2339 | |||
2340 | /* disabled/not-implemented port */ | 2197 | /* disabled/not-implemented port */ |
2341 | else | 2198 | if (!(hpriv->port_map & (1 << i))) |
2342 | ap->ops = &ata_dummy_port_ops; | 2199 | ap->ops = &ata_dummy_port_ops; |
2343 | } | 2200 | } |
2344 | 2201 | ||
diff --git a/drivers/ata/ata_generic.c b/drivers/ata/ata_generic.c index 20534202fc79..47aeccd52fa9 100644 --- a/drivers/ata/ata_generic.c +++ b/drivers/ata/ata_generic.c | |||
@@ -95,53 +95,13 @@ static int generic_set_mode(struct ata_link *link, struct ata_device **unused) | |||
95 | } | 95 | } |
96 | 96 | ||
97 | static struct scsi_host_template generic_sht = { | 97 | static struct scsi_host_template generic_sht = { |
98 | .module = THIS_MODULE, | 98 | ATA_BMDMA_SHT(DRV_NAME), |
99 | .name = DRV_NAME, | ||
100 | .ioctl = ata_scsi_ioctl, | ||
101 | .queuecommand = ata_scsi_queuecmd, | ||
102 | .can_queue = ATA_DEF_QUEUE, | ||
103 | .this_id = ATA_SHT_THIS_ID, | ||
104 | .sg_tablesize = LIBATA_MAX_PRD, | ||
105 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
106 | .emulated = ATA_SHT_EMULATED, | ||
107 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
108 | .proc_name = DRV_NAME, | ||
109 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
110 | .slave_configure = ata_scsi_slave_config, | ||
111 | .slave_destroy = ata_scsi_slave_destroy, | ||
112 | .bios_param = ata_std_bios_param, | ||
113 | }; | 99 | }; |
114 | 100 | ||
115 | static struct ata_port_operations generic_port_ops = { | 101 | static struct ata_port_operations generic_port_ops = { |
116 | .set_mode = generic_set_mode, | 102 | .inherits = &ata_bmdma_port_ops, |
117 | |||
118 | .tf_load = ata_tf_load, | ||
119 | .tf_read = ata_tf_read, | ||
120 | .check_status = ata_check_status, | ||
121 | .exec_command = ata_exec_command, | ||
122 | .dev_select = ata_std_dev_select, | ||
123 | |||
124 | .bmdma_setup = ata_bmdma_setup, | ||
125 | .bmdma_start = ata_bmdma_start, | ||
126 | .bmdma_stop = ata_bmdma_stop, | ||
127 | .bmdma_status = ata_bmdma_status, | ||
128 | |||
129 | .data_xfer = ata_data_xfer, | ||
130 | |||
131 | .freeze = ata_bmdma_freeze, | ||
132 | .thaw = ata_bmdma_thaw, | ||
133 | .error_handler = ata_bmdma_error_handler, | ||
134 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
135 | .cable_detect = ata_cable_unknown, | 103 | .cable_detect = ata_cable_unknown, |
136 | 104 | .set_mode = generic_set_mode, | |
137 | .qc_prep = ata_qc_prep, | ||
138 | .qc_issue = ata_qc_issue_prot, | ||
139 | |||
140 | .irq_handler = ata_interrupt, | ||
141 | .irq_clear = ata_bmdma_irq_clear, | ||
142 | .irq_on = ata_irq_on, | ||
143 | |||
144 | .port_start = ata_sff_port_start, | ||
145 | }; | 105 | }; |
146 | 106 | ||
147 | static int all_generic_ide; /* Set to claim all devices */ | 107 | static int all_generic_ide; /* Set to claim all devices */ |
@@ -160,7 +120,6 @@ static int ata_generic_init_one(struct pci_dev *dev, const struct pci_device_id | |||
160 | { | 120 | { |
161 | u16 command; | 121 | u16 command; |
162 | static const struct ata_port_info info = { | 122 | static const struct ata_port_info info = { |
163 | .sht = &generic_sht, | ||
164 | .flags = ATA_FLAG_SLAVE_POSS, | 123 | .flags = ATA_FLAG_SLAVE_POSS, |
165 | .pio_mask = 0x1f, | 124 | .pio_mask = 0x1f, |
166 | .mwdma_mask = 0x07, | 125 | .mwdma_mask = 0x07, |
@@ -191,9 +150,9 @@ static int ata_generic_init_one(struct pci_dev *dev, const struct pci_device_id | |||
191 | return -ENODEV; | 150 | return -ENODEV; |
192 | 151 | ||
193 | if (dev->vendor == PCI_VENDOR_ID_AL) | 152 | if (dev->vendor == PCI_VENDOR_ID_AL) |
194 | ata_pci_clear_simplex(dev); | 153 | ata_pci_bmdma_clear_simplex(dev); |
195 | 154 | ||
196 | return ata_pci_init_one(dev, ppi); | 155 | return ata_pci_sff_init_one(dev, ppi, &generic_sht, NULL); |
197 | } | 156 | } |
198 | 157 | ||
199 | static struct pci_device_id ata_generic[] = { | 158 | static struct pci_device_id ata_generic[] = { |
diff --git a/drivers/ata/ata_piix.c b/drivers/ata/ata_piix.c index fae8404254c0..b7c38eeb498f 100644 --- a/drivers/ata/ata_piix.c +++ b/drivers/ata/ata_piix.c | |||
@@ -100,13 +100,11 @@ enum { | |||
100 | PIIX_IOCFG = 0x54, /* IDE I/O configuration register */ | 100 | PIIX_IOCFG = 0x54, /* IDE I/O configuration register */ |
101 | ICH5_PMR = 0x90, /* port mapping register */ | 101 | ICH5_PMR = 0x90, /* port mapping register */ |
102 | ICH5_PCS = 0x92, /* port control and status */ | 102 | ICH5_PCS = 0x92, /* port control and status */ |
103 | PIIX_SCC = 0x0A, /* sub-class code register */ | ||
104 | PIIX_SIDPR_BAR = 5, | 103 | PIIX_SIDPR_BAR = 5, |
105 | PIIX_SIDPR_LEN = 16, | 104 | PIIX_SIDPR_LEN = 16, |
106 | PIIX_SIDPR_IDX = 0, | 105 | PIIX_SIDPR_IDX = 0, |
107 | PIIX_SIDPR_DATA = 4, | 106 | PIIX_SIDPR_DATA = 4, |
108 | 107 | ||
109 | PIIX_FLAG_AHCI = (1 << 27), /* AHCI possible */ | ||
110 | PIIX_FLAG_CHECKINTR = (1 << 28), /* make sure PCI INTx enabled */ | 108 | PIIX_FLAG_CHECKINTR = (1 << 28), /* make sure PCI INTx enabled */ |
111 | PIIX_FLAG_SIDPR = (1 << 29), /* SATA idx/data pair regs */ | 109 | PIIX_FLAG_SIDPR = (1 << 29), /* SATA idx/data pair regs */ |
112 | 110 | ||
@@ -140,12 +138,11 @@ enum piix_controller_ids { | |||
140 | ich_pata_100, /* ICH up to UDMA 100 */ | 138 | ich_pata_100, /* ICH up to UDMA 100 */ |
141 | ich5_sata, | 139 | ich5_sata, |
142 | ich6_sata, | 140 | ich6_sata, |
143 | ich6_sata_ahci, | 141 | ich6m_sata, |
144 | ich6m_sata_ahci, | 142 | ich8_sata, |
145 | ich8_sata_ahci, | ||
146 | ich8_2port_sata, | 143 | ich8_2port_sata, |
147 | ich8m_apple_sata_ahci, /* locks up on second port enable */ | 144 | ich8m_apple_sata, /* locks up on second port enable */ |
148 | tolapai_sata_ahci, | 145 | tolapai_sata, |
149 | piix_pata_vmw, /* PIIX4 for VMware, spurious DMA_ERR */ | 146 | piix_pata_vmw, /* PIIX4 for VMware, spurious DMA_ERR */ |
150 | }; | 147 | }; |
151 | 148 | ||
@@ -162,7 +159,7 @@ struct piix_host_priv { | |||
162 | 159 | ||
163 | static int piix_init_one(struct pci_dev *pdev, | 160 | static int piix_init_one(struct pci_dev *pdev, |
164 | const struct pci_device_id *ent); | 161 | const struct pci_device_id *ent); |
165 | static void piix_pata_error_handler(struct ata_port *ap); | 162 | static int piix_pata_prereset(struct ata_link *link, unsigned long deadline); |
166 | static void piix_set_piomode(struct ata_port *ap, struct ata_device *adev); | 163 | static void piix_set_piomode(struct ata_port *ap, struct ata_device *adev); |
167 | static void piix_set_dmamode(struct ata_port *ap, struct ata_device *adev); | 164 | static void piix_set_dmamode(struct ata_port *ap, struct ata_device *adev); |
168 | static void ich_set_dmamode(struct ata_port *ap, struct ata_device *adev); | 165 | static void ich_set_dmamode(struct ata_port *ap, struct ata_device *adev); |
@@ -170,7 +167,6 @@ static int ich_pata_cable_detect(struct ata_port *ap); | |||
170 | static u8 piix_vmw_bmdma_status(struct ata_port *ap); | 167 | static u8 piix_vmw_bmdma_status(struct ata_port *ap); |
171 | static int piix_sidpr_scr_read(struct ata_port *ap, unsigned int reg, u32 *val); | 168 | static int piix_sidpr_scr_read(struct ata_port *ap, unsigned int reg, u32 *val); |
172 | static int piix_sidpr_scr_write(struct ata_port *ap, unsigned int reg, u32 val); | 169 | static int piix_sidpr_scr_write(struct ata_port *ap, unsigned int reg, u32 val); |
173 | static void piix_sidpr_error_handler(struct ata_port *ap); | ||
174 | #ifdef CONFIG_PM | 170 | #ifdef CONFIG_PM |
175 | static int piix_pci_device_suspend(struct pci_dev *pdev, pm_message_t mesg); | 171 | static int piix_pci_device_suspend(struct pci_dev *pdev, pm_message_t mesg); |
176 | static int piix_pci_device_resume(struct pci_dev *pdev); | 172 | static int piix_pci_device_resume(struct pci_dev *pdev); |
@@ -236,25 +232,27 @@ static const struct pci_device_id piix_pci_tbl[] = { | |||
236 | /* 82801FB/FW (ICH6/ICH6W) */ | 232 | /* 82801FB/FW (ICH6/ICH6W) */ |
237 | { 0x8086, 0x2651, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich6_sata }, | 233 | { 0x8086, 0x2651, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich6_sata }, |
238 | /* 82801FR/FRW (ICH6R/ICH6RW) */ | 234 | /* 82801FR/FRW (ICH6R/ICH6RW) */ |
239 | { 0x8086, 0x2652, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich6_sata_ahci }, | 235 | { 0x8086, 0x2652, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich6_sata }, |
240 | /* 82801FBM ICH6M (ICH6R with only port 0 and 2 implemented) */ | 236 | /* 82801FBM ICH6M (ICH6R with only port 0 and 2 implemented). |
241 | { 0x8086, 0x2653, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich6m_sata_ahci }, | 237 | * Attach iff the controller is in IDE mode. */ |
238 | { 0x8086, 0x2653, PCI_ANY_ID, PCI_ANY_ID, | ||
239 | PCI_CLASS_STORAGE_IDE << 8, 0xffff00, ich6m_sata }, | ||
242 | /* 82801GB/GR/GH (ICH7, identical to ICH6) */ | 240 | /* 82801GB/GR/GH (ICH7, identical to ICH6) */ |
243 | { 0x8086, 0x27c0, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich6_sata_ahci }, | 241 | { 0x8086, 0x27c0, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich6_sata }, |
244 | /* 2801GBM/GHM (ICH7M, identical to ICH6M) */ | 242 | /* 2801GBM/GHM (ICH7M, identical to ICH6M) */ |
245 | { 0x8086, 0x27c4, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich6m_sata_ahci }, | 243 | { 0x8086, 0x27c4, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich6m_sata }, |
246 | /* Enterprise Southbridge 2 (631xESB/632xESB) */ | 244 | /* Enterprise Southbridge 2 (631xESB/632xESB) */ |
247 | { 0x8086, 0x2680, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich6_sata_ahci }, | 245 | { 0x8086, 0x2680, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich6_sata }, |
248 | /* SATA Controller 1 IDE (ICH8) */ | 246 | /* SATA Controller 1 IDE (ICH8) */ |
249 | { 0x8086, 0x2820, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_sata_ahci }, | 247 | { 0x8086, 0x2820, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_sata }, |
250 | /* SATA Controller 2 IDE (ICH8) */ | 248 | /* SATA Controller 2 IDE (ICH8) */ |
251 | { 0x8086, 0x2825, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_2port_sata }, | 249 | { 0x8086, 0x2825, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_2port_sata }, |
252 | /* Mobile SATA Controller IDE (ICH8M) */ | 250 | /* Mobile SATA Controller IDE (ICH8M) */ |
253 | { 0x8086, 0x2828, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_sata_ahci }, | 251 | { 0x8086, 0x2828, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_sata }, |
254 | /* Mobile SATA Controller IDE (ICH8M), Apple */ | 252 | /* Mobile SATA Controller IDE (ICH8M), Apple */ |
255 | { 0x8086, 0x2828, 0x106b, 0x00a0, 0, 0, ich8m_apple_sata_ahci }, | 253 | { 0x8086, 0x2828, 0x106b, 0x00a0, 0, 0, ich8m_apple_sata }, |
256 | /* SATA Controller IDE (ICH9) */ | 254 | /* SATA Controller IDE (ICH9) */ |
257 | { 0x8086, 0x2920, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_sata_ahci }, | 255 | { 0x8086, 0x2920, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_sata }, |
258 | /* SATA Controller IDE (ICH9) */ | 256 | /* SATA Controller IDE (ICH9) */ |
259 | { 0x8086, 0x2921, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_2port_sata }, | 257 | { 0x8086, 0x2921, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_2port_sata }, |
260 | /* SATA Controller IDE (ICH9) */ | 258 | /* SATA Controller IDE (ICH9) */ |
@@ -264,15 +262,15 @@ static const struct pci_device_id piix_pci_tbl[] = { | |||
264 | /* SATA Controller IDE (ICH9M) */ | 262 | /* SATA Controller IDE (ICH9M) */ |
265 | { 0x8086, 0x292d, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_2port_sata }, | 263 | { 0x8086, 0x292d, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_2port_sata }, |
266 | /* SATA Controller IDE (ICH9M) */ | 264 | /* SATA Controller IDE (ICH9M) */ |
267 | { 0x8086, 0x292e, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_sata_ahci }, | 265 | { 0x8086, 0x292e, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_sata }, |
268 | /* SATA Controller IDE (Tolapai) */ | 266 | /* SATA Controller IDE (Tolapai) */ |
269 | { 0x8086, 0x5028, PCI_ANY_ID, PCI_ANY_ID, 0, 0, tolapai_sata_ahci }, | 267 | { 0x8086, 0x5028, PCI_ANY_ID, PCI_ANY_ID, 0, 0, tolapai_sata }, |
270 | /* SATA Controller IDE (ICH10) */ | 268 | /* SATA Controller IDE (ICH10) */ |
271 | { 0x8086, 0x3a00, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_sata_ahci }, | 269 | { 0x8086, 0x3a00, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_sata }, |
272 | /* SATA Controller IDE (ICH10) */ | 270 | /* SATA Controller IDE (ICH10) */ |
273 | { 0x8086, 0x3a06, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_2port_sata }, | 271 | { 0x8086, 0x3a06, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_2port_sata }, |
274 | /* SATA Controller IDE (ICH10) */ | 272 | /* SATA Controller IDE (ICH10) */ |
275 | { 0x8086, 0x3a20, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_sata_ahci }, | 273 | { 0x8086, 0x3a20, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_sata }, |
276 | /* SATA Controller IDE (ICH10) */ | 274 | /* SATA Controller IDE (ICH10) */ |
277 | { 0x8086, 0x3a26, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_2port_sata }, | 275 | { 0x8086, 0x3a26, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_2port_sata }, |
278 | 276 | ||
@@ -291,170 +289,37 @@ static struct pci_driver piix_pci_driver = { | |||
291 | }; | 289 | }; |
292 | 290 | ||
293 | static struct scsi_host_template piix_sht = { | 291 | static struct scsi_host_template piix_sht = { |
294 | .module = THIS_MODULE, | 292 | ATA_BMDMA_SHT(DRV_NAME), |
295 | .name = DRV_NAME, | ||
296 | .ioctl = ata_scsi_ioctl, | ||
297 | .queuecommand = ata_scsi_queuecmd, | ||
298 | .can_queue = ATA_DEF_QUEUE, | ||
299 | .this_id = ATA_SHT_THIS_ID, | ||
300 | .sg_tablesize = LIBATA_MAX_PRD, | ||
301 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
302 | .emulated = ATA_SHT_EMULATED, | ||
303 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
304 | .proc_name = DRV_NAME, | ||
305 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
306 | .slave_configure = ata_scsi_slave_config, | ||
307 | .slave_destroy = ata_scsi_slave_destroy, | ||
308 | .bios_param = ata_std_bios_param, | ||
309 | }; | 293 | }; |
310 | 294 | ||
311 | static const struct ata_port_operations piix_pata_ops = { | 295 | static struct ata_port_operations piix_pata_ops = { |
296 | .inherits = &ata_bmdma_port_ops, | ||
297 | .cable_detect = ata_cable_40wire, | ||
312 | .set_piomode = piix_set_piomode, | 298 | .set_piomode = piix_set_piomode, |
313 | .set_dmamode = piix_set_dmamode, | 299 | .set_dmamode = piix_set_dmamode, |
314 | .mode_filter = ata_pci_default_filter, | 300 | .prereset = piix_pata_prereset, |
315 | 301 | }; | |
316 | .tf_load = ata_tf_load, | ||
317 | .tf_read = ata_tf_read, | ||
318 | .check_status = ata_check_status, | ||
319 | .exec_command = ata_exec_command, | ||
320 | .dev_select = ata_std_dev_select, | ||
321 | |||
322 | .bmdma_setup = ata_bmdma_setup, | ||
323 | .bmdma_start = ata_bmdma_start, | ||
324 | .bmdma_stop = ata_bmdma_stop, | ||
325 | .bmdma_status = ata_bmdma_status, | ||
326 | .qc_prep = ata_qc_prep, | ||
327 | .qc_issue = ata_qc_issue_prot, | ||
328 | .data_xfer = ata_data_xfer, | ||
329 | |||
330 | .freeze = ata_bmdma_freeze, | ||
331 | .thaw = ata_bmdma_thaw, | ||
332 | .error_handler = piix_pata_error_handler, | ||
333 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
334 | .cable_detect = ata_cable_40wire, | ||
335 | |||
336 | .irq_clear = ata_bmdma_irq_clear, | ||
337 | .irq_on = ata_irq_on, | ||
338 | 302 | ||
339 | .port_start = ata_port_start, | 303 | static struct ata_port_operations piix_vmw_ops = { |
304 | .inherits = &piix_pata_ops, | ||
305 | .bmdma_status = piix_vmw_bmdma_status, | ||
340 | }; | 306 | }; |
341 | 307 | ||
342 | static const struct ata_port_operations ich_pata_ops = { | 308 | static struct ata_port_operations ich_pata_ops = { |
343 | .set_piomode = piix_set_piomode, | 309 | .inherits = &piix_pata_ops, |
344 | .set_dmamode = ich_set_dmamode, | ||
345 | .mode_filter = ata_pci_default_filter, | ||
346 | |||
347 | .tf_load = ata_tf_load, | ||
348 | .tf_read = ata_tf_read, | ||
349 | .check_status = ata_check_status, | ||
350 | .exec_command = ata_exec_command, | ||
351 | .dev_select = ata_std_dev_select, | ||
352 | |||
353 | .bmdma_setup = ata_bmdma_setup, | ||
354 | .bmdma_start = ata_bmdma_start, | ||
355 | .bmdma_stop = ata_bmdma_stop, | ||
356 | .bmdma_status = ata_bmdma_status, | ||
357 | .qc_prep = ata_qc_prep, | ||
358 | .qc_issue = ata_qc_issue_prot, | ||
359 | .data_xfer = ata_data_xfer, | ||
360 | |||
361 | .freeze = ata_bmdma_freeze, | ||
362 | .thaw = ata_bmdma_thaw, | ||
363 | .error_handler = piix_pata_error_handler, | ||
364 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
365 | .cable_detect = ich_pata_cable_detect, | 310 | .cable_detect = ich_pata_cable_detect, |
366 | 311 | .set_dmamode = ich_set_dmamode, | |
367 | .irq_clear = ata_bmdma_irq_clear, | ||
368 | .irq_on = ata_irq_on, | ||
369 | |||
370 | .port_start = ata_port_start, | ||
371 | }; | 312 | }; |
372 | 313 | ||
373 | static const struct ata_port_operations piix_sata_ops = { | 314 | static struct ata_port_operations piix_sata_ops = { |
374 | .tf_load = ata_tf_load, | 315 | .inherits = &ata_bmdma_port_ops, |
375 | .tf_read = ata_tf_read, | ||
376 | .check_status = ata_check_status, | ||
377 | .exec_command = ata_exec_command, | ||
378 | .dev_select = ata_std_dev_select, | ||
379 | |||
380 | .bmdma_setup = ata_bmdma_setup, | ||
381 | .bmdma_start = ata_bmdma_start, | ||
382 | .bmdma_stop = ata_bmdma_stop, | ||
383 | .bmdma_status = ata_bmdma_status, | ||
384 | .qc_prep = ata_qc_prep, | ||
385 | .qc_issue = ata_qc_issue_prot, | ||
386 | .data_xfer = ata_data_xfer, | ||
387 | |||
388 | .freeze = ata_bmdma_freeze, | ||
389 | .thaw = ata_bmdma_thaw, | ||
390 | .error_handler = ata_bmdma_error_handler, | ||
391 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
392 | |||
393 | .irq_clear = ata_bmdma_irq_clear, | ||
394 | .irq_on = ata_irq_on, | ||
395 | |||
396 | .port_start = ata_port_start, | ||
397 | }; | 316 | }; |
398 | 317 | ||
399 | static const struct ata_port_operations piix_vmw_ops = { | 318 | static struct ata_port_operations piix_sidpr_sata_ops = { |
400 | .set_piomode = piix_set_piomode, | 319 | .inherits = &piix_sata_ops, |
401 | .set_dmamode = piix_set_dmamode, | 320 | .hardreset = sata_std_hardreset, |
402 | .mode_filter = ata_pci_default_filter, | ||
403 | |||
404 | .tf_load = ata_tf_load, | ||
405 | .tf_read = ata_tf_read, | ||
406 | .check_status = ata_check_status, | ||
407 | .exec_command = ata_exec_command, | ||
408 | .dev_select = ata_std_dev_select, | ||
409 | |||
410 | .bmdma_setup = ata_bmdma_setup, | ||
411 | .bmdma_start = ata_bmdma_start, | ||
412 | .bmdma_stop = ata_bmdma_stop, | ||
413 | .bmdma_status = piix_vmw_bmdma_status, | ||
414 | .qc_prep = ata_qc_prep, | ||
415 | .qc_issue = ata_qc_issue_prot, | ||
416 | .data_xfer = ata_data_xfer, | ||
417 | |||
418 | .freeze = ata_bmdma_freeze, | ||
419 | .thaw = ata_bmdma_thaw, | ||
420 | .error_handler = piix_pata_error_handler, | ||
421 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
422 | .cable_detect = ata_cable_40wire, | ||
423 | |||
424 | .irq_handler = ata_interrupt, | ||
425 | .irq_clear = ata_bmdma_irq_clear, | ||
426 | .irq_on = ata_irq_on, | ||
427 | |||
428 | .port_start = ata_port_start, | ||
429 | }; | ||
430 | |||
431 | static const struct ata_port_operations piix_sidpr_sata_ops = { | ||
432 | .tf_load = ata_tf_load, | ||
433 | .tf_read = ata_tf_read, | ||
434 | .check_status = ata_check_status, | ||
435 | .exec_command = ata_exec_command, | ||
436 | .dev_select = ata_std_dev_select, | ||
437 | |||
438 | .bmdma_setup = ata_bmdma_setup, | ||
439 | .bmdma_start = ata_bmdma_start, | ||
440 | .bmdma_stop = ata_bmdma_stop, | ||
441 | .bmdma_status = ata_bmdma_status, | ||
442 | .qc_prep = ata_qc_prep, | ||
443 | .qc_issue = ata_qc_issue_prot, | ||
444 | .data_xfer = ata_data_xfer, | ||
445 | |||
446 | .scr_read = piix_sidpr_scr_read, | 321 | .scr_read = piix_sidpr_scr_read, |
447 | .scr_write = piix_sidpr_scr_write, | 322 | .scr_write = piix_sidpr_scr_write, |
448 | |||
449 | .freeze = ata_bmdma_freeze, | ||
450 | .thaw = ata_bmdma_thaw, | ||
451 | .error_handler = piix_sidpr_error_handler, | ||
452 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
453 | |||
454 | .irq_clear = ata_bmdma_irq_clear, | ||
455 | .irq_on = ata_irq_on, | ||
456 | |||
457 | .port_start = ata_port_start, | ||
458 | }; | 323 | }; |
459 | 324 | ||
460 | static const struct piix_map_db ich5_map_db = { | 325 | static const struct piix_map_db ich5_map_db = { |
@@ -553,12 +418,11 @@ static const struct piix_map_db tolapai_map_db = { | |||
553 | static const struct piix_map_db *piix_map_db_table[] = { | 418 | static const struct piix_map_db *piix_map_db_table[] = { |
554 | [ich5_sata] = &ich5_map_db, | 419 | [ich5_sata] = &ich5_map_db, |
555 | [ich6_sata] = &ich6_map_db, | 420 | [ich6_sata] = &ich6_map_db, |
556 | [ich6_sata_ahci] = &ich6_map_db, | 421 | [ich6m_sata] = &ich6m_map_db, |
557 | [ich6m_sata_ahci] = &ich6m_map_db, | 422 | [ich8_sata] = &ich8_map_db, |
558 | [ich8_sata_ahci] = &ich8_map_db, | ||
559 | [ich8_2port_sata] = &ich8_2port_map_db, | 423 | [ich8_2port_sata] = &ich8_2port_map_db, |
560 | [ich8m_apple_sata_ahci] = &ich8m_apple_map_db, | 424 | [ich8m_apple_sata] = &ich8m_apple_map_db, |
561 | [tolapai_sata_ahci] = &tolapai_map_db, | 425 | [tolapai_sata] = &tolapai_map_db, |
562 | }; | 426 | }; |
563 | 427 | ||
564 | static struct ata_port_info piix_port_info[] = { | 428 | static struct ata_port_info piix_port_info[] = { |
@@ -624,28 +488,18 @@ static struct ata_port_info piix_port_info[] = { | |||
624 | .port_ops = &piix_sata_ops, | 488 | .port_ops = &piix_sata_ops, |
625 | }, | 489 | }, |
626 | 490 | ||
627 | [ich6_sata_ahci] = | 491 | [ich6m_sata] = |
628 | { | ||
629 | .flags = PIIX_SATA_FLAGS | PIIX_FLAG_AHCI, | ||
630 | .pio_mask = 0x1f, /* pio0-4 */ | ||
631 | .mwdma_mask = 0x07, /* mwdma0-2 */ | ||
632 | .udma_mask = ATA_UDMA6, | ||
633 | .port_ops = &piix_sata_ops, | ||
634 | }, | ||
635 | |||
636 | [ich6m_sata_ahci] = | ||
637 | { | 492 | { |
638 | .flags = PIIX_SATA_FLAGS | PIIX_FLAG_AHCI, | 493 | .flags = PIIX_SATA_FLAGS, |
639 | .pio_mask = 0x1f, /* pio0-4 */ | 494 | .pio_mask = 0x1f, /* pio0-4 */ |
640 | .mwdma_mask = 0x07, /* mwdma0-2 */ | 495 | .mwdma_mask = 0x07, /* mwdma0-2 */ |
641 | .udma_mask = ATA_UDMA6, | 496 | .udma_mask = ATA_UDMA6, |
642 | .port_ops = &piix_sata_ops, | 497 | .port_ops = &piix_sata_ops, |
643 | }, | 498 | }, |
644 | 499 | ||
645 | [ich8_sata_ahci] = | 500 | [ich8_sata] = |
646 | { | 501 | { |
647 | .flags = PIIX_SATA_FLAGS | PIIX_FLAG_AHCI | | 502 | .flags = PIIX_SATA_FLAGS | PIIX_FLAG_SIDPR, |
648 | PIIX_FLAG_SIDPR, | ||
649 | .pio_mask = 0x1f, /* pio0-4 */ | 503 | .pio_mask = 0x1f, /* pio0-4 */ |
650 | .mwdma_mask = 0x07, /* mwdma0-2 */ | 504 | .mwdma_mask = 0x07, /* mwdma0-2 */ |
651 | .udma_mask = ATA_UDMA6, | 505 | .udma_mask = ATA_UDMA6, |
@@ -654,27 +508,25 @@ static struct ata_port_info piix_port_info[] = { | |||
654 | 508 | ||
655 | [ich8_2port_sata] = | 509 | [ich8_2port_sata] = |
656 | { | 510 | { |
657 | .flags = PIIX_SATA_FLAGS | PIIX_FLAG_AHCI | | 511 | .flags = PIIX_SATA_FLAGS | PIIX_FLAG_SIDPR, |
658 | PIIX_FLAG_SIDPR, | ||
659 | .pio_mask = 0x1f, /* pio0-4 */ | 512 | .pio_mask = 0x1f, /* pio0-4 */ |
660 | .mwdma_mask = 0x07, /* mwdma0-2 */ | 513 | .mwdma_mask = 0x07, /* mwdma0-2 */ |
661 | .udma_mask = ATA_UDMA6, | 514 | .udma_mask = ATA_UDMA6, |
662 | .port_ops = &piix_sata_ops, | 515 | .port_ops = &piix_sata_ops, |
663 | }, | 516 | }, |
664 | 517 | ||
665 | [tolapai_sata_ahci] = | 518 | [tolapai_sata] = |
666 | { | 519 | { |
667 | .flags = PIIX_SATA_FLAGS | PIIX_FLAG_AHCI, | 520 | .flags = PIIX_SATA_FLAGS, |
668 | .pio_mask = 0x1f, /* pio0-4 */ | 521 | .pio_mask = 0x1f, /* pio0-4 */ |
669 | .mwdma_mask = 0x07, /* mwdma0-2 */ | 522 | .mwdma_mask = 0x07, /* mwdma0-2 */ |
670 | .udma_mask = ATA_UDMA6, | 523 | .udma_mask = ATA_UDMA6, |
671 | .port_ops = &piix_sata_ops, | 524 | .port_ops = &piix_sata_ops, |
672 | }, | 525 | }, |
673 | 526 | ||
674 | [ich8m_apple_sata_ahci] = | 527 | [ich8m_apple_sata] = |
675 | { | 528 | { |
676 | .flags = PIIX_SATA_FLAGS | PIIX_FLAG_AHCI | | 529 | .flags = PIIX_SATA_FLAGS | PIIX_FLAG_SIDPR, |
677 | PIIX_FLAG_SIDPR, | ||
678 | .pio_mask = 0x1f, /* pio0-4 */ | 530 | .pio_mask = 0x1f, /* pio0-4 */ |
679 | .mwdma_mask = 0x07, /* mwdma0-2 */ | 531 | .mwdma_mask = 0x07, /* mwdma0-2 */ |
680 | .udma_mask = ATA_UDMA6, | 532 | .udma_mask = ATA_UDMA6, |
@@ -683,7 +535,6 @@ static struct ata_port_info piix_port_info[] = { | |||
683 | 535 | ||
684 | [piix_pata_vmw] = | 536 | [piix_pata_vmw] = |
685 | { | 537 | { |
686 | .sht = &piix_sht, | ||
687 | .flags = PIIX_PATA_FLAGS, | 538 | .flags = PIIX_PATA_FLAGS, |
688 | .pio_mask = 0x1f, /* pio0-4 */ | 539 | .pio_mask = 0x1f, /* pio0-4 */ |
689 | .mwdma_mask = 0x06, /* mwdma1-2 ?? CHECK 0 should be ok but slow */ | 540 | .mwdma_mask = 0x06, /* mwdma1-2 ?? CHECK 0 should be ok but slow */ |
@@ -776,13 +627,7 @@ static int piix_pata_prereset(struct ata_link *link, unsigned long deadline) | |||
776 | 627 | ||
777 | if (!pci_test_config_bits(pdev, &piix_enable_bits[ap->port_no])) | 628 | if (!pci_test_config_bits(pdev, &piix_enable_bits[ap->port_no])) |
778 | return -ENOENT; | 629 | return -ENOENT; |
779 | return ata_std_prereset(link, deadline); | 630 | return ata_sff_prereset(link, deadline); |
780 | } | ||
781 | |||
782 | static void piix_pata_error_handler(struct ata_port *ap) | ||
783 | { | ||
784 | ata_bmdma_drive_eh(ap, piix_pata_prereset, ata_std_softreset, NULL, | ||
785 | ata_std_postreset); | ||
786 | } | 631 | } |
787 | 632 | ||
788 | /** | 633 | /** |
@@ -1168,35 +1013,6 @@ static int piix_sidpr_scr_write(struct ata_port *ap, unsigned int reg, u32 val) | |||
1168 | return 0; | 1013 | return 0; |
1169 | } | 1014 | } |
1170 | 1015 | ||
1171 | static int piix_sidpr_hardreset(struct ata_link *link, unsigned int *class, | ||
1172 | unsigned long deadline) | ||
1173 | { | ||
1174 | const unsigned long *timing = sata_ehc_deb_timing(&link->eh_context); | ||
1175 | int rc; | ||
1176 | |||
1177 | /* do hardreset */ | ||
1178 | rc = sata_link_hardreset(link, timing, deadline); | ||
1179 | if (rc) { | ||
1180 | ata_link_printk(link, KERN_ERR, | ||
1181 | "COMRESET failed (errno=%d)\n", rc); | ||
1182 | return rc; | ||
1183 | } | ||
1184 | |||
1185 | /* TODO: phy layer with polling, timeouts, etc. */ | ||
1186 | if (ata_link_offline(link)) { | ||
1187 | *class = ATA_DEV_NONE; | ||
1188 | return 0; | ||
1189 | } | ||
1190 | |||
1191 | return -EAGAIN; | ||
1192 | } | ||
1193 | |||
1194 | static void piix_sidpr_error_handler(struct ata_port *ap) | ||
1195 | { | ||
1196 | ata_bmdma_drive_eh(ap, ata_std_prereset, ata_std_softreset, | ||
1197 | piix_sidpr_hardreset, ata_std_postreset); | ||
1198 | } | ||
1199 | |||
1200 | #ifdef CONFIG_PM | 1016 | #ifdef CONFIG_PM |
1201 | static int piix_broken_suspend(void) | 1017 | static int piix_broken_suspend(void) |
1202 | { | 1018 | { |
@@ -1633,6 +1449,16 @@ static int __devinit piix_init_one(struct pci_dev *pdev, | |||
1633 | if (rc) | 1449 | if (rc) |
1634 | return rc; | 1450 | return rc; |
1635 | 1451 | ||
1452 | /* ICH6R may be driven by either ata_piix or ahci driver | ||
1453 | * regardless of BIOS configuration. Make sure AHCI mode is | ||
1454 | * off. | ||
1455 | */ | ||
1456 | if (pdev->vendor == PCI_VENDOR_ID_INTEL && pdev->device == 0x2652) { | ||
1457 | int rc = piix_disable_ahci(pdev); | ||
1458 | if (rc) | ||
1459 | return rc; | ||
1460 | } | ||
1461 | |||
1636 | /* SATA map init can change port_info, do it before prepping host */ | 1462 | /* SATA map init can change port_info, do it before prepping host */ |
1637 | hpriv = devm_kzalloc(dev, sizeof(*hpriv), GFP_KERNEL); | 1463 | hpriv = devm_kzalloc(dev, sizeof(*hpriv), GFP_KERNEL); |
1638 | if (!hpriv) | 1464 | if (!hpriv) |
@@ -1642,22 +1468,12 @@ static int __devinit piix_init_one(struct pci_dev *pdev, | |||
1642 | hpriv->map = piix_init_sata_map(pdev, port_info, | 1468 | hpriv->map = piix_init_sata_map(pdev, port_info, |
1643 | piix_map_db_table[ent->driver_data]); | 1469 | piix_map_db_table[ent->driver_data]); |
1644 | 1470 | ||
1645 | rc = ata_pci_prepare_sff_host(pdev, ppi, &host); | 1471 | rc = ata_pci_sff_prepare_host(pdev, ppi, &host); |
1646 | if (rc) | 1472 | if (rc) |
1647 | return rc; | 1473 | return rc; |
1648 | host->private_data = hpriv; | 1474 | host->private_data = hpriv; |
1649 | 1475 | ||
1650 | /* initialize controller */ | 1476 | /* initialize controller */ |
1651 | if (port_flags & PIIX_FLAG_AHCI) { | ||
1652 | u8 tmp; | ||
1653 | pci_read_config_byte(pdev, PIIX_SCC, &tmp); | ||
1654 | if (tmp == PIIX_AHCI_DEVICE) { | ||
1655 | rc = piix_disable_ahci(pdev); | ||
1656 | if (rc) | ||
1657 | return rc; | ||
1658 | } | ||
1659 | } | ||
1660 | |||
1661 | if (port_flags & ATA_FLAG_SATA) { | 1477 | if (port_flags & ATA_FLAG_SATA) { |
1662 | piix_init_pcs(host, piix_map_db_table[ent->driver_data]); | 1478 | piix_init_pcs(host, piix_map_db_table[ent->driver_data]); |
1663 | piix_init_sidpr(host); | 1479 | piix_init_sidpr(host); |
@@ -1686,7 +1502,7 @@ static int __devinit piix_init_one(struct pci_dev *pdev, | |||
1686 | } | 1502 | } |
1687 | 1503 | ||
1688 | pci_set_master(pdev); | 1504 | pci_set_master(pdev); |
1689 | return ata_pci_activate_sff_host(host, ata_interrupt, &piix_sht); | 1505 | return ata_pci_sff_activate_host(host, ata_sff_interrupt, &piix_sht); |
1690 | } | 1506 | } |
1691 | 1507 | ||
1692 | static int __init piix_init(void) | 1508 | static int __init piix_init(void) |
diff --git a/drivers/ata/libata-acpi.c b/drivers/ata/libata-acpi.c index bf98a566adac..8c1cfc645c85 100644 --- a/drivers/ata/libata-acpi.c +++ b/drivers/ata/libata-acpi.c | |||
@@ -77,7 +77,7 @@ void ata_acpi_associate_sata_port(struct ata_port *ap) | |||
77 | { | 77 | { |
78 | WARN_ON(!(ap->flags & ATA_FLAG_ACPI_SATA)); | 78 | WARN_ON(!(ap->flags & ATA_FLAG_ACPI_SATA)); |
79 | 79 | ||
80 | if (!ap->nr_pmp_links) { | 80 | if (!sata_pmp_attached(ap)) { |
81 | acpi_integer adr = SATA_ADR(ap->port_no, NO_PORT_MULT); | 81 | acpi_integer adr = SATA_ADR(ap->port_no, NO_PORT_MULT); |
82 | 82 | ||
83 | ap->link.device->acpi_handle = | 83 | ap->link.device->acpi_handle = |
@@ -839,7 +839,8 @@ void ata_acpi_on_resume(struct ata_port *ap) | |||
839 | */ | 839 | */ |
840 | ata_link_for_each_dev(dev, &ap->link) { | 840 | ata_link_for_each_dev(dev, &ap->link) { |
841 | ata_acpi_clear_gtf(dev); | 841 | ata_acpi_clear_gtf(dev); |
842 | if (ata_dev_get_GTF(dev, NULL) >= 0) | 842 | if (ata_dev_enabled(dev) && |
843 | ata_dev_get_GTF(dev, NULL) >= 0) | ||
843 | dev->flags |= ATA_DFLAG_ACPI_PENDING; | 844 | dev->flags |= ATA_DFLAG_ACPI_PENDING; |
844 | } | 845 | } |
845 | } else { | 846 | } else { |
@@ -849,7 +850,8 @@ void ata_acpi_on_resume(struct ata_port *ap) | |||
849 | */ | 850 | */ |
850 | ata_link_for_each_dev(dev, &ap->link) { | 851 | ata_link_for_each_dev(dev, &ap->link) { |
851 | ata_acpi_clear_gtf(dev); | 852 | ata_acpi_clear_gtf(dev); |
852 | dev->flags |= ATA_DFLAG_ACPI_PENDING; | 853 | if (ata_dev_enabled(dev)) |
854 | dev->flags |= ATA_DFLAG_ACPI_PENDING; | ||
853 | } | 855 | } |
854 | } | 856 | } |
855 | } | 857 | } |
diff --git a/drivers/ata/libata-core.c b/drivers/ata/libata-core.c index be95fdb69726..733eb94d055e 100644 --- a/drivers/ata/libata-core.c +++ b/drivers/ata/libata-core.c | |||
@@ -46,7 +46,6 @@ | |||
46 | #include <linux/init.h> | 46 | #include <linux/init.h> |
47 | #include <linux/list.h> | 47 | #include <linux/list.h> |
48 | #include <linux/mm.h> | 48 | #include <linux/mm.h> |
49 | #include <linux/highmem.h> | ||
50 | #include <linux/spinlock.h> | 49 | #include <linux/spinlock.h> |
51 | #include <linux/blkdev.h> | 50 | #include <linux/blkdev.h> |
52 | #include <linux/delay.h> | 51 | #include <linux/delay.h> |
@@ -74,6 +73,19 @@ const unsigned long sata_deb_timing_normal[] = { 5, 100, 2000 }; | |||
74 | const unsigned long sata_deb_timing_hotplug[] = { 25, 500, 2000 }; | 73 | const unsigned long sata_deb_timing_hotplug[] = { 25, 500, 2000 }; |
75 | const unsigned long sata_deb_timing_long[] = { 100, 2000, 5000 }; | 74 | const unsigned long sata_deb_timing_long[] = { 100, 2000, 5000 }; |
76 | 75 | ||
76 | const struct ata_port_operations ata_base_port_ops = { | ||
77 | .prereset = ata_std_prereset, | ||
78 | .postreset = ata_std_postreset, | ||
79 | .error_handler = ata_std_error_handler, | ||
80 | }; | ||
81 | |||
82 | const struct ata_port_operations sata_port_ops = { | ||
83 | .inherits = &ata_base_port_ops, | ||
84 | |||
85 | .qc_defer = ata_std_qc_defer, | ||
86 | .hardreset = sata_std_hardreset, | ||
87 | }; | ||
88 | |||
77 | static unsigned int ata_dev_init_params(struct ata_device *dev, | 89 | static unsigned int ata_dev_init_params(struct ata_device *dev, |
78 | u16 heads, u16 sectors); | 90 | u16 heads, u16 sectors); |
79 | static unsigned int ata_dev_set_xfermode(struct ata_device *dev); | 91 | static unsigned int ata_dev_set_xfermode(struct ata_device *dev); |
@@ -1043,50 +1055,6 @@ static void ata_lpm_disable(struct ata_host *host) | |||
1043 | } | 1055 | } |
1044 | #endif /* CONFIG_PM */ | 1056 | #endif /* CONFIG_PM */ |
1045 | 1057 | ||
1046 | |||
1047 | /** | ||
1048 | * ata_devchk - PATA device presence detection | ||
1049 | * @ap: ATA channel to examine | ||
1050 | * @device: Device to examine (starting at zero) | ||
1051 | * | ||
1052 | * This technique was originally described in | ||
1053 | * Hale Landis's ATADRVR (www.ata-atapi.com), and | ||
1054 | * later found its way into the ATA/ATAPI spec. | ||
1055 | * | ||
1056 | * Write a pattern to the ATA shadow registers, | ||
1057 | * and if a device is present, it will respond by | ||
1058 | * correctly storing and echoing back the | ||
1059 | * ATA shadow register contents. | ||
1060 | * | ||
1061 | * LOCKING: | ||
1062 | * caller. | ||
1063 | */ | ||
1064 | |||
1065 | static unsigned int ata_devchk(struct ata_port *ap, unsigned int device) | ||
1066 | { | ||
1067 | struct ata_ioports *ioaddr = &ap->ioaddr; | ||
1068 | u8 nsect, lbal; | ||
1069 | |||
1070 | ap->ops->dev_select(ap, device); | ||
1071 | |||
1072 | iowrite8(0x55, ioaddr->nsect_addr); | ||
1073 | iowrite8(0xaa, ioaddr->lbal_addr); | ||
1074 | |||
1075 | iowrite8(0xaa, ioaddr->nsect_addr); | ||
1076 | iowrite8(0x55, ioaddr->lbal_addr); | ||
1077 | |||
1078 | iowrite8(0x55, ioaddr->nsect_addr); | ||
1079 | iowrite8(0xaa, ioaddr->lbal_addr); | ||
1080 | |||
1081 | nsect = ioread8(ioaddr->nsect_addr); | ||
1082 | lbal = ioread8(ioaddr->lbal_addr); | ||
1083 | |||
1084 | if ((nsect == 0x55) && (lbal == 0xaa)) | ||
1085 | return 1; /* we found a device */ | ||
1086 | |||
1087 | return 0; /* nothing found */ | ||
1088 | } | ||
1089 | |||
1090 | /** | 1058 | /** |
1091 | * ata_dev_classify - determine device type based on ATA-spec signature | 1059 | * ata_dev_classify - determine device type based on ATA-spec signature |
1092 | * @tf: ATA taskfile register set for device to be identified | 1060 | * @tf: ATA taskfile register set for device to be identified |
@@ -1147,75 +1115,6 @@ unsigned int ata_dev_classify(const struct ata_taskfile *tf) | |||
1147 | } | 1115 | } |
1148 | 1116 | ||
1149 | /** | 1117 | /** |
1150 | * ata_dev_try_classify - Parse returned ATA device signature | ||
1151 | * @dev: ATA device to classify (starting at zero) | ||
1152 | * @present: device seems present | ||
1153 | * @r_err: Value of error register on completion | ||
1154 | * | ||
1155 | * After an event -- SRST, E.D.D., or SATA COMRESET -- occurs, | ||
1156 | * an ATA/ATAPI-defined set of values is placed in the ATA | ||
1157 | * shadow registers, indicating the results of device detection | ||
1158 | * and diagnostics. | ||
1159 | * | ||
1160 | * Select the ATA device, and read the values from the ATA shadow | ||
1161 | * registers. Then parse according to the Error register value, | ||
1162 | * and the spec-defined values examined by ata_dev_classify(). | ||
1163 | * | ||
1164 | * LOCKING: | ||
1165 | * caller. | ||
1166 | * | ||
1167 | * RETURNS: | ||
1168 | * Device type - %ATA_DEV_ATA, %ATA_DEV_ATAPI or %ATA_DEV_NONE. | ||
1169 | */ | ||
1170 | unsigned int ata_dev_try_classify(struct ata_device *dev, int present, | ||
1171 | u8 *r_err) | ||
1172 | { | ||
1173 | struct ata_port *ap = dev->link->ap; | ||
1174 | struct ata_taskfile tf; | ||
1175 | unsigned int class; | ||
1176 | u8 err; | ||
1177 | |||
1178 | ap->ops->dev_select(ap, dev->devno); | ||
1179 | |||
1180 | memset(&tf, 0, sizeof(tf)); | ||
1181 | |||
1182 | ap->ops->tf_read(ap, &tf); | ||
1183 | err = tf.feature; | ||
1184 | if (r_err) | ||
1185 | *r_err = err; | ||
1186 | |||
1187 | /* see if device passed diags: continue and warn later */ | ||
1188 | if (err == 0) | ||
1189 | /* diagnostic fail : do nothing _YET_ */ | ||
1190 | dev->horkage |= ATA_HORKAGE_DIAGNOSTIC; | ||
1191 | else if (err == 1) | ||
1192 | /* do nothing */ ; | ||
1193 | else if ((dev->devno == 0) && (err == 0x81)) | ||
1194 | /* do nothing */ ; | ||
1195 | else | ||
1196 | return ATA_DEV_NONE; | ||
1197 | |||
1198 | /* determine if device is ATA or ATAPI */ | ||
1199 | class = ata_dev_classify(&tf); | ||
1200 | |||
1201 | if (class == ATA_DEV_UNKNOWN) { | ||
1202 | /* If the device failed diagnostic, it's likely to | ||
1203 | * have reported incorrect device signature too. | ||
1204 | * Assume ATA device if the device seems present but | ||
1205 | * device signature is invalid with diagnostic | ||
1206 | * failure. | ||
1207 | */ | ||
1208 | if (present && (dev->horkage & ATA_HORKAGE_DIAGNOSTIC)) | ||
1209 | class = ATA_DEV_ATA; | ||
1210 | else | ||
1211 | class = ATA_DEV_NONE; | ||
1212 | } else if ((class == ATA_DEV_ATA) && (ata_chk_status(ap) == 0)) | ||
1213 | class = ATA_DEV_NONE; | ||
1214 | |||
1215 | return class; | ||
1216 | } | ||
1217 | |||
1218 | /** | ||
1219 | * ata_id_string - Convert IDENTIFY DEVICE page into string | 1118 | * ata_id_string - Convert IDENTIFY DEVICE page into string |
1220 | * @id: IDENTIFY DEVICE results we will examine | 1119 | * @id: IDENTIFY DEVICE results we will examine |
1221 | * @s: string into which data is output | 1120 | * @s: string into which data is output |
@@ -1293,7 +1192,7 @@ static u64 ata_id_n_sectors(const u16 *id) | |||
1293 | } | 1192 | } |
1294 | } | 1193 | } |
1295 | 1194 | ||
1296 | static u64 ata_tf_to_lba48(struct ata_taskfile *tf) | 1195 | u64 ata_tf_to_lba48(const struct ata_taskfile *tf) |
1297 | { | 1196 | { |
1298 | u64 sectors = 0; | 1197 | u64 sectors = 0; |
1299 | 1198 | ||
@@ -1304,10 +1203,10 @@ static u64 ata_tf_to_lba48(struct ata_taskfile *tf) | |||
1304 | sectors |= (tf->lbam & 0xff) << 8; | 1203 | sectors |= (tf->lbam & 0xff) << 8; |
1305 | sectors |= (tf->lbal & 0xff); | 1204 | sectors |= (tf->lbal & 0xff); |
1306 | 1205 | ||
1307 | return ++sectors; | 1206 | return sectors; |
1308 | } | 1207 | } |
1309 | 1208 | ||
1310 | static u64 ata_tf_to_lba(struct ata_taskfile *tf) | 1209 | u64 ata_tf_to_lba(const struct ata_taskfile *tf) |
1311 | { | 1210 | { |
1312 | u64 sectors = 0; | 1211 | u64 sectors = 0; |
1313 | 1212 | ||
@@ -1316,7 +1215,7 @@ static u64 ata_tf_to_lba(struct ata_taskfile *tf) | |||
1316 | sectors |= (tf->lbam & 0xff) << 8; | 1215 | sectors |= (tf->lbam & 0xff) << 8; |
1317 | sectors |= (tf->lbal & 0xff); | 1216 | sectors |= (tf->lbal & 0xff); |
1318 | 1217 | ||
1319 | return ++sectors; | 1218 | return sectors; |
1320 | } | 1219 | } |
1321 | 1220 | ||
1322 | /** | 1221 | /** |
@@ -1361,9 +1260,9 @@ static int ata_read_native_max_address(struct ata_device *dev, u64 *max_sectors) | |||
1361 | } | 1260 | } |
1362 | 1261 | ||
1363 | if (lba48) | 1262 | if (lba48) |
1364 | *max_sectors = ata_tf_to_lba48(&tf); | 1263 | *max_sectors = ata_tf_to_lba48(&tf) + 1; |
1365 | else | 1264 | else |
1366 | *max_sectors = ata_tf_to_lba(&tf); | 1265 | *max_sectors = ata_tf_to_lba(&tf) + 1; |
1367 | if (dev->horkage & ATA_HORKAGE_HPA_SIZE) | 1266 | if (dev->horkage & ATA_HORKAGE_HPA_SIZE) |
1368 | (*max_sectors)--; | 1267 | (*max_sectors)--; |
1369 | return 0; | 1268 | return 0; |
@@ -1523,89 +1422,6 @@ static int ata_hpa_resize(struct ata_device *dev) | |||
1523 | } | 1422 | } |
1524 | 1423 | ||
1525 | /** | 1424 | /** |
1526 | * ata_noop_dev_select - Select device 0/1 on ATA bus | ||
1527 | * @ap: ATA channel to manipulate | ||
1528 | * @device: ATA device (numbered from zero) to select | ||
1529 | * | ||
1530 | * This function performs no actual function. | ||
1531 | * | ||
1532 | * May be used as the dev_select() entry in ata_port_operations. | ||
1533 | * | ||
1534 | * LOCKING: | ||
1535 | * caller. | ||
1536 | */ | ||
1537 | void ata_noop_dev_select(struct ata_port *ap, unsigned int device) | ||
1538 | { | ||
1539 | } | ||
1540 | |||
1541 | |||
1542 | /** | ||
1543 | * ata_std_dev_select - Select device 0/1 on ATA bus | ||
1544 | * @ap: ATA channel to manipulate | ||
1545 | * @device: ATA device (numbered from zero) to select | ||
1546 | * | ||
1547 | * Use the method defined in the ATA specification to | ||
1548 | * make either device 0, or device 1, active on the | ||
1549 | * ATA channel. Works with both PIO and MMIO. | ||
1550 | * | ||
1551 | * May be used as the dev_select() entry in ata_port_operations. | ||
1552 | * | ||
1553 | * LOCKING: | ||
1554 | * caller. | ||
1555 | */ | ||
1556 | |||
1557 | void ata_std_dev_select(struct ata_port *ap, unsigned int device) | ||
1558 | { | ||
1559 | u8 tmp; | ||
1560 | |||
1561 | if (device == 0) | ||
1562 | tmp = ATA_DEVICE_OBS; | ||
1563 | else | ||
1564 | tmp = ATA_DEVICE_OBS | ATA_DEV1; | ||
1565 | |||
1566 | iowrite8(tmp, ap->ioaddr.device_addr); | ||
1567 | ata_pause(ap); /* needed; also flushes, for mmio */ | ||
1568 | } | ||
1569 | |||
1570 | /** | ||
1571 | * ata_dev_select - Select device 0/1 on ATA bus | ||
1572 | * @ap: ATA channel to manipulate | ||
1573 | * @device: ATA device (numbered from zero) to select | ||
1574 | * @wait: non-zero to wait for Status register BSY bit to clear | ||
1575 | * @can_sleep: non-zero if context allows sleeping | ||
1576 | * | ||
1577 | * Use the method defined in the ATA specification to | ||
1578 | * make either device 0, or device 1, active on the | ||
1579 | * ATA channel. | ||
1580 | * | ||
1581 | * This is a high-level version of ata_std_dev_select(), | ||
1582 | * which additionally provides the services of inserting | ||
1583 | * the proper pauses and status polling, where needed. | ||
1584 | * | ||
1585 | * LOCKING: | ||
1586 | * caller. | ||
1587 | */ | ||
1588 | |||
1589 | void ata_dev_select(struct ata_port *ap, unsigned int device, | ||
1590 | unsigned int wait, unsigned int can_sleep) | ||
1591 | { | ||
1592 | if (ata_msg_probe(ap)) | ||
1593 | ata_port_printk(ap, KERN_INFO, "ata_dev_select: ENTER, " | ||
1594 | "device %u, wait %u\n", device, wait); | ||
1595 | |||
1596 | if (wait) | ||
1597 | ata_wait_idle(ap); | ||
1598 | |||
1599 | ap->ops->dev_select(ap, device); | ||
1600 | |||
1601 | if (wait) { | ||
1602 | if (can_sleep && ap->link.device[device].class == ATA_DEV_ATAPI) | ||
1603 | msleep(150); | ||
1604 | ata_wait_idle(ap); | ||
1605 | } | ||
1606 | } | ||
1607 | |||
1608 | /** | ||
1609 | * ata_dump_id - IDENTIFY DEVICE info debugging output | 1425 | * ata_dump_id - IDENTIFY DEVICE info debugging output |
1610 | * @id: IDENTIFY DEVICE page to dump | 1426 | * @id: IDENTIFY DEVICE page to dump |
1611 | * | 1427 | * |
@@ -1732,8 +1548,7 @@ unsigned long ata_id_xfermask(const u16 *id) | |||
1732 | * LOCKING: | 1548 | * LOCKING: |
1733 | * Inherited from caller. | 1549 | * Inherited from caller. |
1734 | */ | 1550 | */ |
1735 | static void ata_pio_queue_task(struct ata_port *ap, void *data, | 1551 | void ata_pio_queue_task(struct ata_port *ap, void *data, unsigned long delay) |
1736 | unsigned long delay) | ||
1737 | { | 1552 | { |
1738 | ap->port_task_data = data; | 1553 | ap->port_task_data = data; |
1739 | 1554 | ||
@@ -2097,7 +1912,6 @@ int ata_dev_read_id(struct ata_device *dev, unsigned int *p_class, | |||
2097 | if (ata_msg_ctl(ap)) | 1912 | if (ata_msg_ctl(ap)) |
2098 | ata_dev_printk(dev, KERN_DEBUG, "%s: ENTER\n", __func__); | 1913 | ata_dev_printk(dev, KERN_DEBUG, "%s: ENTER\n", __func__); |
2099 | 1914 | ||
2100 | ata_dev_select(ap, dev->devno, 1, 1); /* select device 0/1 */ | ||
2101 | retry: | 1915 | retry: |
2102 | ata_tf_init(dev, &tf); | 1916 | ata_tf_init(dev, &tf); |
2103 | 1917 | ||
@@ -2464,7 +2278,7 @@ int ata_dev_configure(struct ata_device *dev) | |||
2464 | * changed notifications and ATAPI ANs. | 2278 | * changed notifications and ATAPI ANs. |
2465 | */ | 2279 | */ |
2466 | if ((ap->flags & ATA_FLAG_AN) && ata_id_has_atapi_AN(id) && | 2280 | if ((ap->flags & ATA_FLAG_AN) && ata_id_has_atapi_AN(id) && |
2467 | (!ap->nr_pmp_links || | 2281 | (!sata_pmp_attached(ap) || |
2468 | sata_scr_read(&ap->link, SCR_NOTIFICATION, &sntf) == 0)) { | 2282 | sata_scr_read(&ap->link, SCR_NOTIFICATION, &sntf) == 0)) { |
2469 | unsigned int err_mask; | 2283 | unsigned int err_mask; |
2470 | 2284 | ||
@@ -2558,9 +2372,6 @@ int ata_dev_configure(struct ata_device *dev) | |||
2558 | } | 2372 | } |
2559 | } | 2373 | } |
2560 | 2374 | ||
2561 | if (ata_msg_probe(ap)) | ||
2562 | ata_dev_printk(dev, KERN_DEBUG, "%s: EXIT, drv_stat = 0x%x\n", | ||
2563 | __func__, ata_chk_status(ap)); | ||
2564 | return 0; | 2375 | return 0; |
2565 | 2376 | ||
2566 | err_out_nosup: | 2377 | err_out_nosup: |
@@ -3321,16 +3132,21 @@ static int ata_dev_set_mode(struct ata_device *dev) | |||
3321 | if (rc) | 3132 | if (rc) |
3322 | return rc; | 3133 | return rc; |
3323 | 3134 | ||
3324 | /* Old CFA may refuse this command, which is just fine */ | 3135 | if (dev->xfer_shift == ATA_SHIFT_PIO) { |
3325 | if (dev->xfer_shift == ATA_SHIFT_PIO && ata_id_is_cfa(dev->id)) | 3136 | /* Old CFA may refuse this command, which is just fine */ |
3326 | ign_dev_err = 1; | 3137 | if (ata_id_is_cfa(dev->id)) |
3327 | 3138 | ign_dev_err = 1; | |
3328 | /* Some very old devices and some bad newer ones fail any kind of | 3139 | /* Catch several broken garbage emulations plus some pre |
3329 | SET_XFERMODE request but support PIO0-2 timings and no IORDY */ | 3140 | ATA devices */ |
3330 | if (dev->xfer_shift == ATA_SHIFT_PIO && !ata_id_has_iordy(dev->id) && | 3141 | if (ata_id_major_version(dev->id) == 0 && |
3331 | dev->pio_mode <= XFER_PIO_2) | 3142 | dev->pio_mode <= XFER_PIO_2) |
3332 | ign_dev_err = 1; | 3143 | ign_dev_err = 1; |
3333 | 3144 | /* Some very old devices and some bad newer ones fail | |
3145 | any kind of SET_XFERMODE request but support PIO0-2 | ||
3146 | timings and no IORDY */ | ||
3147 | if (!ata_id_has_iordy(dev->id) && dev->pio_mode <= XFER_PIO_2) | ||
3148 | ign_dev_err = 1; | ||
3149 | } | ||
3334 | /* Early MWDMA devices do DMA but don't allow DMA mode setting. | 3150 | /* Early MWDMA devices do DMA but don't allow DMA mode setting. |
3335 | Don't fail an MWDMA0 set IFF the device indicates it is in MWDMA0 */ | 3151 | Don't fail an MWDMA0 set IFF the device indicates it is in MWDMA0 */ |
3336 | if (dev->xfer_shift == ATA_SHIFT_MWDMA && | 3152 | if (dev->xfer_shift == ATA_SHIFT_MWDMA && |
@@ -3474,170 +3290,73 @@ int ata_do_set_mode(struct ata_link *link, struct ata_device **r_failed_dev) | |||
3474 | } | 3290 | } |
3475 | 3291 | ||
3476 | /** | 3292 | /** |
3477 | * ata_tf_to_host - issue ATA taskfile to host controller | 3293 | * ata_wait_ready - wait for link to become ready |
3478 | * @ap: port to which command is being issued | 3294 | * @link: link to be waited on |
3479 | * @tf: ATA taskfile register set | ||
3480 | * | ||
3481 | * Issues ATA taskfile register set to ATA host controller, | ||
3482 | * with proper synchronization with interrupt handler and | ||
3483 | * other threads. | ||
3484 | * | ||
3485 | * LOCKING: | ||
3486 | * spin_lock_irqsave(host lock) | ||
3487 | */ | ||
3488 | |||
3489 | static inline void ata_tf_to_host(struct ata_port *ap, | ||
3490 | const struct ata_taskfile *tf) | ||
3491 | { | ||
3492 | ap->ops->tf_load(ap, tf); | ||
3493 | ap->ops->exec_command(ap, tf); | ||
3494 | } | ||
3495 | |||
3496 | /** | ||
3497 | * ata_busy_sleep - sleep until BSY clears, or timeout | ||
3498 | * @ap: port containing status register to be polled | ||
3499 | * @tmout_pat: impatience timeout | ||
3500 | * @tmout: overall timeout | ||
3501 | * | ||
3502 | * Sleep until ATA Status register bit BSY clears, | ||
3503 | * or a timeout occurs. | ||
3504 | * | ||
3505 | * LOCKING: | ||
3506 | * Kernel thread context (may sleep). | ||
3507 | * | ||
3508 | * RETURNS: | ||
3509 | * 0 on success, -errno otherwise. | ||
3510 | */ | ||
3511 | int ata_busy_sleep(struct ata_port *ap, | ||
3512 | unsigned long tmout_pat, unsigned long tmout) | ||
3513 | { | ||
3514 | unsigned long timer_start, timeout; | ||
3515 | u8 status; | ||
3516 | |||
3517 | status = ata_busy_wait(ap, ATA_BUSY, 300); | ||
3518 | timer_start = jiffies; | ||
3519 | timeout = timer_start + tmout_pat; | ||
3520 | while (status != 0xff && (status & ATA_BUSY) && | ||
3521 | time_before(jiffies, timeout)) { | ||
3522 | msleep(50); | ||
3523 | status = ata_busy_wait(ap, ATA_BUSY, 3); | ||
3524 | } | ||
3525 | |||
3526 | if (status != 0xff && (status & ATA_BUSY)) | ||
3527 | ata_port_printk(ap, KERN_WARNING, | ||
3528 | "port is slow to respond, please be patient " | ||
3529 | "(Status 0x%x)\n", status); | ||
3530 | |||
3531 | timeout = timer_start + tmout; | ||
3532 | while (status != 0xff && (status & ATA_BUSY) && | ||
3533 | time_before(jiffies, timeout)) { | ||
3534 | msleep(50); | ||
3535 | status = ata_chk_status(ap); | ||
3536 | } | ||
3537 | |||
3538 | if (status == 0xff) | ||
3539 | return -ENODEV; | ||
3540 | |||
3541 | if (status & ATA_BUSY) { | ||
3542 | ata_port_printk(ap, KERN_ERR, "port failed to respond " | ||
3543 | "(%lu secs, Status 0x%x)\n", | ||
3544 | tmout / HZ, status); | ||
3545 | return -EBUSY; | ||
3546 | } | ||
3547 | |||
3548 | return 0; | ||
3549 | } | ||
3550 | |||
3551 | /** | ||
3552 | * ata_wait_after_reset - wait before checking status after reset | ||
3553 | * @ap: port containing status register to be polled | ||
3554 | * @deadline: deadline jiffies for the operation | 3295 | * @deadline: deadline jiffies for the operation |
3296 | * @check_ready: callback to check link readiness | ||
3555 | * | 3297 | * |
3556 | * After reset, we need to pause a while before reading status. | 3298 | * Wait for @link to become ready. @check_ready should return |
3557 | * Also, certain combination of controller and device report 0xff | 3299 | * positive number if @link is ready, 0 if it isn't, -ENODEV if |
3558 | * for some duration (e.g. until SATA PHY is up and running) | 3300 | * link doesn't seem to be occupied, other errno for other error |
3559 | * which is interpreted as empty port in ATA world. This | 3301 | * conditions. |
3560 | * function also waits for such devices to get out of 0xff | ||
3561 | * status. | ||
3562 | * | 3302 | * |
3563 | * LOCKING: | 3303 | * Transient -ENODEV conditions are allowed for |
3564 | * Kernel thread context (may sleep). | 3304 | * ATA_TMOUT_FF_WAIT. |
3565 | */ | ||
3566 | void ata_wait_after_reset(struct ata_port *ap, unsigned long deadline) | ||
3567 | { | ||
3568 | unsigned long until = jiffies + ATA_TMOUT_FF_WAIT; | ||
3569 | |||
3570 | if (time_before(until, deadline)) | ||
3571 | deadline = until; | ||
3572 | |||
3573 | /* Spec mandates ">= 2ms" before checking status. We wait | ||
3574 | * 150ms, because that was the magic delay used for ATAPI | ||
3575 | * devices in Hale Landis's ATADRVR, for the period of time | ||
3576 | * between when the ATA command register is written, and then | ||
3577 | * status is checked. Because waiting for "a while" before | ||
3578 | * checking status is fine, post SRST, we perform this magic | ||
3579 | * delay here as well. | ||
3580 | * | ||
3581 | * Old drivers/ide uses the 2mS rule and then waits for ready. | ||
3582 | */ | ||
3583 | msleep(150); | ||
3584 | |||
3585 | /* Wait for 0xff to clear. Some SATA devices take a long time | ||
3586 | * to clear 0xff after reset. For example, HHD424020F7SV00 | ||
3587 | * iVDR needs >= 800ms while. Quantum GoVault needs even more | ||
3588 | * than that. | ||
3589 | * | ||
3590 | * Note that some PATA controllers (pata_ali) explode if | ||
3591 | * status register is read more than once when there's no | ||
3592 | * device attached. | ||
3593 | */ | ||
3594 | if (ap->flags & ATA_FLAG_SATA) { | ||
3595 | while (1) { | ||
3596 | u8 status = ata_chk_status(ap); | ||
3597 | |||
3598 | if (status != 0xff || time_after(jiffies, deadline)) | ||
3599 | return; | ||
3600 | |||
3601 | msleep(50); | ||
3602 | } | ||
3603 | } | ||
3604 | } | ||
3605 | |||
3606 | /** | ||
3607 | * ata_wait_ready - sleep until BSY clears, or timeout | ||
3608 | * @ap: port containing status register to be polled | ||
3609 | * @deadline: deadline jiffies for the operation | ||
3610 | * | ||
3611 | * Sleep until ATA Status register bit BSY clears, or timeout | ||
3612 | * occurs. | ||
3613 | * | 3305 | * |
3614 | * LOCKING: | 3306 | * LOCKING: |
3615 | * Kernel thread context (may sleep). | 3307 | * EH context. |
3616 | * | 3308 | * |
3617 | * RETURNS: | 3309 | * RETURNS: |
3618 | * 0 on success, -errno otherwise. | 3310 | * 0 if @linke is ready before @deadline; otherwise, -errno. |
3619 | */ | 3311 | */ |
3620 | int ata_wait_ready(struct ata_port *ap, unsigned long deadline) | 3312 | int ata_wait_ready(struct ata_link *link, unsigned long deadline, |
3313 | int (*check_ready)(struct ata_link *link)) | ||
3621 | { | 3314 | { |
3622 | unsigned long start = jiffies; | 3315 | unsigned long start = jiffies; |
3316 | unsigned long nodev_deadline = start + ATA_TMOUT_FF_WAIT; | ||
3623 | int warned = 0; | 3317 | int warned = 0; |
3624 | 3318 | ||
3319 | if (time_after(nodev_deadline, deadline)) | ||
3320 | nodev_deadline = deadline; | ||
3321 | |||
3625 | while (1) { | 3322 | while (1) { |
3626 | u8 status = ata_chk_status(ap); | ||
3627 | unsigned long now = jiffies; | 3323 | unsigned long now = jiffies; |
3324 | int ready, tmp; | ||
3628 | 3325 | ||
3629 | if (!(status & ATA_BUSY)) | 3326 | ready = tmp = check_ready(link); |
3327 | if (ready > 0) | ||
3630 | return 0; | 3328 | return 0; |
3631 | if (!ata_link_online(&ap->link) && status == 0xff) | 3329 | |
3632 | return -ENODEV; | 3330 | /* -ENODEV could be transient. Ignore -ENODEV if link |
3331 | * is online. Also, some SATA devices take a long | ||
3332 | * time to clear 0xff after reset. For example, | ||
3333 | * HHD424020F7SV00 iVDR needs >= 800ms while Quantum | ||
3334 | * GoVault needs even more than that. Wait for | ||
3335 | * ATA_TMOUT_FF_WAIT on -ENODEV if link isn't offline. | ||
3336 | * | ||
3337 | * Note that some PATA controllers (pata_ali) explode | ||
3338 | * if status register is read more than once when | ||
3339 | * there's no device attached. | ||
3340 | */ | ||
3341 | if (ready == -ENODEV) { | ||
3342 | if (ata_link_online(link)) | ||
3343 | ready = 0; | ||
3344 | else if ((link->ap->flags & ATA_FLAG_SATA) && | ||
3345 | !ata_link_offline(link) && | ||
3346 | time_before(now, nodev_deadline)) | ||
3347 | ready = 0; | ||
3348 | } | ||
3349 | |||
3350 | if (ready) | ||
3351 | return ready; | ||
3633 | if (time_after(now, deadline)) | 3352 | if (time_after(now, deadline)) |
3634 | return -EBUSY; | 3353 | return -EBUSY; |
3635 | 3354 | ||
3636 | if (!warned && time_after(now, start + 5 * HZ) && | 3355 | if (!warned && time_after(now, start + 5 * HZ) && |
3637 | (deadline - now > 3 * HZ)) { | 3356 | (deadline - now > 3 * HZ)) { |
3638 | ata_port_printk(ap, KERN_WARNING, | 3357 | ata_link_printk(link, KERN_WARNING, |
3639 | "port is slow to respond, please be patient " | 3358 | "link is slow to respond, please be patient " |
3640 | "(Status 0x%x)\n", status); | 3359 | "(ready=%d)\n", tmp); |
3641 | warned = 1; | 3360 | warned = 1; |
3642 | } | 3361 | } |
3643 | 3362 | ||
@@ -3645,179 +3364,26 @@ int ata_wait_ready(struct ata_port *ap, unsigned long deadline) | |||
3645 | } | 3364 | } |
3646 | } | 3365 | } |
3647 | 3366 | ||
3648 | static int ata_bus_post_reset(struct ata_port *ap, unsigned int devmask, | ||
3649 | unsigned long deadline) | ||
3650 | { | ||
3651 | struct ata_ioports *ioaddr = &ap->ioaddr; | ||
3652 | unsigned int dev0 = devmask & (1 << 0); | ||
3653 | unsigned int dev1 = devmask & (1 << 1); | ||
3654 | int rc, ret = 0; | ||
3655 | |||
3656 | /* if device 0 was found in ata_devchk, wait for its | ||
3657 | * BSY bit to clear | ||
3658 | */ | ||
3659 | if (dev0) { | ||
3660 | rc = ata_wait_ready(ap, deadline); | ||
3661 | if (rc) { | ||
3662 | if (rc != -ENODEV) | ||
3663 | return rc; | ||
3664 | ret = rc; | ||
3665 | } | ||
3666 | } | ||
3667 | |||
3668 | /* if device 1 was found in ata_devchk, wait for register | ||
3669 | * access briefly, then wait for BSY to clear. | ||
3670 | */ | ||
3671 | if (dev1) { | ||
3672 | int i; | ||
3673 | |||
3674 | ap->ops->dev_select(ap, 1); | ||
3675 | |||
3676 | /* Wait for register access. Some ATAPI devices fail | ||
3677 | * to set nsect/lbal after reset, so don't waste too | ||
3678 | * much time on it. We're gonna wait for !BSY anyway. | ||
3679 | */ | ||
3680 | for (i = 0; i < 2; i++) { | ||
3681 | u8 nsect, lbal; | ||
3682 | |||
3683 | nsect = ioread8(ioaddr->nsect_addr); | ||
3684 | lbal = ioread8(ioaddr->lbal_addr); | ||
3685 | if ((nsect == 1) && (lbal == 1)) | ||
3686 | break; | ||
3687 | msleep(50); /* give drive a breather */ | ||
3688 | } | ||
3689 | |||
3690 | rc = ata_wait_ready(ap, deadline); | ||
3691 | if (rc) { | ||
3692 | if (rc != -ENODEV) | ||
3693 | return rc; | ||
3694 | ret = rc; | ||
3695 | } | ||
3696 | } | ||
3697 | |||
3698 | /* is all this really necessary? */ | ||
3699 | ap->ops->dev_select(ap, 0); | ||
3700 | if (dev1) | ||
3701 | ap->ops->dev_select(ap, 1); | ||
3702 | if (dev0) | ||
3703 | ap->ops->dev_select(ap, 0); | ||
3704 | |||
3705 | return ret; | ||
3706 | } | ||
3707 | |||
3708 | static int ata_bus_softreset(struct ata_port *ap, unsigned int devmask, | ||
3709 | unsigned long deadline) | ||
3710 | { | ||
3711 | struct ata_ioports *ioaddr = &ap->ioaddr; | ||
3712 | |||
3713 | DPRINTK("ata%u: bus reset via SRST\n", ap->print_id); | ||
3714 | |||
3715 | /* software reset. causes dev0 to be selected */ | ||
3716 | iowrite8(ap->ctl, ioaddr->ctl_addr); | ||
3717 | udelay(20); /* FIXME: flush */ | ||
3718 | iowrite8(ap->ctl | ATA_SRST, ioaddr->ctl_addr); | ||
3719 | udelay(20); /* FIXME: flush */ | ||
3720 | iowrite8(ap->ctl, ioaddr->ctl_addr); | ||
3721 | |||
3722 | /* wait a while before checking status */ | ||
3723 | ata_wait_after_reset(ap, deadline); | ||
3724 | |||
3725 | /* Before we perform post reset processing we want to see if | ||
3726 | * the bus shows 0xFF because the odd clown forgets the D7 | ||
3727 | * pulldown resistor. | ||
3728 | */ | ||
3729 | if (ata_chk_status(ap) == 0xFF) | ||
3730 | return -ENODEV; | ||
3731 | |||
3732 | return ata_bus_post_reset(ap, devmask, deadline); | ||
3733 | } | ||
3734 | |||
3735 | /** | 3367 | /** |
3736 | * ata_bus_reset - reset host port and associated ATA channel | 3368 | * ata_wait_after_reset - wait for link to become ready after reset |
3737 | * @ap: port to reset | 3369 | * @link: link to be waited on |
3370 | * @deadline: deadline jiffies for the operation | ||
3371 | * @check_ready: callback to check link readiness | ||
3738 | * | 3372 | * |
3739 | * This is typically the first time we actually start issuing | 3373 | * Wait for @link to become ready after reset. |
3740 | * commands to the ATA channel. We wait for BSY to clear, then | ||
3741 | * issue EXECUTE DEVICE DIAGNOSTIC command, polling for its | ||
3742 | * result. Determine what devices, if any, are on the channel | ||
3743 | * by looking at the device 0/1 error register. Look at the signature | ||
3744 | * stored in each device's taskfile registers, to determine if | ||
3745 | * the device is ATA or ATAPI. | ||
3746 | * | 3374 | * |
3747 | * LOCKING: | 3375 | * LOCKING: |
3748 | * PCI/etc. bus probe sem. | 3376 | * EH context. |
3749 | * Obtains host lock. | ||
3750 | * | 3377 | * |
3751 | * SIDE EFFECTS: | 3378 | * RETURNS: |
3752 | * Sets ATA_FLAG_DISABLED if bus reset fails. | 3379 | * 0 if @linke is ready before @deadline; otherwise, -errno. |
3753 | */ | 3380 | */ |
3754 | 3381 | extern int ata_wait_after_reset(struct ata_link *link, unsigned long deadline, | |
3755 | void ata_bus_reset(struct ata_port *ap) | 3382 | int (*check_ready)(struct ata_link *link)) |
3756 | { | 3383 | { |
3757 | struct ata_device *device = ap->link.device; | 3384 | msleep(ATA_WAIT_AFTER_RESET_MSECS); |
3758 | struct ata_ioports *ioaddr = &ap->ioaddr; | ||
3759 | unsigned int slave_possible = ap->flags & ATA_FLAG_SLAVE_POSS; | ||
3760 | u8 err; | ||
3761 | unsigned int dev0, dev1 = 0, devmask = 0; | ||
3762 | int rc; | ||
3763 | |||
3764 | DPRINTK("ENTER, host %u, port %u\n", ap->print_id, ap->port_no); | ||
3765 | |||
3766 | /* determine if device 0/1 are present */ | ||
3767 | if (ap->flags & ATA_FLAG_SATA_RESET) | ||
3768 | dev0 = 1; | ||
3769 | else { | ||
3770 | dev0 = ata_devchk(ap, 0); | ||
3771 | if (slave_possible) | ||
3772 | dev1 = ata_devchk(ap, 1); | ||
3773 | } | ||
3774 | |||
3775 | if (dev0) | ||
3776 | devmask |= (1 << 0); | ||
3777 | if (dev1) | ||
3778 | devmask |= (1 << 1); | ||
3779 | 3385 | ||
3780 | /* select device 0 again */ | 3386 | return ata_wait_ready(link, deadline, check_ready); |
3781 | ap->ops->dev_select(ap, 0); | ||
3782 | |||
3783 | /* issue bus reset */ | ||
3784 | if (ap->flags & ATA_FLAG_SRST) { | ||
3785 | rc = ata_bus_softreset(ap, devmask, jiffies + 40 * HZ); | ||
3786 | if (rc && rc != -ENODEV) | ||
3787 | goto err_out; | ||
3788 | } | ||
3789 | |||
3790 | /* | ||
3791 | * determine by signature whether we have ATA or ATAPI devices | ||
3792 | */ | ||
3793 | device[0].class = ata_dev_try_classify(&device[0], dev0, &err); | ||
3794 | if ((slave_possible) && (err != 0x81)) | ||
3795 | device[1].class = ata_dev_try_classify(&device[1], dev1, &err); | ||
3796 | |||
3797 | /* is double-select really necessary? */ | ||
3798 | if (device[1].class != ATA_DEV_NONE) | ||
3799 | ap->ops->dev_select(ap, 1); | ||
3800 | if (device[0].class != ATA_DEV_NONE) | ||
3801 | ap->ops->dev_select(ap, 0); | ||
3802 | |||
3803 | /* if no devices were detected, disable this port */ | ||
3804 | if ((device[0].class == ATA_DEV_NONE) && | ||
3805 | (device[1].class == ATA_DEV_NONE)) | ||
3806 | goto err_out; | ||
3807 | |||
3808 | if (ap->flags & (ATA_FLAG_SATA_RESET | ATA_FLAG_SRST)) { | ||
3809 | /* set up device control for ATA_FLAG_SATA_RESET */ | ||
3810 | iowrite8(ap->ctl, ioaddr->ctl_addr); | ||
3811 | } | ||
3812 | |||
3813 | DPRINTK("EXIT\n"); | ||
3814 | return; | ||
3815 | |||
3816 | err_out: | ||
3817 | ata_port_printk(ap, KERN_ERR, "disabling port\n"); | ||
3818 | ata_port_disable(ap); | ||
3819 | |||
3820 | DPRINTK("EXIT\n"); | ||
3821 | } | 3387 | } |
3822 | 3388 | ||
3823 | /** | 3389 | /** |
@@ -3906,7 +3472,7 @@ int sata_link_debounce(struct ata_link *link, const unsigned long *params, | |||
3906 | int sata_link_resume(struct ata_link *link, const unsigned long *params, | 3472 | int sata_link_resume(struct ata_link *link, const unsigned long *params, |
3907 | unsigned long deadline) | 3473 | unsigned long deadline) |
3908 | { | 3474 | { |
3909 | u32 scontrol; | 3475 | u32 scontrol, serror; |
3910 | int rc; | 3476 | int rc; |
3911 | 3477 | ||
3912 | if ((rc = sata_scr_read(link, SCR_CONTROL, &scontrol))) | 3478 | if ((rc = sata_scr_read(link, SCR_CONTROL, &scontrol))) |
@@ -3922,7 +3488,25 @@ int sata_link_resume(struct ata_link *link, const unsigned long *params, | |||
3922 | */ | 3488 | */ |
3923 | msleep(200); | 3489 | msleep(200); |
3924 | 3490 | ||
3925 | return sata_link_debounce(link, params, deadline); | 3491 | if ((rc = sata_link_debounce(link, params, deadline))) |
3492 | return rc; | ||
3493 | |||
3494 | /* Clear SError. PMP and some host PHYs require this to | ||
3495 | * operate and clearing should be done before checking PHY | ||
3496 | * online status to avoid race condition (hotplugging between | ||
3497 | * link resume and status check). | ||
3498 | */ | ||
3499 | if (!(rc = sata_scr_read(link, SCR_ERROR, &serror))) | ||
3500 | rc = sata_scr_write(link, SCR_ERROR, serror); | ||
3501 | if (rc == 0 || rc == -EINVAL) { | ||
3502 | unsigned long flags; | ||
3503 | |||
3504 | spin_lock_irqsave(link->ap->lock, flags); | ||
3505 | link->eh_info.serror = 0; | ||
3506 | spin_unlock_irqrestore(link->ap->lock, flags); | ||
3507 | rc = 0; | ||
3508 | } | ||
3509 | return rc; | ||
3926 | } | 3510 | } |
3927 | 3511 | ||
3928 | /** | 3512 | /** |
@@ -3949,17 +3533,6 @@ int ata_std_prereset(struct ata_link *link, unsigned long deadline) | |||
3949 | const unsigned long *timing = sata_ehc_deb_timing(ehc); | 3533 | const unsigned long *timing = sata_ehc_deb_timing(ehc); |
3950 | int rc; | 3534 | int rc; |
3951 | 3535 | ||
3952 | /* handle link resume */ | ||
3953 | if ((ehc->i.flags & ATA_EHI_RESUME_LINK) && | ||
3954 | (link->flags & ATA_LFLAG_HRST_TO_RESUME)) | ||
3955 | ehc->i.action |= ATA_EH_HARDRESET; | ||
3956 | |||
3957 | /* Some PMPs don't work with only SRST, force hardreset if PMP | ||
3958 | * is supported. | ||
3959 | */ | ||
3960 | if (ap->flags & ATA_FLAG_PMP) | ||
3961 | ehc->i.action |= ATA_EH_HARDRESET; | ||
3962 | |||
3963 | /* if we're about to do hardreset, nothing more to do */ | 3536 | /* if we're about to do hardreset, nothing more to do */ |
3964 | if (ehc->i.action & ATA_EH_HARDRESET) | 3537 | if (ehc->i.action & ATA_EH_HARDRESET) |
3965 | return 0; | 3538 | return 0; |
@@ -3973,88 +3546,30 @@ int ata_std_prereset(struct ata_link *link, unsigned long deadline) | |||
3973 | "link for reset (errno=%d)\n", rc); | 3546 | "link for reset (errno=%d)\n", rc); |
3974 | } | 3547 | } |
3975 | 3548 | ||
3976 | /* Wait for !BSY if the controller can wait for the first D2H | 3549 | /* no point in trying softreset on offline link */ |
3977 | * Reg FIS and we don't know that no device is attached. | 3550 | if (ata_link_offline(link)) |
3978 | */ | 3551 | ehc->i.action &= ~ATA_EH_SOFTRESET; |
3979 | if (!(link->flags & ATA_LFLAG_SKIP_D2H_BSY) && !ata_link_offline(link)) { | ||
3980 | rc = ata_wait_ready(ap, deadline); | ||
3981 | if (rc && rc != -ENODEV) { | ||
3982 | ata_link_printk(link, KERN_WARNING, "device not ready " | ||
3983 | "(errno=%d), forcing hardreset\n", rc); | ||
3984 | ehc->i.action |= ATA_EH_HARDRESET; | ||
3985 | } | ||
3986 | } | ||
3987 | 3552 | ||
3988 | return 0; | 3553 | return 0; |
3989 | } | 3554 | } |
3990 | 3555 | ||
3991 | /** | 3556 | /** |
3992 | * ata_std_softreset - reset host port via ATA SRST | ||
3993 | * @link: ATA link to reset | ||
3994 | * @classes: resulting classes of attached devices | ||
3995 | * @deadline: deadline jiffies for the operation | ||
3996 | * | ||
3997 | * Reset host port using ATA SRST. | ||
3998 | * | ||
3999 | * LOCKING: | ||
4000 | * Kernel thread context (may sleep) | ||
4001 | * | ||
4002 | * RETURNS: | ||
4003 | * 0 on success, -errno otherwise. | ||
4004 | */ | ||
4005 | int ata_std_softreset(struct ata_link *link, unsigned int *classes, | ||
4006 | unsigned long deadline) | ||
4007 | { | ||
4008 | struct ata_port *ap = link->ap; | ||
4009 | unsigned int slave_possible = ap->flags & ATA_FLAG_SLAVE_POSS; | ||
4010 | unsigned int devmask = 0; | ||
4011 | int rc; | ||
4012 | u8 err; | ||
4013 | |||
4014 | DPRINTK("ENTER\n"); | ||
4015 | |||
4016 | if (ata_link_offline(link)) { | ||
4017 | classes[0] = ATA_DEV_NONE; | ||
4018 | goto out; | ||
4019 | } | ||
4020 | |||
4021 | /* determine if device 0/1 are present */ | ||
4022 | if (ata_devchk(ap, 0)) | ||
4023 | devmask |= (1 << 0); | ||
4024 | if (slave_possible && ata_devchk(ap, 1)) | ||
4025 | devmask |= (1 << 1); | ||
4026 | |||
4027 | /* select device 0 again */ | ||
4028 | ap->ops->dev_select(ap, 0); | ||
4029 | |||
4030 | /* issue bus reset */ | ||
4031 | DPRINTK("about to softreset, devmask=%x\n", devmask); | ||
4032 | rc = ata_bus_softreset(ap, devmask, deadline); | ||
4033 | /* if link is occupied, -ENODEV too is an error */ | ||
4034 | if (rc && (rc != -ENODEV || sata_scr_valid(link))) { | ||
4035 | ata_link_printk(link, KERN_ERR, "SRST failed (errno=%d)\n", rc); | ||
4036 | return rc; | ||
4037 | } | ||
4038 | |||
4039 | /* determine by signature whether we have ATA or ATAPI devices */ | ||
4040 | classes[0] = ata_dev_try_classify(&link->device[0], | ||
4041 | devmask & (1 << 0), &err); | ||
4042 | if (slave_possible && err != 0x81) | ||
4043 | classes[1] = ata_dev_try_classify(&link->device[1], | ||
4044 | devmask & (1 << 1), &err); | ||
4045 | |||
4046 | out: | ||
4047 | DPRINTK("EXIT, classes[0]=%u [1]=%u\n", classes[0], classes[1]); | ||
4048 | return 0; | ||
4049 | } | ||
4050 | |||
4051 | /** | ||
4052 | * sata_link_hardreset - reset link via SATA phy reset | 3557 | * sata_link_hardreset - reset link via SATA phy reset |
4053 | * @link: link to reset | 3558 | * @link: link to reset |
4054 | * @timing: timing parameters { interval, duratinon, timeout } in msec | 3559 | * @timing: timing parameters { interval, duratinon, timeout } in msec |
4055 | * @deadline: deadline jiffies for the operation | 3560 | * @deadline: deadline jiffies for the operation |
3561 | * @online: optional out parameter indicating link onlineness | ||
3562 | * @check_ready: optional callback to check link readiness | ||
4056 | * | 3563 | * |
4057 | * SATA phy-reset @link using DET bits of SControl register. | 3564 | * SATA phy-reset @link using DET bits of SControl register. |
3565 | * After hardreset, link readiness is waited upon using | ||
3566 | * ata_wait_ready() if @check_ready is specified. LLDs are | ||
3567 | * allowed to not specify @check_ready and wait itself after this | ||
3568 | * function returns. Device classification is LLD's | ||
3569 | * responsibility. | ||
3570 | * | ||
3571 | * *@online is set to one iff reset succeeded and @link is online | ||
3572 | * after reset. | ||
4058 | * | 3573 | * |
4059 | * LOCKING: | 3574 | * LOCKING: |
4060 | * Kernel thread context (may sleep) | 3575 | * Kernel thread context (may sleep) |
@@ -4063,13 +3578,17 @@ int ata_std_softreset(struct ata_link *link, unsigned int *classes, | |||
4063 | * 0 on success, -errno otherwise. | 3578 | * 0 on success, -errno otherwise. |
4064 | */ | 3579 | */ |
4065 | int sata_link_hardreset(struct ata_link *link, const unsigned long *timing, | 3580 | int sata_link_hardreset(struct ata_link *link, const unsigned long *timing, |
4066 | unsigned long deadline) | 3581 | unsigned long deadline, |
3582 | bool *online, int (*check_ready)(struct ata_link *)) | ||
4067 | { | 3583 | { |
4068 | u32 scontrol; | 3584 | u32 scontrol; |
4069 | int rc; | 3585 | int rc; |
4070 | 3586 | ||
4071 | DPRINTK("ENTER\n"); | 3587 | DPRINTK("ENTER\n"); |
4072 | 3588 | ||
3589 | if (online) | ||
3590 | *online = false; | ||
3591 | |||
4073 | if (sata_set_spd_needed(link)) { | 3592 | if (sata_set_spd_needed(link)) { |
4074 | /* SATA spec says nothing about how to reconfigure | 3593 | /* SATA spec says nothing about how to reconfigure |
4075 | * spd. To be on the safe side, turn off phy during | 3594 | * spd. To be on the safe side, turn off phy during |
@@ -4103,77 +3622,69 @@ int sata_link_hardreset(struct ata_link *link, const unsigned long *timing, | |||
4103 | 3622 | ||
4104 | /* bring link back */ | 3623 | /* bring link back */ |
4105 | rc = sata_link_resume(link, timing, deadline); | 3624 | rc = sata_link_resume(link, timing, deadline); |
3625 | if (rc) | ||
3626 | goto out; | ||
3627 | /* if link is offline nothing more to do */ | ||
3628 | if (ata_link_offline(link)) | ||
3629 | goto out; | ||
3630 | |||
3631 | /* Link is online. From this point, -ENODEV too is an error. */ | ||
3632 | if (online) | ||
3633 | *online = true; | ||
3634 | |||
3635 | if (sata_pmp_supported(link->ap) && ata_is_host_link(link)) { | ||
3636 | /* If PMP is supported, we have to do follow-up SRST. | ||
3637 | * Some PMPs don't send D2H Reg FIS after hardreset if | ||
3638 | * the first port is empty. Wait only for | ||
3639 | * ATA_TMOUT_PMP_SRST_WAIT. | ||
3640 | */ | ||
3641 | if (check_ready) { | ||
3642 | unsigned long pmp_deadline; | ||
3643 | |||
3644 | pmp_deadline = jiffies + ATA_TMOUT_PMP_SRST_WAIT; | ||
3645 | if (time_after(pmp_deadline, deadline)) | ||
3646 | pmp_deadline = deadline; | ||
3647 | ata_wait_ready(link, pmp_deadline, check_ready); | ||
3648 | } | ||
3649 | rc = -EAGAIN; | ||
3650 | goto out; | ||
3651 | } | ||
3652 | |||
3653 | rc = 0; | ||
3654 | if (check_ready) | ||
3655 | rc = ata_wait_ready(link, deadline, check_ready); | ||
4106 | out: | 3656 | out: |
3657 | if (rc && rc != -EAGAIN) | ||
3658 | ata_link_printk(link, KERN_ERR, | ||
3659 | "COMRESET failed (errno=%d)\n", rc); | ||
4107 | DPRINTK("EXIT, rc=%d\n", rc); | 3660 | DPRINTK("EXIT, rc=%d\n", rc); |
4108 | return rc; | 3661 | return rc; |
4109 | } | 3662 | } |
4110 | 3663 | ||
4111 | /** | 3664 | /** |
4112 | * sata_std_hardreset - reset host port via SATA phy reset | 3665 | * sata_std_hardreset - COMRESET w/o waiting or classification |
4113 | * @link: link to reset | 3666 | * @link: link to reset |
4114 | * @class: resulting class of attached device | 3667 | * @class: resulting class of attached device |
4115 | * @deadline: deadline jiffies for the operation | 3668 | * @deadline: deadline jiffies for the operation |
4116 | * | 3669 | * |
4117 | * SATA phy-reset host port using DET bits of SControl register, | 3670 | * Standard SATA COMRESET w/o waiting or classification. |
4118 | * wait for !BSY and classify the attached device. | ||
4119 | * | 3671 | * |
4120 | * LOCKING: | 3672 | * LOCKING: |
4121 | * Kernel thread context (may sleep) | 3673 | * Kernel thread context (may sleep) |
4122 | * | 3674 | * |
4123 | * RETURNS: | 3675 | * RETURNS: |
4124 | * 0 on success, -errno otherwise. | 3676 | * 0 if link offline, -EAGAIN if link online, -errno on errors. |
4125 | */ | 3677 | */ |
4126 | int sata_std_hardreset(struct ata_link *link, unsigned int *class, | 3678 | int sata_std_hardreset(struct ata_link *link, unsigned int *class, |
4127 | unsigned long deadline) | 3679 | unsigned long deadline) |
4128 | { | 3680 | { |
4129 | struct ata_port *ap = link->ap; | ||
4130 | const unsigned long *timing = sata_ehc_deb_timing(&link->eh_context); | 3681 | const unsigned long *timing = sata_ehc_deb_timing(&link->eh_context); |
3682 | bool online; | ||
4131 | int rc; | 3683 | int rc; |
4132 | 3684 | ||
4133 | DPRINTK("ENTER\n"); | ||
4134 | |||
4135 | /* do hardreset */ | 3685 | /* do hardreset */ |
4136 | rc = sata_link_hardreset(link, timing, deadline); | 3686 | rc = sata_link_hardreset(link, timing, deadline, &online, NULL); |
4137 | if (rc) { | 3687 | return online ? -EAGAIN : rc; |
4138 | ata_link_printk(link, KERN_ERR, | ||
4139 | "COMRESET failed (errno=%d)\n", rc); | ||
4140 | return rc; | ||
4141 | } | ||
4142 | |||
4143 | /* TODO: phy layer with polling, timeouts, etc. */ | ||
4144 | if (ata_link_offline(link)) { | ||
4145 | *class = ATA_DEV_NONE; | ||
4146 | DPRINTK("EXIT, link offline\n"); | ||
4147 | return 0; | ||
4148 | } | ||
4149 | |||
4150 | /* wait a while before checking status */ | ||
4151 | ata_wait_after_reset(ap, deadline); | ||
4152 | |||
4153 | /* If PMP is supported, we have to do follow-up SRST. Note | ||
4154 | * that some PMPs don't send D2H Reg FIS after hardreset at | ||
4155 | * all if the first port is empty. Wait for it just for a | ||
4156 | * second and request follow-up SRST. | ||
4157 | */ | ||
4158 | if (ap->flags & ATA_FLAG_PMP) { | ||
4159 | ata_wait_ready(ap, jiffies + HZ); | ||
4160 | return -EAGAIN; | ||
4161 | } | ||
4162 | |||
4163 | rc = ata_wait_ready(ap, deadline); | ||
4164 | /* link occupied, -ENODEV too is an error */ | ||
4165 | if (rc) { | ||
4166 | ata_link_printk(link, KERN_ERR, | ||
4167 | "COMRESET failed (errno=%d)\n", rc); | ||
4168 | return rc; | ||
4169 | } | ||
4170 | |||
4171 | ap->ops->dev_select(ap, 0); /* probably unnecessary */ | ||
4172 | |||
4173 | *class = ata_dev_try_classify(link->device, 1, NULL); | ||
4174 | |||
4175 | DPRINTK("EXIT, class=%u\n", *class); | ||
4176 | return 0; | ||
4177 | } | 3688 | } |
4178 | 3689 | ||
4179 | /** | 3690 | /** |
@@ -4190,35 +3701,11 @@ int sata_std_hardreset(struct ata_link *link, unsigned int *class, | |||
4190 | */ | 3701 | */ |
4191 | void ata_std_postreset(struct ata_link *link, unsigned int *classes) | 3702 | void ata_std_postreset(struct ata_link *link, unsigned int *classes) |
4192 | { | 3703 | { |
4193 | struct ata_port *ap = link->ap; | ||
4194 | u32 serror; | ||
4195 | |||
4196 | DPRINTK("ENTER\n"); | 3704 | DPRINTK("ENTER\n"); |
4197 | 3705 | ||
4198 | /* print link status */ | 3706 | /* print link status */ |
4199 | sata_print_link_status(link); | 3707 | sata_print_link_status(link); |
4200 | 3708 | ||
4201 | /* clear SError */ | ||
4202 | if (sata_scr_read(link, SCR_ERROR, &serror) == 0) | ||
4203 | sata_scr_write(link, SCR_ERROR, serror); | ||
4204 | link->eh_info.serror = 0; | ||
4205 | |||
4206 | /* is double-select really necessary? */ | ||
4207 | if (classes[0] != ATA_DEV_NONE) | ||
4208 | ap->ops->dev_select(ap, 1); | ||
4209 | if (classes[1] != ATA_DEV_NONE) | ||
4210 | ap->ops->dev_select(ap, 0); | ||
4211 | |||
4212 | /* bail out if no device is present */ | ||
4213 | if (classes[0] == ATA_DEV_NONE && classes[1] == ATA_DEV_NONE) { | ||
4214 | DPRINTK("EXIT, no device\n"); | ||
4215 | return; | ||
4216 | } | ||
4217 | |||
4218 | /* set up device control */ | ||
4219 | if (ap->ioaddr.ctl_addr) | ||
4220 | iowrite8(ap->ctl, ap->ioaddr.ctl_addr); | ||
4221 | |||
4222 | DPRINTK("EXIT\n"); | 3709 | DPRINTK("EXIT\n"); |
4223 | } | 3710 | } |
4224 | 3711 | ||
@@ -4528,6 +4015,53 @@ static int ata_is_40wire(struct ata_device *dev) | |||
4528 | } | 4015 | } |
4529 | 4016 | ||
4530 | /** | 4017 | /** |
4018 | * cable_is_40wire - 40/80/SATA decider | ||
4019 | * @ap: port to consider | ||
4020 | * | ||
4021 | * This function encapsulates the policy for speed management | ||
4022 | * in one place. At the moment we don't cache the result but | ||
4023 | * there is a good case for setting ap->cbl to the result when | ||
4024 | * we are called with unknown cables (and figuring out if it | ||
4025 | * impacts hotplug at all). | ||
4026 | * | ||
4027 | * Return 1 if the cable appears to be 40 wire. | ||
4028 | */ | ||
4029 | |||
4030 | static int cable_is_40wire(struct ata_port *ap) | ||
4031 | { | ||
4032 | struct ata_link *link; | ||
4033 | struct ata_device *dev; | ||
4034 | |||
4035 | /* If the controller thinks we are 40 wire, we are */ | ||
4036 | if (ap->cbl == ATA_CBL_PATA40) | ||
4037 | return 1; | ||
4038 | /* If the controller thinks we are 80 wire, we are */ | ||
4039 | if (ap->cbl == ATA_CBL_PATA80 || ap->cbl == ATA_CBL_SATA) | ||
4040 | return 0; | ||
4041 | /* If the system is known to be 40 wire short cable (eg laptop), | ||
4042 | then we allow 80 wire modes even if the drive isn't sure */ | ||
4043 | if (ap->cbl == ATA_CBL_PATA40_SHORT) | ||
4044 | return 0; | ||
4045 | /* If the controller doesn't know we scan | ||
4046 | |||
4047 | - Note: We look for all 40 wire detects at this point. | ||
4048 | Any 80 wire detect is taken to be 80 wire cable | ||
4049 | because | ||
4050 | - In many setups only the one drive (slave if present) | ||
4051 | will give a valid detect | ||
4052 | - If you have a non detect capable drive you don't | ||
4053 | want it to colour the choice | ||
4054 | */ | ||
4055 | ata_port_for_each_link(link, ap) { | ||
4056 | ata_link_for_each_dev(dev, link) { | ||
4057 | if (!ata_is_40wire(dev)) | ||
4058 | return 0; | ||
4059 | } | ||
4060 | } | ||
4061 | return 1; | ||
4062 | } | ||
4063 | |||
4064 | /** | ||
4531 | * ata_dev_xfermask - Compute supported xfermask of the given device | 4065 | * ata_dev_xfermask - Compute supported xfermask of the given device |
4532 | * @dev: Device to compute xfermask for | 4066 | * @dev: Device to compute xfermask for |
4533 | * | 4067 | * |
@@ -4595,10 +4129,7 @@ static void ata_dev_xfermask(struct ata_device *dev) | |||
4595 | */ | 4129 | */ |
4596 | if (xfer_mask & (0xF8 << ATA_SHIFT_UDMA)) | 4130 | if (xfer_mask & (0xF8 << ATA_SHIFT_UDMA)) |
4597 | /* UDMA/44 or higher would be available */ | 4131 | /* UDMA/44 or higher would be available */ |
4598 | if ((ap->cbl == ATA_CBL_PATA40) || | 4132 | if (cable_is_40wire(ap)) { |
4599 | (ata_is_40wire(dev) && | ||
4600 | (ap->cbl == ATA_CBL_PATA_UNK || | ||
4601 | ap->cbl == ATA_CBL_PATA80))) { | ||
4602 | ata_dev_printk(dev, KERN_WARNING, | 4133 | ata_dev_printk(dev, KERN_WARNING, |
4603 | "limited to UDMA/33 due to 40-wire cable\n"); | 4134 | "limited to UDMA/33 due to 40-wire cable\n"); |
4604 | xfer_mask &= ~(0xF8 << ATA_SHIFT_UDMA); | 4135 | xfer_mask &= ~(0xF8 << ATA_SHIFT_UDMA); |
@@ -4759,112 +4290,6 @@ void ata_sg_clean(struct ata_queued_cmd *qc) | |||
4759 | } | 4290 | } |
4760 | 4291 | ||
4761 | /** | 4292 | /** |
4762 | * ata_fill_sg - Fill PCI IDE PRD table | ||
4763 | * @qc: Metadata associated with taskfile to be transferred | ||
4764 | * | ||
4765 | * Fill PCI IDE PRD (scatter-gather) table with segments | ||
4766 | * associated with the current disk command. | ||
4767 | * | ||
4768 | * LOCKING: | ||
4769 | * spin_lock_irqsave(host lock) | ||
4770 | * | ||
4771 | */ | ||
4772 | static void ata_fill_sg(struct ata_queued_cmd *qc) | ||
4773 | { | ||
4774 | struct ata_port *ap = qc->ap; | ||
4775 | struct scatterlist *sg; | ||
4776 | unsigned int si, pi; | ||
4777 | |||
4778 | pi = 0; | ||
4779 | for_each_sg(qc->sg, sg, qc->n_elem, si) { | ||
4780 | u32 addr, offset; | ||
4781 | u32 sg_len, len; | ||
4782 | |||
4783 | /* determine if physical DMA addr spans 64K boundary. | ||
4784 | * Note h/w doesn't support 64-bit, so we unconditionally | ||
4785 | * truncate dma_addr_t to u32. | ||
4786 | */ | ||
4787 | addr = (u32) sg_dma_address(sg); | ||
4788 | sg_len = sg_dma_len(sg); | ||
4789 | |||
4790 | while (sg_len) { | ||
4791 | offset = addr & 0xffff; | ||
4792 | len = sg_len; | ||
4793 | if ((offset + sg_len) > 0x10000) | ||
4794 | len = 0x10000 - offset; | ||
4795 | |||
4796 | ap->prd[pi].addr = cpu_to_le32(addr); | ||
4797 | ap->prd[pi].flags_len = cpu_to_le32(len & 0xffff); | ||
4798 | VPRINTK("PRD[%u] = (0x%X, 0x%X)\n", pi, addr, len); | ||
4799 | |||
4800 | pi++; | ||
4801 | sg_len -= len; | ||
4802 | addr += len; | ||
4803 | } | ||
4804 | } | ||
4805 | |||
4806 | ap->prd[pi - 1].flags_len |= cpu_to_le32(ATA_PRD_EOT); | ||
4807 | } | ||
4808 | |||
4809 | /** | ||
4810 | * ata_fill_sg_dumb - Fill PCI IDE PRD table | ||
4811 | * @qc: Metadata associated with taskfile to be transferred | ||
4812 | * | ||
4813 | * Fill PCI IDE PRD (scatter-gather) table with segments | ||
4814 | * associated with the current disk command. Perform the fill | ||
4815 | * so that we avoid writing any length 64K records for | ||
4816 | * controllers that don't follow the spec. | ||
4817 | * | ||
4818 | * LOCKING: | ||
4819 | * spin_lock_irqsave(host lock) | ||
4820 | * | ||
4821 | */ | ||
4822 | static void ata_fill_sg_dumb(struct ata_queued_cmd *qc) | ||
4823 | { | ||
4824 | struct ata_port *ap = qc->ap; | ||
4825 | struct scatterlist *sg; | ||
4826 | unsigned int si, pi; | ||
4827 | |||
4828 | pi = 0; | ||
4829 | for_each_sg(qc->sg, sg, qc->n_elem, si) { | ||
4830 | u32 addr, offset; | ||
4831 | u32 sg_len, len, blen; | ||
4832 | |||
4833 | /* determine if physical DMA addr spans 64K boundary. | ||
4834 | * Note h/w doesn't support 64-bit, so we unconditionally | ||
4835 | * truncate dma_addr_t to u32. | ||
4836 | */ | ||
4837 | addr = (u32) sg_dma_address(sg); | ||
4838 | sg_len = sg_dma_len(sg); | ||
4839 | |||
4840 | while (sg_len) { | ||
4841 | offset = addr & 0xffff; | ||
4842 | len = sg_len; | ||
4843 | if ((offset + sg_len) > 0x10000) | ||
4844 | len = 0x10000 - offset; | ||
4845 | |||
4846 | blen = len & 0xffff; | ||
4847 | ap->prd[pi].addr = cpu_to_le32(addr); | ||
4848 | if (blen == 0) { | ||
4849 | /* Some PATA chipsets like the CS5530 can't | ||
4850 | cope with 0x0000 meaning 64K as the spec says */ | ||
4851 | ap->prd[pi].flags_len = cpu_to_le32(0x8000); | ||
4852 | blen = 0x8000; | ||
4853 | ap->prd[++pi].addr = cpu_to_le32(addr + 0x8000); | ||
4854 | } | ||
4855 | ap->prd[pi].flags_len = cpu_to_le32(blen); | ||
4856 | VPRINTK("PRD[%u] = (0x%X, 0x%X)\n", pi, addr, len); | ||
4857 | |||
4858 | pi++; | ||
4859 | sg_len -= len; | ||
4860 | addr += len; | ||
4861 | } | ||
4862 | } | ||
4863 | |||
4864 | ap->prd[pi - 1].flags_len |= cpu_to_le32(ATA_PRD_EOT); | ||
4865 | } | ||
4866 | |||
4867 | /** | ||
4868 | * ata_check_atapi_dma - Check whether ATAPI DMA can be supported | 4293 | * ata_check_atapi_dma - Check whether ATAPI DMA can be supported |
4869 | * @qc: Metadata associated with taskfile to check | 4294 | * @qc: Metadata associated with taskfile to check |
4870 | * | 4295 | * |
@@ -4924,40 +4349,6 @@ int ata_std_qc_defer(struct ata_queued_cmd *qc) | |||
4924 | return ATA_DEFER_LINK; | 4349 | return ATA_DEFER_LINK; |
4925 | } | 4350 | } |
4926 | 4351 | ||
4927 | /** | ||
4928 | * ata_qc_prep - Prepare taskfile for submission | ||
4929 | * @qc: Metadata associated with taskfile to be prepared | ||
4930 | * | ||
4931 | * Prepare ATA taskfile for submission. | ||
4932 | * | ||
4933 | * LOCKING: | ||
4934 | * spin_lock_irqsave(host lock) | ||
4935 | */ | ||
4936 | void ata_qc_prep(struct ata_queued_cmd *qc) | ||
4937 | { | ||
4938 | if (!(qc->flags & ATA_QCFLAG_DMAMAP)) | ||
4939 | return; | ||
4940 | |||
4941 | ata_fill_sg(qc); | ||
4942 | } | ||
4943 | |||
4944 | /** | ||
4945 | * ata_dumb_qc_prep - Prepare taskfile for submission | ||
4946 | * @qc: Metadata associated with taskfile to be prepared | ||
4947 | * | ||
4948 | * Prepare ATA taskfile for submission. | ||
4949 | * | ||
4950 | * LOCKING: | ||
4951 | * spin_lock_irqsave(host lock) | ||
4952 | */ | ||
4953 | void ata_dumb_qc_prep(struct ata_queued_cmd *qc) | ||
4954 | { | ||
4955 | if (!(qc->flags & ATA_QCFLAG_DMAMAP)) | ||
4956 | return; | ||
4957 | |||
4958 | ata_fill_sg_dumb(qc); | ||
4959 | } | ||
4960 | |||
4961 | void ata_noop_qc_prep(struct ata_queued_cmd *qc) { } | 4352 | void ata_noop_qc_prep(struct ata_queued_cmd *qc) { } |
4962 | 4353 | ||
4963 | /** | 4354 | /** |
@@ -5036,698 +4427,6 @@ void swap_buf_le16(u16 *buf, unsigned int buf_words) | |||
5036 | } | 4427 | } |
5037 | 4428 | ||
5038 | /** | 4429 | /** |
5039 | * ata_data_xfer - Transfer data by PIO | ||
5040 | * @dev: device to target | ||
5041 | * @buf: data buffer | ||
5042 | * @buflen: buffer length | ||
5043 | * @rw: read/write | ||
5044 | * | ||
5045 | * Transfer data from/to the device data register by PIO. | ||
5046 | * | ||
5047 | * LOCKING: | ||
5048 | * Inherited from caller. | ||
5049 | * | ||
5050 | * RETURNS: | ||
5051 | * Bytes consumed. | ||
5052 | */ | ||
5053 | unsigned int ata_data_xfer(struct ata_device *dev, unsigned char *buf, | ||
5054 | unsigned int buflen, int rw) | ||
5055 | { | ||
5056 | struct ata_port *ap = dev->link->ap; | ||
5057 | void __iomem *data_addr = ap->ioaddr.data_addr; | ||
5058 | unsigned int words = buflen >> 1; | ||
5059 | |||
5060 | /* Transfer multiple of 2 bytes */ | ||
5061 | if (rw == READ) | ||
5062 | ioread16_rep(data_addr, buf, words); | ||
5063 | else | ||
5064 | iowrite16_rep(data_addr, buf, words); | ||
5065 | |||
5066 | /* Transfer trailing 1 byte, if any. */ | ||
5067 | if (unlikely(buflen & 0x01)) { | ||
5068 | __le16 align_buf[1] = { 0 }; | ||
5069 | unsigned char *trailing_buf = buf + buflen - 1; | ||
5070 | |||
5071 | if (rw == READ) { | ||
5072 | align_buf[0] = cpu_to_le16(ioread16(data_addr)); | ||
5073 | memcpy(trailing_buf, align_buf, 1); | ||
5074 | } else { | ||
5075 | memcpy(align_buf, trailing_buf, 1); | ||
5076 | iowrite16(le16_to_cpu(align_buf[0]), data_addr); | ||
5077 | } | ||
5078 | words++; | ||
5079 | } | ||
5080 | |||
5081 | return words << 1; | ||
5082 | } | ||
5083 | |||
5084 | /** | ||
5085 | * ata_data_xfer_noirq - Transfer data by PIO | ||
5086 | * @dev: device to target | ||
5087 | * @buf: data buffer | ||
5088 | * @buflen: buffer length | ||
5089 | * @rw: read/write | ||
5090 | * | ||
5091 | * Transfer data from/to the device data register by PIO. Do the | ||
5092 | * transfer with interrupts disabled. | ||
5093 | * | ||
5094 | * LOCKING: | ||
5095 | * Inherited from caller. | ||
5096 | * | ||
5097 | * RETURNS: | ||
5098 | * Bytes consumed. | ||
5099 | */ | ||
5100 | unsigned int ata_data_xfer_noirq(struct ata_device *dev, unsigned char *buf, | ||
5101 | unsigned int buflen, int rw) | ||
5102 | { | ||
5103 | unsigned long flags; | ||
5104 | unsigned int consumed; | ||
5105 | |||
5106 | local_irq_save(flags); | ||
5107 | consumed = ata_data_xfer(dev, buf, buflen, rw); | ||
5108 | local_irq_restore(flags); | ||
5109 | |||
5110 | return consumed; | ||
5111 | } | ||
5112 | |||
5113 | |||
5114 | /** | ||
5115 | * ata_pio_sector - Transfer a sector of data. | ||
5116 | * @qc: Command on going | ||
5117 | * | ||
5118 | * Transfer qc->sect_size bytes of data from/to the ATA device. | ||
5119 | * | ||
5120 | * LOCKING: | ||
5121 | * Inherited from caller. | ||
5122 | */ | ||
5123 | |||
5124 | static void ata_pio_sector(struct ata_queued_cmd *qc) | ||
5125 | { | ||
5126 | int do_write = (qc->tf.flags & ATA_TFLAG_WRITE); | ||
5127 | struct ata_port *ap = qc->ap; | ||
5128 | struct page *page; | ||
5129 | unsigned int offset; | ||
5130 | unsigned char *buf; | ||
5131 | |||
5132 | if (qc->curbytes == qc->nbytes - qc->sect_size) | ||
5133 | ap->hsm_task_state = HSM_ST_LAST; | ||
5134 | |||
5135 | page = sg_page(qc->cursg); | ||
5136 | offset = qc->cursg->offset + qc->cursg_ofs; | ||
5137 | |||
5138 | /* get the current page and offset */ | ||
5139 | page = nth_page(page, (offset >> PAGE_SHIFT)); | ||
5140 | offset %= PAGE_SIZE; | ||
5141 | |||
5142 | DPRINTK("data %s\n", qc->tf.flags & ATA_TFLAG_WRITE ? "write" : "read"); | ||
5143 | |||
5144 | if (PageHighMem(page)) { | ||
5145 | unsigned long flags; | ||
5146 | |||
5147 | /* FIXME: use a bounce buffer */ | ||
5148 | local_irq_save(flags); | ||
5149 | buf = kmap_atomic(page, KM_IRQ0); | ||
5150 | |||
5151 | /* do the actual data transfer */ | ||
5152 | ap->ops->data_xfer(qc->dev, buf + offset, qc->sect_size, do_write); | ||
5153 | |||
5154 | kunmap_atomic(buf, KM_IRQ0); | ||
5155 | local_irq_restore(flags); | ||
5156 | } else { | ||
5157 | buf = page_address(page); | ||
5158 | ap->ops->data_xfer(qc->dev, buf + offset, qc->sect_size, do_write); | ||
5159 | } | ||
5160 | |||
5161 | qc->curbytes += qc->sect_size; | ||
5162 | qc->cursg_ofs += qc->sect_size; | ||
5163 | |||
5164 | if (qc->cursg_ofs == qc->cursg->length) { | ||
5165 | qc->cursg = sg_next(qc->cursg); | ||
5166 | qc->cursg_ofs = 0; | ||
5167 | } | ||
5168 | } | ||
5169 | |||
5170 | /** | ||
5171 | * ata_pio_sectors - Transfer one or many sectors. | ||
5172 | * @qc: Command on going | ||
5173 | * | ||
5174 | * Transfer one or many sectors of data from/to the | ||
5175 | * ATA device for the DRQ request. | ||
5176 | * | ||
5177 | * LOCKING: | ||
5178 | * Inherited from caller. | ||
5179 | */ | ||
5180 | |||
5181 | static void ata_pio_sectors(struct ata_queued_cmd *qc) | ||
5182 | { | ||
5183 | if (is_multi_taskfile(&qc->tf)) { | ||
5184 | /* READ/WRITE MULTIPLE */ | ||
5185 | unsigned int nsect; | ||
5186 | |||
5187 | WARN_ON(qc->dev->multi_count == 0); | ||
5188 | |||
5189 | nsect = min((qc->nbytes - qc->curbytes) / qc->sect_size, | ||
5190 | qc->dev->multi_count); | ||
5191 | while (nsect--) | ||
5192 | ata_pio_sector(qc); | ||
5193 | } else | ||
5194 | ata_pio_sector(qc); | ||
5195 | |||
5196 | ata_altstatus(qc->ap); /* flush */ | ||
5197 | } | ||
5198 | |||
5199 | /** | ||
5200 | * atapi_send_cdb - Write CDB bytes to hardware | ||
5201 | * @ap: Port to which ATAPI device is attached. | ||
5202 | * @qc: Taskfile currently active | ||
5203 | * | ||
5204 | * When device has indicated its readiness to accept | ||
5205 | * a CDB, this function is called. Send the CDB. | ||
5206 | * | ||
5207 | * LOCKING: | ||
5208 | * caller. | ||
5209 | */ | ||
5210 | |||
5211 | static void atapi_send_cdb(struct ata_port *ap, struct ata_queued_cmd *qc) | ||
5212 | { | ||
5213 | /* send SCSI cdb */ | ||
5214 | DPRINTK("send cdb\n"); | ||
5215 | WARN_ON(qc->dev->cdb_len < 12); | ||
5216 | |||
5217 | ap->ops->data_xfer(qc->dev, qc->cdb, qc->dev->cdb_len, 1); | ||
5218 | ata_altstatus(ap); /* flush */ | ||
5219 | |||
5220 | switch (qc->tf.protocol) { | ||
5221 | case ATAPI_PROT_PIO: | ||
5222 | ap->hsm_task_state = HSM_ST; | ||
5223 | break; | ||
5224 | case ATAPI_PROT_NODATA: | ||
5225 | ap->hsm_task_state = HSM_ST_LAST; | ||
5226 | break; | ||
5227 | case ATAPI_PROT_DMA: | ||
5228 | ap->hsm_task_state = HSM_ST_LAST; | ||
5229 | /* initiate bmdma */ | ||
5230 | ap->ops->bmdma_start(qc); | ||
5231 | break; | ||
5232 | } | ||
5233 | } | ||
5234 | |||
5235 | /** | ||
5236 | * __atapi_pio_bytes - Transfer data from/to the ATAPI device. | ||
5237 | * @qc: Command on going | ||
5238 | * @bytes: number of bytes | ||
5239 | * | ||
5240 | * Transfer Transfer data from/to the ATAPI device. | ||
5241 | * | ||
5242 | * LOCKING: | ||
5243 | * Inherited from caller. | ||
5244 | * | ||
5245 | */ | ||
5246 | static int __atapi_pio_bytes(struct ata_queued_cmd *qc, unsigned int bytes) | ||
5247 | { | ||
5248 | int rw = (qc->tf.flags & ATA_TFLAG_WRITE) ? WRITE : READ; | ||
5249 | struct ata_port *ap = qc->ap; | ||
5250 | struct ata_device *dev = qc->dev; | ||
5251 | struct ata_eh_info *ehi = &dev->link->eh_info; | ||
5252 | struct scatterlist *sg; | ||
5253 | struct page *page; | ||
5254 | unsigned char *buf; | ||
5255 | unsigned int offset, count, consumed; | ||
5256 | |||
5257 | next_sg: | ||
5258 | sg = qc->cursg; | ||
5259 | if (unlikely(!sg)) { | ||
5260 | ata_ehi_push_desc(ehi, "unexpected or too much trailing data " | ||
5261 | "buf=%u cur=%u bytes=%u", | ||
5262 | qc->nbytes, qc->curbytes, bytes); | ||
5263 | return -1; | ||
5264 | } | ||
5265 | |||
5266 | page = sg_page(sg); | ||
5267 | offset = sg->offset + qc->cursg_ofs; | ||
5268 | |||
5269 | /* get the current page and offset */ | ||
5270 | page = nth_page(page, (offset >> PAGE_SHIFT)); | ||
5271 | offset %= PAGE_SIZE; | ||
5272 | |||
5273 | /* don't overrun current sg */ | ||
5274 | count = min(sg->length - qc->cursg_ofs, bytes); | ||
5275 | |||
5276 | /* don't cross page boundaries */ | ||
5277 | count = min(count, (unsigned int)PAGE_SIZE - offset); | ||
5278 | |||
5279 | DPRINTK("data %s\n", qc->tf.flags & ATA_TFLAG_WRITE ? "write" : "read"); | ||
5280 | |||
5281 | if (PageHighMem(page)) { | ||
5282 | unsigned long flags; | ||
5283 | |||
5284 | /* FIXME: use bounce buffer */ | ||
5285 | local_irq_save(flags); | ||
5286 | buf = kmap_atomic(page, KM_IRQ0); | ||
5287 | |||
5288 | /* do the actual data transfer */ | ||
5289 | consumed = ap->ops->data_xfer(dev, buf + offset, count, rw); | ||
5290 | |||
5291 | kunmap_atomic(buf, KM_IRQ0); | ||
5292 | local_irq_restore(flags); | ||
5293 | } else { | ||
5294 | buf = page_address(page); | ||
5295 | consumed = ap->ops->data_xfer(dev, buf + offset, count, rw); | ||
5296 | } | ||
5297 | |||
5298 | bytes -= min(bytes, consumed); | ||
5299 | qc->curbytes += count; | ||
5300 | qc->cursg_ofs += count; | ||
5301 | |||
5302 | if (qc->cursg_ofs == sg->length) { | ||
5303 | qc->cursg = sg_next(qc->cursg); | ||
5304 | qc->cursg_ofs = 0; | ||
5305 | } | ||
5306 | |||
5307 | /* consumed can be larger than count only for the last transfer */ | ||
5308 | WARN_ON(qc->cursg && count != consumed); | ||
5309 | |||
5310 | if (bytes) | ||
5311 | goto next_sg; | ||
5312 | return 0; | ||
5313 | } | ||
5314 | |||
5315 | /** | ||
5316 | * atapi_pio_bytes - Transfer data from/to the ATAPI device. | ||
5317 | * @qc: Command on going | ||
5318 | * | ||
5319 | * Transfer Transfer data from/to the ATAPI device. | ||
5320 | * | ||
5321 | * LOCKING: | ||
5322 | * Inherited from caller. | ||
5323 | */ | ||
5324 | |||
5325 | static void atapi_pio_bytes(struct ata_queued_cmd *qc) | ||
5326 | { | ||
5327 | struct ata_port *ap = qc->ap; | ||
5328 | struct ata_device *dev = qc->dev; | ||
5329 | struct ata_eh_info *ehi = &dev->link->eh_info; | ||
5330 | unsigned int ireason, bc_lo, bc_hi, bytes; | ||
5331 | int i_write, do_write = (qc->tf.flags & ATA_TFLAG_WRITE) ? 1 : 0; | ||
5332 | |||
5333 | /* Abuse qc->result_tf for temp storage of intermediate TF | ||
5334 | * here to save some kernel stack usage. | ||
5335 | * For normal completion, qc->result_tf is not relevant. For | ||
5336 | * error, qc->result_tf is later overwritten by ata_qc_complete(). | ||
5337 | * So, the correctness of qc->result_tf is not affected. | ||
5338 | */ | ||
5339 | ap->ops->tf_read(ap, &qc->result_tf); | ||
5340 | ireason = qc->result_tf.nsect; | ||
5341 | bc_lo = qc->result_tf.lbam; | ||
5342 | bc_hi = qc->result_tf.lbah; | ||
5343 | bytes = (bc_hi << 8) | bc_lo; | ||
5344 | |||
5345 | /* shall be cleared to zero, indicating xfer of data */ | ||
5346 | if (unlikely(ireason & (1 << 0))) | ||
5347 | goto atapi_check; | ||
5348 | |||
5349 | /* make sure transfer direction matches expected */ | ||
5350 | i_write = ((ireason & (1 << 1)) == 0) ? 1 : 0; | ||
5351 | if (unlikely(do_write != i_write)) | ||
5352 | goto atapi_check; | ||
5353 | |||
5354 | if (unlikely(!bytes)) | ||
5355 | goto atapi_check; | ||
5356 | |||
5357 | VPRINTK("ata%u: xfering %d bytes\n", ap->print_id, bytes); | ||
5358 | |||
5359 | if (unlikely(__atapi_pio_bytes(qc, bytes))) | ||
5360 | goto err_out; | ||
5361 | ata_altstatus(ap); /* flush */ | ||
5362 | |||
5363 | return; | ||
5364 | |||
5365 | atapi_check: | ||
5366 | ata_ehi_push_desc(ehi, "ATAPI check failed (ireason=0x%x bytes=%u)", | ||
5367 | ireason, bytes); | ||
5368 | err_out: | ||
5369 | qc->err_mask |= AC_ERR_HSM; | ||
5370 | ap->hsm_task_state = HSM_ST_ERR; | ||
5371 | } | ||
5372 | |||
5373 | /** | ||
5374 | * ata_hsm_ok_in_wq - Check if the qc can be handled in the workqueue. | ||
5375 | * @ap: the target ata_port | ||
5376 | * @qc: qc on going | ||
5377 | * | ||
5378 | * RETURNS: | ||
5379 | * 1 if ok in workqueue, 0 otherwise. | ||
5380 | */ | ||
5381 | |||
5382 | static inline int ata_hsm_ok_in_wq(struct ata_port *ap, struct ata_queued_cmd *qc) | ||
5383 | { | ||
5384 | if (qc->tf.flags & ATA_TFLAG_POLLING) | ||
5385 | return 1; | ||
5386 | |||
5387 | if (ap->hsm_task_state == HSM_ST_FIRST) { | ||
5388 | if (qc->tf.protocol == ATA_PROT_PIO && | ||
5389 | (qc->tf.flags & ATA_TFLAG_WRITE)) | ||
5390 | return 1; | ||
5391 | |||
5392 | if (ata_is_atapi(qc->tf.protocol) && | ||
5393 | !(qc->dev->flags & ATA_DFLAG_CDB_INTR)) | ||
5394 | return 1; | ||
5395 | } | ||
5396 | |||
5397 | return 0; | ||
5398 | } | ||
5399 | |||
5400 | /** | ||
5401 | * ata_hsm_qc_complete - finish a qc running on standard HSM | ||
5402 | * @qc: Command to complete | ||
5403 | * @in_wq: 1 if called from workqueue, 0 otherwise | ||
5404 | * | ||
5405 | * Finish @qc which is running on standard HSM. | ||
5406 | * | ||
5407 | * LOCKING: | ||
5408 | * If @in_wq is zero, spin_lock_irqsave(host lock). | ||
5409 | * Otherwise, none on entry and grabs host lock. | ||
5410 | */ | ||
5411 | static void ata_hsm_qc_complete(struct ata_queued_cmd *qc, int in_wq) | ||
5412 | { | ||
5413 | struct ata_port *ap = qc->ap; | ||
5414 | unsigned long flags; | ||
5415 | |||
5416 | if (ap->ops->error_handler) { | ||
5417 | if (in_wq) { | ||
5418 | spin_lock_irqsave(ap->lock, flags); | ||
5419 | |||
5420 | /* EH might have kicked in while host lock is | ||
5421 | * released. | ||
5422 | */ | ||
5423 | qc = ata_qc_from_tag(ap, qc->tag); | ||
5424 | if (qc) { | ||
5425 | if (likely(!(qc->err_mask & AC_ERR_HSM))) { | ||
5426 | ap->ops->irq_on(ap); | ||
5427 | ata_qc_complete(qc); | ||
5428 | } else | ||
5429 | ata_port_freeze(ap); | ||
5430 | } | ||
5431 | |||
5432 | spin_unlock_irqrestore(ap->lock, flags); | ||
5433 | } else { | ||
5434 | if (likely(!(qc->err_mask & AC_ERR_HSM))) | ||
5435 | ata_qc_complete(qc); | ||
5436 | else | ||
5437 | ata_port_freeze(ap); | ||
5438 | } | ||
5439 | } else { | ||
5440 | if (in_wq) { | ||
5441 | spin_lock_irqsave(ap->lock, flags); | ||
5442 | ap->ops->irq_on(ap); | ||
5443 | ata_qc_complete(qc); | ||
5444 | spin_unlock_irqrestore(ap->lock, flags); | ||
5445 | } else | ||
5446 | ata_qc_complete(qc); | ||
5447 | } | ||
5448 | } | ||
5449 | |||
5450 | /** | ||
5451 | * ata_hsm_move - move the HSM to the next state. | ||
5452 | * @ap: the target ata_port | ||
5453 | * @qc: qc on going | ||
5454 | * @status: current device status | ||
5455 | * @in_wq: 1 if called from workqueue, 0 otherwise | ||
5456 | * | ||
5457 | * RETURNS: | ||
5458 | * 1 when poll next status needed, 0 otherwise. | ||
5459 | */ | ||
5460 | int ata_hsm_move(struct ata_port *ap, struct ata_queued_cmd *qc, | ||
5461 | u8 status, int in_wq) | ||
5462 | { | ||
5463 | unsigned long flags = 0; | ||
5464 | int poll_next; | ||
5465 | |||
5466 | WARN_ON((qc->flags & ATA_QCFLAG_ACTIVE) == 0); | ||
5467 | |||
5468 | /* Make sure ata_qc_issue_prot() does not throw things | ||
5469 | * like DMA polling into the workqueue. Notice that | ||
5470 | * in_wq is not equivalent to (qc->tf.flags & ATA_TFLAG_POLLING). | ||
5471 | */ | ||
5472 | WARN_ON(in_wq != ata_hsm_ok_in_wq(ap, qc)); | ||
5473 | |||
5474 | fsm_start: | ||
5475 | DPRINTK("ata%u: protocol %d task_state %d (dev_stat 0x%X)\n", | ||
5476 | ap->print_id, qc->tf.protocol, ap->hsm_task_state, status); | ||
5477 | |||
5478 | switch (ap->hsm_task_state) { | ||
5479 | case HSM_ST_FIRST: | ||
5480 | /* Send first data block or PACKET CDB */ | ||
5481 | |||
5482 | /* If polling, we will stay in the work queue after | ||
5483 | * sending the data. Otherwise, interrupt handler | ||
5484 | * takes over after sending the data. | ||
5485 | */ | ||
5486 | poll_next = (qc->tf.flags & ATA_TFLAG_POLLING); | ||
5487 | |||
5488 | /* check device status */ | ||
5489 | if (unlikely((status & ATA_DRQ) == 0)) { | ||
5490 | /* handle BSY=0, DRQ=0 as error */ | ||
5491 | if (likely(status & (ATA_ERR | ATA_DF))) | ||
5492 | /* device stops HSM for abort/error */ | ||
5493 | qc->err_mask |= AC_ERR_DEV; | ||
5494 | else | ||
5495 | /* HSM violation. Let EH handle this */ | ||
5496 | qc->err_mask |= AC_ERR_HSM; | ||
5497 | |||
5498 | ap->hsm_task_state = HSM_ST_ERR; | ||
5499 | goto fsm_start; | ||
5500 | } | ||
5501 | |||
5502 | /* Device should not ask for data transfer (DRQ=1) | ||
5503 | * when it finds something wrong. | ||
5504 | * We ignore DRQ here and stop the HSM by | ||
5505 | * changing hsm_task_state to HSM_ST_ERR and | ||
5506 | * let the EH abort the command or reset the device. | ||
5507 | */ | ||
5508 | if (unlikely(status & (ATA_ERR | ATA_DF))) { | ||
5509 | /* Some ATAPI tape drives forget to clear the ERR bit | ||
5510 | * when doing the next command (mostly request sense). | ||
5511 | * We ignore ERR here to workaround and proceed sending | ||
5512 | * the CDB. | ||
5513 | */ | ||
5514 | if (!(qc->dev->horkage & ATA_HORKAGE_STUCK_ERR)) { | ||
5515 | ata_port_printk(ap, KERN_WARNING, | ||
5516 | "DRQ=1 with device error, " | ||
5517 | "dev_stat 0x%X\n", status); | ||
5518 | qc->err_mask |= AC_ERR_HSM; | ||
5519 | ap->hsm_task_state = HSM_ST_ERR; | ||
5520 | goto fsm_start; | ||
5521 | } | ||
5522 | } | ||
5523 | |||
5524 | /* Send the CDB (atapi) or the first data block (ata pio out). | ||
5525 | * During the state transition, interrupt handler shouldn't | ||
5526 | * be invoked before the data transfer is complete and | ||
5527 | * hsm_task_state is changed. Hence, the following locking. | ||
5528 | */ | ||
5529 | if (in_wq) | ||
5530 | spin_lock_irqsave(ap->lock, flags); | ||
5531 | |||
5532 | if (qc->tf.protocol == ATA_PROT_PIO) { | ||
5533 | /* PIO data out protocol. | ||
5534 | * send first data block. | ||
5535 | */ | ||
5536 | |||
5537 | /* ata_pio_sectors() might change the state | ||
5538 | * to HSM_ST_LAST. so, the state is changed here | ||
5539 | * before ata_pio_sectors(). | ||
5540 | */ | ||
5541 | ap->hsm_task_state = HSM_ST; | ||
5542 | ata_pio_sectors(qc); | ||
5543 | } else | ||
5544 | /* send CDB */ | ||
5545 | atapi_send_cdb(ap, qc); | ||
5546 | |||
5547 | if (in_wq) | ||
5548 | spin_unlock_irqrestore(ap->lock, flags); | ||
5549 | |||
5550 | /* if polling, ata_pio_task() handles the rest. | ||
5551 | * otherwise, interrupt handler takes over from here. | ||
5552 | */ | ||
5553 | break; | ||
5554 | |||
5555 | case HSM_ST: | ||
5556 | /* complete command or read/write the data register */ | ||
5557 | if (qc->tf.protocol == ATAPI_PROT_PIO) { | ||
5558 | /* ATAPI PIO protocol */ | ||
5559 | if ((status & ATA_DRQ) == 0) { | ||
5560 | /* No more data to transfer or device error. | ||
5561 | * Device error will be tagged in HSM_ST_LAST. | ||
5562 | */ | ||
5563 | ap->hsm_task_state = HSM_ST_LAST; | ||
5564 | goto fsm_start; | ||
5565 | } | ||
5566 | |||
5567 | /* Device should not ask for data transfer (DRQ=1) | ||
5568 | * when it finds something wrong. | ||
5569 | * We ignore DRQ here and stop the HSM by | ||
5570 | * changing hsm_task_state to HSM_ST_ERR and | ||
5571 | * let the EH abort the command or reset the device. | ||
5572 | */ | ||
5573 | if (unlikely(status & (ATA_ERR | ATA_DF))) { | ||
5574 | ata_port_printk(ap, KERN_WARNING, "DRQ=1 with " | ||
5575 | "device error, dev_stat 0x%X\n", | ||
5576 | status); | ||
5577 | qc->err_mask |= AC_ERR_HSM; | ||
5578 | ap->hsm_task_state = HSM_ST_ERR; | ||
5579 | goto fsm_start; | ||
5580 | } | ||
5581 | |||
5582 | atapi_pio_bytes(qc); | ||
5583 | |||
5584 | if (unlikely(ap->hsm_task_state == HSM_ST_ERR)) | ||
5585 | /* bad ireason reported by device */ | ||
5586 | goto fsm_start; | ||
5587 | |||
5588 | } else { | ||
5589 | /* ATA PIO protocol */ | ||
5590 | if (unlikely((status & ATA_DRQ) == 0)) { | ||
5591 | /* handle BSY=0, DRQ=0 as error */ | ||
5592 | if (likely(status & (ATA_ERR | ATA_DF))) | ||
5593 | /* device stops HSM for abort/error */ | ||
5594 | qc->err_mask |= AC_ERR_DEV; | ||
5595 | else | ||
5596 | /* HSM violation. Let EH handle this. | ||
5597 | * Phantom devices also trigger this | ||
5598 | * condition. Mark hint. | ||
5599 | */ | ||
5600 | qc->err_mask |= AC_ERR_HSM | | ||
5601 | AC_ERR_NODEV_HINT; | ||
5602 | |||
5603 | ap->hsm_task_state = HSM_ST_ERR; | ||
5604 | goto fsm_start; | ||
5605 | } | ||
5606 | |||
5607 | /* For PIO reads, some devices may ask for | ||
5608 | * data transfer (DRQ=1) alone with ERR=1. | ||
5609 | * We respect DRQ here and transfer one | ||
5610 | * block of junk data before changing the | ||
5611 | * hsm_task_state to HSM_ST_ERR. | ||
5612 | * | ||
5613 | * For PIO writes, ERR=1 DRQ=1 doesn't make | ||
5614 | * sense since the data block has been | ||
5615 | * transferred to the device. | ||
5616 | */ | ||
5617 | if (unlikely(status & (ATA_ERR | ATA_DF))) { | ||
5618 | /* data might be corrputed */ | ||
5619 | qc->err_mask |= AC_ERR_DEV; | ||
5620 | |||
5621 | if (!(qc->tf.flags & ATA_TFLAG_WRITE)) { | ||
5622 | ata_pio_sectors(qc); | ||
5623 | status = ata_wait_idle(ap); | ||
5624 | } | ||
5625 | |||
5626 | if (status & (ATA_BUSY | ATA_DRQ)) | ||
5627 | qc->err_mask |= AC_ERR_HSM; | ||
5628 | |||
5629 | /* ata_pio_sectors() might change the | ||
5630 | * state to HSM_ST_LAST. so, the state | ||
5631 | * is changed after ata_pio_sectors(). | ||
5632 | */ | ||
5633 | ap->hsm_task_state = HSM_ST_ERR; | ||
5634 | goto fsm_start; | ||
5635 | } | ||
5636 | |||
5637 | ata_pio_sectors(qc); | ||
5638 | |||
5639 | if (ap->hsm_task_state == HSM_ST_LAST && | ||
5640 | (!(qc->tf.flags & ATA_TFLAG_WRITE))) { | ||
5641 | /* all data read */ | ||
5642 | status = ata_wait_idle(ap); | ||
5643 | goto fsm_start; | ||
5644 | } | ||
5645 | } | ||
5646 | |||
5647 | poll_next = 1; | ||
5648 | break; | ||
5649 | |||
5650 | case HSM_ST_LAST: | ||
5651 | if (unlikely(!ata_ok(status))) { | ||
5652 | qc->err_mask |= __ac_err_mask(status); | ||
5653 | ap->hsm_task_state = HSM_ST_ERR; | ||
5654 | goto fsm_start; | ||
5655 | } | ||
5656 | |||
5657 | /* no more data to transfer */ | ||
5658 | DPRINTK("ata%u: dev %u command complete, drv_stat 0x%x\n", | ||
5659 | ap->print_id, qc->dev->devno, status); | ||
5660 | |||
5661 | WARN_ON(qc->err_mask); | ||
5662 | |||
5663 | ap->hsm_task_state = HSM_ST_IDLE; | ||
5664 | |||
5665 | /* complete taskfile transaction */ | ||
5666 | ata_hsm_qc_complete(qc, in_wq); | ||
5667 | |||
5668 | poll_next = 0; | ||
5669 | break; | ||
5670 | |||
5671 | case HSM_ST_ERR: | ||
5672 | /* make sure qc->err_mask is available to | ||
5673 | * know what's wrong and recover | ||
5674 | */ | ||
5675 | WARN_ON(qc->err_mask == 0); | ||
5676 | |||
5677 | ap->hsm_task_state = HSM_ST_IDLE; | ||
5678 | |||
5679 | /* complete taskfile transaction */ | ||
5680 | ata_hsm_qc_complete(qc, in_wq); | ||
5681 | |||
5682 | poll_next = 0; | ||
5683 | break; | ||
5684 | default: | ||
5685 | poll_next = 0; | ||
5686 | BUG(); | ||
5687 | } | ||
5688 | |||
5689 | return poll_next; | ||
5690 | } | ||
5691 | |||
5692 | static void ata_pio_task(struct work_struct *work) | ||
5693 | { | ||
5694 | struct ata_port *ap = | ||
5695 | container_of(work, struct ata_port, port_task.work); | ||
5696 | struct ata_queued_cmd *qc = ap->port_task_data; | ||
5697 | u8 status; | ||
5698 | int poll_next; | ||
5699 | |||
5700 | fsm_start: | ||
5701 | WARN_ON(ap->hsm_task_state == HSM_ST_IDLE); | ||
5702 | |||
5703 | /* | ||
5704 | * This is purely heuristic. This is a fast path. | ||
5705 | * Sometimes when we enter, BSY will be cleared in | ||
5706 | * a chk-status or two. If not, the drive is probably seeking | ||
5707 | * or something. Snooze for a couple msecs, then | ||
5708 | * chk-status again. If still busy, queue delayed work. | ||
5709 | */ | ||
5710 | status = ata_busy_wait(ap, ATA_BUSY, 5); | ||
5711 | if (status & ATA_BUSY) { | ||
5712 | msleep(2); | ||
5713 | status = ata_busy_wait(ap, ATA_BUSY, 10); | ||
5714 | if (status & ATA_BUSY) { | ||
5715 | ata_pio_queue_task(ap, qc, ATA_SHORT_PAUSE); | ||
5716 | return; | ||
5717 | } | ||
5718 | } | ||
5719 | |||
5720 | /* move the HSM */ | ||
5721 | poll_next = ata_hsm_move(ap, qc, status, 1); | ||
5722 | |||
5723 | /* another command or interrupt handler | ||
5724 | * may be running at this point. | ||
5725 | */ | ||
5726 | if (poll_next) | ||
5727 | goto fsm_start; | ||
5728 | } | ||
5729 | |||
5730 | /** | ||
5731 | * ata_qc_new - Request an available ATA command, for queueing | 4430 | * ata_qc_new - Request an available ATA command, for queueing |
5732 | * @ap: Port associated with device @dev | 4431 | * @ap: Port associated with device @dev |
5733 | * @dev: Device from whom we request an available command structure | 4432 | * @dev: Device from whom we request an available command structure |
@@ -5850,7 +4549,7 @@ static void fill_result_tf(struct ata_queued_cmd *qc) | |||
5850 | struct ata_port *ap = qc->ap; | 4549 | struct ata_port *ap = qc->ap; |
5851 | 4550 | ||
5852 | qc->result_tf.flags = qc->tf.flags; | 4551 | qc->result_tf.flags = qc->tf.flags; |
5853 | ap->ops->tf_read(ap, &qc->result_tf); | 4552 | ap->ops->qc_fill_rtf(qc); |
5854 | } | 4553 | } |
5855 | 4554 | ||
5856 | static void ata_verify_xfer(struct ata_queued_cmd *qc) | 4555 | static void ata_verify_xfer(struct ata_queued_cmd *qc) |
@@ -5960,7 +4659,6 @@ void ata_qc_complete(struct ata_queued_cmd *qc) | |||
5960 | * ata_qc_complete_multiple - Complete multiple qcs successfully | 4659 | * ata_qc_complete_multiple - Complete multiple qcs successfully |
5961 | * @ap: port in question | 4660 | * @ap: port in question |
5962 | * @qc_active: new qc_active mask | 4661 | * @qc_active: new qc_active mask |
5963 | * @finish_qc: LLDD callback invoked before completing a qc | ||
5964 | * | 4662 | * |
5965 | * Complete in-flight commands. This functions is meant to be | 4663 | * Complete in-flight commands. This functions is meant to be |
5966 | * called from low-level driver's interrupt routine to complete | 4664 | * called from low-level driver's interrupt routine to complete |
@@ -5973,8 +4671,7 @@ void ata_qc_complete(struct ata_queued_cmd *qc) | |||
5973 | * RETURNS: | 4671 | * RETURNS: |
5974 | * Number of completed commands on success, -errno otherwise. | 4672 | * Number of completed commands on success, -errno otherwise. |
5975 | */ | 4673 | */ |
5976 | int ata_qc_complete_multiple(struct ata_port *ap, u32 qc_active, | 4674 | int ata_qc_complete_multiple(struct ata_port *ap, u32 qc_active) |
5977 | void (*finish_qc)(struct ata_queued_cmd *)) | ||
5978 | { | 4675 | { |
5979 | int nr_done = 0; | 4676 | int nr_done = 0; |
5980 | u32 done_mask; | 4677 | u32 done_mask; |
@@ -5995,8 +4692,6 @@ int ata_qc_complete_multiple(struct ata_port *ap, u32 qc_active, | |||
5995 | continue; | 4692 | continue; |
5996 | 4693 | ||
5997 | if ((qc = ata_qc_from_tag(ap, i))) { | 4694 | if ((qc = ata_qc_from_tag(ap, i))) { |
5998 | if (finish_qc) | ||
5999 | finish_qc(qc); | ||
6000 | ata_qc_complete(qc); | 4695 | ata_qc_complete(qc); |
6001 | nr_done++; | 4696 | nr_done++; |
6002 | } | 4697 | } |
@@ -6055,9 +4750,9 @@ void ata_qc_issue(struct ata_queued_cmd *qc) | |||
6055 | if (ata_sg_setup(qc)) | 4750 | if (ata_sg_setup(qc)) |
6056 | goto sg_err; | 4751 | goto sg_err; |
6057 | 4752 | ||
6058 | /* if device is sleeping, schedule softreset and abort the link */ | 4753 | /* if device is sleeping, schedule reset and abort the link */ |
6059 | if (unlikely(qc->dev->flags & ATA_DFLAG_SLEEPING)) { | 4754 | if (unlikely(qc->dev->flags & ATA_DFLAG_SLEEPING)) { |
6060 | link->eh_info.action |= ATA_EH_SOFTRESET; | 4755 | link->eh_info.action |= ATA_EH_RESET; |
6061 | ata_ehi_push_desc(&link->eh_info, "waking up from sleep"); | 4756 | ata_ehi_push_desc(&link->eh_info, "waking up from sleep"); |
6062 | ata_link_abort(link); | 4757 | ata_link_abort(link); |
6063 | return; | 4758 | return; |
@@ -6077,285 +4772,6 @@ err: | |||
6077 | } | 4772 | } |
6078 | 4773 | ||
6079 | /** | 4774 | /** |
6080 | * ata_qc_issue_prot - issue taskfile to device in proto-dependent manner | ||
6081 | * @qc: command to issue to device | ||
6082 | * | ||
6083 | * Using various libata functions and hooks, this function | ||
6084 | * starts an ATA command. ATA commands are grouped into | ||
6085 | * classes called "protocols", and issuing each type of protocol | ||
6086 | * is slightly different. | ||
6087 | * | ||
6088 | * May be used as the qc_issue() entry in ata_port_operations. | ||
6089 | * | ||
6090 | * LOCKING: | ||
6091 | * spin_lock_irqsave(host lock) | ||
6092 | * | ||
6093 | * RETURNS: | ||
6094 | * Zero on success, AC_ERR_* mask on failure | ||
6095 | */ | ||
6096 | |||
6097 | unsigned int ata_qc_issue_prot(struct ata_queued_cmd *qc) | ||
6098 | { | ||
6099 | struct ata_port *ap = qc->ap; | ||
6100 | |||
6101 | /* Use polling pio if the LLD doesn't handle | ||
6102 | * interrupt driven pio and atapi CDB interrupt. | ||
6103 | */ | ||
6104 | if (ap->flags & ATA_FLAG_PIO_POLLING) { | ||
6105 | switch (qc->tf.protocol) { | ||
6106 | case ATA_PROT_PIO: | ||
6107 | case ATA_PROT_NODATA: | ||
6108 | case ATAPI_PROT_PIO: | ||
6109 | case ATAPI_PROT_NODATA: | ||
6110 | qc->tf.flags |= ATA_TFLAG_POLLING; | ||
6111 | break; | ||
6112 | case ATAPI_PROT_DMA: | ||
6113 | if (qc->dev->flags & ATA_DFLAG_CDB_INTR) | ||
6114 | /* see ata_dma_blacklisted() */ | ||
6115 | BUG(); | ||
6116 | break; | ||
6117 | default: | ||
6118 | break; | ||
6119 | } | ||
6120 | } | ||
6121 | |||
6122 | /* select the device */ | ||
6123 | ata_dev_select(ap, qc->dev->devno, 1, 0); | ||
6124 | |||
6125 | /* start the command */ | ||
6126 | switch (qc->tf.protocol) { | ||
6127 | case ATA_PROT_NODATA: | ||
6128 | if (qc->tf.flags & ATA_TFLAG_POLLING) | ||
6129 | ata_qc_set_polling(qc); | ||
6130 | |||
6131 | ata_tf_to_host(ap, &qc->tf); | ||
6132 | ap->hsm_task_state = HSM_ST_LAST; | ||
6133 | |||
6134 | if (qc->tf.flags & ATA_TFLAG_POLLING) | ||
6135 | ata_pio_queue_task(ap, qc, 0); | ||
6136 | |||
6137 | break; | ||
6138 | |||
6139 | case ATA_PROT_DMA: | ||
6140 | WARN_ON(qc->tf.flags & ATA_TFLAG_POLLING); | ||
6141 | |||
6142 | ap->ops->tf_load(ap, &qc->tf); /* load tf registers */ | ||
6143 | ap->ops->bmdma_setup(qc); /* set up bmdma */ | ||
6144 | ap->ops->bmdma_start(qc); /* initiate bmdma */ | ||
6145 | ap->hsm_task_state = HSM_ST_LAST; | ||
6146 | break; | ||
6147 | |||
6148 | case ATA_PROT_PIO: | ||
6149 | if (qc->tf.flags & ATA_TFLAG_POLLING) | ||
6150 | ata_qc_set_polling(qc); | ||
6151 | |||
6152 | ata_tf_to_host(ap, &qc->tf); | ||
6153 | |||
6154 | if (qc->tf.flags & ATA_TFLAG_WRITE) { | ||
6155 | /* PIO data out protocol */ | ||
6156 | ap->hsm_task_state = HSM_ST_FIRST; | ||
6157 | ata_pio_queue_task(ap, qc, 0); | ||
6158 | |||
6159 | /* always send first data block using | ||
6160 | * the ata_pio_task() codepath. | ||
6161 | */ | ||
6162 | } else { | ||
6163 | /* PIO data in protocol */ | ||
6164 | ap->hsm_task_state = HSM_ST; | ||
6165 | |||
6166 | if (qc->tf.flags & ATA_TFLAG_POLLING) | ||
6167 | ata_pio_queue_task(ap, qc, 0); | ||
6168 | |||
6169 | /* if polling, ata_pio_task() handles the rest. | ||
6170 | * otherwise, interrupt handler takes over from here. | ||
6171 | */ | ||
6172 | } | ||
6173 | |||
6174 | break; | ||
6175 | |||
6176 | case ATAPI_PROT_PIO: | ||
6177 | case ATAPI_PROT_NODATA: | ||
6178 | if (qc->tf.flags & ATA_TFLAG_POLLING) | ||
6179 | ata_qc_set_polling(qc); | ||
6180 | |||
6181 | ata_tf_to_host(ap, &qc->tf); | ||
6182 | |||
6183 | ap->hsm_task_state = HSM_ST_FIRST; | ||
6184 | |||
6185 | /* send cdb by polling if no cdb interrupt */ | ||
6186 | if ((!(qc->dev->flags & ATA_DFLAG_CDB_INTR)) || | ||
6187 | (qc->tf.flags & ATA_TFLAG_POLLING)) | ||
6188 | ata_pio_queue_task(ap, qc, 0); | ||
6189 | break; | ||
6190 | |||
6191 | case ATAPI_PROT_DMA: | ||
6192 | WARN_ON(qc->tf.flags & ATA_TFLAG_POLLING); | ||
6193 | |||
6194 | ap->ops->tf_load(ap, &qc->tf); /* load tf registers */ | ||
6195 | ap->ops->bmdma_setup(qc); /* set up bmdma */ | ||
6196 | ap->hsm_task_state = HSM_ST_FIRST; | ||
6197 | |||
6198 | /* send cdb by polling if no cdb interrupt */ | ||
6199 | if (!(qc->dev->flags & ATA_DFLAG_CDB_INTR)) | ||
6200 | ata_pio_queue_task(ap, qc, 0); | ||
6201 | break; | ||
6202 | |||
6203 | default: | ||
6204 | WARN_ON(1); | ||
6205 | return AC_ERR_SYSTEM; | ||
6206 | } | ||
6207 | |||
6208 | return 0; | ||
6209 | } | ||
6210 | |||
6211 | /** | ||
6212 | * ata_host_intr - Handle host interrupt for given (port, task) | ||
6213 | * @ap: Port on which interrupt arrived (possibly...) | ||
6214 | * @qc: Taskfile currently active in engine | ||
6215 | * | ||
6216 | * Handle host interrupt for given queued command. Currently, | ||
6217 | * only DMA interrupts are handled. All other commands are | ||
6218 | * handled via polling with interrupts disabled (nIEN bit). | ||
6219 | * | ||
6220 | * LOCKING: | ||
6221 | * spin_lock_irqsave(host lock) | ||
6222 | * | ||
6223 | * RETURNS: | ||
6224 | * One if interrupt was handled, zero if not (shared irq). | ||
6225 | */ | ||
6226 | |||
6227 | inline unsigned int ata_host_intr(struct ata_port *ap, | ||
6228 | struct ata_queued_cmd *qc) | ||
6229 | { | ||
6230 | struct ata_eh_info *ehi = &ap->link.eh_info; | ||
6231 | u8 status, host_stat = 0; | ||
6232 | |||
6233 | VPRINTK("ata%u: protocol %d task_state %d\n", | ||
6234 | ap->print_id, qc->tf.protocol, ap->hsm_task_state); | ||
6235 | |||
6236 | /* Check whether we are expecting interrupt in this state */ | ||
6237 | switch (ap->hsm_task_state) { | ||
6238 | case HSM_ST_FIRST: | ||
6239 | /* Some pre-ATAPI-4 devices assert INTRQ | ||
6240 | * at this state when ready to receive CDB. | ||
6241 | */ | ||
6242 | |||
6243 | /* Check the ATA_DFLAG_CDB_INTR flag is enough here. | ||
6244 | * The flag was turned on only for atapi devices. No | ||
6245 | * need to check ata_is_atapi(qc->tf.protocol) again. | ||
6246 | */ | ||
6247 | if (!(qc->dev->flags & ATA_DFLAG_CDB_INTR)) | ||
6248 | goto idle_irq; | ||
6249 | break; | ||
6250 | case HSM_ST_LAST: | ||
6251 | if (qc->tf.protocol == ATA_PROT_DMA || | ||
6252 | qc->tf.protocol == ATAPI_PROT_DMA) { | ||
6253 | /* check status of DMA engine */ | ||
6254 | host_stat = ap->ops->bmdma_status(ap); | ||
6255 | VPRINTK("ata%u: host_stat 0x%X\n", | ||
6256 | ap->print_id, host_stat); | ||
6257 | |||
6258 | /* if it's not our irq... */ | ||
6259 | if (!(host_stat & ATA_DMA_INTR)) | ||
6260 | goto idle_irq; | ||
6261 | |||
6262 | /* before we do anything else, clear DMA-Start bit */ | ||
6263 | ap->ops->bmdma_stop(qc); | ||
6264 | |||
6265 | if (unlikely(host_stat & ATA_DMA_ERR)) { | ||
6266 | /* error when transfering data to/from memory */ | ||
6267 | qc->err_mask |= AC_ERR_HOST_BUS; | ||
6268 | ap->hsm_task_state = HSM_ST_ERR; | ||
6269 | } | ||
6270 | } | ||
6271 | break; | ||
6272 | case HSM_ST: | ||
6273 | break; | ||
6274 | default: | ||
6275 | goto idle_irq; | ||
6276 | } | ||
6277 | |||
6278 | /* check altstatus */ | ||
6279 | status = ata_altstatus(ap); | ||
6280 | if (status & ATA_BUSY) | ||
6281 | goto idle_irq; | ||
6282 | |||
6283 | /* check main status, clearing INTRQ */ | ||
6284 | status = ata_chk_status(ap); | ||
6285 | if (unlikely(status & ATA_BUSY)) | ||
6286 | goto idle_irq; | ||
6287 | |||
6288 | /* ack bmdma irq events */ | ||
6289 | ap->ops->irq_clear(ap); | ||
6290 | |||
6291 | ata_hsm_move(ap, qc, status, 0); | ||
6292 | |||
6293 | if (unlikely(qc->err_mask) && (qc->tf.protocol == ATA_PROT_DMA || | ||
6294 | qc->tf.protocol == ATAPI_PROT_DMA)) | ||
6295 | ata_ehi_push_desc(ehi, "BMDMA stat 0x%x", host_stat); | ||
6296 | |||
6297 | return 1; /* irq handled */ | ||
6298 | |||
6299 | idle_irq: | ||
6300 | ap->stats.idle_irq++; | ||
6301 | |||
6302 | #ifdef ATA_IRQ_TRAP | ||
6303 | if ((ap->stats.idle_irq % 1000) == 0) { | ||
6304 | ata_chk_status(ap); | ||
6305 | ap->ops->irq_clear(ap); | ||
6306 | ata_port_printk(ap, KERN_WARNING, "irq trap\n"); | ||
6307 | return 1; | ||
6308 | } | ||
6309 | #endif | ||
6310 | return 0; /* irq not handled */ | ||
6311 | } | ||
6312 | |||
6313 | /** | ||
6314 | * ata_interrupt - Default ATA host interrupt handler | ||
6315 | * @irq: irq line (unused) | ||
6316 | * @dev_instance: pointer to our ata_host information structure | ||
6317 | * | ||
6318 | * Default interrupt handler for PCI IDE devices. Calls | ||
6319 | * ata_host_intr() for each port that is not disabled. | ||
6320 | * | ||
6321 | * LOCKING: | ||
6322 | * Obtains host lock during operation. | ||
6323 | * | ||
6324 | * RETURNS: | ||
6325 | * IRQ_NONE or IRQ_HANDLED. | ||
6326 | */ | ||
6327 | |||
6328 | irqreturn_t ata_interrupt(int irq, void *dev_instance) | ||
6329 | { | ||
6330 | struct ata_host *host = dev_instance; | ||
6331 | unsigned int i; | ||
6332 | unsigned int handled = 0; | ||
6333 | unsigned long flags; | ||
6334 | |||
6335 | /* TODO: make _irqsave conditional on x86 PCI IDE legacy mode */ | ||
6336 | spin_lock_irqsave(&host->lock, flags); | ||
6337 | |||
6338 | for (i = 0; i < host->n_ports; i++) { | ||
6339 | struct ata_port *ap; | ||
6340 | |||
6341 | ap = host->ports[i]; | ||
6342 | if (ap && | ||
6343 | !(ap->flags & ATA_FLAG_DISABLED)) { | ||
6344 | struct ata_queued_cmd *qc; | ||
6345 | |||
6346 | qc = ata_qc_from_tag(ap, ap->link.active_tag); | ||
6347 | if (qc && (!(qc->tf.flags & ATA_TFLAG_POLLING)) && | ||
6348 | (qc->flags & ATA_QCFLAG_ACTIVE)) | ||
6349 | handled |= ata_host_intr(ap, qc); | ||
6350 | } | ||
6351 | } | ||
6352 | |||
6353 | spin_unlock_irqrestore(&host->lock, flags); | ||
6354 | |||
6355 | return IRQ_RETVAL(handled); | ||
6356 | } | ||
6357 | |||
6358 | /** | ||
6359 | * sata_scr_valid - test whether SCRs are accessible | 4775 | * sata_scr_valid - test whether SCRs are accessible |
6360 | * @link: ATA link to test SCR accessibility for | 4776 | * @link: ATA link to test SCR accessibility for |
6361 | * | 4777 | * |
@@ -6513,32 +4929,6 @@ int ata_link_offline(struct ata_link *link) | |||
6513 | return 0; | 4929 | return 0; |
6514 | } | 4930 | } |
6515 | 4931 | ||
6516 | int ata_flush_cache(struct ata_device *dev) | ||
6517 | { | ||
6518 | unsigned int err_mask; | ||
6519 | u8 cmd; | ||
6520 | |||
6521 | if (!ata_try_flush_cache(dev)) | ||
6522 | return 0; | ||
6523 | |||
6524 | if (dev->flags & ATA_DFLAG_FLUSH_EXT) | ||
6525 | cmd = ATA_CMD_FLUSH_EXT; | ||
6526 | else | ||
6527 | cmd = ATA_CMD_FLUSH; | ||
6528 | |||
6529 | /* This is wrong. On a failed flush we get back the LBA of the lost | ||
6530 | sector and we should (assuming it wasn't aborted as unknown) issue | ||
6531 | a further flush command to continue the writeback until it | ||
6532 | does not error */ | ||
6533 | err_mask = ata_do_simple_cmd(dev, cmd); | ||
6534 | if (err_mask) { | ||
6535 | ata_dev_printk(dev, KERN_ERR, "failed to flush cache\n"); | ||
6536 | return -EIO; | ||
6537 | } | ||
6538 | |||
6539 | return 0; | ||
6540 | } | ||
6541 | |||
6542 | #ifdef CONFIG_PM | 4932 | #ifdef CONFIG_PM |
6543 | static int ata_host_request_pm(struct ata_host *host, pm_message_t mesg, | 4933 | static int ata_host_request_pm(struct ata_host *host, pm_message_t mesg, |
6544 | unsigned int action, unsigned int ehi_flags, | 4934 | unsigned int action, unsigned int ehi_flags, |
@@ -6634,7 +5024,7 @@ int ata_host_suspend(struct ata_host *host, pm_message_t mesg) | |||
6634 | */ | 5024 | */ |
6635 | void ata_host_resume(struct ata_host *host) | 5025 | void ata_host_resume(struct ata_host *host) |
6636 | { | 5026 | { |
6637 | ata_host_request_pm(host, PMSG_ON, ATA_EH_SOFTRESET, | 5027 | ata_host_request_pm(host, PMSG_ON, ATA_EH_RESET, |
6638 | ATA_EHI_NO_AUTOPSY | ATA_EHI_QUIET, 0); | 5028 | ATA_EHI_NO_AUTOPSY | ATA_EHI_QUIET, 0); |
6639 | host->dev->power.power_state = PMSG_ON; | 5029 | host->dev->power.power_state = PMSG_ON; |
6640 | 5030 | ||
@@ -6809,7 +5199,9 @@ struct ata_port *ata_port_alloc(struct ata_host *host) | |||
6809 | ap->msg_enable = ATA_MSG_DRV | ATA_MSG_ERR | ATA_MSG_WARN; | 5199 | ap->msg_enable = ATA_MSG_DRV | ATA_MSG_ERR | ATA_MSG_WARN; |
6810 | #endif | 5200 | #endif |
6811 | 5201 | ||
5202 | #ifdef CONFIG_ATA_SFF | ||
6812 | INIT_DELAYED_WORK(&ap->port_task, ata_pio_task); | 5203 | INIT_DELAYED_WORK(&ap->port_task, ata_pio_task); |
5204 | #endif | ||
6813 | INIT_DELAYED_WORK(&ap->hotplug_task, ata_scsi_hotplug); | 5205 | INIT_DELAYED_WORK(&ap->hotplug_task, ata_scsi_hotplug); |
6814 | INIT_WORK(&ap->scsi_rescan_task, ata_scsi_dev_rescan); | 5206 | INIT_WORK(&ap->scsi_rescan_task, ata_scsi_dev_rescan); |
6815 | INIT_LIST_HEAD(&ap->eh_done_q); | 5207 | INIT_LIST_HEAD(&ap->eh_done_q); |
@@ -6959,8 +5351,6 @@ struct ata_host *ata_host_alloc_pinfo(struct device *dev, | |||
6959 | 5351 | ||
6960 | if (!host->ops && (pi->port_ops != &ata_dummy_port_ops)) | 5352 | if (!host->ops && (pi->port_ops != &ata_dummy_port_ops)) |
6961 | host->ops = pi->port_ops; | 5353 | host->ops = pi->port_ops; |
6962 | if (!host->private_data && pi->private_data) | ||
6963 | host->private_data = pi->private_data; | ||
6964 | } | 5354 | } |
6965 | 5355 | ||
6966 | return host; | 5356 | return host; |
@@ -6985,6 +5375,56 @@ static void ata_host_stop(struct device *gendev, void *res) | |||
6985 | } | 5375 | } |
6986 | 5376 | ||
6987 | /** | 5377 | /** |
5378 | * ata_finalize_port_ops - finalize ata_port_operations | ||
5379 | * @ops: ata_port_operations to finalize | ||
5380 | * | ||
5381 | * An ata_port_operations can inherit from another ops and that | ||
5382 | * ops can again inherit from another. This can go on as many | ||
5383 | * times as necessary as long as there is no loop in the | ||
5384 | * inheritance chain. | ||
5385 | * | ||
5386 | * Ops tables are finalized when the host is started. NULL or | ||
5387 | * unspecified entries are inherited from the closet ancestor | ||
5388 | * which has the method and the entry is populated with it. | ||
5389 | * After finalization, the ops table directly points to all the | ||
5390 | * methods and ->inherits is no longer necessary and cleared. | ||
5391 | * | ||
5392 | * Using ATA_OP_NULL, inheriting ops can force a method to NULL. | ||
5393 | * | ||
5394 | * LOCKING: | ||
5395 | * None. | ||
5396 | */ | ||
5397 | static void ata_finalize_port_ops(struct ata_port_operations *ops) | ||
5398 | { | ||
5399 | static spinlock_t lock = SPIN_LOCK_UNLOCKED; | ||
5400 | const struct ata_port_operations *cur; | ||
5401 | void **begin = (void **)ops; | ||
5402 | void **end = (void **)&ops->inherits; | ||
5403 | void **pp; | ||
5404 | |||
5405 | if (!ops || !ops->inherits) | ||
5406 | return; | ||
5407 | |||
5408 | spin_lock(&lock); | ||
5409 | |||
5410 | for (cur = ops->inherits; cur; cur = cur->inherits) { | ||
5411 | void **inherit = (void **)cur; | ||
5412 | |||
5413 | for (pp = begin; pp < end; pp++, inherit++) | ||
5414 | if (!*pp) | ||
5415 | *pp = *inherit; | ||
5416 | } | ||
5417 | |||
5418 | for (pp = begin; pp < end; pp++) | ||
5419 | if (IS_ERR(*pp)) | ||
5420 | *pp = NULL; | ||
5421 | |||
5422 | ops->inherits = NULL; | ||
5423 | |||
5424 | spin_unlock(&lock); | ||
5425 | } | ||
5426 | |||
5427 | /** | ||
6988 | * ata_host_start - start and freeze ports of an ATA host | 5428 | * ata_host_start - start and freeze ports of an ATA host |
6989 | * @host: ATA host to start ports for | 5429 | * @host: ATA host to start ports for |
6990 | * | 5430 | * |
@@ -7009,9 +5449,13 @@ int ata_host_start(struct ata_host *host) | |||
7009 | if (host->flags & ATA_HOST_STARTED) | 5449 | if (host->flags & ATA_HOST_STARTED) |
7010 | return 0; | 5450 | return 0; |
7011 | 5451 | ||
5452 | ata_finalize_port_ops(host->ops); | ||
5453 | |||
7012 | for (i = 0; i < host->n_ports; i++) { | 5454 | for (i = 0; i < host->n_ports; i++) { |
7013 | struct ata_port *ap = host->ports[i]; | 5455 | struct ata_port *ap = host->ports[i]; |
7014 | 5456 | ||
5457 | ata_finalize_port_ops(ap->ops); | ||
5458 | |||
7015 | if (!host->ops && !ata_port_is_dummy(ap)) | 5459 | if (!host->ops && !ata_port_is_dummy(ap)) |
7016 | host->ops = ap->ops; | 5460 | host->ops = ap->ops; |
7017 | 5461 | ||
@@ -7073,7 +5517,7 @@ int ata_host_start(struct ata_host *host) | |||
7073 | */ | 5517 | */ |
7074 | /* KILLME - the only user left is ipr */ | 5518 | /* KILLME - the only user left is ipr */ |
7075 | void ata_host_init(struct ata_host *host, struct device *dev, | 5519 | void ata_host_init(struct ata_host *host, struct device *dev, |
7076 | unsigned long flags, const struct ata_port_operations *ops) | 5520 | unsigned long flags, struct ata_port_operations *ops) |
7077 | { | 5521 | { |
7078 | spin_lock_init(&host->lock); | 5522 | spin_lock_init(&host->lock); |
7079 | host->dev = dev; | 5523 | host->dev = dev; |
@@ -7169,9 +5613,8 @@ int ata_host_register(struct ata_host *host, struct scsi_host_template *sht) | |||
7169 | /* kick EH for boot probing */ | 5613 | /* kick EH for boot probing */ |
7170 | spin_lock_irqsave(ap->lock, flags); | 5614 | spin_lock_irqsave(ap->lock, flags); |
7171 | 5615 | ||
7172 | ehi->probe_mask = | 5616 | ehi->probe_mask |= ATA_ALL_DEVICES; |
7173 | (1 << ata_link_max_devices(&ap->link)) - 1; | 5617 | ehi->action |= ATA_EH_RESET; |
7174 | ehi->action |= ATA_EH_SOFTRESET; | ||
7175 | ehi->flags |= ATA_EHI_NO_AUTOPSY | ATA_EHI_QUIET; | 5618 | ehi->flags |= ATA_EHI_NO_AUTOPSY | ATA_EHI_QUIET; |
7176 | 5619 | ||
7177 | ap->pflags &= ~ATA_PFLAG_INITIALIZING; | 5620 | ap->pflags &= ~ATA_PFLAG_INITIALIZING; |
@@ -7336,33 +5779,6 @@ void ata_host_detach(struct ata_host *host) | |||
7336 | ata_acpi_dissociate(host); | 5779 | ata_acpi_dissociate(host); |
7337 | } | 5780 | } |
7338 | 5781 | ||
7339 | /** | ||
7340 | * ata_std_ports - initialize ioaddr with standard port offsets. | ||
7341 | * @ioaddr: IO address structure to be initialized | ||
7342 | * | ||
7343 | * Utility function which initializes data_addr, error_addr, | ||
7344 | * feature_addr, nsect_addr, lbal_addr, lbam_addr, lbah_addr, | ||
7345 | * device_addr, status_addr, and command_addr to standard offsets | ||
7346 | * relative to cmd_addr. | ||
7347 | * | ||
7348 | * Does not set ctl_addr, altstatus_addr, bmdma_addr, or scr_addr. | ||
7349 | */ | ||
7350 | |||
7351 | void ata_std_ports(struct ata_ioports *ioaddr) | ||
7352 | { | ||
7353 | ioaddr->data_addr = ioaddr->cmd_addr + ATA_REG_DATA; | ||
7354 | ioaddr->error_addr = ioaddr->cmd_addr + ATA_REG_ERR; | ||
7355 | ioaddr->feature_addr = ioaddr->cmd_addr + ATA_REG_FEATURE; | ||
7356 | ioaddr->nsect_addr = ioaddr->cmd_addr + ATA_REG_NSECT; | ||
7357 | ioaddr->lbal_addr = ioaddr->cmd_addr + ATA_REG_LBAL; | ||
7358 | ioaddr->lbam_addr = ioaddr->cmd_addr + ATA_REG_LBAM; | ||
7359 | ioaddr->lbah_addr = ioaddr->cmd_addr + ATA_REG_LBAH; | ||
7360 | ioaddr->device_addr = ioaddr->cmd_addr + ATA_REG_DEVICE; | ||
7361 | ioaddr->status_addr = ioaddr->cmd_addr + ATA_REG_STATUS; | ||
7362 | ioaddr->command_addr = ioaddr->cmd_addr + ATA_REG_CMD; | ||
7363 | } | ||
7364 | |||
7365 | |||
7366 | #ifdef CONFIG_PCI | 5782 | #ifdef CONFIG_PCI |
7367 | 5783 | ||
7368 | /** | 5784 | /** |
@@ -7749,33 +6165,20 @@ u32 ata_wait_register(void __iomem *reg, u32 mask, u32 val, | |||
7749 | /* | 6165 | /* |
7750 | * Dummy port_ops | 6166 | * Dummy port_ops |
7751 | */ | 6167 | */ |
7752 | static void ata_dummy_noret(struct ata_port *ap) { } | 6168 | static unsigned int ata_dummy_qc_issue(struct ata_queued_cmd *qc) |
7753 | static int ata_dummy_ret0(struct ata_port *ap) { return 0; } | ||
7754 | static void ata_dummy_qc_noret(struct ata_queued_cmd *qc) { } | ||
7755 | |||
7756 | static u8 ata_dummy_check_status(struct ata_port *ap) | ||
7757 | { | 6169 | { |
7758 | return ATA_DRDY; | 6170 | return AC_ERR_SYSTEM; |
7759 | } | 6171 | } |
7760 | 6172 | ||
7761 | static unsigned int ata_dummy_qc_issue(struct ata_queued_cmd *qc) | 6173 | static void ata_dummy_error_handler(struct ata_port *ap) |
7762 | { | 6174 | { |
7763 | return AC_ERR_SYSTEM; | 6175 | /* truly dummy */ |
7764 | } | 6176 | } |
7765 | 6177 | ||
7766 | const struct ata_port_operations ata_dummy_port_ops = { | 6178 | struct ata_port_operations ata_dummy_port_ops = { |
7767 | .check_status = ata_dummy_check_status, | ||
7768 | .check_altstatus = ata_dummy_check_status, | ||
7769 | .dev_select = ata_noop_dev_select, | ||
7770 | .qc_prep = ata_noop_qc_prep, | 6179 | .qc_prep = ata_noop_qc_prep, |
7771 | .qc_issue = ata_dummy_qc_issue, | 6180 | .qc_issue = ata_dummy_qc_issue, |
7772 | .freeze = ata_dummy_noret, | 6181 | .error_handler = ata_dummy_error_handler, |
7773 | .thaw = ata_dummy_noret, | ||
7774 | .error_handler = ata_dummy_noret, | ||
7775 | .post_internal_cmd = ata_dummy_qc_noret, | ||
7776 | .irq_clear = ata_dummy_noret, | ||
7777 | .port_start = ata_dummy_ret0, | ||
7778 | .port_stop = ata_dummy_noret, | ||
7779 | }; | 6182 | }; |
7780 | 6183 | ||
7781 | const struct ata_port_info ata_dummy_port_info = { | 6184 | const struct ata_port_info ata_dummy_port_info = { |
@@ -7791,10 +6194,11 @@ const struct ata_port_info ata_dummy_port_info = { | |||
7791 | EXPORT_SYMBOL_GPL(sata_deb_timing_normal); | 6194 | EXPORT_SYMBOL_GPL(sata_deb_timing_normal); |
7792 | EXPORT_SYMBOL_GPL(sata_deb_timing_hotplug); | 6195 | EXPORT_SYMBOL_GPL(sata_deb_timing_hotplug); |
7793 | EXPORT_SYMBOL_GPL(sata_deb_timing_long); | 6196 | EXPORT_SYMBOL_GPL(sata_deb_timing_long); |
6197 | EXPORT_SYMBOL_GPL(ata_base_port_ops); | ||
6198 | EXPORT_SYMBOL_GPL(sata_port_ops); | ||
7794 | EXPORT_SYMBOL_GPL(ata_dummy_port_ops); | 6199 | EXPORT_SYMBOL_GPL(ata_dummy_port_ops); |
7795 | EXPORT_SYMBOL_GPL(ata_dummy_port_info); | 6200 | EXPORT_SYMBOL_GPL(ata_dummy_port_info); |
7796 | EXPORT_SYMBOL_GPL(ata_std_bios_param); | 6201 | EXPORT_SYMBOL_GPL(ata_std_bios_param); |
7797 | EXPORT_SYMBOL_GPL(ata_std_ports); | ||
7798 | EXPORT_SYMBOL_GPL(ata_host_init); | 6202 | EXPORT_SYMBOL_GPL(ata_host_init); |
7799 | EXPORT_SYMBOL_GPL(ata_host_alloc); | 6203 | EXPORT_SYMBOL_GPL(ata_host_alloc); |
7800 | EXPORT_SYMBOL_GPL(ata_host_alloc_pinfo); | 6204 | EXPORT_SYMBOL_GPL(ata_host_alloc_pinfo); |
@@ -7803,14 +6207,8 @@ EXPORT_SYMBOL_GPL(ata_host_register); | |||
7803 | EXPORT_SYMBOL_GPL(ata_host_activate); | 6207 | EXPORT_SYMBOL_GPL(ata_host_activate); |
7804 | EXPORT_SYMBOL_GPL(ata_host_detach); | 6208 | EXPORT_SYMBOL_GPL(ata_host_detach); |
7805 | EXPORT_SYMBOL_GPL(ata_sg_init); | 6209 | EXPORT_SYMBOL_GPL(ata_sg_init); |
7806 | EXPORT_SYMBOL_GPL(ata_hsm_move); | ||
7807 | EXPORT_SYMBOL_GPL(ata_qc_complete); | 6210 | EXPORT_SYMBOL_GPL(ata_qc_complete); |
7808 | EXPORT_SYMBOL_GPL(ata_qc_complete_multiple); | 6211 | EXPORT_SYMBOL_GPL(ata_qc_complete_multiple); |
7809 | EXPORT_SYMBOL_GPL(ata_qc_issue_prot); | ||
7810 | EXPORT_SYMBOL_GPL(ata_tf_load); | ||
7811 | EXPORT_SYMBOL_GPL(ata_tf_read); | ||
7812 | EXPORT_SYMBOL_GPL(ata_noop_dev_select); | ||
7813 | EXPORT_SYMBOL_GPL(ata_std_dev_select); | ||
7814 | EXPORT_SYMBOL_GPL(sata_print_link_status); | 6212 | EXPORT_SYMBOL_GPL(sata_print_link_status); |
7815 | EXPORT_SYMBOL_GPL(atapi_cmd_type); | 6213 | EXPORT_SYMBOL_GPL(atapi_cmd_type); |
7816 | EXPORT_SYMBOL_GPL(ata_tf_to_fis); | 6214 | EXPORT_SYMBOL_GPL(ata_tf_to_fis); |
@@ -7822,37 +6220,17 @@ EXPORT_SYMBOL_GPL(ata_xfer_mode2mask); | |||
7822 | EXPORT_SYMBOL_GPL(ata_xfer_mode2shift); | 6220 | EXPORT_SYMBOL_GPL(ata_xfer_mode2shift); |
7823 | EXPORT_SYMBOL_GPL(ata_mode_string); | 6221 | EXPORT_SYMBOL_GPL(ata_mode_string); |
7824 | EXPORT_SYMBOL_GPL(ata_id_xfermask); | 6222 | EXPORT_SYMBOL_GPL(ata_id_xfermask); |
7825 | EXPORT_SYMBOL_GPL(ata_check_status); | ||
7826 | EXPORT_SYMBOL_GPL(ata_altstatus); | ||
7827 | EXPORT_SYMBOL_GPL(ata_exec_command); | ||
7828 | EXPORT_SYMBOL_GPL(ata_port_start); | 6223 | EXPORT_SYMBOL_GPL(ata_port_start); |
7829 | EXPORT_SYMBOL_GPL(ata_sff_port_start); | ||
7830 | EXPORT_SYMBOL_GPL(ata_interrupt); | ||
7831 | EXPORT_SYMBOL_GPL(ata_do_set_mode); | 6224 | EXPORT_SYMBOL_GPL(ata_do_set_mode); |
7832 | EXPORT_SYMBOL_GPL(ata_data_xfer); | ||
7833 | EXPORT_SYMBOL_GPL(ata_data_xfer_noirq); | ||
7834 | EXPORT_SYMBOL_GPL(ata_std_qc_defer); | 6225 | EXPORT_SYMBOL_GPL(ata_std_qc_defer); |
7835 | EXPORT_SYMBOL_GPL(ata_qc_prep); | ||
7836 | EXPORT_SYMBOL_GPL(ata_dumb_qc_prep); | ||
7837 | EXPORT_SYMBOL_GPL(ata_noop_qc_prep); | 6226 | EXPORT_SYMBOL_GPL(ata_noop_qc_prep); |
7838 | EXPORT_SYMBOL_GPL(ata_bmdma_setup); | ||
7839 | EXPORT_SYMBOL_GPL(ata_bmdma_start); | ||
7840 | EXPORT_SYMBOL_GPL(ata_bmdma_irq_clear); | ||
7841 | EXPORT_SYMBOL_GPL(ata_bmdma_status); | ||
7842 | EXPORT_SYMBOL_GPL(ata_bmdma_stop); | ||
7843 | EXPORT_SYMBOL_GPL(ata_bmdma_freeze); | ||
7844 | EXPORT_SYMBOL_GPL(ata_bmdma_thaw); | ||
7845 | EXPORT_SYMBOL_GPL(ata_bmdma_drive_eh); | ||
7846 | EXPORT_SYMBOL_GPL(ata_bmdma_error_handler); | ||
7847 | EXPORT_SYMBOL_GPL(ata_bmdma_post_internal_cmd); | ||
7848 | EXPORT_SYMBOL_GPL(ata_port_probe); | 6227 | EXPORT_SYMBOL_GPL(ata_port_probe); |
7849 | EXPORT_SYMBOL_GPL(ata_dev_disable); | 6228 | EXPORT_SYMBOL_GPL(ata_dev_disable); |
7850 | EXPORT_SYMBOL_GPL(sata_set_spd); | 6229 | EXPORT_SYMBOL_GPL(sata_set_spd); |
6230 | EXPORT_SYMBOL_GPL(ata_wait_after_reset); | ||
7851 | EXPORT_SYMBOL_GPL(sata_link_debounce); | 6231 | EXPORT_SYMBOL_GPL(sata_link_debounce); |
7852 | EXPORT_SYMBOL_GPL(sata_link_resume); | 6232 | EXPORT_SYMBOL_GPL(sata_link_resume); |
7853 | EXPORT_SYMBOL_GPL(ata_bus_reset); | ||
7854 | EXPORT_SYMBOL_GPL(ata_std_prereset); | 6233 | EXPORT_SYMBOL_GPL(ata_std_prereset); |
7855 | EXPORT_SYMBOL_GPL(ata_std_softreset); | ||
7856 | EXPORT_SYMBOL_GPL(sata_link_hardreset); | 6234 | EXPORT_SYMBOL_GPL(sata_link_hardreset); |
7857 | EXPORT_SYMBOL_GPL(sata_std_hardreset); | 6235 | EXPORT_SYMBOL_GPL(sata_std_hardreset); |
7858 | EXPORT_SYMBOL_GPL(ata_std_postreset); | 6236 | EXPORT_SYMBOL_GPL(ata_std_postreset); |
@@ -7861,15 +6239,11 @@ EXPORT_SYMBOL_GPL(ata_dev_pair); | |||
7861 | EXPORT_SYMBOL_GPL(ata_port_disable); | 6239 | EXPORT_SYMBOL_GPL(ata_port_disable); |
7862 | EXPORT_SYMBOL_GPL(ata_ratelimit); | 6240 | EXPORT_SYMBOL_GPL(ata_ratelimit); |
7863 | EXPORT_SYMBOL_GPL(ata_wait_register); | 6241 | EXPORT_SYMBOL_GPL(ata_wait_register); |
7864 | EXPORT_SYMBOL_GPL(ata_busy_sleep); | ||
7865 | EXPORT_SYMBOL_GPL(ata_wait_after_reset); | ||
7866 | EXPORT_SYMBOL_GPL(ata_wait_ready); | ||
7867 | EXPORT_SYMBOL_GPL(ata_scsi_ioctl); | 6242 | EXPORT_SYMBOL_GPL(ata_scsi_ioctl); |
7868 | EXPORT_SYMBOL_GPL(ata_scsi_queuecmd); | 6243 | EXPORT_SYMBOL_GPL(ata_scsi_queuecmd); |
7869 | EXPORT_SYMBOL_GPL(ata_scsi_slave_config); | 6244 | EXPORT_SYMBOL_GPL(ata_scsi_slave_config); |
7870 | EXPORT_SYMBOL_GPL(ata_scsi_slave_destroy); | 6245 | EXPORT_SYMBOL_GPL(ata_scsi_slave_destroy); |
7871 | EXPORT_SYMBOL_GPL(ata_scsi_change_queue_depth); | 6246 | EXPORT_SYMBOL_GPL(ata_scsi_change_queue_depth); |
7872 | EXPORT_SYMBOL_GPL(ata_host_intr); | ||
7873 | EXPORT_SYMBOL_GPL(sata_scr_valid); | 6247 | EXPORT_SYMBOL_GPL(sata_scr_valid); |
7874 | EXPORT_SYMBOL_GPL(sata_scr_read); | 6248 | EXPORT_SYMBOL_GPL(sata_scr_read); |
7875 | EXPORT_SYMBOL_GPL(sata_scr_write); | 6249 | EXPORT_SYMBOL_GPL(sata_scr_write); |
@@ -7892,11 +6266,6 @@ EXPORT_SYMBOL_GPL(ata_timing_cycle2mode); | |||
7892 | 6266 | ||
7893 | #ifdef CONFIG_PCI | 6267 | #ifdef CONFIG_PCI |
7894 | EXPORT_SYMBOL_GPL(pci_test_config_bits); | 6268 | EXPORT_SYMBOL_GPL(pci_test_config_bits); |
7895 | EXPORT_SYMBOL_GPL(ata_pci_init_sff_host); | ||
7896 | EXPORT_SYMBOL_GPL(ata_pci_init_bmdma); | ||
7897 | EXPORT_SYMBOL_GPL(ata_pci_prepare_sff_host); | ||
7898 | EXPORT_SYMBOL_GPL(ata_pci_activate_sff_host); | ||
7899 | EXPORT_SYMBOL_GPL(ata_pci_init_one); | ||
7900 | EXPORT_SYMBOL_GPL(ata_pci_remove_one); | 6269 | EXPORT_SYMBOL_GPL(ata_pci_remove_one); |
7901 | #ifdef CONFIG_PM | 6270 | #ifdef CONFIG_PM |
7902 | EXPORT_SYMBOL_GPL(ata_pci_device_do_suspend); | 6271 | EXPORT_SYMBOL_GPL(ata_pci_device_do_suspend); |
@@ -7904,16 +6273,8 @@ EXPORT_SYMBOL_GPL(ata_pci_device_do_resume); | |||
7904 | EXPORT_SYMBOL_GPL(ata_pci_device_suspend); | 6273 | EXPORT_SYMBOL_GPL(ata_pci_device_suspend); |
7905 | EXPORT_SYMBOL_GPL(ata_pci_device_resume); | 6274 | EXPORT_SYMBOL_GPL(ata_pci_device_resume); |
7906 | #endif /* CONFIG_PM */ | 6275 | #endif /* CONFIG_PM */ |
7907 | EXPORT_SYMBOL_GPL(ata_pci_default_filter); | ||
7908 | EXPORT_SYMBOL_GPL(ata_pci_clear_simplex); | ||
7909 | #endif /* CONFIG_PCI */ | 6276 | #endif /* CONFIG_PCI */ |
7910 | 6277 | ||
7911 | EXPORT_SYMBOL_GPL(sata_pmp_qc_defer_cmd_switch); | ||
7912 | EXPORT_SYMBOL_GPL(sata_pmp_std_prereset); | ||
7913 | EXPORT_SYMBOL_GPL(sata_pmp_std_hardreset); | ||
7914 | EXPORT_SYMBOL_GPL(sata_pmp_std_postreset); | ||
7915 | EXPORT_SYMBOL_GPL(sata_pmp_do_eh); | ||
7916 | |||
7917 | EXPORT_SYMBOL_GPL(__ata_ehi_push_desc); | 6278 | EXPORT_SYMBOL_GPL(__ata_ehi_push_desc); |
7918 | EXPORT_SYMBOL_GPL(ata_ehi_push_desc); | 6279 | EXPORT_SYMBOL_GPL(ata_ehi_push_desc); |
7919 | EXPORT_SYMBOL_GPL(ata_ehi_clear_desc); | 6280 | EXPORT_SYMBOL_GPL(ata_ehi_clear_desc); |
@@ -7931,8 +6292,7 @@ EXPORT_SYMBOL_GPL(ata_eh_thaw_port); | |||
7931 | EXPORT_SYMBOL_GPL(ata_eh_qc_complete); | 6292 | EXPORT_SYMBOL_GPL(ata_eh_qc_complete); |
7932 | EXPORT_SYMBOL_GPL(ata_eh_qc_retry); | 6293 | EXPORT_SYMBOL_GPL(ata_eh_qc_retry); |
7933 | EXPORT_SYMBOL_GPL(ata_do_eh); | 6294 | EXPORT_SYMBOL_GPL(ata_do_eh); |
7934 | EXPORT_SYMBOL_GPL(ata_irq_on); | 6295 | EXPORT_SYMBOL_GPL(ata_std_error_handler); |
7935 | EXPORT_SYMBOL_GPL(ata_dev_try_classify); | ||
7936 | 6296 | ||
7937 | EXPORT_SYMBOL_GPL(ata_cable_40wire); | 6297 | EXPORT_SYMBOL_GPL(ata_cable_40wire); |
7938 | EXPORT_SYMBOL_GPL(ata_cable_80wire); | 6298 | EXPORT_SYMBOL_GPL(ata_cable_80wire); |
diff --git a/drivers/ata/libata-eh.c b/drivers/ata/libata-eh.c index a5830329eda4..d94359a24d41 100644 --- a/drivers/ata/libata-eh.c +++ b/drivers/ata/libata-eh.c | |||
@@ -873,9 +873,9 @@ int sata_async_notification(struct ata_port *ap) | |||
873 | if (rc == 0) | 873 | if (rc == 0) |
874 | sata_scr_write(&ap->link, SCR_NOTIFICATION, sntf); | 874 | sata_scr_write(&ap->link, SCR_NOTIFICATION, sntf); |
875 | 875 | ||
876 | if (!ap->nr_pmp_links || rc) { | 876 | if (!sata_pmp_attached(ap) || rc) { |
877 | /* PMP is not attached or SNTF is not available */ | 877 | /* PMP is not attached or SNTF is not available */ |
878 | if (!ap->nr_pmp_links) { | 878 | if (!sata_pmp_attached(ap)) { |
879 | /* PMP is not attached. Check whether ATAPI | 879 | /* PMP is not attached. Check whether ATAPI |
880 | * AN is configured. If so, notify media | 880 | * AN is configured. If so, notify media |
881 | * change. | 881 | * change. |
@@ -1079,19 +1079,6 @@ void ata_eh_about_to_do(struct ata_link *link, struct ata_device *dev, | |||
1079 | 1079 | ||
1080 | spin_lock_irqsave(ap->lock, flags); | 1080 | spin_lock_irqsave(ap->lock, flags); |
1081 | 1081 | ||
1082 | /* Reset is represented by combination of actions and EHI | ||
1083 | * flags. Suck in all related bits before clearing eh_info to | ||
1084 | * avoid losing requested action. | ||
1085 | */ | ||
1086 | if (action & ATA_EH_RESET_MASK) { | ||
1087 | ehc->i.action |= ehi->action & ATA_EH_RESET_MASK; | ||
1088 | ehc->i.flags |= ehi->flags & ATA_EHI_RESET_MODIFIER_MASK; | ||
1089 | |||
1090 | /* make sure all reset actions are cleared & clear EHI flags */ | ||
1091 | action |= ATA_EH_RESET_MASK; | ||
1092 | ehi->flags &= ~ATA_EHI_RESET_MODIFIER_MASK; | ||
1093 | } | ||
1094 | |||
1095 | ata_eh_clear_action(link, dev, ehi, action); | 1082 | ata_eh_clear_action(link, dev, ehi, action); |
1096 | 1083 | ||
1097 | if (!(ehc->i.flags & ATA_EHI_QUIET)) | 1084 | if (!(ehc->i.flags & ATA_EHI_QUIET)) |
@@ -1117,12 +1104,6 @@ void ata_eh_done(struct ata_link *link, struct ata_device *dev, | |||
1117 | { | 1104 | { |
1118 | struct ata_eh_context *ehc = &link->eh_context; | 1105 | struct ata_eh_context *ehc = &link->eh_context; |
1119 | 1106 | ||
1120 | /* if reset is complete, clear all reset actions & reset modifier */ | ||
1121 | if (action & ATA_EH_RESET_MASK) { | ||
1122 | action |= ATA_EH_RESET_MASK; | ||
1123 | ehc->i.flags &= ~ATA_EHI_RESET_MODIFIER_MASK; | ||
1124 | } | ||
1125 | |||
1126 | ata_eh_clear_action(link, dev, &ehc->i, action); | 1107 | ata_eh_clear_action(link, dev, &ehc->i, action); |
1127 | } | 1108 | } |
1128 | 1109 | ||
@@ -1329,20 +1310,20 @@ static void ata_eh_analyze_serror(struct ata_link *link) | |||
1329 | 1310 | ||
1330 | if (serror & SERR_PERSISTENT) { | 1311 | if (serror & SERR_PERSISTENT) { |
1331 | err_mask |= AC_ERR_ATA_BUS; | 1312 | err_mask |= AC_ERR_ATA_BUS; |
1332 | action |= ATA_EH_HARDRESET; | 1313 | action |= ATA_EH_RESET; |
1333 | } | 1314 | } |
1334 | if (serror & | 1315 | if (serror & |
1335 | (SERR_DATA_RECOVERED | SERR_COMM_RECOVERED | SERR_DATA)) { | 1316 | (SERR_DATA_RECOVERED | SERR_COMM_RECOVERED | SERR_DATA)) { |
1336 | err_mask |= AC_ERR_ATA_BUS; | 1317 | err_mask |= AC_ERR_ATA_BUS; |
1337 | action |= ATA_EH_SOFTRESET; | 1318 | action |= ATA_EH_RESET; |
1338 | } | 1319 | } |
1339 | if (serror & SERR_PROTOCOL) { | 1320 | if (serror & SERR_PROTOCOL) { |
1340 | err_mask |= AC_ERR_HSM; | 1321 | err_mask |= AC_ERR_HSM; |
1341 | action |= ATA_EH_SOFTRESET; | 1322 | action |= ATA_EH_RESET; |
1342 | } | 1323 | } |
1343 | if (serror & SERR_INTERNAL) { | 1324 | if (serror & SERR_INTERNAL) { |
1344 | err_mask |= AC_ERR_SYSTEM; | 1325 | err_mask |= AC_ERR_SYSTEM; |
1345 | action |= ATA_EH_HARDRESET; | 1326 | action |= ATA_EH_RESET; |
1346 | } | 1327 | } |
1347 | 1328 | ||
1348 | /* Determine whether a hotplug event has occurred. Both | 1329 | /* Determine whether a hotplug event has occurred. Both |
@@ -1448,7 +1429,7 @@ static unsigned int ata_eh_analyze_tf(struct ata_queued_cmd *qc, | |||
1448 | 1429 | ||
1449 | if ((stat & (ATA_BUSY | ATA_DRQ | ATA_DRDY)) != ATA_DRDY) { | 1430 | if ((stat & (ATA_BUSY | ATA_DRQ | ATA_DRDY)) != ATA_DRDY) { |
1450 | qc->err_mask |= AC_ERR_HSM; | 1431 | qc->err_mask |= AC_ERR_HSM; |
1451 | return ATA_EH_SOFTRESET; | 1432 | return ATA_EH_RESET; |
1452 | } | 1433 | } |
1453 | 1434 | ||
1454 | if (stat & (ATA_ERR | ATA_DF)) | 1435 | if (stat & (ATA_ERR | ATA_DF)) |
@@ -1484,7 +1465,7 @@ static unsigned int ata_eh_analyze_tf(struct ata_queued_cmd *qc, | |||
1484 | } | 1465 | } |
1485 | 1466 | ||
1486 | if (qc->err_mask & (AC_ERR_HSM | AC_ERR_TIMEOUT | AC_ERR_ATA_BUS)) | 1467 | if (qc->err_mask & (AC_ERR_HSM | AC_ERR_TIMEOUT | AC_ERR_ATA_BUS)) |
1487 | action |= ATA_EH_SOFTRESET; | 1468 | action |= ATA_EH_RESET; |
1488 | 1469 | ||
1489 | return action; | 1470 | return action; |
1490 | } | 1471 | } |
@@ -1685,7 +1666,7 @@ static unsigned int ata_eh_speed_down(struct ata_device *dev, | |||
1685 | if (verdict & ATA_EH_SPDN_SPEED_DOWN) { | 1666 | if (verdict & ATA_EH_SPDN_SPEED_DOWN) { |
1686 | /* speed down SATA link speed if possible */ | 1667 | /* speed down SATA link speed if possible */ |
1687 | if (sata_down_spd_limit(link) == 0) { | 1668 | if (sata_down_spd_limit(link) == 0) { |
1688 | action |= ATA_EH_HARDRESET; | 1669 | action |= ATA_EH_RESET; |
1689 | goto done; | 1670 | goto done; |
1690 | } | 1671 | } |
1691 | 1672 | ||
@@ -1705,7 +1686,7 @@ static unsigned int ata_eh_speed_down(struct ata_device *dev, | |||
1705 | dev->spdn_cnt++; | 1686 | dev->spdn_cnt++; |
1706 | 1687 | ||
1707 | if (ata_down_xfermask_limit(dev, sel) == 0) { | 1688 | if (ata_down_xfermask_limit(dev, sel) == 0) { |
1708 | action |= ATA_EH_SOFTRESET; | 1689 | action |= ATA_EH_RESET; |
1709 | goto done; | 1690 | goto done; |
1710 | } | 1691 | } |
1711 | } | 1692 | } |
@@ -1719,7 +1700,7 @@ static unsigned int ata_eh_speed_down(struct ata_device *dev, | |||
1719 | (dev->xfer_shift != ATA_SHIFT_PIO)) { | 1700 | (dev->xfer_shift != ATA_SHIFT_PIO)) { |
1720 | if (ata_down_xfermask_limit(dev, ATA_DNXFER_FORCE_PIO) == 0) { | 1701 | if (ata_down_xfermask_limit(dev, ATA_DNXFER_FORCE_PIO) == 0) { |
1721 | dev->spdn_cnt = 0; | 1702 | dev->spdn_cnt = 0; |
1722 | action |= ATA_EH_SOFTRESET; | 1703 | action |= ATA_EH_RESET; |
1723 | goto done; | 1704 | goto done; |
1724 | } | 1705 | } |
1725 | } | 1706 | } |
@@ -1764,9 +1745,9 @@ static void ata_eh_link_autopsy(struct ata_link *link) | |||
1764 | ehc->i.serror |= serror; | 1745 | ehc->i.serror |= serror; |
1765 | ata_eh_analyze_serror(link); | 1746 | ata_eh_analyze_serror(link); |
1766 | } else if (rc != -EOPNOTSUPP) { | 1747 | } else if (rc != -EOPNOTSUPP) { |
1767 | /* SError read failed, force hardreset and probing */ | 1748 | /* SError read failed, force reset and probing */ |
1768 | ata_ehi_schedule_probe(&ehc->i); | 1749 | ehc->i.probe_mask |= ATA_ALL_DEVICES; |
1769 | ehc->i.action |= ATA_EH_HARDRESET; | 1750 | ehc->i.action |= ATA_EH_RESET; |
1770 | ehc->i.err_mask |= AC_ERR_OTHER; | 1751 | ehc->i.err_mask |= AC_ERR_OTHER; |
1771 | } | 1752 | } |
1772 | 1753 | ||
@@ -1804,6 +1785,11 @@ static void ata_eh_link_autopsy(struct ata_link *link) | |||
1804 | if (qc->flags & ATA_QCFLAG_SENSE_VALID) | 1785 | if (qc->flags & ATA_QCFLAG_SENSE_VALID) |
1805 | qc->err_mask &= ~(AC_ERR_DEV | AC_ERR_OTHER); | 1786 | qc->err_mask &= ~(AC_ERR_DEV | AC_ERR_OTHER); |
1806 | 1787 | ||
1788 | /* determine whether the command is worth retrying */ | ||
1789 | if (!(qc->err_mask & AC_ERR_INVALID) && | ||
1790 | ((qc->flags & ATA_QCFLAG_IO) || qc->err_mask != AC_ERR_DEV)) | ||
1791 | qc->flags |= ATA_QCFLAG_RETRY; | ||
1792 | |||
1807 | /* accumulate error info */ | 1793 | /* accumulate error info */ |
1808 | ehc->i.dev = qc->dev; | 1794 | ehc->i.dev = qc->dev; |
1809 | all_err_mask |= qc->err_mask; | 1795 | all_err_mask |= qc->err_mask; |
@@ -1814,7 +1800,7 @@ static void ata_eh_link_autopsy(struct ata_link *link) | |||
1814 | /* enforce default EH actions */ | 1800 | /* enforce default EH actions */ |
1815 | if (ap->pflags & ATA_PFLAG_FROZEN || | 1801 | if (ap->pflags & ATA_PFLAG_FROZEN || |
1816 | all_err_mask & (AC_ERR_HSM | AC_ERR_TIMEOUT)) | 1802 | all_err_mask & (AC_ERR_HSM | AC_ERR_TIMEOUT)) |
1817 | ehc->i.action |= ATA_EH_SOFTRESET; | 1803 | ehc->i.action |= ATA_EH_RESET; |
1818 | else if (((eflags & ATA_EFLAG_IS_IO) && all_err_mask) || | 1804 | else if (((eflags & ATA_EFLAG_IS_IO) && all_err_mask) || |
1819 | (!(eflags & ATA_EFLAG_IS_IO) && (all_err_mask & ~AC_ERR_DEV))) | 1805 | (!(eflags & ATA_EFLAG_IS_IO) && (all_err_mask & ~AC_ERR_DEV))) |
1820 | ehc->i.action |= ATA_EH_REVALIDATE; | 1806 | ehc->i.action |= ATA_EH_REVALIDATE; |
@@ -1867,7 +1853,7 @@ void ata_eh_autopsy(struct ata_port *ap) | |||
1867 | /* Autopsy of fanout ports can affect host link autopsy. | 1853 | /* Autopsy of fanout ports can affect host link autopsy. |
1868 | * Perform host link autopsy last. | 1854 | * Perform host link autopsy last. |
1869 | */ | 1855 | */ |
1870 | if (ap->nr_pmp_links) | 1856 | if (sata_pmp_attached(ap)) |
1871 | ata_eh_link_autopsy(&ap->link); | 1857 | ata_eh_link_autopsy(&ap->link); |
1872 | } | 1858 | } |
1873 | 1859 | ||
@@ -2066,41 +2052,29 @@ static int ata_do_reset(struct ata_link *link, ata_reset_fn_t reset, | |||
2066 | classes[dev->devno] = ATA_DEV_UNKNOWN; | 2052 | classes[dev->devno] = ATA_DEV_UNKNOWN; |
2067 | 2053 | ||
2068 | rc = reset(link, classes, deadline); | 2054 | rc = reset(link, classes, deadline); |
2069 | if (rc) | ||
2070 | return rc; | ||
2071 | 2055 | ||
2072 | /* If any class isn't ATA_DEV_UNKNOWN, consider classification | 2056 | /* convert all ATA_DEV_UNKNOWN to ATA_DEV_NONE */ |
2073 | * is complete and convert all ATA_DEV_UNKNOWN to | ||
2074 | * ATA_DEV_NONE. | ||
2075 | */ | ||
2076 | ata_link_for_each_dev(dev, link) | 2057 | ata_link_for_each_dev(dev, link) |
2077 | if (classes[dev->devno] != ATA_DEV_UNKNOWN) | 2058 | if (classes[dev->devno] == ATA_DEV_UNKNOWN) |
2078 | break; | 2059 | classes[dev->devno] = ATA_DEV_NONE; |
2079 | |||
2080 | if (dev) { | ||
2081 | ata_link_for_each_dev(dev, link) { | ||
2082 | if (classes[dev->devno] == ATA_DEV_UNKNOWN) | ||
2083 | classes[dev->devno] = ATA_DEV_NONE; | ||
2084 | } | ||
2085 | } | ||
2086 | 2060 | ||
2087 | return 0; | 2061 | return rc; |
2088 | } | 2062 | } |
2089 | 2063 | ||
2090 | static int ata_eh_followup_srst_needed(struct ata_link *link, | 2064 | static int ata_eh_followup_srst_needed(struct ata_link *link, |
2091 | int rc, int classify, | 2065 | int rc, int classify, |
2092 | const unsigned int *classes) | 2066 | const unsigned int *classes) |
2093 | { | 2067 | { |
2094 | if (link->flags & ATA_LFLAG_NO_SRST) | 2068 | if ((link->flags & ATA_LFLAG_NO_SRST) || ata_link_offline(link)) |
2095 | return 0; | 2069 | return 0; |
2096 | if (rc == -EAGAIN) | 2070 | if (rc == -EAGAIN) { |
2097 | return 1; | 2071 | if (classify) |
2072 | return 1; | ||
2073 | rc = 0; | ||
2074 | } | ||
2098 | if (rc != 0) | 2075 | if (rc != 0) |
2099 | return 0; | 2076 | return 0; |
2100 | if ((link->ap->flags & ATA_FLAG_PMP) && ata_is_host_link(link)) | 2077 | if (sata_pmp_supported(link->ap) && ata_is_host_link(link)) |
2101 | return 1; | ||
2102 | if (classify && !(link->flags & ATA_LFLAG_ASSUME_CLASS) && | ||
2103 | classes[0] == ATA_DEV_UNKNOWN) | ||
2104 | return 1; | 2078 | return 1; |
2105 | return 0; | 2079 | return 0; |
2106 | } | 2080 | } |
@@ -2118,7 +2092,6 @@ int ata_eh_reset(struct ata_link *link, int classify, | |||
2118 | int try = 0; | 2092 | int try = 0; |
2119 | struct ata_device *dev; | 2093 | struct ata_device *dev; |
2120 | unsigned long deadline, now; | 2094 | unsigned long deadline, now; |
2121 | unsigned int tmp_action; | ||
2122 | ata_reset_fn_t reset; | 2095 | ata_reset_fn_t reset; |
2123 | unsigned long flags; | 2096 | unsigned long flags; |
2124 | u32 sstatus; | 2097 | u32 sstatus; |
@@ -2129,7 +2102,7 @@ int ata_eh_reset(struct ata_link *link, int classify, | |||
2129 | ap->pflags |= ATA_PFLAG_RESETTING; | 2102 | ap->pflags |= ATA_PFLAG_RESETTING; |
2130 | spin_unlock_irqrestore(ap->lock, flags); | 2103 | spin_unlock_irqrestore(ap->lock, flags); |
2131 | 2104 | ||
2132 | ata_eh_about_to_do(link, NULL, ehc->i.action & ATA_EH_RESET_MASK); | 2105 | ata_eh_about_to_do(link, NULL, ATA_EH_RESET); |
2133 | 2106 | ||
2134 | ata_link_for_each_dev(dev, link) { | 2107 | ata_link_for_each_dev(dev, link) { |
2135 | /* If we issue an SRST then an ATA drive (not ATAPI) | 2108 | /* If we issue an SRST then an ATA drive (not ATAPI) |
@@ -2159,17 +2132,20 @@ int ata_eh_reset(struct ata_link *link, int classify, | |||
2159 | goto done; | 2132 | goto done; |
2160 | } | 2133 | } |
2161 | 2134 | ||
2162 | /* Determine which reset to use and record in ehc->i.action. | 2135 | /* prefer hardreset */ |
2163 | * prereset() may examine and modify it. | 2136 | ehc->i.action &= ~ATA_EH_RESET; |
2164 | */ | 2137 | if (hardreset) { |
2165 | if (softreset && (!hardreset || (!(lflags & ATA_LFLAG_NO_SRST) && | 2138 | reset = hardreset; |
2166 | !sata_set_spd_needed(link) && | 2139 | ehc->i.action = ATA_EH_HARDRESET; |
2167 | !(ehc->i.action & ATA_EH_HARDRESET)))) | 2140 | } else if (softreset) { |
2168 | tmp_action = ATA_EH_SOFTRESET; | 2141 | reset = softreset; |
2169 | else | 2142 | ehc->i.action = ATA_EH_SOFTRESET; |
2170 | tmp_action = ATA_EH_HARDRESET; | 2143 | } else { |
2171 | 2144 | ata_link_printk(link, KERN_ERR, "BUG: no reset method, " | |
2172 | ehc->i.action = (ehc->i.action & ~ATA_EH_RESET_MASK) | tmp_action; | 2145 | "please report to linux-ide@vger.kernel.org\n"); |
2146 | dump_stack(); | ||
2147 | return -EINVAL; | ||
2148 | } | ||
2173 | 2149 | ||
2174 | if (prereset) { | 2150 | if (prereset) { |
2175 | rc = prereset(link, jiffies + ATA_EH_PRERESET_TIMEOUT); | 2151 | rc = prereset(link, jiffies + ATA_EH_PRERESET_TIMEOUT); |
@@ -2177,7 +2153,7 @@ int ata_eh_reset(struct ata_link *link, int classify, | |||
2177 | if (rc == -ENOENT) { | 2153 | if (rc == -ENOENT) { |
2178 | ata_link_printk(link, KERN_DEBUG, | 2154 | ata_link_printk(link, KERN_DEBUG, |
2179 | "port disabled. ignoring.\n"); | 2155 | "port disabled. ignoring.\n"); |
2180 | ehc->i.action &= ~ATA_EH_RESET_MASK; | 2156 | ehc->i.action &= ~ATA_EH_RESET; |
2181 | 2157 | ||
2182 | ata_link_for_each_dev(dev, link) | 2158 | ata_link_for_each_dev(dev, link) |
2183 | classes[dev->devno] = ATA_DEV_NONE; | 2159 | classes[dev->devno] = ATA_DEV_NONE; |
@@ -2190,12 +2166,8 @@ int ata_eh_reset(struct ata_link *link, int classify, | |||
2190 | } | 2166 | } |
2191 | } | 2167 | } |
2192 | 2168 | ||
2193 | /* prereset() might have modified ehc->i.action */ | 2169 | /* prereset() might have cleared ATA_EH_RESET */ |
2194 | if (ehc->i.action & ATA_EH_HARDRESET) | 2170 | if (!(ehc->i.action & ATA_EH_RESET)) { |
2195 | reset = hardreset; | ||
2196 | else if (ehc->i.action & ATA_EH_SOFTRESET) | ||
2197 | reset = softreset; | ||
2198 | else { | ||
2199 | /* prereset told us not to reset, bang classes and return */ | 2171 | /* prereset told us not to reset, bang classes and return */ |
2200 | ata_link_for_each_dev(dev, link) | 2172 | ata_link_for_each_dev(dev, link) |
2201 | classes[dev->devno] = ATA_DEV_NONE; | 2173 | classes[dev->devno] = ATA_DEV_NONE; |
@@ -2203,14 +2175,6 @@ int ata_eh_reset(struct ata_link *link, int classify, | |||
2203 | goto out; | 2175 | goto out; |
2204 | } | 2176 | } |
2205 | 2177 | ||
2206 | /* did prereset() screw up? if so, fix up to avoid oopsing */ | ||
2207 | if (!reset) { | ||
2208 | if (softreset) | ||
2209 | reset = softreset; | ||
2210 | else | ||
2211 | reset = hardreset; | ||
2212 | } | ||
2213 | |||
2214 | retry: | 2178 | retry: |
2215 | deadline = jiffies + ata_eh_reset_timeouts[try++]; | 2179 | deadline = jiffies + ata_eh_reset_timeouts[try++]; |
2216 | 2180 | ||
@@ -2240,7 +2204,7 @@ int ata_eh_reset(struct ata_link *link, int classify, | |||
2240 | goto fail; | 2204 | goto fail; |
2241 | } | 2205 | } |
2242 | 2206 | ||
2243 | ata_eh_about_to_do(link, NULL, ATA_EH_RESET_MASK); | 2207 | ata_eh_about_to_do(link, NULL, ATA_EH_RESET); |
2244 | rc = ata_do_reset(link, reset, classes, deadline); | 2208 | rc = ata_do_reset(link, reset, classes, deadline); |
2245 | } | 2209 | } |
2246 | 2210 | ||
@@ -2248,21 +2212,6 @@ int ata_eh_reset(struct ata_link *link, int classify, | |||
2248 | if (rc && rc != -EAGAIN) | 2212 | if (rc && rc != -EAGAIN) |
2249 | goto fail; | 2213 | goto fail; |
2250 | 2214 | ||
2251 | /* was classification successful? */ | ||
2252 | if (classify && classes[0] == ATA_DEV_UNKNOWN && | ||
2253 | !(lflags & ATA_LFLAG_ASSUME_CLASS)) { | ||
2254 | if (try < max_tries) { | ||
2255 | ata_link_printk(link, KERN_WARNING, | ||
2256 | "classification failed\n"); | ||
2257 | rc = -EINVAL; | ||
2258 | goto fail; | ||
2259 | } | ||
2260 | |||
2261 | ata_link_printk(link, KERN_WARNING, | ||
2262 | "classfication failed, assuming ATA\n"); | ||
2263 | lflags |= ATA_LFLAG_ASSUME_ATA; | ||
2264 | } | ||
2265 | |||
2266 | done: | 2215 | done: |
2267 | ata_link_for_each_dev(dev, link) { | 2216 | ata_link_for_each_dev(dev, link) { |
2268 | /* After the reset, the device state is PIO 0 and the | 2217 | /* After the reset, the device state is PIO 0 and the |
@@ -2290,7 +2239,7 @@ int ata_eh_reset(struct ata_link *link, int classify, | |||
2290 | postreset(link, classes); | 2239 | postreset(link, classes); |
2291 | 2240 | ||
2292 | /* reset successful, schedule revalidation */ | 2241 | /* reset successful, schedule revalidation */ |
2293 | ata_eh_done(link, NULL, ehc->i.action & ATA_EH_RESET_MASK); | 2242 | ata_eh_done(link, NULL, ATA_EH_RESET); |
2294 | ehc->i.action |= ATA_EH_REVALIDATE; | 2243 | ehc->i.action |= ATA_EH_REVALIDATE; |
2295 | 2244 | ||
2296 | rc = 0; | 2245 | rc = 0; |
@@ -2305,6 +2254,11 @@ int ata_eh_reset(struct ata_link *link, int classify, | |||
2305 | return rc; | 2254 | return rc; |
2306 | 2255 | ||
2307 | fail: | 2256 | fail: |
2257 | /* if SCR isn't accessible on a fan-out port, PMP needs to be reset */ | ||
2258 | if (!ata_is_host_link(link) && | ||
2259 | sata_scr_read(link, SCR_STATUS, &sstatus)) | ||
2260 | rc = -ERESTART; | ||
2261 | |||
2308 | if (rc == -ERESTART || try >= max_tries) | 2262 | if (rc == -ERESTART || try >= max_tries) |
2309 | goto out; | 2263 | goto out; |
2310 | 2264 | ||
@@ -2515,6 +2469,7 @@ static int ata_link_nr_vacant(struct ata_link *link) | |||
2515 | 2469 | ||
2516 | static int ata_eh_skip_recovery(struct ata_link *link) | 2470 | static int ata_eh_skip_recovery(struct ata_link *link) |
2517 | { | 2471 | { |
2472 | struct ata_port *ap = link->ap; | ||
2518 | struct ata_eh_context *ehc = &link->eh_context; | 2473 | struct ata_eh_context *ehc = &link->eh_context; |
2519 | struct ata_device *dev; | 2474 | struct ata_device *dev; |
2520 | 2475 | ||
@@ -2522,9 +2477,13 @@ static int ata_eh_skip_recovery(struct ata_link *link) | |||
2522 | if (link->flags & ATA_LFLAG_DISABLED) | 2477 | if (link->flags & ATA_LFLAG_DISABLED) |
2523 | return 1; | 2478 | return 1; |
2524 | 2479 | ||
2525 | /* thaw frozen port, resume link and recover failed devices */ | 2480 | /* thaw frozen port and recover failed devices */ |
2526 | if ((link->ap->pflags & ATA_PFLAG_FROZEN) || | 2481 | if ((ap->pflags & ATA_PFLAG_FROZEN) || ata_link_nr_enabled(link)) |
2527 | (ehc->i.flags & ATA_EHI_RESUME_LINK) || ata_link_nr_enabled(link)) | 2482 | return 0; |
2483 | |||
2484 | /* reset at least once if reset is requested */ | ||
2485 | if ((ehc->i.action & ATA_EH_RESET) && | ||
2486 | !(ehc->i.flags & ATA_EHI_DID_RESET)) | ||
2528 | return 0; | 2487 | return 0; |
2529 | 2488 | ||
2530 | /* skip if class codes for all vacant slots are ATA_DEV_NONE */ | 2489 | /* skip if class codes for all vacant slots are ATA_DEV_NONE */ |
@@ -2548,7 +2507,7 @@ static int ata_eh_schedule_probe(struct ata_device *dev) | |||
2548 | ata_eh_detach_dev(dev); | 2507 | ata_eh_detach_dev(dev); |
2549 | ata_dev_init(dev); | 2508 | ata_dev_init(dev); |
2550 | ehc->did_probe_mask |= (1 << dev->devno); | 2509 | ehc->did_probe_mask |= (1 << dev->devno); |
2551 | ehc->i.action |= ATA_EH_SOFTRESET; | 2510 | ehc->i.action |= ATA_EH_RESET; |
2552 | ehc->saved_xfer_mode[dev->devno] = 0; | 2511 | ehc->saved_xfer_mode[dev->devno] = 0; |
2553 | ehc->saved_ncq_enabled &= ~(1 << dev->devno); | 2512 | ehc->saved_ncq_enabled &= ~(1 << dev->devno); |
2554 | 2513 | ||
@@ -2592,12 +2551,7 @@ static int ata_eh_handle_dev_fail(struct ata_device *dev, int err) | |||
2592 | 2551 | ||
2593 | return 1; | 2552 | return 1; |
2594 | } else { | 2553 | } else { |
2595 | /* soft didn't work? be haaaaard */ | 2554 | ehc->i.action |= ATA_EH_RESET; |
2596 | if (ehc->i.flags & ATA_EHI_DID_RESET) | ||
2597 | ehc->i.action |= ATA_EH_HARDRESET; | ||
2598 | else | ||
2599 | ehc->i.action |= ATA_EH_SOFTRESET; | ||
2600 | |||
2601 | return 0; | 2555 | return 0; |
2602 | } | 2556 | } |
2603 | } | 2557 | } |
@@ -2690,7 +2644,7 @@ int ata_eh_recover(struct ata_port *ap, ata_prereset_fn_t prereset, | |||
2690 | ehc->i.action = 0; | 2644 | ehc->i.action = 0; |
2691 | 2645 | ||
2692 | /* do we need to reset? */ | 2646 | /* do we need to reset? */ |
2693 | if (ehc->i.action & ATA_EH_RESET_MASK) | 2647 | if (ehc->i.action & ATA_EH_RESET) |
2694 | reset = 1; | 2648 | reset = 1; |
2695 | 2649 | ||
2696 | ata_link_for_each_dev(dev, link) | 2650 | ata_link_for_each_dev(dev, link) |
@@ -2702,13 +2656,13 @@ int ata_eh_recover(struct ata_port *ap, ata_prereset_fn_t prereset, | |||
2702 | /* if PMP is attached, this function only deals with | 2656 | /* if PMP is attached, this function only deals with |
2703 | * downstream links, port should stay thawed. | 2657 | * downstream links, port should stay thawed. |
2704 | */ | 2658 | */ |
2705 | if (!ap->nr_pmp_links) | 2659 | if (!sata_pmp_attached(ap)) |
2706 | ata_eh_freeze_port(ap); | 2660 | ata_eh_freeze_port(ap); |
2707 | 2661 | ||
2708 | ata_port_for_each_link(link, ap) { | 2662 | ata_port_for_each_link(link, ap) { |
2709 | struct ata_eh_context *ehc = &link->eh_context; | 2663 | struct ata_eh_context *ehc = &link->eh_context; |
2710 | 2664 | ||
2711 | if (!(ehc->i.action & ATA_EH_RESET_MASK)) | 2665 | if (!(ehc->i.action & ATA_EH_RESET)) |
2712 | continue; | 2666 | continue; |
2713 | 2667 | ||
2714 | rc = ata_eh_reset(link, ata_link_nr_vacant(link), | 2668 | rc = ata_eh_reset(link, ata_link_nr_vacant(link), |
@@ -2721,7 +2675,7 @@ int ata_eh_recover(struct ata_port *ap, ata_prereset_fn_t prereset, | |||
2721 | } | 2675 | } |
2722 | } | 2676 | } |
2723 | 2677 | ||
2724 | if (!ap->nr_pmp_links) | 2678 | if (!sata_pmp_attached(ap)) |
2725 | ata_eh_thaw_port(ap); | 2679 | ata_eh_thaw_port(ap); |
2726 | } | 2680 | } |
2727 | 2681 | ||
@@ -2765,7 +2719,7 @@ dev_fail: | |||
2765 | /* PMP reset requires working host port. | 2719 | /* PMP reset requires working host port. |
2766 | * Can't retry if it's frozen. | 2720 | * Can't retry if it's frozen. |
2767 | */ | 2721 | */ |
2768 | if (ap->nr_pmp_links) | 2722 | if (sata_pmp_attached(ap)) |
2769 | goto out; | 2723 | goto out; |
2770 | break; | 2724 | break; |
2771 | } | 2725 | } |
@@ -2817,18 +2771,11 @@ void ata_eh_finish(struct ata_port *ap) | |||
2817 | /* FIXME: Once EH migration is complete, | 2771 | /* FIXME: Once EH migration is complete, |
2818 | * generate sense data in this function, | 2772 | * generate sense data in this function, |
2819 | * considering both err_mask and tf. | 2773 | * considering both err_mask and tf. |
2820 | * | ||
2821 | * There's no point in retrying invalid | ||
2822 | * (detected by libata) and non-IO device | ||
2823 | * errors (rejected by device). Finish them | ||
2824 | * immediately. | ||
2825 | */ | 2774 | */ |
2826 | if ((qc->err_mask & AC_ERR_INVALID) || | 2775 | if (qc->flags & ATA_QCFLAG_RETRY) |
2827 | (!(qc->flags & ATA_QCFLAG_IO) && | ||
2828 | qc->err_mask == AC_ERR_DEV)) | ||
2829 | ata_eh_qc_complete(qc); | ||
2830 | else | ||
2831 | ata_eh_qc_retry(qc); | 2776 | ata_eh_qc_retry(qc); |
2777 | else | ||
2778 | ata_eh_qc_complete(qc); | ||
2832 | } else { | 2779 | } else { |
2833 | if (qc->flags & ATA_QCFLAG_SENSE_VALID) { | 2780 | if (qc->flags & ATA_QCFLAG_SENSE_VALID) { |
2834 | ata_eh_qc_complete(qc); | 2781 | ata_eh_qc_complete(qc); |
@@ -2848,6 +2795,7 @@ void ata_eh_finish(struct ata_port *ap) | |||
2848 | /** | 2795 | /** |
2849 | * ata_do_eh - do standard error handling | 2796 | * ata_do_eh - do standard error handling |
2850 | * @ap: host port to handle error for | 2797 | * @ap: host port to handle error for |
2798 | * | ||
2851 | * @prereset: prereset method (can be NULL) | 2799 | * @prereset: prereset method (can be NULL) |
2852 | * @softreset: softreset method (can be NULL) | 2800 | * @softreset: softreset method (can be NULL) |
2853 | * @hardreset: hardreset method (can be NULL) | 2801 | * @hardreset: hardreset method (can be NULL) |
@@ -2878,6 +2826,27 @@ void ata_do_eh(struct ata_port *ap, ata_prereset_fn_t prereset, | |||
2878 | ata_eh_finish(ap); | 2826 | ata_eh_finish(ap); |
2879 | } | 2827 | } |
2880 | 2828 | ||
2829 | /** | ||
2830 | * ata_std_error_handler - standard error handler | ||
2831 | * @ap: host port to handle error for | ||
2832 | * | ||
2833 | * Standard error handler | ||
2834 | * | ||
2835 | * LOCKING: | ||
2836 | * Kernel thread context (may sleep). | ||
2837 | */ | ||
2838 | void ata_std_error_handler(struct ata_port *ap) | ||
2839 | { | ||
2840 | struct ata_port_operations *ops = ap->ops; | ||
2841 | ata_reset_fn_t hardreset = ops->hardreset; | ||
2842 | |||
2843 | /* ignore built-in hardreset if SCR access is not available */ | ||
2844 | if (ata_is_builtin_hardreset(hardreset) && !sata_scr_valid(&ap->link)) | ||
2845 | hardreset = NULL; | ||
2846 | |||
2847 | ata_do_eh(ap, ops->prereset, ops->softreset, hardreset, ops->postreset); | ||
2848 | } | ||
2849 | |||
2881 | #ifdef CONFIG_PM | 2850 | #ifdef CONFIG_PM |
2882 | /** | 2851 | /** |
2883 | * ata_eh_handle_port_suspend - perform port suspend operation | 2852 | * ata_eh_handle_port_suspend - perform port suspend operation |
diff --git a/drivers/ata/libata-pmp.c b/drivers/ata/libata-pmp.c index d91f5090ba9d..ff1822a7da38 100644 --- a/drivers/ata/libata-pmp.c +++ b/drivers/ata/libata-pmp.c | |||
@@ -11,6 +11,14 @@ | |||
11 | #include <linux/libata.h> | 11 | #include <linux/libata.h> |
12 | #include "libata.h" | 12 | #include "libata.h" |
13 | 13 | ||
14 | const struct ata_port_operations sata_pmp_port_ops = { | ||
15 | .inherits = &sata_port_ops, | ||
16 | .pmp_prereset = ata_std_prereset, | ||
17 | .pmp_hardreset = sata_std_hardreset, | ||
18 | .pmp_postreset = ata_std_postreset, | ||
19 | .error_handler = sata_pmp_error_handler, | ||
20 | }; | ||
21 | |||
14 | /** | 22 | /** |
15 | * sata_pmp_read - read PMP register | 23 | * sata_pmp_read - read PMP register |
16 | * @link: link to read PMP register for | 24 | * @link: link to read PMP register for |
@@ -176,140 +184,6 @@ int sata_pmp_scr_write(struct ata_link *link, int reg, u32 val) | |||
176 | } | 184 | } |
177 | 185 | ||
178 | /** | 186 | /** |
179 | * sata_pmp_std_prereset - prepare PMP link for reset | ||
180 | * @link: link to be reset | ||
181 | * @deadline: deadline jiffies for the operation | ||
182 | * | ||
183 | * @link is about to be reset. Initialize it. | ||
184 | * | ||
185 | * LOCKING: | ||
186 | * Kernel thread context (may sleep) | ||
187 | * | ||
188 | * RETURNS: | ||
189 | * 0 on success, -errno otherwise. | ||
190 | */ | ||
191 | int sata_pmp_std_prereset(struct ata_link *link, unsigned long deadline) | ||
192 | { | ||
193 | struct ata_eh_context *ehc = &link->eh_context; | ||
194 | const unsigned long *timing = sata_ehc_deb_timing(ehc); | ||
195 | int rc; | ||
196 | |||
197 | /* force HRST? */ | ||
198 | if (link->flags & ATA_LFLAG_NO_SRST) | ||
199 | ehc->i.action |= ATA_EH_HARDRESET; | ||
200 | |||
201 | /* handle link resume */ | ||
202 | if ((ehc->i.flags & ATA_EHI_RESUME_LINK) && | ||
203 | (link->flags & ATA_LFLAG_HRST_TO_RESUME)) | ||
204 | ehc->i.action |= ATA_EH_HARDRESET; | ||
205 | |||
206 | /* if we're about to do hardreset, nothing more to do */ | ||
207 | if (ehc->i.action & ATA_EH_HARDRESET) | ||
208 | return 0; | ||
209 | |||
210 | /* resume link */ | ||
211 | rc = sata_link_resume(link, timing, deadline); | ||
212 | if (rc) { | ||
213 | /* phy resume failed */ | ||
214 | ata_link_printk(link, KERN_WARNING, "failed to resume link " | ||
215 | "for reset (errno=%d)\n", rc); | ||
216 | return rc; | ||
217 | } | ||
218 | |||
219 | /* clear SError bits including .X which blocks the port when set */ | ||
220 | rc = sata_scr_write(link, SCR_ERROR, 0xffffffff); | ||
221 | if (rc) { | ||
222 | ata_link_printk(link, KERN_ERR, | ||
223 | "failed to clear SError (errno=%d)\n", rc); | ||
224 | return rc; | ||
225 | } | ||
226 | |||
227 | return 0; | ||
228 | } | ||
229 | |||
230 | /** | ||
231 | * sata_pmp_std_hardreset - standard hardreset method for PMP link | ||
232 | * @link: link to be reset | ||
233 | * @class: resulting class of attached device | ||
234 | * @deadline: deadline jiffies for the operation | ||
235 | * | ||
236 | * Hardreset PMP port @link. Note that this function doesn't | ||
237 | * wait for BSY clearance. There simply isn't a generic way to | ||
238 | * wait the event. Instead, this function return -EAGAIN thus | ||
239 | * telling libata-EH to followup with softreset. | ||
240 | * | ||
241 | * LOCKING: | ||
242 | * Kernel thread context (may sleep) | ||
243 | * | ||
244 | * RETURNS: | ||
245 | * 0 on success, -errno otherwise. | ||
246 | */ | ||
247 | int sata_pmp_std_hardreset(struct ata_link *link, unsigned int *class, | ||
248 | unsigned long deadline) | ||
249 | { | ||
250 | const unsigned long *timing = sata_ehc_deb_timing(&link->eh_context); | ||
251 | u32 tmp; | ||
252 | int rc; | ||
253 | |||
254 | DPRINTK("ENTER\n"); | ||
255 | |||
256 | /* do hardreset */ | ||
257 | rc = sata_link_hardreset(link, timing, deadline); | ||
258 | if (rc) { | ||
259 | ata_link_printk(link, KERN_ERR, | ||
260 | "COMRESET failed (errno=%d)\n", rc); | ||
261 | goto out; | ||
262 | } | ||
263 | |||
264 | /* clear SError bits including .X which blocks the port when set */ | ||
265 | rc = sata_scr_write(link, SCR_ERROR, 0xffffffff); | ||
266 | if (rc) { | ||
267 | ata_link_printk(link, KERN_ERR, "failed to clear SError " | ||
268 | "during hardreset (errno=%d)\n", rc); | ||
269 | goto out; | ||
270 | } | ||
271 | |||
272 | /* if device is present, follow up with srst to wait for !BSY */ | ||
273 | if (ata_link_online(link)) | ||
274 | rc = -EAGAIN; | ||
275 | out: | ||
276 | /* if SCR isn't accessible, we need to reset the PMP */ | ||
277 | if (rc && rc != -EAGAIN && sata_scr_read(link, SCR_STATUS, &tmp)) | ||
278 | rc = -ERESTART; | ||
279 | |||
280 | DPRINTK("EXIT, rc=%d\n", rc); | ||
281 | return rc; | ||
282 | } | ||
283 | |||
284 | /** | ||
285 | * ata_std_postreset - standard postreset method for PMP link | ||
286 | * @link: the target ata_link | ||
287 | * @classes: classes of attached devices | ||
288 | * | ||
289 | * This function is invoked after a successful reset. Note that | ||
290 | * the device might have been reset more than once using | ||
291 | * different reset methods before postreset is invoked. | ||
292 | * | ||
293 | * LOCKING: | ||
294 | * Kernel thread context (may sleep) | ||
295 | */ | ||
296 | void sata_pmp_std_postreset(struct ata_link *link, unsigned int *class) | ||
297 | { | ||
298 | u32 serror; | ||
299 | |||
300 | DPRINTK("ENTER\n"); | ||
301 | |||
302 | /* clear SError */ | ||
303 | if (sata_scr_read(link, SCR_ERROR, &serror) == 0) | ||
304 | sata_scr_write(link, SCR_ERROR, serror); | ||
305 | |||
306 | /* print link status */ | ||
307 | sata_print_link_status(link); | ||
308 | |||
309 | DPRINTK("EXIT\n"); | ||
310 | } | ||
311 | |||
312 | /** | ||
313 | * sata_pmp_read_gscr - read GSCR block of SATA PMP | 187 | * sata_pmp_read_gscr - read GSCR block of SATA PMP |
314 | * @dev: PMP device | 188 | * @dev: PMP device |
315 | * @gscr: buffer to read GSCR block into | 189 | * @gscr: buffer to read GSCR block into |
@@ -444,9 +318,8 @@ static int sata_pmp_init_links(struct ata_port *ap, int nr_ports) | |||
444 | struct ata_eh_context *ehc = &link->eh_context; | 318 | struct ata_eh_context *ehc = &link->eh_context; |
445 | 319 | ||
446 | link->flags = 0; | 320 | link->flags = 0; |
447 | ehc->i.probe_mask |= 1; | 321 | ehc->i.probe_mask |= ATA_ALL_DEVICES; |
448 | ehc->i.action |= ATA_EH_SOFTRESET; | 322 | ehc->i.action |= ATA_EH_RESET; |
449 | ehc->i.flags |= ATA_EHI_RESUME_LINK; | ||
450 | } | 323 | } |
451 | 324 | ||
452 | return 0; | 325 | return 0; |
@@ -462,9 +335,6 @@ static void sata_pmp_quirks(struct ata_port *ap) | |||
462 | if (vendor == 0x1095 && devid == 0x3726) { | 335 | if (vendor == 0x1095 && devid == 0x3726) { |
463 | /* sil3726 quirks */ | 336 | /* sil3726 quirks */ |
464 | ata_port_for_each_link(link, ap) { | 337 | ata_port_for_each_link(link, ap) { |
465 | /* SError.N need a kick in the ass to get working */ | ||
466 | link->flags |= ATA_LFLAG_HRST_TO_RESUME; | ||
467 | |||
468 | /* class code report is unreliable */ | 338 | /* class code report is unreliable */ |
469 | if (link->pmp < 5) | 339 | if (link->pmp < 5) |
470 | link->flags |= ATA_LFLAG_ASSUME_ATA; | 340 | link->flags |= ATA_LFLAG_ASSUME_ATA; |
@@ -477,9 +347,6 @@ static void sata_pmp_quirks(struct ata_port *ap) | |||
477 | } else if (vendor == 0x1095 && devid == 0x4723) { | 347 | } else if (vendor == 0x1095 && devid == 0x4723) { |
478 | /* sil4723 quirks */ | 348 | /* sil4723 quirks */ |
479 | ata_port_for_each_link(link, ap) { | 349 | ata_port_for_each_link(link, ap) { |
480 | /* SError.N need a kick in the ass to get working */ | ||
481 | link->flags |= ATA_LFLAG_HRST_TO_RESUME; | ||
482 | |||
483 | /* class code report is unreliable */ | 350 | /* class code report is unreliable */ |
484 | if (link->pmp < 2) | 351 | if (link->pmp < 2) |
485 | link->flags |= ATA_LFLAG_ASSUME_ATA; | 352 | link->flags |= ATA_LFLAG_ASSUME_ATA; |
@@ -492,9 +359,6 @@ static void sata_pmp_quirks(struct ata_port *ap) | |||
492 | } else if (vendor == 0x1095 && devid == 0x4726) { | 359 | } else if (vendor == 0x1095 && devid == 0x4726) { |
493 | /* sil4726 quirks */ | 360 | /* sil4726 quirks */ |
494 | ata_port_for_each_link(link, ap) { | 361 | ata_port_for_each_link(link, ap) { |
495 | /* SError.N need a kick in the ass to get working */ | ||
496 | link->flags |= ATA_LFLAG_HRST_TO_RESUME; | ||
497 | |||
498 | /* Class code report is unreliable and SRST | 362 | /* Class code report is unreliable and SRST |
499 | * times out under certain configurations. | 363 | * times out under certain configurations. |
500 | * Config device can be at port 0 or 5 and | 364 | * Config device can be at port 0 or 5 and |
@@ -522,13 +386,6 @@ static void sata_pmp_quirks(struct ata_port *ap) | |||
522 | * otherwise. Don't try hard to recover it. | 386 | * otherwise. Don't try hard to recover it. |
523 | */ | 387 | */ |
524 | ap->pmp_link[ap->nr_pmp_links - 1].flags |= ATA_LFLAG_NO_RETRY; | 388 | ap->pmp_link[ap->nr_pmp_links - 1].flags |= ATA_LFLAG_NO_RETRY; |
525 | } else if (vendor == 0x11ab && devid == 0x4140) { | ||
526 | /* Marvell 88SM4140 quirks. Fan-out ports require PHY | ||
527 | * reset to work; other than that, it behaves very | ||
528 | * nicely. | ||
529 | */ | ||
530 | ata_port_for_each_link(link, ap) | ||
531 | link->flags |= ATA_LFLAG_HRST_TO_RESUME; | ||
532 | } | 389 | } |
533 | } | 390 | } |
534 | 391 | ||
@@ -554,7 +411,7 @@ int sata_pmp_attach(struct ata_device *dev) | |||
554 | int rc; | 411 | int rc; |
555 | 412 | ||
556 | /* is it hanging off the right place? */ | 413 | /* is it hanging off the right place? */ |
557 | if (!(ap->flags & ATA_FLAG_PMP)) { | 414 | if (!sata_pmp_supported(ap)) { |
558 | ata_dev_printk(dev, KERN_ERR, | 415 | ata_dev_printk(dev, KERN_ERR, |
559 | "host does not support Port Multiplier\n"); | 416 | "host does not support Port Multiplier\n"); |
560 | return -EINVAL; | 417 | return -EINVAL; |
@@ -840,13 +697,12 @@ static int sata_pmp_eh_recover_pmp(struct ata_port *ap, | |||
840 | retry: | 697 | retry: |
841 | ehc->classes[0] = ATA_DEV_UNKNOWN; | 698 | ehc->classes[0] = ATA_DEV_UNKNOWN; |
842 | 699 | ||
843 | if (ehc->i.action & ATA_EH_RESET_MASK) { | 700 | if (ehc->i.action & ATA_EH_RESET) { |
844 | struct ata_link *tlink; | 701 | struct ata_link *tlink; |
845 | 702 | ||
846 | ata_eh_freeze_port(ap); | 703 | ata_eh_freeze_port(ap); |
847 | 704 | ||
848 | /* reset */ | 705 | /* reset */ |
849 | ehc->i.action = ATA_EH_HARDRESET; | ||
850 | rc = ata_eh_reset(link, 0, prereset, softreset, hardreset, | 706 | rc = ata_eh_reset(link, 0, prereset, softreset, hardreset, |
851 | postreset); | 707 | postreset); |
852 | if (rc) { | 708 | if (rc) { |
@@ -858,8 +714,12 @@ static int sata_pmp_eh_recover_pmp(struct ata_port *ap, | |||
858 | ata_eh_thaw_port(ap); | 714 | ata_eh_thaw_port(ap); |
859 | 715 | ||
860 | /* PMP is reset, SErrors cannot be trusted, scan all */ | 716 | /* PMP is reset, SErrors cannot be trusted, scan all */ |
861 | ata_port_for_each_link(tlink, ap) | 717 | ata_port_for_each_link(tlink, ap) { |
862 | ata_ehi_schedule_probe(&tlink->eh_context.i); | 718 | struct ata_eh_context *ehc = &tlink->eh_context; |
719 | |||
720 | ehc->i.probe_mask |= ATA_ALL_DEVICES; | ||
721 | ehc->i.action |= ATA_EH_RESET; | ||
722 | } | ||
863 | } | 723 | } |
864 | 724 | ||
865 | /* If revalidation is requested, revalidate and reconfigure; | 725 | /* If revalidation is requested, revalidate and reconfigure; |
@@ -874,7 +734,7 @@ static int sata_pmp_eh_recover_pmp(struct ata_port *ap, | |||
874 | tries--; | 734 | tries--; |
875 | 735 | ||
876 | if (rc == -ENODEV) { | 736 | if (rc == -ENODEV) { |
877 | ehc->i.probe_mask |= 1; | 737 | ehc->i.probe_mask |= ATA_ALL_DEVICES; |
878 | detach = 1; | 738 | detach = 1; |
879 | /* give it just two more chances */ | 739 | /* give it just two more chances */ |
880 | tries = min(tries, 2); | 740 | tries = min(tries, 2); |
@@ -890,11 +750,11 @@ static int sata_pmp_eh_recover_pmp(struct ata_port *ap, | |||
890 | reval_failed = 1; | 750 | reval_failed = 1; |
891 | 751 | ||
892 | ata_dev_printk(dev, KERN_WARNING, | 752 | ata_dev_printk(dev, KERN_WARNING, |
893 | "retrying hardreset%s\n", | 753 | "retrying reset%s\n", |
894 | sleep ? " in 5 secs" : ""); | 754 | sleep ? " in 5 secs" : ""); |
895 | if (sleep) | 755 | if (sleep) |
896 | ssleep(5); | 756 | ssleep(5); |
897 | ehc->i.action |= ATA_EH_HARDRESET; | 757 | ehc->i.action |= ATA_EH_RESET; |
898 | goto retry; | 758 | goto retry; |
899 | } else { | 759 | } else { |
900 | ata_dev_printk(dev, KERN_ERR, "failed to recover PMP " | 760 | ata_dev_printk(dev, KERN_ERR, "failed to recover PMP " |
@@ -938,10 +798,8 @@ static int sata_pmp_eh_handle_disabled_links(struct ata_port *ap) | |||
938 | /* Some PMPs require hardreset sequence to get | 798 | /* Some PMPs require hardreset sequence to get |
939 | * SError.N working. | 799 | * SError.N working. |
940 | */ | 800 | */ |
941 | if ((link->flags & ATA_LFLAG_HRST_TO_RESUME) && | 801 | sata_link_hardreset(link, sata_deb_timing_normal, |
942 | (link->eh_context.i.flags & ATA_EHI_RESUME_LINK)) | 802 | jiffies + ATA_TMOUT_INTERNAL_QUICK, NULL, NULL); |
943 | sata_link_hardreset(link, sata_deb_timing_normal, | ||
944 | jiffies + ATA_TMOUT_INTERNAL_QUICK); | ||
945 | 803 | ||
946 | /* unconditionally clear SError.N */ | 804 | /* unconditionally clear SError.N */ |
947 | rc = sata_scr_write(link, SCR_ERROR, SERR_PHYRDY_CHG); | 805 | rc = sata_scr_write(link, SCR_ERROR, SERR_PHYRDY_CHG); |
@@ -987,14 +845,6 @@ static int sata_pmp_handle_link_fail(struct ata_link *link, int *link_tries) | |||
987 | /** | 845 | /** |
988 | * sata_pmp_eh_recover - recover PMP-enabled port | 846 | * sata_pmp_eh_recover - recover PMP-enabled port |
989 | * @ap: ATA port to recover | 847 | * @ap: ATA port to recover |
990 | * @prereset: prereset method (can be NULL) | ||
991 | * @softreset: softreset method | ||
992 | * @hardreset: hardreset method | ||
993 | * @postreset: postreset method (can be NULL) | ||
994 | * @pmp_prereset: PMP prereset method (can be NULL) | ||
995 | * @pmp_softreset: PMP softreset method (can be NULL) | ||
996 | * @pmp_hardreset: PMP hardreset method (can be NULL) | ||
997 | * @pmp_postreset: PMP postreset method (can be NULL) | ||
998 | * | 848 | * |
999 | * Drive EH recovery operation for PMP enabled port @ap. This | 849 | * Drive EH recovery operation for PMP enabled port @ap. This |
1000 | * function recovers host and PMP ports with proper retrials and | 850 | * function recovers host and PMP ports with proper retrials and |
@@ -1007,12 +857,9 @@ static int sata_pmp_handle_link_fail(struct ata_link *link, int *link_tries) | |||
1007 | * RETURNS: | 857 | * RETURNS: |
1008 | * 0 on success, -errno on failure. | 858 | * 0 on success, -errno on failure. |
1009 | */ | 859 | */ |
1010 | static int sata_pmp_eh_recover(struct ata_port *ap, | 860 | static int sata_pmp_eh_recover(struct ata_port *ap) |
1011 | ata_prereset_fn_t prereset, ata_reset_fn_t softreset, | ||
1012 | ata_reset_fn_t hardreset, ata_postreset_fn_t postreset, | ||
1013 | ata_prereset_fn_t pmp_prereset, ata_reset_fn_t pmp_softreset, | ||
1014 | ata_reset_fn_t pmp_hardreset, ata_postreset_fn_t pmp_postreset) | ||
1015 | { | 861 | { |
862 | struct ata_port_operations *ops = ap->ops; | ||
1016 | int pmp_tries, link_tries[SATA_PMP_MAX_PORTS]; | 863 | int pmp_tries, link_tries[SATA_PMP_MAX_PORTS]; |
1017 | struct ata_link *pmp_link = &ap->link; | 864 | struct ata_link *pmp_link = &ap->link; |
1018 | struct ata_device *pmp_dev = pmp_link->device; | 865 | struct ata_device *pmp_dev = pmp_link->device; |
@@ -1029,9 +876,9 @@ static int sata_pmp_eh_recover(struct ata_port *ap, | |||
1029 | 876 | ||
1030 | retry: | 877 | retry: |
1031 | /* PMP attached? */ | 878 | /* PMP attached? */ |
1032 | if (!ap->nr_pmp_links) { | 879 | if (!sata_pmp_attached(ap)) { |
1033 | rc = ata_eh_recover(ap, prereset, softreset, hardreset, | 880 | rc = ata_eh_recover(ap, ops->prereset, ops->softreset, |
1034 | postreset, NULL); | 881 | ops->hardreset, ops->postreset, NULL); |
1035 | if (rc) { | 882 | if (rc) { |
1036 | ata_link_for_each_dev(dev, &ap->link) | 883 | ata_link_for_each_dev(dev, &ap->link) |
1037 | ata_dev_disable(dev); | 884 | ata_dev_disable(dev); |
@@ -1049,8 +896,8 @@ static int sata_pmp_eh_recover(struct ata_port *ap, | |||
1049 | } | 896 | } |
1050 | 897 | ||
1051 | /* recover pmp */ | 898 | /* recover pmp */ |
1052 | rc = sata_pmp_eh_recover_pmp(ap, prereset, softreset, hardreset, | 899 | rc = sata_pmp_eh_recover_pmp(ap, ops->prereset, ops->softreset, |
1053 | postreset); | 900 | ops->hardreset, ops->postreset); |
1054 | if (rc) | 901 | if (rc) |
1055 | goto pmp_fail; | 902 | goto pmp_fail; |
1056 | 903 | ||
@@ -1060,8 +907,8 @@ static int sata_pmp_eh_recover(struct ata_port *ap, | |||
1060 | goto pmp_fail; | 907 | goto pmp_fail; |
1061 | 908 | ||
1062 | /* recover links */ | 909 | /* recover links */ |
1063 | rc = ata_eh_recover(ap, pmp_prereset, pmp_softreset, pmp_hardreset, | 910 | rc = ata_eh_recover(ap, ops->pmp_prereset, ops->pmp_softreset, |
1064 | pmp_postreset, &link); | 911 | ops->pmp_hardreset, ops->pmp_postreset, &link); |
1065 | if (rc) | 912 | if (rc) |
1066 | goto link_fail; | 913 | goto link_fail; |
1067 | 914 | ||
@@ -1124,7 +971,7 @@ static int sata_pmp_eh_recover(struct ata_port *ap, | |||
1124 | 971 | ||
1125 | link_fail: | 972 | link_fail: |
1126 | if (sata_pmp_handle_link_fail(link, link_tries)) { | 973 | if (sata_pmp_handle_link_fail(link, link_tries)) { |
1127 | pmp_ehc->i.action |= ATA_EH_HARDRESET; | 974 | pmp_ehc->i.action |= ATA_EH_RESET; |
1128 | goto retry; | 975 | goto retry; |
1129 | } | 976 | } |
1130 | 977 | ||
@@ -1136,13 +983,13 @@ static int sata_pmp_eh_recover(struct ata_port *ap, | |||
1136 | if (ap->pflags & ATA_PFLAG_UNLOADING) | 983 | if (ap->pflags & ATA_PFLAG_UNLOADING) |
1137 | return rc; | 984 | return rc; |
1138 | 985 | ||
1139 | if (!ap->nr_pmp_links) | 986 | if (!sata_pmp_attached(ap)) |
1140 | goto retry; | 987 | goto retry; |
1141 | 988 | ||
1142 | if (--pmp_tries) { | 989 | if (--pmp_tries) { |
1143 | ata_port_printk(ap, KERN_WARNING, | 990 | ata_port_printk(ap, KERN_WARNING, |
1144 | "failed to recover PMP, retrying in 5 secs\n"); | 991 | "failed to recover PMP, retrying in 5 secs\n"); |
1145 | pmp_ehc->i.action |= ATA_EH_HARDRESET; | 992 | pmp_ehc->i.action |= ATA_EH_RESET; |
1146 | ssleep(5); | 993 | ssleep(5); |
1147 | goto retry; | 994 | goto retry; |
1148 | } | 995 | } |
@@ -1157,16 +1004,8 @@ static int sata_pmp_eh_recover(struct ata_port *ap, | |||
1157 | } | 1004 | } |
1158 | 1005 | ||
1159 | /** | 1006 | /** |
1160 | * sata_pmp_do_eh - do standard error handling for PMP-enabled host | 1007 | * sata_pmp_error_handler - do standard error handling for PMP-enabled host |
1161 | * @ap: host port to handle error for | 1008 | * @ap: host port to handle error for |
1162 | * @prereset: prereset method (can be NULL) | ||
1163 | * @softreset: softreset method | ||
1164 | * @hardreset: hardreset method | ||
1165 | * @postreset: postreset method (can be NULL) | ||
1166 | * @pmp_prereset: PMP prereset method (can be NULL) | ||
1167 | * @pmp_softreset: PMP softreset method (can be NULL) | ||
1168 | * @pmp_hardreset: PMP hardreset method (can be NULL) | ||
1169 | * @pmp_postreset: PMP postreset method (can be NULL) | ||
1170 | * | 1009 | * |
1171 | * Perform standard error handling sequence for PMP-enabled host | 1010 | * Perform standard error handling sequence for PMP-enabled host |
1172 | * @ap. | 1011 | * @ap. |
@@ -1174,16 +1013,14 @@ static int sata_pmp_eh_recover(struct ata_port *ap, | |||
1174 | * LOCKING: | 1013 | * LOCKING: |
1175 | * Kernel thread context (may sleep). | 1014 | * Kernel thread context (may sleep). |
1176 | */ | 1015 | */ |
1177 | void sata_pmp_do_eh(struct ata_port *ap, | 1016 | void sata_pmp_error_handler(struct ata_port *ap) |
1178 | ata_prereset_fn_t prereset, ata_reset_fn_t softreset, | ||
1179 | ata_reset_fn_t hardreset, ata_postreset_fn_t postreset, | ||
1180 | ata_prereset_fn_t pmp_prereset, ata_reset_fn_t pmp_softreset, | ||
1181 | ata_reset_fn_t pmp_hardreset, ata_postreset_fn_t pmp_postreset) | ||
1182 | { | 1017 | { |
1183 | ata_eh_autopsy(ap); | 1018 | ata_eh_autopsy(ap); |
1184 | ata_eh_report(ap); | 1019 | ata_eh_report(ap); |
1185 | sata_pmp_eh_recover(ap, prereset, softreset, hardreset, postreset, | 1020 | sata_pmp_eh_recover(ap); |
1186 | pmp_prereset, pmp_softreset, pmp_hardreset, | ||
1187 | pmp_postreset); | ||
1188 | ata_eh_finish(ap); | 1021 | ata_eh_finish(ap); |
1189 | } | 1022 | } |
1023 | |||
1024 | EXPORT_SYMBOL_GPL(sata_pmp_port_ops); | ||
1025 | EXPORT_SYMBOL_GPL(sata_pmp_qc_defer_cmd_switch); | ||
1026 | EXPORT_SYMBOL_GPL(sata_pmp_error_handler); | ||
diff --git a/drivers/ata/libata-scsi.c b/drivers/ata/libata-scsi.c index 15795394b0a8..fedf62de9460 100644 --- a/drivers/ata/libata-scsi.c +++ b/drivers/ata/libata-scsi.c | |||
@@ -2393,7 +2393,9 @@ static void atapi_request_sense(struct ata_queued_cmd *qc) | |||
2393 | /* FIXME: is this needed? */ | 2393 | /* FIXME: is this needed? */ |
2394 | memset(cmd->sense_buffer, 0, SCSI_SENSE_BUFFERSIZE); | 2394 | memset(cmd->sense_buffer, 0, SCSI_SENSE_BUFFERSIZE); |
2395 | 2395 | ||
2396 | ap->ops->tf_read(ap, &qc->tf); | 2396 | #ifdef CONFIG_ATA_SFF |
2397 | ap->ops->sff_tf_read(ap, &qc->tf); | ||
2398 | #endif | ||
2397 | 2399 | ||
2398 | /* fill these in, for the case where they are -not- overwritten */ | 2400 | /* fill these in, for the case where they are -not- overwritten */ |
2399 | cmd->sense_buffer[0] = 0x70; | 2401 | cmd->sense_buffer[0] = 0x70; |
@@ -2615,7 +2617,7 @@ static unsigned int atapi_xlat(struct ata_queued_cmd *qc) | |||
2615 | 2617 | ||
2616 | static struct ata_device *ata_find_dev(struct ata_port *ap, int devno) | 2618 | static struct ata_device *ata_find_dev(struct ata_port *ap, int devno) |
2617 | { | 2619 | { |
2618 | if (ap->nr_pmp_links == 0) { | 2620 | if (!sata_pmp_attached(ap)) { |
2619 | if (likely(devno < ata_link_max_devices(&ap->link))) | 2621 | if (likely(devno < ata_link_max_devices(&ap->link))) |
2620 | return &ap->link.device[devno]; | 2622 | return &ap->link.device[devno]; |
2621 | } else { | 2623 | } else { |
@@ -2632,7 +2634,7 @@ static struct ata_device *__ata_scsi_find_dev(struct ata_port *ap, | |||
2632 | int devno; | 2634 | int devno; |
2633 | 2635 | ||
2634 | /* skip commands not addressed to targets we simulate */ | 2636 | /* skip commands not addressed to targets we simulate */ |
2635 | if (ap->nr_pmp_links == 0) { | 2637 | if (!sata_pmp_attached(ap)) { |
2636 | if (unlikely(scsidev->channel || scsidev->lun)) | 2638 | if (unlikely(scsidev->channel || scsidev->lun)) |
2637 | return NULL; | 2639 | return NULL; |
2638 | devno = scsidev->id; | 2640 | devno = scsidev->id; |
@@ -3490,7 +3492,7 @@ static int ata_scsi_user_scan(struct Scsi_Host *shost, unsigned int channel, | |||
3490 | if (lun != SCAN_WILD_CARD && lun) | 3492 | if (lun != SCAN_WILD_CARD && lun) |
3491 | return -EINVAL; | 3493 | return -EINVAL; |
3492 | 3494 | ||
3493 | if (ap->nr_pmp_links == 0) { | 3495 | if (!sata_pmp_attached(ap)) { |
3494 | if (channel != SCAN_WILD_CARD && channel) | 3496 | if (channel != SCAN_WILD_CARD && channel) |
3495 | return -EINVAL; | 3497 | return -EINVAL; |
3496 | devno = id; | 3498 | devno = id; |
@@ -3507,8 +3509,8 @@ static int ata_scsi_user_scan(struct Scsi_Host *shost, unsigned int channel, | |||
3507 | 3509 | ||
3508 | ata_port_for_each_link(link, ap) { | 3510 | ata_port_for_each_link(link, ap) { |
3509 | struct ata_eh_info *ehi = &link->eh_info; | 3511 | struct ata_eh_info *ehi = &link->eh_info; |
3510 | ehi->probe_mask |= (1 << ata_link_max_devices(link)) - 1; | 3512 | ehi->probe_mask |= ATA_ALL_DEVICES; |
3511 | ehi->action |= ATA_EH_SOFTRESET; | 3513 | ehi->action |= ATA_EH_RESET; |
3512 | } | 3514 | } |
3513 | } else { | 3515 | } else { |
3514 | struct ata_device *dev = ata_find_dev(ap, devno); | 3516 | struct ata_device *dev = ata_find_dev(ap, devno); |
@@ -3516,8 +3518,7 @@ static int ata_scsi_user_scan(struct Scsi_Host *shost, unsigned int channel, | |||
3516 | if (dev) { | 3518 | if (dev) { |
3517 | struct ata_eh_info *ehi = &dev->link->eh_info; | 3519 | struct ata_eh_info *ehi = &dev->link->eh_info; |
3518 | ehi->probe_mask |= 1 << dev->devno; | 3520 | ehi->probe_mask |= 1 << dev->devno; |
3519 | ehi->action |= ATA_EH_SOFTRESET; | 3521 | ehi->action |= ATA_EH_RESET; |
3520 | ehi->flags |= ATA_EHI_RESUME_LINK; | ||
3521 | } else | 3522 | } else |
3522 | rc = -EINVAL; | 3523 | rc = -EINVAL; |
3523 | } | 3524 | } |
diff --git a/drivers/ata/libata-sff.c b/drivers/ata/libata-sff.c index 20dc572fb45a..15499522e642 100644 --- a/drivers/ata/libata-sff.c +++ b/drivers/ata/libata-sff.c | |||
@@ -35,11 +35,377 @@ | |||
35 | #include <linux/kernel.h> | 35 | #include <linux/kernel.h> |
36 | #include <linux/pci.h> | 36 | #include <linux/pci.h> |
37 | #include <linux/libata.h> | 37 | #include <linux/libata.h> |
38 | #include <linux/highmem.h> | ||
38 | 39 | ||
39 | #include "libata.h" | 40 | #include "libata.h" |
40 | 41 | ||
42 | const struct ata_port_operations ata_sff_port_ops = { | ||
43 | .inherits = &ata_base_port_ops, | ||
44 | |||
45 | .qc_prep = ata_sff_qc_prep, | ||
46 | .qc_issue = ata_sff_qc_issue, | ||
47 | .qc_fill_rtf = ata_sff_qc_fill_rtf, | ||
48 | |||
49 | .freeze = ata_sff_freeze, | ||
50 | .thaw = ata_sff_thaw, | ||
51 | .prereset = ata_sff_prereset, | ||
52 | .softreset = ata_sff_softreset, | ||
53 | .hardreset = sata_sff_hardreset, | ||
54 | .postreset = ata_sff_postreset, | ||
55 | .error_handler = ata_sff_error_handler, | ||
56 | .post_internal_cmd = ata_sff_post_internal_cmd, | ||
57 | |||
58 | .sff_dev_select = ata_sff_dev_select, | ||
59 | .sff_check_status = ata_sff_check_status, | ||
60 | .sff_tf_load = ata_sff_tf_load, | ||
61 | .sff_tf_read = ata_sff_tf_read, | ||
62 | .sff_exec_command = ata_sff_exec_command, | ||
63 | .sff_data_xfer = ata_sff_data_xfer, | ||
64 | .sff_irq_on = ata_sff_irq_on, | ||
65 | .sff_irq_clear = ata_sff_irq_clear, | ||
66 | |||
67 | .port_start = ata_sff_port_start, | ||
68 | }; | ||
69 | |||
70 | const struct ata_port_operations ata_bmdma_port_ops = { | ||
71 | .inherits = &ata_sff_port_ops, | ||
72 | |||
73 | .mode_filter = ata_bmdma_mode_filter, | ||
74 | |||
75 | .bmdma_setup = ata_bmdma_setup, | ||
76 | .bmdma_start = ata_bmdma_start, | ||
77 | .bmdma_stop = ata_bmdma_stop, | ||
78 | .bmdma_status = ata_bmdma_status, | ||
79 | }; | ||
80 | |||
81 | /** | ||
82 | * ata_fill_sg - Fill PCI IDE PRD table | ||
83 | * @qc: Metadata associated with taskfile to be transferred | ||
84 | * | ||
85 | * Fill PCI IDE PRD (scatter-gather) table with segments | ||
86 | * associated with the current disk command. | ||
87 | * | ||
88 | * LOCKING: | ||
89 | * spin_lock_irqsave(host lock) | ||
90 | * | ||
91 | */ | ||
92 | static void ata_fill_sg(struct ata_queued_cmd *qc) | ||
93 | { | ||
94 | struct ata_port *ap = qc->ap; | ||
95 | struct scatterlist *sg; | ||
96 | unsigned int si, pi; | ||
97 | |||
98 | pi = 0; | ||
99 | for_each_sg(qc->sg, sg, qc->n_elem, si) { | ||
100 | u32 addr, offset; | ||
101 | u32 sg_len, len; | ||
102 | |||
103 | /* determine if physical DMA addr spans 64K boundary. | ||
104 | * Note h/w doesn't support 64-bit, so we unconditionally | ||
105 | * truncate dma_addr_t to u32. | ||
106 | */ | ||
107 | addr = (u32) sg_dma_address(sg); | ||
108 | sg_len = sg_dma_len(sg); | ||
109 | |||
110 | while (sg_len) { | ||
111 | offset = addr & 0xffff; | ||
112 | len = sg_len; | ||
113 | if ((offset + sg_len) > 0x10000) | ||
114 | len = 0x10000 - offset; | ||
115 | |||
116 | ap->prd[pi].addr = cpu_to_le32(addr); | ||
117 | ap->prd[pi].flags_len = cpu_to_le32(len & 0xffff); | ||
118 | VPRINTK("PRD[%u] = (0x%X, 0x%X)\n", pi, addr, len); | ||
119 | |||
120 | pi++; | ||
121 | sg_len -= len; | ||
122 | addr += len; | ||
123 | } | ||
124 | } | ||
125 | |||
126 | ap->prd[pi - 1].flags_len |= cpu_to_le32(ATA_PRD_EOT); | ||
127 | } | ||
128 | |||
129 | /** | ||
130 | * ata_fill_sg_dumb - Fill PCI IDE PRD table | ||
131 | * @qc: Metadata associated with taskfile to be transferred | ||
132 | * | ||
133 | * Fill PCI IDE PRD (scatter-gather) table with segments | ||
134 | * associated with the current disk command. Perform the fill | ||
135 | * so that we avoid writing any length 64K records for | ||
136 | * controllers that don't follow the spec. | ||
137 | * | ||
138 | * LOCKING: | ||
139 | * spin_lock_irqsave(host lock) | ||
140 | * | ||
141 | */ | ||
142 | static void ata_fill_sg_dumb(struct ata_queued_cmd *qc) | ||
143 | { | ||
144 | struct ata_port *ap = qc->ap; | ||
145 | struct scatterlist *sg; | ||
146 | unsigned int si, pi; | ||
147 | |||
148 | pi = 0; | ||
149 | for_each_sg(qc->sg, sg, qc->n_elem, si) { | ||
150 | u32 addr, offset; | ||
151 | u32 sg_len, len, blen; | ||
152 | |||
153 | /* determine if physical DMA addr spans 64K boundary. | ||
154 | * Note h/w doesn't support 64-bit, so we unconditionally | ||
155 | * truncate dma_addr_t to u32. | ||
156 | */ | ||
157 | addr = (u32) sg_dma_address(sg); | ||
158 | sg_len = sg_dma_len(sg); | ||
159 | |||
160 | while (sg_len) { | ||
161 | offset = addr & 0xffff; | ||
162 | len = sg_len; | ||
163 | if ((offset + sg_len) > 0x10000) | ||
164 | len = 0x10000 - offset; | ||
165 | |||
166 | blen = len & 0xffff; | ||
167 | ap->prd[pi].addr = cpu_to_le32(addr); | ||
168 | if (blen == 0) { | ||
169 | /* Some PATA chipsets like the CS5530 can't | ||
170 | cope with 0x0000 meaning 64K as the spec says */ | ||
171 | ap->prd[pi].flags_len = cpu_to_le32(0x8000); | ||
172 | blen = 0x8000; | ||
173 | ap->prd[++pi].addr = cpu_to_le32(addr + 0x8000); | ||
174 | } | ||
175 | ap->prd[pi].flags_len = cpu_to_le32(blen); | ||
176 | VPRINTK("PRD[%u] = (0x%X, 0x%X)\n", pi, addr, len); | ||
177 | |||
178 | pi++; | ||
179 | sg_len -= len; | ||
180 | addr += len; | ||
181 | } | ||
182 | } | ||
183 | |||
184 | ap->prd[pi - 1].flags_len |= cpu_to_le32(ATA_PRD_EOT); | ||
185 | } | ||
186 | |||
187 | /** | ||
188 | * ata_sff_qc_prep - Prepare taskfile for submission | ||
189 | * @qc: Metadata associated with taskfile to be prepared | ||
190 | * | ||
191 | * Prepare ATA taskfile for submission. | ||
192 | * | ||
193 | * LOCKING: | ||
194 | * spin_lock_irqsave(host lock) | ||
195 | */ | ||
196 | void ata_sff_qc_prep(struct ata_queued_cmd *qc) | ||
197 | { | ||
198 | if (!(qc->flags & ATA_QCFLAG_DMAMAP)) | ||
199 | return; | ||
200 | |||
201 | ata_fill_sg(qc); | ||
202 | } | ||
203 | |||
204 | /** | ||
205 | * ata_sff_dumb_qc_prep - Prepare taskfile for submission | ||
206 | * @qc: Metadata associated with taskfile to be prepared | ||
207 | * | ||
208 | * Prepare ATA taskfile for submission. | ||
209 | * | ||
210 | * LOCKING: | ||
211 | * spin_lock_irqsave(host lock) | ||
212 | */ | ||
213 | void ata_sff_dumb_qc_prep(struct ata_queued_cmd *qc) | ||
214 | { | ||
215 | if (!(qc->flags & ATA_QCFLAG_DMAMAP)) | ||
216 | return; | ||
217 | |||
218 | ata_fill_sg_dumb(qc); | ||
219 | } | ||
220 | |||
221 | /** | ||
222 | * ata_sff_check_status - Read device status reg & clear interrupt | ||
223 | * @ap: port where the device is | ||
224 | * | ||
225 | * Reads ATA taskfile status register for currently-selected device | ||
226 | * and return its value. This also clears pending interrupts | ||
227 | * from this device | ||
228 | * | ||
229 | * LOCKING: | ||
230 | * Inherited from caller. | ||
231 | */ | ||
232 | u8 ata_sff_check_status(struct ata_port *ap) | ||
233 | { | ||
234 | return ioread8(ap->ioaddr.status_addr); | ||
235 | } | ||
236 | |||
237 | /** | ||
238 | * ata_sff_altstatus - Read device alternate status reg | ||
239 | * @ap: port where the device is | ||
240 | * | ||
241 | * Reads ATA taskfile alternate status register for | ||
242 | * currently-selected device and return its value. | ||
243 | * | ||
244 | * Note: may NOT be used as the check_altstatus() entry in | ||
245 | * ata_port_operations. | ||
246 | * | ||
247 | * LOCKING: | ||
248 | * Inherited from caller. | ||
249 | */ | ||
250 | u8 ata_sff_altstatus(struct ata_port *ap) | ||
251 | { | ||
252 | if (ap->ops->sff_check_altstatus) | ||
253 | return ap->ops->sff_check_altstatus(ap); | ||
254 | |||
255 | return ioread8(ap->ioaddr.altstatus_addr); | ||
256 | } | ||
257 | |||
258 | /** | ||
259 | * ata_sff_busy_sleep - sleep until BSY clears, or timeout | ||
260 | * @ap: port containing status register to be polled | ||
261 | * @tmout_pat: impatience timeout | ||
262 | * @tmout: overall timeout | ||
263 | * | ||
264 | * Sleep until ATA Status register bit BSY clears, | ||
265 | * or a timeout occurs. | ||
266 | * | ||
267 | * LOCKING: | ||
268 | * Kernel thread context (may sleep). | ||
269 | * | ||
270 | * RETURNS: | ||
271 | * 0 on success, -errno otherwise. | ||
272 | */ | ||
273 | int ata_sff_busy_sleep(struct ata_port *ap, | ||
274 | unsigned long tmout_pat, unsigned long tmout) | ||
275 | { | ||
276 | unsigned long timer_start, timeout; | ||
277 | u8 status; | ||
278 | |||
279 | status = ata_sff_busy_wait(ap, ATA_BUSY, 300); | ||
280 | timer_start = jiffies; | ||
281 | timeout = timer_start + tmout_pat; | ||
282 | while (status != 0xff && (status & ATA_BUSY) && | ||
283 | time_before(jiffies, timeout)) { | ||
284 | msleep(50); | ||
285 | status = ata_sff_busy_wait(ap, ATA_BUSY, 3); | ||
286 | } | ||
287 | |||
288 | if (status != 0xff && (status & ATA_BUSY)) | ||
289 | ata_port_printk(ap, KERN_WARNING, | ||
290 | "port is slow to respond, please be patient " | ||
291 | "(Status 0x%x)\n", status); | ||
292 | |||
293 | timeout = timer_start + tmout; | ||
294 | while (status != 0xff && (status & ATA_BUSY) && | ||
295 | time_before(jiffies, timeout)) { | ||
296 | msleep(50); | ||
297 | status = ap->ops->sff_check_status(ap); | ||
298 | } | ||
299 | |||
300 | if (status == 0xff) | ||
301 | return -ENODEV; | ||
302 | |||
303 | if (status & ATA_BUSY) { | ||
304 | ata_port_printk(ap, KERN_ERR, "port failed to respond " | ||
305 | "(%lu secs, Status 0x%x)\n", | ||
306 | tmout / HZ, status); | ||
307 | return -EBUSY; | ||
308 | } | ||
309 | |||
310 | return 0; | ||
311 | } | ||
312 | |||
313 | static int ata_sff_check_ready(struct ata_link *link) | ||
314 | { | ||
315 | u8 status = link->ap->ops->sff_check_status(link->ap); | ||
316 | |||
317 | if (!(status & ATA_BUSY)) | ||
318 | return 1; | ||
319 | if (status == 0xff) | ||
320 | return -ENODEV; | ||
321 | return 0; | ||
322 | } | ||
323 | |||
324 | /** | ||
325 | * ata_sff_wait_ready - sleep until BSY clears, or timeout | ||
326 | * @link: SFF link to wait ready status for | ||
327 | * @deadline: deadline jiffies for the operation | ||
328 | * | ||
329 | * Sleep until ATA Status register bit BSY clears, or timeout | ||
330 | * occurs. | ||
331 | * | ||
332 | * LOCKING: | ||
333 | * Kernel thread context (may sleep). | ||
334 | * | ||
335 | * RETURNS: | ||
336 | * 0 on success, -errno otherwise. | ||
337 | */ | ||
338 | int ata_sff_wait_ready(struct ata_link *link, unsigned long deadline) | ||
339 | { | ||
340 | return ata_wait_ready(link, deadline, ata_sff_check_ready); | ||
341 | } | ||
342 | |||
343 | /** | ||
344 | * ata_sff_dev_select - Select device 0/1 on ATA bus | ||
345 | * @ap: ATA channel to manipulate | ||
346 | * @device: ATA device (numbered from zero) to select | ||
347 | * | ||
348 | * Use the method defined in the ATA specification to | ||
349 | * make either device 0, or device 1, active on the | ||
350 | * ATA channel. Works with both PIO and MMIO. | ||
351 | * | ||
352 | * May be used as the dev_select() entry in ata_port_operations. | ||
353 | * | ||
354 | * LOCKING: | ||
355 | * caller. | ||
356 | */ | ||
357 | void ata_sff_dev_select(struct ata_port *ap, unsigned int device) | ||
358 | { | ||
359 | u8 tmp; | ||
360 | |||
361 | if (device == 0) | ||
362 | tmp = ATA_DEVICE_OBS; | ||
363 | else | ||
364 | tmp = ATA_DEVICE_OBS | ATA_DEV1; | ||
365 | |||
366 | iowrite8(tmp, ap->ioaddr.device_addr); | ||
367 | ata_sff_pause(ap); /* needed; also flushes, for mmio */ | ||
368 | } | ||
369 | |||
370 | /** | ||
371 | * ata_dev_select - Select device 0/1 on ATA bus | ||
372 | * @ap: ATA channel to manipulate | ||
373 | * @device: ATA device (numbered from zero) to select | ||
374 | * @wait: non-zero to wait for Status register BSY bit to clear | ||
375 | * @can_sleep: non-zero if context allows sleeping | ||
376 | * | ||
377 | * Use the method defined in the ATA specification to | ||
378 | * make either device 0, or device 1, active on the | ||
379 | * ATA channel. | ||
380 | * | ||
381 | * This is a high-level version of ata_sff_dev_select(), which | ||
382 | * additionally provides the services of inserting the proper | ||
383 | * pauses and status polling, where needed. | ||
384 | * | ||
385 | * LOCKING: | ||
386 | * caller. | ||
387 | */ | ||
388 | void ata_dev_select(struct ata_port *ap, unsigned int device, | ||
389 | unsigned int wait, unsigned int can_sleep) | ||
390 | { | ||
391 | if (ata_msg_probe(ap)) | ||
392 | ata_port_printk(ap, KERN_INFO, "ata_dev_select: ENTER, " | ||
393 | "device %u, wait %u\n", device, wait); | ||
394 | |||
395 | if (wait) | ||
396 | ata_wait_idle(ap); | ||
397 | |||
398 | ap->ops->sff_dev_select(ap, device); | ||
399 | |||
400 | if (wait) { | ||
401 | if (can_sleep && ap->link.device[device].class == ATA_DEV_ATAPI) | ||
402 | msleep(150); | ||
403 | ata_wait_idle(ap); | ||
404 | } | ||
405 | } | ||
406 | |||
41 | /** | 407 | /** |
42 | * ata_irq_on - Enable interrupts on a port. | 408 | * ata_sff_irq_on - Enable interrupts on a port. |
43 | * @ap: Port on which interrupts are enabled. | 409 | * @ap: Port on which interrupts are enabled. |
44 | * | 410 | * |
45 | * Enable interrupts on a legacy IDE device using MMIO or PIO, | 411 | * Enable interrupts on a legacy IDE device using MMIO or PIO, |
@@ -48,7 +414,7 @@ | |||
48 | * LOCKING: | 414 | * LOCKING: |
49 | * Inherited from caller. | 415 | * Inherited from caller. |
50 | */ | 416 | */ |
51 | u8 ata_irq_on(struct ata_port *ap) | 417 | u8 ata_sff_irq_on(struct ata_port *ap) |
52 | { | 418 | { |
53 | struct ata_ioports *ioaddr = &ap->ioaddr; | 419 | struct ata_ioports *ioaddr = &ap->ioaddr; |
54 | u8 tmp; | 420 | u8 tmp; |
@@ -60,13 +426,34 @@ u8 ata_irq_on(struct ata_port *ap) | |||
60 | iowrite8(ap->ctl, ioaddr->ctl_addr); | 426 | iowrite8(ap->ctl, ioaddr->ctl_addr); |
61 | tmp = ata_wait_idle(ap); | 427 | tmp = ata_wait_idle(ap); |
62 | 428 | ||
63 | ap->ops->irq_clear(ap); | 429 | ap->ops->sff_irq_clear(ap); |
64 | 430 | ||
65 | return tmp; | 431 | return tmp; |
66 | } | 432 | } |
67 | 433 | ||
68 | /** | 434 | /** |
69 | * ata_tf_load - send taskfile registers to host controller | 435 | * ata_sff_irq_clear - Clear PCI IDE BMDMA interrupt. |
436 | * @ap: Port associated with this ATA transaction. | ||
437 | * | ||
438 | * Clear interrupt and error flags in DMA status register. | ||
439 | * | ||
440 | * May be used as the irq_clear() entry in ata_port_operations. | ||
441 | * | ||
442 | * LOCKING: | ||
443 | * spin_lock_irqsave(host lock) | ||
444 | */ | ||
445 | void ata_sff_irq_clear(struct ata_port *ap) | ||
446 | { | ||
447 | void __iomem *mmio = ap->ioaddr.bmdma_addr; | ||
448 | |||
449 | if (!mmio) | ||
450 | return; | ||
451 | |||
452 | iowrite8(ioread8(mmio + ATA_DMA_STATUS), mmio + ATA_DMA_STATUS); | ||
453 | } | ||
454 | |||
455 | /** | ||
456 | * ata_sff_tf_load - send taskfile registers to host controller | ||
70 | * @ap: Port to which output is sent | 457 | * @ap: Port to which output is sent |
71 | * @tf: ATA taskfile register set | 458 | * @tf: ATA taskfile register set |
72 | * | 459 | * |
@@ -75,8 +462,7 @@ u8 ata_irq_on(struct ata_port *ap) | |||
75 | * LOCKING: | 462 | * LOCKING: |
76 | * Inherited from caller. | 463 | * Inherited from caller. |
77 | */ | 464 | */ |
78 | 465 | void ata_sff_tf_load(struct ata_port *ap, const struct ata_taskfile *tf) | |
79 | void ata_tf_load(struct ata_port *ap, const struct ata_taskfile *tf) | ||
80 | { | 466 | { |
81 | struct ata_ioports *ioaddr = &ap->ioaddr; | 467 | struct ata_ioports *ioaddr = &ap->ioaddr; |
82 | unsigned int is_addr = tf->flags & ATA_TFLAG_ISADDR; | 468 | unsigned int is_addr = tf->flags & ATA_TFLAG_ISADDR; |
@@ -126,26 +512,7 @@ void ata_tf_load(struct ata_port *ap, const struct ata_taskfile *tf) | |||
126 | } | 512 | } |
127 | 513 | ||
128 | /** | 514 | /** |
129 | * ata_exec_command - issue ATA command to host controller | 515 | * ata_sff_tf_read - input device's ATA taskfile shadow registers |
130 | * @ap: port to which command is being issued | ||
131 | * @tf: ATA taskfile register set | ||
132 | * | ||
133 | * Issues ATA command, with proper synchronization with interrupt | ||
134 | * handler / other threads. | ||
135 | * | ||
136 | * LOCKING: | ||
137 | * spin_lock_irqsave(host lock) | ||
138 | */ | ||
139 | void ata_exec_command(struct ata_port *ap, const struct ata_taskfile *tf) | ||
140 | { | ||
141 | DPRINTK("ata%u: cmd 0x%X\n", ap->print_id, tf->command); | ||
142 | |||
143 | iowrite8(tf->command, ap->ioaddr.command_addr); | ||
144 | ata_pause(ap); | ||
145 | } | ||
146 | |||
147 | /** | ||
148 | * ata_tf_read - input device's ATA taskfile shadow registers | ||
149 | * @ap: Port from which input is read | 516 | * @ap: Port from which input is read |
150 | * @tf: ATA taskfile register set for storing input | 517 | * @tf: ATA taskfile register set for storing input |
151 | * | 518 | * |
@@ -157,11 +524,11 @@ void ata_exec_command(struct ata_port *ap, const struct ata_taskfile *tf) | |||
157 | * LOCKING: | 524 | * LOCKING: |
158 | * Inherited from caller. | 525 | * Inherited from caller. |
159 | */ | 526 | */ |
160 | void ata_tf_read(struct ata_port *ap, struct ata_taskfile *tf) | 527 | void ata_sff_tf_read(struct ata_port *ap, struct ata_taskfile *tf) |
161 | { | 528 | { |
162 | struct ata_ioports *ioaddr = &ap->ioaddr; | 529 | struct ata_ioports *ioaddr = &ap->ioaddr; |
163 | 530 | ||
164 | tf->command = ata_check_status(ap); | 531 | tf->command = ata_sff_check_status(ap); |
165 | tf->feature = ioread8(ioaddr->error_addr); | 532 | tf->feature = ioread8(ioaddr->error_addr); |
166 | tf->nsect = ioread8(ioaddr->nsect_addr); | 533 | tf->nsect = ioread8(ioaddr->nsect_addr); |
167 | tf->lbal = ioread8(ioaddr->lbal_addr); | 534 | tf->lbal = ioread8(ioaddr->lbal_addr); |
@@ -185,165 +552,1028 @@ void ata_tf_read(struct ata_port *ap, struct ata_taskfile *tf) | |||
185 | } | 552 | } |
186 | 553 | ||
187 | /** | 554 | /** |
188 | * ata_check_status - Read device status reg & clear interrupt | 555 | * ata_sff_exec_command - issue ATA command to host controller |
189 | * @ap: port where the device is | 556 | * @ap: port to which command is being issued |
557 | * @tf: ATA taskfile register set | ||
190 | * | 558 | * |
191 | * Reads ATA taskfile status register for currently-selected device | 559 | * Issues ATA command, with proper synchronization with interrupt |
192 | * and return its value. This also clears pending interrupts | 560 | * handler / other threads. |
193 | * from this device | ||
194 | * | 561 | * |
195 | * LOCKING: | 562 | * LOCKING: |
196 | * Inherited from caller. | 563 | * spin_lock_irqsave(host lock) |
197 | */ | 564 | */ |
198 | u8 ata_check_status(struct ata_port *ap) | 565 | void ata_sff_exec_command(struct ata_port *ap, const struct ata_taskfile *tf) |
199 | { | 566 | { |
200 | return ioread8(ap->ioaddr.status_addr); | 567 | DPRINTK("ata%u: cmd 0x%X\n", ap->print_id, tf->command); |
568 | |||
569 | iowrite8(tf->command, ap->ioaddr.command_addr); | ||
570 | ata_sff_pause(ap); | ||
201 | } | 571 | } |
202 | 572 | ||
203 | /** | 573 | /** |
204 | * ata_altstatus - Read device alternate status reg | 574 | * ata_tf_to_host - issue ATA taskfile to host controller |
205 | * @ap: port where the device is | 575 | * @ap: port to which command is being issued |
576 | * @tf: ATA taskfile register set | ||
206 | * | 577 | * |
207 | * Reads ATA taskfile alternate status register for | 578 | * Issues ATA taskfile register set to ATA host controller, |
208 | * currently-selected device and return its value. | 579 | * with proper synchronization with interrupt handler and |
580 | * other threads. | ||
209 | * | 581 | * |
210 | * Note: may NOT be used as the check_altstatus() entry in | 582 | * LOCKING: |
211 | * ata_port_operations. | 583 | * spin_lock_irqsave(host lock) |
584 | */ | ||
585 | static inline void ata_tf_to_host(struct ata_port *ap, | ||
586 | const struct ata_taskfile *tf) | ||
587 | { | ||
588 | ap->ops->sff_tf_load(ap, tf); | ||
589 | ap->ops->sff_exec_command(ap, tf); | ||
590 | } | ||
591 | |||
592 | /** | ||
593 | * ata_sff_data_xfer - Transfer data by PIO | ||
594 | * @dev: device to target | ||
595 | * @buf: data buffer | ||
596 | * @buflen: buffer length | ||
597 | * @rw: read/write | ||
598 | * | ||
599 | * Transfer data from/to the device data register by PIO. | ||
212 | * | 600 | * |
213 | * LOCKING: | 601 | * LOCKING: |
214 | * Inherited from caller. | 602 | * Inherited from caller. |
603 | * | ||
604 | * RETURNS: | ||
605 | * Bytes consumed. | ||
215 | */ | 606 | */ |
216 | u8 ata_altstatus(struct ata_port *ap) | 607 | unsigned int ata_sff_data_xfer(struct ata_device *dev, unsigned char *buf, |
608 | unsigned int buflen, int rw) | ||
217 | { | 609 | { |
218 | if (ap->ops->check_altstatus) | 610 | struct ata_port *ap = dev->link->ap; |
219 | return ap->ops->check_altstatus(ap); | 611 | void __iomem *data_addr = ap->ioaddr.data_addr; |
612 | unsigned int words = buflen >> 1; | ||
220 | 613 | ||
221 | return ioread8(ap->ioaddr.altstatus_addr); | 614 | /* Transfer multiple of 2 bytes */ |
615 | if (rw == READ) | ||
616 | ioread16_rep(data_addr, buf, words); | ||
617 | else | ||
618 | iowrite16_rep(data_addr, buf, words); | ||
619 | |||
620 | /* Transfer trailing 1 byte, if any. */ | ||
621 | if (unlikely(buflen & 0x01)) { | ||
622 | __le16 align_buf[1] = { 0 }; | ||
623 | unsigned char *trailing_buf = buf + buflen - 1; | ||
624 | |||
625 | if (rw == READ) { | ||
626 | align_buf[0] = cpu_to_le16(ioread16(data_addr)); | ||
627 | memcpy(trailing_buf, align_buf, 1); | ||
628 | } else { | ||
629 | memcpy(align_buf, trailing_buf, 1); | ||
630 | iowrite16(le16_to_cpu(align_buf[0]), data_addr); | ||
631 | } | ||
632 | words++; | ||
633 | } | ||
634 | |||
635 | return words << 1; | ||
222 | } | 636 | } |
223 | 637 | ||
224 | /** | 638 | /** |
225 | * ata_bmdma_setup - Set up PCI IDE BMDMA transaction | 639 | * ata_sff_data_xfer_noirq - Transfer data by PIO |
226 | * @qc: Info associated with this ATA transaction. | 640 | * @dev: device to target |
641 | * @buf: data buffer | ||
642 | * @buflen: buffer length | ||
643 | * @rw: read/write | ||
644 | * | ||
645 | * Transfer data from/to the device data register by PIO. Do the | ||
646 | * transfer with interrupts disabled. | ||
227 | * | 647 | * |
228 | * LOCKING: | 648 | * LOCKING: |
229 | * spin_lock_irqsave(host lock) | 649 | * Inherited from caller. |
650 | * | ||
651 | * RETURNS: | ||
652 | * Bytes consumed. | ||
230 | */ | 653 | */ |
231 | void ata_bmdma_setup(struct ata_queued_cmd *qc) | 654 | unsigned int ata_sff_data_xfer_noirq(struct ata_device *dev, unsigned char *buf, |
655 | unsigned int buflen, int rw) | ||
656 | { | ||
657 | unsigned long flags; | ||
658 | unsigned int consumed; | ||
659 | |||
660 | local_irq_save(flags); | ||
661 | consumed = ata_sff_data_xfer(dev, buf, buflen, rw); | ||
662 | local_irq_restore(flags); | ||
663 | |||
664 | return consumed; | ||
665 | } | ||
666 | |||
667 | /** | ||
668 | * ata_pio_sector - Transfer a sector of data. | ||
669 | * @qc: Command on going | ||
670 | * | ||
671 | * Transfer qc->sect_size bytes of data from/to the ATA device. | ||
672 | * | ||
673 | * LOCKING: | ||
674 | * Inherited from caller. | ||
675 | */ | ||
676 | static void ata_pio_sector(struct ata_queued_cmd *qc) | ||
232 | { | 677 | { |
678 | int do_write = (qc->tf.flags & ATA_TFLAG_WRITE); | ||
233 | struct ata_port *ap = qc->ap; | 679 | struct ata_port *ap = qc->ap; |
234 | unsigned int rw = (qc->tf.flags & ATA_TFLAG_WRITE); | 680 | struct page *page; |
235 | u8 dmactl; | 681 | unsigned int offset; |
682 | unsigned char *buf; | ||
236 | 683 | ||
237 | /* load PRD table addr. */ | 684 | if (qc->curbytes == qc->nbytes - qc->sect_size) |
238 | mb(); /* make sure PRD table writes are visible to controller */ | 685 | ap->hsm_task_state = HSM_ST_LAST; |
239 | iowrite32(ap->prd_dma, ap->ioaddr.bmdma_addr + ATA_DMA_TABLE_OFS); | ||
240 | 686 | ||
241 | /* specify data direction, triple-check start bit is clear */ | 687 | page = sg_page(qc->cursg); |
242 | dmactl = ioread8(ap->ioaddr.bmdma_addr + ATA_DMA_CMD); | 688 | offset = qc->cursg->offset + qc->cursg_ofs; |
243 | dmactl &= ~(ATA_DMA_WR | ATA_DMA_START); | ||
244 | if (!rw) | ||
245 | dmactl |= ATA_DMA_WR; | ||
246 | iowrite8(dmactl, ap->ioaddr.bmdma_addr + ATA_DMA_CMD); | ||
247 | 689 | ||
248 | /* issue r/w command */ | 690 | /* get the current page and offset */ |
249 | ap->ops->exec_command(ap, &qc->tf); | 691 | page = nth_page(page, (offset >> PAGE_SHIFT)); |
692 | offset %= PAGE_SIZE; | ||
693 | |||
694 | DPRINTK("data %s\n", qc->tf.flags & ATA_TFLAG_WRITE ? "write" : "read"); | ||
695 | |||
696 | if (PageHighMem(page)) { | ||
697 | unsigned long flags; | ||
698 | |||
699 | /* FIXME: use a bounce buffer */ | ||
700 | local_irq_save(flags); | ||
701 | buf = kmap_atomic(page, KM_IRQ0); | ||
702 | |||
703 | /* do the actual data transfer */ | ||
704 | ap->ops->sff_data_xfer(qc->dev, buf + offset, qc->sect_size, | ||
705 | do_write); | ||
706 | |||
707 | kunmap_atomic(buf, KM_IRQ0); | ||
708 | local_irq_restore(flags); | ||
709 | } else { | ||
710 | buf = page_address(page); | ||
711 | ap->ops->sff_data_xfer(qc->dev, buf + offset, qc->sect_size, | ||
712 | do_write); | ||
713 | } | ||
714 | |||
715 | qc->curbytes += qc->sect_size; | ||
716 | qc->cursg_ofs += qc->sect_size; | ||
717 | |||
718 | if (qc->cursg_ofs == qc->cursg->length) { | ||
719 | qc->cursg = sg_next(qc->cursg); | ||
720 | qc->cursg_ofs = 0; | ||
721 | } | ||
250 | } | 722 | } |
251 | 723 | ||
252 | /** | 724 | /** |
253 | * ata_bmdma_start - Start a PCI IDE BMDMA transaction | 725 | * ata_pio_sectors - Transfer one or many sectors. |
254 | * @qc: Info associated with this ATA transaction. | 726 | * @qc: Command on going |
727 | * | ||
728 | * Transfer one or many sectors of data from/to the | ||
729 | * ATA device for the DRQ request. | ||
255 | * | 730 | * |
256 | * LOCKING: | 731 | * LOCKING: |
257 | * spin_lock_irqsave(host lock) | 732 | * Inherited from caller. |
258 | */ | 733 | */ |
259 | void ata_bmdma_start(struct ata_queued_cmd *qc) | 734 | static void ata_pio_sectors(struct ata_queued_cmd *qc) |
735 | { | ||
736 | if (is_multi_taskfile(&qc->tf)) { | ||
737 | /* READ/WRITE MULTIPLE */ | ||
738 | unsigned int nsect; | ||
739 | |||
740 | WARN_ON(qc->dev->multi_count == 0); | ||
741 | |||
742 | nsect = min((qc->nbytes - qc->curbytes) / qc->sect_size, | ||
743 | qc->dev->multi_count); | ||
744 | while (nsect--) | ||
745 | ata_pio_sector(qc); | ||
746 | } else | ||
747 | ata_pio_sector(qc); | ||
748 | |||
749 | ata_sff_altstatus(qc->ap); /* flush */ | ||
750 | } | ||
751 | |||
752 | /** | ||
753 | * atapi_send_cdb - Write CDB bytes to hardware | ||
754 | * @ap: Port to which ATAPI device is attached. | ||
755 | * @qc: Taskfile currently active | ||
756 | * | ||
757 | * When device has indicated its readiness to accept | ||
758 | * a CDB, this function is called. Send the CDB. | ||
759 | * | ||
760 | * LOCKING: | ||
761 | * caller. | ||
762 | */ | ||
763 | static void atapi_send_cdb(struct ata_port *ap, struct ata_queued_cmd *qc) | ||
764 | { | ||
765 | /* send SCSI cdb */ | ||
766 | DPRINTK("send cdb\n"); | ||
767 | WARN_ON(qc->dev->cdb_len < 12); | ||
768 | |||
769 | ap->ops->sff_data_xfer(qc->dev, qc->cdb, qc->dev->cdb_len, 1); | ||
770 | ata_sff_altstatus(ap); /* flush */ | ||
771 | |||
772 | switch (qc->tf.protocol) { | ||
773 | case ATAPI_PROT_PIO: | ||
774 | ap->hsm_task_state = HSM_ST; | ||
775 | break; | ||
776 | case ATAPI_PROT_NODATA: | ||
777 | ap->hsm_task_state = HSM_ST_LAST; | ||
778 | break; | ||
779 | case ATAPI_PROT_DMA: | ||
780 | ap->hsm_task_state = HSM_ST_LAST; | ||
781 | /* initiate bmdma */ | ||
782 | ap->ops->bmdma_start(qc); | ||
783 | break; | ||
784 | } | ||
785 | } | ||
786 | |||
787 | /** | ||
788 | * __atapi_pio_bytes - Transfer data from/to the ATAPI device. | ||
789 | * @qc: Command on going | ||
790 | * @bytes: number of bytes | ||
791 | * | ||
792 | * Transfer Transfer data from/to the ATAPI device. | ||
793 | * | ||
794 | * LOCKING: | ||
795 | * Inherited from caller. | ||
796 | * | ||
797 | */ | ||
798 | static int __atapi_pio_bytes(struct ata_queued_cmd *qc, unsigned int bytes) | ||
260 | { | 799 | { |
800 | int rw = (qc->tf.flags & ATA_TFLAG_WRITE) ? WRITE : READ; | ||
261 | struct ata_port *ap = qc->ap; | 801 | struct ata_port *ap = qc->ap; |
262 | u8 dmactl; | 802 | struct ata_device *dev = qc->dev; |
803 | struct ata_eh_info *ehi = &dev->link->eh_info; | ||
804 | struct scatterlist *sg; | ||
805 | struct page *page; | ||
806 | unsigned char *buf; | ||
807 | unsigned int offset, count, consumed; | ||
808 | |||
809 | next_sg: | ||
810 | sg = qc->cursg; | ||
811 | if (unlikely(!sg)) { | ||
812 | ata_ehi_push_desc(ehi, "unexpected or too much trailing data " | ||
813 | "buf=%u cur=%u bytes=%u", | ||
814 | qc->nbytes, qc->curbytes, bytes); | ||
815 | return -1; | ||
816 | } | ||
263 | 817 | ||
264 | /* start host DMA transaction */ | 818 | page = sg_page(sg); |
265 | dmactl = ioread8(ap->ioaddr.bmdma_addr + ATA_DMA_CMD); | 819 | offset = sg->offset + qc->cursg_ofs; |
266 | iowrite8(dmactl | ATA_DMA_START, ap->ioaddr.bmdma_addr + ATA_DMA_CMD); | ||
267 | 820 | ||
268 | /* Strictly, one may wish to issue an ioread8() here, to | 821 | /* get the current page and offset */ |
269 | * flush the mmio write. However, control also passes | 822 | page = nth_page(page, (offset >> PAGE_SHIFT)); |
270 | * to the hardware at this point, and it will interrupt | 823 | offset %= PAGE_SIZE; |
271 | * us when we are to resume control. So, in effect, | 824 | |
272 | * we don't care when the mmio write flushes. | 825 | /* don't overrun current sg */ |
273 | * Further, a read of the DMA status register _immediately_ | 826 | count = min(sg->length - qc->cursg_ofs, bytes); |
274 | * following the write may not be what certain flaky hardware | 827 | |
275 | * is expected, so I think it is best to not add a readb() | 828 | /* don't cross page boundaries */ |
276 | * without first all the MMIO ATA cards/mobos. | 829 | count = min(count, (unsigned int)PAGE_SIZE - offset); |
277 | * Or maybe I'm just being paranoid. | 830 | |
278 | * | 831 | DPRINTK("data %s\n", qc->tf.flags & ATA_TFLAG_WRITE ? "write" : "read"); |
279 | * FIXME: The posting of this write means I/O starts are | 832 | |
280 | * unneccessarily delayed for MMIO | 833 | if (PageHighMem(page)) { |
834 | unsigned long flags; | ||
835 | |||
836 | /* FIXME: use bounce buffer */ | ||
837 | local_irq_save(flags); | ||
838 | buf = kmap_atomic(page, KM_IRQ0); | ||
839 | |||
840 | /* do the actual data transfer */ | ||
841 | consumed = ap->ops->sff_data_xfer(dev, buf + offset, count, rw); | ||
842 | |||
843 | kunmap_atomic(buf, KM_IRQ0); | ||
844 | local_irq_restore(flags); | ||
845 | } else { | ||
846 | buf = page_address(page); | ||
847 | consumed = ap->ops->sff_data_xfer(dev, buf + offset, count, rw); | ||
848 | } | ||
849 | |||
850 | bytes -= min(bytes, consumed); | ||
851 | qc->curbytes += count; | ||
852 | qc->cursg_ofs += count; | ||
853 | |||
854 | if (qc->cursg_ofs == sg->length) { | ||
855 | qc->cursg = sg_next(qc->cursg); | ||
856 | qc->cursg_ofs = 0; | ||
857 | } | ||
858 | |||
859 | /* consumed can be larger than count only for the last transfer */ | ||
860 | WARN_ON(qc->cursg && count != consumed); | ||
861 | |||
862 | if (bytes) | ||
863 | goto next_sg; | ||
864 | return 0; | ||
865 | } | ||
866 | |||
867 | /** | ||
868 | * atapi_pio_bytes - Transfer data from/to the ATAPI device. | ||
869 | * @qc: Command on going | ||
870 | * | ||
871 | * Transfer Transfer data from/to the ATAPI device. | ||
872 | * | ||
873 | * LOCKING: | ||
874 | * Inherited from caller. | ||
875 | */ | ||
876 | static void atapi_pio_bytes(struct ata_queued_cmd *qc) | ||
877 | { | ||
878 | struct ata_port *ap = qc->ap; | ||
879 | struct ata_device *dev = qc->dev; | ||
880 | struct ata_eh_info *ehi = &dev->link->eh_info; | ||
881 | unsigned int ireason, bc_lo, bc_hi, bytes; | ||
882 | int i_write, do_write = (qc->tf.flags & ATA_TFLAG_WRITE) ? 1 : 0; | ||
883 | |||
884 | /* Abuse qc->result_tf for temp storage of intermediate TF | ||
885 | * here to save some kernel stack usage. | ||
886 | * For normal completion, qc->result_tf is not relevant. For | ||
887 | * error, qc->result_tf is later overwritten by ata_qc_complete(). | ||
888 | * So, the correctness of qc->result_tf is not affected. | ||
281 | */ | 889 | */ |
890 | ap->ops->sff_tf_read(ap, &qc->result_tf); | ||
891 | ireason = qc->result_tf.nsect; | ||
892 | bc_lo = qc->result_tf.lbam; | ||
893 | bc_hi = qc->result_tf.lbah; | ||
894 | bytes = (bc_hi << 8) | bc_lo; | ||
895 | |||
896 | /* shall be cleared to zero, indicating xfer of data */ | ||
897 | if (unlikely(ireason & (1 << 0))) | ||
898 | goto atapi_check; | ||
899 | |||
900 | /* make sure transfer direction matches expected */ | ||
901 | i_write = ((ireason & (1 << 1)) == 0) ? 1 : 0; | ||
902 | if (unlikely(do_write != i_write)) | ||
903 | goto atapi_check; | ||
904 | |||
905 | if (unlikely(!bytes)) | ||
906 | goto atapi_check; | ||
907 | |||
908 | VPRINTK("ata%u: xfering %d bytes\n", ap->print_id, bytes); | ||
909 | |||
910 | if (unlikely(__atapi_pio_bytes(qc, bytes))) | ||
911 | goto err_out; | ||
912 | ata_sff_altstatus(ap); /* flush */ | ||
913 | |||
914 | return; | ||
915 | |||
916 | atapi_check: | ||
917 | ata_ehi_push_desc(ehi, "ATAPI check failed (ireason=0x%x bytes=%u)", | ||
918 | ireason, bytes); | ||
919 | err_out: | ||
920 | qc->err_mask |= AC_ERR_HSM; | ||
921 | ap->hsm_task_state = HSM_ST_ERR; | ||
282 | } | 922 | } |
283 | 923 | ||
284 | /** | 924 | /** |
285 | * ata_bmdma_irq_clear - Clear PCI IDE BMDMA interrupt. | 925 | * ata_hsm_ok_in_wq - Check if the qc can be handled in the workqueue. |
286 | * @ap: Port associated with this ATA transaction. | 926 | * @ap: the target ata_port |
927 | * @qc: qc on going | ||
287 | * | 928 | * |
288 | * Clear interrupt and error flags in DMA status register. | 929 | * RETURNS: |
930 | * 1 if ok in workqueue, 0 otherwise. | ||
931 | */ | ||
932 | static inline int ata_hsm_ok_in_wq(struct ata_port *ap, struct ata_queued_cmd *qc) | ||
933 | { | ||
934 | if (qc->tf.flags & ATA_TFLAG_POLLING) | ||
935 | return 1; | ||
936 | |||
937 | if (ap->hsm_task_state == HSM_ST_FIRST) { | ||
938 | if (qc->tf.protocol == ATA_PROT_PIO && | ||
939 | (qc->tf.flags & ATA_TFLAG_WRITE)) | ||
940 | return 1; | ||
941 | |||
942 | if (ata_is_atapi(qc->tf.protocol) && | ||
943 | !(qc->dev->flags & ATA_DFLAG_CDB_INTR)) | ||
944 | return 1; | ||
945 | } | ||
946 | |||
947 | return 0; | ||
948 | } | ||
949 | |||
950 | /** | ||
951 | * ata_hsm_qc_complete - finish a qc running on standard HSM | ||
952 | * @qc: Command to complete | ||
953 | * @in_wq: 1 if called from workqueue, 0 otherwise | ||
289 | * | 954 | * |
290 | * May be used as the irq_clear() entry in ata_port_operations. | 955 | * Finish @qc which is running on standard HSM. |
291 | * | 956 | * |
292 | * LOCKING: | 957 | * LOCKING: |
293 | * spin_lock_irqsave(host lock) | 958 | * If @in_wq is zero, spin_lock_irqsave(host lock). |
959 | * Otherwise, none on entry and grabs host lock. | ||
294 | */ | 960 | */ |
295 | void ata_bmdma_irq_clear(struct ata_port *ap) | 961 | static void ata_hsm_qc_complete(struct ata_queued_cmd *qc, int in_wq) |
296 | { | 962 | { |
297 | void __iomem *mmio = ap->ioaddr.bmdma_addr; | 963 | struct ata_port *ap = qc->ap; |
964 | unsigned long flags; | ||
298 | 965 | ||
299 | if (!mmio) | 966 | if (ap->ops->error_handler) { |
300 | return; | 967 | if (in_wq) { |
968 | spin_lock_irqsave(ap->lock, flags); | ||
969 | |||
970 | /* EH might have kicked in while host lock is | ||
971 | * released. | ||
972 | */ | ||
973 | qc = ata_qc_from_tag(ap, qc->tag); | ||
974 | if (qc) { | ||
975 | if (likely(!(qc->err_mask & AC_ERR_HSM))) { | ||
976 | ap->ops->sff_irq_on(ap); | ||
977 | ata_qc_complete(qc); | ||
978 | } else | ||
979 | ata_port_freeze(ap); | ||
980 | } | ||
981 | |||
982 | spin_unlock_irqrestore(ap->lock, flags); | ||
983 | } else { | ||
984 | if (likely(!(qc->err_mask & AC_ERR_HSM))) | ||
985 | ata_qc_complete(qc); | ||
986 | else | ||
987 | ata_port_freeze(ap); | ||
988 | } | ||
989 | } else { | ||
990 | if (in_wq) { | ||
991 | spin_lock_irqsave(ap->lock, flags); | ||
992 | ap->ops->sff_irq_on(ap); | ||
993 | ata_qc_complete(qc); | ||
994 | spin_unlock_irqrestore(ap->lock, flags); | ||
995 | } else | ||
996 | ata_qc_complete(qc); | ||
997 | } | ||
998 | } | ||
301 | 999 | ||
302 | iowrite8(ioread8(mmio + ATA_DMA_STATUS), mmio + ATA_DMA_STATUS); | 1000 | /** |
1001 | * ata_sff_hsm_move - move the HSM to the next state. | ||
1002 | * @ap: the target ata_port | ||
1003 | * @qc: qc on going | ||
1004 | * @status: current device status | ||
1005 | * @in_wq: 1 if called from workqueue, 0 otherwise | ||
1006 | * | ||
1007 | * RETURNS: | ||
1008 | * 1 when poll next status needed, 0 otherwise. | ||
1009 | */ | ||
1010 | int ata_sff_hsm_move(struct ata_port *ap, struct ata_queued_cmd *qc, | ||
1011 | u8 status, int in_wq) | ||
1012 | { | ||
1013 | unsigned long flags = 0; | ||
1014 | int poll_next; | ||
1015 | |||
1016 | WARN_ON((qc->flags & ATA_QCFLAG_ACTIVE) == 0); | ||
1017 | |||
1018 | /* Make sure ata_sff_qc_issue() does not throw things | ||
1019 | * like DMA polling into the workqueue. Notice that | ||
1020 | * in_wq is not equivalent to (qc->tf.flags & ATA_TFLAG_POLLING). | ||
1021 | */ | ||
1022 | WARN_ON(in_wq != ata_hsm_ok_in_wq(ap, qc)); | ||
1023 | |||
1024 | fsm_start: | ||
1025 | DPRINTK("ata%u: protocol %d task_state %d (dev_stat 0x%X)\n", | ||
1026 | ap->print_id, qc->tf.protocol, ap->hsm_task_state, status); | ||
1027 | |||
1028 | switch (ap->hsm_task_state) { | ||
1029 | case HSM_ST_FIRST: | ||
1030 | /* Send first data block or PACKET CDB */ | ||
1031 | |||
1032 | /* If polling, we will stay in the work queue after | ||
1033 | * sending the data. Otherwise, interrupt handler | ||
1034 | * takes over after sending the data. | ||
1035 | */ | ||
1036 | poll_next = (qc->tf.flags & ATA_TFLAG_POLLING); | ||
1037 | |||
1038 | /* check device status */ | ||
1039 | if (unlikely((status & ATA_DRQ) == 0)) { | ||
1040 | /* handle BSY=0, DRQ=0 as error */ | ||
1041 | if (likely(status & (ATA_ERR | ATA_DF))) | ||
1042 | /* device stops HSM for abort/error */ | ||
1043 | qc->err_mask |= AC_ERR_DEV; | ||
1044 | else | ||
1045 | /* HSM violation. Let EH handle this */ | ||
1046 | qc->err_mask |= AC_ERR_HSM; | ||
1047 | |||
1048 | ap->hsm_task_state = HSM_ST_ERR; | ||
1049 | goto fsm_start; | ||
1050 | } | ||
1051 | |||
1052 | /* Device should not ask for data transfer (DRQ=1) | ||
1053 | * when it finds something wrong. | ||
1054 | * We ignore DRQ here and stop the HSM by | ||
1055 | * changing hsm_task_state to HSM_ST_ERR and | ||
1056 | * let the EH abort the command or reset the device. | ||
1057 | */ | ||
1058 | if (unlikely(status & (ATA_ERR | ATA_DF))) { | ||
1059 | /* Some ATAPI tape drives forget to clear the ERR bit | ||
1060 | * when doing the next command (mostly request sense). | ||
1061 | * We ignore ERR here to workaround and proceed sending | ||
1062 | * the CDB. | ||
1063 | */ | ||
1064 | if (!(qc->dev->horkage & ATA_HORKAGE_STUCK_ERR)) { | ||
1065 | ata_port_printk(ap, KERN_WARNING, | ||
1066 | "DRQ=1 with device error, " | ||
1067 | "dev_stat 0x%X\n", status); | ||
1068 | qc->err_mask |= AC_ERR_HSM; | ||
1069 | ap->hsm_task_state = HSM_ST_ERR; | ||
1070 | goto fsm_start; | ||
1071 | } | ||
1072 | } | ||
1073 | |||
1074 | /* Send the CDB (atapi) or the first data block (ata pio out). | ||
1075 | * During the state transition, interrupt handler shouldn't | ||
1076 | * be invoked before the data transfer is complete and | ||
1077 | * hsm_task_state is changed. Hence, the following locking. | ||
1078 | */ | ||
1079 | if (in_wq) | ||
1080 | spin_lock_irqsave(ap->lock, flags); | ||
1081 | |||
1082 | if (qc->tf.protocol == ATA_PROT_PIO) { | ||
1083 | /* PIO data out protocol. | ||
1084 | * send first data block. | ||
1085 | */ | ||
1086 | |||
1087 | /* ata_pio_sectors() might change the state | ||
1088 | * to HSM_ST_LAST. so, the state is changed here | ||
1089 | * before ata_pio_sectors(). | ||
1090 | */ | ||
1091 | ap->hsm_task_state = HSM_ST; | ||
1092 | ata_pio_sectors(qc); | ||
1093 | } else | ||
1094 | /* send CDB */ | ||
1095 | atapi_send_cdb(ap, qc); | ||
1096 | |||
1097 | if (in_wq) | ||
1098 | spin_unlock_irqrestore(ap->lock, flags); | ||
1099 | |||
1100 | /* if polling, ata_pio_task() handles the rest. | ||
1101 | * otherwise, interrupt handler takes over from here. | ||
1102 | */ | ||
1103 | break; | ||
1104 | |||
1105 | case HSM_ST: | ||
1106 | /* complete command or read/write the data register */ | ||
1107 | if (qc->tf.protocol == ATAPI_PROT_PIO) { | ||
1108 | /* ATAPI PIO protocol */ | ||
1109 | if ((status & ATA_DRQ) == 0) { | ||
1110 | /* No more data to transfer or device error. | ||
1111 | * Device error will be tagged in HSM_ST_LAST. | ||
1112 | */ | ||
1113 | ap->hsm_task_state = HSM_ST_LAST; | ||
1114 | goto fsm_start; | ||
1115 | } | ||
1116 | |||
1117 | /* Device should not ask for data transfer (DRQ=1) | ||
1118 | * when it finds something wrong. | ||
1119 | * We ignore DRQ here and stop the HSM by | ||
1120 | * changing hsm_task_state to HSM_ST_ERR and | ||
1121 | * let the EH abort the command or reset the device. | ||
1122 | */ | ||
1123 | if (unlikely(status & (ATA_ERR | ATA_DF))) { | ||
1124 | ata_port_printk(ap, KERN_WARNING, "DRQ=1 with " | ||
1125 | "device error, dev_stat 0x%X\n", | ||
1126 | status); | ||
1127 | qc->err_mask |= AC_ERR_HSM; | ||
1128 | ap->hsm_task_state = HSM_ST_ERR; | ||
1129 | goto fsm_start; | ||
1130 | } | ||
1131 | |||
1132 | atapi_pio_bytes(qc); | ||
1133 | |||
1134 | if (unlikely(ap->hsm_task_state == HSM_ST_ERR)) | ||
1135 | /* bad ireason reported by device */ | ||
1136 | goto fsm_start; | ||
1137 | |||
1138 | } else { | ||
1139 | /* ATA PIO protocol */ | ||
1140 | if (unlikely((status & ATA_DRQ) == 0)) { | ||
1141 | /* handle BSY=0, DRQ=0 as error */ | ||
1142 | if (likely(status & (ATA_ERR | ATA_DF))) | ||
1143 | /* device stops HSM for abort/error */ | ||
1144 | qc->err_mask |= AC_ERR_DEV; | ||
1145 | else | ||
1146 | /* HSM violation. Let EH handle this. | ||
1147 | * Phantom devices also trigger this | ||
1148 | * condition. Mark hint. | ||
1149 | */ | ||
1150 | qc->err_mask |= AC_ERR_HSM | | ||
1151 | AC_ERR_NODEV_HINT; | ||
1152 | |||
1153 | ap->hsm_task_state = HSM_ST_ERR; | ||
1154 | goto fsm_start; | ||
1155 | } | ||
1156 | |||
1157 | /* For PIO reads, some devices may ask for | ||
1158 | * data transfer (DRQ=1) alone with ERR=1. | ||
1159 | * We respect DRQ here and transfer one | ||
1160 | * block of junk data before changing the | ||
1161 | * hsm_task_state to HSM_ST_ERR. | ||
1162 | * | ||
1163 | * For PIO writes, ERR=1 DRQ=1 doesn't make | ||
1164 | * sense since the data block has been | ||
1165 | * transferred to the device. | ||
1166 | */ | ||
1167 | if (unlikely(status & (ATA_ERR | ATA_DF))) { | ||
1168 | /* data might be corrputed */ | ||
1169 | qc->err_mask |= AC_ERR_DEV; | ||
1170 | |||
1171 | if (!(qc->tf.flags & ATA_TFLAG_WRITE)) { | ||
1172 | ata_pio_sectors(qc); | ||
1173 | status = ata_wait_idle(ap); | ||
1174 | } | ||
1175 | |||
1176 | if (status & (ATA_BUSY | ATA_DRQ)) | ||
1177 | qc->err_mask |= AC_ERR_HSM; | ||
1178 | |||
1179 | /* ata_pio_sectors() might change the | ||
1180 | * state to HSM_ST_LAST. so, the state | ||
1181 | * is changed after ata_pio_sectors(). | ||
1182 | */ | ||
1183 | ap->hsm_task_state = HSM_ST_ERR; | ||
1184 | goto fsm_start; | ||
1185 | } | ||
1186 | |||
1187 | ata_pio_sectors(qc); | ||
1188 | |||
1189 | if (ap->hsm_task_state == HSM_ST_LAST && | ||
1190 | (!(qc->tf.flags & ATA_TFLAG_WRITE))) { | ||
1191 | /* all data read */ | ||
1192 | status = ata_wait_idle(ap); | ||
1193 | goto fsm_start; | ||
1194 | } | ||
1195 | } | ||
1196 | |||
1197 | poll_next = 1; | ||
1198 | break; | ||
1199 | |||
1200 | case HSM_ST_LAST: | ||
1201 | if (unlikely(!ata_ok(status))) { | ||
1202 | qc->err_mask |= __ac_err_mask(status); | ||
1203 | ap->hsm_task_state = HSM_ST_ERR; | ||
1204 | goto fsm_start; | ||
1205 | } | ||
1206 | |||
1207 | /* no more data to transfer */ | ||
1208 | DPRINTK("ata%u: dev %u command complete, drv_stat 0x%x\n", | ||
1209 | ap->print_id, qc->dev->devno, status); | ||
1210 | |||
1211 | WARN_ON(qc->err_mask); | ||
1212 | |||
1213 | ap->hsm_task_state = HSM_ST_IDLE; | ||
1214 | |||
1215 | /* complete taskfile transaction */ | ||
1216 | ata_hsm_qc_complete(qc, in_wq); | ||
1217 | |||
1218 | poll_next = 0; | ||
1219 | break; | ||
1220 | |||
1221 | case HSM_ST_ERR: | ||
1222 | /* make sure qc->err_mask is available to | ||
1223 | * know what's wrong and recover | ||
1224 | */ | ||
1225 | WARN_ON(qc->err_mask == 0); | ||
1226 | |||
1227 | ap->hsm_task_state = HSM_ST_IDLE; | ||
1228 | |||
1229 | /* complete taskfile transaction */ | ||
1230 | ata_hsm_qc_complete(qc, in_wq); | ||
1231 | |||
1232 | poll_next = 0; | ||
1233 | break; | ||
1234 | default: | ||
1235 | poll_next = 0; | ||
1236 | BUG(); | ||
1237 | } | ||
1238 | |||
1239 | return poll_next; | ||
1240 | } | ||
1241 | |||
1242 | void ata_pio_task(struct work_struct *work) | ||
1243 | { | ||
1244 | struct ata_port *ap = | ||
1245 | container_of(work, struct ata_port, port_task.work); | ||
1246 | struct ata_queued_cmd *qc = ap->port_task_data; | ||
1247 | u8 status; | ||
1248 | int poll_next; | ||
1249 | |||
1250 | fsm_start: | ||
1251 | WARN_ON(ap->hsm_task_state == HSM_ST_IDLE); | ||
1252 | |||
1253 | /* | ||
1254 | * This is purely heuristic. This is a fast path. | ||
1255 | * Sometimes when we enter, BSY will be cleared in | ||
1256 | * a chk-status or two. If not, the drive is probably seeking | ||
1257 | * or something. Snooze for a couple msecs, then | ||
1258 | * chk-status again. If still busy, queue delayed work. | ||
1259 | */ | ||
1260 | status = ata_sff_busy_wait(ap, ATA_BUSY, 5); | ||
1261 | if (status & ATA_BUSY) { | ||
1262 | msleep(2); | ||
1263 | status = ata_sff_busy_wait(ap, ATA_BUSY, 10); | ||
1264 | if (status & ATA_BUSY) { | ||
1265 | ata_pio_queue_task(ap, qc, ATA_SHORT_PAUSE); | ||
1266 | return; | ||
1267 | } | ||
1268 | } | ||
1269 | |||
1270 | /* move the HSM */ | ||
1271 | poll_next = ata_sff_hsm_move(ap, qc, status, 1); | ||
1272 | |||
1273 | /* another command or interrupt handler | ||
1274 | * may be running at this point. | ||
1275 | */ | ||
1276 | if (poll_next) | ||
1277 | goto fsm_start; | ||
303 | } | 1278 | } |
304 | 1279 | ||
305 | /** | 1280 | /** |
306 | * ata_bmdma_status - Read PCI IDE BMDMA status | 1281 | * ata_sff_qc_issue - issue taskfile to device in proto-dependent manner |
307 | * @ap: Port associated with this ATA transaction. | 1282 | * @qc: command to issue to device |
308 | * | 1283 | * |
309 | * Read and return BMDMA status register. | 1284 | * Using various libata functions and hooks, this function |
1285 | * starts an ATA command. ATA commands are grouped into | ||
1286 | * classes called "protocols", and issuing each type of protocol | ||
1287 | * is slightly different. | ||
310 | * | 1288 | * |
311 | * May be used as the bmdma_status() entry in ata_port_operations. | 1289 | * May be used as the qc_issue() entry in ata_port_operations. |
312 | * | 1290 | * |
313 | * LOCKING: | 1291 | * LOCKING: |
314 | * spin_lock_irqsave(host lock) | 1292 | * spin_lock_irqsave(host lock) |
1293 | * | ||
1294 | * RETURNS: | ||
1295 | * Zero on success, AC_ERR_* mask on failure | ||
315 | */ | 1296 | */ |
316 | u8 ata_bmdma_status(struct ata_port *ap) | 1297 | unsigned int ata_sff_qc_issue(struct ata_queued_cmd *qc) |
317 | { | 1298 | { |
318 | return ioread8(ap->ioaddr.bmdma_addr + ATA_DMA_STATUS); | 1299 | struct ata_port *ap = qc->ap; |
1300 | |||
1301 | /* Use polling pio if the LLD doesn't handle | ||
1302 | * interrupt driven pio and atapi CDB interrupt. | ||
1303 | */ | ||
1304 | if (ap->flags & ATA_FLAG_PIO_POLLING) { | ||
1305 | switch (qc->tf.protocol) { | ||
1306 | case ATA_PROT_PIO: | ||
1307 | case ATA_PROT_NODATA: | ||
1308 | case ATAPI_PROT_PIO: | ||
1309 | case ATAPI_PROT_NODATA: | ||
1310 | qc->tf.flags |= ATA_TFLAG_POLLING; | ||
1311 | break; | ||
1312 | case ATAPI_PROT_DMA: | ||
1313 | if (qc->dev->flags & ATA_DFLAG_CDB_INTR) | ||
1314 | /* see ata_dma_blacklisted() */ | ||
1315 | BUG(); | ||
1316 | break; | ||
1317 | default: | ||
1318 | break; | ||
1319 | } | ||
1320 | } | ||
1321 | |||
1322 | /* select the device */ | ||
1323 | ata_dev_select(ap, qc->dev->devno, 1, 0); | ||
1324 | |||
1325 | /* start the command */ | ||
1326 | switch (qc->tf.protocol) { | ||
1327 | case ATA_PROT_NODATA: | ||
1328 | if (qc->tf.flags & ATA_TFLAG_POLLING) | ||
1329 | ata_qc_set_polling(qc); | ||
1330 | |||
1331 | ata_tf_to_host(ap, &qc->tf); | ||
1332 | ap->hsm_task_state = HSM_ST_LAST; | ||
1333 | |||
1334 | if (qc->tf.flags & ATA_TFLAG_POLLING) | ||
1335 | ata_pio_queue_task(ap, qc, 0); | ||
1336 | |||
1337 | break; | ||
1338 | |||
1339 | case ATA_PROT_DMA: | ||
1340 | WARN_ON(qc->tf.flags & ATA_TFLAG_POLLING); | ||
1341 | |||
1342 | ap->ops->sff_tf_load(ap, &qc->tf); /* load tf registers */ | ||
1343 | ap->ops->bmdma_setup(qc); /* set up bmdma */ | ||
1344 | ap->ops->bmdma_start(qc); /* initiate bmdma */ | ||
1345 | ap->hsm_task_state = HSM_ST_LAST; | ||
1346 | break; | ||
1347 | |||
1348 | case ATA_PROT_PIO: | ||
1349 | if (qc->tf.flags & ATA_TFLAG_POLLING) | ||
1350 | ata_qc_set_polling(qc); | ||
1351 | |||
1352 | ata_tf_to_host(ap, &qc->tf); | ||
1353 | |||
1354 | if (qc->tf.flags & ATA_TFLAG_WRITE) { | ||
1355 | /* PIO data out protocol */ | ||
1356 | ap->hsm_task_state = HSM_ST_FIRST; | ||
1357 | ata_pio_queue_task(ap, qc, 0); | ||
1358 | |||
1359 | /* always send first data block using | ||
1360 | * the ata_pio_task() codepath. | ||
1361 | */ | ||
1362 | } else { | ||
1363 | /* PIO data in protocol */ | ||
1364 | ap->hsm_task_state = HSM_ST; | ||
1365 | |||
1366 | if (qc->tf.flags & ATA_TFLAG_POLLING) | ||
1367 | ata_pio_queue_task(ap, qc, 0); | ||
1368 | |||
1369 | /* if polling, ata_pio_task() handles the rest. | ||
1370 | * otherwise, interrupt handler takes over from here. | ||
1371 | */ | ||
1372 | } | ||
1373 | |||
1374 | break; | ||
1375 | |||
1376 | case ATAPI_PROT_PIO: | ||
1377 | case ATAPI_PROT_NODATA: | ||
1378 | if (qc->tf.flags & ATA_TFLAG_POLLING) | ||
1379 | ata_qc_set_polling(qc); | ||
1380 | |||
1381 | ata_tf_to_host(ap, &qc->tf); | ||
1382 | |||
1383 | ap->hsm_task_state = HSM_ST_FIRST; | ||
1384 | |||
1385 | /* send cdb by polling if no cdb interrupt */ | ||
1386 | if ((!(qc->dev->flags & ATA_DFLAG_CDB_INTR)) || | ||
1387 | (qc->tf.flags & ATA_TFLAG_POLLING)) | ||
1388 | ata_pio_queue_task(ap, qc, 0); | ||
1389 | break; | ||
1390 | |||
1391 | case ATAPI_PROT_DMA: | ||
1392 | WARN_ON(qc->tf.flags & ATA_TFLAG_POLLING); | ||
1393 | |||
1394 | ap->ops->sff_tf_load(ap, &qc->tf); /* load tf registers */ | ||
1395 | ap->ops->bmdma_setup(qc); /* set up bmdma */ | ||
1396 | ap->hsm_task_state = HSM_ST_FIRST; | ||
1397 | |||
1398 | /* send cdb by polling if no cdb interrupt */ | ||
1399 | if (!(qc->dev->flags & ATA_DFLAG_CDB_INTR)) | ||
1400 | ata_pio_queue_task(ap, qc, 0); | ||
1401 | break; | ||
1402 | |||
1403 | default: | ||
1404 | WARN_ON(1); | ||
1405 | return AC_ERR_SYSTEM; | ||
1406 | } | ||
1407 | |||
1408 | return 0; | ||
319 | } | 1409 | } |
320 | 1410 | ||
321 | /** | 1411 | /** |
322 | * ata_bmdma_stop - Stop PCI IDE BMDMA transfer | 1412 | * ata_sff_qc_fill_rtf - fill result TF using ->sff_tf_read |
323 | * @qc: Command we are ending DMA for | 1413 | * @qc: qc to fill result TF for |
324 | * | 1414 | * |
325 | * Clears the ATA_DMA_START flag in the dma control register | 1415 | * @qc is finished and result TF needs to be filled. Fill it |
1416 | * using ->sff_tf_read. | ||
326 | * | 1417 | * |
327 | * May be used as the bmdma_stop() entry in ata_port_operations. | 1418 | * LOCKING: |
1419 | * spin_lock_irqsave(host lock) | ||
1420 | * | ||
1421 | * RETURNS: | ||
1422 | * true indicating that result TF is successfully filled. | ||
1423 | */ | ||
1424 | bool ata_sff_qc_fill_rtf(struct ata_queued_cmd *qc) | ||
1425 | { | ||
1426 | qc->ap->ops->sff_tf_read(qc->ap, &qc->result_tf); | ||
1427 | return true; | ||
1428 | } | ||
1429 | |||
1430 | /** | ||
1431 | * ata_sff_host_intr - Handle host interrupt for given (port, task) | ||
1432 | * @ap: Port on which interrupt arrived (possibly...) | ||
1433 | * @qc: Taskfile currently active in engine | ||
1434 | * | ||
1435 | * Handle host interrupt for given queued command. Currently, | ||
1436 | * only DMA interrupts are handled. All other commands are | ||
1437 | * handled via polling with interrupts disabled (nIEN bit). | ||
328 | * | 1438 | * |
329 | * LOCKING: | 1439 | * LOCKING: |
330 | * spin_lock_irqsave(host lock) | 1440 | * spin_lock_irqsave(host lock) |
1441 | * | ||
1442 | * RETURNS: | ||
1443 | * One if interrupt was handled, zero if not (shared irq). | ||
331 | */ | 1444 | */ |
332 | void ata_bmdma_stop(struct ata_queued_cmd *qc) | 1445 | inline unsigned int ata_sff_host_intr(struct ata_port *ap, |
1446 | struct ata_queued_cmd *qc) | ||
333 | { | 1447 | { |
334 | struct ata_port *ap = qc->ap; | 1448 | struct ata_eh_info *ehi = &ap->link.eh_info; |
335 | void __iomem *mmio = ap->ioaddr.bmdma_addr; | 1449 | u8 status, host_stat = 0; |
336 | 1450 | ||
337 | /* clear start/stop bit */ | 1451 | VPRINTK("ata%u: protocol %d task_state %d\n", |
338 | iowrite8(ioread8(mmio + ATA_DMA_CMD) & ~ATA_DMA_START, | 1452 | ap->print_id, qc->tf.protocol, ap->hsm_task_state); |
339 | mmio + ATA_DMA_CMD); | ||
340 | 1453 | ||
341 | /* one-PIO-cycle guaranteed wait, per spec, for HDMA1:0 transition */ | 1454 | /* Check whether we are expecting interrupt in this state */ |
342 | ata_altstatus(ap); /* dummy read */ | 1455 | switch (ap->hsm_task_state) { |
1456 | case HSM_ST_FIRST: | ||
1457 | /* Some pre-ATAPI-4 devices assert INTRQ | ||
1458 | * at this state when ready to receive CDB. | ||
1459 | */ | ||
1460 | |||
1461 | /* Check the ATA_DFLAG_CDB_INTR flag is enough here. | ||
1462 | * The flag was turned on only for atapi devices. No | ||
1463 | * need to check ata_is_atapi(qc->tf.protocol) again. | ||
1464 | */ | ||
1465 | if (!(qc->dev->flags & ATA_DFLAG_CDB_INTR)) | ||
1466 | goto idle_irq; | ||
1467 | break; | ||
1468 | case HSM_ST_LAST: | ||
1469 | if (qc->tf.protocol == ATA_PROT_DMA || | ||
1470 | qc->tf.protocol == ATAPI_PROT_DMA) { | ||
1471 | /* check status of DMA engine */ | ||
1472 | host_stat = ap->ops->bmdma_status(ap); | ||
1473 | VPRINTK("ata%u: host_stat 0x%X\n", | ||
1474 | ap->print_id, host_stat); | ||
1475 | |||
1476 | /* if it's not our irq... */ | ||
1477 | if (!(host_stat & ATA_DMA_INTR)) | ||
1478 | goto idle_irq; | ||
1479 | |||
1480 | /* before we do anything else, clear DMA-Start bit */ | ||
1481 | ap->ops->bmdma_stop(qc); | ||
1482 | |||
1483 | if (unlikely(host_stat & ATA_DMA_ERR)) { | ||
1484 | /* error when transfering data to/from memory */ | ||
1485 | qc->err_mask |= AC_ERR_HOST_BUS; | ||
1486 | ap->hsm_task_state = HSM_ST_ERR; | ||
1487 | } | ||
1488 | } | ||
1489 | break; | ||
1490 | case HSM_ST: | ||
1491 | break; | ||
1492 | default: | ||
1493 | goto idle_irq; | ||
1494 | } | ||
1495 | |||
1496 | /* check altstatus */ | ||
1497 | status = ata_sff_altstatus(ap); | ||
1498 | if (status & ATA_BUSY) | ||
1499 | goto idle_irq; | ||
1500 | |||
1501 | /* check main status, clearing INTRQ */ | ||
1502 | status = ap->ops->sff_check_status(ap); | ||
1503 | if (unlikely(status & ATA_BUSY)) | ||
1504 | goto idle_irq; | ||
1505 | |||
1506 | /* ack bmdma irq events */ | ||
1507 | ap->ops->sff_irq_clear(ap); | ||
1508 | |||
1509 | ata_sff_hsm_move(ap, qc, status, 0); | ||
1510 | |||
1511 | if (unlikely(qc->err_mask) && (qc->tf.protocol == ATA_PROT_DMA || | ||
1512 | qc->tf.protocol == ATAPI_PROT_DMA)) | ||
1513 | ata_ehi_push_desc(ehi, "BMDMA stat 0x%x", host_stat); | ||
1514 | |||
1515 | return 1; /* irq handled */ | ||
1516 | |||
1517 | idle_irq: | ||
1518 | ap->stats.idle_irq++; | ||
1519 | |||
1520 | #ifdef ATA_IRQ_TRAP | ||
1521 | if ((ap->stats.idle_irq % 1000) == 0) { | ||
1522 | ap->ops->sff_check_status(ap); | ||
1523 | ap->ops->sff_irq_clear(ap); | ||
1524 | ata_port_printk(ap, KERN_WARNING, "irq trap\n"); | ||
1525 | return 1; | ||
1526 | } | ||
1527 | #endif | ||
1528 | return 0; /* irq not handled */ | ||
1529 | } | ||
1530 | |||
1531 | /** | ||
1532 | * ata_sff_interrupt - Default ATA host interrupt handler | ||
1533 | * @irq: irq line (unused) | ||
1534 | * @dev_instance: pointer to our ata_host information structure | ||
1535 | * | ||
1536 | * Default interrupt handler for PCI IDE devices. Calls | ||
1537 | * ata_sff_host_intr() for each port that is not disabled. | ||
1538 | * | ||
1539 | * LOCKING: | ||
1540 | * Obtains host lock during operation. | ||
1541 | * | ||
1542 | * RETURNS: | ||
1543 | * IRQ_NONE or IRQ_HANDLED. | ||
1544 | */ | ||
1545 | irqreturn_t ata_sff_interrupt(int irq, void *dev_instance) | ||
1546 | { | ||
1547 | struct ata_host *host = dev_instance; | ||
1548 | unsigned int i; | ||
1549 | unsigned int handled = 0; | ||
1550 | unsigned long flags; | ||
1551 | |||
1552 | /* TODO: make _irqsave conditional on x86 PCI IDE legacy mode */ | ||
1553 | spin_lock_irqsave(&host->lock, flags); | ||
1554 | |||
1555 | for (i = 0; i < host->n_ports; i++) { | ||
1556 | struct ata_port *ap; | ||
1557 | |||
1558 | ap = host->ports[i]; | ||
1559 | if (ap && | ||
1560 | !(ap->flags & ATA_FLAG_DISABLED)) { | ||
1561 | struct ata_queued_cmd *qc; | ||
1562 | |||
1563 | qc = ata_qc_from_tag(ap, ap->link.active_tag); | ||
1564 | if (qc && (!(qc->tf.flags & ATA_TFLAG_POLLING)) && | ||
1565 | (qc->flags & ATA_QCFLAG_ACTIVE)) | ||
1566 | handled |= ata_sff_host_intr(ap, qc); | ||
1567 | } | ||
1568 | } | ||
1569 | |||
1570 | spin_unlock_irqrestore(&host->lock, flags); | ||
1571 | |||
1572 | return IRQ_RETVAL(handled); | ||
343 | } | 1573 | } |
344 | 1574 | ||
345 | /** | 1575 | /** |
346 | * ata_bmdma_freeze - Freeze BMDMA controller port | 1576 | * ata_sff_freeze - Freeze SFF controller port |
347 | * @ap: port to freeze | 1577 | * @ap: port to freeze |
348 | * | 1578 | * |
349 | * Freeze BMDMA controller port. | 1579 | * Freeze BMDMA controller port. |
@@ -351,7 +1581,7 @@ void ata_bmdma_stop(struct ata_queued_cmd *qc) | |||
351 | * LOCKING: | 1581 | * LOCKING: |
352 | * Inherited from caller. | 1582 | * Inherited from caller. |
353 | */ | 1583 | */ |
354 | void ata_bmdma_freeze(struct ata_port *ap) | 1584 | void ata_sff_freeze(struct ata_port *ap) |
355 | { | 1585 | { |
356 | struct ata_ioports *ioaddr = &ap->ioaddr; | 1586 | struct ata_ioports *ioaddr = &ap->ioaddr; |
357 | 1587 | ||
@@ -365,51 +1595,412 @@ void ata_bmdma_freeze(struct ata_port *ap) | |||
365 | * ATA_NIEN manipulation. Also, many controllers fail to mask | 1595 | * ATA_NIEN manipulation. Also, many controllers fail to mask |
366 | * previously pending IRQ on ATA_NIEN assertion. Clear it. | 1596 | * previously pending IRQ on ATA_NIEN assertion. Clear it. |
367 | */ | 1597 | */ |
368 | ata_chk_status(ap); | 1598 | ap->ops->sff_check_status(ap); |
369 | 1599 | ||
370 | ap->ops->irq_clear(ap); | 1600 | ap->ops->sff_irq_clear(ap); |
371 | } | 1601 | } |
372 | 1602 | ||
373 | /** | 1603 | /** |
374 | * ata_bmdma_thaw - Thaw BMDMA controller port | 1604 | * ata_sff_thaw - Thaw SFF controller port |
375 | * @ap: port to thaw | 1605 | * @ap: port to thaw |
376 | * | 1606 | * |
377 | * Thaw BMDMA controller port. | 1607 | * Thaw SFF controller port. |
378 | * | 1608 | * |
379 | * LOCKING: | 1609 | * LOCKING: |
380 | * Inherited from caller. | 1610 | * Inherited from caller. |
381 | */ | 1611 | */ |
382 | void ata_bmdma_thaw(struct ata_port *ap) | 1612 | void ata_sff_thaw(struct ata_port *ap) |
383 | { | 1613 | { |
384 | /* clear & re-enable interrupts */ | 1614 | /* clear & re-enable interrupts */ |
385 | ata_chk_status(ap); | 1615 | ap->ops->sff_check_status(ap); |
386 | ap->ops->irq_clear(ap); | 1616 | ap->ops->sff_irq_clear(ap); |
387 | ap->ops->irq_on(ap); | 1617 | ap->ops->sff_irq_on(ap); |
1618 | } | ||
1619 | |||
1620 | /** | ||
1621 | * ata_sff_prereset - prepare SFF link for reset | ||
1622 | * @link: SFF link to be reset | ||
1623 | * @deadline: deadline jiffies for the operation | ||
1624 | * | ||
1625 | * SFF link @link is about to be reset. Initialize it. It first | ||
1626 | * calls ata_std_prereset() and wait for !BSY if the port is | ||
1627 | * being softreset. | ||
1628 | * | ||
1629 | * LOCKING: | ||
1630 | * Kernel thread context (may sleep) | ||
1631 | * | ||
1632 | * RETURNS: | ||
1633 | * 0 on success, -errno otherwise. | ||
1634 | */ | ||
1635 | int ata_sff_prereset(struct ata_link *link, unsigned long deadline) | ||
1636 | { | ||
1637 | struct ata_eh_context *ehc = &link->eh_context; | ||
1638 | int rc; | ||
1639 | |||
1640 | rc = ata_std_prereset(link, deadline); | ||
1641 | if (rc) | ||
1642 | return rc; | ||
1643 | |||
1644 | /* if we're about to do hardreset, nothing more to do */ | ||
1645 | if (ehc->i.action & ATA_EH_HARDRESET) | ||
1646 | return 0; | ||
1647 | |||
1648 | /* wait for !BSY if we don't know that no device is attached */ | ||
1649 | if (!ata_link_offline(link)) { | ||
1650 | rc = ata_sff_wait_ready(link, deadline); | ||
1651 | if (rc && rc != -ENODEV) { | ||
1652 | ata_link_printk(link, KERN_WARNING, "device not ready " | ||
1653 | "(errno=%d), forcing hardreset\n", rc); | ||
1654 | ehc->i.action |= ATA_EH_HARDRESET; | ||
1655 | } | ||
1656 | } | ||
1657 | |||
1658 | return 0; | ||
1659 | } | ||
1660 | |||
1661 | /** | ||
1662 | * ata_devchk - PATA device presence detection | ||
1663 | * @ap: ATA channel to examine | ||
1664 | * @device: Device to examine (starting at zero) | ||
1665 | * | ||
1666 | * This technique was originally described in | ||
1667 | * Hale Landis's ATADRVR (www.ata-atapi.com), and | ||
1668 | * later found its way into the ATA/ATAPI spec. | ||
1669 | * | ||
1670 | * Write a pattern to the ATA shadow registers, | ||
1671 | * and if a device is present, it will respond by | ||
1672 | * correctly storing and echoing back the | ||
1673 | * ATA shadow register contents. | ||
1674 | * | ||
1675 | * LOCKING: | ||
1676 | * caller. | ||
1677 | */ | ||
1678 | static unsigned int ata_devchk(struct ata_port *ap, unsigned int device) | ||
1679 | { | ||
1680 | struct ata_ioports *ioaddr = &ap->ioaddr; | ||
1681 | u8 nsect, lbal; | ||
1682 | |||
1683 | ap->ops->sff_dev_select(ap, device); | ||
1684 | |||
1685 | iowrite8(0x55, ioaddr->nsect_addr); | ||
1686 | iowrite8(0xaa, ioaddr->lbal_addr); | ||
1687 | |||
1688 | iowrite8(0xaa, ioaddr->nsect_addr); | ||
1689 | iowrite8(0x55, ioaddr->lbal_addr); | ||
1690 | |||
1691 | iowrite8(0x55, ioaddr->nsect_addr); | ||
1692 | iowrite8(0xaa, ioaddr->lbal_addr); | ||
1693 | |||
1694 | nsect = ioread8(ioaddr->nsect_addr); | ||
1695 | lbal = ioread8(ioaddr->lbal_addr); | ||
1696 | |||
1697 | if ((nsect == 0x55) && (lbal == 0xaa)) | ||
1698 | return 1; /* we found a device */ | ||
1699 | |||
1700 | return 0; /* nothing found */ | ||
1701 | } | ||
1702 | |||
1703 | /** | ||
1704 | * ata_sff_dev_classify - Parse returned ATA device signature | ||
1705 | * @dev: ATA device to classify (starting at zero) | ||
1706 | * @present: device seems present | ||
1707 | * @r_err: Value of error register on completion | ||
1708 | * | ||
1709 | * After an event -- SRST, E.D.D., or SATA COMRESET -- occurs, | ||
1710 | * an ATA/ATAPI-defined set of values is placed in the ATA | ||
1711 | * shadow registers, indicating the results of device detection | ||
1712 | * and diagnostics. | ||
1713 | * | ||
1714 | * Select the ATA device, and read the values from the ATA shadow | ||
1715 | * registers. Then parse according to the Error register value, | ||
1716 | * and the spec-defined values examined by ata_dev_classify(). | ||
1717 | * | ||
1718 | * LOCKING: | ||
1719 | * caller. | ||
1720 | * | ||
1721 | * RETURNS: | ||
1722 | * Device type - %ATA_DEV_ATA, %ATA_DEV_ATAPI or %ATA_DEV_NONE. | ||
1723 | */ | ||
1724 | unsigned int ata_sff_dev_classify(struct ata_device *dev, int present, | ||
1725 | u8 *r_err) | ||
1726 | { | ||
1727 | struct ata_port *ap = dev->link->ap; | ||
1728 | struct ata_taskfile tf; | ||
1729 | unsigned int class; | ||
1730 | u8 err; | ||
1731 | |||
1732 | ap->ops->sff_dev_select(ap, dev->devno); | ||
1733 | |||
1734 | memset(&tf, 0, sizeof(tf)); | ||
1735 | |||
1736 | ap->ops->sff_tf_read(ap, &tf); | ||
1737 | err = tf.feature; | ||
1738 | if (r_err) | ||
1739 | *r_err = err; | ||
1740 | |||
1741 | /* see if device passed diags: continue and warn later */ | ||
1742 | if (err == 0) | ||
1743 | /* diagnostic fail : do nothing _YET_ */ | ||
1744 | dev->horkage |= ATA_HORKAGE_DIAGNOSTIC; | ||
1745 | else if (err == 1) | ||
1746 | /* do nothing */ ; | ||
1747 | else if ((dev->devno == 0) && (err == 0x81)) | ||
1748 | /* do nothing */ ; | ||
1749 | else | ||
1750 | return ATA_DEV_NONE; | ||
1751 | |||
1752 | /* determine if device is ATA or ATAPI */ | ||
1753 | class = ata_dev_classify(&tf); | ||
1754 | |||
1755 | if (class == ATA_DEV_UNKNOWN) { | ||
1756 | /* If the device failed diagnostic, it's likely to | ||
1757 | * have reported incorrect device signature too. | ||
1758 | * Assume ATA device if the device seems present but | ||
1759 | * device signature is invalid with diagnostic | ||
1760 | * failure. | ||
1761 | */ | ||
1762 | if (present && (dev->horkage & ATA_HORKAGE_DIAGNOSTIC)) | ||
1763 | class = ATA_DEV_ATA; | ||
1764 | else | ||
1765 | class = ATA_DEV_NONE; | ||
1766 | } else if ((class == ATA_DEV_ATA) && | ||
1767 | (ap->ops->sff_check_status(ap) == 0)) | ||
1768 | class = ATA_DEV_NONE; | ||
1769 | |||
1770 | return class; | ||
1771 | } | ||
1772 | |||
1773 | /** | ||
1774 | * ata_sff_wait_after_reset - wait for devices to become ready after reset | ||
1775 | * @link: SFF link which is just reset | ||
1776 | * @devmask: mask of present devices | ||
1777 | * @deadline: deadline jiffies for the operation | ||
1778 | * | ||
1779 | * Wait devices attached to SFF @link to become ready after | ||
1780 | * reset. It contains preceding 150ms wait to avoid accessing TF | ||
1781 | * status register too early. | ||
1782 | * | ||
1783 | * LOCKING: | ||
1784 | * Kernel thread context (may sleep). | ||
1785 | * | ||
1786 | * RETURNS: | ||
1787 | * 0 on success, -ENODEV if some or all of devices in @devmask | ||
1788 | * don't seem to exist. -errno on other errors. | ||
1789 | */ | ||
1790 | int ata_sff_wait_after_reset(struct ata_link *link, unsigned int devmask, | ||
1791 | unsigned long deadline) | ||
1792 | { | ||
1793 | struct ata_port *ap = link->ap; | ||
1794 | struct ata_ioports *ioaddr = &ap->ioaddr; | ||
1795 | unsigned int dev0 = devmask & (1 << 0); | ||
1796 | unsigned int dev1 = devmask & (1 << 1); | ||
1797 | int rc, ret = 0; | ||
1798 | |||
1799 | msleep(ATA_WAIT_AFTER_RESET_MSECS); | ||
1800 | |||
1801 | /* always check readiness of the master device */ | ||
1802 | rc = ata_sff_wait_ready(link, deadline); | ||
1803 | /* -ENODEV means the odd clown forgot the D7 pulldown resistor | ||
1804 | * and TF status is 0xff, bail out on it too. | ||
1805 | */ | ||
1806 | if (rc) | ||
1807 | return rc; | ||
1808 | |||
1809 | /* if device 1 was found in ata_devchk, wait for register | ||
1810 | * access briefly, then wait for BSY to clear. | ||
1811 | */ | ||
1812 | if (dev1) { | ||
1813 | int i; | ||
1814 | |||
1815 | ap->ops->sff_dev_select(ap, 1); | ||
1816 | |||
1817 | /* Wait for register access. Some ATAPI devices fail | ||
1818 | * to set nsect/lbal after reset, so don't waste too | ||
1819 | * much time on it. We're gonna wait for !BSY anyway. | ||
1820 | */ | ||
1821 | for (i = 0; i < 2; i++) { | ||
1822 | u8 nsect, lbal; | ||
1823 | |||
1824 | nsect = ioread8(ioaddr->nsect_addr); | ||
1825 | lbal = ioread8(ioaddr->lbal_addr); | ||
1826 | if ((nsect == 1) && (lbal == 1)) | ||
1827 | break; | ||
1828 | msleep(50); /* give drive a breather */ | ||
1829 | } | ||
1830 | |||
1831 | rc = ata_sff_wait_ready(link, deadline); | ||
1832 | if (rc) { | ||
1833 | if (rc != -ENODEV) | ||
1834 | return rc; | ||
1835 | ret = rc; | ||
1836 | } | ||
1837 | } | ||
1838 | |||
1839 | /* is all this really necessary? */ | ||
1840 | ap->ops->sff_dev_select(ap, 0); | ||
1841 | if (dev1) | ||
1842 | ap->ops->sff_dev_select(ap, 1); | ||
1843 | if (dev0) | ||
1844 | ap->ops->sff_dev_select(ap, 0); | ||
1845 | |||
1846 | return ret; | ||
1847 | } | ||
1848 | |||
1849 | static int ata_bus_softreset(struct ata_port *ap, unsigned int devmask, | ||
1850 | unsigned long deadline) | ||
1851 | { | ||
1852 | struct ata_ioports *ioaddr = &ap->ioaddr; | ||
1853 | |||
1854 | DPRINTK("ata%u: bus reset via SRST\n", ap->print_id); | ||
1855 | |||
1856 | /* software reset. causes dev0 to be selected */ | ||
1857 | iowrite8(ap->ctl, ioaddr->ctl_addr); | ||
1858 | udelay(20); /* FIXME: flush */ | ||
1859 | iowrite8(ap->ctl | ATA_SRST, ioaddr->ctl_addr); | ||
1860 | udelay(20); /* FIXME: flush */ | ||
1861 | iowrite8(ap->ctl, ioaddr->ctl_addr); | ||
1862 | |||
1863 | /* wait the port to become ready */ | ||
1864 | return ata_sff_wait_after_reset(&ap->link, devmask, deadline); | ||
1865 | } | ||
1866 | |||
1867 | /** | ||
1868 | * ata_sff_softreset - reset host port via ATA SRST | ||
1869 | * @link: ATA link to reset | ||
1870 | * @classes: resulting classes of attached devices | ||
1871 | * @deadline: deadline jiffies for the operation | ||
1872 | * | ||
1873 | * Reset host port using ATA SRST. | ||
1874 | * | ||
1875 | * LOCKING: | ||
1876 | * Kernel thread context (may sleep) | ||
1877 | * | ||
1878 | * RETURNS: | ||
1879 | * 0 on success, -errno otherwise. | ||
1880 | */ | ||
1881 | int ata_sff_softreset(struct ata_link *link, unsigned int *classes, | ||
1882 | unsigned long deadline) | ||
1883 | { | ||
1884 | struct ata_port *ap = link->ap; | ||
1885 | unsigned int slave_possible = ap->flags & ATA_FLAG_SLAVE_POSS; | ||
1886 | unsigned int devmask = 0; | ||
1887 | int rc; | ||
1888 | u8 err; | ||
1889 | |||
1890 | DPRINTK("ENTER\n"); | ||
1891 | |||
1892 | /* determine if device 0/1 are present */ | ||
1893 | if (ata_devchk(ap, 0)) | ||
1894 | devmask |= (1 << 0); | ||
1895 | if (slave_possible && ata_devchk(ap, 1)) | ||
1896 | devmask |= (1 << 1); | ||
1897 | |||
1898 | /* select device 0 again */ | ||
1899 | ap->ops->sff_dev_select(ap, 0); | ||
1900 | |||
1901 | /* issue bus reset */ | ||
1902 | DPRINTK("about to softreset, devmask=%x\n", devmask); | ||
1903 | rc = ata_bus_softreset(ap, devmask, deadline); | ||
1904 | /* if link is occupied, -ENODEV too is an error */ | ||
1905 | if (rc && (rc != -ENODEV || sata_scr_valid(link))) { | ||
1906 | ata_link_printk(link, KERN_ERR, "SRST failed (errno=%d)\n", rc); | ||
1907 | return rc; | ||
1908 | } | ||
1909 | |||
1910 | /* determine by signature whether we have ATA or ATAPI devices */ | ||
1911 | classes[0] = ata_sff_dev_classify(&link->device[0], | ||
1912 | devmask & (1 << 0), &err); | ||
1913 | if (slave_possible && err != 0x81) | ||
1914 | classes[1] = ata_sff_dev_classify(&link->device[1], | ||
1915 | devmask & (1 << 1), &err); | ||
1916 | |||
1917 | DPRINTK("EXIT, classes[0]=%u [1]=%u\n", classes[0], classes[1]); | ||
1918 | return 0; | ||
1919 | } | ||
1920 | |||
1921 | /** | ||
1922 | * sata_sff_hardreset - reset host port via SATA phy reset | ||
1923 | * @link: link to reset | ||
1924 | * @class: resulting class of attached device | ||
1925 | * @deadline: deadline jiffies for the operation | ||
1926 | * | ||
1927 | * SATA phy-reset host port using DET bits of SControl register, | ||
1928 | * wait for !BSY and classify the attached device. | ||
1929 | * | ||
1930 | * LOCKING: | ||
1931 | * Kernel thread context (may sleep) | ||
1932 | * | ||
1933 | * RETURNS: | ||
1934 | * 0 on success, -errno otherwise. | ||
1935 | */ | ||
1936 | int sata_sff_hardreset(struct ata_link *link, unsigned int *class, | ||
1937 | unsigned long deadline) | ||
1938 | { | ||
1939 | struct ata_eh_context *ehc = &link->eh_context; | ||
1940 | const unsigned long *timing = sata_ehc_deb_timing(ehc); | ||
1941 | bool online; | ||
1942 | int rc; | ||
1943 | |||
1944 | rc = sata_link_hardreset(link, timing, deadline, &online, | ||
1945 | ata_sff_check_ready); | ||
1946 | if (online) | ||
1947 | *class = ata_sff_dev_classify(link->device, 1, NULL); | ||
1948 | |||
1949 | DPRINTK("EXIT, class=%u\n", *class); | ||
1950 | return rc; | ||
388 | } | 1951 | } |
389 | 1952 | ||
390 | /** | 1953 | /** |
391 | * ata_bmdma_drive_eh - Perform EH with given methods for BMDMA controller | 1954 | * ata_sff_postreset - SFF postreset callback |
1955 | * @link: the target SFF ata_link | ||
1956 | * @classes: classes of attached devices | ||
1957 | * | ||
1958 | * This function is invoked after a successful reset. It first | ||
1959 | * calls ata_std_postreset() and performs SFF specific postreset | ||
1960 | * processing. | ||
1961 | * | ||
1962 | * LOCKING: | ||
1963 | * Kernel thread context (may sleep) | ||
1964 | */ | ||
1965 | void ata_sff_postreset(struct ata_link *link, unsigned int *classes) | ||
1966 | { | ||
1967 | struct ata_port *ap = link->ap; | ||
1968 | |||
1969 | ata_std_postreset(link, classes); | ||
1970 | |||
1971 | /* is double-select really necessary? */ | ||
1972 | if (classes[0] != ATA_DEV_NONE) | ||
1973 | ap->ops->sff_dev_select(ap, 1); | ||
1974 | if (classes[1] != ATA_DEV_NONE) | ||
1975 | ap->ops->sff_dev_select(ap, 0); | ||
1976 | |||
1977 | /* bail out if no device is present */ | ||
1978 | if (classes[0] == ATA_DEV_NONE && classes[1] == ATA_DEV_NONE) { | ||
1979 | DPRINTK("EXIT, no device\n"); | ||
1980 | return; | ||
1981 | } | ||
1982 | |||
1983 | /* set up device control */ | ||
1984 | if (ap->ioaddr.ctl_addr) | ||
1985 | iowrite8(ap->ctl, ap->ioaddr.ctl_addr); | ||
1986 | } | ||
1987 | |||
1988 | /** | ||
1989 | * ata_sff_error_handler - Stock error handler for BMDMA controller | ||
392 | * @ap: port to handle error for | 1990 | * @ap: port to handle error for |
393 | * @prereset: prereset method (can be NULL) | ||
394 | * @softreset: softreset method (can be NULL) | ||
395 | * @hardreset: hardreset method (can be NULL) | ||
396 | * @postreset: postreset method (can be NULL) | ||
397 | * | 1991 | * |
398 | * Handle error for ATA BMDMA controller. It can handle both | 1992 | * Stock error handler for SFF controller. It can handle both |
399 | * PATA and SATA controllers. Many controllers should be able to | 1993 | * PATA and SATA controllers. Many controllers should be able to |
400 | * use this EH as-is or with some added handling before and | 1994 | * use this EH as-is or with some added handling before and |
401 | * after. | 1995 | * after. |
402 | * | 1996 | * |
403 | * This function is intended to be used for constructing | ||
404 | * ->error_handler callback by low level drivers. | ||
405 | * | ||
406 | * LOCKING: | 1997 | * LOCKING: |
407 | * Kernel thread context (may sleep) | 1998 | * Kernel thread context (may sleep) |
408 | */ | 1999 | */ |
409 | void ata_bmdma_drive_eh(struct ata_port *ap, ata_prereset_fn_t prereset, | 2000 | void ata_sff_error_handler(struct ata_port *ap) |
410 | ata_reset_fn_t softreset, ata_reset_fn_t hardreset, | ||
411 | ata_postreset_fn_t postreset) | ||
412 | { | 2001 | { |
2002 | ata_reset_fn_t softreset = ap->ops->softreset; | ||
2003 | ata_reset_fn_t hardreset = ap->ops->hardreset; | ||
413 | struct ata_queued_cmd *qc; | 2004 | struct ata_queued_cmd *qc; |
414 | unsigned long flags; | 2005 | unsigned long flags; |
415 | int thaw = 0; | 2006 | int thaw = 0; |
@@ -423,7 +2014,8 @@ void ata_bmdma_drive_eh(struct ata_port *ap, ata_prereset_fn_t prereset, | |||
423 | 2014 | ||
424 | ap->hsm_task_state = HSM_ST_IDLE; | 2015 | ap->hsm_task_state = HSM_ST_IDLE; |
425 | 2016 | ||
426 | if (qc && (qc->tf.protocol == ATA_PROT_DMA || | 2017 | if (ap->ioaddr.bmdma_addr && |
2018 | qc && (qc->tf.protocol == ATA_PROT_DMA || | ||
427 | qc->tf.protocol == ATAPI_PROT_DMA)) { | 2019 | qc->tf.protocol == ATAPI_PROT_DMA)) { |
428 | u8 host_stat; | 2020 | u8 host_stat; |
429 | 2021 | ||
@@ -442,9 +2034,9 @@ void ata_bmdma_drive_eh(struct ata_port *ap, ata_prereset_fn_t prereset, | |||
442 | ap->ops->bmdma_stop(qc); | 2034 | ap->ops->bmdma_stop(qc); |
443 | } | 2035 | } |
444 | 2036 | ||
445 | ata_altstatus(ap); | 2037 | ata_sff_altstatus(ap); |
446 | ata_chk_status(ap); | 2038 | ap->ops->sff_check_status(ap); |
447 | ap->ops->irq_clear(ap); | 2039 | ap->ops->sff_irq_clear(ap); |
448 | 2040 | ||
449 | spin_unlock_irqrestore(ap->lock, flags); | 2041 | spin_unlock_irqrestore(ap->lock, flags); |
450 | 2042 | ||
@@ -452,40 +2044,27 @@ void ata_bmdma_drive_eh(struct ata_port *ap, ata_prereset_fn_t prereset, | |||
452 | ata_eh_thaw_port(ap); | 2044 | ata_eh_thaw_port(ap); |
453 | 2045 | ||
454 | /* PIO and DMA engines have been stopped, perform recovery */ | 2046 | /* PIO and DMA engines have been stopped, perform recovery */ |
455 | ata_do_eh(ap, prereset, softreset, hardreset, postreset); | ||
456 | } | ||
457 | |||
458 | /** | ||
459 | * ata_bmdma_error_handler - Stock error handler for BMDMA controller | ||
460 | * @ap: port to handle error for | ||
461 | * | ||
462 | * Stock error handler for BMDMA controller. | ||
463 | * | ||
464 | * LOCKING: | ||
465 | * Kernel thread context (may sleep) | ||
466 | */ | ||
467 | void ata_bmdma_error_handler(struct ata_port *ap) | ||
468 | { | ||
469 | ata_reset_fn_t softreset = NULL, hardreset = NULL; | ||
470 | 2047 | ||
471 | if (ap->ioaddr.ctl_addr) | 2048 | /* Ignore ata_sff_softreset if ctl isn't accessible and |
472 | softreset = ata_std_softreset; | 2049 | * built-in hardresets if SCR access isn't available. |
473 | if (sata_scr_valid(&ap->link)) | 2050 | */ |
474 | hardreset = sata_std_hardreset; | 2051 | if (softreset == ata_sff_softreset && !ap->ioaddr.ctl_addr) |
2052 | softreset = NULL; | ||
2053 | if (ata_is_builtin_hardreset(hardreset) && !sata_scr_valid(&ap->link)) | ||
2054 | hardreset = NULL; | ||
475 | 2055 | ||
476 | ata_bmdma_drive_eh(ap, ata_std_prereset, softreset, hardreset, | 2056 | ata_do_eh(ap, ap->ops->prereset, softreset, hardreset, |
477 | ata_std_postreset); | 2057 | ap->ops->postreset); |
478 | } | 2058 | } |
479 | 2059 | ||
480 | /** | 2060 | /** |
481 | * ata_bmdma_post_internal_cmd - Stock post_internal_cmd for | 2061 | * ata_sff_post_internal_cmd - Stock post_internal_cmd for SFF controller |
482 | * BMDMA controller | ||
483 | * @qc: internal command to clean up | 2062 | * @qc: internal command to clean up |
484 | * | 2063 | * |
485 | * LOCKING: | 2064 | * LOCKING: |
486 | * Kernel thread context (may sleep) | 2065 | * Kernel thread context (may sleep) |
487 | */ | 2066 | */ |
488 | void ata_bmdma_post_internal_cmd(struct ata_queued_cmd *qc) | 2067 | void ata_sff_post_internal_cmd(struct ata_queued_cmd *qc) |
489 | { | 2068 | { |
490 | if (qc->ap->ioaddr.bmdma_addr) | 2069 | if (qc->ap->ioaddr.bmdma_addr) |
491 | ata_bmdma_stop(qc); | 2070 | ata_bmdma_stop(qc); |
@@ -504,7 +2083,6 @@ void ata_bmdma_post_internal_cmd(struct ata_queued_cmd *qc) | |||
504 | * LOCKING: | 2083 | * LOCKING: |
505 | * Inherited from caller. | 2084 | * Inherited from caller. |
506 | */ | 2085 | */ |
507 | |||
508 | int ata_sff_port_start(struct ata_port *ap) | 2086 | int ata_sff_port_start(struct ata_port *ap) |
509 | { | 2087 | { |
510 | if (ap->ioaddr.bmdma_addr) | 2088 | if (ap->ioaddr.bmdma_addr) |
@@ -512,24 +2090,262 @@ int ata_sff_port_start(struct ata_port *ap) | |||
512 | return 0; | 2090 | return 0; |
513 | } | 2091 | } |
514 | 2092 | ||
515 | #ifdef CONFIG_PCI | 2093 | /** |
2094 | * ata_sff_std_ports - initialize ioaddr with standard port offsets. | ||
2095 | * @ioaddr: IO address structure to be initialized | ||
2096 | * | ||
2097 | * Utility function which initializes data_addr, error_addr, | ||
2098 | * feature_addr, nsect_addr, lbal_addr, lbam_addr, lbah_addr, | ||
2099 | * device_addr, status_addr, and command_addr to standard offsets | ||
2100 | * relative to cmd_addr. | ||
2101 | * | ||
2102 | * Does not set ctl_addr, altstatus_addr, bmdma_addr, or scr_addr. | ||
2103 | */ | ||
2104 | void ata_sff_std_ports(struct ata_ioports *ioaddr) | ||
2105 | { | ||
2106 | ioaddr->data_addr = ioaddr->cmd_addr + ATA_REG_DATA; | ||
2107 | ioaddr->error_addr = ioaddr->cmd_addr + ATA_REG_ERR; | ||
2108 | ioaddr->feature_addr = ioaddr->cmd_addr + ATA_REG_FEATURE; | ||
2109 | ioaddr->nsect_addr = ioaddr->cmd_addr + ATA_REG_NSECT; | ||
2110 | ioaddr->lbal_addr = ioaddr->cmd_addr + ATA_REG_LBAL; | ||
2111 | ioaddr->lbam_addr = ioaddr->cmd_addr + ATA_REG_LBAM; | ||
2112 | ioaddr->lbah_addr = ioaddr->cmd_addr + ATA_REG_LBAH; | ||
2113 | ioaddr->device_addr = ioaddr->cmd_addr + ATA_REG_DEVICE; | ||
2114 | ioaddr->status_addr = ioaddr->cmd_addr + ATA_REG_STATUS; | ||
2115 | ioaddr->command_addr = ioaddr->cmd_addr + ATA_REG_CMD; | ||
2116 | } | ||
516 | 2117 | ||
517 | static int ata_resources_present(struct pci_dev *pdev, int port) | 2118 | unsigned long ata_bmdma_mode_filter(struct ata_device *adev, |
2119 | unsigned long xfer_mask) | ||
518 | { | 2120 | { |
519 | int i; | 2121 | /* Filter out DMA modes if the device has been configured by |
2122 | the BIOS as PIO only */ | ||
520 | 2123 | ||
521 | /* Check the PCI resources for this channel are enabled */ | 2124 | if (adev->link->ap->ioaddr.bmdma_addr == NULL) |
522 | port = port * 2; | 2125 | xfer_mask &= ~(ATA_MASK_MWDMA | ATA_MASK_UDMA); |
523 | for (i = 0; i < 2; i ++) { | 2126 | return xfer_mask; |
524 | if (pci_resource_start(pdev, port + i) == 0 || | 2127 | } |
525 | pci_resource_len(pdev, port + i) == 0) | 2128 | |
526 | return 0; | 2129 | /** |
2130 | * ata_bmdma_setup - Set up PCI IDE BMDMA transaction | ||
2131 | * @qc: Info associated with this ATA transaction. | ||
2132 | * | ||
2133 | * LOCKING: | ||
2134 | * spin_lock_irqsave(host lock) | ||
2135 | */ | ||
2136 | void ata_bmdma_setup(struct ata_queued_cmd *qc) | ||
2137 | { | ||
2138 | struct ata_port *ap = qc->ap; | ||
2139 | unsigned int rw = (qc->tf.flags & ATA_TFLAG_WRITE); | ||
2140 | u8 dmactl; | ||
2141 | |||
2142 | /* load PRD table addr. */ | ||
2143 | mb(); /* make sure PRD table writes are visible to controller */ | ||
2144 | iowrite32(ap->prd_dma, ap->ioaddr.bmdma_addr + ATA_DMA_TABLE_OFS); | ||
2145 | |||
2146 | /* specify data direction, triple-check start bit is clear */ | ||
2147 | dmactl = ioread8(ap->ioaddr.bmdma_addr + ATA_DMA_CMD); | ||
2148 | dmactl &= ~(ATA_DMA_WR | ATA_DMA_START); | ||
2149 | if (!rw) | ||
2150 | dmactl |= ATA_DMA_WR; | ||
2151 | iowrite8(dmactl, ap->ioaddr.bmdma_addr + ATA_DMA_CMD); | ||
2152 | |||
2153 | /* issue r/w command */ | ||
2154 | ap->ops->sff_exec_command(ap, &qc->tf); | ||
2155 | } | ||
2156 | |||
2157 | /** | ||
2158 | * ata_bmdma_start - Start a PCI IDE BMDMA transaction | ||
2159 | * @qc: Info associated with this ATA transaction. | ||
2160 | * | ||
2161 | * LOCKING: | ||
2162 | * spin_lock_irqsave(host lock) | ||
2163 | */ | ||
2164 | void ata_bmdma_start(struct ata_queued_cmd *qc) | ||
2165 | { | ||
2166 | struct ata_port *ap = qc->ap; | ||
2167 | u8 dmactl; | ||
2168 | |||
2169 | /* start host DMA transaction */ | ||
2170 | dmactl = ioread8(ap->ioaddr.bmdma_addr + ATA_DMA_CMD); | ||
2171 | iowrite8(dmactl | ATA_DMA_START, ap->ioaddr.bmdma_addr + ATA_DMA_CMD); | ||
2172 | |||
2173 | /* Strictly, one may wish to issue an ioread8() here, to | ||
2174 | * flush the mmio write. However, control also passes | ||
2175 | * to the hardware at this point, and it will interrupt | ||
2176 | * us when we are to resume control. So, in effect, | ||
2177 | * we don't care when the mmio write flushes. | ||
2178 | * Further, a read of the DMA status register _immediately_ | ||
2179 | * following the write may not be what certain flaky hardware | ||
2180 | * is expected, so I think it is best to not add a readb() | ||
2181 | * without first all the MMIO ATA cards/mobos. | ||
2182 | * Or maybe I'm just being paranoid. | ||
2183 | * | ||
2184 | * FIXME: The posting of this write means I/O starts are | ||
2185 | * unneccessarily delayed for MMIO | ||
2186 | */ | ||
2187 | } | ||
2188 | |||
2189 | /** | ||
2190 | * ata_bmdma_stop - Stop PCI IDE BMDMA transfer | ||
2191 | * @qc: Command we are ending DMA for | ||
2192 | * | ||
2193 | * Clears the ATA_DMA_START flag in the dma control register | ||
2194 | * | ||
2195 | * May be used as the bmdma_stop() entry in ata_port_operations. | ||
2196 | * | ||
2197 | * LOCKING: | ||
2198 | * spin_lock_irqsave(host lock) | ||
2199 | */ | ||
2200 | void ata_bmdma_stop(struct ata_queued_cmd *qc) | ||
2201 | { | ||
2202 | struct ata_port *ap = qc->ap; | ||
2203 | void __iomem *mmio = ap->ioaddr.bmdma_addr; | ||
2204 | |||
2205 | /* clear start/stop bit */ | ||
2206 | iowrite8(ioread8(mmio + ATA_DMA_CMD) & ~ATA_DMA_START, | ||
2207 | mmio + ATA_DMA_CMD); | ||
2208 | |||
2209 | /* one-PIO-cycle guaranteed wait, per spec, for HDMA1:0 transition */ | ||
2210 | ata_sff_altstatus(ap); /* dummy read */ | ||
2211 | } | ||
2212 | |||
2213 | /** | ||
2214 | * ata_bmdma_status - Read PCI IDE BMDMA status | ||
2215 | * @ap: Port associated with this ATA transaction. | ||
2216 | * | ||
2217 | * Read and return BMDMA status register. | ||
2218 | * | ||
2219 | * May be used as the bmdma_status() entry in ata_port_operations. | ||
2220 | * | ||
2221 | * LOCKING: | ||
2222 | * spin_lock_irqsave(host lock) | ||
2223 | */ | ||
2224 | u8 ata_bmdma_status(struct ata_port *ap) | ||
2225 | { | ||
2226 | return ioread8(ap->ioaddr.bmdma_addr + ATA_DMA_STATUS); | ||
2227 | } | ||
2228 | |||
2229 | /** | ||
2230 | * ata_bus_reset - reset host port and associated ATA channel | ||
2231 | * @ap: port to reset | ||
2232 | * | ||
2233 | * This is typically the first time we actually start issuing | ||
2234 | * commands to the ATA channel. We wait for BSY to clear, then | ||
2235 | * issue EXECUTE DEVICE DIAGNOSTIC command, polling for its | ||
2236 | * result. Determine what devices, if any, are on the channel | ||
2237 | * by looking at the device 0/1 error register. Look at the signature | ||
2238 | * stored in each device's taskfile registers, to determine if | ||
2239 | * the device is ATA or ATAPI. | ||
2240 | * | ||
2241 | * LOCKING: | ||
2242 | * PCI/etc. bus probe sem. | ||
2243 | * Obtains host lock. | ||
2244 | * | ||
2245 | * SIDE EFFECTS: | ||
2246 | * Sets ATA_FLAG_DISABLED if bus reset fails. | ||
2247 | * | ||
2248 | * DEPRECATED: | ||
2249 | * This function is only for drivers which still use old EH and | ||
2250 | * will be removed soon. | ||
2251 | */ | ||
2252 | void ata_bus_reset(struct ata_port *ap) | ||
2253 | { | ||
2254 | struct ata_device *device = ap->link.device; | ||
2255 | struct ata_ioports *ioaddr = &ap->ioaddr; | ||
2256 | unsigned int slave_possible = ap->flags & ATA_FLAG_SLAVE_POSS; | ||
2257 | u8 err; | ||
2258 | unsigned int dev0, dev1 = 0, devmask = 0; | ||
2259 | int rc; | ||
2260 | |||
2261 | DPRINTK("ENTER, host %u, port %u\n", ap->print_id, ap->port_no); | ||
2262 | |||
2263 | /* determine if device 0/1 are present */ | ||
2264 | if (ap->flags & ATA_FLAG_SATA_RESET) | ||
2265 | dev0 = 1; | ||
2266 | else { | ||
2267 | dev0 = ata_devchk(ap, 0); | ||
2268 | if (slave_possible) | ||
2269 | dev1 = ata_devchk(ap, 1); | ||
527 | } | 2270 | } |
528 | return 1; | 2271 | |
2272 | if (dev0) | ||
2273 | devmask |= (1 << 0); | ||
2274 | if (dev1) | ||
2275 | devmask |= (1 << 1); | ||
2276 | |||
2277 | /* select device 0 again */ | ||
2278 | ap->ops->sff_dev_select(ap, 0); | ||
2279 | |||
2280 | /* issue bus reset */ | ||
2281 | if (ap->flags & ATA_FLAG_SRST) { | ||
2282 | rc = ata_bus_softreset(ap, devmask, jiffies + 40 * HZ); | ||
2283 | if (rc && rc != -ENODEV) | ||
2284 | goto err_out; | ||
2285 | } | ||
2286 | |||
2287 | /* | ||
2288 | * determine by signature whether we have ATA or ATAPI devices | ||
2289 | */ | ||
2290 | device[0].class = ata_sff_dev_classify(&device[0], dev0, &err); | ||
2291 | if ((slave_possible) && (err != 0x81)) | ||
2292 | device[1].class = ata_sff_dev_classify(&device[1], dev1, &err); | ||
2293 | |||
2294 | /* is double-select really necessary? */ | ||
2295 | if (device[1].class != ATA_DEV_NONE) | ||
2296 | ap->ops->sff_dev_select(ap, 1); | ||
2297 | if (device[0].class != ATA_DEV_NONE) | ||
2298 | ap->ops->sff_dev_select(ap, 0); | ||
2299 | |||
2300 | /* if no devices were detected, disable this port */ | ||
2301 | if ((device[0].class == ATA_DEV_NONE) && | ||
2302 | (device[1].class == ATA_DEV_NONE)) | ||
2303 | goto err_out; | ||
2304 | |||
2305 | if (ap->flags & (ATA_FLAG_SATA_RESET | ATA_FLAG_SRST)) { | ||
2306 | /* set up device control for ATA_FLAG_SATA_RESET */ | ||
2307 | iowrite8(ap->ctl, ioaddr->ctl_addr); | ||
2308 | } | ||
2309 | |||
2310 | DPRINTK("EXIT\n"); | ||
2311 | return; | ||
2312 | |||
2313 | err_out: | ||
2314 | ata_port_printk(ap, KERN_ERR, "disabling port\n"); | ||
2315 | ata_port_disable(ap); | ||
2316 | |||
2317 | DPRINTK("EXIT\n"); | ||
529 | } | 2318 | } |
530 | 2319 | ||
2320 | #ifdef CONFIG_PCI | ||
2321 | |||
531 | /** | 2322 | /** |
532 | * ata_pci_init_bmdma - acquire PCI BMDMA resources and init ATA host | 2323 | * ata_pci_bmdma_clear_simplex - attempt to kick device out of simplex |
2324 | * @pdev: PCI device | ||
2325 | * | ||
2326 | * Some PCI ATA devices report simplex mode but in fact can be told to | ||
2327 | * enter non simplex mode. This implements the necessary logic to | ||
2328 | * perform the task on such devices. Calling it on other devices will | ||
2329 | * have -undefined- behaviour. | ||
2330 | */ | ||
2331 | int ata_pci_bmdma_clear_simplex(struct pci_dev *pdev) | ||
2332 | { | ||
2333 | unsigned long bmdma = pci_resource_start(pdev, 4); | ||
2334 | u8 simplex; | ||
2335 | |||
2336 | if (bmdma == 0) | ||
2337 | return -ENOENT; | ||
2338 | |||
2339 | simplex = inb(bmdma + 0x02); | ||
2340 | outb(simplex & 0x60, bmdma + 0x02); | ||
2341 | simplex = inb(bmdma + 0x02); | ||
2342 | if (simplex & 0x80) | ||
2343 | return -EOPNOTSUPP; | ||
2344 | return 0; | ||
2345 | } | ||
2346 | |||
2347 | /** | ||
2348 | * ata_pci_bmdma_init - acquire PCI BMDMA resources and init ATA host | ||
533 | * @host: target ATA host | 2349 | * @host: target ATA host |
534 | * | 2350 | * |
535 | * Acquire PCI BMDMA resources and initialize @host accordingly. | 2351 | * Acquire PCI BMDMA resources and initialize @host accordingly. |
@@ -540,7 +2356,7 @@ static int ata_resources_present(struct pci_dev *pdev, int port) | |||
540 | * RETURNS: | 2356 | * RETURNS: |
541 | * 0 on success, -errno otherwise. | 2357 | * 0 on success, -errno otherwise. |
542 | */ | 2358 | */ |
543 | int ata_pci_init_bmdma(struct ata_host *host) | 2359 | int ata_pci_bmdma_init(struct ata_host *host) |
544 | { | 2360 | { |
545 | struct device *gdev = host->dev; | 2361 | struct device *gdev = host->dev; |
546 | struct pci_dev *pdev = to_pci_dev(gdev); | 2362 | struct pci_dev *pdev = to_pci_dev(gdev); |
@@ -585,8 +2401,22 @@ int ata_pci_init_bmdma(struct ata_host *host) | |||
585 | return 0; | 2401 | return 0; |
586 | } | 2402 | } |
587 | 2403 | ||
2404 | static int ata_resources_present(struct pci_dev *pdev, int port) | ||
2405 | { | ||
2406 | int i; | ||
2407 | |||
2408 | /* Check the PCI resources for this channel are enabled */ | ||
2409 | port = port * 2; | ||
2410 | for (i = 0; i < 2; i ++) { | ||
2411 | if (pci_resource_start(pdev, port + i) == 0 || | ||
2412 | pci_resource_len(pdev, port + i) == 0) | ||
2413 | return 0; | ||
2414 | } | ||
2415 | return 1; | ||
2416 | } | ||
2417 | |||
588 | /** | 2418 | /** |
589 | * ata_pci_init_sff_host - acquire native PCI ATA resources and init host | 2419 | * ata_pci_sff_init_host - acquire native PCI ATA resources and init host |
590 | * @host: target ATA host | 2420 | * @host: target ATA host |
591 | * | 2421 | * |
592 | * Acquire native PCI ATA resources for @host and initialize the | 2422 | * Acquire native PCI ATA resources for @host and initialize the |
@@ -604,7 +2434,7 @@ int ata_pci_init_bmdma(struct ata_host *host) | |||
604 | * 0 if at least one port is initialized, -ENODEV if no port is | 2434 | * 0 if at least one port is initialized, -ENODEV if no port is |
605 | * available. | 2435 | * available. |
606 | */ | 2436 | */ |
607 | int ata_pci_init_sff_host(struct ata_host *host) | 2437 | int ata_pci_sff_init_host(struct ata_host *host) |
608 | { | 2438 | { |
609 | struct device *gdev = host->dev; | 2439 | struct device *gdev = host->dev; |
610 | struct pci_dev *pdev = to_pci_dev(gdev); | 2440 | struct pci_dev *pdev = to_pci_dev(gdev); |
@@ -646,7 +2476,7 @@ int ata_pci_init_sff_host(struct ata_host *host) | |||
646 | ap->ioaddr.altstatus_addr = | 2476 | ap->ioaddr.altstatus_addr = |
647 | ap->ioaddr.ctl_addr = (void __iomem *) | 2477 | ap->ioaddr.ctl_addr = (void __iomem *) |
648 | ((unsigned long)iomap[base + 1] | ATA_PCI_CTL_OFS); | 2478 | ((unsigned long)iomap[base + 1] | ATA_PCI_CTL_OFS); |
649 | ata_std_ports(&ap->ioaddr); | 2479 | ata_sff_std_ports(&ap->ioaddr); |
650 | 2480 | ||
651 | ata_port_desc(ap, "cmd 0x%llx ctl 0x%llx", | 2481 | ata_port_desc(ap, "cmd 0x%llx ctl 0x%llx", |
652 | (unsigned long long)pci_resource_start(pdev, base), | 2482 | (unsigned long long)pci_resource_start(pdev, base), |
@@ -664,7 +2494,7 @@ int ata_pci_init_sff_host(struct ata_host *host) | |||
664 | } | 2494 | } |
665 | 2495 | ||
666 | /** | 2496 | /** |
667 | * ata_pci_prepare_sff_host - helper to prepare native PCI ATA host | 2497 | * ata_pci_sff_prepare_host - helper to prepare native PCI ATA host |
668 | * @pdev: target PCI device | 2498 | * @pdev: target PCI device |
669 | * @ppi: array of port_info, must be enough for two ports | 2499 | * @ppi: array of port_info, must be enough for two ports |
670 | * @r_host: out argument for the initialized ATA host | 2500 | * @r_host: out argument for the initialized ATA host |
@@ -678,7 +2508,7 @@ int ata_pci_init_sff_host(struct ata_host *host) | |||
678 | * RETURNS: | 2508 | * RETURNS: |
679 | * 0 on success, -errno otherwise. | 2509 | * 0 on success, -errno otherwise. |
680 | */ | 2510 | */ |
681 | int ata_pci_prepare_sff_host(struct pci_dev *pdev, | 2511 | int ata_pci_sff_prepare_host(struct pci_dev *pdev, |
682 | const struct ata_port_info * const * ppi, | 2512 | const struct ata_port_info * const * ppi, |
683 | struct ata_host **r_host) | 2513 | struct ata_host **r_host) |
684 | { | 2514 | { |
@@ -696,12 +2526,12 @@ int ata_pci_prepare_sff_host(struct pci_dev *pdev, | |||
696 | goto err_out; | 2526 | goto err_out; |
697 | } | 2527 | } |
698 | 2528 | ||
699 | rc = ata_pci_init_sff_host(host); | 2529 | rc = ata_pci_sff_init_host(host); |
700 | if (rc) | 2530 | if (rc) |
701 | goto err_out; | 2531 | goto err_out; |
702 | 2532 | ||
703 | /* init DMA related stuff */ | 2533 | /* init DMA related stuff */ |
704 | rc = ata_pci_init_bmdma(host); | 2534 | rc = ata_pci_bmdma_init(host); |
705 | if (rc) | 2535 | if (rc) |
706 | goto err_bmdma; | 2536 | goto err_bmdma; |
707 | 2537 | ||
@@ -722,7 +2552,7 @@ int ata_pci_prepare_sff_host(struct pci_dev *pdev, | |||
722 | } | 2552 | } |
723 | 2553 | ||
724 | /** | 2554 | /** |
725 | * ata_pci_activate_sff_host - start SFF host, request IRQ and register it | 2555 | * ata_pci_sff_activate_host - start SFF host, request IRQ and register it |
726 | * @host: target SFF ATA host | 2556 | * @host: target SFF ATA host |
727 | * @irq_handler: irq_handler used when requesting IRQ(s) | 2557 | * @irq_handler: irq_handler used when requesting IRQ(s) |
728 | * @sht: scsi_host_template to use when registering the host | 2558 | * @sht: scsi_host_template to use when registering the host |
@@ -737,7 +2567,7 @@ int ata_pci_prepare_sff_host(struct pci_dev *pdev, | |||
737 | * RETURNS: | 2567 | * RETURNS: |
738 | * 0 on success, -errno otherwise. | 2568 | * 0 on success, -errno otherwise. |
739 | */ | 2569 | */ |
740 | int ata_pci_activate_sff_host(struct ata_host *host, | 2570 | int ata_pci_sff_activate_host(struct ata_host *host, |
741 | irq_handler_t irq_handler, | 2571 | irq_handler_t irq_handler, |
742 | struct scsi_host_template *sht) | 2572 | struct scsi_host_template *sht) |
743 | { | 2573 | { |
@@ -815,9 +2645,11 @@ int ata_pci_activate_sff_host(struct ata_host *host, | |||
815 | } | 2645 | } |
816 | 2646 | ||
817 | /** | 2647 | /** |
818 | * ata_pci_init_one - Initialize/register PCI IDE host controller | 2648 | * ata_pci_sff_init_one - Initialize/register PCI IDE host controller |
819 | * @pdev: Controller to be initialized | 2649 | * @pdev: Controller to be initialized |
820 | * @ppi: array of port_info, must be enough for two ports | 2650 | * @ppi: array of port_info, must be enough for two ports |
2651 | * @sht: scsi_host_template to use when registering the host | ||
2652 | * @host_priv: host private_data | ||
821 | * | 2653 | * |
822 | * This is a helper function which can be called from a driver's | 2654 | * This is a helper function which can be called from a driver's |
823 | * xxx_init_one() probe function if the hardware uses traditional | 2655 | * xxx_init_one() probe function if the hardware uses traditional |
@@ -837,8 +2669,9 @@ int ata_pci_activate_sff_host(struct ata_host *host, | |||
837 | * RETURNS: | 2669 | * RETURNS: |
838 | * Zero on success, negative on errno-based value on error. | 2670 | * Zero on success, negative on errno-based value on error. |
839 | */ | 2671 | */ |
840 | int ata_pci_init_one(struct pci_dev *pdev, | 2672 | int ata_pci_sff_init_one(struct pci_dev *pdev, |
841 | const struct ata_port_info * const * ppi) | 2673 | const struct ata_port_info * const * ppi, |
2674 | struct scsi_host_template *sht, void *host_priv) | ||
842 | { | 2675 | { |
843 | struct device *dev = &pdev->dev; | 2676 | struct device *dev = &pdev->dev; |
844 | const struct ata_port_info *pi = NULL; | 2677 | const struct ata_port_info *pi = NULL; |
@@ -869,13 +2702,13 @@ int ata_pci_init_one(struct pci_dev *pdev, | |||
869 | goto out; | 2702 | goto out; |
870 | 2703 | ||
871 | /* prepare and activate SFF host */ | 2704 | /* prepare and activate SFF host */ |
872 | rc = ata_pci_prepare_sff_host(pdev, ppi, &host); | 2705 | rc = ata_pci_sff_prepare_host(pdev, ppi, &host); |
873 | if (rc) | 2706 | if (rc) |
874 | goto out; | 2707 | goto out; |
2708 | host->private_data = host_priv; | ||
875 | 2709 | ||
876 | pci_set_master(pdev); | 2710 | pci_set_master(pdev); |
877 | rc = ata_pci_activate_sff_host(host, pi->port_ops->irq_handler, | 2711 | rc = ata_pci_sff_activate_host(host, ata_sff_interrupt, sht); |
878 | pi->sht); | ||
879 | out: | 2712 | out: |
880 | if (rc == 0) | 2713 | if (rc == 0) |
881 | devres_remove_group(&pdev->dev, NULL); | 2714 | devres_remove_group(&pdev->dev, NULL); |
@@ -885,41 +2718,52 @@ int ata_pci_init_one(struct pci_dev *pdev, | |||
885 | return rc; | 2718 | return rc; |
886 | } | 2719 | } |
887 | 2720 | ||
888 | /** | ||
889 | * ata_pci_clear_simplex - attempt to kick device out of simplex | ||
890 | * @pdev: PCI device | ||
891 | * | ||
892 | * Some PCI ATA devices report simplex mode but in fact can be told to | ||
893 | * enter non simplex mode. This implements the necessary logic to | ||
894 | * perform the task on such devices. Calling it on other devices will | ||
895 | * have -undefined- behaviour. | ||
896 | */ | ||
897 | |||
898 | int ata_pci_clear_simplex(struct pci_dev *pdev) | ||
899 | { | ||
900 | unsigned long bmdma = pci_resource_start(pdev, 4); | ||
901 | u8 simplex; | ||
902 | |||
903 | if (bmdma == 0) | ||
904 | return -ENOENT; | ||
905 | |||
906 | simplex = inb(bmdma + 0x02); | ||
907 | outb(simplex & 0x60, bmdma + 0x02); | ||
908 | simplex = inb(bmdma + 0x02); | ||
909 | if (simplex & 0x80) | ||
910 | return -EOPNOTSUPP; | ||
911 | return 0; | ||
912 | } | ||
913 | |||
914 | unsigned long ata_pci_default_filter(struct ata_device *adev, unsigned long xfer_mask) | ||
915 | { | ||
916 | /* Filter out DMA modes if the device has been configured by | ||
917 | the BIOS as PIO only */ | ||
918 | |||
919 | if (adev->link->ap->ioaddr.bmdma_addr == NULL) | ||
920 | xfer_mask &= ~(ATA_MASK_MWDMA | ATA_MASK_UDMA); | ||
921 | return xfer_mask; | ||
922 | } | ||
923 | |||
924 | #endif /* CONFIG_PCI */ | 2721 | #endif /* CONFIG_PCI */ |
925 | 2722 | ||
2723 | EXPORT_SYMBOL_GPL(ata_sff_port_ops); | ||
2724 | EXPORT_SYMBOL_GPL(ata_bmdma_port_ops); | ||
2725 | EXPORT_SYMBOL_GPL(ata_sff_qc_prep); | ||
2726 | EXPORT_SYMBOL_GPL(ata_sff_dumb_qc_prep); | ||
2727 | EXPORT_SYMBOL_GPL(ata_sff_dev_select); | ||
2728 | EXPORT_SYMBOL_GPL(ata_sff_check_status); | ||
2729 | EXPORT_SYMBOL_GPL(ata_sff_altstatus); | ||
2730 | EXPORT_SYMBOL_GPL(ata_sff_busy_sleep); | ||
2731 | EXPORT_SYMBOL_GPL(ata_sff_wait_ready); | ||
2732 | EXPORT_SYMBOL_GPL(ata_sff_tf_load); | ||
2733 | EXPORT_SYMBOL_GPL(ata_sff_tf_read); | ||
2734 | EXPORT_SYMBOL_GPL(ata_sff_exec_command); | ||
2735 | EXPORT_SYMBOL_GPL(ata_sff_data_xfer); | ||
2736 | EXPORT_SYMBOL_GPL(ata_sff_data_xfer_noirq); | ||
2737 | EXPORT_SYMBOL_GPL(ata_sff_irq_on); | ||
2738 | EXPORT_SYMBOL_GPL(ata_sff_irq_clear); | ||
2739 | EXPORT_SYMBOL_GPL(ata_sff_hsm_move); | ||
2740 | EXPORT_SYMBOL_GPL(ata_sff_qc_issue); | ||
2741 | EXPORT_SYMBOL_GPL(ata_sff_qc_fill_rtf); | ||
2742 | EXPORT_SYMBOL_GPL(ata_sff_host_intr); | ||
2743 | EXPORT_SYMBOL_GPL(ata_sff_interrupt); | ||
2744 | EXPORT_SYMBOL_GPL(ata_sff_freeze); | ||
2745 | EXPORT_SYMBOL_GPL(ata_sff_thaw); | ||
2746 | EXPORT_SYMBOL_GPL(ata_sff_prereset); | ||
2747 | EXPORT_SYMBOL_GPL(ata_sff_dev_classify); | ||
2748 | EXPORT_SYMBOL_GPL(ata_sff_wait_after_reset); | ||
2749 | EXPORT_SYMBOL_GPL(ata_sff_softreset); | ||
2750 | EXPORT_SYMBOL_GPL(sata_sff_hardreset); | ||
2751 | EXPORT_SYMBOL_GPL(ata_sff_postreset); | ||
2752 | EXPORT_SYMBOL_GPL(ata_sff_error_handler); | ||
2753 | EXPORT_SYMBOL_GPL(ata_sff_post_internal_cmd); | ||
2754 | EXPORT_SYMBOL_GPL(ata_sff_port_start); | ||
2755 | EXPORT_SYMBOL_GPL(ata_sff_std_ports); | ||
2756 | EXPORT_SYMBOL_GPL(ata_bmdma_mode_filter); | ||
2757 | EXPORT_SYMBOL_GPL(ata_bmdma_setup); | ||
2758 | EXPORT_SYMBOL_GPL(ata_bmdma_start); | ||
2759 | EXPORT_SYMBOL_GPL(ata_bmdma_stop); | ||
2760 | EXPORT_SYMBOL_GPL(ata_bmdma_status); | ||
2761 | EXPORT_SYMBOL_GPL(ata_bus_reset); | ||
2762 | #ifdef CONFIG_PCI | ||
2763 | EXPORT_SYMBOL_GPL(ata_pci_bmdma_clear_simplex); | ||
2764 | EXPORT_SYMBOL_GPL(ata_pci_bmdma_init); | ||
2765 | EXPORT_SYMBOL_GPL(ata_pci_sff_init_host); | ||
2766 | EXPORT_SYMBOL_GPL(ata_pci_sff_prepare_host); | ||
2767 | EXPORT_SYMBOL_GPL(ata_pci_sff_activate_host); | ||
2768 | EXPORT_SYMBOL_GPL(ata_pci_sff_init_one); | ||
2769 | #endif /* CONFIG_PCI */ | ||
diff --git a/drivers/ata/libata.h b/drivers/ata/libata.h index aa884f71a12a..4aeeabb10a47 100644 --- a/drivers/ata/libata.h +++ b/drivers/ata/libata.h | |||
@@ -38,6 +38,17 @@ struct ata_scsi_args { | |||
38 | void (*done)(struct scsi_cmnd *); | 38 | void (*done)(struct scsi_cmnd *); |
39 | }; | 39 | }; |
40 | 40 | ||
41 | static inline int ata_is_builtin_hardreset(ata_reset_fn_t reset) | ||
42 | { | ||
43 | if (reset == sata_std_hardreset) | ||
44 | return 1; | ||
45 | #ifdef CONFIG_ATA_SFF | ||
46 | if (reset == sata_sff_hardreset) | ||
47 | return 1; | ||
48 | #endif | ||
49 | return 0; | ||
50 | } | ||
51 | |||
41 | /* libata-core.c */ | 52 | /* libata-core.c */ |
42 | enum { | 53 | enum { |
43 | /* flags for ata_dev_read_id() */ | 54 | /* flags for ata_dev_read_id() */ |
@@ -61,12 +72,16 @@ extern int libata_fua; | |||
61 | extern int libata_noacpi; | 72 | extern int libata_noacpi; |
62 | extern int libata_allow_tpm; | 73 | extern int libata_allow_tpm; |
63 | extern void ata_force_cbl(struct ata_port *ap); | 74 | extern void ata_force_cbl(struct ata_port *ap); |
75 | extern u64 ata_tf_to_lba(const struct ata_taskfile *tf); | ||
76 | extern u64 ata_tf_to_lba48(const struct ata_taskfile *tf); | ||
64 | extern struct ata_queued_cmd *ata_qc_new_init(struct ata_device *dev); | 77 | extern struct ata_queued_cmd *ata_qc_new_init(struct ata_device *dev); |
65 | extern int ata_build_rw_tf(struct ata_taskfile *tf, struct ata_device *dev, | 78 | extern int ata_build_rw_tf(struct ata_taskfile *tf, struct ata_device *dev, |
66 | u64 block, u32 n_block, unsigned int tf_flags, | 79 | u64 block, u32 n_block, unsigned int tf_flags, |
67 | unsigned int tag); | 80 | unsigned int tag); |
68 | extern u64 ata_tf_read_block(struct ata_taskfile *tf, struct ata_device *dev); | 81 | extern u64 ata_tf_read_block(struct ata_taskfile *tf, struct ata_device *dev); |
69 | extern void ata_dev_disable(struct ata_device *dev); | 82 | extern void ata_dev_disable(struct ata_device *dev); |
83 | extern void ata_pio_queue_task(struct ata_port *ap, void *data, | ||
84 | unsigned long delay); | ||
70 | extern void ata_port_flush_task(struct ata_port *ap); | 85 | extern void ata_port_flush_task(struct ata_port *ap); |
71 | extern unsigned ata_exec_internal(struct ata_device *dev, | 86 | extern unsigned ata_exec_internal(struct ata_device *dev, |
72 | struct ata_taskfile *tf, const u8 *cdb, | 87 | struct ata_taskfile *tf, const u8 *cdb, |
@@ -77,6 +92,8 @@ extern unsigned ata_exec_internal_sg(struct ata_device *dev, | |||
77 | int dma_dir, struct scatterlist *sg, | 92 | int dma_dir, struct scatterlist *sg, |
78 | unsigned int n_elem, unsigned long timeout); | 93 | unsigned int n_elem, unsigned long timeout); |
79 | extern unsigned int ata_do_simple_cmd(struct ata_device *dev, u8 cmd); | 94 | extern unsigned int ata_do_simple_cmd(struct ata_device *dev, u8 cmd); |
95 | extern int ata_wait_ready(struct ata_link *link, unsigned long deadline, | ||
96 | int (*check_ready)(struct ata_link *link)); | ||
80 | extern int ata_dev_read_id(struct ata_device *dev, unsigned int *p_class, | 97 | extern int ata_dev_read_id(struct ata_device *dev, unsigned int *p_class, |
81 | unsigned int flags, u16 *id); | 98 | unsigned int flags, u16 *id); |
82 | extern int ata_dev_reread_id(struct ata_device *dev, unsigned int readid_flags); | 99 | extern int ata_dev_reread_id(struct ata_device *dev, unsigned int readid_flags); |
@@ -91,10 +108,7 @@ extern void ata_qc_free(struct ata_queued_cmd *qc); | |||
91 | extern void ata_qc_issue(struct ata_queued_cmd *qc); | 108 | extern void ata_qc_issue(struct ata_queued_cmd *qc); |
92 | extern void __ata_qc_complete(struct ata_queued_cmd *qc); | 109 | extern void __ata_qc_complete(struct ata_queued_cmd *qc); |
93 | extern int ata_check_atapi_dma(struct ata_queued_cmd *qc); | 110 | extern int ata_check_atapi_dma(struct ata_queued_cmd *qc); |
94 | extern void ata_dev_select(struct ata_port *ap, unsigned int device, | ||
95 | unsigned int wait, unsigned int can_sleep); | ||
96 | extern void swap_buf_le16(u16 *buf, unsigned int buf_words); | 111 | extern void swap_buf_le16(u16 *buf, unsigned int buf_words); |
97 | extern int ata_flush_cache(struct ata_device *dev); | ||
98 | extern void ata_dev_init(struct ata_device *dev); | 112 | extern void ata_dev_init(struct ata_device *dev); |
99 | extern void ata_link_init(struct ata_port *ap, struct ata_link *link, int pmp); | 113 | extern void ata_link_init(struct ata_port *ap, struct ata_link *link, int pmp); |
100 | extern int sata_link_init_spd(struct ata_link *link); | 114 | extern int sata_link_init_spd(struct ata_link *link); |
@@ -165,11 +179,6 @@ extern void ata_schedule_scsi_eh(struct Scsi_Host *shost); | |||
165 | extern void ata_scsi_dev_rescan(struct work_struct *work); | 179 | extern void ata_scsi_dev_rescan(struct work_struct *work); |
166 | extern int ata_bus_probe(struct ata_port *ap); | 180 | extern int ata_bus_probe(struct ata_port *ap); |
167 | 181 | ||
168 | /* libata-pmp.c */ | ||
169 | extern int sata_pmp_scr_read(struct ata_link *link, int reg, u32 *val); | ||
170 | extern int sata_pmp_scr_write(struct ata_link *link, int reg, u32 val); | ||
171 | extern int sata_pmp_attach(struct ata_device *dev); | ||
172 | |||
173 | /* libata-eh.c */ | 182 | /* libata-eh.c */ |
174 | extern enum scsi_eh_timer_return ata_scsi_timed_out(struct scsi_cmnd *cmd); | 183 | extern enum scsi_eh_timer_return ata_scsi_timed_out(struct scsi_cmnd *cmd); |
175 | extern void ata_scsi_error(struct Scsi_Host *host); | 184 | extern void ata_scsi_error(struct Scsi_Host *host); |
@@ -193,8 +202,34 @@ extern int ata_eh_recover(struct ata_port *ap, ata_prereset_fn_t prereset, | |||
193 | struct ata_link **r_failed_disk); | 202 | struct ata_link **r_failed_disk); |
194 | extern void ata_eh_finish(struct ata_port *ap); | 203 | extern void ata_eh_finish(struct ata_port *ap); |
195 | 204 | ||
205 | /* libata-pmp.c */ | ||
206 | #ifdef CONFIG_SATA_PMP | ||
207 | extern int sata_pmp_scr_read(struct ata_link *link, int reg, u32 *val); | ||
208 | extern int sata_pmp_scr_write(struct ata_link *link, int reg, u32 val); | ||
209 | extern int sata_pmp_attach(struct ata_device *dev); | ||
210 | #else /* CONFIG_SATA_PMP */ | ||
211 | static inline int sata_pmp_scr_read(struct ata_link *link, int reg, u32 *val) | ||
212 | { | ||
213 | return -EINVAL; | ||
214 | } | ||
215 | |||
216 | static inline int sata_pmp_scr_write(struct ata_link *link, int reg, u32 val) | ||
217 | { | ||
218 | return -EINVAL; | ||
219 | } | ||
220 | |||
221 | static inline int sata_pmp_attach(struct ata_device *dev) | ||
222 | { | ||
223 | return -EINVAL; | ||
224 | } | ||
225 | #endif /* CONFIG_SATA_PMP */ | ||
226 | |||
196 | /* libata-sff.c */ | 227 | /* libata-sff.c */ |
228 | #ifdef CONFIG_ATA_SFF | ||
229 | extern void ata_dev_select(struct ata_port *ap, unsigned int device, | ||
230 | unsigned int wait, unsigned int can_sleep); | ||
197 | extern u8 ata_irq_on(struct ata_port *ap); | 231 | extern u8 ata_irq_on(struct ata_port *ap); |
198 | 232 | extern void ata_pio_task(struct work_struct *work); | |
233 | #endif /* CONFIG_ATA_SFF */ | ||
199 | 234 | ||
200 | #endif /* __LIBATA_H__ */ | 235 | #endif /* __LIBATA_H__ */ |
diff --git a/drivers/ata/pata_acpi.c b/drivers/ata/pata_acpi.c index bdc3b9d7395c..c5f91e629945 100644 --- a/drivers/ata/pata_acpi.c +++ b/drivers/ata/pata_acpi.c | |||
@@ -47,7 +47,7 @@ static int pacpi_pre_reset(struct ata_link *link, unsigned long deadline) | |||
47 | if (ap->acpi_handle == NULL || ata_acpi_gtm(ap, &acpi->gtm) < 0) | 47 | if (ap->acpi_handle == NULL || ata_acpi_gtm(ap, &acpi->gtm) < 0) |
48 | return -ENODEV; | 48 | return -ENODEV; |
49 | 49 | ||
50 | return ata_std_prereset(link, deadline); | 50 | return ata_sff_prereset(link, deadline); |
51 | } | 51 | } |
52 | 52 | ||
53 | /** | 53 | /** |
@@ -68,20 +68,6 @@ static int pacpi_cable_detect(struct ata_port *ap) | |||
68 | } | 68 | } |
69 | 69 | ||
70 | /** | 70 | /** |
71 | * pacpi_error_handler - Setup and error handler | ||
72 | * @ap: Port to handle | ||
73 | * | ||
74 | * LOCKING: | ||
75 | * None (inherited from caller). | ||
76 | */ | ||
77 | |||
78 | static void pacpi_error_handler(struct ata_port *ap) | ||
79 | { | ||
80 | ata_bmdma_drive_eh(ap, pacpi_pre_reset, ata_std_softreset, NULL, | ||
81 | ata_std_postreset); | ||
82 | } | ||
83 | |||
84 | /** | ||
85 | * pacpi_discover_modes - filter non ACPI modes | 71 | * pacpi_discover_modes - filter non ACPI modes |
86 | * @adev: ATA device | 72 | * @adev: ATA device |
87 | * @mask: proposed modes | 73 | * @mask: proposed modes |
@@ -120,7 +106,7 @@ static unsigned long pacpi_discover_modes(struct ata_port *ap, struct ata_device | |||
120 | static unsigned long pacpi_mode_filter(struct ata_device *adev, unsigned long mask) | 106 | static unsigned long pacpi_mode_filter(struct ata_device *adev, unsigned long mask) |
121 | { | 107 | { |
122 | struct pata_acpi *acpi = adev->link->ap->private_data; | 108 | struct pata_acpi *acpi = adev->link->ap->private_data; |
123 | return ata_pci_default_filter(adev, mask & acpi->mask[adev->devno]); | 109 | return ata_bmdma_mode_filter(adev, mask & acpi->mask[adev->devno]); |
124 | } | 110 | } |
125 | 111 | ||
126 | /** | 112 | /** |
@@ -176,7 +162,7 @@ static void pacpi_set_dmamode(struct ata_port *ap, struct ata_device *adev) | |||
176 | } | 162 | } |
177 | 163 | ||
178 | /** | 164 | /** |
179 | * pacpi_qc_issue_prot - command issue | 165 | * pacpi_qc_issue - command issue |
180 | * @qc: command pending | 166 | * @qc: command pending |
181 | * | 167 | * |
182 | * Called when the libata layer is about to issue a command. We wrap | 168 | * Called when the libata layer is about to issue a command. We wrap |
@@ -184,14 +170,14 @@ static void pacpi_set_dmamode(struct ata_port *ap, struct ata_device *adev) | |||
184 | * neccessary. | 170 | * neccessary. |
185 | */ | 171 | */ |
186 | 172 | ||
187 | static unsigned int pacpi_qc_issue_prot(struct ata_queued_cmd *qc) | 173 | static unsigned int pacpi_qc_issue(struct ata_queued_cmd *qc) |
188 | { | 174 | { |
189 | struct ata_port *ap = qc->ap; | 175 | struct ata_port *ap = qc->ap; |
190 | struct ata_device *adev = qc->dev; | 176 | struct ata_device *adev = qc->dev; |
191 | struct pata_acpi *acpi = ap->private_data; | 177 | struct pata_acpi *acpi = ap->private_data; |
192 | 178 | ||
193 | if (acpi->gtm.flags & 0x10) | 179 | if (acpi->gtm.flags & 0x10) |
194 | return ata_qc_issue_prot(qc); | 180 | return ata_sff_qc_issue(qc); |
195 | 181 | ||
196 | if (adev != acpi->last) { | 182 | if (adev != acpi->last) { |
197 | pacpi_set_piomode(ap, adev); | 183 | pacpi_set_piomode(ap, adev); |
@@ -199,7 +185,7 @@ static unsigned int pacpi_qc_issue_prot(struct ata_queued_cmd *qc) | |||
199 | pacpi_set_dmamode(ap, adev); | 185 | pacpi_set_dmamode(ap, adev); |
200 | acpi->last = adev; | 186 | acpi->last = adev; |
201 | } | 187 | } |
202 | return ata_qc_issue_prot(qc); | 188 | return ata_sff_qc_issue(qc); |
203 | } | 189 | } |
204 | 190 | ||
205 | /** | 191 | /** |
@@ -232,57 +218,17 @@ static int pacpi_port_start(struct ata_port *ap) | |||
232 | } | 218 | } |
233 | 219 | ||
234 | static struct scsi_host_template pacpi_sht = { | 220 | static struct scsi_host_template pacpi_sht = { |
235 | .module = THIS_MODULE, | 221 | ATA_BMDMA_SHT(DRV_NAME), |
236 | .name = DRV_NAME, | ||
237 | .ioctl = ata_scsi_ioctl, | ||
238 | .queuecommand = ata_scsi_queuecmd, | ||
239 | .can_queue = ATA_DEF_QUEUE, | ||
240 | .this_id = ATA_SHT_THIS_ID, | ||
241 | .sg_tablesize = LIBATA_MAX_PRD, | ||
242 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
243 | .emulated = ATA_SHT_EMULATED, | ||
244 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
245 | .proc_name = DRV_NAME, | ||
246 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
247 | .slave_configure = ata_scsi_slave_config, | ||
248 | .slave_destroy = ata_scsi_slave_destroy, | ||
249 | /* Use standard CHS mapping rules */ | ||
250 | .bios_param = ata_std_bios_param, | ||
251 | }; | 222 | }; |
252 | 223 | ||
253 | static const struct ata_port_operations pacpi_ops = { | 224 | static struct ata_port_operations pacpi_ops = { |
225 | .inherits = &ata_bmdma_port_ops, | ||
226 | .qc_issue = pacpi_qc_issue, | ||
227 | .cable_detect = pacpi_cable_detect, | ||
228 | .mode_filter = pacpi_mode_filter, | ||
254 | .set_piomode = pacpi_set_piomode, | 229 | .set_piomode = pacpi_set_piomode, |
255 | .set_dmamode = pacpi_set_dmamode, | 230 | .set_dmamode = pacpi_set_dmamode, |
256 | .mode_filter = pacpi_mode_filter, | 231 | .prereset = pacpi_pre_reset, |
257 | |||
258 | /* Task file is PCI ATA format, use helpers */ | ||
259 | .tf_load = ata_tf_load, | ||
260 | .tf_read = ata_tf_read, | ||
261 | .check_status = ata_check_status, | ||
262 | .exec_command = ata_exec_command, | ||
263 | .dev_select = ata_std_dev_select, | ||
264 | |||
265 | .freeze = ata_bmdma_freeze, | ||
266 | .thaw = ata_bmdma_thaw, | ||
267 | .error_handler = pacpi_error_handler, | ||
268 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
269 | .cable_detect = pacpi_cable_detect, | ||
270 | |||
271 | /* BMDMA handling is PCI ATA format, use helpers */ | ||
272 | .bmdma_setup = ata_bmdma_setup, | ||
273 | .bmdma_start = ata_bmdma_start, | ||
274 | .bmdma_stop = ata_bmdma_stop, | ||
275 | .bmdma_status = ata_bmdma_status, | ||
276 | .qc_prep = ata_qc_prep, | ||
277 | .qc_issue = pacpi_qc_issue_prot, | ||
278 | .data_xfer = ata_data_xfer, | ||
279 | |||
280 | /* Timeout handling */ | ||
281 | .irq_handler = ata_interrupt, | ||
282 | .irq_clear = ata_bmdma_irq_clear, | ||
283 | .irq_on = ata_irq_on, | ||
284 | |||
285 | /* Generic PATA PCI ATA helpers */ | ||
286 | .port_start = pacpi_port_start, | 232 | .port_start = pacpi_port_start, |
287 | }; | 233 | }; |
288 | 234 | ||
@@ -304,7 +250,6 @@ static const struct ata_port_operations pacpi_ops = { | |||
304 | static int pacpi_init_one (struct pci_dev *pdev, const struct pci_device_id *id) | 250 | static int pacpi_init_one (struct pci_dev *pdev, const struct pci_device_id *id) |
305 | { | 251 | { |
306 | static const struct ata_port_info info = { | 252 | static const struct ata_port_info info = { |
307 | .sht = &pacpi_sht, | ||
308 | .flags = ATA_FLAG_SLAVE_POSS | ATA_FLAG_SRST, | 253 | .flags = ATA_FLAG_SLAVE_POSS | ATA_FLAG_SRST, |
309 | 254 | ||
310 | .pio_mask = 0x1f, | 255 | .pio_mask = 0x1f, |
@@ -314,7 +259,7 @@ static int pacpi_init_one (struct pci_dev *pdev, const struct pci_device_id *id) | |||
314 | .port_ops = &pacpi_ops, | 259 | .port_ops = &pacpi_ops, |
315 | }; | 260 | }; |
316 | const struct ata_port_info *ppi[] = { &info, NULL }; | 261 | const struct ata_port_info *ppi[] = { &info, NULL }; |
317 | return ata_pci_init_one(pdev, ppi); | 262 | return ata_pci_sff_init_one(pdev, ppi, &pacpi_sht, NULL); |
318 | } | 263 | } |
319 | 264 | ||
320 | static const struct pci_device_id pacpi_pci_tbl[] = { | 265 | static const struct pci_device_id pacpi_pci_tbl[] = { |
diff --git a/drivers/ata/pata_ali.c b/drivers/ata/pata_ali.c index 511a830b6256..fcabe46f262b 100644 --- a/drivers/ata/pata_ali.c +++ b/drivers/ata/pata_ali.c | |||
@@ -121,7 +121,7 @@ static unsigned long ali_20_filter(struct ata_device *adev, unsigned long mask) | |||
121 | ata_id_c_string(adev->id, model_num, ATA_ID_PROD, sizeof(model_num)); | 121 | ata_id_c_string(adev->id, model_num, ATA_ID_PROD, sizeof(model_num)); |
122 | if (strstr(model_num, "WDC")) | 122 | if (strstr(model_num, "WDC")) |
123 | return mask &= ~ATA_MASK_UDMA; | 123 | return mask &= ~ATA_MASK_UDMA; |
124 | return ata_pci_default_filter(adev, mask); | 124 | return ata_bmdma_mode_filter(adev, mask); |
125 | } | 125 | } |
126 | 126 | ||
127 | /** | 127 | /** |
@@ -339,21 +339,7 @@ static int ali_check_atapi_dma(struct ata_queued_cmd *qc) | |||
339 | } | 339 | } |
340 | 340 | ||
341 | static struct scsi_host_template ali_sht = { | 341 | static struct scsi_host_template ali_sht = { |
342 | .module = THIS_MODULE, | 342 | ATA_BMDMA_SHT(DRV_NAME), |
343 | .name = DRV_NAME, | ||
344 | .ioctl = ata_scsi_ioctl, | ||
345 | .queuecommand = ata_scsi_queuecmd, | ||
346 | .can_queue = ATA_DEF_QUEUE, | ||
347 | .this_id = ATA_SHT_THIS_ID, | ||
348 | .sg_tablesize = LIBATA_MAX_PRD, | ||
349 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
350 | .emulated = ATA_SHT_EMULATED, | ||
351 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
352 | .proc_name = DRV_NAME, | ||
353 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
354 | .slave_configure = ata_scsi_slave_config, | ||
355 | .slave_destroy = ata_scsi_slave_destroy, | ||
356 | .bios_param = ata_std_bios_param, | ||
357 | }; | 343 | }; |
358 | 344 | ||
359 | /* | 345 | /* |
@@ -361,29 +347,15 @@ static struct scsi_host_template ali_sht = { | |||
361 | */ | 347 | */ |
362 | 348 | ||
363 | static struct ata_port_operations ali_early_port_ops = { | 349 | static struct ata_port_operations ali_early_port_ops = { |
364 | .set_piomode = ali_set_piomode, | 350 | .inherits = &ata_sff_port_ops, |
365 | .tf_load = ata_tf_load, | ||
366 | .tf_read = ata_tf_read, | ||
367 | .check_status = ata_check_status, | ||
368 | .exec_command = ata_exec_command, | ||
369 | .dev_select = ata_std_dev_select, | ||
370 | |||
371 | .freeze = ata_bmdma_freeze, | ||
372 | .thaw = ata_bmdma_thaw, | ||
373 | .error_handler = ata_bmdma_error_handler, | ||
374 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
375 | .cable_detect = ata_cable_40wire, | 351 | .cable_detect = ata_cable_40wire, |
352 | .set_piomode = ali_set_piomode, | ||
353 | }; | ||
376 | 354 | ||
377 | .qc_prep = ata_qc_prep, | 355 | static const struct ata_port_operations ali_dma_base_ops = { |
378 | .qc_issue = ata_qc_issue_prot, | 356 | .inherits = &ata_bmdma_port_ops, |
379 | 357 | .set_piomode = ali_set_piomode, | |
380 | .data_xfer = ata_data_xfer, | 358 | .set_dmamode = ali_set_dmamode, |
381 | |||
382 | .irq_handler = ata_interrupt, | ||
383 | .irq_clear = ata_bmdma_irq_clear, | ||
384 | .irq_on = ata_irq_on, | ||
385 | |||
386 | .port_start = ata_sff_port_start, | ||
387 | }; | 359 | }; |
388 | 360 | ||
389 | /* | 361 | /* |
@@ -391,115 +363,31 @@ static struct ata_port_operations ali_early_port_ops = { | |||
391 | * detect | 363 | * detect |
392 | */ | 364 | */ |
393 | static struct ata_port_operations ali_20_port_ops = { | 365 | static struct ata_port_operations ali_20_port_ops = { |
394 | .set_piomode = ali_set_piomode, | 366 | .inherits = &ali_dma_base_ops, |
395 | .set_dmamode = ali_set_dmamode, | 367 | .cable_detect = ata_cable_40wire, |
396 | .mode_filter = ali_20_filter, | 368 | .mode_filter = ali_20_filter, |
397 | |||
398 | .tf_load = ata_tf_load, | ||
399 | .tf_read = ata_tf_read, | ||
400 | .check_atapi_dma = ali_check_atapi_dma, | 369 | .check_atapi_dma = ali_check_atapi_dma, |
401 | .check_status = ata_check_status, | ||
402 | .exec_command = ata_exec_command, | ||
403 | .dev_select = ata_std_dev_select, | ||
404 | .dev_config = ali_lock_sectors, | 370 | .dev_config = ali_lock_sectors, |
405 | |||
406 | .freeze = ata_bmdma_freeze, | ||
407 | .thaw = ata_bmdma_thaw, | ||
408 | .error_handler = ata_bmdma_error_handler, | ||
409 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
410 | .cable_detect = ata_cable_40wire, | ||
411 | |||
412 | .bmdma_setup = ata_bmdma_setup, | ||
413 | .bmdma_start = ata_bmdma_start, | ||
414 | .bmdma_stop = ata_bmdma_stop, | ||
415 | .bmdma_status = ata_bmdma_status, | ||
416 | |||
417 | .qc_prep = ata_qc_prep, | ||
418 | .qc_issue = ata_qc_issue_prot, | ||
419 | |||
420 | .data_xfer = ata_data_xfer, | ||
421 | |||
422 | .irq_handler = ata_interrupt, | ||
423 | .irq_clear = ata_bmdma_irq_clear, | ||
424 | .irq_on = ata_irq_on, | ||
425 | |||
426 | .port_start = ata_sff_port_start, | ||
427 | }; | 371 | }; |
428 | 372 | ||
429 | /* | 373 | /* |
430 | * Port operations for DMA capable ALi with cable detect | 374 | * Port operations for DMA capable ALi with cable detect |
431 | */ | 375 | */ |
432 | static struct ata_port_operations ali_c2_port_ops = { | 376 | static struct ata_port_operations ali_c2_port_ops = { |
433 | .set_piomode = ali_set_piomode, | 377 | .inherits = &ali_dma_base_ops, |
434 | .set_dmamode = ali_set_dmamode, | ||
435 | .mode_filter = ata_pci_default_filter, | ||
436 | .tf_load = ata_tf_load, | ||
437 | .tf_read = ata_tf_read, | ||
438 | .check_atapi_dma = ali_check_atapi_dma, | 378 | .check_atapi_dma = ali_check_atapi_dma, |
439 | .check_status = ata_check_status, | ||
440 | .exec_command = ata_exec_command, | ||
441 | .dev_select = ata_std_dev_select, | ||
442 | .dev_config = ali_lock_sectors, | ||
443 | |||
444 | .freeze = ata_bmdma_freeze, | ||
445 | .thaw = ata_bmdma_thaw, | ||
446 | .error_handler = ata_bmdma_error_handler, | ||
447 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
448 | .cable_detect = ali_c2_cable_detect, | 379 | .cable_detect = ali_c2_cable_detect, |
449 | 380 | .dev_config = ali_lock_sectors, | |
450 | .bmdma_setup = ata_bmdma_setup, | ||
451 | .bmdma_start = ata_bmdma_start, | ||
452 | .bmdma_stop = ata_bmdma_stop, | ||
453 | .bmdma_status = ata_bmdma_status, | ||
454 | |||
455 | .qc_prep = ata_qc_prep, | ||
456 | .qc_issue = ata_qc_issue_prot, | ||
457 | |||
458 | .data_xfer = ata_data_xfer, | ||
459 | |||
460 | .irq_handler = ata_interrupt, | ||
461 | .irq_clear = ata_bmdma_irq_clear, | ||
462 | .irq_on = ata_irq_on, | ||
463 | |||
464 | .port_start = ata_sff_port_start, | ||
465 | }; | 381 | }; |
466 | 382 | ||
467 | /* | 383 | /* |
468 | * Port operations for DMA capable ALi with cable detect and LBA48 | 384 | * Port operations for DMA capable ALi with cable detect and LBA48 |
469 | */ | 385 | */ |
470 | static struct ata_port_operations ali_c5_port_ops = { | 386 | static struct ata_port_operations ali_c5_port_ops = { |
471 | .set_piomode = ali_set_piomode, | 387 | .inherits = &ali_dma_base_ops, |
472 | .set_dmamode = ali_set_dmamode, | ||
473 | .mode_filter = ata_pci_default_filter, | ||
474 | .tf_load = ata_tf_load, | ||
475 | .tf_read = ata_tf_read, | ||
476 | .check_atapi_dma = ali_check_atapi_dma, | 388 | .check_atapi_dma = ali_check_atapi_dma, |
477 | .check_status = ata_check_status, | ||
478 | .exec_command = ata_exec_command, | ||
479 | .dev_select = ata_std_dev_select, | ||
480 | .dev_config = ali_warn_atapi_dma, | 389 | .dev_config = ali_warn_atapi_dma, |
481 | |||
482 | .freeze = ata_bmdma_freeze, | ||
483 | .thaw = ata_bmdma_thaw, | ||
484 | .error_handler = ata_bmdma_error_handler, | ||
485 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
486 | .cable_detect = ali_c2_cable_detect, | 390 | .cable_detect = ali_c2_cable_detect, |
487 | |||
488 | .bmdma_setup = ata_bmdma_setup, | ||
489 | .bmdma_start = ata_bmdma_start, | ||
490 | .bmdma_stop = ata_bmdma_stop, | ||
491 | .bmdma_status = ata_bmdma_status, | ||
492 | |||
493 | .qc_prep = ata_qc_prep, | ||
494 | .qc_issue = ata_qc_issue_prot, | ||
495 | |||
496 | .data_xfer = ata_data_xfer, | ||
497 | |||
498 | .irq_handler = ata_interrupt, | ||
499 | .irq_clear = ata_bmdma_irq_clear, | ||
500 | .irq_on = ata_irq_on, | ||
501 | |||
502 | .port_start = ata_sff_port_start, | ||
503 | }; | 391 | }; |
504 | 392 | ||
505 | 393 | ||
@@ -561,7 +449,7 @@ static void ali_init_chipset(struct pci_dev *pdev) | |||
561 | } | 449 | } |
562 | pci_dev_put(isa_bridge); | 450 | pci_dev_put(isa_bridge); |
563 | pci_dev_put(north); | 451 | pci_dev_put(north); |
564 | ata_pci_clear_simplex(pdev); | 452 | ata_pci_bmdma_clear_simplex(pdev); |
565 | } | 453 | } |
566 | /** | 454 | /** |
567 | * ali_init_one - discovery callback | 455 | * ali_init_one - discovery callback |
@@ -575,14 +463,12 @@ static void ali_init_chipset(struct pci_dev *pdev) | |||
575 | static int ali_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | 463 | static int ali_init_one(struct pci_dev *pdev, const struct pci_device_id *id) |
576 | { | 464 | { |
577 | static const struct ata_port_info info_early = { | 465 | static const struct ata_port_info info_early = { |
578 | .sht = &ali_sht, | ||
579 | .flags = ATA_FLAG_SLAVE_POSS, | 466 | .flags = ATA_FLAG_SLAVE_POSS, |
580 | .pio_mask = 0x1f, | 467 | .pio_mask = 0x1f, |
581 | .port_ops = &ali_early_port_ops | 468 | .port_ops = &ali_early_port_ops |
582 | }; | 469 | }; |
583 | /* Revision 0x20 added DMA */ | 470 | /* Revision 0x20 added DMA */ |
584 | static const struct ata_port_info info_20 = { | 471 | static const struct ata_port_info info_20 = { |
585 | .sht = &ali_sht, | ||
586 | .flags = ATA_FLAG_SLAVE_POSS | ATA_FLAG_PIO_LBA48, | 472 | .flags = ATA_FLAG_SLAVE_POSS | ATA_FLAG_PIO_LBA48, |
587 | .pio_mask = 0x1f, | 473 | .pio_mask = 0x1f, |
588 | .mwdma_mask = 0x07, | 474 | .mwdma_mask = 0x07, |
@@ -590,7 +476,6 @@ static int ali_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
590 | }; | 476 | }; |
591 | /* Revision 0x20 with support logic added UDMA */ | 477 | /* Revision 0x20 with support logic added UDMA */ |
592 | static const struct ata_port_info info_20_udma = { | 478 | static const struct ata_port_info info_20_udma = { |
593 | .sht = &ali_sht, | ||
594 | .flags = ATA_FLAG_SLAVE_POSS | ATA_FLAG_PIO_LBA48, | 479 | .flags = ATA_FLAG_SLAVE_POSS | ATA_FLAG_PIO_LBA48, |
595 | .pio_mask = 0x1f, | 480 | .pio_mask = 0x1f, |
596 | .mwdma_mask = 0x07, | 481 | .mwdma_mask = 0x07, |
@@ -599,7 +484,6 @@ static int ali_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
599 | }; | 484 | }; |
600 | /* Revision 0xC2 adds UDMA66 */ | 485 | /* Revision 0xC2 adds UDMA66 */ |
601 | static const struct ata_port_info info_c2 = { | 486 | static const struct ata_port_info info_c2 = { |
602 | .sht = &ali_sht, | ||
603 | .flags = ATA_FLAG_SLAVE_POSS | ATA_FLAG_PIO_LBA48, | 487 | .flags = ATA_FLAG_SLAVE_POSS | ATA_FLAG_PIO_LBA48, |
604 | .pio_mask = 0x1f, | 488 | .pio_mask = 0x1f, |
605 | .mwdma_mask = 0x07, | 489 | .mwdma_mask = 0x07, |
@@ -608,7 +492,6 @@ static int ali_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
608 | }; | 492 | }; |
609 | /* Revision 0xC3 is UDMA66 for now */ | 493 | /* Revision 0xC3 is UDMA66 for now */ |
610 | static const struct ata_port_info info_c3 = { | 494 | static const struct ata_port_info info_c3 = { |
611 | .sht = &ali_sht, | ||
612 | .flags = ATA_FLAG_SLAVE_POSS | ATA_FLAG_PIO_LBA48, | 495 | .flags = ATA_FLAG_SLAVE_POSS | ATA_FLAG_PIO_LBA48, |
613 | .pio_mask = 0x1f, | 496 | .pio_mask = 0x1f, |
614 | .mwdma_mask = 0x07, | 497 | .mwdma_mask = 0x07, |
@@ -617,7 +500,6 @@ static int ali_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
617 | }; | 500 | }; |
618 | /* Revision 0xC4 is UDMA100 */ | 501 | /* Revision 0xC4 is UDMA100 */ |
619 | static const struct ata_port_info info_c4 = { | 502 | static const struct ata_port_info info_c4 = { |
620 | .sht = &ali_sht, | ||
621 | .flags = ATA_FLAG_SLAVE_POSS | ATA_FLAG_PIO_LBA48, | 503 | .flags = ATA_FLAG_SLAVE_POSS | ATA_FLAG_PIO_LBA48, |
622 | .pio_mask = 0x1f, | 504 | .pio_mask = 0x1f, |
623 | .mwdma_mask = 0x07, | 505 | .mwdma_mask = 0x07, |
@@ -626,7 +508,6 @@ static int ali_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
626 | }; | 508 | }; |
627 | /* Revision 0xC5 is UDMA133 with LBA48 DMA */ | 509 | /* Revision 0xC5 is UDMA133 with LBA48 DMA */ |
628 | static const struct ata_port_info info_c5 = { | 510 | static const struct ata_port_info info_c5 = { |
629 | .sht = &ali_sht, | ||
630 | .flags = ATA_FLAG_SLAVE_POSS, | 511 | .flags = ATA_FLAG_SLAVE_POSS, |
631 | .pio_mask = 0x1f, | 512 | .pio_mask = 0x1f, |
632 | .mwdma_mask = 0x07, | 513 | .mwdma_mask = 0x07, |
@@ -637,6 +518,11 @@ static int ali_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
637 | const struct ata_port_info *ppi[] = { NULL, NULL }; | 518 | const struct ata_port_info *ppi[] = { NULL, NULL }; |
638 | u8 tmp; | 519 | u8 tmp; |
639 | struct pci_dev *isa_bridge; | 520 | struct pci_dev *isa_bridge; |
521 | int rc; | ||
522 | |||
523 | rc = pcim_enable_device(pdev); | ||
524 | if (rc) | ||
525 | return rc; | ||
640 | 526 | ||
641 | /* | 527 | /* |
642 | * The chipset revision selects the driver operations and | 528 | * The chipset revision selects the driver operations and |
@@ -666,14 +552,21 @@ static int ali_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
666 | ppi[0] = &info_20_udma; | 552 | ppi[0] = &info_20_udma; |
667 | pci_dev_put(isa_bridge); | 553 | pci_dev_put(isa_bridge); |
668 | } | 554 | } |
669 | return ata_pci_init_one(pdev, ppi); | 555 | return ata_pci_sff_init_one(pdev, ppi, &ali_sht, NULL); |
670 | } | 556 | } |
671 | 557 | ||
672 | #ifdef CONFIG_PM | 558 | #ifdef CONFIG_PM |
673 | static int ali_reinit_one(struct pci_dev *pdev) | 559 | static int ali_reinit_one(struct pci_dev *pdev) |
674 | { | 560 | { |
561 | struct ata_host *host = dev_get_drvdata(&pdev->dev); | ||
562 | int rc; | ||
563 | |||
564 | rc = ata_pci_device_do_resume(pdev); | ||
565 | if (rc) | ||
566 | return rc; | ||
675 | ali_init_chipset(pdev); | 567 | ali_init_chipset(pdev); |
676 | return ata_pci_device_resume(pdev); | 568 | ata_host_resume(host); |
569 | return 0; | ||
677 | } | 570 | } |
678 | #endif | 571 | #endif |
679 | 572 | ||
diff --git a/drivers/ata/pata_amd.c b/drivers/ata/pata_amd.c index 4b8d9b592ca4..26665c396485 100644 --- a/drivers/ata/pata_amd.c +++ b/drivers/ata/pata_amd.c | |||
@@ -56,7 +56,9 @@ static void timing_setup(struct ata_port *ap, struct ata_device *adev, int offse | |||
56 | u8 t; | 56 | u8 t; |
57 | 57 | ||
58 | T = 1000000000 / amd_clock; | 58 | T = 1000000000 / amd_clock; |
59 | UT = T / min_t(int, max_t(int, clock, 1), 2); | 59 | UT = T; |
60 | if (clock >= 2) | ||
61 | UT = T / 2; | ||
60 | 62 | ||
61 | if (ata_timing_compute(adev, speed, &at, T, UT) < 0) { | 63 | if (ata_timing_compute(adev, speed, &at, T, UT) < 0) { |
62 | dev_printk(KERN_ERR, &pdev->dev, "unknown mode %d.\n", speed); | 64 | dev_printk(KERN_ERR, &pdev->dev, "unknown mode %d.\n", speed); |
@@ -141,13 +143,7 @@ static int amd_pre_reset(struct ata_link *link, unsigned long deadline) | |||
141 | if (!pci_test_config_bits(pdev, &amd_enable_bits[ap->port_no])) | 143 | if (!pci_test_config_bits(pdev, &amd_enable_bits[ap->port_no])) |
142 | return -ENOENT; | 144 | return -ENOENT; |
143 | 145 | ||
144 | return ata_std_prereset(link, deadline); | 146 | return ata_sff_prereset(link, deadline); |
145 | } | ||
146 | |||
147 | static void amd_error_handler(struct ata_port *ap) | ||
148 | { | ||
149 | ata_bmdma_drive_eh(ap, amd_pre_reset, ata_std_softreset, NULL, | ||
150 | ata_std_postreset); | ||
151 | } | 147 | } |
152 | 148 | ||
153 | static int amd_cable_detect(struct ata_port *ap) | 149 | static int amd_cable_detect(struct ata_port *ap) |
@@ -297,14 +293,7 @@ static int nv_pre_reset(struct ata_link *link, unsigned long deadline) | |||
297 | if (!pci_test_config_bits(pdev, &nv_enable_bits[ap->port_no])) | 293 | if (!pci_test_config_bits(pdev, &nv_enable_bits[ap->port_no])) |
298 | return -ENOENT; | 294 | return -ENOENT; |
299 | 295 | ||
300 | return ata_std_prereset(link, deadline); | 296 | return ata_sff_prereset(link, deadline); |
301 | } | ||
302 | |||
303 | static void nv_error_handler(struct ata_port *ap) | ||
304 | { | ||
305 | ata_bmdma_drive_eh(ap, nv_pre_reset, | ||
306 | ata_std_softreset, NULL, | ||
307 | ata_std_postreset); | ||
308 | } | 297 | } |
309 | 298 | ||
310 | /** | 299 | /** |
@@ -353,228 +342,66 @@ static void nv_host_stop(struct ata_host *host) | |||
353 | } | 342 | } |
354 | 343 | ||
355 | static struct scsi_host_template amd_sht = { | 344 | static struct scsi_host_template amd_sht = { |
356 | .module = THIS_MODULE, | 345 | ATA_BMDMA_SHT(DRV_NAME), |
357 | .name = DRV_NAME, | 346 | }; |
358 | .ioctl = ata_scsi_ioctl, | 347 | |
359 | .queuecommand = ata_scsi_queuecmd, | 348 | static const struct ata_port_operations amd_base_port_ops = { |
360 | .can_queue = ATA_DEF_QUEUE, | 349 | .inherits = &ata_bmdma_port_ops, |
361 | .this_id = ATA_SHT_THIS_ID, | 350 | .prereset = amd_pre_reset, |
362 | .sg_tablesize = LIBATA_MAX_PRD, | ||
363 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
364 | .emulated = ATA_SHT_EMULATED, | ||
365 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
366 | .proc_name = DRV_NAME, | ||
367 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
368 | .slave_configure = ata_scsi_slave_config, | ||
369 | .slave_destroy = ata_scsi_slave_destroy, | ||
370 | .bios_param = ata_std_bios_param, | ||
371 | }; | 351 | }; |
372 | 352 | ||
373 | static struct ata_port_operations amd33_port_ops = { | 353 | static struct ata_port_operations amd33_port_ops = { |
354 | .inherits = &amd_base_port_ops, | ||
355 | .cable_detect = ata_cable_40wire, | ||
374 | .set_piomode = amd33_set_piomode, | 356 | .set_piomode = amd33_set_piomode, |
375 | .set_dmamode = amd33_set_dmamode, | 357 | .set_dmamode = amd33_set_dmamode, |
376 | .mode_filter = ata_pci_default_filter, | ||
377 | .tf_load = ata_tf_load, | ||
378 | .tf_read = ata_tf_read, | ||
379 | .check_status = ata_check_status, | ||
380 | .exec_command = ata_exec_command, | ||
381 | .dev_select = ata_std_dev_select, | ||
382 | |||
383 | .freeze = ata_bmdma_freeze, | ||
384 | .thaw = ata_bmdma_thaw, | ||
385 | .error_handler = amd_error_handler, | ||
386 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
387 | .cable_detect = ata_cable_40wire, | ||
388 | |||
389 | .bmdma_setup = ata_bmdma_setup, | ||
390 | .bmdma_start = ata_bmdma_start, | ||
391 | .bmdma_stop = ata_bmdma_stop, | ||
392 | .bmdma_status = ata_bmdma_status, | ||
393 | |||
394 | .qc_prep = ata_qc_prep, | ||
395 | .qc_issue = ata_qc_issue_prot, | ||
396 | |||
397 | .data_xfer = ata_data_xfer, | ||
398 | |||
399 | .irq_handler = ata_interrupt, | ||
400 | .irq_clear = ata_bmdma_irq_clear, | ||
401 | .irq_on = ata_irq_on, | ||
402 | |||
403 | .port_start = ata_sff_port_start, | ||
404 | }; | 358 | }; |
405 | 359 | ||
406 | static struct ata_port_operations amd66_port_ops = { | 360 | static struct ata_port_operations amd66_port_ops = { |
361 | .inherits = &amd_base_port_ops, | ||
362 | .cable_detect = ata_cable_unknown, | ||
407 | .set_piomode = amd66_set_piomode, | 363 | .set_piomode = amd66_set_piomode, |
408 | .set_dmamode = amd66_set_dmamode, | 364 | .set_dmamode = amd66_set_dmamode, |
409 | .mode_filter = ata_pci_default_filter, | ||
410 | .tf_load = ata_tf_load, | ||
411 | .tf_read = ata_tf_read, | ||
412 | .check_status = ata_check_status, | ||
413 | .exec_command = ata_exec_command, | ||
414 | .dev_select = ata_std_dev_select, | ||
415 | |||
416 | .freeze = ata_bmdma_freeze, | ||
417 | .thaw = ata_bmdma_thaw, | ||
418 | .error_handler = amd_error_handler, | ||
419 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
420 | .cable_detect = ata_cable_unknown, | ||
421 | |||
422 | .bmdma_setup = ata_bmdma_setup, | ||
423 | .bmdma_start = ata_bmdma_start, | ||
424 | .bmdma_stop = ata_bmdma_stop, | ||
425 | .bmdma_status = ata_bmdma_status, | ||
426 | |||
427 | .qc_prep = ata_qc_prep, | ||
428 | .qc_issue = ata_qc_issue_prot, | ||
429 | |||
430 | .data_xfer = ata_data_xfer, | ||
431 | |||
432 | .irq_handler = ata_interrupt, | ||
433 | .irq_clear = ata_bmdma_irq_clear, | ||
434 | .irq_on = ata_irq_on, | ||
435 | |||
436 | .port_start = ata_sff_port_start, | ||
437 | }; | 365 | }; |
438 | 366 | ||
439 | static struct ata_port_operations amd100_port_ops = { | 367 | static struct ata_port_operations amd100_port_ops = { |
368 | .inherits = &amd_base_port_ops, | ||
369 | .cable_detect = ata_cable_unknown, | ||
440 | .set_piomode = amd100_set_piomode, | 370 | .set_piomode = amd100_set_piomode, |
441 | .set_dmamode = amd100_set_dmamode, | 371 | .set_dmamode = amd100_set_dmamode, |
442 | .mode_filter = ata_pci_default_filter, | ||
443 | .tf_load = ata_tf_load, | ||
444 | .tf_read = ata_tf_read, | ||
445 | .check_status = ata_check_status, | ||
446 | .exec_command = ata_exec_command, | ||
447 | .dev_select = ata_std_dev_select, | ||
448 | |||
449 | .freeze = ata_bmdma_freeze, | ||
450 | .thaw = ata_bmdma_thaw, | ||
451 | .error_handler = amd_error_handler, | ||
452 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
453 | .cable_detect = ata_cable_unknown, | ||
454 | |||
455 | .bmdma_setup = ata_bmdma_setup, | ||
456 | .bmdma_start = ata_bmdma_start, | ||
457 | .bmdma_stop = ata_bmdma_stop, | ||
458 | .bmdma_status = ata_bmdma_status, | ||
459 | |||
460 | .qc_prep = ata_qc_prep, | ||
461 | .qc_issue = ata_qc_issue_prot, | ||
462 | |||
463 | .data_xfer = ata_data_xfer, | ||
464 | |||
465 | .irq_handler = ata_interrupt, | ||
466 | .irq_clear = ata_bmdma_irq_clear, | ||
467 | .irq_on = ata_irq_on, | ||
468 | |||
469 | .port_start = ata_sff_port_start, | ||
470 | }; | 372 | }; |
471 | 373 | ||
472 | static struct ata_port_operations amd133_port_ops = { | 374 | static struct ata_port_operations amd133_port_ops = { |
375 | .inherits = &amd_base_port_ops, | ||
376 | .cable_detect = amd_cable_detect, | ||
473 | .set_piomode = amd133_set_piomode, | 377 | .set_piomode = amd133_set_piomode, |
474 | .set_dmamode = amd133_set_dmamode, | 378 | .set_dmamode = amd133_set_dmamode, |
475 | .mode_filter = ata_pci_default_filter, | 379 | }; |
476 | .tf_load = ata_tf_load, | ||
477 | .tf_read = ata_tf_read, | ||
478 | .check_status = ata_check_status, | ||
479 | .exec_command = ata_exec_command, | ||
480 | .dev_select = ata_std_dev_select, | ||
481 | |||
482 | .freeze = ata_bmdma_freeze, | ||
483 | .thaw = ata_bmdma_thaw, | ||
484 | .error_handler = amd_error_handler, | ||
485 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
486 | .cable_detect = amd_cable_detect, | ||
487 | |||
488 | .bmdma_setup = ata_bmdma_setup, | ||
489 | .bmdma_start = ata_bmdma_start, | ||
490 | .bmdma_stop = ata_bmdma_stop, | ||
491 | .bmdma_status = ata_bmdma_status, | ||
492 | |||
493 | .qc_prep = ata_qc_prep, | ||
494 | .qc_issue = ata_qc_issue_prot, | ||
495 | |||
496 | .data_xfer = ata_data_xfer, | ||
497 | |||
498 | .irq_handler = ata_interrupt, | ||
499 | .irq_clear = ata_bmdma_irq_clear, | ||
500 | .irq_on = ata_irq_on, | ||
501 | 380 | ||
502 | .port_start = ata_sff_port_start, | 381 | static const struct ata_port_operations nv_base_port_ops = { |
382 | .inherits = &ata_bmdma_port_ops, | ||
383 | .cable_detect = ata_cable_ignore, | ||
384 | .mode_filter = nv_mode_filter, | ||
385 | .prereset = nv_pre_reset, | ||
386 | .host_stop = nv_host_stop, | ||
503 | }; | 387 | }; |
504 | 388 | ||
505 | static struct ata_port_operations nv100_port_ops = { | 389 | static struct ata_port_operations nv100_port_ops = { |
390 | .inherits = &nv_base_port_ops, | ||
506 | .set_piomode = nv100_set_piomode, | 391 | .set_piomode = nv100_set_piomode, |
507 | .set_dmamode = nv100_set_dmamode, | 392 | .set_dmamode = nv100_set_dmamode, |
508 | .tf_load = ata_tf_load, | ||
509 | .tf_read = ata_tf_read, | ||
510 | .check_status = ata_check_status, | ||
511 | .exec_command = ata_exec_command, | ||
512 | .dev_select = ata_std_dev_select, | ||
513 | |||
514 | .freeze = ata_bmdma_freeze, | ||
515 | .thaw = ata_bmdma_thaw, | ||
516 | .error_handler = nv_error_handler, | ||
517 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
518 | .cable_detect = ata_cable_ignore, | ||
519 | .mode_filter = nv_mode_filter, | ||
520 | |||
521 | .bmdma_setup = ata_bmdma_setup, | ||
522 | .bmdma_start = ata_bmdma_start, | ||
523 | .bmdma_stop = ata_bmdma_stop, | ||
524 | .bmdma_status = ata_bmdma_status, | ||
525 | |||
526 | .qc_prep = ata_qc_prep, | ||
527 | .qc_issue = ata_qc_issue_prot, | ||
528 | |||
529 | .data_xfer = ata_data_xfer, | ||
530 | |||
531 | .irq_handler = ata_interrupt, | ||
532 | .irq_clear = ata_bmdma_irq_clear, | ||
533 | .irq_on = ata_irq_on, | ||
534 | |||
535 | .port_start = ata_sff_port_start, | ||
536 | .host_stop = nv_host_stop, | ||
537 | }; | 393 | }; |
538 | 394 | ||
539 | static struct ata_port_operations nv133_port_ops = { | 395 | static struct ata_port_operations nv133_port_ops = { |
396 | .inherits = &nv_base_port_ops, | ||
540 | .set_piomode = nv133_set_piomode, | 397 | .set_piomode = nv133_set_piomode, |
541 | .set_dmamode = nv133_set_dmamode, | 398 | .set_dmamode = nv133_set_dmamode, |
542 | .tf_load = ata_tf_load, | ||
543 | .tf_read = ata_tf_read, | ||
544 | .check_status = ata_check_status, | ||
545 | .exec_command = ata_exec_command, | ||
546 | .dev_select = ata_std_dev_select, | ||
547 | |||
548 | .freeze = ata_bmdma_freeze, | ||
549 | .thaw = ata_bmdma_thaw, | ||
550 | .error_handler = nv_error_handler, | ||
551 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
552 | .cable_detect = ata_cable_ignore, | ||
553 | .mode_filter = nv_mode_filter, | ||
554 | |||
555 | .bmdma_setup = ata_bmdma_setup, | ||
556 | .bmdma_start = ata_bmdma_start, | ||
557 | .bmdma_stop = ata_bmdma_stop, | ||
558 | .bmdma_status = ata_bmdma_status, | ||
559 | |||
560 | .qc_prep = ata_qc_prep, | ||
561 | .qc_issue = ata_qc_issue_prot, | ||
562 | |||
563 | .data_xfer = ata_data_xfer, | ||
564 | |||
565 | .irq_handler = ata_interrupt, | ||
566 | .irq_clear = ata_bmdma_irq_clear, | ||
567 | .irq_on = ata_irq_on, | ||
568 | |||
569 | .port_start = ata_sff_port_start, | ||
570 | .host_stop = nv_host_stop, | ||
571 | }; | 399 | }; |
572 | 400 | ||
573 | static int amd_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | 401 | static int amd_init_one(struct pci_dev *pdev, const struct pci_device_id *id) |
574 | { | 402 | { |
575 | static const struct ata_port_info info[10] = { | 403 | static const struct ata_port_info info[10] = { |
576 | { /* 0: AMD 7401 */ | 404 | { /* 0: AMD 7401 */ |
577 | .sht = &amd_sht, | ||
578 | .flags = ATA_FLAG_SLAVE_POSS, | 405 | .flags = ATA_FLAG_SLAVE_POSS, |
579 | .pio_mask = 0x1f, | 406 | .pio_mask = 0x1f, |
580 | .mwdma_mask = 0x07, /* No SWDMA */ | 407 | .mwdma_mask = 0x07, /* No SWDMA */ |
@@ -582,7 +409,6 @@ static int amd_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
582 | .port_ops = &amd33_port_ops | 409 | .port_ops = &amd33_port_ops |
583 | }, | 410 | }, |
584 | { /* 1: Early AMD7409 - no swdma */ | 411 | { /* 1: Early AMD7409 - no swdma */ |
585 | .sht = &amd_sht, | ||
586 | .flags = ATA_FLAG_SLAVE_POSS, | 412 | .flags = ATA_FLAG_SLAVE_POSS, |
587 | .pio_mask = 0x1f, | 413 | .pio_mask = 0x1f, |
588 | .mwdma_mask = 0x07, | 414 | .mwdma_mask = 0x07, |
@@ -590,7 +416,6 @@ static int amd_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
590 | .port_ops = &amd66_port_ops | 416 | .port_ops = &amd66_port_ops |
591 | }, | 417 | }, |
592 | { /* 2: AMD 7409, no swdma errata */ | 418 | { /* 2: AMD 7409, no swdma errata */ |
593 | .sht = &amd_sht, | ||
594 | .flags = ATA_FLAG_SLAVE_POSS, | 419 | .flags = ATA_FLAG_SLAVE_POSS, |
595 | .pio_mask = 0x1f, | 420 | .pio_mask = 0x1f, |
596 | .mwdma_mask = 0x07, | 421 | .mwdma_mask = 0x07, |
@@ -598,7 +423,6 @@ static int amd_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
598 | .port_ops = &amd66_port_ops | 423 | .port_ops = &amd66_port_ops |
599 | }, | 424 | }, |
600 | { /* 3: AMD 7411 */ | 425 | { /* 3: AMD 7411 */ |
601 | .sht = &amd_sht, | ||
602 | .flags = ATA_FLAG_SLAVE_POSS, | 426 | .flags = ATA_FLAG_SLAVE_POSS, |
603 | .pio_mask = 0x1f, | 427 | .pio_mask = 0x1f, |
604 | .mwdma_mask = 0x07, | 428 | .mwdma_mask = 0x07, |
@@ -606,7 +430,6 @@ static int amd_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
606 | .port_ops = &amd100_port_ops | 430 | .port_ops = &amd100_port_ops |
607 | }, | 431 | }, |
608 | { /* 4: AMD 7441 */ | 432 | { /* 4: AMD 7441 */ |
609 | .sht = &amd_sht, | ||
610 | .flags = ATA_FLAG_SLAVE_POSS, | 433 | .flags = ATA_FLAG_SLAVE_POSS, |
611 | .pio_mask = 0x1f, | 434 | .pio_mask = 0x1f, |
612 | .mwdma_mask = 0x07, | 435 | .mwdma_mask = 0x07, |
@@ -614,7 +437,6 @@ static int amd_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
614 | .port_ops = &amd100_port_ops | 437 | .port_ops = &amd100_port_ops |
615 | }, | 438 | }, |
616 | { /* 5: AMD 8111*/ | 439 | { /* 5: AMD 8111*/ |
617 | .sht = &amd_sht, | ||
618 | .flags = ATA_FLAG_SLAVE_POSS, | 440 | .flags = ATA_FLAG_SLAVE_POSS, |
619 | .pio_mask = 0x1f, | 441 | .pio_mask = 0x1f, |
620 | .mwdma_mask = 0x07, | 442 | .mwdma_mask = 0x07, |
@@ -622,7 +444,6 @@ static int amd_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
622 | .port_ops = &amd133_port_ops | 444 | .port_ops = &amd133_port_ops |
623 | }, | 445 | }, |
624 | { /* 6: AMD 8111 UDMA 100 (Serenade) */ | 446 | { /* 6: AMD 8111 UDMA 100 (Serenade) */ |
625 | .sht = &amd_sht, | ||
626 | .flags = ATA_FLAG_SLAVE_POSS, | 447 | .flags = ATA_FLAG_SLAVE_POSS, |
627 | .pio_mask = 0x1f, | 448 | .pio_mask = 0x1f, |
628 | .mwdma_mask = 0x07, | 449 | .mwdma_mask = 0x07, |
@@ -630,7 +451,6 @@ static int amd_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
630 | .port_ops = &amd133_port_ops | 451 | .port_ops = &amd133_port_ops |
631 | }, | 452 | }, |
632 | { /* 7: Nvidia Nforce */ | 453 | { /* 7: Nvidia Nforce */ |
633 | .sht = &amd_sht, | ||
634 | .flags = ATA_FLAG_SLAVE_POSS, | 454 | .flags = ATA_FLAG_SLAVE_POSS, |
635 | .pio_mask = 0x1f, | 455 | .pio_mask = 0x1f, |
636 | .mwdma_mask = 0x07, | 456 | .mwdma_mask = 0x07, |
@@ -638,7 +458,6 @@ static int amd_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
638 | .port_ops = &nv100_port_ops | 458 | .port_ops = &nv100_port_ops |
639 | }, | 459 | }, |
640 | { /* 8: Nvidia Nforce2 and later */ | 460 | { /* 8: Nvidia Nforce2 and later */ |
641 | .sht = &amd_sht, | ||
642 | .flags = ATA_FLAG_SLAVE_POSS, | 461 | .flags = ATA_FLAG_SLAVE_POSS, |
643 | .pio_mask = 0x1f, | 462 | .pio_mask = 0x1f, |
644 | .mwdma_mask = 0x07, | 463 | .mwdma_mask = 0x07, |
@@ -646,7 +465,6 @@ static int amd_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
646 | .port_ops = &nv133_port_ops | 465 | .port_ops = &nv133_port_ops |
647 | }, | 466 | }, |
648 | { /* 9: AMD CS5536 (Geode companion) */ | 467 | { /* 9: AMD CS5536 (Geode companion) */ |
649 | .sht = &amd_sht, | ||
650 | .flags = ATA_FLAG_SLAVE_POSS, | 468 | .flags = ATA_FLAG_SLAVE_POSS, |
651 | .pio_mask = 0x1f, | 469 | .pio_mask = 0x1f, |
652 | .mwdma_mask = 0x07, | 470 | .mwdma_mask = 0x07, |
@@ -654,15 +472,20 @@ static int amd_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
654 | .port_ops = &amd100_port_ops | 472 | .port_ops = &amd100_port_ops |
655 | } | 473 | } |
656 | }; | 474 | }; |
657 | struct ata_port_info pi; | 475 | const struct ata_port_info *ppi[] = { NULL, NULL }; |
658 | const struct ata_port_info *ppi[] = { &pi, NULL }; | ||
659 | static int printed_version; | 476 | static int printed_version; |
660 | int type = id->driver_data; | 477 | int type = id->driver_data; |
478 | void *hpriv = NULL; | ||
661 | u8 fifo; | 479 | u8 fifo; |
480 | int rc; | ||
662 | 481 | ||
663 | if (!printed_version++) | 482 | if (!printed_version++) |
664 | dev_printk(KERN_DEBUG, &pdev->dev, "version " DRV_VERSION "\n"); | 483 | dev_printk(KERN_DEBUG, &pdev->dev, "version " DRV_VERSION "\n"); |
665 | 484 | ||
485 | rc = pcim_enable_device(pdev); | ||
486 | if (rc) | ||
487 | return rc; | ||
488 | |||
666 | pci_read_config_byte(pdev, 0x41, &fifo); | 489 | pci_read_config_byte(pdev, 0x41, &fifo); |
667 | 490 | ||
668 | /* Check for AMD7409 without swdma errata and if found adjust type */ | 491 | /* Check for AMD7409 without swdma errata and if found adjust type */ |
@@ -677,10 +500,10 @@ static int amd_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
677 | /* | 500 | /* |
678 | * Okay, type is determined now. Apply type-specific workarounds. | 501 | * Okay, type is determined now. Apply type-specific workarounds. |
679 | */ | 502 | */ |
680 | pi = info[type]; | 503 | ppi[0] = &info[type]; |
681 | 504 | ||
682 | if (type < 3) | 505 | if (type < 3) |
683 | ata_pci_clear_simplex(pdev); | 506 | ata_pci_bmdma_clear_simplex(pdev); |
684 | 507 | ||
685 | /* Check for AMD7411 */ | 508 | /* Check for AMD7411 */ |
686 | if (type == 3) | 509 | if (type == 3) |
@@ -696,16 +519,23 @@ static int amd_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
696 | u32 udma; | 519 | u32 udma; |
697 | 520 | ||
698 | pci_read_config_dword(pdev, 0x60, &udma); | 521 | pci_read_config_dword(pdev, 0x60, &udma); |
699 | pi.private_data = (void *)(unsigned long)udma; | 522 | hpriv = (void *)(unsigned long)udma; |
700 | } | 523 | } |
701 | 524 | ||
702 | /* And fire it up */ | 525 | /* And fire it up */ |
703 | return ata_pci_init_one(pdev, ppi); | 526 | return ata_pci_sff_init_one(pdev, ppi, &amd_sht, hpriv); |
704 | } | 527 | } |
705 | 528 | ||
706 | #ifdef CONFIG_PM | 529 | #ifdef CONFIG_PM |
707 | static int amd_reinit_one(struct pci_dev *pdev) | 530 | static int amd_reinit_one(struct pci_dev *pdev) |
708 | { | 531 | { |
532 | struct ata_host *host = dev_get_drvdata(&pdev->dev); | ||
533 | int rc; | ||
534 | |||
535 | rc = ata_pci_device_do_resume(pdev); | ||
536 | if (rc) | ||
537 | return rc; | ||
538 | |||
709 | if (pdev->vendor == PCI_VENDOR_ID_AMD) { | 539 | if (pdev->vendor == PCI_VENDOR_ID_AMD) { |
710 | u8 fifo; | 540 | u8 fifo; |
711 | pci_read_config_byte(pdev, 0x41, &fifo); | 541 | pci_read_config_byte(pdev, 0x41, &fifo); |
@@ -716,9 +546,11 @@ static int amd_reinit_one(struct pci_dev *pdev) | |||
716 | pci_write_config_byte(pdev, 0x41, fifo | 0xF0); | 546 | pci_write_config_byte(pdev, 0x41, fifo | 0xF0); |
717 | if (pdev->device == PCI_DEVICE_ID_AMD_VIPER_7409 || | 547 | if (pdev->device == PCI_DEVICE_ID_AMD_VIPER_7409 || |
718 | pdev->device == PCI_DEVICE_ID_AMD_COBRA_7401) | 548 | pdev->device == PCI_DEVICE_ID_AMD_COBRA_7401) |
719 | ata_pci_clear_simplex(pdev); | 549 | ata_pci_bmdma_clear_simplex(pdev); |
720 | } | 550 | } |
721 | return ata_pci_device_resume(pdev); | 551 | |
552 | ata_host_resume(host); | ||
553 | return 0; | ||
722 | } | 554 | } |
723 | #endif | 555 | #endif |
724 | 556 | ||
diff --git a/drivers/ata/pata_artop.c b/drivers/ata/pata_artop.c index d4218310327b..0f513bc11193 100644 --- a/drivers/ata/pata_artop.c +++ b/drivers/ata/pata_artop.c | |||
@@ -52,22 +52,7 @@ static int artop6210_pre_reset(struct ata_link *link, unsigned long deadline) | |||
52 | if (!pci_test_config_bits(pdev, &artop_enable_bits[ap->port_no])) | 52 | if (!pci_test_config_bits(pdev, &artop_enable_bits[ap->port_no])) |
53 | return -ENOENT; | 53 | return -ENOENT; |
54 | 54 | ||
55 | return ata_std_prereset(link, deadline); | 55 | return ata_sff_prereset(link, deadline); |
56 | } | ||
57 | |||
58 | /** | ||
59 | * artop6210_error_handler - Probe specified port on PATA host controller | ||
60 | * @ap: Port to probe | ||
61 | * | ||
62 | * LOCKING: | ||
63 | * None (inherited from caller). | ||
64 | */ | ||
65 | |||
66 | static void artop6210_error_handler(struct ata_port *ap) | ||
67 | { | ||
68 | ata_bmdma_drive_eh(ap, artop6210_pre_reset, | ||
69 | ata_std_softreset, NULL, | ||
70 | ata_std_postreset); | ||
71 | } | 56 | } |
72 | 57 | ||
73 | /** | 58 | /** |
@@ -93,7 +78,7 @@ static int artop6260_pre_reset(struct ata_link *link, unsigned long deadline) | |||
93 | if (pdev->device % 1 && !pci_test_config_bits(pdev, &artop_enable_bits[ap->port_no])) | 78 | if (pdev->device % 1 && !pci_test_config_bits(pdev, &artop_enable_bits[ap->port_no])) |
94 | return -ENOENT; | 79 | return -ENOENT; |
95 | 80 | ||
96 | return ata_std_prereset(link, deadline); | 81 | return ata_sff_prereset(link, deadline); |
97 | } | 82 | } |
98 | 83 | ||
99 | /** | 84 | /** |
@@ -114,21 +99,6 @@ static int artop6260_cable_detect(struct ata_port *ap) | |||
114 | } | 99 | } |
115 | 100 | ||
116 | /** | 101 | /** |
117 | * artop6260_error_handler - Probe specified port on PATA host controller | ||
118 | * @ap: Port to probe | ||
119 | * | ||
120 | * LOCKING: | ||
121 | * None (inherited from caller). | ||
122 | */ | ||
123 | |||
124 | static void artop6260_error_handler(struct ata_port *ap) | ||
125 | { | ||
126 | ata_bmdma_drive_eh(ap, artop6260_pre_reset, | ||
127 | ata_std_softreset, NULL, | ||
128 | ata_std_postreset); | ||
129 | } | ||
130 | |||
131 | /** | ||
132 | * artop6210_load_piomode - Load a set of PATA PIO timings | 102 | * artop6210_load_piomode - Load a set of PATA PIO timings |
133 | * @ap: Port whose timings we are configuring | 103 | * @ap: Port whose timings we are configuring |
134 | * @adev: Device | 104 | * @adev: Device |
@@ -314,85 +284,23 @@ static void artop6260_set_dmamode (struct ata_port *ap, struct ata_device *adev) | |||
314 | } | 284 | } |
315 | 285 | ||
316 | static struct scsi_host_template artop_sht = { | 286 | static struct scsi_host_template artop_sht = { |
317 | .module = THIS_MODULE, | 287 | ATA_BMDMA_SHT(DRV_NAME), |
318 | .name = DRV_NAME, | ||
319 | .ioctl = ata_scsi_ioctl, | ||
320 | .queuecommand = ata_scsi_queuecmd, | ||
321 | .can_queue = ATA_DEF_QUEUE, | ||
322 | .this_id = ATA_SHT_THIS_ID, | ||
323 | .sg_tablesize = LIBATA_MAX_PRD, | ||
324 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
325 | .emulated = ATA_SHT_EMULATED, | ||
326 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
327 | .proc_name = DRV_NAME, | ||
328 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
329 | .slave_configure = ata_scsi_slave_config, | ||
330 | .slave_destroy = ata_scsi_slave_destroy, | ||
331 | .bios_param = ata_std_bios_param, | ||
332 | }; | 288 | }; |
333 | 289 | ||
334 | static const struct ata_port_operations artop6210_ops = { | 290 | static struct ata_port_operations artop6210_ops = { |
291 | .inherits = &ata_bmdma_port_ops, | ||
292 | .cable_detect = ata_cable_40wire, | ||
335 | .set_piomode = artop6210_set_piomode, | 293 | .set_piomode = artop6210_set_piomode, |
336 | .set_dmamode = artop6210_set_dmamode, | 294 | .set_dmamode = artop6210_set_dmamode, |
337 | .mode_filter = ata_pci_default_filter, | 295 | .prereset = artop6210_pre_reset, |
338 | |||
339 | .tf_load = ata_tf_load, | ||
340 | .tf_read = ata_tf_read, | ||
341 | .check_status = ata_check_status, | ||
342 | .exec_command = ata_exec_command, | ||
343 | .dev_select = ata_std_dev_select, | ||
344 | |||
345 | .freeze = ata_bmdma_freeze, | ||
346 | .thaw = ata_bmdma_thaw, | ||
347 | .error_handler = artop6210_error_handler, | ||
348 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
349 | .cable_detect = ata_cable_40wire, | ||
350 | |||
351 | .bmdma_setup = ata_bmdma_setup, | ||
352 | .bmdma_start = ata_bmdma_start, | ||
353 | .bmdma_stop = ata_bmdma_stop, | ||
354 | .bmdma_status = ata_bmdma_status, | ||
355 | .qc_prep = ata_qc_prep, | ||
356 | .qc_issue = ata_qc_issue_prot, | ||
357 | |||
358 | .data_xfer = ata_data_xfer, | ||
359 | |||
360 | .irq_handler = ata_interrupt, | ||
361 | .irq_clear = ata_bmdma_irq_clear, | ||
362 | .irq_on = ata_irq_on, | ||
363 | |||
364 | .port_start = ata_sff_port_start, | ||
365 | }; | 296 | }; |
366 | 297 | ||
367 | static const struct ata_port_operations artop6260_ops = { | 298 | static struct ata_port_operations artop6260_ops = { |
299 | .inherits = &ata_bmdma_port_ops, | ||
300 | .cable_detect = artop6260_cable_detect, | ||
368 | .set_piomode = artop6260_set_piomode, | 301 | .set_piomode = artop6260_set_piomode, |
369 | .set_dmamode = artop6260_set_dmamode, | 302 | .set_dmamode = artop6260_set_dmamode, |
370 | 303 | .prereset = artop6260_pre_reset, | |
371 | .tf_load = ata_tf_load, | ||
372 | .tf_read = ata_tf_read, | ||
373 | .check_status = ata_check_status, | ||
374 | .exec_command = ata_exec_command, | ||
375 | .dev_select = ata_std_dev_select, | ||
376 | |||
377 | .freeze = ata_bmdma_freeze, | ||
378 | .thaw = ata_bmdma_thaw, | ||
379 | .error_handler = artop6260_error_handler, | ||
380 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
381 | .cable_detect = artop6260_cable_detect, | ||
382 | |||
383 | .bmdma_setup = ata_bmdma_setup, | ||
384 | .bmdma_start = ata_bmdma_start, | ||
385 | .bmdma_stop = ata_bmdma_stop, | ||
386 | .bmdma_status = ata_bmdma_status, | ||
387 | .qc_prep = ata_qc_prep, | ||
388 | .qc_issue = ata_qc_issue_prot, | ||
389 | .data_xfer = ata_data_xfer, | ||
390 | |||
391 | .irq_handler = ata_interrupt, | ||
392 | .irq_clear = ata_bmdma_irq_clear, | ||
393 | .irq_on = ata_irq_on, | ||
394 | |||
395 | .port_start = ata_sff_port_start, | ||
396 | }; | 304 | }; |
397 | 305 | ||
398 | 306 | ||
@@ -414,7 +322,6 @@ static int artop_init_one (struct pci_dev *pdev, const struct pci_device_id *id) | |||
414 | { | 322 | { |
415 | static int printed_version; | 323 | static int printed_version; |
416 | static const struct ata_port_info info_6210 = { | 324 | static const struct ata_port_info info_6210 = { |
417 | .sht = &artop_sht, | ||
418 | .flags = ATA_FLAG_SLAVE_POSS, | 325 | .flags = ATA_FLAG_SLAVE_POSS, |
419 | .pio_mask = 0x1f, /* pio0-4 */ | 326 | .pio_mask = 0x1f, /* pio0-4 */ |
420 | .mwdma_mask = 0x07, /* mwdma0-2 */ | 327 | .mwdma_mask = 0x07, /* mwdma0-2 */ |
@@ -422,7 +329,6 @@ static int artop_init_one (struct pci_dev *pdev, const struct pci_device_id *id) | |||
422 | .port_ops = &artop6210_ops, | 329 | .port_ops = &artop6210_ops, |
423 | }; | 330 | }; |
424 | static const struct ata_port_info info_626x = { | 331 | static const struct ata_port_info info_626x = { |
425 | .sht = &artop_sht, | ||
426 | .flags = ATA_FLAG_SLAVE_POSS, | 332 | .flags = ATA_FLAG_SLAVE_POSS, |
427 | .pio_mask = 0x1f, /* pio0-4 */ | 333 | .pio_mask = 0x1f, /* pio0-4 */ |
428 | .mwdma_mask = 0x07, /* mwdma0-2 */ | 334 | .mwdma_mask = 0x07, /* mwdma0-2 */ |
@@ -430,7 +336,6 @@ static int artop_init_one (struct pci_dev *pdev, const struct pci_device_id *id) | |||
430 | .port_ops = &artop6260_ops, | 336 | .port_ops = &artop6260_ops, |
431 | }; | 337 | }; |
432 | static const struct ata_port_info info_628x = { | 338 | static const struct ata_port_info info_628x = { |
433 | .sht = &artop_sht, | ||
434 | .flags = ATA_FLAG_SLAVE_POSS, | 339 | .flags = ATA_FLAG_SLAVE_POSS, |
435 | .pio_mask = 0x1f, /* pio0-4 */ | 340 | .pio_mask = 0x1f, /* pio0-4 */ |
436 | .mwdma_mask = 0x07, /* mwdma0-2 */ | 341 | .mwdma_mask = 0x07, /* mwdma0-2 */ |
@@ -438,7 +343,6 @@ static int artop_init_one (struct pci_dev *pdev, const struct pci_device_id *id) | |||
438 | .port_ops = &artop6260_ops, | 343 | .port_ops = &artop6260_ops, |
439 | }; | 344 | }; |
440 | static const struct ata_port_info info_628x_fast = { | 345 | static const struct ata_port_info info_628x_fast = { |
441 | .sht = &artop_sht, | ||
442 | .flags = ATA_FLAG_SLAVE_POSS, | 346 | .flags = ATA_FLAG_SLAVE_POSS, |
443 | .pio_mask = 0x1f, /* pio0-4 */ | 347 | .pio_mask = 0x1f, /* pio0-4 */ |
444 | .mwdma_mask = 0x07, /* mwdma0-2 */ | 348 | .mwdma_mask = 0x07, /* mwdma0-2 */ |
@@ -446,11 +350,16 @@ static int artop_init_one (struct pci_dev *pdev, const struct pci_device_id *id) | |||
446 | .port_ops = &artop6260_ops, | 350 | .port_ops = &artop6260_ops, |
447 | }; | 351 | }; |
448 | const struct ata_port_info *ppi[] = { NULL, NULL }; | 352 | const struct ata_port_info *ppi[] = { NULL, NULL }; |
353 | int rc; | ||
449 | 354 | ||
450 | if (!printed_version++) | 355 | if (!printed_version++) |
451 | dev_printk(KERN_DEBUG, &pdev->dev, | 356 | dev_printk(KERN_DEBUG, &pdev->dev, |
452 | "version " DRV_VERSION "\n"); | 357 | "version " DRV_VERSION "\n"); |
453 | 358 | ||
359 | rc = pcim_enable_device(pdev); | ||
360 | if (rc) | ||
361 | return rc; | ||
362 | |||
454 | if (id->driver_data == 0) { /* 6210 variant */ | 363 | if (id->driver_data == 0) { /* 6210 variant */ |
455 | ppi[0] = &info_6210; | 364 | ppi[0] = &info_6210; |
456 | ppi[1] = &ata_dummy_port_info; | 365 | ppi[1] = &ata_dummy_port_info; |
@@ -491,7 +400,7 @@ static int artop_init_one (struct pci_dev *pdev, const struct pci_device_id *id) | |||
491 | 400 | ||
492 | BUG_ON(ppi[0] == NULL); | 401 | BUG_ON(ppi[0] == NULL); |
493 | 402 | ||
494 | return ata_pci_init_one(pdev, ppi); | 403 | return ata_pci_sff_init_one(pdev, ppi, &artop_sht, NULL); |
495 | } | 404 | } |
496 | 405 | ||
497 | static const struct pci_device_id artop_pci_tbl[] = { | 406 | static const struct pci_device_id artop_pci_tbl[] = { |
diff --git a/drivers/ata/pata_at32.c b/drivers/ata/pata_at32.c index db057b183d60..3e8651d78952 100644 --- a/drivers/ata/pata_at32.c +++ b/drivers/ata/pata_at32.c | |||
@@ -166,52 +166,14 @@ static void pata_at32_set_piomode(struct ata_port *ap, struct ata_device *adev) | |||
166 | } | 166 | } |
167 | } | 167 | } |
168 | 168 | ||
169 | static void pata_at32_irq_clear(struct ata_port *ap) | ||
170 | { | ||
171 | /* No DMA controller yet */ | ||
172 | } | ||
173 | |||
174 | static struct scsi_host_template at32_sht = { | 169 | static struct scsi_host_template at32_sht = { |
175 | .module = THIS_MODULE, | 170 | ATA_PIO_SHT(DRV_NAME), |
176 | .name = DRV_NAME, | ||
177 | .ioctl = ata_scsi_ioctl, | ||
178 | .queuecommand = ata_scsi_queuecmd, | ||
179 | .can_queue = ATA_DEF_QUEUE, | ||
180 | .this_id = ATA_SHT_THIS_ID, | ||
181 | .sg_tablesize = LIBATA_MAX_PRD, | ||
182 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
183 | .emulated = ATA_SHT_EMULATED, | ||
184 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
185 | .proc_name = DRV_NAME, | ||
186 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
187 | .slave_configure = ata_scsi_slave_config, | ||
188 | .slave_destroy = ata_scsi_slave_destroy, | ||
189 | .bios_param = ata_std_bios_param, | ||
190 | }; | 171 | }; |
191 | 172 | ||
192 | static struct ata_port_operations at32_port_ops = { | 173 | static struct ata_port_operations at32_port_ops = { |
193 | .set_piomode = pata_at32_set_piomode, | 174 | .inherits = &ata_sff_port_ops, |
194 | .tf_load = ata_tf_load, | ||
195 | .tf_read = ata_tf_read, | ||
196 | .exec_command = ata_exec_command, | ||
197 | .check_status = ata_check_status, | ||
198 | .dev_select = ata_std_dev_select, | ||
199 | |||
200 | .freeze = ata_bmdma_freeze, | ||
201 | .thaw = ata_bmdma_thaw, | ||
202 | .error_handler = ata_bmdma_error_handler, | ||
203 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
204 | .cable_detect = ata_cable_40wire, | 175 | .cable_detect = ata_cable_40wire, |
205 | 176 | .set_piomode = pata_at32_set_piomode, | |
206 | .qc_prep = ata_qc_prep, | ||
207 | .qc_issue = ata_qc_issue_prot, | ||
208 | |||
209 | .data_xfer = ata_data_xfer, | ||
210 | |||
211 | .irq_clear = pata_at32_irq_clear, | ||
212 | .irq_on = ata_irq_on, | ||
213 | |||
214 | .port_start = ata_sff_port_start, | ||
215 | }; | 177 | }; |
216 | 178 | ||
217 | static int __init pata_at32_init_one(struct device *dev, | 179 | static int __init pata_at32_init_one(struct device *dev, |
@@ -261,7 +223,7 @@ static int __init pata_at32_init_one(struct device *dev, | |||
261 | host->private_data = info; | 223 | host->private_data = info; |
262 | 224 | ||
263 | /* Register ATA device and return */ | 225 | /* Register ATA device and return */ |
264 | return ata_host_activate(host, info->irq, ata_interrupt, | 226 | return ata_host_activate(host, info->irq, ata_sff_interrupt, |
265 | IRQF_SHARED | IRQF_TRIGGER_RISING, | 227 | IRQF_SHARED | IRQF_TRIGGER_RISING, |
266 | &at32_sht); | 228 | &at32_sht); |
267 | } | 229 | } |
diff --git a/drivers/ata/pata_atiixp.c b/drivers/ata/pata_atiixp.c index 408bdc1a9776..78738fb4223b 100644 --- a/drivers/ata/pata_atiixp.c +++ b/drivers/ata/pata_atiixp.c | |||
@@ -45,12 +45,7 @@ static int atiixp_pre_reset(struct ata_link *link, unsigned long deadline) | |||
45 | if (!pci_test_config_bits(pdev, &atiixp_enable_bits[ap->port_no])) | 45 | if (!pci_test_config_bits(pdev, &atiixp_enable_bits[ap->port_no])) |
46 | return -ENOENT; | 46 | return -ENOENT; |
47 | 47 | ||
48 | return ata_std_prereset(link, deadline); | 48 | return ata_sff_prereset(link, deadline); |
49 | } | ||
50 | |||
51 | static void atiixp_error_handler(struct ata_port *ap) | ||
52 | { | ||
53 | ata_bmdma_drive_eh(ap, atiixp_pre_reset, ata_std_softreset, NULL, ata_std_postreset); | ||
54 | } | 49 | } |
55 | 50 | ||
56 | static int atiixp_cable_detect(struct ata_port *ap) | 51 | static int atiixp_cable_detect(struct ata_port *ap) |
@@ -221,60 +216,26 @@ static void atiixp_bmdma_stop(struct ata_queued_cmd *qc) | |||
221 | } | 216 | } |
222 | 217 | ||
223 | static struct scsi_host_template atiixp_sht = { | 218 | static struct scsi_host_template atiixp_sht = { |
224 | .module = THIS_MODULE, | 219 | ATA_BMDMA_SHT(DRV_NAME), |
225 | .name = DRV_NAME, | ||
226 | .ioctl = ata_scsi_ioctl, | ||
227 | .queuecommand = ata_scsi_queuecmd, | ||
228 | .can_queue = ATA_DEF_QUEUE, | ||
229 | .this_id = ATA_SHT_THIS_ID, | ||
230 | .sg_tablesize = LIBATA_DUMB_MAX_PRD, | 220 | .sg_tablesize = LIBATA_DUMB_MAX_PRD, |
231 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
232 | .emulated = ATA_SHT_EMULATED, | ||
233 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
234 | .proc_name = DRV_NAME, | ||
235 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
236 | .slave_configure = ata_scsi_slave_config, | ||
237 | .slave_destroy = ata_scsi_slave_destroy, | ||
238 | .bios_param = ata_std_bios_param, | ||
239 | }; | 221 | }; |
240 | 222 | ||
241 | static struct ata_port_operations atiixp_port_ops = { | 223 | static struct ata_port_operations atiixp_port_ops = { |
242 | .set_piomode = atiixp_set_piomode, | 224 | .inherits = &ata_bmdma_port_ops, |
243 | .set_dmamode = atiixp_set_dmamode, | ||
244 | .mode_filter = ata_pci_default_filter, | ||
245 | .tf_load = ata_tf_load, | ||
246 | .tf_read = ata_tf_read, | ||
247 | .check_status = ata_check_status, | ||
248 | .exec_command = ata_exec_command, | ||
249 | .dev_select = ata_std_dev_select, | ||
250 | |||
251 | .freeze = ata_bmdma_freeze, | ||
252 | .thaw = ata_bmdma_thaw, | ||
253 | .error_handler = atiixp_error_handler, | ||
254 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
255 | .cable_detect = atiixp_cable_detect, | ||
256 | 225 | ||
257 | .bmdma_setup = ata_bmdma_setup, | 226 | .qc_prep = ata_sff_dumb_qc_prep, |
258 | .bmdma_start = atiixp_bmdma_start, | 227 | .bmdma_start = atiixp_bmdma_start, |
259 | .bmdma_stop = atiixp_bmdma_stop, | 228 | .bmdma_stop = atiixp_bmdma_stop, |
260 | .bmdma_status = ata_bmdma_status, | ||
261 | 229 | ||
262 | .qc_prep = ata_dumb_qc_prep, | 230 | .cable_detect = atiixp_cable_detect, |
263 | .qc_issue = ata_qc_issue_prot, | 231 | .set_piomode = atiixp_set_piomode, |
264 | 232 | .set_dmamode = atiixp_set_dmamode, | |
265 | .data_xfer = ata_data_xfer, | 233 | .prereset = atiixp_pre_reset, |
266 | |||
267 | .irq_handler = ata_interrupt, | ||
268 | .irq_clear = ata_bmdma_irq_clear, | ||
269 | .irq_on = ata_irq_on, | ||
270 | |||
271 | .port_start = ata_sff_port_start, | ||
272 | }; | 234 | }; |
273 | 235 | ||
274 | static int atiixp_init_one(struct pci_dev *dev, const struct pci_device_id *id) | 236 | static int atiixp_init_one(struct pci_dev *dev, const struct pci_device_id *id) |
275 | { | 237 | { |
276 | static const struct ata_port_info info = { | 238 | static const struct ata_port_info info = { |
277 | .sht = &atiixp_sht, | ||
278 | .flags = ATA_FLAG_SLAVE_POSS, | 239 | .flags = ATA_FLAG_SLAVE_POSS, |
279 | .pio_mask = 0x1f, | 240 | .pio_mask = 0x1f, |
280 | .mwdma_mask = 0x06, /* No MWDMA0 support */ | 241 | .mwdma_mask = 0x06, /* No MWDMA0 support */ |
@@ -282,7 +243,7 @@ static int atiixp_init_one(struct pci_dev *dev, const struct pci_device_id *id) | |||
282 | .port_ops = &atiixp_port_ops | 243 | .port_ops = &atiixp_port_ops |
283 | }; | 244 | }; |
284 | const struct ata_port_info *ppi[] = { &info, NULL }; | 245 | const struct ata_port_info *ppi[] = { &info, NULL }; |
285 | return ata_pci_init_one(dev, ppi); | 246 | return ata_pci_sff_init_one(dev, ppi, &atiixp_sht, NULL); |
286 | } | 247 | } |
287 | 248 | ||
288 | static const struct pci_device_id atiixp[] = { | 249 | static const struct pci_device_id atiixp[] = { |
diff --git a/drivers/ata/pata_bf54x.c b/drivers/ata/pata_bf54x.c index 7f87f105c2f6..0a5ad98635b1 100644 --- a/drivers/ata/pata_bf54x.c +++ b/drivers/ata/pata_bf54x.c | |||
@@ -674,7 +674,7 @@ static void read_atapi_data(void __iomem *base, | |||
674 | * @ap: Port to which output is sent | 674 | * @ap: Port to which output is sent |
675 | * @tf: ATA taskfile register set | 675 | * @tf: ATA taskfile register set |
676 | * | 676 | * |
677 | * Note: Original code is ata_tf_load(). | 677 | * Note: Original code is ata_sff_tf_load(). |
678 | */ | 678 | */ |
679 | 679 | ||
680 | static void bfin_tf_load(struct ata_port *ap, const struct ata_taskfile *tf) | 680 | static void bfin_tf_load(struct ata_port *ap, const struct ata_taskfile *tf) |
@@ -745,7 +745,7 @@ static u8 bfin_check_status(struct ata_port *ap) | |||
745 | * @ap: Port from which input is read | 745 | * @ap: Port from which input is read |
746 | * @tf: ATA taskfile register set for storing input | 746 | * @tf: ATA taskfile register set for storing input |
747 | * | 747 | * |
748 | * Note: Original code is ata_tf_read(). | 748 | * Note: Original code is ata_sff_tf_read(). |
749 | */ | 749 | */ |
750 | 750 | ||
751 | static void bfin_tf_read(struct ata_port *ap, struct ata_taskfile *tf) | 751 | static void bfin_tf_read(struct ata_port *ap, struct ata_taskfile *tf) |
@@ -775,7 +775,7 @@ static void bfin_tf_read(struct ata_port *ap, struct ata_taskfile *tf) | |||
775 | * @ap: port to which command is being issued | 775 | * @ap: port to which command is being issued |
776 | * @tf: ATA taskfile register set | 776 | * @tf: ATA taskfile register set |
777 | * | 777 | * |
778 | * Note: Original code is ata_exec_command(). | 778 | * Note: Original code is ata_sff_exec_command(). |
779 | */ | 779 | */ |
780 | 780 | ||
781 | static void bfin_exec_command(struct ata_port *ap, | 781 | static void bfin_exec_command(struct ata_port *ap, |
@@ -785,7 +785,7 @@ static void bfin_exec_command(struct ata_port *ap, | |||
785 | dev_dbg(ap->dev, "ata%u: cmd 0x%X\n", ap->print_id, tf->command); | 785 | dev_dbg(ap->dev, "ata%u: cmd 0x%X\n", ap->print_id, tf->command); |
786 | 786 | ||
787 | write_atapi_register(base, ATA_REG_CMD, tf->command); | 787 | write_atapi_register(base, ATA_REG_CMD, tf->command); |
788 | ata_pause(ap); | 788 | ata_sff_pause(ap); |
789 | } | 789 | } |
790 | 790 | ||
791 | /** | 791 | /** |
@@ -800,14 +800,14 @@ static u8 bfin_check_altstatus(struct ata_port *ap) | |||
800 | } | 800 | } |
801 | 801 | ||
802 | /** | 802 | /** |
803 | * bfin_std_dev_select - Select device 0/1 on ATA bus | 803 | * bfin_dev_select - Select device 0/1 on ATA bus |
804 | * @ap: ATA channel to manipulate | 804 | * @ap: ATA channel to manipulate |
805 | * @device: ATA device (numbered from zero) to select | 805 | * @device: ATA device (numbered from zero) to select |
806 | * | 806 | * |
807 | * Note: Original code is ata_std_dev_select(). | 807 | * Note: Original code is ata_sff_dev_select(). |
808 | */ | 808 | */ |
809 | 809 | ||
810 | static void bfin_std_dev_select(struct ata_port *ap, unsigned int device) | 810 | static void bfin_dev_select(struct ata_port *ap, unsigned int device) |
811 | { | 811 | { |
812 | void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr; | 812 | void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr; |
813 | u8 tmp; | 813 | u8 tmp; |
@@ -818,7 +818,7 @@ static void bfin_std_dev_select(struct ata_port *ap, unsigned int device) | |||
818 | tmp = ATA_DEVICE_OBS | ATA_DEV1; | 818 | tmp = ATA_DEVICE_OBS | ATA_DEV1; |
819 | 819 | ||
820 | write_atapi_register(base, ATA_REG_DEVICE, tmp); | 820 | write_atapi_register(base, ATA_REG_DEVICE, tmp); |
821 | ata_pause(ap); | 821 | ata_sff_pause(ap); |
822 | } | 822 | } |
823 | 823 | ||
824 | /** | 824 | /** |
@@ -977,7 +977,7 @@ static unsigned int bfin_devchk(struct ata_port *ap, | |||
977 | void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr; | 977 | void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr; |
978 | u8 nsect, lbal; | 978 | u8 nsect, lbal; |
979 | 979 | ||
980 | bfin_std_dev_select(ap, device); | 980 | bfin_dev_select(ap, device); |
981 | 981 | ||
982 | write_atapi_register(base, ATA_REG_NSECT, 0x55); | 982 | write_atapi_register(base, ATA_REG_NSECT, 0x55); |
983 | write_atapi_register(base, ATA_REG_LBAL, 0xaa); | 983 | write_atapi_register(base, ATA_REG_LBAL, 0xaa); |
@@ -1014,7 +1014,7 @@ static void bfin_bus_post_reset(struct ata_port *ap, unsigned int devmask) | |||
1014 | * BSY bit to clear | 1014 | * BSY bit to clear |
1015 | */ | 1015 | */ |
1016 | if (dev0) | 1016 | if (dev0) |
1017 | ata_busy_sleep(ap, ATA_TMOUT_BOOT_QUICK, ATA_TMOUT_BOOT); | 1017 | ata_sff_busy_sleep(ap, ATA_TMOUT_BOOT_QUICK, ATA_TMOUT_BOOT); |
1018 | 1018 | ||
1019 | /* if device 1 was found in ata_devchk, wait for | 1019 | /* if device 1 was found in ata_devchk, wait for |
1020 | * register access, then wait for BSY to clear | 1020 | * register access, then wait for BSY to clear |
@@ -1023,7 +1023,7 @@ static void bfin_bus_post_reset(struct ata_port *ap, unsigned int devmask) | |||
1023 | while (dev1) { | 1023 | while (dev1) { |
1024 | u8 nsect, lbal; | 1024 | u8 nsect, lbal; |
1025 | 1025 | ||
1026 | bfin_std_dev_select(ap, 1); | 1026 | bfin_dev_select(ap, 1); |
1027 | nsect = read_atapi_register(base, ATA_REG_NSECT); | 1027 | nsect = read_atapi_register(base, ATA_REG_NSECT); |
1028 | lbal = read_atapi_register(base, ATA_REG_LBAL); | 1028 | lbal = read_atapi_register(base, ATA_REG_LBAL); |
1029 | if ((nsect == 1) && (lbal == 1)) | 1029 | if ((nsect == 1) && (lbal == 1)) |
@@ -1035,14 +1035,14 @@ static void bfin_bus_post_reset(struct ata_port *ap, unsigned int devmask) | |||
1035 | msleep(50); /* give drive a breather */ | 1035 | msleep(50); /* give drive a breather */ |
1036 | } | 1036 | } |
1037 | if (dev1) | 1037 | if (dev1) |
1038 | ata_busy_sleep(ap, ATA_TMOUT_BOOT_QUICK, ATA_TMOUT_BOOT); | 1038 | ata_sff_busy_sleep(ap, ATA_TMOUT_BOOT_QUICK, ATA_TMOUT_BOOT); |
1039 | 1039 | ||
1040 | /* is all this really necessary? */ | 1040 | /* is all this really necessary? */ |
1041 | bfin_std_dev_select(ap, 0); | 1041 | bfin_dev_select(ap, 0); |
1042 | if (dev1) | 1042 | if (dev1) |
1043 | bfin_std_dev_select(ap, 1); | 1043 | bfin_dev_select(ap, 1); |
1044 | if (dev0) | 1044 | if (dev0) |
1045 | bfin_std_dev_select(ap, 0); | 1045 | bfin_dev_select(ap, 0); |
1046 | } | 1046 | } |
1047 | 1047 | ||
1048 | /** | 1048 | /** |
@@ -1088,26 +1088,21 @@ static unsigned int bfin_bus_softreset(struct ata_port *ap, | |||
1088 | } | 1088 | } |
1089 | 1089 | ||
1090 | /** | 1090 | /** |
1091 | * bfin_std_softreset - reset host port via ATA SRST | 1091 | * bfin_softreset - reset host port via ATA SRST |
1092 | * @ap: port to reset | 1092 | * @ap: port to reset |
1093 | * @classes: resulting classes of attached devices | 1093 | * @classes: resulting classes of attached devices |
1094 | * | 1094 | * |
1095 | * Note: Original code is ata_std_softreset(). | 1095 | * Note: Original code is ata_sff_softreset(). |
1096 | */ | 1096 | */ |
1097 | 1097 | ||
1098 | static int bfin_std_softreset(struct ata_link *link, unsigned int *classes, | 1098 | static int bfin_softreset(struct ata_link *link, unsigned int *classes, |
1099 | unsigned long deadline) | 1099 | unsigned long deadline) |
1100 | { | 1100 | { |
1101 | struct ata_port *ap = link->ap; | 1101 | struct ata_port *ap = link->ap; |
1102 | unsigned int slave_possible = ap->flags & ATA_FLAG_SLAVE_POSS; | 1102 | unsigned int slave_possible = ap->flags & ATA_FLAG_SLAVE_POSS; |
1103 | unsigned int devmask = 0, err_mask; | 1103 | unsigned int devmask = 0, err_mask; |
1104 | u8 err; | 1104 | u8 err; |
1105 | 1105 | ||
1106 | if (ata_link_offline(link)) { | ||
1107 | classes[0] = ATA_DEV_NONE; | ||
1108 | goto out; | ||
1109 | } | ||
1110 | |||
1111 | /* determine if device 0/1 are present */ | 1106 | /* determine if device 0/1 are present */ |
1112 | if (bfin_devchk(ap, 0)) | 1107 | if (bfin_devchk(ap, 0)) |
1113 | devmask |= (1 << 0); | 1108 | devmask |= (1 << 0); |
@@ -1115,7 +1110,7 @@ static int bfin_std_softreset(struct ata_link *link, unsigned int *classes, | |||
1115 | devmask |= (1 << 1); | 1110 | devmask |= (1 << 1); |
1116 | 1111 | ||
1117 | /* select device 0 again */ | 1112 | /* select device 0 again */ |
1118 | bfin_std_dev_select(ap, 0); | 1113 | bfin_dev_select(ap, 0); |
1119 | 1114 | ||
1120 | /* issue bus reset */ | 1115 | /* issue bus reset */ |
1121 | err_mask = bfin_bus_softreset(ap, devmask); | 1116 | err_mask = bfin_bus_softreset(ap, devmask); |
@@ -1126,13 +1121,12 @@ static int bfin_std_softreset(struct ata_link *link, unsigned int *classes, | |||
1126 | } | 1121 | } |
1127 | 1122 | ||
1128 | /* determine by signature whether we have ATA or ATAPI devices */ | 1123 | /* determine by signature whether we have ATA or ATAPI devices */ |
1129 | classes[0] = ata_dev_try_classify(&ap->link.device[0], | 1124 | classes[0] = ata_sff_dev_classify(&ap->link.device[0], |
1130 | devmask & (1 << 0), &err); | 1125 | devmask & (1 << 0), &err); |
1131 | if (slave_possible && err != 0x81) | 1126 | if (slave_possible && err != 0x81) |
1132 | classes[1] = ata_dev_try_classify(&ap->link.device[1], | 1127 | classes[1] = ata_sff_dev_classify(&ap->link.device[1], |
1133 | devmask & (1 << 1), &err); | 1128 | devmask & (1 << 1), &err); |
1134 | 1129 | ||
1135 | out: | ||
1136 | return 0; | 1130 | return 0; |
1137 | } | 1131 | } |
1138 | 1132 | ||
@@ -1167,7 +1161,7 @@ static unsigned char bfin_bmdma_status(struct ata_port *ap) | |||
1167 | * @buflen: buffer length | 1161 | * @buflen: buffer length |
1168 | * @write_data: read/write | 1162 | * @write_data: read/write |
1169 | * | 1163 | * |
1170 | * Note: Original code is ata_data_xfer(). | 1164 | * Note: Original code is ata_sff_data_xfer(). |
1171 | */ | 1165 | */ |
1172 | 1166 | ||
1173 | static unsigned int bfin_data_xfer(struct ata_device *dev, unsigned char *buf, | 1167 | static unsigned int bfin_data_xfer(struct ata_device *dev, unsigned char *buf, |
@@ -1206,7 +1200,7 @@ static unsigned int bfin_data_xfer(struct ata_device *dev, unsigned char *buf, | |||
1206 | * bfin_irq_clear - Clear ATAPI interrupt. | 1200 | * bfin_irq_clear - Clear ATAPI interrupt. |
1207 | * @ap: Port associated with this ATA transaction. | 1201 | * @ap: Port associated with this ATA transaction. |
1208 | * | 1202 | * |
1209 | * Note: Original code is ata_bmdma_irq_clear(). | 1203 | * Note: Original code is ata_sff_irq_clear(). |
1210 | */ | 1204 | */ |
1211 | 1205 | ||
1212 | static void bfin_irq_clear(struct ata_port *ap) | 1206 | static void bfin_irq_clear(struct ata_port *ap) |
@@ -1223,7 +1217,7 @@ static void bfin_irq_clear(struct ata_port *ap) | |||
1223 | * bfin_irq_on - Enable interrupts on a port. | 1217 | * bfin_irq_on - Enable interrupts on a port. |
1224 | * @ap: Port on which interrupts are enabled. | 1218 | * @ap: Port on which interrupts are enabled. |
1225 | * | 1219 | * |
1226 | * Note: Original code is ata_irq_on(). | 1220 | * Note: Original code is ata_sff_irq_on(). |
1227 | */ | 1221 | */ |
1228 | 1222 | ||
1229 | static unsigned char bfin_irq_on(struct ata_port *ap) | 1223 | static unsigned char bfin_irq_on(struct ata_port *ap) |
@@ -1244,13 +1238,13 @@ static unsigned char bfin_irq_on(struct ata_port *ap) | |||
1244 | } | 1238 | } |
1245 | 1239 | ||
1246 | /** | 1240 | /** |
1247 | * bfin_bmdma_freeze - Freeze DMA controller port | 1241 | * bfin_freeze - Freeze DMA controller port |
1248 | * @ap: port to freeze | 1242 | * @ap: port to freeze |
1249 | * | 1243 | * |
1250 | * Note: Original code is ata_bmdma_freeze(). | 1244 | * Note: Original code is ata_sff_freeze(). |
1251 | */ | 1245 | */ |
1252 | 1246 | ||
1253 | static void bfin_bmdma_freeze(struct ata_port *ap) | 1247 | static void bfin_freeze(struct ata_port *ap) |
1254 | { | 1248 | { |
1255 | void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr; | 1249 | void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr; |
1256 | 1250 | ||
@@ -1264,19 +1258,19 @@ static void bfin_bmdma_freeze(struct ata_port *ap) | |||
1264 | * ATA_NIEN manipulation. Also, many controllers fail to mask | 1258 | * ATA_NIEN manipulation. Also, many controllers fail to mask |
1265 | * previously pending IRQ on ATA_NIEN assertion. Clear it. | 1259 | * previously pending IRQ on ATA_NIEN assertion. Clear it. |
1266 | */ | 1260 | */ |
1267 | ata_chk_status(ap); | 1261 | ap->ops->sff_check_status(ap); |
1268 | 1262 | ||
1269 | bfin_irq_clear(ap); | 1263 | bfin_irq_clear(ap); |
1270 | } | 1264 | } |
1271 | 1265 | ||
1272 | /** | 1266 | /** |
1273 | * bfin_bmdma_thaw - Thaw DMA controller port | 1267 | * bfin_thaw - Thaw DMA controller port |
1274 | * @ap: port to thaw | 1268 | * @ap: port to thaw |
1275 | * | 1269 | * |
1276 | * Note: Original code is ata_bmdma_thaw(). | 1270 | * Note: Original code is ata_sff_thaw(). |
1277 | */ | 1271 | */ |
1278 | 1272 | ||
1279 | void bfin_bmdma_thaw(struct ata_port *ap) | 1273 | void bfin_thaw(struct ata_port *ap) |
1280 | { | 1274 | { |
1281 | bfin_check_status(ap); | 1275 | bfin_check_status(ap); |
1282 | bfin_irq_clear(ap); | 1276 | bfin_irq_clear(ap); |
@@ -1284,14 +1278,14 @@ void bfin_bmdma_thaw(struct ata_port *ap) | |||
1284 | } | 1278 | } |
1285 | 1279 | ||
1286 | /** | 1280 | /** |
1287 | * bfin_std_postreset - standard postreset callback | 1281 | * bfin_postreset - standard postreset callback |
1288 | * @ap: the target ata_port | 1282 | * @ap: the target ata_port |
1289 | * @classes: classes of attached devices | 1283 | * @classes: classes of attached devices |
1290 | * | 1284 | * |
1291 | * Note: Original code is ata_std_postreset(). | 1285 | * Note: Original code is ata_sff_postreset(). |
1292 | */ | 1286 | */ |
1293 | 1287 | ||
1294 | static void bfin_std_postreset(struct ata_link *link, unsigned int *classes) | 1288 | static void bfin_postreset(struct ata_link *link, unsigned int *classes) |
1295 | { | 1289 | { |
1296 | struct ata_port *ap = link->ap; | 1290 | struct ata_port *ap = link->ap; |
1297 | void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr; | 1291 | void __iomem *base = (void __iomem *)ap->ioaddr.ctl_addr; |
@@ -1301,9 +1295,9 @@ static void bfin_std_postreset(struct ata_link *link, unsigned int *classes) | |||
1301 | 1295 | ||
1302 | /* is double-select really necessary? */ | 1296 | /* is double-select really necessary? */ |
1303 | if (classes[0] != ATA_DEV_NONE) | 1297 | if (classes[0] != ATA_DEV_NONE) |
1304 | bfin_std_dev_select(ap, 1); | 1298 | bfin_dev_select(ap, 1); |
1305 | if (classes[1] != ATA_DEV_NONE) | 1299 | if (classes[1] != ATA_DEV_NONE) |
1306 | bfin_std_dev_select(ap, 0); | 1300 | bfin_dev_select(ap, 0); |
1307 | 1301 | ||
1308 | /* bail out if no device is present */ | 1302 | /* bail out if no device is present */ |
1309 | if (classes[0] == ATA_DEV_NONE && classes[1] == ATA_DEV_NONE) { | 1303 | if (classes[0] == ATA_DEV_NONE && classes[1] == ATA_DEV_NONE) { |
@@ -1314,17 +1308,6 @@ static void bfin_std_postreset(struct ata_link *link, unsigned int *classes) | |||
1314 | write_atapi_register(base, ATA_REG_CTRL, ap->ctl); | 1308 | write_atapi_register(base, ATA_REG_CTRL, ap->ctl); |
1315 | } | 1309 | } |
1316 | 1310 | ||
1317 | /** | ||
1318 | * bfin_error_handler - Stock error handler for DMA controller | ||
1319 | * @ap: port to handle error for | ||
1320 | */ | ||
1321 | |||
1322 | static void bfin_error_handler(struct ata_port *ap) | ||
1323 | { | ||
1324 | ata_bmdma_drive_eh(ap, ata_std_prereset, bfin_std_softreset, NULL, | ||
1325 | bfin_std_postreset); | ||
1326 | } | ||
1327 | |||
1328 | static void bfin_port_stop(struct ata_port *ap) | 1311 | static void bfin_port_stop(struct ata_port *ap) |
1329 | { | 1312 | { |
1330 | dev_dbg(ap->dev, "in atapi port stop\n"); | 1313 | dev_dbg(ap->dev, "in atapi port stop\n"); |
@@ -1357,51 +1340,40 @@ static int bfin_port_start(struct ata_port *ap) | |||
1357 | } | 1340 | } |
1358 | 1341 | ||
1359 | static struct scsi_host_template bfin_sht = { | 1342 | static struct scsi_host_template bfin_sht = { |
1360 | .module = THIS_MODULE, | 1343 | ATA_BASE_SHT(DRV_NAME), |
1361 | .name = DRV_NAME, | ||
1362 | .ioctl = ata_scsi_ioctl, | ||
1363 | .queuecommand = ata_scsi_queuecmd, | ||
1364 | .can_queue = ATA_DEF_QUEUE, | ||
1365 | .this_id = ATA_SHT_THIS_ID, | ||
1366 | .sg_tablesize = SG_NONE, | 1344 | .sg_tablesize = SG_NONE, |
1367 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
1368 | .emulated = ATA_SHT_EMULATED, | ||
1369 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
1370 | .proc_name = DRV_NAME, | ||
1371 | .dma_boundary = ATA_DMA_BOUNDARY, | 1345 | .dma_boundary = ATA_DMA_BOUNDARY, |
1372 | .slave_configure = ata_scsi_slave_config, | ||
1373 | .slave_destroy = ata_scsi_slave_destroy, | ||
1374 | .bios_param = ata_std_bios_param, | ||
1375 | }; | 1346 | }; |
1376 | 1347 | ||
1377 | static const struct ata_port_operations bfin_pata_ops = { | 1348 | static const struct ata_port_operations bfin_pata_ops = { |
1349 | .inherits = &ata_sff_port_ops, | ||
1350 | |||
1378 | .set_piomode = bfin_set_piomode, | 1351 | .set_piomode = bfin_set_piomode, |
1379 | .set_dmamode = bfin_set_dmamode, | 1352 | .set_dmamode = bfin_set_dmamode, |
1380 | 1353 | ||
1381 | .tf_load = bfin_tf_load, | 1354 | .sff_tf_load = bfin_tf_load, |
1382 | .tf_read = bfin_tf_read, | 1355 | .sff_tf_read = bfin_tf_read, |
1383 | .exec_command = bfin_exec_command, | 1356 | .sff_exec_command = bfin_exec_command, |
1384 | .check_status = bfin_check_status, | 1357 | .sff_check_status = bfin_check_status, |
1385 | .check_altstatus = bfin_check_altstatus, | 1358 | .sff_check_altstatus = bfin_check_altstatus, |
1386 | .dev_select = bfin_std_dev_select, | 1359 | .sff_dev_select = bfin_dev_select, |
1387 | 1360 | ||
1388 | .bmdma_setup = bfin_bmdma_setup, | 1361 | .bmdma_setup = bfin_bmdma_setup, |
1389 | .bmdma_start = bfin_bmdma_start, | 1362 | .bmdma_start = bfin_bmdma_start, |
1390 | .bmdma_stop = bfin_bmdma_stop, | 1363 | .bmdma_stop = bfin_bmdma_stop, |
1391 | .bmdma_status = bfin_bmdma_status, | 1364 | .bmdma_status = bfin_bmdma_status, |
1392 | .data_xfer = bfin_data_xfer, | 1365 | .sff_data_xfer = bfin_data_xfer, |
1393 | 1366 | ||
1394 | .qc_prep = ata_noop_qc_prep, | 1367 | .qc_prep = ata_noop_qc_prep, |
1395 | .qc_issue = ata_qc_issue_prot, | ||
1396 | 1368 | ||
1397 | .freeze = bfin_bmdma_freeze, | 1369 | .freeze = bfin_freeze, |
1398 | .thaw = bfin_bmdma_thaw, | 1370 | .thaw = bfin_thaw, |
1399 | .error_handler = bfin_error_handler, | 1371 | .softreset = bfin_softreset, |
1372 | .postreset = bfin_postreset, | ||
1400 | .post_internal_cmd = bfin_bmdma_stop, | 1373 | .post_internal_cmd = bfin_bmdma_stop, |
1401 | 1374 | ||
1402 | .irq_handler = ata_interrupt, | 1375 | .sff_irq_clear = bfin_irq_clear, |
1403 | .irq_clear = bfin_irq_clear, | 1376 | .sff_irq_on = bfin_irq_on, |
1404 | .irq_on = bfin_irq_on, | ||
1405 | 1377 | ||
1406 | .port_start = bfin_port_start, | 1378 | .port_start = bfin_port_start, |
1407 | .port_stop = bfin_port_stop, | 1379 | .port_stop = bfin_port_stop, |
@@ -1409,7 +1381,6 @@ static const struct ata_port_operations bfin_pata_ops = { | |||
1409 | 1381 | ||
1410 | static struct ata_port_info bfin_port_info[] = { | 1382 | static struct ata_port_info bfin_port_info[] = { |
1411 | { | 1383 | { |
1412 | .sht = &bfin_sht, | ||
1413 | .flags = ATA_FLAG_SLAVE_POSS | 1384 | .flags = ATA_FLAG_SLAVE_POSS |
1414 | | ATA_FLAG_MMIO | 1385 | | ATA_FLAG_MMIO |
1415 | | ATA_FLAG_NO_LEGACY, | 1386 | | ATA_FLAG_NO_LEGACY, |
@@ -1536,7 +1507,7 @@ static int __devinit bfin_atapi_probe(struct platform_device *pdev) | |||
1536 | } | 1507 | } |
1537 | 1508 | ||
1538 | if (ata_host_activate(host, platform_get_irq(pdev, 0), | 1509 | if (ata_host_activate(host, platform_get_irq(pdev, 0), |
1539 | ata_interrupt, IRQF_SHARED, &bfin_sht) != 0) { | 1510 | ata_sff_interrupt, IRQF_SHARED, &bfin_sht) != 0) { |
1540 | peripheral_free_list(atapi_io_port); | 1511 | peripheral_free_list(atapi_io_port); |
1541 | dev_err(&pdev->dev, "Fail to attach ATAPI device\n"); | 1512 | dev_err(&pdev->dev, "Fail to attach ATAPI device\n"); |
1542 | return -ENODEV; | 1513 | return -ENODEV; |
diff --git a/drivers/ata/pata_cmd640.c b/drivers/ata/pata_cmd640.c index 43d198f90968..2de30b990278 100644 --- a/drivers/ata/pata_cmd640.c +++ b/drivers/ata/pata_cmd640.c | |||
@@ -107,8 +107,8 @@ static void cmd640_set_piomode(struct ata_port *ap, struct ata_device *adev) | |||
107 | pci_write_config_byte(pdev, arttim + 1, (t.active << 4) | t.recover); | 107 | pci_write_config_byte(pdev, arttim + 1, (t.active << 4) | t.recover); |
108 | } else { | 108 | } else { |
109 | /* Save the shared timings for channel, they will be loaded | 109 | /* Save the shared timings for channel, they will be loaded |
110 | by qc_issue_prot. Reloading the setup time is expensive | 110 | by qc_issue. Reloading the setup time is expensive so we |
111 | so we keep a merged one loaded */ | 111 | keep a merged one loaded */ |
112 | pci_read_config_byte(pdev, ARTIM23, ®); | 112 | pci_read_config_byte(pdev, ARTIM23, ®); |
113 | reg &= 0x3F; | 113 | reg &= 0x3F; |
114 | reg |= t.setup; | 114 | reg |= t.setup; |
@@ -119,14 +119,14 @@ static void cmd640_set_piomode(struct ata_port *ap, struct ata_device *adev) | |||
119 | 119 | ||
120 | 120 | ||
121 | /** | 121 | /** |
122 | * cmd640_qc_issue_prot - command preparation hook | 122 | * cmd640_qc_issue - command preparation hook |
123 | * @qc: Command to be issued | 123 | * @qc: Command to be issued |
124 | * | 124 | * |
125 | * Channel 1 has shared timings. We must reprogram the | 125 | * Channel 1 has shared timings. We must reprogram the |
126 | * clock each drive 2/3 switch we do. | 126 | * clock each drive 2/3 switch we do. |
127 | */ | 127 | */ |
128 | 128 | ||
129 | static unsigned int cmd640_qc_issue_prot(struct ata_queued_cmd *qc) | 129 | static unsigned int cmd640_qc_issue(struct ata_queued_cmd *qc) |
130 | { | 130 | { |
131 | struct ata_port *ap = qc->ap; | 131 | struct ata_port *ap = qc->ap; |
132 | struct ata_device *adev = qc->dev; | 132 | struct ata_device *adev = qc->dev; |
@@ -137,7 +137,7 @@ static unsigned int cmd640_qc_issue_prot(struct ata_queued_cmd *qc) | |||
137 | pci_write_config_byte(pdev, DRWTIM23, timing->reg58[adev->devno]); | 137 | pci_write_config_byte(pdev, DRWTIM23, timing->reg58[adev->devno]); |
138 | timing->last = adev->devno; | 138 | timing->last = adev->devno; |
139 | } | 139 | } |
140 | return ata_qc_issue_prot(qc); | 140 | return ata_sff_qc_issue(qc); |
141 | } | 141 | } |
142 | 142 | ||
143 | /** | 143 | /** |
@@ -166,53 +166,16 @@ static int cmd640_port_start(struct ata_port *ap) | |||
166 | } | 166 | } |
167 | 167 | ||
168 | static struct scsi_host_template cmd640_sht = { | 168 | static struct scsi_host_template cmd640_sht = { |
169 | .module = THIS_MODULE, | 169 | ATA_BMDMA_SHT(DRV_NAME), |
170 | .name = DRV_NAME, | ||
171 | .ioctl = ata_scsi_ioctl, | ||
172 | .queuecommand = ata_scsi_queuecmd, | ||
173 | .can_queue = ATA_DEF_QUEUE, | ||
174 | .this_id = ATA_SHT_THIS_ID, | ||
175 | .sg_tablesize = LIBATA_MAX_PRD, | ||
176 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
177 | .emulated = ATA_SHT_EMULATED, | ||
178 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
179 | .proc_name = DRV_NAME, | ||
180 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
181 | .slave_configure = ata_scsi_slave_config, | ||
182 | .slave_destroy = ata_scsi_slave_destroy, | ||
183 | .bios_param = ata_std_bios_param, | ||
184 | }; | 170 | }; |
185 | 171 | ||
186 | static struct ata_port_operations cmd640_port_ops = { | 172 | static struct ata_port_operations cmd640_port_ops = { |
187 | .set_piomode = cmd640_set_piomode, | 173 | .inherits = &ata_bmdma_port_ops, |
188 | .mode_filter = ata_pci_default_filter, | 174 | /* In theory xfer_noirq is not needed once we kill the prefetcher */ |
189 | .tf_load = ata_tf_load, | 175 | .sff_data_xfer = ata_sff_data_xfer_noirq, |
190 | .tf_read = ata_tf_read, | 176 | .qc_issue = cmd640_qc_issue, |
191 | .check_status = ata_check_status, | ||
192 | .exec_command = ata_exec_command, | ||
193 | .dev_select = ata_std_dev_select, | ||
194 | |||
195 | .freeze = ata_bmdma_freeze, | ||
196 | .thaw = ata_bmdma_thaw, | ||
197 | .error_handler = ata_bmdma_error_handler, | ||
198 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
199 | .cable_detect = ata_cable_40wire, | 177 | .cable_detect = ata_cable_40wire, |
200 | 178 | .set_piomode = cmd640_set_piomode, | |
201 | .bmdma_setup = ata_bmdma_setup, | ||
202 | .bmdma_start = ata_bmdma_start, | ||
203 | .bmdma_stop = ata_bmdma_stop, | ||
204 | .bmdma_status = ata_bmdma_status, | ||
205 | |||
206 | .qc_prep = ata_qc_prep, | ||
207 | .qc_issue = cmd640_qc_issue_prot, | ||
208 | |||
209 | /* In theory this is not needed once we kill the prefetcher */ | ||
210 | .data_xfer = ata_data_xfer_noirq, | ||
211 | |||
212 | .irq_handler = ata_interrupt, | ||
213 | .irq_clear = ata_bmdma_irq_clear, | ||
214 | .irq_on = ata_irq_on, | ||
215 | |||
216 | .port_start = cmd640_port_start, | 179 | .port_start = cmd640_port_start, |
217 | }; | 180 | }; |
218 | 181 | ||
@@ -248,26 +211,36 @@ static void cmd640_hardware_init(struct pci_dev *pdev) | |||
248 | static int cmd640_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | 211 | static int cmd640_init_one(struct pci_dev *pdev, const struct pci_device_id *id) |
249 | { | 212 | { |
250 | static const struct ata_port_info info = { | 213 | static const struct ata_port_info info = { |
251 | .sht = &cmd640_sht, | ||
252 | .flags = ATA_FLAG_SLAVE_POSS, | 214 | .flags = ATA_FLAG_SLAVE_POSS, |
253 | .pio_mask = 0x1f, | 215 | .pio_mask = 0x1f, |
254 | .port_ops = &cmd640_port_ops | 216 | .port_ops = &cmd640_port_ops |
255 | }; | 217 | }; |
256 | const struct ata_port_info *ppi[] = { &info, NULL }; | 218 | const struct ata_port_info *ppi[] = { &info, NULL }; |
219 | int rc; | ||
220 | |||
221 | rc = pcim_enable_device(pdev); | ||
222 | if (rc) | ||
223 | return rc; | ||
257 | 224 | ||
258 | cmd640_hardware_init(pdev); | 225 | cmd640_hardware_init(pdev); |
259 | return ata_pci_init_one(pdev, ppi); | 226 | |
227 | return ata_pci_sff_init_one(pdev, ppi, &cmd640_sht, NULL); | ||
260 | } | 228 | } |
261 | 229 | ||
230 | #ifdef CONFIG_PM | ||
262 | static int cmd640_reinit_one(struct pci_dev *pdev) | 231 | static int cmd640_reinit_one(struct pci_dev *pdev) |
263 | { | 232 | { |
233 | struct ata_host *host = dev_get_drvdata(&pdev->dev); | ||
234 | int rc; | ||
235 | |||
236 | rc = ata_pci_device_do_resume(pdev); | ||
237 | if (rc) | ||
238 | return rc; | ||
264 | cmd640_hardware_init(pdev); | 239 | cmd640_hardware_init(pdev); |
265 | #ifdef CONFIG_PM | 240 | ata_host_resume(host); |
266 | return ata_pci_device_resume(pdev); | ||
267 | #else | ||
268 | return 0; | 241 | return 0; |
269 | #endif | ||
270 | } | 242 | } |
243 | #endif | ||
271 | 244 | ||
272 | static const struct pci_device_id cmd640[] = { | 245 | static const struct pci_device_id cmd640[] = { |
273 | { PCI_VDEVICE(CMD, 0x640), 0 }, | 246 | { PCI_VDEVICE(CMD, 0x640), 0 }, |
@@ -281,8 +254,8 @@ static struct pci_driver cmd640_pci_driver = { | |||
281 | .remove = ata_pci_remove_one, | 254 | .remove = ata_pci_remove_one, |
282 | #ifdef CONFIG_PM | 255 | #ifdef CONFIG_PM |
283 | .suspend = ata_pci_device_suspend, | 256 | .suspend = ata_pci_device_suspend, |
284 | #endif | ||
285 | .resume = cmd640_reinit_one, | 257 | .resume = cmd640_reinit_one, |
258 | #endif | ||
286 | }; | 259 | }; |
287 | 260 | ||
288 | static int __init cmd640_init(void) | 261 | static int __init cmd640_init(void) |
diff --git a/drivers/ata/pata_cmd64x.c b/drivers/ata/pata_cmd64x.c index 7acbbd9ee469..ddd09b7d98c9 100644 --- a/drivers/ata/pata_cmd64x.c +++ b/drivers/ata/pata_cmd64x.c | |||
@@ -266,120 +266,30 @@ static void cmd646r1_bmdma_stop(struct ata_queued_cmd *qc) | |||
266 | } | 266 | } |
267 | 267 | ||
268 | static struct scsi_host_template cmd64x_sht = { | 268 | static struct scsi_host_template cmd64x_sht = { |
269 | .module = THIS_MODULE, | 269 | ATA_BMDMA_SHT(DRV_NAME), |
270 | .name = DRV_NAME, | ||
271 | .ioctl = ata_scsi_ioctl, | ||
272 | .queuecommand = ata_scsi_queuecmd, | ||
273 | .can_queue = ATA_DEF_QUEUE, | ||
274 | .this_id = ATA_SHT_THIS_ID, | ||
275 | .sg_tablesize = LIBATA_MAX_PRD, | ||
276 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
277 | .emulated = ATA_SHT_EMULATED, | ||
278 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
279 | .proc_name = DRV_NAME, | ||
280 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
281 | .slave_configure = ata_scsi_slave_config, | ||
282 | .slave_destroy = ata_scsi_slave_destroy, | ||
283 | .bios_param = ata_std_bios_param, | ||
284 | }; | 270 | }; |
285 | 271 | ||
286 | static struct ata_port_operations cmd64x_port_ops = { | 272 | static const struct ata_port_operations cmd64x_base_ops = { |
273 | .inherits = &ata_bmdma_port_ops, | ||
287 | .set_piomode = cmd64x_set_piomode, | 274 | .set_piomode = cmd64x_set_piomode, |
288 | .set_dmamode = cmd64x_set_dmamode, | 275 | .set_dmamode = cmd64x_set_dmamode, |
289 | .mode_filter = ata_pci_default_filter, | ||
290 | .tf_load = ata_tf_load, | ||
291 | .tf_read = ata_tf_read, | ||
292 | .check_status = ata_check_status, | ||
293 | .exec_command = ata_exec_command, | ||
294 | .dev_select = ata_std_dev_select, | ||
295 | |||
296 | .freeze = ata_bmdma_freeze, | ||
297 | .thaw = ata_bmdma_thaw, | ||
298 | .error_handler = ata_bmdma_error_handler, | ||
299 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
300 | .cable_detect = ata_cable_40wire, | ||
301 | |||
302 | .bmdma_setup = ata_bmdma_setup, | ||
303 | .bmdma_start = ata_bmdma_start, | ||
304 | .bmdma_stop = ata_bmdma_stop, | ||
305 | .bmdma_status = ata_bmdma_status, | ||
306 | |||
307 | .qc_prep = ata_qc_prep, | ||
308 | .qc_issue = ata_qc_issue_prot, | ||
309 | |||
310 | .data_xfer = ata_data_xfer, | ||
311 | |||
312 | .irq_handler = ata_interrupt, | ||
313 | .irq_clear = ata_bmdma_irq_clear, | ||
314 | .irq_on = ata_irq_on, | ||
315 | |||
316 | .port_start = ata_port_start, | ||
317 | }; | 276 | }; |
318 | 277 | ||
319 | static struct ata_port_operations cmd646r1_port_ops = { | 278 | static struct ata_port_operations cmd64x_port_ops = { |
320 | .set_piomode = cmd64x_set_piomode, | 279 | .inherits = &cmd64x_base_ops, |
321 | .set_dmamode = cmd64x_set_dmamode, | ||
322 | .mode_filter = ata_pci_default_filter, | ||
323 | .tf_load = ata_tf_load, | ||
324 | .tf_read = ata_tf_read, | ||
325 | .check_status = ata_check_status, | ||
326 | .exec_command = ata_exec_command, | ||
327 | .dev_select = ata_std_dev_select, | ||
328 | |||
329 | .freeze = ata_bmdma_freeze, | ||
330 | .thaw = ata_bmdma_thaw, | ||
331 | .error_handler = ata_bmdma_error_handler, | ||
332 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
333 | .cable_detect = ata_cable_40wire, | 280 | .cable_detect = ata_cable_40wire, |
281 | }; | ||
334 | 282 | ||
335 | .bmdma_setup = ata_bmdma_setup, | 283 | static struct ata_port_operations cmd646r1_port_ops = { |
336 | .bmdma_start = ata_bmdma_start, | 284 | .inherits = &cmd64x_base_ops, |
337 | .bmdma_stop = cmd646r1_bmdma_stop, | 285 | .bmdma_stop = cmd646r1_bmdma_stop, |
338 | .bmdma_status = ata_bmdma_status, | 286 | .cable_detect = ata_cable_40wire, |
339 | |||
340 | .qc_prep = ata_qc_prep, | ||
341 | .qc_issue = ata_qc_issue_prot, | ||
342 | |||
343 | .data_xfer = ata_data_xfer, | ||
344 | |||
345 | .irq_handler = ata_interrupt, | ||
346 | .irq_clear = ata_bmdma_irq_clear, | ||
347 | .irq_on = ata_irq_on, | ||
348 | |||
349 | .port_start = ata_port_start, | ||
350 | }; | 287 | }; |
351 | 288 | ||
352 | static struct ata_port_operations cmd648_port_ops = { | 289 | static struct ata_port_operations cmd648_port_ops = { |
353 | .set_piomode = cmd64x_set_piomode, | 290 | .inherits = &cmd64x_base_ops, |
354 | .set_dmamode = cmd64x_set_dmamode, | ||
355 | .mode_filter = ata_pci_default_filter, | ||
356 | .tf_load = ata_tf_load, | ||
357 | .tf_read = ata_tf_read, | ||
358 | .check_status = ata_check_status, | ||
359 | .exec_command = ata_exec_command, | ||
360 | .dev_select = ata_std_dev_select, | ||
361 | |||
362 | .freeze = ata_bmdma_freeze, | ||
363 | .thaw = ata_bmdma_thaw, | ||
364 | .error_handler = ata_bmdma_error_handler, | ||
365 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
366 | .cable_detect = cmd648_cable_detect, | ||
367 | |||
368 | .bmdma_setup = ata_bmdma_setup, | ||
369 | .bmdma_start = ata_bmdma_start, | ||
370 | .bmdma_stop = cmd648_bmdma_stop, | 291 | .bmdma_stop = cmd648_bmdma_stop, |
371 | .bmdma_status = ata_bmdma_status, | 292 | .cable_detect = cmd648_cable_detect, |
372 | |||
373 | .qc_prep = ata_qc_prep, | ||
374 | .qc_issue = ata_qc_issue_prot, | ||
375 | |||
376 | .data_xfer = ata_data_xfer, | ||
377 | |||
378 | .irq_handler = ata_interrupt, | ||
379 | .irq_clear = ata_bmdma_irq_clear, | ||
380 | .irq_on = ata_irq_on, | ||
381 | |||
382 | .port_start = ata_port_start, | ||
383 | }; | 293 | }; |
384 | 294 | ||
385 | static int cmd64x_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | 295 | static int cmd64x_init_one(struct pci_dev *pdev, const struct pci_device_id *id) |
@@ -388,21 +298,18 @@ static int cmd64x_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
388 | 298 | ||
389 | static const struct ata_port_info cmd_info[6] = { | 299 | static const struct ata_port_info cmd_info[6] = { |
390 | { /* CMD 643 - no UDMA */ | 300 | { /* CMD 643 - no UDMA */ |
391 | .sht = &cmd64x_sht, | ||
392 | .flags = ATA_FLAG_SLAVE_POSS, | 301 | .flags = ATA_FLAG_SLAVE_POSS, |
393 | .pio_mask = 0x1f, | 302 | .pio_mask = 0x1f, |
394 | .mwdma_mask = 0x07, | 303 | .mwdma_mask = 0x07, |
395 | .port_ops = &cmd64x_port_ops | 304 | .port_ops = &cmd64x_port_ops |
396 | }, | 305 | }, |
397 | { /* CMD 646 with broken UDMA */ | 306 | { /* CMD 646 with broken UDMA */ |
398 | .sht = &cmd64x_sht, | ||
399 | .flags = ATA_FLAG_SLAVE_POSS, | 307 | .flags = ATA_FLAG_SLAVE_POSS, |
400 | .pio_mask = 0x1f, | 308 | .pio_mask = 0x1f, |
401 | .mwdma_mask = 0x07, | 309 | .mwdma_mask = 0x07, |
402 | .port_ops = &cmd64x_port_ops | 310 | .port_ops = &cmd64x_port_ops |
403 | }, | 311 | }, |
404 | { /* CMD 646 with working UDMA */ | 312 | { /* CMD 646 with working UDMA */ |
405 | .sht = &cmd64x_sht, | ||
406 | .flags = ATA_FLAG_SLAVE_POSS, | 313 | .flags = ATA_FLAG_SLAVE_POSS, |
407 | .pio_mask = 0x1f, | 314 | .pio_mask = 0x1f, |
408 | .mwdma_mask = 0x07, | 315 | .mwdma_mask = 0x07, |
@@ -410,14 +317,12 @@ static int cmd64x_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
410 | .port_ops = &cmd64x_port_ops | 317 | .port_ops = &cmd64x_port_ops |
411 | }, | 318 | }, |
412 | { /* CMD 646 rev 1 */ | 319 | { /* CMD 646 rev 1 */ |
413 | .sht = &cmd64x_sht, | ||
414 | .flags = ATA_FLAG_SLAVE_POSS, | 320 | .flags = ATA_FLAG_SLAVE_POSS, |
415 | .pio_mask = 0x1f, | 321 | .pio_mask = 0x1f, |
416 | .mwdma_mask = 0x07, | 322 | .mwdma_mask = 0x07, |
417 | .port_ops = &cmd646r1_port_ops | 323 | .port_ops = &cmd646r1_port_ops |
418 | }, | 324 | }, |
419 | { /* CMD 648 */ | 325 | { /* CMD 648 */ |
420 | .sht = &cmd64x_sht, | ||
421 | .flags = ATA_FLAG_SLAVE_POSS, | 326 | .flags = ATA_FLAG_SLAVE_POSS, |
422 | .pio_mask = 0x1f, | 327 | .pio_mask = 0x1f, |
423 | .mwdma_mask = 0x07, | 328 | .mwdma_mask = 0x07, |
@@ -425,7 +330,6 @@ static int cmd64x_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
425 | .port_ops = &cmd648_port_ops | 330 | .port_ops = &cmd648_port_ops |
426 | }, | 331 | }, |
427 | { /* CMD 649 */ | 332 | { /* CMD 649 */ |
428 | .sht = &cmd64x_sht, | ||
429 | .flags = ATA_FLAG_SLAVE_POSS, | 333 | .flags = ATA_FLAG_SLAVE_POSS, |
430 | .pio_mask = 0x1f, | 334 | .pio_mask = 0x1f, |
431 | .mwdma_mask = 0x07, | 335 | .mwdma_mask = 0x07, |
@@ -435,12 +339,17 @@ static int cmd64x_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
435 | }; | 339 | }; |
436 | const struct ata_port_info *ppi[] = { &cmd_info[id->driver_data], NULL }; | 340 | const struct ata_port_info *ppi[] = { &cmd_info[id->driver_data], NULL }; |
437 | u8 mrdmode; | 341 | u8 mrdmode; |
342 | int rc; | ||
343 | |||
344 | rc = pcim_enable_device(pdev); | ||
345 | if (rc) | ||
346 | return rc; | ||
438 | 347 | ||
439 | pci_read_config_dword(pdev, PCI_CLASS_REVISION, &class_rev); | 348 | pci_read_config_dword(pdev, PCI_CLASS_REVISION, &class_rev); |
440 | class_rev &= 0xFF; | 349 | class_rev &= 0xFF; |
441 | 350 | ||
442 | if (id->driver_data == 0) /* 643 */ | 351 | if (id->driver_data == 0) /* 643 */ |
443 | ata_pci_clear_simplex(pdev); | 352 | ata_pci_bmdma_clear_simplex(pdev); |
444 | 353 | ||
445 | if (pdev->device == PCI_DEVICE_ID_CMD_646) { | 354 | if (pdev->device == PCI_DEVICE_ID_CMD_646) { |
446 | /* Does UDMA work ? */ | 355 | /* Does UDMA work ? */ |
@@ -464,13 +373,20 @@ static int cmd64x_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
464 | pci_write_config_byte(pdev, UDIDETCR0, 0xF0); | 373 | pci_write_config_byte(pdev, UDIDETCR0, 0xF0); |
465 | #endif | 374 | #endif |
466 | 375 | ||
467 | return ata_pci_init_one(pdev, ppi); | 376 | return ata_pci_sff_init_one(pdev, ppi, &cmd64x_sht, NULL); |
468 | } | 377 | } |
469 | 378 | ||
470 | #ifdef CONFIG_PM | 379 | #ifdef CONFIG_PM |
471 | static int cmd64x_reinit_one(struct pci_dev *pdev) | 380 | static int cmd64x_reinit_one(struct pci_dev *pdev) |
472 | { | 381 | { |
382 | struct ata_host *host = dev_get_drvdata(&pdev->dev); | ||
473 | u8 mrdmode; | 383 | u8 mrdmode; |
384 | int rc; | ||
385 | |||
386 | rc = ata_pci_device_do_resume(pdev); | ||
387 | if (rc) | ||
388 | return rc; | ||
389 | |||
474 | pci_write_config_byte(pdev, PCI_LATENCY_TIMER, 64); | 390 | pci_write_config_byte(pdev, PCI_LATENCY_TIMER, 64); |
475 | pci_read_config_byte(pdev, MRDMODE, &mrdmode); | 391 | pci_read_config_byte(pdev, MRDMODE, &mrdmode); |
476 | mrdmode &= ~ 0x30; /* IRQ set up */ | 392 | mrdmode &= ~ 0x30; /* IRQ set up */ |
@@ -479,7 +395,8 @@ static int cmd64x_reinit_one(struct pci_dev *pdev) | |||
479 | #ifdef CONFIG_PPC | 395 | #ifdef CONFIG_PPC |
480 | pci_write_config_byte(pdev, UDIDETCR0, 0xF0); | 396 | pci_write_config_byte(pdev, UDIDETCR0, 0xF0); |
481 | #endif | 397 | #endif |
482 | return ata_pci_device_resume(pdev); | 398 | ata_host_resume(host); |
399 | return 0; | ||
483 | } | 400 | } |
484 | #endif | 401 | #endif |
485 | 402 | ||
diff --git a/drivers/ata/pata_cs5520.c b/drivers/ata/pata_cs5520.c index 7ed279b0a12e..1186bcd2781c 100644 --- a/drivers/ata/pata_cs5520.c +++ b/drivers/ata/pata_cs5520.c | |||
@@ -140,51 +140,16 @@ static void cs5520_set_piomode(struct ata_port *ap, struct ata_device *adev) | |||
140 | } | 140 | } |
141 | 141 | ||
142 | static struct scsi_host_template cs5520_sht = { | 142 | static struct scsi_host_template cs5520_sht = { |
143 | .module = THIS_MODULE, | 143 | ATA_BMDMA_SHT(DRV_NAME), |
144 | .name = DRV_NAME, | ||
145 | .ioctl = ata_scsi_ioctl, | ||
146 | .queuecommand = ata_scsi_queuecmd, | ||
147 | .can_queue = ATA_DEF_QUEUE, | ||
148 | .this_id = ATA_SHT_THIS_ID, | ||
149 | .sg_tablesize = LIBATA_DUMB_MAX_PRD, | 144 | .sg_tablesize = LIBATA_DUMB_MAX_PRD, |
150 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
151 | .emulated = ATA_SHT_EMULATED, | ||
152 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
153 | .proc_name = DRV_NAME, | ||
154 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
155 | .slave_configure = ata_scsi_slave_config, | ||
156 | .slave_destroy = ata_scsi_slave_destroy, | ||
157 | .bios_param = ata_std_bios_param, | ||
158 | }; | 145 | }; |
159 | 146 | ||
160 | static struct ata_port_operations cs5520_port_ops = { | 147 | static struct ata_port_operations cs5520_port_ops = { |
148 | .inherits = &ata_bmdma_port_ops, | ||
149 | .qc_prep = ata_sff_dumb_qc_prep, | ||
150 | .cable_detect = ata_cable_40wire, | ||
161 | .set_piomode = cs5520_set_piomode, | 151 | .set_piomode = cs5520_set_piomode, |
162 | .set_dmamode = cs5520_set_dmamode, | 152 | .set_dmamode = cs5520_set_dmamode, |
163 | |||
164 | .tf_load = ata_tf_load, | ||
165 | .tf_read = ata_tf_read, | ||
166 | .check_status = ata_check_status, | ||
167 | .exec_command = ata_exec_command, | ||
168 | .dev_select = ata_std_dev_select, | ||
169 | |||
170 | .freeze = ata_bmdma_freeze, | ||
171 | .thaw = ata_bmdma_thaw, | ||
172 | .error_handler = ata_bmdma_error_handler, | ||
173 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
174 | .cable_detect = ata_cable_40wire, | ||
175 | |||
176 | .bmdma_setup = ata_bmdma_setup, | ||
177 | .bmdma_start = ata_bmdma_start, | ||
178 | .bmdma_stop = ata_bmdma_stop, | ||
179 | .bmdma_status = ata_bmdma_status, | ||
180 | .qc_prep = ata_dumb_qc_prep, | ||
181 | .qc_issue = ata_qc_issue_prot, | ||
182 | .data_xfer = ata_data_xfer, | ||
183 | |||
184 | .irq_clear = ata_bmdma_irq_clear, | ||
185 | .irq_on = ata_irq_on, | ||
186 | |||
187 | .port_start = ata_sff_port_start, | ||
188 | }; | 153 | }; |
189 | 154 | ||
190 | static int __devinit cs5520_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | 155 | static int __devinit cs5520_init_one(struct pci_dev *pdev, const struct pci_device_id *id) |
@@ -203,6 +168,10 @@ static int __devinit cs5520_init_one(struct pci_dev *pdev, const struct pci_devi | |||
203 | struct ata_ioports *ioaddr; | 168 | struct ata_ioports *ioaddr; |
204 | int i, rc; | 169 | int i, rc; |
205 | 170 | ||
171 | rc = pcim_enable_device(pdev); | ||
172 | if (rc) | ||
173 | return rc; | ||
174 | |||
206 | /* IDE port enable bits */ | 175 | /* IDE port enable bits */ |
207 | pci_read_config_byte(pdev, 0x60, &pcicfg); | 176 | pci_read_config_byte(pdev, 0x60, &pcicfg); |
208 | 177 | ||
@@ -258,7 +227,7 @@ static int __devinit cs5520_init_one(struct pci_dev *pdev, const struct pci_devi | |||
258 | ioaddr->ctl_addr = iomap[1]; | 227 | ioaddr->ctl_addr = iomap[1]; |
259 | ioaddr->altstatus_addr = iomap[1]; | 228 | ioaddr->altstatus_addr = iomap[1]; |
260 | ioaddr->bmdma_addr = iomap[4]; | 229 | ioaddr->bmdma_addr = iomap[4]; |
261 | ata_std_ports(ioaddr); | 230 | ata_sff_std_ports(ioaddr); |
262 | 231 | ||
263 | ata_port_desc(host->ports[0], | 232 | ata_port_desc(host->ports[0], |
264 | "cmd 0x%x ctl 0x%x", cmd_port[0], ctl_port[0]); | 233 | "cmd 0x%x ctl 0x%x", cmd_port[0], ctl_port[0]); |
@@ -269,7 +238,7 @@ static int __devinit cs5520_init_one(struct pci_dev *pdev, const struct pci_devi | |||
269 | ioaddr->ctl_addr = iomap[3]; | 238 | ioaddr->ctl_addr = iomap[3]; |
270 | ioaddr->altstatus_addr = iomap[3]; | 239 | ioaddr->altstatus_addr = iomap[3]; |
271 | ioaddr->bmdma_addr = iomap[4] + 8; | 240 | ioaddr->bmdma_addr = iomap[4] + 8; |
272 | ata_std_ports(ioaddr); | 241 | ata_sff_std_ports(ioaddr); |
273 | 242 | ||
274 | ata_port_desc(host->ports[1], | 243 | ata_port_desc(host->ports[1], |
275 | "cmd 0x%x ctl 0x%x", cmd_port[1], ctl_port[1]); | 244 | "cmd 0x%x ctl 0x%x", cmd_port[1], ctl_port[1]); |
@@ -289,7 +258,7 @@ static int __devinit cs5520_init_one(struct pci_dev *pdev, const struct pci_devi | |||
289 | continue; | 258 | continue; |
290 | 259 | ||
291 | rc = devm_request_irq(&pdev->dev, irq[ap->port_no], | 260 | rc = devm_request_irq(&pdev->dev, irq[ap->port_no], |
292 | ata_interrupt, 0, DRV_NAME, host); | 261 | ata_sff_interrupt, 0, DRV_NAME, host); |
293 | if (rc) | 262 | if (rc) |
294 | return rc; | 263 | return rc; |
295 | 264 | ||
@@ -310,11 +279,20 @@ static int __devinit cs5520_init_one(struct pci_dev *pdev, const struct pci_devi | |||
310 | 279 | ||
311 | static int cs5520_reinit_one(struct pci_dev *pdev) | 280 | static int cs5520_reinit_one(struct pci_dev *pdev) |
312 | { | 281 | { |
282 | struct ata_host *host = dev_get_drvdata(&pdev->dev); | ||
313 | u8 pcicfg; | 283 | u8 pcicfg; |
284 | int rc; | ||
285 | |||
286 | rc = ata_pci_device_do_resume(pdev); | ||
287 | if (rc) | ||
288 | return rc; | ||
289 | |||
314 | pci_read_config_byte(pdev, 0x60, &pcicfg); | 290 | pci_read_config_byte(pdev, 0x60, &pcicfg); |
315 | if ((pcicfg & 0x40) == 0) | 291 | if ((pcicfg & 0x40) == 0) |
316 | pci_write_config_byte(pdev, 0x60, pcicfg | 0x40); | 292 | pci_write_config_byte(pdev, 0x60, pcicfg | 0x40); |
317 | return ata_pci_device_resume(pdev); | 293 | |
294 | ata_host_resume(host); | ||
295 | return 0; | ||
318 | } | 296 | } |
319 | 297 | ||
320 | /** | 298 | /** |
diff --git a/drivers/ata/pata_cs5530.c b/drivers/ata/pata_cs5530.c index e1818fdd9159..744beebaaf49 100644 --- a/drivers/ata/pata_cs5530.c +++ b/drivers/ata/pata_cs5530.c | |||
@@ -133,7 +133,7 @@ static void cs5530_set_dmamode(struct ata_port *ap, struct ata_device *adev) | |||
133 | } | 133 | } |
134 | 134 | ||
135 | /** | 135 | /** |
136 | * cs5530_qc_issue_prot - command issue | 136 | * cs5530_qc_issue - command issue |
137 | * @qc: command pending | 137 | * @qc: command pending |
138 | * | 138 | * |
139 | * Called when the libata layer is about to issue a command. We wrap | 139 | * Called when the libata layer is about to issue a command. We wrap |
@@ -142,7 +142,7 @@ static void cs5530_set_dmamode(struct ata_port *ap, struct ata_device *adev) | |||
142 | * one MWDMA/UDMA bit. | 142 | * one MWDMA/UDMA bit. |
143 | */ | 143 | */ |
144 | 144 | ||
145 | static unsigned int cs5530_qc_issue_prot(struct ata_queued_cmd *qc) | 145 | static unsigned int cs5530_qc_issue(struct ata_queued_cmd *qc) |
146 | { | 146 | { |
147 | struct ata_port *ap = qc->ap; | 147 | struct ata_port *ap = qc->ap; |
148 | struct ata_device *adev = qc->dev; | 148 | struct ata_device *adev = qc->dev; |
@@ -157,59 +157,23 @@ static unsigned int cs5530_qc_issue_prot(struct ata_queued_cmd *qc) | |||
157 | cs5530_set_dmamode(ap, adev); | 157 | cs5530_set_dmamode(ap, adev); |
158 | } | 158 | } |
159 | 159 | ||
160 | return ata_qc_issue_prot(qc); | 160 | return ata_sff_qc_issue(qc); |
161 | } | 161 | } |
162 | 162 | ||
163 | static struct scsi_host_template cs5530_sht = { | 163 | static struct scsi_host_template cs5530_sht = { |
164 | .module = THIS_MODULE, | 164 | ATA_BMDMA_SHT(DRV_NAME), |
165 | .name = DRV_NAME, | 165 | .sg_tablesize = LIBATA_DUMB_MAX_PRD, |
166 | .ioctl = ata_scsi_ioctl, | ||
167 | .queuecommand = ata_scsi_queuecmd, | ||
168 | .can_queue = ATA_DEF_QUEUE, | ||
169 | .this_id = ATA_SHT_THIS_ID, | ||
170 | .sg_tablesize = LIBATA_DUMB_MAX_PRD, | ||
171 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
172 | .emulated = ATA_SHT_EMULATED, | ||
173 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
174 | .proc_name = DRV_NAME, | ||
175 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
176 | .slave_configure = ata_scsi_slave_config, | ||
177 | .slave_destroy = ata_scsi_slave_destroy, | ||
178 | .bios_param = ata_std_bios_param, | ||
179 | }; | 166 | }; |
180 | 167 | ||
181 | static struct ata_port_operations cs5530_port_ops = { | 168 | static struct ata_port_operations cs5530_port_ops = { |
182 | .set_piomode = cs5530_set_piomode, | 169 | .inherits = &ata_bmdma_port_ops, |
183 | .set_dmamode = cs5530_set_dmamode, | ||
184 | .mode_filter = ata_pci_default_filter, | ||
185 | |||
186 | .tf_load = ata_tf_load, | ||
187 | .tf_read = ata_tf_read, | ||
188 | .check_status = ata_check_status, | ||
189 | .exec_command = ata_exec_command, | ||
190 | .dev_select = ata_std_dev_select, | ||
191 | |||
192 | .bmdma_setup = ata_bmdma_setup, | ||
193 | .bmdma_start = ata_bmdma_start, | ||
194 | .bmdma_stop = ata_bmdma_stop, | ||
195 | .bmdma_status = ata_bmdma_status, | ||
196 | |||
197 | .freeze = ata_bmdma_freeze, | ||
198 | .thaw = ata_bmdma_thaw, | ||
199 | .error_handler = ata_bmdma_error_handler, | ||
200 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
201 | .cable_detect = ata_cable_40wire, | ||
202 | |||
203 | .qc_prep = ata_dumb_qc_prep, | ||
204 | .qc_issue = cs5530_qc_issue_prot, | ||
205 | 170 | ||
206 | .data_xfer = ata_data_xfer, | 171 | .qc_prep = ata_sff_dumb_qc_prep, |
172 | .qc_issue = cs5530_qc_issue, | ||
207 | 173 | ||
208 | .irq_handler = ata_interrupt, | 174 | .cable_detect = ata_cable_40wire, |
209 | .irq_clear = ata_bmdma_irq_clear, | 175 | .set_piomode = cs5530_set_piomode, |
210 | .irq_on = ata_irq_on, | 176 | .set_dmamode = cs5530_set_dmamode, |
211 | |||
212 | .port_start = ata_sff_port_start, | ||
213 | }; | 177 | }; |
214 | 178 | ||
215 | static const struct dmi_system_id palmax_dmi_table[] = { | 179 | static const struct dmi_system_id palmax_dmi_table[] = { |
@@ -334,7 +298,6 @@ fail_put: | |||
334 | static int cs5530_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | 298 | static int cs5530_init_one(struct pci_dev *pdev, const struct pci_device_id *id) |
335 | { | 299 | { |
336 | static const struct ata_port_info info = { | 300 | static const struct ata_port_info info = { |
337 | .sht = &cs5530_sht, | ||
338 | .flags = ATA_FLAG_SLAVE_POSS, | 301 | .flags = ATA_FLAG_SLAVE_POSS, |
339 | .pio_mask = 0x1f, | 302 | .pio_mask = 0x1f, |
340 | .mwdma_mask = 0x07, | 303 | .mwdma_mask = 0x07, |
@@ -343,12 +306,16 @@ static int cs5530_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
343 | }; | 306 | }; |
344 | /* The docking connector doesn't do UDMA, and it seems not MWDMA */ | 307 | /* The docking connector doesn't do UDMA, and it seems not MWDMA */ |
345 | static const struct ata_port_info info_palmax_secondary = { | 308 | static const struct ata_port_info info_palmax_secondary = { |
346 | .sht = &cs5530_sht, | ||
347 | .flags = ATA_FLAG_SLAVE_POSS, | 309 | .flags = ATA_FLAG_SLAVE_POSS, |
348 | .pio_mask = 0x1f, | 310 | .pio_mask = 0x1f, |
349 | .port_ops = &cs5530_port_ops | 311 | .port_ops = &cs5530_port_ops |
350 | }; | 312 | }; |
351 | const struct ata_port_info *ppi[] = { &info, NULL }; | 313 | const struct ata_port_info *ppi[] = { &info, NULL }; |
314 | int rc; | ||
315 | |||
316 | rc = pcim_enable_device(pdev); | ||
317 | if (rc) | ||
318 | return rc; | ||
352 | 319 | ||
353 | /* Chip initialisation */ | 320 | /* Chip initialisation */ |
354 | if (cs5530_init_chip()) | 321 | if (cs5530_init_chip()) |
@@ -358,16 +325,25 @@ static int cs5530_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
358 | ppi[1] = &info_palmax_secondary; | 325 | ppi[1] = &info_palmax_secondary; |
359 | 326 | ||
360 | /* Now kick off ATA set up */ | 327 | /* Now kick off ATA set up */ |
361 | return ata_pci_init_one(pdev, ppi); | 328 | return ata_pci_sff_init_one(pdev, ppi, &cs5530_sht, NULL); |
362 | } | 329 | } |
363 | 330 | ||
364 | #ifdef CONFIG_PM | 331 | #ifdef CONFIG_PM |
365 | static int cs5530_reinit_one(struct pci_dev *pdev) | 332 | static int cs5530_reinit_one(struct pci_dev *pdev) |
366 | { | 333 | { |
334 | struct ata_host *host = dev_get_drvdata(&pdev->dev); | ||
335 | int rc; | ||
336 | |||
337 | rc = ata_pci_device_do_resume(pdev); | ||
338 | if (rc) | ||
339 | return rc; | ||
340 | |||
367 | /* If we fail on resume we are doomed */ | 341 | /* If we fail on resume we are doomed */ |
368 | if (cs5530_init_chip()) | 342 | if (cs5530_init_chip()) |
369 | BUG(); | 343 | return -EIO; |
370 | return ata_pci_device_resume(pdev); | 344 | |
345 | ata_host_resume(host); | ||
346 | return 0; | ||
371 | } | 347 | } |
372 | #endif /* CONFIG_PM */ | 348 | #endif /* CONFIG_PM */ |
373 | 349 | ||
diff --git a/drivers/ata/pata_cs5535.c b/drivers/ata/pata_cs5535.c index 01324530d052..f1b6556f0483 100644 --- a/drivers/ata/pata_cs5535.c +++ b/drivers/ata/pata_cs5535.c | |||
@@ -158,55 +158,14 @@ static void cs5535_set_dmamode(struct ata_port *ap, struct ata_device *adev) | |||
158 | } | 158 | } |
159 | 159 | ||
160 | static struct scsi_host_template cs5535_sht = { | 160 | static struct scsi_host_template cs5535_sht = { |
161 | .module = THIS_MODULE, | 161 | ATA_BMDMA_SHT(DRV_NAME), |
162 | .name = DRV_NAME, | ||
163 | .ioctl = ata_scsi_ioctl, | ||
164 | .queuecommand = ata_scsi_queuecmd, | ||
165 | .can_queue = ATA_DEF_QUEUE, | ||
166 | .this_id = ATA_SHT_THIS_ID, | ||
167 | .sg_tablesize = LIBATA_MAX_PRD, | ||
168 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
169 | .emulated = ATA_SHT_EMULATED, | ||
170 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
171 | .proc_name = DRV_NAME, | ||
172 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
173 | .slave_configure = ata_scsi_slave_config, | ||
174 | .slave_destroy = ata_scsi_slave_destroy, | ||
175 | .bios_param = ata_std_bios_param, | ||
176 | }; | 162 | }; |
177 | 163 | ||
178 | static struct ata_port_operations cs5535_port_ops = { | 164 | static struct ata_port_operations cs5535_port_ops = { |
165 | .inherits = &ata_bmdma_port_ops, | ||
166 | .cable_detect = cs5535_cable_detect, | ||
179 | .set_piomode = cs5535_set_piomode, | 167 | .set_piomode = cs5535_set_piomode, |
180 | .set_dmamode = cs5535_set_dmamode, | 168 | .set_dmamode = cs5535_set_dmamode, |
181 | .mode_filter = ata_pci_default_filter, | ||
182 | |||
183 | .tf_load = ata_tf_load, | ||
184 | .tf_read = ata_tf_read, | ||
185 | .check_status = ata_check_status, | ||
186 | .exec_command = ata_exec_command, | ||
187 | .dev_select = ata_std_dev_select, | ||
188 | |||
189 | .freeze = ata_bmdma_freeze, | ||
190 | .thaw = ata_bmdma_thaw, | ||
191 | .error_handler = ata_bmdma_error_handler, | ||
192 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
193 | .cable_detect = cs5535_cable_detect, | ||
194 | |||
195 | .bmdma_setup = ata_bmdma_setup, | ||
196 | .bmdma_start = ata_bmdma_start, | ||
197 | .bmdma_stop = ata_bmdma_stop, | ||
198 | .bmdma_status = ata_bmdma_status, | ||
199 | |||
200 | .qc_prep = ata_qc_prep, | ||
201 | .qc_issue = ata_qc_issue_prot, | ||
202 | |||
203 | .data_xfer = ata_data_xfer, | ||
204 | |||
205 | .irq_handler = ata_interrupt, | ||
206 | .irq_clear = ata_bmdma_irq_clear, | ||
207 | .irq_on = ata_irq_on, | ||
208 | |||
209 | .port_start = ata_sff_port_start, | ||
210 | }; | 169 | }; |
211 | 170 | ||
212 | /** | 171 | /** |
@@ -222,7 +181,6 @@ static struct ata_port_operations cs5535_port_ops = { | |||
222 | static int cs5535_init_one(struct pci_dev *dev, const struct pci_device_id *id) | 181 | static int cs5535_init_one(struct pci_dev *dev, const struct pci_device_id *id) |
223 | { | 182 | { |
224 | static const struct ata_port_info info = { | 183 | static const struct ata_port_info info = { |
225 | .sht = &cs5535_sht, | ||
226 | .flags = ATA_FLAG_SLAVE_POSS, | 184 | .flags = ATA_FLAG_SLAVE_POSS, |
227 | .pio_mask = 0x1f, | 185 | .pio_mask = 0x1f, |
228 | .mwdma_mask = 0x07, | 186 | .mwdma_mask = 0x07, |
@@ -241,7 +199,7 @@ static int cs5535_init_one(struct pci_dev *dev, const struct pci_device_id *id) | |||
241 | rdmsr(ATAC_CH0D1_PIO, timings, dummy); | 199 | rdmsr(ATAC_CH0D1_PIO, timings, dummy); |
242 | if (CS5535_BAD_PIO(timings)) | 200 | if (CS5535_BAD_PIO(timings)) |
243 | wrmsr(ATAC_CH0D1_PIO, 0xF7F4F7F4UL, 0); | 201 | wrmsr(ATAC_CH0D1_PIO, 0xF7F4F7F4UL, 0); |
244 | return ata_pci_init_one(dev, ppi); | 202 | return ata_pci_sff_init_one(dev, ppi, &cs5535_sht, NULL); |
245 | } | 203 | } |
246 | 204 | ||
247 | static const struct pci_device_id cs5535[] = { | 205 | static const struct pci_device_id cs5535[] = { |
diff --git a/drivers/ata/pata_cs5536.c b/drivers/ata/pata_cs5536.c index 1c4ff9b52b5c..73f8332cb679 100644 --- a/drivers/ata/pata_cs5536.c +++ b/drivers/ata/pata_cs5536.c | |||
@@ -221,55 +221,14 @@ static void cs5536_set_dmamode(struct ata_port *ap, struct ata_device *adev) | |||
221 | } | 221 | } |
222 | 222 | ||
223 | static struct scsi_host_template cs5536_sht = { | 223 | static struct scsi_host_template cs5536_sht = { |
224 | .module = THIS_MODULE, | 224 | ATA_BMDMA_SHT(DRV_NAME), |
225 | .name = DRV_NAME, | ||
226 | .ioctl = ata_scsi_ioctl, | ||
227 | .queuecommand = ata_scsi_queuecmd, | ||
228 | .can_queue = ATA_DEF_QUEUE, | ||
229 | .this_id = ATA_SHT_THIS_ID, | ||
230 | .sg_tablesize = LIBATA_MAX_PRD, | ||
231 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
232 | .emulated = ATA_SHT_EMULATED, | ||
233 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
234 | .proc_name = DRV_NAME, | ||
235 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
236 | .slave_configure = ata_scsi_slave_config, | ||
237 | .slave_destroy = ata_scsi_slave_destroy, | ||
238 | .bios_param = ata_std_bios_param, | ||
239 | }; | 225 | }; |
240 | 226 | ||
241 | static struct ata_port_operations cs5536_port_ops = { | 227 | static struct ata_port_operations cs5536_port_ops = { |
228 | .inherits = &ata_bmdma_port_ops, | ||
229 | .cable_detect = cs5536_cable_detect, | ||
242 | .set_piomode = cs5536_set_piomode, | 230 | .set_piomode = cs5536_set_piomode, |
243 | .set_dmamode = cs5536_set_dmamode, | 231 | .set_dmamode = cs5536_set_dmamode, |
244 | .mode_filter = ata_pci_default_filter, | ||
245 | |||
246 | .tf_load = ata_tf_load, | ||
247 | .tf_read = ata_tf_read, | ||
248 | .check_status = ata_check_status, | ||
249 | .exec_command = ata_exec_command, | ||
250 | .dev_select = ata_std_dev_select, | ||
251 | |||
252 | .freeze = ata_bmdma_freeze, | ||
253 | .thaw = ata_bmdma_thaw, | ||
254 | .error_handler = ata_bmdma_error_handler, | ||
255 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
256 | .cable_detect = cs5536_cable_detect, | ||
257 | |||
258 | .bmdma_setup = ata_bmdma_setup, | ||
259 | .bmdma_start = ata_bmdma_start, | ||
260 | .bmdma_stop = ata_bmdma_stop, | ||
261 | .bmdma_status = ata_bmdma_status, | ||
262 | |||
263 | .qc_prep = ata_qc_prep, | ||
264 | .qc_issue = ata_qc_issue_prot, | ||
265 | |||
266 | .data_xfer = ata_data_xfer, | ||
267 | |||
268 | .irq_handler = ata_interrupt, | ||
269 | .irq_clear = ata_bmdma_irq_clear, | ||
270 | .irq_on = ata_irq_on, | ||
271 | |||
272 | .port_start = ata_port_start, | ||
273 | }; | 232 | }; |
274 | 233 | ||
275 | /** | 234 | /** |
@@ -282,7 +241,6 @@ static struct ata_port_operations cs5536_port_ops = { | |||
282 | static int cs5536_init_one(struct pci_dev *dev, const struct pci_device_id *id) | 241 | static int cs5536_init_one(struct pci_dev *dev, const struct pci_device_id *id) |
283 | { | 242 | { |
284 | static const struct ata_port_info info = { | 243 | static const struct ata_port_info info = { |
285 | .sht = &cs5536_sht, | ||
286 | .flags = ATA_FLAG_SLAVE_POSS, | 244 | .flags = ATA_FLAG_SLAVE_POSS, |
287 | .pio_mask = 0x1f, | 245 | .pio_mask = 0x1f, |
288 | .mwdma_mask = 0x07, | 246 | .mwdma_mask = 0x07, |
@@ -303,7 +261,7 @@ static int cs5536_init_one(struct pci_dev *dev, const struct pci_device_id *id) | |||
303 | return -ENODEV; | 261 | return -ENODEV; |
304 | } | 262 | } |
305 | 263 | ||
306 | return ata_pci_init_one(dev, ppi); | 264 | return ata_pci_sff_init_one(dev, ppi, &cs5536_sht, NULL); |
307 | } | 265 | } |
308 | 266 | ||
309 | static const struct pci_device_id cs5536[] = { | 267 | static const struct pci_device_id cs5536[] = { |
diff --git a/drivers/ata/pata_cypress.c b/drivers/ata/pata_cypress.c index fc5f9c4e5d87..a9c3218e22fd 100644 --- a/drivers/ata/pata_cypress.c +++ b/drivers/ata/pata_cypress.c | |||
@@ -110,61 +110,19 @@ static void cy82c693_set_dmamode(struct ata_port *ap, struct ata_device *adev) | |||
110 | } | 110 | } |
111 | 111 | ||
112 | static struct scsi_host_template cy82c693_sht = { | 112 | static struct scsi_host_template cy82c693_sht = { |
113 | .module = THIS_MODULE, | 113 | ATA_BMDMA_SHT(DRV_NAME), |
114 | .name = DRV_NAME, | ||
115 | .ioctl = ata_scsi_ioctl, | ||
116 | .queuecommand = ata_scsi_queuecmd, | ||
117 | .can_queue = ATA_DEF_QUEUE, | ||
118 | .this_id = ATA_SHT_THIS_ID, | ||
119 | .sg_tablesize = LIBATA_MAX_PRD, | ||
120 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
121 | .emulated = ATA_SHT_EMULATED, | ||
122 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
123 | .proc_name = DRV_NAME, | ||
124 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
125 | .slave_configure = ata_scsi_slave_config, | ||
126 | .slave_destroy = ata_scsi_slave_destroy, | ||
127 | .bios_param = ata_std_bios_param, | ||
128 | }; | 114 | }; |
129 | 115 | ||
130 | static struct ata_port_operations cy82c693_port_ops = { | 116 | static struct ata_port_operations cy82c693_port_ops = { |
117 | .inherits = &ata_bmdma_port_ops, | ||
118 | .cable_detect = ata_cable_40wire, | ||
131 | .set_piomode = cy82c693_set_piomode, | 119 | .set_piomode = cy82c693_set_piomode, |
132 | .set_dmamode = cy82c693_set_dmamode, | 120 | .set_dmamode = cy82c693_set_dmamode, |
133 | .mode_filter = ata_pci_default_filter, | ||
134 | |||
135 | .tf_load = ata_tf_load, | ||
136 | .tf_read = ata_tf_read, | ||
137 | .check_status = ata_check_status, | ||
138 | .exec_command = ata_exec_command, | ||
139 | .dev_select = ata_std_dev_select, | ||
140 | |||
141 | .freeze = ata_bmdma_freeze, | ||
142 | .thaw = ata_bmdma_thaw, | ||
143 | .error_handler = ata_bmdma_error_handler, | ||
144 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
145 | .cable_detect = ata_cable_40wire, | ||
146 | |||
147 | .bmdma_setup = ata_bmdma_setup, | ||
148 | .bmdma_start = ata_bmdma_start, | ||
149 | .bmdma_stop = ata_bmdma_stop, | ||
150 | .bmdma_status = ata_bmdma_status, | ||
151 | |||
152 | .qc_prep = ata_qc_prep, | ||
153 | .qc_issue = ata_qc_issue_prot, | ||
154 | |||
155 | .data_xfer = ata_data_xfer, | ||
156 | |||
157 | .irq_handler = ata_interrupt, | ||
158 | .irq_clear = ata_bmdma_irq_clear, | ||
159 | .irq_on = ata_irq_on, | ||
160 | |||
161 | .port_start = ata_sff_port_start, | ||
162 | }; | 121 | }; |
163 | 122 | ||
164 | static int cy82c693_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | 123 | static int cy82c693_init_one(struct pci_dev *pdev, const struct pci_device_id *id) |
165 | { | 124 | { |
166 | static const struct ata_port_info info = { | 125 | static const struct ata_port_info info = { |
167 | .sht = &cy82c693_sht, | ||
168 | .flags = ATA_FLAG_SLAVE_POSS, | 126 | .flags = ATA_FLAG_SLAVE_POSS, |
169 | .pio_mask = 0x1f, | 127 | .pio_mask = 0x1f, |
170 | .mwdma_mask = 0x07, | 128 | .mwdma_mask = 0x07, |
@@ -178,7 +136,7 @@ static int cy82c693_init_one(struct pci_dev *pdev, const struct pci_device_id *i | |||
178 | if (PCI_FUNC(pdev->devfn) != 1) | 136 | if (PCI_FUNC(pdev->devfn) != 1) |
179 | return -ENODEV; | 137 | return -ENODEV; |
180 | 138 | ||
181 | return ata_pci_init_one(pdev, ppi); | 139 | return ata_pci_sff_init_one(pdev, ppi, &cy82c693_sht, NULL); |
182 | } | 140 | } |
183 | 141 | ||
184 | static const struct pci_device_id cy82c693[] = { | 142 | static const struct pci_device_id cy82c693[] = { |
diff --git a/drivers/ata/pata_efar.c b/drivers/ata/pata_efar.c index dc33220fe5b2..9fba82976ba6 100644 --- a/drivers/ata/pata_efar.c +++ b/drivers/ata/pata_efar.c | |||
@@ -45,20 +45,7 @@ static int efar_pre_reset(struct ata_link *link, unsigned long deadline) | |||
45 | if (!pci_test_config_bits(pdev, &efar_enable_bits[ap->port_no])) | 45 | if (!pci_test_config_bits(pdev, &efar_enable_bits[ap->port_no])) |
46 | return -ENOENT; | 46 | return -ENOENT; |
47 | 47 | ||
48 | return ata_std_prereset(link, deadline); | 48 | return ata_sff_prereset(link, deadline); |
49 | } | ||
50 | |||
51 | /** | ||
52 | * efar_probe_reset - Probe specified port on PATA host controller | ||
53 | * @ap: Port to probe | ||
54 | * | ||
55 | * LOCKING: | ||
56 | * None (inherited from caller). | ||
57 | */ | ||
58 | |||
59 | static void efar_error_handler(struct ata_port *ap) | ||
60 | { | ||
61 | ata_bmdma_drive_eh(ap, efar_pre_reset, ata_std_softreset, NULL, ata_std_postreset); | ||
62 | } | 49 | } |
63 | 50 | ||
64 | /** | 51 | /** |
@@ -233,53 +220,15 @@ static void efar_set_dmamode (struct ata_port *ap, struct ata_device *adev) | |||
233 | } | 220 | } |
234 | 221 | ||
235 | static struct scsi_host_template efar_sht = { | 222 | static struct scsi_host_template efar_sht = { |
236 | .module = THIS_MODULE, | 223 | ATA_BMDMA_SHT(DRV_NAME), |
237 | .name = DRV_NAME, | ||
238 | .ioctl = ata_scsi_ioctl, | ||
239 | .queuecommand = ata_scsi_queuecmd, | ||
240 | .can_queue = ATA_DEF_QUEUE, | ||
241 | .this_id = ATA_SHT_THIS_ID, | ||
242 | .sg_tablesize = LIBATA_MAX_PRD, | ||
243 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
244 | .emulated = ATA_SHT_EMULATED, | ||
245 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
246 | .proc_name = DRV_NAME, | ||
247 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
248 | .slave_configure = ata_scsi_slave_config, | ||
249 | .slave_destroy = ata_scsi_slave_destroy, | ||
250 | .bios_param = ata_std_bios_param, | ||
251 | }; | 224 | }; |
252 | 225 | ||
253 | static const struct ata_port_operations efar_ops = { | 226 | static struct ata_port_operations efar_ops = { |
227 | .inherits = &ata_bmdma_port_ops, | ||
228 | .cable_detect = efar_cable_detect, | ||
254 | .set_piomode = efar_set_piomode, | 229 | .set_piomode = efar_set_piomode, |
255 | .set_dmamode = efar_set_dmamode, | 230 | .set_dmamode = efar_set_dmamode, |
256 | .mode_filter = ata_pci_default_filter, | 231 | .prereset = efar_pre_reset, |
257 | |||
258 | .tf_load = ata_tf_load, | ||
259 | .tf_read = ata_tf_read, | ||
260 | .check_status = ata_check_status, | ||
261 | .exec_command = ata_exec_command, | ||
262 | .dev_select = ata_std_dev_select, | ||
263 | |||
264 | .freeze = ata_bmdma_freeze, | ||
265 | .thaw = ata_bmdma_thaw, | ||
266 | .error_handler = efar_error_handler, | ||
267 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
268 | .cable_detect = efar_cable_detect, | ||
269 | |||
270 | .bmdma_setup = ata_bmdma_setup, | ||
271 | .bmdma_start = ata_bmdma_start, | ||
272 | .bmdma_stop = ata_bmdma_stop, | ||
273 | .bmdma_status = ata_bmdma_status, | ||
274 | .qc_prep = ata_qc_prep, | ||
275 | .qc_issue = ata_qc_issue_prot, | ||
276 | .data_xfer = ata_data_xfer, | ||
277 | |||
278 | .irq_handler = ata_interrupt, | ||
279 | .irq_clear = ata_bmdma_irq_clear, | ||
280 | .irq_on = ata_irq_on, | ||
281 | |||
282 | .port_start = ata_sff_port_start, | ||
283 | }; | 232 | }; |
284 | 233 | ||
285 | 234 | ||
@@ -301,7 +250,6 @@ static int efar_init_one (struct pci_dev *pdev, const struct pci_device_id *ent) | |||
301 | { | 250 | { |
302 | static int printed_version; | 251 | static int printed_version; |
303 | static const struct ata_port_info info = { | 252 | static const struct ata_port_info info = { |
304 | .sht = &efar_sht, | ||
305 | .flags = ATA_FLAG_SLAVE_POSS, | 253 | .flags = ATA_FLAG_SLAVE_POSS, |
306 | .pio_mask = 0x1f, /* pio0-4 */ | 254 | .pio_mask = 0x1f, /* pio0-4 */ |
307 | .mwdma_mask = 0x07, /* mwdma1-2 */ | 255 | .mwdma_mask = 0x07, /* mwdma1-2 */ |
@@ -314,7 +262,7 @@ static int efar_init_one (struct pci_dev *pdev, const struct pci_device_id *ent) | |||
314 | dev_printk(KERN_DEBUG, &pdev->dev, | 262 | dev_printk(KERN_DEBUG, &pdev->dev, |
315 | "version " DRV_VERSION "\n"); | 263 | "version " DRV_VERSION "\n"); |
316 | 264 | ||
317 | return ata_pci_init_one(pdev, ppi); | 265 | return ata_pci_sff_init_one(pdev, ppi, &efar_sht, NULL); |
318 | } | 266 | } |
319 | 267 | ||
320 | static const struct pci_device_id efar_pci_tbl[] = { | 268 | static const struct pci_device_id efar_pci_tbl[] = { |
diff --git a/drivers/ata/pata_hpt366.c b/drivers/ata/pata_hpt366.c index a742efa0da2b..f2b83eabc7c7 100644 --- a/drivers/ata/pata_hpt366.c +++ b/drivers/ata/pata_hpt366.c | |||
@@ -184,7 +184,7 @@ static unsigned long hpt366_filter(struct ata_device *adev, unsigned long mask) | |||
184 | if (hpt_dma_blacklisted(adev, "UDMA4", bad_ata66_4)) | 184 | if (hpt_dma_blacklisted(adev, "UDMA4", bad_ata66_4)) |
185 | mask &= ~(0xF0 << ATA_SHIFT_UDMA); | 185 | mask &= ~(0xF0 << ATA_SHIFT_UDMA); |
186 | } | 186 | } |
187 | return ata_pci_default_filter(adev, mask); | 187 | return ata_bmdma_mode_filter(adev, mask); |
188 | } | 188 | } |
189 | 189 | ||
190 | /** | 190 | /** |
@@ -290,21 +290,7 @@ static void hpt366_set_dmamode(struct ata_port *ap, struct ata_device *adev) | |||
290 | } | 290 | } |
291 | 291 | ||
292 | static struct scsi_host_template hpt36x_sht = { | 292 | static struct scsi_host_template hpt36x_sht = { |
293 | .module = THIS_MODULE, | 293 | ATA_BMDMA_SHT(DRV_NAME), |
294 | .name = DRV_NAME, | ||
295 | .ioctl = ata_scsi_ioctl, | ||
296 | .queuecommand = ata_scsi_queuecmd, | ||
297 | .can_queue = ATA_DEF_QUEUE, | ||
298 | .this_id = ATA_SHT_THIS_ID, | ||
299 | .sg_tablesize = LIBATA_MAX_PRD, | ||
300 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
301 | .emulated = ATA_SHT_EMULATED, | ||
302 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
303 | .proc_name = DRV_NAME, | ||
304 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
305 | .slave_configure = ata_scsi_slave_config, | ||
306 | .slave_destroy = ata_scsi_slave_destroy, | ||
307 | .bios_param = ata_std_bios_param, | ||
308 | }; | 294 | }; |
309 | 295 | ||
310 | /* | 296 | /* |
@@ -312,37 +298,11 @@ static struct scsi_host_template hpt36x_sht = { | |||
312 | */ | 298 | */ |
313 | 299 | ||
314 | static struct ata_port_operations hpt366_port_ops = { | 300 | static struct ata_port_operations hpt366_port_ops = { |
301 | .inherits = &ata_bmdma_port_ops, | ||
302 | .cable_detect = hpt36x_cable_detect, | ||
303 | .mode_filter = hpt366_filter, | ||
315 | .set_piomode = hpt366_set_piomode, | 304 | .set_piomode = hpt366_set_piomode, |
316 | .set_dmamode = hpt366_set_dmamode, | 305 | .set_dmamode = hpt366_set_dmamode, |
317 | .mode_filter = hpt366_filter, | ||
318 | |||
319 | .tf_load = ata_tf_load, | ||
320 | .tf_read = ata_tf_read, | ||
321 | .check_status = ata_check_status, | ||
322 | .exec_command = ata_exec_command, | ||
323 | .dev_select = ata_std_dev_select, | ||
324 | |||
325 | .freeze = ata_bmdma_freeze, | ||
326 | .thaw = ata_bmdma_thaw, | ||
327 | .error_handler = ata_bmdma_error_handler, | ||
328 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
329 | .cable_detect = hpt36x_cable_detect, | ||
330 | |||
331 | .bmdma_setup = ata_bmdma_setup, | ||
332 | .bmdma_start = ata_bmdma_start, | ||
333 | .bmdma_stop = ata_bmdma_stop, | ||
334 | .bmdma_status = ata_bmdma_status, | ||
335 | |||
336 | .qc_prep = ata_qc_prep, | ||
337 | .qc_issue = ata_qc_issue_prot, | ||
338 | |||
339 | .data_xfer = ata_data_xfer, | ||
340 | |||
341 | .irq_handler = ata_interrupt, | ||
342 | .irq_clear = ata_bmdma_irq_clear, | ||
343 | .irq_on = ata_irq_on, | ||
344 | |||
345 | .port_start = ata_sff_port_start, | ||
346 | }; | 306 | }; |
347 | 307 | ||
348 | /** | 308 | /** |
@@ -390,18 +350,22 @@ static void hpt36x_init_chipset(struct pci_dev *dev) | |||
390 | static int hpt36x_init_one(struct pci_dev *dev, const struct pci_device_id *id) | 350 | static int hpt36x_init_one(struct pci_dev *dev, const struct pci_device_id *id) |
391 | { | 351 | { |
392 | static const struct ata_port_info info_hpt366 = { | 352 | static const struct ata_port_info info_hpt366 = { |
393 | .sht = &hpt36x_sht, | ||
394 | .flags = ATA_FLAG_SLAVE_POSS, | 353 | .flags = ATA_FLAG_SLAVE_POSS, |
395 | .pio_mask = 0x1f, | 354 | .pio_mask = 0x1f, |
396 | .mwdma_mask = 0x07, | 355 | .mwdma_mask = 0x07, |
397 | .udma_mask = ATA_UDMA4, | 356 | .udma_mask = ATA_UDMA4, |
398 | .port_ops = &hpt366_port_ops | 357 | .port_ops = &hpt366_port_ops |
399 | }; | 358 | }; |
400 | struct ata_port_info info = info_hpt366; | 359 | const struct ata_port_info *ppi[] = { &info_hpt366, NULL }; |
401 | const struct ata_port_info *ppi[] = { &info, NULL }; | ||
402 | 360 | ||
361 | void *hpriv = NULL; | ||
403 | u32 class_rev; | 362 | u32 class_rev; |
404 | u32 reg1; | 363 | u32 reg1; |
364 | int rc; | ||
365 | |||
366 | rc = pcim_enable_device(dev); | ||
367 | if (rc) | ||
368 | return rc; | ||
405 | 369 | ||
406 | pci_read_config_dword(dev, PCI_CLASS_REVISION, &class_rev); | 370 | pci_read_config_dword(dev, PCI_CLASS_REVISION, &class_rev); |
407 | class_rev &= 0xFF; | 371 | class_rev &= 0xFF; |
@@ -419,24 +383,31 @@ static int hpt36x_init_one(struct pci_dev *dev, const struct pci_device_id *id) | |||
419 | /* info_hpt366 is safe against re-entry so we can scribble on it */ | 383 | /* info_hpt366 is safe against re-entry so we can scribble on it */ |
420 | switch((reg1 & 0x700) >> 8) { | 384 | switch((reg1 & 0x700) >> 8) { |
421 | case 5: | 385 | case 5: |
422 | info.private_data = &hpt366_40; | 386 | hpriv = &hpt366_40; |
423 | break; | 387 | break; |
424 | case 9: | 388 | case 9: |
425 | info.private_data = &hpt366_25; | 389 | hpriv = &hpt366_25; |
426 | break; | 390 | break; |
427 | default: | 391 | default: |
428 | info.private_data = &hpt366_33; | 392 | hpriv = &hpt366_33; |
429 | break; | 393 | break; |
430 | } | 394 | } |
431 | /* Now kick off ATA set up */ | 395 | /* Now kick off ATA set up */ |
432 | return ata_pci_init_one(dev, ppi); | 396 | return ata_pci_sff_init_one(dev, ppi, &hpt36x_sht, hpriv); |
433 | } | 397 | } |
434 | 398 | ||
435 | #ifdef CONFIG_PM | 399 | #ifdef CONFIG_PM |
436 | static int hpt36x_reinit_one(struct pci_dev *dev) | 400 | static int hpt36x_reinit_one(struct pci_dev *dev) |
437 | { | 401 | { |
402 | struct ata_host *host = dev_get_drvdata(&dev->dev); | ||
403 | int rc; | ||
404 | |||
405 | rc = ata_pci_device_do_resume(dev); | ||
406 | if (rc) | ||
407 | return rc; | ||
438 | hpt36x_init_chipset(dev); | 408 | hpt36x_init_chipset(dev); |
439 | return ata_pci_device_resume(dev); | 409 | ata_host_resume(host); |
410 | return 0; | ||
440 | } | 411 | } |
441 | #endif | 412 | #endif |
442 | 413 | ||
diff --git a/drivers/ata/pata_hpt37x.c b/drivers/ata/pata_hpt37x.c index 9a10878b2ad8..42163998de9a 100644 --- a/drivers/ata/pata_hpt37x.c +++ b/drivers/ata/pata_hpt37x.c | |||
@@ -283,7 +283,7 @@ static unsigned long hpt370_filter(struct ata_device *adev, unsigned long mask) | |||
283 | if (hpt_dma_blacklisted(adev, "UDMA100", bad_ata100_5)) | 283 | if (hpt_dma_blacklisted(adev, "UDMA100", bad_ata100_5)) |
284 | mask &= ~(0xE0 << ATA_SHIFT_UDMA); | 284 | mask &= ~(0xE0 << ATA_SHIFT_UDMA); |
285 | } | 285 | } |
286 | return ata_pci_default_filter(adev, mask); | 286 | return ata_bmdma_mode_filter(adev, mask); |
287 | } | 287 | } |
288 | 288 | ||
289 | /** | 289 | /** |
@@ -299,7 +299,7 @@ static unsigned long hpt370a_filter(struct ata_device *adev, unsigned long mask) | |||
299 | if (hpt_dma_blacklisted(adev, "UDMA100", bad_ata100_5)) | 299 | if (hpt_dma_blacklisted(adev, "UDMA100", bad_ata100_5)) |
300 | mask &= ~(0xE0 << ATA_SHIFT_UDMA); | 300 | mask &= ~(0xE0 << ATA_SHIFT_UDMA); |
301 | } | 301 | } |
302 | return ata_pci_default_filter(adev, mask); | 302 | return ata_bmdma_mode_filter(adev, mask); |
303 | } | 303 | } |
304 | 304 | ||
305 | /** | 305 | /** |
@@ -338,22 +338,10 @@ static int hpt37x_pre_reset(struct ata_link *link, unsigned long deadline) | |||
338 | pci_write_config_byte(pdev, 0x50 + 4 * ap->port_no, 0x37); | 338 | pci_write_config_byte(pdev, 0x50 + 4 * ap->port_no, 0x37); |
339 | udelay(100); | 339 | udelay(100); |
340 | 340 | ||
341 | return ata_std_prereset(link, deadline); | 341 | return ata_sff_prereset(link, deadline); |
342 | } | 342 | } |
343 | 343 | ||
344 | /** | 344 | static int hpt374_fn1_pre_reset(struct ata_link *link, unsigned long deadline) |
345 | * hpt37x_error_handler - reset the hpt374 | ||
346 | * @ap: ATA port to reset | ||
347 | * | ||
348 | * Perform probe for HPT37x, except for HPT374 channel 2 | ||
349 | */ | ||
350 | |||
351 | static void hpt37x_error_handler(struct ata_port *ap) | ||
352 | { | ||
353 | ata_bmdma_drive_eh(ap, hpt37x_pre_reset, ata_std_softreset, NULL, ata_std_postreset); | ||
354 | } | ||
355 | |||
356 | static int hpt374_pre_reset(struct ata_link *link, unsigned long deadline) | ||
357 | { | 345 | { |
358 | static const struct pci_bits hpt37x_enable_bits[] = { | 346 | static const struct pci_bits hpt37x_enable_bits[] = { |
359 | { 0x50, 1, 0x04, 0x04 }, | 347 | { 0x50, 1, 0x04, 0x04 }, |
@@ -386,26 +374,7 @@ static int hpt374_pre_reset(struct ata_link *link, unsigned long deadline) | |||
386 | pci_write_config_byte(pdev, 0x50 + 4 * ap->port_no, 0x37); | 374 | pci_write_config_byte(pdev, 0x50 + 4 * ap->port_no, 0x37); |
387 | udelay(100); | 375 | udelay(100); |
388 | 376 | ||
389 | return ata_std_prereset(link, deadline); | 377 | return ata_sff_prereset(link, deadline); |
390 | } | ||
391 | |||
392 | /** | ||
393 | * hpt374_error_handler - reset the hpt374 | ||
394 | * @classes: | ||
395 | * | ||
396 | * The 374 cable detect is a little different due to the extra | ||
397 | * channels. The function 0 channels work like usual but function 1 | ||
398 | * is special | ||
399 | */ | ||
400 | |||
401 | static void hpt374_error_handler(struct ata_port *ap) | ||
402 | { | ||
403 | struct pci_dev *pdev = to_pci_dev(ap->host->dev); | ||
404 | |||
405 | if (!(PCI_FUNC(pdev->devfn) & 1)) | ||
406 | hpt37x_error_handler(ap); | ||
407 | else | ||
408 | ata_bmdma_drive_eh(ap, hpt374_pre_reset, ata_std_softreset, NULL, ata_std_postreset); | ||
409 | } | 378 | } |
410 | 379 | ||
411 | /** | 380 | /** |
@@ -619,21 +588,7 @@ static void hpt37x_bmdma_stop(struct ata_queued_cmd *qc) | |||
619 | 588 | ||
620 | 589 | ||
621 | static struct scsi_host_template hpt37x_sht = { | 590 | static struct scsi_host_template hpt37x_sht = { |
622 | .module = THIS_MODULE, | 591 | ATA_BMDMA_SHT(DRV_NAME), |
623 | .name = DRV_NAME, | ||
624 | .ioctl = ata_scsi_ioctl, | ||
625 | .queuecommand = ata_scsi_queuecmd, | ||
626 | .can_queue = ATA_DEF_QUEUE, | ||
627 | .this_id = ATA_SHT_THIS_ID, | ||
628 | .sg_tablesize = LIBATA_MAX_PRD, | ||
629 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
630 | .emulated = ATA_SHT_EMULATED, | ||
631 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
632 | .proc_name = DRV_NAME, | ||
633 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
634 | .slave_configure = ata_scsi_slave_config, | ||
635 | .slave_destroy = ata_scsi_slave_destroy, | ||
636 | .bios_param = ata_std_bios_param, | ||
637 | }; | 592 | }; |
638 | 593 | ||
639 | /* | 594 | /* |
@@ -641,36 +596,15 @@ static struct scsi_host_template hpt37x_sht = { | |||
641 | */ | 596 | */ |
642 | 597 | ||
643 | static struct ata_port_operations hpt370_port_ops = { | 598 | static struct ata_port_operations hpt370_port_ops = { |
644 | .set_piomode = hpt370_set_piomode, | 599 | .inherits = &ata_bmdma_port_ops, |
645 | .set_dmamode = hpt370_set_dmamode, | ||
646 | .mode_filter = hpt370_filter, | ||
647 | |||
648 | .tf_load = ata_tf_load, | ||
649 | .tf_read = ata_tf_read, | ||
650 | .check_status = ata_check_status, | ||
651 | .exec_command = ata_exec_command, | ||
652 | .dev_select = ata_std_dev_select, | ||
653 | 600 | ||
654 | .freeze = ata_bmdma_freeze, | ||
655 | .thaw = ata_bmdma_thaw, | ||
656 | .error_handler = hpt37x_error_handler, | ||
657 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
658 | |||
659 | .bmdma_setup = ata_bmdma_setup, | ||
660 | .bmdma_start = hpt370_bmdma_start, | 601 | .bmdma_start = hpt370_bmdma_start, |
661 | .bmdma_stop = hpt370_bmdma_stop, | 602 | .bmdma_stop = hpt370_bmdma_stop, |
662 | .bmdma_status = ata_bmdma_status, | ||
663 | |||
664 | .qc_prep = ata_qc_prep, | ||
665 | .qc_issue = ata_qc_issue_prot, | ||
666 | |||
667 | .data_xfer = ata_data_xfer, | ||
668 | |||
669 | .irq_handler = ata_interrupt, | ||
670 | .irq_clear = ata_bmdma_irq_clear, | ||
671 | .irq_on = ata_irq_on, | ||
672 | 603 | ||
673 | .port_start = ata_sff_port_start, | 604 | .mode_filter = hpt370_filter, |
605 | .set_piomode = hpt370_set_piomode, | ||
606 | .set_dmamode = hpt370_set_dmamode, | ||
607 | .prereset = hpt37x_pre_reset, | ||
674 | }; | 608 | }; |
675 | 609 | ||
676 | /* | 610 | /* |
@@ -678,36 +612,8 @@ static struct ata_port_operations hpt370_port_ops = { | |||
678 | */ | 612 | */ |
679 | 613 | ||
680 | static struct ata_port_operations hpt370a_port_ops = { | 614 | static struct ata_port_operations hpt370a_port_ops = { |
681 | .set_piomode = hpt370_set_piomode, | 615 | .inherits = &hpt370_port_ops, |
682 | .set_dmamode = hpt370_set_dmamode, | ||
683 | .mode_filter = hpt370a_filter, | 616 | .mode_filter = hpt370a_filter, |
684 | |||
685 | .tf_load = ata_tf_load, | ||
686 | .tf_read = ata_tf_read, | ||
687 | .check_status = ata_check_status, | ||
688 | .exec_command = ata_exec_command, | ||
689 | .dev_select = ata_std_dev_select, | ||
690 | |||
691 | .freeze = ata_bmdma_freeze, | ||
692 | .thaw = ata_bmdma_thaw, | ||
693 | .error_handler = hpt37x_error_handler, | ||
694 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
695 | |||
696 | .bmdma_setup = ata_bmdma_setup, | ||
697 | .bmdma_start = hpt370_bmdma_start, | ||
698 | .bmdma_stop = hpt370_bmdma_stop, | ||
699 | .bmdma_status = ata_bmdma_status, | ||
700 | |||
701 | .qc_prep = ata_qc_prep, | ||
702 | .qc_issue = ata_qc_issue_prot, | ||
703 | |||
704 | .data_xfer = ata_data_xfer, | ||
705 | |||
706 | .irq_handler = ata_interrupt, | ||
707 | .irq_clear = ata_bmdma_irq_clear, | ||
708 | .irq_on = ata_irq_on, | ||
709 | |||
710 | .port_start = ata_sff_port_start, | ||
711 | }; | 617 | }; |
712 | 618 | ||
713 | /* | 619 | /* |
@@ -716,74 +622,23 @@ static struct ata_port_operations hpt370a_port_ops = { | |||
716 | */ | 622 | */ |
717 | 623 | ||
718 | static struct ata_port_operations hpt372_port_ops = { | 624 | static struct ata_port_operations hpt372_port_ops = { |
719 | .set_piomode = hpt372_set_piomode, | 625 | .inherits = &ata_bmdma_port_ops, |
720 | .set_dmamode = hpt372_set_dmamode, | ||
721 | .mode_filter = ata_pci_default_filter, | ||
722 | |||
723 | .tf_load = ata_tf_load, | ||
724 | .tf_read = ata_tf_read, | ||
725 | .check_status = ata_check_status, | ||
726 | .exec_command = ata_exec_command, | ||
727 | .dev_select = ata_std_dev_select, | ||
728 | 626 | ||
729 | .freeze = ata_bmdma_freeze, | ||
730 | .thaw = ata_bmdma_thaw, | ||
731 | .error_handler = hpt37x_error_handler, | ||
732 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
733 | |||
734 | .bmdma_setup = ata_bmdma_setup, | ||
735 | .bmdma_start = ata_bmdma_start, | ||
736 | .bmdma_stop = hpt37x_bmdma_stop, | 627 | .bmdma_stop = hpt37x_bmdma_stop, |
737 | .bmdma_status = ata_bmdma_status, | ||
738 | |||
739 | .qc_prep = ata_qc_prep, | ||
740 | .qc_issue = ata_qc_issue_prot, | ||
741 | |||
742 | .data_xfer = ata_data_xfer, | ||
743 | |||
744 | .irq_handler = ata_interrupt, | ||
745 | .irq_clear = ata_bmdma_irq_clear, | ||
746 | .irq_on = ata_irq_on, | ||
747 | 628 | ||
748 | .port_start = ata_sff_port_start, | 629 | .set_piomode = hpt372_set_piomode, |
630 | .set_dmamode = hpt372_set_dmamode, | ||
631 | .prereset = hpt37x_pre_reset, | ||
749 | }; | 632 | }; |
750 | 633 | ||
751 | /* | 634 | /* |
752 | * Configuration for HPT374. Mode setting works like 372 and friends | 635 | * Configuration for HPT374. Mode setting works like 372 and friends |
753 | * but we have a different cable detection procedure. | 636 | * but we have a different cable detection procedure for function 1. |
754 | */ | 637 | */ |
755 | 638 | ||
756 | static struct ata_port_operations hpt374_port_ops = { | 639 | static struct ata_port_operations hpt374_fn1_port_ops = { |
757 | .set_piomode = hpt372_set_piomode, | 640 | .inherits = &hpt372_port_ops, |
758 | .set_dmamode = hpt372_set_dmamode, | 641 | .prereset = hpt374_fn1_pre_reset, |
759 | .mode_filter = ata_pci_default_filter, | ||
760 | |||
761 | .tf_load = ata_tf_load, | ||
762 | .tf_read = ata_tf_read, | ||
763 | .check_status = ata_check_status, | ||
764 | .exec_command = ata_exec_command, | ||
765 | .dev_select = ata_std_dev_select, | ||
766 | |||
767 | .freeze = ata_bmdma_freeze, | ||
768 | .thaw = ata_bmdma_thaw, | ||
769 | .error_handler = hpt374_error_handler, | ||
770 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
771 | |||
772 | .bmdma_setup = ata_bmdma_setup, | ||
773 | .bmdma_start = ata_bmdma_start, | ||
774 | .bmdma_stop = hpt37x_bmdma_stop, | ||
775 | .bmdma_status = ata_bmdma_status, | ||
776 | |||
777 | .qc_prep = ata_qc_prep, | ||
778 | .qc_issue = ata_qc_issue_prot, | ||
779 | |||
780 | .data_xfer = ata_data_xfer, | ||
781 | |||
782 | .irq_handler = ata_interrupt, | ||
783 | .irq_clear = ata_bmdma_irq_clear, | ||
784 | .irq_on = ata_irq_on, | ||
785 | |||
786 | .port_start = ata_sff_port_start, | ||
787 | }; | 642 | }; |
788 | 643 | ||
789 | /** | 644 | /** |
@@ -897,7 +752,6 @@ static int hpt37x_init_one(struct pci_dev *dev, const struct pci_device_id *id) | |||
897 | { | 752 | { |
898 | /* HPT370 - UDMA100 */ | 753 | /* HPT370 - UDMA100 */ |
899 | static const struct ata_port_info info_hpt370 = { | 754 | static const struct ata_port_info info_hpt370 = { |
900 | .sht = &hpt37x_sht, | ||
901 | .flags = ATA_FLAG_SLAVE_POSS, | 755 | .flags = ATA_FLAG_SLAVE_POSS, |
902 | .pio_mask = 0x1f, | 756 | .pio_mask = 0x1f, |
903 | .mwdma_mask = 0x07, | 757 | .mwdma_mask = 0x07, |
@@ -906,7 +760,6 @@ static int hpt37x_init_one(struct pci_dev *dev, const struct pci_device_id *id) | |||
906 | }; | 760 | }; |
907 | /* HPT370A - UDMA100 */ | 761 | /* HPT370A - UDMA100 */ |
908 | static const struct ata_port_info info_hpt370a = { | 762 | static const struct ata_port_info info_hpt370a = { |
909 | .sht = &hpt37x_sht, | ||
910 | .flags = ATA_FLAG_SLAVE_POSS, | 763 | .flags = ATA_FLAG_SLAVE_POSS, |
911 | .pio_mask = 0x1f, | 764 | .pio_mask = 0x1f, |
912 | .mwdma_mask = 0x07, | 765 | .mwdma_mask = 0x07, |
@@ -915,7 +768,6 @@ static int hpt37x_init_one(struct pci_dev *dev, const struct pci_device_id *id) | |||
915 | }; | 768 | }; |
916 | /* HPT370 - UDMA100 */ | 769 | /* HPT370 - UDMA100 */ |
917 | static const struct ata_port_info info_hpt370_33 = { | 770 | static const struct ata_port_info info_hpt370_33 = { |
918 | .sht = &hpt37x_sht, | ||
919 | .flags = ATA_FLAG_SLAVE_POSS, | 771 | .flags = ATA_FLAG_SLAVE_POSS, |
920 | .pio_mask = 0x1f, | 772 | .pio_mask = 0x1f, |
921 | .mwdma_mask = 0x07, | 773 | .mwdma_mask = 0x07, |
@@ -924,7 +776,6 @@ static int hpt37x_init_one(struct pci_dev *dev, const struct pci_device_id *id) | |||
924 | }; | 776 | }; |
925 | /* HPT370A - UDMA100 */ | 777 | /* HPT370A - UDMA100 */ |
926 | static const struct ata_port_info info_hpt370a_33 = { | 778 | static const struct ata_port_info info_hpt370a_33 = { |
927 | .sht = &hpt37x_sht, | ||
928 | .flags = ATA_FLAG_SLAVE_POSS, | 779 | .flags = ATA_FLAG_SLAVE_POSS, |
929 | .pio_mask = 0x1f, | 780 | .pio_mask = 0x1f, |
930 | .mwdma_mask = 0x07, | 781 | .mwdma_mask = 0x07, |
@@ -933,28 +784,31 @@ static int hpt37x_init_one(struct pci_dev *dev, const struct pci_device_id *id) | |||
933 | }; | 784 | }; |
934 | /* HPT371, 372 and friends - UDMA133 */ | 785 | /* HPT371, 372 and friends - UDMA133 */ |
935 | static const struct ata_port_info info_hpt372 = { | 786 | static const struct ata_port_info info_hpt372 = { |
936 | .sht = &hpt37x_sht, | ||
937 | .flags = ATA_FLAG_SLAVE_POSS, | 787 | .flags = ATA_FLAG_SLAVE_POSS, |
938 | .pio_mask = 0x1f, | 788 | .pio_mask = 0x1f, |
939 | .mwdma_mask = 0x07, | 789 | .mwdma_mask = 0x07, |
940 | .udma_mask = ATA_UDMA6, | 790 | .udma_mask = ATA_UDMA6, |
941 | .port_ops = &hpt372_port_ops | 791 | .port_ops = &hpt372_port_ops |
942 | }; | 792 | }; |
943 | /* HPT374 - UDMA100 */ | 793 | /* HPT374 - UDMA100, function 1 uses different prereset method */ |
944 | static const struct ata_port_info info_hpt374 = { | 794 | static const struct ata_port_info info_hpt374_fn0 = { |
945 | .sht = &hpt37x_sht, | ||
946 | .flags = ATA_FLAG_SLAVE_POSS, | 795 | .flags = ATA_FLAG_SLAVE_POSS, |
947 | .pio_mask = 0x1f, | 796 | .pio_mask = 0x1f, |
948 | .mwdma_mask = 0x07, | 797 | .mwdma_mask = 0x07, |
949 | .udma_mask = ATA_UDMA5, | 798 | .udma_mask = ATA_UDMA5, |
950 | .port_ops = &hpt374_port_ops | 799 | .port_ops = &hpt372_port_ops |
800 | }; | ||
801 | static const struct ata_port_info info_hpt374_fn1 = { | ||
802 | .flags = ATA_FLAG_SLAVE_POSS, | ||
803 | .pio_mask = 0x1f, | ||
804 | .mwdma_mask = 0x07, | ||
805 | .udma_mask = ATA_UDMA5, | ||
806 | .port_ops = &hpt374_fn1_port_ops | ||
951 | }; | 807 | }; |
952 | 808 | ||
953 | static const int MHz[4] = { 33, 40, 50, 66 }; | 809 | static const int MHz[4] = { 33, 40, 50, 66 }; |
954 | const struct ata_port_info *port; | ||
955 | void *private_data = NULL; | 810 | void *private_data = NULL; |
956 | struct ata_port_info port_info; | 811 | const struct ata_port_info *ppi[] = { NULL, NULL }; |
957 | const struct ata_port_info *ppi[] = { &port_info, NULL }; | ||
958 | 812 | ||
959 | u8 irqmask; | 813 | u8 irqmask; |
960 | u32 class_rev; | 814 | u32 class_rev; |
@@ -966,6 +820,11 @@ static int hpt37x_init_one(struct pci_dev *dev, const struct pci_device_id *id) | |||
966 | 820 | ||
967 | const struct hpt_chip *chip_table; | 821 | const struct hpt_chip *chip_table; |
968 | int clock_slot; | 822 | int clock_slot; |
823 | int rc; | ||
824 | |||
825 | rc = pcim_enable_device(dev); | ||
826 | if (rc) | ||
827 | return rc; | ||
969 | 828 | ||
970 | pci_read_config_dword(dev, PCI_CLASS_REVISION, &class_rev); | 829 | pci_read_config_dword(dev, PCI_CLASS_REVISION, &class_rev); |
971 | class_rev &= 0xFF; | 830 | class_rev &= 0xFF; |
@@ -981,17 +840,17 @@ static int hpt37x_init_one(struct pci_dev *dev, const struct pci_device_id *id) | |||
981 | 840 | ||
982 | switch(class_rev) { | 841 | switch(class_rev) { |
983 | case 3: | 842 | case 3: |
984 | port = &info_hpt370; | 843 | ppi[0] = &info_hpt370; |
985 | chip_table = &hpt370; | 844 | chip_table = &hpt370; |
986 | prefer_dpll = 0; | 845 | prefer_dpll = 0; |
987 | break; | 846 | break; |
988 | case 4: | 847 | case 4: |
989 | port = &info_hpt370a; | 848 | ppi[0] = &info_hpt370a; |
990 | chip_table = &hpt370a; | 849 | chip_table = &hpt370a; |
991 | prefer_dpll = 0; | 850 | prefer_dpll = 0; |
992 | break; | 851 | break; |
993 | case 5: | 852 | case 5: |
994 | port = &info_hpt372; | 853 | ppi[0] = &info_hpt372; |
995 | chip_table = &hpt372; | 854 | chip_table = &hpt372; |
996 | break; | 855 | break; |
997 | default: | 856 | default: |
@@ -1004,21 +863,21 @@ static int hpt37x_init_one(struct pci_dev *dev, const struct pci_device_id *id) | |||
1004 | /* 372N if rev >= 2*/ | 863 | /* 372N if rev >= 2*/ |
1005 | if (class_rev >= 2) | 864 | if (class_rev >= 2) |
1006 | return -ENODEV; | 865 | return -ENODEV; |
1007 | port = &info_hpt372; | 866 | ppi[0] = &info_hpt372; |
1008 | chip_table = &hpt372a; | 867 | chip_table = &hpt372a; |
1009 | break; | 868 | break; |
1010 | case PCI_DEVICE_ID_TTI_HPT302: | 869 | case PCI_DEVICE_ID_TTI_HPT302: |
1011 | /* 302N if rev > 1 */ | 870 | /* 302N if rev > 1 */ |
1012 | if (class_rev > 1) | 871 | if (class_rev > 1) |
1013 | return -ENODEV; | 872 | return -ENODEV; |
1014 | port = &info_hpt372; | 873 | ppi[0] = &info_hpt372; |
1015 | /* Check this */ | 874 | /* Check this */ |
1016 | chip_table = &hpt302; | 875 | chip_table = &hpt302; |
1017 | break; | 876 | break; |
1018 | case PCI_DEVICE_ID_TTI_HPT371: | 877 | case PCI_DEVICE_ID_TTI_HPT371: |
1019 | if (class_rev > 1) | 878 | if (class_rev > 1) |
1020 | return -ENODEV; | 879 | return -ENODEV; |
1021 | port = &info_hpt372; | 880 | ppi[0] = &info_hpt372; |
1022 | chip_table = &hpt371; | 881 | chip_table = &hpt371; |
1023 | /* Single channel device, master is not present | 882 | /* Single channel device, master is not present |
1024 | but the BIOS (or us for non x86) must mark it | 883 | but the BIOS (or us for non x86) must mark it |
@@ -1029,7 +888,10 @@ static int hpt37x_init_one(struct pci_dev *dev, const struct pci_device_id *id) | |||
1029 | break; | 888 | break; |
1030 | case PCI_DEVICE_ID_TTI_HPT374: | 889 | case PCI_DEVICE_ID_TTI_HPT374: |
1031 | chip_table = &hpt374; | 890 | chip_table = &hpt374; |
1032 | port = &info_hpt374; | 891 | if (!(PCI_FUNC(dev->devfn) & 1)) |
892 | *ppi = &info_hpt374_fn0; | ||
893 | else | ||
894 | *ppi = &info_hpt374_fn1; | ||
1033 | break; | 895 | break; |
1034 | default: | 896 | default: |
1035 | printk(KERN_ERR "pata_hpt37x: PCI table is bogus please report (%d).\n", dev->device); | 897 | printk(KERN_ERR "pata_hpt37x: PCI table is bogus please report (%d).\n", dev->device); |
@@ -1108,7 +970,7 @@ static int hpt37x_init_one(struct pci_dev *dev, const struct pci_device_id *id) | |||
1108 | int dpll, adjust; | 970 | int dpll, adjust; |
1109 | 971 | ||
1110 | /* Compute DPLL */ | 972 | /* Compute DPLL */ |
1111 | dpll = (port->udma_mask & 0xC0) ? 3 : 2; | 973 | dpll = (ppi[0]->udma_mask & 0xC0) ? 3 : 2; |
1112 | 974 | ||
1113 | f_low = (MHz[clock_slot] * 48) / MHz[dpll]; | 975 | f_low = (MHz[clock_slot] * 48) / MHz[dpll]; |
1114 | f_high = f_low + 2; | 976 | f_high = f_low + 2; |
@@ -1148,19 +1010,16 @@ static int hpt37x_init_one(struct pci_dev *dev, const struct pci_device_id *id) | |||
1148 | * about lack of UDMA133 support on lower clocks | 1010 | * about lack of UDMA133 support on lower clocks |
1149 | */ | 1011 | */ |
1150 | 1012 | ||
1151 | if (clock_slot < 2 && port == &info_hpt370) | 1013 | if (clock_slot < 2 && ppi[0] == &info_hpt370) |
1152 | port = &info_hpt370_33; | 1014 | ppi[0] = &info_hpt370_33; |
1153 | if (clock_slot < 2 && port == &info_hpt370a) | 1015 | if (clock_slot < 2 && ppi[0] == &info_hpt370a) |
1154 | port = &info_hpt370a_33; | 1016 | ppi[0] = &info_hpt370a_33; |
1155 | printk(KERN_INFO "pata_hpt37x: %s using %dMHz bus clock.\n", | 1017 | printk(KERN_INFO "pata_hpt37x: %s using %dMHz bus clock.\n", |
1156 | chip_table->name, MHz[clock_slot]); | 1018 | chip_table->name, MHz[clock_slot]); |
1157 | } | 1019 | } |
1158 | 1020 | ||
1159 | /* Now kick off ATA set up */ | 1021 | /* Now kick off ATA set up */ |
1160 | port_info = *port; | 1022 | return ata_pci_sff_init_one(dev, ppi, &hpt37x_sht, private_data); |
1161 | port_info.private_data = private_data; | ||
1162 | |||
1163 | return ata_pci_init_one(dev, ppi); | ||
1164 | } | 1023 | } |
1165 | 1024 | ||
1166 | static const struct pci_device_id hpt37x[] = { | 1025 | static const struct pci_device_id hpt37x[] = { |
diff --git a/drivers/ata/pata_hpt3x2n.c b/drivers/ata/pata_hpt3x2n.c index 9f1c084f846f..d5c9fd7b82bb 100644 --- a/drivers/ata/pata_hpt3x2n.c +++ b/drivers/ata/pata_hpt3x2n.c | |||
@@ -148,7 +148,7 @@ static int hpt3x2n_cable_detect(struct ata_port *ap) | |||
148 | * Reset the hardware and state machine, | 148 | * Reset the hardware and state machine, |
149 | */ | 149 | */ |
150 | 150 | ||
151 | static int hpt3xn_pre_reset(struct ata_link *link, unsigned long deadline) | 151 | static int hpt3x2n_pre_reset(struct ata_link *link, unsigned long deadline) |
152 | { | 152 | { |
153 | struct ata_port *ap = link->ap; | 153 | struct ata_port *ap = link->ap; |
154 | struct pci_dev *pdev = to_pci_dev(ap->host->dev); | 154 | struct pci_dev *pdev = to_pci_dev(ap->host->dev); |
@@ -156,19 +156,7 @@ static int hpt3xn_pre_reset(struct ata_link *link, unsigned long deadline) | |||
156 | pci_write_config_byte(pdev, 0x50 + 4 * ap->port_no, 0x37); | 156 | pci_write_config_byte(pdev, 0x50 + 4 * ap->port_no, 0x37); |
157 | udelay(100); | 157 | udelay(100); |
158 | 158 | ||
159 | return ata_std_prereset(link, deadline); | 159 | return ata_sff_prereset(link, deadline); |
160 | } | ||
161 | |||
162 | /** | ||
163 | * hpt3x2n_error_handler - probe the hpt3x2n bus | ||
164 | * @ap: ATA port to reset | ||
165 | * | ||
166 | * Perform the probe reset handling for the 3x2N | ||
167 | */ | ||
168 | |||
169 | static void hpt3x2n_error_handler(struct ata_port *ap) | ||
170 | { | ||
171 | ata_bmdma_drive_eh(ap, hpt3xn_pre_reset, ata_std_softreset, NULL, ata_std_postreset); | ||
172 | } | 160 | } |
173 | 161 | ||
174 | /** | 162 | /** |
@@ -320,7 +308,7 @@ static int hpt3x2n_use_dpll(struct ata_port *ap, int writing) | |||
320 | return 0; | 308 | return 0; |
321 | } | 309 | } |
322 | 310 | ||
323 | static unsigned int hpt3x2n_qc_issue_prot(struct ata_queued_cmd *qc) | 311 | static unsigned int hpt3x2n_qc_issue(struct ata_queued_cmd *qc) |
324 | { | 312 | { |
325 | struct ata_taskfile *tf = &qc->tf; | 313 | struct ata_taskfile *tf = &qc->tf; |
326 | struct ata_port *ap = qc->ap; | 314 | struct ata_port *ap = qc->ap; |
@@ -335,25 +323,11 @@ static unsigned int hpt3x2n_qc_issue_prot(struct ata_queued_cmd *qc) | |||
335 | hpt3x2n_set_clock(ap, 0x23); | 323 | hpt3x2n_set_clock(ap, 0x23); |
336 | } | 324 | } |
337 | } | 325 | } |
338 | return ata_qc_issue_prot(qc); | 326 | return ata_sff_qc_issue(qc); |
339 | } | 327 | } |
340 | 328 | ||
341 | static struct scsi_host_template hpt3x2n_sht = { | 329 | static struct scsi_host_template hpt3x2n_sht = { |
342 | .module = THIS_MODULE, | 330 | ATA_BMDMA_SHT(DRV_NAME), |
343 | .name = DRV_NAME, | ||
344 | .ioctl = ata_scsi_ioctl, | ||
345 | .queuecommand = ata_scsi_queuecmd, | ||
346 | .can_queue = ATA_DEF_QUEUE, | ||
347 | .this_id = ATA_SHT_THIS_ID, | ||
348 | .sg_tablesize = LIBATA_MAX_PRD, | ||
349 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
350 | .emulated = ATA_SHT_EMULATED, | ||
351 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
352 | .proc_name = DRV_NAME, | ||
353 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
354 | .slave_configure = ata_scsi_slave_config, | ||
355 | .slave_destroy = ata_scsi_slave_destroy, | ||
356 | .bios_param = ata_std_bios_param, | ||
357 | }; | 331 | }; |
358 | 332 | ||
359 | /* | 333 | /* |
@@ -361,37 +335,15 @@ static struct scsi_host_template hpt3x2n_sht = { | |||
361 | */ | 335 | */ |
362 | 336 | ||
363 | static struct ata_port_operations hpt3x2n_port_ops = { | 337 | static struct ata_port_operations hpt3x2n_port_ops = { |
364 | .set_piomode = hpt3x2n_set_piomode, | 338 | .inherits = &ata_bmdma_port_ops, |
365 | .set_dmamode = hpt3x2n_set_dmamode, | ||
366 | .mode_filter = ata_pci_default_filter, | ||
367 | |||
368 | .tf_load = ata_tf_load, | ||
369 | .tf_read = ata_tf_read, | ||
370 | .check_status = ata_check_status, | ||
371 | .exec_command = ata_exec_command, | ||
372 | .dev_select = ata_std_dev_select, | ||
373 | |||
374 | .freeze = ata_bmdma_freeze, | ||
375 | .thaw = ata_bmdma_thaw, | ||
376 | .error_handler = hpt3x2n_error_handler, | ||
377 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
378 | .cable_detect = hpt3x2n_cable_detect, | ||
379 | 339 | ||
380 | .bmdma_setup = ata_bmdma_setup, | ||
381 | .bmdma_start = ata_bmdma_start, | ||
382 | .bmdma_stop = hpt3x2n_bmdma_stop, | 340 | .bmdma_stop = hpt3x2n_bmdma_stop, |
383 | .bmdma_status = ata_bmdma_status, | 341 | .qc_issue = hpt3x2n_qc_issue, |
384 | |||
385 | .qc_prep = ata_qc_prep, | ||
386 | .qc_issue = hpt3x2n_qc_issue_prot, | ||
387 | |||
388 | .data_xfer = ata_data_xfer, | ||
389 | |||
390 | .irq_handler = ata_interrupt, | ||
391 | .irq_clear = ata_bmdma_irq_clear, | ||
392 | .irq_on = ata_irq_on, | ||
393 | 342 | ||
394 | .port_start = ata_sff_port_start, | 343 | .cable_detect = hpt3x2n_cable_detect, |
344 | .set_piomode = hpt3x2n_set_piomode, | ||
345 | .set_dmamode = hpt3x2n_set_dmamode, | ||
346 | .prereset = hpt3x2n_pre_reset, | ||
395 | }; | 347 | }; |
396 | 348 | ||
397 | /** | 349 | /** |
@@ -488,15 +440,13 @@ static int hpt3x2n_init_one(struct pci_dev *dev, const struct pci_device_id *id) | |||
488 | { | 440 | { |
489 | /* HPT372N and friends - UDMA133 */ | 441 | /* HPT372N and friends - UDMA133 */ |
490 | static const struct ata_port_info info = { | 442 | static const struct ata_port_info info = { |
491 | .sht = &hpt3x2n_sht, | ||
492 | .flags = ATA_FLAG_SLAVE_POSS, | 443 | .flags = ATA_FLAG_SLAVE_POSS, |
493 | .pio_mask = 0x1f, | 444 | .pio_mask = 0x1f, |
494 | .mwdma_mask = 0x07, | 445 | .mwdma_mask = 0x07, |
495 | .udma_mask = ATA_UDMA6, | 446 | .udma_mask = ATA_UDMA6, |
496 | .port_ops = &hpt3x2n_port_ops | 447 | .port_ops = &hpt3x2n_port_ops |
497 | }; | 448 | }; |
498 | struct ata_port_info port = info; | 449 | const struct ata_port_info *ppi[] = { &info, NULL }; |
499 | const struct ata_port_info *ppi[] = { &port, NULL }; | ||
500 | 450 | ||
501 | u8 irqmask; | 451 | u8 irqmask; |
502 | u32 class_rev; | 452 | u32 class_rev; |
@@ -505,6 +455,12 @@ static int hpt3x2n_init_one(struct pci_dev *dev, const struct pci_device_id *id) | |||
505 | unsigned int f_low, f_high; | 455 | unsigned int f_low, f_high; |
506 | int adjust; | 456 | int adjust; |
507 | unsigned long iobase = pci_resource_start(dev, 4); | 457 | unsigned long iobase = pci_resource_start(dev, 4); |
458 | void *hpriv = NULL; | ||
459 | int rc; | ||
460 | |||
461 | rc = pcim_enable_device(dev); | ||
462 | if (rc) | ||
463 | return rc; | ||
508 | 464 | ||
509 | pci_read_config_dword(dev, PCI_CLASS_REVISION, &class_rev); | 465 | pci_read_config_dword(dev, PCI_CLASS_REVISION, &class_rev); |
510 | class_rev &= 0xFF; | 466 | class_rev &= 0xFF; |
@@ -586,9 +542,8 @@ static int hpt3x2n_init_one(struct pci_dev *dev, const struct pci_device_id *id) | |||
586 | pci_mhz); | 542 | pci_mhz); |
587 | /* Set our private data up. We only need a few flags so we use | 543 | /* Set our private data up. We only need a few flags so we use |
588 | it directly */ | 544 | it directly */ |
589 | port.private_data = NULL; | ||
590 | if (pci_mhz > 60) { | 545 | if (pci_mhz > 60) { |
591 | port.private_data = (void *)PCI66; | 546 | hpriv = (void *)PCI66; |
592 | /* | 547 | /* |
593 | * On HPT371N, if ATA clock is 66 MHz we must set bit 2 in | 548 | * On HPT371N, if ATA clock is 66 MHz we must set bit 2 in |
594 | * the MISC. register to stretch the UltraDMA Tss timing. | 549 | * the MISC. register to stretch the UltraDMA Tss timing. |
@@ -599,7 +554,7 @@ static int hpt3x2n_init_one(struct pci_dev *dev, const struct pci_device_id *id) | |||
599 | } | 554 | } |
600 | 555 | ||
601 | /* Now kick off ATA set up */ | 556 | /* Now kick off ATA set up */ |
602 | return ata_pci_init_one(dev, ppi); | 557 | return ata_pci_sff_init_one(dev, ppi, &hpt3x2n_sht, hpriv); |
603 | } | 558 | } |
604 | 559 | ||
605 | static const struct pci_device_id hpt3x2n[] = { | 560 | static const struct pci_device_id hpt3x2n[] = { |
diff --git a/drivers/ata/pata_hpt3x3.c b/drivers/ata/pata_hpt3x3.c index cb8bdb6887de..f11a320337c0 100644 --- a/drivers/ata/pata_hpt3x3.c +++ b/drivers/ata/pata_hpt3x3.c | |||
@@ -102,58 +102,17 @@ static int hpt3x3_atapi_dma(struct ata_queued_cmd *qc) | |||
102 | } | 102 | } |
103 | 103 | ||
104 | static struct scsi_host_template hpt3x3_sht = { | 104 | static struct scsi_host_template hpt3x3_sht = { |
105 | .module = THIS_MODULE, | 105 | ATA_BMDMA_SHT(DRV_NAME), |
106 | .name = DRV_NAME, | ||
107 | .ioctl = ata_scsi_ioctl, | ||
108 | .queuecommand = ata_scsi_queuecmd, | ||
109 | .can_queue = ATA_DEF_QUEUE, | ||
110 | .this_id = ATA_SHT_THIS_ID, | ||
111 | .sg_tablesize = LIBATA_MAX_PRD, | ||
112 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
113 | .emulated = ATA_SHT_EMULATED, | ||
114 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
115 | .proc_name = DRV_NAME, | ||
116 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
117 | .slave_configure = ata_scsi_slave_config, | ||
118 | .slave_destroy = ata_scsi_slave_destroy, | ||
119 | .bios_param = ata_std_bios_param, | ||
120 | }; | 106 | }; |
121 | 107 | ||
122 | static struct ata_port_operations hpt3x3_port_ops = { | 108 | static struct ata_port_operations hpt3x3_port_ops = { |
109 | .inherits = &ata_bmdma_port_ops, | ||
110 | .check_atapi_dma= hpt3x3_atapi_dma, | ||
111 | .cable_detect = ata_cable_40wire, | ||
123 | .set_piomode = hpt3x3_set_piomode, | 112 | .set_piomode = hpt3x3_set_piomode, |
124 | #if defined(CONFIG_PATA_HPT3X3_DMA) | 113 | #if defined(CONFIG_PATA_HPT3X3_DMA) |
125 | .set_dmamode = hpt3x3_set_dmamode, | 114 | .set_dmamode = hpt3x3_set_dmamode, |
126 | #endif | 115 | #endif |
127 | .mode_filter = ata_pci_default_filter, | ||
128 | |||
129 | .tf_load = ata_tf_load, | ||
130 | .tf_read = ata_tf_read, | ||
131 | .check_status = ata_check_status, | ||
132 | .exec_command = ata_exec_command, | ||
133 | .dev_select = ata_std_dev_select, | ||
134 | |||
135 | .freeze = ata_bmdma_freeze, | ||
136 | .thaw = ata_bmdma_thaw, | ||
137 | .error_handler = ata_bmdma_error_handler, | ||
138 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
139 | .cable_detect = ata_cable_40wire, | ||
140 | |||
141 | .bmdma_setup = ata_bmdma_setup, | ||
142 | .bmdma_start = ata_bmdma_start, | ||
143 | .bmdma_stop = ata_bmdma_stop, | ||
144 | .bmdma_status = ata_bmdma_status, | ||
145 | .check_atapi_dma= hpt3x3_atapi_dma, | ||
146 | |||
147 | .qc_prep = ata_qc_prep, | ||
148 | .qc_issue = ata_qc_issue_prot, | ||
149 | |||
150 | .data_xfer = ata_data_xfer, | ||
151 | |||
152 | .irq_handler = ata_interrupt, | ||
153 | .irq_clear = ata_bmdma_irq_clear, | ||
154 | .irq_on = ata_irq_on, | ||
155 | |||
156 | .port_start = ata_sff_port_start, | ||
157 | }; | 116 | }; |
158 | 117 | ||
159 | /** | 118 | /** |
@@ -189,7 +148,6 @@ static int hpt3x3_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
189 | { | 148 | { |
190 | static int printed_version; | 149 | static int printed_version; |
191 | static const struct ata_port_info info = { | 150 | static const struct ata_port_info info = { |
192 | .sht = &hpt3x3_sht, | ||
193 | .flags = ATA_FLAG_SLAVE_POSS, | 151 | .flags = ATA_FLAG_SLAVE_POSS, |
194 | .pio_mask = 0x1f, | 152 | .pio_mask = 0x1f, |
195 | #if defined(CONFIG_PATA_HPT3X3_DMA) | 153 | #if defined(CONFIG_PATA_HPT3X3_DMA) |
@@ -244,15 +202,15 @@ static int hpt3x3_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
244 | ioaddr->altstatus_addr = | 202 | ioaddr->altstatus_addr = |
245 | ioaddr->ctl_addr = base + offset_ctl[i]; | 203 | ioaddr->ctl_addr = base + offset_ctl[i]; |
246 | ioaddr->scr_addr = NULL; | 204 | ioaddr->scr_addr = NULL; |
247 | ata_std_ports(ioaddr); | 205 | ata_sff_std_ports(ioaddr); |
248 | ioaddr->bmdma_addr = base + 8 * i; | 206 | ioaddr->bmdma_addr = base + 8 * i; |
249 | 207 | ||
250 | ata_port_pbar_desc(ap, 4, -1, "ioport"); | 208 | ata_port_pbar_desc(ap, 4, -1, "ioport"); |
251 | ata_port_pbar_desc(ap, 4, offset_cmd[i], "cmd"); | 209 | ata_port_pbar_desc(ap, 4, offset_cmd[i], "cmd"); |
252 | } | 210 | } |
253 | pci_set_master(pdev); | 211 | pci_set_master(pdev); |
254 | return ata_host_activate(host, pdev->irq, ata_interrupt, IRQF_SHARED, | 212 | return ata_host_activate(host, pdev->irq, ata_sff_interrupt, |
255 | &hpt3x3_sht); | 213 | IRQF_SHARED, &hpt3x3_sht); |
256 | } | 214 | } |
257 | 215 | ||
258 | #ifdef CONFIG_PM | 216 | #ifdef CONFIG_PM |
diff --git a/drivers/ata/pata_icside.c b/drivers/ata/pata_icside.c index f97068be2d79..17138436423d 100644 --- a/drivers/ata/pata_icside.c +++ b/drivers/ata/pata_icside.c | |||
@@ -250,7 +250,7 @@ static void pata_icside_bmdma_setup(struct ata_queued_cmd *qc) | |||
250 | set_dma_mode(state->dma, write ? DMA_MODE_WRITE : DMA_MODE_READ); | 250 | set_dma_mode(state->dma, write ? DMA_MODE_WRITE : DMA_MODE_READ); |
251 | 251 | ||
252 | /* issue r/w command */ | 252 | /* issue r/w command */ |
253 | ap->ops->exec_command(ap, &qc->tf); | 253 | ap->ops->sff_exec_command(ap, &qc->tf); |
254 | } | 254 | } |
255 | 255 | ||
256 | static void pata_icside_bmdma_start(struct ata_queued_cmd *qc) | 256 | static void pata_icside_bmdma_start(struct ata_queued_cmd *qc) |
@@ -270,7 +270,7 @@ static void pata_icside_bmdma_stop(struct ata_queued_cmd *qc) | |||
270 | disable_dma(state->dma); | 270 | disable_dma(state->dma); |
271 | 271 | ||
272 | /* see ata_bmdma_stop */ | 272 | /* see ata_bmdma_stop */ |
273 | ata_altstatus(ap); | 273 | ata_sff_altstatus(ap); |
274 | } | 274 | } |
275 | 275 | ||
276 | static u8 pata_icside_bmdma_status(struct ata_port *ap) | 276 | static u8 pata_icside_bmdma_status(struct ata_port *ap) |
@@ -305,35 +305,18 @@ static int icside_dma_init(struct pata_icside_info *info) | |||
305 | 305 | ||
306 | 306 | ||
307 | static struct scsi_host_template pata_icside_sht = { | 307 | static struct scsi_host_template pata_icside_sht = { |
308 | .module = THIS_MODULE, | 308 | ATA_BASE_SHT(DRV_NAME), |
309 | .name = DRV_NAME, | ||
310 | .ioctl = ata_scsi_ioctl, | ||
311 | .queuecommand = ata_scsi_queuecmd, | ||
312 | .can_queue = ATA_DEF_QUEUE, | ||
313 | .this_id = ATA_SHT_THIS_ID, | ||
314 | .sg_tablesize = PATA_ICSIDE_MAX_SG, | 309 | .sg_tablesize = PATA_ICSIDE_MAX_SG, |
315 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
316 | .emulated = ATA_SHT_EMULATED, | ||
317 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
318 | .proc_name = DRV_NAME, | ||
319 | .dma_boundary = ~0, /* no dma boundaries */ | 310 | .dma_boundary = ~0, /* no dma boundaries */ |
320 | .slave_configure = ata_scsi_slave_config, | ||
321 | .slave_destroy = ata_scsi_slave_destroy, | ||
322 | .bios_param = ata_std_bios_param, | ||
323 | }; | 311 | }; |
324 | 312 | ||
325 | /* wish this was exported from libata-core */ | ||
326 | static void ata_dummy_noret(struct ata_port *port) | ||
327 | { | ||
328 | } | ||
329 | |||
330 | static void pata_icside_postreset(struct ata_link *link, unsigned int *classes) | 313 | static void pata_icside_postreset(struct ata_link *link, unsigned int *classes) |
331 | { | 314 | { |
332 | struct ata_port *ap = link->ap; | 315 | struct ata_port *ap = link->ap; |
333 | struct pata_icside_state *state = ap->host->private_data; | 316 | struct pata_icside_state *state = ap->host->private_data; |
334 | 317 | ||
335 | if (classes[0] != ATA_DEV_NONE || classes[1] != ATA_DEV_NONE) | 318 | if (classes[0] != ATA_DEV_NONE || classes[1] != ATA_DEV_NONE) |
336 | return ata_std_postreset(link, classes); | 319 | return ata_sff_postreset(link, classes); |
337 | 320 | ||
338 | state->port[ap->port_no].disabled = 1; | 321 | state->port[ap->port_no].disabled = 1; |
339 | 322 | ||
@@ -349,42 +332,20 @@ static void pata_icside_postreset(struct ata_link *link, unsigned int *classes) | |||
349 | } | 332 | } |
350 | } | 333 | } |
351 | 334 | ||
352 | static void pata_icside_error_handler(struct ata_port *ap) | ||
353 | { | ||
354 | ata_bmdma_drive_eh(ap, ata_std_prereset, ata_std_softreset, NULL, | ||
355 | pata_icside_postreset); | ||
356 | } | ||
357 | |||
358 | static struct ata_port_operations pata_icside_port_ops = { | 335 | static struct ata_port_operations pata_icside_port_ops = { |
359 | .set_dmamode = pata_icside_set_dmamode, | 336 | .inherits = &ata_sff_port_ops, |
360 | |||
361 | .tf_load = ata_tf_load, | ||
362 | .tf_read = ata_tf_read, | ||
363 | .exec_command = ata_exec_command, | ||
364 | .check_status = ata_check_status, | ||
365 | .dev_select = ata_std_dev_select, | ||
366 | |||
367 | .cable_detect = ata_cable_40wire, | ||
368 | |||
369 | .bmdma_setup = pata_icside_bmdma_setup, | ||
370 | .bmdma_start = pata_icside_bmdma_start, | ||
371 | |||
372 | .data_xfer = ata_data_xfer_noirq, | ||
373 | |||
374 | /* no need to build any PRD tables for DMA */ | 337 | /* no need to build any PRD tables for DMA */ |
375 | .qc_prep = ata_noop_qc_prep, | 338 | .qc_prep = ata_noop_qc_prep, |
376 | .qc_issue = ata_qc_issue_prot, | 339 | .sff_data_xfer = ata_sff_data_xfer_noirq, |
377 | 340 | .bmdma_setup = pata_icside_bmdma_setup, | |
378 | .freeze = ata_bmdma_freeze, | 341 | .bmdma_start = pata_icside_bmdma_start, |
379 | .thaw = ata_bmdma_thaw, | ||
380 | .error_handler = pata_icside_error_handler, | ||
381 | .post_internal_cmd = pata_icside_bmdma_stop, | ||
382 | |||
383 | .irq_clear = ata_dummy_noret, | ||
384 | .irq_on = ata_irq_on, | ||
385 | |||
386 | .bmdma_stop = pata_icside_bmdma_stop, | 342 | .bmdma_stop = pata_icside_bmdma_stop, |
387 | .bmdma_status = pata_icside_bmdma_status, | 343 | .bmdma_status = pata_icside_bmdma_status, |
344 | |||
345 | .cable_detect = ata_cable_40wire, | ||
346 | .set_dmamode = pata_icside_set_dmamode, | ||
347 | .postreset = pata_icside_postreset, | ||
348 | .post_internal_cmd = pata_icside_bmdma_stop, | ||
388 | }; | 349 | }; |
389 | 350 | ||
390 | static void __devinit | 351 | static void __devinit |
@@ -520,7 +481,7 @@ static int __devinit pata_icside_add_ports(struct pata_icside_info *info) | |||
520 | pata_icside_setup_ioaddr(ap, info->base, info, info->port[i]); | 481 | pata_icside_setup_ioaddr(ap, info->base, info, info->port[i]); |
521 | } | 482 | } |
522 | 483 | ||
523 | return ata_host_activate(host, ec->irq, ata_interrupt, 0, | 484 | return ata_host_activate(host, ec->irq, ata_sff_interrupt, 0, |
524 | &pata_icside_sht); | 485 | &pata_icside_sht); |
525 | } | 486 | } |
526 | 487 | ||
diff --git a/drivers/ata/pata_isapnp.c b/drivers/ata/pata_isapnp.c index 4320e7986321..6a111baab523 100644 --- a/drivers/ata/pata_isapnp.c +++ b/drivers/ata/pata_isapnp.c | |||
@@ -20,45 +20,12 @@ | |||
20 | #define DRV_VERSION "0.2.2" | 20 | #define DRV_VERSION "0.2.2" |
21 | 21 | ||
22 | static struct scsi_host_template isapnp_sht = { | 22 | static struct scsi_host_template isapnp_sht = { |
23 | .module = THIS_MODULE, | 23 | ATA_PIO_SHT(DRV_NAME), |
24 | .name = DRV_NAME, | ||
25 | .ioctl = ata_scsi_ioctl, | ||
26 | .queuecommand = ata_scsi_queuecmd, | ||
27 | .can_queue = ATA_DEF_QUEUE, | ||
28 | .this_id = ATA_SHT_THIS_ID, | ||
29 | .sg_tablesize = LIBATA_MAX_PRD, | ||
30 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
31 | .emulated = ATA_SHT_EMULATED, | ||
32 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
33 | .proc_name = DRV_NAME, | ||
34 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
35 | .slave_configure = ata_scsi_slave_config, | ||
36 | .slave_destroy = ata_scsi_slave_destroy, | ||
37 | .bios_param = ata_std_bios_param, | ||
38 | }; | 24 | }; |
39 | 25 | ||
40 | static struct ata_port_operations isapnp_port_ops = { | 26 | static struct ata_port_operations isapnp_port_ops = { |
41 | .tf_load = ata_tf_load, | 27 | .inherits = &ata_sff_port_ops, |
42 | .tf_read = ata_tf_read, | ||
43 | .check_status = ata_check_status, | ||
44 | .exec_command = ata_exec_command, | ||
45 | .dev_select = ata_std_dev_select, | ||
46 | |||
47 | .freeze = ata_bmdma_freeze, | ||
48 | .thaw = ata_bmdma_thaw, | ||
49 | .error_handler = ata_bmdma_error_handler, | ||
50 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
51 | .cable_detect = ata_cable_40wire, | 28 | .cable_detect = ata_cable_40wire, |
52 | |||
53 | .qc_prep = ata_qc_prep, | ||
54 | .qc_issue = ata_qc_issue_prot, | ||
55 | |||
56 | .data_xfer = ata_data_xfer, | ||
57 | |||
58 | .irq_clear = ata_bmdma_irq_clear, | ||
59 | .irq_on = ata_irq_on, | ||
60 | |||
61 | .port_start = ata_sff_port_start, | ||
62 | }; | 29 | }; |
63 | 30 | ||
64 | /** | 31 | /** |
@@ -83,7 +50,7 @@ static int isapnp_init_one(struct pnp_dev *idev, const struct pnp_device_id *dev | |||
83 | 50 | ||
84 | if (pnp_irq_valid(idev, 0)) { | 51 | if (pnp_irq_valid(idev, 0)) { |
85 | irq = pnp_irq(idev, 0); | 52 | irq = pnp_irq(idev, 0); |
86 | handler = ata_interrupt; | 53 | handler = ata_sff_interrupt; |
87 | } | 54 | } |
88 | 55 | ||
89 | /* allocate host */ | 56 | /* allocate host */ |
@@ -111,7 +78,7 @@ static int isapnp_init_one(struct pnp_dev *idev, const struct pnp_device_id *dev | |||
111 | ap->ioaddr.ctl_addr = ctl_addr; | 78 | ap->ioaddr.ctl_addr = ctl_addr; |
112 | } | 79 | } |
113 | 80 | ||
114 | ata_std_ports(&ap->ioaddr); | 81 | ata_sff_std_ports(&ap->ioaddr); |
115 | 82 | ||
116 | ata_port_desc(ap, "cmd 0x%llx ctl 0x%llx", | 83 | ata_port_desc(ap, "cmd 0x%llx ctl 0x%llx", |
117 | (unsigned long long)pnp_port_start(idev, 0), | 84 | (unsigned long long)pnp_port_start(idev, 0), |
diff --git a/drivers/ata/pata_it8213.c b/drivers/ata/pata_it8213.c index e0c2cc29d0ca..c113d7c079c8 100644 --- a/drivers/ata/pata_it8213.c +++ b/drivers/ata/pata_it8213.c | |||
@@ -40,20 +40,7 @@ static int it8213_pre_reset(struct ata_link *link, unsigned long deadline) | |||
40 | if (!pci_test_config_bits(pdev, &it8213_enable_bits[ap->port_no])) | 40 | if (!pci_test_config_bits(pdev, &it8213_enable_bits[ap->port_no])) |
41 | return -ENOENT; | 41 | return -ENOENT; |
42 | 42 | ||
43 | return ata_std_prereset(link, deadline); | 43 | return ata_sff_prereset(link, deadline); |
44 | } | ||
45 | |||
46 | /** | ||
47 | * it8213_error_handler - Probe specified port on PATA host controller | ||
48 | * @ap: Port to probe | ||
49 | * | ||
50 | * LOCKING: | ||
51 | * None (inherited from caller). | ||
52 | */ | ||
53 | |||
54 | static void it8213_error_handler(struct ata_port *ap) | ||
55 | { | ||
56 | ata_bmdma_drive_eh(ap, it8213_pre_reset, ata_std_softreset, NULL, ata_std_postreset); | ||
57 | } | 44 | } |
58 | 45 | ||
59 | /** | 46 | /** |
@@ -243,53 +230,16 @@ static void it8213_set_dmamode (struct ata_port *ap, struct ata_device *adev) | |||
243 | } | 230 | } |
244 | 231 | ||
245 | static struct scsi_host_template it8213_sht = { | 232 | static struct scsi_host_template it8213_sht = { |
246 | .module = THIS_MODULE, | 233 | ATA_BMDMA_SHT(DRV_NAME), |
247 | .name = DRV_NAME, | ||
248 | .ioctl = ata_scsi_ioctl, | ||
249 | .queuecommand = ata_scsi_queuecmd, | ||
250 | .can_queue = ATA_DEF_QUEUE, | ||
251 | .this_id = ATA_SHT_THIS_ID, | ||
252 | .sg_tablesize = LIBATA_MAX_PRD, | ||
253 | .max_sectors = ATA_MAX_SECTORS, | ||
254 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
255 | .emulated = ATA_SHT_EMULATED, | ||
256 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
257 | .proc_name = DRV_NAME, | ||
258 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
259 | .slave_configure = ata_scsi_slave_config, | ||
260 | .bios_param = ata_std_bios_param, | ||
261 | }; | 234 | }; |
262 | 235 | ||
263 | static const struct ata_port_operations it8213_ops = { | 236 | |
237 | static struct ata_port_operations it8213_ops = { | ||
238 | .inherits = &ata_bmdma_port_ops, | ||
239 | .cable_detect = it8213_cable_detect, | ||
264 | .set_piomode = it8213_set_piomode, | 240 | .set_piomode = it8213_set_piomode, |
265 | .set_dmamode = it8213_set_dmamode, | 241 | .set_dmamode = it8213_set_dmamode, |
266 | .mode_filter = ata_pci_default_filter, | 242 | .prereset = it8213_pre_reset, |
267 | |||
268 | .tf_load = ata_tf_load, | ||
269 | .tf_read = ata_tf_read, | ||
270 | .check_status = ata_check_status, | ||
271 | .exec_command = ata_exec_command, | ||
272 | .dev_select = ata_std_dev_select, | ||
273 | |||
274 | .freeze = ata_bmdma_freeze, | ||
275 | .thaw = ata_bmdma_thaw, | ||
276 | .error_handler = it8213_error_handler, | ||
277 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
278 | .cable_detect = it8213_cable_detect, | ||
279 | |||
280 | .bmdma_setup = ata_bmdma_setup, | ||
281 | .bmdma_start = ata_bmdma_start, | ||
282 | .bmdma_stop = ata_bmdma_stop, | ||
283 | .bmdma_status = ata_bmdma_status, | ||
284 | .qc_prep = ata_qc_prep, | ||
285 | .qc_issue = ata_qc_issue_prot, | ||
286 | .data_xfer = ata_data_xfer, | ||
287 | |||
288 | .irq_handler = ata_interrupt, | ||
289 | .irq_clear = ata_bmdma_irq_clear, | ||
290 | .irq_on = ata_irq_on, | ||
291 | |||
292 | .port_start = ata_sff_port_start, | ||
293 | }; | 243 | }; |
294 | 244 | ||
295 | 245 | ||
@@ -311,7 +261,6 @@ static int it8213_init_one (struct pci_dev *pdev, const struct pci_device_id *en | |||
311 | { | 261 | { |
312 | static int printed_version; | 262 | static int printed_version; |
313 | static const struct ata_port_info info = { | 263 | static const struct ata_port_info info = { |
314 | .sht = &it8213_sht, | ||
315 | .flags = ATA_FLAG_SLAVE_POSS, | 264 | .flags = ATA_FLAG_SLAVE_POSS, |
316 | .pio_mask = 0x1f, /* pio0-4 */ | 265 | .pio_mask = 0x1f, /* pio0-4 */ |
317 | .mwdma_mask = 0x07, /* mwdma0-2 */ | 266 | .mwdma_mask = 0x07, /* mwdma0-2 */ |
@@ -325,7 +274,7 @@ static int it8213_init_one (struct pci_dev *pdev, const struct pci_device_id *en | |||
325 | dev_printk(KERN_DEBUG, &pdev->dev, | 274 | dev_printk(KERN_DEBUG, &pdev->dev, |
326 | "version " DRV_VERSION "\n"); | 275 | "version " DRV_VERSION "\n"); |
327 | 276 | ||
328 | return ata_pci_init_one(pdev, ppi); | 277 | return ata_pci_sff_init_one(pdev, ppi, &it8213_sht, NULL); |
329 | } | 278 | } |
330 | 279 | ||
331 | static const struct pci_device_id it8213_pci_tbl[] = { | 280 | static const struct pci_device_id it8213_pci_tbl[] = { |
diff --git a/drivers/ata/pata_it821x.c b/drivers/ata/pata_it821x.c index 257951d03dbb..e10816931b2f 100644 --- a/drivers/ata/pata_it821x.c +++ b/drivers/ata/pata_it821x.c | |||
@@ -395,11 +395,11 @@ static void it821x_passthru_dev_select(struct ata_port *ap, | |||
395 | it821x_program(ap, adev, itdev->pio[adev->devno]); | 395 | it821x_program(ap, adev, itdev->pio[adev->devno]); |
396 | itdev->last_device = device; | 396 | itdev->last_device = device; |
397 | } | 397 | } |
398 | ata_std_dev_select(ap, device); | 398 | ata_sff_dev_select(ap, device); |
399 | } | 399 | } |
400 | 400 | ||
401 | /** | 401 | /** |
402 | * it821x_smart_qc_issue_prot - wrap qc issue prot | 402 | * it821x_smart_qc_issue - wrap qc issue prot |
403 | * @qc: command | 403 | * @qc: command |
404 | * | 404 | * |
405 | * Wrap the command issue sequence for the IT821x. We need to | 405 | * Wrap the command issue sequence for the IT821x. We need to |
@@ -407,7 +407,7 @@ static void it821x_passthru_dev_select(struct ata_port *ap, | |||
407 | * usual happenings kick off | 407 | * usual happenings kick off |
408 | */ | 408 | */ |
409 | 409 | ||
410 | static unsigned int it821x_smart_qc_issue_prot(struct ata_queued_cmd *qc) | 410 | static unsigned int it821x_smart_qc_issue(struct ata_queued_cmd *qc) |
411 | { | 411 | { |
412 | switch(qc->tf.command) | 412 | switch(qc->tf.command) |
413 | { | 413 | { |
@@ -427,14 +427,14 @@ static unsigned int it821x_smart_qc_issue_prot(struct ata_queued_cmd *qc) | |||
427 | case ATA_CMD_ID_ATA: | 427 | case ATA_CMD_ID_ATA: |
428 | /* Arguably should just no-op this one */ | 428 | /* Arguably should just no-op this one */ |
429 | case ATA_CMD_SET_FEATURES: | 429 | case ATA_CMD_SET_FEATURES: |
430 | return ata_qc_issue_prot(qc); | 430 | return ata_sff_qc_issue(qc); |
431 | } | 431 | } |
432 | printk(KERN_DEBUG "it821x: can't process command 0x%02X\n", qc->tf.command); | 432 | printk(KERN_DEBUG "it821x: can't process command 0x%02X\n", qc->tf.command); |
433 | return AC_ERR_DEV; | 433 | return AC_ERR_DEV; |
434 | } | 434 | } |
435 | 435 | ||
436 | /** | 436 | /** |
437 | * it821x_passthru_qc_issue_prot - wrap qc issue prot | 437 | * it821x_passthru_qc_issue - wrap qc issue prot |
438 | * @qc: command | 438 | * @qc: command |
439 | * | 439 | * |
440 | * Wrap the command issue sequence for the IT821x. We need to | 440 | * Wrap the command issue sequence for the IT821x. We need to |
@@ -442,10 +442,10 @@ static unsigned int it821x_smart_qc_issue_prot(struct ata_queued_cmd *qc) | |||
442 | * usual happenings kick off | 442 | * usual happenings kick off |
443 | */ | 443 | */ |
444 | 444 | ||
445 | static unsigned int it821x_passthru_qc_issue_prot(struct ata_queued_cmd *qc) | 445 | static unsigned int it821x_passthru_qc_issue(struct ata_queued_cmd *qc) |
446 | { | 446 | { |
447 | it821x_passthru_dev_select(qc->ap, qc->dev->devno); | 447 | it821x_passthru_dev_select(qc->ap, qc->dev->devno); |
448 | return ata_qc_issue_prot(qc); | 448 | return ata_sff_qc_issue(qc); |
449 | } | 449 | } |
450 | 450 | ||
451 | /** | 451 | /** |
@@ -632,89 +632,34 @@ static int it821x_port_start(struct ata_port *ap) | |||
632 | } | 632 | } |
633 | 633 | ||
634 | static struct scsi_host_template it821x_sht = { | 634 | static struct scsi_host_template it821x_sht = { |
635 | .module = THIS_MODULE, | 635 | ATA_BMDMA_SHT(DRV_NAME), |
636 | .name = DRV_NAME, | ||
637 | .ioctl = ata_scsi_ioctl, | ||
638 | .queuecommand = ata_scsi_queuecmd, | ||
639 | .can_queue = ATA_DEF_QUEUE, | ||
640 | .this_id = ATA_SHT_THIS_ID, | ||
641 | .sg_tablesize = LIBATA_MAX_PRD, | ||
642 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
643 | .emulated = ATA_SHT_EMULATED, | ||
644 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
645 | .proc_name = DRV_NAME, | ||
646 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
647 | .slave_configure = ata_scsi_slave_config, | ||
648 | .slave_destroy = ata_scsi_slave_destroy, | ||
649 | .bios_param = ata_std_bios_param, | ||
650 | }; | 636 | }; |
651 | 637 | ||
652 | static struct ata_port_operations it821x_smart_port_ops = { | 638 | static struct ata_port_operations it821x_smart_port_ops = { |
653 | .set_mode = it821x_smart_set_mode, | 639 | .inherits = &ata_bmdma_port_ops, |
654 | .tf_load = ata_tf_load, | ||
655 | .tf_read = ata_tf_read, | ||
656 | .mode_filter = ata_pci_default_filter, | ||
657 | 640 | ||
658 | .check_status = ata_check_status, | ||
659 | .check_atapi_dma= it821x_check_atapi_dma, | 641 | .check_atapi_dma= it821x_check_atapi_dma, |
660 | .exec_command = ata_exec_command, | 642 | .qc_issue = it821x_smart_qc_issue, |
661 | .dev_select = ata_std_dev_select, | ||
662 | .dev_config = it821x_dev_config, | ||
663 | 643 | ||
664 | .freeze = ata_bmdma_freeze, | ||
665 | .thaw = ata_bmdma_thaw, | ||
666 | .error_handler = ata_bmdma_error_handler, | ||
667 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
668 | .cable_detect = it821x_ident_hack, | 644 | .cable_detect = it821x_ident_hack, |
669 | 645 | .set_mode = it821x_smart_set_mode, | |
670 | .bmdma_setup = ata_bmdma_setup, | 646 | .dev_config = it821x_dev_config, |
671 | .bmdma_start = ata_bmdma_start, | ||
672 | .bmdma_stop = ata_bmdma_stop, | ||
673 | .bmdma_status = ata_bmdma_status, | ||
674 | |||
675 | .qc_prep = ata_qc_prep, | ||
676 | .qc_issue = it821x_smart_qc_issue_prot, | ||
677 | |||
678 | .data_xfer = ata_data_xfer, | ||
679 | |||
680 | .irq_handler = ata_interrupt, | ||
681 | .irq_clear = ata_bmdma_irq_clear, | ||
682 | .irq_on = ata_irq_on, | ||
683 | 647 | ||
684 | .port_start = it821x_port_start, | 648 | .port_start = it821x_port_start, |
685 | }; | 649 | }; |
686 | 650 | ||
687 | static struct ata_port_operations it821x_passthru_port_ops = { | 651 | static struct ata_port_operations it821x_passthru_port_ops = { |
688 | .set_piomode = it821x_passthru_set_piomode, | 652 | .inherits = &ata_bmdma_port_ops, |
689 | .set_dmamode = it821x_passthru_set_dmamode, | ||
690 | .mode_filter = ata_pci_default_filter, | ||
691 | 653 | ||
692 | .tf_load = ata_tf_load, | ||
693 | .tf_read = ata_tf_read, | ||
694 | .check_status = ata_check_status, | ||
695 | .exec_command = ata_exec_command, | ||
696 | .check_atapi_dma= it821x_check_atapi_dma, | 654 | .check_atapi_dma= it821x_check_atapi_dma, |
697 | .dev_select = it821x_passthru_dev_select, | 655 | .sff_dev_select = it821x_passthru_dev_select, |
698 | |||
699 | .freeze = ata_bmdma_freeze, | ||
700 | .thaw = ata_bmdma_thaw, | ||
701 | .error_handler = ata_bmdma_error_handler, | ||
702 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
703 | .cable_detect = ata_cable_unknown, | ||
704 | |||
705 | .bmdma_setup = ata_bmdma_setup, | ||
706 | .bmdma_start = it821x_passthru_bmdma_start, | 656 | .bmdma_start = it821x_passthru_bmdma_start, |
707 | .bmdma_stop = it821x_passthru_bmdma_stop, | 657 | .bmdma_stop = it821x_passthru_bmdma_stop, |
708 | .bmdma_status = ata_bmdma_status, | 658 | .qc_issue = it821x_passthru_qc_issue, |
709 | |||
710 | .qc_prep = ata_qc_prep, | ||
711 | .qc_issue = it821x_passthru_qc_issue_prot, | ||
712 | |||
713 | .data_xfer = ata_data_xfer, | ||
714 | 659 | ||
715 | .irq_clear = ata_bmdma_irq_clear, | 660 | .cable_detect = ata_cable_unknown, |
716 | .irq_handler = ata_interrupt, | 661 | .set_piomode = it821x_passthru_set_piomode, |
717 | .irq_on = ata_irq_on, | 662 | .set_dmamode = it821x_passthru_set_dmamode, |
718 | 663 | ||
719 | .port_start = it821x_port_start, | 664 | .port_start = it821x_port_start, |
720 | }; | 665 | }; |
@@ -742,14 +687,12 @@ static int it821x_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
742 | u8 conf; | 687 | u8 conf; |
743 | 688 | ||
744 | static const struct ata_port_info info_smart = { | 689 | static const struct ata_port_info info_smart = { |
745 | .sht = &it821x_sht, | ||
746 | .flags = ATA_FLAG_SLAVE_POSS, | 690 | .flags = ATA_FLAG_SLAVE_POSS, |
747 | .pio_mask = 0x1f, | 691 | .pio_mask = 0x1f, |
748 | .mwdma_mask = 0x07, | 692 | .mwdma_mask = 0x07, |
749 | .port_ops = &it821x_smart_port_ops | 693 | .port_ops = &it821x_smart_port_ops |
750 | }; | 694 | }; |
751 | static const struct ata_port_info info_passthru = { | 695 | static const struct ata_port_info info_passthru = { |
752 | .sht = &it821x_sht, | ||
753 | .flags = ATA_FLAG_SLAVE_POSS, | 696 | .flags = ATA_FLAG_SLAVE_POSS, |
754 | .pio_mask = 0x1f, | 697 | .pio_mask = 0x1f, |
755 | .mwdma_mask = 0x07, | 698 | .mwdma_mask = 0x07, |
@@ -759,6 +702,11 @@ static int it821x_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
759 | 702 | ||
760 | const struct ata_port_info *ppi[] = { NULL, NULL }; | 703 | const struct ata_port_info *ppi[] = { NULL, NULL }; |
761 | static char *mode[2] = { "pass through", "smart" }; | 704 | static char *mode[2] = { "pass through", "smart" }; |
705 | int rc; | ||
706 | |||
707 | rc = pcim_enable_device(pdev); | ||
708 | if (rc) | ||
709 | return rc; | ||
762 | 710 | ||
763 | /* Force the card into bypass mode if so requested */ | 711 | /* Force the card into bypass mode if so requested */ |
764 | if (it8212_noraid) { | 712 | if (it8212_noraid) { |
@@ -774,16 +722,23 @@ static int it821x_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
774 | else | 722 | else |
775 | ppi[0] = &info_smart; | 723 | ppi[0] = &info_smart; |
776 | 724 | ||
777 | return ata_pci_init_one(pdev, ppi); | 725 | return ata_pci_sff_init_one(pdev, ppi, &it821x_sht, NULL); |
778 | } | 726 | } |
779 | 727 | ||
780 | #ifdef CONFIG_PM | 728 | #ifdef CONFIG_PM |
781 | static int it821x_reinit_one(struct pci_dev *pdev) | 729 | static int it821x_reinit_one(struct pci_dev *pdev) |
782 | { | 730 | { |
731 | struct ata_host *host = dev_get_drvdata(&pdev->dev); | ||
732 | int rc; | ||
733 | |||
734 | rc = ata_pci_device_do_resume(pdev); | ||
735 | if (rc) | ||
736 | return rc; | ||
783 | /* Resume - turn raid back off if need be */ | 737 | /* Resume - turn raid back off if need be */ |
784 | if (it8212_noraid) | 738 | if (it8212_noraid) |
785 | it821x_disable_raid(pdev); | 739 | it821x_disable_raid(pdev); |
786 | return ata_pci_device_resume(pdev); | 740 | ata_host_resume(host); |
741 | return rc; | ||
787 | } | 742 | } |
788 | #endif | 743 | #endif |
789 | 744 | ||
diff --git a/drivers/ata/pata_ixp4xx_cf.c b/drivers/ata/pata_ixp4xx_cf.c index 030878fedeb5..8a175f23b907 100644 --- a/drivers/ata/pata_ixp4xx_cf.c +++ b/drivers/ata/pata_ixp4xx_cf.c | |||
@@ -88,48 +88,14 @@ static unsigned int ixp4xx_mmio_data_xfer(struct ata_device *dev, | |||
88 | } | 88 | } |
89 | 89 | ||
90 | static struct scsi_host_template ixp4xx_sht = { | 90 | static struct scsi_host_template ixp4xx_sht = { |
91 | .module = THIS_MODULE, | 91 | ATA_PIO_SHT(DRV_NAME), |
92 | .name = DRV_NAME, | ||
93 | .ioctl = ata_scsi_ioctl, | ||
94 | .queuecommand = ata_scsi_queuecmd, | ||
95 | .can_queue = ATA_DEF_QUEUE, | ||
96 | .this_id = ATA_SHT_THIS_ID, | ||
97 | .sg_tablesize = LIBATA_MAX_PRD, | ||
98 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
99 | .emulated = ATA_SHT_EMULATED, | ||
100 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
101 | .proc_name = DRV_NAME, | ||
102 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
103 | .slave_configure = ata_scsi_slave_config, | ||
104 | .slave_destroy = ata_scsi_slave_destroy, | ||
105 | .bios_param = ata_std_bios_param, | ||
106 | }; | 92 | }; |
107 | 93 | ||
108 | static struct ata_port_operations ixp4xx_port_ops = { | 94 | static struct ata_port_operations ixp4xx_port_ops = { |
109 | .set_mode = ixp4xx_set_mode, | 95 | .inherits = &ata_sff_port_ops, |
110 | .mode_filter = ata_pci_default_filter, | 96 | .sff_data_xfer = ixp4xx_mmio_data_xfer, |
111 | |||
112 | .tf_load = ata_tf_load, | ||
113 | .tf_read = ata_tf_read, | ||
114 | .exec_command = ata_exec_command, | ||
115 | .check_status = ata_check_status, | ||
116 | .dev_select = ata_std_dev_select, | ||
117 | |||
118 | .freeze = ata_bmdma_freeze, | ||
119 | .thaw = ata_bmdma_thaw, | ||
120 | .error_handler = ata_bmdma_error_handler, | ||
121 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
122 | |||
123 | .qc_prep = ata_qc_prep, | ||
124 | .qc_issue = ata_qc_issue_prot, | ||
125 | .data_xfer = ixp4xx_mmio_data_xfer, | ||
126 | .cable_detect = ata_cable_40wire, | 97 | .cable_detect = ata_cable_40wire, |
127 | 98 | .set_mode = ixp4xx_set_mode, | |
128 | .irq_handler = ata_interrupt, | ||
129 | .irq_clear = ata_bmdma_irq_clear, | ||
130 | .irq_on = ata_irq_on, | ||
131 | |||
132 | .port_start = ata_port_start, | ||
133 | }; | 99 | }; |
134 | 100 | ||
135 | static void ixp4xx_setup_port(struct ata_port *ap, | 101 | static void ixp4xx_setup_port(struct ata_port *ap, |
@@ -144,7 +110,7 @@ static void ixp4xx_setup_port(struct ata_port *ap, | |||
144 | ioaddr->altstatus_addr = data->cs1 + 0x06; | 110 | ioaddr->altstatus_addr = data->cs1 + 0x06; |
145 | ioaddr->ctl_addr = data->cs1 + 0x06; | 111 | ioaddr->ctl_addr = data->cs1 + 0x06; |
146 | 112 | ||
147 | ata_std_ports(ioaddr); | 113 | ata_sff_std_ports(ioaddr); |
148 | 114 | ||
149 | #ifndef __ARMEB__ | 115 | #ifndef __ARMEB__ |
150 | 116 | ||
@@ -220,7 +186,7 @@ static __devinit int ixp4xx_pata_probe(struct platform_device *pdev) | |||
220 | dev_printk(KERN_INFO, &pdev->dev, "version " DRV_VERSION "\n"); | 186 | dev_printk(KERN_INFO, &pdev->dev, "version " DRV_VERSION "\n"); |
221 | 187 | ||
222 | /* activate host */ | 188 | /* activate host */ |
223 | return ata_host_activate(host, irq, ata_interrupt, 0, &ixp4xx_sht); | 189 | return ata_host_activate(host, irq, ata_sff_interrupt, 0, &ixp4xx_sht); |
224 | } | 190 | } |
225 | 191 | ||
226 | static __devexit int ixp4xx_pata_remove(struct platform_device *dev) | 192 | static __devexit int ixp4xx_pata_remove(struct platform_device *dev) |
diff --git a/drivers/ata/pata_jmicron.c b/drivers/ata/pata_jmicron.c index 00bbbbd50e97..73b7596816b4 100644 --- a/drivers/ata/pata_jmicron.c +++ b/drivers/ata/pata_jmicron.c | |||
@@ -102,73 +102,18 @@ static int jmicron_pre_reset(struct ata_link *link, unsigned long deadline) | |||
102 | ap->cbl = ATA_CBL_SATA; | 102 | ap->cbl = ATA_CBL_SATA; |
103 | break; | 103 | break; |
104 | } | 104 | } |
105 | return ata_std_prereset(link, deadline); | 105 | return ata_sff_prereset(link, deadline); |
106 | } | ||
107 | |||
108 | /** | ||
109 | * jmicron_error_handler - Setup and error handler | ||
110 | * @ap: Port to handle | ||
111 | * | ||
112 | * LOCKING: | ||
113 | * None (inherited from caller). | ||
114 | */ | ||
115 | |||
116 | static void jmicron_error_handler(struct ata_port *ap) | ||
117 | { | ||
118 | ata_bmdma_drive_eh(ap, jmicron_pre_reset, ata_std_softreset, NULL, | ||
119 | ata_std_postreset); | ||
120 | } | 106 | } |
121 | 107 | ||
122 | /* No PIO or DMA methods needed for this device */ | 108 | /* No PIO or DMA methods needed for this device */ |
123 | 109 | ||
124 | static struct scsi_host_template jmicron_sht = { | 110 | static struct scsi_host_template jmicron_sht = { |
125 | .module = THIS_MODULE, | 111 | ATA_BMDMA_SHT(DRV_NAME), |
126 | .name = DRV_NAME, | ||
127 | .ioctl = ata_scsi_ioctl, | ||
128 | .queuecommand = ata_scsi_queuecmd, | ||
129 | .can_queue = ATA_DEF_QUEUE, | ||
130 | .this_id = ATA_SHT_THIS_ID, | ||
131 | .sg_tablesize = LIBATA_MAX_PRD, | ||
132 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
133 | .emulated = ATA_SHT_EMULATED, | ||
134 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
135 | .proc_name = DRV_NAME, | ||
136 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
137 | .slave_configure = ata_scsi_slave_config, | ||
138 | .slave_destroy = ata_scsi_slave_destroy, | ||
139 | /* Use standard CHS mapping rules */ | ||
140 | .bios_param = ata_std_bios_param, | ||
141 | }; | 112 | }; |
142 | 113 | ||
143 | static const struct ata_port_operations jmicron_ops = { | 114 | static struct ata_port_operations jmicron_ops = { |
144 | /* Task file is PCI ATA format, use helpers */ | 115 | .inherits = &ata_bmdma_port_ops, |
145 | .tf_load = ata_tf_load, | 116 | .prereset = jmicron_pre_reset, |
146 | .tf_read = ata_tf_read, | ||
147 | .check_status = ata_check_status, | ||
148 | .exec_command = ata_exec_command, | ||
149 | .dev_select = ata_std_dev_select, | ||
150 | |||
151 | .freeze = ata_bmdma_freeze, | ||
152 | .thaw = ata_bmdma_thaw, | ||
153 | .error_handler = jmicron_error_handler, | ||
154 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
155 | |||
156 | /* BMDMA handling is PCI ATA format, use helpers */ | ||
157 | .bmdma_setup = ata_bmdma_setup, | ||
158 | .bmdma_start = ata_bmdma_start, | ||
159 | .bmdma_stop = ata_bmdma_stop, | ||
160 | .bmdma_status = ata_bmdma_status, | ||
161 | .qc_prep = ata_qc_prep, | ||
162 | .qc_issue = ata_qc_issue_prot, | ||
163 | .data_xfer = ata_data_xfer, | ||
164 | |||
165 | /* IRQ-related hooks */ | ||
166 | .irq_handler = ata_interrupt, | ||
167 | .irq_clear = ata_bmdma_irq_clear, | ||
168 | .irq_on = ata_irq_on, | ||
169 | |||
170 | /* Generic PATA PCI ATA helpers */ | ||
171 | .port_start = ata_port_start, | ||
172 | }; | 117 | }; |
173 | 118 | ||
174 | 119 | ||
@@ -189,7 +134,6 @@ static const struct ata_port_operations jmicron_ops = { | |||
189 | static int jmicron_init_one (struct pci_dev *pdev, const struct pci_device_id *id) | 134 | static int jmicron_init_one (struct pci_dev *pdev, const struct pci_device_id *id) |
190 | { | 135 | { |
191 | static const struct ata_port_info info = { | 136 | static const struct ata_port_info info = { |
192 | .sht = &jmicron_sht, | ||
193 | .flags = ATA_FLAG_SLAVE_POSS, | 137 | .flags = ATA_FLAG_SLAVE_POSS, |
194 | 138 | ||
195 | .pio_mask = 0x1f, | 139 | .pio_mask = 0x1f, |
@@ -200,7 +144,7 @@ static int jmicron_init_one (struct pci_dev *pdev, const struct pci_device_id *i | |||
200 | }; | 144 | }; |
201 | const struct ata_port_info *ppi[] = { &info, NULL }; | 145 | const struct ata_port_info *ppi[] = { &info, NULL }; |
202 | 146 | ||
203 | return ata_pci_init_one(pdev, ppi); | 147 | return ata_pci_sff_init_one(pdev, ppi, &jmicron_sht, NULL); |
204 | } | 148 | } |
205 | 149 | ||
206 | static const struct pci_device_id jmicron_pci_tbl[] = { | 150 | static const struct pci_device_id jmicron_pci_tbl[] = { |
diff --git a/drivers/ata/pata_legacy.c b/drivers/ata/pata_legacy.c index 50fe08ebe23c..7af4b29cc422 100644 --- a/drivers/ata/pata_legacy.c +++ b/drivers/ata/pata_legacy.c | |||
@@ -208,21 +208,12 @@ static int legacy_set_mode(struct ata_link *link, struct ata_device **unused) | |||
208 | } | 208 | } |
209 | 209 | ||
210 | static struct scsi_host_template legacy_sht = { | 210 | static struct scsi_host_template legacy_sht = { |
211 | .module = THIS_MODULE, | 211 | ATA_PIO_SHT(DRV_NAME), |
212 | .name = DRV_NAME, | 212 | }; |
213 | .ioctl = ata_scsi_ioctl, | 213 | |
214 | .queuecommand = ata_scsi_queuecmd, | 214 | static const struct ata_port_operations legacy_base_port_ops = { |
215 | .can_queue = ATA_DEF_QUEUE, | 215 | .inherits = &ata_sff_port_ops, |
216 | .this_id = ATA_SHT_THIS_ID, | 216 | .cable_detect = ata_cable_40wire, |
217 | .sg_tablesize = LIBATA_MAX_PRD, | ||
218 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
219 | .emulated = ATA_SHT_EMULATED, | ||
220 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
221 | .proc_name = DRV_NAME, | ||
222 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
223 | .slave_configure = ata_scsi_slave_config, | ||
224 | .slave_destroy = ata_scsi_slave_destroy, | ||
225 | .bios_param = ata_std_bios_param, | ||
226 | }; | 217 | }; |
227 | 218 | ||
228 | /* | 219 | /* |
@@ -234,55 +225,14 @@ static struct scsi_host_template legacy_sht = { | |||
234 | */ | 225 | */ |
235 | 226 | ||
236 | static struct ata_port_operations simple_port_ops = { | 227 | static struct ata_port_operations simple_port_ops = { |
237 | .tf_load = ata_tf_load, | 228 | .inherits = &legacy_base_port_ops, |
238 | .tf_read = ata_tf_read, | 229 | .sff_data_xfer = ata_sff_data_xfer_noirq, |
239 | .check_status = ata_check_status, | ||
240 | .exec_command = ata_exec_command, | ||
241 | .dev_select = ata_std_dev_select, | ||
242 | |||
243 | .freeze = ata_bmdma_freeze, | ||
244 | .thaw = ata_bmdma_thaw, | ||
245 | .error_handler = ata_bmdma_error_handler, | ||
246 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
247 | .cable_detect = ata_cable_40wire, | ||
248 | |||
249 | .qc_prep = ata_qc_prep, | ||
250 | .qc_issue = ata_qc_issue_prot, | ||
251 | |||
252 | .data_xfer = ata_data_xfer_noirq, | ||
253 | |||
254 | .irq_handler = ata_interrupt, | ||
255 | .irq_clear = ata_bmdma_irq_clear, | ||
256 | .irq_on = ata_irq_on, | ||
257 | |||
258 | .port_start = ata_sff_port_start, | ||
259 | }; | 230 | }; |
260 | 231 | ||
261 | static struct ata_port_operations legacy_port_ops = { | 232 | static struct ata_port_operations legacy_port_ops = { |
233 | .inherits = &legacy_base_port_ops, | ||
234 | .sff_data_xfer = ata_sff_data_xfer_noirq, | ||
262 | .set_mode = legacy_set_mode, | 235 | .set_mode = legacy_set_mode, |
263 | |||
264 | .tf_load = ata_tf_load, | ||
265 | .tf_read = ata_tf_read, | ||
266 | .check_status = ata_check_status, | ||
267 | .exec_command = ata_exec_command, | ||
268 | .dev_select = ata_std_dev_select, | ||
269 | .cable_detect = ata_cable_40wire, | ||
270 | |||
271 | .freeze = ata_bmdma_freeze, | ||
272 | .thaw = ata_bmdma_thaw, | ||
273 | .error_handler = ata_bmdma_error_handler, | ||
274 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
275 | |||
276 | .qc_prep = ata_qc_prep, | ||
277 | .qc_issue = ata_qc_issue_prot, | ||
278 | |||
279 | .data_xfer = ata_data_xfer_noirq, | ||
280 | |||
281 | .irq_handler = ata_interrupt, | ||
282 | .irq_clear = ata_bmdma_irq_clear, | ||
283 | .irq_on = ata_irq_on, | ||
284 | |||
285 | .port_start = ata_sff_port_start, | ||
286 | }; | 236 | }; |
287 | 237 | ||
288 | /* | 238 | /* |
@@ -367,36 +317,15 @@ static unsigned int pdc_data_xfer_vlb(struct ata_device *dev, | |||
367 | } | 317 | } |
368 | local_irq_restore(flags); | 318 | local_irq_restore(flags); |
369 | } else | 319 | } else |
370 | buflen = ata_data_xfer_noirq(dev, buf, buflen, rw); | 320 | buflen = ata_sff_data_xfer_noirq(dev, buf, buflen, rw); |
371 | 321 | ||
372 | return buflen; | 322 | return buflen; |
373 | } | 323 | } |
374 | 324 | ||
375 | static struct ata_port_operations pdc20230_port_ops = { | 325 | static struct ata_port_operations pdc20230_port_ops = { |
326 | .inherits = &legacy_base_port_ops, | ||
376 | .set_piomode = pdc20230_set_piomode, | 327 | .set_piomode = pdc20230_set_piomode, |
377 | 328 | .sff_data_xfer = pdc_data_xfer_vlb, | |
378 | .tf_load = ata_tf_load, | ||
379 | .tf_read = ata_tf_read, | ||
380 | .check_status = ata_check_status, | ||
381 | .exec_command = ata_exec_command, | ||
382 | .dev_select = ata_std_dev_select, | ||
383 | |||
384 | .freeze = ata_bmdma_freeze, | ||
385 | .thaw = ata_bmdma_thaw, | ||
386 | .error_handler = ata_bmdma_error_handler, | ||
387 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
388 | .cable_detect = ata_cable_40wire, | ||
389 | |||
390 | .qc_prep = ata_qc_prep, | ||
391 | .qc_issue = ata_qc_issue_prot, | ||
392 | |||
393 | .data_xfer = pdc_data_xfer_vlb, | ||
394 | |||
395 | .irq_handler = ata_interrupt, | ||
396 | .irq_clear = ata_bmdma_irq_clear, | ||
397 | .irq_on = ata_irq_on, | ||
398 | |||
399 | .port_start = ata_sff_port_start, | ||
400 | }; | 329 | }; |
401 | 330 | ||
402 | /* | 331 | /* |
@@ -427,30 +356,8 @@ static void ht6560a_set_piomode(struct ata_port *ap, struct ata_device *adev) | |||
427 | } | 356 | } |
428 | 357 | ||
429 | static struct ata_port_operations ht6560a_port_ops = { | 358 | static struct ata_port_operations ht6560a_port_ops = { |
359 | .inherits = &legacy_base_port_ops, | ||
430 | .set_piomode = ht6560a_set_piomode, | 360 | .set_piomode = ht6560a_set_piomode, |
431 | |||
432 | .tf_load = ata_tf_load, | ||
433 | .tf_read = ata_tf_read, | ||
434 | .check_status = ata_check_status, | ||
435 | .exec_command = ata_exec_command, | ||
436 | .dev_select = ata_std_dev_select, | ||
437 | |||
438 | .freeze = ata_bmdma_freeze, | ||
439 | .thaw = ata_bmdma_thaw, | ||
440 | .error_handler = ata_bmdma_error_handler, | ||
441 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
442 | .cable_detect = ata_cable_40wire, | ||
443 | |||
444 | .qc_prep = ata_qc_prep, | ||
445 | .qc_issue = ata_qc_issue_prot, | ||
446 | |||
447 | .data_xfer = ata_data_xfer, /* Check vlb/noirq */ | ||
448 | |||
449 | .irq_handler = ata_interrupt, | ||
450 | .irq_clear = ata_bmdma_irq_clear, | ||
451 | .irq_on = ata_irq_on, | ||
452 | |||
453 | .port_start = ata_sff_port_start, | ||
454 | }; | 361 | }; |
455 | 362 | ||
456 | /* | 363 | /* |
@@ -492,30 +399,8 @@ static void ht6560b_set_piomode(struct ata_port *ap, struct ata_device *adev) | |||
492 | } | 399 | } |
493 | 400 | ||
494 | static struct ata_port_operations ht6560b_port_ops = { | 401 | static struct ata_port_operations ht6560b_port_ops = { |
402 | .inherits = &legacy_base_port_ops, | ||
495 | .set_piomode = ht6560b_set_piomode, | 403 | .set_piomode = ht6560b_set_piomode, |
496 | |||
497 | .tf_load = ata_tf_load, | ||
498 | .tf_read = ata_tf_read, | ||
499 | .check_status = ata_check_status, | ||
500 | .exec_command = ata_exec_command, | ||
501 | .dev_select = ata_std_dev_select, | ||
502 | |||
503 | .freeze = ata_bmdma_freeze, | ||
504 | .thaw = ata_bmdma_thaw, | ||
505 | .error_handler = ata_bmdma_error_handler, | ||
506 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
507 | .cable_detect = ata_cable_40wire, | ||
508 | |||
509 | .qc_prep = ata_qc_prep, | ||
510 | .qc_issue = ata_qc_issue_prot, | ||
511 | |||
512 | .data_xfer = ata_data_xfer, /* FIXME: Check 32bit and noirq */ | ||
513 | |||
514 | .irq_handler = ata_interrupt, | ||
515 | .irq_clear = ata_bmdma_irq_clear, | ||
516 | .irq_on = ata_irq_on, | ||
517 | |||
518 | .port_start = ata_sff_port_start, | ||
519 | }; | 404 | }; |
520 | 405 | ||
521 | /* | 406 | /* |
@@ -613,30 +498,8 @@ static void opti82c611a_set_piomode(struct ata_port *ap, | |||
613 | 498 | ||
614 | 499 | ||
615 | static struct ata_port_operations opti82c611a_port_ops = { | 500 | static struct ata_port_operations opti82c611a_port_ops = { |
501 | .inherits = &legacy_base_port_ops, | ||
616 | .set_piomode = opti82c611a_set_piomode, | 502 | .set_piomode = opti82c611a_set_piomode, |
617 | |||
618 | .tf_load = ata_tf_load, | ||
619 | .tf_read = ata_tf_read, | ||
620 | .check_status = ata_check_status, | ||
621 | .exec_command = ata_exec_command, | ||
622 | .dev_select = ata_std_dev_select, | ||
623 | |||
624 | .freeze = ata_bmdma_freeze, | ||
625 | .thaw = ata_bmdma_thaw, | ||
626 | .error_handler = ata_bmdma_error_handler, | ||
627 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
628 | .cable_detect = ata_cable_40wire, | ||
629 | |||
630 | .qc_prep = ata_qc_prep, | ||
631 | .qc_issue = ata_qc_issue_prot, | ||
632 | |||
633 | .data_xfer = ata_data_xfer, | ||
634 | |||
635 | .irq_handler = ata_interrupt, | ||
636 | .irq_clear = ata_bmdma_irq_clear, | ||
637 | .irq_on = ata_irq_on, | ||
638 | |||
639 | .port_start = ata_sff_port_start, | ||
640 | }; | 503 | }; |
641 | 504 | ||
642 | /* | 505 | /* |
@@ -716,7 +579,7 @@ static void opti82c46x_set_piomode(struct ata_port *ap, struct ata_device *adev) | |||
716 | } | 579 | } |
717 | 580 | ||
718 | /** | 581 | /** |
719 | * opt82c465mv_qc_issue_prot - command issue | 582 | * opt82c465mv_qc_issue - command issue |
720 | * @qc: command pending | 583 | * @qc: command pending |
721 | * | 584 | * |
722 | * Called when the libata layer is about to issue a command. We wrap | 585 | * Called when the libata layer is about to issue a command. We wrap |
@@ -730,7 +593,7 @@ static void opti82c46x_set_piomode(struct ata_port *ap, struct ata_device *adev) | |||
730 | * FIXME: dual channel needs ->serialize support | 593 | * FIXME: dual channel needs ->serialize support |
731 | */ | 594 | */ |
732 | 595 | ||
733 | static unsigned int opti82c46x_qc_issue_prot(struct ata_queued_cmd *qc) | 596 | static unsigned int opti82c46x_qc_issue(struct ata_queued_cmd *qc) |
734 | { | 597 | { |
735 | struct ata_port *ap = qc->ap; | 598 | struct ata_port *ap = qc->ap; |
736 | struct ata_device *adev = qc->dev; | 599 | struct ata_device *adev = qc->dev; |
@@ -741,34 +604,13 @@ static unsigned int opti82c46x_qc_issue_prot(struct ata_queued_cmd *qc) | |||
741 | && ap->host->private_data != NULL) | 604 | && ap->host->private_data != NULL) |
742 | opti82c46x_set_piomode(ap, adev); | 605 | opti82c46x_set_piomode(ap, adev); |
743 | 606 | ||
744 | return ata_qc_issue_prot(qc); | 607 | return ata_sff_qc_issue(qc); |
745 | } | 608 | } |
746 | 609 | ||
747 | static struct ata_port_operations opti82c46x_port_ops = { | 610 | static struct ata_port_operations opti82c46x_port_ops = { |
611 | .inherits = &legacy_base_port_ops, | ||
748 | .set_piomode = opti82c46x_set_piomode, | 612 | .set_piomode = opti82c46x_set_piomode, |
749 | 613 | .qc_issue = opti82c46x_qc_issue, | |
750 | .tf_load = ata_tf_load, | ||
751 | .tf_read = ata_tf_read, | ||
752 | .check_status = ata_check_status, | ||
753 | .exec_command = ata_exec_command, | ||
754 | .dev_select = ata_std_dev_select, | ||
755 | |||
756 | .freeze = ata_bmdma_freeze, | ||
757 | .thaw = ata_bmdma_thaw, | ||
758 | .error_handler = ata_bmdma_error_handler, | ||
759 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
760 | .cable_detect = ata_cable_40wire, | ||
761 | |||
762 | .qc_prep = ata_qc_prep, | ||
763 | .qc_issue = opti82c46x_qc_issue_prot, | ||
764 | |||
765 | .data_xfer = ata_data_xfer, | ||
766 | |||
767 | .irq_handler = ata_interrupt, | ||
768 | .irq_clear = ata_bmdma_irq_clear, | ||
769 | .irq_on = ata_irq_on, | ||
770 | |||
771 | .port_start = ata_sff_port_start, | ||
772 | }; | 614 | }; |
773 | 615 | ||
774 | static void qdi6500_set_piomode(struct ata_port *ap, struct ata_device *adev) | 616 | static void qdi6500_set_piomode(struct ata_port *ap, struct ata_device *adev) |
@@ -802,7 +644,7 @@ static void qdi6500_set_piomode(struct ata_port *ap, struct ata_device *adev) | |||
802 | * @irq: interrupt line | 644 | * @irq: interrupt line |
803 | * | 645 | * |
804 | * In dual channel mode the 6580 has one clock per channel and we have | 646 | * In dual channel mode the 6580 has one clock per channel and we have |
805 | * to software clockswitch in qc_issue_prot. | 647 | * to software clockswitch in qc_issue. |
806 | */ | 648 | */ |
807 | 649 | ||
808 | static void qdi6580dp_set_piomode(struct ata_port *ap, struct ata_device *adev) | 650 | static void qdi6580dp_set_piomode(struct ata_port *ap, struct ata_device *adev) |
@@ -868,14 +710,14 @@ static void qdi6580_set_piomode(struct ata_port *ap, struct ata_device *adev) | |||
868 | } | 710 | } |
869 | 711 | ||
870 | /** | 712 | /** |
871 | * qdi_qc_issue_prot - command issue | 713 | * qdi_qc_issue - command issue |
872 | * @qc: command pending | 714 | * @qc: command pending |
873 | * | 715 | * |
874 | * Called when the libata layer is about to issue a command. We wrap | 716 | * Called when the libata layer is about to issue a command. We wrap |
875 | * this interface so that we can load the correct ATA timings. | 717 | * this interface so that we can load the correct ATA timings. |
876 | */ | 718 | */ |
877 | 719 | ||
878 | static unsigned int qdi_qc_issue_prot(struct ata_queued_cmd *qc) | 720 | static unsigned int qdi_qc_issue(struct ata_queued_cmd *qc) |
879 | { | 721 | { |
880 | struct ata_port *ap = qc->ap; | 722 | struct ata_port *ap = qc->ap; |
881 | struct ata_device *adev = qc->dev; | 723 | struct ata_device *adev = qc->dev; |
@@ -888,7 +730,7 @@ static unsigned int qdi_qc_issue_prot(struct ata_queued_cmd *qc) | |||
888 | 2 * ap->port_no); | 730 | 2 * ap->port_no); |
889 | } | 731 | } |
890 | } | 732 | } |
891 | return ata_qc_issue_prot(qc); | 733 | return ata_sff_qc_issue(qc); |
892 | } | 734 | } |
893 | 735 | ||
894 | static unsigned int vlb32_data_xfer(struct ata_device *adev, unsigned char *buf, | 736 | static unsigned int vlb32_data_xfer(struct ata_device *adev, unsigned char *buf, |
@@ -917,7 +759,7 @@ static unsigned int vlb32_data_xfer(struct ata_device *adev, unsigned char *buf, | |||
917 | } | 759 | } |
918 | return (buflen + 3) & ~3; | 760 | return (buflen + 3) & ~3; |
919 | } else | 761 | } else |
920 | return ata_data_xfer(adev, buf, buflen, rw); | 762 | return ata_sff_data_xfer(adev, buf, buflen, rw); |
921 | } | 763 | } |
922 | 764 | ||
923 | static int qdi_port(struct platform_device *dev, | 765 | static int qdi_port(struct platform_device *dev, |
@@ -930,84 +772,22 @@ static int qdi_port(struct platform_device *dev, | |||
930 | } | 772 | } |
931 | 773 | ||
932 | static struct ata_port_operations qdi6500_port_ops = { | 774 | static struct ata_port_operations qdi6500_port_ops = { |
775 | .inherits = &legacy_base_port_ops, | ||
933 | .set_piomode = qdi6500_set_piomode, | 776 | .set_piomode = qdi6500_set_piomode, |
934 | 777 | .qc_issue = qdi_qc_issue, | |
935 | .tf_load = ata_tf_load, | 778 | .sff_data_xfer = vlb32_data_xfer, |
936 | .tf_read = ata_tf_read, | ||
937 | .check_status = ata_check_status, | ||
938 | .exec_command = ata_exec_command, | ||
939 | .dev_select = ata_std_dev_select, | ||
940 | |||
941 | .freeze = ata_bmdma_freeze, | ||
942 | .thaw = ata_bmdma_thaw, | ||
943 | .error_handler = ata_bmdma_error_handler, | ||
944 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
945 | .cable_detect = ata_cable_40wire, | ||
946 | |||
947 | .qc_prep = ata_qc_prep, | ||
948 | .qc_issue = qdi_qc_issue_prot, | ||
949 | |||
950 | .data_xfer = vlb32_data_xfer, | ||
951 | |||
952 | .irq_handler = ata_interrupt, | ||
953 | .irq_clear = ata_bmdma_irq_clear, | ||
954 | .irq_on = ata_irq_on, | ||
955 | |||
956 | .port_start = ata_sff_port_start, | ||
957 | }; | 779 | }; |
958 | 780 | ||
959 | static struct ata_port_operations qdi6580_port_ops = { | 781 | static struct ata_port_operations qdi6580_port_ops = { |
782 | .inherits = &legacy_base_port_ops, | ||
960 | .set_piomode = qdi6580_set_piomode, | 783 | .set_piomode = qdi6580_set_piomode, |
961 | 784 | .sff_data_xfer = vlb32_data_xfer, | |
962 | .tf_load = ata_tf_load, | ||
963 | .tf_read = ata_tf_read, | ||
964 | .check_status = ata_check_status, | ||
965 | .exec_command = ata_exec_command, | ||
966 | .dev_select = ata_std_dev_select, | ||
967 | |||
968 | .freeze = ata_bmdma_freeze, | ||
969 | .thaw = ata_bmdma_thaw, | ||
970 | .error_handler = ata_bmdma_error_handler, | ||
971 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
972 | .cable_detect = ata_cable_40wire, | ||
973 | |||
974 | .qc_prep = ata_qc_prep, | ||
975 | .qc_issue = ata_qc_issue_prot, | ||
976 | |||
977 | .data_xfer = vlb32_data_xfer, | ||
978 | |||
979 | .irq_handler = ata_interrupt, | ||
980 | .irq_clear = ata_bmdma_irq_clear, | ||
981 | .irq_on = ata_irq_on, | ||
982 | |||
983 | .port_start = ata_sff_port_start, | ||
984 | }; | 785 | }; |
985 | 786 | ||
986 | static struct ata_port_operations qdi6580dp_port_ops = { | 787 | static struct ata_port_operations qdi6580dp_port_ops = { |
788 | .inherits = &legacy_base_port_ops, | ||
987 | .set_piomode = qdi6580dp_set_piomode, | 789 | .set_piomode = qdi6580dp_set_piomode, |
988 | 790 | .sff_data_xfer = vlb32_data_xfer, | |
989 | .tf_load = ata_tf_load, | ||
990 | .tf_read = ata_tf_read, | ||
991 | .check_status = ata_check_status, | ||
992 | .exec_command = ata_exec_command, | ||
993 | .dev_select = ata_std_dev_select, | ||
994 | |||
995 | .freeze = ata_bmdma_freeze, | ||
996 | .thaw = ata_bmdma_thaw, | ||
997 | .error_handler = ata_bmdma_error_handler, | ||
998 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
999 | .cable_detect = ata_cable_40wire, | ||
1000 | |||
1001 | .qc_prep = ata_qc_prep, | ||
1002 | .qc_issue = qdi_qc_issue_prot, | ||
1003 | |||
1004 | .data_xfer = vlb32_data_xfer, | ||
1005 | |||
1006 | .irq_handler = ata_interrupt, | ||
1007 | .irq_clear = ata_bmdma_irq_clear, | ||
1008 | .irq_on = ata_irq_on, | ||
1009 | |||
1010 | .port_start = ata_sff_port_start, | ||
1011 | }; | 791 | }; |
1012 | 792 | ||
1013 | static DEFINE_SPINLOCK(winbond_lock); | 793 | static DEFINE_SPINLOCK(winbond_lock); |
@@ -1076,29 +856,9 @@ static int winbond_port(struct platform_device *dev, | |||
1076 | } | 856 | } |
1077 | 857 | ||
1078 | static struct ata_port_operations winbond_port_ops = { | 858 | static struct ata_port_operations winbond_port_ops = { |
859 | .inherits = &legacy_base_port_ops, | ||
1079 | .set_piomode = winbond_set_piomode, | 860 | .set_piomode = winbond_set_piomode, |
1080 | 861 | .sff_data_xfer = vlb32_data_xfer, | |
1081 | .tf_load = ata_tf_load, | ||
1082 | .tf_read = ata_tf_read, | ||
1083 | .check_status = ata_check_status, | ||
1084 | .exec_command = ata_exec_command, | ||
1085 | .dev_select = ata_std_dev_select, | ||
1086 | |||
1087 | .freeze = ata_bmdma_freeze, | ||
1088 | .thaw = ata_bmdma_thaw, | ||
1089 | .error_handler = ata_bmdma_error_handler, | ||
1090 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
1091 | .cable_detect = ata_cable_40wire, | ||
1092 | |||
1093 | .qc_prep = ata_qc_prep, | ||
1094 | .qc_issue = ata_qc_issue_prot, | ||
1095 | |||
1096 | .data_xfer = vlb32_data_xfer, | ||
1097 | |||
1098 | .irq_clear = ata_bmdma_irq_clear, | ||
1099 | .irq_on = ata_irq_on, | ||
1100 | |||
1101 | .port_start = ata_sff_port_start, | ||
1102 | }; | 862 | }; |
1103 | 863 | ||
1104 | static struct legacy_controller controllers[] = { | 864 | static struct legacy_controller controllers[] = { |
@@ -1256,13 +1016,13 @@ static __init int legacy_init_one(struct legacy_probe *probe) | |||
1256 | ap->ioaddr.cmd_addr = io_addr; | 1016 | ap->ioaddr.cmd_addr = io_addr; |
1257 | ap->ioaddr.altstatus_addr = ctrl_addr; | 1017 | ap->ioaddr.altstatus_addr = ctrl_addr; |
1258 | ap->ioaddr.ctl_addr = ctrl_addr; | 1018 | ap->ioaddr.ctl_addr = ctrl_addr; |
1259 | ata_std_ports(&ap->ioaddr); | 1019 | ata_sff_std_ports(&ap->ioaddr); |
1260 | ap->host->private_data = ld; | 1020 | ap->host->private_data = ld; |
1261 | 1021 | ||
1262 | ata_port_desc(ap, "cmd 0x%lx ctl 0x%lx", io, io + 0x0206); | 1022 | ata_port_desc(ap, "cmd 0x%lx ctl 0x%lx", io, io + 0x0206); |
1263 | 1023 | ||
1264 | ret = ata_host_activate(host, probe->irq, ata_interrupt, 0, | 1024 | ret = ata_host_activate(host, probe->irq, ata_sff_interrupt, 0, |
1265 | &legacy_sht); | 1025 | &legacy_sht); |
1266 | if (ret) | 1026 | if (ret) |
1267 | goto fail; | 1027 | goto fail; |
1268 | ld->platform_dev = pdev; | 1028 | ld->platform_dev = pdev; |
diff --git a/drivers/ata/pata_marvell.c b/drivers/ata/pata_marvell.c index a81f25d87235..24a011b25024 100644 --- a/drivers/ata/pata_marvell.c +++ b/drivers/ata/pata_marvell.c | |||
@@ -55,7 +55,7 @@ static int marvell_pre_reset(struct ata_link *link, unsigned long deadline) | |||
55 | (!(devices & 0x10))) /* PATA enable ? */ | 55 | (!(devices & 0x10))) /* PATA enable ? */ |
56 | return -ENOENT; | 56 | return -ENOENT; |
57 | 57 | ||
58 | return ata_std_prereset(link, deadline); | 58 | return ata_sff_prereset(link, deadline); |
59 | } | 59 | } |
60 | 60 | ||
61 | static int marvell_cable_detect(struct ata_port *ap) | 61 | static int marvell_cable_detect(struct ata_port *ap) |
@@ -75,71 +75,16 @@ static int marvell_cable_detect(struct ata_port *ap) | |||
75 | return 0; /* Our BUG macro needs the right markup */ | 75 | return 0; /* Our BUG macro needs the right markup */ |
76 | } | 76 | } |
77 | 77 | ||
78 | /** | ||
79 | * marvell_error_handler - Setup and error handler | ||
80 | * @ap: Port to handle | ||
81 | * | ||
82 | * LOCKING: | ||
83 | * None (inherited from caller). | ||
84 | */ | ||
85 | |||
86 | static void marvell_error_handler(struct ata_port *ap) | ||
87 | { | ||
88 | ata_bmdma_drive_eh(ap, marvell_pre_reset, ata_std_softreset, NULL, | ||
89 | ata_std_postreset); | ||
90 | } | ||
91 | |||
92 | /* No PIO or DMA methods needed for this device */ | 78 | /* No PIO or DMA methods needed for this device */ |
93 | 79 | ||
94 | static struct scsi_host_template marvell_sht = { | 80 | static struct scsi_host_template marvell_sht = { |
95 | .module = THIS_MODULE, | 81 | ATA_BMDMA_SHT(DRV_NAME), |
96 | .name = DRV_NAME, | ||
97 | .ioctl = ata_scsi_ioctl, | ||
98 | .queuecommand = ata_scsi_queuecmd, | ||
99 | .can_queue = ATA_DEF_QUEUE, | ||
100 | .this_id = ATA_SHT_THIS_ID, | ||
101 | .sg_tablesize = LIBATA_MAX_PRD, | ||
102 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
103 | .emulated = ATA_SHT_EMULATED, | ||
104 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
105 | .proc_name = DRV_NAME, | ||
106 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
107 | .slave_configure = ata_scsi_slave_config, | ||
108 | .slave_destroy = ata_scsi_slave_destroy, | ||
109 | /* Use standard CHS mapping rules */ | ||
110 | .bios_param = ata_std_bios_param, | ||
111 | }; | 82 | }; |
112 | 83 | ||
113 | static const struct ata_port_operations marvell_ops = { | 84 | static struct ata_port_operations marvell_ops = { |
114 | /* Task file is PCI ATA format, use helpers */ | 85 | .inherits = &ata_bmdma_port_ops, |
115 | .tf_load = ata_tf_load, | ||
116 | .tf_read = ata_tf_read, | ||
117 | .check_status = ata_check_status, | ||
118 | .exec_command = ata_exec_command, | ||
119 | .dev_select = ata_std_dev_select, | ||
120 | |||
121 | .freeze = ata_bmdma_freeze, | ||
122 | .thaw = ata_bmdma_thaw, | ||
123 | .error_handler = marvell_error_handler, | ||
124 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
125 | .cable_detect = marvell_cable_detect, | 86 | .cable_detect = marvell_cable_detect, |
126 | 87 | .prereset = marvell_pre_reset, | |
127 | /* BMDMA handling is PCI ATA format, use helpers */ | ||
128 | .bmdma_setup = ata_bmdma_setup, | ||
129 | .bmdma_start = ata_bmdma_start, | ||
130 | .bmdma_stop = ata_bmdma_stop, | ||
131 | .bmdma_status = ata_bmdma_status, | ||
132 | .qc_prep = ata_qc_prep, | ||
133 | .qc_issue = ata_qc_issue_prot, | ||
134 | .data_xfer = ata_data_xfer, | ||
135 | |||
136 | /* Timeout handling */ | ||
137 | .irq_handler = ata_interrupt, | ||
138 | .irq_clear = ata_bmdma_irq_clear, | ||
139 | .irq_on = ata_irq_on, | ||
140 | |||
141 | /* Generic PATA PCI ATA helpers */ | ||
142 | .port_start = ata_sff_port_start, | ||
143 | }; | 88 | }; |
144 | 89 | ||
145 | 90 | ||
@@ -160,7 +105,6 @@ static const struct ata_port_operations marvell_ops = { | |||
160 | static int marvell_init_one (struct pci_dev *pdev, const struct pci_device_id *id) | 105 | static int marvell_init_one (struct pci_dev *pdev, const struct pci_device_id *id) |
161 | { | 106 | { |
162 | static const struct ata_port_info info = { | 107 | static const struct ata_port_info info = { |
163 | .sht = &marvell_sht, | ||
164 | .flags = ATA_FLAG_SLAVE_POSS, | 108 | .flags = ATA_FLAG_SLAVE_POSS, |
165 | 109 | ||
166 | .pio_mask = 0x1f, | 110 | .pio_mask = 0x1f, |
@@ -170,7 +114,6 @@ static int marvell_init_one (struct pci_dev *pdev, const struct pci_device_id *i | |||
170 | .port_ops = &marvell_ops, | 114 | .port_ops = &marvell_ops, |
171 | }; | 115 | }; |
172 | static const struct ata_port_info info_sata = { | 116 | static const struct ata_port_info info_sata = { |
173 | .sht = &marvell_sht, | ||
174 | /* Slave possible as its magically mapped not real */ | 117 | /* Slave possible as its magically mapped not real */ |
175 | .flags = ATA_FLAG_SLAVE_POSS, | 118 | .flags = ATA_FLAG_SLAVE_POSS, |
176 | 119 | ||
@@ -185,7 +128,7 @@ static int marvell_init_one (struct pci_dev *pdev, const struct pci_device_id *i | |||
185 | if (pdev->device == 0x6101) | 128 | if (pdev->device == 0x6101) |
186 | ppi[1] = &ata_dummy_port_info; | 129 | ppi[1] = &ata_dummy_port_info; |
187 | 130 | ||
188 | return ata_pci_init_one(pdev, ppi); | 131 | return ata_pci_sff_init_one(pdev, ppi, &marvell_sht, NULL); |
189 | } | 132 | } |
190 | 133 | ||
191 | static const struct pci_device_id marvell_pci_tbl[] = { | 134 | static const struct pci_device_id marvell_pci_tbl[] = { |
diff --git a/drivers/ata/pata_mpc52xx.c b/drivers/ata/pata_mpc52xx.c index 5413ebfa72e5..bc79df6e7cb0 100644 --- a/drivers/ata/pata_mpc52xx.c +++ b/drivers/ata/pata_mpc52xx.c | |||
@@ -252,53 +252,19 @@ mpc52xx_ata_dev_select(struct ata_port *ap, unsigned int device) | |||
252 | if (device != priv->csel) | 252 | if (device != priv->csel) |
253 | mpc52xx_ata_apply_timings(priv, device); | 253 | mpc52xx_ata_apply_timings(priv, device); |
254 | 254 | ||
255 | ata_std_dev_select(ap,device); | 255 | ata_sff_dev_select(ap,device); |
256 | } | 256 | } |
257 | 257 | ||
258 | static void | ||
259 | mpc52xx_ata_error_handler(struct ata_port *ap) | ||
260 | { | ||
261 | ata_bmdma_drive_eh(ap, ata_std_prereset, ata_std_softreset, NULL, | ||
262 | ata_std_postreset); | ||
263 | } | ||
264 | |||
265 | |||
266 | |||
267 | static struct scsi_host_template mpc52xx_ata_sht = { | 258 | static struct scsi_host_template mpc52xx_ata_sht = { |
268 | .module = THIS_MODULE, | 259 | ATA_PIO_SHT(DRV_NAME), |
269 | .name = DRV_NAME, | ||
270 | .ioctl = ata_scsi_ioctl, | ||
271 | .queuecommand = ata_scsi_queuecmd, | ||
272 | .can_queue = ATA_DEF_QUEUE, | ||
273 | .this_id = ATA_SHT_THIS_ID, | ||
274 | .sg_tablesize = LIBATA_MAX_PRD, | ||
275 | .max_sectors = ATA_MAX_SECTORS, | ||
276 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
277 | .emulated = ATA_SHT_EMULATED, | ||
278 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
279 | .proc_name = DRV_NAME, | ||
280 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
281 | .slave_configure = ata_scsi_slave_config, | ||
282 | .bios_param = ata_std_bios_param, | ||
283 | }; | 260 | }; |
284 | 261 | ||
285 | static struct ata_port_operations mpc52xx_ata_port_ops = { | 262 | static struct ata_port_operations mpc52xx_ata_port_ops = { |
286 | .set_piomode = mpc52xx_ata_set_piomode, | 263 | .inherits = &ata_sff_port_ops, |
287 | .dev_select = mpc52xx_ata_dev_select, | 264 | .sff_dev_select = mpc52xx_ata_dev_select, |
288 | .tf_load = ata_tf_load, | ||
289 | .tf_read = ata_tf_read, | ||
290 | .check_status = ata_check_status, | ||
291 | .exec_command = ata_exec_command, | ||
292 | .freeze = ata_bmdma_freeze, | ||
293 | .thaw = ata_bmdma_thaw, | ||
294 | .error_handler = mpc52xx_ata_error_handler, | ||
295 | .cable_detect = ata_cable_40wire, | 265 | .cable_detect = ata_cable_40wire, |
296 | .qc_prep = ata_qc_prep, | 266 | .set_piomode = mpc52xx_ata_set_piomode, |
297 | .qc_issue = ata_qc_issue_prot, | 267 | .post_internal_cmd = ATA_OP_NULL, |
298 | .data_xfer = ata_data_xfer, | ||
299 | .irq_clear = ata_bmdma_irq_clear, | ||
300 | .irq_on = ata_irq_on, | ||
301 | .port_start = ata_port_start, | ||
302 | }; | 268 | }; |
303 | 269 | ||
304 | static int __devinit | 270 | static int __devinit |
@@ -339,7 +305,7 @@ mpc52xx_ata_init_one(struct device *dev, struct mpc52xx_ata_priv *priv, | |||
339 | ata_port_desc(ap, "ata_regs 0x%lx", raw_ata_regs); | 305 | ata_port_desc(ap, "ata_regs 0x%lx", raw_ata_regs); |
340 | 306 | ||
341 | /* activate host */ | 307 | /* activate host */ |
342 | return ata_host_activate(host, priv->ata_irq, ata_interrupt, 0, | 308 | return ata_host_activate(host, priv->ata_irq, ata_sff_interrupt, 0, |
343 | &mpc52xx_ata_sht); | 309 | &mpc52xx_ata_sht); |
344 | } | 310 | } |
345 | 311 | ||
diff --git a/drivers/ata/pata_mpiix.c b/drivers/ata/pata_mpiix.c index c0d9e0cf208c..7d7e3fdab71f 100644 --- a/drivers/ata/pata_mpiix.c +++ b/drivers/ata/pata_mpiix.c | |||
@@ -55,21 +55,7 @@ static int mpiix_pre_reset(struct ata_link *link, unsigned long deadline) | |||
55 | if (!pci_test_config_bits(pdev, &mpiix_enable_bits)) | 55 | if (!pci_test_config_bits(pdev, &mpiix_enable_bits)) |
56 | return -ENOENT; | 56 | return -ENOENT; |
57 | 57 | ||
58 | return ata_std_prereset(link, deadline); | 58 | return ata_sff_prereset(link, deadline); |
59 | } | ||
60 | |||
61 | /** | ||
62 | * mpiix_error_handler - probe reset | ||
63 | * @ap: ATA port | ||
64 | * | ||
65 | * Perform the ATA probe and bus reset sequence plus specific handling | ||
66 | * for this hardware. The MPIIX has the enable bits in a different place | ||
67 | * to PIIX4 and friends. As a pure PIO device it has no cable detect | ||
68 | */ | ||
69 | |||
70 | static void mpiix_error_handler(struct ata_port *ap) | ||
71 | { | ||
72 | ata_bmdma_drive_eh(ap, mpiix_pre_reset, ata_std_softreset, NULL, ata_std_postreset); | ||
73 | } | 59 | } |
74 | 60 | ||
75 | /** | 61 | /** |
@@ -83,8 +69,8 @@ static void mpiix_error_handler(struct ata_port *ap) | |||
83 | * | 69 | * |
84 | * This would get very ugly because we can only program timing for one | 70 | * This would get very ugly because we can only program timing for one |
85 | * device at a time, the other gets PIO0. Fortunately libata calls | 71 | * device at a time, the other gets PIO0. Fortunately libata calls |
86 | * our qc_issue_prot command before a command is issued so we can | 72 | * our qc_issue command before a command is issued so we can flip the |
87 | * flip the timings back and forth to reduce the pain. | 73 | * timings back and forth to reduce the pain. |
88 | */ | 74 | */ |
89 | 75 | ||
90 | static void mpiix_set_piomode(struct ata_port *ap, struct ata_device *adev) | 76 | static void mpiix_set_piomode(struct ata_port *ap, struct ata_device *adev) |
@@ -124,7 +110,7 @@ static void mpiix_set_piomode(struct ata_port *ap, struct ata_device *adev) | |||
124 | } | 110 | } |
125 | 111 | ||
126 | /** | 112 | /** |
127 | * mpiix_qc_issue_prot - command issue | 113 | * mpiix_qc_issue - command issue |
128 | * @qc: command pending | 114 | * @qc: command pending |
129 | * | 115 | * |
130 | * Called when the libata layer is about to issue a command. We wrap | 116 | * Called when the libata layer is about to issue a command. We wrap |
@@ -134,7 +120,7 @@ static void mpiix_set_piomode(struct ata_port *ap, struct ata_device *adev) | |||
134 | * be made PIO0. | 120 | * be made PIO0. |
135 | */ | 121 | */ |
136 | 122 | ||
137 | static unsigned int mpiix_qc_issue_prot(struct ata_queued_cmd *qc) | 123 | static unsigned int mpiix_qc_issue(struct ata_queued_cmd *qc) |
138 | { | 124 | { |
139 | struct ata_port *ap = qc->ap; | 125 | struct ata_port *ap = qc->ap; |
140 | struct ata_device *adev = qc->dev; | 126 | struct ata_device *adev = qc->dev; |
@@ -147,50 +133,19 @@ static unsigned int mpiix_qc_issue_prot(struct ata_queued_cmd *qc) | |||
147 | if (adev->pio_mode && adev != ap->private_data) | 133 | if (adev->pio_mode && adev != ap->private_data) |
148 | mpiix_set_piomode(ap, adev); | 134 | mpiix_set_piomode(ap, adev); |
149 | 135 | ||
150 | return ata_qc_issue_prot(qc); | 136 | return ata_sff_qc_issue(qc); |
151 | } | 137 | } |
152 | 138 | ||
153 | static struct scsi_host_template mpiix_sht = { | 139 | static struct scsi_host_template mpiix_sht = { |
154 | .module = THIS_MODULE, | 140 | ATA_PIO_SHT(DRV_NAME), |
155 | .name = DRV_NAME, | ||
156 | .ioctl = ata_scsi_ioctl, | ||
157 | .queuecommand = ata_scsi_queuecmd, | ||
158 | .can_queue = ATA_DEF_QUEUE, | ||
159 | .this_id = ATA_SHT_THIS_ID, | ||
160 | .sg_tablesize = LIBATA_MAX_PRD, | ||
161 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
162 | .emulated = ATA_SHT_EMULATED, | ||
163 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
164 | .proc_name = DRV_NAME, | ||
165 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
166 | .slave_configure = ata_scsi_slave_config, | ||
167 | .slave_destroy = ata_scsi_slave_destroy, | ||
168 | .bios_param = ata_std_bios_param, | ||
169 | }; | 141 | }; |
170 | 142 | ||
171 | static struct ata_port_operations mpiix_port_ops = { | 143 | static struct ata_port_operations mpiix_port_ops = { |
172 | .set_piomode = mpiix_set_piomode, | 144 | .inherits = &ata_sff_port_ops, |
173 | 145 | .qc_issue = mpiix_qc_issue, | |
174 | .tf_load = ata_tf_load, | ||
175 | .tf_read = ata_tf_read, | ||
176 | .check_status = ata_check_status, | ||
177 | .exec_command = ata_exec_command, | ||
178 | .dev_select = ata_std_dev_select, | ||
179 | |||
180 | .freeze = ata_bmdma_freeze, | ||
181 | .thaw = ata_bmdma_thaw, | ||
182 | .error_handler = mpiix_error_handler, | ||
183 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
184 | .cable_detect = ata_cable_40wire, | 146 | .cable_detect = ata_cable_40wire, |
185 | 147 | .set_piomode = mpiix_set_piomode, | |
186 | .qc_prep = ata_qc_prep, | 148 | .prereset = mpiix_pre_reset, |
187 | .qc_issue = mpiix_qc_issue_prot, | ||
188 | .data_xfer = ata_data_xfer, | ||
189 | |||
190 | .irq_clear = ata_bmdma_irq_clear, | ||
191 | .irq_on = ata_irq_on, | ||
192 | |||
193 | .port_start = ata_sff_port_start, | ||
194 | }; | 149 | }; |
195 | 150 | ||
196 | static int mpiix_init_one(struct pci_dev *dev, const struct pci_device_id *id) | 151 | static int mpiix_init_one(struct pci_dev *dev, const struct pci_device_id *id) |
@@ -252,10 +207,10 @@ static int mpiix_init_one(struct pci_dev *dev, const struct pci_device_id *id) | |||
252 | ap->ioaddr.altstatus_addr = ctl_addr; | 207 | ap->ioaddr.altstatus_addr = ctl_addr; |
253 | 208 | ||
254 | /* Let libata fill in the port details */ | 209 | /* Let libata fill in the port details */ |
255 | ata_std_ports(&ap->ioaddr); | 210 | ata_sff_std_ports(&ap->ioaddr); |
256 | 211 | ||
257 | /* activate host */ | 212 | /* activate host */ |
258 | return ata_host_activate(host, irq, ata_interrupt, IRQF_SHARED, | 213 | return ata_host_activate(host, irq, ata_sff_interrupt, IRQF_SHARED, |
259 | &mpiix_sht); | 214 | &mpiix_sht); |
260 | } | 215 | } |
261 | 216 | ||
diff --git a/drivers/ata/pata_netcell.c b/drivers/ata/pata_netcell.c index 25c922abd554..d9719c8b9dbe 100644 --- a/drivers/ata/pata_netcell.c +++ b/drivers/ata/pata_netcell.c | |||
@@ -21,54 +21,12 @@ | |||
21 | /* No PIO or DMA methods needed for this device */ | 21 | /* No PIO or DMA methods needed for this device */ |
22 | 22 | ||
23 | static struct scsi_host_template netcell_sht = { | 23 | static struct scsi_host_template netcell_sht = { |
24 | .module = THIS_MODULE, | 24 | ATA_BMDMA_SHT(DRV_NAME), |
25 | .name = DRV_NAME, | ||
26 | .ioctl = ata_scsi_ioctl, | ||
27 | .queuecommand = ata_scsi_queuecmd, | ||
28 | .can_queue = ATA_DEF_QUEUE, | ||
29 | .this_id = ATA_SHT_THIS_ID, | ||
30 | .sg_tablesize = LIBATA_MAX_PRD, | ||
31 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
32 | .emulated = ATA_SHT_EMULATED, | ||
33 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
34 | .proc_name = DRV_NAME, | ||
35 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
36 | .slave_configure = ata_scsi_slave_config, | ||
37 | .slave_destroy = ata_scsi_slave_destroy, | ||
38 | /* Use standard CHS mapping rules */ | ||
39 | .bios_param = ata_std_bios_param, | ||
40 | }; | 25 | }; |
41 | 26 | ||
42 | static const struct ata_port_operations netcell_ops = { | 27 | static struct ata_port_operations netcell_ops = { |
43 | /* Task file is PCI ATA format, use helpers */ | 28 | .inherits = &ata_bmdma_port_ops, |
44 | .tf_load = ata_tf_load, | ||
45 | .tf_read = ata_tf_read, | ||
46 | .check_status = ata_check_status, | ||
47 | .exec_command = ata_exec_command, | ||
48 | .dev_select = ata_std_dev_select, | ||
49 | |||
50 | .freeze = ata_bmdma_freeze, | ||
51 | .thaw = ata_bmdma_thaw, | ||
52 | .error_handler = ata_bmdma_error_handler, | ||
53 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
54 | .cable_detect = ata_cable_80wire, | 29 | .cable_detect = ata_cable_80wire, |
55 | |||
56 | /* BMDMA handling is PCI ATA format, use helpers */ | ||
57 | .bmdma_setup = ata_bmdma_setup, | ||
58 | .bmdma_start = ata_bmdma_start, | ||
59 | .bmdma_stop = ata_bmdma_stop, | ||
60 | .bmdma_status = ata_bmdma_status, | ||
61 | .qc_prep = ata_qc_prep, | ||
62 | .qc_issue = ata_qc_issue_prot, | ||
63 | .data_xfer = ata_data_xfer, | ||
64 | |||
65 | /* IRQ-related hooks */ | ||
66 | .irq_handler = ata_interrupt, | ||
67 | .irq_clear = ata_bmdma_irq_clear, | ||
68 | .irq_on = ata_irq_on, | ||
69 | |||
70 | /* Generic PATA PCI ATA helpers */ | ||
71 | .port_start = ata_sff_port_start, | ||
72 | }; | 30 | }; |
73 | 31 | ||
74 | 32 | ||
@@ -90,7 +48,6 @@ static int netcell_init_one (struct pci_dev *pdev, const struct pci_device_id *e | |||
90 | { | 48 | { |
91 | static int printed_version; | 49 | static int printed_version; |
92 | static const struct ata_port_info info = { | 50 | static const struct ata_port_info info = { |
93 | .sht = &netcell_sht, | ||
94 | .flags = ATA_FLAG_SLAVE_POSS, | 51 | .flags = ATA_FLAG_SLAVE_POSS, |
95 | /* Actually we don't really care about these as the | 52 | /* Actually we don't really care about these as the |
96 | firmware deals with it */ | 53 | firmware deals with it */ |
@@ -100,16 +57,21 @@ static int netcell_init_one (struct pci_dev *pdev, const struct pci_device_id *e | |||
100 | .port_ops = &netcell_ops, | 57 | .port_ops = &netcell_ops, |
101 | }; | 58 | }; |
102 | const struct ata_port_info *port_info[] = { &info, NULL }; | 59 | const struct ata_port_info *port_info[] = { &info, NULL }; |
60 | int rc; | ||
103 | 61 | ||
104 | if (!printed_version++) | 62 | if (!printed_version++) |
105 | dev_printk(KERN_DEBUG, &pdev->dev, | 63 | dev_printk(KERN_DEBUG, &pdev->dev, |
106 | "version " DRV_VERSION "\n"); | 64 | "version " DRV_VERSION "\n"); |
107 | 65 | ||
66 | rc = pcim_enable_device(pdev); | ||
67 | if (rc) | ||
68 | return rc; | ||
69 | |||
108 | /* Any chip specific setup/optimisation/messages here */ | 70 | /* Any chip specific setup/optimisation/messages here */ |
109 | ata_pci_clear_simplex(pdev); | 71 | ata_pci_bmdma_clear_simplex(pdev); |
110 | 72 | ||
111 | /* And let the library code do the work */ | 73 | /* And let the library code do the work */ |
112 | return ata_pci_init_one(pdev, port_info); | 74 | return ata_pci_sff_init_one(pdev, port_info, &netcell_sht, NULL); |
113 | } | 75 | } |
114 | 76 | ||
115 | static const struct pci_device_id netcell_pci_tbl[] = { | 77 | static const struct pci_device_id netcell_pci_tbl[] = { |
diff --git a/drivers/ata/pata_ninja32.c b/drivers/ata/pata_ninja32.c index 15dd649f89ee..565e67cd13fa 100644 --- a/drivers/ata/pata_ninja32.c +++ b/drivers/ata/pata_ninja32.c | |||
@@ -73,60 +73,20 @@ static void ninja32_dev_select(struct ata_port *ap, unsigned int device) | |||
73 | struct ata_device *adev = &ap->link.device[device]; | 73 | struct ata_device *adev = &ap->link.device[device]; |
74 | if (ap->private_data != adev) { | 74 | if (ap->private_data != adev) { |
75 | iowrite8(0xd6, ap->ioaddr.bmdma_addr + 0x1f); | 75 | iowrite8(0xd6, ap->ioaddr.bmdma_addr + 0x1f); |
76 | ata_std_dev_select(ap, device); | 76 | ata_sff_dev_select(ap, device); |
77 | ninja32_set_piomode(ap, adev); | 77 | ninja32_set_piomode(ap, adev); |
78 | } | 78 | } |
79 | } | 79 | } |
80 | 80 | ||
81 | static struct scsi_host_template ninja32_sht = { | 81 | static struct scsi_host_template ninja32_sht = { |
82 | .module = THIS_MODULE, | 82 | ATA_BMDMA_SHT(DRV_NAME), |
83 | .name = DRV_NAME, | ||
84 | .ioctl = ata_scsi_ioctl, | ||
85 | .queuecommand = ata_scsi_queuecmd, | ||
86 | .can_queue = ATA_DEF_QUEUE, | ||
87 | .this_id = ATA_SHT_THIS_ID, | ||
88 | .sg_tablesize = LIBATA_MAX_PRD, | ||
89 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
90 | .emulated = ATA_SHT_EMULATED, | ||
91 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
92 | .proc_name = DRV_NAME, | ||
93 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
94 | .slave_configure = ata_scsi_slave_config, | ||
95 | .slave_destroy = ata_scsi_slave_destroy, | ||
96 | .bios_param = ata_std_bios_param, | ||
97 | }; | 83 | }; |
98 | 84 | ||
99 | static struct ata_port_operations ninja32_port_ops = { | 85 | static struct ata_port_operations ninja32_port_ops = { |
100 | .set_piomode = ninja32_set_piomode, | 86 | .inherits = &ata_bmdma_port_ops, |
101 | .mode_filter = ata_pci_default_filter, | 87 | .sff_dev_select = ninja32_dev_select, |
102 | |||
103 | .tf_load = ata_tf_load, | ||
104 | .tf_read = ata_tf_read, | ||
105 | .check_status = ata_check_status, | ||
106 | .exec_command = ata_exec_command, | ||
107 | .dev_select = ninja32_dev_select, | ||
108 | |||
109 | .freeze = ata_bmdma_freeze, | ||
110 | .thaw = ata_bmdma_thaw, | ||
111 | .error_handler = ata_bmdma_error_handler, | ||
112 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
113 | .cable_detect = ata_cable_40wire, | 88 | .cable_detect = ata_cable_40wire, |
114 | 89 | .set_piomode = ninja32_set_piomode, | |
115 | .bmdma_setup = ata_bmdma_setup, | ||
116 | .bmdma_start = ata_bmdma_start, | ||
117 | .bmdma_stop = ata_bmdma_stop, | ||
118 | .bmdma_status = ata_bmdma_status, | ||
119 | |||
120 | .qc_prep = ata_qc_prep, | ||
121 | .qc_issue = ata_qc_issue_prot, | ||
122 | |||
123 | .data_xfer = ata_data_xfer, | ||
124 | |||
125 | .irq_handler = ata_interrupt, | ||
126 | .irq_clear = ata_bmdma_irq_clear, | ||
127 | .irq_on = ata_irq_on, | ||
128 | |||
129 | .port_start = ata_sff_port_start, | ||
130 | }; | 90 | }; |
131 | 91 | ||
132 | static int ninja32_init_one(struct pci_dev *dev, const struct pci_device_id *id) | 92 | static int ninja32_init_one(struct pci_dev *dev, const struct pci_device_id *id) |
@@ -172,7 +132,7 @@ static int ninja32_init_one(struct pci_dev *dev, const struct pci_device_id *id) | |||
172 | ap->ioaddr.ctl_addr = base + 0x1E; | 132 | ap->ioaddr.ctl_addr = base + 0x1E; |
173 | ap->ioaddr.altstatus_addr = base + 0x1E; | 133 | ap->ioaddr.altstatus_addr = base + 0x1E; |
174 | ap->ioaddr.bmdma_addr = base; | 134 | ap->ioaddr.bmdma_addr = base; |
175 | ata_std_ports(&ap->ioaddr); | 135 | ata_sff_std_ports(&ap->ioaddr); |
176 | 136 | ||
177 | iowrite8(0x05, base + 0x01); /* Enable interrupt lines */ | 137 | iowrite8(0x05, base + 0x01); /* Enable interrupt lines */ |
178 | iowrite8(0xBE, base + 0x02); /* Burst, ?? setup */ | 138 | iowrite8(0xBE, base + 0x02); /* Burst, ?? setup */ |
@@ -182,7 +142,7 @@ static int ninja32_init_one(struct pci_dev *dev, const struct pci_device_id *id) | |||
182 | iowrite8(0xa4, base + 0x1c); /* Unknown */ | 142 | iowrite8(0xa4, base + 0x1c); /* Unknown */ |
183 | iowrite8(0x83, base + 0x1d); /* BMDMA control: WAIT0 */ | 143 | iowrite8(0x83, base + 0x1d); /* BMDMA control: WAIT0 */ |
184 | /* FIXME: Should we disable them at remove ? */ | 144 | /* FIXME: Should we disable them at remove ? */ |
185 | return ata_host_activate(host, dev->irq, ata_interrupt, | 145 | return ata_host_activate(host, dev->irq, ata_sff_interrupt, |
186 | IRQF_SHARED, &ninja32_sht); | 146 | IRQF_SHARED, &ninja32_sht); |
187 | } | 147 | } |
188 | 148 | ||
diff --git a/drivers/ata/pata_ns87410.c b/drivers/ata/pata_ns87410.c index 9fe66fd75017..76d2455bc453 100644 --- a/drivers/ata/pata_ns87410.c +++ b/drivers/ata/pata_ns87410.c | |||
@@ -50,21 +50,7 @@ static int ns87410_pre_reset(struct ata_link *link, unsigned long deadline) | |||
50 | if (!pci_test_config_bits(pdev, &ns87410_enable_bits[ap->port_no])) | 50 | if (!pci_test_config_bits(pdev, &ns87410_enable_bits[ap->port_no])) |
51 | return -ENOENT; | 51 | return -ENOENT; |
52 | 52 | ||
53 | return ata_std_prereset(link, deadline); | 53 | return ata_sff_prereset(link, deadline); |
54 | } | ||
55 | |||
56 | /** | ||
57 | * ns87410_error_handler - probe reset | ||
58 | * @ap: ATA port | ||
59 | * | ||
60 | * Perform the ATA probe and bus reset sequence plus specific handling | ||
61 | * for this hardware. The MPIIX has the enable bits in a different place | ||
62 | * to PIIX4 and friends. As a pure PIO device it has no cable detect | ||
63 | */ | ||
64 | |||
65 | static void ns87410_error_handler(struct ata_port *ap) | ||
66 | { | ||
67 | ata_bmdma_drive_eh(ap, ns87410_pre_reset, ata_std_softreset, NULL, ata_std_postreset); | ||
68 | } | 54 | } |
69 | 55 | ||
70 | /** | 56 | /** |
@@ -119,7 +105,7 @@ static void ns87410_set_piomode(struct ata_port *ap, struct ata_device *adev) | |||
119 | } | 105 | } |
120 | 106 | ||
121 | /** | 107 | /** |
122 | * ns87410_qc_issue_prot - command issue | 108 | * ns87410_qc_issue - command issue |
123 | * @qc: command pending | 109 | * @qc: command pending |
124 | * | 110 | * |
125 | * Called when the libata layer is about to issue a command. We wrap | 111 | * Called when the libata layer is about to issue a command. We wrap |
@@ -127,7 +113,7 @@ static void ns87410_set_piomode(struct ata_port *ap, struct ata_device *adev) | |||
127 | * necessary. | 113 | * necessary. |
128 | */ | 114 | */ |
129 | 115 | ||
130 | static unsigned int ns87410_qc_issue_prot(struct ata_queued_cmd *qc) | 116 | static unsigned int ns87410_qc_issue(struct ata_queued_cmd *qc) |
131 | { | 117 | { |
132 | struct ata_port *ap = qc->ap; | 118 | struct ata_port *ap = qc->ap; |
133 | struct ata_device *adev = qc->dev; | 119 | struct ata_device *adev = qc->dev; |
@@ -140,64 +126,30 @@ static unsigned int ns87410_qc_issue_prot(struct ata_queued_cmd *qc) | |||
140 | if (adev->pio_mode && adev != ap->private_data) | 126 | if (adev->pio_mode && adev != ap->private_data) |
141 | ns87410_set_piomode(ap, adev); | 127 | ns87410_set_piomode(ap, adev); |
142 | 128 | ||
143 | return ata_qc_issue_prot(qc); | 129 | return ata_sff_qc_issue(qc); |
144 | } | 130 | } |
145 | 131 | ||
146 | static struct scsi_host_template ns87410_sht = { | 132 | static struct scsi_host_template ns87410_sht = { |
147 | .module = THIS_MODULE, | 133 | ATA_PIO_SHT(DRV_NAME), |
148 | .name = DRV_NAME, | ||
149 | .ioctl = ata_scsi_ioctl, | ||
150 | .queuecommand = ata_scsi_queuecmd, | ||
151 | .can_queue = ATA_DEF_QUEUE, | ||
152 | .this_id = ATA_SHT_THIS_ID, | ||
153 | .sg_tablesize = LIBATA_MAX_PRD, | ||
154 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
155 | .emulated = ATA_SHT_EMULATED, | ||
156 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
157 | .proc_name = DRV_NAME, | ||
158 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
159 | .slave_configure = ata_scsi_slave_config, | ||
160 | .slave_destroy = ata_scsi_slave_destroy, | ||
161 | .bios_param = ata_std_bios_param, | ||
162 | }; | 134 | }; |
163 | 135 | ||
164 | static struct ata_port_operations ns87410_port_ops = { | 136 | static struct ata_port_operations ns87410_port_ops = { |
165 | .set_piomode = ns87410_set_piomode, | 137 | .inherits = &ata_sff_port_ops, |
166 | 138 | .qc_issue = ns87410_qc_issue, | |
167 | .tf_load = ata_tf_load, | ||
168 | .tf_read = ata_tf_read, | ||
169 | .check_status = ata_check_status, | ||
170 | .exec_command = ata_exec_command, | ||
171 | .dev_select = ata_std_dev_select, | ||
172 | |||
173 | .freeze = ata_bmdma_freeze, | ||
174 | .thaw = ata_bmdma_thaw, | ||
175 | .error_handler = ns87410_error_handler, | ||
176 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
177 | .cable_detect = ata_cable_40wire, | 139 | .cable_detect = ata_cable_40wire, |
178 | 140 | .set_piomode = ns87410_set_piomode, | |
179 | .qc_prep = ata_qc_prep, | 141 | .prereset = ns87410_pre_reset, |
180 | .qc_issue = ns87410_qc_issue_prot, | ||
181 | |||
182 | .data_xfer = ata_data_xfer, | ||
183 | |||
184 | .irq_handler = ata_interrupt, | ||
185 | .irq_clear = ata_bmdma_irq_clear, | ||
186 | .irq_on = ata_irq_on, | ||
187 | |||
188 | .port_start = ata_sff_port_start, | ||
189 | }; | 142 | }; |
190 | 143 | ||
191 | static int ns87410_init_one(struct pci_dev *dev, const struct pci_device_id *id) | 144 | static int ns87410_init_one(struct pci_dev *dev, const struct pci_device_id *id) |
192 | { | 145 | { |
193 | static const struct ata_port_info info = { | 146 | static const struct ata_port_info info = { |
194 | .sht = &ns87410_sht, | ||
195 | .flags = ATA_FLAG_SLAVE_POSS, | 147 | .flags = ATA_FLAG_SLAVE_POSS, |
196 | .pio_mask = 0x0F, | 148 | .pio_mask = 0x0F, |
197 | .port_ops = &ns87410_port_ops | 149 | .port_ops = &ns87410_port_ops |
198 | }; | 150 | }; |
199 | const struct ata_port_info *ppi[] = { &info, NULL }; | 151 | const struct ata_port_info *ppi[] = { &info, NULL }; |
200 | return ata_pci_init_one(dev, ppi); | 152 | return ata_pci_sff_init_one(dev, ppi, &ns87410_sht, NULL); |
201 | } | 153 | } |
202 | 154 | ||
203 | static const struct pci_device_id ns87410[] = { | 155 | static const struct pci_device_id ns87410[] = { |
diff --git a/drivers/ata/pata_ns87415.c b/drivers/ata/pata_ns87415.c index d0e2e50823b1..ae92b0049bd5 100644 --- a/drivers/ata/pata_ns87415.c +++ b/drivers/ata/pata_ns87415.c | |||
@@ -138,7 +138,7 @@ static void ns87415_bmdma_setup(struct ata_queued_cmd *qc) | |||
138 | dmactl |= ATA_DMA_WR; | 138 | dmactl |= ATA_DMA_WR; |
139 | iowrite8(dmactl, ap->ioaddr.bmdma_addr + ATA_DMA_CMD); | 139 | iowrite8(dmactl, ap->ioaddr.bmdma_addr + ATA_DMA_CMD); |
140 | /* issue r/w command */ | 140 | /* issue r/w command */ |
141 | ap->ops->exec_command(ap, &qc->tf); | 141 | ap->ops->sff_exec_command(ap, &qc->tf); |
142 | } | 142 | } |
143 | 143 | ||
144 | /** | 144 | /** |
@@ -172,14 +172,14 @@ static void ns87415_bmdma_stop(struct ata_queued_cmd *qc) | |||
172 | } | 172 | } |
173 | 173 | ||
174 | /** | 174 | /** |
175 | * ns87415_bmdma_irq_clear - Clear interrupt | 175 | * ns87415_irq_clear - Clear interrupt |
176 | * @ap: Channel to clear | 176 | * @ap: Channel to clear |
177 | * | 177 | * |
178 | * Erratum: Due to a chip bug regisers 02 and 0A bit 1 and 2 (the | 178 | * Erratum: Due to a chip bug regisers 02 and 0A bit 1 and 2 (the |
179 | * error bits) are reset by writing to register 00 or 08. | 179 | * error bits) are reset by writing to register 00 or 08. |
180 | */ | 180 | */ |
181 | 181 | ||
182 | static void ns87415_bmdma_irq_clear(struct ata_port *ap) | 182 | static void ns87415_irq_clear(struct ata_port *ap) |
183 | { | 183 | { |
184 | void __iomem *mmio = ap->ioaddr.bmdma_addr; | 184 | void __iomem *mmio = ap->ioaddr.bmdma_addr; |
185 | 185 | ||
@@ -297,90 +297,32 @@ static u8 ns87560_bmdma_status(struct ata_port *ap) | |||
297 | { | 297 | { |
298 | return ns87560_read_buggy(ap->ioaddr.bmdma_addr + ATA_DMA_STATUS); | 298 | return ns87560_read_buggy(ap->ioaddr.bmdma_addr + ATA_DMA_STATUS); |
299 | } | 299 | } |
300 | |||
301 | static const struct ata_port_operations ns87560_pata_ops = { | ||
302 | .set_piomode = ns87415_set_piomode, | ||
303 | .mode_filter = ata_pci_default_filter, | ||
304 | |||
305 | .tf_load = ata_tf_load, | ||
306 | .tf_read = ns87560_tf_read, | ||
307 | .check_status = ns87560_check_status, | ||
308 | .check_atapi_dma = ns87415_check_atapi_dma, | ||
309 | .exec_command = ata_exec_command, | ||
310 | .dev_select = ata_std_dev_select, | ||
311 | |||
312 | .freeze = ata_bmdma_freeze, | ||
313 | .thaw = ata_bmdma_thaw, | ||
314 | .error_handler = ata_bmdma_error_handler, | ||
315 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
316 | .cable_detect = ata_cable_40wire, | ||
317 | |||
318 | .bmdma_setup = ns87415_bmdma_setup, | ||
319 | .bmdma_start = ns87415_bmdma_start, | ||
320 | .bmdma_stop = ns87415_bmdma_stop, | ||
321 | .bmdma_status = ns87560_bmdma_status, | ||
322 | .qc_prep = ata_qc_prep, | ||
323 | .qc_issue = ata_qc_issue_prot, | ||
324 | .data_xfer = ata_data_xfer, | ||
325 | |||
326 | .irq_handler = ata_interrupt, | ||
327 | .irq_clear = ns87415_bmdma_irq_clear, | ||
328 | .irq_on = ata_irq_on, | ||
329 | |||
330 | .port_start = ata_sff_port_start, | ||
331 | }; | ||
332 | |||
333 | #endif /* 87560 SuperIO Support */ | 300 | #endif /* 87560 SuperIO Support */ |
334 | 301 | ||
302 | static struct ata_port_operations ns87415_pata_ops = { | ||
303 | .inherits = &ata_bmdma_port_ops, | ||
335 | 304 | ||
336 | static const struct ata_port_operations ns87415_pata_ops = { | ||
337 | .set_piomode = ns87415_set_piomode, | ||
338 | .mode_filter = ata_pci_default_filter, | ||
339 | |||
340 | .tf_load = ata_tf_load, | ||
341 | .tf_read = ata_tf_read, | ||
342 | .check_status = ata_check_status, | ||
343 | .check_atapi_dma = ns87415_check_atapi_dma, | 305 | .check_atapi_dma = ns87415_check_atapi_dma, |
344 | .exec_command = ata_exec_command, | ||
345 | .dev_select = ata_std_dev_select, | ||
346 | |||
347 | .freeze = ata_bmdma_freeze, | ||
348 | .thaw = ata_bmdma_thaw, | ||
349 | .error_handler = ata_bmdma_error_handler, | ||
350 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
351 | .cable_detect = ata_cable_40wire, | ||
352 | |||
353 | .bmdma_setup = ns87415_bmdma_setup, | 306 | .bmdma_setup = ns87415_bmdma_setup, |
354 | .bmdma_start = ns87415_bmdma_start, | 307 | .bmdma_start = ns87415_bmdma_start, |
355 | .bmdma_stop = ns87415_bmdma_stop, | 308 | .bmdma_stop = ns87415_bmdma_stop, |
356 | .bmdma_status = ata_bmdma_status, | 309 | .sff_irq_clear = ns87415_irq_clear, |
357 | .qc_prep = ata_qc_prep, | ||
358 | .qc_issue = ata_qc_issue_prot, | ||
359 | .data_xfer = ata_data_xfer, | ||
360 | 310 | ||
361 | .irq_handler = ata_interrupt, | 311 | .cable_detect = ata_cable_40wire, |
362 | .irq_clear = ns87415_bmdma_irq_clear, | 312 | .set_piomode = ns87415_set_piomode, |
363 | .irq_on = ata_irq_on, | 313 | }; |
364 | 314 | ||
365 | .port_start = ata_sff_port_start, | 315 | #if defined(CONFIG_SUPERIO) |
316 | static struct ata_port_operations ns87560_pata_ops = { | ||
317 | .inherits = &ns87415_pata_ops, | ||
318 | .sff_tf_read = ns87560_tf_read, | ||
319 | .sff_check_status = ns87560_check_status, | ||
320 | .bmdma_status = ns87560_bmdma_status, | ||
366 | }; | 321 | }; |
322 | #endif | ||
367 | 323 | ||
368 | static struct scsi_host_template ns87415_sht = { | 324 | static struct scsi_host_template ns87415_sht = { |
369 | .module = THIS_MODULE, | 325 | ATA_BMDMA_SHT(DRV_NAME), |
370 | .name = DRV_NAME, | ||
371 | .ioctl = ata_scsi_ioctl, | ||
372 | .queuecommand = ata_scsi_queuecmd, | ||
373 | .can_queue = ATA_DEF_QUEUE, | ||
374 | .this_id = ATA_SHT_THIS_ID, | ||
375 | .sg_tablesize = LIBATA_MAX_PRD, | ||
376 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
377 | .emulated = ATA_SHT_EMULATED, | ||
378 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
379 | .proc_name = DRV_NAME, | ||
380 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
381 | .slave_configure = ata_scsi_slave_config, | ||
382 | .slave_destroy = ata_scsi_slave_destroy, | ||
383 | .bios_param = ata_std_bios_param, | ||
384 | }; | 326 | }; |
385 | 327 | ||
386 | 328 | ||
@@ -403,16 +345,15 @@ static int ns87415_init_one (struct pci_dev *pdev, const struct pci_device_id *e | |||
403 | { | 345 | { |
404 | static int printed_version; | 346 | static int printed_version; |
405 | static const struct ata_port_info info = { | 347 | static const struct ata_port_info info = { |
406 | .sht = &ns87415_sht, | ||
407 | .flags = ATA_FLAG_SLAVE_POSS, | 348 | .flags = ATA_FLAG_SLAVE_POSS, |
408 | .pio_mask = 0x1f, /* pio0-4 */ | 349 | .pio_mask = 0x1f, /* pio0-4 */ |
409 | .mwdma_mask = 0x07, /* mwdma0-2 */ | 350 | .mwdma_mask = 0x07, /* mwdma0-2 */ |
410 | .port_ops = &ns87415_pata_ops, | 351 | .port_ops = &ns87415_pata_ops, |
411 | }; | 352 | }; |
412 | const struct ata_port_info *ppi[] = { &info, NULL }; | 353 | const struct ata_port_info *ppi[] = { &info, NULL }; |
354 | int rc; | ||
413 | #if defined(CONFIG_SUPERIO) | 355 | #if defined(CONFIG_SUPERIO) |
414 | static const struct ata_port_info info87560 = { | 356 | static const struct ata_port_info info87560 = { |
415 | .sht = &ns87415_sht, | ||
416 | .flags = ATA_FLAG_SLAVE_POSS, | 357 | .flags = ATA_FLAG_SLAVE_POSS, |
417 | .pio_mask = 0x1f, /* pio0-4 */ | 358 | .pio_mask = 0x1f, /* pio0-4 */ |
418 | .mwdma_mask = 0x07, /* mwdma0-2 */ | 359 | .mwdma_mask = 0x07, /* mwdma0-2 */ |
@@ -425,11 +366,16 @@ static int ns87415_init_one (struct pci_dev *pdev, const struct pci_device_id *e | |||
425 | if (!printed_version++) | 366 | if (!printed_version++) |
426 | dev_printk(KERN_DEBUG, &pdev->dev, | 367 | dev_printk(KERN_DEBUG, &pdev->dev, |
427 | "version " DRV_VERSION "\n"); | 368 | "version " DRV_VERSION "\n"); |
369 | |||
370 | rc = pcim_enable_device(pdev); | ||
371 | if (rc) | ||
372 | return rc; | ||
373 | |||
428 | /* Select 512 byte sectors */ | 374 | /* Select 512 byte sectors */ |
429 | pci_write_config_byte(pdev, 0x55, 0xEE); | 375 | pci_write_config_byte(pdev, 0x55, 0xEE); |
430 | /* Select PIO0 8bit clocking */ | 376 | /* Select PIO0 8bit clocking */ |
431 | pci_write_config_byte(pdev, 0x54, 0xB7); | 377 | pci_write_config_byte(pdev, 0x54, 0xB7); |
432 | return ata_pci_init_one(pdev, ppi); | 378 | return ata_pci_sff_init_one(pdev, ppi, &ns87415_sht, NULL); |
433 | } | 379 | } |
434 | 380 | ||
435 | static const struct pci_device_id ns87415_pci_tbl[] = { | 381 | static const struct pci_device_id ns87415_pci_tbl[] = { |
diff --git a/drivers/ata/pata_oldpiix.c b/drivers/ata/pata_oldpiix.c index 44da09ace52c..e678af383d13 100644 --- a/drivers/ata/pata_oldpiix.c +++ b/drivers/ata/pata_oldpiix.c | |||
@@ -47,21 +47,7 @@ static int oldpiix_pre_reset(struct ata_link *link, unsigned long deadline) | |||
47 | if (!pci_test_config_bits(pdev, &oldpiix_enable_bits[ap->port_no])) | 47 | if (!pci_test_config_bits(pdev, &oldpiix_enable_bits[ap->port_no])) |
48 | return -ENOENT; | 48 | return -ENOENT; |
49 | 49 | ||
50 | return ata_std_prereset(link, deadline); | 50 | return ata_sff_prereset(link, deadline); |
51 | } | ||
52 | |||
53 | /** | ||
54 | * oldpiix_pata_error_handler - Probe specified port on PATA host controller | ||
55 | * @ap: Port to probe | ||
56 | * @classes: | ||
57 | * | ||
58 | * LOCKING: | ||
59 | * None (inherited from caller). | ||
60 | */ | ||
61 | |||
62 | static void oldpiix_pata_error_handler(struct ata_port *ap) | ||
63 | { | ||
64 | ata_bmdma_drive_eh(ap, oldpiix_pre_reset, ata_std_softreset, NULL, ata_std_postreset); | ||
65 | } | 51 | } |
66 | 52 | ||
67 | /** | 53 | /** |
@@ -195,7 +181,7 @@ static void oldpiix_set_dmamode (struct ata_port *ap, struct ata_device *adev) | |||
195 | } | 181 | } |
196 | 182 | ||
197 | /** | 183 | /** |
198 | * oldpiix_qc_issue_prot - command issue | 184 | * oldpiix_qc_issue - command issue |
199 | * @qc: command pending | 185 | * @qc: command pending |
200 | * | 186 | * |
201 | * Called when the libata layer is about to issue a command. We wrap | 187 | * Called when the libata layer is about to issue a command. We wrap |
@@ -205,7 +191,7 @@ static void oldpiix_set_dmamode (struct ata_port *ap, struct ata_device *adev) | |||
205 | * be made PIO0. | 191 | * be made PIO0. |
206 | */ | 192 | */ |
207 | 193 | ||
208 | static unsigned int oldpiix_qc_issue_prot(struct ata_queued_cmd *qc) | 194 | static unsigned int oldpiix_qc_issue(struct ata_queued_cmd *qc) |
209 | { | 195 | { |
210 | struct ata_port *ap = qc->ap; | 196 | struct ata_port *ap = qc->ap; |
211 | struct ata_device *adev = qc->dev; | 197 | struct ata_device *adev = qc->dev; |
@@ -215,58 +201,21 @@ static unsigned int oldpiix_qc_issue_prot(struct ata_queued_cmd *qc) | |||
215 | if (adev->dma_mode) | 201 | if (adev->dma_mode) |
216 | oldpiix_set_dmamode(ap, adev); | 202 | oldpiix_set_dmamode(ap, adev); |
217 | } | 203 | } |
218 | return ata_qc_issue_prot(qc); | 204 | return ata_sff_qc_issue(qc); |
219 | } | 205 | } |
220 | 206 | ||
221 | 207 | ||
222 | static struct scsi_host_template oldpiix_sht = { | 208 | static struct scsi_host_template oldpiix_sht = { |
223 | .module = THIS_MODULE, | 209 | ATA_BMDMA_SHT(DRV_NAME), |
224 | .name = DRV_NAME, | ||
225 | .ioctl = ata_scsi_ioctl, | ||
226 | .queuecommand = ata_scsi_queuecmd, | ||
227 | .can_queue = ATA_DEF_QUEUE, | ||
228 | .this_id = ATA_SHT_THIS_ID, | ||
229 | .sg_tablesize = LIBATA_MAX_PRD, | ||
230 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
231 | .emulated = ATA_SHT_EMULATED, | ||
232 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
233 | .proc_name = DRV_NAME, | ||
234 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
235 | .slave_configure = ata_scsi_slave_config, | ||
236 | .slave_destroy = ata_scsi_slave_destroy, | ||
237 | .bios_param = ata_std_bios_param, | ||
238 | }; | 210 | }; |
239 | 211 | ||
240 | static const struct ata_port_operations oldpiix_pata_ops = { | 212 | static struct ata_port_operations oldpiix_pata_ops = { |
213 | .inherits = &ata_bmdma_port_ops, | ||
214 | .qc_issue = oldpiix_qc_issue, | ||
215 | .cable_detect = ata_cable_40wire, | ||
241 | .set_piomode = oldpiix_set_piomode, | 216 | .set_piomode = oldpiix_set_piomode, |
242 | .set_dmamode = oldpiix_set_dmamode, | 217 | .set_dmamode = oldpiix_set_dmamode, |
243 | .mode_filter = ata_pci_default_filter, | 218 | .prereset = oldpiix_pre_reset, |
244 | |||
245 | .tf_load = ata_tf_load, | ||
246 | .tf_read = ata_tf_read, | ||
247 | .check_status = ata_check_status, | ||
248 | .exec_command = ata_exec_command, | ||
249 | .dev_select = ata_std_dev_select, | ||
250 | |||
251 | .freeze = ata_bmdma_freeze, | ||
252 | .thaw = ata_bmdma_thaw, | ||
253 | .error_handler = oldpiix_pata_error_handler, | ||
254 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
255 | .cable_detect = ata_cable_40wire, | ||
256 | |||
257 | .bmdma_setup = ata_bmdma_setup, | ||
258 | .bmdma_start = ata_bmdma_start, | ||
259 | .bmdma_stop = ata_bmdma_stop, | ||
260 | .bmdma_status = ata_bmdma_status, | ||
261 | .qc_prep = ata_qc_prep, | ||
262 | .qc_issue = oldpiix_qc_issue_prot, | ||
263 | .data_xfer = ata_data_xfer, | ||
264 | |||
265 | .irq_handler = ata_interrupt, | ||
266 | .irq_clear = ata_bmdma_irq_clear, | ||
267 | .irq_on = ata_irq_on, | ||
268 | |||
269 | .port_start = ata_sff_port_start, | ||
270 | }; | 219 | }; |
271 | 220 | ||
272 | 221 | ||
@@ -289,7 +238,6 @@ static int oldpiix_init_one (struct pci_dev *pdev, const struct pci_device_id *e | |||
289 | { | 238 | { |
290 | static int printed_version; | 239 | static int printed_version; |
291 | static const struct ata_port_info info = { | 240 | static const struct ata_port_info info = { |
292 | .sht = &oldpiix_sht, | ||
293 | .flags = ATA_FLAG_SLAVE_POSS, | 241 | .flags = ATA_FLAG_SLAVE_POSS, |
294 | .pio_mask = 0x1f, /* pio0-4 */ | 242 | .pio_mask = 0x1f, /* pio0-4 */ |
295 | .mwdma_mask = 0x07, /* mwdma1-2 */ | 243 | .mwdma_mask = 0x07, /* mwdma1-2 */ |
@@ -301,7 +249,7 @@ static int oldpiix_init_one (struct pci_dev *pdev, const struct pci_device_id *e | |||
301 | dev_printk(KERN_DEBUG, &pdev->dev, | 249 | dev_printk(KERN_DEBUG, &pdev->dev, |
302 | "version " DRV_VERSION "\n"); | 250 | "version " DRV_VERSION "\n"); |
303 | 251 | ||
304 | return ata_pci_init_one(pdev, ppi); | 252 | return ata_pci_sff_init_one(pdev, ppi, &oldpiix_sht, NULL); |
305 | } | 253 | } |
306 | 254 | ||
307 | static const struct pci_device_id oldpiix_pci_tbl[] = { | 255 | static const struct pci_device_id oldpiix_pci_tbl[] = { |
diff --git a/drivers/ata/pata_opti.c b/drivers/ata/pata_opti.c index 8f79447b6151..fb2cf661b0e8 100644 --- a/drivers/ata/pata_opti.c +++ b/drivers/ata/pata_opti.c | |||
@@ -64,22 +64,7 @@ static int opti_pre_reset(struct ata_link *link, unsigned long deadline) | |||
64 | if (!pci_test_config_bits(pdev, &opti_enable_bits[ap->port_no])) | 64 | if (!pci_test_config_bits(pdev, &opti_enable_bits[ap->port_no])) |
65 | return -ENOENT; | 65 | return -ENOENT; |
66 | 66 | ||
67 | return ata_std_prereset(link, deadline); | 67 | return ata_sff_prereset(link, deadline); |
68 | } | ||
69 | |||
70 | /** | ||
71 | * opti_probe_reset - probe reset | ||
72 | * @ap: ATA port | ||
73 | * | ||
74 | * Perform the ATA probe and bus reset sequence plus specific handling | ||
75 | * for this hardware. The Opti needs little handling - we have no UDMA66 | ||
76 | * capability that needs cable detection. All we must do is check the port | ||
77 | * is enabled. | ||
78 | */ | ||
79 | |||
80 | static void opti_error_handler(struct ata_port *ap) | ||
81 | { | ||
82 | ata_bmdma_drive_eh(ap, opti_pre_reset, ata_std_softreset, NULL, ata_std_postreset); | ||
83 | } | 68 | } |
84 | 69 | ||
85 | /** | 70 | /** |
@@ -165,58 +150,19 @@ static void opti_set_piomode(struct ata_port *ap, struct ata_device *adev) | |||
165 | } | 150 | } |
166 | 151 | ||
167 | static struct scsi_host_template opti_sht = { | 152 | static struct scsi_host_template opti_sht = { |
168 | .module = THIS_MODULE, | 153 | ATA_PIO_SHT(DRV_NAME), |
169 | .name = DRV_NAME, | ||
170 | .ioctl = ata_scsi_ioctl, | ||
171 | .queuecommand = ata_scsi_queuecmd, | ||
172 | .can_queue = ATA_DEF_QUEUE, | ||
173 | .this_id = ATA_SHT_THIS_ID, | ||
174 | .sg_tablesize = LIBATA_MAX_PRD, | ||
175 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
176 | .emulated = ATA_SHT_EMULATED, | ||
177 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
178 | .proc_name = DRV_NAME, | ||
179 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
180 | .slave_configure = ata_scsi_slave_config, | ||
181 | .slave_destroy = ata_scsi_slave_destroy, | ||
182 | .bios_param = ata_std_bios_param, | ||
183 | }; | 154 | }; |
184 | 155 | ||
185 | static struct ata_port_operations opti_port_ops = { | 156 | static struct ata_port_operations opti_port_ops = { |
186 | .set_piomode = opti_set_piomode, | 157 | .inherits = &ata_sff_port_ops, |
187 | .tf_load = ata_tf_load, | ||
188 | .tf_read = ata_tf_read, | ||
189 | .check_status = ata_check_status, | ||
190 | .exec_command = ata_exec_command, | ||
191 | .dev_select = ata_std_dev_select, | ||
192 | |||
193 | .freeze = ata_bmdma_freeze, | ||
194 | .thaw = ata_bmdma_thaw, | ||
195 | .error_handler = opti_error_handler, | ||
196 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
197 | .cable_detect = ata_cable_40wire, | 158 | .cable_detect = ata_cable_40wire, |
198 | 159 | .set_piomode = opti_set_piomode, | |
199 | .bmdma_setup = ata_bmdma_setup, | 160 | .prereset = opti_pre_reset, |
200 | .bmdma_start = ata_bmdma_start, | ||
201 | .bmdma_stop = ata_bmdma_stop, | ||
202 | .bmdma_status = ata_bmdma_status, | ||
203 | |||
204 | .qc_prep = ata_qc_prep, | ||
205 | .qc_issue = ata_qc_issue_prot, | ||
206 | |||
207 | .data_xfer = ata_data_xfer, | ||
208 | |||
209 | .irq_handler = ata_interrupt, | ||
210 | .irq_clear = ata_bmdma_irq_clear, | ||
211 | .irq_on = ata_irq_on, | ||
212 | |||
213 | .port_start = ata_sff_port_start, | ||
214 | }; | 161 | }; |
215 | 162 | ||
216 | static int opti_init_one(struct pci_dev *dev, const struct pci_device_id *id) | 163 | static int opti_init_one(struct pci_dev *dev, const struct pci_device_id *id) |
217 | { | 164 | { |
218 | static const struct ata_port_info info = { | 165 | static const struct ata_port_info info = { |
219 | .sht = &opti_sht, | ||
220 | .flags = ATA_FLAG_SLAVE_POSS, | 166 | .flags = ATA_FLAG_SLAVE_POSS, |
221 | .pio_mask = 0x1f, | 167 | .pio_mask = 0x1f, |
222 | .port_ops = &opti_port_ops | 168 | .port_ops = &opti_port_ops |
@@ -227,7 +173,7 @@ static int opti_init_one(struct pci_dev *dev, const struct pci_device_id *id) | |||
227 | if (!printed_version++) | 173 | if (!printed_version++) |
228 | dev_printk(KERN_DEBUG, &dev->dev, "version " DRV_VERSION "\n"); | 174 | dev_printk(KERN_DEBUG, &dev->dev, "version " DRV_VERSION "\n"); |
229 | 175 | ||
230 | return ata_pci_init_one(dev, ppi); | 176 | return ata_pci_sff_init_one(dev, ppi, &opti_sht, NULL); |
231 | } | 177 | } |
232 | 178 | ||
233 | static const struct pci_device_id opti[] = { | 179 | static const struct pci_device_id opti[] = { |
diff --git a/drivers/ata/pata_optidma.c b/drivers/ata/pata_optidma.c index f9b485a487ae..4cd744456313 100644 --- a/drivers/ata/pata_optidma.c +++ b/drivers/ata/pata_optidma.c | |||
@@ -64,22 +64,7 @@ static int optidma_pre_reset(struct ata_link *link, unsigned long deadline) | |||
64 | if (ap->port_no && !pci_test_config_bits(pdev, &optidma_enable_bits)) | 64 | if (ap->port_no && !pci_test_config_bits(pdev, &optidma_enable_bits)) |
65 | return -ENOENT; | 65 | return -ENOENT; |
66 | 66 | ||
67 | return ata_std_prereset(link, deadline); | 67 | return ata_sff_prereset(link, deadline); |
68 | } | ||
69 | |||
70 | /** | ||
71 | * optidma_probe_reset - probe reset | ||
72 | * @ap: ATA port | ||
73 | * | ||
74 | * Perform the ATA probe and bus reset sequence plus specific handling | ||
75 | * for this hardware. The Opti needs little handling - we have no UDMA66 | ||
76 | * capability that needs cable detection. All we must do is check the port | ||
77 | * is enabled. | ||
78 | */ | ||
79 | |||
80 | static void optidma_error_handler(struct ata_port *ap) | ||
81 | { | ||
82 | ata_bmdma_drive_eh(ap, optidma_pre_reset, ata_std_softreset, NULL, ata_std_postreset); | ||
83 | } | 68 | } |
84 | 69 | ||
85 | /** | 70 | /** |
@@ -350,89 +335,22 @@ static int optidma_set_mode(struct ata_link *link, struct ata_device **r_failed) | |||
350 | } | 335 | } |
351 | 336 | ||
352 | static struct scsi_host_template optidma_sht = { | 337 | static struct scsi_host_template optidma_sht = { |
353 | .module = THIS_MODULE, | 338 | ATA_BMDMA_SHT(DRV_NAME), |
354 | .name = DRV_NAME, | ||
355 | .ioctl = ata_scsi_ioctl, | ||
356 | .queuecommand = ata_scsi_queuecmd, | ||
357 | .can_queue = ATA_DEF_QUEUE, | ||
358 | .this_id = ATA_SHT_THIS_ID, | ||
359 | .sg_tablesize = LIBATA_MAX_PRD, | ||
360 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
361 | .emulated = ATA_SHT_EMULATED, | ||
362 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
363 | .proc_name = DRV_NAME, | ||
364 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
365 | .slave_configure = ata_scsi_slave_config, | ||
366 | .slave_destroy = ata_scsi_slave_destroy, | ||
367 | .bios_param = ata_std_bios_param, | ||
368 | }; | 339 | }; |
369 | 340 | ||
370 | static struct ata_port_operations optidma_port_ops = { | 341 | static struct ata_port_operations optidma_port_ops = { |
342 | .inherits = &ata_bmdma_port_ops, | ||
343 | .cable_detect = ata_cable_40wire, | ||
371 | .set_piomode = optidma_set_pio_mode, | 344 | .set_piomode = optidma_set_pio_mode, |
372 | .set_dmamode = optidma_set_dma_mode, | 345 | .set_dmamode = optidma_set_dma_mode, |
373 | |||
374 | .tf_load = ata_tf_load, | ||
375 | .tf_read = ata_tf_read, | ||
376 | .check_status = ata_check_status, | ||
377 | .exec_command = ata_exec_command, | ||
378 | .dev_select = ata_std_dev_select, | ||
379 | |||
380 | .freeze = ata_bmdma_freeze, | ||
381 | .thaw = ata_bmdma_thaw, | ||
382 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
383 | .error_handler = optidma_error_handler, | ||
384 | .set_mode = optidma_set_mode, | 346 | .set_mode = optidma_set_mode, |
385 | .cable_detect = ata_cable_40wire, | 347 | .prereset = optidma_pre_reset, |
386 | |||
387 | .bmdma_setup = ata_bmdma_setup, | ||
388 | .bmdma_start = ata_bmdma_start, | ||
389 | .bmdma_stop = ata_bmdma_stop, | ||
390 | .bmdma_status = ata_bmdma_status, | ||
391 | |||
392 | .qc_prep = ata_qc_prep, | ||
393 | .qc_issue = ata_qc_issue_prot, | ||
394 | |||
395 | .data_xfer = ata_data_xfer, | ||
396 | |||
397 | .irq_handler = ata_interrupt, | ||
398 | .irq_clear = ata_bmdma_irq_clear, | ||
399 | .irq_on = ata_irq_on, | ||
400 | |||
401 | .port_start = ata_sff_port_start, | ||
402 | }; | 348 | }; |
403 | 349 | ||
404 | static struct ata_port_operations optiplus_port_ops = { | 350 | static struct ata_port_operations optiplus_port_ops = { |
351 | .inherits = &optidma_port_ops, | ||
405 | .set_piomode = optiplus_set_pio_mode, | 352 | .set_piomode = optiplus_set_pio_mode, |
406 | .set_dmamode = optiplus_set_dma_mode, | 353 | .set_dmamode = optiplus_set_dma_mode, |
407 | |||
408 | .tf_load = ata_tf_load, | ||
409 | .tf_read = ata_tf_read, | ||
410 | .check_status = ata_check_status, | ||
411 | .exec_command = ata_exec_command, | ||
412 | .dev_select = ata_std_dev_select, | ||
413 | |||
414 | .freeze = ata_bmdma_freeze, | ||
415 | .thaw = ata_bmdma_thaw, | ||
416 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
417 | .error_handler = optidma_error_handler, | ||
418 | .set_mode = optidma_set_mode, | ||
419 | .cable_detect = ata_cable_40wire, | ||
420 | |||
421 | .bmdma_setup = ata_bmdma_setup, | ||
422 | .bmdma_start = ata_bmdma_start, | ||
423 | .bmdma_stop = ata_bmdma_stop, | ||
424 | .bmdma_status = ata_bmdma_status, | ||
425 | |||
426 | .qc_prep = ata_qc_prep, | ||
427 | .qc_issue = ata_qc_issue_prot, | ||
428 | |||
429 | .data_xfer = ata_data_xfer, | ||
430 | |||
431 | .irq_handler = ata_interrupt, | ||
432 | .irq_clear = ata_bmdma_irq_clear, | ||
433 | .irq_on = ata_irq_on, | ||
434 | |||
435 | .port_start = ata_sff_port_start, | ||
436 | }; | 354 | }; |
437 | 355 | ||
438 | /** | 356 | /** |
@@ -481,14 +399,12 @@ done_nomsg: /* Wrong chip revision */ | |||
481 | static int optidma_init_one(struct pci_dev *dev, const struct pci_device_id *id) | 399 | static int optidma_init_one(struct pci_dev *dev, const struct pci_device_id *id) |
482 | { | 400 | { |
483 | static const struct ata_port_info info_82c700 = { | 401 | static const struct ata_port_info info_82c700 = { |
484 | .sht = &optidma_sht, | ||
485 | .flags = ATA_FLAG_SLAVE_POSS, | 402 | .flags = ATA_FLAG_SLAVE_POSS, |
486 | .pio_mask = 0x1f, | 403 | .pio_mask = 0x1f, |
487 | .mwdma_mask = 0x07, | 404 | .mwdma_mask = 0x07, |
488 | .port_ops = &optidma_port_ops | 405 | .port_ops = &optidma_port_ops |
489 | }; | 406 | }; |
490 | static const struct ata_port_info info_82c700_udma = { | 407 | static const struct ata_port_info info_82c700_udma = { |
491 | .sht = &optidma_sht, | ||
492 | .flags = ATA_FLAG_SLAVE_POSS, | 408 | .flags = ATA_FLAG_SLAVE_POSS, |
493 | .pio_mask = 0x1f, | 409 | .pio_mask = 0x1f, |
494 | .mwdma_mask = 0x07, | 410 | .mwdma_mask = 0x07, |
@@ -497,10 +413,15 @@ static int optidma_init_one(struct pci_dev *dev, const struct pci_device_id *id) | |||
497 | }; | 413 | }; |
498 | const struct ata_port_info *ppi[] = { &info_82c700, NULL }; | 414 | const struct ata_port_info *ppi[] = { &info_82c700, NULL }; |
499 | static int printed_version; | 415 | static int printed_version; |
416 | int rc; | ||
500 | 417 | ||
501 | if (!printed_version++) | 418 | if (!printed_version++) |
502 | dev_printk(KERN_DEBUG, &dev->dev, "version " DRV_VERSION "\n"); | 419 | dev_printk(KERN_DEBUG, &dev->dev, "version " DRV_VERSION "\n"); |
503 | 420 | ||
421 | rc = pcim_enable_device(dev); | ||
422 | if (rc) | ||
423 | return rc; | ||
424 | |||
504 | /* Fixed location chipset magic */ | 425 | /* Fixed location chipset magic */ |
505 | inw(0x1F1); | 426 | inw(0x1F1); |
506 | inw(0x1F1); | 427 | inw(0x1F1); |
@@ -509,7 +430,7 @@ static int optidma_init_one(struct pci_dev *dev, const struct pci_device_id *id) | |||
509 | if (optiplus_with_udma(dev)) | 430 | if (optiplus_with_udma(dev)) |
510 | ppi[0] = &info_82c700_udma; | 431 | ppi[0] = &info_82c700_udma; |
511 | 432 | ||
512 | return ata_pci_init_one(dev, ppi); | 433 | return ata_pci_sff_init_one(dev, ppi, &optidma_sht, NULL); |
513 | } | 434 | } |
514 | 435 | ||
515 | static const struct pci_device_id optidma[] = { | 436 | static const struct pci_device_id optidma[] = { |
diff --git a/drivers/ata/pata_pcmcia.c b/drivers/ata/pata_pcmcia.c index 3e7f6a9da28b..3d39f9dfec5a 100644 --- a/drivers/ata/pata_pcmcia.c +++ b/drivers/ata/pata_pcmcia.c | |||
@@ -128,71 +128,21 @@ static unsigned int ata_data_xfer_8bit(struct ata_device *dev, | |||
128 | 128 | ||
129 | 129 | ||
130 | static struct scsi_host_template pcmcia_sht = { | 130 | static struct scsi_host_template pcmcia_sht = { |
131 | .module = THIS_MODULE, | 131 | ATA_PIO_SHT(DRV_NAME), |
132 | .name = DRV_NAME, | ||
133 | .ioctl = ata_scsi_ioctl, | ||
134 | .queuecommand = ata_scsi_queuecmd, | ||
135 | .can_queue = ATA_DEF_QUEUE, | ||
136 | .this_id = ATA_SHT_THIS_ID, | ||
137 | .sg_tablesize = LIBATA_MAX_PRD, | ||
138 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
139 | .emulated = ATA_SHT_EMULATED, | ||
140 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
141 | .proc_name = DRV_NAME, | ||
142 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
143 | .slave_configure = ata_scsi_slave_config, | ||
144 | .slave_destroy = ata_scsi_slave_destroy, | ||
145 | .bios_param = ata_std_bios_param, | ||
146 | }; | 132 | }; |
147 | 133 | ||
148 | static struct ata_port_operations pcmcia_port_ops = { | 134 | static struct ata_port_operations pcmcia_port_ops = { |
149 | .set_mode = pcmcia_set_mode, | 135 | .inherits = &ata_sff_port_ops, |
150 | .tf_load = ata_tf_load, | 136 | .sff_data_xfer = ata_sff_data_xfer_noirq, |
151 | .tf_read = ata_tf_read, | ||
152 | .check_status = ata_check_status, | ||
153 | .exec_command = ata_exec_command, | ||
154 | .dev_select = ata_std_dev_select, | ||
155 | |||
156 | .freeze = ata_bmdma_freeze, | ||
157 | .thaw = ata_bmdma_thaw, | ||
158 | .error_handler = ata_bmdma_error_handler, | ||
159 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
160 | .cable_detect = ata_cable_40wire, | 137 | .cable_detect = ata_cable_40wire, |
161 | 138 | .set_mode = pcmcia_set_mode, | |
162 | .qc_prep = ata_qc_prep, | ||
163 | .qc_issue = ata_qc_issue_prot, | ||
164 | |||
165 | .data_xfer = ata_data_xfer_noirq, | ||
166 | |||
167 | .irq_clear = ata_bmdma_irq_clear, | ||
168 | .irq_on = ata_irq_on, | ||
169 | |||
170 | .port_start = ata_sff_port_start, | ||
171 | }; | 139 | }; |
172 | 140 | ||
173 | static struct ata_port_operations pcmcia_8bit_port_ops = { | 141 | static struct ata_port_operations pcmcia_8bit_port_ops = { |
174 | .set_mode = pcmcia_set_mode_8bit, | 142 | .inherits = &ata_sff_port_ops, |
175 | .tf_load = ata_tf_load, | 143 | .sff_data_xfer = ata_data_xfer_8bit, |
176 | .tf_read = ata_tf_read, | ||
177 | .check_status = ata_check_status, | ||
178 | .exec_command = ata_exec_command, | ||
179 | .dev_select = ata_std_dev_select, | ||
180 | |||
181 | .freeze = ata_bmdma_freeze, | ||
182 | .thaw = ata_bmdma_thaw, | ||
183 | .error_handler = ata_bmdma_error_handler, | ||
184 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
185 | .cable_detect = ata_cable_40wire, | 144 | .cable_detect = ata_cable_40wire, |
186 | 145 | .set_mode = pcmcia_set_mode_8bit, | |
187 | .qc_prep = ata_qc_prep, | ||
188 | .qc_issue = ata_qc_issue_prot, | ||
189 | |||
190 | .data_xfer = ata_data_xfer_8bit, | ||
191 | |||
192 | .irq_clear = ata_bmdma_irq_clear, | ||
193 | .irq_on = ata_irq_on, | ||
194 | |||
195 | .port_start = ata_sff_port_start, | ||
196 | }; | 146 | }; |
197 | 147 | ||
198 | #define CS_CHECK(fn, ret) \ | 148 | #define CS_CHECK(fn, ret) \ |
@@ -373,13 +323,13 @@ next_entry: | |||
373 | ap->ioaddr.cmd_addr = io_addr + 0x10 * p; | 323 | ap->ioaddr.cmd_addr = io_addr + 0x10 * p; |
374 | ap->ioaddr.altstatus_addr = ctl_addr + 0x10 * p; | 324 | ap->ioaddr.altstatus_addr = ctl_addr + 0x10 * p; |
375 | ap->ioaddr.ctl_addr = ctl_addr + 0x10 * p; | 325 | ap->ioaddr.ctl_addr = ctl_addr + 0x10 * p; |
376 | ata_std_ports(&ap->ioaddr); | 326 | ata_sff_std_ports(&ap->ioaddr); |
377 | 327 | ||
378 | ata_port_desc(ap, "cmd 0x%lx ctl 0x%lx", io_base, ctl_base); | 328 | ata_port_desc(ap, "cmd 0x%lx ctl 0x%lx", io_base, ctl_base); |
379 | } | 329 | } |
380 | 330 | ||
381 | /* activate */ | 331 | /* activate */ |
382 | ret = ata_host_activate(host, pdev->irq.AssignedIRQ, ata_interrupt, | 332 | ret = ata_host_activate(host, pdev->irq.AssignedIRQ, ata_sff_interrupt, |
383 | IRQF_SHARED, &pcmcia_sht); | 333 | IRQF_SHARED, &pcmcia_sht); |
384 | if (ret) | 334 | if (ret) |
385 | goto failed; | 335 | goto failed; |
diff --git a/drivers/ata/pata_pdc2027x.c b/drivers/ata/pata_pdc2027x.c index 511c89b9bae8..0e1c2c1134d3 100644 --- a/drivers/ata/pata_pdc2027x.c +++ b/drivers/ata/pata_pdc2027x.c | |||
@@ -63,7 +63,7 @@ enum { | |||
63 | }; | 63 | }; |
64 | 64 | ||
65 | static int pdc2027x_init_one(struct pci_dev *pdev, const struct pci_device_id *ent); | 65 | static int pdc2027x_init_one(struct pci_dev *pdev, const struct pci_device_id *ent); |
66 | static void pdc2027x_error_handler(struct ata_port *ap); | 66 | static int pdc2027x_prereset(struct ata_link *link, unsigned long deadline); |
67 | static void pdc2027x_set_piomode(struct ata_port *ap, struct ata_device *adev); | 67 | static void pdc2027x_set_piomode(struct ata_port *ap, struct ata_device *adev); |
68 | static void pdc2027x_set_dmamode(struct ata_port *ap, struct ata_device *adev); | 68 | static void pdc2027x_set_dmamode(struct ata_port *ap, struct ata_device *adev); |
69 | static int pdc2027x_check_atapi_dma(struct ata_queued_cmd *qc); | 69 | static int pdc2027x_check_atapi_dma(struct ata_queued_cmd *qc); |
@@ -129,84 +129,22 @@ static struct pci_driver pdc2027x_pci_driver = { | |||
129 | }; | 129 | }; |
130 | 130 | ||
131 | static struct scsi_host_template pdc2027x_sht = { | 131 | static struct scsi_host_template pdc2027x_sht = { |
132 | .module = THIS_MODULE, | 132 | ATA_BMDMA_SHT(DRV_NAME), |
133 | .name = DRV_NAME, | ||
134 | .ioctl = ata_scsi_ioctl, | ||
135 | .queuecommand = ata_scsi_queuecmd, | ||
136 | .can_queue = ATA_DEF_QUEUE, | ||
137 | .this_id = ATA_SHT_THIS_ID, | ||
138 | .sg_tablesize = LIBATA_MAX_PRD, | ||
139 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
140 | .emulated = ATA_SHT_EMULATED, | ||
141 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
142 | .proc_name = DRV_NAME, | ||
143 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
144 | .slave_configure = ata_scsi_slave_config, | ||
145 | .slave_destroy = ata_scsi_slave_destroy, | ||
146 | .bios_param = ata_std_bios_param, | ||
147 | }; | 133 | }; |
148 | 134 | ||
149 | static struct ata_port_operations pdc2027x_pata100_ops = { | 135 | static struct ata_port_operations pdc2027x_pata100_ops = { |
150 | .mode_filter = ata_pci_default_filter, | 136 | .inherits = &ata_bmdma_port_ops, |
151 | |||
152 | .tf_load = ata_tf_load, | ||
153 | .tf_read = ata_tf_read, | ||
154 | .check_status = ata_check_status, | ||
155 | .exec_command = ata_exec_command, | ||
156 | .dev_select = ata_std_dev_select, | ||
157 | |||
158 | .check_atapi_dma = pdc2027x_check_atapi_dma, | 137 | .check_atapi_dma = pdc2027x_check_atapi_dma, |
159 | .bmdma_setup = ata_bmdma_setup, | ||
160 | .bmdma_start = ata_bmdma_start, | ||
161 | .bmdma_stop = ata_bmdma_stop, | ||
162 | .bmdma_status = ata_bmdma_status, | ||
163 | .qc_prep = ata_qc_prep, | ||
164 | .qc_issue = ata_qc_issue_prot, | ||
165 | .data_xfer = ata_data_xfer, | ||
166 | |||
167 | .freeze = ata_bmdma_freeze, | ||
168 | .thaw = ata_bmdma_thaw, | ||
169 | .error_handler = pdc2027x_error_handler, | ||
170 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
171 | .cable_detect = pdc2027x_cable_detect, | 138 | .cable_detect = pdc2027x_cable_detect, |
172 | 139 | .prereset = pdc2027x_prereset, | |
173 | .irq_clear = ata_bmdma_irq_clear, | ||
174 | .irq_on = ata_irq_on, | ||
175 | |||
176 | .port_start = ata_sff_port_start, | ||
177 | }; | 140 | }; |
178 | 141 | ||
179 | static struct ata_port_operations pdc2027x_pata133_ops = { | 142 | static struct ata_port_operations pdc2027x_pata133_ops = { |
143 | .inherits = &pdc2027x_pata100_ops, | ||
144 | .mode_filter = pdc2027x_mode_filter, | ||
180 | .set_piomode = pdc2027x_set_piomode, | 145 | .set_piomode = pdc2027x_set_piomode, |
181 | .set_dmamode = pdc2027x_set_dmamode, | 146 | .set_dmamode = pdc2027x_set_dmamode, |
182 | .set_mode = pdc2027x_set_mode, | 147 | .set_mode = pdc2027x_set_mode, |
183 | .mode_filter = pdc2027x_mode_filter, | ||
184 | |||
185 | .tf_load = ata_tf_load, | ||
186 | .tf_read = ata_tf_read, | ||
187 | .check_status = ata_check_status, | ||
188 | .exec_command = ata_exec_command, | ||
189 | .dev_select = ata_std_dev_select, | ||
190 | |||
191 | .check_atapi_dma = pdc2027x_check_atapi_dma, | ||
192 | .bmdma_setup = ata_bmdma_setup, | ||
193 | .bmdma_start = ata_bmdma_start, | ||
194 | .bmdma_stop = ata_bmdma_stop, | ||
195 | .bmdma_status = ata_bmdma_status, | ||
196 | .qc_prep = ata_qc_prep, | ||
197 | .qc_issue = ata_qc_issue_prot, | ||
198 | .data_xfer = ata_data_xfer, | ||
199 | |||
200 | .freeze = ata_bmdma_freeze, | ||
201 | .thaw = ata_bmdma_thaw, | ||
202 | .error_handler = pdc2027x_error_handler, | ||
203 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
204 | .cable_detect = pdc2027x_cable_detect, | ||
205 | |||
206 | .irq_clear = ata_bmdma_irq_clear, | ||
207 | .irq_on = ata_irq_on, | ||
208 | |||
209 | .port_start = ata_sff_port_start, | ||
210 | }; | 148 | }; |
211 | 149 | ||
212 | static struct ata_port_info pdc2027x_port_info[] = { | 150 | static struct ata_port_info pdc2027x_port_info[] = { |
@@ -310,22 +248,7 @@ static int pdc2027x_prereset(struct ata_link *link, unsigned long deadline) | |||
310 | /* Check whether port enabled */ | 248 | /* Check whether port enabled */ |
311 | if (!pdc2027x_port_enabled(link->ap)) | 249 | if (!pdc2027x_port_enabled(link->ap)) |
312 | return -ENOENT; | 250 | return -ENOENT; |
313 | return ata_std_prereset(link, deadline); | 251 | return ata_sff_prereset(link, deadline); |
314 | } | ||
315 | |||
316 | /** | ||
317 | * pdc2027x_error_handler - Perform reset on PATA port and classify | ||
318 | * @ap: Port to reset | ||
319 | * | ||
320 | * Reset PATA phy and classify attached devices. | ||
321 | * | ||
322 | * LOCKING: | ||
323 | * None (inherited from caller). | ||
324 | */ | ||
325 | |||
326 | static void pdc2027x_error_handler(struct ata_port *ap) | ||
327 | { | ||
328 | ata_bmdma_drive_eh(ap, pdc2027x_prereset, ata_std_softreset, NULL, ata_std_postreset); | ||
329 | } | 252 | } |
330 | 253 | ||
331 | /** | 254 | /** |
@@ -342,7 +265,7 @@ static unsigned long pdc2027x_mode_filter(struct ata_device *adev, unsigned long | |||
342 | struct ata_device *pair = ata_dev_pair(adev); | 265 | struct ata_device *pair = ata_dev_pair(adev); |
343 | 266 | ||
344 | if (adev->class != ATA_DEV_ATA || adev->devno == 0 || pair == NULL) | 267 | if (adev->class != ATA_DEV_ATA || adev->devno == 0 || pair == NULL) |
345 | return ata_pci_default_filter(adev, mask); | 268 | return ata_bmdma_mode_filter(adev, mask); |
346 | 269 | ||
347 | /* Check for slave of a Maxtor at UDMA6 */ | 270 | /* Check for slave of a Maxtor at UDMA6 */ |
348 | ata_id_c_string(pair->id, model_num, ATA_ID_PROD, | 271 | ata_id_c_string(pair->id, model_num, ATA_ID_PROD, |
@@ -351,7 +274,7 @@ static unsigned long pdc2027x_mode_filter(struct ata_device *adev, unsigned long | |||
351 | if (strstr(model_num, "Maxtor") == NULL && pair->dma_mode == XFER_UDMA_6) | 274 | if (strstr(model_num, "Maxtor") == NULL && pair->dma_mode == XFER_UDMA_6) |
352 | mask &= ~ (1 << (6 + ATA_SHIFT_UDMA)); | 275 | mask &= ~ (1 << (6 + ATA_SHIFT_UDMA)); |
353 | 276 | ||
354 | return ata_pci_default_filter(adev, mask); | 277 | return ata_bmdma_mode_filter(adev, mask); |
355 | } | 278 | } |
356 | 279 | ||
357 | /** | 280 | /** |
@@ -836,8 +759,8 @@ static int __devinit pdc2027x_init_one(struct pci_dev *pdev, const struct pci_de | |||
836 | return -EIO; | 759 | return -EIO; |
837 | 760 | ||
838 | pci_set_master(pdev); | 761 | pci_set_master(pdev); |
839 | return ata_host_activate(host, pdev->irq, ata_interrupt, IRQF_SHARED, | 762 | return ata_host_activate(host, pdev->irq, ata_sff_interrupt, |
840 | &pdc2027x_sht); | 763 | IRQF_SHARED, &pdc2027x_sht); |
841 | } | 764 | } |
842 | 765 | ||
843 | /** | 766 | /** |
diff --git a/drivers/ata/pata_pdc202xx_old.c b/drivers/ata/pata_pdc202xx_old.c index 3ed866723e0c..d2673060bc8d 100644 --- a/drivers/ata/pata_pdc202xx_old.c +++ b/drivers/ata/pata_pdc202xx_old.c | |||
@@ -262,94 +262,34 @@ static int pdc2026x_check_atapi_dma(struct ata_queued_cmd *qc) | |||
262 | } | 262 | } |
263 | 263 | ||
264 | static struct scsi_host_template pdc202xx_sht = { | 264 | static struct scsi_host_template pdc202xx_sht = { |
265 | .module = THIS_MODULE, | 265 | ATA_BMDMA_SHT(DRV_NAME), |
266 | .name = DRV_NAME, | ||
267 | .ioctl = ata_scsi_ioctl, | ||
268 | .queuecommand = ata_scsi_queuecmd, | ||
269 | .can_queue = ATA_DEF_QUEUE, | ||
270 | .this_id = ATA_SHT_THIS_ID, | ||
271 | .sg_tablesize = LIBATA_MAX_PRD, | ||
272 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
273 | .emulated = ATA_SHT_EMULATED, | ||
274 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
275 | .proc_name = DRV_NAME, | ||
276 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
277 | .slave_configure = ata_scsi_slave_config, | ||
278 | .slave_destroy = ata_scsi_slave_destroy, | ||
279 | .bios_param = ata_std_bios_param, | ||
280 | }; | 266 | }; |
281 | 267 | ||
282 | static struct ata_port_operations pdc2024x_port_ops = { | 268 | static struct ata_port_operations pdc2024x_port_ops = { |
283 | .set_piomode = pdc202xx_set_piomode, | 269 | .inherits = &ata_bmdma_port_ops, |
284 | .set_dmamode = pdc202xx_set_dmamode, | 270 | |
285 | .mode_filter = ata_pci_default_filter, | 271 | .cable_detect = ata_cable_40wire, |
286 | .tf_load = ata_tf_load, | 272 | .set_piomode = pdc202xx_set_piomode, |
287 | .tf_read = ata_tf_read, | 273 | .set_dmamode = pdc202xx_set_dmamode, |
288 | .check_status = ata_check_status, | ||
289 | .exec_command = ata_exec_command, | ||
290 | .dev_select = ata_std_dev_select, | ||
291 | |||
292 | .freeze = ata_bmdma_freeze, | ||
293 | .thaw = ata_bmdma_thaw, | ||
294 | .error_handler = ata_bmdma_error_handler, | ||
295 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
296 | .cable_detect = ata_cable_40wire, | ||
297 | |||
298 | .bmdma_setup = ata_bmdma_setup, | ||
299 | .bmdma_start = ata_bmdma_start, | ||
300 | .bmdma_stop = ata_bmdma_stop, | ||
301 | .bmdma_status = ata_bmdma_status, | ||
302 | |||
303 | .qc_prep = ata_qc_prep, | ||
304 | .qc_issue = ata_qc_issue_prot, | ||
305 | .data_xfer = ata_data_xfer, | ||
306 | |||
307 | .irq_handler = ata_interrupt, | ||
308 | .irq_clear = ata_bmdma_irq_clear, | ||
309 | .irq_on = ata_irq_on, | ||
310 | |||
311 | .port_start = ata_sff_port_start, | ||
312 | }; | 274 | }; |
313 | 275 | ||
314 | static struct ata_port_operations pdc2026x_port_ops = { | 276 | static struct ata_port_operations pdc2026x_port_ops = { |
315 | .set_piomode = pdc202xx_set_piomode, | 277 | .inherits = &pdc2024x_port_ops, |
316 | .set_dmamode = pdc202xx_set_dmamode, | 278 | |
317 | .mode_filter = ata_pci_default_filter, | 279 | .check_atapi_dma = pdc2026x_check_atapi_dma, |
318 | .tf_load = ata_tf_load, | 280 | .bmdma_start = pdc2026x_bmdma_start, |
319 | .tf_read = ata_tf_read, | 281 | .bmdma_stop = pdc2026x_bmdma_stop, |
320 | .check_status = ata_check_status, | 282 | |
321 | .exec_command = ata_exec_command, | 283 | .cable_detect = pdc2026x_cable_detect, |
322 | .dev_select = ata_std_dev_select, | 284 | .dev_config = pdc2026x_dev_config, |
323 | .dev_config = pdc2026x_dev_config, | 285 | |
324 | 286 | .port_start = pdc2026x_port_start, | |
325 | .freeze = ata_bmdma_freeze, | ||
326 | .thaw = ata_bmdma_thaw, | ||
327 | .error_handler = ata_bmdma_error_handler, | ||
328 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
329 | .cable_detect = pdc2026x_cable_detect, | ||
330 | |||
331 | .check_atapi_dma= pdc2026x_check_atapi_dma, | ||
332 | .bmdma_setup = ata_bmdma_setup, | ||
333 | .bmdma_start = pdc2026x_bmdma_start, | ||
334 | .bmdma_stop = pdc2026x_bmdma_stop, | ||
335 | .bmdma_status = ata_bmdma_status, | ||
336 | |||
337 | .qc_prep = ata_qc_prep, | ||
338 | .qc_issue = ata_qc_issue_prot, | ||
339 | .data_xfer = ata_data_xfer, | ||
340 | |||
341 | .irq_handler = ata_interrupt, | ||
342 | .irq_clear = ata_bmdma_irq_clear, | ||
343 | .irq_on = ata_irq_on, | ||
344 | |||
345 | .port_start = pdc2026x_port_start, | ||
346 | }; | 287 | }; |
347 | 288 | ||
348 | static int pdc202xx_init_one(struct pci_dev *dev, const struct pci_device_id *id) | 289 | static int pdc202xx_init_one(struct pci_dev *dev, const struct pci_device_id *id) |
349 | { | 290 | { |
350 | static const struct ata_port_info info[3] = { | 291 | static const struct ata_port_info info[3] = { |
351 | { | 292 | { |
352 | .sht = &pdc202xx_sht, | ||
353 | .flags = ATA_FLAG_SLAVE_POSS, | 293 | .flags = ATA_FLAG_SLAVE_POSS, |
354 | .pio_mask = 0x1f, | 294 | .pio_mask = 0x1f, |
355 | .mwdma_mask = 0x07, | 295 | .mwdma_mask = 0x07, |
@@ -357,7 +297,6 @@ static int pdc202xx_init_one(struct pci_dev *dev, const struct pci_device_id *id | |||
357 | .port_ops = &pdc2024x_port_ops | 297 | .port_ops = &pdc2024x_port_ops |
358 | }, | 298 | }, |
359 | { | 299 | { |
360 | .sht = &pdc202xx_sht, | ||
361 | .flags = ATA_FLAG_SLAVE_POSS, | 300 | .flags = ATA_FLAG_SLAVE_POSS, |
362 | .pio_mask = 0x1f, | 301 | .pio_mask = 0x1f, |
363 | .mwdma_mask = 0x07, | 302 | .mwdma_mask = 0x07, |
@@ -365,7 +304,6 @@ static int pdc202xx_init_one(struct pci_dev *dev, const struct pci_device_id *id | |||
365 | .port_ops = &pdc2026x_port_ops | 304 | .port_ops = &pdc2026x_port_ops |
366 | }, | 305 | }, |
367 | { | 306 | { |
368 | .sht = &pdc202xx_sht, | ||
369 | .flags = ATA_FLAG_SLAVE_POSS, | 307 | .flags = ATA_FLAG_SLAVE_POSS, |
370 | .pio_mask = 0x1f, | 308 | .pio_mask = 0x1f, |
371 | .mwdma_mask = 0x07, | 309 | .mwdma_mask = 0x07, |
@@ -386,7 +324,7 @@ static int pdc202xx_init_one(struct pci_dev *dev, const struct pci_device_id *id | |||
386 | return -ENODEV; | 324 | return -ENODEV; |
387 | } | 325 | } |
388 | } | 326 | } |
389 | return ata_pci_init_one(dev, ppi); | 327 | return ata_pci_sff_init_one(dev, ppi, &pdc202xx_sht, NULL); |
390 | } | 328 | } |
391 | 329 | ||
392 | static const struct pci_device_id pdc202xx[] = { | 330 | static const struct pci_device_id pdc202xx[] = { |
diff --git a/drivers/ata/pata_platform.c b/drivers/ata/pata_platform.c index aad7adc6ea56..6527c56c34a3 100644 --- a/drivers/ata/pata_platform.c +++ b/drivers/ata/pata_platform.c | |||
@@ -46,50 +46,16 @@ static int pata_platform_set_mode(struct ata_link *link, struct ata_device **unu | |||
46 | return 0; | 46 | return 0; |
47 | } | 47 | } |
48 | 48 | ||
49 | static int ata_dummy_ret0(struct ata_port *ap) { return 0; } | ||
50 | |||
51 | static struct scsi_host_template pata_platform_sht = { | 49 | static struct scsi_host_template pata_platform_sht = { |
52 | .module = THIS_MODULE, | 50 | ATA_PIO_SHT(DRV_NAME), |
53 | .name = DRV_NAME, | ||
54 | .ioctl = ata_scsi_ioctl, | ||
55 | .queuecommand = ata_scsi_queuecmd, | ||
56 | .can_queue = ATA_DEF_QUEUE, | ||
57 | .this_id = ATA_SHT_THIS_ID, | ||
58 | .sg_tablesize = LIBATA_MAX_PRD, | ||
59 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
60 | .emulated = ATA_SHT_EMULATED, | ||
61 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
62 | .proc_name = DRV_NAME, | ||
63 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
64 | .slave_configure = ata_scsi_slave_config, | ||
65 | .slave_destroy = ata_scsi_slave_destroy, | ||
66 | .bios_param = ata_std_bios_param, | ||
67 | }; | 51 | }; |
68 | 52 | ||
69 | static struct ata_port_operations pata_platform_port_ops = { | 53 | static struct ata_port_operations pata_platform_port_ops = { |
70 | .set_mode = pata_platform_set_mode, | 54 | .inherits = &ata_sff_port_ops, |
71 | 55 | .sff_data_xfer = ata_sff_data_xfer_noirq, | |
72 | .tf_load = ata_tf_load, | ||
73 | .tf_read = ata_tf_read, | ||
74 | .check_status = ata_check_status, | ||
75 | .exec_command = ata_exec_command, | ||
76 | .dev_select = ata_std_dev_select, | ||
77 | |||
78 | .freeze = ata_bmdma_freeze, | ||
79 | .thaw = ata_bmdma_thaw, | ||
80 | .error_handler = ata_bmdma_error_handler, | ||
81 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
82 | .cable_detect = ata_cable_unknown, | 56 | .cable_detect = ata_cable_unknown, |
83 | 57 | .set_mode = pata_platform_set_mode, | |
84 | .qc_prep = ata_qc_prep, | 58 | .port_start = ATA_OP_NULL, |
85 | .qc_issue = ata_qc_issue_prot, | ||
86 | |||
87 | .data_xfer = ata_data_xfer_noirq, | ||
88 | |||
89 | .irq_clear = ata_bmdma_irq_clear, | ||
90 | .irq_on = ata_irq_on, | ||
91 | |||
92 | .port_start = ata_dummy_ret0, | ||
93 | }; | 59 | }; |
94 | 60 | ||
95 | static void pata_platform_setup_port(struct ata_ioports *ioaddr, | 61 | static void pata_platform_setup_port(struct ata_ioports *ioaddr, |
@@ -210,7 +176,7 @@ int __devinit __pata_platform_probe(struct device *dev, | |||
210 | (unsigned long long)ctl_res->start); | 176 | (unsigned long long)ctl_res->start); |
211 | 177 | ||
212 | /* activate */ | 178 | /* activate */ |
213 | return ata_host_activate(host, irq, irq ? ata_interrupt : NULL, | 179 | return ata_host_activate(host, irq, irq ? ata_sff_interrupt : NULL, |
214 | irq_flags, &pata_platform_sht); | 180 | irq_flags, &pata_platform_sht); |
215 | } | 181 | } |
216 | EXPORT_SYMBOL_GPL(__pata_platform_probe); | 182 | EXPORT_SYMBOL_GPL(__pata_platform_probe); |
diff --git a/drivers/ata/pata_qdi.c b/drivers/ata/pata_qdi.c index 9f308ed76cc8..bf45cf017753 100644 --- a/drivers/ata/pata_qdi.c +++ b/drivers/ata/pata_qdi.c | |||
@@ -102,14 +102,14 @@ static void qdi6580_set_piomode(struct ata_port *ap, struct ata_device *adev) | |||
102 | } | 102 | } |
103 | 103 | ||
104 | /** | 104 | /** |
105 | * qdi_qc_issue_prot - command issue | 105 | * qdi_qc_issue - command issue |
106 | * @qc: command pending | 106 | * @qc: command pending |
107 | * | 107 | * |
108 | * Called when the libata layer is about to issue a command. We wrap | 108 | * Called when the libata layer is about to issue a command. We wrap |
109 | * this interface so that we can load the correct ATA timings. | 109 | * this interface so that we can load the correct ATA timings. |
110 | */ | 110 | */ |
111 | 111 | ||
112 | static unsigned int qdi_qc_issue_prot(struct ata_queued_cmd *qc) | 112 | static unsigned int qdi_qc_issue(struct ata_queued_cmd *qc) |
113 | { | 113 | { |
114 | struct ata_port *ap = qc->ap; | 114 | struct ata_port *ap = qc->ap; |
115 | struct ata_device *adev = qc->dev; | 115 | struct ata_device *adev = qc->dev; |
@@ -121,7 +121,7 @@ static unsigned int qdi_qc_issue_prot(struct ata_queued_cmd *qc) | |||
121 | outb(qdi->clock[adev->devno], qdi->timing); | 121 | outb(qdi->clock[adev->devno], qdi->timing); |
122 | } | 122 | } |
123 | } | 123 | } |
124 | return ata_qc_issue_prot(qc); | 124 | return ata_sff_qc_issue(qc); |
125 | } | 125 | } |
126 | 126 | ||
127 | static unsigned int qdi_data_xfer(struct ata_device *dev, unsigned char *buf, | 127 | static unsigned int qdi_data_xfer(struct ata_device *dev, unsigned char *buf, |
@@ -148,79 +148,26 @@ static unsigned int qdi_data_xfer(struct ata_device *dev, unsigned char *buf, | |||
148 | buflen += 4 - slop; | 148 | buflen += 4 - slop; |
149 | } | 149 | } |
150 | } else | 150 | } else |
151 | buflen = ata_data_xfer(dev, buf, buflen, rw); | 151 | buflen = ata_sff_data_xfer(dev, buf, buflen, rw); |
152 | 152 | ||
153 | return buflen; | 153 | return buflen; |
154 | } | 154 | } |
155 | 155 | ||
156 | static struct scsi_host_template qdi_sht = { | 156 | static struct scsi_host_template qdi_sht = { |
157 | .module = THIS_MODULE, | 157 | ATA_PIO_SHT(DRV_NAME), |
158 | .name = DRV_NAME, | ||
159 | .ioctl = ata_scsi_ioctl, | ||
160 | .queuecommand = ata_scsi_queuecmd, | ||
161 | .can_queue = ATA_DEF_QUEUE, | ||
162 | .this_id = ATA_SHT_THIS_ID, | ||
163 | .sg_tablesize = LIBATA_MAX_PRD, | ||
164 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
165 | .emulated = ATA_SHT_EMULATED, | ||
166 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
167 | .proc_name = DRV_NAME, | ||
168 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
169 | .slave_configure = ata_scsi_slave_config, | ||
170 | .slave_destroy = ata_scsi_slave_destroy, | ||
171 | .bios_param = ata_std_bios_param, | ||
172 | }; | 158 | }; |
173 | 159 | ||
174 | static struct ata_port_operations qdi6500_port_ops = { | 160 | static struct ata_port_operations qdi6500_port_ops = { |
175 | .set_piomode = qdi6500_set_piomode, | 161 | .inherits = &ata_sff_port_ops, |
176 | 162 | .qc_issue = qdi_qc_issue, | |
177 | .tf_load = ata_tf_load, | 163 | .sff_data_xfer = qdi_data_xfer, |
178 | .tf_read = ata_tf_read, | ||
179 | .check_status = ata_check_status, | ||
180 | .exec_command = ata_exec_command, | ||
181 | .dev_select = ata_std_dev_select, | ||
182 | |||
183 | .freeze = ata_bmdma_freeze, | ||
184 | .thaw = ata_bmdma_thaw, | ||
185 | .error_handler = ata_bmdma_error_handler, | ||
186 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
187 | .cable_detect = ata_cable_40wire, | 164 | .cable_detect = ata_cable_40wire, |
188 | 165 | .set_piomode = qdi6500_set_piomode, | |
189 | .qc_prep = ata_qc_prep, | ||
190 | .qc_issue = qdi_qc_issue_prot, | ||
191 | |||
192 | .data_xfer = qdi_data_xfer, | ||
193 | |||
194 | .irq_clear = ata_bmdma_irq_clear, | ||
195 | .irq_on = ata_irq_on, | ||
196 | |||
197 | .port_start = ata_sff_port_start, | ||
198 | }; | 166 | }; |
199 | 167 | ||
200 | static struct ata_port_operations qdi6580_port_ops = { | 168 | static struct ata_port_operations qdi6580_port_ops = { |
169 | .inherits = &qdi6500_port_ops, | ||
201 | .set_piomode = qdi6580_set_piomode, | 170 | .set_piomode = qdi6580_set_piomode, |
202 | |||
203 | .tf_load = ata_tf_load, | ||
204 | .tf_read = ata_tf_read, | ||
205 | .check_status = ata_check_status, | ||
206 | .exec_command = ata_exec_command, | ||
207 | .dev_select = ata_std_dev_select, | ||
208 | |||
209 | .freeze = ata_bmdma_freeze, | ||
210 | .thaw = ata_bmdma_thaw, | ||
211 | .error_handler = ata_bmdma_error_handler, | ||
212 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
213 | .cable_detect = ata_cable_40wire, | ||
214 | |||
215 | .qc_prep = ata_qc_prep, | ||
216 | .qc_issue = qdi_qc_issue_prot, | ||
217 | |||
218 | .data_xfer = qdi_data_xfer, | ||
219 | |||
220 | .irq_clear = ata_bmdma_irq_clear, | ||
221 | .irq_on = ata_irq_on, | ||
222 | |||
223 | .port_start = ata_sff_port_start, | ||
224 | }; | 171 | }; |
225 | 172 | ||
226 | /** | 173 | /** |
@@ -276,7 +223,7 @@ static __init int qdi_init_one(unsigned long port, int type, unsigned long io, i | |||
276 | ap->ioaddr.cmd_addr = io_addr; | 223 | ap->ioaddr.cmd_addr = io_addr; |
277 | ap->ioaddr.altstatus_addr = ctl_addr; | 224 | ap->ioaddr.altstatus_addr = ctl_addr; |
278 | ap->ioaddr.ctl_addr = ctl_addr; | 225 | ap->ioaddr.ctl_addr = ctl_addr; |
279 | ata_std_ports(&ap->ioaddr); | 226 | ata_sff_std_ports(&ap->ioaddr); |
280 | 227 | ||
281 | ata_port_desc(ap, "cmd %lx ctl %lx", io, ctl); | 228 | ata_port_desc(ap, "cmd %lx ctl %lx", io, ctl); |
282 | 229 | ||
@@ -292,7 +239,7 @@ static __init int qdi_init_one(unsigned long port, int type, unsigned long io, i | |||
292 | printk(KERN_INFO DRV_NAME": qd%d at 0x%lx.\n", type, io); | 239 | printk(KERN_INFO DRV_NAME": qd%d at 0x%lx.\n", type, io); |
293 | 240 | ||
294 | /* activate */ | 241 | /* activate */ |
295 | ret = ata_host_activate(host, irq, ata_interrupt, 0, &qdi_sht); | 242 | ret = ata_host_activate(host, irq, ata_sff_interrupt, 0, &qdi_sht); |
296 | if (ret) | 243 | if (ret) |
297 | goto fail; | 244 | goto fail; |
298 | 245 | ||
diff --git a/drivers/ata/pata_radisys.c b/drivers/ata/pata_radisys.c index 8109b08fc024..1c0d9fa7ee54 100644 --- a/drivers/ata/pata_radisys.c +++ b/drivers/ata/pata_radisys.c | |||
@@ -156,7 +156,7 @@ static void radisys_set_dmamode (struct ata_port *ap, struct ata_device *adev) | |||
156 | } | 156 | } |
157 | 157 | ||
158 | /** | 158 | /** |
159 | * radisys_qc_issue_prot - command issue | 159 | * radisys_qc_issue - command issue |
160 | * @qc: command pending | 160 | * @qc: command pending |
161 | * | 161 | * |
162 | * Called when the libata layer is about to issue a command. We wrap | 162 | * Called when the libata layer is about to issue a command. We wrap |
@@ -166,7 +166,7 @@ static void radisys_set_dmamode (struct ata_port *ap, struct ata_device *adev) | |||
166 | * be made PIO0. | 166 | * be made PIO0. |
167 | */ | 167 | */ |
168 | 168 | ||
169 | static unsigned int radisys_qc_issue_prot(struct ata_queued_cmd *qc) | 169 | static unsigned int radisys_qc_issue(struct ata_queued_cmd *qc) |
170 | { | 170 | { |
171 | struct ata_port *ap = qc->ap; | 171 | struct ata_port *ap = qc->ap; |
172 | struct ata_device *adev = qc->dev; | 172 | struct ata_device *adev = qc->dev; |
@@ -180,58 +180,20 @@ static unsigned int radisys_qc_issue_prot(struct ata_queued_cmd *qc) | |||
180 | radisys_set_piomode(ap, adev); | 180 | radisys_set_piomode(ap, adev); |
181 | } | 181 | } |
182 | } | 182 | } |
183 | return ata_qc_issue_prot(qc); | 183 | return ata_sff_qc_issue(qc); |
184 | } | 184 | } |
185 | 185 | ||
186 | 186 | ||
187 | static struct scsi_host_template radisys_sht = { | 187 | static struct scsi_host_template radisys_sht = { |
188 | .module = THIS_MODULE, | 188 | ATA_BMDMA_SHT(DRV_NAME), |
189 | .name = DRV_NAME, | ||
190 | .ioctl = ata_scsi_ioctl, | ||
191 | .queuecommand = ata_scsi_queuecmd, | ||
192 | .can_queue = ATA_DEF_QUEUE, | ||
193 | .this_id = ATA_SHT_THIS_ID, | ||
194 | .sg_tablesize = LIBATA_MAX_PRD, | ||
195 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
196 | .emulated = ATA_SHT_EMULATED, | ||
197 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
198 | .proc_name = DRV_NAME, | ||
199 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
200 | .slave_configure = ata_scsi_slave_config, | ||
201 | .slave_destroy = ata_scsi_slave_destroy, | ||
202 | .bios_param = ata_std_bios_param, | ||
203 | }; | 189 | }; |
204 | 190 | ||
205 | static const struct ata_port_operations radisys_pata_ops = { | 191 | static struct ata_port_operations radisys_pata_ops = { |
192 | .inherits = &ata_bmdma_port_ops, | ||
193 | .qc_issue = radisys_qc_issue, | ||
194 | .cable_detect = ata_cable_unknown, | ||
206 | .set_piomode = radisys_set_piomode, | 195 | .set_piomode = radisys_set_piomode, |
207 | .set_dmamode = radisys_set_dmamode, | 196 | .set_dmamode = radisys_set_dmamode, |
208 | .mode_filter = ata_pci_default_filter, | ||
209 | |||
210 | .tf_load = ata_tf_load, | ||
211 | .tf_read = ata_tf_read, | ||
212 | .check_status = ata_check_status, | ||
213 | .exec_command = ata_exec_command, | ||
214 | .dev_select = ata_std_dev_select, | ||
215 | |||
216 | .freeze = ata_bmdma_freeze, | ||
217 | .thaw = ata_bmdma_thaw, | ||
218 | .error_handler = ata_bmdma_error_handler, | ||
219 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
220 | .cable_detect = ata_cable_unknown, | ||
221 | |||
222 | .bmdma_setup = ata_bmdma_setup, | ||
223 | .bmdma_start = ata_bmdma_start, | ||
224 | .bmdma_stop = ata_bmdma_stop, | ||
225 | .bmdma_status = ata_bmdma_status, | ||
226 | .qc_prep = ata_qc_prep, | ||
227 | .qc_issue = radisys_qc_issue_prot, | ||
228 | .data_xfer = ata_data_xfer, | ||
229 | |||
230 | .irq_handler = ata_interrupt, | ||
231 | .irq_clear = ata_bmdma_irq_clear, | ||
232 | .irq_on = ata_irq_on, | ||
233 | |||
234 | .port_start = ata_sff_port_start, | ||
235 | }; | 197 | }; |
236 | 198 | ||
237 | 199 | ||
@@ -254,7 +216,6 @@ static int radisys_init_one (struct pci_dev *pdev, const struct pci_device_id *e | |||
254 | { | 216 | { |
255 | static int printed_version; | 217 | static int printed_version; |
256 | static const struct ata_port_info info = { | 218 | static const struct ata_port_info info = { |
257 | .sht = &radisys_sht, | ||
258 | .flags = ATA_FLAG_SLAVE_POSS, | 219 | .flags = ATA_FLAG_SLAVE_POSS, |
259 | .pio_mask = 0x1f, /* pio0-4 */ | 220 | .pio_mask = 0x1f, /* pio0-4 */ |
260 | .mwdma_mask = 0x07, /* mwdma1-2 */ | 221 | .mwdma_mask = 0x07, /* mwdma1-2 */ |
@@ -267,7 +228,7 @@ static int radisys_init_one (struct pci_dev *pdev, const struct pci_device_id *e | |||
267 | dev_printk(KERN_DEBUG, &pdev->dev, | 228 | dev_printk(KERN_DEBUG, &pdev->dev, |
268 | "version " DRV_VERSION "\n"); | 229 | "version " DRV_VERSION "\n"); |
269 | 230 | ||
270 | return ata_pci_init_one(pdev, ppi); | 231 | return ata_pci_sff_init_one(pdev, ppi, &radisys_sht, NULL); |
271 | } | 232 | } |
272 | 233 | ||
273 | static const struct pci_device_id radisys_pci_tbl[] = { | 234 | static const struct pci_device_id radisys_pci_tbl[] = { |
diff --git a/drivers/ata/pata_rb500_cf.c b/drivers/ata/pata_rb500_cf.c index 4ce9b03fe6c8..800ae4601f44 100644 --- a/drivers/ata/pata_rb500_cf.c +++ b/drivers/ata/pata_rb500_cf.c | |||
@@ -57,7 +57,7 @@ static inline void rb500_pata_finish_io(struct ata_port *ap) | |||
57 | struct ata_host *ah = ap->host; | 57 | struct ata_host *ah = ap->host; |
58 | struct rb500_cf_info *info = ah->private_data; | 58 | struct rb500_cf_info *info = ah->private_data; |
59 | 59 | ||
60 | ata_altstatus(ap); | 60 | ata_sff_altstatus(ap); |
61 | ndelay(RB500_CF_IO_DELAY); | 61 | ndelay(RB500_CF_IO_DELAY); |
62 | 62 | ||
63 | set_irq_type(info->irq, IRQ_TYPE_LEVEL_HIGH); | 63 | set_irq_type(info->irq, IRQ_TYPE_LEVEL_HIGH); |
@@ -109,7 +109,7 @@ static irqreturn_t rb500_pata_irq_handler(int irq, void *dev_instance) | |||
109 | if (gpio_get_value(info->gpio_line)) { | 109 | if (gpio_get_value(info->gpio_line)) { |
110 | set_irq_type(info->irq, IRQ_TYPE_LEVEL_LOW); | 110 | set_irq_type(info->irq, IRQ_TYPE_LEVEL_LOW); |
111 | if (!info->frozen) | 111 | if (!info->frozen) |
112 | ata_interrupt(info->irq, dev_instance); | 112 | ata_sff_interrupt(info->irq, dev_instance); |
113 | } else { | 113 | } else { |
114 | set_irq_type(info->irq, IRQ_TYPE_LEVEL_HIGH); | 114 | set_irq_type(info->irq, IRQ_TYPE_LEVEL_HIGH); |
115 | } | 115 | } |
@@ -117,58 +117,18 @@ static irqreturn_t rb500_pata_irq_handler(int irq, void *dev_instance) | |||
117 | return IRQ_HANDLED; | 117 | return IRQ_HANDLED; |
118 | } | 118 | } |
119 | 119 | ||
120 | static void rb500_pata_irq_clear(struct ata_port *ap) | ||
121 | { | ||
122 | } | ||
123 | |||
124 | static int rb500_pata_port_start(struct ata_port *ap) | ||
125 | { | ||
126 | return 0; | ||
127 | } | ||
128 | |||
129 | static struct ata_port_operations rb500_pata_port_ops = { | 120 | static struct ata_port_operations rb500_pata_port_ops = { |
130 | .tf_load = ata_tf_load, | 121 | .inherits = &ata_sff_port_ops, |
131 | .tf_read = ata_tf_read, | 122 | .sff_exec_command = rb500_pata_exec_command, |
132 | 123 | .sff_data_xfer = rb500_pata_data_xfer, | |
133 | .exec_command = rb500_pata_exec_command, | ||
134 | .check_status = ata_check_status, | ||
135 | .dev_select = ata_std_dev_select, | ||
136 | |||
137 | .data_xfer = rb500_pata_data_xfer, | ||
138 | |||
139 | .qc_prep = ata_qc_prep, | ||
140 | .qc_issue = ata_qc_issue_prot, | ||
141 | |||
142 | .freeze = rb500_pata_freeze, | 124 | .freeze = rb500_pata_freeze, |
143 | .thaw = rb500_pata_thaw, | 125 | .thaw = rb500_pata_thaw, |
144 | .error_handler = ata_bmdma_error_handler, | ||
145 | |||
146 | .irq_handler = rb500_pata_irq_handler, | ||
147 | .irq_clear = rb500_pata_irq_clear, | ||
148 | .irq_on = ata_irq_on, | ||
149 | |||
150 | .port_start = rb500_pata_port_start, | ||
151 | }; | 126 | }; |
152 | 127 | ||
153 | /* ------------------------------------------------------------------------ */ | 128 | /* ------------------------------------------------------------------------ */ |
154 | 129 | ||
155 | static struct scsi_host_template rb500_pata_sht = { | 130 | static struct scsi_host_template rb500_pata_sht = { |
156 | .module = THIS_MODULE, | 131 | ATA_PIO_SHT(DRV_NAME), |
157 | .name = DRV_NAME, | ||
158 | .ioctl = ata_scsi_ioctl, | ||
159 | .queuecommand = ata_scsi_queuecmd, | ||
160 | .slave_configure = ata_scsi_slave_config, | ||
161 | .slave_destroy = ata_scsi_slave_destroy, | ||
162 | .bios_param = ata_std_bios_param, | ||
163 | .proc_name = DRV_NAME, | ||
164 | |||
165 | .can_queue = ATA_DEF_QUEUE, | ||
166 | .this_id = ATA_SHT_THIS_ID, | ||
167 | .sg_tablesize = LIBATA_MAX_PRD, | ||
168 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
169 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
170 | .emulated = ATA_SHT_EMULATED, | ||
171 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
172 | }; | 132 | }; |
173 | 133 | ||
174 | /* ------------------------------------------------------------------------ */ | 134 | /* ------------------------------------------------------------------------ */ |
@@ -188,7 +148,7 @@ static void rb500_pata_setup_ports(struct ata_host *ah) | |||
188 | ap->ioaddr.ctl_addr = info->iobase + RB500_CF_REG_CTRL; | 148 | ap->ioaddr.ctl_addr = info->iobase + RB500_CF_REG_CTRL; |
189 | ap->ioaddr.altstatus_addr = info->iobase + RB500_CF_REG_CTRL; | 149 | ap->ioaddr.altstatus_addr = info->iobase + RB500_CF_REG_CTRL; |
190 | 150 | ||
191 | ata_std_ports(&ap->ioaddr); | 151 | ata_sff_std_ports(&ap->ioaddr); |
192 | 152 | ||
193 | ap->ioaddr.data_addr = info->iobase + RB500_CF_REG_DATA; | 153 | ap->ioaddr.data_addr = info->iobase + RB500_CF_REG_DATA; |
194 | } | 154 | } |
diff --git a/drivers/ata/pata_rz1000.c b/drivers/ata/pata_rz1000.c index ba8a31c55edb..7dfd1f3f6f3a 100644 --- a/drivers/ata/pata_rz1000.c +++ b/drivers/ata/pata_rz1000.c | |||
@@ -53,53 +53,13 @@ static int rz1000_set_mode(struct ata_link *link, struct ata_device **unused) | |||
53 | 53 | ||
54 | 54 | ||
55 | static struct scsi_host_template rz1000_sht = { | 55 | static struct scsi_host_template rz1000_sht = { |
56 | .module = THIS_MODULE, | 56 | ATA_PIO_SHT(DRV_NAME), |
57 | .name = DRV_NAME, | ||
58 | .ioctl = ata_scsi_ioctl, | ||
59 | .queuecommand = ata_scsi_queuecmd, | ||
60 | .can_queue = ATA_DEF_QUEUE, | ||
61 | .this_id = ATA_SHT_THIS_ID, | ||
62 | .sg_tablesize = LIBATA_MAX_PRD, | ||
63 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
64 | .emulated = ATA_SHT_EMULATED, | ||
65 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
66 | .proc_name = DRV_NAME, | ||
67 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
68 | .slave_configure = ata_scsi_slave_config, | ||
69 | .slave_destroy = ata_scsi_slave_destroy, | ||
70 | .bios_param = ata_std_bios_param, | ||
71 | }; | 57 | }; |
72 | 58 | ||
73 | static struct ata_port_operations rz1000_port_ops = { | 59 | static struct ata_port_operations rz1000_port_ops = { |
74 | .set_mode = rz1000_set_mode, | 60 | .inherits = &ata_sff_port_ops, |
75 | |||
76 | .tf_load = ata_tf_load, | ||
77 | .tf_read = ata_tf_read, | ||
78 | .check_status = ata_check_status, | ||
79 | .exec_command = ata_exec_command, | ||
80 | .dev_select = ata_std_dev_select, | ||
81 | |||
82 | .bmdma_setup = ata_bmdma_setup, | ||
83 | .bmdma_start = ata_bmdma_start, | ||
84 | .bmdma_stop = ata_bmdma_stop, | ||
85 | .bmdma_status = ata_bmdma_status, | ||
86 | |||
87 | .qc_prep = ata_qc_prep, | ||
88 | .qc_issue = ata_qc_issue_prot, | ||
89 | |||
90 | .data_xfer = ata_data_xfer, | ||
91 | |||
92 | .freeze = ata_bmdma_freeze, | ||
93 | .thaw = ata_bmdma_thaw, | ||
94 | .error_handler = ata_bmdma_error_handler, | ||
95 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
96 | .cable_detect = ata_cable_40wire, | 61 | .cable_detect = ata_cable_40wire, |
97 | 62 | .set_mode = rz1000_set_mode, | |
98 | .irq_handler = ata_interrupt, | ||
99 | .irq_clear = ata_bmdma_irq_clear, | ||
100 | .irq_on = ata_irq_on, | ||
101 | |||
102 | .port_start = ata_sff_port_start, | ||
103 | }; | 63 | }; |
104 | 64 | ||
105 | static int rz1000_fifo_disable(struct pci_dev *pdev) | 65 | static int rz1000_fifo_disable(struct pci_dev *pdev) |
@@ -129,7 +89,6 @@ static int rz1000_init_one (struct pci_dev *pdev, const struct pci_device_id *en | |||
129 | { | 89 | { |
130 | static int printed_version; | 90 | static int printed_version; |
131 | static const struct ata_port_info info = { | 91 | static const struct ata_port_info info = { |
132 | .sht = &rz1000_sht, | ||
133 | .flags = ATA_FLAG_SLAVE_POSS, | 92 | .flags = ATA_FLAG_SLAVE_POSS, |
134 | .pio_mask = 0x1f, | 93 | .pio_mask = 0x1f, |
135 | .port_ops = &rz1000_port_ops | 94 | .port_ops = &rz1000_port_ops |
@@ -140,7 +99,7 @@ static int rz1000_init_one (struct pci_dev *pdev, const struct pci_device_id *en | |||
140 | printk(KERN_DEBUG DRV_NAME " version " DRV_VERSION "\n"); | 99 | printk(KERN_DEBUG DRV_NAME " version " DRV_VERSION "\n"); |
141 | 100 | ||
142 | if (rz1000_fifo_disable(pdev) == 0) | 101 | if (rz1000_fifo_disable(pdev) == 0) |
143 | return ata_pci_init_one(pdev, ppi); | 102 | return ata_pci_sff_init_one(pdev, ppi, &rz1000_sht, NULL); |
144 | 103 | ||
145 | printk(KERN_ERR DRV_NAME ": failed to disable read-ahead on chipset..\n"); | 104 | printk(KERN_ERR DRV_NAME ": failed to disable read-ahead on chipset..\n"); |
146 | /* Not safe to use so skip */ | 105 | /* Not safe to use so skip */ |
diff --git a/drivers/ata/pata_sc1200.c b/drivers/ata/pata_sc1200.c index 725a8586cd6e..cbab397e3db7 100644 --- a/drivers/ata/pata_sc1200.c +++ b/drivers/ata/pata_sc1200.c | |||
@@ -151,7 +151,7 @@ static void sc1200_set_dmamode(struct ata_port *ap, struct ata_device *adev) | |||
151 | } | 151 | } |
152 | 152 | ||
153 | /** | 153 | /** |
154 | * sc1200_qc_issue_prot - command issue | 154 | * sc1200_qc_issue - command issue |
155 | * @qc: command pending | 155 | * @qc: command pending |
156 | * | 156 | * |
157 | * Called when the libata layer is about to issue a command. We wrap | 157 | * Called when the libata layer is about to issue a command. We wrap |
@@ -160,7 +160,7 @@ static void sc1200_set_dmamode(struct ata_port *ap, struct ata_device *adev) | |||
160 | * one MWDMA/UDMA bit. | 160 | * one MWDMA/UDMA bit. |
161 | */ | 161 | */ |
162 | 162 | ||
163 | static unsigned int sc1200_qc_issue_prot(struct ata_queued_cmd *qc) | 163 | static unsigned int sc1200_qc_issue(struct ata_queued_cmd *qc) |
164 | { | 164 | { |
165 | struct ata_port *ap = qc->ap; | 165 | struct ata_port *ap = qc->ap; |
166 | struct ata_device *adev = qc->dev; | 166 | struct ata_device *adev = qc->dev; |
@@ -175,59 +175,21 @@ static unsigned int sc1200_qc_issue_prot(struct ata_queued_cmd *qc) | |||
175 | sc1200_set_dmamode(ap, adev); | 175 | sc1200_set_dmamode(ap, adev); |
176 | } | 176 | } |
177 | 177 | ||
178 | return ata_qc_issue_prot(qc); | 178 | return ata_sff_qc_issue(qc); |
179 | } | 179 | } |
180 | 180 | ||
181 | static struct scsi_host_template sc1200_sht = { | 181 | static struct scsi_host_template sc1200_sht = { |
182 | .module = THIS_MODULE, | 182 | ATA_BMDMA_SHT(DRV_NAME), |
183 | .name = DRV_NAME, | 183 | .sg_tablesize = LIBATA_DUMB_MAX_PRD, |
184 | .ioctl = ata_scsi_ioctl, | ||
185 | .queuecommand = ata_scsi_queuecmd, | ||
186 | .can_queue = ATA_DEF_QUEUE, | ||
187 | .this_id = ATA_SHT_THIS_ID, | ||
188 | .sg_tablesize = LIBATA_DUMB_MAX_PRD, | ||
189 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
190 | .emulated = ATA_SHT_EMULATED, | ||
191 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
192 | .proc_name = DRV_NAME, | ||
193 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
194 | .slave_configure = ata_scsi_slave_config, | ||
195 | .slave_destroy = ata_scsi_slave_destroy, | ||
196 | .bios_param = ata_std_bios_param, | ||
197 | }; | 184 | }; |
198 | 185 | ||
199 | static struct ata_port_operations sc1200_port_ops = { | 186 | static struct ata_port_operations sc1200_port_ops = { |
187 | .inherits = &ata_bmdma_port_ops, | ||
188 | .qc_prep = ata_sff_dumb_qc_prep, | ||
189 | .qc_issue = sc1200_qc_issue, | ||
190 | .cable_detect = ata_cable_40wire, | ||
200 | .set_piomode = sc1200_set_piomode, | 191 | .set_piomode = sc1200_set_piomode, |
201 | .set_dmamode = sc1200_set_dmamode, | 192 | .set_dmamode = sc1200_set_dmamode, |
202 | .mode_filter = ata_pci_default_filter, | ||
203 | |||
204 | .tf_load = ata_tf_load, | ||
205 | .tf_read = ata_tf_read, | ||
206 | .check_status = ata_check_status, | ||
207 | .exec_command = ata_exec_command, | ||
208 | .dev_select = ata_std_dev_select, | ||
209 | |||
210 | .freeze = ata_bmdma_freeze, | ||
211 | .thaw = ata_bmdma_thaw, | ||
212 | .error_handler = ata_bmdma_error_handler, | ||
213 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
214 | .cable_detect = ata_cable_40wire, | ||
215 | |||
216 | .bmdma_setup = ata_bmdma_setup, | ||
217 | .bmdma_start = ata_bmdma_start, | ||
218 | .bmdma_stop = ata_bmdma_stop, | ||
219 | .bmdma_status = ata_bmdma_status, | ||
220 | |||
221 | .qc_prep = ata_dumb_qc_prep, | ||
222 | .qc_issue = sc1200_qc_issue_prot, | ||
223 | |||
224 | .data_xfer = ata_data_xfer, | ||
225 | |||
226 | .irq_handler = ata_interrupt, | ||
227 | .irq_clear = ata_bmdma_irq_clear, | ||
228 | .irq_on = ata_irq_on, | ||
229 | |||
230 | .port_start = ata_sff_port_start, | ||
231 | }; | 193 | }; |
232 | 194 | ||
233 | /** | 195 | /** |
@@ -242,7 +204,6 @@ static struct ata_port_operations sc1200_port_ops = { | |||
242 | static int sc1200_init_one(struct pci_dev *dev, const struct pci_device_id *id) | 204 | static int sc1200_init_one(struct pci_dev *dev, const struct pci_device_id *id) |
243 | { | 205 | { |
244 | static const struct ata_port_info info = { | 206 | static const struct ata_port_info info = { |
245 | .sht = &sc1200_sht, | ||
246 | .flags = ATA_FLAG_SLAVE_POSS, | 207 | .flags = ATA_FLAG_SLAVE_POSS, |
247 | .pio_mask = 0x1f, | 208 | .pio_mask = 0x1f, |
248 | .mwdma_mask = 0x07, | 209 | .mwdma_mask = 0x07, |
@@ -252,7 +213,7 @@ static int sc1200_init_one(struct pci_dev *dev, const struct pci_device_id *id) | |||
252 | /* Can't enable port 2 yet, see top comments */ | 213 | /* Can't enable port 2 yet, see top comments */ |
253 | const struct ata_port_info *ppi[] = { &info, &ata_dummy_port_info }; | 214 | const struct ata_port_info *ppi[] = { &info, &ata_dummy_port_info }; |
254 | 215 | ||
255 | return ata_pci_init_one(dev, ppi); | 216 | return ata_pci_sff_init_one(dev, ppi, &sc1200_sht, NULL); |
256 | } | 217 | } |
257 | 218 | ||
258 | static const struct pci_device_id sc1200[] = { | 219 | static const struct pci_device_id sc1200[] = { |
diff --git a/drivers/ata/pata_scc.c b/drivers/ata/pata_scc.c index 6c016deeaed8..e965b251ca24 100644 --- a/drivers/ata/pata_scc.c +++ b/drivers/ata/pata_scc.c | |||
@@ -266,7 +266,7 @@ unsigned long scc_mode_filter(struct ata_device *adev, unsigned long mask) | |||
266 | printk(KERN_INFO "%s: limit ATAPI UDMA to UDMA4\n", DRV_NAME); | 266 | printk(KERN_INFO "%s: limit ATAPI UDMA to UDMA4\n", DRV_NAME); |
267 | mask &= ~(0xE0 << ATA_SHIFT_UDMA); | 267 | mask &= ~(0xE0 << ATA_SHIFT_UDMA); |
268 | } | 268 | } |
269 | return ata_pci_default_filter(adev, mask); | 269 | return ata_bmdma_mode_filter(adev, mask); |
270 | } | 270 | } |
271 | 271 | ||
272 | /** | 272 | /** |
@@ -274,7 +274,7 @@ unsigned long scc_mode_filter(struct ata_device *adev, unsigned long mask) | |||
274 | * @ap: Port to which output is sent | 274 | * @ap: Port to which output is sent |
275 | * @tf: ATA taskfile register set | 275 | * @tf: ATA taskfile register set |
276 | * | 276 | * |
277 | * Note: Original code is ata_tf_load(). | 277 | * Note: Original code is ata_sff_tf_load(). |
278 | */ | 278 | */ |
279 | 279 | ||
280 | static void scc_tf_load (struct ata_port *ap, const struct ata_taskfile *tf) | 280 | static void scc_tf_load (struct ata_port *ap, const struct ata_taskfile *tf) |
@@ -341,7 +341,7 @@ static u8 scc_check_status (struct ata_port *ap) | |||
341 | * @ap: Port from which input is read | 341 | * @ap: Port from which input is read |
342 | * @tf: ATA taskfile register set for storing input | 342 | * @tf: ATA taskfile register set for storing input |
343 | * | 343 | * |
344 | * Note: Original code is ata_tf_read(). | 344 | * Note: Original code is ata_sff_tf_read(). |
345 | */ | 345 | */ |
346 | 346 | ||
347 | static void scc_tf_read (struct ata_port *ap, struct ata_taskfile *tf) | 347 | static void scc_tf_read (struct ata_port *ap, struct ata_taskfile *tf) |
@@ -373,7 +373,7 @@ static void scc_tf_read (struct ata_port *ap, struct ata_taskfile *tf) | |||
373 | * @ap: port to which command is being issued | 373 | * @ap: port to which command is being issued |
374 | * @tf: ATA taskfile register set | 374 | * @tf: ATA taskfile register set |
375 | * | 375 | * |
376 | * Note: Original code is ata_exec_command(). | 376 | * Note: Original code is ata_sff_exec_command(). |
377 | */ | 377 | */ |
378 | 378 | ||
379 | static void scc_exec_command (struct ata_port *ap, | 379 | static void scc_exec_command (struct ata_port *ap, |
@@ -382,7 +382,7 @@ static void scc_exec_command (struct ata_port *ap, | |||
382 | DPRINTK("ata%u: cmd 0x%X\n", ap->print_id, tf->command); | 382 | DPRINTK("ata%u: cmd 0x%X\n", ap->print_id, tf->command); |
383 | 383 | ||
384 | out_be32(ap->ioaddr.command_addr, tf->command); | 384 | out_be32(ap->ioaddr.command_addr, tf->command); |
385 | ata_pause(ap); | 385 | ata_sff_pause(ap); |
386 | } | 386 | } |
387 | 387 | ||
388 | /** | 388 | /** |
@@ -396,14 +396,14 @@ static u8 scc_check_altstatus (struct ata_port *ap) | |||
396 | } | 396 | } |
397 | 397 | ||
398 | /** | 398 | /** |
399 | * scc_std_dev_select - Select device 0/1 on ATA bus | 399 | * scc_dev_select - Select device 0/1 on ATA bus |
400 | * @ap: ATA channel to manipulate | 400 | * @ap: ATA channel to manipulate |
401 | * @device: ATA device (numbered from zero) to select | 401 | * @device: ATA device (numbered from zero) to select |
402 | * | 402 | * |
403 | * Note: Original code is ata_std_dev_select(). | 403 | * Note: Original code is ata_sff_dev_select(). |
404 | */ | 404 | */ |
405 | 405 | ||
406 | static void scc_std_dev_select (struct ata_port *ap, unsigned int device) | 406 | static void scc_dev_select (struct ata_port *ap, unsigned int device) |
407 | { | 407 | { |
408 | u8 tmp; | 408 | u8 tmp; |
409 | 409 | ||
@@ -413,7 +413,7 @@ static void scc_std_dev_select (struct ata_port *ap, unsigned int device) | |||
413 | tmp = ATA_DEVICE_OBS | ATA_DEV1; | 413 | tmp = ATA_DEVICE_OBS | ATA_DEV1; |
414 | 414 | ||
415 | out_be32(ap->ioaddr.device_addr, tmp); | 415 | out_be32(ap->ioaddr.device_addr, tmp); |
416 | ata_pause(ap); | 416 | ata_sff_pause(ap); |
417 | } | 417 | } |
418 | 418 | ||
419 | /** | 419 | /** |
@@ -441,7 +441,7 @@ static void scc_bmdma_setup (struct ata_queued_cmd *qc) | |||
441 | out_be32(mmio + SCC_DMA_CMD, dmactl); | 441 | out_be32(mmio + SCC_DMA_CMD, dmactl); |
442 | 442 | ||
443 | /* issue r/w command */ | 443 | /* issue r/w command */ |
444 | ap->ops->exec_command(ap, &qc->tf); | 444 | ap->ops->sff_exec_command(ap, &qc->tf); |
445 | } | 445 | } |
446 | 446 | ||
447 | /** | 447 | /** |
@@ -476,7 +476,7 @@ static unsigned int scc_devchk (struct ata_port *ap, | |||
476 | struct ata_ioports *ioaddr = &ap->ioaddr; | 476 | struct ata_ioports *ioaddr = &ap->ioaddr; |
477 | u8 nsect, lbal; | 477 | u8 nsect, lbal; |
478 | 478 | ||
479 | ap->ops->dev_select(ap, device); | 479 | ap->ops->sff_dev_select(ap, device); |
480 | 480 | ||
481 | out_be32(ioaddr->nsect_addr, 0x55); | 481 | out_be32(ioaddr->nsect_addr, 0x55); |
482 | out_be32(ioaddr->lbal_addr, 0xaa); | 482 | out_be32(ioaddr->lbal_addr, 0xaa); |
@@ -497,57 +497,78 @@ static unsigned int scc_devchk (struct ata_port *ap, | |||
497 | } | 497 | } |
498 | 498 | ||
499 | /** | 499 | /** |
500 | * scc_bus_post_reset - PATA device post reset | 500 | * scc_wait_after_reset - wait for devices to become ready after reset |
501 | * | 501 | * |
502 | * Note: Original code is ata_bus_post_reset(). | 502 | * Note: Original code is ata_sff_wait_after_reset |
503 | */ | 503 | */ |
504 | 504 | ||
505 | static int scc_bus_post_reset(struct ata_port *ap, unsigned int devmask, | 505 | int scc_wait_after_reset(struct ata_link *link, unsigned int devmask, |
506 | unsigned long deadline) | 506 | unsigned long deadline) |
507 | { | 507 | { |
508 | struct ata_port *ap = link->ap; | ||
508 | struct ata_ioports *ioaddr = &ap->ioaddr; | 509 | struct ata_ioports *ioaddr = &ap->ioaddr; |
509 | unsigned int dev0 = devmask & (1 << 0); | 510 | unsigned int dev0 = devmask & (1 << 0); |
510 | unsigned int dev1 = devmask & (1 << 1); | 511 | unsigned int dev1 = devmask & (1 << 1); |
511 | int rc; | 512 | int rc, ret = 0; |
512 | 513 | ||
513 | /* if device 0 was found in ata_devchk, wait for its | 514 | /* Spec mandates ">= 2ms" before checking status. We wait |
514 | * BSY bit to clear | 515 | * 150ms, because that was the magic delay used for ATAPI |
516 | * devices in Hale Landis's ATADRVR, for the period of time | ||
517 | * between when the ATA command register is written, and then | ||
518 | * status is checked. Because waiting for "a while" before | ||
519 | * checking status is fine, post SRST, we perform this magic | ||
520 | * delay here as well. | ||
521 | * | ||
522 | * Old drivers/ide uses the 2mS rule and then waits for ready. | ||
515 | */ | 523 | */ |
516 | if (dev0) { | 524 | msleep(150); |
517 | rc = ata_wait_ready(ap, deadline); | ||
518 | if (rc && rc != -ENODEV) | ||
519 | return rc; | ||
520 | } | ||
521 | 525 | ||
522 | /* if device 1 was found in ata_devchk, wait for | 526 | /* always check readiness of the master device */ |
523 | * register access, then wait for BSY to clear | 527 | rc = ata_sff_wait_ready(link, deadline); |
528 | /* -ENODEV means the odd clown forgot the D7 pulldown resistor | ||
529 | * and TF status is 0xff, bail out on it too. | ||
524 | */ | 530 | */ |
525 | while (dev1) { | 531 | if (rc) |
526 | u8 nsect, lbal; | 532 | return rc; |
527 | 533 | ||
528 | ap->ops->dev_select(ap, 1); | 534 | /* if device 1 was found in ata_devchk, wait for register |
529 | nsect = in_be32(ioaddr->nsect_addr); | 535 | * access briefly, then wait for BSY to clear. |
530 | lbal = in_be32(ioaddr->lbal_addr); | 536 | */ |
531 | if ((nsect == 1) && (lbal == 1)) | ||
532 | break; | ||
533 | if (time_after(jiffies, deadline)) | ||
534 | return -EBUSY; | ||
535 | msleep(50); /* give drive a breather */ | ||
536 | } | ||
537 | if (dev1) { | 537 | if (dev1) { |
538 | rc = ata_wait_ready(ap, deadline); | 538 | int i; |
539 | if (rc && rc != -ENODEV) | 539 | |
540 | return rc; | 540 | ap->ops->sff_dev_select(ap, 1); |
541 | |||
542 | /* Wait for register access. Some ATAPI devices fail | ||
543 | * to set nsect/lbal after reset, so don't waste too | ||
544 | * much time on it. We're gonna wait for !BSY anyway. | ||
545 | */ | ||
546 | for (i = 0; i < 2; i++) { | ||
547 | u8 nsect, lbal; | ||
548 | |||
549 | nsect = in_be32(ioaddr->nsect_addr); | ||
550 | lbal = in_be32(ioaddr->lbal_addr); | ||
551 | if ((nsect == 1) && (lbal == 1)) | ||
552 | break; | ||
553 | msleep(50); /* give drive a breather */ | ||
554 | } | ||
555 | |||
556 | rc = ata_sff_wait_ready(link, deadline); | ||
557 | if (rc) { | ||
558 | if (rc != -ENODEV) | ||
559 | return rc; | ||
560 | ret = rc; | ||
561 | } | ||
541 | } | 562 | } |
542 | 563 | ||
543 | /* is all this really necessary? */ | 564 | /* is all this really necessary? */ |
544 | ap->ops->dev_select(ap, 0); | 565 | ap->ops->sff_dev_select(ap, 0); |
545 | if (dev1) | 566 | if (dev1) |
546 | ap->ops->dev_select(ap, 1); | 567 | ap->ops->sff_dev_select(ap, 1); |
547 | if (dev0) | 568 | if (dev0) |
548 | ap->ops->dev_select(ap, 0); | 569 | ap->ops->sff_dev_select(ap, 0); |
549 | 570 | ||
550 | return 0; | 571 | return ret; |
551 | } | 572 | } |
552 | 573 | ||
553 | /** | 574 | /** |
@@ -570,32 +591,22 @@ static unsigned int scc_bus_softreset(struct ata_port *ap, unsigned int devmask, | |||
570 | udelay(20); | 591 | udelay(20); |
571 | out_be32(ioaddr->ctl_addr, ap->ctl); | 592 | out_be32(ioaddr->ctl_addr, ap->ctl); |
572 | 593 | ||
573 | /* wait a while before checking status */ | 594 | scc_wait_after_reset(&ap->link, devmask, deadline); |
574 | ata_wait_after_reset(ap, deadline); | ||
575 | |||
576 | /* Before we perform post reset processing we want to see if | ||
577 | * the bus shows 0xFF because the odd clown forgets the D7 | ||
578 | * pulldown resistor. | ||
579 | */ | ||
580 | if (scc_check_status(ap) == 0xFF) | ||
581 | return 0; | ||
582 | |||
583 | scc_bus_post_reset(ap, devmask, deadline); | ||
584 | 595 | ||
585 | return 0; | 596 | return 0; |
586 | } | 597 | } |
587 | 598 | ||
588 | /** | 599 | /** |
589 | * scc_std_softreset - reset host port via ATA SRST | 600 | * scc_softreset - reset host port via ATA SRST |
590 | * @ap: port to reset | 601 | * @ap: port to reset |
591 | * @classes: resulting classes of attached devices | 602 | * @classes: resulting classes of attached devices |
592 | * @deadline: deadline jiffies for the operation | 603 | * @deadline: deadline jiffies for the operation |
593 | * | 604 | * |
594 | * Note: Original code is ata_std_softreset(). | 605 | * Note: Original code is ata_sff_softreset(). |
595 | */ | 606 | */ |
596 | 607 | ||
597 | static int scc_std_softreset(struct ata_link *link, unsigned int *classes, | 608 | static int scc_softreset(struct ata_link *link, unsigned int *classes, |
598 | unsigned long deadline) | 609 | unsigned long deadline) |
599 | { | 610 | { |
600 | struct ata_port *ap = link->ap; | 611 | struct ata_port *ap = link->ap; |
601 | unsigned int slave_possible = ap->flags & ATA_FLAG_SLAVE_POSS; | 612 | unsigned int slave_possible = ap->flags & ATA_FLAG_SLAVE_POSS; |
@@ -604,11 +615,6 @@ static int scc_std_softreset(struct ata_link *link, unsigned int *classes, | |||
604 | 615 | ||
605 | DPRINTK("ENTER\n"); | 616 | DPRINTK("ENTER\n"); |
606 | 617 | ||
607 | if (ata_link_offline(link)) { | ||
608 | classes[0] = ATA_DEV_NONE; | ||
609 | goto out; | ||
610 | } | ||
611 | |||
612 | /* determine if device 0/1 are present */ | 618 | /* determine if device 0/1 are present */ |
613 | if (scc_devchk(ap, 0)) | 619 | if (scc_devchk(ap, 0)) |
614 | devmask |= (1 << 0); | 620 | devmask |= (1 << 0); |
@@ -616,7 +622,7 @@ static int scc_std_softreset(struct ata_link *link, unsigned int *classes, | |||
616 | devmask |= (1 << 1); | 622 | devmask |= (1 << 1); |
617 | 623 | ||
618 | /* select device 0 again */ | 624 | /* select device 0 again */ |
619 | ap->ops->dev_select(ap, 0); | 625 | ap->ops->sff_dev_select(ap, 0); |
620 | 626 | ||
621 | /* issue bus reset */ | 627 | /* issue bus reset */ |
622 | DPRINTK("about to softreset, devmask=%x\n", devmask); | 628 | DPRINTK("about to softreset, devmask=%x\n", devmask); |
@@ -628,13 +634,12 @@ static int scc_std_softreset(struct ata_link *link, unsigned int *classes, | |||
628 | } | 634 | } |
629 | 635 | ||
630 | /* determine by signature whether we have ATA or ATAPI devices */ | 636 | /* determine by signature whether we have ATA or ATAPI devices */ |
631 | classes[0] = ata_dev_try_classify(&ap->link.device[0], | 637 | classes[0] = ata_sff_dev_classify(&ap->link.device[0], |
632 | devmask & (1 << 0), &err); | 638 | devmask & (1 << 0), &err); |
633 | if (slave_possible && err != 0x81) | 639 | if (slave_possible && err != 0x81) |
634 | classes[1] = ata_dev_try_classify(&ap->link.device[1], | 640 | classes[1] = ata_sff_dev_classify(&ap->link.device[1], |
635 | devmask & (1 << 1), &err); | 641 | devmask & (1 << 1), &err); |
636 | 642 | ||
637 | out: | ||
638 | DPRINTK("EXIT, classes[0]=%u [1]=%u\n", classes[0], classes[1]); | 643 | DPRINTK("EXIT, classes[0]=%u [1]=%u\n", classes[0], classes[1]); |
639 | return 0; | 644 | return 0; |
640 | } | 645 | } |
@@ -695,7 +700,7 @@ static void scc_bmdma_stop (struct ata_queued_cmd *qc) | |||
695 | printk(KERN_WARNING "%s: Internal Bus Error\n", DRV_NAME); | 700 | printk(KERN_WARNING "%s: Internal Bus Error\n", DRV_NAME); |
696 | out_be32(bmid_base + SCC_DMA_INTST, INTSTS_BMSINT); | 701 | out_be32(bmid_base + SCC_DMA_INTST, INTSTS_BMSINT); |
697 | /* TBD: SW reset */ | 702 | /* TBD: SW reset */ |
698 | scc_std_softreset(&ap->link, &classes, deadline); | 703 | scc_softreset(&ap->link, &classes, deadline); |
699 | continue; | 704 | continue; |
700 | } | 705 | } |
701 | 706 | ||
@@ -721,7 +726,7 @@ static void scc_bmdma_stop (struct ata_queued_cmd *qc) | |||
721 | in_be32(bmid_base + SCC_DMA_CMD) & ~ATA_DMA_START); | 726 | in_be32(bmid_base + SCC_DMA_CMD) & ~ATA_DMA_START); |
722 | 727 | ||
723 | /* one-PIO-cycle guaranteed wait, per spec, for HDMA1:0 transition */ | 728 | /* one-PIO-cycle guaranteed wait, per spec, for HDMA1:0 transition */ |
724 | ata_altstatus(ap); /* dummy read */ | 729 | ata_sff_altstatus(ap); /* dummy read */ |
725 | } | 730 | } |
726 | 731 | ||
727 | /** | 732 | /** |
@@ -742,7 +747,7 @@ static u8 scc_bmdma_status (struct ata_port *ap) | |||
742 | return host_stat; | 747 | return host_stat; |
743 | 748 | ||
744 | /* errata A252,A308 workaround: Step4 */ | 749 | /* errata A252,A308 workaround: Step4 */ |
745 | if ((ata_altstatus(ap) & ATA_ERR) && (int_status & INTSTS_INTRQ)) | 750 | if ((ata_sff_altstatus(ap) & ATA_ERR) && (int_status & INTSTS_INTRQ)) |
746 | return (host_stat | ATA_DMA_INTR); | 751 | return (host_stat | ATA_DMA_INTR); |
747 | 752 | ||
748 | /* errata A308 workaround Step5 */ | 753 | /* errata A308 workaround Step5 */ |
@@ -773,7 +778,7 @@ static u8 scc_bmdma_status (struct ata_port *ap) | |||
773 | * @buflen: buffer length | 778 | * @buflen: buffer length |
774 | * @rw: read/write | 779 | * @rw: read/write |
775 | * | 780 | * |
776 | * Note: Original code is ata_data_xfer(). | 781 | * Note: Original code is ata_sff_data_xfer(). |
777 | */ | 782 | */ |
778 | 783 | ||
779 | static unsigned int scc_data_xfer (struct ata_device *dev, unsigned char *buf, | 784 | static unsigned int scc_data_xfer (struct ata_device *dev, unsigned char *buf, |
@@ -782,28 +787,28 @@ static unsigned int scc_data_xfer (struct ata_device *dev, unsigned char *buf, | |||
782 | struct ata_port *ap = dev->link->ap; | 787 | struct ata_port *ap = dev->link->ap; |
783 | unsigned int words = buflen >> 1; | 788 | unsigned int words = buflen >> 1; |
784 | unsigned int i; | 789 | unsigned int i; |
785 | u16 *buf16 = (u16 *) buf; | 790 | __le16 *buf16 = (__le16 *) buf; |
786 | void __iomem *mmio = ap->ioaddr.data_addr; | 791 | void __iomem *mmio = ap->ioaddr.data_addr; |
787 | 792 | ||
788 | /* Transfer multiple of 2 bytes */ | 793 | /* Transfer multiple of 2 bytes */ |
789 | if (rw == READ) | 794 | if (rw == READ) |
790 | for (i = 0; i < words; i++) | 795 | for (i = 0; i < words; i++) |
791 | buf16[i] = le16_to_cpu(in_be32(mmio)); | 796 | buf16[i] = cpu_to_le16(in_be32(mmio)); |
792 | else | 797 | else |
793 | for (i = 0; i < words; i++) | 798 | for (i = 0; i < words; i++) |
794 | out_be32(mmio, cpu_to_le16(buf16[i])); | 799 | out_be32(mmio, le16_to_cpu(buf16[i])); |
795 | 800 | ||
796 | /* Transfer trailing 1 byte, if any. */ | 801 | /* Transfer trailing 1 byte, if any. */ |
797 | if (unlikely(buflen & 0x01)) { | 802 | if (unlikely(buflen & 0x01)) { |
798 | u16 align_buf[1] = { 0 }; | 803 | __le16 align_buf[1] = { 0 }; |
799 | unsigned char *trailing_buf = buf + buflen - 1; | 804 | unsigned char *trailing_buf = buf + buflen - 1; |
800 | 805 | ||
801 | if (rw == READ) { | 806 | if (rw == READ) { |
802 | align_buf[0] = le16_to_cpu(in_be32(mmio)); | 807 | align_buf[0] = cpu_to_le16(in_be32(mmio)); |
803 | memcpy(trailing_buf, align_buf, 1); | 808 | memcpy(trailing_buf, align_buf, 1); |
804 | } else { | 809 | } else { |
805 | memcpy(align_buf, trailing_buf, 1); | 810 | memcpy(align_buf, trailing_buf, 1); |
806 | out_be32(mmio, cpu_to_le16(align_buf[0])); | 811 | out_be32(mmio, le16_to_cpu(align_buf[0])); |
807 | } | 812 | } |
808 | words++; | 813 | words++; |
809 | } | 814 | } |
@@ -815,7 +820,7 @@ static unsigned int scc_data_xfer (struct ata_device *dev, unsigned char *buf, | |||
815 | * scc_irq_on - Enable interrupts on a port. | 820 | * scc_irq_on - Enable interrupts on a port. |
816 | * @ap: Port on which interrupts are enabled. | 821 | * @ap: Port on which interrupts are enabled. |
817 | * | 822 | * |
818 | * Note: Original code is ata_irq_on(). | 823 | * Note: Original code is ata_sff_irq_on(). |
819 | */ | 824 | */ |
820 | 825 | ||
821 | static u8 scc_irq_on (struct ata_port *ap) | 826 | static u8 scc_irq_on (struct ata_port *ap) |
@@ -829,19 +834,19 @@ static u8 scc_irq_on (struct ata_port *ap) | |||
829 | out_be32(ioaddr->ctl_addr, ap->ctl); | 834 | out_be32(ioaddr->ctl_addr, ap->ctl); |
830 | tmp = ata_wait_idle(ap); | 835 | tmp = ata_wait_idle(ap); |
831 | 836 | ||
832 | ap->ops->irq_clear(ap); | 837 | ap->ops->sff_irq_clear(ap); |
833 | 838 | ||
834 | return tmp; | 839 | return tmp; |
835 | } | 840 | } |
836 | 841 | ||
837 | /** | 842 | /** |
838 | * scc_bmdma_freeze - Freeze BMDMA controller port | 843 | * scc_freeze - Freeze BMDMA controller port |
839 | * @ap: port to freeze | 844 | * @ap: port to freeze |
840 | * | 845 | * |
841 | * Note: Original code is ata_bmdma_freeze(). | 846 | * Note: Original code is ata_sff_freeze(). |
842 | */ | 847 | */ |
843 | 848 | ||
844 | static void scc_bmdma_freeze (struct ata_port *ap) | 849 | static void scc_freeze (struct ata_port *ap) |
845 | { | 850 | { |
846 | struct ata_ioports *ioaddr = &ap->ioaddr; | 851 | struct ata_ioports *ioaddr = &ap->ioaddr; |
847 | 852 | ||
@@ -854,9 +859,9 @@ static void scc_bmdma_freeze (struct ata_port *ap) | |||
854 | * ATA_NIEN manipulation. Also, many controllers fail to mask | 859 | * ATA_NIEN manipulation. Also, many controllers fail to mask |
855 | * previously pending IRQ on ATA_NIEN assertion. Clear it. | 860 | * previously pending IRQ on ATA_NIEN assertion. Clear it. |
856 | */ | 861 | */ |
857 | ata_chk_status(ap); | 862 | ap->ops->sff_check_status(ap); |
858 | 863 | ||
859 | ap->ops->irq_clear(ap); | 864 | ap->ops->sff_irq_clear(ap); |
860 | } | 865 | } |
861 | 866 | ||
862 | /** | 867 | /** |
@@ -868,18 +873,18 @@ static void scc_bmdma_freeze (struct ata_port *ap) | |||
868 | static int scc_pata_prereset(struct ata_link *link, unsigned long deadline) | 873 | static int scc_pata_prereset(struct ata_link *link, unsigned long deadline) |
869 | { | 874 | { |
870 | link->ap->cbl = ATA_CBL_PATA80; | 875 | link->ap->cbl = ATA_CBL_PATA80; |
871 | return ata_std_prereset(link, deadline); | 876 | return ata_sff_prereset(link, deadline); |
872 | } | 877 | } |
873 | 878 | ||
874 | /** | 879 | /** |
875 | * scc_std_postreset - standard postreset callback | 880 | * scc_postreset - standard postreset callback |
876 | * @ap: the target ata_port | 881 | * @ap: the target ata_port |
877 | * @classes: classes of attached devices | 882 | * @classes: classes of attached devices |
878 | * | 883 | * |
879 | * Note: Original code is ata_std_postreset(). | 884 | * Note: Original code is ata_sff_postreset(). |
880 | */ | 885 | */ |
881 | 886 | ||
882 | static void scc_std_postreset(struct ata_link *link, unsigned int *classes) | 887 | static void scc_postreset(struct ata_link *link, unsigned int *classes) |
883 | { | 888 | { |
884 | struct ata_port *ap = link->ap; | 889 | struct ata_port *ap = link->ap; |
885 | 890 | ||
@@ -887,9 +892,9 @@ static void scc_std_postreset(struct ata_link *link, unsigned int *classes) | |||
887 | 892 | ||
888 | /* is double-select really necessary? */ | 893 | /* is double-select really necessary? */ |
889 | if (classes[0] != ATA_DEV_NONE) | 894 | if (classes[0] != ATA_DEV_NONE) |
890 | ap->ops->dev_select(ap, 1); | 895 | ap->ops->sff_dev_select(ap, 1); |
891 | if (classes[1] != ATA_DEV_NONE) | 896 | if (classes[1] != ATA_DEV_NONE) |
892 | ap->ops->dev_select(ap, 0); | 897 | ap->ops->sff_dev_select(ap, 0); |
893 | 898 | ||
894 | /* bail out if no device is present */ | 899 | /* bail out if no device is present */ |
895 | if (classes[0] == ATA_DEV_NONE && classes[1] == ATA_DEV_NONE) { | 900 | if (classes[0] == ATA_DEV_NONE && classes[1] == ATA_DEV_NONE) { |
@@ -905,24 +910,13 @@ static void scc_std_postreset(struct ata_link *link, unsigned int *classes) | |||
905 | } | 910 | } |
906 | 911 | ||
907 | /** | 912 | /** |
908 | * scc_error_handler - Stock error handler for BMDMA controller | 913 | * scc_irq_clear - Clear PCI IDE BMDMA interrupt. |
909 | * @ap: port to handle error for | ||
910 | */ | ||
911 | |||
912 | static void scc_error_handler (struct ata_port *ap) | ||
913 | { | ||
914 | ata_bmdma_drive_eh(ap, scc_pata_prereset, scc_std_softreset, NULL, | ||
915 | scc_std_postreset); | ||
916 | } | ||
917 | |||
918 | /** | ||
919 | * scc_bmdma_irq_clear - Clear PCI IDE BMDMA interrupt. | ||
920 | * @ap: Port associated with this ATA transaction. | 914 | * @ap: Port associated with this ATA transaction. |
921 | * | 915 | * |
922 | * Note: Original code is ata_bmdma_irq_clear(). | 916 | * Note: Original code is ata_sff_irq_clear(). |
923 | */ | 917 | */ |
924 | 918 | ||
925 | static void scc_bmdma_irq_clear (struct ata_port *ap) | 919 | static void scc_irq_clear (struct ata_port *ap) |
926 | { | 920 | { |
927 | void __iomem *mmio = ap->ioaddr.bmdma_addr; | 921 | void __iomem *mmio = ap->ioaddr.bmdma_addr; |
928 | 922 | ||
@@ -968,52 +962,37 @@ static void scc_port_stop (struct ata_port *ap) | |||
968 | } | 962 | } |
969 | 963 | ||
970 | static struct scsi_host_template scc_sht = { | 964 | static struct scsi_host_template scc_sht = { |
971 | .module = THIS_MODULE, | 965 | ATA_BMDMA_SHT(DRV_NAME), |
972 | .name = DRV_NAME, | ||
973 | .ioctl = ata_scsi_ioctl, | ||
974 | .queuecommand = ata_scsi_queuecmd, | ||
975 | .can_queue = ATA_DEF_QUEUE, | ||
976 | .this_id = ATA_SHT_THIS_ID, | ||
977 | .sg_tablesize = LIBATA_MAX_PRD, | ||
978 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
979 | .emulated = ATA_SHT_EMULATED, | ||
980 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
981 | .proc_name = DRV_NAME, | ||
982 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
983 | .slave_configure = ata_scsi_slave_config, | ||
984 | .slave_destroy = ata_scsi_slave_destroy, | ||
985 | .bios_param = ata_std_bios_param, | ||
986 | }; | 966 | }; |
987 | 967 | ||
988 | static const struct ata_port_operations scc_pata_ops = { | 968 | static struct ata_port_operations scc_pata_ops = { |
969 | .inherits = &ata_bmdma_port_ops, | ||
970 | |||
989 | .set_piomode = scc_set_piomode, | 971 | .set_piomode = scc_set_piomode, |
990 | .set_dmamode = scc_set_dmamode, | 972 | .set_dmamode = scc_set_dmamode, |
991 | .mode_filter = scc_mode_filter, | 973 | .mode_filter = scc_mode_filter, |
992 | 974 | ||
993 | .tf_load = scc_tf_load, | 975 | .sff_tf_load = scc_tf_load, |
994 | .tf_read = scc_tf_read, | 976 | .sff_tf_read = scc_tf_read, |
995 | .exec_command = scc_exec_command, | 977 | .sff_exec_command = scc_exec_command, |
996 | .check_status = scc_check_status, | 978 | .sff_check_status = scc_check_status, |
997 | .check_altstatus = scc_check_altstatus, | 979 | .sff_check_altstatus = scc_check_altstatus, |
998 | .dev_select = scc_std_dev_select, | 980 | .sff_dev_select = scc_dev_select, |
999 | 981 | ||
1000 | .bmdma_setup = scc_bmdma_setup, | 982 | .bmdma_setup = scc_bmdma_setup, |
1001 | .bmdma_start = scc_bmdma_start, | 983 | .bmdma_start = scc_bmdma_start, |
1002 | .bmdma_stop = scc_bmdma_stop, | 984 | .bmdma_stop = scc_bmdma_stop, |
1003 | .bmdma_status = scc_bmdma_status, | 985 | .bmdma_status = scc_bmdma_status, |
1004 | .data_xfer = scc_data_xfer, | 986 | .sff_data_xfer = scc_data_xfer, |
1005 | |||
1006 | .qc_prep = ata_qc_prep, | ||
1007 | .qc_issue = ata_qc_issue_prot, | ||
1008 | |||
1009 | .freeze = scc_bmdma_freeze, | ||
1010 | .thaw = ata_bmdma_thaw, | ||
1011 | 987 | ||
1012 | .error_handler = scc_error_handler, | 988 | .freeze = scc_freeze, |
989 | .prereset = scc_pata_prereset, | ||
990 | .softreset = scc_softreset, | ||
991 | .postreset = scc_postreset, | ||
1013 | .post_internal_cmd = scc_bmdma_stop, | 992 | .post_internal_cmd = scc_bmdma_stop, |
1014 | 993 | ||
1015 | .irq_clear = scc_bmdma_irq_clear, | 994 | .sff_irq_clear = scc_irq_clear, |
1016 | .irq_on = scc_irq_on, | 995 | .sff_irq_on = scc_irq_on, |
1017 | 996 | ||
1018 | .port_start = scc_port_start, | 997 | .port_start = scc_port_start, |
1019 | .port_stop = scc_port_stop, | 998 | .port_stop = scc_port_stop, |
@@ -1166,8 +1145,8 @@ static int scc_init_one (struct pci_dev *pdev, const struct pci_device_id *ent) | |||
1166 | if (rc) | 1145 | if (rc) |
1167 | return rc; | 1146 | return rc; |
1168 | 1147 | ||
1169 | return ata_host_activate(host, pdev->irq, ata_interrupt, IRQF_SHARED, | 1148 | return ata_host_activate(host, pdev->irq, ata_sff_interrupt, |
1170 | &scc_sht); | 1149 | IRQF_SHARED, &scc_sht); |
1171 | } | 1150 | } |
1172 | 1151 | ||
1173 | static struct pci_driver scc_pci_driver = { | 1152 | static struct pci_driver scc_pci_driver = { |
diff --git a/drivers/ata/pata_serverworks.c b/drivers/ata/pata_serverworks.c index a589c0fa0dbb..ffd26d0dc50d 100644 --- a/drivers/ata/pata_serverworks.c +++ b/drivers/ata/pata_serverworks.c | |||
@@ -199,7 +199,7 @@ static unsigned long serverworks_osb4_filter(struct ata_device *adev, unsigned l | |||
199 | { | 199 | { |
200 | if (adev->class == ATA_DEV_ATA) | 200 | if (adev->class == ATA_DEV_ATA) |
201 | mask &= ~ATA_MASK_UDMA; | 201 | mask &= ~ATA_MASK_UDMA; |
202 | return ata_pci_default_filter(adev, mask); | 202 | return ata_bmdma_mode_filter(adev, mask); |
203 | } | 203 | } |
204 | 204 | ||
205 | 205 | ||
@@ -219,7 +219,7 @@ static unsigned long serverworks_csb_filter(struct ata_device *adev, unsigned lo | |||
219 | 219 | ||
220 | /* Disk, UDMA */ | 220 | /* Disk, UDMA */ |
221 | if (adev->class != ATA_DEV_ATA) | 221 | if (adev->class != ATA_DEV_ATA) |
222 | return ata_pci_default_filter(adev, mask); | 222 | return ata_bmdma_mode_filter(adev, mask); |
223 | 223 | ||
224 | /* Actually do need to check */ | 224 | /* Actually do need to check */ |
225 | ata_id_c_string(adev->id, model_num, ATA_ID_PROD, sizeof(model_num)); | 225 | ata_id_c_string(adev->id, model_num, ATA_ID_PROD, sizeof(model_num)); |
@@ -228,7 +228,7 @@ static unsigned long serverworks_csb_filter(struct ata_device *adev, unsigned lo | |||
228 | if (!strcmp(p, model_num)) | 228 | if (!strcmp(p, model_num)) |
229 | mask &= ~(0xE0 << ATA_SHIFT_UDMA); | 229 | mask &= ~(0xE0 << ATA_SHIFT_UDMA); |
230 | } | 230 | } |
231 | return ata_pci_default_filter(adev, mask); | 231 | return ata_bmdma_mode_filter(adev, mask); |
232 | } | 232 | } |
233 | 233 | ||
234 | /** | 234 | /** |
@@ -298,89 +298,20 @@ static void serverworks_set_dmamode(struct ata_port *ap, struct ata_device *adev | |||
298 | } | 298 | } |
299 | 299 | ||
300 | static struct scsi_host_template serverworks_sht = { | 300 | static struct scsi_host_template serverworks_sht = { |
301 | .module = THIS_MODULE, | 301 | ATA_BMDMA_SHT(DRV_NAME), |
302 | .name = DRV_NAME, | ||
303 | .ioctl = ata_scsi_ioctl, | ||
304 | .queuecommand = ata_scsi_queuecmd, | ||
305 | .can_queue = ATA_DEF_QUEUE, | ||
306 | .this_id = ATA_SHT_THIS_ID, | ||
307 | .sg_tablesize = LIBATA_MAX_PRD, | ||
308 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
309 | .emulated = ATA_SHT_EMULATED, | ||
310 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
311 | .proc_name = DRV_NAME, | ||
312 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
313 | .slave_configure = ata_scsi_slave_config, | ||
314 | .slave_destroy = ata_scsi_slave_destroy, | ||
315 | .bios_param = ata_std_bios_param, | ||
316 | }; | 302 | }; |
317 | 303 | ||
318 | static struct ata_port_operations serverworks_osb4_port_ops = { | 304 | static struct ata_port_operations serverworks_osb4_port_ops = { |
305 | .inherits = &ata_bmdma_port_ops, | ||
306 | .cable_detect = serverworks_cable_detect, | ||
307 | .mode_filter = serverworks_osb4_filter, | ||
319 | .set_piomode = serverworks_set_piomode, | 308 | .set_piomode = serverworks_set_piomode, |
320 | .set_dmamode = serverworks_set_dmamode, | 309 | .set_dmamode = serverworks_set_dmamode, |
321 | .mode_filter = serverworks_osb4_filter, | ||
322 | |||
323 | .tf_load = ata_tf_load, | ||
324 | .tf_read = ata_tf_read, | ||
325 | .check_status = ata_check_status, | ||
326 | .exec_command = ata_exec_command, | ||
327 | .dev_select = ata_std_dev_select, | ||
328 | |||
329 | .freeze = ata_bmdma_freeze, | ||
330 | .thaw = ata_bmdma_thaw, | ||
331 | .error_handler = ata_bmdma_error_handler, | ||
332 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
333 | .cable_detect = serverworks_cable_detect, | ||
334 | |||
335 | .bmdma_setup = ata_bmdma_setup, | ||
336 | .bmdma_start = ata_bmdma_start, | ||
337 | .bmdma_stop = ata_bmdma_stop, | ||
338 | .bmdma_status = ata_bmdma_status, | ||
339 | |||
340 | .qc_prep = ata_qc_prep, | ||
341 | .qc_issue = ata_qc_issue_prot, | ||
342 | |||
343 | .data_xfer = ata_data_xfer, | ||
344 | |||
345 | .irq_handler = ata_interrupt, | ||
346 | .irq_clear = ata_bmdma_irq_clear, | ||
347 | .irq_on = ata_irq_on, | ||
348 | |||
349 | .port_start = ata_sff_port_start, | ||
350 | }; | 310 | }; |
351 | 311 | ||
352 | static struct ata_port_operations serverworks_csb_port_ops = { | 312 | static struct ata_port_operations serverworks_csb_port_ops = { |
353 | .set_piomode = serverworks_set_piomode, | 313 | .inherits = &serverworks_osb4_port_ops, |
354 | .set_dmamode = serverworks_set_dmamode, | ||
355 | .mode_filter = serverworks_csb_filter, | 314 | .mode_filter = serverworks_csb_filter, |
356 | |||
357 | .tf_load = ata_tf_load, | ||
358 | .tf_read = ata_tf_read, | ||
359 | .check_status = ata_check_status, | ||
360 | .exec_command = ata_exec_command, | ||
361 | .dev_select = ata_std_dev_select, | ||
362 | |||
363 | .freeze = ata_bmdma_freeze, | ||
364 | .thaw = ata_bmdma_thaw, | ||
365 | .error_handler = ata_bmdma_error_handler, | ||
366 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
367 | .cable_detect = serverworks_cable_detect, | ||
368 | |||
369 | .bmdma_setup = ata_bmdma_setup, | ||
370 | .bmdma_start = ata_bmdma_start, | ||
371 | .bmdma_stop = ata_bmdma_stop, | ||
372 | .bmdma_status = ata_bmdma_status, | ||
373 | |||
374 | .qc_prep = ata_qc_prep, | ||
375 | .qc_issue = ata_qc_issue_prot, | ||
376 | |||
377 | .data_xfer = ata_data_xfer, | ||
378 | |||
379 | .irq_handler = ata_interrupt, | ||
380 | .irq_clear = ata_bmdma_irq_clear, | ||
381 | .irq_on = ata_irq_on, | ||
382 | |||
383 | .port_start = ata_sff_port_start, | ||
384 | }; | 315 | }; |
385 | 316 | ||
386 | static int serverworks_fixup_osb4(struct pci_dev *pdev) | 317 | static int serverworks_fixup_osb4(struct pci_dev *pdev) |
@@ -468,28 +399,24 @@ static int serverworks_init_one(struct pci_dev *pdev, const struct pci_device_id | |||
468 | { | 399 | { |
469 | static const struct ata_port_info info[4] = { | 400 | static const struct ata_port_info info[4] = { |
470 | { /* OSB4 */ | 401 | { /* OSB4 */ |
471 | .sht = &serverworks_sht, | ||
472 | .flags = ATA_FLAG_SLAVE_POSS, | 402 | .flags = ATA_FLAG_SLAVE_POSS, |
473 | .pio_mask = 0x1f, | 403 | .pio_mask = 0x1f, |
474 | .mwdma_mask = 0x07, | 404 | .mwdma_mask = 0x07, |
475 | .udma_mask = 0x07, | 405 | .udma_mask = 0x07, |
476 | .port_ops = &serverworks_osb4_port_ops | 406 | .port_ops = &serverworks_osb4_port_ops |
477 | }, { /* OSB4 no UDMA */ | 407 | }, { /* OSB4 no UDMA */ |
478 | .sht = &serverworks_sht, | ||
479 | .flags = ATA_FLAG_SLAVE_POSS, | 408 | .flags = ATA_FLAG_SLAVE_POSS, |
480 | .pio_mask = 0x1f, | 409 | .pio_mask = 0x1f, |
481 | .mwdma_mask = 0x07, | 410 | .mwdma_mask = 0x07, |
482 | .udma_mask = 0x00, | 411 | .udma_mask = 0x00, |
483 | .port_ops = &serverworks_osb4_port_ops | 412 | .port_ops = &serverworks_osb4_port_ops |
484 | }, { /* CSB5 */ | 413 | }, { /* CSB5 */ |
485 | .sht = &serverworks_sht, | ||
486 | .flags = ATA_FLAG_SLAVE_POSS, | 414 | .flags = ATA_FLAG_SLAVE_POSS, |
487 | .pio_mask = 0x1f, | 415 | .pio_mask = 0x1f, |
488 | .mwdma_mask = 0x07, | 416 | .mwdma_mask = 0x07, |
489 | .udma_mask = ATA_UDMA4, | 417 | .udma_mask = ATA_UDMA4, |
490 | .port_ops = &serverworks_csb_port_ops | 418 | .port_ops = &serverworks_csb_port_ops |
491 | }, { /* CSB5 - later revisions*/ | 419 | }, { /* CSB5 - later revisions*/ |
492 | .sht = &serverworks_sht, | ||
493 | .flags = ATA_FLAG_SLAVE_POSS, | 420 | .flags = ATA_FLAG_SLAVE_POSS, |
494 | .pio_mask = 0x1f, | 421 | .pio_mask = 0x1f, |
495 | .mwdma_mask = 0x07, | 422 | .mwdma_mask = 0x07, |
@@ -498,6 +425,11 @@ static int serverworks_init_one(struct pci_dev *pdev, const struct pci_device_id | |||
498 | } | 425 | } |
499 | }; | 426 | }; |
500 | const struct ata_port_info *ppi[] = { &info[id->driver_data], NULL }; | 427 | const struct ata_port_info *ppi[] = { &info[id->driver_data], NULL }; |
428 | int rc; | ||
429 | |||
430 | rc = pcim_enable_device(pdev); | ||
431 | if (rc) | ||
432 | return rc; | ||
501 | 433 | ||
502 | /* Force master latency timer to 64 PCI clocks */ | 434 | /* Force master latency timer to 64 PCI clocks */ |
503 | pci_write_config_byte(pdev, PCI_LATENCY_TIMER, 0x40); | 435 | pci_write_config_byte(pdev, PCI_LATENCY_TIMER, 0x40); |
@@ -527,24 +459,30 @@ static int serverworks_init_one(struct pci_dev *pdev, const struct pci_device_id | |||
527 | serverworks_fixup_ht1000(pdev); | 459 | serverworks_fixup_ht1000(pdev); |
528 | 460 | ||
529 | if (pdev->device == PCI_DEVICE_ID_SERVERWORKS_CSB5IDE) | 461 | if (pdev->device == PCI_DEVICE_ID_SERVERWORKS_CSB5IDE) |
530 | ata_pci_clear_simplex(pdev); | 462 | ata_pci_bmdma_clear_simplex(pdev); |
531 | 463 | ||
532 | return ata_pci_init_one(pdev, ppi); | 464 | return ata_pci_sff_init_one(pdev, ppi, &serverworks_sht, NULL); |
533 | } | 465 | } |
534 | 466 | ||
535 | #ifdef CONFIG_PM | 467 | #ifdef CONFIG_PM |
536 | static int serverworks_reinit_one(struct pci_dev *pdev) | 468 | static int serverworks_reinit_one(struct pci_dev *pdev) |
537 | { | 469 | { |
470 | struct ata_host *host = dev_get_drvdata(&pdev->dev); | ||
471 | int rc; | ||
472 | |||
473 | rc = ata_pci_device_do_resume(pdev); | ||
474 | if (rc) | ||
475 | return rc; | ||
476 | |||
538 | /* Force master latency timer to 64 PCI clocks */ | 477 | /* Force master latency timer to 64 PCI clocks */ |
539 | pci_write_config_byte(pdev, PCI_LATENCY_TIMER, 0x40); | 478 | pci_write_config_byte(pdev, PCI_LATENCY_TIMER, 0x40); |
540 | 479 | ||
541 | switch (pdev->device) | 480 | switch (pdev->device) { |
542 | { | ||
543 | case PCI_DEVICE_ID_SERVERWORKS_OSB4IDE: | 481 | case PCI_DEVICE_ID_SERVERWORKS_OSB4IDE: |
544 | serverworks_fixup_osb4(pdev); | 482 | serverworks_fixup_osb4(pdev); |
545 | break; | 483 | break; |
546 | case PCI_DEVICE_ID_SERVERWORKS_CSB5IDE: | 484 | case PCI_DEVICE_ID_SERVERWORKS_CSB5IDE: |
547 | ata_pci_clear_simplex(pdev); | 485 | ata_pci_bmdma_clear_simplex(pdev); |
548 | /* fall through */ | 486 | /* fall through */ |
549 | case PCI_DEVICE_ID_SERVERWORKS_CSB6IDE: | 487 | case PCI_DEVICE_ID_SERVERWORKS_CSB6IDE: |
550 | case PCI_DEVICE_ID_SERVERWORKS_CSB6IDE2: | 488 | case PCI_DEVICE_ID_SERVERWORKS_CSB6IDE2: |
@@ -554,7 +492,9 @@ static int serverworks_reinit_one(struct pci_dev *pdev) | |||
554 | serverworks_fixup_ht1000(pdev); | 492 | serverworks_fixup_ht1000(pdev); |
555 | break; | 493 | break; |
556 | } | 494 | } |
557 | return ata_pci_device_resume(pdev); | 495 | |
496 | ata_host_resume(host); | ||
497 | return 0; | ||
558 | } | 498 | } |
559 | #endif | 499 | #endif |
560 | 500 | ||
diff --git a/drivers/ata/pata_sil680.c b/drivers/ata/pata_sil680.c index 7c5b2dd9a1a1..720b8645f58a 100644 --- a/drivers/ata/pata_sil680.c +++ b/drivers/ata/pata_sil680.c | |||
@@ -192,54 +192,14 @@ static void sil680_set_dmamode(struct ata_port *ap, struct ata_device *adev) | |||
192 | } | 192 | } |
193 | 193 | ||
194 | static struct scsi_host_template sil680_sht = { | 194 | static struct scsi_host_template sil680_sht = { |
195 | .module = THIS_MODULE, | 195 | ATA_BMDMA_SHT(DRV_NAME), |
196 | .name = DRV_NAME, | ||
197 | .ioctl = ata_scsi_ioctl, | ||
198 | .queuecommand = ata_scsi_queuecmd, | ||
199 | .can_queue = ATA_DEF_QUEUE, | ||
200 | .this_id = ATA_SHT_THIS_ID, | ||
201 | .sg_tablesize = LIBATA_MAX_PRD, | ||
202 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
203 | .emulated = ATA_SHT_EMULATED, | ||
204 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
205 | .proc_name = DRV_NAME, | ||
206 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
207 | .slave_configure = ata_scsi_slave_config, | ||
208 | .slave_destroy = ata_scsi_slave_destroy, | ||
209 | .bios_param = ata_std_bios_param, | ||
210 | }; | 196 | }; |
211 | 197 | ||
212 | static struct ata_port_operations sil680_port_ops = { | 198 | static struct ata_port_operations sil680_port_ops = { |
199 | .inherits = &ata_bmdma_port_ops, | ||
200 | .cable_detect = sil680_cable_detect, | ||
213 | .set_piomode = sil680_set_piomode, | 201 | .set_piomode = sil680_set_piomode, |
214 | .set_dmamode = sil680_set_dmamode, | 202 | .set_dmamode = sil680_set_dmamode, |
215 | .mode_filter = ata_pci_default_filter, | ||
216 | .tf_load = ata_tf_load, | ||
217 | .tf_read = ata_tf_read, | ||
218 | .check_status = ata_check_status, | ||
219 | .exec_command = ata_exec_command, | ||
220 | .dev_select = ata_std_dev_select, | ||
221 | |||
222 | .freeze = ata_bmdma_freeze, | ||
223 | .thaw = ata_bmdma_thaw, | ||
224 | .error_handler = ata_bmdma_error_handler, | ||
225 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
226 | .cable_detect = sil680_cable_detect, | ||
227 | |||
228 | .bmdma_setup = ata_bmdma_setup, | ||
229 | .bmdma_start = ata_bmdma_start, | ||
230 | .bmdma_stop = ata_bmdma_stop, | ||
231 | .bmdma_status = ata_bmdma_status, | ||
232 | |||
233 | .qc_prep = ata_qc_prep, | ||
234 | .qc_issue = ata_qc_issue_prot, | ||
235 | |||
236 | .data_xfer = ata_data_xfer, | ||
237 | |||
238 | .irq_handler = ata_interrupt, | ||
239 | .irq_clear = ata_bmdma_irq_clear, | ||
240 | .irq_on = ata_irq_on, | ||
241 | |||
242 | .port_start = ata_sff_port_start, | ||
243 | }; | 203 | }; |
244 | 204 | ||
245 | /** | 205 | /** |
@@ -322,7 +282,6 @@ static int __devinit sil680_init_one(struct pci_dev *pdev, | |||
322 | const struct pci_device_id *id) | 282 | const struct pci_device_id *id) |
323 | { | 283 | { |
324 | static const struct ata_port_info info = { | 284 | static const struct ata_port_info info = { |
325 | .sht = &sil680_sht, | ||
326 | .flags = ATA_FLAG_SLAVE_POSS, | 285 | .flags = ATA_FLAG_SLAVE_POSS, |
327 | .pio_mask = 0x1f, | 286 | .pio_mask = 0x1f, |
328 | .mwdma_mask = 0x07, | 287 | .mwdma_mask = 0x07, |
@@ -330,7 +289,6 @@ static int __devinit sil680_init_one(struct pci_dev *pdev, | |||
330 | .port_ops = &sil680_port_ops | 289 | .port_ops = &sil680_port_ops |
331 | }; | 290 | }; |
332 | static const struct ata_port_info info_slow = { | 291 | static const struct ata_port_info info_slow = { |
333 | .sht = &sil680_sht, | ||
334 | .flags = ATA_FLAG_SLAVE_POSS, | 292 | .flags = ATA_FLAG_SLAVE_POSS, |
335 | .pio_mask = 0x1f, | 293 | .pio_mask = 0x1f, |
336 | .mwdma_mask = 0x07, | 294 | .mwdma_mask = 0x07, |
@@ -346,6 +304,10 @@ static int __devinit sil680_init_one(struct pci_dev *pdev, | |||
346 | if (!printed_version++) | 304 | if (!printed_version++) |
347 | dev_printk(KERN_DEBUG, &pdev->dev, "version " DRV_VERSION "\n"); | 305 | dev_printk(KERN_DEBUG, &pdev->dev, "version " DRV_VERSION "\n"); |
348 | 306 | ||
307 | rc = pcim_enable_device(pdev); | ||
308 | if (rc) | ||
309 | return rc; | ||
310 | |||
349 | switch (sil680_init_chip(pdev, &try_mmio)) { | 311 | switch (sil680_init_chip(pdev, &try_mmio)) { |
350 | case 0: | 312 | case 0: |
351 | ppi[0] = &info_slow; | 313 | ppi[0] = &info_slow; |
@@ -388,28 +350,33 @@ static int __devinit sil680_init_one(struct pci_dev *pdev, | |||
388 | host->ports[0]->ioaddr.cmd_addr = mmio_base + 0x80; | 350 | host->ports[0]->ioaddr.cmd_addr = mmio_base + 0x80; |
389 | host->ports[0]->ioaddr.ctl_addr = mmio_base + 0x8a; | 351 | host->ports[0]->ioaddr.ctl_addr = mmio_base + 0x8a; |
390 | host->ports[0]->ioaddr.altstatus_addr = mmio_base + 0x8a; | 352 | host->ports[0]->ioaddr.altstatus_addr = mmio_base + 0x8a; |
391 | ata_std_ports(&host->ports[0]->ioaddr); | 353 | ata_sff_std_ports(&host->ports[0]->ioaddr); |
392 | host->ports[1]->ioaddr.bmdma_addr = mmio_base + 0x08; | 354 | host->ports[1]->ioaddr.bmdma_addr = mmio_base + 0x08; |
393 | host->ports[1]->ioaddr.cmd_addr = mmio_base + 0xc0; | 355 | host->ports[1]->ioaddr.cmd_addr = mmio_base + 0xc0; |
394 | host->ports[1]->ioaddr.ctl_addr = mmio_base + 0xca; | 356 | host->ports[1]->ioaddr.ctl_addr = mmio_base + 0xca; |
395 | host->ports[1]->ioaddr.altstatus_addr = mmio_base + 0xca; | 357 | host->ports[1]->ioaddr.altstatus_addr = mmio_base + 0xca; |
396 | ata_std_ports(&host->ports[1]->ioaddr); | 358 | ata_sff_std_ports(&host->ports[1]->ioaddr); |
397 | 359 | ||
398 | /* Register & activate */ | 360 | /* Register & activate */ |
399 | return ata_host_activate(host, pdev->irq, ata_interrupt, IRQF_SHARED, | 361 | return ata_host_activate(host, pdev->irq, ata_sff_interrupt, |
400 | &sil680_sht); | 362 | IRQF_SHARED, &sil680_sht); |
401 | 363 | ||
402 | use_ioports: | 364 | use_ioports: |
403 | return ata_pci_init_one(pdev, ppi); | 365 | return ata_pci_sff_init_one(pdev, ppi, &sil680_sht, NULL); |
404 | } | 366 | } |
405 | 367 | ||
406 | #ifdef CONFIG_PM | 368 | #ifdef CONFIG_PM |
407 | static int sil680_reinit_one(struct pci_dev *pdev) | 369 | static int sil680_reinit_one(struct pci_dev *pdev) |
408 | { | 370 | { |
409 | int try_mmio; | 371 | struct ata_host *host = dev_get_drvdata(&pdev->dev); |
372 | int try_mmio, rc; | ||
410 | 373 | ||
374 | rc = ata_pci_device_do_resume(pdev); | ||
375 | if (rc) | ||
376 | return rc; | ||
411 | sil680_init_chip(pdev, &try_mmio); | 377 | sil680_init_chip(pdev, &try_mmio); |
412 | return ata_pci_device_resume(pdev); | 378 | ata_host_resume(host); |
379 | return 0; | ||
413 | } | 380 | } |
414 | #endif | 381 | #endif |
415 | 382 | ||
diff --git a/drivers/ata/pata_sis.c b/drivers/ata/pata_sis.c index dc7e91562e43..e82c66e8d31b 100644 --- a/drivers/ata/pata_sis.c +++ b/drivers/ata/pata_sis.c | |||
@@ -156,24 +156,11 @@ static int sis_pre_reset(struct ata_link *link, unsigned long deadline) | |||
156 | /* Clear the FIFO settings. We can't enable the FIFO until | 156 | /* Clear the FIFO settings. We can't enable the FIFO until |
157 | we know we are poking at a disk */ | 157 | we know we are poking at a disk */ |
158 | pci_write_config_byte(pdev, 0x4B, 0); | 158 | pci_write_config_byte(pdev, 0x4B, 0); |
159 | return ata_std_prereset(link, deadline); | 159 | return ata_sff_prereset(link, deadline); |
160 | } | 160 | } |
161 | 161 | ||
162 | 162 | ||
163 | /** | 163 | /** |
164 | * sis_error_handler - Probe specified port on PATA host controller | ||
165 | * @ap: Port to probe | ||
166 | * | ||
167 | * LOCKING: | ||
168 | * None (inherited from caller). | ||
169 | */ | ||
170 | |||
171 | static void sis_error_handler(struct ata_port *ap) | ||
172 | { | ||
173 | ata_bmdma_drive_eh(ap, sis_pre_reset, ata_std_softreset, NULL, ata_std_postreset); | ||
174 | } | ||
175 | |||
176 | /** | ||
177 | * sis_set_fifo - Set RWP fifo bits for this device | 164 | * sis_set_fifo - Set RWP fifo bits for this device |
178 | * @ap: Port | 165 | * @ap: Port |
179 | * @adev: Device | 166 | * @adev: Device |
@@ -514,217 +501,57 @@ static void sis_133_set_dmamode (struct ata_port *ap, struct ata_device *adev) | |||
514 | } | 501 | } |
515 | 502 | ||
516 | static struct scsi_host_template sis_sht = { | 503 | static struct scsi_host_template sis_sht = { |
517 | .module = THIS_MODULE, | 504 | ATA_BMDMA_SHT(DRV_NAME), |
518 | .name = DRV_NAME, | ||
519 | .ioctl = ata_scsi_ioctl, | ||
520 | .queuecommand = ata_scsi_queuecmd, | ||
521 | .can_queue = ATA_DEF_QUEUE, | ||
522 | .this_id = ATA_SHT_THIS_ID, | ||
523 | .sg_tablesize = LIBATA_MAX_PRD, | ||
524 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
525 | .emulated = ATA_SHT_EMULATED, | ||
526 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
527 | .proc_name = DRV_NAME, | ||
528 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
529 | .slave_configure = ata_scsi_slave_config, | ||
530 | .slave_destroy = ata_scsi_slave_destroy, | ||
531 | .bios_param = ata_std_bios_param, | ||
532 | }; | 505 | }; |
533 | 506 | ||
534 | static const struct ata_port_operations sis_133_ops = { | 507 | static struct ata_port_operations sis_133_for_sata_ops = { |
508 | .inherits = &ata_bmdma_port_ops, | ||
535 | .set_piomode = sis_133_set_piomode, | 509 | .set_piomode = sis_133_set_piomode, |
536 | .set_dmamode = sis_133_set_dmamode, | 510 | .set_dmamode = sis_133_set_dmamode, |
537 | .mode_filter = ata_pci_default_filter, | ||
538 | |||
539 | .tf_load = ata_tf_load, | ||
540 | .tf_read = ata_tf_read, | ||
541 | .check_status = ata_check_status, | ||
542 | .exec_command = ata_exec_command, | ||
543 | .dev_select = ata_std_dev_select, | ||
544 | |||
545 | .freeze = ata_bmdma_freeze, | ||
546 | .thaw = ata_bmdma_thaw, | ||
547 | .error_handler = sis_error_handler, | ||
548 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
549 | .cable_detect = sis_133_cable_detect, | 511 | .cable_detect = sis_133_cable_detect, |
512 | }; | ||
550 | 513 | ||
551 | .bmdma_setup = ata_bmdma_setup, | 514 | static struct ata_port_operations sis_base_ops = { |
552 | .bmdma_start = ata_bmdma_start, | 515 | .inherits = &ata_bmdma_port_ops, |
553 | .bmdma_stop = ata_bmdma_stop, | 516 | .prereset = sis_pre_reset, |
554 | .bmdma_status = ata_bmdma_status, | ||
555 | .qc_prep = ata_qc_prep, | ||
556 | .qc_issue = ata_qc_issue_prot, | ||
557 | .data_xfer = ata_data_xfer, | ||
558 | |||
559 | .irq_handler = ata_interrupt, | ||
560 | .irq_clear = ata_bmdma_irq_clear, | ||
561 | .irq_on = ata_irq_on, | ||
562 | |||
563 | .port_start = ata_sff_port_start, | ||
564 | }; | 517 | }; |
565 | 518 | ||
566 | static const struct ata_port_operations sis_133_for_sata_ops = { | 519 | static struct ata_port_operations sis_133_ops = { |
520 | .inherits = &sis_base_ops, | ||
567 | .set_piomode = sis_133_set_piomode, | 521 | .set_piomode = sis_133_set_piomode, |
568 | .set_dmamode = sis_133_set_dmamode, | 522 | .set_dmamode = sis_133_set_dmamode, |
569 | .mode_filter = ata_pci_default_filter, | ||
570 | |||
571 | .tf_load = ata_tf_load, | ||
572 | .tf_read = ata_tf_read, | ||
573 | .check_status = ata_check_status, | ||
574 | .exec_command = ata_exec_command, | ||
575 | .dev_select = ata_std_dev_select, | ||
576 | |||
577 | .freeze = ata_bmdma_freeze, | ||
578 | .thaw = ata_bmdma_thaw, | ||
579 | .error_handler = ata_bmdma_error_handler, | ||
580 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
581 | .cable_detect = sis_133_cable_detect, | 523 | .cable_detect = sis_133_cable_detect, |
582 | |||
583 | .bmdma_setup = ata_bmdma_setup, | ||
584 | .bmdma_start = ata_bmdma_start, | ||
585 | .bmdma_stop = ata_bmdma_stop, | ||
586 | .bmdma_status = ata_bmdma_status, | ||
587 | .qc_prep = ata_qc_prep, | ||
588 | .qc_issue = ata_qc_issue_prot, | ||
589 | .data_xfer = ata_data_xfer, | ||
590 | |||
591 | .irq_handler = ata_interrupt, | ||
592 | .irq_clear = ata_bmdma_irq_clear, | ||
593 | .irq_on = ata_irq_on, | ||
594 | |||
595 | .port_start = ata_sff_port_start, | ||
596 | }; | 524 | }; |
597 | 525 | ||
598 | static const struct ata_port_operations sis_133_early_ops = { | 526 | static struct ata_port_operations sis_133_early_ops = { |
527 | .inherits = &sis_base_ops, | ||
599 | .set_piomode = sis_100_set_piomode, | 528 | .set_piomode = sis_100_set_piomode, |
600 | .set_dmamode = sis_133_early_set_dmamode, | 529 | .set_dmamode = sis_133_early_set_dmamode, |
601 | .mode_filter = ata_pci_default_filter, | ||
602 | |||
603 | .tf_load = ata_tf_load, | ||
604 | .tf_read = ata_tf_read, | ||
605 | .check_status = ata_check_status, | ||
606 | .exec_command = ata_exec_command, | ||
607 | .dev_select = ata_std_dev_select, | ||
608 | |||
609 | .freeze = ata_bmdma_freeze, | ||
610 | .thaw = ata_bmdma_thaw, | ||
611 | .error_handler = sis_error_handler, | ||
612 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
613 | .cable_detect = sis_66_cable_detect, | 530 | .cable_detect = sis_66_cable_detect, |
614 | |||
615 | .bmdma_setup = ata_bmdma_setup, | ||
616 | .bmdma_start = ata_bmdma_start, | ||
617 | .bmdma_stop = ata_bmdma_stop, | ||
618 | .bmdma_status = ata_bmdma_status, | ||
619 | .qc_prep = ata_qc_prep, | ||
620 | .qc_issue = ata_qc_issue_prot, | ||
621 | .data_xfer = ata_data_xfer, | ||
622 | |||
623 | .irq_handler = ata_interrupt, | ||
624 | .irq_clear = ata_bmdma_irq_clear, | ||
625 | .irq_on = ata_irq_on, | ||
626 | |||
627 | .port_start = ata_sff_port_start, | ||
628 | }; | 531 | }; |
629 | 532 | ||
630 | static const struct ata_port_operations sis_100_ops = { | 533 | static struct ata_port_operations sis_100_ops = { |
534 | .inherits = &sis_base_ops, | ||
631 | .set_piomode = sis_100_set_piomode, | 535 | .set_piomode = sis_100_set_piomode, |
632 | .set_dmamode = sis_100_set_dmamode, | 536 | .set_dmamode = sis_100_set_dmamode, |
633 | .mode_filter = ata_pci_default_filter, | ||
634 | |||
635 | .tf_load = ata_tf_load, | ||
636 | .tf_read = ata_tf_read, | ||
637 | .check_status = ata_check_status, | ||
638 | .exec_command = ata_exec_command, | ||
639 | .dev_select = ata_std_dev_select, | ||
640 | |||
641 | .freeze = ata_bmdma_freeze, | ||
642 | .thaw = ata_bmdma_thaw, | ||
643 | .error_handler = sis_error_handler, | ||
644 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
645 | .cable_detect = sis_66_cable_detect, | 537 | .cable_detect = sis_66_cable_detect, |
646 | |||
647 | .bmdma_setup = ata_bmdma_setup, | ||
648 | .bmdma_start = ata_bmdma_start, | ||
649 | .bmdma_stop = ata_bmdma_stop, | ||
650 | .bmdma_status = ata_bmdma_status, | ||
651 | .qc_prep = ata_qc_prep, | ||
652 | .qc_issue = ata_qc_issue_prot, | ||
653 | .data_xfer = ata_data_xfer, | ||
654 | |||
655 | .irq_handler = ata_interrupt, | ||
656 | .irq_clear = ata_bmdma_irq_clear, | ||
657 | .irq_on = ata_irq_on, | ||
658 | |||
659 | .port_start = ata_sff_port_start, | ||
660 | }; | 538 | }; |
661 | 539 | ||
662 | static const struct ata_port_operations sis_66_ops = { | 540 | static struct ata_port_operations sis_66_ops = { |
541 | .inherits = &sis_base_ops, | ||
663 | .set_piomode = sis_old_set_piomode, | 542 | .set_piomode = sis_old_set_piomode, |
664 | .set_dmamode = sis_66_set_dmamode, | 543 | .set_dmamode = sis_66_set_dmamode, |
665 | .mode_filter = ata_pci_default_filter, | ||
666 | |||
667 | .tf_load = ata_tf_load, | ||
668 | .tf_read = ata_tf_read, | ||
669 | .check_status = ata_check_status, | ||
670 | .exec_command = ata_exec_command, | ||
671 | .dev_select = ata_std_dev_select, | ||
672 | .cable_detect = sis_66_cable_detect, | 544 | .cable_detect = sis_66_cable_detect, |
673 | |||
674 | .freeze = ata_bmdma_freeze, | ||
675 | .thaw = ata_bmdma_thaw, | ||
676 | .error_handler = sis_error_handler, | ||
677 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
678 | |||
679 | .bmdma_setup = ata_bmdma_setup, | ||
680 | .bmdma_start = ata_bmdma_start, | ||
681 | .bmdma_stop = ata_bmdma_stop, | ||
682 | .bmdma_status = ata_bmdma_status, | ||
683 | .qc_prep = ata_qc_prep, | ||
684 | .qc_issue = ata_qc_issue_prot, | ||
685 | .data_xfer = ata_data_xfer, | ||
686 | |||
687 | .irq_handler = ata_interrupt, | ||
688 | .irq_clear = ata_bmdma_irq_clear, | ||
689 | .irq_on = ata_irq_on, | ||
690 | |||
691 | .port_start = ata_sff_port_start, | ||
692 | }; | 545 | }; |
693 | 546 | ||
694 | static const struct ata_port_operations sis_old_ops = { | 547 | static struct ata_port_operations sis_old_ops = { |
548 | .inherits = &sis_base_ops, | ||
695 | .set_piomode = sis_old_set_piomode, | 549 | .set_piomode = sis_old_set_piomode, |
696 | .set_dmamode = sis_old_set_dmamode, | 550 | .set_dmamode = sis_old_set_dmamode, |
697 | .mode_filter = ata_pci_default_filter, | ||
698 | |||
699 | .tf_load = ata_tf_load, | ||
700 | .tf_read = ata_tf_read, | ||
701 | .check_status = ata_check_status, | ||
702 | .exec_command = ata_exec_command, | ||
703 | .dev_select = ata_std_dev_select, | ||
704 | |||
705 | .freeze = ata_bmdma_freeze, | ||
706 | .thaw = ata_bmdma_thaw, | ||
707 | .error_handler = sis_error_handler, | ||
708 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
709 | .cable_detect = ata_cable_40wire, | 551 | .cable_detect = ata_cable_40wire, |
710 | |||
711 | .bmdma_setup = ata_bmdma_setup, | ||
712 | .bmdma_start = ata_bmdma_start, | ||
713 | .bmdma_stop = ata_bmdma_stop, | ||
714 | .bmdma_status = ata_bmdma_status, | ||
715 | .qc_prep = ata_qc_prep, | ||
716 | .qc_issue = ata_qc_issue_prot, | ||
717 | .data_xfer = ata_data_xfer, | ||
718 | |||
719 | .irq_handler = ata_interrupt, | ||
720 | .irq_clear = ata_bmdma_irq_clear, | ||
721 | .irq_on = ata_irq_on, | ||
722 | |||
723 | .port_start = ata_sff_port_start, | ||
724 | }; | 552 | }; |
725 | 553 | ||
726 | static const struct ata_port_info sis_info = { | 554 | static const struct ata_port_info sis_info = { |
727 | .sht = &sis_sht, | ||
728 | .flags = ATA_FLAG_SLAVE_POSS, | 555 | .flags = ATA_FLAG_SLAVE_POSS, |
729 | .pio_mask = 0x1f, /* pio0-4 */ | 556 | .pio_mask = 0x1f, /* pio0-4 */ |
730 | .mwdma_mask = 0x07, | 557 | .mwdma_mask = 0x07, |
@@ -732,7 +559,6 @@ static const struct ata_port_info sis_info = { | |||
732 | .port_ops = &sis_old_ops, | 559 | .port_ops = &sis_old_ops, |
733 | }; | 560 | }; |
734 | static const struct ata_port_info sis_info33 = { | 561 | static const struct ata_port_info sis_info33 = { |
735 | .sht = &sis_sht, | ||
736 | .flags = ATA_FLAG_SLAVE_POSS, | 562 | .flags = ATA_FLAG_SLAVE_POSS, |
737 | .pio_mask = 0x1f, /* pio0-4 */ | 563 | .pio_mask = 0x1f, /* pio0-4 */ |
738 | .mwdma_mask = 0x07, | 564 | .mwdma_mask = 0x07, |
@@ -740,42 +566,36 @@ static const struct ata_port_info sis_info33 = { | |||
740 | .port_ops = &sis_old_ops, | 566 | .port_ops = &sis_old_ops, |
741 | }; | 567 | }; |
742 | static const struct ata_port_info sis_info66 = { | 568 | static const struct ata_port_info sis_info66 = { |
743 | .sht = &sis_sht, | ||
744 | .flags = ATA_FLAG_SLAVE_POSS, | 569 | .flags = ATA_FLAG_SLAVE_POSS, |
745 | .pio_mask = 0x1f, /* pio0-4 */ | 570 | .pio_mask = 0x1f, /* pio0-4 */ |
746 | .udma_mask = ATA_UDMA4, /* UDMA 66 */ | 571 | .udma_mask = ATA_UDMA4, /* UDMA 66 */ |
747 | .port_ops = &sis_66_ops, | 572 | .port_ops = &sis_66_ops, |
748 | }; | 573 | }; |
749 | static const struct ata_port_info sis_info100 = { | 574 | static const struct ata_port_info sis_info100 = { |
750 | .sht = &sis_sht, | ||
751 | .flags = ATA_FLAG_SLAVE_POSS, | 575 | .flags = ATA_FLAG_SLAVE_POSS, |
752 | .pio_mask = 0x1f, /* pio0-4 */ | 576 | .pio_mask = 0x1f, /* pio0-4 */ |
753 | .udma_mask = ATA_UDMA5, | 577 | .udma_mask = ATA_UDMA5, |
754 | .port_ops = &sis_100_ops, | 578 | .port_ops = &sis_100_ops, |
755 | }; | 579 | }; |
756 | static const struct ata_port_info sis_info100_early = { | 580 | static const struct ata_port_info sis_info100_early = { |
757 | .sht = &sis_sht, | ||
758 | .flags = ATA_FLAG_SLAVE_POSS, | 581 | .flags = ATA_FLAG_SLAVE_POSS, |
759 | .udma_mask = ATA_UDMA5, | 582 | .udma_mask = ATA_UDMA5, |
760 | .pio_mask = 0x1f, /* pio0-4 */ | 583 | .pio_mask = 0x1f, /* pio0-4 */ |
761 | .port_ops = &sis_66_ops, | 584 | .port_ops = &sis_66_ops, |
762 | }; | 585 | }; |
763 | static const struct ata_port_info sis_info133 = { | 586 | static const struct ata_port_info sis_info133 = { |
764 | .sht = &sis_sht, | ||
765 | .flags = ATA_FLAG_SLAVE_POSS, | 587 | .flags = ATA_FLAG_SLAVE_POSS, |
766 | .pio_mask = 0x1f, /* pio0-4 */ | 588 | .pio_mask = 0x1f, /* pio0-4 */ |
767 | .udma_mask = ATA_UDMA6, | 589 | .udma_mask = ATA_UDMA6, |
768 | .port_ops = &sis_133_ops, | 590 | .port_ops = &sis_133_ops, |
769 | }; | 591 | }; |
770 | const struct ata_port_info sis_info133_for_sata = { | 592 | const struct ata_port_info sis_info133_for_sata = { |
771 | .sht = &sis_sht, | ||
772 | .flags = ATA_FLAG_SLAVE_POSS | ATA_FLAG_SRST, | 593 | .flags = ATA_FLAG_SLAVE_POSS | ATA_FLAG_SRST, |
773 | .pio_mask = 0x1f, /* pio0-4 */ | 594 | .pio_mask = 0x1f, /* pio0-4 */ |
774 | .udma_mask = ATA_UDMA6, | 595 | .udma_mask = ATA_UDMA6, |
775 | .port_ops = &sis_133_for_sata_ops, | 596 | .port_ops = &sis_133_for_sata_ops, |
776 | }; | 597 | }; |
777 | static const struct ata_port_info sis_info133_early = { | 598 | static const struct ata_port_info sis_info133_early = { |
778 | .sht = &sis_sht, | ||
779 | .flags = ATA_FLAG_SLAVE_POSS, | 599 | .flags = ATA_FLAG_SLAVE_POSS, |
780 | .pio_mask = 0x1f, /* pio0-4 */ | 600 | .pio_mask = 0x1f, /* pio0-4 */ |
781 | .udma_mask = ATA_UDMA6, | 601 | .udma_mask = ATA_UDMA6, |
@@ -857,11 +677,11 @@ static void sis_fixup(struct pci_dev *pdev, struct sis_chipset *sis) | |||
857 | static int sis_init_one (struct pci_dev *pdev, const struct pci_device_id *ent) | 677 | static int sis_init_one (struct pci_dev *pdev, const struct pci_device_id *ent) |
858 | { | 678 | { |
859 | static int printed_version; | 679 | static int printed_version; |
860 | struct ata_port_info port; | 680 | const struct ata_port_info *ppi[] = { NULL, NULL }; |
861 | const struct ata_port_info *ppi[] = { &port, NULL }; | ||
862 | struct pci_dev *host = NULL; | 681 | struct pci_dev *host = NULL; |
863 | struct sis_chipset *chipset = NULL; | 682 | struct sis_chipset *chipset = NULL; |
864 | struct sis_chipset *sets; | 683 | struct sis_chipset *sets; |
684 | int rc; | ||
865 | 685 | ||
866 | static struct sis_chipset sis_chipsets[] = { | 686 | static struct sis_chipset sis_chipsets[] = { |
867 | 687 | ||
@@ -914,8 +734,11 @@ static int sis_init_one (struct pci_dev *pdev, const struct pci_device_id *ent) | |||
914 | dev_printk(KERN_DEBUG, &pdev->dev, | 734 | dev_printk(KERN_DEBUG, &pdev->dev, |
915 | "version " DRV_VERSION "\n"); | 735 | "version " DRV_VERSION "\n"); |
916 | 736 | ||
917 | /* We have to find the bridge first */ | 737 | rc = pcim_enable_device(pdev); |
738 | if (rc) | ||
739 | return rc; | ||
918 | 740 | ||
741 | /* We have to find the bridge first */ | ||
919 | for (sets = &sis_chipsets[0]; sets->device; sets++) { | 742 | for (sets = &sis_chipsets[0]; sets->device; sets++) { |
920 | host = pci_get_device(PCI_VENDOR_ID_SI, sets->device, NULL); | 743 | host = pci_get_device(PCI_VENDOR_ID_SI, sets->device, NULL); |
921 | if (host != NULL) { | 744 | if (host != NULL) { |
@@ -994,12 +817,11 @@ static int sis_init_one (struct pci_dev *pdev, const struct pci_device_id *ent) | |||
994 | if (chipset == NULL) | 817 | if (chipset == NULL) |
995 | return -ENODEV; | 818 | return -ENODEV; |
996 | 819 | ||
997 | port = *chipset->info; | 820 | ppi[0] = chipset->info; |
998 | port.private_data = chipset; | ||
999 | 821 | ||
1000 | sis_fixup(pdev, chipset); | 822 | sis_fixup(pdev, chipset); |
1001 | 823 | ||
1002 | return ata_pci_init_one(pdev, ppi); | 824 | return ata_pci_sff_init_one(pdev, ppi, &sis_sht, chipset); |
1003 | } | 825 | } |
1004 | 826 | ||
1005 | static const struct pci_device_id sis_pci_tbl[] = { | 827 | static const struct pci_device_id sis_pci_tbl[] = { |
diff --git a/drivers/ata/pata_sl82c105.c b/drivers/ata/pata_sl82c105.c index 81ef207f8265..70d94fb28a5f 100644 --- a/drivers/ata/pata_sl82c105.c +++ b/drivers/ata/pata_sl82c105.c | |||
@@ -60,13 +60,7 @@ static int sl82c105_pre_reset(struct ata_link *link, unsigned long deadline) | |||
60 | 60 | ||
61 | if (ap->port_no && !pci_test_config_bits(pdev, &sl82c105_enable_bits[ap->port_no])) | 61 | if (ap->port_no && !pci_test_config_bits(pdev, &sl82c105_enable_bits[ap->port_no])) |
62 | return -ENOENT; | 62 | return -ENOENT; |
63 | return ata_std_prereset(link, deadline); | 63 | return ata_sff_prereset(link, deadline); |
64 | } | ||
65 | |||
66 | |||
67 | static void sl82c105_error_handler(struct ata_port *ap) | ||
68 | { | ||
69 | ata_bmdma_drive_eh(ap, sl82c105_pre_reset, ata_std_softreset, NULL, ata_std_postreset); | ||
70 | } | 64 | } |
71 | 65 | ||
72 | 66 | ||
@@ -235,55 +229,17 @@ static int sl82c105_qc_defer(struct ata_queued_cmd *qc) | |||
235 | } | 229 | } |
236 | 230 | ||
237 | static struct scsi_host_template sl82c105_sht = { | 231 | static struct scsi_host_template sl82c105_sht = { |
238 | .module = THIS_MODULE, | 232 | ATA_BMDMA_SHT(DRV_NAME), |
239 | .name = DRV_NAME, | ||
240 | .ioctl = ata_scsi_ioctl, | ||
241 | .queuecommand = ata_scsi_queuecmd, | ||
242 | .can_queue = ATA_DEF_QUEUE, | ||
243 | .this_id = ATA_SHT_THIS_ID, | ||
244 | .sg_tablesize = LIBATA_MAX_PRD, | ||
245 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
246 | .emulated = ATA_SHT_EMULATED, | ||
247 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
248 | .proc_name = DRV_NAME, | ||
249 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
250 | .slave_configure = ata_scsi_slave_config, | ||
251 | .slave_destroy = ata_scsi_slave_destroy, | ||
252 | .bios_param = ata_std_bios_param, | ||
253 | }; | 233 | }; |
254 | 234 | ||
255 | static struct ata_port_operations sl82c105_port_ops = { | 235 | static struct ata_port_operations sl82c105_port_ops = { |
256 | .set_piomode = sl82c105_set_piomode, | 236 | .inherits = &ata_bmdma_port_ops, |
257 | .mode_filter = ata_pci_default_filter, | 237 | .qc_defer = sl82c105_qc_defer, |
258 | |||
259 | .tf_load = ata_tf_load, | ||
260 | .tf_read = ata_tf_read, | ||
261 | .check_status = ata_check_status, | ||
262 | .exec_command = ata_exec_command, | ||
263 | .dev_select = ata_std_dev_select, | ||
264 | |||
265 | .freeze = ata_bmdma_freeze, | ||
266 | .thaw = ata_bmdma_thaw, | ||
267 | .error_handler = sl82c105_error_handler, | ||
268 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
269 | .cable_detect = ata_cable_40wire, | ||
270 | |||
271 | .bmdma_setup = ata_bmdma_setup, | ||
272 | .bmdma_start = sl82c105_bmdma_start, | 238 | .bmdma_start = sl82c105_bmdma_start, |
273 | .bmdma_stop = sl82c105_bmdma_stop, | 239 | .bmdma_stop = sl82c105_bmdma_stop, |
274 | .bmdma_status = ata_bmdma_status, | 240 | .cable_detect = ata_cable_40wire, |
275 | 241 | .set_piomode = sl82c105_set_piomode, | |
276 | .qc_defer = sl82c105_qc_defer, | 242 | .prereset = sl82c105_pre_reset, |
277 | .qc_prep = ata_qc_prep, | ||
278 | .qc_issue = ata_qc_issue_prot, | ||
279 | |||
280 | .data_xfer = ata_data_xfer, | ||
281 | |||
282 | .irq_handler = ata_interrupt, | ||
283 | .irq_clear = ata_bmdma_irq_clear, | ||
284 | .irq_on = ata_irq_on, | ||
285 | |||
286 | .port_start = ata_sff_port_start, | ||
287 | }; | 243 | }; |
288 | 244 | ||
289 | /** | 245 | /** |
@@ -327,14 +283,12 @@ static int sl82c105_bridge_revision(struct pci_dev *pdev) | |||
327 | static int sl82c105_init_one(struct pci_dev *dev, const struct pci_device_id *id) | 283 | static int sl82c105_init_one(struct pci_dev *dev, const struct pci_device_id *id) |
328 | { | 284 | { |
329 | static const struct ata_port_info info_dma = { | 285 | static const struct ata_port_info info_dma = { |
330 | .sht = &sl82c105_sht, | ||
331 | .flags = ATA_FLAG_SLAVE_POSS, | 286 | .flags = ATA_FLAG_SLAVE_POSS, |
332 | .pio_mask = 0x1f, | 287 | .pio_mask = 0x1f, |
333 | .mwdma_mask = 0x07, | 288 | .mwdma_mask = 0x07, |
334 | .port_ops = &sl82c105_port_ops | 289 | .port_ops = &sl82c105_port_ops |
335 | }; | 290 | }; |
336 | static const struct ata_port_info info_early = { | 291 | static const struct ata_port_info info_early = { |
337 | .sht = &sl82c105_sht, | ||
338 | .flags = ATA_FLAG_SLAVE_POSS, | 292 | .flags = ATA_FLAG_SLAVE_POSS, |
339 | .pio_mask = 0x1f, | 293 | .pio_mask = 0x1f, |
340 | .port_ops = &sl82c105_port_ops | 294 | .port_ops = &sl82c105_port_ops |
@@ -344,6 +298,11 @@ static int sl82c105_init_one(struct pci_dev *dev, const struct pci_device_id *id | |||
344 | NULL }; | 298 | NULL }; |
345 | u32 val; | 299 | u32 val; |
346 | int rev; | 300 | int rev; |
301 | int rc; | ||
302 | |||
303 | rc = pcim_enable_device(dev); | ||
304 | if (rc) | ||
305 | return rc; | ||
347 | 306 | ||
348 | rev = sl82c105_bridge_revision(dev); | 307 | rev = sl82c105_bridge_revision(dev); |
349 | 308 | ||
@@ -358,7 +317,7 @@ static int sl82c105_init_one(struct pci_dev *dev, const struct pci_device_id *id | |||
358 | val |= CTRL_P0EN | CTRL_P0F16 | CTRL_P1F16; | 317 | val |= CTRL_P0EN | CTRL_P0F16 | CTRL_P1F16; |
359 | pci_write_config_dword(dev, 0x40, val); | 318 | pci_write_config_dword(dev, 0x40, val); |
360 | 319 | ||
361 | return ata_pci_init_one(dev, ppi); | 320 | return ata_pci_sff_init_one(dev, ppi, &sl82c105_sht, NULL); |
362 | } | 321 | } |
363 | 322 | ||
364 | static const struct pci_device_id sl82c105[] = { | 323 | static const struct pci_device_id sl82c105[] = { |
diff --git a/drivers/ata/pata_triflex.c b/drivers/ata/pata_triflex.c index 403eafcffe12..b181261f2743 100644 --- a/drivers/ata/pata_triflex.c +++ b/drivers/ata/pata_triflex.c | |||
@@ -66,16 +66,11 @@ static int triflex_prereset(struct ata_link *link, unsigned long deadline) | |||
66 | if (!pci_test_config_bits(pdev, &triflex_enable_bits[ap->port_no])) | 66 | if (!pci_test_config_bits(pdev, &triflex_enable_bits[ap->port_no])) |
67 | return -ENOENT; | 67 | return -ENOENT; |
68 | 68 | ||
69 | return ata_std_prereset(link, deadline); | 69 | return ata_sff_prereset(link, deadline); |
70 | } | 70 | } |
71 | 71 | ||
72 | 72 | ||
73 | 73 | ||
74 | static void triflex_error_handler(struct ata_port *ap) | ||
75 | { | ||
76 | ata_bmdma_drive_eh(ap, triflex_prereset, ata_std_softreset, NULL, ata_std_postreset); | ||
77 | } | ||
78 | |||
79 | /** | 74 | /** |
80 | * triflex_load_timing - timing configuration | 75 | * triflex_load_timing - timing configuration |
81 | * @ap: ATA interface | 76 | * @ap: ATA interface |
@@ -180,60 +175,21 @@ static void triflex_bmdma_stop(struct ata_queued_cmd *qc) | |||
180 | } | 175 | } |
181 | 176 | ||
182 | static struct scsi_host_template triflex_sht = { | 177 | static struct scsi_host_template triflex_sht = { |
183 | .module = THIS_MODULE, | 178 | ATA_BMDMA_SHT(DRV_NAME), |
184 | .name = DRV_NAME, | ||
185 | .ioctl = ata_scsi_ioctl, | ||
186 | .queuecommand = ata_scsi_queuecmd, | ||
187 | .can_queue = ATA_DEF_QUEUE, | ||
188 | .this_id = ATA_SHT_THIS_ID, | ||
189 | .sg_tablesize = LIBATA_MAX_PRD, | ||
190 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
191 | .emulated = ATA_SHT_EMULATED, | ||
192 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
193 | .proc_name = DRV_NAME, | ||
194 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
195 | .slave_configure = ata_scsi_slave_config, | ||
196 | .slave_destroy = ata_scsi_slave_destroy, | ||
197 | .bios_param = ata_std_bios_param, | ||
198 | }; | 179 | }; |
199 | 180 | ||
200 | static struct ata_port_operations triflex_port_ops = { | 181 | static struct ata_port_operations triflex_port_ops = { |
201 | .set_piomode = triflex_set_piomode, | 182 | .inherits = &ata_bmdma_port_ops, |
202 | .mode_filter = ata_pci_default_filter, | ||
203 | |||
204 | .tf_load = ata_tf_load, | ||
205 | .tf_read = ata_tf_read, | ||
206 | .check_status = ata_check_status, | ||
207 | .exec_command = ata_exec_command, | ||
208 | .dev_select = ata_std_dev_select, | ||
209 | |||
210 | .freeze = ata_bmdma_freeze, | ||
211 | .thaw = ata_bmdma_thaw, | ||
212 | .error_handler = triflex_error_handler, | ||
213 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
214 | .cable_detect = ata_cable_40wire, | ||
215 | |||
216 | .bmdma_setup = ata_bmdma_setup, | ||
217 | .bmdma_start = triflex_bmdma_start, | 183 | .bmdma_start = triflex_bmdma_start, |
218 | .bmdma_stop = triflex_bmdma_stop, | 184 | .bmdma_stop = triflex_bmdma_stop, |
219 | .bmdma_status = ata_bmdma_status, | 185 | .cable_detect = ata_cable_40wire, |
220 | 186 | .set_piomode = triflex_set_piomode, | |
221 | .qc_prep = ata_qc_prep, | 187 | .prereset = triflex_prereset, |
222 | .qc_issue = ata_qc_issue_prot, | ||
223 | |||
224 | .data_xfer = ata_data_xfer, | ||
225 | |||
226 | .irq_handler = ata_interrupt, | ||
227 | .irq_clear = ata_bmdma_irq_clear, | ||
228 | .irq_on = ata_irq_on, | ||
229 | |||
230 | .port_start = ata_sff_port_start, | ||
231 | }; | 188 | }; |
232 | 189 | ||
233 | static int triflex_init_one(struct pci_dev *dev, const struct pci_device_id *id) | 190 | static int triflex_init_one(struct pci_dev *dev, const struct pci_device_id *id) |
234 | { | 191 | { |
235 | static const struct ata_port_info info = { | 192 | static const struct ata_port_info info = { |
236 | .sht = &triflex_sht, | ||
237 | .flags = ATA_FLAG_SLAVE_POSS, | 193 | .flags = ATA_FLAG_SLAVE_POSS, |
238 | .pio_mask = 0x1f, | 194 | .pio_mask = 0x1f, |
239 | .mwdma_mask = 0x07, | 195 | .mwdma_mask = 0x07, |
@@ -245,7 +201,7 @@ static int triflex_init_one(struct pci_dev *dev, const struct pci_device_id *id) | |||
245 | if (!printed_version++) | 201 | if (!printed_version++) |
246 | dev_printk(KERN_DEBUG, &dev->dev, "version " DRV_VERSION "\n"); | 202 | dev_printk(KERN_DEBUG, &dev->dev, "version " DRV_VERSION "\n"); |
247 | 203 | ||
248 | return ata_pci_init_one(dev, ppi); | 204 | return ata_pci_sff_init_one(dev, ppi, &triflex_sht, NULL); |
249 | } | 205 | } |
250 | 206 | ||
251 | static const struct pci_device_id triflex[] = { | 207 | static const struct pci_device_id triflex[] = { |
diff --git a/drivers/ata/pata_via.c b/drivers/ata/pata_via.c index d119a68c388f..d4840748fb5c 100644 --- a/drivers/ata/pata_via.c +++ b/drivers/ata/pata_via.c | |||
@@ -210,23 +210,11 @@ static int via_pre_reset(struct ata_link *link, unsigned long deadline) | |||
210 | return -ENOENT; | 210 | return -ENOENT; |
211 | } | 211 | } |
212 | 212 | ||
213 | return ata_std_prereset(link, deadline); | 213 | return ata_sff_prereset(link, deadline); |
214 | } | 214 | } |
215 | 215 | ||
216 | 216 | ||
217 | /** | 217 | /** |
218 | * via_error_handler - reset for VIA chips | ||
219 | * @ap: ATA port | ||
220 | * | ||
221 | * Handle the reset callback for the later chips with cable detect | ||
222 | */ | ||
223 | |||
224 | static void via_error_handler(struct ata_port *ap) | ||
225 | { | ||
226 | ata_bmdma_drive_eh(ap, via_pre_reset, ata_std_softreset, NULL, ata_std_postreset); | ||
227 | } | ||
228 | |||
229 | /** | ||
230 | * via_do_set_mode - set initial PIO mode data | 218 | * via_do_set_mode - set initial PIO mode data |
231 | * @ap: ATA interface | 219 | * @ap: ATA interface |
232 | * @adev: ATA device | 220 | * @adev: ATA device |
@@ -335,89 +323,20 @@ static void via_set_dmamode(struct ata_port *ap, struct ata_device *adev) | |||
335 | } | 323 | } |
336 | 324 | ||
337 | static struct scsi_host_template via_sht = { | 325 | static struct scsi_host_template via_sht = { |
338 | .module = THIS_MODULE, | 326 | ATA_BMDMA_SHT(DRV_NAME), |
339 | .name = DRV_NAME, | ||
340 | .ioctl = ata_scsi_ioctl, | ||
341 | .queuecommand = ata_scsi_queuecmd, | ||
342 | .can_queue = ATA_DEF_QUEUE, | ||
343 | .this_id = ATA_SHT_THIS_ID, | ||
344 | .sg_tablesize = LIBATA_MAX_PRD, | ||
345 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
346 | .emulated = ATA_SHT_EMULATED, | ||
347 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
348 | .proc_name = DRV_NAME, | ||
349 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
350 | .slave_configure = ata_scsi_slave_config, | ||
351 | .slave_destroy = ata_scsi_slave_destroy, | ||
352 | .bios_param = ata_std_bios_param, | ||
353 | }; | 327 | }; |
354 | 328 | ||
355 | static struct ata_port_operations via_port_ops = { | 329 | static struct ata_port_operations via_port_ops = { |
330 | .inherits = &ata_bmdma_port_ops, | ||
331 | .cable_detect = via_cable_detect, | ||
356 | .set_piomode = via_set_piomode, | 332 | .set_piomode = via_set_piomode, |
357 | .set_dmamode = via_set_dmamode, | 333 | .set_dmamode = via_set_dmamode, |
358 | .mode_filter = ata_pci_default_filter, | 334 | .prereset = via_pre_reset, |
359 | |||
360 | .tf_load = ata_tf_load, | ||
361 | .tf_read = ata_tf_read, | ||
362 | .check_status = ata_check_status, | ||
363 | .exec_command = ata_exec_command, | ||
364 | .dev_select = ata_std_dev_select, | ||
365 | |||
366 | .freeze = ata_bmdma_freeze, | ||
367 | .thaw = ata_bmdma_thaw, | ||
368 | .error_handler = via_error_handler, | ||
369 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
370 | .cable_detect = via_cable_detect, | ||
371 | |||
372 | .bmdma_setup = ata_bmdma_setup, | ||
373 | .bmdma_start = ata_bmdma_start, | ||
374 | .bmdma_stop = ata_bmdma_stop, | ||
375 | .bmdma_status = ata_bmdma_status, | ||
376 | |||
377 | .qc_prep = ata_qc_prep, | ||
378 | .qc_issue = ata_qc_issue_prot, | ||
379 | |||
380 | .data_xfer = ata_data_xfer, | ||
381 | |||
382 | .irq_handler = ata_interrupt, | ||
383 | .irq_clear = ata_bmdma_irq_clear, | ||
384 | .irq_on = ata_irq_on, | ||
385 | |||
386 | .port_start = ata_sff_port_start, | ||
387 | }; | 335 | }; |
388 | 336 | ||
389 | static struct ata_port_operations via_port_ops_noirq = { | 337 | static struct ata_port_operations via_port_ops_noirq = { |
390 | .set_piomode = via_set_piomode, | 338 | .inherits = &via_port_ops, |
391 | .set_dmamode = via_set_dmamode, | 339 | .sff_data_xfer = ata_sff_data_xfer_noirq, |
392 | .mode_filter = ata_pci_default_filter, | ||
393 | |||
394 | .tf_load = ata_tf_load, | ||
395 | .tf_read = ata_tf_read, | ||
396 | .check_status = ata_check_status, | ||
397 | .exec_command = ata_exec_command, | ||
398 | .dev_select = ata_std_dev_select, | ||
399 | |||
400 | .freeze = ata_bmdma_freeze, | ||
401 | .thaw = ata_bmdma_thaw, | ||
402 | .error_handler = via_error_handler, | ||
403 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
404 | .cable_detect = via_cable_detect, | ||
405 | |||
406 | .bmdma_setup = ata_bmdma_setup, | ||
407 | .bmdma_start = ata_bmdma_start, | ||
408 | .bmdma_stop = ata_bmdma_stop, | ||
409 | .bmdma_status = ata_bmdma_status, | ||
410 | |||
411 | .qc_prep = ata_qc_prep, | ||
412 | .qc_issue = ata_qc_issue_prot, | ||
413 | |||
414 | .data_xfer = ata_data_xfer_noirq, | ||
415 | |||
416 | .irq_handler = ata_interrupt, | ||
417 | .irq_clear = ata_bmdma_irq_clear, | ||
418 | .irq_on = ata_irq_on, | ||
419 | |||
420 | .port_start = ata_sff_port_start, | ||
421 | }; | 340 | }; |
422 | 341 | ||
423 | /** | 342 | /** |
@@ -467,7 +386,6 @@ static int via_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
467 | { | 386 | { |
468 | /* Early VIA without UDMA support */ | 387 | /* Early VIA without UDMA support */ |
469 | static const struct ata_port_info via_mwdma_info = { | 388 | static const struct ata_port_info via_mwdma_info = { |
470 | .sht = &via_sht, | ||
471 | .flags = ATA_FLAG_SLAVE_POSS, | 389 | .flags = ATA_FLAG_SLAVE_POSS, |
472 | .pio_mask = 0x1f, | 390 | .pio_mask = 0x1f, |
473 | .mwdma_mask = 0x07, | 391 | .mwdma_mask = 0x07, |
@@ -475,7 +393,6 @@ static int via_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
475 | }; | 393 | }; |
476 | /* Ditto with IRQ masking required */ | 394 | /* Ditto with IRQ masking required */ |
477 | static const struct ata_port_info via_mwdma_info_borked = { | 395 | static const struct ata_port_info via_mwdma_info_borked = { |
478 | .sht = &via_sht, | ||
479 | .flags = ATA_FLAG_SLAVE_POSS, | 396 | .flags = ATA_FLAG_SLAVE_POSS, |
480 | .pio_mask = 0x1f, | 397 | .pio_mask = 0x1f, |
481 | .mwdma_mask = 0x07, | 398 | .mwdma_mask = 0x07, |
@@ -483,7 +400,6 @@ static int via_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
483 | }; | 400 | }; |
484 | /* VIA UDMA 33 devices (and borked 66) */ | 401 | /* VIA UDMA 33 devices (and borked 66) */ |
485 | static const struct ata_port_info via_udma33_info = { | 402 | static const struct ata_port_info via_udma33_info = { |
486 | .sht = &via_sht, | ||
487 | .flags = ATA_FLAG_SLAVE_POSS, | 403 | .flags = ATA_FLAG_SLAVE_POSS, |
488 | .pio_mask = 0x1f, | 404 | .pio_mask = 0x1f, |
489 | .mwdma_mask = 0x07, | 405 | .mwdma_mask = 0x07, |
@@ -492,7 +408,6 @@ static int via_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
492 | }; | 408 | }; |
493 | /* VIA UDMA 66 devices */ | 409 | /* VIA UDMA 66 devices */ |
494 | static const struct ata_port_info via_udma66_info = { | 410 | static const struct ata_port_info via_udma66_info = { |
495 | .sht = &via_sht, | ||
496 | .flags = ATA_FLAG_SLAVE_POSS, | 411 | .flags = ATA_FLAG_SLAVE_POSS, |
497 | .pio_mask = 0x1f, | 412 | .pio_mask = 0x1f, |
498 | .mwdma_mask = 0x07, | 413 | .mwdma_mask = 0x07, |
@@ -501,7 +416,6 @@ static int via_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
501 | }; | 416 | }; |
502 | /* VIA UDMA 100 devices */ | 417 | /* VIA UDMA 100 devices */ |
503 | static const struct ata_port_info via_udma100_info = { | 418 | static const struct ata_port_info via_udma100_info = { |
504 | .sht = &via_sht, | ||
505 | .flags = ATA_FLAG_SLAVE_POSS, | 419 | .flags = ATA_FLAG_SLAVE_POSS, |
506 | .pio_mask = 0x1f, | 420 | .pio_mask = 0x1f, |
507 | .mwdma_mask = 0x07, | 421 | .mwdma_mask = 0x07, |
@@ -510,24 +424,27 @@ static int via_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
510 | }; | 424 | }; |
511 | /* UDMA133 with bad AST (All current 133) */ | 425 | /* UDMA133 with bad AST (All current 133) */ |
512 | static const struct ata_port_info via_udma133_info = { | 426 | static const struct ata_port_info via_udma133_info = { |
513 | .sht = &via_sht, | ||
514 | .flags = ATA_FLAG_SLAVE_POSS, | 427 | .flags = ATA_FLAG_SLAVE_POSS, |
515 | .pio_mask = 0x1f, | 428 | .pio_mask = 0x1f, |
516 | .mwdma_mask = 0x07, | 429 | .mwdma_mask = 0x07, |
517 | .udma_mask = ATA_UDMA6, /* FIXME: should check north bridge */ | 430 | .udma_mask = ATA_UDMA6, /* FIXME: should check north bridge */ |
518 | .port_ops = &via_port_ops | 431 | .port_ops = &via_port_ops |
519 | }; | 432 | }; |
520 | struct ata_port_info type; | 433 | const struct ata_port_info *ppi[] = { NULL, NULL }; |
521 | const struct ata_port_info *ppi[] = { &type, NULL }; | ||
522 | struct pci_dev *isa = NULL; | 434 | struct pci_dev *isa = NULL; |
523 | const struct via_isa_bridge *config; | 435 | const struct via_isa_bridge *config; |
524 | static int printed_version; | 436 | static int printed_version; |
525 | u8 enable; | 437 | u8 enable; |
526 | u32 timing; | 438 | u32 timing; |
439 | int rc; | ||
527 | 440 | ||
528 | if (!printed_version++) | 441 | if (!printed_version++) |
529 | dev_printk(KERN_DEBUG, &pdev->dev, "version " DRV_VERSION "\n"); | 442 | dev_printk(KERN_DEBUG, &pdev->dev, "version " DRV_VERSION "\n"); |
530 | 443 | ||
444 | rc = pcim_enable_device(pdev); | ||
445 | if (rc) | ||
446 | return rc; | ||
447 | |||
531 | /* To find out how the IDE will behave and what features we | 448 | /* To find out how the IDE will behave and what features we |
532 | actually have to look at the bridge not the IDE controller */ | 449 | actually have to look at the bridge not the IDE controller */ |
533 | for (config = via_isa_bridges; config->id; config++) | 450 | for (config = via_isa_bridges; config->id; config++) |
@@ -561,25 +478,25 @@ static int via_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
561 | switch(config->flags & VIA_UDMA) { | 478 | switch(config->flags & VIA_UDMA) { |
562 | case VIA_UDMA_NONE: | 479 | case VIA_UDMA_NONE: |
563 | if (config->flags & VIA_NO_UNMASK) | 480 | if (config->flags & VIA_NO_UNMASK) |
564 | type = via_mwdma_info_borked; | 481 | ppi[0] = &via_mwdma_info_borked; |
565 | else | 482 | else |
566 | type = via_mwdma_info; | 483 | ppi[0] = &via_mwdma_info; |
567 | break; | 484 | break; |
568 | case VIA_UDMA_33: | 485 | case VIA_UDMA_33: |
569 | type = via_udma33_info; | 486 | ppi[0] = &via_udma33_info; |
570 | break; | 487 | break; |
571 | case VIA_UDMA_66: | 488 | case VIA_UDMA_66: |
572 | type = via_udma66_info; | 489 | ppi[0] = &via_udma66_info; |
573 | /* The 66 MHz devices require we enable the clock */ | 490 | /* The 66 MHz devices require we enable the clock */ |
574 | pci_read_config_dword(pdev, 0x50, &timing); | 491 | pci_read_config_dword(pdev, 0x50, &timing); |
575 | timing |= 0x80008; | 492 | timing |= 0x80008; |
576 | pci_write_config_dword(pdev, 0x50, timing); | 493 | pci_write_config_dword(pdev, 0x50, timing); |
577 | break; | 494 | break; |
578 | case VIA_UDMA_100: | 495 | case VIA_UDMA_100: |
579 | type = via_udma100_info; | 496 | ppi[0] = &via_udma100_info; |
580 | break; | 497 | break; |
581 | case VIA_UDMA_133: | 498 | case VIA_UDMA_133: |
582 | type = via_udma133_info; | 499 | ppi[0] = &via_udma133_info; |
583 | break; | 500 | break; |
584 | default: | 501 | default: |
585 | WARN_ON(1); | 502 | WARN_ON(1); |
@@ -594,9 +511,7 @@ static int via_init_one(struct pci_dev *pdev, const struct pci_device_id *id) | |||
594 | } | 511 | } |
595 | 512 | ||
596 | /* We have established the device type, now fire it up */ | 513 | /* We have established the device type, now fire it up */ |
597 | type.private_data = (void *)config; | 514 | return ata_pci_sff_init_one(pdev, ppi, &via_sht, (void *)config); |
598 | |||
599 | return ata_pci_init_one(pdev, ppi); | ||
600 | } | 515 | } |
601 | 516 | ||
602 | #ifdef CONFIG_PM | 517 | #ifdef CONFIG_PM |
@@ -615,6 +530,11 @@ static int via_reinit_one(struct pci_dev *pdev) | |||
615 | u32 timing; | 530 | u32 timing; |
616 | struct ata_host *host = dev_get_drvdata(&pdev->dev); | 531 | struct ata_host *host = dev_get_drvdata(&pdev->dev); |
617 | const struct via_isa_bridge *config = host->private_data; | 532 | const struct via_isa_bridge *config = host->private_data; |
533 | int rc; | ||
534 | |||
535 | rc = ata_pci_device_do_resume(pdev); | ||
536 | if (rc) | ||
537 | return rc; | ||
618 | 538 | ||
619 | via_config_fifo(pdev, config->flags); | 539 | via_config_fifo(pdev, config->flags); |
620 | 540 | ||
@@ -630,7 +550,9 @@ static int via_reinit_one(struct pci_dev *pdev) | |||
630 | timing &= ~0x80008; | 550 | timing &= ~0x80008; |
631 | pci_write_config_dword(pdev, 0x50, timing); | 551 | pci_write_config_dword(pdev, 0x50, timing); |
632 | } | 552 | } |
633 | return ata_pci_device_resume(pdev); | 553 | |
554 | ata_host_resume(host); | ||
555 | return 0; | ||
634 | } | 556 | } |
635 | #endif | 557 | #endif |
636 | 558 | ||
diff --git a/drivers/ata/pata_winbond.c b/drivers/ata/pata_winbond.c index 99c92eda217b..6e52a3573fbf 100644 --- a/drivers/ata/pata_winbond.c +++ b/drivers/ata/pata_winbond.c | |||
@@ -116,53 +116,20 @@ static unsigned int winbond_data_xfer(struct ata_device *dev, | |||
116 | buflen += 4 - slop; | 116 | buflen += 4 - slop; |
117 | } | 117 | } |
118 | } else | 118 | } else |
119 | buflen = ata_data_xfer(dev, buf, buflen, rw); | 119 | buflen = ata_sff_data_xfer(dev, buf, buflen, rw); |
120 | 120 | ||
121 | return buflen; | 121 | return buflen; |
122 | } | 122 | } |
123 | 123 | ||
124 | static struct scsi_host_template winbond_sht = { | 124 | static struct scsi_host_template winbond_sht = { |
125 | .module = THIS_MODULE, | 125 | ATA_PIO_SHT(DRV_NAME), |
126 | .name = DRV_NAME, | ||
127 | .ioctl = ata_scsi_ioctl, | ||
128 | .queuecommand = ata_scsi_queuecmd, | ||
129 | .can_queue = ATA_DEF_QUEUE, | ||
130 | .this_id = ATA_SHT_THIS_ID, | ||
131 | .sg_tablesize = LIBATA_MAX_PRD, | ||
132 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
133 | .emulated = ATA_SHT_EMULATED, | ||
134 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
135 | .proc_name = DRV_NAME, | ||
136 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
137 | .slave_configure = ata_scsi_slave_config, | ||
138 | .slave_destroy = ata_scsi_slave_destroy, | ||
139 | .bios_param = ata_std_bios_param, | ||
140 | }; | 126 | }; |
141 | 127 | ||
142 | static struct ata_port_operations winbond_port_ops = { | 128 | static struct ata_port_operations winbond_port_ops = { |
143 | .set_piomode = winbond_set_piomode, | 129 | .inherits = &ata_sff_port_ops, |
144 | 130 | .sff_data_xfer = winbond_data_xfer, | |
145 | .tf_load = ata_tf_load, | ||
146 | .tf_read = ata_tf_read, | ||
147 | .check_status = ata_check_status, | ||
148 | .exec_command = ata_exec_command, | ||
149 | .dev_select = ata_std_dev_select, | ||
150 | |||
151 | .freeze = ata_bmdma_freeze, | ||
152 | .thaw = ata_bmdma_thaw, | ||
153 | .error_handler = ata_bmdma_error_handler, | ||
154 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
155 | .cable_detect = ata_cable_40wire, | 131 | .cable_detect = ata_cable_40wire, |
156 | 132 | .set_piomode = winbond_set_piomode, | |
157 | .qc_prep = ata_qc_prep, | ||
158 | .qc_issue = ata_qc_issue_prot, | ||
159 | |||
160 | .data_xfer = winbond_data_xfer, | ||
161 | |||
162 | .irq_clear = ata_bmdma_irq_clear, | ||
163 | .irq_on = ata_irq_on, | ||
164 | |||
165 | .port_start = ata_sff_port_start, | ||
166 | }; | 133 | }; |
167 | 134 | ||
168 | /** | 135 | /** |
@@ -231,7 +198,7 @@ static __init int winbond_init_one(unsigned long port) | |||
231 | ap->ioaddr.cmd_addr = cmd_addr; | 198 | ap->ioaddr.cmd_addr = cmd_addr; |
232 | ap->ioaddr.altstatus_addr = ctl_addr; | 199 | ap->ioaddr.altstatus_addr = ctl_addr; |
233 | ap->ioaddr.ctl_addr = ctl_addr; | 200 | ap->ioaddr.ctl_addr = ctl_addr; |
234 | ata_std_ports(&ap->ioaddr); | 201 | ata_sff_std_ports(&ap->ioaddr); |
235 | 202 | ||
236 | /* hook in a private data structure per channel */ | 203 | /* hook in a private data structure per channel */ |
237 | host->private_data = &winbond_data[nr_winbond_host]; | 204 | host->private_data = &winbond_data[nr_winbond_host]; |
@@ -239,7 +206,7 @@ static __init int winbond_init_one(unsigned long port) | |||
239 | winbond_data[nr_winbond_host].platform_dev = pdev; | 206 | winbond_data[nr_winbond_host].platform_dev = pdev; |
240 | 207 | ||
241 | /* activate */ | 208 | /* activate */ |
242 | rc = ata_host_activate(host, 14 + i, ata_interrupt, 0, | 209 | rc = ata_host_activate(host, 14 + i, ata_sff_interrupt, 0, |
243 | &winbond_sht); | 210 | &winbond_sht); |
244 | if (rc) | 211 | if (rc) |
245 | goto err_unregister; | 212 | goto err_unregister; |
diff --git a/drivers/ata/pdc_adma.c b/drivers/ata/pdc_adma.c index 8e1b7e9c0ae4..be53545c9f64 100644 --- a/drivers/ata/pdc_adma.c +++ b/drivers/ata/pdc_adma.c | |||
@@ -131,56 +131,33 @@ struct adma_port_priv { | |||
131 | static int adma_ata_init_one(struct pci_dev *pdev, | 131 | static int adma_ata_init_one(struct pci_dev *pdev, |
132 | const struct pci_device_id *ent); | 132 | const struct pci_device_id *ent); |
133 | static int adma_port_start(struct ata_port *ap); | 133 | static int adma_port_start(struct ata_port *ap); |
134 | static void adma_host_stop(struct ata_host *host); | ||
135 | static void adma_port_stop(struct ata_port *ap); | 134 | static void adma_port_stop(struct ata_port *ap); |
136 | static void adma_qc_prep(struct ata_queued_cmd *qc); | 135 | static void adma_qc_prep(struct ata_queued_cmd *qc); |
137 | static unsigned int adma_qc_issue(struct ata_queued_cmd *qc); | 136 | static unsigned int adma_qc_issue(struct ata_queued_cmd *qc); |
138 | static int adma_check_atapi_dma(struct ata_queued_cmd *qc); | 137 | static int adma_check_atapi_dma(struct ata_queued_cmd *qc); |
139 | static void adma_bmdma_stop(struct ata_queued_cmd *qc); | ||
140 | static u8 adma_bmdma_status(struct ata_port *ap); | ||
141 | static void adma_irq_clear(struct ata_port *ap); | ||
142 | static void adma_freeze(struct ata_port *ap); | 138 | static void adma_freeze(struct ata_port *ap); |
143 | static void adma_thaw(struct ata_port *ap); | 139 | static void adma_thaw(struct ata_port *ap); |
144 | static void adma_error_handler(struct ata_port *ap); | 140 | static int adma_prereset(struct ata_link *link, unsigned long deadline); |
145 | 141 | ||
146 | static struct scsi_host_template adma_ata_sht = { | 142 | static struct scsi_host_template adma_ata_sht = { |
147 | .module = THIS_MODULE, | 143 | ATA_BASE_SHT(DRV_NAME), |
148 | .name = DRV_NAME, | ||
149 | .ioctl = ata_scsi_ioctl, | ||
150 | .queuecommand = ata_scsi_queuecmd, | ||
151 | .slave_configure = ata_scsi_slave_config, | ||
152 | .slave_destroy = ata_scsi_slave_destroy, | ||
153 | .bios_param = ata_std_bios_param, | ||
154 | .proc_name = DRV_NAME, | ||
155 | .can_queue = ATA_DEF_QUEUE, | ||
156 | .this_id = ATA_SHT_THIS_ID, | ||
157 | .sg_tablesize = LIBATA_MAX_PRD, | 144 | .sg_tablesize = LIBATA_MAX_PRD, |
158 | .dma_boundary = ADMA_DMA_BOUNDARY, | 145 | .dma_boundary = ADMA_DMA_BOUNDARY, |
159 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
160 | .use_clustering = ENABLE_CLUSTERING, | ||
161 | .emulated = ATA_SHT_EMULATED, | ||
162 | }; | 146 | }; |
163 | 147 | ||
164 | static const struct ata_port_operations adma_ata_ops = { | 148 | static struct ata_port_operations adma_ata_ops = { |
165 | .tf_load = ata_tf_load, | 149 | .inherits = &ata_sff_port_ops, |
166 | .tf_read = ata_tf_read, | 150 | |
167 | .exec_command = ata_exec_command, | ||
168 | .check_status = ata_check_status, | ||
169 | .dev_select = ata_std_dev_select, | ||
170 | .check_atapi_dma = adma_check_atapi_dma, | 151 | .check_atapi_dma = adma_check_atapi_dma, |
171 | .data_xfer = ata_data_xfer, | ||
172 | .qc_prep = adma_qc_prep, | 152 | .qc_prep = adma_qc_prep, |
173 | .qc_issue = adma_qc_issue, | 153 | .qc_issue = adma_qc_issue, |
154 | |||
174 | .freeze = adma_freeze, | 155 | .freeze = adma_freeze, |
175 | .thaw = adma_thaw, | 156 | .thaw = adma_thaw, |
176 | .error_handler = adma_error_handler, | 157 | .prereset = adma_prereset, |
177 | .irq_clear = adma_irq_clear, | 158 | |
178 | .irq_on = ata_irq_on, | ||
179 | .port_start = adma_port_start, | 159 | .port_start = adma_port_start, |
180 | .port_stop = adma_port_stop, | 160 | .port_stop = adma_port_stop, |
181 | .host_stop = adma_host_stop, | ||
182 | .bmdma_stop = adma_bmdma_stop, | ||
183 | .bmdma_status = adma_bmdma_status, | ||
184 | }; | 161 | }; |
185 | 162 | ||
186 | static struct ata_port_info adma_port_info[] = { | 163 | static struct ata_port_info adma_port_info[] = { |
@@ -213,21 +190,6 @@ static int adma_check_atapi_dma(struct ata_queued_cmd *qc) | |||
213 | return 1; /* ATAPI DMA not yet supported */ | 190 | return 1; /* ATAPI DMA not yet supported */ |
214 | } | 191 | } |
215 | 192 | ||
216 | static void adma_bmdma_stop(struct ata_queued_cmd *qc) | ||
217 | { | ||
218 | /* nothing */ | ||
219 | } | ||
220 | |||
221 | static u8 adma_bmdma_status(struct ata_port *ap) | ||
222 | { | ||
223 | return 0; | ||
224 | } | ||
225 | |||
226 | static void adma_irq_clear(struct ata_port *ap) | ||
227 | { | ||
228 | /* nothing */ | ||
229 | } | ||
230 | |||
231 | static void adma_reset_engine(struct ata_port *ap) | 193 | static void adma_reset_engine(struct ata_port *ap) |
232 | { | 194 | { |
233 | void __iomem *chan = ADMA_PORT_REGS(ap); | 195 | void __iomem *chan = ADMA_PORT_REGS(ap); |
@@ -246,7 +208,7 @@ static void adma_reinit_engine(struct ata_port *ap) | |||
246 | 208 | ||
247 | /* mask/clear ATA interrupts */ | 209 | /* mask/clear ATA interrupts */ |
248 | writeb(ATA_NIEN, ap->ioaddr.ctl_addr); | 210 | writeb(ATA_NIEN, ap->ioaddr.ctl_addr); |
249 | ata_check_status(ap); | 211 | ata_sff_check_status(ap); |
250 | 212 | ||
251 | /* reset the ADMA engine */ | 213 | /* reset the ADMA engine */ |
252 | adma_reset_engine(ap); | 214 | adma_reset_engine(ap); |
@@ -281,7 +243,7 @@ static void adma_freeze(struct ata_port *ap) | |||
281 | 243 | ||
282 | /* mask/clear ATA interrupts */ | 244 | /* mask/clear ATA interrupts */ |
283 | writeb(ATA_NIEN, ap->ioaddr.ctl_addr); | 245 | writeb(ATA_NIEN, ap->ioaddr.ctl_addr); |
284 | ata_check_status(ap); | 246 | ata_sff_check_status(ap); |
285 | 247 | ||
286 | /* reset ADMA to idle state */ | 248 | /* reset ADMA to idle state */ |
287 | writew(aPIOMD4 | aNIEN | aRSTADM, chan + ADMA_CONTROL); | 249 | writew(aPIOMD4 | aNIEN | aRSTADM, chan + ADMA_CONTROL); |
@@ -304,13 +266,7 @@ static int adma_prereset(struct ata_link *link, unsigned long deadline) | |||
304 | pp->state = adma_state_mmio; | 266 | pp->state = adma_state_mmio; |
305 | adma_reinit_engine(ap); | 267 | adma_reinit_engine(ap); |
306 | 268 | ||
307 | return ata_std_prereset(link, deadline); | 269 | return ata_sff_prereset(link, deadline); |
308 | } | ||
309 | |||
310 | static void adma_error_handler(struct ata_port *ap) | ||
311 | { | ||
312 | ata_do_eh(ap, adma_prereset, ata_std_softreset, NULL, | ||
313 | ata_std_postreset); | ||
314 | } | 270 | } |
315 | 271 | ||
316 | static int adma_fill_sg(struct ata_queued_cmd *qc) | 272 | static int adma_fill_sg(struct ata_queued_cmd *qc) |
@@ -366,7 +322,7 @@ static void adma_qc_prep(struct ata_queued_cmd *qc) | |||
366 | 322 | ||
367 | adma_enter_reg_mode(qc->ap); | 323 | adma_enter_reg_mode(qc->ap); |
368 | if (qc->tf.protocol != ATA_PROT_DMA) { | 324 | if (qc->tf.protocol != ATA_PROT_DMA) { |
369 | ata_qc_prep(qc); | 325 | ata_sff_qc_prep(qc); |
370 | return; | 326 | return; |
371 | } | 327 | } |
372 | 328 | ||
@@ -465,7 +421,7 @@ static unsigned int adma_qc_issue(struct ata_queued_cmd *qc) | |||
465 | } | 421 | } |
466 | 422 | ||
467 | pp->state = adma_state_mmio; | 423 | pp->state = adma_state_mmio; |
468 | return ata_qc_issue_prot(qc); | 424 | return ata_sff_qc_issue(qc); |
469 | } | 425 | } |
470 | 426 | ||
471 | static inline unsigned int adma_intr_pkt(struct ata_host *host) | 427 | static inline unsigned int adma_intr_pkt(struct ata_host *host) |
@@ -536,7 +492,7 @@ static inline unsigned int adma_intr_mmio(struct ata_host *host) | |||
536 | if (qc && (!(qc->tf.flags & ATA_TFLAG_POLLING))) { | 492 | if (qc && (!(qc->tf.flags & ATA_TFLAG_POLLING))) { |
537 | 493 | ||
538 | /* check main status, clearing INTRQ */ | 494 | /* check main status, clearing INTRQ */ |
539 | u8 status = ata_check_status(ap); | 495 | u8 status = ata_sff_check_status(ap); |
540 | if ((status & ATA_BUSY)) | 496 | if ((status & ATA_BUSY)) |
541 | continue; | 497 | continue; |
542 | DPRINTK("ata%u: protocol %d (dev_stat 0x%X)\n", | 498 | DPRINTK("ata%u: protocol %d (dev_stat 0x%X)\n", |
@@ -633,14 +589,6 @@ static void adma_port_stop(struct ata_port *ap) | |||
633 | adma_reset_engine(ap); | 589 | adma_reset_engine(ap); |
634 | } | 590 | } |
635 | 591 | ||
636 | static void adma_host_stop(struct ata_host *host) | ||
637 | { | ||
638 | unsigned int port_no; | ||
639 | |||
640 | for (port_no = 0; port_no < ADMA_PORTS; ++port_no) | ||
641 | adma_reset_engine(host->ports[port_no]); | ||
642 | } | ||
643 | |||
644 | static void adma_host_init(struct ata_host *host, unsigned int chip_id) | 592 | static void adma_host_init(struct ata_host *host, unsigned int chip_id) |
645 | { | 593 | { |
646 | unsigned int port_no; | 594 | unsigned int port_no; |
diff --git a/drivers/ata/sata_fsl.c b/drivers/ata/sata_fsl.c index 9d1e3cad4aa9..fddd346b1d57 100644 --- a/drivers/ata/sata_fsl.c +++ b/drivers/ata/sata_fsl.c | |||
@@ -35,7 +35,6 @@ enum { | |||
35 | SATA_FSL_HOST_FLAGS = (ATA_FLAG_SATA | ATA_FLAG_NO_LEGACY | | 35 | SATA_FSL_HOST_FLAGS = (ATA_FLAG_SATA | ATA_FLAG_NO_LEGACY | |
36 | ATA_FLAG_MMIO | ATA_FLAG_PIO_DMA | | 36 | ATA_FLAG_MMIO | ATA_FLAG_PIO_DMA | |
37 | ATA_FLAG_NCQ), | 37 | ATA_FLAG_NCQ), |
38 | SATA_FSL_HOST_LFLAGS = ATA_LFLAG_SKIP_D2H_BSY, | ||
39 | 38 | ||
40 | SATA_FSL_MAX_CMDS = SATA_FSL_QUEUE_DEPTH, | 39 | SATA_FSL_MAX_CMDS = SATA_FSL_QUEUE_DEPTH, |
41 | SATA_FSL_CMD_HDR_SIZE = 16, /* 4 DWORDS */ | 40 | SATA_FSL_CMD_HDR_SIZE = 16, /* 4 DWORDS */ |
@@ -245,17 +244,6 @@ struct sata_fsl_port_priv { | |||
245 | dma_addr_t cmdslot_paddr; | 244 | dma_addr_t cmdslot_paddr; |
246 | struct command_desc *cmdentry; | 245 | struct command_desc *cmdentry; |
247 | dma_addr_t cmdentry_paddr; | 246 | dma_addr_t cmdentry_paddr; |
248 | |||
249 | /* | ||
250 | * SATA FSL controller has a Status FIS which should contain the | ||
251 | * received D2H FIS & taskfile registers. This SFIS is present in | ||
252 | * the command descriptor, and to have a ready reference to it, | ||
253 | * we are caching it here, quite similar to what is done in H/W on | ||
254 | * AHCI compliant devices by copying taskfile fields to a 32-bit | ||
255 | * register. | ||
256 | */ | ||
257 | |||
258 | struct ata_taskfile tf; | ||
259 | }; | 247 | }; |
260 | 248 | ||
261 | /* | 249 | /* |
@@ -465,6 +453,20 @@ static unsigned int sata_fsl_qc_issue(struct ata_queued_cmd *qc) | |||
465 | return 0; | 453 | return 0; |
466 | } | 454 | } |
467 | 455 | ||
456 | static bool sata_fsl_qc_fill_rtf(struct ata_queued_cmd *qc) | ||
457 | { | ||
458 | struct sata_fsl_port_priv *pp = qc->ap->private_data; | ||
459 | struct sata_fsl_host_priv *host_priv = qc->ap->host->private_data; | ||
460 | void __iomem *hcr_base = host_priv->hcr_base; | ||
461 | unsigned int tag = sata_fsl_tag(qc->tag, hcr_base); | ||
462 | struct command_desc *cd; | ||
463 | |||
464 | cd = pp->cmdentry + tag; | ||
465 | |||
466 | ata_tf_from_fis(cd->sfis, &qc->result_tf); | ||
467 | return true; | ||
468 | } | ||
469 | |||
468 | static int sata_fsl_scr_write(struct ata_port *ap, unsigned int sc_reg_in, | 470 | static int sata_fsl_scr_write(struct ata_port *ap, unsigned int sc_reg_in, |
469 | u32 val) | 471 | u32 val) |
470 | { | 472 | { |
@@ -556,38 +558,6 @@ static void sata_fsl_thaw(struct ata_port *ap) | |||
556 | ioread32(hcr_base + HCONTROL), ioread32(hcr_base + HSTATUS)); | 558 | ioread32(hcr_base + HCONTROL), ioread32(hcr_base + HSTATUS)); |
557 | } | 559 | } |
558 | 560 | ||
559 | /* | ||
560 | * NOTE : 1st D2H FIS from device does not update sfis in command descriptor. | ||
561 | */ | ||
562 | static inline void sata_fsl_cache_taskfile_from_d2h_fis(struct ata_queued_cmd | ||
563 | *qc, | ||
564 | struct ata_port *ap) | ||
565 | { | ||
566 | struct sata_fsl_port_priv *pp = ap->private_data; | ||
567 | struct sata_fsl_host_priv *host_priv = ap->host->private_data; | ||
568 | void __iomem *hcr_base = host_priv->hcr_base; | ||
569 | unsigned int tag = sata_fsl_tag(qc->tag, hcr_base); | ||
570 | struct command_desc *cd; | ||
571 | |||
572 | cd = pp->cmdentry + tag; | ||
573 | |||
574 | ata_tf_from_fis(cd->sfis, &pp->tf); | ||
575 | } | ||
576 | |||
577 | static u8 sata_fsl_check_status(struct ata_port *ap) | ||
578 | { | ||
579 | struct sata_fsl_port_priv *pp = ap->private_data; | ||
580 | |||
581 | return pp->tf.command; | ||
582 | } | ||
583 | |||
584 | static void sata_fsl_tf_read(struct ata_port *ap, struct ata_taskfile *tf) | ||
585 | { | ||
586 | struct sata_fsl_port_priv *pp = ap->private_data; | ||
587 | |||
588 | *tf = pp->tf; | ||
589 | } | ||
590 | |||
591 | static int sata_fsl_port_start(struct ata_port *ap) | 561 | static int sata_fsl_port_start(struct ata_port *ap) |
592 | { | 562 | { |
593 | struct device *dev = ap->host->dev; | 563 | struct device *dev = ap->host->dev; |
@@ -708,6 +678,15 @@ static unsigned int sata_fsl_dev_classify(struct ata_port *ap) | |||
708 | return ata_dev_classify(&tf); | 678 | return ata_dev_classify(&tf); |
709 | } | 679 | } |
710 | 680 | ||
681 | static int sata_fsl_prereset(struct ata_linke *link, unsigned long deadline) | ||
682 | { | ||
683 | /* FIXME: Never skip softreset, sata_fsl_softreset() is | ||
684 | * combination of soft and hard resets. sata_fsl_softreset() | ||
685 | * needs to be splitted into soft and hard resets. | ||
686 | */ | ||
687 | return 0; | ||
688 | } | ||
689 | |||
711 | static int sata_fsl_softreset(struct ata_link *link, unsigned int *class, | 690 | static int sata_fsl_softreset(struct ata_link *link, unsigned int *class, |
712 | unsigned long deadline) | 691 | unsigned long deadline) |
713 | { | 692 | { |
@@ -913,16 +892,6 @@ err: | |||
913 | return -EIO; | 892 | return -EIO; |
914 | } | 893 | } |
915 | 894 | ||
916 | static void sata_fsl_error_handler(struct ata_port *ap) | ||
917 | { | ||
918 | |||
919 | DPRINTK("in xx_error_handler\n"); | ||
920 | |||
921 | /* perform recovery */ | ||
922 | ata_do_eh(ap, ata_std_prereset, sata_fsl_softreset, sata_std_hardreset, | ||
923 | ata_std_postreset); | ||
924 | } | ||
925 | |||
926 | static void sata_fsl_post_internal_cmd(struct ata_queued_cmd *qc) | 895 | static void sata_fsl_post_internal_cmd(struct ata_queued_cmd *qc) |
927 | { | 896 | { |
928 | if (qc->flags & ATA_QCFLAG_FAILED) | 897 | if (qc->flags & ATA_QCFLAG_FAILED) |
@@ -934,11 +903,6 @@ static void sata_fsl_post_internal_cmd(struct ata_queued_cmd *qc) | |||
934 | } | 903 | } |
935 | } | 904 | } |
936 | 905 | ||
937 | static void sata_fsl_irq_clear(struct ata_port *ap) | ||
938 | { | ||
939 | /* unused */ | ||
940 | } | ||
941 | |||
942 | static void sata_fsl_error_intr(struct ata_port *ap) | 906 | static void sata_fsl_error_intr(struct ata_port *ap) |
943 | { | 907 | { |
944 | struct ata_link *link = &ap->link; | 908 | struct ata_link *link = &ap->link; |
@@ -996,7 +960,7 @@ static void sata_fsl_error_intr(struct ata_port *ap) | |||
996 | /* handle fatal errors */ | 960 | /* handle fatal errors */ |
997 | if (hstatus & FATAL_ERROR_DECODE) { | 961 | if (hstatus & FATAL_ERROR_DECODE) { |
998 | err_mask |= AC_ERR_ATA_BUS; | 962 | err_mask |= AC_ERR_ATA_BUS; |
999 | action |= ATA_EH_SOFTRESET; | 963 | action |= ATA_EH_RESET; |
1000 | /* how will fatal error interrupts be completed ?? */ | 964 | /* how will fatal error interrupts be completed ?? */ |
1001 | freeze = 1; | 965 | freeze = 1; |
1002 | } | 966 | } |
@@ -1013,10 +977,9 @@ static void sata_fsl_error_intr(struct ata_port *ap) | |||
1013 | /* record error info */ | 977 | /* record error info */ |
1014 | qc = ata_qc_from_tag(ap, link->active_tag); | 978 | qc = ata_qc_from_tag(ap, link->active_tag); |
1015 | 979 | ||
1016 | if (qc) { | 980 | if (qc) |
1017 | sata_fsl_cache_taskfile_from_d2h_fis(qc, qc->ap); | ||
1018 | qc->err_mask |= err_mask; | 981 | qc->err_mask |= err_mask; |
1019 | } else | 982 | else |
1020 | ehi->err_mask |= err_mask; | 983 | ehi->err_mask |= err_mask; |
1021 | 984 | ||
1022 | ehi->action |= action; | 985 | ehi->action |= action; |
@@ -1029,14 +992,6 @@ static void sata_fsl_error_intr(struct ata_port *ap) | |||
1029 | ata_port_abort(ap); | 992 | ata_port_abort(ap); |
1030 | } | 993 | } |
1031 | 994 | ||
1032 | static void sata_fsl_qc_complete(struct ata_queued_cmd *qc) | ||
1033 | { | ||
1034 | if (qc->flags & ATA_QCFLAG_RESULT_TF) { | ||
1035 | DPRINTK("xx_qc_complete called\n"); | ||
1036 | sata_fsl_cache_taskfile_from_d2h_fis(qc, qc->ap); | ||
1037 | } | ||
1038 | } | ||
1039 | |||
1040 | static void sata_fsl_host_intr(struct ata_port *ap) | 995 | static void sata_fsl_host_intr(struct ata_port *ap) |
1041 | { | 996 | { |
1042 | struct ata_link *link = &ap->link; | 997 | struct ata_link *link = &ap->link; |
@@ -1077,10 +1032,8 @@ static void sata_fsl_host_intr(struct ata_port *ap) | |||
1077 | for (i = 0; i < SATA_FSL_QUEUE_DEPTH; i++) { | 1032 | for (i = 0; i < SATA_FSL_QUEUE_DEPTH; i++) { |
1078 | if (qc_active & (1 << i)) { | 1033 | if (qc_active & (1 << i)) { |
1079 | qc = ata_qc_from_tag(ap, i); | 1034 | qc = ata_qc_from_tag(ap, i); |
1080 | if (qc) { | 1035 | if (qc) |
1081 | sata_fsl_qc_complete(qc); | ||
1082 | ata_qc_complete(qc); | 1036 | ata_qc_complete(qc); |
1083 | } | ||
1084 | DPRINTK | 1037 | DPRINTK |
1085 | ("completing ncq cmd,tag=%d,CC=0x%x,CA=0x%x\n", | 1038 | ("completing ncq cmd,tag=%d,CC=0x%x,CA=0x%x\n", |
1086 | i, ioread32(hcr_base + CC), | 1039 | i, ioread32(hcr_base + CC), |
@@ -1096,10 +1049,8 @@ static void sata_fsl_host_intr(struct ata_port *ap) | |||
1096 | DPRINTK("completing non-ncq cmd, tag=%d,CC=0x%x\n", | 1049 | DPRINTK("completing non-ncq cmd, tag=%d,CC=0x%x\n", |
1097 | link->active_tag, ioread32(hcr_base + CC)); | 1050 | link->active_tag, ioread32(hcr_base + CC)); |
1098 | 1051 | ||
1099 | if (qc) { | 1052 | if (qc) |
1100 | sata_fsl_qc_complete(qc); | ||
1101 | ata_qc_complete(qc); | 1053 | ata_qc_complete(qc); |
1102 | } | ||
1103 | } else { | 1054 | } else { |
1104 | /* Spurious Interrupt!! */ | 1055 | /* Spurious Interrupt!! */ |
1105 | DPRINTK("spurious interrupt!!, CC = 0x%x\n", | 1056 | DPRINTK("spurious interrupt!!, CC = 0x%x\n", |
@@ -1197,41 +1148,26 @@ static int sata_fsl_init_controller(struct ata_host *host) | |||
1197 | * scsi mid-layer and libata interface structures | 1148 | * scsi mid-layer and libata interface structures |
1198 | */ | 1149 | */ |
1199 | static struct scsi_host_template sata_fsl_sht = { | 1150 | static struct scsi_host_template sata_fsl_sht = { |
1200 | .module = THIS_MODULE, | 1151 | ATA_NCQ_SHT("sata_fsl"), |
1201 | .name = "sata_fsl", | ||
1202 | .ioctl = ata_scsi_ioctl, | ||
1203 | .queuecommand = ata_scsi_queuecmd, | ||
1204 | .change_queue_depth = ata_scsi_change_queue_depth, | ||
1205 | .can_queue = SATA_FSL_QUEUE_DEPTH, | 1152 | .can_queue = SATA_FSL_QUEUE_DEPTH, |
1206 | .this_id = ATA_SHT_THIS_ID, | ||
1207 | .sg_tablesize = SATA_FSL_MAX_PRD_USABLE, | 1153 | .sg_tablesize = SATA_FSL_MAX_PRD_USABLE, |
1208 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
1209 | .emulated = ATA_SHT_EMULATED, | ||
1210 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
1211 | .proc_name = "sata_fsl", | ||
1212 | .dma_boundary = ATA_DMA_BOUNDARY, | 1154 | .dma_boundary = ATA_DMA_BOUNDARY, |
1213 | .slave_configure = ata_scsi_slave_config, | ||
1214 | .slave_destroy = ata_scsi_slave_destroy, | ||
1215 | .bios_param = ata_std_bios_param, | ||
1216 | }; | 1155 | }; |
1217 | 1156 | ||
1218 | static const struct ata_port_operations sata_fsl_ops = { | 1157 | static const struct ata_port_operations sata_fsl_ops = { |
1219 | .check_status = sata_fsl_check_status, | 1158 | .inherits = &sata_port_ops, |
1220 | .check_altstatus = sata_fsl_check_status, | ||
1221 | .dev_select = ata_noop_dev_select, | ||
1222 | |||
1223 | .tf_read = sata_fsl_tf_read, | ||
1224 | 1159 | ||
1225 | .qc_prep = sata_fsl_qc_prep, | 1160 | .qc_prep = sata_fsl_qc_prep, |
1226 | .qc_issue = sata_fsl_qc_issue, | 1161 | .qc_issue = sata_fsl_qc_issue, |
1227 | .irq_clear = sata_fsl_irq_clear, | 1162 | .qc_fill_rtf = sata_fsl_qc_fill_rtf, |
1228 | 1163 | ||
1229 | .scr_read = sata_fsl_scr_read, | 1164 | .scr_read = sata_fsl_scr_read, |
1230 | .scr_write = sata_fsl_scr_write, | 1165 | .scr_write = sata_fsl_scr_write, |
1231 | 1166 | ||
1232 | .freeze = sata_fsl_freeze, | 1167 | .freeze = sata_fsl_freeze, |
1233 | .thaw = sata_fsl_thaw, | 1168 | .thaw = sata_fsl_thaw, |
1234 | .error_handler = sata_fsl_error_handler, | 1169 | .prereset = sata_fsl_prereset, |
1170 | .softreset = sata_fsl_softreset, | ||
1235 | .post_internal_cmd = sata_fsl_post_internal_cmd, | 1171 | .post_internal_cmd = sata_fsl_post_internal_cmd, |
1236 | 1172 | ||
1237 | .port_start = sata_fsl_port_start, | 1173 | .port_start = sata_fsl_port_start, |
@@ -1241,7 +1177,6 @@ static const struct ata_port_operations sata_fsl_ops = { | |||
1241 | static const struct ata_port_info sata_fsl_port_info[] = { | 1177 | static const struct ata_port_info sata_fsl_port_info[] = { |
1242 | { | 1178 | { |
1243 | .flags = SATA_FSL_HOST_FLAGS, | 1179 | .flags = SATA_FSL_HOST_FLAGS, |
1244 | .link_flags = SATA_FSL_HOST_LFLAGS, | ||
1245 | .pio_mask = 0x1f, /* pio 0-4 */ | 1180 | .pio_mask = 0x1f, /* pio 0-4 */ |
1246 | .udma_mask = 0x7f, /* udma 0-6 */ | 1181 | .udma_mask = 0x7f, /* udma 0-6 */ |
1247 | .port_ops = &sata_fsl_ops, | 1182 | .port_ops = &sata_fsl_ops, |
@@ -1297,11 +1232,6 @@ static int sata_fsl_probe(struct of_device *ofdev, | |||
1297 | /* host->iomap is not used currently */ | 1232 | /* host->iomap is not used currently */ |
1298 | host->private_data = host_priv; | 1233 | host->private_data = host_priv; |
1299 | 1234 | ||
1300 | /* setup port(s) */ | ||
1301 | |||
1302 | host->ports[0]->ioaddr.cmd_addr = host_priv->hcr_base; | ||
1303 | host->ports[0]->ioaddr.scr_addr = host_priv->ssr_base; | ||
1304 | |||
1305 | /* initialize host controller */ | 1235 | /* initialize host controller */ |
1306 | sata_fsl_init_controller(host); | 1236 | sata_fsl_init_controller(host); |
1307 | 1237 | ||
diff --git a/drivers/ata/sata_inic162x.c b/drivers/ata/sata_inic162x.c index 59e65edc5820..d27bb9a2568f 100644 --- a/drivers/ata/sata_inic162x.c +++ b/drivers/ata/sata_inic162x.c | |||
@@ -109,21 +109,7 @@ struct inic_port_priv { | |||
109 | }; | 109 | }; |
110 | 110 | ||
111 | static struct scsi_host_template inic_sht = { | 111 | static struct scsi_host_template inic_sht = { |
112 | .module = THIS_MODULE, | 112 | ATA_BMDMA_SHT(DRV_NAME), |
113 | .name = DRV_NAME, | ||
114 | .ioctl = ata_scsi_ioctl, | ||
115 | .queuecommand = ata_scsi_queuecmd, | ||
116 | .can_queue = ATA_DEF_QUEUE, | ||
117 | .this_id = ATA_SHT_THIS_ID, | ||
118 | .sg_tablesize = LIBATA_MAX_PRD, | ||
119 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
120 | .emulated = ATA_SHT_EMULATED, | ||
121 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
122 | .proc_name = DRV_NAME, | ||
123 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
124 | .slave_configure = ata_scsi_slave_config, | ||
125 | .slave_destroy = ata_scsi_slave_destroy, | ||
126 | .bios_param = ata_std_bios_param, | ||
127 | }; | 113 | }; |
128 | 114 | ||
129 | static const int scr_map[] = { | 115 | static const int scr_map[] = { |
@@ -236,7 +222,7 @@ static void inic_bmdma_setup(struct ata_queued_cmd *qc) | |||
236 | writeb(pp->cached_prdctl, port_base + PORT_PRD_CTL); | 222 | writeb(pp->cached_prdctl, port_base + PORT_PRD_CTL); |
237 | 223 | ||
238 | /* issue r/w command */ | 224 | /* issue r/w command */ |
239 | ap->ops->exec_command(ap, &qc->tf); | 225 | ap->ops->sff_exec_command(ap, &qc->tf); |
240 | } | 226 | } |
241 | 227 | ||
242 | static void inic_bmdma_start(struct ata_queued_cmd *qc) | 228 | static void inic_bmdma_start(struct ata_queued_cmd *qc) |
@@ -266,11 +252,6 @@ static u8 inic_bmdma_status(struct ata_port *ap) | |||
266 | return ATA_DMA_INTR; | 252 | return ATA_DMA_INTR; |
267 | } | 253 | } |
268 | 254 | ||
269 | static void inic_irq_clear(struct ata_port *ap) | ||
270 | { | ||
271 | /* noop */ | ||
272 | } | ||
273 | |||
274 | static void inic_host_intr(struct ata_port *ap) | 255 | static void inic_host_intr(struct ata_port *ap) |
275 | { | 256 | { |
276 | void __iomem *port_base = inic_port_base(ap); | 257 | void __iomem *port_base = inic_port_base(ap); |
@@ -286,14 +267,14 @@ static void inic_host_intr(struct ata_port *ap) | |||
286 | ata_qc_from_tag(ap, ap->link.active_tag); | 267 | ata_qc_from_tag(ap, ap->link.active_tag); |
287 | 268 | ||
288 | if (unlikely(!qc || (qc->tf.flags & ATA_TFLAG_POLLING))) { | 269 | if (unlikely(!qc || (qc->tf.flags & ATA_TFLAG_POLLING))) { |
289 | ata_chk_status(ap); /* clear ATA interrupt */ | 270 | ap->ops->sff_check_status(ap); /* clear ATA interrupt */ |
290 | return; | 271 | return; |
291 | } | 272 | } |
292 | 273 | ||
293 | if (likely(ata_host_intr(ap, qc))) | 274 | if (likely(ata_sff_host_intr(ap, qc))) |
294 | return; | 275 | return; |
295 | 276 | ||
296 | ata_chk_status(ap); /* clear ATA interrupt */ | 277 | ap->ops->sff_check_status(ap); /* clear ATA interrupt */ |
297 | ata_port_printk(ap, KERN_WARNING, "unhandled " | 278 | ata_port_printk(ap, KERN_WARNING, "unhandled " |
298 | "interrupt, irq_stat=%x\n", irq_stat); | 279 | "interrupt, irq_stat=%x\n", irq_stat); |
299 | return; | 280 | return; |
@@ -370,12 +351,12 @@ static unsigned int inic_qc_issue(struct ata_queued_cmd *qc) | |||
370 | */ | 351 | */ |
371 | if (unlikely(qc->tf.command == ATA_CMD_ID_ATA || | 352 | if (unlikely(qc->tf.command == ATA_CMD_ID_ATA || |
372 | qc->tf.command == ATA_CMD_ID_ATAPI)) { | 353 | qc->tf.command == ATA_CMD_ID_ATAPI)) { |
373 | u8 stat = ata_chk_status(ap); | 354 | u8 stat = ap->ops->sff_check_status(ap); |
374 | if (stat == 0x7f || stat == 0xff) | 355 | if (stat == 0x7f || stat == 0xff) |
375 | return AC_ERR_HSM; | 356 | return AC_ERR_HSM; |
376 | } | 357 | } |
377 | 358 | ||
378 | return ata_qc_issue_prot(qc); | 359 | return ata_sff_qc_issue(qc); |
379 | } | 360 | } |
380 | 361 | ||
381 | static void inic_freeze(struct ata_port *ap) | 362 | static void inic_freeze(struct ata_port *ap) |
@@ -384,7 +365,7 @@ static void inic_freeze(struct ata_port *ap) | |||
384 | 365 | ||
385 | __inic_set_pirq_mask(ap, PIRQ_MASK_FREEZE); | 366 | __inic_set_pirq_mask(ap, PIRQ_MASK_FREEZE); |
386 | 367 | ||
387 | ata_chk_status(ap); | 368 | ap->ops->sff_check_status(ap); |
388 | writeb(0xff, port_base + PORT_IRQ_STAT); | 369 | writeb(0xff, port_base + PORT_IRQ_STAT); |
389 | 370 | ||
390 | readb(port_base + PORT_IRQ_STAT); /* flush */ | 371 | readb(port_base + PORT_IRQ_STAT); /* flush */ |
@@ -394,7 +375,7 @@ static void inic_thaw(struct ata_port *ap) | |||
394 | { | 375 | { |
395 | void __iomem *port_base = inic_port_base(ap); | 376 | void __iomem *port_base = inic_port_base(ap); |
396 | 377 | ||
397 | ata_chk_status(ap); | 378 | ap->ops->sff_check_status(ap); |
398 | writeb(0xff, port_base + PORT_IRQ_STAT); | 379 | writeb(0xff, port_base + PORT_IRQ_STAT); |
399 | 380 | ||
400 | __inic_set_pirq_mask(ap, PIRQ_MASK_OTHER); | 381 | __inic_set_pirq_mask(ap, PIRQ_MASK_OTHER); |
@@ -436,10 +417,8 @@ static int inic_hardreset(struct ata_link *link, unsigned int *class, | |||
436 | if (ata_link_online(link)) { | 417 | if (ata_link_online(link)) { |
437 | struct ata_taskfile tf; | 418 | struct ata_taskfile tf; |
438 | 419 | ||
439 | /* wait a while before checking status */ | 420 | /* wait for link to become ready */ |
440 | ata_wait_after_reset(ap, deadline); | 421 | rc = ata_sff_wait_after_reset(link, 1, deadline); |
441 | |||
442 | rc = ata_wait_ready(ap, deadline); | ||
443 | /* link occupied, -ENODEV too is an error */ | 422 | /* link occupied, -ENODEV too is an error */ |
444 | if (rc) { | 423 | if (rc) { |
445 | ata_link_printk(link, KERN_WARNING, "device not ready " | 424 | ata_link_printk(link, KERN_WARNING, "device not ready " |
@@ -447,10 +426,8 @@ static int inic_hardreset(struct ata_link *link, unsigned int *class, | |||
447 | return rc; | 426 | return rc; |
448 | } | 427 | } |
449 | 428 | ||
450 | ata_tf_read(ap, &tf); | 429 | ata_sff_tf_read(ap, &tf); |
451 | *class = ata_dev_classify(&tf); | 430 | *class = ata_dev_classify(&tf); |
452 | if (*class == ATA_DEV_UNKNOWN) | ||
453 | *class = ATA_DEV_NONE; | ||
454 | } | 431 | } |
455 | 432 | ||
456 | return 0; | 433 | return 0; |
@@ -471,8 +448,7 @@ static void inic_error_handler(struct ata_port *ap) | |||
471 | spin_unlock_irqrestore(ap->lock, flags); | 448 | spin_unlock_irqrestore(ap->lock, flags); |
472 | 449 | ||
473 | /* PIO and DMA engines have been stopped, perform recovery */ | 450 | /* PIO and DMA engines have been stopped, perform recovery */ |
474 | ata_do_eh(ap, ata_std_prereset, NULL, inic_hardreset, | 451 | ata_std_error_handler(ap); |
475 | ata_std_postreset); | ||
476 | } | 452 | } |
477 | 453 | ||
478 | static void inic_post_internal_cmd(struct ata_queued_cmd *qc) | 454 | static void inic_post_internal_cmd(struct ata_queued_cmd *qc) |
@@ -541,35 +517,26 @@ static int inic_port_start(struct ata_port *ap) | |||
541 | } | 517 | } |
542 | 518 | ||
543 | static struct ata_port_operations inic_port_ops = { | 519 | static struct ata_port_operations inic_port_ops = { |
544 | .tf_load = ata_tf_load, | 520 | .inherits = &ata_sff_port_ops, |
545 | .tf_read = ata_tf_read, | ||
546 | .check_status = ata_check_status, | ||
547 | .exec_command = ata_exec_command, | ||
548 | .dev_select = ata_std_dev_select, | ||
549 | |||
550 | .scr_read = inic_scr_read, | ||
551 | .scr_write = inic_scr_write, | ||
552 | 521 | ||
553 | .bmdma_setup = inic_bmdma_setup, | 522 | .bmdma_setup = inic_bmdma_setup, |
554 | .bmdma_start = inic_bmdma_start, | 523 | .bmdma_start = inic_bmdma_start, |
555 | .bmdma_stop = inic_bmdma_stop, | 524 | .bmdma_stop = inic_bmdma_stop, |
556 | .bmdma_status = inic_bmdma_status, | 525 | .bmdma_status = inic_bmdma_status, |
557 | |||
558 | .irq_clear = inic_irq_clear, | ||
559 | .irq_on = ata_irq_on, | ||
560 | |||
561 | .qc_prep = ata_qc_prep, | ||
562 | .qc_issue = inic_qc_issue, | 526 | .qc_issue = inic_qc_issue, |
563 | .data_xfer = ata_data_xfer, | ||
564 | 527 | ||
565 | .freeze = inic_freeze, | 528 | .freeze = inic_freeze, |
566 | .thaw = inic_thaw, | 529 | .thaw = inic_thaw, |
530 | .softreset = ATA_OP_NULL, /* softreset is broken */ | ||
531 | .hardreset = inic_hardreset, | ||
567 | .error_handler = inic_error_handler, | 532 | .error_handler = inic_error_handler, |
568 | .post_internal_cmd = inic_post_internal_cmd, | 533 | .post_internal_cmd = inic_post_internal_cmd, |
569 | .dev_config = inic_dev_config, | 534 | .dev_config = inic_dev_config, |
570 | 535 | ||
571 | .port_resume = inic_port_resume, | 536 | .scr_read = inic_scr_read, |
537 | .scr_write = inic_scr_write, | ||
572 | 538 | ||
539 | .port_resume = inic_port_resume, | ||
573 | .port_start = inic_port_start, | 540 | .port_start = inic_port_start, |
574 | }; | 541 | }; |
575 | 542 | ||
@@ -692,7 +659,7 @@ static int inic_init_one(struct pci_dev *pdev, const struct pci_device_id *ent) | |||
692 | ((unsigned long)iomap[2 * i + 1] | ATA_PCI_CTL_OFS); | 659 | ((unsigned long)iomap[2 * i + 1] | ATA_PCI_CTL_OFS); |
693 | port->scr_addr = iomap[MMIO_BAR] + offset + PORT_SCR; | 660 | port->scr_addr = iomap[MMIO_BAR] + offset + PORT_SCR; |
694 | 661 | ||
695 | ata_std_ports(port); | 662 | ata_sff_std_ports(port); |
696 | 663 | ||
697 | ata_port_pbar_desc(ap, MMIO_BAR, -1, "mmio"); | 664 | ata_port_pbar_desc(ap, MMIO_BAR, -1, "mmio"); |
698 | ata_port_pbar_desc(ap, MMIO_BAR, offset, "port"); | 665 | ata_port_pbar_desc(ap, MMIO_BAR, offset, "port"); |
diff --git a/drivers/ata/sata_mv.c b/drivers/ata/sata_mv.c index 6ebebde8454a..05ff8c776497 100644 --- a/drivers/ata/sata_mv.c +++ b/drivers/ata/sata_mv.c | |||
@@ -1,6 +1,7 @@ | |||
1 | /* | 1 | /* |
2 | * sata_mv.c - Marvell SATA support | 2 | * sata_mv.c - Marvell SATA support |
3 | * | 3 | * |
4 | * Copyright 2008: Marvell Corporation, all rights reserved. | ||
4 | * Copyright 2005: EMC Corporation, all rights reserved. | 5 | * Copyright 2005: EMC Corporation, all rights reserved. |
5 | * Copyright 2005 Red Hat, Inc. All rights reserved. | 6 | * Copyright 2005 Red Hat, Inc. All rights reserved. |
6 | * | 7 | * |
@@ -39,7 +40,9 @@ | |||
39 | 40 | ||
40 | 5) Investigate problems with PCI Message Signalled Interrupts (MSI). | 41 | 5) Investigate problems with PCI Message Signalled Interrupts (MSI). |
41 | 42 | ||
42 | 6) Add port multiplier support (intermediate) | 43 | 6) Cache frequently-accessed registers in mv_port_priv to reduce overhead. |
44 | |||
45 | 7) Fix/reenable hot plug/unplug (should happen as a side-effect of (2) above). | ||
43 | 46 | ||
44 | 8) Develop a low-power-consumption strategy, and implement it. | 47 | 8) Develop a low-power-consumption strategy, and implement it. |
45 | 48 | ||
@@ -61,7 +64,6 @@ | |||
61 | 64 | ||
62 | */ | 65 | */ |
63 | 66 | ||
64 | |||
65 | #include <linux/kernel.h> | 67 | #include <linux/kernel.h> |
66 | #include <linux/module.h> | 68 | #include <linux/module.h> |
67 | #include <linux/pci.h> | 69 | #include <linux/pci.h> |
@@ -131,7 +133,7 @@ enum { | |||
131 | MV_FLAG_DUAL_HC = (1 << 30), /* two SATA Host Controllers */ | 133 | MV_FLAG_DUAL_HC = (1 << 30), /* two SATA Host Controllers */ |
132 | MV_FLAG_IRQ_COALESCE = (1 << 29), /* IRQ coalescing capability */ | 134 | MV_FLAG_IRQ_COALESCE = (1 << 29), /* IRQ coalescing capability */ |
133 | /* SoC integrated controllers, no PCI interface */ | 135 | /* SoC integrated controllers, no PCI interface */ |
134 | MV_FLAG_SOC = (1 << 28), | 136 | MV_FLAG_SOC = (1 << 28), |
135 | 137 | ||
136 | MV_COMMON_FLAGS = ATA_FLAG_SATA | ATA_FLAG_NO_LEGACY | | 138 | MV_COMMON_FLAGS = ATA_FLAG_SATA | ATA_FLAG_NO_LEGACY | |
137 | ATA_FLAG_MMIO | ATA_FLAG_NO_ATAPI | | 139 | ATA_FLAG_MMIO | ATA_FLAG_NO_ATAPI | |
@@ -141,6 +143,7 @@ enum { | |||
141 | CRQB_FLAG_READ = (1 << 0), | 143 | CRQB_FLAG_READ = (1 << 0), |
142 | CRQB_TAG_SHIFT = 1, | 144 | CRQB_TAG_SHIFT = 1, |
143 | CRQB_IOID_SHIFT = 6, /* CRQB Gen-II/IIE IO Id shift */ | 145 | CRQB_IOID_SHIFT = 6, /* CRQB Gen-II/IIE IO Id shift */ |
146 | CRQB_PMP_SHIFT = 12, /* CRQB Gen-II/IIE PMP shift */ | ||
144 | CRQB_HOSTQ_SHIFT = 17, /* CRQB Gen-II/IIE HostQueTag shift */ | 147 | CRQB_HOSTQ_SHIFT = 17, /* CRQB Gen-II/IIE HostQueTag shift */ |
145 | CRQB_CMD_ADDR_SHIFT = 8, | 148 | CRQB_CMD_ADDR_SHIFT = 8, |
146 | CRQB_CMD_CS = (0x2 << 11), | 149 | CRQB_CMD_CS = (0x2 << 11), |
@@ -199,7 +202,7 @@ enum { | |||
199 | TWSI_INT = (1 << 24), | 202 | TWSI_INT = (1 << 24), |
200 | HC_MAIN_RSVD = (0x7f << 25), /* bits 31-25 */ | 203 | HC_MAIN_RSVD = (0x7f << 25), /* bits 31-25 */ |
201 | HC_MAIN_RSVD_5 = (0x1fff << 19), /* bits 31-19 */ | 204 | HC_MAIN_RSVD_5 = (0x1fff << 19), /* bits 31-19 */ |
202 | HC_MAIN_RSVD_SOC = (0x3fffffb << 6), /* bits 31-9, 7-6 */ | 205 | HC_MAIN_RSVD_SOC = (0x3fffffb << 6), /* bits 31-9, 7-6 */ |
203 | HC_MAIN_MASKED_IRQS = (TRAN_LO_DONE | TRAN_HI_DONE | | 206 | HC_MAIN_MASKED_IRQS = (TRAN_LO_DONE | TRAN_HI_DONE | |
204 | PORTS_0_7_COAL_DONE | GPIO_INT | TWSI_INT | | 207 | PORTS_0_7_COAL_DONE | GPIO_INT | TWSI_INT | |
205 | HC_MAIN_RSVD), | 208 | HC_MAIN_RSVD), |
@@ -223,13 +226,24 @@ enum { | |||
223 | SATA_STATUS_OFS = 0x300, /* ctrl, err regs follow status */ | 226 | SATA_STATUS_OFS = 0x300, /* ctrl, err regs follow status */ |
224 | SATA_ACTIVE_OFS = 0x350, | 227 | SATA_ACTIVE_OFS = 0x350, |
225 | SATA_FIS_IRQ_CAUSE_OFS = 0x364, | 228 | SATA_FIS_IRQ_CAUSE_OFS = 0x364, |
229 | |||
230 | LTMODE_OFS = 0x30c, | ||
231 | LTMODE_BIT8 = (1 << 8), /* unknown, but necessary */ | ||
232 | |||
226 | PHY_MODE3 = 0x310, | 233 | PHY_MODE3 = 0x310, |
227 | PHY_MODE4 = 0x314, | 234 | PHY_MODE4 = 0x314, |
228 | PHY_MODE2 = 0x330, | 235 | PHY_MODE2 = 0x330, |
236 | SATA_IFCTL_OFS = 0x344, | ||
237 | SATA_IFSTAT_OFS = 0x34c, | ||
238 | VENDOR_UNIQUE_FIS_OFS = 0x35c, | ||
239 | |||
240 | FIS_CFG_OFS = 0x360, | ||
241 | FIS_CFG_SINGLE_SYNC = (1 << 16), /* SYNC on DMA activation */ | ||
242 | |||
229 | MV5_PHY_MODE = 0x74, | 243 | MV5_PHY_MODE = 0x74, |
230 | MV5_LT_MODE = 0x30, | 244 | MV5_LT_MODE = 0x30, |
231 | MV5_PHY_CTL = 0x0C, | 245 | MV5_PHY_CTL = 0x0C, |
232 | SATA_INTERFACE_CTL = 0x050, | 246 | SATA_INTERFACE_CFG = 0x050, |
233 | 247 | ||
234 | MV_M2_PREAMP_MASK = 0x7e0, | 248 | MV_M2_PREAMP_MASK = 0x7e0, |
235 | 249 | ||
@@ -240,6 +254,8 @@ enum { | |||
240 | EDMA_CFG_NCQ_GO_ON_ERR = (1 << 14), /* continue on error */ | 254 | EDMA_CFG_NCQ_GO_ON_ERR = (1 << 14), /* continue on error */ |
241 | EDMA_CFG_RD_BRST_EXT = (1 << 11), /* read burst 512B */ | 255 | EDMA_CFG_RD_BRST_EXT = (1 << 11), /* read burst 512B */ |
242 | EDMA_CFG_WR_BUFF_LEN = (1 << 13), /* write buffer 512B */ | 256 | EDMA_CFG_WR_BUFF_LEN = (1 << 13), /* write buffer 512B */ |
257 | EDMA_CFG_EDMA_FBS = (1 << 16), /* EDMA FIS-Based Switching */ | ||
258 | EDMA_CFG_FBS = (1 << 26), /* FIS-Based Switching */ | ||
243 | 259 | ||
244 | EDMA_ERR_IRQ_CAUSE_OFS = 0x8, | 260 | EDMA_ERR_IRQ_CAUSE_OFS = 0x8, |
245 | EDMA_ERR_IRQ_MASK_OFS = 0xc, | 261 | EDMA_ERR_IRQ_MASK_OFS = 0xc, |
@@ -282,7 +298,9 @@ enum { | |||
282 | EDMA_ERR_IRQ_TRANSIENT = EDMA_ERR_LNK_CTRL_RX_0 | | 298 | EDMA_ERR_IRQ_TRANSIENT = EDMA_ERR_LNK_CTRL_RX_0 | |
283 | EDMA_ERR_LNK_CTRL_RX_1 | | 299 | EDMA_ERR_LNK_CTRL_RX_1 | |
284 | EDMA_ERR_LNK_CTRL_RX_3 | | 300 | EDMA_ERR_LNK_CTRL_RX_3 | |
285 | EDMA_ERR_LNK_CTRL_TX, | 301 | EDMA_ERR_LNK_CTRL_TX | |
302 | /* temporary, until we fix hotplug: */ | ||
303 | (EDMA_ERR_DEV_DCON | EDMA_ERR_DEV_CON), | ||
286 | 304 | ||
287 | EDMA_EH_FREEZE = EDMA_ERR_D_PAR | | 305 | EDMA_EH_FREEZE = EDMA_ERR_D_PAR | |
288 | EDMA_ERR_PRD_PAR | | 306 | EDMA_ERR_PRD_PAR | |
@@ -298,6 +316,7 @@ enum { | |||
298 | EDMA_ERR_LNK_DATA_RX | | 316 | EDMA_ERR_LNK_DATA_RX | |
299 | EDMA_ERR_LNK_DATA_TX | | 317 | EDMA_ERR_LNK_DATA_TX | |
300 | EDMA_ERR_TRANS_PROTO, | 318 | EDMA_ERR_TRANS_PROTO, |
319 | |||
301 | EDMA_EH_FREEZE_5 = EDMA_ERR_D_PAR | | 320 | EDMA_EH_FREEZE_5 = EDMA_ERR_D_PAR | |
302 | EDMA_ERR_PRD_PAR | | 321 | EDMA_ERR_PRD_PAR | |
303 | EDMA_ERR_DEV_DCON | | 322 | EDMA_ERR_DEV_DCON | |
@@ -344,7 +363,6 @@ enum { | |||
344 | /* Port private flags (pp_flags) */ | 363 | /* Port private flags (pp_flags) */ |
345 | MV_PP_FLAG_EDMA_EN = (1 << 0), /* is EDMA engine enabled? */ | 364 | MV_PP_FLAG_EDMA_EN = (1 << 0), /* is EDMA engine enabled? */ |
346 | MV_PP_FLAG_NCQ_EN = (1 << 1), /* is EDMA set up for NCQ? */ | 365 | MV_PP_FLAG_NCQ_EN = (1 << 1), /* is EDMA set up for NCQ? */ |
347 | MV_PP_FLAG_HAD_A_RESET = (1 << 2), /* 1st hard reset complete? */ | ||
348 | }; | 366 | }; |
349 | 367 | ||
350 | #define IS_GEN_I(hpriv) ((hpriv)->hp_flags & MV_HP_GEN_I) | 368 | #define IS_GEN_I(hpriv) ((hpriv)->hp_flags & MV_HP_GEN_I) |
@@ -461,7 +479,6 @@ struct mv_hw_ops { | |||
461 | void (*reset_bus)(struct ata_host *host, void __iomem *mmio); | 479 | void (*reset_bus)(struct ata_host *host, void __iomem *mmio); |
462 | }; | 480 | }; |
463 | 481 | ||
464 | static void mv_irq_clear(struct ata_port *ap); | ||
465 | static int mv_scr_read(struct ata_port *ap, unsigned int sc_reg_in, u32 *val); | 482 | static int mv_scr_read(struct ata_port *ap, unsigned int sc_reg_in, u32 *val); |
466 | static int mv_scr_write(struct ata_port *ap, unsigned int sc_reg_in, u32 val); | 483 | static int mv_scr_write(struct ata_port *ap, unsigned int sc_reg_in, u32 val); |
467 | static int mv5_scr_read(struct ata_port *ap, unsigned int sc_reg_in, u32 *val); | 484 | static int mv5_scr_read(struct ata_port *ap, unsigned int sc_reg_in, u32 *val); |
@@ -471,7 +488,8 @@ static void mv_port_stop(struct ata_port *ap); | |||
471 | static void mv_qc_prep(struct ata_queued_cmd *qc); | 488 | static void mv_qc_prep(struct ata_queued_cmd *qc); |
472 | static void mv_qc_prep_iie(struct ata_queued_cmd *qc); | 489 | static void mv_qc_prep_iie(struct ata_queued_cmd *qc); |
473 | static unsigned int mv_qc_issue(struct ata_queued_cmd *qc); | 490 | static unsigned int mv_qc_issue(struct ata_queued_cmd *qc); |
474 | static void mv_error_handler(struct ata_port *ap); | 491 | static int mv_hardreset(struct ata_link *link, unsigned int *class, |
492 | unsigned long deadline); | ||
475 | static void mv_eh_freeze(struct ata_port *ap); | 493 | static void mv_eh_freeze(struct ata_port *ap); |
476 | static void mv_eh_thaw(struct ata_port *ap); | 494 | static void mv_eh_thaw(struct ata_port *ap); |
477 | static void mv6_dev_config(struct ata_device *dev); | 495 | static void mv6_dev_config(struct ata_device *dev); |
@@ -504,72 +522,46 @@ static void mv_soc_reset_flash(struct mv_host_priv *hpriv, | |||
504 | void __iomem *mmio); | 522 | void __iomem *mmio); |
505 | static void mv_soc_reset_bus(struct ata_host *host, void __iomem *mmio); | 523 | static void mv_soc_reset_bus(struct ata_host *host, void __iomem *mmio); |
506 | static void mv_reset_pci_bus(struct ata_host *host, void __iomem *mmio); | 524 | static void mv_reset_pci_bus(struct ata_host *host, void __iomem *mmio); |
507 | static void mv_channel_reset(struct mv_host_priv *hpriv, void __iomem *mmio, | 525 | static void mv_reset_channel(struct mv_host_priv *hpriv, void __iomem *mmio, |
508 | unsigned int port_no); | 526 | unsigned int port_no); |
509 | static void mv_edma_cfg(struct mv_port_priv *pp, struct mv_host_priv *hpriv, | 527 | static int mv_stop_edma(struct ata_port *ap); |
510 | void __iomem *port_mmio, int want_ncq); | 528 | static int mv_stop_edma_engine(void __iomem *port_mmio); |
511 | static int __mv_stop_dma(struct ata_port *ap); | 529 | static void mv_edma_cfg(struct ata_port *ap, int want_ncq); |
530 | |||
531 | static void mv_pmp_select(struct ata_port *ap, int pmp); | ||
532 | static int mv_pmp_hardreset(struct ata_link *link, unsigned int *class, | ||
533 | unsigned long deadline); | ||
534 | static int mv_softreset(struct ata_link *link, unsigned int *class, | ||
535 | unsigned long deadline); | ||
512 | 536 | ||
513 | /* .sg_tablesize is (MV_MAX_SG_CT / 2) in the structures below | 537 | /* .sg_tablesize is (MV_MAX_SG_CT / 2) in the structures below |
514 | * because we have to allow room for worst case splitting of | 538 | * because we have to allow room for worst case splitting of |
515 | * PRDs for 64K boundaries in mv_fill_sg(). | 539 | * PRDs for 64K boundaries in mv_fill_sg(). |
516 | */ | 540 | */ |
517 | static struct scsi_host_template mv5_sht = { | 541 | static struct scsi_host_template mv5_sht = { |
518 | .module = THIS_MODULE, | 542 | ATA_BASE_SHT(DRV_NAME), |
519 | .name = DRV_NAME, | ||
520 | .ioctl = ata_scsi_ioctl, | ||
521 | .queuecommand = ata_scsi_queuecmd, | ||
522 | .can_queue = ATA_DEF_QUEUE, | ||
523 | .this_id = ATA_SHT_THIS_ID, | ||
524 | .sg_tablesize = MV_MAX_SG_CT / 2, | 543 | .sg_tablesize = MV_MAX_SG_CT / 2, |
525 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
526 | .emulated = ATA_SHT_EMULATED, | ||
527 | .use_clustering = 1, | ||
528 | .proc_name = DRV_NAME, | ||
529 | .dma_boundary = MV_DMA_BOUNDARY, | 544 | .dma_boundary = MV_DMA_BOUNDARY, |
530 | .slave_configure = ata_scsi_slave_config, | ||
531 | .slave_destroy = ata_scsi_slave_destroy, | ||
532 | .bios_param = ata_std_bios_param, | ||
533 | }; | 545 | }; |
534 | 546 | ||
535 | static struct scsi_host_template mv6_sht = { | 547 | static struct scsi_host_template mv6_sht = { |
536 | .module = THIS_MODULE, | 548 | ATA_NCQ_SHT(DRV_NAME), |
537 | .name = DRV_NAME, | ||
538 | .ioctl = ata_scsi_ioctl, | ||
539 | .queuecommand = ata_scsi_queuecmd, | ||
540 | .change_queue_depth = ata_scsi_change_queue_depth, | ||
541 | .can_queue = MV_MAX_Q_DEPTH - 1, | 549 | .can_queue = MV_MAX_Q_DEPTH - 1, |
542 | .this_id = ATA_SHT_THIS_ID, | ||
543 | .sg_tablesize = MV_MAX_SG_CT / 2, | 550 | .sg_tablesize = MV_MAX_SG_CT / 2, |
544 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
545 | .emulated = ATA_SHT_EMULATED, | ||
546 | .use_clustering = 1, | ||
547 | .proc_name = DRV_NAME, | ||
548 | .dma_boundary = MV_DMA_BOUNDARY, | 551 | .dma_boundary = MV_DMA_BOUNDARY, |
549 | .slave_configure = ata_scsi_slave_config, | ||
550 | .slave_destroy = ata_scsi_slave_destroy, | ||
551 | .bios_param = ata_std_bios_param, | ||
552 | }; | 552 | }; |
553 | 553 | ||
554 | static const struct ata_port_operations mv5_ops = { | 554 | static struct ata_port_operations mv5_ops = { |
555 | .tf_load = ata_tf_load, | 555 | .inherits = &ata_sff_port_ops, |
556 | .tf_read = ata_tf_read, | ||
557 | .check_status = ata_check_status, | ||
558 | .exec_command = ata_exec_command, | ||
559 | .dev_select = ata_std_dev_select, | ||
560 | |||
561 | .cable_detect = ata_cable_sata, | ||
562 | 556 | ||
563 | .qc_prep = mv_qc_prep, | 557 | .qc_prep = mv_qc_prep, |
564 | .qc_issue = mv_qc_issue, | 558 | .qc_issue = mv_qc_issue, |
565 | .data_xfer = ata_data_xfer, | ||
566 | |||
567 | .irq_clear = mv_irq_clear, | ||
568 | .irq_on = ata_irq_on, | ||
569 | 559 | ||
570 | .error_handler = mv_error_handler, | ||
571 | .freeze = mv_eh_freeze, | 560 | .freeze = mv_eh_freeze, |
572 | .thaw = mv_eh_thaw, | 561 | .thaw = mv_eh_thaw, |
562 | .hardreset = mv_hardreset, | ||
563 | .error_handler = ata_std_error_handler, /* avoid SFF EH */ | ||
564 | .post_internal_cmd = ATA_OP_NULL, | ||
573 | 565 | ||
574 | .scr_read = mv5_scr_read, | 566 | .scr_read = mv5_scr_read, |
575 | .scr_write = mv5_scr_write, | 567 | .scr_write = mv5_scr_write, |
@@ -578,61 +570,24 @@ static const struct ata_port_operations mv5_ops = { | |||
578 | .port_stop = mv_port_stop, | 570 | .port_stop = mv_port_stop, |
579 | }; | 571 | }; |
580 | 572 | ||
581 | static const struct ata_port_operations mv6_ops = { | 573 | static struct ata_port_operations mv6_ops = { |
574 | .inherits = &mv5_ops, | ||
575 | .qc_defer = sata_pmp_qc_defer_cmd_switch, | ||
582 | .dev_config = mv6_dev_config, | 576 | .dev_config = mv6_dev_config, |
583 | .tf_load = ata_tf_load, | ||
584 | .tf_read = ata_tf_read, | ||
585 | .check_status = ata_check_status, | ||
586 | .exec_command = ata_exec_command, | ||
587 | .dev_select = ata_std_dev_select, | ||
588 | |||
589 | .cable_detect = ata_cable_sata, | ||
590 | |||
591 | .qc_prep = mv_qc_prep, | ||
592 | .qc_issue = mv_qc_issue, | ||
593 | .data_xfer = ata_data_xfer, | ||
594 | |||
595 | .irq_clear = mv_irq_clear, | ||
596 | .irq_on = ata_irq_on, | ||
597 | |||
598 | .error_handler = mv_error_handler, | ||
599 | .freeze = mv_eh_freeze, | ||
600 | .thaw = mv_eh_thaw, | ||
601 | .qc_defer = ata_std_qc_defer, | ||
602 | |||
603 | .scr_read = mv_scr_read, | 577 | .scr_read = mv_scr_read, |
604 | .scr_write = mv_scr_write, | 578 | .scr_write = mv_scr_write, |
605 | 579 | ||
606 | .port_start = mv_port_start, | 580 | .pmp_hardreset = mv_pmp_hardreset, |
607 | .port_stop = mv_port_stop, | 581 | .pmp_softreset = mv_softreset, |
582 | .softreset = mv_softreset, | ||
583 | .error_handler = sata_pmp_error_handler, | ||
608 | }; | 584 | }; |
609 | 585 | ||
610 | static const struct ata_port_operations mv_iie_ops = { | 586 | static struct ata_port_operations mv_iie_ops = { |
611 | .tf_load = ata_tf_load, | 587 | .inherits = &mv6_ops, |
612 | .tf_read = ata_tf_read, | 588 | .qc_defer = ata_std_qc_defer, /* FIS-based switching */ |
613 | .check_status = ata_check_status, | 589 | .dev_config = ATA_OP_NULL, |
614 | .exec_command = ata_exec_command, | ||
615 | .dev_select = ata_std_dev_select, | ||
616 | |||
617 | .cable_detect = ata_cable_sata, | ||
618 | |||
619 | .qc_prep = mv_qc_prep_iie, | 590 | .qc_prep = mv_qc_prep_iie, |
620 | .qc_issue = mv_qc_issue, | ||
621 | .data_xfer = ata_data_xfer, | ||
622 | |||
623 | .irq_clear = mv_irq_clear, | ||
624 | .irq_on = ata_irq_on, | ||
625 | |||
626 | .error_handler = mv_error_handler, | ||
627 | .freeze = mv_eh_freeze, | ||
628 | .thaw = mv_eh_thaw, | ||
629 | .qc_defer = ata_std_qc_defer, | ||
630 | |||
631 | .scr_read = mv_scr_read, | ||
632 | .scr_write = mv_scr_write, | ||
633 | |||
634 | .port_start = mv_port_start, | ||
635 | .port_stop = mv_port_stop, | ||
636 | }; | 591 | }; |
637 | 592 | ||
638 | static const struct ata_port_info mv_port_info[] = { | 593 | static const struct ata_port_info mv_port_info[] = { |
@@ -656,6 +611,7 @@ static const struct ata_port_info mv_port_info[] = { | |||
656 | }, | 611 | }, |
657 | { /* chip_604x */ | 612 | { /* chip_604x */ |
658 | .flags = MV_COMMON_FLAGS | MV_6XXX_FLAGS | | 613 | .flags = MV_COMMON_FLAGS | MV_6XXX_FLAGS | |
614 | ATA_FLAG_PMP | ATA_FLAG_ACPI_SATA | | ||
659 | ATA_FLAG_NCQ, | 615 | ATA_FLAG_NCQ, |
660 | .pio_mask = 0x1f, /* pio0-4 */ | 616 | .pio_mask = 0x1f, /* pio0-4 */ |
661 | .udma_mask = ATA_UDMA6, | 617 | .udma_mask = ATA_UDMA6, |
@@ -663,6 +619,7 @@ static const struct ata_port_info mv_port_info[] = { | |||
663 | }, | 619 | }, |
664 | { /* chip_608x */ | 620 | { /* chip_608x */ |
665 | .flags = MV_COMMON_FLAGS | MV_6XXX_FLAGS | | 621 | .flags = MV_COMMON_FLAGS | MV_6XXX_FLAGS | |
622 | ATA_FLAG_PMP | ATA_FLAG_ACPI_SATA | | ||
666 | ATA_FLAG_NCQ | MV_FLAG_DUAL_HC, | 623 | ATA_FLAG_NCQ | MV_FLAG_DUAL_HC, |
667 | .pio_mask = 0x1f, /* pio0-4 */ | 624 | .pio_mask = 0x1f, /* pio0-4 */ |
668 | .udma_mask = ATA_UDMA6, | 625 | .udma_mask = ATA_UDMA6, |
@@ -670,6 +627,7 @@ static const struct ata_port_info mv_port_info[] = { | |||
670 | }, | 627 | }, |
671 | { /* chip_6042 */ | 628 | { /* chip_6042 */ |
672 | .flags = MV_COMMON_FLAGS | MV_6XXX_FLAGS | | 629 | .flags = MV_COMMON_FLAGS | MV_6XXX_FLAGS | |
630 | ATA_FLAG_PMP | ATA_FLAG_ACPI_SATA | | ||
673 | ATA_FLAG_NCQ, | 631 | ATA_FLAG_NCQ, |
674 | .pio_mask = 0x1f, /* pio0-4 */ | 632 | .pio_mask = 0x1f, /* pio0-4 */ |
675 | .udma_mask = ATA_UDMA6, | 633 | .udma_mask = ATA_UDMA6, |
@@ -677,16 +635,19 @@ static const struct ata_port_info mv_port_info[] = { | |||
677 | }, | 635 | }, |
678 | { /* chip_7042 */ | 636 | { /* chip_7042 */ |
679 | .flags = MV_COMMON_FLAGS | MV_6XXX_FLAGS | | 637 | .flags = MV_COMMON_FLAGS | MV_6XXX_FLAGS | |
638 | ATA_FLAG_PMP | ATA_FLAG_ACPI_SATA | | ||
680 | ATA_FLAG_NCQ, | 639 | ATA_FLAG_NCQ, |
681 | .pio_mask = 0x1f, /* pio0-4 */ | 640 | .pio_mask = 0x1f, /* pio0-4 */ |
682 | .udma_mask = ATA_UDMA6, | 641 | .udma_mask = ATA_UDMA6, |
683 | .port_ops = &mv_iie_ops, | 642 | .port_ops = &mv_iie_ops, |
684 | }, | 643 | }, |
685 | { /* chip_soc */ | 644 | { /* chip_soc */ |
686 | .flags = MV_COMMON_FLAGS | MV_FLAG_SOC, | 645 | .flags = MV_COMMON_FLAGS | MV_6XXX_FLAGS | |
687 | .pio_mask = 0x1f, /* pio0-4 */ | 646 | ATA_FLAG_PMP | ATA_FLAG_ACPI_SATA | |
688 | .udma_mask = ATA_UDMA6, | 647 | ATA_FLAG_NCQ | MV_FLAG_SOC, |
689 | .port_ops = &mv_iie_ops, | 648 | .pio_mask = 0x1f, /* pio0-4 */ |
649 | .udma_mask = ATA_UDMA6, | ||
650 | .port_ops = &mv_iie_ops, | ||
690 | }, | 651 | }, |
691 | }; | 652 | }; |
692 | 653 | ||
@@ -785,6 +746,14 @@ static inline void __iomem *mv_port_base(void __iomem *base, unsigned int port) | |||
785 | (mv_hardport_from_port(port) * MV_PORT_REG_SZ); | 746 | (mv_hardport_from_port(port) * MV_PORT_REG_SZ); |
786 | } | 747 | } |
787 | 748 | ||
749 | static void __iomem *mv5_phy_base(void __iomem *mmio, unsigned int port) | ||
750 | { | ||
751 | void __iomem *hc_mmio = mv_hc_base_from_port(mmio, port); | ||
752 | unsigned long ofs = (mv_hardport_from_port(port) + 1) * 0x100UL; | ||
753 | |||
754 | return hc_mmio + ofs; | ||
755 | } | ||
756 | |||
788 | static inline void __iomem *mv_host_base(struct ata_host *host) | 757 | static inline void __iomem *mv_host_base(struct ata_host *host) |
789 | { | 758 | { |
790 | struct mv_host_priv *hpriv = host->private_data; | 759 | struct mv_host_priv *hpriv = host->private_data; |
@@ -801,10 +770,6 @@ static inline int mv_get_hc_count(unsigned long port_flags) | |||
801 | return ((port_flags & MV_FLAG_DUAL_HC) ? 2 : 1); | 770 | return ((port_flags & MV_FLAG_DUAL_HC) ? 2 : 1); |
802 | } | 771 | } |
803 | 772 | ||
804 | static void mv_irq_clear(struct ata_port *ap) | ||
805 | { | ||
806 | } | ||
807 | |||
808 | static void mv_set_edma_ptrs(void __iomem *port_mmio, | 773 | static void mv_set_edma_ptrs(void __iomem *port_mmio, |
809 | struct mv_host_priv *hpriv, | 774 | struct mv_host_priv *hpriv, |
810 | struct mv_port_priv *pp) | 775 | struct mv_port_priv *pp) |
@@ -864,7 +829,7 @@ static void mv_start_dma(struct ata_port *ap, void __iomem *port_mmio, | |||
864 | if (pp->pp_flags & MV_PP_FLAG_EDMA_EN) { | 829 | if (pp->pp_flags & MV_PP_FLAG_EDMA_EN) { |
865 | int using_ncq = ((pp->pp_flags & MV_PP_FLAG_NCQ_EN) != 0); | 830 | int using_ncq = ((pp->pp_flags & MV_PP_FLAG_NCQ_EN) != 0); |
866 | if (want_ncq != using_ncq) | 831 | if (want_ncq != using_ncq) |
867 | __mv_stop_dma(ap); | 832 | mv_stop_edma(ap); |
868 | } | 833 | } |
869 | if (!(pp->pp_flags & MV_PP_FLAG_EDMA_EN)) { | 834 | if (!(pp->pp_flags & MV_PP_FLAG_EDMA_EN)) { |
870 | struct mv_host_priv *hpriv = ap->host->private_data; | 835 | struct mv_host_priv *hpriv = ap->host->private_data; |
@@ -885,7 +850,7 @@ static void mv_start_dma(struct ata_port *ap, void __iomem *port_mmio, | |||
885 | hc_mmio + HC_IRQ_CAUSE_OFS); | 850 | hc_mmio + HC_IRQ_CAUSE_OFS); |
886 | } | 851 | } |
887 | 852 | ||
888 | mv_edma_cfg(pp, hpriv, port_mmio, want_ncq); | 853 | mv_edma_cfg(ap, want_ncq); |
889 | 854 | ||
890 | /* clear FIS IRQ Cause */ | 855 | /* clear FIS IRQ Cause */ |
891 | writelfl(0, port_mmio + SATA_FIS_IRQ_CAUSE_OFS); | 856 | writelfl(0, port_mmio + SATA_FIS_IRQ_CAUSE_OFS); |
@@ -899,58 +864,42 @@ static void mv_start_dma(struct ata_port *ap, void __iomem *port_mmio, | |||
899 | } | 864 | } |
900 | 865 | ||
901 | /** | 866 | /** |
902 | * __mv_stop_dma - Disable eDMA engine | 867 | * mv_stop_edma_engine - Disable eDMA engine |
903 | * @ap: ATA channel to manipulate | 868 | * @port_mmio: io base address |
904 | * | ||
905 | * Verify the local cache of the eDMA state is accurate with a | ||
906 | * WARN_ON. | ||
907 | * | 869 | * |
908 | * LOCKING: | 870 | * LOCKING: |
909 | * Inherited from caller. | 871 | * Inherited from caller. |
910 | */ | 872 | */ |
911 | static int __mv_stop_dma(struct ata_port *ap) | 873 | static int mv_stop_edma_engine(void __iomem *port_mmio) |
912 | { | 874 | { |
913 | void __iomem *port_mmio = mv_ap_base(ap); | 875 | int i; |
914 | struct mv_port_priv *pp = ap->private_data; | ||
915 | u32 reg; | ||
916 | int i, err = 0; | ||
917 | 876 | ||
918 | if (pp->pp_flags & MV_PP_FLAG_EDMA_EN) { | 877 | /* Disable eDMA. The disable bit auto clears. */ |
919 | /* Disable EDMA if active. The disable bit auto clears. | 878 | writelfl(EDMA_DS, port_mmio + EDMA_CMD_OFS); |
920 | */ | ||
921 | writelfl(EDMA_DS, port_mmio + EDMA_CMD_OFS); | ||
922 | pp->pp_flags &= ~MV_PP_FLAG_EDMA_EN; | ||
923 | } else { | ||
924 | WARN_ON(EDMA_EN & readl(port_mmio + EDMA_CMD_OFS)); | ||
925 | } | ||
926 | 879 | ||
927 | /* now properly wait for the eDMA to stop */ | 880 | /* Wait for the chip to confirm eDMA is off. */ |
928 | for (i = 1000; i > 0; i--) { | 881 | for (i = 10000; i > 0; i--) { |
929 | reg = readl(port_mmio + EDMA_CMD_OFS); | 882 | u32 reg = readl(port_mmio + EDMA_CMD_OFS); |
930 | if (!(reg & EDMA_EN)) | 883 | if (!(reg & EDMA_EN)) |
931 | break; | 884 | return 0; |
932 | 885 | udelay(10); | |
933 | udelay(100); | ||
934 | } | ||
935 | |||
936 | if (reg & EDMA_EN) { | ||
937 | ata_port_printk(ap, KERN_ERR, "Unable to stop eDMA\n"); | ||
938 | err = -EIO; | ||
939 | } | 886 | } |
940 | 887 | return -EIO; | |
941 | return err; | ||
942 | } | 888 | } |
943 | 889 | ||
944 | static int mv_stop_dma(struct ata_port *ap) | 890 | static int mv_stop_edma(struct ata_port *ap) |
945 | { | 891 | { |
946 | unsigned long flags; | 892 | void __iomem *port_mmio = mv_ap_base(ap); |
947 | int rc; | 893 | struct mv_port_priv *pp = ap->private_data; |
948 | |||
949 | spin_lock_irqsave(&ap->host->lock, flags); | ||
950 | rc = __mv_stop_dma(ap); | ||
951 | spin_unlock_irqrestore(&ap->host->lock, flags); | ||
952 | 894 | ||
953 | return rc; | 895 | if (!(pp->pp_flags & MV_PP_FLAG_EDMA_EN)) |
896 | return 0; | ||
897 | pp->pp_flags &= ~MV_PP_FLAG_EDMA_EN; | ||
898 | if (mv_stop_edma_engine(port_mmio)) { | ||
899 | ata_port_printk(ap, KERN_ERR, "Unable to stop eDMA\n"); | ||
900 | return -EIO; | ||
901 | } | ||
902 | return 0; | ||
954 | } | 903 | } |
955 | 904 | ||
956 | #ifdef ATA_DEBUG | 905 | #ifdef ATA_DEBUG |
@@ -1074,18 +1023,50 @@ static int mv_scr_write(struct ata_port *ap, unsigned int sc_reg_in, u32 val) | |||
1074 | static void mv6_dev_config(struct ata_device *adev) | 1023 | static void mv6_dev_config(struct ata_device *adev) |
1075 | { | 1024 | { |
1076 | /* | 1025 | /* |
1026 | * Deal with Gen-II ("mv6") hardware quirks/restrictions: | ||
1027 | * | ||
1028 | * Gen-II does not support NCQ over a port multiplier | ||
1029 | * (no FIS-based switching). | ||
1030 | * | ||
1077 | * We don't have hob_nsect when doing NCQ commands on Gen-II. | 1031 | * We don't have hob_nsect when doing NCQ commands on Gen-II. |
1078 | * See mv_qc_prep() for more info. | 1032 | * See mv_qc_prep() for more info. |
1079 | */ | 1033 | */ |
1080 | if (adev->flags & ATA_DFLAG_NCQ) | 1034 | if (adev->flags & ATA_DFLAG_NCQ) { |
1081 | if (adev->max_sectors > ATA_MAX_SECTORS) | 1035 | if (sata_pmp_attached(adev->link->ap)) |
1036 | adev->flags &= ~ATA_DFLAG_NCQ; | ||
1037 | else if (adev->max_sectors > ATA_MAX_SECTORS) | ||
1082 | adev->max_sectors = ATA_MAX_SECTORS; | 1038 | adev->max_sectors = ATA_MAX_SECTORS; |
1039 | } | ||
1083 | } | 1040 | } |
1084 | 1041 | ||
1085 | static void mv_edma_cfg(struct mv_port_priv *pp, struct mv_host_priv *hpriv, | 1042 | static void mv_config_fbs(void __iomem *port_mmio, int enable_fbs) |
1086 | void __iomem *port_mmio, int want_ncq) | 1043 | { |
1044 | u32 old_fcfg, new_fcfg, old_ltmode, new_ltmode; | ||
1045 | /* | ||
1046 | * Various bit settings required for operation | ||
1047 | * in FIS-based switching (fbs) mode on GenIIe: | ||
1048 | */ | ||
1049 | old_fcfg = readl(port_mmio + FIS_CFG_OFS); | ||
1050 | old_ltmode = readl(port_mmio + LTMODE_OFS); | ||
1051 | if (enable_fbs) { | ||
1052 | new_fcfg = old_fcfg | FIS_CFG_SINGLE_SYNC; | ||
1053 | new_ltmode = old_ltmode | LTMODE_BIT8; | ||
1054 | } else { /* disable fbs */ | ||
1055 | new_fcfg = old_fcfg & ~FIS_CFG_SINGLE_SYNC; | ||
1056 | new_ltmode = old_ltmode & ~LTMODE_BIT8; | ||
1057 | } | ||
1058 | if (new_fcfg != old_fcfg) | ||
1059 | writelfl(new_fcfg, port_mmio + FIS_CFG_OFS); | ||
1060 | if (new_ltmode != old_ltmode) | ||
1061 | writelfl(new_ltmode, port_mmio + LTMODE_OFS); | ||
1062 | } | ||
1063 | |||
1064 | static void mv_edma_cfg(struct ata_port *ap, int want_ncq) | ||
1087 | { | 1065 | { |
1088 | u32 cfg; | 1066 | u32 cfg; |
1067 | struct mv_port_priv *pp = ap->private_data; | ||
1068 | struct mv_host_priv *hpriv = ap->host->private_data; | ||
1069 | void __iomem *port_mmio = mv_ap_base(ap); | ||
1089 | 1070 | ||
1090 | /* set up non-NCQ EDMA configuration */ | 1071 | /* set up non-NCQ EDMA configuration */ |
1091 | cfg = EDMA_CFG_Q_DEPTH; /* always 0x1f for *all* chips */ | 1072 | cfg = EDMA_CFG_Q_DEPTH; /* always 0x1f for *all* chips */ |
@@ -1101,6 +1082,13 @@ static void mv_edma_cfg(struct mv_port_priv *pp, struct mv_host_priv *hpriv, | |||
1101 | cfg |= (1 << 22); /* enab 4-entry host queue cache */ | 1082 | cfg |= (1 << 22); /* enab 4-entry host queue cache */ |
1102 | cfg |= (1 << 18); /* enab early completion */ | 1083 | cfg |= (1 << 18); /* enab early completion */ |
1103 | cfg |= (1 << 17); /* enab cut-through (dis stor&forwrd) */ | 1084 | cfg |= (1 << 17); /* enab cut-through (dis stor&forwrd) */ |
1085 | |||
1086 | if (want_ncq && sata_pmp_attached(ap)) { | ||
1087 | cfg |= EDMA_CFG_EDMA_FBS; /* FIS-based switching */ | ||
1088 | mv_config_fbs(port_mmio, 1); | ||
1089 | } else { | ||
1090 | mv_config_fbs(port_mmio, 0); | ||
1091 | } | ||
1104 | } | 1092 | } |
1105 | 1093 | ||
1106 | if (want_ncq) { | 1094 | if (want_ncq) { |
@@ -1156,8 +1144,6 @@ static int mv_port_start(struct ata_port *ap) | |||
1156 | struct device *dev = ap->host->dev; | 1144 | struct device *dev = ap->host->dev; |
1157 | struct mv_host_priv *hpriv = ap->host->private_data; | 1145 | struct mv_host_priv *hpriv = ap->host->private_data; |
1158 | struct mv_port_priv *pp; | 1146 | struct mv_port_priv *pp; |
1159 | void __iomem *port_mmio = mv_ap_base(ap); | ||
1160 | unsigned long flags; | ||
1161 | int tag; | 1147 | int tag; |
1162 | 1148 | ||
1163 | pp = devm_kzalloc(dev, sizeof(*pp), GFP_KERNEL); | 1149 | pp = devm_kzalloc(dev, sizeof(*pp), GFP_KERNEL); |
@@ -1190,18 +1176,6 @@ static int mv_port_start(struct ata_port *ap) | |||
1190 | pp->sg_tbl_dma[tag] = pp->sg_tbl_dma[0]; | 1176 | pp->sg_tbl_dma[tag] = pp->sg_tbl_dma[0]; |
1191 | } | 1177 | } |
1192 | } | 1178 | } |
1193 | |||
1194 | spin_lock_irqsave(&ap->host->lock, flags); | ||
1195 | |||
1196 | mv_edma_cfg(pp, hpriv, port_mmio, 0); | ||
1197 | mv_set_edma_ptrs(port_mmio, hpriv, pp); | ||
1198 | |||
1199 | spin_unlock_irqrestore(&ap->host->lock, flags); | ||
1200 | |||
1201 | /* Don't turn on EDMA here...do it before DMA commands only. Else | ||
1202 | * we'll be unable to send non-data, PIO, etc due to restricted access | ||
1203 | * to shadow regs. | ||
1204 | */ | ||
1205 | return 0; | 1179 | return 0; |
1206 | 1180 | ||
1207 | out_port_free_dma_mem: | 1181 | out_port_free_dma_mem: |
@@ -1220,7 +1194,7 @@ out_port_free_dma_mem: | |||
1220 | */ | 1194 | */ |
1221 | static void mv_port_stop(struct ata_port *ap) | 1195 | static void mv_port_stop(struct ata_port *ap) |
1222 | { | 1196 | { |
1223 | mv_stop_dma(ap); | 1197 | mv_stop_edma(ap); |
1224 | mv_port_free_dma_mem(ap); | 1198 | mv_port_free_dma_mem(ap); |
1225 | } | 1199 | } |
1226 | 1200 | ||
@@ -1306,6 +1280,7 @@ static void mv_qc_prep(struct ata_queued_cmd *qc) | |||
1306 | flags |= CRQB_FLAG_READ; | 1280 | flags |= CRQB_FLAG_READ; |
1307 | WARN_ON(MV_MAX_Q_DEPTH <= qc->tag); | 1281 | WARN_ON(MV_MAX_Q_DEPTH <= qc->tag); |
1308 | flags |= qc->tag << CRQB_TAG_SHIFT; | 1282 | flags |= qc->tag << CRQB_TAG_SHIFT; |
1283 | flags |= (qc->dev->link->pmp & 0xf) << CRQB_PMP_SHIFT; | ||
1309 | 1284 | ||
1310 | /* get current queue index from software */ | 1285 | /* get current queue index from software */ |
1311 | in_index = pp->req_idx & MV_MAX_Q_DEPTH_MASK; | 1286 | in_index = pp->req_idx & MV_MAX_Q_DEPTH_MASK; |
@@ -1390,14 +1365,14 @@ static void mv_qc_prep_iie(struct ata_queued_cmd *qc) | |||
1390 | (qc->tf.protocol != ATA_PROT_NCQ)) | 1365 | (qc->tf.protocol != ATA_PROT_NCQ)) |
1391 | return; | 1366 | return; |
1392 | 1367 | ||
1393 | /* Fill in Gen IIE command request block | 1368 | /* Fill in Gen IIE command request block */ |
1394 | */ | ||
1395 | if (!(qc->tf.flags & ATA_TFLAG_WRITE)) | 1369 | if (!(qc->tf.flags & ATA_TFLAG_WRITE)) |
1396 | flags |= CRQB_FLAG_READ; | 1370 | flags |= CRQB_FLAG_READ; |
1397 | 1371 | ||
1398 | WARN_ON(MV_MAX_Q_DEPTH <= qc->tag); | 1372 | WARN_ON(MV_MAX_Q_DEPTH <= qc->tag); |
1399 | flags |= qc->tag << CRQB_TAG_SHIFT; | 1373 | flags |= qc->tag << CRQB_TAG_SHIFT; |
1400 | flags |= qc->tag << CRQB_HOSTQ_SHIFT; | 1374 | flags |= qc->tag << CRQB_HOSTQ_SHIFT; |
1375 | flags |= (qc->dev->link->pmp & 0xf) << CRQB_PMP_SHIFT; | ||
1401 | 1376 | ||
1402 | /* get current queue index from software */ | 1377 | /* get current queue index from software */ |
1403 | in_index = pp->req_idx & MV_MAX_Q_DEPTH_MASK; | 1378 | in_index = pp->req_idx & MV_MAX_Q_DEPTH_MASK; |
@@ -1455,12 +1430,14 @@ static unsigned int mv_qc_issue(struct ata_queued_cmd *qc) | |||
1455 | 1430 | ||
1456 | if ((qc->tf.protocol != ATA_PROT_DMA) && | 1431 | if ((qc->tf.protocol != ATA_PROT_DMA) && |
1457 | (qc->tf.protocol != ATA_PROT_NCQ)) { | 1432 | (qc->tf.protocol != ATA_PROT_NCQ)) { |
1458 | /* We're about to send a non-EDMA capable command to the | 1433 | /* |
1434 | * We're about to send a non-EDMA capable command to the | ||
1459 | * port. Turn off EDMA so there won't be problems accessing | 1435 | * port. Turn off EDMA so there won't be problems accessing |
1460 | * shadow block, etc registers. | 1436 | * shadow block, etc registers. |
1461 | */ | 1437 | */ |
1462 | __mv_stop_dma(ap); | 1438 | mv_stop_edma(ap); |
1463 | return ata_qc_issue_prot(qc); | 1439 | mv_pmp_select(ap, qc->dev->link->pmp); |
1440 | return ata_sff_qc_issue(qc); | ||
1464 | } | 1441 | } |
1465 | 1442 | ||
1466 | mv_start_dma(ap, port_mmio, pp, qc->tf.protocol); | 1443 | mv_start_dma(ap, port_mmio, pp, qc->tf.protocol); |
@@ -1482,10 +1459,10 @@ static unsigned int mv_qc_issue(struct ata_queued_cmd *qc) | |||
1482 | * @reset_allowed: bool: 0 == don't trigger from reset here | 1459 | * @reset_allowed: bool: 0 == don't trigger from reset here |
1483 | * | 1460 | * |
1484 | * In most cases, just clear the interrupt and move on. However, | 1461 | * In most cases, just clear the interrupt and move on. However, |
1485 | * some cases require an eDMA reset, which is done right before | 1462 | * some cases require an eDMA reset, which also performs a COMRESET. |
1486 | * the COMRESET in mv_phy_reset(). The SERR case requires a | 1463 | * The SERR case requires a clear of pending errors in the SATA |
1487 | * clear of pending errors in the SATA SERROR register. Finally, | 1464 | * SERROR register. Finally, if the port disabled DMA, |
1488 | * if the port disabled DMA, update our cached copy to match. | 1465 | * update our cached copy to match. |
1489 | * | 1466 | * |
1490 | * LOCKING: | 1467 | * LOCKING: |
1491 | * Inherited from caller. | 1468 | * Inherited from caller. |
@@ -1524,14 +1501,14 @@ static void mv_err_intr(struct ata_port *ap, struct ata_queued_cmd *qc) | |||
1524 | EDMA_ERR_CRQB_PAR | EDMA_ERR_CRPB_PAR | | 1501 | EDMA_ERR_CRQB_PAR | EDMA_ERR_CRPB_PAR | |
1525 | EDMA_ERR_INTRL_PAR)) { | 1502 | EDMA_ERR_INTRL_PAR)) { |
1526 | err_mask |= AC_ERR_ATA_BUS; | 1503 | err_mask |= AC_ERR_ATA_BUS; |
1527 | action |= ATA_EH_HARDRESET; | 1504 | action |= ATA_EH_RESET; |
1528 | ata_ehi_push_desc(ehi, "parity error"); | 1505 | ata_ehi_push_desc(ehi, "parity error"); |
1529 | } | 1506 | } |
1530 | if (edma_err_cause & (EDMA_ERR_DEV_DCON | EDMA_ERR_DEV_CON)) { | 1507 | if (edma_err_cause & (EDMA_ERR_DEV_DCON | EDMA_ERR_DEV_CON)) { |
1531 | ata_ehi_hotplugged(ehi); | 1508 | ata_ehi_hotplugged(ehi); |
1532 | ata_ehi_push_desc(ehi, edma_err_cause & EDMA_ERR_DEV_DCON ? | 1509 | ata_ehi_push_desc(ehi, edma_err_cause & EDMA_ERR_DEV_DCON ? |
1533 | "dev disconnect" : "dev connect"); | 1510 | "dev disconnect" : "dev connect"); |
1534 | action |= ATA_EH_HARDRESET; | 1511 | action |= ATA_EH_RESET; |
1535 | } | 1512 | } |
1536 | 1513 | ||
1537 | if (IS_GEN_I(hpriv)) { | 1514 | if (IS_GEN_I(hpriv)) { |
@@ -1555,7 +1532,7 @@ static void mv_err_intr(struct ata_port *ap, struct ata_queued_cmd *qc) | |||
1555 | sata_scr_read(&ap->link, SCR_ERROR, &serr); | 1532 | sata_scr_read(&ap->link, SCR_ERROR, &serr); |
1556 | sata_scr_write_flush(&ap->link, SCR_ERROR, serr); | 1533 | sata_scr_write_flush(&ap->link, SCR_ERROR, serr); |
1557 | err_mask = AC_ERR_ATA_BUS; | 1534 | err_mask = AC_ERR_ATA_BUS; |
1558 | action |= ATA_EH_HARDRESET; | 1535 | action |= ATA_EH_RESET; |
1559 | } | 1536 | } |
1560 | } | 1537 | } |
1561 | 1538 | ||
@@ -1564,7 +1541,7 @@ static void mv_err_intr(struct ata_port *ap, struct ata_queued_cmd *qc) | |||
1564 | 1541 | ||
1565 | if (!err_mask) { | 1542 | if (!err_mask) { |
1566 | err_mask = AC_ERR_OTHER; | 1543 | err_mask = AC_ERR_OTHER; |
1567 | action |= ATA_EH_HARDRESET; | 1544 | action |= ATA_EH_RESET; |
1568 | } | 1545 | } |
1569 | 1546 | ||
1570 | ehi->serror |= serr; | 1547 | ehi->serror |= serr; |
@@ -1723,9 +1700,9 @@ static void mv_host_intr(struct ata_host *host, u32 relevant, unsigned int hc) | |||
1723 | pp = ap->private_data; | 1700 | pp = ap->private_data; |
1724 | 1701 | ||
1725 | shift = port << 1; /* (port * 2) */ | 1702 | shift = port << 1; /* (port * 2) */ |
1726 | if (port >= MV_PORTS_PER_HC) { | 1703 | if (port >= MV_PORTS_PER_HC) |
1727 | shift++; /* skip bit 8 in the HC Main IRQ reg */ | 1704 | shift++; /* skip bit 8 in the HC Main IRQ reg */ |
1728 | } | 1705 | |
1729 | have_err_bits = ((PORT0_ERR << shift) & relevant); | 1706 | have_err_bits = ((PORT0_ERR << shift) & relevant); |
1730 | 1707 | ||
1731 | if (unlikely(have_err_bits)) { | 1708 | if (unlikely(have_err_bits)) { |
@@ -1780,7 +1757,7 @@ static void mv_pci_error(struct ata_host *host, void __iomem *mmio) | |||
1780 | ata_ehi_push_desc(ehi, | 1757 | ata_ehi_push_desc(ehi, |
1781 | "PCI err cause 0x%08x", err_cause); | 1758 | "PCI err cause 0x%08x", err_cause); |
1782 | err_mask = AC_ERR_HOST_BUS; | 1759 | err_mask = AC_ERR_HOST_BUS; |
1783 | ehi->action = ATA_EH_HARDRESET; | 1760 | ehi->action = ATA_EH_RESET; |
1784 | qc = ata_qc_from_tag(ap, ap->link.active_tag); | 1761 | qc = ata_qc_from_tag(ap, ap->link.active_tag); |
1785 | if (qc) | 1762 | if (qc) |
1786 | qc->err_mask |= err_mask; | 1763 | qc->err_mask |= err_mask; |
@@ -1814,6 +1791,7 @@ static irqreturn_t mv_interrupt(int irq, void *dev_instance) | |||
1814 | void __iomem *mmio = hpriv->base; | 1791 | void __iomem *mmio = hpriv->base; |
1815 | u32 irq_stat, irq_mask; | 1792 | u32 irq_stat, irq_mask; |
1816 | 1793 | ||
1794 | /* Note to self: &host->lock == &ap->host->lock == ap->lock */ | ||
1817 | spin_lock(&host->lock); | 1795 | spin_lock(&host->lock); |
1818 | 1796 | ||
1819 | irq_stat = readl(hpriv->main_cause_reg_addr); | 1797 | irq_stat = readl(hpriv->main_cause_reg_addr); |
@@ -1847,14 +1825,6 @@ out_unlock: | |||
1847 | return IRQ_RETVAL(handled); | 1825 | return IRQ_RETVAL(handled); |
1848 | } | 1826 | } |
1849 | 1827 | ||
1850 | static void __iomem *mv5_phy_base(void __iomem *mmio, unsigned int port) | ||
1851 | { | ||
1852 | void __iomem *hc_mmio = mv_hc_base_from_port(mmio, port); | ||
1853 | unsigned long ofs = (mv_hardport_from_port(port) + 1) * 0x100UL; | ||
1854 | |||
1855 | return hc_mmio + ofs; | ||
1856 | } | ||
1857 | |||
1858 | static unsigned int mv5_scr_offset(unsigned int sc_reg_in) | 1828 | static unsigned int mv5_scr_offset(unsigned int sc_reg_in) |
1859 | { | 1829 | { |
1860 | unsigned int ofs; | 1830 | unsigned int ofs; |
@@ -1980,9 +1950,12 @@ static void mv5_reset_hc_port(struct mv_host_priv *hpriv, void __iomem *mmio, | |||
1980 | { | 1950 | { |
1981 | void __iomem *port_mmio = mv_port_base(mmio, port); | 1951 | void __iomem *port_mmio = mv_port_base(mmio, port); |
1982 | 1952 | ||
1983 | writelfl(EDMA_DS, port_mmio + EDMA_CMD_OFS); | 1953 | /* |
1984 | 1954 | * The datasheet warns against setting ATA_RST when EDMA is active | |
1985 | mv_channel_reset(hpriv, mmio, port); | 1955 | * (but doesn't say what the problem might be). So we first try |
1956 | * to disable the EDMA engine before doing the ATA_RST operation. | ||
1957 | */ | ||
1958 | mv_reset_channel(hpriv, mmio, port); | ||
1986 | 1959 | ||
1987 | ZERO(0x028); /* command */ | 1960 | ZERO(0x028); /* command */ |
1988 | writel(0x11f, port_mmio + EDMA_CFG_OFS); | 1961 | writel(0x11f, port_mmio + EDMA_CFG_OFS); |
@@ -2132,6 +2105,13 @@ static int mv6_reset_hc(struct mv_host_priv *hpriv, void __iomem *mmio, | |||
2132 | printk(KERN_ERR DRV_NAME ": can't clear global reset\n"); | 2105 | printk(KERN_ERR DRV_NAME ": can't clear global reset\n"); |
2133 | rc = 1; | 2106 | rc = 1; |
2134 | } | 2107 | } |
2108 | /* | ||
2109 | * Temporary: wait 3 seconds before port-probing can happen, | ||
2110 | * so that we don't miss finding sleepy SilXXXX port-multipliers. | ||
2111 | * This can go away once hotplug is fully/correctly implemented. | ||
2112 | */ | ||
2113 | if (rc == 0) | ||
2114 | msleep(3000); | ||
2135 | done: | 2115 | done: |
2136 | return rc; | 2116 | return rc; |
2137 | } | 2117 | } |
@@ -2200,14 +2180,15 @@ static void mv6_phy_errata(struct mv_host_priv *hpriv, void __iomem *mmio, | |||
2200 | m4 = readl(port_mmio + PHY_MODE4); | 2180 | m4 = readl(port_mmio + PHY_MODE4); |
2201 | 2181 | ||
2202 | if (hp_flags & MV_HP_ERRATA_60X1B2) | 2182 | if (hp_flags & MV_HP_ERRATA_60X1B2) |
2203 | tmp = readl(port_mmio + 0x310); | 2183 | tmp = readl(port_mmio + PHY_MODE3); |
2204 | 2184 | ||
2185 | /* workaround for errata FEr SATA#10 (part 1) */ | ||
2205 | m4 = (m4 & ~(1 << 1)) | (1 << 0); | 2186 | m4 = (m4 & ~(1 << 1)) | (1 << 0); |
2206 | 2187 | ||
2207 | writel(m4, port_mmio + PHY_MODE4); | 2188 | writel(m4, port_mmio + PHY_MODE4); |
2208 | 2189 | ||
2209 | if (hp_flags & MV_HP_ERRATA_60X1B2) | 2190 | if (hp_flags & MV_HP_ERRATA_60X1B2) |
2210 | writel(tmp, port_mmio + 0x310); | 2191 | writel(tmp, port_mmio + PHY_MODE3); |
2211 | } | 2192 | } |
2212 | 2193 | ||
2213 | /* Revert values of pre-emphasis and signal amps to the saved ones */ | 2194 | /* Revert values of pre-emphasis and signal amps to the saved ones */ |
@@ -2255,9 +2236,12 @@ static void mv_soc_reset_hc_port(struct mv_host_priv *hpriv, | |||
2255 | { | 2236 | { |
2256 | void __iomem *port_mmio = mv_port_base(mmio, port); | 2237 | void __iomem *port_mmio = mv_port_base(mmio, port); |
2257 | 2238 | ||
2258 | writelfl(EDMA_DS, port_mmio + EDMA_CMD_OFS); | 2239 | /* |
2259 | 2240 | * The datasheet warns against setting ATA_RST when EDMA is active | |
2260 | mv_channel_reset(hpriv, mmio, port); | 2241 | * (but doesn't say what the problem might be). So we first try |
2242 | * to disable the EDMA engine before doing the ATA_RST operation. | ||
2243 | */ | ||
2244 | mv_reset_channel(hpriv, mmio, port); | ||
2261 | 2245 | ||
2262 | ZERO(0x028); /* command */ | 2246 | ZERO(0x028); /* command */ |
2263 | writel(0x101f, port_mmio + EDMA_CFG_OFS); | 2247 | writel(0x101f, port_mmio + EDMA_CFG_OFS); |
@@ -2314,25 +2298,39 @@ static void mv_soc_reset_bus(struct ata_host *host, void __iomem *mmio) | |||
2314 | return; | 2298 | return; |
2315 | } | 2299 | } |
2316 | 2300 | ||
2317 | static void mv_channel_reset(struct mv_host_priv *hpriv, void __iomem *mmio, | 2301 | static void mv_setup_ifctl(void __iomem *port_mmio, int want_gen2i) |
2302 | { | ||
2303 | u32 ifctl = readl(port_mmio + SATA_INTERFACE_CFG); | ||
2304 | |||
2305 | ifctl = (ifctl & 0xf7f) | 0x9b1000; /* from chip spec */ | ||
2306 | if (want_gen2i) | ||
2307 | ifctl |= (1 << 7); /* enable gen2i speed */ | ||
2308 | writelfl(ifctl, port_mmio + SATA_INTERFACE_CFG); | ||
2309 | } | ||
2310 | |||
2311 | /* | ||
2312 | * Caller must ensure that EDMA is not active, | ||
2313 | * by first doing mv_stop_edma() where needed. | ||
2314 | */ | ||
2315 | static void mv_reset_channel(struct mv_host_priv *hpriv, void __iomem *mmio, | ||
2318 | unsigned int port_no) | 2316 | unsigned int port_no) |
2319 | { | 2317 | { |
2320 | void __iomem *port_mmio = mv_port_base(mmio, port_no); | 2318 | void __iomem *port_mmio = mv_port_base(mmio, port_no); |
2321 | 2319 | ||
2320 | mv_stop_edma_engine(port_mmio); | ||
2322 | writelfl(ATA_RST, port_mmio + EDMA_CMD_OFS); | 2321 | writelfl(ATA_RST, port_mmio + EDMA_CMD_OFS); |
2323 | 2322 | ||
2324 | if (IS_GEN_II(hpriv)) { | 2323 | if (!IS_GEN_I(hpriv)) { |
2325 | u32 ifctl = readl(port_mmio + SATA_INTERFACE_CTL); | 2324 | /* Enable 3.0gb/s link speed */ |
2326 | ifctl |= (1 << 7); /* enable gen2i speed */ | 2325 | mv_setup_ifctl(port_mmio, 1); |
2327 | ifctl = (ifctl & 0xfff) | 0x9b1000; /* from chip spec */ | ||
2328 | writelfl(ifctl, port_mmio + SATA_INTERFACE_CTL); | ||
2329 | } | 2326 | } |
2330 | 2327 | /* | |
2331 | udelay(25); /* allow reset propagation */ | 2328 | * Strobing ATA_RST here causes a hard reset of the SATA transport, |
2332 | 2329 | * link, and physical layers. It resets all SATA interface registers | |
2333 | /* Spec never mentions clearing the bit. Marvell's driver does | 2330 | * (except for SATA_INTERFACE_CFG), and issues a COMRESET to the dev. |
2334 | * clear the bit, however. | ||
2335 | */ | 2331 | */ |
2332 | writelfl(ATA_RST, port_mmio + EDMA_CMD_OFS); | ||
2333 | udelay(25); /* allow reset propagation */ | ||
2336 | writelfl(0, port_mmio + EDMA_CMD_OFS); | 2334 | writelfl(0, port_mmio + EDMA_CMD_OFS); |
2337 | 2335 | ||
2338 | hpriv->ops->phy_errata(hpriv, mmio, port_no); | 2336 | hpriv->ops->phy_errata(hpriv, mmio, port_no); |
@@ -2341,136 +2339,32 @@ static void mv_channel_reset(struct mv_host_priv *hpriv, void __iomem *mmio, | |||
2341 | mdelay(1); | 2339 | mdelay(1); |
2342 | } | 2340 | } |
2343 | 2341 | ||
2344 | /** | 2342 | static void mv_pmp_select(struct ata_port *ap, int pmp) |
2345 | * mv_phy_reset - Perform eDMA reset followed by COMRESET | ||
2346 | * @ap: ATA channel to manipulate | ||
2347 | * | ||
2348 | * Part of this is taken from __sata_phy_reset and modified to | ||
2349 | * not sleep since this routine gets called from interrupt level. | ||
2350 | * | ||
2351 | * LOCKING: | ||
2352 | * Inherited from caller. This is coded to safe to call at | ||
2353 | * interrupt level, i.e. it does not sleep. | ||
2354 | */ | ||
2355 | static void mv_phy_reset(struct ata_port *ap, unsigned int *class, | ||
2356 | unsigned long deadline) | ||
2357 | { | 2343 | { |
2358 | struct mv_port_priv *pp = ap->private_data; | 2344 | if (sata_pmp_supported(ap)) { |
2359 | struct mv_host_priv *hpriv = ap->host->private_data; | 2345 | void __iomem *port_mmio = mv_ap_base(ap); |
2360 | void __iomem *port_mmio = mv_ap_base(ap); | 2346 | u32 reg = readl(port_mmio + SATA_IFCTL_OFS); |
2361 | int retry = 5; | 2347 | int old = reg & 0xf; |
2362 | u32 sstatus; | ||
2363 | |||
2364 | VPRINTK("ENTER, port %u, mmio 0x%p\n", ap->port_no, port_mmio); | ||
2365 | |||
2366 | #ifdef DEBUG | ||
2367 | { | ||
2368 | u32 sstatus, serror, scontrol; | ||
2369 | |||
2370 | mv_scr_read(ap, SCR_STATUS, &sstatus); | ||
2371 | mv_scr_read(ap, SCR_ERROR, &serror); | ||
2372 | mv_scr_read(ap, SCR_CONTROL, &scontrol); | ||
2373 | DPRINTK("S-regs after ATA_RST: SStat 0x%08x SErr 0x%08x " | ||
2374 | "SCtrl 0x%08x\n", sstatus, serror, scontrol); | ||
2375 | } | ||
2376 | #endif | ||
2377 | |||
2378 | /* Issue COMRESET via SControl */ | ||
2379 | comreset_retry: | ||
2380 | sata_scr_write_flush(&ap->link, SCR_CONTROL, 0x301); | ||
2381 | msleep(1); | ||
2382 | |||
2383 | sata_scr_write_flush(&ap->link, SCR_CONTROL, 0x300); | ||
2384 | msleep(20); | ||
2385 | 2348 | ||
2386 | do { | 2349 | if (old != pmp) { |
2387 | sata_scr_read(&ap->link, SCR_STATUS, &sstatus); | 2350 | reg = (reg & ~0xf) | pmp; |
2388 | if (((sstatus & 0x3) == 3) || ((sstatus & 0x3) == 0)) | 2351 | writelfl(reg, port_mmio + SATA_IFCTL_OFS); |
2389 | break; | 2352 | } |
2390 | |||
2391 | msleep(1); | ||
2392 | } while (time_before(jiffies, deadline)); | ||
2393 | |||
2394 | /* work around errata */ | ||
2395 | if (IS_GEN_II(hpriv) && | ||
2396 | (sstatus != 0x0) && (sstatus != 0x113) && (sstatus != 0x123) && | ||
2397 | (retry-- > 0)) | ||
2398 | goto comreset_retry; | ||
2399 | |||
2400 | #ifdef DEBUG | ||
2401 | { | ||
2402 | u32 sstatus, serror, scontrol; | ||
2403 | |||
2404 | mv_scr_read(ap, SCR_STATUS, &sstatus); | ||
2405 | mv_scr_read(ap, SCR_ERROR, &serror); | ||
2406 | mv_scr_read(ap, SCR_CONTROL, &scontrol); | ||
2407 | DPRINTK("S-regs after PHY wake: SStat 0x%08x SErr 0x%08x " | ||
2408 | "SCtrl 0x%08x\n", sstatus, serror, scontrol); | ||
2409 | } | ||
2410 | #endif | ||
2411 | |||
2412 | if (ata_link_offline(&ap->link)) { | ||
2413 | *class = ATA_DEV_NONE; | ||
2414 | return; | ||
2415 | } | ||
2416 | |||
2417 | /* even after SStatus reflects that device is ready, | ||
2418 | * it seems to take a while for link to be fully | ||
2419 | * established (and thus Status no longer 0x80/0x7F), | ||
2420 | * so we poll a bit for that, here. | ||
2421 | */ | ||
2422 | retry = 20; | ||
2423 | while (1) { | ||
2424 | u8 drv_stat = ata_check_status(ap); | ||
2425 | if ((drv_stat != 0x80) && (drv_stat != 0x7f)) | ||
2426 | break; | ||
2427 | msleep(500); | ||
2428 | if (retry-- <= 0) | ||
2429 | break; | ||
2430 | if (time_after(jiffies, deadline)) | ||
2431 | break; | ||
2432 | } | 2353 | } |
2433 | |||
2434 | /* FIXME: if we passed the deadline, the following | ||
2435 | * code probably produces an invalid result | ||
2436 | */ | ||
2437 | |||
2438 | /* finally, read device signature from TF registers */ | ||
2439 | *class = ata_dev_try_classify(ap->link.device, 1, NULL); | ||
2440 | |||
2441 | writelfl(0, port_mmio + EDMA_ERR_IRQ_CAUSE_OFS); | ||
2442 | |||
2443 | WARN_ON(pp->pp_flags & MV_PP_FLAG_EDMA_EN); | ||
2444 | |||
2445 | VPRINTK("EXIT\n"); | ||
2446 | } | 2354 | } |
2447 | 2355 | ||
2448 | static int mv_prereset(struct ata_link *link, unsigned long deadline) | 2356 | static int mv_pmp_hardreset(struct ata_link *link, unsigned int *class, |
2357 | unsigned long deadline) | ||
2449 | { | 2358 | { |
2450 | struct ata_port *ap = link->ap; | 2359 | mv_pmp_select(link->ap, sata_srst_pmp(link)); |
2451 | struct mv_port_priv *pp = ap->private_data; | 2360 | return sata_std_hardreset(link, class, deadline); |
2452 | struct ata_eh_context *ehc = &link->eh_context; | 2361 | } |
2453 | int rc; | ||
2454 | |||
2455 | rc = mv_stop_dma(ap); | ||
2456 | if (rc) | ||
2457 | ehc->i.action |= ATA_EH_HARDRESET; | ||
2458 | |||
2459 | if (!(pp->pp_flags & MV_PP_FLAG_HAD_A_RESET)) { | ||
2460 | pp->pp_flags |= MV_PP_FLAG_HAD_A_RESET; | ||
2461 | ehc->i.action |= ATA_EH_HARDRESET; | ||
2462 | } | ||
2463 | |||
2464 | /* if we're about to do hardreset, nothing more to do */ | ||
2465 | if (ehc->i.action & ATA_EH_HARDRESET) | ||
2466 | return 0; | ||
2467 | |||
2468 | if (ata_link_online(link)) | ||
2469 | rc = ata_wait_ready(ap, deadline); | ||
2470 | else | ||
2471 | rc = -ENODEV; | ||
2472 | 2362 | ||
2473 | return rc; | 2363 | static int mv_softreset(struct ata_link *link, unsigned int *class, |
2364 | unsigned long deadline) | ||
2365 | { | ||
2366 | mv_pmp_select(link->ap, sata_srst_pmp(link)); | ||
2367 | return ata_sff_softreset(link, class, deadline); | ||
2474 | } | 2368 | } |
2475 | 2369 | ||
2476 | static int mv_hardreset(struct ata_link *link, unsigned int *class, | 2370 | static int mv_hardreset(struct ata_link *link, unsigned int *class, |
@@ -2478,43 +2372,34 @@ static int mv_hardreset(struct ata_link *link, unsigned int *class, | |||
2478 | { | 2372 | { |
2479 | struct ata_port *ap = link->ap; | 2373 | struct ata_port *ap = link->ap; |
2480 | struct mv_host_priv *hpriv = ap->host->private_data; | 2374 | struct mv_host_priv *hpriv = ap->host->private_data; |
2375 | struct mv_port_priv *pp = ap->private_data; | ||
2481 | void __iomem *mmio = hpriv->base; | 2376 | void __iomem *mmio = hpriv->base; |
2377 | int rc, attempts = 0, extra = 0; | ||
2378 | u32 sstatus; | ||
2379 | bool online; | ||
2482 | 2380 | ||
2483 | mv_stop_dma(ap); | 2381 | mv_reset_channel(hpriv, mmio, ap->port_no); |
2484 | 2382 | pp->pp_flags &= ~MV_PP_FLAG_EDMA_EN; | |
2485 | mv_channel_reset(hpriv, mmio, ap->port_no); | ||
2486 | |||
2487 | mv_phy_reset(ap, class, deadline); | ||
2488 | |||
2489 | return 0; | ||
2490 | } | ||
2491 | |||
2492 | static void mv_postreset(struct ata_link *link, unsigned int *classes) | ||
2493 | { | ||
2494 | struct ata_port *ap = link->ap; | ||
2495 | u32 serr; | ||
2496 | |||
2497 | /* print link status */ | ||
2498 | sata_print_link_status(link); | ||
2499 | |||
2500 | /* clear SError */ | ||
2501 | sata_scr_read(link, SCR_ERROR, &serr); | ||
2502 | sata_scr_write_flush(link, SCR_ERROR, serr); | ||
2503 | 2383 | ||
2504 | /* bail out if no device is present */ | 2384 | /* Workaround for errata FEr SATA#10 (part 2) */ |
2505 | if (classes[0] == ATA_DEV_NONE && classes[1] == ATA_DEV_NONE) { | 2385 | do { |
2506 | DPRINTK("EXIT, no device\n"); | 2386 | const unsigned long *timing = |
2507 | return; | 2387 | sata_ehc_deb_timing(&link->eh_context); |
2508 | } | ||
2509 | 2388 | ||
2510 | /* set up device control */ | 2389 | rc = sata_link_hardreset(link, timing, deadline + extra, |
2511 | iowrite8(ap->ctl, ap->ioaddr.ctl_addr); | 2390 | &online, NULL); |
2512 | } | 2391 | if (rc) |
2392 | return rc; | ||
2393 | sata_scr_read(link, SCR_STATUS, &sstatus); | ||
2394 | if (!IS_GEN_I(hpriv) && ++attempts >= 5 && sstatus == 0x121) { | ||
2395 | /* Force 1.5gb/s link speed and try again */ | ||
2396 | mv_setup_ifctl(mv_ap_base(ap), 0); | ||
2397 | if (time_after(jiffies + HZ, deadline)) | ||
2398 | extra = HZ; /* only extend it once, max */ | ||
2399 | } | ||
2400 | } while (sstatus != 0x0 && sstatus != 0x113 && sstatus != 0x123); | ||
2513 | 2401 | ||
2514 | static void mv_error_handler(struct ata_port *ap) | 2402 | return rc; |
2515 | { | ||
2516 | ata_do_eh(ap, mv_prereset, ata_std_softreset, | ||
2517 | mv_hardreset, mv_postreset); | ||
2518 | } | 2403 | } |
2519 | 2404 | ||
2520 | static void mv_eh_freeze(struct ata_port *ap) | 2405 | static void mv_eh_freeze(struct ata_port *ap) |
@@ -2808,19 +2693,6 @@ static int mv_init_host(struct ata_host *host, unsigned int board_idx) | |||
2808 | hpriv->ops->enable_leds(hpriv, mmio); | 2693 | hpriv->ops->enable_leds(hpriv, mmio); |
2809 | 2694 | ||
2810 | for (port = 0; port < host->n_ports; port++) { | 2695 | for (port = 0; port < host->n_ports; port++) { |
2811 | if (IS_GEN_II(hpriv)) { | ||
2812 | void __iomem *port_mmio = mv_port_base(mmio, port); | ||
2813 | |||
2814 | u32 ifctl = readl(port_mmio + SATA_INTERFACE_CTL); | ||
2815 | ifctl |= (1 << 7); /* enable gen2i speed */ | ||
2816 | ifctl = (ifctl & 0xfff) | 0x9b1000; /* from chip spec */ | ||
2817 | writelfl(ifctl, port_mmio + SATA_INTERFACE_CTL); | ||
2818 | } | ||
2819 | |||
2820 | hpriv->ops->phy_errata(hpriv, mmio, port); | ||
2821 | } | ||
2822 | |||
2823 | for (port = 0; port < host->n_ports; port++) { | ||
2824 | struct ata_port *ap = host->ports[port]; | 2696 | struct ata_port *ap = host->ports[port]; |
2825 | void __iomem *port_mmio = mv_port_base(mmio, port); | 2697 | void __iomem *port_mmio = mv_port_base(mmio, port); |
2826 | 2698 | ||
@@ -3192,7 +3064,7 @@ MODULE_DESCRIPTION("SCSI low-level driver for Marvell SATA controllers"); | |||
3192 | MODULE_LICENSE("GPL"); | 3064 | MODULE_LICENSE("GPL"); |
3193 | MODULE_DEVICE_TABLE(pci, mv_pci_tbl); | 3065 | MODULE_DEVICE_TABLE(pci, mv_pci_tbl); |
3194 | MODULE_VERSION(DRV_VERSION); | 3066 | MODULE_VERSION(DRV_VERSION); |
3195 | MODULE_ALIAS("platform:sata_mv"); | 3067 | MODULE_ALIAS("platform:" DRV_NAME); |
3196 | 3068 | ||
3197 | #ifdef CONFIG_PCI | 3069 | #ifdef CONFIG_PCI |
3198 | module_param(msi, int, 0444); | 3070 | module_param(msi, int, 0444); |
diff --git a/drivers/ata/sata_nv.c b/drivers/ata/sata_nv.c index ed5473bf7a0a..109b07495721 100644 --- a/drivers/ata/sata_nv.c +++ b/drivers/ata/sata_nv.c | |||
@@ -309,7 +309,8 @@ static void nv_nf2_freeze(struct ata_port *ap); | |||
309 | static void nv_nf2_thaw(struct ata_port *ap); | 309 | static void nv_nf2_thaw(struct ata_port *ap); |
310 | static void nv_ck804_freeze(struct ata_port *ap); | 310 | static void nv_ck804_freeze(struct ata_port *ap); |
311 | static void nv_ck804_thaw(struct ata_port *ap); | 311 | static void nv_ck804_thaw(struct ata_port *ap); |
312 | static void nv_error_handler(struct ata_port *ap); | 312 | static int nv_hardreset(struct ata_link *link, unsigned int *class, |
313 | unsigned long deadline); | ||
313 | static int nv_adma_slave_config(struct scsi_device *sdev); | 314 | static int nv_adma_slave_config(struct scsi_device *sdev); |
314 | static int nv_adma_check_atapi_dma(struct ata_queued_cmd *qc); | 315 | static int nv_adma_check_atapi_dma(struct ata_queued_cmd *qc); |
315 | static void nv_adma_qc_prep(struct ata_queued_cmd *qc); | 316 | static void nv_adma_qc_prep(struct ata_queued_cmd *qc); |
@@ -385,157 +386,60 @@ static struct pci_driver nv_pci_driver = { | |||
385 | }; | 386 | }; |
386 | 387 | ||
387 | static struct scsi_host_template nv_sht = { | 388 | static struct scsi_host_template nv_sht = { |
388 | .module = THIS_MODULE, | 389 | ATA_BMDMA_SHT(DRV_NAME), |
389 | .name = DRV_NAME, | ||
390 | .ioctl = ata_scsi_ioctl, | ||
391 | .queuecommand = ata_scsi_queuecmd, | ||
392 | .can_queue = ATA_DEF_QUEUE, | ||
393 | .this_id = ATA_SHT_THIS_ID, | ||
394 | .sg_tablesize = LIBATA_MAX_PRD, | ||
395 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
396 | .emulated = ATA_SHT_EMULATED, | ||
397 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
398 | .proc_name = DRV_NAME, | ||
399 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
400 | .slave_configure = ata_scsi_slave_config, | ||
401 | .slave_destroy = ata_scsi_slave_destroy, | ||
402 | .bios_param = ata_std_bios_param, | ||
403 | }; | 390 | }; |
404 | 391 | ||
405 | static struct scsi_host_template nv_adma_sht = { | 392 | static struct scsi_host_template nv_adma_sht = { |
406 | .module = THIS_MODULE, | 393 | ATA_NCQ_SHT(DRV_NAME), |
407 | .name = DRV_NAME, | ||
408 | .ioctl = ata_scsi_ioctl, | ||
409 | .queuecommand = ata_scsi_queuecmd, | ||
410 | .change_queue_depth = ata_scsi_change_queue_depth, | ||
411 | .can_queue = NV_ADMA_MAX_CPBS, | 394 | .can_queue = NV_ADMA_MAX_CPBS, |
412 | .this_id = ATA_SHT_THIS_ID, | ||
413 | .sg_tablesize = NV_ADMA_SGTBL_TOTAL_LEN, | 395 | .sg_tablesize = NV_ADMA_SGTBL_TOTAL_LEN, |
414 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
415 | .emulated = ATA_SHT_EMULATED, | ||
416 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
417 | .proc_name = DRV_NAME, | ||
418 | .dma_boundary = NV_ADMA_DMA_BOUNDARY, | 396 | .dma_boundary = NV_ADMA_DMA_BOUNDARY, |
419 | .slave_configure = nv_adma_slave_config, | 397 | .slave_configure = nv_adma_slave_config, |
420 | .slave_destroy = ata_scsi_slave_destroy, | ||
421 | .bios_param = ata_std_bios_param, | ||
422 | }; | 398 | }; |
423 | 399 | ||
424 | static struct scsi_host_template nv_swncq_sht = { | 400 | static struct scsi_host_template nv_swncq_sht = { |
425 | .module = THIS_MODULE, | 401 | ATA_NCQ_SHT(DRV_NAME), |
426 | .name = DRV_NAME, | ||
427 | .ioctl = ata_scsi_ioctl, | ||
428 | .queuecommand = ata_scsi_queuecmd, | ||
429 | .change_queue_depth = ata_scsi_change_queue_depth, | ||
430 | .can_queue = ATA_MAX_QUEUE, | 402 | .can_queue = ATA_MAX_QUEUE, |
431 | .this_id = ATA_SHT_THIS_ID, | ||
432 | .sg_tablesize = LIBATA_MAX_PRD, | 403 | .sg_tablesize = LIBATA_MAX_PRD, |
433 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
434 | .emulated = ATA_SHT_EMULATED, | ||
435 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
436 | .proc_name = DRV_NAME, | ||
437 | .dma_boundary = ATA_DMA_BOUNDARY, | 404 | .dma_boundary = ATA_DMA_BOUNDARY, |
438 | .slave_configure = nv_swncq_slave_config, | 405 | .slave_configure = nv_swncq_slave_config, |
439 | .slave_destroy = ata_scsi_slave_destroy, | ||
440 | .bios_param = ata_std_bios_param, | ||
441 | }; | 406 | }; |
442 | 407 | ||
443 | static const struct ata_port_operations nv_generic_ops = { | 408 | static struct ata_port_operations nv_generic_ops = { |
444 | .tf_load = ata_tf_load, | 409 | .inherits = &ata_bmdma_port_ops, |
445 | .tf_read = ata_tf_read, | 410 | .hardreset = nv_hardreset, |
446 | .exec_command = ata_exec_command, | ||
447 | .check_status = ata_check_status, | ||
448 | .dev_select = ata_std_dev_select, | ||
449 | .bmdma_setup = ata_bmdma_setup, | ||
450 | .bmdma_start = ata_bmdma_start, | ||
451 | .bmdma_stop = ata_bmdma_stop, | ||
452 | .bmdma_status = ata_bmdma_status, | ||
453 | .qc_prep = ata_qc_prep, | ||
454 | .qc_issue = ata_qc_issue_prot, | ||
455 | .freeze = ata_bmdma_freeze, | ||
456 | .thaw = ata_bmdma_thaw, | ||
457 | .error_handler = nv_error_handler, | ||
458 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
459 | .data_xfer = ata_data_xfer, | ||
460 | .irq_clear = ata_bmdma_irq_clear, | ||
461 | .irq_on = ata_irq_on, | ||
462 | .scr_read = nv_scr_read, | 411 | .scr_read = nv_scr_read, |
463 | .scr_write = nv_scr_write, | 412 | .scr_write = nv_scr_write, |
464 | .port_start = ata_port_start, | ||
465 | }; | 413 | }; |
466 | 414 | ||
467 | static const struct ata_port_operations nv_nf2_ops = { | 415 | static struct ata_port_operations nv_nf2_ops = { |
468 | .tf_load = ata_tf_load, | 416 | .inherits = &nv_generic_ops, |
469 | .tf_read = ata_tf_read, | ||
470 | .exec_command = ata_exec_command, | ||
471 | .check_status = ata_check_status, | ||
472 | .dev_select = ata_std_dev_select, | ||
473 | .bmdma_setup = ata_bmdma_setup, | ||
474 | .bmdma_start = ata_bmdma_start, | ||
475 | .bmdma_stop = ata_bmdma_stop, | ||
476 | .bmdma_status = ata_bmdma_status, | ||
477 | .qc_prep = ata_qc_prep, | ||
478 | .qc_issue = ata_qc_issue_prot, | ||
479 | .freeze = nv_nf2_freeze, | 417 | .freeze = nv_nf2_freeze, |
480 | .thaw = nv_nf2_thaw, | 418 | .thaw = nv_nf2_thaw, |
481 | .error_handler = nv_error_handler, | ||
482 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
483 | .data_xfer = ata_data_xfer, | ||
484 | .irq_clear = ata_bmdma_irq_clear, | ||
485 | .irq_on = ata_irq_on, | ||
486 | .scr_read = nv_scr_read, | ||
487 | .scr_write = nv_scr_write, | ||
488 | .port_start = ata_port_start, | ||
489 | }; | 419 | }; |
490 | 420 | ||
491 | static const struct ata_port_operations nv_ck804_ops = { | 421 | static struct ata_port_operations nv_ck804_ops = { |
492 | .tf_load = ata_tf_load, | 422 | .inherits = &nv_generic_ops, |
493 | .tf_read = ata_tf_read, | ||
494 | .exec_command = ata_exec_command, | ||
495 | .check_status = ata_check_status, | ||
496 | .dev_select = ata_std_dev_select, | ||
497 | .bmdma_setup = ata_bmdma_setup, | ||
498 | .bmdma_start = ata_bmdma_start, | ||
499 | .bmdma_stop = ata_bmdma_stop, | ||
500 | .bmdma_status = ata_bmdma_status, | ||
501 | .qc_prep = ata_qc_prep, | ||
502 | .qc_issue = ata_qc_issue_prot, | ||
503 | .freeze = nv_ck804_freeze, | 423 | .freeze = nv_ck804_freeze, |
504 | .thaw = nv_ck804_thaw, | 424 | .thaw = nv_ck804_thaw, |
505 | .error_handler = nv_error_handler, | ||
506 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
507 | .data_xfer = ata_data_xfer, | ||
508 | .irq_clear = ata_bmdma_irq_clear, | ||
509 | .irq_on = ata_irq_on, | ||
510 | .scr_read = nv_scr_read, | ||
511 | .scr_write = nv_scr_write, | ||
512 | .port_start = ata_port_start, | ||
513 | .host_stop = nv_ck804_host_stop, | 425 | .host_stop = nv_ck804_host_stop, |
514 | }; | 426 | }; |
515 | 427 | ||
516 | static const struct ata_port_operations nv_adma_ops = { | 428 | static struct ata_port_operations nv_adma_ops = { |
517 | .tf_load = ata_tf_load, | 429 | .inherits = &nv_generic_ops, |
518 | .tf_read = nv_adma_tf_read, | 430 | |
519 | .check_atapi_dma = nv_adma_check_atapi_dma, | 431 | .check_atapi_dma = nv_adma_check_atapi_dma, |
520 | .exec_command = ata_exec_command, | 432 | .sff_tf_read = nv_adma_tf_read, |
521 | .check_status = ata_check_status, | ||
522 | .dev_select = ata_std_dev_select, | ||
523 | .bmdma_setup = ata_bmdma_setup, | ||
524 | .bmdma_start = ata_bmdma_start, | ||
525 | .bmdma_stop = ata_bmdma_stop, | ||
526 | .bmdma_status = ata_bmdma_status, | ||
527 | .qc_defer = ata_std_qc_defer, | 433 | .qc_defer = ata_std_qc_defer, |
528 | .qc_prep = nv_adma_qc_prep, | 434 | .qc_prep = nv_adma_qc_prep, |
529 | .qc_issue = nv_adma_qc_issue, | 435 | .qc_issue = nv_adma_qc_issue, |
436 | .sff_irq_clear = nv_adma_irq_clear, | ||
437 | |||
530 | .freeze = nv_adma_freeze, | 438 | .freeze = nv_adma_freeze, |
531 | .thaw = nv_adma_thaw, | 439 | .thaw = nv_adma_thaw, |
532 | .error_handler = nv_adma_error_handler, | 440 | .error_handler = nv_adma_error_handler, |
533 | .post_internal_cmd = nv_adma_post_internal_cmd, | 441 | .post_internal_cmd = nv_adma_post_internal_cmd, |
534 | .data_xfer = ata_data_xfer, | 442 | |
535 | .irq_clear = nv_adma_irq_clear, | ||
536 | .irq_on = ata_irq_on, | ||
537 | .scr_read = nv_scr_read, | ||
538 | .scr_write = nv_scr_write, | ||
539 | .port_start = nv_adma_port_start, | 443 | .port_start = nv_adma_port_start, |
540 | .port_stop = nv_adma_port_stop, | 444 | .port_stop = nv_adma_port_stop, |
541 | #ifdef CONFIG_PM | 445 | #ifdef CONFIG_PM |
@@ -545,28 +449,17 @@ static const struct ata_port_operations nv_adma_ops = { | |||
545 | .host_stop = nv_adma_host_stop, | 449 | .host_stop = nv_adma_host_stop, |
546 | }; | 450 | }; |
547 | 451 | ||
548 | static const struct ata_port_operations nv_swncq_ops = { | 452 | static struct ata_port_operations nv_swncq_ops = { |
549 | .tf_load = ata_tf_load, | 453 | .inherits = &nv_generic_ops, |
550 | .tf_read = ata_tf_read, | 454 | |
551 | .exec_command = ata_exec_command, | ||
552 | .check_status = ata_check_status, | ||
553 | .dev_select = ata_std_dev_select, | ||
554 | .bmdma_setup = ata_bmdma_setup, | ||
555 | .bmdma_start = ata_bmdma_start, | ||
556 | .bmdma_stop = ata_bmdma_stop, | ||
557 | .bmdma_status = ata_bmdma_status, | ||
558 | .qc_defer = ata_std_qc_defer, | 455 | .qc_defer = ata_std_qc_defer, |
559 | .qc_prep = nv_swncq_qc_prep, | 456 | .qc_prep = nv_swncq_qc_prep, |
560 | .qc_issue = nv_swncq_qc_issue, | 457 | .qc_issue = nv_swncq_qc_issue, |
458 | |||
561 | .freeze = nv_mcp55_freeze, | 459 | .freeze = nv_mcp55_freeze, |
562 | .thaw = nv_mcp55_thaw, | 460 | .thaw = nv_mcp55_thaw, |
563 | .error_handler = nv_swncq_error_handler, | 461 | .error_handler = nv_swncq_error_handler, |
564 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | 462 | |
565 | .data_xfer = ata_data_xfer, | ||
566 | .irq_clear = ata_bmdma_irq_clear, | ||
567 | .irq_on = ata_irq_on, | ||
568 | .scr_read = nv_scr_read, | ||
569 | .scr_write = nv_scr_write, | ||
570 | #ifdef CONFIG_PM | 463 | #ifdef CONFIG_PM |
571 | .port_suspend = nv_swncq_port_suspend, | 464 | .port_suspend = nv_swncq_port_suspend, |
572 | .port_resume = nv_swncq_port_resume, | 465 | .port_resume = nv_swncq_port_resume, |
@@ -574,63 +467,61 @@ static const struct ata_port_operations nv_swncq_ops = { | |||
574 | .port_start = nv_swncq_port_start, | 467 | .port_start = nv_swncq_port_start, |
575 | }; | 468 | }; |
576 | 469 | ||
470 | struct nv_pi_priv { | ||
471 | irq_handler_t irq_handler; | ||
472 | struct scsi_host_template *sht; | ||
473 | }; | ||
474 | |||
475 | #define NV_PI_PRIV(_irq_handler, _sht) \ | ||
476 | &(struct nv_pi_priv){ .irq_handler = _irq_handler, .sht = _sht } | ||
477 | |||
577 | static const struct ata_port_info nv_port_info[] = { | 478 | static const struct ata_port_info nv_port_info[] = { |
578 | /* generic */ | 479 | /* generic */ |
579 | { | 480 | { |
580 | .sht = &nv_sht, | ||
581 | .flags = ATA_FLAG_SATA | ATA_FLAG_NO_LEGACY, | 481 | .flags = ATA_FLAG_SATA | ATA_FLAG_NO_LEGACY, |
582 | .link_flags = ATA_LFLAG_HRST_TO_RESUME, | ||
583 | .pio_mask = NV_PIO_MASK, | 482 | .pio_mask = NV_PIO_MASK, |
584 | .mwdma_mask = NV_MWDMA_MASK, | 483 | .mwdma_mask = NV_MWDMA_MASK, |
585 | .udma_mask = NV_UDMA_MASK, | 484 | .udma_mask = NV_UDMA_MASK, |
586 | .port_ops = &nv_generic_ops, | 485 | .port_ops = &nv_generic_ops, |
587 | .irq_handler = nv_generic_interrupt, | 486 | .private_data = NV_PI_PRIV(nv_generic_interrupt, &nv_sht), |
588 | }, | 487 | }, |
589 | /* nforce2/3 */ | 488 | /* nforce2/3 */ |
590 | { | 489 | { |
591 | .sht = &nv_sht, | ||
592 | .flags = ATA_FLAG_SATA | ATA_FLAG_NO_LEGACY, | 490 | .flags = ATA_FLAG_SATA | ATA_FLAG_NO_LEGACY, |
593 | .link_flags = ATA_LFLAG_HRST_TO_RESUME, | ||
594 | .pio_mask = NV_PIO_MASK, | 491 | .pio_mask = NV_PIO_MASK, |
595 | .mwdma_mask = NV_MWDMA_MASK, | 492 | .mwdma_mask = NV_MWDMA_MASK, |
596 | .udma_mask = NV_UDMA_MASK, | 493 | .udma_mask = NV_UDMA_MASK, |
597 | .port_ops = &nv_nf2_ops, | 494 | .port_ops = &nv_nf2_ops, |
598 | .irq_handler = nv_nf2_interrupt, | 495 | .private_data = NV_PI_PRIV(nv_nf2_interrupt, &nv_sht), |
599 | }, | 496 | }, |
600 | /* ck804 */ | 497 | /* ck804 */ |
601 | { | 498 | { |
602 | .sht = &nv_sht, | ||
603 | .flags = ATA_FLAG_SATA | ATA_FLAG_NO_LEGACY, | 499 | .flags = ATA_FLAG_SATA | ATA_FLAG_NO_LEGACY, |
604 | .link_flags = ATA_LFLAG_HRST_TO_RESUME, | ||
605 | .pio_mask = NV_PIO_MASK, | 500 | .pio_mask = NV_PIO_MASK, |
606 | .mwdma_mask = NV_MWDMA_MASK, | 501 | .mwdma_mask = NV_MWDMA_MASK, |
607 | .udma_mask = NV_UDMA_MASK, | 502 | .udma_mask = NV_UDMA_MASK, |
608 | .port_ops = &nv_ck804_ops, | 503 | .port_ops = &nv_ck804_ops, |
609 | .irq_handler = nv_ck804_interrupt, | 504 | .private_data = NV_PI_PRIV(nv_ck804_interrupt, &nv_sht), |
610 | }, | 505 | }, |
611 | /* ADMA */ | 506 | /* ADMA */ |
612 | { | 507 | { |
613 | .sht = &nv_adma_sht, | ||
614 | .flags = ATA_FLAG_SATA | ATA_FLAG_NO_LEGACY | | 508 | .flags = ATA_FLAG_SATA | ATA_FLAG_NO_LEGACY | |
615 | ATA_FLAG_MMIO | ATA_FLAG_NCQ, | 509 | ATA_FLAG_MMIO | ATA_FLAG_NCQ, |
616 | .link_flags = ATA_LFLAG_HRST_TO_RESUME, | ||
617 | .pio_mask = NV_PIO_MASK, | 510 | .pio_mask = NV_PIO_MASK, |
618 | .mwdma_mask = NV_MWDMA_MASK, | 511 | .mwdma_mask = NV_MWDMA_MASK, |
619 | .udma_mask = NV_UDMA_MASK, | 512 | .udma_mask = NV_UDMA_MASK, |
620 | .port_ops = &nv_adma_ops, | 513 | .port_ops = &nv_adma_ops, |
621 | .irq_handler = nv_adma_interrupt, | 514 | .private_data = NV_PI_PRIV(nv_adma_interrupt, &nv_adma_sht), |
622 | }, | 515 | }, |
623 | /* SWNCQ */ | 516 | /* SWNCQ */ |
624 | { | 517 | { |
625 | .sht = &nv_swncq_sht, | ||
626 | .flags = ATA_FLAG_SATA | ATA_FLAG_NO_LEGACY | | 518 | .flags = ATA_FLAG_SATA | ATA_FLAG_NO_LEGACY | |
627 | ATA_FLAG_NCQ, | 519 | ATA_FLAG_NCQ, |
628 | .link_flags = ATA_LFLAG_HRST_TO_RESUME, | ||
629 | .pio_mask = NV_PIO_MASK, | 520 | .pio_mask = NV_PIO_MASK, |
630 | .mwdma_mask = NV_MWDMA_MASK, | 521 | .mwdma_mask = NV_MWDMA_MASK, |
631 | .udma_mask = NV_UDMA_MASK, | 522 | .udma_mask = NV_UDMA_MASK, |
632 | .port_ops = &nv_swncq_ops, | 523 | .port_ops = &nv_swncq_ops, |
633 | .irq_handler = nv_swncq_interrupt, | 524 | .private_data = NV_PI_PRIV(nv_swncq_interrupt, &nv_swncq_sht), |
634 | }, | 525 | }, |
635 | }; | 526 | }; |
636 | 527 | ||
@@ -640,8 +531,8 @@ MODULE_LICENSE("GPL"); | |||
640 | MODULE_DEVICE_TABLE(pci, nv_pci_tbl); | 531 | MODULE_DEVICE_TABLE(pci, nv_pci_tbl); |
641 | MODULE_VERSION(DRV_VERSION); | 532 | MODULE_VERSION(DRV_VERSION); |
642 | 533 | ||
643 | static int adma_enabled = 1; | 534 | static int adma_enabled; |
644 | static int swncq_enabled; | 535 | static int swncq_enabled = 1; |
645 | 536 | ||
646 | static void nv_adma_register_mode(struct ata_port *ap) | 537 | static void nv_adma_register_mode(struct ata_port *ap) |
647 | { | 538 | { |
@@ -839,7 +730,7 @@ static void nv_adma_tf_read(struct ata_port *ap, struct ata_taskfile *tf) | |||
839 | ADMA mode could abort outstanding commands. */ | 730 | ADMA mode could abort outstanding commands. */ |
840 | nv_adma_register_mode(ap); | 731 | nv_adma_register_mode(ap); |
841 | 732 | ||
842 | ata_tf_read(ap, tf); | 733 | ata_sff_tf_read(ap, tf); |
843 | } | 734 | } |
844 | 735 | ||
845 | static unsigned int nv_adma_tf_to_cpb(struct ata_taskfile *tf, __le16 *cpb) | 736 | static unsigned int nv_adma_tf_to_cpb(struct ata_taskfile *tf, __le16 *cpb) |
@@ -929,7 +820,7 @@ static int nv_adma_check_cpb(struct ata_port *ap, int cpb_num, int force_err) | |||
929 | "notifier for tag %d with no cmd?\n", | 820 | "notifier for tag %d with no cmd?\n", |
930 | cpb_num); | 821 | cpb_num); |
931 | ehi->err_mask |= AC_ERR_HSM; | 822 | ehi->err_mask |= AC_ERR_HSM; |
932 | ehi->action |= ATA_EH_SOFTRESET; | 823 | ehi->action |= ATA_EH_RESET; |
933 | ata_port_freeze(ap); | 824 | ata_port_freeze(ap); |
934 | return 1; | 825 | return 1; |
935 | } | 826 | } |
@@ -953,12 +844,12 @@ static int nv_host_intr(struct ata_port *ap, u8 irq_stat) | |||
953 | 844 | ||
954 | /* DEV interrupt w/ no active qc? */ | 845 | /* DEV interrupt w/ no active qc? */ |
955 | if (unlikely(!qc || (qc->tf.flags & ATA_TFLAG_POLLING))) { | 846 | if (unlikely(!qc || (qc->tf.flags & ATA_TFLAG_POLLING))) { |
956 | ata_check_status(ap); | 847 | ata_sff_check_status(ap); |
957 | return 1; | 848 | return 1; |
958 | } | 849 | } |
959 | 850 | ||
960 | /* handle interrupt */ | 851 | /* handle interrupt */ |
961 | return ata_host_intr(ap, qc); | 852 | return ata_sff_host_intr(ap, qc); |
962 | } | 853 | } |
963 | 854 | ||
964 | static irqreturn_t nv_adma_interrupt(int irq, void *dev_instance) | 855 | static irqreturn_t nv_adma_interrupt(int irq, void *dev_instance) |
@@ -1137,7 +1028,7 @@ static void nv_adma_irq_clear(struct ata_port *ap) | |||
1137 | u32 notifier_clears[2]; | 1028 | u32 notifier_clears[2]; |
1138 | 1029 | ||
1139 | if (pp->flags & NV_ADMA_ATAPI_SETUP_COMPLETE) { | 1030 | if (pp->flags & NV_ADMA_ATAPI_SETUP_COMPLETE) { |
1140 | ata_bmdma_irq_clear(ap); | 1031 | ata_sff_irq_clear(ap); |
1141 | return; | 1032 | return; |
1142 | } | 1033 | } |
1143 | 1034 | ||
@@ -1168,7 +1059,7 @@ static void nv_adma_post_internal_cmd(struct ata_queued_cmd *qc) | |||
1168 | struct nv_adma_port_priv *pp = qc->ap->private_data; | 1059 | struct nv_adma_port_priv *pp = qc->ap->private_data; |
1169 | 1060 | ||
1170 | if (pp->flags & NV_ADMA_PORT_REGISTER_MODE) | 1061 | if (pp->flags & NV_ADMA_PORT_REGISTER_MODE) |
1171 | ata_bmdma_post_internal_cmd(qc); | 1062 | ata_sff_post_internal_cmd(qc); |
1172 | } | 1063 | } |
1173 | 1064 | ||
1174 | static int nv_adma_port_start(struct ata_port *ap) | 1065 | static int nv_adma_port_start(struct ata_port *ap) |
@@ -1445,7 +1336,7 @@ static void nv_adma_qc_prep(struct ata_queued_cmd *qc) | |||
1445 | BUG_ON(!(pp->flags & NV_ADMA_ATAPI_SETUP_COMPLETE) && | 1336 | BUG_ON(!(pp->flags & NV_ADMA_ATAPI_SETUP_COMPLETE) && |
1446 | (qc->flags & ATA_QCFLAG_DMAMAP)); | 1337 | (qc->flags & ATA_QCFLAG_DMAMAP)); |
1447 | nv_adma_register_mode(qc->ap); | 1338 | nv_adma_register_mode(qc->ap); |
1448 | ata_qc_prep(qc); | 1339 | ata_sff_qc_prep(qc); |
1449 | return; | 1340 | return; |
1450 | } | 1341 | } |
1451 | 1342 | ||
@@ -1504,7 +1395,7 @@ static unsigned int nv_adma_qc_issue(struct ata_queued_cmd *qc) | |||
1504 | BUG_ON(!(pp->flags & NV_ADMA_ATAPI_SETUP_COMPLETE) && | 1395 | BUG_ON(!(pp->flags & NV_ADMA_ATAPI_SETUP_COMPLETE) && |
1505 | (qc->flags & ATA_QCFLAG_DMAMAP)); | 1396 | (qc->flags & ATA_QCFLAG_DMAMAP)); |
1506 | nv_adma_register_mode(qc->ap); | 1397 | nv_adma_register_mode(qc->ap); |
1507 | return ata_qc_issue_prot(qc); | 1398 | return ata_sff_qc_issue(qc); |
1508 | } else | 1399 | } else |
1509 | nv_adma_mode(qc->ap); | 1400 | nv_adma_mode(qc->ap); |
1510 | 1401 | ||
@@ -1545,11 +1436,11 @@ static irqreturn_t nv_generic_interrupt(int irq, void *dev_instance) | |||
1545 | 1436 | ||
1546 | qc = ata_qc_from_tag(ap, ap->link.active_tag); | 1437 | qc = ata_qc_from_tag(ap, ap->link.active_tag); |
1547 | if (qc && (!(qc->tf.flags & ATA_TFLAG_POLLING))) | 1438 | if (qc && (!(qc->tf.flags & ATA_TFLAG_POLLING))) |
1548 | handled += ata_host_intr(ap, qc); | 1439 | handled += ata_sff_host_intr(ap, qc); |
1549 | else | 1440 | else |
1550 | // No request pending? Clear interrupt status | 1441 | // No request pending? Clear interrupt status |
1551 | // anyway, in case there's one pending. | 1442 | // anyway, in case there's one pending. |
1552 | ap->ops->check_status(ap); | 1443 | ap->ops->sff_check_status(ap); |
1553 | } | 1444 | } |
1554 | 1445 | ||
1555 | } | 1446 | } |
@@ -1680,7 +1571,7 @@ static void nv_mcp55_freeze(struct ata_port *ap) | |||
1680 | mask = readl(mmio_base + NV_INT_ENABLE_MCP55); | 1571 | mask = readl(mmio_base + NV_INT_ENABLE_MCP55); |
1681 | mask &= ~(NV_INT_ALL_MCP55 << shift); | 1572 | mask &= ~(NV_INT_ALL_MCP55 << shift); |
1682 | writel(mask, mmio_base + NV_INT_ENABLE_MCP55); | 1573 | writel(mask, mmio_base + NV_INT_ENABLE_MCP55); |
1683 | ata_bmdma_freeze(ap); | 1574 | ata_sff_freeze(ap); |
1684 | } | 1575 | } |
1685 | 1576 | ||
1686 | static void nv_mcp55_thaw(struct ata_port *ap) | 1577 | static void nv_mcp55_thaw(struct ata_port *ap) |
@@ -1694,7 +1585,7 @@ static void nv_mcp55_thaw(struct ata_port *ap) | |||
1694 | mask = readl(mmio_base + NV_INT_ENABLE_MCP55); | 1585 | mask = readl(mmio_base + NV_INT_ENABLE_MCP55); |
1695 | mask |= (NV_INT_MASK_MCP55 << shift); | 1586 | mask |= (NV_INT_MASK_MCP55 << shift); |
1696 | writel(mask, mmio_base + NV_INT_ENABLE_MCP55); | 1587 | writel(mask, mmio_base + NV_INT_ENABLE_MCP55); |
1697 | ata_bmdma_thaw(ap); | 1588 | ata_sff_thaw(ap); |
1698 | } | 1589 | } |
1699 | 1590 | ||
1700 | static int nv_hardreset(struct ata_link *link, unsigned int *class, | 1591 | static int nv_hardreset(struct ata_link *link, unsigned int *class, |
@@ -1706,13 +1597,7 @@ static int nv_hardreset(struct ata_link *link, unsigned int *class, | |||
1706 | * some controllers. Don't classify on hardreset. For more | 1597 | * some controllers. Don't classify on hardreset. For more |
1707 | * info, see http://bugzilla.kernel.org/show_bug.cgi?id=3352 | 1598 | * info, see http://bugzilla.kernel.org/show_bug.cgi?id=3352 |
1708 | */ | 1599 | */ |
1709 | return sata_std_hardreset(link, &dummy, deadline); | 1600 | return sata_sff_hardreset(link, &dummy, deadline); |
1710 | } | ||
1711 | |||
1712 | static void nv_error_handler(struct ata_port *ap) | ||
1713 | { | ||
1714 | ata_bmdma_drive_eh(ap, ata_std_prereset, ata_std_softreset, | ||
1715 | nv_hardreset, ata_std_postreset); | ||
1716 | } | 1601 | } |
1717 | 1602 | ||
1718 | static void nv_adma_error_handler(struct ata_port *ap) | 1603 | static void nv_adma_error_handler(struct ata_port *ap) |
@@ -1768,8 +1653,7 @@ static void nv_adma_error_handler(struct ata_port *ap) | |||
1768 | readw(mmio + NV_ADMA_CTL); /* flush posted write */ | 1653 | readw(mmio + NV_ADMA_CTL); /* flush posted write */ |
1769 | } | 1654 | } |
1770 | 1655 | ||
1771 | ata_bmdma_drive_eh(ap, ata_std_prereset, ata_std_softreset, | 1656 | ata_sff_error_handler(ap); |
1772 | nv_hardreset, ata_std_postreset); | ||
1773 | } | 1657 | } |
1774 | 1658 | ||
1775 | static void nv_swncq_qc_to_dq(struct ata_port *ap, struct ata_queued_cmd *qc) | 1659 | static void nv_swncq_qc_to_dq(struct ata_port *ap, struct ata_queued_cmd *qc) |
@@ -1855,7 +1739,7 @@ static void nv_swncq_ncq_stop(struct ata_port *ap) | |||
1855 | pp->dhfis_bits, pp->dmafis_bits, pp->sdbfis_bits); | 1739 | pp->dhfis_bits, pp->dmafis_bits, pp->sdbfis_bits); |
1856 | 1740 | ||
1857 | ata_port_printk(ap, KERN_ERR, "ATA_REG 0x%X ERR_REG 0x%X\n", | 1741 | ata_port_printk(ap, KERN_ERR, "ATA_REG 0x%X ERR_REG 0x%X\n", |
1858 | ap->ops->check_status(ap), | 1742 | ap->ops->sff_check_status(ap), |
1859 | ioread8(ap->ioaddr.error_addr)); | 1743 | ioread8(ap->ioaddr.error_addr)); |
1860 | 1744 | ||
1861 | sactive = readl(pp->sactive_block); | 1745 | sactive = readl(pp->sactive_block); |
@@ -1881,7 +1765,7 @@ static void nv_swncq_ncq_stop(struct ata_port *ap) | |||
1881 | } | 1765 | } |
1882 | 1766 | ||
1883 | nv_swncq_pp_reinit(ap); | 1767 | nv_swncq_pp_reinit(ap); |
1884 | ap->ops->irq_clear(ap); | 1768 | ap->ops->sff_irq_clear(ap); |
1885 | __ata_bmdma_stop(ap); | 1769 | __ata_bmdma_stop(ap); |
1886 | nv_swncq_irq_clear(ap, 0xffff); | 1770 | nv_swncq_irq_clear(ap, 0xffff); |
1887 | } | 1771 | } |
@@ -1892,11 +1776,10 @@ static void nv_swncq_error_handler(struct ata_port *ap) | |||
1892 | 1776 | ||
1893 | if (ap->link.sactive) { | 1777 | if (ap->link.sactive) { |
1894 | nv_swncq_ncq_stop(ap); | 1778 | nv_swncq_ncq_stop(ap); |
1895 | ehc->i.action |= ATA_EH_HARDRESET; | 1779 | ehc->i.action |= ATA_EH_RESET; |
1896 | } | 1780 | } |
1897 | 1781 | ||
1898 | ata_bmdma_drive_eh(ap, ata_std_prereset, ata_std_softreset, | 1782 | ata_sff_error_handler(ap); |
1899 | nv_hardreset, ata_std_postreset); | ||
1900 | } | 1783 | } |
1901 | 1784 | ||
1902 | #ifdef CONFIG_PM | 1785 | #ifdef CONFIG_PM |
@@ -2042,7 +1925,7 @@ static int nv_swncq_port_start(struct ata_port *ap) | |||
2042 | static void nv_swncq_qc_prep(struct ata_queued_cmd *qc) | 1925 | static void nv_swncq_qc_prep(struct ata_queued_cmd *qc) |
2043 | { | 1926 | { |
2044 | if (qc->tf.protocol != ATA_PROT_NCQ) { | 1927 | if (qc->tf.protocol != ATA_PROT_NCQ) { |
2045 | ata_qc_prep(qc); | 1928 | ata_sff_qc_prep(qc); |
2046 | return; | 1929 | return; |
2047 | } | 1930 | } |
2048 | 1931 | ||
@@ -2104,8 +1987,8 @@ static unsigned int nv_swncq_issue_atacmd(struct ata_port *ap, | |||
2104 | pp->dmafis_bits &= ~(1 << qc->tag); | 1987 | pp->dmafis_bits &= ~(1 << qc->tag); |
2105 | pp->qc_active |= (0x1 << qc->tag); | 1988 | pp->qc_active |= (0x1 << qc->tag); |
2106 | 1989 | ||
2107 | ap->ops->tf_load(ap, &qc->tf); /* load tf registers */ | 1990 | ap->ops->sff_tf_load(ap, &qc->tf); /* load tf registers */ |
2108 | ap->ops->exec_command(ap, &qc->tf); | 1991 | ap->ops->sff_exec_command(ap, &qc->tf); |
2109 | 1992 | ||
2110 | DPRINTK("Issued tag %u\n", qc->tag); | 1993 | DPRINTK("Issued tag %u\n", qc->tag); |
2111 | 1994 | ||
@@ -2118,7 +2001,7 @@ static unsigned int nv_swncq_qc_issue(struct ata_queued_cmd *qc) | |||
2118 | struct nv_swncq_port_priv *pp = ap->private_data; | 2001 | struct nv_swncq_port_priv *pp = ap->private_data; |
2119 | 2002 | ||
2120 | if (qc->tf.protocol != ATA_PROT_NCQ) | 2003 | if (qc->tf.protocol != ATA_PROT_NCQ) |
2121 | return ata_qc_issue_prot(qc); | 2004 | return ata_sff_qc_issue(qc); |
2122 | 2005 | ||
2123 | DPRINTK("Enter\n"); | 2006 | DPRINTK("Enter\n"); |
2124 | 2007 | ||
@@ -2173,11 +2056,11 @@ static int nv_swncq_sdbfis(struct ata_port *ap) | |||
2173 | ata_ehi_clear_desc(ehi); | 2056 | ata_ehi_clear_desc(ehi); |
2174 | ata_ehi_push_desc(ehi, "BMDMA stat 0x%x", host_stat); | 2057 | ata_ehi_push_desc(ehi, "BMDMA stat 0x%x", host_stat); |
2175 | ehi->err_mask |= AC_ERR_HOST_BUS; | 2058 | ehi->err_mask |= AC_ERR_HOST_BUS; |
2176 | ehi->action |= ATA_EH_SOFTRESET; | 2059 | ehi->action |= ATA_EH_RESET; |
2177 | return -EINVAL; | 2060 | return -EINVAL; |
2178 | } | 2061 | } |
2179 | 2062 | ||
2180 | ap->ops->irq_clear(ap); | 2063 | ap->ops->sff_irq_clear(ap); |
2181 | __ata_bmdma_stop(ap); | 2064 | __ata_bmdma_stop(ap); |
2182 | 2065 | ||
2183 | sactive = readl(pp->sactive_block); | 2066 | sactive = readl(pp->sactive_block); |
@@ -2188,7 +2071,7 @@ static int nv_swncq_sdbfis(struct ata_port *ap) | |||
2188 | ata_ehi_push_desc(ehi, "illegal SWNCQ:qc_active transition" | 2071 | ata_ehi_push_desc(ehi, "illegal SWNCQ:qc_active transition" |
2189 | "(%08x->%08x)", pp->qc_active, sactive); | 2072 | "(%08x->%08x)", pp->qc_active, sactive); |
2190 | ehi->err_mask |= AC_ERR_HSM; | 2073 | ehi->err_mask |= AC_ERR_HSM; |
2191 | ehi->action |= ATA_EH_HARDRESET; | 2074 | ehi->action |= ATA_EH_RESET; |
2192 | return -EINVAL; | 2075 | return -EINVAL; |
2193 | } | 2076 | } |
2194 | for (i = 0; i < ATA_MAX_QUEUE; i++) { | 2077 | for (i = 0; i < ATA_MAX_QUEUE; i++) { |
@@ -2299,7 +2182,7 @@ static void nv_swncq_host_interrupt(struct ata_port *ap, u16 fis) | |||
2299 | u8 ata_stat; | 2182 | u8 ata_stat; |
2300 | int rc = 0; | 2183 | int rc = 0; |
2301 | 2184 | ||
2302 | ata_stat = ap->ops->check_status(ap); | 2185 | ata_stat = ap->ops->sff_check_status(ap); |
2303 | nv_swncq_irq_clear(ap, fis); | 2186 | nv_swncq_irq_clear(ap, fis); |
2304 | if (!fis) | 2187 | if (!fis) |
2305 | return; | 2188 | return; |
@@ -2324,7 +2207,7 @@ static void nv_swncq_host_interrupt(struct ata_port *ap, u16 fis) | |||
2324 | ata_ehi_push_desc(ehi, "Ata error. fis:0x%X", fis); | 2207 | ata_ehi_push_desc(ehi, "Ata error. fis:0x%X", fis); |
2325 | ehi->err_mask |= AC_ERR_DEV; | 2208 | ehi->err_mask |= AC_ERR_DEV; |
2326 | ehi->serror |= serror; | 2209 | ehi->serror |= serror; |
2327 | ehi->action |= ATA_EH_SOFTRESET; | 2210 | ehi->action |= ATA_EH_RESET; |
2328 | ata_port_freeze(ap); | 2211 | ata_port_freeze(ap); |
2329 | return; | 2212 | return; |
2330 | } | 2213 | } |
@@ -2356,13 +2239,13 @@ static void nv_swncq_host_interrupt(struct ata_port *ap, u16 fis) | |||
2356 | if (pp->ncq_flags & (ncq_saw_sdb | ncq_saw_backout)) { | 2239 | if (pp->ncq_flags & (ncq_saw_sdb | ncq_saw_backout)) { |
2357 | ata_ehi_push_desc(ehi, "illegal fis transaction"); | 2240 | ata_ehi_push_desc(ehi, "illegal fis transaction"); |
2358 | ehi->err_mask |= AC_ERR_HSM; | 2241 | ehi->err_mask |= AC_ERR_HSM; |
2359 | ehi->action |= ATA_EH_HARDRESET; | 2242 | ehi->action |= ATA_EH_RESET; |
2360 | goto irq_error; | 2243 | goto irq_error; |
2361 | } | 2244 | } |
2362 | 2245 | ||
2363 | if (!(fis & NV_SWNCQ_IRQ_DMASETUP) && | 2246 | if (!(fis & NV_SWNCQ_IRQ_DMASETUP) && |
2364 | !(pp->ncq_flags & ncq_saw_dmas)) { | 2247 | !(pp->ncq_flags & ncq_saw_dmas)) { |
2365 | ata_stat = ap->ops->check_status(ap); | 2248 | ata_stat = ap->ops->sff_check_status(ap); |
2366 | if (ata_stat & ATA_BUSY) | 2249 | if (ata_stat & ATA_BUSY) |
2367 | goto irq_exit; | 2250 | goto irq_exit; |
2368 | 2251 | ||
@@ -2429,6 +2312,7 @@ static int nv_init_one(struct pci_dev *pdev, const struct pci_device_id *ent) | |||
2429 | { | 2312 | { |
2430 | static int printed_version; | 2313 | static int printed_version; |
2431 | const struct ata_port_info *ppi[] = { NULL, NULL }; | 2314 | const struct ata_port_info *ppi[] = { NULL, NULL }; |
2315 | struct nv_pi_priv *ipriv; | ||
2432 | struct ata_host *host; | 2316 | struct ata_host *host; |
2433 | struct nv_host_priv *hpriv; | 2317 | struct nv_host_priv *hpriv; |
2434 | int rc; | 2318 | int rc; |
@@ -2465,7 +2349,8 @@ static int nv_init_one(struct pci_dev *pdev, const struct pci_device_id *ent) | |||
2465 | } | 2349 | } |
2466 | 2350 | ||
2467 | ppi[0] = &nv_port_info[type]; | 2351 | ppi[0] = &nv_port_info[type]; |
2468 | rc = ata_pci_prepare_sff_host(pdev, ppi, &host); | 2352 | ipriv = ppi[0]->private_data; |
2353 | rc = ata_pci_sff_prepare_host(pdev, ppi, &host); | ||
2469 | if (rc) | 2354 | if (rc) |
2470 | return rc; | 2355 | return rc; |
2471 | 2356 | ||
@@ -2503,8 +2388,8 @@ static int nv_init_one(struct pci_dev *pdev, const struct pci_device_id *ent) | |||
2503 | nv_swncq_host_init(host); | 2388 | nv_swncq_host_init(host); |
2504 | 2389 | ||
2505 | pci_set_master(pdev); | 2390 | pci_set_master(pdev); |
2506 | return ata_host_activate(host, pdev->irq, ppi[0]->irq_handler, | 2391 | return ata_host_activate(host, pdev->irq, ipriv->irq_handler, |
2507 | IRQF_SHARED, ppi[0]->sht); | 2392 | IRQF_SHARED, ipriv->sht); |
2508 | } | 2393 | } |
2509 | 2394 | ||
2510 | #ifdef CONFIG_PM | 2395 | #ifdef CONFIG_PM |
@@ -2600,5 +2485,5 @@ module_exit(nv_exit); | |||
2600 | module_param_named(adma, adma_enabled, bool, 0444); | 2485 | module_param_named(adma, adma_enabled, bool, 0444); |
2601 | MODULE_PARM_DESC(adma, "Enable use of ADMA (Default: true)"); | 2486 | MODULE_PARM_DESC(adma, "Enable use of ADMA (Default: true)"); |
2602 | module_param_named(swncq, swncq_enabled, bool, 0444); | 2487 | module_param_named(swncq, swncq_enabled, bool, 0444); |
2603 | MODULE_PARM_DESC(swncq, "Enable use of SWNCQ (Default: false)"); | 2488 | MODULE_PARM_DESC(swncq, "Enable use of SWNCQ (Default: true)"); |
2604 | 2489 | ||
diff --git a/drivers/ata/sata_promise.c b/drivers/ata/sata_promise.c index 11c1afea2db2..5a10dc5048ad 100644 --- a/drivers/ata/sata_promise.c +++ b/drivers/ata/sata_promise.c | |||
@@ -143,103 +143,57 @@ static void pdc_exec_command_mmio(struct ata_port *ap, const struct ata_taskfile | |||
143 | static int pdc_check_atapi_dma(struct ata_queued_cmd *qc); | 143 | static int pdc_check_atapi_dma(struct ata_queued_cmd *qc); |
144 | static int pdc_old_sata_check_atapi_dma(struct ata_queued_cmd *qc); | 144 | static int pdc_old_sata_check_atapi_dma(struct ata_queued_cmd *qc); |
145 | static void pdc_irq_clear(struct ata_port *ap); | 145 | static void pdc_irq_clear(struct ata_port *ap); |
146 | static unsigned int pdc_qc_issue_prot(struct ata_queued_cmd *qc); | 146 | static unsigned int pdc_qc_issue(struct ata_queued_cmd *qc); |
147 | static void pdc_freeze(struct ata_port *ap); | 147 | static void pdc_freeze(struct ata_port *ap); |
148 | static void pdc_sata_freeze(struct ata_port *ap); | 148 | static void pdc_sata_freeze(struct ata_port *ap); |
149 | static void pdc_thaw(struct ata_port *ap); | 149 | static void pdc_thaw(struct ata_port *ap); |
150 | static void pdc_sata_thaw(struct ata_port *ap); | 150 | static void pdc_sata_thaw(struct ata_port *ap); |
151 | static void pdc_pata_error_handler(struct ata_port *ap); | 151 | static void pdc_error_handler(struct ata_port *ap); |
152 | static void pdc_sata_error_handler(struct ata_port *ap); | ||
153 | static void pdc_post_internal_cmd(struct ata_queued_cmd *qc); | 152 | static void pdc_post_internal_cmd(struct ata_queued_cmd *qc); |
154 | static int pdc_pata_cable_detect(struct ata_port *ap); | 153 | static int pdc_pata_cable_detect(struct ata_port *ap); |
155 | static int pdc_sata_cable_detect(struct ata_port *ap); | 154 | static int pdc_sata_cable_detect(struct ata_port *ap); |
156 | 155 | ||
157 | static struct scsi_host_template pdc_ata_sht = { | 156 | static struct scsi_host_template pdc_ata_sht = { |
158 | .module = THIS_MODULE, | 157 | ATA_BASE_SHT(DRV_NAME), |
159 | .name = DRV_NAME, | ||
160 | .ioctl = ata_scsi_ioctl, | ||
161 | .queuecommand = ata_scsi_queuecmd, | ||
162 | .can_queue = ATA_DEF_QUEUE, | ||
163 | .this_id = ATA_SHT_THIS_ID, | ||
164 | .sg_tablesize = PDC_MAX_PRD, | 158 | .sg_tablesize = PDC_MAX_PRD, |
165 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
166 | .emulated = ATA_SHT_EMULATED, | ||
167 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
168 | .proc_name = DRV_NAME, | ||
169 | .dma_boundary = ATA_DMA_BOUNDARY, | 159 | .dma_boundary = ATA_DMA_BOUNDARY, |
170 | .slave_configure = ata_scsi_slave_config, | ||
171 | .slave_destroy = ata_scsi_slave_destroy, | ||
172 | .bios_param = ata_std_bios_param, | ||
173 | }; | 160 | }; |
174 | 161 | ||
175 | static const struct ata_port_operations pdc_sata_ops = { | 162 | static const struct ata_port_operations pdc_common_ops = { |
176 | .tf_load = pdc_tf_load_mmio, | 163 | .inherits = &ata_sff_port_ops, |
177 | .tf_read = ata_tf_read, | ||
178 | .check_status = ata_check_status, | ||
179 | .exec_command = pdc_exec_command_mmio, | ||
180 | .dev_select = ata_std_dev_select, | ||
181 | .check_atapi_dma = pdc_check_atapi_dma, | ||
182 | 164 | ||
165 | .sff_tf_load = pdc_tf_load_mmio, | ||
166 | .sff_exec_command = pdc_exec_command_mmio, | ||
167 | .check_atapi_dma = pdc_check_atapi_dma, | ||
183 | .qc_prep = pdc_qc_prep, | 168 | .qc_prep = pdc_qc_prep, |
184 | .qc_issue = pdc_qc_issue_prot, | 169 | .qc_issue = pdc_qc_issue, |
185 | .freeze = pdc_sata_freeze, | 170 | .sff_irq_clear = pdc_irq_clear, |
186 | .thaw = pdc_sata_thaw, | ||
187 | .error_handler = pdc_sata_error_handler, | ||
188 | .post_internal_cmd = pdc_post_internal_cmd, | ||
189 | .cable_detect = pdc_sata_cable_detect, | ||
190 | .data_xfer = ata_data_xfer, | ||
191 | .irq_clear = pdc_irq_clear, | ||
192 | .irq_on = ata_irq_on, | ||
193 | 171 | ||
194 | .scr_read = pdc_sata_scr_read, | 172 | .post_internal_cmd = pdc_post_internal_cmd, |
195 | .scr_write = pdc_sata_scr_write, | 173 | .error_handler = pdc_error_handler, |
196 | .port_start = pdc_sata_port_start, | ||
197 | }; | 174 | }; |
198 | 175 | ||
199 | /* First-generation chips need a more restrictive ->check_atapi_dma op */ | 176 | static struct ata_port_operations pdc_sata_ops = { |
200 | static const struct ata_port_operations pdc_old_sata_ops = { | 177 | .inherits = &pdc_common_ops, |
201 | .tf_load = pdc_tf_load_mmio, | 178 | .cable_detect = pdc_sata_cable_detect, |
202 | .tf_read = ata_tf_read, | ||
203 | .check_status = ata_check_status, | ||
204 | .exec_command = pdc_exec_command_mmio, | ||
205 | .dev_select = ata_std_dev_select, | ||
206 | .check_atapi_dma = pdc_old_sata_check_atapi_dma, | ||
207 | |||
208 | .qc_prep = pdc_qc_prep, | ||
209 | .qc_issue = pdc_qc_issue_prot, | ||
210 | .freeze = pdc_sata_freeze, | 179 | .freeze = pdc_sata_freeze, |
211 | .thaw = pdc_sata_thaw, | 180 | .thaw = pdc_sata_thaw, |
212 | .error_handler = pdc_sata_error_handler, | ||
213 | .post_internal_cmd = pdc_post_internal_cmd, | ||
214 | .cable_detect = pdc_sata_cable_detect, | ||
215 | .data_xfer = ata_data_xfer, | ||
216 | .irq_clear = pdc_irq_clear, | ||
217 | .irq_on = ata_irq_on, | ||
218 | |||
219 | .scr_read = pdc_sata_scr_read, | 181 | .scr_read = pdc_sata_scr_read, |
220 | .scr_write = pdc_sata_scr_write, | 182 | .scr_write = pdc_sata_scr_write, |
221 | .port_start = pdc_sata_port_start, | 183 | .port_start = pdc_sata_port_start, |
222 | }; | 184 | }; |
223 | 185 | ||
224 | static const struct ata_port_operations pdc_pata_ops = { | 186 | /* First-generation chips need a more restrictive ->check_atapi_dma op */ |
225 | .tf_load = pdc_tf_load_mmio, | 187 | static struct ata_port_operations pdc_old_sata_ops = { |
226 | .tf_read = ata_tf_read, | 188 | .inherits = &pdc_sata_ops, |
227 | .check_status = ata_check_status, | 189 | .check_atapi_dma = pdc_old_sata_check_atapi_dma, |
228 | .exec_command = pdc_exec_command_mmio, | 190 | }; |
229 | .dev_select = ata_std_dev_select, | ||
230 | .check_atapi_dma = pdc_check_atapi_dma, | ||
231 | 191 | ||
232 | .qc_prep = pdc_qc_prep, | 192 | static struct ata_port_operations pdc_pata_ops = { |
233 | .qc_issue = pdc_qc_issue_prot, | 193 | .inherits = &pdc_common_ops, |
194 | .cable_detect = pdc_pata_cable_detect, | ||
234 | .freeze = pdc_freeze, | 195 | .freeze = pdc_freeze, |
235 | .thaw = pdc_thaw, | 196 | .thaw = pdc_thaw, |
236 | .error_handler = pdc_pata_error_handler, | ||
237 | .post_internal_cmd = pdc_post_internal_cmd, | ||
238 | .cable_detect = pdc_pata_cable_detect, | ||
239 | .data_xfer = ata_data_xfer, | ||
240 | .irq_clear = pdc_irq_clear, | ||
241 | .irq_on = ata_irq_on, | ||
242 | |||
243 | .port_start = pdc_common_port_start, | 197 | .port_start = pdc_common_port_start, |
244 | }; | 198 | }; |
245 | 199 | ||
@@ -451,7 +405,7 @@ static void pdc_atapi_pkt(struct ata_queued_cmd *qc) | |||
451 | u8 *cdb = qc->cdb; | 405 | u8 *cdb = qc->cdb; |
452 | struct pdc_port_priv *pp = ap->private_data; | 406 | struct pdc_port_priv *pp = ap->private_data; |
453 | u8 *buf = pp->pkt; | 407 | u8 *buf = pp->pkt; |
454 | u32 *buf32 = (u32 *) buf; | 408 | __le32 *buf32 = (__le32 *) buf; |
455 | unsigned int dev_sel, feature; | 409 | unsigned int dev_sel, feature; |
456 | 410 | ||
457 | /* set control bits (byte 0), zero delay seq id (byte 3), | 411 | /* set control bits (byte 0), zero delay seq id (byte 3), |
@@ -738,24 +692,12 @@ static void pdc_sata_thaw(struct ata_port *ap) | |||
738 | readl(host_mmio + hotplug_offset); /* flush */ | 692 | readl(host_mmio + hotplug_offset); /* flush */ |
739 | } | 693 | } |
740 | 694 | ||
741 | static void pdc_common_error_handler(struct ata_port *ap, ata_reset_fn_t hardreset) | 695 | static void pdc_error_handler(struct ata_port *ap) |
742 | { | 696 | { |
743 | if (!(ap->pflags & ATA_PFLAG_FROZEN)) | 697 | if (!(ap->pflags & ATA_PFLAG_FROZEN)) |
744 | pdc_reset_port(ap); | 698 | pdc_reset_port(ap); |
745 | 699 | ||
746 | /* perform recovery */ | 700 | ata_std_error_handler(ap); |
747 | ata_do_eh(ap, ata_std_prereset, ata_std_softreset, hardreset, | ||
748 | ata_std_postreset); | ||
749 | } | ||
750 | |||
751 | static void pdc_pata_error_handler(struct ata_port *ap) | ||
752 | { | ||
753 | pdc_common_error_handler(ap, NULL); | ||
754 | } | ||
755 | |||
756 | static void pdc_sata_error_handler(struct ata_port *ap) | ||
757 | { | ||
758 | pdc_common_error_handler(ap, sata_std_hardreset); | ||
759 | } | 701 | } |
760 | 702 | ||
761 | static void pdc_post_internal_cmd(struct ata_queued_cmd *qc) | 703 | static void pdc_post_internal_cmd(struct ata_queued_cmd *qc) |
@@ -952,7 +894,7 @@ static inline void pdc_packet_start(struct ata_queued_cmd *qc) | |||
952 | readl(ap->ioaddr.cmd_addr + PDC_PKT_SUBMIT); /* flush */ | 894 | readl(ap->ioaddr.cmd_addr + PDC_PKT_SUBMIT); /* flush */ |
953 | } | 895 | } |
954 | 896 | ||
955 | static unsigned int pdc_qc_issue_prot(struct ata_queued_cmd *qc) | 897 | static unsigned int pdc_qc_issue(struct ata_queued_cmd *qc) |
956 | { | 898 | { |
957 | switch (qc->tf.protocol) { | 899 | switch (qc->tf.protocol) { |
958 | case ATAPI_PROT_NODATA: | 900 | case ATAPI_PROT_NODATA: |
@@ -972,20 +914,20 @@ static unsigned int pdc_qc_issue_prot(struct ata_queued_cmd *qc) | |||
972 | break; | 914 | break; |
973 | } | 915 | } |
974 | 916 | ||
975 | return ata_qc_issue_prot(qc); | 917 | return ata_sff_qc_issue(qc); |
976 | } | 918 | } |
977 | 919 | ||
978 | static void pdc_tf_load_mmio(struct ata_port *ap, const struct ata_taskfile *tf) | 920 | static void pdc_tf_load_mmio(struct ata_port *ap, const struct ata_taskfile *tf) |
979 | { | 921 | { |
980 | WARN_ON(tf->protocol == ATA_PROT_DMA || tf->protocol == ATAPI_PROT_DMA); | 922 | WARN_ON(tf->protocol == ATA_PROT_DMA || tf->protocol == ATAPI_PROT_DMA); |
981 | ata_tf_load(ap, tf); | 923 | ata_sff_tf_load(ap, tf); |
982 | } | 924 | } |
983 | 925 | ||
984 | static void pdc_exec_command_mmio(struct ata_port *ap, | 926 | static void pdc_exec_command_mmio(struct ata_port *ap, |
985 | const struct ata_taskfile *tf) | 927 | const struct ata_taskfile *tf) |
986 | { | 928 | { |
987 | WARN_ON(tf->protocol == ATA_PROT_DMA || tf->protocol == ATAPI_PROT_DMA); | 929 | WARN_ON(tf->protocol == ATA_PROT_DMA || tf->protocol == ATAPI_PROT_DMA); |
988 | ata_exec_command(ap, tf); | 930 | ata_sff_exec_command(ap, tf); |
989 | } | 931 | } |
990 | 932 | ||
991 | static int pdc_check_atapi_dma(struct ata_queued_cmd *qc) | 933 | static int pdc_check_atapi_dma(struct ata_queued_cmd *qc) |
diff --git a/drivers/ata/sata_qstor.c b/drivers/ata/sata_qstor.c index 91cc12c82040..1600107047cf 100644 --- a/drivers/ata/sata_qstor.c +++ b/drivers/ata/sata_qstor.c | |||
@@ -121,50 +121,38 @@ static unsigned int qs_qc_issue(struct ata_queued_cmd *qc); | |||
121 | static int qs_check_atapi_dma(struct ata_queued_cmd *qc); | 121 | static int qs_check_atapi_dma(struct ata_queued_cmd *qc); |
122 | static void qs_bmdma_stop(struct ata_queued_cmd *qc); | 122 | static void qs_bmdma_stop(struct ata_queued_cmd *qc); |
123 | static u8 qs_bmdma_status(struct ata_port *ap); | 123 | static u8 qs_bmdma_status(struct ata_port *ap); |
124 | static void qs_irq_clear(struct ata_port *ap); | ||
125 | static void qs_freeze(struct ata_port *ap); | 124 | static void qs_freeze(struct ata_port *ap); |
126 | static void qs_thaw(struct ata_port *ap); | 125 | static void qs_thaw(struct ata_port *ap); |
126 | static int qs_prereset(struct ata_link *link, unsigned long deadline); | ||
127 | static void qs_error_handler(struct ata_port *ap); | 127 | static void qs_error_handler(struct ata_port *ap); |
128 | 128 | ||
129 | static struct scsi_host_template qs_ata_sht = { | 129 | static struct scsi_host_template qs_ata_sht = { |
130 | .module = THIS_MODULE, | 130 | ATA_BASE_SHT(DRV_NAME), |
131 | .name = DRV_NAME, | ||
132 | .ioctl = ata_scsi_ioctl, | ||
133 | .queuecommand = ata_scsi_queuecmd, | ||
134 | .can_queue = ATA_DEF_QUEUE, | ||
135 | .this_id = ATA_SHT_THIS_ID, | ||
136 | .sg_tablesize = QS_MAX_PRD, | 131 | .sg_tablesize = QS_MAX_PRD, |
137 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
138 | .emulated = ATA_SHT_EMULATED, | ||
139 | .use_clustering = ENABLE_CLUSTERING, | ||
140 | .proc_name = DRV_NAME, | ||
141 | .dma_boundary = QS_DMA_BOUNDARY, | 132 | .dma_boundary = QS_DMA_BOUNDARY, |
142 | .slave_configure = ata_scsi_slave_config, | ||
143 | .slave_destroy = ata_scsi_slave_destroy, | ||
144 | .bios_param = ata_std_bios_param, | ||
145 | }; | 133 | }; |
146 | 134 | ||
147 | static const struct ata_port_operations qs_ata_ops = { | 135 | static struct ata_port_operations qs_ata_ops = { |
148 | .tf_load = ata_tf_load, | 136 | .inherits = &ata_sff_port_ops, |
149 | .tf_read = ata_tf_read, | 137 | |
150 | .check_status = ata_check_status, | ||
151 | .check_atapi_dma = qs_check_atapi_dma, | 138 | .check_atapi_dma = qs_check_atapi_dma, |
152 | .exec_command = ata_exec_command, | 139 | .bmdma_stop = qs_bmdma_stop, |
153 | .dev_select = ata_std_dev_select, | 140 | .bmdma_status = qs_bmdma_status, |
154 | .qc_prep = qs_qc_prep, | 141 | .qc_prep = qs_qc_prep, |
155 | .qc_issue = qs_qc_issue, | 142 | .qc_issue = qs_qc_issue, |
156 | .data_xfer = ata_data_xfer, | 143 | |
157 | .freeze = qs_freeze, | 144 | .freeze = qs_freeze, |
158 | .thaw = qs_thaw, | 145 | .thaw = qs_thaw, |
146 | .prereset = qs_prereset, | ||
147 | .softreset = ATA_OP_NULL, | ||
159 | .error_handler = qs_error_handler, | 148 | .error_handler = qs_error_handler, |
160 | .irq_clear = qs_irq_clear, | 149 | .post_internal_cmd = ATA_OP_NULL, |
161 | .irq_on = ata_irq_on, | 150 | |
162 | .scr_read = qs_scr_read, | 151 | .scr_read = qs_scr_read, |
163 | .scr_write = qs_scr_write, | 152 | .scr_write = qs_scr_write, |
153 | |||
164 | .port_start = qs_port_start, | 154 | .port_start = qs_port_start, |
165 | .host_stop = qs_host_stop, | 155 | .host_stop = qs_host_stop, |
166 | .bmdma_stop = qs_bmdma_stop, | ||
167 | .bmdma_status = qs_bmdma_status, | ||
168 | }; | 156 | }; |
169 | 157 | ||
170 | static const struct ata_port_info qs_port_info[] = { | 158 | static const struct ata_port_info qs_port_info[] = { |
@@ -211,11 +199,6 @@ static u8 qs_bmdma_status(struct ata_port *ap) | |||
211 | return 0; | 199 | return 0; |
212 | } | 200 | } |
213 | 201 | ||
214 | static void qs_irq_clear(struct ata_port *ap) | ||
215 | { | ||
216 | /* nothing */ | ||
217 | } | ||
218 | |||
219 | static inline void qs_enter_reg_mode(struct ata_port *ap) | 202 | static inline void qs_enter_reg_mode(struct ata_port *ap) |
220 | { | 203 | { |
221 | u8 __iomem *chan = qs_mmio_base(ap->host) + (ap->port_no * 0x4000); | 204 | u8 __iomem *chan = qs_mmio_base(ap->host) + (ap->port_no * 0x4000); |
@@ -256,7 +239,7 @@ static int qs_prereset(struct ata_link *link, unsigned long deadline) | |||
256 | struct ata_port *ap = link->ap; | 239 | struct ata_port *ap = link->ap; |
257 | 240 | ||
258 | qs_reset_channel_logic(ap); | 241 | qs_reset_channel_logic(ap); |
259 | return ata_std_prereset(link, deadline); | 242 | return ata_sff_prereset(link, deadline); |
260 | } | 243 | } |
261 | 244 | ||
262 | static int qs_scr_read(struct ata_port *ap, unsigned int sc_reg, u32 *val) | 245 | static int qs_scr_read(struct ata_port *ap, unsigned int sc_reg, u32 *val) |
@@ -270,8 +253,7 @@ static int qs_scr_read(struct ata_port *ap, unsigned int sc_reg, u32 *val) | |||
270 | static void qs_error_handler(struct ata_port *ap) | 253 | static void qs_error_handler(struct ata_port *ap) |
271 | { | 254 | { |
272 | qs_enter_reg_mode(ap); | 255 | qs_enter_reg_mode(ap); |
273 | ata_do_eh(ap, qs_prereset, NULL, sata_std_hardreset, | 256 | ata_std_error_handler(ap); |
274 | ata_std_postreset); | ||
275 | } | 257 | } |
276 | 258 | ||
277 | static int qs_scr_write(struct ata_port *ap, unsigned int sc_reg, u32 val) | 259 | static int qs_scr_write(struct ata_port *ap, unsigned int sc_reg, u32 val) |
@@ -321,7 +303,7 @@ static void qs_qc_prep(struct ata_queued_cmd *qc) | |||
321 | 303 | ||
322 | qs_enter_reg_mode(qc->ap); | 304 | qs_enter_reg_mode(qc->ap); |
323 | if (qc->tf.protocol != ATA_PROT_DMA) { | 305 | if (qc->tf.protocol != ATA_PROT_DMA) { |
324 | ata_qc_prep(qc); | 306 | ata_sff_qc_prep(qc); |
325 | return; | 307 | return; |
326 | } | 308 | } |
327 | 309 | ||
@@ -380,7 +362,7 @@ static unsigned int qs_qc_issue(struct ata_queued_cmd *qc) | |||
380 | } | 362 | } |
381 | 363 | ||
382 | pp->state = qs_state_mmio; | 364 | pp->state = qs_state_mmio; |
383 | return ata_qc_issue_prot(qc); | 365 | return ata_sff_qc_issue(qc); |
384 | } | 366 | } |
385 | 367 | ||
386 | static void qs_do_or_die(struct ata_queued_cmd *qc, u8 status) | 368 | static void qs_do_or_die(struct ata_queued_cmd *qc, u8 status) |
@@ -469,7 +451,7 @@ static inline unsigned int qs_intr_mmio(struct ata_host *host) | |||
469 | * and pretend we knew it was ours.. (ugh). | 451 | * and pretend we knew it was ours.. (ugh). |
470 | * This does not affect packet mode. | 452 | * This does not affect packet mode. |
471 | */ | 453 | */ |
472 | ata_check_status(ap); | 454 | ata_sff_check_status(ap); |
473 | handled = 1; | 455 | handled = 1; |
474 | continue; | 456 | continue; |
475 | } | 457 | } |
@@ -477,7 +459,7 @@ static inline unsigned int qs_intr_mmio(struct ata_host *host) | |||
477 | if (!pp || pp->state != qs_state_mmio) | 459 | if (!pp || pp->state != qs_state_mmio) |
478 | continue; | 460 | continue; |
479 | if (!(qc->tf.flags & ATA_TFLAG_POLLING)) | 461 | if (!(qc->tf.flags & ATA_TFLAG_POLLING)) |
480 | handled |= ata_host_intr(ap, qc); | 462 | handled |= ata_sff_host_intr(ap, qc); |
481 | } | 463 | } |
482 | } | 464 | } |
483 | return handled; | 465 | return handled; |
diff --git a/drivers/ata/sata_sil.c b/drivers/ata/sata_sil.c index 0b8191b52f97..88bf4212590f 100644 --- a/drivers/ata/sata_sil.c +++ b/drivers/ata/sata_sil.c | |||
@@ -60,7 +60,6 @@ enum { | |||
60 | 60 | ||
61 | SIL_DFL_PORT_FLAGS = ATA_FLAG_SATA | ATA_FLAG_NO_LEGACY | | 61 | SIL_DFL_PORT_FLAGS = ATA_FLAG_SATA | ATA_FLAG_NO_LEGACY | |
62 | ATA_FLAG_MMIO, | 62 | ATA_FLAG_MMIO, |
63 | SIL_DFL_LINK_FLAGS = ATA_LFLAG_HRST_TO_RESUME, | ||
64 | 63 | ||
65 | /* | 64 | /* |
66 | * Controller IDs | 65 | * Controller IDs |
@@ -168,54 +167,23 @@ static struct pci_driver sil_pci_driver = { | |||
168 | }; | 167 | }; |
169 | 168 | ||
170 | static struct scsi_host_template sil_sht = { | 169 | static struct scsi_host_template sil_sht = { |
171 | .module = THIS_MODULE, | 170 | ATA_BMDMA_SHT(DRV_NAME), |
172 | .name = DRV_NAME, | ||
173 | .ioctl = ata_scsi_ioctl, | ||
174 | .queuecommand = ata_scsi_queuecmd, | ||
175 | .can_queue = ATA_DEF_QUEUE, | ||
176 | .this_id = ATA_SHT_THIS_ID, | ||
177 | .sg_tablesize = LIBATA_MAX_PRD, | ||
178 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
179 | .emulated = ATA_SHT_EMULATED, | ||
180 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
181 | .proc_name = DRV_NAME, | ||
182 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
183 | .slave_configure = ata_scsi_slave_config, | ||
184 | .slave_destroy = ata_scsi_slave_destroy, | ||
185 | .bios_param = ata_std_bios_param, | ||
186 | }; | 171 | }; |
187 | 172 | ||
188 | static const struct ata_port_operations sil_ops = { | 173 | static struct ata_port_operations sil_ops = { |
174 | .inherits = &ata_bmdma_port_ops, | ||
189 | .dev_config = sil_dev_config, | 175 | .dev_config = sil_dev_config, |
190 | .tf_load = ata_tf_load, | ||
191 | .tf_read = ata_tf_read, | ||
192 | .check_status = ata_check_status, | ||
193 | .exec_command = ata_exec_command, | ||
194 | .dev_select = ata_std_dev_select, | ||
195 | .set_mode = sil_set_mode, | 176 | .set_mode = sil_set_mode, |
196 | .bmdma_setup = ata_bmdma_setup, | ||
197 | .bmdma_start = ata_bmdma_start, | ||
198 | .bmdma_stop = ata_bmdma_stop, | ||
199 | .bmdma_status = ata_bmdma_status, | ||
200 | .qc_prep = ata_qc_prep, | ||
201 | .qc_issue = ata_qc_issue_prot, | ||
202 | .data_xfer = ata_data_xfer, | ||
203 | .freeze = sil_freeze, | 177 | .freeze = sil_freeze, |
204 | .thaw = sil_thaw, | 178 | .thaw = sil_thaw, |
205 | .error_handler = ata_bmdma_error_handler, | ||
206 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
207 | .irq_clear = ata_bmdma_irq_clear, | ||
208 | .irq_on = ata_irq_on, | ||
209 | .scr_read = sil_scr_read, | 179 | .scr_read = sil_scr_read, |
210 | .scr_write = sil_scr_write, | 180 | .scr_write = sil_scr_write, |
211 | .port_start = ata_port_start, | ||
212 | }; | 181 | }; |
213 | 182 | ||
214 | static const struct ata_port_info sil_port_info[] = { | 183 | static const struct ata_port_info sil_port_info[] = { |
215 | /* sil_3112 */ | 184 | /* sil_3112 */ |
216 | { | 185 | { |
217 | .flags = SIL_DFL_PORT_FLAGS | SIL_FLAG_MOD15WRITE, | 186 | .flags = SIL_DFL_PORT_FLAGS | SIL_FLAG_MOD15WRITE, |
218 | .link_flags = SIL_DFL_LINK_FLAGS, | ||
219 | .pio_mask = 0x1f, /* pio0-4 */ | 187 | .pio_mask = 0x1f, /* pio0-4 */ |
220 | .mwdma_mask = 0x07, /* mwdma0-2 */ | 188 | .mwdma_mask = 0x07, /* mwdma0-2 */ |
221 | .udma_mask = ATA_UDMA5, | 189 | .udma_mask = ATA_UDMA5, |
@@ -225,7 +193,6 @@ static const struct ata_port_info sil_port_info[] = { | |||
225 | { | 193 | { |
226 | .flags = SIL_DFL_PORT_FLAGS | SIL_FLAG_MOD15WRITE | | 194 | .flags = SIL_DFL_PORT_FLAGS | SIL_FLAG_MOD15WRITE | |
227 | SIL_FLAG_NO_SATA_IRQ, | 195 | SIL_FLAG_NO_SATA_IRQ, |
228 | .link_flags = SIL_DFL_LINK_FLAGS, | ||
229 | .pio_mask = 0x1f, /* pio0-4 */ | 196 | .pio_mask = 0x1f, /* pio0-4 */ |
230 | .mwdma_mask = 0x07, /* mwdma0-2 */ | 197 | .mwdma_mask = 0x07, /* mwdma0-2 */ |
231 | .udma_mask = ATA_UDMA5, | 198 | .udma_mask = ATA_UDMA5, |
@@ -234,7 +201,6 @@ static const struct ata_port_info sil_port_info[] = { | |||
234 | /* sil_3512 */ | 201 | /* sil_3512 */ |
235 | { | 202 | { |
236 | .flags = SIL_DFL_PORT_FLAGS | SIL_FLAG_RERR_ON_DMA_ACT, | 203 | .flags = SIL_DFL_PORT_FLAGS | SIL_FLAG_RERR_ON_DMA_ACT, |
237 | .link_flags = SIL_DFL_LINK_FLAGS, | ||
238 | .pio_mask = 0x1f, /* pio0-4 */ | 204 | .pio_mask = 0x1f, /* pio0-4 */ |
239 | .mwdma_mask = 0x07, /* mwdma0-2 */ | 205 | .mwdma_mask = 0x07, /* mwdma0-2 */ |
240 | .udma_mask = ATA_UDMA5, | 206 | .udma_mask = ATA_UDMA5, |
@@ -243,7 +209,6 @@ static const struct ata_port_info sil_port_info[] = { | |||
243 | /* sil_3114 */ | 209 | /* sil_3114 */ |
244 | { | 210 | { |
245 | .flags = SIL_DFL_PORT_FLAGS | SIL_FLAG_RERR_ON_DMA_ACT, | 211 | .flags = SIL_DFL_PORT_FLAGS | SIL_FLAG_RERR_ON_DMA_ACT, |
246 | .link_flags = SIL_DFL_LINK_FLAGS, | ||
247 | .pio_mask = 0x1f, /* pio0-4 */ | 212 | .pio_mask = 0x1f, /* pio0-4 */ |
248 | .mwdma_mask = 0x07, /* mwdma0-2 */ | 213 | .mwdma_mask = 0x07, /* mwdma0-2 */ |
249 | .udma_mask = ATA_UDMA5, | 214 | .udma_mask = ATA_UDMA5, |
@@ -404,7 +369,7 @@ static void sil_host_intr(struct ata_port *ap, u32 bmdma2) | |||
404 | 369 | ||
405 | if (unlikely(!qc || (qc->tf.flags & ATA_TFLAG_POLLING))) { | 370 | if (unlikely(!qc || (qc->tf.flags & ATA_TFLAG_POLLING))) { |
406 | /* this sometimes happens, just clear IRQ */ | 371 | /* this sometimes happens, just clear IRQ */ |
407 | ata_chk_status(ap); | 372 | ap->ops->sff_check_status(ap); |
408 | return; | 373 | return; |
409 | } | 374 | } |
410 | 375 | ||
@@ -440,15 +405,15 @@ static void sil_host_intr(struct ata_port *ap, u32 bmdma2) | |||
440 | } | 405 | } |
441 | 406 | ||
442 | /* check main status, clearing INTRQ */ | 407 | /* check main status, clearing INTRQ */ |
443 | status = ata_chk_status(ap); | 408 | status = ap->ops->sff_check_status(ap); |
444 | if (unlikely(status & ATA_BUSY)) | 409 | if (unlikely(status & ATA_BUSY)) |
445 | goto err_hsm; | 410 | goto err_hsm; |
446 | 411 | ||
447 | /* ack bmdma irq events */ | 412 | /* ack bmdma irq events */ |
448 | ata_bmdma_irq_clear(ap); | 413 | ata_sff_irq_clear(ap); |
449 | 414 | ||
450 | /* kick HSM in the ass */ | 415 | /* kick HSM in the ass */ |
451 | ata_hsm_move(ap, qc, status, 0); | 416 | ata_sff_hsm_move(ap, qc, status, 0); |
452 | 417 | ||
453 | if (unlikely(qc->err_mask) && ata_is_dma(qc->tf.protocol)) | 418 | if (unlikely(qc->err_mask) && ata_is_dma(qc->tf.protocol)) |
454 | ata_ehi_push_desc(ehi, "BMDMA2 stat 0x%x", bmdma2); | 419 | ata_ehi_push_desc(ehi, "BMDMA2 stat 0x%x", bmdma2); |
@@ -515,8 +480,8 @@ static void sil_thaw(struct ata_port *ap) | |||
515 | u32 tmp; | 480 | u32 tmp; |
516 | 481 | ||
517 | /* clear IRQ */ | 482 | /* clear IRQ */ |
518 | ata_chk_status(ap); | 483 | ap->ops->sff_check_status(ap); |
519 | ata_bmdma_irq_clear(ap); | 484 | ata_sff_irq_clear(ap); |
520 | 485 | ||
521 | /* turn on SATA IRQ if supported */ | 486 | /* turn on SATA IRQ if supported */ |
522 | if (!(ap->flags & SIL_FLAG_NO_SATA_IRQ)) | 487 | if (!(ap->flags & SIL_FLAG_NO_SATA_IRQ)) |
@@ -690,7 +655,7 @@ static int sil_init_one(struct pci_dev *pdev, const struct pci_device_id *ent) | |||
690 | ioaddr->ctl_addr = mmio_base + sil_port[i].ctl; | 655 | ioaddr->ctl_addr = mmio_base + sil_port[i].ctl; |
691 | ioaddr->bmdma_addr = mmio_base + sil_port[i].bmdma; | 656 | ioaddr->bmdma_addr = mmio_base + sil_port[i].bmdma; |
692 | ioaddr->scr_addr = mmio_base + sil_port[i].scr; | 657 | ioaddr->scr_addr = mmio_base + sil_port[i].scr; |
693 | ata_std_ports(ioaddr); | 658 | ata_sff_std_ports(ioaddr); |
694 | 659 | ||
695 | ata_port_pbar_desc(ap, SIL_MMIO_BAR, -1, "mmio"); | 660 | ata_port_pbar_desc(ap, SIL_MMIO_BAR, -1, "mmio"); |
696 | ata_port_pbar_desc(ap, SIL_MMIO_BAR, sil_port[i].tf, "tf"); | 661 | ata_port_pbar_desc(ap, SIL_MMIO_BAR, sil_port[i].tf, "tf"); |
diff --git a/drivers/ata/sata_sil24.c b/drivers/ata/sata_sil24.c index df7988df7908..27a110110077 100644 --- a/drivers/ata/sata_sil24.c +++ b/drivers/ata/sata_sil24.c | |||
@@ -254,7 +254,6 @@ enum { | |||
254 | ATA_FLAG_MMIO | ATA_FLAG_PIO_DMA | | 254 | ATA_FLAG_MMIO | ATA_FLAG_PIO_DMA | |
255 | ATA_FLAG_NCQ | ATA_FLAG_ACPI_SATA | | 255 | ATA_FLAG_NCQ | ATA_FLAG_ACPI_SATA | |
256 | ATA_FLAG_AN | ATA_FLAG_PMP, | 256 | ATA_FLAG_AN | ATA_FLAG_PMP, |
257 | SIL24_COMMON_LFLAGS = ATA_LFLAG_SKIP_D2H_BSY, | ||
258 | SIL24_FLAG_PCIX_IRQ_WOC = (1 << 24), /* IRQ loss errata on PCI-X */ | 257 | SIL24_FLAG_PCIX_IRQ_WOC = (1 << 24), /* IRQ loss errata on PCI-X */ |
259 | 258 | ||
260 | IRQ_STAT_4PORTS = 0xf, | 259 | IRQ_STAT_4PORTS = 0xf, |
@@ -286,45 +285,45 @@ static struct sil24_cerr_info { | |||
286 | "device error via D2H FIS" }, | 285 | "device error via D2H FIS" }, |
287 | [PORT_CERR_SDB] = { AC_ERR_DEV, 0, | 286 | [PORT_CERR_SDB] = { AC_ERR_DEV, 0, |
288 | "device error via SDB FIS" }, | 287 | "device error via SDB FIS" }, |
289 | [PORT_CERR_DATA] = { AC_ERR_ATA_BUS, ATA_EH_SOFTRESET, | 288 | [PORT_CERR_DATA] = { AC_ERR_ATA_BUS, ATA_EH_RESET, |
290 | "error in data FIS" }, | 289 | "error in data FIS" }, |
291 | [PORT_CERR_SEND] = { AC_ERR_ATA_BUS, ATA_EH_SOFTRESET, | 290 | [PORT_CERR_SEND] = { AC_ERR_ATA_BUS, ATA_EH_RESET, |
292 | "failed to transmit command FIS" }, | 291 | "failed to transmit command FIS" }, |
293 | [PORT_CERR_INCONSISTENT] = { AC_ERR_HSM, ATA_EH_SOFTRESET, | 292 | [PORT_CERR_INCONSISTENT] = { AC_ERR_HSM, ATA_EH_RESET, |
294 | "protocol mismatch" }, | 293 | "protocol mismatch" }, |
295 | [PORT_CERR_DIRECTION] = { AC_ERR_HSM, ATA_EH_SOFTRESET, | 294 | [PORT_CERR_DIRECTION] = { AC_ERR_HSM, ATA_EH_RESET, |
296 | "data directon mismatch" }, | 295 | "data directon mismatch" }, |
297 | [PORT_CERR_UNDERRUN] = { AC_ERR_HSM, ATA_EH_SOFTRESET, | 296 | [PORT_CERR_UNDERRUN] = { AC_ERR_HSM, ATA_EH_RESET, |
298 | "ran out of SGEs while writing" }, | 297 | "ran out of SGEs while writing" }, |
299 | [PORT_CERR_OVERRUN] = { AC_ERR_HSM, ATA_EH_SOFTRESET, | 298 | [PORT_CERR_OVERRUN] = { AC_ERR_HSM, ATA_EH_RESET, |
300 | "ran out of SGEs while reading" }, | 299 | "ran out of SGEs while reading" }, |
301 | [PORT_CERR_PKT_PROT] = { AC_ERR_HSM, ATA_EH_SOFTRESET, | 300 | [PORT_CERR_PKT_PROT] = { AC_ERR_HSM, ATA_EH_RESET, |
302 | "invalid data directon for ATAPI CDB" }, | 301 | "invalid data directon for ATAPI CDB" }, |
303 | [PORT_CERR_SGT_BOUNDARY] = { AC_ERR_SYSTEM, ATA_EH_SOFTRESET, | 302 | [PORT_CERR_SGT_BOUNDARY] = { AC_ERR_SYSTEM, ATA_EH_RESET, |
304 | "SGT not on qword boundary" }, | 303 | "SGT not on qword boundary" }, |
305 | [PORT_CERR_SGT_TGTABRT] = { AC_ERR_HOST_BUS, ATA_EH_SOFTRESET, | 304 | [PORT_CERR_SGT_TGTABRT] = { AC_ERR_HOST_BUS, ATA_EH_RESET, |
306 | "PCI target abort while fetching SGT" }, | 305 | "PCI target abort while fetching SGT" }, |
307 | [PORT_CERR_SGT_MSTABRT] = { AC_ERR_HOST_BUS, ATA_EH_SOFTRESET, | 306 | [PORT_CERR_SGT_MSTABRT] = { AC_ERR_HOST_BUS, ATA_EH_RESET, |
308 | "PCI master abort while fetching SGT" }, | 307 | "PCI master abort while fetching SGT" }, |
309 | [PORT_CERR_SGT_PCIPERR] = { AC_ERR_HOST_BUS, ATA_EH_SOFTRESET, | 308 | [PORT_CERR_SGT_PCIPERR] = { AC_ERR_HOST_BUS, ATA_EH_RESET, |
310 | "PCI parity error while fetching SGT" }, | 309 | "PCI parity error while fetching SGT" }, |
311 | [PORT_CERR_CMD_BOUNDARY] = { AC_ERR_SYSTEM, ATA_EH_SOFTRESET, | 310 | [PORT_CERR_CMD_BOUNDARY] = { AC_ERR_SYSTEM, ATA_EH_RESET, |
312 | "PRB not on qword boundary" }, | 311 | "PRB not on qword boundary" }, |
313 | [PORT_CERR_CMD_TGTABRT] = { AC_ERR_HOST_BUS, ATA_EH_SOFTRESET, | 312 | [PORT_CERR_CMD_TGTABRT] = { AC_ERR_HOST_BUS, ATA_EH_RESET, |
314 | "PCI target abort while fetching PRB" }, | 313 | "PCI target abort while fetching PRB" }, |
315 | [PORT_CERR_CMD_MSTABRT] = { AC_ERR_HOST_BUS, ATA_EH_SOFTRESET, | 314 | [PORT_CERR_CMD_MSTABRT] = { AC_ERR_HOST_BUS, ATA_EH_RESET, |
316 | "PCI master abort while fetching PRB" }, | 315 | "PCI master abort while fetching PRB" }, |
317 | [PORT_CERR_CMD_PCIPERR] = { AC_ERR_HOST_BUS, ATA_EH_SOFTRESET, | 316 | [PORT_CERR_CMD_PCIPERR] = { AC_ERR_HOST_BUS, ATA_EH_RESET, |
318 | "PCI parity error while fetching PRB" }, | 317 | "PCI parity error while fetching PRB" }, |
319 | [PORT_CERR_XFR_UNDEF] = { AC_ERR_HOST_BUS, ATA_EH_SOFTRESET, | 318 | [PORT_CERR_XFR_UNDEF] = { AC_ERR_HOST_BUS, ATA_EH_RESET, |
320 | "undefined error while transferring data" }, | 319 | "undefined error while transferring data" }, |
321 | [PORT_CERR_XFR_TGTABRT] = { AC_ERR_HOST_BUS, ATA_EH_SOFTRESET, | 320 | [PORT_CERR_XFR_TGTABRT] = { AC_ERR_HOST_BUS, ATA_EH_RESET, |
322 | "PCI target abort while transferring data" }, | 321 | "PCI target abort while transferring data" }, |
323 | [PORT_CERR_XFR_MSTABRT] = { AC_ERR_HOST_BUS, ATA_EH_SOFTRESET, | 322 | [PORT_CERR_XFR_MSTABRT] = { AC_ERR_HOST_BUS, ATA_EH_RESET, |
324 | "PCI master abort while transferring data" }, | 323 | "PCI master abort while transferring data" }, |
325 | [PORT_CERR_XFR_PCIPERR] = { AC_ERR_HOST_BUS, ATA_EH_SOFTRESET, | 324 | [PORT_CERR_XFR_PCIPERR] = { AC_ERR_HOST_BUS, ATA_EH_RESET, |
326 | "PCI parity error while transferring data" }, | 325 | "PCI parity error while transferring data" }, |
327 | [PORT_CERR_SENDSERVICE] = { AC_ERR_HSM, ATA_EH_SOFTRESET, | 326 | [PORT_CERR_SENDSERVICE] = { AC_ERR_HSM, ATA_EH_RESET, |
328 | "FIS received while sending service FIS" }, | 327 | "FIS received while sending service FIS" }, |
329 | }; | 328 | }; |
330 | 329 | ||
@@ -337,23 +336,26 @@ static struct sil24_cerr_info { | |||
337 | struct sil24_port_priv { | 336 | struct sil24_port_priv { |
338 | union sil24_cmd_block *cmd_block; /* 32 cmd blocks */ | 337 | union sil24_cmd_block *cmd_block; /* 32 cmd blocks */ |
339 | dma_addr_t cmd_block_dma; /* DMA base addr for them */ | 338 | dma_addr_t cmd_block_dma; /* DMA base addr for them */ |
340 | struct ata_taskfile tf; /* Cached taskfile registers */ | ||
341 | int do_port_rst; | 339 | int do_port_rst; |
342 | }; | 340 | }; |
343 | 341 | ||
344 | static void sil24_dev_config(struct ata_device *dev); | 342 | static void sil24_dev_config(struct ata_device *dev); |
345 | static u8 sil24_check_status(struct ata_port *ap); | ||
346 | static int sil24_scr_read(struct ata_port *ap, unsigned sc_reg, u32 *val); | 343 | static int sil24_scr_read(struct ata_port *ap, unsigned sc_reg, u32 *val); |
347 | static int sil24_scr_write(struct ata_port *ap, unsigned sc_reg, u32 val); | 344 | static int sil24_scr_write(struct ata_port *ap, unsigned sc_reg, u32 val); |
348 | static void sil24_tf_read(struct ata_port *ap, struct ata_taskfile *tf); | ||
349 | static int sil24_qc_defer(struct ata_queued_cmd *qc); | 345 | static int sil24_qc_defer(struct ata_queued_cmd *qc); |
350 | static void sil24_qc_prep(struct ata_queued_cmd *qc); | 346 | static void sil24_qc_prep(struct ata_queued_cmd *qc); |
351 | static unsigned int sil24_qc_issue(struct ata_queued_cmd *qc); | 347 | static unsigned int sil24_qc_issue(struct ata_queued_cmd *qc); |
352 | static void sil24_irq_clear(struct ata_port *ap); | 348 | static bool sil24_qc_fill_rtf(struct ata_queued_cmd *qc); |
353 | static void sil24_pmp_attach(struct ata_port *ap); | 349 | static void sil24_pmp_attach(struct ata_port *ap); |
354 | static void sil24_pmp_detach(struct ata_port *ap); | 350 | static void sil24_pmp_detach(struct ata_port *ap); |
355 | static void sil24_freeze(struct ata_port *ap); | 351 | static void sil24_freeze(struct ata_port *ap); |
356 | static void sil24_thaw(struct ata_port *ap); | 352 | static void sil24_thaw(struct ata_port *ap); |
353 | static int sil24_softreset(struct ata_link *link, unsigned int *class, | ||
354 | unsigned long deadline); | ||
355 | static int sil24_hardreset(struct ata_link *link, unsigned int *class, | ||
356 | unsigned long deadline); | ||
357 | static int sil24_pmp_hardreset(struct ata_link *link, unsigned int *class, | ||
358 | unsigned long deadline); | ||
357 | static void sil24_error_handler(struct ata_port *ap); | 359 | static void sil24_error_handler(struct ata_port *ap); |
358 | static void sil24_post_internal_cmd(struct ata_queued_cmd *qc); | 360 | static void sil24_post_internal_cmd(struct ata_queued_cmd *qc); |
359 | static int sil24_port_start(struct ata_port *ap); | 361 | static int sil24_port_start(struct ata_port *ap); |
@@ -386,52 +388,36 @@ static struct pci_driver sil24_pci_driver = { | |||
386 | }; | 388 | }; |
387 | 389 | ||
388 | static struct scsi_host_template sil24_sht = { | 390 | static struct scsi_host_template sil24_sht = { |
389 | .module = THIS_MODULE, | 391 | ATA_NCQ_SHT(DRV_NAME), |
390 | .name = DRV_NAME, | ||
391 | .ioctl = ata_scsi_ioctl, | ||
392 | .queuecommand = ata_scsi_queuecmd, | ||
393 | .change_queue_depth = ata_scsi_change_queue_depth, | ||
394 | .can_queue = SIL24_MAX_CMDS, | 392 | .can_queue = SIL24_MAX_CMDS, |
395 | .this_id = ATA_SHT_THIS_ID, | ||
396 | .sg_tablesize = SIL24_MAX_SGE, | 393 | .sg_tablesize = SIL24_MAX_SGE, |
397 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
398 | .emulated = ATA_SHT_EMULATED, | ||
399 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
400 | .proc_name = DRV_NAME, | ||
401 | .dma_boundary = ATA_DMA_BOUNDARY, | 394 | .dma_boundary = ATA_DMA_BOUNDARY, |
402 | .slave_configure = ata_scsi_slave_config, | ||
403 | .slave_destroy = ata_scsi_slave_destroy, | ||
404 | .bios_param = ata_std_bios_param, | ||
405 | }; | 395 | }; |
406 | 396 | ||
407 | static const struct ata_port_operations sil24_ops = { | 397 | static struct ata_port_operations sil24_ops = { |
408 | .dev_config = sil24_dev_config, | 398 | .inherits = &sata_pmp_port_ops, |
409 | |||
410 | .check_status = sil24_check_status, | ||
411 | .check_altstatus = sil24_check_status, | ||
412 | .dev_select = ata_noop_dev_select, | ||
413 | |||
414 | .tf_read = sil24_tf_read, | ||
415 | 399 | ||
416 | .qc_defer = sil24_qc_defer, | 400 | .qc_defer = sil24_qc_defer, |
417 | .qc_prep = sil24_qc_prep, | 401 | .qc_prep = sil24_qc_prep, |
418 | .qc_issue = sil24_qc_issue, | 402 | .qc_issue = sil24_qc_issue, |
403 | .qc_fill_rtf = sil24_qc_fill_rtf, | ||
419 | 404 | ||
420 | .irq_clear = sil24_irq_clear, | 405 | .freeze = sil24_freeze, |
406 | .thaw = sil24_thaw, | ||
407 | .softreset = sil24_softreset, | ||
408 | .hardreset = sil24_hardreset, | ||
409 | .pmp_softreset = sil24_softreset, | ||
410 | .pmp_hardreset = sil24_pmp_hardreset, | ||
411 | .error_handler = sil24_error_handler, | ||
412 | .post_internal_cmd = sil24_post_internal_cmd, | ||
413 | .dev_config = sil24_dev_config, | ||
421 | 414 | ||
422 | .scr_read = sil24_scr_read, | 415 | .scr_read = sil24_scr_read, |
423 | .scr_write = sil24_scr_write, | 416 | .scr_write = sil24_scr_write, |
424 | |||
425 | .pmp_attach = sil24_pmp_attach, | 417 | .pmp_attach = sil24_pmp_attach, |
426 | .pmp_detach = sil24_pmp_detach, | 418 | .pmp_detach = sil24_pmp_detach, |
427 | 419 | ||
428 | .freeze = sil24_freeze, | ||
429 | .thaw = sil24_thaw, | ||
430 | .error_handler = sil24_error_handler, | ||
431 | .post_internal_cmd = sil24_post_internal_cmd, | ||
432 | |||
433 | .port_start = sil24_port_start, | 420 | .port_start = sil24_port_start, |
434 | |||
435 | #ifdef CONFIG_PM | 421 | #ifdef CONFIG_PM |
436 | .port_resume = sil24_port_resume, | 422 | .port_resume = sil24_port_resume, |
437 | #endif | 423 | #endif |
@@ -449,7 +435,6 @@ static const struct ata_port_info sil24_port_info[] = { | |||
449 | { | 435 | { |
450 | .flags = SIL24_COMMON_FLAGS | SIL24_NPORTS2FLAG(4) | | 436 | .flags = SIL24_COMMON_FLAGS | SIL24_NPORTS2FLAG(4) | |
451 | SIL24_FLAG_PCIX_IRQ_WOC, | 437 | SIL24_FLAG_PCIX_IRQ_WOC, |
452 | .link_flags = SIL24_COMMON_LFLAGS, | ||
453 | .pio_mask = 0x1f, /* pio0-4 */ | 438 | .pio_mask = 0x1f, /* pio0-4 */ |
454 | .mwdma_mask = 0x07, /* mwdma0-2 */ | 439 | .mwdma_mask = 0x07, /* mwdma0-2 */ |
455 | .udma_mask = ATA_UDMA5, /* udma0-5 */ | 440 | .udma_mask = ATA_UDMA5, /* udma0-5 */ |
@@ -458,7 +443,6 @@ static const struct ata_port_info sil24_port_info[] = { | |||
458 | /* sil_3132 */ | 443 | /* sil_3132 */ |
459 | { | 444 | { |
460 | .flags = SIL24_COMMON_FLAGS | SIL24_NPORTS2FLAG(2), | 445 | .flags = SIL24_COMMON_FLAGS | SIL24_NPORTS2FLAG(2), |
461 | .link_flags = SIL24_COMMON_LFLAGS, | ||
462 | .pio_mask = 0x1f, /* pio0-4 */ | 446 | .pio_mask = 0x1f, /* pio0-4 */ |
463 | .mwdma_mask = 0x07, /* mwdma0-2 */ | 447 | .mwdma_mask = 0x07, /* mwdma0-2 */ |
464 | .udma_mask = ATA_UDMA5, /* udma0-5 */ | 448 | .udma_mask = ATA_UDMA5, /* udma0-5 */ |
@@ -467,7 +451,6 @@ static const struct ata_port_info sil24_port_info[] = { | |||
467 | /* sil_3131/sil_3531 */ | 451 | /* sil_3131/sil_3531 */ |
468 | { | 452 | { |
469 | .flags = SIL24_COMMON_FLAGS | SIL24_NPORTS2FLAG(1), | 453 | .flags = SIL24_COMMON_FLAGS | SIL24_NPORTS2FLAG(1), |
470 | .link_flags = SIL24_COMMON_LFLAGS, | ||
471 | .pio_mask = 0x1f, /* pio0-4 */ | 454 | .pio_mask = 0x1f, /* pio0-4 */ |
472 | .mwdma_mask = 0x07, /* mwdma0-2 */ | 455 | .mwdma_mask = 0x07, /* mwdma0-2 */ |
473 | .udma_mask = ATA_UDMA5, /* udma0-5 */ | 456 | .udma_mask = ATA_UDMA5, /* udma0-5 */ |
@@ -482,9 +465,19 @@ static int sil24_tag(int tag) | |||
482 | return tag; | 465 | return tag; |
483 | } | 466 | } |
484 | 467 | ||
468 | static unsigned long sil24_port_offset(struct ata_port *ap) | ||
469 | { | ||
470 | return ap->port_no * PORT_REGS_SIZE; | ||
471 | } | ||
472 | |||
473 | static void __iomem *sil24_port_base(struct ata_port *ap) | ||
474 | { | ||
475 | return ap->host->iomap[SIL24_PORT_BAR] + sil24_port_offset(ap); | ||
476 | } | ||
477 | |||
485 | static void sil24_dev_config(struct ata_device *dev) | 478 | static void sil24_dev_config(struct ata_device *dev) |
486 | { | 479 | { |
487 | void __iomem *port = dev->link->ap->ioaddr.cmd_addr; | 480 | void __iomem *port = sil24_port_base(dev->link->ap); |
488 | 481 | ||
489 | if (dev->cdb_len == 16) | 482 | if (dev->cdb_len == 16) |
490 | writel(PORT_CS_CDB16, port + PORT_CTRL_STAT); | 483 | writel(PORT_CS_CDB16, port + PORT_CTRL_STAT); |
@@ -494,7 +487,7 @@ static void sil24_dev_config(struct ata_device *dev) | |||
494 | 487 | ||
495 | static void sil24_read_tf(struct ata_port *ap, int tag, struct ata_taskfile *tf) | 488 | static void sil24_read_tf(struct ata_port *ap, int tag, struct ata_taskfile *tf) |
496 | { | 489 | { |
497 | void __iomem *port = ap->ioaddr.cmd_addr; | 490 | void __iomem *port = sil24_port_base(ap); |
498 | struct sil24_prb __iomem *prb; | 491 | struct sil24_prb __iomem *prb; |
499 | u8 fis[6 * 4]; | 492 | u8 fis[6 * 4]; |
500 | 493 | ||
@@ -503,12 +496,6 @@ static void sil24_read_tf(struct ata_port *ap, int tag, struct ata_taskfile *tf) | |||
503 | ata_tf_from_fis(fis, tf); | 496 | ata_tf_from_fis(fis, tf); |
504 | } | 497 | } |
505 | 498 | ||
506 | static u8 sil24_check_status(struct ata_port *ap) | ||
507 | { | ||
508 | struct sil24_port_priv *pp = ap->private_data; | ||
509 | return pp->tf.command; | ||
510 | } | ||
511 | |||
512 | static int sil24_scr_map[] = { | 499 | static int sil24_scr_map[] = { |
513 | [SCR_CONTROL] = 0, | 500 | [SCR_CONTROL] = 0, |
514 | [SCR_STATUS] = 1, | 501 | [SCR_STATUS] = 1, |
@@ -518,7 +505,7 @@ static int sil24_scr_map[] = { | |||
518 | 505 | ||
519 | static int sil24_scr_read(struct ata_port *ap, unsigned sc_reg, u32 *val) | 506 | static int sil24_scr_read(struct ata_port *ap, unsigned sc_reg, u32 *val) |
520 | { | 507 | { |
521 | void __iomem *scr_addr = ap->ioaddr.scr_addr; | 508 | void __iomem *scr_addr = sil24_port_base(ap) + PORT_SCONTROL; |
522 | 509 | ||
523 | if (sc_reg < ARRAY_SIZE(sil24_scr_map)) { | 510 | if (sc_reg < ARRAY_SIZE(sil24_scr_map)) { |
524 | void __iomem *addr; | 511 | void __iomem *addr; |
@@ -531,7 +518,7 @@ static int sil24_scr_read(struct ata_port *ap, unsigned sc_reg, u32 *val) | |||
531 | 518 | ||
532 | static int sil24_scr_write(struct ata_port *ap, unsigned sc_reg, u32 val) | 519 | static int sil24_scr_write(struct ata_port *ap, unsigned sc_reg, u32 val) |
533 | { | 520 | { |
534 | void __iomem *scr_addr = ap->ioaddr.scr_addr; | 521 | void __iomem *scr_addr = sil24_port_base(ap) + PORT_SCONTROL; |
535 | 522 | ||
536 | if (sc_reg < ARRAY_SIZE(sil24_scr_map)) { | 523 | if (sc_reg < ARRAY_SIZE(sil24_scr_map)) { |
537 | void __iomem *addr; | 524 | void __iomem *addr; |
@@ -542,15 +529,9 @@ static int sil24_scr_write(struct ata_port *ap, unsigned sc_reg, u32 val) | |||
542 | return -EINVAL; | 529 | return -EINVAL; |
543 | } | 530 | } |
544 | 531 | ||
545 | static void sil24_tf_read(struct ata_port *ap, struct ata_taskfile *tf) | ||
546 | { | ||
547 | struct sil24_port_priv *pp = ap->private_data; | ||
548 | *tf = pp->tf; | ||
549 | } | ||
550 | |||
551 | static void sil24_config_port(struct ata_port *ap) | 532 | static void sil24_config_port(struct ata_port *ap) |
552 | { | 533 | { |
553 | void __iomem *port = ap->ioaddr.cmd_addr; | 534 | void __iomem *port = sil24_port_base(ap); |
554 | 535 | ||
555 | /* configure IRQ WoC */ | 536 | /* configure IRQ WoC */ |
556 | if (ap->flags & SIL24_FLAG_PCIX_IRQ_WOC) | 537 | if (ap->flags & SIL24_FLAG_PCIX_IRQ_WOC) |
@@ -575,7 +556,7 @@ static void sil24_config_port(struct ata_port *ap) | |||
575 | 556 | ||
576 | static void sil24_config_pmp(struct ata_port *ap, int attached) | 557 | static void sil24_config_pmp(struct ata_port *ap, int attached) |
577 | { | 558 | { |
578 | void __iomem *port = ap->ioaddr.cmd_addr; | 559 | void __iomem *port = sil24_port_base(ap); |
579 | 560 | ||
580 | if (attached) | 561 | if (attached) |
581 | writel(PORT_CS_PMP_EN, port + PORT_CTRL_STAT); | 562 | writel(PORT_CS_PMP_EN, port + PORT_CTRL_STAT); |
@@ -585,7 +566,7 @@ static void sil24_config_pmp(struct ata_port *ap, int attached) | |||
585 | 566 | ||
586 | static void sil24_clear_pmp(struct ata_port *ap) | 567 | static void sil24_clear_pmp(struct ata_port *ap) |
587 | { | 568 | { |
588 | void __iomem *port = ap->ioaddr.cmd_addr; | 569 | void __iomem *port = sil24_port_base(ap); |
589 | int i; | 570 | int i; |
590 | 571 | ||
591 | writel(PORT_CS_PMP_RESUME, port + PORT_CTRL_CLR); | 572 | writel(PORT_CS_PMP_RESUME, port + PORT_CTRL_CLR); |
@@ -600,12 +581,12 @@ static void sil24_clear_pmp(struct ata_port *ap) | |||
600 | 581 | ||
601 | static int sil24_init_port(struct ata_port *ap) | 582 | static int sil24_init_port(struct ata_port *ap) |
602 | { | 583 | { |
603 | void __iomem *port = ap->ioaddr.cmd_addr; | 584 | void __iomem *port = sil24_port_base(ap); |
604 | struct sil24_port_priv *pp = ap->private_data; | 585 | struct sil24_port_priv *pp = ap->private_data; |
605 | u32 tmp; | 586 | u32 tmp; |
606 | 587 | ||
607 | /* clear PMP error status */ | 588 | /* clear PMP error status */ |
608 | if (ap->nr_pmp_links) | 589 | if (sata_pmp_attached(ap)) |
609 | sil24_clear_pmp(ap); | 590 | sil24_clear_pmp(ap); |
610 | 591 | ||
611 | writel(PORT_CS_INIT, port + PORT_CTRL_STAT); | 592 | writel(PORT_CS_INIT, port + PORT_CTRL_STAT); |
@@ -616,7 +597,7 @@ static int sil24_init_port(struct ata_port *ap) | |||
616 | 597 | ||
617 | if ((tmp & (PORT_CS_INIT | PORT_CS_RDY)) != PORT_CS_RDY) { | 598 | if ((tmp & (PORT_CS_INIT | PORT_CS_RDY)) != PORT_CS_RDY) { |
618 | pp->do_port_rst = 1; | 599 | pp->do_port_rst = 1; |
619 | ap->link.eh_context.i.action |= ATA_EH_HARDRESET; | 600 | ap->link.eh_context.i.action |= ATA_EH_RESET; |
620 | return -EIO; | 601 | return -EIO; |
621 | } | 602 | } |
622 | 603 | ||
@@ -628,7 +609,7 @@ static int sil24_exec_polled_cmd(struct ata_port *ap, int pmp, | |||
628 | int is_cmd, u32 ctrl, | 609 | int is_cmd, u32 ctrl, |
629 | unsigned long timeout_msec) | 610 | unsigned long timeout_msec) |
630 | { | 611 | { |
631 | void __iomem *port = ap->ioaddr.cmd_addr; | 612 | void __iomem *port = sil24_port_base(ap); |
632 | struct sil24_port_priv *pp = ap->private_data; | 613 | struct sil24_port_priv *pp = ap->private_data; |
633 | struct sil24_prb *prb = &pp->cmd_block[0].ata.prb; | 614 | struct sil24_prb *prb = &pp->cmd_block[0].ata.prb; |
634 | dma_addr_t paddr = pp->cmd_block_dma; | 615 | dma_addr_t paddr = pp->cmd_block_dma; |
@@ -670,10 +651,11 @@ static int sil24_exec_polled_cmd(struct ata_port *ap, int pmp, | |||
670 | return rc; | 651 | return rc; |
671 | } | 652 | } |
672 | 653 | ||
673 | static int sil24_do_softreset(struct ata_link *link, unsigned int *class, | 654 | static int sil24_softreset(struct ata_link *link, unsigned int *class, |
674 | int pmp, unsigned long deadline) | 655 | unsigned long deadline) |
675 | { | 656 | { |
676 | struct ata_port *ap = link->ap; | 657 | struct ata_port *ap = link->ap; |
658 | int pmp = sata_srst_pmp(link); | ||
677 | unsigned long timeout_msec = 0; | 659 | unsigned long timeout_msec = 0; |
678 | struct ata_taskfile tf; | 660 | struct ata_taskfile tf; |
679 | const char *reason; | 661 | const char *reason; |
@@ -681,12 +663,6 @@ static int sil24_do_softreset(struct ata_link *link, unsigned int *class, | |||
681 | 663 | ||
682 | DPRINTK("ENTER\n"); | 664 | DPRINTK("ENTER\n"); |
683 | 665 | ||
684 | if (ata_link_offline(link)) { | ||
685 | DPRINTK("PHY reports no device\n"); | ||
686 | *class = ATA_DEV_NONE; | ||
687 | goto out; | ||
688 | } | ||
689 | |||
690 | /* put the port into known state */ | 666 | /* put the port into known state */ |
691 | if (sil24_init_port(ap)) { | 667 | if (sil24_init_port(ap)) { |
692 | reason = "port not ready"; | 668 | reason = "port not ready"; |
@@ -711,10 +687,6 @@ static int sil24_do_softreset(struct ata_link *link, unsigned int *class, | |||
711 | sil24_read_tf(ap, 0, &tf); | 687 | sil24_read_tf(ap, 0, &tf); |
712 | *class = ata_dev_classify(&tf); | 688 | *class = ata_dev_classify(&tf); |
713 | 689 | ||
714 | if (*class == ATA_DEV_UNKNOWN) | ||
715 | *class = ATA_DEV_NONE; | ||
716 | |||
717 | out: | ||
718 | DPRINTK("EXIT, class=%u\n", *class); | 690 | DPRINTK("EXIT, class=%u\n", *class); |
719 | return 0; | 691 | return 0; |
720 | 692 | ||
@@ -723,17 +695,11 @@ static int sil24_do_softreset(struct ata_link *link, unsigned int *class, | |||
723 | return -EIO; | 695 | return -EIO; |
724 | } | 696 | } |
725 | 697 | ||
726 | static int sil24_softreset(struct ata_link *link, unsigned int *class, | ||
727 | unsigned long deadline) | ||
728 | { | ||
729 | return sil24_do_softreset(link, class, SATA_PMP_CTRL_PORT, deadline); | ||
730 | } | ||
731 | |||
732 | static int sil24_hardreset(struct ata_link *link, unsigned int *class, | 698 | static int sil24_hardreset(struct ata_link *link, unsigned int *class, |
733 | unsigned long deadline) | 699 | unsigned long deadline) |
734 | { | 700 | { |
735 | struct ata_port *ap = link->ap; | 701 | struct ata_port *ap = link->ap; |
736 | void __iomem *port = ap->ioaddr.cmd_addr; | 702 | void __iomem *port = sil24_port_base(ap); |
737 | struct sil24_port_priv *pp = ap->private_data; | 703 | struct sil24_port_priv *pp = ap->private_data; |
738 | int did_port_rst = 0; | 704 | int did_port_rst = 0; |
739 | const char *reason; | 705 | const char *reason; |
@@ -911,7 +877,7 @@ static unsigned int sil24_qc_issue(struct ata_queued_cmd *qc) | |||
911 | { | 877 | { |
912 | struct ata_port *ap = qc->ap; | 878 | struct ata_port *ap = qc->ap; |
913 | struct sil24_port_priv *pp = ap->private_data; | 879 | struct sil24_port_priv *pp = ap->private_data; |
914 | void __iomem *port = ap->ioaddr.cmd_addr; | 880 | void __iomem *port = sil24_port_base(ap); |
915 | unsigned int tag = sil24_tag(qc->tag); | 881 | unsigned int tag = sil24_tag(qc->tag); |
916 | dma_addr_t paddr; | 882 | dma_addr_t paddr; |
917 | void __iomem *activate; | 883 | void __iomem *activate; |
@@ -925,9 +891,10 @@ static unsigned int sil24_qc_issue(struct ata_queued_cmd *qc) | |||
925 | return 0; | 891 | return 0; |
926 | } | 892 | } |
927 | 893 | ||
928 | static void sil24_irq_clear(struct ata_port *ap) | 894 | static bool sil24_qc_fill_rtf(struct ata_queued_cmd *qc) |
929 | { | 895 | { |
930 | /* unused */ | 896 | sil24_read_tf(qc->ap, qc->tag, &qc->result_tf); |
897 | return true; | ||
931 | } | 898 | } |
932 | 899 | ||
933 | static void sil24_pmp_attach(struct ata_port *ap) | 900 | static void sil24_pmp_attach(struct ata_port *ap) |
@@ -942,12 +909,6 @@ static void sil24_pmp_detach(struct ata_port *ap) | |||
942 | sil24_config_pmp(ap, 0); | 909 | sil24_config_pmp(ap, 0); |
943 | } | 910 | } |
944 | 911 | ||
945 | static int sil24_pmp_softreset(struct ata_link *link, unsigned int *class, | ||
946 | unsigned long deadline) | ||
947 | { | ||
948 | return sil24_do_softreset(link, class, link->pmp, deadline); | ||
949 | } | ||
950 | |||
951 | static int sil24_pmp_hardreset(struct ata_link *link, unsigned int *class, | 912 | static int sil24_pmp_hardreset(struct ata_link *link, unsigned int *class, |
952 | unsigned long deadline) | 913 | unsigned long deadline) |
953 | { | 914 | { |
@@ -960,12 +921,12 @@ static int sil24_pmp_hardreset(struct ata_link *link, unsigned int *class, | |||
960 | return rc; | 921 | return rc; |
961 | } | 922 | } |
962 | 923 | ||
963 | return sata_pmp_std_hardreset(link, class, deadline); | 924 | return sata_std_hardreset(link, class, deadline); |
964 | } | 925 | } |
965 | 926 | ||
966 | static void sil24_freeze(struct ata_port *ap) | 927 | static void sil24_freeze(struct ata_port *ap) |
967 | { | 928 | { |
968 | void __iomem *port = ap->ioaddr.cmd_addr; | 929 | void __iomem *port = sil24_port_base(ap); |
969 | 930 | ||
970 | /* Port-wide IRQ mask in HOST_CTRL doesn't really work, clear | 931 | /* Port-wide IRQ mask in HOST_CTRL doesn't really work, clear |
971 | * PORT_IRQ_ENABLE instead. | 932 | * PORT_IRQ_ENABLE instead. |
@@ -975,7 +936,7 @@ static void sil24_freeze(struct ata_port *ap) | |||
975 | 936 | ||
976 | static void sil24_thaw(struct ata_port *ap) | 937 | static void sil24_thaw(struct ata_port *ap) |
977 | { | 938 | { |
978 | void __iomem *port = ap->ioaddr.cmd_addr; | 939 | void __iomem *port = sil24_port_base(ap); |
979 | u32 tmp; | 940 | u32 tmp; |
980 | 941 | ||
981 | /* clear IRQ */ | 942 | /* clear IRQ */ |
@@ -988,7 +949,7 @@ static void sil24_thaw(struct ata_port *ap) | |||
988 | 949 | ||
989 | static void sil24_error_intr(struct ata_port *ap) | 950 | static void sil24_error_intr(struct ata_port *ap) |
990 | { | 951 | { |
991 | void __iomem *port = ap->ioaddr.cmd_addr; | 952 | void __iomem *port = sil24_port_base(ap); |
992 | struct sil24_port_priv *pp = ap->private_data; | 953 | struct sil24_port_priv *pp = ap->private_data; |
993 | struct ata_queued_cmd *qc = NULL; | 954 | struct ata_queued_cmd *qc = NULL; |
994 | struct ata_link *link; | 955 | struct ata_link *link; |
@@ -1022,7 +983,7 @@ static void sil24_error_intr(struct ata_port *ap) | |||
1022 | 983 | ||
1023 | if (irq_stat & PORT_IRQ_UNK_FIS) { | 984 | if (irq_stat & PORT_IRQ_UNK_FIS) { |
1024 | ehi->err_mask |= AC_ERR_HSM; | 985 | ehi->err_mask |= AC_ERR_HSM; |
1025 | ehi->action |= ATA_EH_SOFTRESET; | 986 | ehi->action |= ATA_EH_RESET; |
1026 | ata_ehi_push_desc(ehi, "unknown FIS"); | 987 | ata_ehi_push_desc(ehi, "unknown FIS"); |
1027 | freeze = 1; | 988 | freeze = 1; |
1028 | } | 989 | } |
@@ -1043,14 +1004,14 @@ static void sil24_error_intr(struct ata_port *ap) | |||
1043 | */ | 1004 | */ |
1044 | if (ap->nr_active_links >= 3) { | 1005 | if (ap->nr_active_links >= 3) { |
1045 | ehi->err_mask |= AC_ERR_OTHER; | 1006 | ehi->err_mask |= AC_ERR_OTHER; |
1046 | ehi->action |= ATA_EH_HARDRESET; | 1007 | ehi->action |= ATA_EH_RESET; |
1047 | ata_ehi_push_desc(ehi, "PMP DMA CS errata"); | 1008 | ata_ehi_push_desc(ehi, "PMP DMA CS errata"); |
1048 | pp->do_port_rst = 1; | 1009 | pp->do_port_rst = 1; |
1049 | freeze = 1; | 1010 | freeze = 1; |
1050 | } | 1011 | } |
1051 | 1012 | ||
1052 | /* find out the offending link and qc */ | 1013 | /* find out the offending link and qc */ |
1053 | if (ap->nr_pmp_links) { | 1014 | if (sata_pmp_attached(ap)) { |
1054 | context = readl(port + PORT_CONTEXT); | 1015 | context = readl(port + PORT_CONTEXT); |
1055 | pmp = (context >> 5) & 0xf; | 1016 | pmp = (context >> 5) & 0xf; |
1056 | 1017 | ||
@@ -1064,7 +1025,7 @@ static void sil24_error_intr(struct ata_port *ap) | |||
1064 | irq_stat); | 1025 | irq_stat); |
1065 | } else { | 1026 | } else { |
1066 | err_mask |= AC_ERR_HSM; | 1027 | err_mask |= AC_ERR_HSM; |
1067 | action |= ATA_EH_HARDRESET; | 1028 | action |= ATA_EH_RESET; |
1068 | freeze = 1; | 1029 | freeze = 1; |
1069 | } | 1030 | } |
1070 | } else | 1031 | } else |
@@ -1078,28 +1039,27 @@ static void sil24_error_intr(struct ata_port *ap) | |||
1078 | if (ci && ci->desc) { | 1039 | if (ci && ci->desc) { |
1079 | err_mask |= ci->err_mask; | 1040 | err_mask |= ci->err_mask; |
1080 | action |= ci->action; | 1041 | action |= ci->action; |
1081 | if (action & ATA_EH_RESET_MASK) | 1042 | if (action & ATA_EH_RESET) |
1082 | freeze = 1; | 1043 | freeze = 1; |
1083 | ata_ehi_push_desc(ehi, "%s", ci->desc); | 1044 | ata_ehi_push_desc(ehi, "%s", ci->desc); |
1084 | } else { | 1045 | } else { |
1085 | err_mask |= AC_ERR_OTHER; | 1046 | err_mask |= AC_ERR_OTHER; |
1086 | action |= ATA_EH_SOFTRESET; | 1047 | action |= ATA_EH_RESET; |
1087 | freeze = 1; | 1048 | freeze = 1; |
1088 | ata_ehi_push_desc(ehi, "unknown command error %d", | 1049 | ata_ehi_push_desc(ehi, "unknown command error %d", |
1089 | cerr); | 1050 | cerr); |
1090 | } | 1051 | } |
1091 | 1052 | ||
1092 | /* record error info */ | 1053 | /* record error info */ |
1093 | if (qc) { | 1054 | if (qc) |
1094 | sil24_read_tf(ap, qc->tag, &pp->tf); | ||
1095 | qc->err_mask |= err_mask; | 1055 | qc->err_mask |= err_mask; |
1096 | } else | 1056 | else |
1097 | ehi->err_mask |= err_mask; | 1057 | ehi->err_mask |= err_mask; |
1098 | 1058 | ||
1099 | ehi->action |= action; | 1059 | ehi->action |= action; |
1100 | 1060 | ||
1101 | /* if PMP, resume */ | 1061 | /* if PMP, resume */ |
1102 | if (ap->nr_pmp_links) | 1062 | if (sata_pmp_attached(ap)) |
1103 | writel(PORT_CS_PMP_RESUME, port + PORT_CTRL_STAT); | 1063 | writel(PORT_CS_PMP_RESUME, port + PORT_CTRL_STAT); |
1104 | } | 1064 | } |
1105 | 1065 | ||
@@ -1114,18 +1074,9 @@ static void sil24_error_intr(struct ata_port *ap) | |||
1114 | } | 1074 | } |
1115 | } | 1075 | } |
1116 | 1076 | ||
1117 | static void sil24_finish_qc(struct ata_queued_cmd *qc) | ||
1118 | { | ||
1119 | struct ata_port *ap = qc->ap; | ||
1120 | struct sil24_port_priv *pp = ap->private_data; | ||
1121 | |||
1122 | if (qc->flags & ATA_QCFLAG_RESULT_TF) | ||
1123 | sil24_read_tf(ap, qc->tag, &pp->tf); | ||
1124 | } | ||
1125 | |||
1126 | static inline void sil24_host_intr(struct ata_port *ap) | 1077 | static inline void sil24_host_intr(struct ata_port *ap) |
1127 | { | 1078 | { |
1128 | void __iomem *port = ap->ioaddr.cmd_addr; | 1079 | void __iomem *port = sil24_port_base(ap); |
1129 | u32 slot_stat, qc_active; | 1080 | u32 slot_stat, qc_active; |
1130 | int rc; | 1081 | int rc; |
1131 | 1082 | ||
@@ -1147,13 +1098,13 @@ static inline void sil24_host_intr(struct ata_port *ap) | |||
1147 | } | 1098 | } |
1148 | 1099 | ||
1149 | qc_active = slot_stat & ~HOST_SSTAT_ATTN; | 1100 | qc_active = slot_stat & ~HOST_SSTAT_ATTN; |
1150 | rc = ata_qc_complete_multiple(ap, qc_active, sil24_finish_qc); | 1101 | rc = ata_qc_complete_multiple(ap, qc_active); |
1151 | if (rc > 0) | 1102 | if (rc > 0) |
1152 | return; | 1103 | return; |
1153 | if (rc < 0) { | 1104 | if (rc < 0) { |
1154 | struct ata_eh_info *ehi = &ap->link.eh_info; | 1105 | struct ata_eh_info *ehi = &ap->link.eh_info; |
1155 | ehi->err_mask |= AC_ERR_HSM; | 1106 | ehi->err_mask |= AC_ERR_HSM; |
1156 | ehi->action |= ATA_EH_SOFTRESET; | 1107 | ehi->action |= ATA_EH_RESET; |
1157 | ata_port_freeze(ap); | 1108 | ata_port_freeze(ap); |
1158 | return; | 1109 | return; |
1159 | } | 1110 | } |
@@ -1209,11 +1160,7 @@ static void sil24_error_handler(struct ata_port *ap) | |||
1209 | if (sil24_init_port(ap)) | 1160 | if (sil24_init_port(ap)) |
1210 | ata_eh_freeze_port(ap); | 1161 | ata_eh_freeze_port(ap); |
1211 | 1162 | ||
1212 | /* perform recovery */ | 1163 | sata_pmp_error_handler(ap); |
1213 | sata_pmp_do_eh(ap, ata_std_prereset, sil24_softreset, sil24_hardreset, | ||
1214 | ata_std_postreset, sata_pmp_std_prereset, | ||
1215 | sil24_pmp_softreset, sil24_pmp_hardreset, | ||
1216 | sata_pmp_std_postreset); | ||
1217 | 1164 | ||
1218 | pp->do_port_rst = 0; | 1165 | pp->do_port_rst = 0; |
1219 | } | 1166 | } |
@@ -1239,8 +1186,6 @@ static int sil24_port_start(struct ata_port *ap) | |||
1239 | if (!pp) | 1186 | if (!pp) |
1240 | return -ENOMEM; | 1187 | return -ENOMEM; |
1241 | 1188 | ||
1242 | pp->tf.command = ATA_DRDY; | ||
1243 | |||
1244 | cb = dmam_alloc_coherent(dev, cb_size, &cb_dma, GFP_KERNEL); | 1189 | cb = dmam_alloc_coherent(dev, cb_size, &cb_dma, GFP_KERNEL); |
1245 | if (!cb) | 1190 | if (!cb) |
1246 | return -ENOMEM; | 1191 | return -ENOMEM; |
@@ -1251,6 +1196,9 @@ static int sil24_port_start(struct ata_port *ap) | |||
1251 | 1196 | ||
1252 | ap->private_data = pp; | 1197 | ap->private_data = pp; |
1253 | 1198 | ||
1199 | ata_port_pbar_desc(ap, SIL24_HOST_BAR, -1, "host"); | ||
1200 | ata_port_pbar_desc(ap, SIL24_PORT_BAR, sil24_port_offset(ap), "port"); | ||
1201 | |||
1254 | return 0; | 1202 | return 0; |
1255 | } | 1203 | } |
1256 | 1204 | ||
@@ -1269,7 +1217,8 @@ static void sil24_init_controller(struct ata_host *host) | |||
1269 | /* init ports */ | 1217 | /* init ports */ |
1270 | for (i = 0; i < host->n_ports; i++) { | 1218 | for (i = 0; i < host->n_ports; i++) { |
1271 | struct ata_port *ap = host->ports[i]; | 1219 | struct ata_port *ap = host->ports[i]; |
1272 | void __iomem *port = ap->ioaddr.cmd_addr; | 1220 | void __iomem *port = sil24_port_base(ap); |
1221 | |||
1273 | 1222 | ||
1274 | /* Initial PHY setting */ | 1223 | /* Initial PHY setting */ |
1275 | writel(0x20c, port + PORT_PHY_CFG); | 1224 | writel(0x20c, port + PORT_PHY_CFG); |
@@ -1302,7 +1251,7 @@ static int sil24_init_one(struct pci_dev *pdev, const struct pci_device_id *ent) | |||
1302 | const struct ata_port_info *ppi[] = { &pi, NULL }; | 1251 | const struct ata_port_info *ppi[] = { &pi, NULL }; |
1303 | void __iomem * const *iomap; | 1252 | void __iomem * const *iomap; |
1304 | struct ata_host *host; | 1253 | struct ata_host *host; |
1305 | int i, rc; | 1254 | int rc; |
1306 | u32 tmp; | 1255 | u32 tmp; |
1307 | 1256 | ||
1308 | /* cause link error if sil24_cmd_block is sized wrongly */ | 1257 | /* cause link error if sil24_cmd_block is sized wrongly */ |
@@ -1342,18 +1291,6 @@ static int sil24_init_one(struct pci_dev *pdev, const struct pci_device_id *ent) | |||
1342 | return -ENOMEM; | 1291 | return -ENOMEM; |
1343 | host->iomap = iomap; | 1292 | host->iomap = iomap; |
1344 | 1293 | ||
1345 | for (i = 0; i < host->n_ports; i++) { | ||
1346 | struct ata_port *ap = host->ports[i]; | ||
1347 | size_t offset = ap->port_no * PORT_REGS_SIZE; | ||
1348 | void __iomem *port = iomap[SIL24_PORT_BAR] + offset; | ||
1349 | |||
1350 | host->ports[i]->ioaddr.cmd_addr = port; | ||
1351 | host->ports[i]->ioaddr.scr_addr = port + PORT_SCONTROL; | ||
1352 | |||
1353 | ata_port_pbar_desc(ap, SIL24_HOST_BAR, -1, "host"); | ||
1354 | ata_port_pbar_desc(ap, SIL24_PORT_BAR, offset, "port"); | ||
1355 | } | ||
1356 | |||
1357 | /* configure and activate the device */ | 1294 | /* configure and activate the device */ |
1358 | if (!pci_set_dma_mask(pdev, DMA_64BIT_MASK)) { | 1295 | if (!pci_set_dma_mask(pdev, DMA_64BIT_MASK)) { |
1359 | rc = pci_set_consistent_dma_mask(pdev, DMA_64BIT_MASK); | 1296 | rc = pci_set_consistent_dma_mask(pdev, DMA_64BIT_MASK); |
diff --git a/drivers/ata/sata_sis.c b/drivers/ata/sata_sis.c index a01260a56432..6b8e45ba32e8 100644 --- a/drivers/ata/sata_sis.c +++ b/drivers/ata/sata_sis.c | |||
@@ -86,45 +86,13 @@ static struct pci_driver sis_pci_driver = { | |||
86 | }; | 86 | }; |
87 | 87 | ||
88 | static struct scsi_host_template sis_sht = { | 88 | static struct scsi_host_template sis_sht = { |
89 | .module = THIS_MODULE, | 89 | ATA_BMDMA_SHT(DRV_NAME), |
90 | .name = DRV_NAME, | ||
91 | .ioctl = ata_scsi_ioctl, | ||
92 | .queuecommand = ata_scsi_queuecmd, | ||
93 | .can_queue = ATA_DEF_QUEUE, | ||
94 | .this_id = ATA_SHT_THIS_ID, | ||
95 | .sg_tablesize = LIBATA_MAX_PRD, | ||
96 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
97 | .emulated = ATA_SHT_EMULATED, | ||
98 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
99 | .proc_name = DRV_NAME, | ||
100 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
101 | .slave_configure = ata_scsi_slave_config, | ||
102 | .slave_destroy = ata_scsi_slave_destroy, | ||
103 | .bios_param = ata_std_bios_param, | ||
104 | }; | 90 | }; |
105 | 91 | ||
106 | static const struct ata_port_operations sis_ops = { | 92 | static struct ata_port_operations sis_ops = { |
107 | .tf_load = ata_tf_load, | 93 | .inherits = &ata_bmdma_port_ops, |
108 | .tf_read = ata_tf_read, | ||
109 | .check_status = ata_check_status, | ||
110 | .exec_command = ata_exec_command, | ||
111 | .dev_select = ata_std_dev_select, | ||
112 | .bmdma_setup = ata_bmdma_setup, | ||
113 | .bmdma_start = ata_bmdma_start, | ||
114 | .bmdma_stop = ata_bmdma_stop, | ||
115 | .bmdma_status = ata_bmdma_status, | ||
116 | .qc_prep = ata_qc_prep, | ||
117 | .qc_issue = ata_qc_issue_prot, | ||
118 | .data_xfer = ata_data_xfer, | ||
119 | .freeze = ata_bmdma_freeze, | ||
120 | .thaw = ata_bmdma_thaw, | ||
121 | .error_handler = ata_bmdma_error_handler, | ||
122 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
123 | .irq_clear = ata_bmdma_irq_clear, | ||
124 | .irq_on = ata_irq_on, | ||
125 | .scr_read = sis_scr_read, | 94 | .scr_read = sis_scr_read, |
126 | .scr_write = sis_scr_write, | 95 | .scr_write = sis_scr_write, |
127 | .port_start = ata_port_start, | ||
128 | }; | 96 | }; |
129 | 97 | ||
130 | static const struct ata_port_info sis_port_info = { | 98 | static const struct ata_port_info sis_port_info = { |
@@ -341,7 +309,7 @@ static int sis_init_one(struct pci_dev *pdev, const struct pci_device_id *ent) | |||
341 | break; | 309 | break; |
342 | } | 310 | } |
343 | 311 | ||
344 | rc = ata_pci_prepare_sff_host(pdev, ppi, &host); | 312 | rc = ata_pci_sff_prepare_host(pdev, ppi, &host); |
345 | if (rc) | 313 | if (rc) |
346 | return rc; | 314 | return rc; |
347 | 315 | ||
@@ -359,8 +327,8 @@ static int sis_init_one(struct pci_dev *pdev, const struct pci_device_id *ent) | |||
359 | 327 | ||
360 | pci_set_master(pdev); | 328 | pci_set_master(pdev); |
361 | pci_intx(pdev, 1); | 329 | pci_intx(pdev, 1); |
362 | return ata_host_activate(host, pdev->irq, ata_interrupt, IRQF_SHARED, | 330 | return ata_host_activate(host, pdev->irq, ata_sff_interrupt, |
363 | &sis_sht); | 331 | IRQF_SHARED, &sis_sht); |
364 | } | 332 | } |
365 | 333 | ||
366 | static int __init sis_init(void) | 334 | static int __init sis_init(void) |
diff --git a/drivers/ata/sata_svw.c b/drivers/ata/sata_svw.c index 019e367b59fc..16aa6839aa5a 100644 --- a/drivers/ata/sata_svw.c +++ b/drivers/ata/sata_svw.c | |||
@@ -233,7 +233,7 @@ static void k2_bmdma_setup_mmio(struct ata_queued_cmd *qc) | |||
233 | 233 | ||
234 | /* issue r/w command if this is not a ATA DMA command*/ | 234 | /* issue r/w command if this is not a ATA DMA command*/ |
235 | if (qc->tf.protocol != ATA_PROT_DMA) | 235 | if (qc->tf.protocol != ATA_PROT_DMA) |
236 | ap->ops->exec_command(ap, &qc->tf); | 236 | ap->ops->sff_exec_command(ap, &qc->tf); |
237 | } | 237 | } |
238 | 238 | ||
239 | /** | 239 | /** |
@@ -269,7 +269,7 @@ static void k2_bmdma_start_mmio(struct ata_queued_cmd *qc) | |||
269 | and the start command. */ | 269 | and the start command. */ |
270 | /* issue r/w command if the access is to ATA*/ | 270 | /* issue r/w command if the access is to ATA*/ |
271 | if (qc->tf.protocol == ATA_PROT_DMA) | 271 | if (qc->tf.protocol == ATA_PROT_DMA) |
272 | ap->ops->exec_command(ap, &qc->tf); | 272 | ap->ops->sff_exec_command(ap, &qc->tf); |
273 | } | 273 | } |
274 | 274 | ||
275 | 275 | ||
@@ -327,50 +327,23 @@ static int k2_sata_proc_info(struct Scsi_Host *shost, char *page, char **start, | |||
327 | 327 | ||
328 | 328 | ||
329 | static struct scsi_host_template k2_sata_sht = { | 329 | static struct scsi_host_template k2_sata_sht = { |
330 | .module = THIS_MODULE, | 330 | ATA_BMDMA_SHT(DRV_NAME), |
331 | .name = DRV_NAME, | ||
332 | .ioctl = ata_scsi_ioctl, | ||
333 | .queuecommand = ata_scsi_queuecmd, | ||
334 | .can_queue = ATA_DEF_QUEUE, | ||
335 | .this_id = ATA_SHT_THIS_ID, | ||
336 | .sg_tablesize = LIBATA_MAX_PRD, | ||
337 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
338 | .emulated = ATA_SHT_EMULATED, | ||
339 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
340 | .proc_name = DRV_NAME, | ||
341 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
342 | .slave_configure = ata_scsi_slave_config, | ||
343 | .slave_destroy = ata_scsi_slave_destroy, | ||
344 | #ifdef CONFIG_PPC_OF | 331 | #ifdef CONFIG_PPC_OF |
345 | .proc_info = k2_sata_proc_info, | 332 | .proc_info = k2_sata_proc_info, |
346 | #endif | 333 | #endif |
347 | .bios_param = ata_std_bios_param, | ||
348 | }; | 334 | }; |
349 | 335 | ||
350 | 336 | ||
351 | static const struct ata_port_operations k2_sata_ops = { | 337 | static struct ata_port_operations k2_sata_ops = { |
352 | .tf_load = k2_sata_tf_load, | 338 | .inherits = &ata_bmdma_port_ops, |
353 | .tf_read = k2_sata_tf_read, | 339 | .sff_tf_load = k2_sata_tf_load, |
354 | .check_status = k2_stat_check_status, | 340 | .sff_tf_read = k2_sata_tf_read, |
355 | .exec_command = ata_exec_command, | 341 | .sff_check_status = k2_stat_check_status, |
356 | .dev_select = ata_std_dev_select, | ||
357 | .check_atapi_dma = k2_sata_check_atapi_dma, | 342 | .check_atapi_dma = k2_sata_check_atapi_dma, |
358 | .bmdma_setup = k2_bmdma_setup_mmio, | 343 | .bmdma_setup = k2_bmdma_setup_mmio, |
359 | .bmdma_start = k2_bmdma_start_mmio, | 344 | .bmdma_start = k2_bmdma_start_mmio, |
360 | .bmdma_stop = ata_bmdma_stop, | ||
361 | .bmdma_status = ata_bmdma_status, | ||
362 | .qc_prep = ata_qc_prep, | ||
363 | .qc_issue = ata_qc_issue_prot, | ||
364 | .data_xfer = ata_data_xfer, | ||
365 | .freeze = ata_bmdma_freeze, | ||
366 | .thaw = ata_bmdma_thaw, | ||
367 | .error_handler = ata_bmdma_error_handler, | ||
368 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
369 | .irq_clear = ata_bmdma_irq_clear, | ||
370 | .irq_on = ata_irq_on, | ||
371 | .scr_read = k2_sata_scr_read, | 345 | .scr_read = k2_sata_scr_read, |
372 | .scr_write = k2_sata_scr_write, | 346 | .scr_write = k2_sata_scr_write, |
373 | .port_start = ata_port_start, | ||
374 | }; | 347 | }; |
375 | 348 | ||
376 | static const struct ata_port_info k2_port_info[] = { | 349 | static const struct ata_port_info k2_port_info[] = { |
@@ -519,8 +492,8 @@ static int k2_sata_init_one(struct pci_dev *pdev, const struct pci_device_id *en | |||
519 | writel(0x0, mmio_base + K2_SATA_SIM_OFFSET); | 492 | writel(0x0, mmio_base + K2_SATA_SIM_OFFSET); |
520 | 493 | ||
521 | pci_set_master(pdev); | 494 | pci_set_master(pdev); |
522 | return ata_host_activate(host, pdev->irq, ata_interrupt, IRQF_SHARED, | 495 | return ata_host_activate(host, pdev->irq, ata_sff_interrupt, |
523 | &k2_sata_sht); | 496 | IRQF_SHARED, &k2_sata_sht); |
524 | } | 497 | } |
525 | 498 | ||
526 | /* 0x240 is device ID for Apple K2 device | 499 | /* 0x240 is device ID for Apple K2 device |
diff --git a/drivers/ata/sata_sx4.c b/drivers/ata/sata_sx4.c index e3d56bc6726d..ec04b8d3c791 100644 --- a/drivers/ata/sata_sx4.c +++ b/drivers/ata/sata_sx4.c | |||
@@ -232,40 +232,30 @@ static void pdc20621_get_from_dimm(struct ata_host *host, | |||
232 | static void pdc20621_put_to_dimm(struct ata_host *host, | 232 | static void pdc20621_put_to_dimm(struct ata_host *host, |
233 | void *psource, u32 offset, u32 size); | 233 | void *psource, u32 offset, u32 size); |
234 | static void pdc20621_irq_clear(struct ata_port *ap); | 234 | static void pdc20621_irq_clear(struct ata_port *ap); |
235 | static unsigned int pdc20621_qc_issue_prot(struct ata_queued_cmd *qc); | 235 | static unsigned int pdc20621_qc_issue(struct ata_queued_cmd *qc); |
236 | 236 | ||
237 | 237 | ||
238 | static struct scsi_host_template pdc_sata_sht = { | 238 | static struct scsi_host_template pdc_sata_sht = { |
239 | .module = THIS_MODULE, | 239 | ATA_BASE_SHT(DRV_NAME), |
240 | .name = DRV_NAME, | ||
241 | .ioctl = ata_scsi_ioctl, | ||
242 | .queuecommand = ata_scsi_queuecmd, | ||
243 | .can_queue = ATA_DEF_QUEUE, | ||
244 | .this_id = ATA_SHT_THIS_ID, | ||
245 | .sg_tablesize = LIBATA_MAX_PRD, | 240 | .sg_tablesize = LIBATA_MAX_PRD, |
246 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
247 | .emulated = ATA_SHT_EMULATED, | ||
248 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
249 | .proc_name = DRV_NAME, | ||
250 | .dma_boundary = ATA_DMA_BOUNDARY, | 241 | .dma_boundary = ATA_DMA_BOUNDARY, |
251 | .slave_configure = ata_scsi_slave_config, | ||
252 | .slave_destroy = ata_scsi_slave_destroy, | ||
253 | .bios_param = ata_std_bios_param, | ||
254 | }; | 242 | }; |
255 | 243 | ||
256 | static const struct ata_port_operations pdc_20621_ops = { | 244 | /* TODO: inherit from base port_ops after converting to new EH */ |
257 | .tf_load = pdc_tf_load_mmio, | 245 | static struct ata_port_operations pdc_20621_ops = { |
258 | .tf_read = ata_tf_read, | 246 | .sff_tf_load = pdc_tf_load_mmio, |
259 | .check_status = ata_check_status, | 247 | .sff_tf_read = ata_sff_tf_read, |
260 | .exec_command = pdc_exec_command_mmio, | 248 | .sff_check_status = ata_sff_check_status, |
261 | .dev_select = ata_std_dev_select, | 249 | .sff_exec_command = pdc_exec_command_mmio, |
250 | .sff_dev_select = ata_sff_dev_select, | ||
262 | .phy_reset = pdc_20621_phy_reset, | 251 | .phy_reset = pdc_20621_phy_reset, |
263 | .qc_prep = pdc20621_qc_prep, | 252 | .qc_prep = pdc20621_qc_prep, |
264 | .qc_issue = pdc20621_qc_issue_prot, | 253 | .qc_issue = pdc20621_qc_issue, |
265 | .data_xfer = ata_data_xfer, | 254 | .qc_fill_rtf = ata_sff_qc_fill_rtf, |
255 | .sff_data_xfer = ata_sff_data_xfer, | ||
266 | .eng_timeout = pdc_eng_timeout, | 256 | .eng_timeout = pdc_eng_timeout, |
267 | .irq_clear = pdc20621_irq_clear, | 257 | .sff_irq_clear = pdc20621_irq_clear, |
268 | .irq_on = ata_irq_on, | 258 | .sff_irq_on = ata_sff_irq_on, |
269 | .port_start = pdc_port_start, | 259 | .port_start = pdc_port_start, |
270 | }; | 260 | }; |
271 | 261 | ||
@@ -475,7 +465,7 @@ static void pdc20621_dma_prep(struct ata_queued_cmd *qc) | |||
475 | void __iomem *dimm_mmio = ap->host->iomap[PDC_DIMM_BAR]; | 465 | void __iomem *dimm_mmio = ap->host->iomap[PDC_DIMM_BAR]; |
476 | unsigned int portno = ap->port_no; | 466 | unsigned int portno = ap->port_no; |
477 | unsigned int i, si, idx, total_len = 0, sgt_len; | 467 | unsigned int i, si, idx, total_len = 0, sgt_len; |
478 | u32 *buf = (u32 *) &pp->dimm_buf[PDC_DIMM_HEADER_SZ]; | 468 | __le32 *buf = (__le32 *) &pp->dimm_buf[PDC_DIMM_HEADER_SZ]; |
479 | 469 | ||
480 | WARN_ON(!(qc->flags & ATA_QCFLAG_DMAMAP)); | 470 | WARN_ON(!(qc->flags & ATA_QCFLAG_DMAMAP)); |
481 | 471 | ||
@@ -693,7 +683,7 @@ static void pdc20621_packet_start(struct ata_queued_cmd *qc) | |||
693 | } | 683 | } |
694 | } | 684 | } |
695 | 685 | ||
696 | static unsigned int pdc20621_qc_issue_prot(struct ata_queued_cmd *qc) | 686 | static unsigned int pdc20621_qc_issue(struct ata_queued_cmd *qc) |
697 | { | 687 | { |
698 | switch (qc->tf.protocol) { | 688 | switch (qc->tf.protocol) { |
699 | case ATA_PROT_DMA: | 689 | case ATA_PROT_DMA: |
@@ -709,7 +699,7 @@ static unsigned int pdc20621_qc_issue_prot(struct ata_queued_cmd *qc) | |||
709 | break; | 699 | break; |
710 | } | 700 | } |
711 | 701 | ||
712 | return ata_qc_issue_prot(qc); | 702 | return ata_sff_qc_issue(qc); |
713 | } | 703 | } |
714 | 704 | ||
715 | static inline unsigned int pdc20621_host_intr(struct ata_port *ap, | 705 | static inline unsigned int pdc20621_host_intr(struct ata_port *ap, |
@@ -781,7 +771,7 @@ static inline unsigned int pdc20621_host_intr(struct ata_port *ap, | |||
781 | /* command completion, but no data xfer */ | 771 | /* command completion, but no data xfer */ |
782 | } else if (qc->tf.protocol == ATA_PROT_NODATA) { | 772 | } else if (qc->tf.protocol == ATA_PROT_NODATA) { |
783 | 773 | ||
784 | status = ata_busy_wait(ap, ATA_BUSY | ATA_DRQ, 1000); | 774 | status = ata_sff_busy_wait(ap, ATA_BUSY | ATA_DRQ, 1000); |
785 | DPRINTK("BUS_NODATA (drv_stat 0x%X)\n", status); | 775 | DPRINTK("BUS_NODATA (drv_stat 0x%X)\n", status); |
786 | qc->err_mask |= ac_err_mask(status); | 776 | qc->err_mask |= ac_err_mask(status); |
787 | ata_qc_complete(qc); | 777 | ata_qc_complete(qc); |
@@ -890,7 +880,7 @@ static void pdc_eng_timeout(struct ata_port *ap) | |||
890 | break; | 880 | break; |
891 | 881 | ||
892 | default: | 882 | default: |
893 | drv_stat = ata_busy_wait(ap, ATA_BUSY | ATA_DRQ, 1000); | 883 | drv_stat = ata_sff_busy_wait(ap, ATA_BUSY | ATA_DRQ, 1000); |
894 | 884 | ||
895 | ata_port_printk(ap, KERN_ERR, | 885 | ata_port_printk(ap, KERN_ERR, |
896 | "unknown timeout, cmd 0x%x stat 0x%x\n", | 886 | "unknown timeout, cmd 0x%x stat 0x%x\n", |
@@ -909,7 +899,7 @@ static void pdc_tf_load_mmio(struct ata_port *ap, const struct ata_taskfile *tf) | |||
909 | { | 899 | { |
910 | WARN_ON(tf->protocol == ATA_PROT_DMA || | 900 | WARN_ON(tf->protocol == ATA_PROT_DMA || |
911 | tf->protocol == ATA_PROT_NODATA); | 901 | tf->protocol == ATA_PROT_NODATA); |
912 | ata_tf_load(ap, tf); | 902 | ata_sff_tf_load(ap, tf); |
913 | } | 903 | } |
914 | 904 | ||
915 | 905 | ||
@@ -917,7 +907,7 @@ static void pdc_exec_command_mmio(struct ata_port *ap, const struct ata_taskfile | |||
917 | { | 907 | { |
918 | WARN_ON(tf->protocol == ATA_PROT_DMA || | 908 | WARN_ON(tf->protocol == ATA_PROT_DMA || |
919 | tf->protocol == ATA_PROT_NODATA); | 909 | tf->protocol == ATA_PROT_NODATA); |
920 | ata_exec_command(ap, tf); | 910 | ata_sff_exec_command(ap, tf); |
921 | } | 911 | } |
922 | 912 | ||
923 | 913 | ||
diff --git a/drivers/ata/sata_uli.c b/drivers/ata/sata_uli.c index e710e71b7b92..f277cea904ce 100644 --- a/drivers/ata/sata_uli.c +++ b/drivers/ata/sata_uli.c | |||
@@ -76,50 +76,13 @@ static struct pci_driver uli_pci_driver = { | |||
76 | }; | 76 | }; |
77 | 77 | ||
78 | static struct scsi_host_template uli_sht = { | 78 | static struct scsi_host_template uli_sht = { |
79 | .module = THIS_MODULE, | 79 | ATA_BMDMA_SHT(DRV_NAME), |
80 | .name = DRV_NAME, | ||
81 | .ioctl = ata_scsi_ioctl, | ||
82 | .queuecommand = ata_scsi_queuecmd, | ||
83 | .can_queue = ATA_DEF_QUEUE, | ||
84 | .this_id = ATA_SHT_THIS_ID, | ||
85 | .sg_tablesize = LIBATA_MAX_PRD, | ||
86 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
87 | .emulated = ATA_SHT_EMULATED, | ||
88 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
89 | .proc_name = DRV_NAME, | ||
90 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
91 | .slave_configure = ata_scsi_slave_config, | ||
92 | .slave_destroy = ata_scsi_slave_destroy, | ||
93 | .bios_param = ata_std_bios_param, | ||
94 | }; | 80 | }; |
95 | 81 | ||
96 | static const struct ata_port_operations uli_ops = { | 82 | static struct ata_port_operations uli_ops = { |
97 | .tf_load = ata_tf_load, | 83 | .inherits = &ata_bmdma_port_ops, |
98 | .tf_read = ata_tf_read, | ||
99 | .check_status = ata_check_status, | ||
100 | .exec_command = ata_exec_command, | ||
101 | .dev_select = ata_std_dev_select, | ||
102 | |||
103 | .bmdma_setup = ata_bmdma_setup, | ||
104 | .bmdma_start = ata_bmdma_start, | ||
105 | .bmdma_stop = ata_bmdma_stop, | ||
106 | .bmdma_status = ata_bmdma_status, | ||
107 | .qc_prep = ata_qc_prep, | ||
108 | .qc_issue = ata_qc_issue_prot, | ||
109 | .data_xfer = ata_data_xfer, | ||
110 | |||
111 | .freeze = ata_bmdma_freeze, | ||
112 | .thaw = ata_bmdma_thaw, | ||
113 | .error_handler = ata_bmdma_error_handler, | ||
114 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
115 | |||
116 | .irq_clear = ata_bmdma_irq_clear, | ||
117 | .irq_on = ata_irq_on, | ||
118 | |||
119 | .scr_read = uli_scr_read, | 84 | .scr_read = uli_scr_read, |
120 | .scr_write = uli_scr_write, | 85 | .scr_write = uli_scr_write, |
121 | |||
122 | .port_start = ata_port_start, | ||
123 | }; | 86 | }; |
124 | 87 | ||
125 | static const struct ata_port_info uli_port_info = { | 88 | static const struct ata_port_info uli_port_info = { |
@@ -212,11 +175,11 @@ static int uli_init_one(struct pci_dev *pdev, const struct pci_device_id *ent) | |||
212 | host->private_data = hpriv; | 175 | host->private_data = hpriv; |
213 | 176 | ||
214 | /* the first two ports are standard SFF */ | 177 | /* the first two ports are standard SFF */ |
215 | rc = ata_pci_init_sff_host(host); | 178 | rc = ata_pci_sff_init_host(host); |
216 | if (rc) | 179 | if (rc) |
217 | return rc; | 180 | return rc; |
218 | 181 | ||
219 | rc = ata_pci_init_bmdma(host); | 182 | rc = ata_pci_bmdma_init(host); |
220 | if (rc) | 183 | if (rc) |
221 | return rc; | 184 | return rc; |
222 | 185 | ||
@@ -237,7 +200,7 @@ static int uli_init_one(struct pci_dev *pdev, const struct pci_device_id *ent) | |||
237 | ((unsigned long)iomap[1] | ATA_PCI_CTL_OFS) + 4; | 200 | ((unsigned long)iomap[1] | ATA_PCI_CTL_OFS) + 4; |
238 | ioaddr->bmdma_addr = iomap[4] + 16; | 201 | ioaddr->bmdma_addr = iomap[4] + 16; |
239 | hpriv->scr_cfg_addr[2] = ULI5287_BASE + ULI5287_OFFS*4; | 202 | hpriv->scr_cfg_addr[2] = ULI5287_BASE + ULI5287_OFFS*4; |
240 | ata_std_ports(ioaddr); | 203 | ata_sff_std_ports(ioaddr); |
241 | 204 | ||
242 | ata_port_desc(host->ports[2], | 205 | ata_port_desc(host->ports[2], |
243 | "cmd 0x%llx ctl 0x%llx bmdma 0x%llx", | 206 | "cmd 0x%llx ctl 0x%llx bmdma 0x%llx", |
@@ -252,7 +215,7 @@ static int uli_init_one(struct pci_dev *pdev, const struct pci_device_id *ent) | |||
252 | ((unsigned long)iomap[3] | ATA_PCI_CTL_OFS) + 4; | 215 | ((unsigned long)iomap[3] | ATA_PCI_CTL_OFS) + 4; |
253 | ioaddr->bmdma_addr = iomap[4] + 24; | 216 | ioaddr->bmdma_addr = iomap[4] + 24; |
254 | hpriv->scr_cfg_addr[3] = ULI5287_BASE + ULI5287_OFFS*5; | 217 | hpriv->scr_cfg_addr[3] = ULI5287_BASE + ULI5287_OFFS*5; |
255 | ata_std_ports(ioaddr); | 218 | ata_sff_std_ports(ioaddr); |
256 | 219 | ||
257 | ata_port_desc(host->ports[2], | 220 | ata_port_desc(host->ports[2], |
258 | "cmd 0x%llx ctl 0x%llx bmdma 0x%llx", | 221 | "cmd 0x%llx ctl 0x%llx bmdma 0x%llx", |
@@ -279,8 +242,8 @@ static int uli_init_one(struct pci_dev *pdev, const struct pci_device_id *ent) | |||
279 | 242 | ||
280 | pci_set_master(pdev); | 243 | pci_set_master(pdev); |
281 | pci_intx(pdev, 1); | 244 | pci_intx(pdev, 1); |
282 | return ata_host_activate(host, pdev->irq, ata_interrupt, IRQF_SHARED, | 245 | return ata_host_activate(host, pdev->irq, ata_sff_interrupt, |
283 | &uli_sht); | 246 | IRQF_SHARED, &uli_sht); |
284 | } | 247 | } |
285 | 248 | ||
286 | static int __init uli_init(void) | 249 | static int __init uli_init(void) |
diff --git a/drivers/ata/sata_via.c b/drivers/ata/sata_via.c index 0d03f44824fb..96deeb354e16 100644 --- a/drivers/ata/sata_via.c +++ b/drivers/ata/sata_via.c | |||
@@ -71,7 +71,7 @@ static int svia_init_one(struct pci_dev *pdev, const struct pci_device_id *ent); | |||
71 | static int svia_scr_read(struct ata_port *ap, unsigned int sc_reg, u32 *val); | 71 | static int svia_scr_read(struct ata_port *ap, unsigned int sc_reg, u32 *val); |
72 | static int svia_scr_write(struct ata_port *ap, unsigned int sc_reg, u32 val); | 72 | static int svia_scr_write(struct ata_port *ap, unsigned int sc_reg, u32 val); |
73 | static void svia_noop_freeze(struct ata_port *ap); | 73 | static void svia_noop_freeze(struct ata_port *ap); |
74 | static void vt6420_error_handler(struct ata_port *ap); | 74 | static int vt6420_prereset(struct ata_link *link, unsigned long deadline); |
75 | static int vt6421_pata_cable_detect(struct ata_port *ap); | 75 | static int vt6421_pata_cable_detect(struct ata_port *ap); |
76 | static void vt6421_set_pio_mode(struct ata_port *ap, struct ata_device *adev); | 76 | static void vt6421_set_pio_mode(struct ata_port *ap, struct ata_device *adev); |
77 | static void vt6421_set_dma_mode(struct ata_port *ap, struct ata_device *adev); | 77 | static void vt6421_set_dma_mode(struct ata_port *ap, struct ata_device *adev); |
@@ -100,110 +100,26 @@ static struct pci_driver svia_pci_driver = { | |||
100 | }; | 100 | }; |
101 | 101 | ||
102 | static struct scsi_host_template svia_sht = { | 102 | static struct scsi_host_template svia_sht = { |
103 | .module = THIS_MODULE, | 103 | ATA_BMDMA_SHT(DRV_NAME), |
104 | .name = DRV_NAME, | ||
105 | .ioctl = ata_scsi_ioctl, | ||
106 | .queuecommand = ata_scsi_queuecmd, | ||
107 | .can_queue = ATA_DEF_QUEUE, | ||
108 | .this_id = ATA_SHT_THIS_ID, | ||
109 | .sg_tablesize = LIBATA_MAX_PRD, | ||
110 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
111 | .emulated = ATA_SHT_EMULATED, | ||
112 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
113 | .proc_name = DRV_NAME, | ||
114 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
115 | .slave_configure = ata_scsi_slave_config, | ||
116 | .slave_destroy = ata_scsi_slave_destroy, | ||
117 | .bios_param = ata_std_bios_param, | ||
118 | }; | 104 | }; |
119 | 105 | ||
120 | static const struct ata_port_operations vt6420_sata_ops = { | 106 | static struct ata_port_operations vt6420_sata_ops = { |
121 | .tf_load = ata_tf_load, | 107 | .inherits = &ata_bmdma_port_ops, |
122 | .tf_read = ata_tf_read, | ||
123 | .check_status = ata_check_status, | ||
124 | .exec_command = ata_exec_command, | ||
125 | .dev_select = ata_std_dev_select, | ||
126 | |||
127 | .bmdma_setup = ata_bmdma_setup, | ||
128 | .bmdma_start = ata_bmdma_start, | ||
129 | .bmdma_stop = ata_bmdma_stop, | ||
130 | .bmdma_status = ata_bmdma_status, | ||
131 | |||
132 | .qc_prep = ata_qc_prep, | ||
133 | .qc_issue = ata_qc_issue_prot, | ||
134 | .data_xfer = ata_data_xfer, | ||
135 | |||
136 | .freeze = svia_noop_freeze, | 108 | .freeze = svia_noop_freeze, |
137 | .thaw = ata_bmdma_thaw, | 109 | .prereset = vt6420_prereset, |
138 | .error_handler = vt6420_error_handler, | ||
139 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
140 | |||
141 | .irq_clear = ata_bmdma_irq_clear, | ||
142 | .irq_on = ata_irq_on, | ||
143 | |||
144 | .port_start = ata_port_start, | ||
145 | }; | 110 | }; |
146 | 111 | ||
147 | static const struct ata_port_operations vt6421_pata_ops = { | 112 | static struct ata_port_operations vt6421_pata_ops = { |
113 | .inherits = &ata_bmdma_port_ops, | ||
114 | .cable_detect = vt6421_pata_cable_detect, | ||
148 | .set_piomode = vt6421_set_pio_mode, | 115 | .set_piomode = vt6421_set_pio_mode, |
149 | .set_dmamode = vt6421_set_dma_mode, | 116 | .set_dmamode = vt6421_set_dma_mode, |
150 | |||
151 | .tf_load = ata_tf_load, | ||
152 | .tf_read = ata_tf_read, | ||
153 | .check_status = ata_check_status, | ||
154 | .exec_command = ata_exec_command, | ||
155 | .dev_select = ata_std_dev_select, | ||
156 | |||
157 | .bmdma_setup = ata_bmdma_setup, | ||
158 | .bmdma_start = ata_bmdma_start, | ||
159 | .bmdma_stop = ata_bmdma_stop, | ||
160 | .bmdma_status = ata_bmdma_status, | ||
161 | |||
162 | .qc_prep = ata_qc_prep, | ||
163 | .qc_issue = ata_qc_issue_prot, | ||
164 | .data_xfer = ata_data_xfer, | ||
165 | |||
166 | .freeze = ata_bmdma_freeze, | ||
167 | .thaw = ata_bmdma_thaw, | ||
168 | .error_handler = ata_bmdma_error_handler, | ||
169 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
170 | .cable_detect = vt6421_pata_cable_detect, | ||
171 | |||
172 | .irq_clear = ata_bmdma_irq_clear, | ||
173 | .irq_on = ata_irq_on, | ||
174 | |||
175 | .port_start = ata_port_start, | ||
176 | }; | 117 | }; |
177 | 118 | ||
178 | static const struct ata_port_operations vt6421_sata_ops = { | 119 | static struct ata_port_operations vt6421_sata_ops = { |
179 | .tf_load = ata_tf_load, | 120 | .inherits = &ata_bmdma_port_ops, |
180 | .tf_read = ata_tf_read, | ||
181 | .check_status = ata_check_status, | ||
182 | .exec_command = ata_exec_command, | ||
183 | .dev_select = ata_std_dev_select, | ||
184 | |||
185 | .bmdma_setup = ata_bmdma_setup, | ||
186 | .bmdma_start = ata_bmdma_start, | ||
187 | .bmdma_stop = ata_bmdma_stop, | ||
188 | .bmdma_status = ata_bmdma_status, | ||
189 | |||
190 | .qc_prep = ata_qc_prep, | ||
191 | .qc_issue = ata_qc_issue_prot, | ||
192 | .data_xfer = ata_data_xfer, | ||
193 | |||
194 | .freeze = ata_bmdma_freeze, | ||
195 | .thaw = ata_bmdma_thaw, | ||
196 | .error_handler = ata_bmdma_error_handler, | ||
197 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
198 | .cable_detect = ata_cable_sata, | ||
199 | |||
200 | .irq_clear = ata_bmdma_irq_clear, | ||
201 | .irq_on = ata_irq_on, | ||
202 | |||
203 | .scr_read = svia_scr_read, | 121 | .scr_read = svia_scr_read, |
204 | .scr_write = svia_scr_write, | 122 | .scr_write = svia_scr_write, |
205 | |||
206 | .port_start = ata_port_start, | ||
207 | }; | 123 | }; |
208 | 124 | ||
209 | static const struct ata_port_info vt6420_port_info = { | 125 | static const struct ata_port_info vt6420_port_info = { |
@@ -257,8 +173,8 @@ static void svia_noop_freeze(struct ata_port *ap) | |||
257 | /* Some VIA controllers choke if ATA_NIEN is manipulated in | 173 | /* Some VIA controllers choke if ATA_NIEN is manipulated in |
258 | * certain way. Leave it alone and just clear pending IRQ. | 174 | * certain way. Leave it alone and just clear pending IRQ. |
259 | */ | 175 | */ |
260 | ata_chk_status(ap); | 176 | ap->ops->sff_check_status(ap); |
261 | ata_bmdma_irq_clear(ap); | 177 | ata_sff_irq_clear(ap); |
262 | } | 178 | } |
263 | 179 | ||
264 | /** | 180 | /** |
@@ -320,23 +236,17 @@ static int vt6420_prereset(struct ata_link *link, unsigned long deadline) | |||
320 | 236 | ||
321 | if (!online) { | 237 | if (!online) { |
322 | /* tell EH to bail */ | 238 | /* tell EH to bail */ |
323 | ehc->i.action &= ~ATA_EH_RESET_MASK; | 239 | ehc->i.action &= ~ATA_EH_RESET; |
324 | return 0; | 240 | return 0; |
325 | } | 241 | } |
326 | 242 | ||
327 | skip_scr: | 243 | skip_scr: |
328 | /* wait for !BSY */ | 244 | /* wait for !BSY */ |
329 | ata_wait_ready(ap, deadline); | 245 | ata_sff_wait_ready(link, deadline); |
330 | 246 | ||
331 | return 0; | 247 | return 0; |
332 | } | 248 | } |
333 | 249 | ||
334 | static void vt6420_error_handler(struct ata_port *ap) | ||
335 | { | ||
336 | ata_bmdma_drive_eh(ap, vt6420_prereset, ata_std_softreset, NULL, | ||
337 | ata_std_postreset); | ||
338 | } | ||
339 | |||
340 | static int vt6421_pata_cable_detect(struct ata_port *ap) | 250 | static int vt6421_pata_cable_detect(struct ata_port *ap) |
341 | { | 251 | { |
342 | struct pci_dev *pdev = to_pci_dev(ap->host->dev); | 252 | struct pci_dev *pdev = to_pci_dev(ap->host->dev); |
@@ -394,7 +304,7 @@ static void vt6421_init_addrs(struct ata_port *ap) | |||
394 | ioaddr->bmdma_addr = bmdma_addr; | 304 | ioaddr->bmdma_addr = bmdma_addr; |
395 | ioaddr->scr_addr = vt6421_scr_addr(iomap[5], ap->port_no); | 305 | ioaddr->scr_addr = vt6421_scr_addr(iomap[5], ap->port_no); |
396 | 306 | ||
397 | ata_std_ports(ioaddr); | 307 | ata_sff_std_ports(ioaddr); |
398 | 308 | ||
399 | ata_port_pbar_desc(ap, ap->port_no, -1, "port"); | 309 | ata_port_pbar_desc(ap, ap->port_no, -1, "port"); |
400 | ata_port_pbar_desc(ap, 4, ap->port_no * 8, "bmdma"); | 310 | ata_port_pbar_desc(ap, 4, ap->port_no * 8, "bmdma"); |
@@ -406,7 +316,7 @@ static int vt6420_prepare_host(struct pci_dev *pdev, struct ata_host **r_host) | |||
406 | struct ata_host *host; | 316 | struct ata_host *host; |
407 | int rc; | 317 | int rc; |
408 | 318 | ||
409 | rc = ata_pci_prepare_sff_host(pdev, ppi, &host); | 319 | rc = ata_pci_sff_prepare_host(pdev, ppi, &host); |
410 | if (rc) | 320 | if (rc) |
411 | return rc; | 321 | return rc; |
412 | *r_host = host; | 322 | *r_host = host; |
@@ -538,8 +448,8 @@ static int svia_init_one(struct pci_dev *pdev, const struct pci_device_id *ent) | |||
538 | svia_configure(pdev); | 448 | svia_configure(pdev); |
539 | 449 | ||
540 | pci_set_master(pdev); | 450 | pci_set_master(pdev); |
541 | return ata_host_activate(host, pdev->irq, ata_interrupt, IRQF_SHARED, | 451 | return ata_host_activate(host, pdev->irq, ata_sff_interrupt, |
542 | &svia_sht); | 452 | IRQF_SHARED, &svia_sht); |
543 | } | 453 | } |
544 | 454 | ||
545 | static int __init svia_init(void) | 455 | static int __init svia_init(void) |
diff --git a/drivers/ata/sata_vsc.c b/drivers/ata/sata_vsc.c index 95ae3ed24a9d..f3d635c0a2e9 100644 --- a/drivers/ata/sata_vsc.c +++ b/drivers/ata/sata_vsc.c | |||
@@ -200,7 +200,7 @@ static void vsc_sata_tf_read(struct ata_port *ap, struct ata_taskfile *tf) | |||
200 | struct ata_ioports *ioaddr = &ap->ioaddr; | 200 | struct ata_ioports *ioaddr = &ap->ioaddr; |
201 | u16 nsect, lbal, lbam, lbah, feature; | 201 | u16 nsect, lbal, lbam, lbah, feature; |
202 | 202 | ||
203 | tf->command = ata_check_status(ap); | 203 | tf->command = ata_sff_check_status(ap); |
204 | tf->device = readw(ioaddr->device_addr); | 204 | tf->device = readw(ioaddr->device_addr); |
205 | feature = readw(ioaddr->error_addr); | 205 | feature = readw(ioaddr->error_addr); |
206 | nsect = readw(ioaddr->nsect_addr); | 206 | nsect = readw(ioaddr->nsect_addr); |
@@ -243,7 +243,7 @@ static void vsc_port_intr(u8 port_status, struct ata_port *ap) | |||
243 | 243 | ||
244 | qc = ata_qc_from_tag(ap, ap->link.active_tag); | 244 | qc = ata_qc_from_tag(ap, ap->link.active_tag); |
245 | if (qc && likely(!(qc->tf.flags & ATA_TFLAG_POLLING))) | 245 | if (qc && likely(!(qc->tf.flags & ATA_TFLAG_POLLING))) |
246 | handled = ata_host_intr(ap, qc); | 246 | handled = ata_sff_host_intr(ap, qc); |
247 | 247 | ||
248 | /* We received an interrupt during a polled command, | 248 | /* We received an interrupt during a polled command, |
249 | * or some other spurious condition. Interrupt reporting | 249 | * or some other spurious condition. Interrupt reporting |
@@ -251,7 +251,7 @@ static void vsc_port_intr(u8 port_status, struct ata_port *ap) | |||
251 | * simply clear the interrupt | 251 | * simply clear the interrupt |
252 | */ | 252 | */ |
253 | if (unlikely(!handled)) | 253 | if (unlikely(!handled)) |
254 | ata_chk_status(ap); | 254 | ap->ops->sff_check_status(ap); |
255 | } | 255 | } |
256 | 256 | ||
257 | /* | 257 | /* |
@@ -300,46 +300,18 @@ out: | |||
300 | 300 | ||
301 | 301 | ||
302 | static struct scsi_host_template vsc_sata_sht = { | 302 | static struct scsi_host_template vsc_sata_sht = { |
303 | .module = THIS_MODULE, | 303 | ATA_BMDMA_SHT(DRV_NAME), |
304 | .name = DRV_NAME, | ||
305 | .ioctl = ata_scsi_ioctl, | ||
306 | .queuecommand = ata_scsi_queuecmd, | ||
307 | .can_queue = ATA_DEF_QUEUE, | ||
308 | .this_id = ATA_SHT_THIS_ID, | ||
309 | .sg_tablesize = LIBATA_MAX_PRD, | ||
310 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, | ||
311 | .emulated = ATA_SHT_EMULATED, | ||
312 | .use_clustering = ATA_SHT_USE_CLUSTERING, | ||
313 | .proc_name = DRV_NAME, | ||
314 | .dma_boundary = ATA_DMA_BOUNDARY, | ||
315 | .slave_configure = ata_scsi_slave_config, | ||
316 | .slave_destroy = ata_scsi_slave_destroy, | ||
317 | .bios_param = ata_std_bios_param, | ||
318 | }; | 304 | }; |
319 | 305 | ||
320 | 306 | ||
321 | static const struct ata_port_operations vsc_sata_ops = { | 307 | static struct ata_port_operations vsc_sata_ops = { |
322 | .tf_load = vsc_sata_tf_load, | 308 | .inherits = &ata_bmdma_port_ops, |
323 | .tf_read = vsc_sata_tf_read, | 309 | .sff_tf_load = vsc_sata_tf_load, |
324 | .exec_command = ata_exec_command, | 310 | .sff_tf_read = vsc_sata_tf_read, |
325 | .check_status = ata_check_status, | ||
326 | .dev_select = ata_std_dev_select, | ||
327 | .bmdma_setup = ata_bmdma_setup, | ||
328 | .bmdma_start = ata_bmdma_start, | ||
329 | .bmdma_stop = ata_bmdma_stop, | ||
330 | .bmdma_status = ata_bmdma_status, | ||
331 | .qc_prep = ata_qc_prep, | ||
332 | .qc_issue = ata_qc_issue_prot, | ||
333 | .data_xfer = ata_data_xfer, | ||
334 | .freeze = vsc_freeze, | 311 | .freeze = vsc_freeze, |
335 | .thaw = vsc_thaw, | 312 | .thaw = vsc_thaw, |
336 | .error_handler = ata_bmdma_error_handler, | ||
337 | .post_internal_cmd = ata_bmdma_post_internal_cmd, | ||
338 | .irq_clear = ata_bmdma_irq_clear, | ||
339 | .irq_on = ata_irq_on, | ||
340 | .scr_read = vsc_sata_scr_read, | 313 | .scr_read = vsc_sata_scr_read, |
341 | .scr_write = vsc_sata_scr_write, | 314 | .scr_write = vsc_sata_scr_write, |
342 | .port_start = ata_port_start, | ||
343 | }; | 315 | }; |
344 | 316 | ||
345 | static void __devinit vsc_sata_setup_port(struct ata_ioports *port, | 317 | static void __devinit vsc_sata_setup_port(struct ata_ioports *port, |
diff --git a/drivers/scsi/ipr.c b/drivers/scsi/ipr.c index c72014a3e7d4..65dc18dea845 100644 --- a/drivers/scsi/ipr.c +++ b/drivers/scsi/ipr.c | |||
@@ -3937,7 +3937,7 @@ static int __ipr_eh_dev_reset(struct scsi_cmnd * scsi_cmd) | |||
3937 | if (ipr_is_gata(res) && res->sata_port) { | 3937 | if (ipr_is_gata(res) && res->sata_port) { |
3938 | ap = res->sata_port->ap; | 3938 | ap = res->sata_port->ap; |
3939 | spin_unlock_irq(scsi_cmd->device->host->host_lock); | 3939 | spin_unlock_irq(scsi_cmd->device->host->host_lock); |
3940 | ata_do_eh(ap, NULL, NULL, ipr_sata_reset, NULL); | 3940 | ata_std_error_handler(ap); |
3941 | spin_lock_irq(scsi_cmd->device->host->host_lock); | 3941 | spin_lock_irq(scsi_cmd->device->host->host_lock); |
3942 | 3942 | ||
3943 | list_for_each_entry(ipr_cmd, &ioa_cfg->pending_q, queue) { | 3943 | list_for_each_entry(ipr_cmd, &ioa_cfg->pending_q, queue) { |
@@ -5041,33 +5041,6 @@ static void ipr_ata_post_internal(struct ata_queued_cmd *qc) | |||
5041 | } | 5041 | } |
5042 | 5042 | ||
5043 | /** | 5043 | /** |
5044 | * ipr_tf_read - Read the current ATA taskfile for the ATA port | ||
5045 | * @ap: ATA port | ||
5046 | * @tf: destination ATA taskfile | ||
5047 | * | ||
5048 | * Return value: | ||
5049 | * none | ||
5050 | **/ | ||
5051 | static void ipr_tf_read(struct ata_port *ap, struct ata_taskfile *tf) | ||
5052 | { | ||
5053 | struct ipr_sata_port *sata_port = ap->private_data; | ||
5054 | struct ipr_ioasa_gata *g = &sata_port->ioasa; | ||
5055 | |||
5056 | tf->feature = g->error; | ||
5057 | tf->nsect = g->nsect; | ||
5058 | tf->lbal = g->lbal; | ||
5059 | tf->lbam = g->lbam; | ||
5060 | tf->lbah = g->lbah; | ||
5061 | tf->device = g->device; | ||
5062 | tf->command = g->status; | ||
5063 | tf->hob_nsect = g->hob_nsect; | ||
5064 | tf->hob_lbal = g->hob_lbal; | ||
5065 | tf->hob_lbam = g->hob_lbam; | ||
5066 | tf->hob_lbah = g->hob_lbah; | ||
5067 | tf->ctl = g->alt_status; | ||
5068 | } | ||
5069 | |||
5070 | /** | ||
5071 | * ipr_copy_sata_tf - Copy a SATA taskfile to an IOA data structure | 5044 | * ipr_copy_sata_tf - Copy a SATA taskfile to an IOA data structure |
5072 | * @regs: destination | 5045 | * @regs: destination |
5073 | * @tf: source ATA taskfile | 5046 | * @tf: source ATA taskfile |
@@ -5245,40 +5218,41 @@ static unsigned int ipr_qc_issue(struct ata_queued_cmd *qc) | |||
5245 | } | 5218 | } |
5246 | 5219 | ||
5247 | /** | 5220 | /** |
5248 | * ipr_ata_check_status - Return last ATA status | 5221 | * ipr_qc_fill_rtf - Read result TF |
5249 | * @ap: ATA port | 5222 | * @qc: ATA queued command |
5250 | * | 5223 | * |
5251 | * Return value: | 5224 | * Return value: |
5252 | * ATA status | 5225 | * true |
5253 | **/ | 5226 | **/ |
5254 | static u8 ipr_ata_check_status(struct ata_port *ap) | 5227 | static bool ipr_qc_fill_rtf(struct ata_queued_cmd *qc) |
5255 | { | 5228 | { |
5256 | struct ipr_sata_port *sata_port = ap->private_data; | 5229 | struct ipr_sata_port *sata_port = qc->ap->private_data; |
5257 | return sata_port->ioasa.status; | 5230 | struct ipr_ioasa_gata *g = &sata_port->ioasa; |
5258 | } | 5231 | struct ata_taskfile *tf = &qc->result_tf; |
5259 | 5232 | ||
5260 | /** | 5233 | tf->feature = g->error; |
5261 | * ipr_ata_check_altstatus - Return last ATA altstatus | 5234 | tf->nsect = g->nsect; |
5262 | * @ap: ATA port | 5235 | tf->lbal = g->lbal; |
5263 | * | 5236 | tf->lbam = g->lbam; |
5264 | * Return value: | 5237 | tf->lbah = g->lbah; |
5265 | * Alt ATA status | 5238 | tf->device = g->device; |
5266 | **/ | 5239 | tf->command = g->status; |
5267 | static u8 ipr_ata_check_altstatus(struct ata_port *ap) | 5240 | tf->hob_nsect = g->hob_nsect; |
5268 | { | 5241 | tf->hob_lbal = g->hob_lbal; |
5269 | struct ipr_sata_port *sata_port = ap->private_data; | 5242 | tf->hob_lbam = g->hob_lbam; |
5270 | return sata_port->ioasa.alt_status; | 5243 | tf->hob_lbah = g->hob_lbah; |
5244 | tf->ctl = g->alt_status; | ||
5245 | |||
5246 | return true; | ||
5271 | } | 5247 | } |
5272 | 5248 | ||
5273 | static struct ata_port_operations ipr_sata_ops = { | 5249 | static struct ata_port_operations ipr_sata_ops = { |
5274 | .check_status = ipr_ata_check_status, | ||
5275 | .check_altstatus = ipr_ata_check_altstatus, | ||
5276 | .dev_select = ata_noop_dev_select, | ||
5277 | .phy_reset = ipr_ata_phy_reset, | 5250 | .phy_reset = ipr_ata_phy_reset, |
5251 | .hardreset = ipr_sata_reset, | ||
5278 | .post_internal_cmd = ipr_ata_post_internal, | 5252 | .post_internal_cmd = ipr_ata_post_internal, |
5279 | .tf_read = ipr_tf_read, | ||
5280 | .qc_prep = ata_noop_qc_prep, | 5253 | .qc_prep = ata_noop_qc_prep, |
5281 | .qc_issue = ipr_qc_issue, | 5254 | .qc_issue = ipr_qc_issue, |
5255 | .qc_fill_rtf = ipr_qc_fill_rtf, | ||
5282 | .port_start = ata_sas_port_start, | 5256 | .port_start = ata_sas_port_start, |
5283 | .port_stop = ata_sas_port_stop | 5257 | .port_stop = ata_sas_port_stop |
5284 | }; | 5258 | }; |
diff --git a/drivers/scsi/libsas/sas_ata.c b/drivers/scsi/libsas/sas_ata.c index b0e5ac372a32..a4811e4106df 100644 --- a/drivers/scsi/libsas/sas_ata.c +++ b/drivers/scsi/libsas/sas_ata.c | |||
@@ -225,10 +225,12 @@ static unsigned int sas_ata_qc_issue(struct ata_queued_cmd *qc) | |||
225 | return 0; | 225 | return 0; |
226 | } | 226 | } |
227 | 227 | ||
228 | static u8 sas_ata_check_status(struct ata_port *ap) | 228 | static bool sas_ata_qc_fill_rtf(struct ata_queued_cmd *qc) |
229 | { | 229 | { |
230 | struct domain_device *dev = ap->private_data; | 230 | struct domain_device *dev = qc->ap->private_data; |
231 | return dev->sata_dev.tf.command; | 231 | |
232 | memcpy(&qc->result_tf, &dev->sata_dev.tf, sizeof(qc->result_tf)); | ||
233 | return true; | ||
232 | } | 234 | } |
233 | 235 | ||
234 | static void sas_ata_phy_reset(struct ata_port *ap) | 236 | static void sas_ata_phy_reset(struct ata_port *ap) |
@@ -292,12 +294,6 @@ static void sas_ata_post_internal(struct ata_queued_cmd *qc) | |||
292 | } | 294 | } |
293 | } | 295 | } |
294 | 296 | ||
295 | static void sas_ata_tf_read(struct ata_port *ap, struct ata_taskfile *tf) | ||
296 | { | ||
297 | struct domain_device *dev = ap->private_data; | ||
298 | memcpy(tf, &dev->sata_dev.tf, sizeof (*tf)); | ||
299 | } | ||
300 | |||
301 | static int sas_ata_scr_write(struct ata_port *ap, unsigned int sc_reg_in, | 297 | static int sas_ata_scr_write(struct ata_port *ap, unsigned int sc_reg_in, |
302 | u32 val) | 298 | u32 val) |
303 | { | 299 | { |
@@ -348,14 +344,11 @@ static int sas_ata_scr_read(struct ata_port *ap, unsigned int sc_reg_in, | |||
348 | } | 344 | } |
349 | 345 | ||
350 | static struct ata_port_operations sas_sata_ops = { | 346 | static struct ata_port_operations sas_sata_ops = { |
351 | .check_status = sas_ata_check_status, | ||
352 | .check_altstatus = sas_ata_check_status, | ||
353 | .dev_select = ata_noop_dev_select, | ||
354 | .phy_reset = sas_ata_phy_reset, | 347 | .phy_reset = sas_ata_phy_reset, |
355 | .post_internal_cmd = sas_ata_post_internal, | 348 | .post_internal_cmd = sas_ata_post_internal, |
356 | .tf_read = sas_ata_tf_read, | ||
357 | .qc_prep = ata_noop_qc_prep, | 349 | .qc_prep = ata_noop_qc_prep, |
358 | .qc_issue = sas_ata_qc_issue, | 350 | .qc_issue = sas_ata_qc_issue, |
351 | .qc_fill_rtf = sas_ata_qc_fill_rtf, | ||
359 | .port_start = ata_sas_port_start, | 352 | .port_start = ata_sas_port_start, |
360 | .port_stop = ata_sas_port_stop, | 353 | .port_stop = ata_sas_port_stop, |
361 | .scr_read = sas_ata_scr_read, | 354 | .scr_read = sas_ata_scr_read, |
diff --git a/include/linux/libata.h b/include/linux/libata.h index 37ee881c42ac..165734a2dd47 100644 --- a/include/linux/libata.h +++ b/include/linux/libata.h | |||
@@ -122,6 +122,8 @@ enum { | |||
122 | 122 | ||
123 | ATAPI_MAX_DRAIN = 16 << 10, | 123 | ATAPI_MAX_DRAIN = 16 << 10, |
124 | 124 | ||
125 | ATA_ALL_DEVICES = (1 << ATA_MAX_DEVICES) - 1, | ||
126 | |||
125 | ATA_SHT_EMULATED = 1, | 127 | ATA_SHT_EMULATED = 1, |
126 | ATA_SHT_CMD_PER_LUN = 1, | 128 | ATA_SHT_CMD_PER_LUN = 1, |
127 | ATA_SHT_THIS_ID = -1, | 129 | ATA_SHT_THIS_ID = -1, |
@@ -163,9 +165,6 @@ enum { | |||
163 | ATA_DEV_NONE = 9, /* no device */ | 165 | ATA_DEV_NONE = 9, /* no device */ |
164 | 166 | ||
165 | /* struct ata_link flags */ | 167 | /* struct ata_link flags */ |
166 | ATA_LFLAG_HRST_TO_RESUME = (1 << 0), /* hardreset to resume link */ | ||
167 | ATA_LFLAG_SKIP_D2H_BSY = (1 << 1), /* can't wait for the first D2H | ||
168 | * Register FIS clearing BSY */ | ||
169 | ATA_LFLAG_NO_SRST = (1 << 2), /* avoid softreset */ | 168 | ATA_LFLAG_NO_SRST = (1 << 2), /* avoid softreset */ |
170 | ATA_LFLAG_ASSUME_ATA = (1 << 3), /* assume ATA class */ | 169 | ATA_LFLAG_ASSUME_ATA = (1 << 3), /* assume ATA class */ |
171 | ATA_LFLAG_ASSUME_SEMB = (1 << 4), /* assume SEMB class */ | 170 | ATA_LFLAG_ASSUME_SEMB = (1 << 4), /* assume SEMB class */ |
@@ -225,6 +224,7 @@ enum { | |||
225 | ATA_QCFLAG_RESULT_TF = (1 << 4), /* result TF requested */ | 224 | ATA_QCFLAG_RESULT_TF = (1 << 4), /* result TF requested */ |
226 | ATA_QCFLAG_CLEAR_EXCL = (1 << 5), /* clear excl_link on completion */ | 225 | ATA_QCFLAG_CLEAR_EXCL = (1 << 5), /* clear excl_link on completion */ |
227 | ATA_QCFLAG_QUIET = (1 << 6), /* don't report device error */ | 226 | ATA_QCFLAG_QUIET = (1 << 6), /* don't report device error */ |
227 | ATA_QCFLAG_RETRY = (1 << 7), /* retry after failure */ | ||
228 | 228 | ||
229 | ATA_QCFLAG_FAILED = (1 << 16), /* cmd failed and is owned by EH */ | 229 | ATA_QCFLAG_FAILED = (1 << 16), /* cmd failed and is owned by EH */ |
230 | ATA_QCFLAG_SENSE_VALID = (1 << 17), /* sense data valid */ | 230 | ATA_QCFLAG_SENSE_VALID = (1 << 17), /* sense data valid */ |
@@ -249,6 +249,25 @@ enum { | |||
249 | */ | 249 | */ |
250 | ATA_TMOUT_FF_WAIT = 4 * HZ / 5, | 250 | ATA_TMOUT_FF_WAIT = 4 * HZ / 5, |
251 | 251 | ||
252 | /* Spec mandates to wait for ">= 2ms" before checking status | ||
253 | * after reset. We wait 150ms, because that was the magic | ||
254 | * delay used for ATAPI devices in Hale Landis's ATADRVR, for | ||
255 | * the period of time between when the ATA command register is | ||
256 | * written, and then status is checked. Because waiting for | ||
257 | * "a while" before checking status is fine, post SRST, we | ||
258 | * perform this magic delay here as well. | ||
259 | * | ||
260 | * Old drivers/ide uses the 2mS rule and then waits for ready. | ||
261 | */ | ||
262 | ATA_WAIT_AFTER_RESET_MSECS = 150, | ||
263 | |||
264 | /* If PMP is supported, we have to do follow-up SRST. As some | ||
265 | * PMPs don't send D2H Reg FIS after hardreset, LLDs are | ||
266 | * advised to wait only for the following duration before | ||
267 | * doing SRST. | ||
268 | */ | ||
269 | ATA_TMOUT_PMP_SRST_WAIT = 1 * HZ, | ||
270 | |||
252 | /* ATA bus states */ | 271 | /* ATA bus states */ |
253 | BUS_UNKNOWN = 0, | 272 | BUS_UNKNOWN = 0, |
254 | BUS_DMA = 1, | 273 | BUS_DMA = 1, |
@@ -292,17 +311,16 @@ enum { | |||
292 | 311 | ||
293 | /* reset / recovery action types */ | 312 | /* reset / recovery action types */ |
294 | ATA_EH_REVALIDATE = (1 << 0), | 313 | ATA_EH_REVALIDATE = (1 << 0), |
295 | ATA_EH_SOFTRESET = (1 << 1), | 314 | ATA_EH_SOFTRESET = (1 << 1), /* meaningful only in ->prereset */ |
296 | ATA_EH_HARDRESET = (1 << 2), | 315 | ATA_EH_HARDRESET = (1 << 2), /* meaningful only in ->prereset */ |
316 | ATA_EH_RESET = ATA_EH_SOFTRESET | ATA_EH_HARDRESET, | ||
297 | ATA_EH_ENABLE_LINK = (1 << 3), | 317 | ATA_EH_ENABLE_LINK = (1 << 3), |
298 | ATA_EH_LPM = (1 << 4), /* link power management action */ | 318 | ATA_EH_LPM = (1 << 4), /* link power management action */ |
299 | 319 | ||
300 | ATA_EH_RESET_MASK = ATA_EH_SOFTRESET | ATA_EH_HARDRESET, | ||
301 | ATA_EH_PERDEV_MASK = ATA_EH_REVALIDATE, | 320 | ATA_EH_PERDEV_MASK = ATA_EH_REVALIDATE, |
302 | 321 | ||
303 | /* ata_eh_info->flags */ | 322 | /* ata_eh_info->flags */ |
304 | ATA_EHI_HOTPLUGGED = (1 << 0), /* could have been hotplugged */ | 323 | ATA_EHI_HOTPLUGGED = (1 << 0), /* could have been hotplugged */ |
305 | ATA_EHI_RESUME_LINK = (1 << 1), /* resume link (reset modifier) */ | ||
306 | ATA_EHI_NO_AUTOPSY = (1 << 2), /* no autopsy */ | 324 | ATA_EHI_NO_AUTOPSY = (1 << 2), /* no autopsy */ |
307 | ATA_EHI_QUIET = (1 << 3), /* be quiet */ | 325 | ATA_EHI_QUIET = (1 << 3), /* be quiet */ |
308 | 326 | ||
@@ -313,7 +331,6 @@ enum { | |||
313 | ATA_EHI_POST_SETMODE = (1 << 20), /* revaildating after setmode */ | 331 | ATA_EHI_POST_SETMODE = (1 << 20), /* revaildating after setmode */ |
314 | 332 | ||
315 | ATA_EHI_DID_RESET = ATA_EHI_DID_SOFTRESET | ATA_EHI_DID_HARDRESET, | 333 | ATA_EHI_DID_RESET = ATA_EHI_DID_SOFTRESET | ATA_EHI_DID_HARDRESET, |
316 | ATA_EHI_RESET_MODIFIER_MASK = ATA_EHI_RESUME_LINK, | ||
317 | 334 | ||
318 | /* max tries if error condition is still set after ->error_handler */ | 335 | /* max tries if error condition is still set after ->error_handler */ |
319 | ATA_EH_MAX_TRIES = 5, | 336 | ATA_EH_MAX_TRIES = 5, |
@@ -352,6 +369,22 @@ enum { | |||
352 | ATAPI_READ_CD = 2, /* READ CD [MSF] */ | 369 | ATAPI_READ_CD = 2, /* READ CD [MSF] */ |
353 | ATAPI_PASS_THRU = 3, /* SAT pass-thru */ | 370 | ATAPI_PASS_THRU = 3, /* SAT pass-thru */ |
354 | ATAPI_MISC = 4, /* the rest */ | 371 | ATAPI_MISC = 4, /* the rest */ |
372 | |||
373 | /* Timing constants */ | ||
374 | ATA_TIMING_SETUP = (1 << 0), | ||
375 | ATA_TIMING_ACT8B = (1 << 1), | ||
376 | ATA_TIMING_REC8B = (1 << 2), | ||
377 | ATA_TIMING_CYC8B = (1 << 3), | ||
378 | ATA_TIMING_8BIT = ATA_TIMING_ACT8B | ATA_TIMING_REC8B | | ||
379 | ATA_TIMING_CYC8B, | ||
380 | ATA_TIMING_ACTIVE = (1 << 4), | ||
381 | ATA_TIMING_RECOVER = (1 << 5), | ||
382 | ATA_TIMING_CYCLE = (1 << 6), | ||
383 | ATA_TIMING_UDMA = (1 << 7), | ||
384 | ATA_TIMING_ALL = ATA_TIMING_SETUP | ATA_TIMING_ACT8B | | ||
385 | ATA_TIMING_REC8B | ATA_TIMING_CYC8B | | ||
386 | ATA_TIMING_ACTIVE | ATA_TIMING_RECOVER | | ||
387 | ATA_TIMING_CYCLE | ATA_TIMING_UDMA, | ||
355 | }; | 388 | }; |
356 | 389 | ||
357 | enum ata_xfer_mask { | 390 | enum ata_xfer_mask { |
@@ -412,6 +445,7 @@ enum link_pm { | |||
412 | }; | 445 | }; |
413 | extern struct class_device_attribute class_device_attr_link_power_management_policy; | 446 | extern struct class_device_attribute class_device_attr_link_power_management_policy; |
414 | 447 | ||
448 | #ifdef CONFIG_ATA_SFF | ||
415 | struct ata_ioports { | 449 | struct ata_ioports { |
416 | void __iomem *cmd_addr; | 450 | void __iomem *cmd_addr; |
417 | void __iomem *data_addr; | 451 | void __iomem *data_addr; |
@@ -429,6 +463,7 @@ struct ata_ioports { | |||
429 | void __iomem *bmdma_addr; | 463 | void __iomem *bmdma_addr; |
430 | void __iomem *scr_addr; | 464 | void __iomem *scr_addr; |
431 | }; | 465 | }; |
466 | #endif /* CONFIG_ATA_SFF */ | ||
432 | 467 | ||
433 | struct ata_host { | 468 | struct ata_host { |
434 | spinlock_t lock; | 469 | spinlock_t lock; |
@@ -436,7 +471,7 @@ struct ata_host { | |||
436 | void __iomem * const *iomap; | 471 | void __iomem * const *iomap; |
437 | unsigned int n_ports; | 472 | unsigned int n_ports; |
438 | void *private_data; | 473 | void *private_data; |
439 | const struct ata_port_operations *ops; | 474 | struct ata_port_operations *ops; |
440 | unsigned long flags; | 475 | unsigned long flags; |
441 | #ifdef CONFIG_ATA_ACPI | 476 | #ifdef CONFIG_ATA_ACPI |
442 | acpi_handle acpi_handle; | 477 | acpi_handle acpi_handle; |
@@ -605,7 +640,7 @@ struct ata_link { | |||
605 | 640 | ||
606 | struct ata_port { | 641 | struct ata_port { |
607 | struct Scsi_Host *scsi_host; /* our co-allocated scsi host */ | 642 | struct Scsi_Host *scsi_host; /* our co-allocated scsi host */ |
608 | const struct ata_port_operations *ops; | 643 | struct ata_port_operations *ops; |
609 | spinlock_t *lock; | 644 | spinlock_t *lock; |
610 | unsigned long flags; /* ATA_FLAG_xxx */ | 645 | unsigned long flags; /* ATA_FLAG_xxx */ |
611 | unsigned int pflags; /* ATA_PFLAG_xxx */ | 646 | unsigned int pflags; /* ATA_PFLAG_xxx */ |
@@ -615,7 +650,9 @@ struct ata_port { | |||
615 | struct ata_prd *prd; /* our SG list */ | 650 | struct ata_prd *prd; /* our SG list */ |
616 | dma_addr_t prd_dma; /* and its DMA mapping */ | 651 | dma_addr_t prd_dma; /* and its DMA mapping */ |
617 | 652 | ||
653 | #ifdef CONFIG_ATA_SFF | ||
618 | struct ata_ioports ioaddr; /* ATA cmd/ctl/dma register blocks */ | 654 | struct ata_ioports ioaddr; /* ATA cmd/ctl/dma register blocks */ |
655 | #endif /* CONFIG_ATA_SFF */ | ||
619 | 656 | ||
620 | u8 ctl; /* cache of ATA control register */ | 657 | u8 ctl; /* cache of ATA control register */ |
621 | u8 last_ctl; /* Cache last written value */ | 658 | u8 last_ctl; /* Cache last written value */ |
@@ -667,81 +704,108 @@ struct ata_port { | |||
667 | u8 sector_buf[ATA_SECT_SIZE]; /* owned by EH */ | 704 | u8 sector_buf[ATA_SECT_SIZE]; /* owned by EH */ |
668 | }; | 705 | }; |
669 | 706 | ||
670 | struct ata_port_operations { | 707 | /* The following initializer overrides a method to NULL whether one of |
671 | void (*dev_config) (struct ata_device *); | 708 | * its parent has the method defined or not. This is equivalent to |
672 | 709 | * ERR_PTR(-ENOENT). Unfortunately, ERR_PTR doesn't render a constant | |
673 | void (*set_piomode) (struct ata_port *, struct ata_device *); | 710 | * expression and thus can't be used as an initializer. |
674 | void (*set_dmamode) (struct ata_port *, struct ata_device *); | 711 | */ |
675 | unsigned long (*mode_filter) (struct ata_device *, unsigned long); | 712 | #define ATA_OP_NULL (void *)(unsigned long)(-ENOENT) |
676 | |||
677 | void (*tf_load) (struct ata_port *ap, const struct ata_taskfile *tf); | ||
678 | void (*tf_read) (struct ata_port *ap, struct ata_taskfile *tf); | ||
679 | |||
680 | void (*exec_command)(struct ata_port *ap, const struct ata_taskfile *tf); | ||
681 | u8 (*check_status)(struct ata_port *ap); | ||
682 | u8 (*check_altstatus)(struct ata_port *ap); | ||
683 | void (*dev_select)(struct ata_port *ap, unsigned int device); | ||
684 | |||
685 | void (*phy_reset) (struct ata_port *ap); /* obsolete */ | ||
686 | int (*set_mode) (struct ata_link *link, struct ata_device **r_failed_dev); | ||
687 | |||
688 | int (*cable_detect) (struct ata_port *ap); | ||
689 | |||
690 | int (*check_atapi_dma) (struct ata_queued_cmd *qc); | ||
691 | |||
692 | void (*bmdma_setup) (struct ata_queued_cmd *qc); | ||
693 | void (*bmdma_start) (struct ata_queued_cmd *qc); | ||
694 | |||
695 | unsigned int (*data_xfer) (struct ata_device *dev, unsigned char *buf, | ||
696 | unsigned int buflen, int rw); | ||
697 | |||
698 | int (*qc_defer) (struct ata_queued_cmd *qc); | ||
699 | void (*qc_prep) (struct ata_queued_cmd *qc); | ||
700 | unsigned int (*qc_issue) (struct ata_queued_cmd *qc); | ||
701 | |||
702 | /* port multiplier */ | ||
703 | void (*pmp_attach) (struct ata_port *ap); | ||
704 | void (*pmp_detach) (struct ata_port *ap); | ||
705 | 713 | ||
706 | /* Error handlers. ->error_handler overrides ->eng_timeout and | 714 | struct ata_port_operations { |
707 | * indicates that new-style EH is in place. | 715 | /* |
716 | * Command execution | ||
708 | */ | 717 | */ |
709 | void (*eng_timeout) (struct ata_port *ap); /* obsolete */ | 718 | int (*qc_defer)(struct ata_queued_cmd *qc); |
710 | 719 | int (*check_atapi_dma)(struct ata_queued_cmd *qc); | |
711 | void (*freeze) (struct ata_port *ap); | 720 | void (*qc_prep)(struct ata_queued_cmd *qc); |
712 | void (*thaw) (struct ata_port *ap); | 721 | unsigned int (*qc_issue)(struct ata_queued_cmd *qc); |
713 | void (*error_handler) (struct ata_port *ap); | 722 | bool (*qc_fill_rtf)(struct ata_queued_cmd *qc); |
714 | void (*post_internal_cmd) (struct ata_queued_cmd *qc); | 723 | |
715 | 724 | /* | |
716 | irq_handler_t irq_handler; | 725 | * Configuration and exception handling |
717 | void (*irq_clear) (struct ata_port *); | 726 | */ |
718 | u8 (*irq_on) (struct ata_port *); | 727 | int (*cable_detect)(struct ata_port *ap); |
719 | 728 | unsigned long (*mode_filter)(struct ata_device *dev, unsigned long xfer_mask); | |
720 | int (*scr_read) (struct ata_port *ap, unsigned int sc_reg, u32 *val); | 729 | void (*set_piomode)(struct ata_port *ap, struct ata_device *dev); |
721 | int (*scr_write) (struct ata_port *ap, unsigned int sc_reg, u32 val); | 730 | void (*set_dmamode)(struct ata_port *ap, struct ata_device *dev); |
731 | int (*set_mode)(struct ata_link *link, struct ata_device **r_failed_dev); | ||
732 | |||
733 | void (*dev_config)(struct ata_device *dev); | ||
734 | |||
735 | void (*freeze)(struct ata_port *ap); | ||
736 | void (*thaw)(struct ata_port *ap); | ||
737 | ata_prereset_fn_t prereset; | ||
738 | ata_reset_fn_t softreset; | ||
739 | ata_reset_fn_t hardreset; | ||
740 | ata_postreset_fn_t postreset; | ||
741 | ata_prereset_fn_t pmp_prereset; | ||
742 | ata_reset_fn_t pmp_softreset; | ||
743 | ata_reset_fn_t pmp_hardreset; | ||
744 | ata_postreset_fn_t pmp_postreset; | ||
745 | void (*error_handler)(struct ata_port *ap); | ||
746 | void (*post_internal_cmd)(struct ata_queued_cmd *qc); | ||
747 | |||
748 | /* | ||
749 | * Optional features | ||
750 | */ | ||
751 | int (*scr_read)(struct ata_port *ap, unsigned int sc_reg, u32 *val); | ||
752 | int (*scr_write)(struct ata_port *ap, unsigned int sc_reg, u32 val); | ||
753 | void (*pmp_attach)(struct ata_port *ap); | ||
754 | void (*pmp_detach)(struct ata_port *ap); | ||
755 | int (*enable_pm)(struct ata_port *ap, enum link_pm policy); | ||
756 | void (*disable_pm)(struct ata_port *ap); | ||
757 | |||
758 | /* | ||
759 | * Start, stop, suspend and resume | ||
760 | */ | ||
761 | int (*port_suspend)(struct ata_port *ap, pm_message_t mesg); | ||
762 | int (*port_resume)(struct ata_port *ap); | ||
763 | int (*port_start)(struct ata_port *ap); | ||
764 | void (*port_stop)(struct ata_port *ap); | ||
765 | void (*host_stop)(struct ata_host *host); | ||
766 | |||
767 | #ifdef CONFIG_ATA_SFF | ||
768 | /* | ||
769 | * SFF / taskfile oriented ops | ||
770 | */ | ||
771 | void (*sff_dev_select)(struct ata_port *ap, unsigned int device); | ||
772 | u8 (*sff_check_status)(struct ata_port *ap); | ||
773 | u8 (*sff_check_altstatus)(struct ata_port *ap); | ||
774 | void (*sff_tf_load)(struct ata_port *ap, const struct ata_taskfile *tf); | ||
775 | void (*sff_tf_read)(struct ata_port *ap, struct ata_taskfile *tf); | ||
776 | void (*sff_exec_command)(struct ata_port *ap, | ||
777 | const struct ata_taskfile *tf); | ||
778 | unsigned int (*sff_data_xfer)(struct ata_device *dev, | ||
779 | unsigned char *buf, unsigned int buflen, int rw); | ||
780 | u8 (*sff_irq_on)(struct ata_port *); | ||
781 | void (*sff_irq_clear)(struct ata_port *); | ||
722 | 782 | ||
723 | int (*port_suspend) (struct ata_port *ap, pm_message_t mesg); | 783 | void (*bmdma_setup)(struct ata_queued_cmd *qc); |
724 | int (*port_resume) (struct ata_port *ap); | 784 | void (*bmdma_start)(struct ata_queued_cmd *qc); |
725 | int (*enable_pm) (struct ata_port *ap, enum link_pm policy); | 785 | void (*bmdma_stop)(struct ata_queued_cmd *qc); |
726 | void (*disable_pm) (struct ata_port *ap); | 786 | u8 (*bmdma_status)(struct ata_port *ap); |
727 | int (*port_start) (struct ata_port *ap); | 787 | #endif /* CONFIG_ATA_SFF */ |
728 | void (*port_stop) (struct ata_port *ap); | ||
729 | 788 | ||
730 | void (*host_stop) (struct ata_host *host); | 789 | /* |
790 | * Obsolete | ||
791 | */ | ||
792 | void (*phy_reset)(struct ata_port *ap); | ||
793 | void (*eng_timeout)(struct ata_port *ap); | ||
731 | 794 | ||
732 | void (*bmdma_stop) (struct ata_queued_cmd *qc); | 795 | /* |
733 | u8 (*bmdma_status) (struct ata_port *ap); | 796 | * ->inherits must be the last field and all the preceding |
797 | * fields must be pointers. | ||
798 | */ | ||
799 | const struct ata_port_operations *inherits; | ||
734 | }; | 800 | }; |
735 | 801 | ||
736 | struct ata_port_info { | 802 | struct ata_port_info { |
737 | struct scsi_host_template *sht; | ||
738 | unsigned long flags; | 803 | unsigned long flags; |
739 | unsigned long link_flags; | 804 | unsigned long link_flags; |
740 | unsigned long pio_mask; | 805 | unsigned long pio_mask; |
741 | unsigned long mwdma_mask; | 806 | unsigned long mwdma_mask; |
742 | unsigned long udma_mask; | 807 | unsigned long udma_mask; |
743 | const struct ata_port_operations *port_ops; | 808 | struct ata_port_operations *port_ops; |
744 | irq_handler_t irq_handler; | ||
745 | void *private_data; | 809 | void *private_data; |
746 | }; | 810 | }; |
747 | 811 | ||
@@ -759,11 +823,14 @@ struct ata_timing { | |||
759 | 823 | ||
760 | #define FIT(v, vmin, vmax) max_t(short, min_t(short, v, vmax), vmin) | 824 | #define FIT(v, vmin, vmax) max_t(short, min_t(short, v, vmax), vmin) |
761 | 825 | ||
826 | /* | ||
827 | * Core layer - drivers/ata/libata-core.c | ||
828 | */ | ||
762 | extern const unsigned long sata_deb_timing_normal[]; | 829 | extern const unsigned long sata_deb_timing_normal[]; |
763 | extern const unsigned long sata_deb_timing_hotplug[]; | 830 | extern const unsigned long sata_deb_timing_hotplug[]; |
764 | extern const unsigned long sata_deb_timing_long[]; | 831 | extern const unsigned long sata_deb_timing_long[]; |
765 | 832 | ||
766 | extern const struct ata_port_operations ata_dummy_port_ops; | 833 | extern struct ata_port_operations ata_dummy_port_ops; |
767 | extern const struct ata_port_info ata_dummy_port_info; | 834 | extern const struct ata_port_info ata_dummy_port_info; |
768 | 835 | ||
769 | static inline const unsigned long * | 836 | static inline const unsigned long * |
@@ -782,22 +849,21 @@ static inline int ata_port_is_dummy(struct ata_port *ap) | |||
782 | 849 | ||
783 | extern void sata_print_link_status(struct ata_link *link); | 850 | extern void sata_print_link_status(struct ata_link *link); |
784 | extern void ata_port_probe(struct ata_port *); | 851 | extern void ata_port_probe(struct ata_port *); |
785 | extern void ata_bus_reset(struct ata_port *ap); | ||
786 | extern int sata_set_spd(struct ata_link *link); | 852 | extern int sata_set_spd(struct ata_link *link); |
853 | extern int ata_std_prereset(struct ata_link *link, unsigned long deadline); | ||
854 | extern int ata_wait_after_reset(struct ata_link *link, unsigned long deadline, | ||
855 | int (*check_ready)(struct ata_link *link)); | ||
787 | extern int sata_link_debounce(struct ata_link *link, | 856 | extern int sata_link_debounce(struct ata_link *link, |
788 | const unsigned long *params, unsigned long deadline); | 857 | const unsigned long *params, unsigned long deadline); |
789 | extern int sata_link_resume(struct ata_link *link, const unsigned long *params, | 858 | extern int sata_link_resume(struct ata_link *link, const unsigned long *params, |
790 | unsigned long deadline); | 859 | unsigned long deadline); |
791 | extern int ata_std_prereset(struct ata_link *link, unsigned long deadline); | ||
792 | extern int ata_std_softreset(struct ata_link *link, unsigned int *classes, | ||
793 | unsigned long deadline); | ||
794 | extern int sata_link_hardreset(struct ata_link *link, | 860 | extern int sata_link_hardreset(struct ata_link *link, |
795 | const unsigned long *timing, unsigned long deadline); | 861 | const unsigned long *timing, unsigned long deadline, |
862 | bool *online, int (*check_ready)(struct ata_link *)); | ||
796 | extern int sata_std_hardreset(struct ata_link *link, unsigned int *class, | 863 | extern int sata_std_hardreset(struct ata_link *link, unsigned int *class, |
797 | unsigned long deadline); | 864 | unsigned long deadline); |
798 | extern void ata_std_postreset(struct ata_link *link, unsigned int *classes); | 865 | extern void ata_std_postreset(struct ata_link *link, unsigned int *classes); |
799 | extern void ata_port_disable(struct ata_port *); | 866 | extern void ata_port_disable(struct ata_port *); |
800 | extern void ata_std_ports(struct ata_ioports *ioaddr); | ||
801 | 867 | ||
802 | extern struct ata_host *ata_host_alloc(struct device *dev, int max_ports); | 868 | extern struct ata_host *ata_host_alloc(struct device *dev, int max_ports); |
803 | extern struct ata_host *ata_host_alloc_pinfo(struct device *dev, | 869 | extern struct ata_host *ata_host_alloc_pinfo(struct device *dev, |
@@ -810,7 +876,7 @@ extern int ata_host_activate(struct ata_host *host, int irq, | |||
810 | struct scsi_host_template *sht); | 876 | struct scsi_host_template *sht); |
811 | extern void ata_host_detach(struct ata_host *host); | 877 | extern void ata_host_detach(struct ata_host *host); |
812 | extern void ata_host_init(struct ata_host *, struct device *, | 878 | extern void ata_host_init(struct ata_host *, struct device *, |
813 | unsigned long, const struct ata_port_operations *); | 879 | unsigned long, struct ata_port_operations *); |
814 | extern int ata_scsi_detect(struct scsi_host_template *sht); | 880 | extern int ata_scsi_detect(struct scsi_host_template *sht); |
815 | extern int ata_scsi_ioctl(struct scsi_device *dev, int cmd, void __user *arg); | 881 | extern int ata_scsi_ioctl(struct scsi_device *dev, int cmd, void __user *arg); |
816 | extern int ata_scsi_queuecmd(struct scsi_cmnd *cmd, void (*done)(struct scsi_cmnd *)); | 882 | extern int ata_scsi_queuecmd(struct scsi_cmnd *cmd, void (*done)(struct scsi_cmnd *)); |
@@ -823,7 +889,6 @@ extern void ata_sas_port_stop(struct ata_port *ap); | |||
823 | extern int ata_sas_slave_configure(struct scsi_device *, struct ata_port *); | 889 | extern int ata_sas_slave_configure(struct scsi_device *, struct ata_port *); |
824 | extern int ata_sas_queuecmd(struct scsi_cmnd *cmd, void (*done)(struct scsi_cmnd *), | 890 | extern int ata_sas_queuecmd(struct scsi_cmnd *cmd, void (*done)(struct scsi_cmnd *), |
825 | struct ata_port *ap); | 891 | struct ata_port *ap); |
826 | extern unsigned int ata_host_intr(struct ata_port *ap, struct ata_queued_cmd *qc); | ||
827 | extern int sata_scr_valid(struct ata_link *link); | 892 | extern int sata_scr_valid(struct ata_link *link); |
828 | extern int sata_scr_read(struct ata_link *link, int reg, u32 *val); | 893 | extern int sata_scr_read(struct ata_link *link, int reg, u32 *val); |
829 | extern int sata_scr_write(struct ata_link *link, int reg, u32 val); | 894 | extern int sata_scr_write(struct ata_link *link, int reg, u32 val); |
@@ -835,21 +900,9 @@ extern int ata_host_suspend(struct ata_host *host, pm_message_t mesg); | |||
835 | extern void ata_host_resume(struct ata_host *host); | 900 | extern void ata_host_resume(struct ata_host *host); |
836 | #endif | 901 | #endif |
837 | extern int ata_ratelimit(void); | 902 | extern int ata_ratelimit(void); |
838 | extern int ata_busy_sleep(struct ata_port *ap, | ||
839 | unsigned long timeout_pat, unsigned long timeout); | ||
840 | extern void ata_wait_after_reset(struct ata_port *ap, unsigned long deadline); | ||
841 | extern int ata_wait_ready(struct ata_port *ap, unsigned long deadline); | ||
842 | extern u32 ata_wait_register(void __iomem *reg, u32 mask, u32 val, | 903 | extern u32 ata_wait_register(void __iomem *reg, u32 mask, u32 val, |
843 | unsigned long interval_msec, | 904 | unsigned long interval_msec, |
844 | unsigned long timeout_msec); | 905 | unsigned long timeout_msec); |
845 | extern unsigned int ata_dev_try_classify(struct ata_device *dev, int present, | ||
846 | u8 *r_err); | ||
847 | |||
848 | /* | ||
849 | * Default driver ops implementations | ||
850 | */ | ||
851 | extern void ata_tf_load(struct ata_port *ap, const struct ata_taskfile *tf); | ||
852 | extern void ata_tf_read(struct ata_port *ap, struct ata_taskfile *tf); | ||
853 | extern int atapi_cmd_type(u8 opcode); | 906 | extern int atapi_cmd_type(u8 opcode); |
854 | extern void ata_tf_to_fis(const struct ata_taskfile *tf, | 907 | extern void ata_tf_to_fis(const struct ata_taskfile *tf, |
855 | u8 pmp, int is_cmd, u8 *fis); | 908 | u8 pmp, int is_cmd, u8 *fis); |
@@ -864,23 +917,9 @@ extern unsigned long ata_xfer_mode2mask(u8 xfer_mode); | |||
864 | extern int ata_xfer_mode2shift(unsigned long xfer_mode); | 917 | extern int ata_xfer_mode2shift(unsigned long xfer_mode); |
865 | extern const char *ata_mode_string(unsigned long xfer_mask); | 918 | extern const char *ata_mode_string(unsigned long xfer_mask); |
866 | extern unsigned long ata_id_xfermask(const u16 *id); | 919 | extern unsigned long ata_id_xfermask(const u16 *id); |
867 | extern void ata_noop_dev_select(struct ata_port *ap, unsigned int device); | ||
868 | extern void ata_std_dev_select(struct ata_port *ap, unsigned int device); | ||
869 | extern u8 ata_check_status(struct ata_port *ap); | ||
870 | extern u8 ata_altstatus(struct ata_port *ap); | ||
871 | extern void ata_exec_command(struct ata_port *ap, const struct ata_taskfile *tf); | ||
872 | extern int ata_port_start(struct ata_port *ap); | 920 | extern int ata_port_start(struct ata_port *ap); |
873 | extern int ata_sff_port_start(struct ata_port *ap); | ||
874 | extern irqreturn_t ata_interrupt(int irq, void *dev_instance); | ||
875 | extern unsigned int ata_data_xfer(struct ata_device *dev, | ||
876 | unsigned char *buf, unsigned int buflen, int rw); | ||
877 | extern unsigned int ata_data_xfer_noirq(struct ata_device *dev, | ||
878 | unsigned char *buf, unsigned int buflen, int rw); | ||
879 | extern int ata_std_qc_defer(struct ata_queued_cmd *qc); | 921 | extern int ata_std_qc_defer(struct ata_queued_cmd *qc); |
880 | extern void ata_dumb_qc_prep(struct ata_queued_cmd *qc); | ||
881 | extern void ata_qc_prep(struct ata_queued_cmd *qc); | ||
882 | extern void ata_noop_qc_prep(struct ata_queued_cmd *qc); | 922 | extern void ata_noop_qc_prep(struct ata_queued_cmd *qc); |
883 | extern unsigned int ata_qc_issue_prot(struct ata_queued_cmd *qc); | ||
884 | extern void ata_sg_init(struct ata_queued_cmd *qc, struct scatterlist *sg, | 923 | extern void ata_sg_init(struct ata_queued_cmd *qc, struct scatterlist *sg, |
885 | unsigned int n_elem); | 924 | unsigned int n_elem); |
886 | extern unsigned int ata_dev_classify(const struct ata_taskfile *tf); | 925 | extern unsigned int ata_dev_classify(const struct ata_taskfile *tf); |
@@ -889,24 +928,8 @@ extern void ata_id_string(const u16 *id, unsigned char *s, | |||
889 | unsigned int ofs, unsigned int len); | 928 | unsigned int ofs, unsigned int len); |
890 | extern void ata_id_c_string(const u16 *id, unsigned char *s, | 929 | extern void ata_id_c_string(const u16 *id, unsigned char *s, |
891 | unsigned int ofs, unsigned int len); | 930 | unsigned int ofs, unsigned int len); |
892 | extern void ata_bmdma_setup(struct ata_queued_cmd *qc); | ||
893 | extern void ata_bmdma_start(struct ata_queued_cmd *qc); | ||
894 | extern void ata_bmdma_stop(struct ata_queued_cmd *qc); | ||
895 | extern u8 ata_bmdma_status(struct ata_port *ap); | ||
896 | extern void ata_bmdma_irq_clear(struct ata_port *ap); | ||
897 | extern void ata_bmdma_freeze(struct ata_port *ap); | ||
898 | extern void ata_bmdma_thaw(struct ata_port *ap); | ||
899 | extern void ata_bmdma_drive_eh(struct ata_port *ap, ata_prereset_fn_t prereset, | ||
900 | ata_reset_fn_t softreset, | ||
901 | ata_reset_fn_t hardreset, | ||
902 | ata_postreset_fn_t postreset); | ||
903 | extern void ata_bmdma_error_handler(struct ata_port *ap); | ||
904 | extern void ata_bmdma_post_internal_cmd(struct ata_queued_cmd *qc); | ||
905 | extern int ata_hsm_move(struct ata_port *ap, struct ata_queued_cmd *qc, | ||
906 | u8 status, int in_wq); | ||
907 | extern void ata_qc_complete(struct ata_queued_cmd *qc); | 931 | extern void ata_qc_complete(struct ata_queued_cmd *qc); |
908 | extern int ata_qc_complete_multiple(struct ata_port *ap, u32 qc_active, | 932 | extern int ata_qc_complete_multiple(struct ata_port *ap, u32 qc_active); |
909 | void (*finish_qc)(struct ata_queued_cmd *)); | ||
910 | extern void ata_scsi_simulate(struct ata_device *dev, struct scsi_cmnd *cmd, | 933 | extern void ata_scsi_simulate(struct ata_device *dev, struct scsi_cmnd *cmd, |
911 | void (*done)(struct scsi_cmnd *)); | 934 | void (*done)(struct scsi_cmnd *)); |
912 | extern int ata_std_bios_param(struct scsi_device *sdev, | 935 | extern int ata_std_bios_param(struct scsi_device *sdev, |
@@ -918,7 +941,6 @@ extern int ata_scsi_change_queue_depth(struct scsi_device *sdev, | |||
918 | int queue_depth); | 941 | int queue_depth); |
919 | extern struct ata_device *ata_dev_pair(struct ata_device *adev); | 942 | extern struct ata_device *ata_dev_pair(struct ata_device *adev); |
920 | extern int ata_do_set_mode(struct ata_link *link, struct ata_device **r_failed_dev); | 943 | extern int ata_do_set_mode(struct ata_link *link, struct ata_device **r_failed_dev); |
921 | extern u8 ata_irq_on(struct ata_port *ap); | ||
922 | 944 | ||
923 | extern int ata_cable_40wire(struct ata_port *ap); | 945 | extern int ata_cable_40wire(struct ata_port *ap); |
924 | extern int ata_cable_80wire(struct ata_port *ap); | 946 | extern int ata_cable_80wire(struct ata_port *ap); |
@@ -926,10 +948,7 @@ extern int ata_cable_sata(struct ata_port *ap); | |||
926 | extern int ata_cable_ignore(struct ata_port *ap); | 948 | extern int ata_cable_ignore(struct ata_port *ap); |
927 | extern int ata_cable_unknown(struct ata_port *ap); | 949 | extern int ata_cable_unknown(struct ata_port *ap); |
928 | 950 | ||
929 | /* | 951 | /* Timing helpers */ |
930 | * Timing helpers | ||
931 | */ | ||
932 | |||
933 | extern unsigned int ata_pio_need_iordy(const struct ata_device *); | 952 | extern unsigned int ata_pio_need_iordy(const struct ata_device *); |
934 | extern const struct ata_timing *ata_timing_find_mode(u8 xfer_mode); | 953 | extern const struct ata_timing *ata_timing_find_mode(u8 xfer_mode); |
935 | extern int ata_timing_compute(struct ata_device *, unsigned short, | 954 | extern int ata_timing_compute(struct ata_device *, unsigned short, |
@@ -939,24 +958,31 @@ extern void ata_timing_merge(const struct ata_timing *, | |||
939 | unsigned int); | 958 | unsigned int); |
940 | extern u8 ata_timing_cycle2mode(unsigned int xfer_shift, int cycle); | 959 | extern u8 ata_timing_cycle2mode(unsigned int xfer_shift, int cycle); |
941 | 960 | ||
942 | enum { | 961 | /* PCI */ |
943 | ATA_TIMING_SETUP = (1 << 0), | 962 | #ifdef CONFIG_PCI |
944 | ATA_TIMING_ACT8B = (1 << 1), | 963 | struct pci_dev; |
945 | ATA_TIMING_REC8B = (1 << 2), | 964 | |
946 | ATA_TIMING_CYC8B = (1 << 3), | 965 | struct pci_bits { |
947 | ATA_TIMING_8BIT = ATA_TIMING_ACT8B | ATA_TIMING_REC8B | | 966 | unsigned int reg; /* PCI config register to read */ |
948 | ATA_TIMING_CYC8B, | 967 | unsigned int width; /* 1 (8 bit), 2 (16 bit), 4 (32 bit) */ |
949 | ATA_TIMING_ACTIVE = (1 << 4), | 968 | unsigned long mask; |
950 | ATA_TIMING_RECOVER = (1 << 5), | 969 | unsigned long val; |
951 | ATA_TIMING_CYCLE = (1 << 6), | ||
952 | ATA_TIMING_UDMA = (1 << 7), | ||
953 | ATA_TIMING_ALL = ATA_TIMING_SETUP | ATA_TIMING_ACT8B | | ||
954 | ATA_TIMING_REC8B | ATA_TIMING_CYC8B | | ||
955 | ATA_TIMING_ACTIVE | ATA_TIMING_RECOVER | | ||
956 | ATA_TIMING_CYCLE | ATA_TIMING_UDMA, | ||
957 | }; | 970 | }; |
958 | 971 | ||
959 | /* libata-acpi.c */ | 972 | extern int pci_test_config_bits(struct pci_dev *pdev, const struct pci_bits *bits); |
973 | extern void ata_pci_remove_one(struct pci_dev *pdev); | ||
974 | |||
975 | #ifdef CONFIG_PM | ||
976 | extern void ata_pci_device_do_suspend(struct pci_dev *pdev, pm_message_t mesg); | ||
977 | extern int __must_check ata_pci_device_do_resume(struct pci_dev *pdev); | ||
978 | extern int ata_pci_device_suspend(struct pci_dev *pdev, pm_message_t mesg); | ||
979 | extern int ata_pci_device_resume(struct pci_dev *pdev); | ||
980 | #endif /* CONFIG_PM */ | ||
981 | #endif /* CONFIG_PCI */ | ||
982 | |||
983 | /* | ||
984 | * ACPI - drivers/ata/libata-acpi.c | ||
985 | */ | ||
960 | #ifdef CONFIG_ATA_ACPI | 986 | #ifdef CONFIG_ATA_ACPI |
961 | static inline const struct ata_acpi_gtm *ata_acpi_init_gtm(struct ata_port *ap) | 987 | static inline const struct ata_acpi_gtm *ata_acpi_init_gtm(struct ata_port *ap) |
962 | { | 988 | { |
@@ -1000,56 +1026,8 @@ static inline int ata_acpi_cbl_80wire(struct ata_port *ap, | |||
1000 | } | 1026 | } |
1001 | #endif | 1027 | #endif |
1002 | 1028 | ||
1003 | #ifdef CONFIG_PCI | ||
1004 | struct pci_dev; | ||
1005 | |||
1006 | extern int ata_pci_init_one(struct pci_dev *pdev, | ||
1007 | const struct ata_port_info * const * ppi); | ||
1008 | extern void ata_pci_remove_one(struct pci_dev *pdev); | ||
1009 | #ifdef CONFIG_PM | ||
1010 | extern void ata_pci_device_do_suspend(struct pci_dev *pdev, pm_message_t mesg); | ||
1011 | extern int __must_check ata_pci_device_do_resume(struct pci_dev *pdev); | ||
1012 | extern int ata_pci_device_suspend(struct pci_dev *pdev, pm_message_t mesg); | ||
1013 | extern int ata_pci_device_resume(struct pci_dev *pdev); | ||
1014 | #endif | ||
1015 | extern int ata_pci_clear_simplex(struct pci_dev *pdev); | ||
1016 | |||
1017 | struct pci_bits { | ||
1018 | unsigned int reg; /* PCI config register to read */ | ||
1019 | unsigned int width; /* 1 (8 bit), 2 (16 bit), 4 (32 bit) */ | ||
1020 | unsigned long mask; | ||
1021 | unsigned long val; | ||
1022 | }; | ||
1023 | |||
1024 | extern int ata_pci_init_sff_host(struct ata_host *host); | ||
1025 | extern int ata_pci_init_bmdma(struct ata_host *host); | ||
1026 | extern int ata_pci_prepare_sff_host(struct pci_dev *pdev, | ||
1027 | const struct ata_port_info * const * ppi, | ||
1028 | struct ata_host **r_host); | ||
1029 | extern int ata_pci_activate_sff_host(struct ata_host *host, | ||
1030 | irq_handler_t irq_handler, | ||
1031 | struct scsi_host_template *sht); | ||
1032 | extern int pci_test_config_bits(struct pci_dev *pdev, const struct pci_bits *bits); | ||
1033 | extern unsigned long ata_pci_default_filter(struct ata_device *dev, | ||
1034 | unsigned long xfer_mask); | ||
1035 | #endif /* CONFIG_PCI */ | ||
1036 | |||
1037 | /* | ||
1038 | * PMP | ||
1039 | */ | ||
1040 | extern int sata_pmp_qc_defer_cmd_switch(struct ata_queued_cmd *qc); | ||
1041 | extern int sata_pmp_std_prereset(struct ata_link *link, unsigned long deadline); | ||
1042 | extern int sata_pmp_std_hardreset(struct ata_link *link, unsigned int *class, | ||
1043 | unsigned long deadline); | ||
1044 | extern void sata_pmp_std_postreset(struct ata_link *link, unsigned int *class); | ||
1045 | extern void sata_pmp_do_eh(struct ata_port *ap, | ||
1046 | ata_prereset_fn_t prereset, ata_reset_fn_t softreset, | ||
1047 | ata_reset_fn_t hardreset, ata_postreset_fn_t postreset, | ||
1048 | ata_prereset_fn_t pmp_prereset, ata_reset_fn_t pmp_softreset, | ||
1049 | ata_reset_fn_t pmp_hardreset, ata_postreset_fn_t pmp_postreset); | ||
1050 | |||
1051 | /* | 1029 | /* |
1052 | * EH | 1030 | * EH - drivers/ata/libata-eh.c |
1053 | */ | 1031 | */ |
1054 | extern void ata_port_schedule_eh(struct ata_port *ap); | 1032 | extern void ata_port_schedule_eh(struct ata_port *ap); |
1055 | extern int ata_link_abort(struct ata_link *link); | 1033 | extern int ata_link_abort(struct ata_link *link); |
@@ -1066,6 +1044,92 @@ extern void ata_eh_qc_retry(struct ata_queued_cmd *qc); | |||
1066 | extern void ata_do_eh(struct ata_port *ap, ata_prereset_fn_t prereset, | 1044 | extern void ata_do_eh(struct ata_port *ap, ata_prereset_fn_t prereset, |
1067 | ata_reset_fn_t softreset, ata_reset_fn_t hardreset, | 1045 | ata_reset_fn_t softreset, ata_reset_fn_t hardreset, |
1068 | ata_postreset_fn_t postreset); | 1046 | ata_postreset_fn_t postreset); |
1047 | extern void ata_std_error_handler(struct ata_port *ap); | ||
1048 | |||
1049 | /* | ||
1050 | * Base operations to inherit from and initializers for sht | ||
1051 | * | ||
1052 | * Operations | ||
1053 | * | ||
1054 | * base : Common to all libata drivers. | ||
1055 | * sata : SATA controllers w/ native interface. | ||
1056 | * pmp : SATA controllers w/ PMP support. | ||
1057 | * sff : SFF ATA controllers w/o BMDMA support. | ||
1058 | * bmdma : SFF ATA controllers w/ BMDMA support. | ||
1059 | * | ||
1060 | * sht initializers | ||
1061 | * | ||
1062 | * BASE : Common to all libata drivers. The user must set | ||
1063 | * sg_tablesize and dma_boundary. | ||
1064 | * PIO : SFF ATA controllers w/ only PIO support. | ||
1065 | * BMDMA : SFF ATA controllers w/ BMDMA support. sg_tablesize and | ||
1066 | * dma_boundary are set to BMDMA limits. | ||
1067 | * NCQ : SATA controllers supporting NCQ. The user must set | ||
1068 | * sg_tablesize, dma_boundary and can_queue. | ||
1069 | */ | ||
1070 | extern const struct ata_port_operations ata_base_port_ops; | ||
1071 | extern const struct ata_port_operations sata_port_ops; | ||
1072 | |||
1073 | #define ATA_BASE_SHT(drv_name) \ | ||
1074 | .module = THIS_MODULE, \ | ||
1075 | .name = drv_name, \ | ||
1076 | .ioctl = ata_scsi_ioctl, \ | ||
1077 | .queuecommand = ata_scsi_queuecmd, \ | ||
1078 | .can_queue = ATA_DEF_QUEUE, \ | ||
1079 | .this_id = ATA_SHT_THIS_ID, \ | ||
1080 | .cmd_per_lun = ATA_SHT_CMD_PER_LUN, \ | ||
1081 | .emulated = ATA_SHT_EMULATED, \ | ||
1082 | .use_clustering = ATA_SHT_USE_CLUSTERING, \ | ||
1083 | .proc_name = drv_name, \ | ||
1084 | .slave_configure = ata_scsi_slave_config, \ | ||
1085 | .slave_destroy = ata_scsi_slave_destroy, \ | ||
1086 | .bios_param = ata_std_bios_param | ||
1087 | |||
1088 | #define ATA_NCQ_SHT(drv_name) \ | ||
1089 | ATA_BASE_SHT(drv_name), \ | ||
1090 | .change_queue_depth = ata_scsi_change_queue_depth | ||
1091 | |||
1092 | /* | ||
1093 | * PMP helpers | ||
1094 | */ | ||
1095 | #ifdef CONFIG_SATA_PMP | ||
1096 | static inline bool sata_pmp_supported(struct ata_port *ap) | ||
1097 | { | ||
1098 | return ap->flags & ATA_FLAG_PMP; | ||
1099 | } | ||
1100 | |||
1101 | static inline bool sata_pmp_attached(struct ata_port *ap) | ||
1102 | { | ||
1103 | return ap->nr_pmp_links != 0; | ||
1104 | } | ||
1105 | |||
1106 | static inline int ata_is_host_link(const struct ata_link *link) | ||
1107 | { | ||
1108 | return link == &link->ap->link; | ||
1109 | } | ||
1110 | #else /* CONFIG_SATA_PMP */ | ||
1111 | static inline bool sata_pmp_supported(struct ata_port *ap) | ||
1112 | { | ||
1113 | return false; | ||
1114 | } | ||
1115 | |||
1116 | static inline bool sata_pmp_attached(struct ata_port *ap) | ||
1117 | { | ||
1118 | return false; | ||
1119 | } | ||
1120 | |||
1121 | static inline int ata_is_host_link(const struct ata_link *link) | ||
1122 | { | ||
1123 | return 1; | ||
1124 | } | ||
1125 | #endif /* CONFIG_SATA_PMP */ | ||
1126 | |||
1127 | static inline int sata_srst_pmp(struct ata_link *link) | ||
1128 | { | ||
1129 | if (sata_pmp_supported(link->ap) && ata_is_host_link(link)) | ||
1130 | return SATA_PMP_CTRL_PORT; | ||
1131 | return link->pmp; | ||
1132 | } | ||
1069 | 1133 | ||
1070 | /* | 1134 | /* |
1071 | * printk helpers | 1135 | * printk helpers |
@@ -1074,7 +1138,7 @@ extern void ata_do_eh(struct ata_port *ap, ata_prereset_fn_t prereset, | |||
1074 | printk("%sata%u: "fmt, lv, (ap)->print_id , ##args) | 1138 | printk("%sata%u: "fmt, lv, (ap)->print_id , ##args) |
1075 | 1139 | ||
1076 | #define ata_link_printk(link, lv, fmt, args...) do { \ | 1140 | #define ata_link_printk(link, lv, fmt, args...) do { \ |
1077 | if ((link)->ap->nr_pmp_links) \ | 1141 | if (sata_pmp_attached((link)->ap)) \ |
1078 | printk("%sata%u.%02u: "fmt, lv, (link)->ap->print_id, \ | 1142 | printk("%sata%u.%02u: "fmt, lv, (link)->ap->print_id, \ |
1079 | (link)->pmp , ##args); \ | 1143 | (link)->pmp , ##args); \ |
1080 | else \ | 1144 | else \ |
@@ -1094,18 +1158,11 @@ extern void ata_ehi_push_desc(struct ata_eh_info *ehi, const char *fmt, ...) | |||
1094 | __attribute__ ((format (printf, 2, 3))); | 1158 | __attribute__ ((format (printf, 2, 3))); |
1095 | extern void ata_ehi_clear_desc(struct ata_eh_info *ehi); | 1159 | extern void ata_ehi_clear_desc(struct ata_eh_info *ehi); |
1096 | 1160 | ||
1097 | static inline void ata_ehi_schedule_probe(struct ata_eh_info *ehi) | ||
1098 | { | ||
1099 | ehi->flags |= ATA_EHI_RESUME_LINK; | ||
1100 | ehi->action |= ATA_EH_SOFTRESET; | ||
1101 | ehi->probe_mask |= (1 << ATA_MAX_DEVICES) - 1; | ||
1102 | } | ||
1103 | |||
1104 | static inline void ata_ehi_hotplugged(struct ata_eh_info *ehi) | 1161 | static inline void ata_ehi_hotplugged(struct ata_eh_info *ehi) |
1105 | { | 1162 | { |
1106 | ata_ehi_schedule_probe(ehi); | 1163 | ehi->probe_mask |= (1 << ATA_MAX_DEVICES) - 1; |
1107 | ehi->flags |= ATA_EHI_HOTPLUGGED; | 1164 | ehi->flags |= ATA_EHI_HOTPLUGGED; |
1108 | ehi->action |= ATA_EH_ENABLE_LINK; | 1165 | ehi->action |= ATA_EH_RESET | ATA_EH_ENABLE_LINK; |
1109 | ehi->err_mask |= AC_ERR_ATA_BUS; | 1166 | ehi->err_mask |= AC_ERR_ATA_BUS; |
1110 | } | 1167 | } |
1111 | 1168 | ||
@@ -1126,7 +1183,7 @@ static inline unsigned int ata_tag_valid(unsigned int tag) | |||
1126 | 1183 | ||
1127 | static inline unsigned int ata_tag_internal(unsigned int tag) | 1184 | static inline unsigned int ata_tag_internal(unsigned int tag) |
1128 | { | 1185 | { |
1129 | return tag == ATA_MAX_QUEUE - 1; | 1186 | return tag == ATA_TAG_INTERNAL; |
1130 | } | 1187 | } |
1131 | 1188 | ||
1132 | /* | 1189 | /* |
@@ -1167,11 +1224,6 @@ static inline unsigned int ata_dev_absent(const struct ata_device *dev) | |||
1167 | /* | 1224 | /* |
1168 | * link helpers | 1225 | * link helpers |
1169 | */ | 1226 | */ |
1170 | static inline int ata_is_host_link(const struct ata_link *link) | ||
1171 | { | ||
1172 | return link == &link->ap->link; | ||
1173 | } | ||
1174 | |||
1175 | static inline int ata_link_max_devices(const struct ata_link *link) | 1227 | static inline int ata_link_max_devices(const struct ata_link *link) |
1176 | { | 1228 | { |
1177 | if (ata_is_host_link(link) && link->ap->flags & ATA_FLAG_SLAVE_POSS) | 1229 | if (ata_is_host_link(link) && link->ap->flags & ATA_FLAG_SLAVE_POSS) |
@@ -1186,7 +1238,7 @@ static inline int ata_link_active(struct ata_link *link) | |||
1186 | 1238 | ||
1187 | static inline struct ata_link *ata_port_first_link(struct ata_port *ap) | 1239 | static inline struct ata_link *ata_port_first_link(struct ata_port *ap) |
1188 | { | 1240 | { |
1189 | if (ap->nr_pmp_links) | 1241 | if (sata_pmp_attached(ap)) |
1190 | return ap->pmp_link; | 1242 | return ap->pmp_link; |
1191 | return &ap->link; | 1243 | return &ap->link; |
1192 | } | 1244 | } |
@@ -1195,8 +1247,8 @@ static inline struct ata_link *ata_port_next_link(struct ata_link *link) | |||
1195 | { | 1247 | { |
1196 | struct ata_port *ap = link->ap; | 1248 | struct ata_port *ap = link->ap; |
1197 | 1249 | ||
1198 | if (link == &ap->link) { | 1250 | if (ata_is_host_link(link)) { |
1199 | if (!ap->nr_pmp_links) | 1251 | if (!sata_pmp_attached(ap)) |
1200 | return NULL; | 1252 | return NULL; |
1201 | return ap->pmp_link; | 1253 | return ap->pmp_link; |
1202 | } | 1254 | } |
@@ -1222,11 +1274,6 @@ static inline struct ata_link *ata_port_next_link(struct ata_link *link) | |||
1222 | for ((dev) = (link)->device + ata_link_max_devices(link) - 1; \ | 1274 | for ((dev) = (link)->device + ata_link_max_devices(link) - 1; \ |
1223 | (dev) >= (link)->device || ((dev) = NULL); (dev)--) | 1275 | (dev) >= (link)->device || ((dev) = NULL); (dev)--) |
1224 | 1276 | ||
1225 | static inline u8 ata_chk_status(struct ata_port *ap) | ||
1226 | { | ||
1227 | return ap->ops->check_status(ap); | ||
1228 | } | ||
1229 | |||
1230 | /** | 1277 | /** |
1231 | * ata_ncq_enabled - Test whether NCQ is enabled | 1278 | * ata_ncq_enabled - Test whether NCQ is enabled |
1232 | * @dev: ATA device to test for | 1279 | * @dev: ATA device to test for |
@@ -1243,74 +1290,6 @@ static inline int ata_ncq_enabled(struct ata_device *dev) | |||
1243 | ATA_DFLAG_NCQ)) == ATA_DFLAG_NCQ; | 1290 | ATA_DFLAG_NCQ)) == ATA_DFLAG_NCQ; |
1244 | } | 1291 | } |
1245 | 1292 | ||
1246 | /** | ||
1247 | * ata_pause - Flush writes and pause 400 nanoseconds. | ||
1248 | * @ap: Port to wait for. | ||
1249 | * | ||
1250 | * LOCKING: | ||
1251 | * Inherited from caller. | ||
1252 | */ | ||
1253 | |||
1254 | static inline void ata_pause(struct ata_port *ap) | ||
1255 | { | ||
1256 | ata_altstatus(ap); | ||
1257 | ndelay(400); | ||
1258 | } | ||
1259 | |||
1260 | |||
1261 | /** | ||
1262 | * ata_busy_wait - Wait for a port status register | ||
1263 | * @ap: Port to wait for. | ||
1264 | * @bits: bits that must be clear | ||
1265 | * @max: number of 10uS waits to perform | ||
1266 | * | ||
1267 | * Waits up to max*10 microseconds for the selected bits in the port's | ||
1268 | * status register to be cleared. | ||
1269 | * Returns final value of status register. | ||
1270 | * | ||
1271 | * LOCKING: | ||
1272 | * Inherited from caller. | ||
1273 | */ | ||
1274 | |||
1275 | static inline u8 ata_busy_wait(struct ata_port *ap, unsigned int bits, | ||
1276 | unsigned int max) | ||
1277 | { | ||
1278 | u8 status; | ||
1279 | |||
1280 | do { | ||
1281 | udelay(10); | ||
1282 | status = ata_chk_status(ap); | ||
1283 | max--; | ||
1284 | } while (status != 0xff && (status & bits) && (max > 0)); | ||
1285 | |||
1286 | return status; | ||
1287 | } | ||
1288 | |||
1289 | |||
1290 | /** | ||
1291 | * ata_wait_idle - Wait for a port to be idle. | ||
1292 | * @ap: Port to wait for. | ||
1293 | * | ||
1294 | * Waits up to 10ms for port's BUSY and DRQ signals to clear. | ||
1295 | * Returns final value of status register. | ||
1296 | * | ||
1297 | * LOCKING: | ||
1298 | * Inherited from caller. | ||
1299 | */ | ||
1300 | |||
1301 | static inline u8 ata_wait_idle(struct ata_port *ap) | ||
1302 | { | ||
1303 | u8 status = ata_busy_wait(ap, ATA_BUSY | ATA_DRQ, 1000); | ||
1304 | |||
1305 | #ifdef ATA_DEBUG | ||
1306 | if (status != 0xff && (status & (ATA_BUSY | ATA_DRQ))) | ||
1307 | ata_port_printk(ap, KERN_DEBUG, "abnormal Status 0x%X\n", | ||
1308 | status); | ||
1309 | #endif | ||
1310 | |||
1311 | return status; | ||
1312 | } | ||
1313 | |||
1314 | static inline void ata_qc_set_polling(struct ata_queued_cmd *qc) | 1293 | static inline void ata_qc_set_polling(struct ata_queued_cmd *qc) |
1315 | { | 1294 | { |
1316 | qc->tf.ctl |= ATA_NIEN; | 1295 | qc->tf.ctl |= ATA_NIEN; |
@@ -1403,4 +1382,171 @@ static inline struct ata_port *ata_shost_to_port(struct Scsi_Host *host) | |||
1403 | return *(struct ata_port **)&host->hostdata[0]; | 1382 | return *(struct ata_port **)&host->hostdata[0]; |
1404 | } | 1383 | } |
1405 | 1384 | ||
1385 | |||
1386 | /************************************************************************** | ||
1387 | * PMP - drivers/ata/libata-pmp.c | ||
1388 | */ | ||
1389 | #ifdef CONFIG_SATA_PMP | ||
1390 | |||
1391 | extern const struct ata_port_operations sata_pmp_port_ops; | ||
1392 | |||
1393 | extern int sata_pmp_qc_defer_cmd_switch(struct ata_queued_cmd *qc); | ||
1394 | extern void sata_pmp_error_handler(struct ata_port *ap); | ||
1395 | |||
1396 | #else /* CONFIG_SATA_PMP */ | ||
1397 | |||
1398 | #define sata_pmp_port_ops sata_port_ops | ||
1399 | #define sata_pmp_qc_defer_cmd_switch ata_std_qc_defer | ||
1400 | #define sata_pmp_error_handler ata_std_error_handler | ||
1401 | |||
1402 | #endif /* CONFIG_SATA_PMP */ | ||
1403 | |||
1404 | |||
1405 | /************************************************************************** | ||
1406 | * SFF - drivers/ata/libata-sff.c | ||
1407 | */ | ||
1408 | #ifdef CONFIG_ATA_SFF | ||
1409 | |||
1410 | extern const struct ata_port_operations ata_sff_port_ops; | ||
1411 | extern const struct ata_port_operations ata_bmdma_port_ops; | ||
1412 | |||
1413 | /* PIO only, sg_tablesize and dma_boundary limits can be removed */ | ||
1414 | #define ATA_PIO_SHT(drv_name) \ | ||
1415 | ATA_BASE_SHT(drv_name), \ | ||
1416 | .sg_tablesize = LIBATA_MAX_PRD, \ | ||
1417 | .dma_boundary = ATA_DMA_BOUNDARY | ||
1418 | |||
1419 | #define ATA_BMDMA_SHT(drv_name) \ | ||
1420 | ATA_BASE_SHT(drv_name), \ | ||
1421 | .sg_tablesize = LIBATA_MAX_PRD, \ | ||
1422 | .dma_boundary = ATA_DMA_BOUNDARY | ||
1423 | |||
1424 | extern void ata_sff_qc_prep(struct ata_queued_cmd *qc); | ||
1425 | extern void ata_sff_dumb_qc_prep(struct ata_queued_cmd *qc); | ||
1426 | extern void ata_sff_dev_select(struct ata_port *ap, unsigned int device); | ||
1427 | extern u8 ata_sff_check_status(struct ata_port *ap); | ||
1428 | extern u8 ata_sff_altstatus(struct ata_port *ap); | ||
1429 | extern int ata_sff_busy_sleep(struct ata_port *ap, | ||
1430 | unsigned long timeout_pat, unsigned long timeout); | ||
1431 | extern int ata_sff_wait_ready(struct ata_link *link, unsigned long deadline); | ||
1432 | extern void ata_sff_tf_load(struct ata_port *ap, const struct ata_taskfile *tf); | ||
1433 | extern void ata_sff_tf_read(struct ata_port *ap, struct ata_taskfile *tf); | ||
1434 | extern void ata_sff_exec_command(struct ata_port *ap, | ||
1435 | const struct ata_taskfile *tf); | ||
1436 | extern unsigned int ata_sff_data_xfer(struct ata_device *dev, | ||
1437 | unsigned char *buf, unsigned int buflen, int rw); | ||
1438 | extern unsigned int ata_sff_data_xfer_noirq(struct ata_device *dev, | ||
1439 | unsigned char *buf, unsigned int buflen, int rw); | ||
1440 | extern u8 ata_sff_irq_on(struct ata_port *ap); | ||
1441 | extern void ata_sff_irq_clear(struct ata_port *ap); | ||
1442 | extern int ata_sff_hsm_move(struct ata_port *ap, struct ata_queued_cmd *qc, | ||
1443 | u8 status, int in_wq); | ||
1444 | extern unsigned int ata_sff_qc_issue(struct ata_queued_cmd *qc); | ||
1445 | extern bool ata_sff_qc_fill_rtf(struct ata_queued_cmd *qc); | ||
1446 | extern unsigned int ata_sff_host_intr(struct ata_port *ap, | ||
1447 | struct ata_queued_cmd *qc); | ||
1448 | extern irqreturn_t ata_sff_interrupt(int irq, void *dev_instance); | ||
1449 | extern void ata_sff_freeze(struct ata_port *ap); | ||
1450 | extern void ata_sff_thaw(struct ata_port *ap); | ||
1451 | extern int ata_sff_prereset(struct ata_link *link, unsigned long deadline); | ||
1452 | extern unsigned int ata_sff_dev_classify(struct ata_device *dev, int present, | ||
1453 | u8 *r_err); | ||
1454 | extern int ata_sff_wait_after_reset(struct ata_link *link, unsigned int devmask, | ||
1455 | unsigned long deadline); | ||
1456 | extern int ata_sff_softreset(struct ata_link *link, unsigned int *classes, | ||
1457 | unsigned long deadline); | ||
1458 | extern int sata_sff_hardreset(struct ata_link *link, unsigned int *class, | ||
1459 | unsigned long deadline); | ||
1460 | extern void ata_sff_postreset(struct ata_link *link, unsigned int *classes); | ||
1461 | extern void ata_sff_error_handler(struct ata_port *ap); | ||
1462 | extern void ata_sff_post_internal_cmd(struct ata_queued_cmd *qc); | ||
1463 | extern int ata_sff_port_start(struct ata_port *ap); | ||
1464 | extern void ata_sff_std_ports(struct ata_ioports *ioaddr); | ||
1465 | extern unsigned long ata_bmdma_mode_filter(struct ata_device *dev, | ||
1466 | unsigned long xfer_mask); | ||
1467 | extern void ata_bmdma_setup(struct ata_queued_cmd *qc); | ||
1468 | extern void ata_bmdma_start(struct ata_queued_cmd *qc); | ||
1469 | extern void ata_bmdma_stop(struct ata_queued_cmd *qc); | ||
1470 | extern u8 ata_bmdma_status(struct ata_port *ap); | ||
1471 | extern void ata_bus_reset(struct ata_port *ap); | ||
1472 | |||
1473 | #ifdef CONFIG_PCI | ||
1474 | extern int ata_pci_bmdma_clear_simplex(struct pci_dev *pdev); | ||
1475 | extern int ata_pci_bmdma_init(struct ata_host *host); | ||
1476 | extern int ata_pci_sff_init_host(struct ata_host *host); | ||
1477 | extern int ata_pci_sff_prepare_host(struct pci_dev *pdev, | ||
1478 | const struct ata_port_info * const * ppi, | ||
1479 | struct ata_host **r_host); | ||
1480 | extern int ata_pci_sff_activate_host(struct ata_host *host, | ||
1481 | irq_handler_t irq_handler, | ||
1482 | struct scsi_host_template *sht); | ||
1483 | extern int ata_pci_sff_init_one(struct pci_dev *pdev, | ||
1484 | const struct ata_port_info * const * ppi, | ||
1485 | struct scsi_host_template *sht, void *host_priv); | ||
1486 | #endif /* CONFIG_PCI */ | ||
1487 | |||
1488 | /** | ||
1489 | * ata_sff_pause - Flush writes and pause 400 nanoseconds. | ||
1490 | * @ap: Port to wait for. | ||
1491 | * | ||
1492 | * LOCKING: | ||
1493 | * Inherited from caller. | ||
1494 | */ | ||
1495 | static inline void ata_sff_pause(struct ata_port *ap) | ||
1496 | { | ||
1497 | ata_sff_altstatus(ap); | ||
1498 | ndelay(400); | ||
1499 | } | ||
1500 | |||
1501 | /** | ||
1502 | * ata_sff_busy_wait - Wait for a port status register | ||
1503 | * @ap: Port to wait for. | ||
1504 | * @bits: bits that must be clear | ||
1505 | * @max: number of 10uS waits to perform | ||
1506 | * | ||
1507 | * Waits up to max*10 microseconds for the selected bits in the port's | ||
1508 | * status register to be cleared. | ||
1509 | * Returns final value of status register. | ||
1510 | * | ||
1511 | * LOCKING: | ||
1512 | * Inherited from caller. | ||
1513 | */ | ||
1514 | static inline u8 ata_sff_busy_wait(struct ata_port *ap, unsigned int bits, | ||
1515 | unsigned int max) | ||
1516 | { | ||
1517 | u8 status; | ||
1518 | |||
1519 | do { | ||
1520 | udelay(10); | ||
1521 | status = ap->ops->sff_check_status(ap); | ||
1522 | max--; | ||
1523 | } while (status != 0xff && (status & bits) && (max > 0)); | ||
1524 | |||
1525 | return status; | ||
1526 | } | ||
1527 | |||
1528 | /** | ||
1529 | * ata_wait_idle - Wait for a port to be idle. | ||
1530 | * @ap: Port to wait for. | ||
1531 | * | ||
1532 | * Waits up to 10ms for port's BUSY and DRQ signals to clear. | ||
1533 | * Returns final value of status register. | ||
1534 | * | ||
1535 | * LOCKING: | ||
1536 | * Inherited from caller. | ||
1537 | */ | ||
1538 | static inline u8 ata_wait_idle(struct ata_port *ap) | ||
1539 | { | ||
1540 | u8 status = ata_sff_busy_wait(ap, ATA_BUSY | ATA_DRQ, 1000); | ||
1541 | |||
1542 | #ifdef ATA_DEBUG | ||
1543 | if (status != 0xff && (status & (ATA_BUSY | ATA_DRQ))) | ||
1544 | ata_port_printk(ap, KERN_DEBUG, "abnormal Status 0x%X\n", | ||
1545 | status); | ||
1546 | #endif | ||
1547 | |||
1548 | return status; | ||
1549 | } | ||
1550 | #endif /* CONFIG_ATA_SFF */ | ||
1551 | |||
1406 | #endif /* __LINUX_LIBATA_H__ */ | 1552 | #endif /* __LINUX_LIBATA_H__ */ |