aboutsummaryrefslogtreecommitdiffstats
path: root/drivers/scsi
diff options
context:
space:
mode:
authorDan Williams <dan.j.williams@intel.com>2011-06-30 20:38:32 -0400
committerDan Williams <dan.j.williams@intel.com>2011-07-03 07:04:52 -0400
commitd9dcb4ba791de2a06b19ac47cd61601cf3d4e208 (patch)
tree4b818b5b14c078703c034489c92e74975be6c06f /drivers/scsi
parent78a6f06e0e82125787d7aa308fe28c2c8381540c (diff)
isci: unify isci_host and scic_sds_controller
Remove the distinction between these two implementations and unify on isci_host (local instances named ihost). Hmmm, we had two 'oem_parameters' instances, one was unused... nice. Reported-by: Christoph Hellwig <hch@lst.de> Signed-off-by: Dan Williams <dan.j.williams@intel.com>
Diffstat (limited to 'drivers/scsi')
-rw-r--r--drivers/scsi/isci/host.c1000
-rw-r--r--drivers/scsi/isci/host.h210
-rw-r--r--drivers/scsi/isci/init.c10
-rw-r--r--drivers/scsi/isci/phy.c40
-rw-r--r--drivers/scsi/isci/port.c89
-rw-r--r--drivers/scsi/isci/port.h4
-rw-r--r--drivers/scsi/isci/port_config.c82
-rw-r--r--drivers/scsi/isci/probe_roms.h2
-rw-r--r--drivers/scsi/isci/remote_device.c76
-rw-r--r--drivers/scsi/isci/remote_device.h6
-rw-r--r--drivers/scsi/isci/remote_node_context.c14
-rw-r--r--drivers/scsi/isci/request.c200
-rw-r--r--drivers/scsi/isci/request.h4
-rw-r--r--drivers/scsi/isci/task.c44
-rw-r--r--drivers/scsi/isci/unsolicited_frame_control.c6
-rw-r--r--drivers/scsi/isci/unsolicited_frame_control.h4
16 files changed, 806 insertions, 985 deletions
diff --git a/drivers/scsi/isci/host.c b/drivers/scsi/isci/host.c
index 45d7f71c609a..bb298f8f609a 100644
--- a/drivers/scsi/isci/host.c
+++ b/drivers/scsi/isci/host.c
@@ -181,35 +181,35 @@ void sci_change_state(struct sci_base_state_machine *sm, u32 next_state)
181} 181}
182 182
183static bool scic_sds_controller_completion_queue_has_entries( 183static bool scic_sds_controller_completion_queue_has_entries(
184 struct scic_sds_controller *scic) 184 struct isci_host *ihost)
185{ 185{
186 u32 get_value = scic->completion_queue_get; 186 u32 get_value = ihost->completion_queue_get;
187 u32 get_index = get_value & SMU_COMPLETION_QUEUE_GET_POINTER_MASK; 187 u32 get_index = get_value & SMU_COMPLETION_QUEUE_GET_POINTER_MASK;
188 188
189 if (NORMALIZE_GET_POINTER_CYCLE_BIT(get_value) == 189 if (NORMALIZE_GET_POINTER_CYCLE_BIT(get_value) ==
190 COMPLETION_QUEUE_CYCLE_BIT(scic->completion_queue[get_index])) 190 COMPLETION_QUEUE_CYCLE_BIT(ihost->completion_queue[get_index]))
191 return true; 191 return true;
192 192
193 return false; 193 return false;
194} 194}
195 195
196static bool scic_sds_controller_isr(struct scic_sds_controller *scic) 196static bool scic_sds_controller_isr(struct isci_host *ihost)
197{ 197{
198 if (scic_sds_controller_completion_queue_has_entries(scic)) { 198 if (scic_sds_controller_completion_queue_has_entries(ihost)) {
199 return true; 199 return true;
200 } else { 200 } else {
201 /* 201 /*
202 * we have a spurious interrupt it could be that we have already 202 * we have a spurious interrupt it could be that we have already
203 * emptied the completion queue from a previous interrupt */ 203 * emptied the completion queue from a previous interrupt */
204 writel(SMU_ISR_COMPLETION, &scic->smu_registers->interrupt_status); 204 writel(SMU_ISR_COMPLETION, &ihost->smu_registers->interrupt_status);
205 205
206 /* 206 /*
207 * There is a race in the hardware that could cause us not to be notified 207 * There is a race in the hardware that could cause us not to be notified
208 * of an interrupt completion if we do not take this step. We will mask 208 * of an interrupt completion if we do not take this step. We will mask
209 * then unmask the interrupts so if there is another interrupt pending 209 * then unmask the interrupts so if there is another interrupt pending
210 * the clearing of the interrupt source we get the next interrupt message. */ 210 * the clearing of the interrupt source we get the next interrupt message. */
211 writel(0xFF000000, &scic->smu_registers->interrupt_mask); 211 writel(0xFF000000, &ihost->smu_registers->interrupt_mask);
212 writel(0, &scic->smu_registers->interrupt_mask); 212 writel(0, &ihost->smu_registers->interrupt_mask);
213 } 213 }
214 214
215 return false; 215 return false;
@@ -219,18 +219,18 @@ irqreturn_t isci_msix_isr(int vec, void *data)
219{ 219{
220 struct isci_host *ihost = data; 220 struct isci_host *ihost = data;
221 221
222 if (scic_sds_controller_isr(&ihost->sci)) 222 if (scic_sds_controller_isr(ihost))
223 tasklet_schedule(&ihost->completion_tasklet); 223 tasklet_schedule(&ihost->completion_tasklet);
224 224
225 return IRQ_HANDLED; 225 return IRQ_HANDLED;
226} 226}
227 227
228static bool scic_sds_controller_error_isr(struct scic_sds_controller *scic) 228static bool scic_sds_controller_error_isr(struct isci_host *ihost)
229{ 229{
230 u32 interrupt_status; 230 u32 interrupt_status;
231 231
232 interrupt_status = 232 interrupt_status =
233 readl(&scic->smu_registers->interrupt_status); 233 readl(&ihost->smu_registers->interrupt_status);
234 interrupt_status &= (SMU_ISR_QUEUE_ERROR | SMU_ISR_QUEUE_SUSPEND); 234 interrupt_status &= (SMU_ISR_QUEUE_ERROR | SMU_ISR_QUEUE_SUSPEND);
235 235
236 if (interrupt_status != 0) { 236 if (interrupt_status != 0) {
@@ -246,28 +246,27 @@ static bool scic_sds_controller_error_isr(struct scic_sds_controller *scic)
246 * then unmask the error interrupts so if there was another interrupt 246 * then unmask the error interrupts so if there was another interrupt
247 * pending we will be notified. 247 * pending we will be notified.
248 * Could we write the value of (SMU_ISR_QUEUE_ERROR | SMU_ISR_QUEUE_SUSPEND)? */ 248 * Could we write the value of (SMU_ISR_QUEUE_ERROR | SMU_ISR_QUEUE_SUSPEND)? */
249 writel(0xff, &scic->smu_registers->interrupt_mask); 249 writel(0xff, &ihost->smu_registers->interrupt_mask);
250 writel(0, &scic->smu_registers->interrupt_mask); 250 writel(0, &ihost->smu_registers->interrupt_mask);
251 251
252 return false; 252 return false;
253} 253}
254 254
255static void scic_sds_controller_task_completion(struct scic_sds_controller *scic, 255static void scic_sds_controller_task_completion(struct isci_host *ihost,
256 u32 completion_entry) 256 u32 completion_entry)
257{ 257{
258 u32 index = SCU_GET_COMPLETION_INDEX(completion_entry); 258 u32 index = SCU_GET_COMPLETION_INDEX(completion_entry);
259 struct isci_host *ihost = scic_to_ihost(scic);
260 struct isci_request *ireq = ihost->reqs[index]; 259 struct isci_request *ireq = ihost->reqs[index];
261 260
262 /* Make sure that we really want to process this IO request */ 261 /* Make sure that we really want to process this IO request */
263 if (test_bit(IREQ_ACTIVE, &ireq->flags) && 262 if (test_bit(IREQ_ACTIVE, &ireq->flags) &&
264 ireq->io_tag != SCI_CONTROLLER_INVALID_IO_TAG && 263 ireq->io_tag != SCI_CONTROLLER_INVALID_IO_TAG &&
265 ISCI_TAG_SEQ(ireq->io_tag) == scic->io_request_sequence[index]) 264 ISCI_TAG_SEQ(ireq->io_tag) == ihost->io_request_sequence[index])
266 /* Yep this is a valid io request pass it along to the io request handler */ 265 /* Yep this is a valid io request pass it along to the io request handler */
267 scic_sds_io_request_tc_completion(ireq, completion_entry); 266 scic_sds_io_request_tc_completion(ireq, completion_entry);
268} 267}
269 268
270static void scic_sds_controller_sdma_completion(struct scic_sds_controller *scic, 269static void scic_sds_controller_sdma_completion(struct isci_host *ihost,
271 u32 completion_entry) 270 u32 completion_entry)
272{ 271{
273 u32 index; 272 u32 index;
@@ -279,8 +278,8 @@ static void scic_sds_controller_sdma_completion(struct scic_sds_controller *scic
279 switch (scu_get_command_request_type(completion_entry)) { 278 switch (scu_get_command_request_type(completion_entry)) {
280 case SCU_CONTEXT_COMMAND_REQUEST_TYPE_POST_TC: 279 case SCU_CONTEXT_COMMAND_REQUEST_TYPE_POST_TC:
281 case SCU_CONTEXT_COMMAND_REQUEST_TYPE_DUMP_TC: 280 case SCU_CONTEXT_COMMAND_REQUEST_TYPE_DUMP_TC:
282 ireq = scic_to_ihost(scic)->reqs[index]; 281 ireq = ihost->reqs[index];
283 dev_warn(scic_to_dev(scic), "%s: %x for io request %p\n", 282 dev_warn(&ihost->pdev->dev, "%s: %x for io request %p\n",
284 __func__, completion_entry, ireq); 283 __func__, completion_entry, ireq);
285 /* @todo For a post TC operation we need to fail the IO 284 /* @todo For a post TC operation we need to fail the IO
286 * request 285 * request
@@ -289,27 +288,26 @@ static void scic_sds_controller_sdma_completion(struct scic_sds_controller *scic
289 case SCU_CONTEXT_COMMAND_REQUEST_TYPE_DUMP_RNC: 288 case SCU_CONTEXT_COMMAND_REQUEST_TYPE_DUMP_RNC:
290 case SCU_CONTEXT_COMMAND_REQUEST_TYPE_OTHER_RNC: 289 case SCU_CONTEXT_COMMAND_REQUEST_TYPE_OTHER_RNC:
291 case SCU_CONTEXT_COMMAND_REQUEST_TYPE_POST_RNC: 290 case SCU_CONTEXT_COMMAND_REQUEST_TYPE_POST_RNC:
292 idev = scic->device_table[index]; 291 idev = ihost->device_table[index];
293 dev_warn(scic_to_dev(scic), "%s: %x for device %p\n", 292 dev_warn(&ihost->pdev->dev, "%s: %x for device %p\n",
294 __func__, completion_entry, idev); 293 __func__, completion_entry, idev);
295 /* @todo For a port RNC operation we need to fail the 294 /* @todo For a port RNC operation we need to fail the
296 * device 295 * device
297 */ 296 */
298 break; 297 break;
299 default: 298 default:
300 dev_warn(scic_to_dev(scic), "%s: unknown completion type %x\n", 299 dev_warn(&ihost->pdev->dev, "%s: unknown completion type %x\n",
301 __func__, completion_entry); 300 __func__, completion_entry);
302 break; 301 break;
303 } 302 }
304} 303}
305 304
306static void scic_sds_controller_unsolicited_frame(struct scic_sds_controller *scic, 305static void scic_sds_controller_unsolicited_frame(struct isci_host *ihost,
307 u32 completion_entry) 306 u32 completion_entry)
308{ 307{
309 u32 index; 308 u32 index;
310 u32 frame_index; 309 u32 frame_index;
311 310
312 struct isci_host *ihost = scic_to_ihost(scic);
313 struct scu_unsolicited_frame_header *frame_header; 311 struct scu_unsolicited_frame_header *frame_header;
314 struct isci_phy *iphy; 312 struct isci_phy *iphy;
315 struct isci_remote_device *idev; 313 struct isci_remote_device *idev;
@@ -318,15 +316,15 @@ static void scic_sds_controller_unsolicited_frame(struct scic_sds_controller *sc
318 316
319 frame_index = SCU_GET_FRAME_INDEX(completion_entry); 317 frame_index = SCU_GET_FRAME_INDEX(completion_entry);
320 318
321 frame_header = scic->uf_control.buffers.array[frame_index].header; 319 frame_header = ihost->uf_control.buffers.array[frame_index].header;
322 scic->uf_control.buffers.array[frame_index].state = UNSOLICITED_FRAME_IN_USE; 320 ihost->uf_control.buffers.array[frame_index].state = UNSOLICITED_FRAME_IN_USE;
323 321
324 if (SCU_GET_FRAME_ERROR(completion_entry)) { 322 if (SCU_GET_FRAME_ERROR(completion_entry)) {
325 /* 323 /*
326 * / @todo If the IAF frame or SIGNATURE FIS frame has an error will 324 * / @todo If the IAF frame or SIGNATURE FIS frame has an error will
327 * / this cause a problem? We expect the phy initialization will 325 * / this cause a problem? We expect the phy initialization will
328 * / fail if there is an error in the frame. */ 326 * / fail if there is an error in the frame. */
329 scic_sds_controller_release_frame(scic, frame_index); 327 scic_sds_controller_release_frame(ihost, frame_index);
330 return; 328 return;
331 } 329 }
332 330
@@ -347,15 +345,15 @@ static void scic_sds_controller_unsolicited_frame(struct scic_sds_controller *sc
347 iphy = &ihost->phys[index]; 345 iphy = &ihost->phys[index];
348 result = scic_sds_phy_frame_handler(iphy, frame_index); 346 result = scic_sds_phy_frame_handler(iphy, frame_index);
349 } else { 347 } else {
350 if (index < scic->remote_node_entries) 348 if (index < ihost->remote_node_entries)
351 idev = scic->device_table[index]; 349 idev = ihost->device_table[index];
352 else 350 else
353 idev = NULL; 351 idev = NULL;
354 352
355 if (idev != NULL) 353 if (idev != NULL)
356 result = scic_sds_remote_device_frame_handler(idev, frame_index); 354 result = scic_sds_remote_device_frame_handler(idev, frame_index);
357 else 355 else
358 scic_sds_controller_release_frame(scic, frame_index); 356 scic_sds_controller_release_frame(ihost, frame_index);
359 } 357 }
360 } 358 }
361 359
@@ -366,10 +364,9 @@ static void scic_sds_controller_unsolicited_frame(struct scic_sds_controller *sc
366 } 364 }
367} 365}
368 366
369static void scic_sds_controller_event_completion(struct scic_sds_controller *scic, 367static void scic_sds_controller_event_completion(struct isci_host *ihost,
370 u32 completion_entry) 368 u32 completion_entry)
371{ 369{
372 struct isci_host *ihost = scic_to_ihost(scic);
373 struct isci_remote_device *idev; 370 struct isci_remote_device *idev;
374 struct isci_request *ireq; 371 struct isci_request *ireq;
375 struct isci_phy *iphy; 372 struct isci_phy *iphy;
@@ -380,11 +377,11 @@ static void scic_sds_controller_event_completion(struct scic_sds_controller *sci
380 switch (scu_get_event_type(completion_entry)) { 377 switch (scu_get_event_type(completion_entry)) {
381 case SCU_EVENT_TYPE_SMU_COMMAND_ERROR: 378 case SCU_EVENT_TYPE_SMU_COMMAND_ERROR:
382 /* / @todo The driver did something wrong and we need to fix the condtion. */ 379 /* / @todo The driver did something wrong and we need to fix the condtion. */
383 dev_err(scic_to_dev(scic), 380 dev_err(&ihost->pdev->dev,
384 "%s: SCIC Controller 0x%p received SMU command error " 381 "%s: SCIC Controller 0x%p received SMU command error "
385 "0x%x\n", 382 "0x%x\n",
386 __func__, 383 __func__,
387 scic, 384 ihost,
388 completion_entry); 385 completion_entry);
389 break; 386 break;
390 387
@@ -394,11 +391,11 @@ static void scic_sds_controller_event_completion(struct scic_sds_controller *sci
394 /* 391 /*
395 * / @todo This is a hardware failure and its likely that we want to 392 * / @todo This is a hardware failure and its likely that we want to
396 * / reset the controller. */ 393 * / reset the controller. */
397 dev_err(scic_to_dev(scic), 394 dev_err(&ihost->pdev->dev,
398 "%s: SCIC Controller 0x%p received fatal controller " 395 "%s: SCIC Controller 0x%p received fatal controller "
399 "event 0x%x\n", 396 "event 0x%x\n",
400 __func__, 397 __func__,
401 scic, 398 ihost,
402 completion_entry); 399 completion_entry);
403 break; 400 break;
404 401
@@ -415,27 +412,27 @@ static void scic_sds_controller_event_completion(struct scic_sds_controller *sci
415 if (ireq != NULL) 412 if (ireq != NULL)
416 scic_sds_io_request_event_handler(ireq, completion_entry); 413 scic_sds_io_request_event_handler(ireq, completion_entry);
417 else 414 else
418 dev_warn(scic_to_dev(scic), 415 dev_warn(&ihost->pdev->dev,
419 "%s: SCIC Controller 0x%p received " 416 "%s: SCIC Controller 0x%p received "
420 "event 0x%x for io request object " 417 "event 0x%x for io request object "
421 "that doesnt exist.\n", 418 "that doesnt exist.\n",
422 __func__, 419 __func__,
423 scic, 420 ihost,
424 completion_entry); 421 completion_entry);
425 422
426 break; 423 break;
427 424
428 case SCU_EVENT_SPECIFIC_IT_NEXUS_TIMEOUT: 425 case SCU_EVENT_SPECIFIC_IT_NEXUS_TIMEOUT:
429 idev = scic->device_table[index]; 426 idev = ihost->device_table[index];
430 if (idev != NULL) 427 if (idev != NULL)
431 scic_sds_remote_device_event_handler(idev, completion_entry); 428 scic_sds_remote_device_event_handler(idev, completion_entry);
432 else 429 else
433 dev_warn(scic_to_dev(scic), 430 dev_warn(&ihost->pdev->dev,
434 "%s: SCIC Controller 0x%p received " 431 "%s: SCIC Controller 0x%p received "
435 "event 0x%x for remote device object " 432 "event 0x%x for remote device object "
436 "that doesnt exist.\n", 433 "that doesnt exist.\n",
437 __func__, 434 __func__,
438 scic, 435 ihost,
439 completion_entry); 436 completion_entry);
440 437
441 break; 438 break;
@@ -459,25 +456,25 @@ static void scic_sds_controller_event_completion(struct scic_sds_controller *sci
459 case SCU_EVENT_TYPE_RNC_SUSPEND_TX: 456 case SCU_EVENT_TYPE_RNC_SUSPEND_TX:
460 case SCU_EVENT_TYPE_RNC_SUSPEND_TX_RX: 457 case SCU_EVENT_TYPE_RNC_SUSPEND_TX_RX:
461 case SCU_EVENT_TYPE_RNC_OPS_MISC: 458 case SCU_EVENT_TYPE_RNC_OPS_MISC:
462 if (index < scic->remote_node_entries) { 459 if (index < ihost->remote_node_entries) {
463 idev = scic->device_table[index]; 460 idev = ihost->device_table[index];
464 461
465 if (idev != NULL) 462 if (idev != NULL)
466 scic_sds_remote_device_event_handler(idev, completion_entry); 463 scic_sds_remote_device_event_handler(idev, completion_entry);
467 } else 464 } else
468 dev_err(scic_to_dev(scic), 465 dev_err(&ihost->pdev->dev,
469 "%s: SCIC Controller 0x%p received event 0x%x " 466 "%s: SCIC Controller 0x%p received event 0x%x "
470 "for remote device object 0x%0x that doesnt " 467 "for remote device object 0x%0x that doesnt "
471 "exist.\n", 468 "exist.\n",
472 __func__, 469 __func__,
473 scic, 470 ihost,
474 completion_entry, 471 completion_entry,
475 index); 472 index);
476 473
477 break; 474 break;
478 475
479 default: 476 default:
480 dev_warn(scic_to_dev(scic), 477 dev_warn(&ihost->pdev->dev,
481 "%s: SCIC Controller received unknown event code %x\n", 478 "%s: SCIC Controller received unknown event code %x\n",
482 __func__, 479 __func__,
483 completion_entry); 480 completion_entry);
@@ -485,7 +482,7 @@ static void scic_sds_controller_event_completion(struct scic_sds_controller *sci
485 } 482 }
486} 483}
487 484
488static void scic_sds_controller_process_completions(struct scic_sds_controller *scic) 485static void scic_sds_controller_process_completions(struct isci_host *ihost)
489{ 486{
490 u32 completion_count = 0; 487 u32 completion_count = 0;
491 u32 completion_entry; 488 u32 completion_entry;
@@ -494,47 +491,47 @@ static void scic_sds_controller_process_completions(struct scic_sds_controller *
494 u32 event_get; 491 u32 event_get;
495 u32 event_cycle; 492 u32 event_cycle;
496 493
497 dev_dbg(scic_to_dev(scic), 494 dev_dbg(&ihost->pdev->dev,
498 "%s: completion queue begining get:0x%08x\n", 495 "%s: completion queue begining get:0x%08x\n",
499 __func__, 496 __func__,
500 scic->completion_queue_get); 497 ihost->completion_queue_get);
501 498
502 /* Get the component parts of the completion queue */ 499 /* Get the component parts of the completion queue */
503 get_index = NORMALIZE_GET_POINTER(scic->completion_queue_get); 500 get_index = NORMALIZE_GET_POINTER(ihost->completion_queue_get);
504 get_cycle = SMU_CQGR_CYCLE_BIT & scic->completion_queue_get; 501 get_cycle = SMU_CQGR_CYCLE_BIT & ihost->completion_queue_get;
505 502
506 event_get = NORMALIZE_EVENT_POINTER(scic->completion_queue_get); 503 event_get = NORMALIZE_EVENT_POINTER(ihost->completion_queue_get);
507 event_cycle = SMU_CQGR_EVENT_CYCLE_BIT & scic->completion_queue_get; 504 event_cycle = SMU_CQGR_EVENT_CYCLE_BIT & ihost->completion_queue_get;
508 505
509 while ( 506 while (
510 NORMALIZE_GET_POINTER_CYCLE_BIT(get_cycle) 507 NORMALIZE_GET_POINTER_CYCLE_BIT(get_cycle)
511 == COMPLETION_QUEUE_CYCLE_BIT(scic->completion_queue[get_index]) 508 == COMPLETION_QUEUE_CYCLE_BIT(ihost->completion_queue[get_index])
512 ) { 509 ) {
513 completion_count++; 510 completion_count++;
514 511
515 completion_entry = scic->completion_queue[get_index]; 512 completion_entry = ihost->completion_queue[get_index];
516 513
517 /* increment the get pointer and check for rollover to toggle the cycle bit */ 514 /* increment the get pointer and check for rollover to toggle the cycle bit */
518 get_cycle ^= ((get_index+1) & SCU_MAX_COMPLETION_QUEUE_ENTRIES) << 515 get_cycle ^= ((get_index+1) & SCU_MAX_COMPLETION_QUEUE_ENTRIES) <<
519 (SMU_COMPLETION_QUEUE_GET_CYCLE_BIT_SHIFT - SCU_MAX_COMPLETION_QUEUE_SHIFT); 516 (SMU_COMPLETION_QUEUE_GET_CYCLE_BIT_SHIFT - SCU_MAX_COMPLETION_QUEUE_SHIFT);
520 get_index = (get_index+1) & (SCU_MAX_COMPLETION_QUEUE_ENTRIES-1); 517 get_index = (get_index+1) & (SCU_MAX_COMPLETION_QUEUE_ENTRIES-1);
521 518
522 dev_dbg(scic_to_dev(scic), 519 dev_dbg(&ihost->pdev->dev,
523 "%s: completion queue entry:0x%08x\n", 520 "%s: completion queue entry:0x%08x\n",
524 __func__, 521 __func__,
525 completion_entry); 522 completion_entry);
526 523
527 switch (SCU_GET_COMPLETION_TYPE(completion_entry)) { 524 switch (SCU_GET_COMPLETION_TYPE(completion_entry)) {
528 case SCU_COMPLETION_TYPE_TASK: 525 case SCU_COMPLETION_TYPE_TASK:
529 scic_sds_controller_task_completion(scic, completion_entry); 526 scic_sds_controller_task_completion(ihost, completion_entry);
530 break; 527 break;
531 528
532 case SCU_COMPLETION_TYPE_SDMA: 529 case SCU_COMPLETION_TYPE_SDMA:
533 scic_sds_controller_sdma_completion(scic, completion_entry); 530 scic_sds_controller_sdma_completion(ihost, completion_entry);
534 break; 531 break;
535 532
536 case SCU_COMPLETION_TYPE_UFI: 533 case SCU_COMPLETION_TYPE_UFI:
537 scic_sds_controller_unsolicited_frame(scic, completion_entry); 534 scic_sds_controller_unsolicited_frame(ihost, completion_entry);
538 break; 535 break;
539 536
540 case SCU_COMPLETION_TYPE_EVENT: 537 case SCU_COMPLETION_TYPE_EVENT:
@@ -543,11 +540,11 @@ static void scic_sds_controller_process_completions(struct scic_sds_controller *
543 (SMU_COMPLETION_QUEUE_GET_EVENT_CYCLE_BIT_SHIFT - SCU_MAX_EVENTS_SHIFT); 540 (SMU_COMPLETION_QUEUE_GET_EVENT_CYCLE_BIT_SHIFT - SCU_MAX_EVENTS_SHIFT);
544 event_get = (event_get+1) & (SCU_MAX_EVENTS-1); 541 event_get = (event_get+1) & (SCU_MAX_EVENTS-1);
545 542
546 scic_sds_controller_event_completion(scic, completion_entry); 543 scic_sds_controller_event_completion(ihost, completion_entry);
547 break; 544 break;
548 } 545 }
549 default: 546 default:
550 dev_warn(scic_to_dev(scic), 547 dev_warn(&ihost->pdev->dev,
551 "%s: SCIC Controller received unknown " 548 "%s: SCIC Controller received unknown "
552 "completion type %x\n", 549 "completion type %x\n",
553 __func__, 550 __func__,
@@ -558,7 +555,7 @@ static void scic_sds_controller_process_completions(struct scic_sds_controller *
558 555
559 /* Update the get register if we completed one or more entries */ 556 /* Update the get register if we completed one or more entries */
560 if (completion_count > 0) { 557 if (completion_count > 0) {
561 scic->completion_queue_get = 558 ihost->completion_queue_get =
562 SMU_CQGR_GEN_BIT(ENABLE) | 559 SMU_CQGR_GEN_BIT(ENABLE) |
563 SMU_CQGR_GEN_BIT(EVENT_ENABLE) | 560 SMU_CQGR_GEN_BIT(EVENT_ENABLE) |
564 event_cycle | 561 event_cycle |
@@ -566,35 +563,35 @@ static void scic_sds_controller_process_completions(struct scic_sds_controller *
566 get_cycle | 563 get_cycle |
567 SMU_CQGR_GEN_VAL(POINTER, get_index); 564 SMU_CQGR_GEN_VAL(POINTER, get_index);
568 565
569 writel(scic->completion_queue_get, 566 writel(ihost->completion_queue_get,
570 &scic->smu_registers->completion_queue_get); 567 &ihost->smu_registers->completion_queue_get);
571 568
572 } 569 }
573 570
574 dev_dbg(scic_to_dev(scic), 571 dev_dbg(&ihost->pdev->dev,
575 "%s: completion queue ending get:0x%08x\n", 572 "%s: completion queue ending get:0x%08x\n",
576 __func__, 573 __func__,
577 scic->completion_queue_get); 574 ihost->completion_queue_get);
578 575
579} 576}
580 577
581static void scic_sds_controller_error_handler(struct scic_sds_controller *scic) 578static void scic_sds_controller_error_handler(struct isci_host *ihost)
582{ 579{
583 u32 interrupt_status; 580 u32 interrupt_status;
584 581
585 interrupt_status = 582 interrupt_status =
586 readl(&scic->smu_registers->interrupt_status); 583 readl(&ihost->smu_registers->interrupt_status);
587 584
588 if ((interrupt_status & SMU_ISR_QUEUE_SUSPEND) && 585 if ((interrupt_status & SMU_ISR_QUEUE_SUSPEND) &&
589 scic_sds_controller_completion_queue_has_entries(scic)) { 586 scic_sds_controller_completion_queue_has_entries(ihost)) {
590 587
591 scic_sds_controller_process_completions(scic); 588 scic_sds_controller_process_completions(ihost);
592 writel(SMU_ISR_QUEUE_SUSPEND, &scic->smu_registers->interrupt_status); 589 writel(SMU_ISR_QUEUE_SUSPEND, &ihost->smu_registers->interrupt_status);
593 } else { 590 } else {
594 dev_err(scic_to_dev(scic), "%s: status: %#x\n", __func__, 591 dev_err(&ihost->pdev->dev, "%s: status: %#x\n", __func__,
595 interrupt_status); 592 interrupt_status);
596 593
597 sci_change_state(&scic->sm, SCIC_FAILED); 594 sci_change_state(&ihost->sm, SCIC_FAILED);
598 595
599 return; 596 return;
600 } 597 }
@@ -602,22 +599,21 @@ static void scic_sds_controller_error_handler(struct scic_sds_controller *scic)
602 /* If we dont process any completions I am not sure that we want to do this. 599 /* If we dont process any completions I am not sure that we want to do this.
603 * We are in the middle of a hardware fault and should probably be reset. 600 * We are in the middle of a hardware fault and should probably be reset.
604 */ 601 */
605 writel(0, &scic->smu_registers->interrupt_mask); 602 writel(0, &ihost->smu_registers->interrupt_mask);
606} 603}
607 604
608irqreturn_t isci_intx_isr(int vec, void *data) 605irqreturn_t isci_intx_isr(int vec, void *data)
609{ 606{
610 irqreturn_t ret = IRQ_NONE; 607 irqreturn_t ret = IRQ_NONE;
611 struct isci_host *ihost = data; 608 struct isci_host *ihost = data;
612 struct scic_sds_controller *scic = &ihost->sci;
613 609
614 if (scic_sds_controller_isr(scic)) { 610 if (scic_sds_controller_isr(ihost)) {
615 writel(SMU_ISR_COMPLETION, &scic->smu_registers->interrupt_status); 611 writel(SMU_ISR_COMPLETION, &ihost->smu_registers->interrupt_status);
616 tasklet_schedule(&ihost->completion_tasklet); 612 tasklet_schedule(&ihost->completion_tasklet);
617 ret = IRQ_HANDLED; 613 ret = IRQ_HANDLED;
618 } else if (scic_sds_controller_error_isr(scic)) { 614 } else if (scic_sds_controller_error_isr(ihost)) {
619 spin_lock(&ihost->scic_lock); 615 spin_lock(&ihost->scic_lock);
620 scic_sds_controller_error_handler(scic); 616 scic_sds_controller_error_handler(ihost);
621 spin_unlock(&ihost->scic_lock); 617 spin_unlock(&ihost->scic_lock);
622 ret = IRQ_HANDLED; 618 ret = IRQ_HANDLED;
623 } 619 }
@@ -629,8 +625,8 @@ irqreturn_t isci_error_isr(int vec, void *data)
629{ 625{
630 struct isci_host *ihost = data; 626 struct isci_host *ihost = data;
631 627
632 if (scic_sds_controller_error_isr(&ihost->sci)) 628 if (scic_sds_controller_error_isr(ihost))
633 scic_sds_controller_error_handler(&ihost->sci); 629 scic_sds_controller_error_handler(ihost);
634 630
635 return IRQ_HANDLED; 631 return IRQ_HANDLED;
636} 632}
@@ -685,11 +681,10 @@ int isci_host_scan_finished(struct Scsi_Host *shost, unsigned long time)
685 * This method returns the number of milliseconds for the suggested start 681 * This method returns the number of milliseconds for the suggested start
686 * operation timeout. 682 * operation timeout.
687 */ 683 */
688static u32 scic_controller_get_suggested_start_timeout( 684static u32 scic_controller_get_suggested_start_timeout(struct isci_host *ihost)
689 struct scic_sds_controller *sc)
690{ 685{
691 /* Validate the user supplied parameters. */ 686 /* Validate the user supplied parameters. */
692 if (sc == NULL) 687 if (!ihost)
693 return 0; 688 return 0;
694 689
695 /* 690 /*
@@ -711,35 +706,32 @@ static u32 scic_controller_get_suggested_start_timeout(
711 + ((SCI_MAX_PHYS - 1) * SCIC_SDS_CONTROLLER_POWER_CONTROL_INTERVAL); 706 + ((SCI_MAX_PHYS - 1) * SCIC_SDS_CONTROLLER_POWER_CONTROL_INTERVAL);
712} 707}
713 708
714static void scic_controller_enable_interrupts( 709static void scic_controller_enable_interrupts(struct isci_host *ihost)
715 struct scic_sds_controller *scic)
716{ 710{
717 BUG_ON(scic->smu_registers == NULL); 711 BUG_ON(ihost->smu_registers == NULL);
718 writel(0, &scic->smu_registers->interrupt_mask); 712 writel(0, &ihost->smu_registers->interrupt_mask);
719} 713}
720 714
721void scic_controller_disable_interrupts( 715void scic_controller_disable_interrupts(struct isci_host *ihost)
722 struct scic_sds_controller *scic)
723{ 716{
724 BUG_ON(scic->smu_registers == NULL); 717 BUG_ON(ihost->smu_registers == NULL);
725 writel(0xffffffff, &scic->smu_registers->interrupt_mask); 718 writel(0xffffffff, &ihost->smu_registers->interrupt_mask);
726} 719}
727 720
728static void scic_sds_controller_enable_port_task_scheduler( 721static void scic_sds_controller_enable_port_task_scheduler(struct isci_host *ihost)
729 struct scic_sds_controller *scic)
730{ 722{
731 u32 port_task_scheduler_value; 723 u32 port_task_scheduler_value;
732 724
733 port_task_scheduler_value = 725 port_task_scheduler_value =
734 readl(&scic->scu_registers->peg0.ptsg.control); 726 readl(&ihost->scu_registers->peg0.ptsg.control);
735 port_task_scheduler_value |= 727 port_task_scheduler_value |=
736 (SCU_PTSGCR_GEN_BIT(ETM_ENABLE) | 728 (SCU_PTSGCR_GEN_BIT(ETM_ENABLE) |
737 SCU_PTSGCR_GEN_BIT(PTSG_ENABLE)); 729 SCU_PTSGCR_GEN_BIT(PTSG_ENABLE));
738 writel(port_task_scheduler_value, 730 writel(port_task_scheduler_value,
739 &scic->scu_registers->peg0.ptsg.control); 731 &ihost->scu_registers->peg0.ptsg.control);
740} 732}
741 733
742static void scic_sds_controller_assign_task_entries(struct scic_sds_controller *scic) 734static void scic_sds_controller_assign_task_entries(struct isci_host *ihost)
743{ 735{
744 u32 task_assignment; 736 u32 task_assignment;
745 737
@@ -749,32 +741,32 @@ static void scic_sds_controller_assign_task_entries(struct scic_sds_controller *
749 */ 741 */
750 742
751 task_assignment = 743 task_assignment =
752 readl(&scic->smu_registers->task_context_assignment[0]); 744 readl(&ihost->smu_registers->task_context_assignment[0]);
753 745
754 task_assignment |= (SMU_TCA_GEN_VAL(STARTING, 0)) | 746 task_assignment |= (SMU_TCA_GEN_VAL(STARTING, 0)) |
755 (SMU_TCA_GEN_VAL(ENDING, scic->task_context_entries - 1)) | 747 (SMU_TCA_GEN_VAL(ENDING, ihost->task_context_entries - 1)) |
756 (SMU_TCA_GEN_BIT(RANGE_CHECK_ENABLE)); 748 (SMU_TCA_GEN_BIT(RANGE_CHECK_ENABLE));
757 749
758 writel(task_assignment, 750 writel(task_assignment,
759 &scic->smu_registers->task_context_assignment[0]); 751 &ihost->smu_registers->task_context_assignment[0]);
760 752
761} 753}
762 754
763static void scic_sds_controller_initialize_completion_queue(struct scic_sds_controller *scic) 755static void scic_sds_controller_initialize_completion_queue(struct isci_host *ihost)
764{ 756{
765 u32 index; 757 u32 index;
766 u32 completion_queue_control_value; 758 u32 completion_queue_control_value;
767 u32 completion_queue_get_value; 759 u32 completion_queue_get_value;
768 u32 completion_queue_put_value; 760 u32 completion_queue_put_value;
769 761
770 scic->completion_queue_get = 0; 762 ihost->completion_queue_get = 0;
771 763
772 completion_queue_control_value = 764 completion_queue_control_value =
773 (SMU_CQC_QUEUE_LIMIT_SET(SCU_MAX_COMPLETION_QUEUE_ENTRIES - 1) | 765 (SMU_CQC_QUEUE_LIMIT_SET(SCU_MAX_COMPLETION_QUEUE_ENTRIES - 1) |
774 SMU_CQC_EVENT_LIMIT_SET(SCU_MAX_EVENTS - 1)); 766 SMU_CQC_EVENT_LIMIT_SET(SCU_MAX_EVENTS - 1));
775 767
776 writel(completion_queue_control_value, 768 writel(completion_queue_control_value,
777 &scic->smu_registers->completion_queue_control); 769 &ihost->smu_registers->completion_queue_control);
778 770
779 771
780 /* Set the completion queue get pointer and enable the queue */ 772 /* Set the completion queue get pointer and enable the queue */
@@ -786,7 +778,7 @@ static void scic_sds_controller_initialize_completion_queue(struct scic_sds_cont
786 ); 778 );
787 779
788 writel(completion_queue_get_value, 780 writel(completion_queue_get_value,
789 &scic->smu_registers->completion_queue_get); 781 &ihost->smu_registers->completion_queue_get);
790 782
791 /* Set the completion queue put pointer */ 783 /* Set the completion queue put pointer */
792 completion_queue_put_value = ( 784 completion_queue_put_value = (
@@ -795,7 +787,7 @@ static void scic_sds_controller_initialize_completion_queue(struct scic_sds_cont
795 ); 787 );
796 788
797 writel(completion_queue_put_value, 789 writel(completion_queue_put_value,
798 &scic->smu_registers->completion_queue_put); 790 &ihost->smu_registers->completion_queue_put);
799 791
800 /* Initialize the cycle bit of the completion queue entries */ 792 /* Initialize the cycle bit of the completion queue entries */
801 for (index = 0; index < SCU_MAX_COMPLETION_QUEUE_ENTRIES; index++) { 793 for (index = 0; index < SCU_MAX_COMPLETION_QUEUE_ENTRIES; index++) {
@@ -803,11 +795,11 @@ static void scic_sds_controller_initialize_completion_queue(struct scic_sds_cont
803 * If get.cycle_bit != completion_queue.cycle_bit 795 * If get.cycle_bit != completion_queue.cycle_bit
804 * its not a valid completion queue entry 796 * its not a valid completion queue entry
805 * so at system start all entries are invalid */ 797 * so at system start all entries are invalid */
806 scic->completion_queue[index] = 0x80000000; 798 ihost->completion_queue[index] = 0x80000000;
807 } 799 }
808} 800}
809 801
810static void scic_sds_controller_initialize_unsolicited_frame_queue(struct scic_sds_controller *scic) 802static void scic_sds_controller_initialize_unsolicited_frame_queue(struct isci_host *ihost)
811{ 803{
812 u32 frame_queue_control_value; 804 u32 frame_queue_control_value;
813 u32 frame_queue_get_value; 805 u32 frame_queue_get_value;
@@ -818,7 +810,7 @@ static void scic_sds_controller_initialize_unsolicited_frame_queue(struct scic_s
818 SCU_UFQC_GEN_VAL(QUEUE_SIZE, SCU_MAX_UNSOLICITED_FRAMES); 810 SCU_UFQC_GEN_VAL(QUEUE_SIZE, SCU_MAX_UNSOLICITED_FRAMES);
819 811
820 writel(frame_queue_control_value, 812 writel(frame_queue_control_value,
821 &scic->scu_registers->sdma.unsolicited_frame_queue_control); 813 &ihost->scu_registers->sdma.unsolicited_frame_queue_control);
822 814
823 /* Setup the get pointer for the unsolicited frame queue */ 815 /* Setup the get pointer for the unsolicited frame queue */
824 frame_queue_get_value = ( 816 frame_queue_get_value = (
@@ -827,11 +819,11 @@ static void scic_sds_controller_initialize_unsolicited_frame_queue(struct scic_s
827 ); 819 );
828 820
829 writel(frame_queue_get_value, 821 writel(frame_queue_get_value,
830 &scic->scu_registers->sdma.unsolicited_frame_get_pointer); 822 &ihost->scu_registers->sdma.unsolicited_frame_get_pointer);
831 /* Setup the put pointer for the unsolicited frame queue */ 823 /* Setup the put pointer for the unsolicited frame queue */
832 frame_queue_put_value = SCU_UFQPP_GEN_VAL(POINTER, 0); 824 frame_queue_put_value = SCU_UFQPP_GEN_VAL(POINTER, 0);
833 writel(frame_queue_put_value, 825 writel(frame_queue_put_value,
834 &scic->scu_registers->sdma.unsolicited_frame_put_pointer); 826 &ihost->scu_registers->sdma.unsolicited_frame_put_pointer);
835} 827}
836 828
837/** 829/**
@@ -846,17 +838,16 @@ static void scic_sds_controller_initialize_unsolicited_frame_queue(struct scic_s
846 * none. 838 * none.
847 */ 839 */
848static void scic_sds_controller_transition_to_ready( 840static void scic_sds_controller_transition_to_ready(
849 struct scic_sds_controller *scic, 841 struct isci_host *ihost,
850 enum sci_status status) 842 enum sci_status status)
851{ 843{
852 struct isci_host *ihost = scic_to_ihost(scic);
853 844
854 if (scic->sm.current_state_id == SCIC_STARTING) { 845 if (ihost->sm.current_state_id == SCIC_STARTING) {
855 /* 846 /*
856 * We move into the ready state, because some of the phys/ports 847 * We move into the ready state, because some of the phys/ports
857 * may be up and operational. 848 * may be up and operational.
858 */ 849 */
859 sci_change_state(&scic->sm, SCIC_READY); 850 sci_change_state(&ihost->sm, SCIC_READY);
860 851
861 isci_host_start_complete(ihost, status); 852 isci_host_start_complete(ihost, status);
862 } 853 }
@@ -892,19 +883,18 @@ static bool is_phy_starting(struct isci_phy *iphy)
892 * controller to the READY state and inform the user 883 * controller to the READY state and inform the user
893 * (scic_cb_controller_start_complete()). 884 * (scic_cb_controller_start_complete()).
894 */ 885 */
895static enum sci_status scic_sds_controller_start_next_phy(struct scic_sds_controller *scic) 886static enum sci_status scic_sds_controller_start_next_phy(struct isci_host *ihost)
896{ 887{
897 struct isci_host *ihost = scic_to_ihost(scic); 888 struct scic_sds_oem_params *oem = &ihost->oem_parameters.sds1;
898 struct scic_sds_oem_params *oem = &scic->oem_parameters.sds1;
899 struct isci_phy *iphy; 889 struct isci_phy *iphy;
900 enum sci_status status; 890 enum sci_status status;
901 891
902 status = SCI_SUCCESS; 892 status = SCI_SUCCESS;
903 893
904 if (scic->phy_startup_timer_pending) 894 if (ihost->phy_startup_timer_pending)
905 return status; 895 return status;
906 896
907 if (scic->next_phy_to_start >= SCI_MAX_PHYS) { 897 if (ihost->next_phy_to_start >= SCI_MAX_PHYS) {
908 bool is_controller_start_complete = true; 898 bool is_controller_start_complete = true;
909 u32 state; 899 u32 state;
910 u8 index; 900 u8 index;
@@ -934,16 +924,16 @@ static enum sci_status scic_sds_controller_start_next_phy(struct scic_sds_contro
934 * The controller has successfully finished the start process. 924 * The controller has successfully finished the start process.
935 * Inform the SCI Core user and transition to the READY state. */ 925 * Inform the SCI Core user and transition to the READY state. */
936 if (is_controller_start_complete == true) { 926 if (is_controller_start_complete == true) {
937 scic_sds_controller_transition_to_ready(scic, SCI_SUCCESS); 927 scic_sds_controller_transition_to_ready(ihost, SCI_SUCCESS);
938 sci_del_timer(&scic->phy_timer); 928 sci_del_timer(&ihost->phy_timer);
939 scic->phy_startup_timer_pending = false; 929 ihost->phy_startup_timer_pending = false;
940 } 930 }
941 } else { 931 } else {
942 iphy = &ihost->phys[scic->next_phy_to_start]; 932 iphy = &ihost->phys[ihost->next_phy_to_start];
943 933
944 if (oem->controller.mode_type == SCIC_PORT_MANUAL_CONFIGURATION_MODE) { 934 if (oem->controller.mode_type == SCIC_PORT_MANUAL_CONFIGURATION_MODE) {
945 if (phy_get_non_dummy_port(iphy) == NULL) { 935 if (phy_get_non_dummy_port(iphy) == NULL) {
946 scic->next_phy_to_start++; 936 ihost->next_phy_to_start++;
947 937
948 /* Caution recursion ahead be forwarned 938 /* Caution recursion ahead be forwarned
949 * 939 *
@@ -954,27 +944,27 @@ static enum sci_status scic_sds_controller_start_next_phy(struct scic_sds_contro
954 * incorrectly for the PORT or it was never 944 * incorrectly for the PORT or it was never
955 * assigned to a PORT 945 * assigned to a PORT
956 */ 946 */
957 return scic_sds_controller_start_next_phy(scic); 947 return scic_sds_controller_start_next_phy(ihost);
958 } 948 }
959 } 949 }
960 950
961 status = scic_sds_phy_start(iphy); 951 status = scic_sds_phy_start(iphy);
962 952
963 if (status == SCI_SUCCESS) { 953 if (status == SCI_SUCCESS) {
964 sci_mod_timer(&scic->phy_timer, 954 sci_mod_timer(&ihost->phy_timer,
965 SCIC_SDS_CONTROLLER_PHY_START_TIMEOUT); 955 SCIC_SDS_CONTROLLER_PHY_START_TIMEOUT);
966 scic->phy_startup_timer_pending = true; 956 ihost->phy_startup_timer_pending = true;
967 } else { 957 } else {
968 dev_warn(scic_to_dev(scic), 958 dev_warn(&ihost->pdev->dev,
969 "%s: Controller stop operation failed " 959 "%s: Controller stop operation failed "
970 "to stop phy %d because of status " 960 "to stop phy %d because of status "
971 "%d.\n", 961 "%d.\n",
972 __func__, 962 __func__,
973 ihost->phys[scic->next_phy_to_start].phy_index, 963 ihost->phys[ihost->next_phy_to_start].phy_index,
974 status); 964 status);
975 } 965 }
976 966
977 scic->next_phy_to_start++; 967 ihost->next_phy_to_start++;
978 } 968 }
979 969
980 return status; 970 return status;
@@ -983,8 +973,7 @@ static enum sci_status scic_sds_controller_start_next_phy(struct scic_sds_contro
983static void phy_startup_timeout(unsigned long data) 973static void phy_startup_timeout(unsigned long data)
984{ 974{
985 struct sci_timer *tmr = (struct sci_timer *)data; 975 struct sci_timer *tmr = (struct sci_timer *)data;
986 struct scic_sds_controller *scic = container_of(tmr, typeof(*scic), phy_timer); 976 struct isci_host *ihost = container_of(tmr, typeof(*ihost), phy_timer);
987 struct isci_host *ihost = scic_to_ihost(scic);
988 unsigned long flags; 977 unsigned long flags;
989 enum sci_status status; 978 enum sci_status status;
990 979
@@ -993,10 +982,10 @@ static void phy_startup_timeout(unsigned long data)
993 if (tmr->cancel) 982 if (tmr->cancel)
994 goto done; 983 goto done;
995 984
996 scic->phy_startup_timer_pending = false; 985 ihost->phy_startup_timer_pending = false;
997 986
998 do { 987 do {
999 status = scic_sds_controller_start_next_phy(scic); 988 status = scic_sds_controller_start_next_phy(ihost);
1000 } while (status != SCI_SUCCESS); 989 } while (status != SCI_SUCCESS);
1001 990
1002done: 991done:
@@ -1008,15 +997,14 @@ static u16 isci_tci_active(struct isci_host *ihost)
1008 return CIRC_CNT(ihost->tci_head, ihost->tci_tail, SCI_MAX_IO_REQUESTS); 997 return CIRC_CNT(ihost->tci_head, ihost->tci_tail, SCI_MAX_IO_REQUESTS);
1009} 998}
1010 999
1011static enum sci_status scic_controller_start(struct scic_sds_controller *scic, 1000static enum sci_status scic_controller_start(struct isci_host *ihost,
1012 u32 timeout) 1001 u32 timeout)
1013{ 1002{
1014 struct isci_host *ihost = scic_to_ihost(scic);
1015 enum sci_status result; 1003 enum sci_status result;
1016 u16 index; 1004 u16 index;
1017 1005
1018 if (scic->sm.current_state_id != SCIC_INITIALIZED) { 1006 if (ihost->sm.current_state_id != SCIC_INITIALIZED) {
1019 dev_warn(scic_to_dev(scic), 1007 dev_warn(&ihost->pdev->dev,
1020 "SCIC Controller start operation requested in " 1008 "SCIC Controller start operation requested in "
1021 "invalid state\n"); 1009 "invalid state\n");
1022 return SCI_FAILURE_INVALID_STATE; 1010 return SCI_FAILURE_INVALID_STATE;
@@ -1026,34 +1014,34 @@ static enum sci_status scic_controller_start(struct scic_sds_controller *scic,
1026 BUILD_BUG_ON(SCI_MAX_IO_REQUESTS > 1 << sizeof(ihost->tci_pool[0]) * 8); 1014 BUILD_BUG_ON(SCI_MAX_IO_REQUESTS > 1 << sizeof(ihost->tci_pool[0]) * 8);
1027 ihost->tci_head = 0; 1015 ihost->tci_head = 0;
1028 ihost->tci_tail = 0; 1016 ihost->tci_tail = 0;
1029 for (index = 0; index < scic->task_context_entries; index++) 1017 for (index = 0; index < ihost->task_context_entries; index++)
1030 isci_tci_free(ihost, index); 1018 isci_tci_free(ihost, index);
1031 1019
1032 /* Build the RNi free pool */ 1020 /* Build the RNi free pool */
1033 scic_sds_remote_node_table_initialize( 1021 scic_sds_remote_node_table_initialize(
1034 &scic->available_remote_nodes, 1022 &ihost->available_remote_nodes,
1035 scic->remote_node_entries); 1023 ihost->remote_node_entries);
1036 1024
1037 /* 1025 /*
1038 * Before anything else lets make sure we will not be 1026 * Before anything else lets make sure we will not be
1039 * interrupted by the hardware. 1027 * interrupted by the hardware.
1040 */ 1028 */
1041 scic_controller_disable_interrupts(scic); 1029 scic_controller_disable_interrupts(ihost);
1042 1030
1043 /* Enable the port task scheduler */ 1031 /* Enable the port task scheduler */
1044 scic_sds_controller_enable_port_task_scheduler(scic); 1032 scic_sds_controller_enable_port_task_scheduler(ihost);
1045 1033
1046 /* Assign all the task entries to scic physical function */ 1034 /* Assign all the task entries to ihost physical function */
1047 scic_sds_controller_assign_task_entries(scic); 1035 scic_sds_controller_assign_task_entries(ihost);
1048 1036
1049 /* Now initialize the completion queue */ 1037 /* Now initialize the completion queue */
1050 scic_sds_controller_initialize_completion_queue(scic); 1038 scic_sds_controller_initialize_completion_queue(ihost);
1051 1039
1052 /* Initialize the unsolicited frame queue for use */ 1040 /* Initialize the unsolicited frame queue for use */
1053 scic_sds_controller_initialize_unsolicited_frame_queue(scic); 1041 scic_sds_controller_initialize_unsolicited_frame_queue(ihost);
1054 1042
1055 /* Start all of the ports on this controller */ 1043 /* Start all of the ports on this controller */
1056 for (index = 0; index < scic->logical_port_entries; index++) { 1044 for (index = 0; index < ihost->logical_port_entries; index++) {
1057 struct isci_port *iport = &ihost->ports[index]; 1045 struct isci_port *iport = &ihost->ports[index];
1058 1046
1059 result = scic_sds_port_start(iport); 1047 result = scic_sds_port_start(iport);
@@ -1061,11 +1049,11 @@ static enum sci_status scic_controller_start(struct scic_sds_controller *scic,
1061 return result; 1049 return result;
1062 } 1050 }
1063 1051
1064 scic_sds_controller_start_next_phy(scic); 1052 scic_sds_controller_start_next_phy(ihost);
1065 1053
1066 sci_mod_timer(&scic->timer, timeout); 1054 sci_mod_timer(&ihost->timer, timeout);
1067 1055
1068 sci_change_state(&scic->sm, SCIC_STARTING); 1056 sci_change_state(&ihost->sm, SCIC_STARTING);
1069 1057
1070 return SCI_SUCCESS; 1058 return SCI_SUCCESS;
1071} 1059}
@@ -1073,35 +1061,35 @@ static enum sci_status scic_controller_start(struct scic_sds_controller *scic,
1073void isci_host_scan_start(struct Scsi_Host *shost) 1061void isci_host_scan_start(struct Scsi_Host *shost)
1074{ 1062{
1075 struct isci_host *ihost = SHOST_TO_SAS_HA(shost)->lldd_ha; 1063 struct isci_host *ihost = SHOST_TO_SAS_HA(shost)->lldd_ha;
1076 unsigned long tmo = scic_controller_get_suggested_start_timeout(&ihost->sci); 1064 unsigned long tmo = scic_controller_get_suggested_start_timeout(ihost);
1077 1065
1078 set_bit(IHOST_START_PENDING, &ihost->flags); 1066 set_bit(IHOST_START_PENDING, &ihost->flags);
1079 1067
1080 spin_lock_irq(&ihost->scic_lock); 1068 spin_lock_irq(&ihost->scic_lock);
1081 scic_controller_start(&ihost->sci, tmo); 1069 scic_controller_start(ihost, tmo);
1082 scic_controller_enable_interrupts(&ihost->sci); 1070 scic_controller_enable_interrupts(ihost);
1083 spin_unlock_irq(&ihost->scic_lock); 1071 spin_unlock_irq(&ihost->scic_lock);
1084} 1072}
1085 1073
1086static void isci_host_stop_complete(struct isci_host *ihost, enum sci_status completion_status) 1074static void isci_host_stop_complete(struct isci_host *ihost, enum sci_status completion_status)
1087{ 1075{
1088 isci_host_change_state(ihost, isci_stopped); 1076 isci_host_change_state(ihost, isci_stopped);
1089 scic_controller_disable_interrupts(&ihost->sci); 1077 scic_controller_disable_interrupts(ihost);
1090 clear_bit(IHOST_STOP_PENDING, &ihost->flags); 1078 clear_bit(IHOST_STOP_PENDING, &ihost->flags);
1091 wake_up(&ihost->eventq); 1079 wake_up(&ihost->eventq);
1092} 1080}
1093 1081
1094static void scic_sds_controller_completion_handler(struct scic_sds_controller *scic) 1082static void scic_sds_controller_completion_handler(struct isci_host *ihost)
1095{ 1083{
1096 /* Empty out the completion queue */ 1084 /* Empty out the completion queue */
1097 if (scic_sds_controller_completion_queue_has_entries(scic)) 1085 if (scic_sds_controller_completion_queue_has_entries(ihost))
1098 scic_sds_controller_process_completions(scic); 1086 scic_sds_controller_process_completions(ihost);
1099 1087
1100 /* Clear the interrupt and enable all interrupts again */ 1088 /* Clear the interrupt and enable all interrupts again */
1101 writel(SMU_ISR_COMPLETION, &scic->smu_registers->interrupt_status); 1089 writel(SMU_ISR_COMPLETION, &ihost->smu_registers->interrupt_status);
1102 /* Could we write the value of SMU_ISR_COMPLETION? */ 1090 /* Could we write the value of SMU_ISR_COMPLETION? */
1103 writel(0xFF000000, &scic->smu_registers->interrupt_mask); 1091 writel(0xFF000000, &ihost->smu_registers->interrupt_mask);
1104 writel(0, &scic->smu_registers->interrupt_mask); 1092 writel(0, &ihost->smu_registers->interrupt_mask);
1105} 1093}
1106 1094
1107/** 1095/**
@@ -1114,7 +1102,7 @@ static void scic_sds_controller_completion_handler(struct scic_sds_controller *s
1114 */ 1102 */
1115static void isci_host_completion_routine(unsigned long data) 1103static void isci_host_completion_routine(unsigned long data)
1116{ 1104{
1117 struct isci_host *isci_host = (struct isci_host *)data; 1105 struct isci_host *ihost = (struct isci_host *)data;
1118 struct list_head completed_request_list; 1106 struct list_head completed_request_list;
1119 struct list_head errored_request_list; 1107 struct list_head errored_request_list;
1120 struct list_head *current_position; 1108 struct list_head *current_position;
@@ -1126,20 +1114,20 @@ static void isci_host_completion_routine(unsigned long data)
1126 INIT_LIST_HEAD(&completed_request_list); 1114 INIT_LIST_HEAD(&completed_request_list);
1127 INIT_LIST_HEAD(&errored_request_list); 1115 INIT_LIST_HEAD(&errored_request_list);
1128 1116
1129 spin_lock_irq(&isci_host->scic_lock); 1117 spin_lock_irq(&ihost->scic_lock);
1130 1118
1131 scic_sds_controller_completion_handler(&isci_host->sci); 1119 scic_sds_controller_completion_handler(ihost);
1132 1120
1133 /* Take the lists of completed I/Os from the host. */ 1121 /* Take the lists of completed I/Os from the host. */
1134 1122
1135 list_splice_init(&isci_host->requests_to_complete, 1123 list_splice_init(&ihost->requests_to_complete,
1136 &completed_request_list); 1124 &completed_request_list);
1137 1125
1138 /* Take the list of errored I/Os from the host. */ 1126 /* Take the list of errored I/Os from the host. */
1139 list_splice_init(&isci_host->requests_to_errorback, 1127 list_splice_init(&ihost->requests_to_errorback,
1140 &errored_request_list); 1128 &errored_request_list);
1141 1129
1142 spin_unlock_irq(&isci_host->scic_lock); 1130 spin_unlock_irq(&ihost->scic_lock);
1143 1131
1144 /* Process any completions in the lists. */ 1132 /* Process any completions in the lists. */
1145 list_for_each_safe(current_position, next_position, 1133 list_for_each_safe(current_position, next_position,
@@ -1150,7 +1138,7 @@ static void isci_host_completion_routine(unsigned long data)
1150 task = isci_request_access_task(request); 1138 task = isci_request_access_task(request);
1151 1139
1152 /* Normal notification (task_done) */ 1140 /* Normal notification (task_done) */
1153 dev_dbg(&isci_host->pdev->dev, 1141 dev_dbg(&ihost->pdev->dev,
1154 "%s: Normal - request/task = %p/%p\n", 1142 "%s: Normal - request/task = %p/%p\n",
1155 __func__, 1143 __func__,
1156 request, 1144 request,
@@ -1169,9 +1157,9 @@ static void isci_host_completion_routine(unsigned long data)
1169 } 1157 }
1170 } 1158 }
1171 1159
1172 spin_lock_irq(&isci_host->scic_lock); 1160 spin_lock_irq(&ihost->scic_lock);
1173 isci_free_tag(isci_host, request->io_tag); 1161 isci_free_tag(ihost, request->io_tag);
1174 spin_unlock_irq(&isci_host->scic_lock); 1162 spin_unlock_irq(&ihost->scic_lock);
1175 } 1163 }
1176 list_for_each_entry_safe(request, next_request, &errored_request_list, 1164 list_for_each_entry_safe(request, next_request, &errored_request_list,
1177 completed_node) { 1165 completed_node) {
@@ -1179,7 +1167,7 @@ static void isci_host_completion_routine(unsigned long data)
1179 task = isci_request_access_task(request); 1167 task = isci_request_access_task(request);
1180 1168
1181 /* Use sas_task_abort */ 1169 /* Use sas_task_abort */
1182 dev_warn(&isci_host->pdev->dev, 1170 dev_warn(&ihost->pdev->dev,
1183 "%s: Error - request/task = %p/%p\n", 1171 "%s: Error - request/task = %p/%p\n",
1184 __func__, 1172 __func__,
1185 request, 1173 request,
@@ -1202,13 +1190,13 @@ static void isci_host_completion_routine(unsigned long data)
1202 * it. 1190 * it.
1203 */ 1191 */
1204 1192
1205 spin_lock_irq(&isci_host->scic_lock); 1193 spin_lock_irq(&ihost->scic_lock);
1206 /* Remove the request from the remote device's list 1194 /* Remove the request from the remote device's list
1207 * of pending requests. 1195 * of pending requests.
1208 */ 1196 */
1209 list_del_init(&request->dev_node); 1197 list_del_init(&request->dev_node);
1210 isci_free_tag(isci_host, request->io_tag); 1198 isci_free_tag(ihost, request->io_tag);
1211 spin_unlock_irq(&isci_host->scic_lock); 1199 spin_unlock_irq(&ihost->scic_lock);
1212 } 1200 }
1213 } 1201 }
1214 1202
@@ -1232,18 +1220,18 @@ static void isci_host_completion_routine(unsigned long data)
1232 * controller is already in the STOPPED state. SCI_FAILURE_INVALID_STATE if the 1220 * controller is already in the STOPPED state. SCI_FAILURE_INVALID_STATE if the
1233 * controller is not either in the STARTED or STOPPED states. 1221 * controller is not either in the STARTED or STOPPED states.
1234 */ 1222 */
1235static enum sci_status scic_controller_stop(struct scic_sds_controller *scic, 1223static enum sci_status scic_controller_stop(struct isci_host *ihost,
1236 u32 timeout) 1224 u32 timeout)
1237{ 1225{
1238 if (scic->sm.current_state_id != SCIC_READY) { 1226 if (ihost->sm.current_state_id != SCIC_READY) {
1239 dev_warn(scic_to_dev(scic), 1227 dev_warn(&ihost->pdev->dev,
1240 "SCIC Controller stop operation requested in " 1228 "SCIC Controller stop operation requested in "
1241 "invalid state\n"); 1229 "invalid state\n");
1242 return SCI_FAILURE_INVALID_STATE; 1230 return SCI_FAILURE_INVALID_STATE;
1243 } 1231 }
1244 1232
1245 sci_mod_timer(&scic->timer, timeout); 1233 sci_mod_timer(&ihost->timer, timeout);
1246 sci_change_state(&scic->sm, SCIC_STOPPING); 1234 sci_change_state(&ihost->sm, SCIC_STOPPING);
1247 return SCI_SUCCESS; 1235 return SCI_SUCCESS;
1248} 1236}
1249 1237
@@ -1259,9 +1247,9 @@ static enum sci_status scic_controller_stop(struct scic_sds_controller *scic,
1259 * SCI_SUCCESS if the reset operation successfully started. SCI_FATAL_ERROR if 1247 * SCI_SUCCESS if the reset operation successfully started. SCI_FATAL_ERROR if
1260 * the controller reset operation is unable to complete. 1248 * the controller reset operation is unable to complete.
1261 */ 1249 */
1262static enum sci_status scic_controller_reset(struct scic_sds_controller *scic) 1250static enum sci_status scic_controller_reset(struct isci_host *ihost)
1263{ 1251{
1264 switch (scic->sm.current_state_id) { 1252 switch (ihost->sm.current_state_id) {
1265 case SCIC_RESET: 1253 case SCIC_RESET:
1266 case SCIC_READY: 1254 case SCIC_READY:
1267 case SCIC_STOPPED: 1255 case SCIC_STOPPED:
@@ -1270,10 +1258,10 @@ static enum sci_status scic_controller_reset(struct scic_sds_controller *scic)
1270 * The reset operation is not a graceful cleanup, just 1258 * The reset operation is not a graceful cleanup, just
1271 * perform the state transition. 1259 * perform the state transition.
1272 */ 1260 */
1273 sci_change_state(&scic->sm, SCIC_RESETTING); 1261 sci_change_state(&ihost->sm, SCIC_RESETTING);
1274 return SCI_SUCCESS; 1262 return SCI_SUCCESS;
1275 default: 1263 default:
1276 dev_warn(scic_to_dev(scic), 1264 dev_warn(&ihost->pdev->dev,
1277 "SCIC Controller reset operation requested in " 1265 "SCIC Controller reset operation requested in "
1278 "invalid state\n"); 1266 "invalid state\n");
1279 return SCI_FAILURE_INVALID_STATE; 1267 return SCI_FAILURE_INVALID_STATE;
@@ -1298,14 +1286,14 @@ void isci_host_deinit(struct isci_host *ihost)
1298 set_bit(IHOST_STOP_PENDING, &ihost->flags); 1286 set_bit(IHOST_STOP_PENDING, &ihost->flags);
1299 1287
1300 spin_lock_irq(&ihost->scic_lock); 1288 spin_lock_irq(&ihost->scic_lock);
1301 scic_controller_stop(&ihost->sci, SCIC_CONTROLLER_STOP_TIMEOUT); 1289 scic_controller_stop(ihost, SCIC_CONTROLLER_STOP_TIMEOUT);
1302 spin_unlock_irq(&ihost->scic_lock); 1290 spin_unlock_irq(&ihost->scic_lock);
1303 1291
1304 wait_for_stop(ihost); 1292 wait_for_stop(ihost);
1305 scic_controller_reset(&ihost->sci); 1293 scic_controller_reset(ihost);
1306 1294
1307 /* Cancel any/all outstanding port timers */ 1295 /* Cancel any/all outstanding port timers */
1308 for (i = 0; i < ihost->sci.logical_port_entries; i++) { 1296 for (i = 0; i < ihost->logical_port_entries; i++) {
1309 struct isci_port *iport = &ihost->ports[i]; 1297 struct isci_port *iport = &ihost->ports[i];
1310 del_timer_sync(&iport->timer.timer); 1298 del_timer_sync(&iport->timer.timer);
1311 } 1299 }
@@ -1316,13 +1304,13 @@ void isci_host_deinit(struct isci_host *ihost)
1316 del_timer_sync(&iphy->sata_timer.timer); 1304 del_timer_sync(&iphy->sata_timer.timer);
1317 } 1305 }
1318 1306
1319 del_timer_sync(&ihost->sci.port_agent.timer.timer); 1307 del_timer_sync(&ihost->port_agent.timer.timer);
1320 1308
1321 del_timer_sync(&ihost->sci.power_control.timer.timer); 1309 del_timer_sync(&ihost->power_control.timer.timer);
1322 1310
1323 del_timer_sync(&ihost->sci.timer.timer); 1311 del_timer_sync(&ihost->timer.timer);
1324 1312
1325 del_timer_sync(&ihost->sci.phy_timer.timer); 1313 del_timer_sync(&ihost->phy_timer.timer);
1326} 1314}
1327 1315
1328static void __iomem *scu_base(struct isci_host *isci_host) 1316static void __iomem *scu_base(struct isci_host *isci_host)
@@ -1369,16 +1357,16 @@ static void isci_user_parameters_get(
1369 1357
1370static void scic_sds_controller_initial_state_enter(struct sci_base_state_machine *sm) 1358static void scic_sds_controller_initial_state_enter(struct sci_base_state_machine *sm)
1371{ 1359{
1372 struct scic_sds_controller *scic = container_of(sm, typeof(*scic), sm); 1360 struct isci_host *ihost = container_of(sm, typeof(*ihost), sm);
1373 1361
1374 sci_change_state(&scic->sm, SCIC_RESET); 1362 sci_change_state(&ihost->sm, SCIC_RESET);
1375} 1363}
1376 1364
1377static inline void scic_sds_controller_starting_state_exit(struct sci_base_state_machine *sm) 1365static inline void scic_sds_controller_starting_state_exit(struct sci_base_state_machine *sm)
1378{ 1366{
1379 struct scic_sds_controller *scic = container_of(sm, typeof(*scic), sm); 1367 struct isci_host *ihost = container_of(sm, typeof(*ihost), sm);
1380 1368
1381 sci_del_timer(&scic->timer); 1369 sci_del_timer(&ihost->timer);
1382} 1370}
1383 1371
1384#define INTERRUPT_COALESCE_TIMEOUT_BASE_RANGE_LOWER_BOUND_NS 853 1372#define INTERRUPT_COALESCE_TIMEOUT_BASE_RANGE_LOWER_BOUND_NS 853
@@ -1405,10 +1393,10 @@ static inline void scic_sds_controller_starting_state_exit(struct sci_base_state
1405 * SCI_SUCCESS The user successfully updated the interrutp coalescence. 1393 * SCI_SUCCESS The user successfully updated the interrutp coalescence.
1406 * SCI_FAILURE_INVALID_PARAMETER_VALUE The user input value is out of range. 1394 * SCI_FAILURE_INVALID_PARAMETER_VALUE The user input value is out of range.
1407 */ 1395 */
1408static enum sci_status scic_controller_set_interrupt_coalescence( 1396static enum sci_status
1409 struct scic_sds_controller *scic_controller, 1397scic_controller_set_interrupt_coalescence(struct isci_host *ihost,
1410 u32 coalesce_number, 1398 u32 coalesce_number,
1411 u32 coalesce_timeout) 1399 u32 coalesce_timeout)
1412{ 1400{
1413 u8 timeout_encode = 0; 1401 u8 timeout_encode = 0;
1414 u32 min = 0; 1402 u32 min = 0;
@@ -1491,11 +1479,11 @@ static enum sci_status scic_controller_set_interrupt_coalescence(
1491 1479
1492 writel(SMU_ICC_GEN_VAL(NUMBER, coalesce_number) | 1480 writel(SMU_ICC_GEN_VAL(NUMBER, coalesce_number) |
1493 SMU_ICC_GEN_VAL(TIMER, timeout_encode), 1481 SMU_ICC_GEN_VAL(TIMER, timeout_encode),
1494 &scic_controller->smu_registers->interrupt_coalesce_control); 1482 &ihost->smu_registers->interrupt_coalesce_control);
1495 1483
1496 1484
1497 scic_controller->interrupt_coalesce_number = (u16)coalesce_number; 1485 ihost->interrupt_coalesce_number = (u16)coalesce_number;
1498 scic_controller->interrupt_coalesce_timeout = coalesce_timeout / 100; 1486 ihost->interrupt_coalesce_timeout = coalesce_timeout / 100;
1499 1487
1500 return SCI_SUCCESS; 1488 return SCI_SUCCESS;
1501} 1489}
@@ -1503,26 +1491,25 @@ static enum sci_status scic_controller_set_interrupt_coalescence(
1503 1491
1504static void scic_sds_controller_ready_state_enter(struct sci_base_state_machine *sm) 1492static void scic_sds_controller_ready_state_enter(struct sci_base_state_machine *sm)
1505{ 1493{
1506 struct scic_sds_controller *scic = container_of(sm, typeof(*scic), sm); 1494 struct isci_host *ihost = container_of(sm, typeof(*ihost), sm);
1507 1495
1508 /* set the default interrupt coalescence number and timeout value. */ 1496 /* set the default interrupt coalescence number and timeout value. */
1509 scic_controller_set_interrupt_coalescence(scic, 0x10, 250); 1497 scic_controller_set_interrupt_coalescence(ihost, 0x10, 250);
1510} 1498}
1511 1499
1512static void scic_sds_controller_ready_state_exit(struct sci_base_state_machine *sm) 1500static void scic_sds_controller_ready_state_exit(struct sci_base_state_machine *sm)
1513{ 1501{
1514 struct scic_sds_controller *scic = container_of(sm, typeof(*scic), sm); 1502 struct isci_host *ihost = container_of(sm, typeof(*ihost), sm);
1515 1503
1516 /* disable interrupt coalescence. */ 1504 /* disable interrupt coalescence. */
1517 scic_controller_set_interrupt_coalescence(scic, 0, 0); 1505 scic_controller_set_interrupt_coalescence(ihost, 0, 0);
1518} 1506}
1519 1507
1520static enum sci_status scic_sds_controller_stop_phys(struct scic_sds_controller *scic) 1508static enum sci_status scic_sds_controller_stop_phys(struct isci_host *ihost)
1521{ 1509{
1522 u32 index; 1510 u32 index;
1523 enum sci_status status; 1511 enum sci_status status;
1524 enum sci_status phy_status; 1512 enum sci_status phy_status;
1525 struct isci_host *ihost = scic_to_ihost(scic);
1526 1513
1527 status = SCI_SUCCESS; 1514 status = SCI_SUCCESS;
1528 1515
@@ -1533,7 +1520,7 @@ static enum sci_status scic_sds_controller_stop_phys(struct scic_sds_controller
1533 phy_status != SCI_FAILURE_INVALID_STATE) { 1520 phy_status != SCI_FAILURE_INVALID_STATE) {
1534 status = SCI_FAILURE; 1521 status = SCI_FAILURE;
1535 1522
1536 dev_warn(scic_to_dev(scic), 1523 dev_warn(&ihost->pdev->dev,
1537 "%s: Controller stop operation failed to stop " 1524 "%s: Controller stop operation failed to stop "
1538 "phy %d because of status %d.\n", 1525 "phy %d because of status %d.\n",
1539 __func__, 1526 __func__,
@@ -1544,14 +1531,13 @@ static enum sci_status scic_sds_controller_stop_phys(struct scic_sds_controller
1544 return status; 1531 return status;
1545} 1532}
1546 1533
1547static enum sci_status scic_sds_controller_stop_ports(struct scic_sds_controller *scic) 1534static enum sci_status scic_sds_controller_stop_ports(struct isci_host *ihost)
1548{ 1535{
1549 u32 index; 1536 u32 index;
1550 enum sci_status port_status; 1537 enum sci_status port_status;
1551 enum sci_status status = SCI_SUCCESS; 1538 enum sci_status status = SCI_SUCCESS;
1552 struct isci_host *ihost = scic_to_ihost(scic);
1553 1539
1554 for (index = 0; index < scic->logical_port_entries; index++) { 1540 for (index = 0; index < ihost->logical_port_entries; index++) {
1555 struct isci_port *iport = &ihost->ports[index]; 1541 struct isci_port *iport = &ihost->ports[index];
1556 1542
1557 port_status = scic_sds_port_stop(iport); 1543 port_status = scic_sds_port_stop(iport);
@@ -1560,7 +1546,7 @@ static enum sci_status scic_sds_controller_stop_ports(struct scic_sds_controller
1560 (port_status != SCI_FAILURE_INVALID_STATE)) { 1546 (port_status != SCI_FAILURE_INVALID_STATE)) {
1561 status = SCI_FAILURE; 1547 status = SCI_FAILURE;
1562 1548
1563 dev_warn(scic_to_dev(scic), 1549 dev_warn(&ihost->pdev->dev,
1564 "%s: Controller stop operation failed to " 1550 "%s: Controller stop operation failed to "
1565 "stop port %d because of status %d.\n", 1551 "stop port %d because of status %d.\n",
1566 __func__, 1552 __func__,
@@ -1572,7 +1558,7 @@ static enum sci_status scic_sds_controller_stop_ports(struct scic_sds_controller
1572 return status; 1558 return status;
1573} 1559}
1574 1560
1575static enum sci_status scic_sds_controller_stop_devices(struct scic_sds_controller *scic) 1561static enum sci_status scic_sds_controller_stop_devices(struct isci_host *ihost)
1576{ 1562{
1577 u32 index; 1563 u32 index;
1578 enum sci_status status; 1564 enum sci_status status;
@@ -1580,19 +1566,19 @@ static enum sci_status scic_sds_controller_stop_devices(struct scic_sds_controll
1580 1566
1581 status = SCI_SUCCESS; 1567 status = SCI_SUCCESS;
1582 1568
1583 for (index = 0; index < scic->remote_node_entries; index++) { 1569 for (index = 0; index < ihost->remote_node_entries; index++) {
1584 if (scic->device_table[index] != NULL) { 1570 if (ihost->device_table[index] != NULL) {
1585 /* / @todo What timeout value do we want to provide to this request? */ 1571 /* / @todo What timeout value do we want to provide to this request? */
1586 device_status = scic_remote_device_stop(scic->device_table[index], 0); 1572 device_status = scic_remote_device_stop(ihost->device_table[index], 0);
1587 1573
1588 if ((device_status != SCI_SUCCESS) && 1574 if ((device_status != SCI_SUCCESS) &&
1589 (device_status != SCI_FAILURE_INVALID_STATE)) { 1575 (device_status != SCI_FAILURE_INVALID_STATE)) {
1590 dev_warn(scic_to_dev(scic), 1576 dev_warn(&ihost->pdev->dev,
1591 "%s: Controller stop operation failed " 1577 "%s: Controller stop operation failed "
1592 "to stop device 0x%p because of " 1578 "to stop device 0x%p because of "
1593 "status %d.\n", 1579 "status %d.\n",
1594 __func__, 1580 __func__,
1595 scic->device_table[index], device_status); 1581 ihost->device_table[index], device_status);
1596 } 1582 }
1597 } 1583 }
1598 } 1584 }
@@ -1602,19 +1588,19 @@ static enum sci_status scic_sds_controller_stop_devices(struct scic_sds_controll
1602 1588
1603static void scic_sds_controller_stopping_state_enter(struct sci_base_state_machine *sm) 1589static void scic_sds_controller_stopping_state_enter(struct sci_base_state_machine *sm)
1604{ 1590{
1605 struct scic_sds_controller *scic = container_of(sm, typeof(*scic), sm); 1591 struct isci_host *ihost = container_of(sm, typeof(*ihost), sm);
1606 1592
1607 /* Stop all of the components for this controller */ 1593 /* Stop all of the components for this controller */
1608 scic_sds_controller_stop_phys(scic); 1594 scic_sds_controller_stop_phys(ihost);
1609 scic_sds_controller_stop_ports(scic); 1595 scic_sds_controller_stop_ports(ihost);
1610 scic_sds_controller_stop_devices(scic); 1596 scic_sds_controller_stop_devices(ihost);
1611} 1597}
1612 1598
1613static void scic_sds_controller_stopping_state_exit(struct sci_base_state_machine *sm) 1599static void scic_sds_controller_stopping_state_exit(struct sci_base_state_machine *sm)
1614{ 1600{
1615 struct scic_sds_controller *scic = container_of(sm, typeof(*scic), sm); 1601 struct isci_host *ihost = container_of(sm, typeof(*ihost), sm);
1616 1602
1617 sci_del_timer(&scic->timer); 1603 sci_del_timer(&ihost->timer);
1618} 1604}
1619 1605
1620 1606
@@ -1623,30 +1609,30 @@ static void scic_sds_controller_stopping_state_exit(struct sci_base_state_machin
1623 * 1609 *
1624 * This method will reset the controller hardware. 1610 * This method will reset the controller hardware.
1625 */ 1611 */
1626static void scic_sds_controller_reset_hardware(struct scic_sds_controller *scic) 1612static void scic_sds_controller_reset_hardware(struct isci_host *ihost)
1627{ 1613{
1628 /* Disable interrupts so we dont take any spurious interrupts */ 1614 /* Disable interrupts so we dont take any spurious interrupts */
1629 scic_controller_disable_interrupts(scic); 1615 scic_controller_disable_interrupts(ihost);
1630 1616
1631 /* Reset the SCU */ 1617 /* Reset the SCU */
1632 writel(0xFFFFFFFF, &scic->smu_registers->soft_reset_control); 1618 writel(0xFFFFFFFF, &ihost->smu_registers->soft_reset_control);
1633 1619
1634 /* Delay for 1ms to before clearing the CQP and UFQPR. */ 1620 /* Delay for 1ms to before clearing the CQP and UFQPR. */
1635 udelay(1000); 1621 udelay(1000);
1636 1622
1637 /* The write to the CQGR clears the CQP */ 1623 /* The write to the CQGR clears the CQP */
1638 writel(0x00000000, &scic->smu_registers->completion_queue_get); 1624 writel(0x00000000, &ihost->smu_registers->completion_queue_get);
1639 1625
1640 /* The write to the UFQGP clears the UFQPR */ 1626 /* The write to the UFQGP clears the UFQPR */
1641 writel(0, &scic->scu_registers->sdma.unsolicited_frame_get_pointer); 1627 writel(0, &ihost->scu_registers->sdma.unsolicited_frame_get_pointer);
1642} 1628}
1643 1629
1644static void scic_sds_controller_resetting_state_enter(struct sci_base_state_machine *sm) 1630static void scic_sds_controller_resetting_state_enter(struct sci_base_state_machine *sm)
1645{ 1631{
1646 struct scic_sds_controller *scic = container_of(sm, typeof(*scic), sm); 1632 struct isci_host *ihost = container_of(sm, typeof(*ihost), sm);
1647 1633
1648 scic_sds_controller_reset_hardware(scic); 1634 scic_sds_controller_reset_hardware(ihost);
1649 sci_change_state(&scic->sm, SCIC_RESET); 1635 sci_change_state(&ihost->sm, SCIC_RESET);
1650} 1636}
1651 1637
1652static const struct sci_base_state scic_sds_controller_state_table[] = { 1638static const struct sci_base_state scic_sds_controller_state_table[] = {
@@ -1674,58 +1660,56 @@ static const struct sci_base_state scic_sds_controller_state_table[] = {
1674 [SCIC_FAILED] = {} 1660 [SCIC_FAILED] = {}
1675}; 1661};
1676 1662
1677static void scic_sds_controller_set_default_config_parameters(struct scic_sds_controller *scic) 1663static void scic_sds_controller_set_default_config_parameters(struct isci_host *ihost)
1678{ 1664{
1679 /* these defaults are overridden by the platform / firmware */ 1665 /* these defaults are overridden by the platform / firmware */
1680 struct isci_host *ihost = scic_to_ihost(scic);
1681 u16 index; 1666 u16 index;
1682 1667
1683 /* Default to APC mode. */ 1668 /* Default to APC mode. */
1684 scic->oem_parameters.sds1.controller.mode_type = SCIC_PORT_AUTOMATIC_CONFIGURATION_MODE; 1669 ihost->oem_parameters.sds1.controller.mode_type = SCIC_PORT_AUTOMATIC_CONFIGURATION_MODE;
1685 1670
1686 /* Default to APC mode. */ 1671 /* Default to APC mode. */
1687 scic->oem_parameters.sds1.controller.max_concurrent_dev_spin_up = 1; 1672 ihost->oem_parameters.sds1.controller.max_concurrent_dev_spin_up = 1;
1688 1673
1689 /* Default to no SSC operation. */ 1674 /* Default to no SSC operation. */
1690 scic->oem_parameters.sds1.controller.do_enable_ssc = false; 1675 ihost->oem_parameters.sds1.controller.do_enable_ssc = false;
1691 1676
1692 /* Initialize all of the port parameter information to narrow ports. */ 1677 /* Initialize all of the port parameter information to narrow ports. */
1693 for (index = 0; index < SCI_MAX_PORTS; index++) { 1678 for (index = 0; index < SCI_MAX_PORTS; index++) {
1694 scic->oem_parameters.sds1.ports[index].phy_mask = 0; 1679 ihost->oem_parameters.sds1.ports[index].phy_mask = 0;
1695 } 1680 }
1696 1681
1697 /* Initialize all of the phy parameter information. */ 1682 /* Initialize all of the phy parameter information. */
1698 for (index = 0; index < SCI_MAX_PHYS; index++) { 1683 for (index = 0; index < SCI_MAX_PHYS; index++) {
1699 /* Default to 6G (i.e. Gen 3) for now. */ 1684 /* Default to 6G (i.e. Gen 3) for now. */
1700 scic->user_parameters.sds1.phys[index].max_speed_generation = 3; 1685 ihost->user_parameters.sds1.phys[index].max_speed_generation = 3;
1701 1686
1702 /* the frequencies cannot be 0 */ 1687 /* the frequencies cannot be 0 */
1703 scic->user_parameters.sds1.phys[index].align_insertion_frequency = 0x7f; 1688 ihost->user_parameters.sds1.phys[index].align_insertion_frequency = 0x7f;
1704 scic->user_parameters.sds1.phys[index].in_connection_align_insertion_frequency = 0xff; 1689 ihost->user_parameters.sds1.phys[index].in_connection_align_insertion_frequency = 0xff;
1705 scic->user_parameters.sds1.phys[index].notify_enable_spin_up_insertion_frequency = 0x33; 1690 ihost->user_parameters.sds1.phys[index].notify_enable_spin_up_insertion_frequency = 0x33;
1706 1691
1707 /* 1692 /*
1708 * Previous Vitesse based expanders had a arbitration issue that 1693 * Previous Vitesse based expanders had a arbitration issue that
1709 * is worked around by having the upper 32-bits of SAS address 1694 * is worked around by having the upper 32-bits of SAS address
1710 * with a value greater then the Vitesse company identifier. 1695 * with a value greater then the Vitesse company identifier.
1711 * Hence, usage of 0x5FCFFFFF. */ 1696 * Hence, usage of 0x5FCFFFFF. */
1712 scic->oem_parameters.sds1.phys[index].sas_address.low = 0x1 + ihost->id; 1697 ihost->oem_parameters.sds1.phys[index].sas_address.low = 0x1 + ihost->id;
1713 scic->oem_parameters.sds1.phys[index].sas_address.high = 0x5FCFFFFF; 1698 ihost->oem_parameters.sds1.phys[index].sas_address.high = 0x5FCFFFFF;
1714 } 1699 }
1715 1700
1716 scic->user_parameters.sds1.stp_inactivity_timeout = 5; 1701 ihost->user_parameters.sds1.stp_inactivity_timeout = 5;
1717 scic->user_parameters.sds1.ssp_inactivity_timeout = 5; 1702 ihost->user_parameters.sds1.ssp_inactivity_timeout = 5;
1718 scic->user_parameters.sds1.stp_max_occupancy_timeout = 5; 1703 ihost->user_parameters.sds1.stp_max_occupancy_timeout = 5;
1719 scic->user_parameters.sds1.ssp_max_occupancy_timeout = 20; 1704 ihost->user_parameters.sds1.ssp_max_occupancy_timeout = 20;
1720 scic->user_parameters.sds1.no_outbound_task_timeout = 20; 1705 ihost->user_parameters.sds1.no_outbound_task_timeout = 20;
1721} 1706}
1722 1707
1723static void controller_timeout(unsigned long data) 1708static void controller_timeout(unsigned long data)
1724{ 1709{
1725 struct sci_timer *tmr = (struct sci_timer *)data; 1710 struct sci_timer *tmr = (struct sci_timer *)data;
1726 struct scic_sds_controller *scic = container_of(tmr, typeof(*scic), timer); 1711 struct isci_host *ihost = container_of(tmr, typeof(*ihost), timer);
1727 struct isci_host *ihost = scic_to_ihost(scic); 1712 struct sci_base_state_machine *sm = &ihost->sm;
1728 struct sci_base_state_machine *sm = &scic->sm;
1729 unsigned long flags; 1713 unsigned long flags;
1730 1714
1731 spin_lock_irqsave(&ihost->scic_lock, flags); 1715 spin_lock_irqsave(&ihost->scic_lock, flags);
@@ -1734,12 +1718,12 @@ static void controller_timeout(unsigned long data)
1734 goto done; 1718 goto done;
1735 1719
1736 if (sm->current_state_id == SCIC_STARTING) 1720 if (sm->current_state_id == SCIC_STARTING)
1737 scic_sds_controller_transition_to_ready(scic, SCI_FAILURE_TIMEOUT); 1721 scic_sds_controller_transition_to_ready(ihost, SCI_FAILURE_TIMEOUT);
1738 else if (sm->current_state_id == SCIC_STOPPING) { 1722 else if (sm->current_state_id == SCIC_STOPPING) {
1739 sci_change_state(sm, SCIC_FAILED); 1723 sci_change_state(sm, SCIC_FAILED);
1740 isci_host_stop_complete(ihost, SCI_FAILURE_TIMEOUT); 1724 isci_host_stop_complete(ihost, SCI_FAILURE_TIMEOUT);
1741 } else /* / @todo Now what do we want to do in this case? */ 1725 } else /* / @todo Now what do we want to do in this case? */
1742 dev_err(scic_to_dev(scic), 1726 dev_err(&ihost->pdev->dev,
1743 "%s: Controller timer fired when controller was not " 1727 "%s: Controller timer fired when controller was not "
1744 "in a state being timed.\n", 1728 "in a state being timed.\n",
1745 __func__); 1729 __func__);
@@ -1764,24 +1748,23 @@ done:
1764 * SCI_FAILURE_UNSUPPORTED_INIT_DATA_VERSION This value is returned if the 1748 * SCI_FAILURE_UNSUPPORTED_INIT_DATA_VERSION This value is returned if the
1765 * controller does not support the supplied initialization data version. 1749 * controller does not support the supplied initialization data version.
1766 */ 1750 */
1767static enum sci_status scic_controller_construct(struct scic_sds_controller *scic, 1751static enum sci_status scic_controller_construct(struct isci_host *ihost,
1768 void __iomem *scu_base, 1752 void __iomem *scu_base,
1769 void __iomem *smu_base) 1753 void __iomem *smu_base)
1770{ 1754{
1771 struct isci_host *ihost = scic_to_ihost(scic);
1772 u8 i; 1755 u8 i;
1773 1756
1774 sci_init_sm(&scic->sm, scic_sds_controller_state_table, SCIC_INITIAL); 1757 sci_init_sm(&ihost->sm, scic_sds_controller_state_table, SCIC_INITIAL);
1775 1758
1776 scic->scu_registers = scu_base; 1759 ihost->scu_registers = scu_base;
1777 scic->smu_registers = smu_base; 1760 ihost->smu_registers = smu_base;
1778 1761
1779 scic_sds_port_configuration_agent_construct(&scic->port_agent); 1762 scic_sds_port_configuration_agent_construct(&ihost->port_agent);
1780 1763
1781 /* Construct the ports for this controller */ 1764 /* Construct the ports for this controller */
1782 for (i = 0; i < SCI_MAX_PORTS; i++) 1765 for (i = 0; i < SCI_MAX_PORTS; i++)
1783 scic_sds_port_construct(&ihost->ports[i], i, scic); 1766 scic_sds_port_construct(&ihost->ports[i], i, ihost);
1784 scic_sds_port_construct(&ihost->ports[i], SCIC_SDS_DUMMY_PORT, scic); 1767 scic_sds_port_construct(&ihost->ports[i], SCIC_SDS_DUMMY_PORT, ihost);
1785 1768
1786 /* Construct the phys for this controller */ 1769 /* Construct the phys for this controller */
1787 for (i = 0; i < SCI_MAX_PHYS; i++) { 1770 for (i = 0; i < SCI_MAX_PHYS; i++) {
@@ -1790,14 +1773,14 @@ static enum sci_status scic_controller_construct(struct scic_sds_controller *sci
1790 &ihost->ports[SCI_MAX_PORTS], i); 1773 &ihost->ports[SCI_MAX_PORTS], i);
1791 } 1774 }
1792 1775
1793 scic->invalid_phy_mask = 0; 1776 ihost->invalid_phy_mask = 0;
1794 1777
1795 sci_init_timer(&scic->timer, controller_timeout); 1778 sci_init_timer(&ihost->timer, controller_timeout);
1796 1779
1797 /* Initialize the User and OEM parameters to default values. */ 1780 /* Initialize the User and OEM parameters to default values. */
1798 scic_sds_controller_set_default_config_parameters(scic); 1781 scic_sds_controller_set_default_config_parameters(ihost);
1799 1782
1800 return scic_controller_reset(scic); 1783 return scic_controller_reset(ihost);
1801} 1784}
1802 1785
1803int scic_oem_parameters_validate(struct scic_sds_oem_params *oem) 1786int scic_oem_parameters_validate(struct scic_sds_oem_params *oem)
@@ -1834,10 +1817,10 @@ int scic_oem_parameters_validate(struct scic_sds_oem_params *oem)
1834 return 0; 1817 return 0;
1835} 1818}
1836 1819
1837static enum sci_status scic_oem_parameters_set(struct scic_sds_controller *scic, 1820static enum sci_status scic_oem_parameters_set(struct isci_host *ihost,
1838 union scic_oem_parameters *scic_parms) 1821 union scic_oem_parameters *scic_parms)
1839{ 1822{
1840 u32 state = scic->sm.current_state_id; 1823 u32 state = ihost->sm.current_state_id;
1841 1824
1842 if (state == SCIC_RESET || 1825 if (state == SCIC_RESET ||
1843 state == SCIC_INITIALIZING || 1826 state == SCIC_INITIALIZING ||
@@ -1845,7 +1828,7 @@ static enum sci_status scic_oem_parameters_set(struct scic_sds_controller *scic,
1845 1828
1846 if (scic_oem_parameters_validate(&scic_parms->sds1)) 1829 if (scic_oem_parameters_validate(&scic_parms->sds1))
1847 return SCI_FAILURE_INVALID_PARAMETER_VALUE; 1830 return SCI_FAILURE_INVALID_PARAMETER_VALUE;
1848 scic->oem_parameters.sds1 = scic_parms->sds1; 1831 ihost->oem_parameters.sds1 = scic_parms->sds1;
1849 1832
1850 return SCI_SUCCESS; 1833 return SCI_SUCCESS;
1851 } 1834 }
@@ -1854,17 +1837,16 @@ static enum sci_status scic_oem_parameters_set(struct scic_sds_controller *scic,
1854} 1837}
1855 1838
1856void scic_oem_parameters_get( 1839void scic_oem_parameters_get(
1857 struct scic_sds_controller *scic, 1840 struct isci_host *ihost,
1858 union scic_oem_parameters *scic_parms) 1841 union scic_oem_parameters *scic_parms)
1859{ 1842{
1860 memcpy(scic_parms, (&scic->oem_parameters), sizeof(*scic_parms)); 1843 memcpy(scic_parms, (&ihost->oem_parameters), sizeof(*scic_parms));
1861} 1844}
1862 1845
1863static void power_control_timeout(unsigned long data) 1846static void power_control_timeout(unsigned long data)
1864{ 1847{
1865 struct sci_timer *tmr = (struct sci_timer *)data; 1848 struct sci_timer *tmr = (struct sci_timer *)data;
1866 struct scic_sds_controller *scic = container_of(tmr, typeof(*scic), power_control.timer); 1849 struct isci_host *ihost = container_of(tmr, typeof(*ihost), power_control.timer);
1867 struct isci_host *ihost = scic_to_ihost(scic);
1868 struct isci_phy *iphy; 1850 struct isci_phy *iphy;
1869 unsigned long flags; 1851 unsigned long flags;
1870 u8 i; 1852 u8 i;
@@ -1874,29 +1856,29 @@ static void power_control_timeout(unsigned long data)
1874 if (tmr->cancel) 1856 if (tmr->cancel)
1875 goto done; 1857 goto done;
1876 1858
1877 scic->power_control.phys_granted_power = 0; 1859 ihost->power_control.phys_granted_power = 0;
1878 1860
1879 if (scic->power_control.phys_waiting == 0) { 1861 if (ihost->power_control.phys_waiting == 0) {
1880 scic->power_control.timer_started = false; 1862 ihost->power_control.timer_started = false;
1881 goto done; 1863 goto done;
1882 } 1864 }
1883 1865
1884 for (i = 0; i < SCI_MAX_PHYS; i++) { 1866 for (i = 0; i < SCI_MAX_PHYS; i++) {
1885 1867
1886 if (scic->power_control.phys_waiting == 0) 1868 if (ihost->power_control.phys_waiting == 0)
1887 break; 1869 break;
1888 1870
1889 iphy = scic->power_control.requesters[i]; 1871 iphy = ihost->power_control.requesters[i];
1890 if (iphy == NULL) 1872 if (iphy == NULL)
1891 continue; 1873 continue;
1892 1874
1893 if (scic->power_control.phys_granted_power >= 1875 if (ihost->power_control.phys_granted_power >=
1894 scic->oem_parameters.sds1.controller.max_concurrent_dev_spin_up) 1876 ihost->oem_parameters.sds1.controller.max_concurrent_dev_spin_up)
1895 break; 1877 break;
1896 1878
1897 scic->power_control.requesters[i] = NULL; 1879 ihost->power_control.requesters[i] = NULL;
1898 scic->power_control.phys_waiting--; 1880 ihost->power_control.phys_waiting--;
1899 scic->power_control.phys_granted_power++; 1881 ihost->power_control.phys_granted_power++;
1900 scic_sds_phy_consume_power_handler(iphy); 1882 scic_sds_phy_consume_power_handler(iphy);
1901 } 1883 }
1902 1884
@@ -1905,7 +1887,7 @@ static void power_control_timeout(unsigned long data)
1905 * timer in case another phy becomes ready. 1887 * timer in case another phy becomes ready.
1906 */ 1888 */
1907 sci_mod_timer(tmr, SCIC_SDS_CONTROLLER_POWER_CONTROL_INTERVAL); 1889 sci_mod_timer(tmr, SCIC_SDS_CONTROLLER_POWER_CONTROL_INTERVAL);
1908 scic->power_control.timer_started = true; 1890 ihost->power_control.timer_started = true;
1909 1891
1910done: 1892done:
1911 spin_unlock_irqrestore(&ihost->scic_lock, flags); 1893 spin_unlock_irqrestore(&ihost->scic_lock, flags);
@@ -1918,31 +1900,31 @@ done:
1918 * 1900 *
1919 */ 1901 */
1920void scic_sds_controller_power_control_queue_insert( 1902void scic_sds_controller_power_control_queue_insert(
1921 struct scic_sds_controller *scic, 1903 struct isci_host *ihost,
1922 struct isci_phy *iphy) 1904 struct isci_phy *iphy)
1923{ 1905{
1924 BUG_ON(iphy == NULL); 1906 BUG_ON(iphy == NULL);
1925 1907
1926 if (scic->power_control.phys_granted_power < 1908 if (ihost->power_control.phys_granted_power <
1927 scic->oem_parameters.sds1.controller.max_concurrent_dev_spin_up) { 1909 ihost->oem_parameters.sds1.controller.max_concurrent_dev_spin_up) {
1928 scic->power_control.phys_granted_power++; 1910 ihost->power_control.phys_granted_power++;
1929 scic_sds_phy_consume_power_handler(iphy); 1911 scic_sds_phy_consume_power_handler(iphy);
1930 1912
1931 /* 1913 /*
1932 * stop and start the power_control timer. When the timer fires, the 1914 * stop and start the power_control timer. When the timer fires, the
1933 * no_of_phys_granted_power will be set to 0 1915 * no_of_phys_granted_power will be set to 0
1934 */ 1916 */
1935 if (scic->power_control.timer_started) 1917 if (ihost->power_control.timer_started)
1936 sci_del_timer(&scic->power_control.timer); 1918 sci_del_timer(&ihost->power_control.timer);
1937 1919
1938 sci_mod_timer(&scic->power_control.timer, 1920 sci_mod_timer(&ihost->power_control.timer,
1939 SCIC_SDS_CONTROLLER_POWER_CONTROL_INTERVAL); 1921 SCIC_SDS_CONTROLLER_POWER_CONTROL_INTERVAL);
1940 scic->power_control.timer_started = true; 1922 ihost->power_control.timer_started = true;
1941 1923
1942 } else { 1924 } else {
1943 /* Add the phy in the waiting list */ 1925 /* Add the phy in the waiting list */
1944 scic->power_control.requesters[iphy->phy_index] = iphy; 1926 ihost->power_control.requesters[iphy->phy_index] = iphy;
1945 scic->power_control.phys_waiting++; 1927 ihost->power_control.phys_waiting++;
1946 } 1928 }
1947} 1929}
1948 1930
@@ -1953,16 +1935,16 @@ void scic_sds_controller_power_control_queue_insert(
1953 * 1935 *
1954 */ 1936 */
1955void scic_sds_controller_power_control_queue_remove( 1937void scic_sds_controller_power_control_queue_remove(
1956 struct scic_sds_controller *scic, 1938 struct isci_host *ihost,
1957 struct isci_phy *iphy) 1939 struct isci_phy *iphy)
1958{ 1940{
1959 BUG_ON(iphy == NULL); 1941 BUG_ON(iphy == NULL);
1960 1942
1961 if (scic->power_control.requesters[iphy->phy_index] != NULL) { 1943 if (ihost->power_control.requesters[iphy->phy_index] != NULL) {
1962 scic->power_control.phys_waiting--; 1944 ihost->power_control.phys_waiting--;
1963 } 1945 }
1964 1946
1965 scic->power_control.requesters[iphy->phy_index] = NULL; 1947 ihost->power_control.requesters[iphy->phy_index] = NULL;
1966} 1948}
1967 1949
1968#define AFE_REGISTER_WRITE_DELAY 10 1950#define AFE_REGISTER_WRITE_DELAY 10
@@ -1970,50 +1952,50 @@ void scic_sds_controller_power_control_queue_remove(
1970/* Initialize the AFE for this phy index. We need to read the AFE setup from 1952/* Initialize the AFE for this phy index. We need to read the AFE setup from
1971 * the OEM parameters 1953 * the OEM parameters
1972 */ 1954 */
1973static void scic_sds_controller_afe_initialization(struct scic_sds_controller *scic) 1955static void scic_sds_controller_afe_initialization(struct isci_host *ihost)
1974{ 1956{
1975 const struct scic_sds_oem_params *oem = &scic->oem_parameters.sds1; 1957 const struct scic_sds_oem_params *oem = &ihost->oem_parameters.sds1;
1976 u32 afe_status; 1958 u32 afe_status;
1977 u32 phy_id; 1959 u32 phy_id;
1978 1960
1979 /* Clear DFX Status registers */ 1961 /* Clear DFX Status registers */
1980 writel(0x0081000f, &scic->scu_registers->afe.afe_dfx_master_control0); 1962 writel(0x0081000f, &ihost->scu_registers->afe.afe_dfx_master_control0);
1981 udelay(AFE_REGISTER_WRITE_DELAY); 1963 udelay(AFE_REGISTER_WRITE_DELAY);
1982 1964
1983 if (is_b0()) { 1965 if (is_b0()) {
1984 /* PM Rx Equalization Save, PM SPhy Rx Acknowledgement 1966 /* PM Rx Equalization Save, PM SPhy Rx Acknowledgement
1985 * Timer, PM Stagger Timer */ 1967 * Timer, PM Stagger Timer */
1986 writel(0x0007BFFF, &scic->scu_registers->afe.afe_pmsn_master_control2); 1968 writel(0x0007BFFF, &ihost->scu_registers->afe.afe_pmsn_master_control2);
1987 udelay(AFE_REGISTER_WRITE_DELAY); 1969 udelay(AFE_REGISTER_WRITE_DELAY);
1988 } 1970 }
1989 1971
1990 /* Configure bias currents to normal */ 1972 /* Configure bias currents to normal */
1991 if (is_a0()) 1973 if (is_a0())
1992 writel(0x00005500, &scic->scu_registers->afe.afe_bias_control); 1974 writel(0x00005500, &ihost->scu_registers->afe.afe_bias_control);
1993 else if (is_a2()) 1975 else if (is_a2())
1994 writel(0x00005A00, &scic->scu_registers->afe.afe_bias_control); 1976 writel(0x00005A00, &ihost->scu_registers->afe.afe_bias_control);
1995 else if (is_b0() || is_c0()) 1977 else if (is_b0() || is_c0())
1996 writel(0x00005F00, &scic->scu_registers->afe.afe_bias_control); 1978 writel(0x00005F00, &ihost->scu_registers->afe.afe_bias_control);
1997 1979
1998 udelay(AFE_REGISTER_WRITE_DELAY); 1980 udelay(AFE_REGISTER_WRITE_DELAY);
1999 1981
2000 /* Enable PLL */ 1982 /* Enable PLL */
2001 if (is_b0() || is_c0()) 1983 if (is_b0() || is_c0())
2002 writel(0x80040A08, &scic->scu_registers->afe.afe_pll_control0); 1984 writel(0x80040A08, &ihost->scu_registers->afe.afe_pll_control0);
2003 else 1985 else
2004 writel(0x80040908, &scic->scu_registers->afe.afe_pll_control0); 1986 writel(0x80040908, &ihost->scu_registers->afe.afe_pll_control0);
2005 1987
2006 udelay(AFE_REGISTER_WRITE_DELAY); 1988 udelay(AFE_REGISTER_WRITE_DELAY);
2007 1989
2008 /* Wait for the PLL to lock */ 1990 /* Wait for the PLL to lock */
2009 do { 1991 do {
2010 afe_status = readl(&scic->scu_registers->afe.afe_common_block_status); 1992 afe_status = readl(&ihost->scu_registers->afe.afe_common_block_status);
2011 udelay(AFE_REGISTER_WRITE_DELAY); 1993 udelay(AFE_REGISTER_WRITE_DELAY);
2012 } while ((afe_status & 0x00001000) == 0); 1994 } while ((afe_status & 0x00001000) == 0);
2013 1995
2014 if (is_a0() || is_a2()) { 1996 if (is_a0() || is_a2()) {
2015 /* Shorten SAS SNW lock time (RxLock timer value from 76 us to 50 us) */ 1997 /* Shorten SAS SNW lock time (RxLock timer value from 76 us to 50 us) */
2016 writel(0x7bcc96ad, &scic->scu_registers->afe.afe_pmsn_master_control0); 1998 writel(0x7bcc96ad, &ihost->scu_registers->afe.afe_pmsn_master_control0);
2017 udelay(AFE_REGISTER_WRITE_DELAY); 1999 udelay(AFE_REGISTER_WRITE_DELAY);
2018 } 2000 }
2019 2001
@@ -2022,26 +2004,26 @@ static void scic_sds_controller_afe_initialization(struct scic_sds_controller *s
2022 2004
2023 if (is_b0()) { 2005 if (is_b0()) {
2024 /* Configure transmitter SSC parameters */ 2006 /* Configure transmitter SSC parameters */
2025 writel(0x00030000, &scic->scu_registers->afe.scu_afe_xcvr[phy_id].afe_tx_ssc_control); 2007 writel(0x00030000, &ihost->scu_registers->afe.scu_afe_xcvr[phy_id].afe_tx_ssc_control);
2026 udelay(AFE_REGISTER_WRITE_DELAY); 2008 udelay(AFE_REGISTER_WRITE_DELAY);
2027 } else if (is_c0()) { 2009 } else if (is_c0()) {
2028 /* Configure transmitter SSC parameters */ 2010 /* Configure transmitter SSC parameters */
2029 writel(0x0003000, &scic->scu_registers->afe.scu_afe_xcvr[phy_id].afe_tx_ssc_control); 2011 writel(0x0003000, &ihost->scu_registers->afe.scu_afe_xcvr[phy_id].afe_tx_ssc_control);
2030 udelay(AFE_REGISTER_WRITE_DELAY); 2012 udelay(AFE_REGISTER_WRITE_DELAY);
2031 2013
2032 /* 2014 /*
2033 * All defaults, except the Receive Word Alignament/Comma Detect 2015 * All defaults, except the Receive Word Alignament/Comma Detect
2034 * Enable....(0xe800) */ 2016 * Enable....(0xe800) */
2035 writel(0x00004500, &scic->scu_registers->afe.scu_afe_xcvr[phy_id].afe_xcvr_control0); 2017 writel(0x00004500, &ihost->scu_registers->afe.scu_afe_xcvr[phy_id].afe_xcvr_control0);
2036 udelay(AFE_REGISTER_WRITE_DELAY); 2018 udelay(AFE_REGISTER_WRITE_DELAY);
2037 } else { 2019 } else {
2038 /* 2020 /*
2039 * All defaults, except the Receive Word Alignament/Comma Detect 2021 * All defaults, except the Receive Word Alignament/Comma Detect
2040 * Enable....(0xe800) */ 2022 * Enable....(0xe800) */
2041 writel(0x00004512, &scic->scu_registers->afe.scu_afe_xcvr[phy_id].afe_xcvr_control0); 2023 writel(0x00004512, &ihost->scu_registers->afe.scu_afe_xcvr[phy_id].afe_xcvr_control0);
2042 udelay(AFE_REGISTER_WRITE_DELAY); 2024 udelay(AFE_REGISTER_WRITE_DELAY);
2043 2025
2044 writel(0x0050100F, &scic->scu_registers->afe.scu_afe_xcvr[phy_id].afe_xcvr_control1); 2026 writel(0x0050100F, &ihost->scu_registers->afe.scu_afe_xcvr[phy_id].afe_xcvr_control1);
2045 udelay(AFE_REGISTER_WRITE_DELAY); 2027 udelay(AFE_REGISTER_WRITE_DELAY);
2046 } 2028 }
2047 2029
@@ -2049,106 +2031,105 @@ static void scic_sds_controller_afe_initialization(struct scic_sds_controller *s
2049 * Power up TX and RX out from power down (PWRDNTX and PWRDNRX) 2031 * Power up TX and RX out from power down (PWRDNTX and PWRDNRX)
2050 * & increase TX int & ext bias 20%....(0xe85c) */ 2032 * & increase TX int & ext bias 20%....(0xe85c) */
2051 if (is_a0()) 2033 if (is_a0())
2052 writel(0x000003D4, &scic->scu_registers->afe.scu_afe_xcvr[phy_id].afe_channel_control); 2034 writel(0x000003D4, &ihost->scu_registers->afe.scu_afe_xcvr[phy_id].afe_channel_control);
2053 else if (is_a2()) 2035 else if (is_a2())
2054 writel(0x000003F0, &scic->scu_registers->afe.scu_afe_xcvr[phy_id].afe_channel_control); 2036 writel(0x000003F0, &ihost->scu_registers->afe.scu_afe_xcvr[phy_id].afe_channel_control);
2055 else if (is_b0()) { 2037 else if (is_b0()) {
2056 /* Power down TX and RX (PWRDNTX and PWRDNRX) */ 2038 /* Power down TX and RX (PWRDNTX and PWRDNRX) */
2057 writel(0x000003D7, &scic->scu_registers->afe.scu_afe_xcvr[phy_id].afe_channel_control); 2039 writel(0x000003D7, &ihost->scu_registers->afe.scu_afe_xcvr[phy_id].afe_channel_control);
2058 udelay(AFE_REGISTER_WRITE_DELAY); 2040 udelay(AFE_REGISTER_WRITE_DELAY);
2059 2041
2060 /* 2042 /*
2061 * Power up TX and RX out from power down (PWRDNTX and PWRDNRX) 2043 * Power up TX and RX out from power down (PWRDNTX and PWRDNRX)
2062 * & increase TX int & ext bias 20%....(0xe85c) */ 2044 * & increase TX int & ext bias 20%....(0xe85c) */
2063 writel(0x000003D4, &scic->scu_registers->afe.scu_afe_xcvr[phy_id].afe_channel_control); 2045 writel(0x000003D4, &ihost->scu_registers->afe.scu_afe_xcvr[phy_id].afe_channel_control);
2064 } else { 2046 } else {
2065 writel(0x000001E7, &scic->scu_registers->afe.scu_afe_xcvr[phy_id].afe_channel_control); 2047 writel(0x000001E7, &ihost->scu_registers->afe.scu_afe_xcvr[phy_id].afe_channel_control);
2066 udelay(AFE_REGISTER_WRITE_DELAY); 2048 udelay(AFE_REGISTER_WRITE_DELAY);
2067 2049
2068 /* 2050 /*
2069 * Power up TX and RX out from power down (PWRDNTX and PWRDNRX) 2051 * Power up TX and RX out from power down (PWRDNTX and PWRDNRX)
2070 * & increase TX int & ext bias 20%....(0xe85c) */ 2052 * & increase TX int & ext bias 20%....(0xe85c) */
2071 writel(0x000001E4, &scic->scu_registers->afe.scu_afe_xcvr[phy_id].afe_channel_control); 2053 writel(0x000001E4, &ihost->scu_registers->afe.scu_afe_xcvr[phy_id].afe_channel_control);
2072 } 2054 }
2073 udelay(AFE_REGISTER_WRITE_DELAY); 2055 udelay(AFE_REGISTER_WRITE_DELAY);
2074 2056
2075 if (is_a0() || is_a2()) { 2057 if (is_a0() || is_a2()) {
2076 /* Enable TX equalization (0xe824) */ 2058 /* Enable TX equalization (0xe824) */
2077 writel(0x00040000, &scic->scu_registers->afe.scu_afe_xcvr[phy_id].afe_tx_control); 2059 writel(0x00040000, &ihost->scu_registers->afe.scu_afe_xcvr[phy_id].afe_tx_control);
2078 udelay(AFE_REGISTER_WRITE_DELAY); 2060 udelay(AFE_REGISTER_WRITE_DELAY);
2079 } 2061 }
2080 2062
2081 /* 2063 /*
2082 * RDPI=0x0(RX Power On), RXOOBDETPDNC=0x0, TPD=0x0(TX Power On), 2064 * RDPI=0x0(RX Power On), RXOOBDETPDNC=0x0, TPD=0x0(TX Power On),
2083 * RDD=0x0(RX Detect Enabled) ....(0xe800) */ 2065 * RDD=0x0(RX Detect Enabled) ....(0xe800) */
2084 writel(0x00004100, &scic->scu_registers->afe.scu_afe_xcvr[phy_id].afe_xcvr_control0); 2066 writel(0x00004100, &ihost->scu_registers->afe.scu_afe_xcvr[phy_id].afe_xcvr_control0);
2085 udelay(AFE_REGISTER_WRITE_DELAY); 2067 udelay(AFE_REGISTER_WRITE_DELAY);
2086 2068
2087 /* Leave DFE/FFE on */ 2069 /* Leave DFE/FFE on */
2088 if (is_a0()) 2070 if (is_a0())
2089 writel(0x3F09983F, &scic->scu_registers->afe.scu_afe_xcvr[phy_id].afe_rx_ssc_control0); 2071 writel(0x3F09983F, &ihost->scu_registers->afe.scu_afe_xcvr[phy_id].afe_rx_ssc_control0);
2090 else if (is_a2()) 2072 else if (is_a2())
2091 writel(0x3F11103F, &scic->scu_registers->afe.scu_afe_xcvr[phy_id].afe_rx_ssc_control0); 2073 writel(0x3F11103F, &ihost->scu_registers->afe.scu_afe_xcvr[phy_id].afe_rx_ssc_control0);
2092 else if (is_b0()) { 2074 else if (is_b0()) {
2093 writel(0x3F11103F, &scic->scu_registers->afe.scu_afe_xcvr[phy_id].afe_rx_ssc_control0); 2075 writel(0x3F11103F, &ihost->scu_registers->afe.scu_afe_xcvr[phy_id].afe_rx_ssc_control0);
2094 udelay(AFE_REGISTER_WRITE_DELAY); 2076 udelay(AFE_REGISTER_WRITE_DELAY);
2095 /* Enable TX equalization (0xe824) */ 2077 /* Enable TX equalization (0xe824) */
2096 writel(0x00040000, &scic->scu_registers->afe.scu_afe_xcvr[phy_id].afe_tx_control); 2078 writel(0x00040000, &ihost->scu_registers->afe.scu_afe_xcvr[phy_id].afe_tx_control);
2097 } else { 2079 } else {
2098 writel(0x0140DF0F, &scic->scu_registers->afe.scu_afe_xcvr[phy_id].afe_rx_ssc_control1); 2080 writel(0x0140DF0F, &ihost->scu_registers->afe.scu_afe_xcvr[phy_id].afe_rx_ssc_control1);
2099 udelay(AFE_REGISTER_WRITE_DELAY); 2081 udelay(AFE_REGISTER_WRITE_DELAY);
2100 2082
2101 writel(0x3F6F103F, &scic->scu_registers->afe.scu_afe_xcvr[phy_id].afe_rx_ssc_control0); 2083 writel(0x3F6F103F, &ihost->scu_registers->afe.scu_afe_xcvr[phy_id].afe_rx_ssc_control0);
2102 udelay(AFE_REGISTER_WRITE_DELAY); 2084 udelay(AFE_REGISTER_WRITE_DELAY);
2103 2085
2104 /* Enable TX equalization (0xe824) */ 2086 /* Enable TX equalization (0xe824) */
2105 writel(0x00040000, &scic->scu_registers->afe.scu_afe_xcvr[phy_id].afe_tx_control); 2087 writel(0x00040000, &ihost->scu_registers->afe.scu_afe_xcvr[phy_id].afe_tx_control);
2106 } 2088 }
2107 2089
2108 udelay(AFE_REGISTER_WRITE_DELAY); 2090 udelay(AFE_REGISTER_WRITE_DELAY);
2109 2091
2110 writel(oem_phy->afe_tx_amp_control0, 2092 writel(oem_phy->afe_tx_amp_control0,
2111 &scic->scu_registers->afe.scu_afe_xcvr[phy_id].afe_tx_amp_control0); 2093 &ihost->scu_registers->afe.scu_afe_xcvr[phy_id].afe_tx_amp_control0);
2112 udelay(AFE_REGISTER_WRITE_DELAY); 2094 udelay(AFE_REGISTER_WRITE_DELAY);
2113 2095
2114 writel(oem_phy->afe_tx_amp_control1, 2096 writel(oem_phy->afe_tx_amp_control1,
2115 &scic->scu_registers->afe.scu_afe_xcvr[phy_id].afe_tx_amp_control1); 2097 &ihost->scu_registers->afe.scu_afe_xcvr[phy_id].afe_tx_amp_control1);
2116 udelay(AFE_REGISTER_WRITE_DELAY); 2098 udelay(AFE_REGISTER_WRITE_DELAY);
2117 2099
2118 writel(oem_phy->afe_tx_amp_control2, 2100 writel(oem_phy->afe_tx_amp_control2,
2119 &scic->scu_registers->afe.scu_afe_xcvr[phy_id].afe_tx_amp_control2); 2101 &ihost->scu_registers->afe.scu_afe_xcvr[phy_id].afe_tx_amp_control2);
2120 udelay(AFE_REGISTER_WRITE_DELAY); 2102 udelay(AFE_REGISTER_WRITE_DELAY);
2121 2103
2122 writel(oem_phy->afe_tx_amp_control3, 2104 writel(oem_phy->afe_tx_amp_control3,
2123 &scic->scu_registers->afe.scu_afe_xcvr[phy_id].afe_tx_amp_control3); 2105 &ihost->scu_registers->afe.scu_afe_xcvr[phy_id].afe_tx_amp_control3);
2124 udelay(AFE_REGISTER_WRITE_DELAY); 2106 udelay(AFE_REGISTER_WRITE_DELAY);
2125 } 2107 }
2126 2108
2127 /* Transfer control to the PEs */ 2109 /* Transfer control to the PEs */
2128 writel(0x00010f00, &scic->scu_registers->afe.afe_dfx_master_control0); 2110 writel(0x00010f00, &ihost->scu_registers->afe.afe_dfx_master_control0);
2129 udelay(AFE_REGISTER_WRITE_DELAY); 2111 udelay(AFE_REGISTER_WRITE_DELAY);
2130} 2112}
2131 2113
2132static void scic_sds_controller_initialize_power_control(struct scic_sds_controller *scic) 2114static void scic_sds_controller_initialize_power_control(struct isci_host *ihost)
2133{ 2115{
2134 sci_init_timer(&scic->power_control.timer, power_control_timeout); 2116 sci_init_timer(&ihost->power_control.timer, power_control_timeout);
2135 2117
2136 memset(scic->power_control.requesters, 0, 2118 memset(ihost->power_control.requesters, 0,
2137 sizeof(scic->power_control.requesters)); 2119 sizeof(ihost->power_control.requesters));
2138 2120
2139 scic->power_control.phys_waiting = 0; 2121 ihost->power_control.phys_waiting = 0;
2140 scic->power_control.phys_granted_power = 0; 2122 ihost->power_control.phys_granted_power = 0;
2141} 2123}
2142 2124
2143static enum sci_status scic_controller_initialize(struct scic_sds_controller *scic) 2125static enum sci_status scic_controller_initialize(struct isci_host *ihost)
2144{ 2126{
2145 struct sci_base_state_machine *sm = &scic->sm; 2127 struct sci_base_state_machine *sm = &ihost->sm;
2146 struct isci_host *ihost = scic_to_ihost(scic);
2147 enum sci_status result = SCI_FAILURE; 2128 enum sci_status result = SCI_FAILURE;
2148 unsigned long i, state, val; 2129 unsigned long i, state, val;
2149 2130
2150 if (scic->sm.current_state_id != SCIC_RESET) { 2131 if (ihost->sm.current_state_id != SCIC_RESET) {
2151 dev_warn(scic_to_dev(scic), 2132 dev_warn(&ihost->pdev->dev,
2152 "SCIC Controller initialize operation requested " 2133 "SCIC Controller initialize operation requested "
2153 "in invalid state\n"); 2134 "in invalid state\n");
2154 return SCI_FAILURE_INVALID_STATE; 2135 return SCI_FAILURE_INVALID_STATE;
@@ -2156,23 +2137,23 @@ static enum sci_status scic_controller_initialize(struct scic_sds_controller *sc
2156 2137
2157 sci_change_state(sm, SCIC_INITIALIZING); 2138 sci_change_state(sm, SCIC_INITIALIZING);
2158 2139
2159 sci_init_timer(&scic->phy_timer, phy_startup_timeout); 2140 sci_init_timer(&ihost->phy_timer, phy_startup_timeout);
2160 2141
2161 scic->next_phy_to_start = 0; 2142 ihost->next_phy_to_start = 0;
2162 scic->phy_startup_timer_pending = false; 2143 ihost->phy_startup_timer_pending = false;
2163 2144
2164 scic_sds_controller_initialize_power_control(scic); 2145 scic_sds_controller_initialize_power_control(ihost);
2165 2146
2166 /* 2147 /*
2167 * There is nothing to do here for B0 since we do not have to 2148 * There is nothing to do here for B0 since we do not have to
2168 * program the AFE registers. 2149 * program the AFE registers.
2169 * / @todo The AFE settings are supposed to be correct for the B0 but 2150 * / @todo The AFE settings are supposed to be correct for the B0 but
2170 * / presently they seem to be wrong. */ 2151 * / presently they seem to be wrong. */
2171 scic_sds_controller_afe_initialization(scic); 2152 scic_sds_controller_afe_initialization(ihost);
2172 2153
2173 2154
2174 /* Take the hardware out of reset */ 2155 /* Take the hardware out of reset */
2175 writel(0, &scic->smu_registers->soft_reset_control); 2156 writel(0, &ihost->smu_registers->soft_reset_control);
2176 2157
2177 /* 2158 /*
2178 * / @todo Provide meaningfull error code for hardware failure 2159 * / @todo Provide meaningfull error code for hardware failure
@@ -2182,7 +2163,7 @@ static enum sci_status scic_controller_initialize(struct scic_sds_controller *sc
2182 2163
2183 /* Loop until the hardware reports success */ 2164 /* Loop until the hardware reports success */
2184 udelay(SCU_CONTEXT_RAM_INIT_STALL_TIME); 2165 udelay(SCU_CONTEXT_RAM_INIT_STALL_TIME);
2185 status = readl(&scic->smu_registers->control_status); 2166 status = readl(&ihost->smu_registers->control_status);
2186 2167
2187 if ((status & SCU_RAM_INIT_COMPLETED) == SCU_RAM_INIT_COMPLETED) 2168 if ((status & SCU_RAM_INIT_COMPLETED) == SCU_RAM_INIT_COMPLETED)
2188 break; 2169 break;
@@ -2193,32 +2174,32 @@ static enum sci_status scic_controller_initialize(struct scic_sds_controller *sc
2193 /* 2174 /*
2194 * Determine what are the actaul device capacities that the 2175 * Determine what are the actaul device capacities that the
2195 * hardware will support */ 2176 * hardware will support */
2196 val = readl(&scic->smu_registers->device_context_capacity); 2177 val = readl(&ihost->smu_registers->device_context_capacity);
2197 2178
2198 /* Record the smaller of the two capacity values */ 2179 /* Record the smaller of the two capacity values */
2199 scic->logical_port_entries = min(smu_max_ports(val), SCI_MAX_PORTS); 2180 ihost->logical_port_entries = min(smu_max_ports(val), SCI_MAX_PORTS);
2200 scic->task_context_entries = min(smu_max_task_contexts(val), SCI_MAX_IO_REQUESTS); 2181 ihost->task_context_entries = min(smu_max_task_contexts(val), SCI_MAX_IO_REQUESTS);
2201 scic->remote_node_entries = min(smu_max_rncs(val), SCI_MAX_REMOTE_DEVICES); 2182 ihost->remote_node_entries = min(smu_max_rncs(val), SCI_MAX_REMOTE_DEVICES);
2202 2183
2203 /* 2184 /*
2204 * Make all PEs that are unassigned match up with the 2185 * Make all PEs that are unassigned match up with the
2205 * logical ports 2186 * logical ports
2206 */ 2187 */
2207 for (i = 0; i < scic->logical_port_entries; i++) { 2188 for (i = 0; i < ihost->logical_port_entries; i++) {
2208 struct scu_port_task_scheduler_group_registers __iomem 2189 struct scu_port_task_scheduler_group_registers __iomem
2209 *ptsg = &scic->scu_registers->peg0.ptsg; 2190 *ptsg = &ihost->scu_registers->peg0.ptsg;
2210 2191
2211 writel(i, &ptsg->protocol_engine[i]); 2192 writel(i, &ptsg->protocol_engine[i]);
2212 } 2193 }
2213 2194
2214 /* Initialize hardware PCI Relaxed ordering in DMA engines */ 2195 /* Initialize hardware PCI Relaxed ordering in DMA engines */
2215 val = readl(&scic->scu_registers->sdma.pdma_configuration); 2196 val = readl(&ihost->scu_registers->sdma.pdma_configuration);
2216 val |= SCU_PDMACR_GEN_BIT(PCI_RELAXED_ORDERING_ENABLE); 2197 val |= SCU_PDMACR_GEN_BIT(PCI_RELAXED_ORDERING_ENABLE);
2217 writel(val, &scic->scu_registers->sdma.pdma_configuration); 2198 writel(val, &ihost->scu_registers->sdma.pdma_configuration);
2218 2199
2219 val = readl(&scic->scu_registers->sdma.cdma_configuration); 2200 val = readl(&ihost->scu_registers->sdma.cdma_configuration);
2220 val |= SCU_CDMACR_GEN_BIT(PCI_RELAXED_ORDERING_ENABLE); 2201 val |= SCU_CDMACR_GEN_BIT(PCI_RELAXED_ORDERING_ENABLE);
2221 writel(val, &scic->scu_registers->sdma.cdma_configuration); 2202 writel(val, &ihost->scu_registers->sdma.cdma_configuration);
2222 2203
2223 /* 2204 /*
2224 * Initialize the PHYs before the PORTs because the PHY registers 2205 * Initialize the PHYs before the PORTs because the PHY registers
@@ -2226,23 +2207,23 @@ static enum sci_status scic_controller_initialize(struct scic_sds_controller *sc
2226 */ 2207 */
2227 for (i = 0; i < SCI_MAX_PHYS; i++) { 2208 for (i = 0; i < SCI_MAX_PHYS; i++) {
2228 result = scic_sds_phy_initialize(&ihost->phys[i], 2209 result = scic_sds_phy_initialize(&ihost->phys[i],
2229 &scic->scu_registers->peg0.pe[i].tl, 2210 &ihost->scu_registers->peg0.pe[i].tl,
2230 &scic->scu_registers->peg0.pe[i].ll); 2211 &ihost->scu_registers->peg0.pe[i].ll);
2231 if (result != SCI_SUCCESS) 2212 if (result != SCI_SUCCESS)
2232 goto out; 2213 goto out;
2233 } 2214 }
2234 2215
2235 for (i = 0; i < scic->logical_port_entries; i++) { 2216 for (i = 0; i < ihost->logical_port_entries; i++) {
2236 result = scic_sds_port_initialize(&ihost->ports[i], 2217 result = scic_sds_port_initialize(&ihost->ports[i],
2237 &scic->scu_registers->peg0.ptsg.port[i], 2218 &ihost->scu_registers->peg0.ptsg.port[i],
2238 &scic->scu_registers->peg0.ptsg.protocol_engine, 2219 &ihost->scu_registers->peg0.ptsg.protocol_engine,
2239 &scic->scu_registers->peg0.viit[i]); 2220 &ihost->scu_registers->peg0.viit[i]);
2240 2221
2241 if (result != SCI_SUCCESS) 2222 if (result != SCI_SUCCESS)
2242 goto out; 2223 goto out;
2243 } 2224 }
2244 2225
2245 result = scic_sds_port_configuration_agent_initialize(scic, &scic->port_agent); 2226 result = scic_sds_port_configuration_agent_initialize(ihost, &ihost->port_agent);
2246 2227
2247 out: 2228 out:
2248 /* Advance the controller state machine */ 2229 /* Advance the controller state machine */
@@ -2256,10 +2237,10 @@ static enum sci_status scic_controller_initialize(struct scic_sds_controller *sc
2256} 2237}
2257 2238
2258static enum sci_status scic_user_parameters_set( 2239static enum sci_status scic_user_parameters_set(
2259 struct scic_sds_controller *scic, 2240 struct isci_host *ihost,
2260 union scic_user_parameters *scic_parms) 2241 union scic_user_parameters *scic_parms)
2261{ 2242{
2262 u32 state = scic->sm.current_state_id; 2243 u32 state = ihost->sm.current_state_id;
2263 2244
2264 if (state == SCIC_RESET || 2245 if (state == SCIC_RESET ||
2265 state == SCIC_INITIALIZING || 2246 state == SCIC_INITIALIZING ||
@@ -2301,7 +2282,7 @@ static enum sci_status scic_user_parameters_set(
2301 (scic_parms->sds1.no_outbound_task_timeout == 0)) 2282 (scic_parms->sds1.no_outbound_task_timeout == 0))
2302 return SCI_FAILURE_INVALID_PARAMETER_VALUE; 2283 return SCI_FAILURE_INVALID_PARAMETER_VALUE;
2303 2284
2304 memcpy(&scic->user_parameters, scic_parms, sizeof(*scic_parms)); 2285 memcpy(&ihost->user_parameters, scic_parms, sizeof(*scic_parms));
2305 2286
2306 return SCI_SUCCESS; 2287 return SCI_SUCCESS;
2307 } 2288 }
@@ -2309,40 +2290,40 @@ static enum sci_status scic_user_parameters_set(
2309 return SCI_FAILURE_INVALID_STATE; 2290 return SCI_FAILURE_INVALID_STATE;
2310} 2291}
2311 2292
2312static int scic_controller_mem_init(struct scic_sds_controller *scic) 2293static int scic_controller_mem_init(struct isci_host *ihost)
2313{ 2294{
2314 struct device *dev = scic_to_dev(scic); 2295 struct device *dev = &ihost->pdev->dev;
2315 dma_addr_t dma; 2296 dma_addr_t dma;
2316 size_t size; 2297 size_t size;
2317 int err; 2298 int err;
2318 2299
2319 size = SCU_MAX_COMPLETION_QUEUE_ENTRIES * sizeof(u32); 2300 size = SCU_MAX_COMPLETION_QUEUE_ENTRIES * sizeof(u32);
2320 scic->completion_queue = dmam_alloc_coherent(dev, size, &dma, GFP_KERNEL); 2301 ihost->completion_queue = dmam_alloc_coherent(dev, size, &dma, GFP_KERNEL);
2321 if (!scic->completion_queue) 2302 if (!ihost->completion_queue)
2322 return -ENOMEM; 2303 return -ENOMEM;
2323 2304
2324 writel(lower_32_bits(dma), &scic->smu_registers->completion_queue_lower); 2305 writel(lower_32_bits(dma), &ihost->smu_registers->completion_queue_lower);
2325 writel(upper_32_bits(dma), &scic->smu_registers->completion_queue_upper); 2306 writel(upper_32_bits(dma), &ihost->smu_registers->completion_queue_upper);
2326 2307
2327 size = scic->remote_node_entries * sizeof(union scu_remote_node_context); 2308 size = ihost->remote_node_entries * sizeof(union scu_remote_node_context);
2328 scic->remote_node_context_table = dmam_alloc_coherent(dev, size, &dma, 2309 ihost->remote_node_context_table = dmam_alloc_coherent(dev, size, &dma,
2329 GFP_KERNEL); 2310 GFP_KERNEL);
2330 if (!scic->remote_node_context_table) 2311 if (!ihost->remote_node_context_table)
2331 return -ENOMEM; 2312 return -ENOMEM;
2332 2313
2333 writel(lower_32_bits(dma), &scic->smu_registers->remote_node_context_lower); 2314 writel(lower_32_bits(dma), &ihost->smu_registers->remote_node_context_lower);
2334 writel(upper_32_bits(dma), &scic->smu_registers->remote_node_context_upper); 2315 writel(upper_32_bits(dma), &ihost->smu_registers->remote_node_context_upper);
2335 2316
2336 size = scic->task_context_entries * sizeof(struct scu_task_context), 2317 size = ihost->task_context_entries * sizeof(struct scu_task_context),
2337 scic->task_context_table = dmam_alloc_coherent(dev, size, &dma, GFP_KERNEL); 2318 ihost->task_context_table = dmam_alloc_coherent(dev, size, &dma, GFP_KERNEL);
2338 if (!scic->task_context_table) 2319 if (!ihost->task_context_table)
2339 return -ENOMEM; 2320 return -ENOMEM;
2340 2321
2341 scic->task_context_dma = dma; 2322 ihost->task_context_dma = dma;
2342 writel(lower_32_bits(dma), &scic->smu_registers->host_task_table_lower); 2323 writel(lower_32_bits(dma), &ihost->smu_registers->host_task_table_lower);
2343 writel(upper_32_bits(dma), &scic->smu_registers->host_task_table_upper); 2324 writel(upper_32_bits(dma), &ihost->smu_registers->host_task_table_upper);
2344 2325
2345 err = scic_sds_unsolicited_frame_control_construct(scic); 2326 err = scic_sds_unsolicited_frame_control_construct(ihost);
2346 if (err) 2327 if (err)
2347 return err; 2328 return err;
2348 2329
@@ -2350,112 +2331,112 @@ static int scic_controller_mem_init(struct scic_sds_controller *scic)
2350 * Inform the silicon as to the location of the UF headers and 2331 * Inform the silicon as to the location of the UF headers and
2351 * address table. 2332 * address table.
2352 */ 2333 */
2353 writel(lower_32_bits(scic->uf_control.headers.physical_address), 2334 writel(lower_32_bits(ihost->uf_control.headers.physical_address),
2354 &scic->scu_registers->sdma.uf_header_base_address_lower); 2335 &ihost->scu_registers->sdma.uf_header_base_address_lower);
2355 writel(upper_32_bits(scic->uf_control.headers.physical_address), 2336 writel(upper_32_bits(ihost->uf_control.headers.physical_address),
2356 &scic->scu_registers->sdma.uf_header_base_address_upper); 2337 &ihost->scu_registers->sdma.uf_header_base_address_upper);
2357 2338
2358 writel(lower_32_bits(scic->uf_control.address_table.physical_address), 2339 writel(lower_32_bits(ihost->uf_control.address_table.physical_address),
2359 &scic->scu_registers->sdma.uf_address_table_lower); 2340 &ihost->scu_registers->sdma.uf_address_table_lower);
2360 writel(upper_32_bits(scic->uf_control.address_table.physical_address), 2341 writel(upper_32_bits(ihost->uf_control.address_table.physical_address),
2361 &scic->scu_registers->sdma.uf_address_table_upper); 2342 &ihost->scu_registers->sdma.uf_address_table_upper);
2362 2343
2363 return 0; 2344 return 0;
2364} 2345}
2365 2346
2366int isci_host_init(struct isci_host *isci_host) 2347int isci_host_init(struct isci_host *ihost)
2367{ 2348{
2368 int err = 0, i; 2349 int err = 0, i;
2369 enum sci_status status; 2350 enum sci_status status;
2370 union scic_oem_parameters oem; 2351 union scic_oem_parameters oem;
2371 union scic_user_parameters scic_user_params; 2352 union scic_user_parameters scic_user_params;
2372 struct isci_pci_info *pci_info = to_pci_info(isci_host->pdev); 2353 struct isci_pci_info *pci_info = to_pci_info(ihost->pdev);
2373 2354
2374 spin_lock_init(&isci_host->state_lock); 2355 spin_lock_init(&ihost->state_lock);
2375 spin_lock_init(&isci_host->scic_lock); 2356 spin_lock_init(&ihost->scic_lock);
2376 init_waitqueue_head(&isci_host->eventq); 2357 init_waitqueue_head(&ihost->eventq);
2377 2358
2378 isci_host_change_state(isci_host, isci_starting); 2359 isci_host_change_state(ihost, isci_starting);
2379 2360
2380 status = scic_controller_construct(&isci_host->sci, scu_base(isci_host), 2361 status = scic_controller_construct(ihost, scu_base(ihost),
2381 smu_base(isci_host)); 2362 smu_base(ihost));
2382 2363
2383 if (status != SCI_SUCCESS) { 2364 if (status != SCI_SUCCESS) {
2384 dev_err(&isci_host->pdev->dev, 2365 dev_err(&ihost->pdev->dev,
2385 "%s: scic_controller_construct failed - status = %x\n", 2366 "%s: scic_controller_construct failed - status = %x\n",
2386 __func__, 2367 __func__,
2387 status); 2368 status);
2388 return -ENODEV; 2369 return -ENODEV;
2389 } 2370 }
2390 2371
2391 isci_host->sas_ha.dev = &isci_host->pdev->dev; 2372 ihost->sas_ha.dev = &ihost->pdev->dev;
2392 isci_host->sas_ha.lldd_ha = isci_host; 2373 ihost->sas_ha.lldd_ha = ihost;
2393 2374
2394 /* 2375 /*
2395 * grab initial values stored in the controller object for OEM and USER 2376 * grab initial values stored in the controller object for OEM and USER
2396 * parameters 2377 * parameters
2397 */ 2378 */
2398 isci_user_parameters_get(isci_host, &scic_user_params); 2379 isci_user_parameters_get(ihost, &scic_user_params);
2399 status = scic_user_parameters_set(&isci_host->sci, 2380 status = scic_user_parameters_set(ihost,
2400 &scic_user_params); 2381 &scic_user_params);
2401 if (status != SCI_SUCCESS) { 2382 if (status != SCI_SUCCESS) {
2402 dev_warn(&isci_host->pdev->dev, 2383 dev_warn(&ihost->pdev->dev,
2403 "%s: scic_user_parameters_set failed\n", 2384 "%s: scic_user_parameters_set failed\n",
2404 __func__); 2385 __func__);
2405 return -ENODEV; 2386 return -ENODEV;
2406 } 2387 }
2407 2388
2408 scic_oem_parameters_get(&isci_host->sci, &oem); 2389 scic_oem_parameters_get(ihost, &oem);
2409 2390
2410 /* grab any OEM parameters specified in orom */ 2391 /* grab any OEM parameters specified in orom */
2411 if (pci_info->orom) { 2392 if (pci_info->orom) {
2412 status = isci_parse_oem_parameters(&oem, 2393 status = isci_parse_oem_parameters(&oem,
2413 pci_info->orom, 2394 pci_info->orom,
2414 isci_host->id); 2395 ihost->id);
2415 if (status != SCI_SUCCESS) { 2396 if (status != SCI_SUCCESS) {
2416 dev_warn(&isci_host->pdev->dev, 2397 dev_warn(&ihost->pdev->dev,
2417 "parsing firmware oem parameters failed\n"); 2398 "parsing firmware oem parameters failed\n");
2418 return -EINVAL; 2399 return -EINVAL;
2419 } 2400 }
2420 } 2401 }
2421 2402
2422 status = scic_oem_parameters_set(&isci_host->sci, &oem); 2403 status = scic_oem_parameters_set(ihost, &oem);
2423 if (status != SCI_SUCCESS) { 2404 if (status != SCI_SUCCESS) {
2424 dev_warn(&isci_host->pdev->dev, 2405 dev_warn(&ihost->pdev->dev,
2425 "%s: scic_oem_parameters_set failed\n", 2406 "%s: scic_oem_parameters_set failed\n",
2426 __func__); 2407 __func__);
2427 return -ENODEV; 2408 return -ENODEV;
2428 } 2409 }
2429 2410
2430 tasklet_init(&isci_host->completion_tasklet, 2411 tasklet_init(&ihost->completion_tasklet,
2431 isci_host_completion_routine, (unsigned long)isci_host); 2412 isci_host_completion_routine, (unsigned long)ihost);
2432 2413
2433 INIT_LIST_HEAD(&isci_host->requests_to_complete); 2414 INIT_LIST_HEAD(&ihost->requests_to_complete);
2434 INIT_LIST_HEAD(&isci_host->requests_to_errorback); 2415 INIT_LIST_HEAD(&ihost->requests_to_errorback);
2435 2416
2436 spin_lock_irq(&isci_host->scic_lock); 2417 spin_lock_irq(&ihost->scic_lock);
2437 status = scic_controller_initialize(&isci_host->sci); 2418 status = scic_controller_initialize(ihost);
2438 spin_unlock_irq(&isci_host->scic_lock); 2419 spin_unlock_irq(&ihost->scic_lock);
2439 if (status != SCI_SUCCESS) { 2420 if (status != SCI_SUCCESS) {
2440 dev_warn(&isci_host->pdev->dev, 2421 dev_warn(&ihost->pdev->dev,
2441 "%s: scic_controller_initialize failed -" 2422 "%s: scic_controller_initialize failed -"
2442 " status = 0x%x\n", 2423 " status = 0x%x\n",
2443 __func__, status); 2424 __func__, status);
2444 return -ENODEV; 2425 return -ENODEV;
2445 } 2426 }
2446 2427
2447 err = scic_controller_mem_init(&isci_host->sci); 2428 err = scic_controller_mem_init(ihost);
2448 if (err) 2429 if (err)
2449 return err; 2430 return err;
2450 2431
2451 for (i = 0; i < SCI_MAX_PORTS; i++) 2432 for (i = 0; i < SCI_MAX_PORTS; i++)
2452 isci_port_init(&isci_host->ports[i], isci_host, i); 2433 isci_port_init(&ihost->ports[i], ihost, i);
2453 2434
2454 for (i = 0; i < SCI_MAX_PHYS; i++) 2435 for (i = 0; i < SCI_MAX_PHYS; i++)
2455 isci_phy_init(&isci_host->phys[i], isci_host, i); 2436 isci_phy_init(&ihost->phys[i], ihost, i);
2456 2437
2457 for (i = 0; i < SCI_MAX_REMOTE_DEVICES; i++) { 2438 for (i = 0; i < SCI_MAX_REMOTE_DEVICES; i++) {
2458 struct isci_remote_device *idev = &isci_host->devices[i]; 2439 struct isci_remote_device *idev = &ihost->devices[i];
2459 2440
2460 INIT_LIST_HEAD(&idev->reqs_in_process); 2441 INIT_LIST_HEAD(&idev->reqs_in_process);
2461 INIT_LIST_HEAD(&idev->node); 2442 INIT_LIST_HEAD(&idev->node);
@@ -2465,63 +2446,62 @@ int isci_host_init(struct isci_host *isci_host)
2465 struct isci_request *ireq; 2446 struct isci_request *ireq;
2466 dma_addr_t dma; 2447 dma_addr_t dma;
2467 2448
2468 ireq = dmam_alloc_coherent(&isci_host->pdev->dev, 2449 ireq = dmam_alloc_coherent(&ihost->pdev->dev,
2469 sizeof(struct isci_request), &dma, 2450 sizeof(struct isci_request), &dma,
2470 GFP_KERNEL); 2451 GFP_KERNEL);
2471 if (!ireq) 2452 if (!ireq)
2472 return -ENOMEM; 2453 return -ENOMEM;
2473 2454
2474 ireq->tc = &isci_host->sci.task_context_table[i]; 2455 ireq->tc = &ihost->task_context_table[i];
2475 ireq->owning_controller = &isci_host->sci; 2456 ireq->owning_controller = ihost;
2476 spin_lock_init(&ireq->state_lock); 2457 spin_lock_init(&ireq->state_lock);
2477 ireq->request_daddr = dma; 2458 ireq->request_daddr = dma;
2478 ireq->isci_host = isci_host; 2459 ireq->isci_host = ihost;
2479 2460 ihost->reqs[i] = ireq;
2480 isci_host->reqs[i] = ireq;
2481 } 2461 }
2482 2462
2483 return 0; 2463 return 0;
2484} 2464}
2485 2465
2486void scic_sds_controller_link_up(struct scic_sds_controller *scic, 2466void scic_sds_controller_link_up(struct isci_host *ihost,
2487 struct isci_port *iport, struct isci_phy *iphy) 2467 struct isci_port *iport, struct isci_phy *iphy)
2488{ 2468{
2489 switch (scic->sm.current_state_id) { 2469 switch (ihost->sm.current_state_id) {
2490 case SCIC_STARTING: 2470 case SCIC_STARTING:
2491 sci_del_timer(&scic->phy_timer); 2471 sci_del_timer(&ihost->phy_timer);
2492 scic->phy_startup_timer_pending = false; 2472 ihost->phy_startup_timer_pending = false;
2493 scic->port_agent.link_up_handler(scic, &scic->port_agent, 2473 ihost->port_agent.link_up_handler(ihost, &ihost->port_agent,
2494 iport, iphy); 2474 iport, iphy);
2495 scic_sds_controller_start_next_phy(scic); 2475 scic_sds_controller_start_next_phy(ihost);
2496 break; 2476 break;
2497 case SCIC_READY: 2477 case SCIC_READY:
2498 scic->port_agent.link_up_handler(scic, &scic->port_agent, 2478 ihost->port_agent.link_up_handler(ihost, &ihost->port_agent,
2499 iport, iphy); 2479 iport, iphy);
2500 break; 2480 break;
2501 default: 2481 default:
2502 dev_dbg(scic_to_dev(scic), 2482 dev_dbg(&ihost->pdev->dev,
2503 "%s: SCIC Controller linkup event from phy %d in " 2483 "%s: SCIC Controller linkup event from phy %d in "
2504 "unexpected state %d\n", __func__, iphy->phy_index, 2484 "unexpected state %d\n", __func__, iphy->phy_index,
2505 scic->sm.current_state_id); 2485 ihost->sm.current_state_id);
2506 } 2486 }
2507} 2487}
2508 2488
2509void scic_sds_controller_link_down(struct scic_sds_controller *scic, 2489void scic_sds_controller_link_down(struct isci_host *ihost,
2510 struct isci_port *iport, struct isci_phy *iphy) 2490 struct isci_port *iport, struct isci_phy *iphy)
2511{ 2491{
2512 switch (scic->sm.current_state_id) { 2492 switch (ihost->sm.current_state_id) {
2513 case SCIC_STARTING: 2493 case SCIC_STARTING:
2514 case SCIC_READY: 2494 case SCIC_READY:
2515 scic->port_agent.link_down_handler(scic, &scic->port_agent, 2495 ihost->port_agent.link_down_handler(ihost, &ihost->port_agent,
2516 iport, iphy); 2496 iport, iphy);
2517 break; 2497 break;
2518 default: 2498 default:
2519 dev_dbg(scic_to_dev(scic), 2499 dev_dbg(&ihost->pdev->dev,
2520 "%s: SCIC Controller linkdown event from phy %d in " 2500 "%s: SCIC Controller linkdown event from phy %d in "
2521 "unexpected state %d\n", 2501 "unexpected state %d\n",
2522 __func__, 2502 __func__,
2523 iphy->phy_index, 2503 iphy->phy_index,
2524 scic->sm.current_state_id); 2504 ihost->sm.current_state_id);
2525 } 2505 }
2526} 2506}
2527 2507
@@ -2530,14 +2510,13 @@ void scic_sds_controller_link_down(struct scic_sds_controller *scic,
2530 * controller are still in the stopping state. 2510 * controller are still in the stopping state.
2531 * 2511 *
2532 */ 2512 */
2533static bool scic_sds_controller_has_remote_devices_stopping( 2513static bool scic_sds_controller_has_remote_devices_stopping(struct isci_host *ihost)
2534 struct scic_sds_controller *controller)
2535{ 2514{
2536 u32 index; 2515 u32 index;
2537 2516
2538 for (index = 0; index < controller->remote_node_entries; index++) { 2517 for (index = 0; index < ihost->remote_node_entries; index++) {
2539 if ((controller->device_table[index] != NULL) && 2518 if ((ihost->device_table[index] != NULL) &&
2540 (controller->device_table[index]->sm.current_state_id == SCI_DEV_STOPPING)) 2519 (ihost->device_table[index]->sm.current_state_id == SCI_DEV_STOPPING))
2541 return true; 2520 return true;
2542 } 2521 }
2543 2522
@@ -2548,20 +2527,20 @@ static bool scic_sds_controller_has_remote_devices_stopping(
2548 * This method is called by the remote device to inform the controller 2527 * This method is called by the remote device to inform the controller
2549 * object that the remote device has stopped. 2528 * object that the remote device has stopped.
2550 */ 2529 */
2551void scic_sds_controller_remote_device_stopped(struct scic_sds_controller *scic, 2530void scic_sds_controller_remote_device_stopped(struct isci_host *ihost,
2552 struct isci_remote_device *idev) 2531 struct isci_remote_device *idev)
2553{ 2532{
2554 if (scic->sm.current_state_id != SCIC_STOPPING) { 2533 if (ihost->sm.current_state_id != SCIC_STOPPING) {
2555 dev_dbg(scic_to_dev(scic), 2534 dev_dbg(&ihost->pdev->dev,
2556 "SCIC Controller 0x%p remote device stopped event " 2535 "SCIC Controller 0x%p remote device stopped event "
2557 "from device 0x%p in unexpected state %d\n", 2536 "from device 0x%p in unexpected state %d\n",
2558 scic, idev, 2537 ihost, idev,
2559 scic->sm.current_state_id); 2538 ihost->sm.current_state_id);
2560 return; 2539 return;
2561 } 2540 }
2562 2541
2563 if (!scic_sds_controller_has_remote_devices_stopping(scic)) { 2542 if (!scic_sds_controller_has_remote_devices_stopping(ihost)) {
2564 sci_change_state(&scic->sm, SCIC_STOPPED); 2543 sci_change_state(&ihost->sm, SCIC_STOPPED);
2565 } 2544 }
2566} 2545}
2567 2546
@@ -2573,32 +2552,32 @@ void scic_sds_controller_remote_device_stopped(struct scic_sds_controller *scic,
2573 * 2552 *
2574 */ 2553 */
2575void scic_sds_controller_post_request( 2554void scic_sds_controller_post_request(
2576 struct scic_sds_controller *scic, 2555 struct isci_host *ihost,
2577 u32 request) 2556 u32 request)
2578{ 2557{
2579 dev_dbg(scic_to_dev(scic), 2558 dev_dbg(&ihost->pdev->dev,
2580 "%s: SCIC Controller 0x%p post request 0x%08x\n", 2559 "%s: SCIC Controller 0x%p post request 0x%08x\n",
2581 __func__, 2560 __func__,
2582 scic, 2561 ihost,
2583 request); 2562 request);
2584 2563
2585 writel(request, &scic->smu_registers->post_context_port); 2564 writel(request, &ihost->smu_registers->post_context_port);
2586} 2565}
2587 2566
2588struct isci_request *scic_request_by_tag(struct scic_sds_controller *scic, u16 io_tag) 2567struct isci_request *scic_request_by_tag(struct isci_host *ihost, u16 io_tag)
2589{ 2568{
2590 u16 task_index; 2569 u16 task_index;
2591 u16 task_sequence; 2570 u16 task_sequence;
2592 2571
2593 task_index = ISCI_TAG_TCI(io_tag); 2572 task_index = ISCI_TAG_TCI(io_tag);
2594 2573
2595 if (task_index < scic->task_context_entries) { 2574 if (task_index < ihost->task_context_entries) {
2596 struct isci_request *ireq = scic_to_ihost(scic)->reqs[task_index]; 2575 struct isci_request *ireq = ihost->reqs[task_index];
2597 2576
2598 if (test_bit(IREQ_ACTIVE, &ireq->flags)) { 2577 if (test_bit(IREQ_ACTIVE, &ireq->flags)) {
2599 task_sequence = ISCI_TAG_SEQ(io_tag); 2578 task_sequence = ISCI_TAG_SEQ(io_tag);
2600 2579
2601 if (task_sequence == scic->io_request_sequence[task_index]) 2580 if (task_sequence == ihost->io_request_sequence[task_index])
2602 return ireq; 2581 return ireq;
2603 } 2582 }
2604 } 2583 }
@@ -2621,7 +2600,7 @@ struct isci_request *scic_request_by_tag(struct scic_sds_controller *scic, u16 i
2621 * node index available. 2600 * node index available.
2622 */ 2601 */
2623enum sci_status scic_sds_controller_allocate_remote_node_context( 2602enum sci_status scic_sds_controller_allocate_remote_node_context(
2624 struct scic_sds_controller *scic, 2603 struct isci_host *ihost,
2625 struct isci_remote_device *idev, 2604 struct isci_remote_device *idev,
2626 u16 *node_id) 2605 u16 *node_id)
2627{ 2606{
@@ -2629,11 +2608,11 @@ enum sci_status scic_sds_controller_allocate_remote_node_context(
2629 u32 remote_node_count = scic_sds_remote_device_node_count(idev); 2608 u32 remote_node_count = scic_sds_remote_device_node_count(idev);
2630 2609
2631 node_index = scic_sds_remote_node_table_allocate_remote_node( 2610 node_index = scic_sds_remote_node_table_allocate_remote_node(
2632 &scic->available_remote_nodes, remote_node_count 2611 &ihost->available_remote_nodes, remote_node_count
2633 ); 2612 );
2634 2613
2635 if (node_index != SCIC_SDS_REMOTE_NODE_CONTEXT_INVALID_INDEX) { 2614 if (node_index != SCIC_SDS_REMOTE_NODE_CONTEXT_INVALID_INDEX) {
2636 scic->device_table[node_index] = idev; 2615 ihost->device_table[node_index] = idev;
2637 2616
2638 *node_id = node_index; 2617 *node_id = node_index;
2639 2618
@@ -2653,17 +2632,17 @@ enum sci_status scic_sds_controller_allocate_remote_node_context(
2653 * 2632 *
2654 */ 2633 */
2655void scic_sds_controller_free_remote_node_context( 2634void scic_sds_controller_free_remote_node_context(
2656 struct scic_sds_controller *scic, 2635 struct isci_host *ihost,
2657 struct isci_remote_device *idev, 2636 struct isci_remote_device *idev,
2658 u16 node_id) 2637 u16 node_id)
2659{ 2638{
2660 u32 remote_node_count = scic_sds_remote_device_node_count(idev); 2639 u32 remote_node_count = scic_sds_remote_device_node_count(idev);
2661 2640
2662 if (scic->device_table[node_id] == idev) { 2641 if (ihost->device_table[node_id] == idev) {
2663 scic->device_table[node_id] = NULL; 2642 ihost->device_table[node_id] = NULL;
2664 2643
2665 scic_sds_remote_node_table_release_remote_node_index( 2644 scic_sds_remote_node_table_release_remote_node_index(
2666 &scic->available_remote_nodes, remote_node_count, node_id 2645 &ihost->available_remote_nodes, remote_node_count, node_id
2667 ); 2646 );
2668 } 2647 }
2669} 2648}
@@ -2677,14 +2656,14 @@ void scic_sds_controller_free_remote_node_context(
2677 * union scu_remote_node_context* 2656 * union scu_remote_node_context*
2678 */ 2657 */
2679union scu_remote_node_context *scic_sds_controller_get_remote_node_context_buffer( 2658union scu_remote_node_context *scic_sds_controller_get_remote_node_context_buffer(
2680 struct scic_sds_controller *scic, 2659 struct isci_host *ihost,
2681 u16 node_id 2660 u16 node_id
2682 ) { 2661 ) {
2683 if ( 2662 if (
2684 (node_id < scic->remote_node_entries) 2663 (node_id < ihost->remote_node_entries)
2685 && (scic->device_table[node_id] != NULL) 2664 && (ihost->device_table[node_id] != NULL)
2686 ) { 2665 ) {
2687 return &scic->remote_node_context_table[node_id]; 2666 return &ihost->remote_node_context_table[node_id];
2688 } 2667 }
2689 2668
2690 return NULL; 2669 return NULL;
@@ -2722,13 +2701,13 @@ void scic_sds_controller_copy_sata_response(
2722 * 2701 *
2723 */ 2702 */
2724void scic_sds_controller_release_frame( 2703void scic_sds_controller_release_frame(
2725 struct scic_sds_controller *scic, 2704 struct isci_host *ihost,
2726 u32 frame_index) 2705 u32 frame_index)
2727{ 2706{
2728 if (scic_sds_unsolicited_frame_control_release_frame( 2707 if (scic_sds_unsolicited_frame_control_release_frame(
2729 &scic->uf_control, frame_index) == true) 2708 &ihost->uf_control, frame_index) == true)
2730 writel(scic->uf_control.get, 2709 writel(ihost->uf_control.get,
2731 &scic->scu_registers->sdma.unsolicited_frame_get_pointer); 2710 &ihost->scu_registers->sdma.unsolicited_frame_get_pointer);
2732} 2711}
2733 2712
2734void isci_tci_free(struct isci_host *ihost, u16 tci) 2713void isci_tci_free(struct isci_host *ihost, u16 tci)
@@ -2757,7 +2736,7 @@ u16 isci_alloc_tag(struct isci_host *ihost)
2757{ 2736{
2758 if (isci_tci_space(ihost)) { 2737 if (isci_tci_space(ihost)) {
2759 u16 tci = isci_tci_alloc(ihost); 2738 u16 tci = isci_tci_alloc(ihost);
2760 u8 seq = ihost->sci.io_request_sequence[tci]; 2739 u8 seq = ihost->io_request_sequence[tci];
2761 2740
2762 return ISCI_TAG(seq, tci); 2741 return ISCI_TAG(seq, tci);
2763 } 2742 }
@@ -2767,7 +2746,6 @@ u16 isci_alloc_tag(struct isci_host *ihost)
2767 2746
2768enum sci_status isci_free_tag(struct isci_host *ihost, u16 io_tag) 2747enum sci_status isci_free_tag(struct isci_host *ihost, u16 io_tag)
2769{ 2748{
2770 struct scic_sds_controller *scic = &ihost->sci;
2771 u16 tci = ISCI_TAG_TCI(io_tag); 2749 u16 tci = ISCI_TAG_TCI(io_tag);
2772 u16 seq = ISCI_TAG_SEQ(io_tag); 2750 u16 seq = ISCI_TAG_SEQ(io_tag);
2773 2751
@@ -2775,8 +2753,8 @@ enum sci_status isci_free_tag(struct isci_host *ihost, u16 io_tag)
2775 if (isci_tci_active(ihost) == 0) 2753 if (isci_tci_active(ihost) == 0)
2776 return SCI_FAILURE_INVALID_IO_TAG; 2754 return SCI_FAILURE_INVALID_IO_TAG;
2777 2755
2778 if (seq == scic->io_request_sequence[tci]) { 2756 if (seq == ihost->io_request_sequence[tci]) {
2779 scic->io_request_sequence[tci] = (seq+1) & (SCI_MAX_SEQ-1); 2757 ihost->io_request_sequence[tci] = (seq+1) & (SCI_MAX_SEQ-1);
2780 2758
2781 isci_tci_free(ihost, tci); 2759 isci_tci_free(ihost, tci);
2782 2760
@@ -2797,23 +2775,23 @@ enum sci_status isci_free_tag(struct isci_host *ihost, u16 io_tag)
2797 * @io_tag: This parameter specifies a previously allocated IO tag that the 2775 * @io_tag: This parameter specifies a previously allocated IO tag that the
2798 * user desires to be utilized for this request. 2776 * user desires to be utilized for this request.
2799 */ 2777 */
2800enum sci_status scic_controller_start_io(struct scic_sds_controller *scic, 2778enum sci_status scic_controller_start_io(struct isci_host *ihost,
2801 struct isci_remote_device *idev, 2779 struct isci_remote_device *idev,
2802 struct isci_request *ireq) 2780 struct isci_request *ireq)
2803{ 2781{
2804 enum sci_status status; 2782 enum sci_status status;
2805 2783
2806 if (scic->sm.current_state_id != SCIC_READY) { 2784 if (ihost->sm.current_state_id != SCIC_READY) {
2807 dev_warn(scic_to_dev(scic), "invalid state to start I/O"); 2785 dev_warn(&ihost->pdev->dev, "invalid state to start I/O");
2808 return SCI_FAILURE_INVALID_STATE; 2786 return SCI_FAILURE_INVALID_STATE;
2809 } 2787 }
2810 2788
2811 status = scic_sds_remote_device_start_io(scic, idev, ireq); 2789 status = scic_sds_remote_device_start_io(ihost, idev, ireq);
2812 if (status != SCI_SUCCESS) 2790 if (status != SCI_SUCCESS)
2813 return status; 2791 return status;
2814 2792
2815 set_bit(IREQ_ACTIVE, &ireq->flags); 2793 set_bit(IREQ_ACTIVE, &ireq->flags);
2816 scic_sds_controller_post_request(scic, scic_sds_request_get_post_context(ireq)); 2794 scic_sds_controller_post_request(ihost, scic_sds_request_get_post_context(ireq));
2817 return SCI_SUCCESS; 2795 return SCI_SUCCESS;
2818} 2796}
2819 2797
@@ -2834,14 +2812,14 @@ enum sci_status scic_controller_start_io(struct scic_sds_controller *scic,
2834 * for the request. Determine the failure situations and return values. 2812 * for the request. Determine the failure situations and return values.
2835 */ 2813 */
2836enum sci_status scic_controller_terminate_request( 2814enum sci_status scic_controller_terminate_request(
2837 struct scic_sds_controller *scic, 2815 struct isci_host *ihost,
2838 struct isci_remote_device *idev, 2816 struct isci_remote_device *idev,
2839 struct isci_request *ireq) 2817 struct isci_request *ireq)
2840{ 2818{
2841 enum sci_status status; 2819 enum sci_status status;
2842 2820
2843 if (scic->sm.current_state_id != SCIC_READY) { 2821 if (ihost->sm.current_state_id != SCIC_READY) {
2844 dev_warn(scic_to_dev(scic), 2822 dev_warn(&ihost->pdev->dev,
2845 "invalid state to terminate request\n"); 2823 "invalid state to terminate request\n");
2846 return SCI_FAILURE_INVALID_STATE; 2824 return SCI_FAILURE_INVALID_STATE;
2847 } 2825 }
@@ -2854,7 +2832,7 @@ enum sci_status scic_controller_terminate_request(
2854 * Utilize the original post context command and or in the POST_TC_ABORT 2832 * Utilize the original post context command and or in the POST_TC_ABORT
2855 * request sub-type. 2833 * request sub-type.
2856 */ 2834 */
2857 scic_sds_controller_post_request(scic, 2835 scic_sds_controller_post_request(ihost,
2858 scic_sds_request_get_post_context(ireq) | 2836 scic_sds_request_get_post_context(ireq) |
2859 SCU_CONTEXT_COMMAND_REQUEST_POST_TC_ABORT); 2837 SCU_CONTEXT_COMMAND_REQUEST_POST_TC_ABORT);
2860 return SCI_SUCCESS; 2838 return SCI_SUCCESS;
@@ -2872,19 +2850,19 @@ enum sci_status scic_controller_terminate_request(
2872 * @io_request: the handle to the io request object to complete. 2850 * @io_request: the handle to the io request object to complete.
2873 */ 2851 */
2874enum sci_status scic_controller_complete_io( 2852enum sci_status scic_controller_complete_io(
2875 struct scic_sds_controller *scic, 2853 struct isci_host *ihost,
2876 struct isci_remote_device *idev, 2854 struct isci_remote_device *idev,
2877 struct isci_request *ireq) 2855 struct isci_request *ireq)
2878{ 2856{
2879 enum sci_status status; 2857 enum sci_status status;
2880 u16 index; 2858 u16 index;
2881 2859
2882 switch (scic->sm.current_state_id) { 2860 switch (ihost->sm.current_state_id) {
2883 case SCIC_STOPPING: 2861 case SCIC_STOPPING:
2884 /* XXX: Implement this function */ 2862 /* XXX: Implement this function */
2885 return SCI_FAILURE; 2863 return SCI_FAILURE;
2886 case SCIC_READY: 2864 case SCIC_READY:
2887 status = scic_sds_remote_device_complete_io(scic, idev, ireq); 2865 status = scic_sds_remote_device_complete_io(ihost, idev, ireq);
2888 if (status != SCI_SUCCESS) 2866 if (status != SCI_SUCCESS)
2889 return status; 2867 return status;
2890 2868
@@ -2892,7 +2870,7 @@ enum sci_status scic_controller_complete_io(
2892 clear_bit(IREQ_ACTIVE, &ireq->flags); 2870 clear_bit(IREQ_ACTIVE, &ireq->flags);
2893 return SCI_SUCCESS; 2871 return SCI_SUCCESS;
2894 default: 2872 default:
2895 dev_warn(scic_to_dev(scic), "invalid state to complete I/O"); 2873 dev_warn(&ihost->pdev->dev, "invalid state to complete I/O");
2896 return SCI_FAILURE_INVALID_STATE; 2874 return SCI_FAILURE_INVALID_STATE;
2897 } 2875 }
2898 2876
@@ -2900,15 +2878,15 @@ enum sci_status scic_controller_complete_io(
2900 2878
2901enum sci_status scic_controller_continue_io(struct isci_request *ireq) 2879enum sci_status scic_controller_continue_io(struct isci_request *ireq)
2902{ 2880{
2903 struct scic_sds_controller *scic = ireq->owning_controller; 2881 struct isci_host *ihost = ireq->owning_controller;
2904 2882
2905 if (scic->sm.current_state_id != SCIC_READY) { 2883 if (ihost->sm.current_state_id != SCIC_READY) {
2906 dev_warn(scic_to_dev(scic), "invalid state to continue I/O"); 2884 dev_warn(&ihost->pdev->dev, "invalid state to continue I/O");
2907 return SCI_FAILURE_INVALID_STATE; 2885 return SCI_FAILURE_INVALID_STATE;
2908 } 2886 }
2909 2887
2910 set_bit(IREQ_ACTIVE, &ireq->flags); 2888 set_bit(IREQ_ACTIVE, &ireq->flags);
2911 scic_sds_controller_post_request(scic, scic_sds_request_get_post_context(ireq)); 2889 scic_sds_controller_post_request(ihost, scic_sds_request_get_post_context(ireq));
2912 return SCI_SUCCESS; 2890 return SCI_SUCCESS;
2913} 2891}
2914 2892
@@ -2922,21 +2900,21 @@ enum sci_status scic_controller_continue_io(struct isci_request *ireq)
2922 * @task_request: the handle to the task request object to start. 2900 * @task_request: the handle to the task request object to start.
2923 */ 2901 */
2924enum sci_task_status scic_controller_start_task( 2902enum sci_task_status scic_controller_start_task(
2925 struct scic_sds_controller *scic, 2903 struct isci_host *ihost,
2926 struct isci_remote_device *idev, 2904 struct isci_remote_device *idev,
2927 struct isci_request *ireq) 2905 struct isci_request *ireq)
2928{ 2906{
2929 enum sci_status status; 2907 enum sci_status status;
2930 2908
2931 if (scic->sm.current_state_id != SCIC_READY) { 2909 if (ihost->sm.current_state_id != SCIC_READY) {
2932 dev_warn(scic_to_dev(scic), 2910 dev_warn(&ihost->pdev->dev,
2933 "%s: SCIC Controller starting task from invalid " 2911 "%s: SCIC Controller starting task from invalid "
2934 "state\n", 2912 "state\n",
2935 __func__); 2913 __func__);
2936 return SCI_TASK_FAILURE_INVALID_STATE; 2914 return SCI_TASK_FAILURE_INVALID_STATE;
2937 } 2915 }
2938 2916
2939 status = scic_sds_remote_device_start_task(scic, idev, ireq); 2917 status = scic_sds_remote_device_start_task(ihost, idev, ireq);
2940 switch (status) { 2918 switch (status) {
2941 case SCI_FAILURE_RESET_DEVICE_PARTIAL_SUCCESS: 2919 case SCI_FAILURE_RESET_DEVICE_PARTIAL_SUCCESS:
2942 set_bit(IREQ_ACTIVE, &ireq->flags); 2920 set_bit(IREQ_ACTIVE, &ireq->flags);
@@ -2950,7 +2928,7 @@ enum sci_task_status scic_controller_start_task(
2950 case SCI_SUCCESS: 2928 case SCI_SUCCESS:
2951 set_bit(IREQ_ACTIVE, &ireq->flags); 2929 set_bit(IREQ_ACTIVE, &ireq->flags);
2952 2930
2953 scic_sds_controller_post_request(scic, 2931 scic_sds_controller_post_request(ihost,
2954 scic_sds_request_get_post_context(ireq)); 2932 scic_sds_request_get_post_context(ireq));
2955 break; 2933 break;
2956 default: 2934 default:
diff --git a/drivers/scsi/isci/host.h b/drivers/scsi/isci/host.h
index ca2e3b0ee0dd..013f672a8fd7 100644
--- a/drivers/scsi/isci/host.h
+++ b/drivers/scsi/isci/host.h
@@ -106,7 +106,7 @@ struct scic_power_control {
106}; 106};
107 107
108struct scic_sds_port_configuration_agent; 108struct scic_sds_port_configuration_agent;
109typedef void (*port_config_fn)(struct scic_sds_controller *, 109typedef void (*port_config_fn)(struct isci_host *,
110 struct scic_sds_port_configuration_agent *, 110 struct scic_sds_port_configuration_agent *,
111 struct isci_port *, struct isci_phy *); 111 struct isci_port *, struct isci_phy *);
112 112
@@ -124,171 +124,66 @@ struct scic_sds_port_configuration_agent {
124}; 124};
125 125
126/** 126/**
127 * struct scic_sds_controller - 127 * isci_host - primary host/controller object
128 * 128 * @timer: timeout start/stop operations
129 * This structure represents the SCU controller object. 129 * @device_table: rni (hw remote node index) to remote device lookup table
130 * @available_remote_nodes: rni allocator
131 * @power_control: manage device spin up
132 * @io_request_sequence: generation number for tci's (task contexts)
133 * @task_context_table: hw task context table
134 * @remote_node_context_table: hw remote node context table
135 * @completion_queue: hw-producer driver-consumer communication ring
136 * @completion_queue_get: tracks the driver 'head' of the ring to notify hw
137 * @logical_port_entries: min({driver|silicon}-supported-port-count)
138 * @remote_node_entries: min({driver|silicon}-supported-node-count)
139 * @task_context_entries: min({driver|silicon}-supported-task-count)
140 * @phy_timer: phy startup timer
141 * @invalid_phy_mask: if an invalid_link_up notification is reported a bit for
142 * the phy index is set so further notifications are not
143 * made. Once the phy reports link up and is made part of a
144 * port then this bit is cleared.
145
130 */ 146 */
131struct scic_sds_controller { 147struct isci_host {
132 /**
133 * This field contains the information for the base controller state
134 * machine.
135 */
136 struct sci_base_state_machine sm; 148 struct sci_base_state_machine sm;
137 149 /* XXX can we time this externally */
138 /**
139 * Timer for controller start/stop operations.
140 */
141 struct sci_timer timer; 150 struct sci_timer timer;
142 151 /* XXX drop reference module params directly */
143 /**
144 * This field contains the user parameters to be utilized for this
145 * core controller object.
146 */
147 union scic_user_parameters user_parameters; 152 union scic_user_parameters user_parameters;
148 153 /* XXX no need to be a union */
149 /**
150 * This field contains the OEM parameters to be utilized for this
151 * core controller object.
152 */
153 union scic_oem_parameters oem_parameters; 154 union scic_oem_parameters oem_parameters;
154
155 /**
156 * This field contains the port configuration agent for this controller.
157 */
158 struct scic_sds_port_configuration_agent port_agent; 155 struct scic_sds_port_configuration_agent port_agent;
159
160 /**
161 * This field is the array of device objects that are currently constructed
162 * for this controller object. This table is used as a fast lookup of device
163 * objects that need to handle device completion notifications from the
164 * hardware. The table is RNi based.
165 */
166 struct isci_remote_device *device_table[SCI_MAX_REMOTE_DEVICES]; 156 struct isci_remote_device *device_table[SCI_MAX_REMOTE_DEVICES];
167
168 /**
169 * This field is the free RNi data structure
170 */
171 struct scic_remote_node_table available_remote_nodes; 157 struct scic_remote_node_table available_remote_nodes;
172
173 /**
174 * This filed is the struct scic_power_control data used to controll when direct
175 * attached devices can consume power.
176 */
177 struct scic_power_control power_control; 158 struct scic_power_control power_control;
178
179 /* sequence number per tci */
180 u8 io_request_sequence[SCI_MAX_IO_REQUESTS]; 159 u8 io_request_sequence[SCI_MAX_IO_REQUESTS];
181
182 /**
183 * This field is a pointer to the memory allocated by the driver for the task
184 * context table. This data is shared between the hardware and software.
185 */
186 struct scu_task_context *task_context_table; 160 struct scu_task_context *task_context_table;
187 dma_addr_t task_context_dma; 161 dma_addr_t task_context_dma;
188
189 /**
190 * This field is a pointer to the memory allocated by the driver for the
191 * remote node context table. This table is shared between the hardware and
192 * software.
193 */
194 union scu_remote_node_context *remote_node_context_table; 162 union scu_remote_node_context *remote_node_context_table;
195
196 /**
197 * This field is a pointer to the completion queue. This memory is
198 * written to by the hardware and read by the software.
199 */
200 u32 *completion_queue; 163 u32 *completion_queue;
201
202 /**
203 * This field is the software copy of the completion queue get pointer. The
204 * controller object writes this value to the hardware after processing the
205 * completion entries.
206 */
207 u32 completion_queue_get; 164 u32 completion_queue_get;
208
209 /**
210 * This field is the minimum of the number of hardware supported port entries
211 * and the software requested port entries.
212 */
213 u32 logical_port_entries; 165 u32 logical_port_entries;
214
215 /**
216 * This field is the minimum number of devices supported by the hardware and
217 * the number of devices requested by the software.
218 */
219 u32 remote_node_entries; 166 u32 remote_node_entries;
220
221 /**
222 * This field is the minimum number of IO requests supported by the hardware
223 * and the number of IO requests requested by the software.
224 */
225 u32 task_context_entries; 167 u32 task_context_entries;
226
227 /**
228 * This object contains all of the unsolicited frame specific
229 * data utilized by the core controller.
230 */
231 struct scic_sds_unsolicited_frame_control uf_control; 168 struct scic_sds_unsolicited_frame_control uf_control;
232 169
233 /* Phy Startup Data */ 170 /* phy startup */
234 /**
235 * Timer for controller phy request startup. On controller start the
236 * controller will start each PHY individually in order of phy index.
237 */
238 struct sci_timer phy_timer; 171 struct sci_timer phy_timer;
239 172 /* XXX kill */
240 /**
241 * This field is set when the phy_timer is running and is cleared when
242 * the phy_timer is stopped.
243 */
244 bool phy_startup_timer_pending; 173 bool phy_startup_timer_pending;
245
246 /**
247 * This field is the index of the next phy start. It is initialized to 0 and
248 * increments for each phy index that is started.
249 */
250 u32 next_phy_to_start; 174 u32 next_phy_to_start;
251
252 /**
253 * This field controlls the invalid link up notifications to the SCI_USER. If
254 * an invalid_link_up notification is reported a bit for the PHY index is set
255 * so further notifications are not made. Once the PHY object reports link up
256 * and is made part of a port then this bit for the PHY index is cleared.
257 */
258 u8 invalid_phy_mask; 175 u8 invalid_phy_mask;
259 176
260 /* 177 /* TODO attempt dynamic interrupt coalescing scheme */
261 * This field saves the current interrupt coalescing number of the controller.
262 */
263 u16 interrupt_coalesce_number; 178 u16 interrupt_coalesce_number;
264
265 /*
266 * This field saves the current interrupt coalescing timeout value in microseconds.
267 */
268 u32 interrupt_coalesce_timeout; 179 u32 interrupt_coalesce_timeout;
269
270 /**
271 * This field is a pointer to the memory mapped register space for the
272 * struct smu_registers.
273 */
274 struct smu_registers __iomem *smu_registers; 180 struct smu_registers __iomem *smu_registers;
275
276 /**
277 * This field is a pointer to the memory mapped register space for the
278 * struct scu_registers.
279 */
280 struct scu_registers __iomem *scu_registers; 181 struct scu_registers __iomem *scu_registers;
281 182
282};
283
284struct isci_host {
285 struct scic_sds_controller sci;
286 u16 tci_head; 183 u16 tci_head;
287 u16 tci_tail; 184 u16 tci_tail;
288 u16 tci_pool[SCI_MAX_IO_REQUESTS]; 185 u16 tci_pool[SCI_MAX_IO_REQUESTS];
289 186
290 union scic_oem_parameters oem_parameters;
291
292 int id; /* unique within a given pci device */ 187 int id; /* unique within a given pci device */
293 struct isci_phy phys[SCI_MAX_PHYS]; 188 struct isci_phy phys[SCI_MAX_PHYS];
294 struct isci_port ports[SCI_MAX_PORTS + 1]; /* includes dummy port */ 189 struct isci_port ports[SCI_MAX_PORTS + 1]; /* includes dummy port */
@@ -464,14 +359,6 @@ static inline struct isci_host *dev_to_ihost(struct domain_device *dev)
464 return dev->port->ha->lldd_ha; 359 return dev->port->ha->lldd_ha;
465} 360}
466 361
467static inline struct isci_host *scic_to_ihost(struct scic_sds_controller *scic)
468{
469 /* XXX delete after merging scic_sds_contoller and isci_host */
470 struct isci_host *ihost = container_of(scic, typeof(*ihost), sci);
471
472 return ihost;
473}
474
475/** 362/**
476 * scic_sds_controller_get_protocol_engine_group() - 363 * scic_sds_controller_get_protocol_engine_group() -
477 * 364 *
@@ -518,11 +405,6 @@ static inline int scic_sds_remote_device_node_count(struct isci_remote_device *i
518#define scic_sds_controller_clear_invalid_phy(controller, phy) \ 405#define scic_sds_controller_clear_invalid_phy(controller, phy) \
519 ((controller)->invalid_phy_mask &= ~(1 << (phy)->phy_index)) 406 ((controller)->invalid_phy_mask &= ~(1 << (phy)->phy_index))
520 407
521static inline struct device *scic_to_dev(struct scic_sds_controller *scic)
522{
523 return &scic_to_ihost(scic)->pdev->dev;
524}
525
526static inline struct device *sciphy_to_dev(struct isci_phy *iphy) 408static inline struct device *sciphy_to_dev(struct isci_phy *iphy)
527{ 409{
528 410
@@ -578,54 +460,54 @@ static inline bool is_c0(void)
578 return isci_si_rev > ISCI_SI_REVB0; 460 return isci_si_rev > ISCI_SI_REVB0;
579} 461}
580 462
581void scic_sds_controller_post_request(struct scic_sds_controller *scic, 463void scic_sds_controller_post_request(struct isci_host *ihost,
582 u32 request); 464 u32 request);
583void scic_sds_controller_release_frame(struct scic_sds_controller *scic, 465void scic_sds_controller_release_frame(struct isci_host *ihost,
584 u32 frame_index); 466 u32 frame_index);
585void scic_sds_controller_copy_sata_response(void *response_buffer, 467void scic_sds_controller_copy_sata_response(void *response_buffer,
586 void *frame_header, 468 void *frame_header,
587 void *frame_buffer); 469 void *frame_buffer);
588enum sci_status scic_sds_controller_allocate_remote_node_context(struct scic_sds_controller *scic, 470enum sci_status scic_sds_controller_allocate_remote_node_context(struct isci_host *ihost,
589 struct isci_remote_device *idev, 471 struct isci_remote_device *idev,
590 u16 *node_id); 472 u16 *node_id);
591void scic_sds_controller_free_remote_node_context( 473void scic_sds_controller_free_remote_node_context(
592 struct scic_sds_controller *scic, 474 struct isci_host *ihost,
593 struct isci_remote_device *idev, 475 struct isci_remote_device *idev,
594 u16 node_id); 476 u16 node_id);
595union scu_remote_node_context *scic_sds_controller_get_remote_node_context_buffer( 477union scu_remote_node_context *scic_sds_controller_get_remote_node_context_buffer(
596 struct scic_sds_controller *scic, 478 struct isci_host *ihost,
597 u16 node_id); 479 u16 node_id);
598 480
599struct isci_request *scic_request_by_tag(struct scic_sds_controller *scic, 481struct isci_request *scic_request_by_tag(struct isci_host *ihost,
600 u16 io_tag); 482 u16 io_tag);
601 483
602void scic_sds_controller_power_control_queue_insert( 484void scic_sds_controller_power_control_queue_insert(
603 struct scic_sds_controller *scic, 485 struct isci_host *ihost,
604 struct isci_phy *iphy); 486 struct isci_phy *iphy);
605 487
606void scic_sds_controller_power_control_queue_remove( 488void scic_sds_controller_power_control_queue_remove(
607 struct scic_sds_controller *scic, 489 struct isci_host *ihost,
608 struct isci_phy *iphy); 490 struct isci_phy *iphy);
609 491
610void scic_sds_controller_link_up( 492void scic_sds_controller_link_up(
611 struct scic_sds_controller *scic, 493 struct isci_host *ihost,
612 struct isci_port *iport, 494 struct isci_port *iport,
613 struct isci_phy *iphy); 495 struct isci_phy *iphy);
614 496
615void scic_sds_controller_link_down( 497void scic_sds_controller_link_down(
616 struct scic_sds_controller *scic, 498 struct isci_host *ihost,
617 struct isci_port *iport, 499 struct isci_port *iport,
618 struct isci_phy *iphy); 500 struct isci_phy *iphy);
619 501
620void scic_sds_controller_remote_device_stopped( 502void scic_sds_controller_remote_device_stopped(
621 struct scic_sds_controller *scic, 503 struct isci_host *ihost,
622 struct isci_remote_device *idev); 504 struct isci_remote_device *idev);
623 505
624void scic_sds_controller_copy_task_context( 506void scic_sds_controller_copy_task_context(
625 struct scic_sds_controller *scic, 507 struct isci_host *ihost,
626 struct isci_request *ireq); 508 struct isci_request *ireq);
627 509
628void scic_sds_controller_register_setup(struct scic_sds_controller *scic); 510void scic_sds_controller_register_setup(struct isci_host *ihost);
629 511
630enum sci_status scic_controller_continue_io(struct isci_request *ireq); 512enum sci_status scic_controller_continue_io(struct isci_request *ireq);
631int isci_host_scan_finished(struct Scsi_Host *, unsigned long); 513int isci_host_scan_finished(struct Scsi_Host *, unsigned long);
@@ -655,25 +537,25 @@ void isci_host_remote_device_start_complete(
655 enum sci_status); 537 enum sci_status);
656 538
657void scic_controller_disable_interrupts( 539void scic_controller_disable_interrupts(
658 struct scic_sds_controller *scic); 540 struct isci_host *ihost);
659 541
660enum sci_status scic_controller_start_io( 542enum sci_status scic_controller_start_io(
661 struct scic_sds_controller *scic, 543 struct isci_host *ihost,
662 struct isci_remote_device *idev, 544 struct isci_remote_device *idev,
663 struct isci_request *ireq); 545 struct isci_request *ireq);
664 546
665enum sci_task_status scic_controller_start_task( 547enum sci_task_status scic_controller_start_task(
666 struct scic_sds_controller *scic, 548 struct isci_host *ihost,
667 struct isci_remote_device *idev, 549 struct isci_remote_device *idev,
668 struct isci_request *ireq); 550 struct isci_request *ireq);
669 551
670enum sci_status scic_controller_terminate_request( 552enum sci_status scic_controller_terminate_request(
671 struct scic_sds_controller *scic, 553 struct isci_host *ihost,
672 struct isci_remote_device *idev, 554 struct isci_remote_device *idev,
673 struct isci_request *ireq); 555 struct isci_request *ireq);
674 556
675enum sci_status scic_controller_complete_io( 557enum sci_status scic_controller_complete_io(
676 struct scic_sds_controller *scic, 558 struct isci_host *ihost,
677 struct isci_remote_device *idev, 559 struct isci_remote_device *idev,
678 struct isci_request *ireq); 560 struct isci_request *ireq);
679 561
@@ -681,6 +563,6 @@ void scic_sds_port_configuration_agent_construct(
681 struct scic_sds_port_configuration_agent *port_agent); 563 struct scic_sds_port_configuration_agent *port_agent);
682 564
683enum sci_status scic_sds_port_configuration_agent_initialize( 565enum sci_status scic_sds_port_configuration_agent_initialize(
684 struct scic_sds_controller *controller, 566 struct isci_host *ihost,
685 struct scic_sds_port_configuration_agent *port_agent); 567 struct scic_sds_port_configuration_agent *port_agent);
686#endif 568#endif
diff --git a/drivers/scsi/isci/init.c b/drivers/scsi/isci/init.c
index bbfb6e563207..68ca1a4f30af 100644
--- a/drivers/scsi/isci/init.c
+++ b/drivers/scsi/isci/init.c
@@ -548,13 +548,13 @@ static int __devinit isci_pci_probe(struct pci_dev *pdev, const struct pci_devic
548 548
549static void __devexit isci_pci_remove(struct pci_dev *pdev) 549static void __devexit isci_pci_remove(struct pci_dev *pdev)
550{ 550{
551 struct isci_host *isci_host; 551 struct isci_host *ihost;
552 int i; 552 int i;
553 553
554 for_each_isci_host(i, isci_host, pdev) { 554 for_each_isci_host(i, ihost, pdev) {
555 isci_unregister(isci_host); 555 isci_unregister(ihost);
556 isci_host_deinit(isci_host); 556 isci_host_deinit(ihost);
557 scic_controller_disable_interrupts(&isci_host->sci); 557 scic_controller_disable_interrupts(ihost);
558 } 558 }
559} 559}
560 560
diff --git a/drivers/scsi/isci/phy.c b/drivers/scsi/isci/phy.c
index fd0e9734e5d0..ca96b5ad0d52 100644
--- a/drivers/scsi/isci/phy.c
+++ b/drivers/scsi/isci/phy.c
@@ -112,13 +112,13 @@ static enum sci_status
112scic_sds_phy_link_layer_initialization(struct isci_phy *iphy, 112scic_sds_phy_link_layer_initialization(struct isci_phy *iphy,
113 struct scu_link_layer_registers __iomem *link_layer_registers) 113 struct scu_link_layer_registers __iomem *link_layer_registers)
114{ 114{
115 struct scic_sds_controller *scic = 115 struct isci_host *ihost =
116 iphy->owning_port->owning_controller; 116 iphy->owning_port->owning_controller;
117 int phy_idx = iphy->phy_index; 117 int phy_idx = iphy->phy_index;
118 struct sci_phy_user_params *phy_user = 118 struct sci_phy_user_params *phy_user =
119 &scic->user_parameters.sds1.phys[phy_idx]; 119 &ihost->user_parameters.sds1.phys[phy_idx];
120 struct sci_phy_oem_params *phy_oem = 120 struct sci_phy_oem_params *phy_oem =
121 &scic->oem_parameters.sds1.phys[phy_idx]; 121 &ihost->oem_parameters.sds1.phys[phy_idx];
122 u32 phy_configuration; 122 u32 phy_configuration;
123 struct scic_phy_cap phy_cap; 123 struct scic_phy_cap phy_cap;
124 u32 parity_check = 0; 124 u32 parity_check = 0;
@@ -169,7 +169,7 @@ scic_sds_phy_link_layer_initialization(struct isci_phy *iphy,
169 phy_cap.gen3_no_ssc = 1; 169 phy_cap.gen3_no_ssc = 1;
170 phy_cap.gen2_no_ssc = 1; 170 phy_cap.gen2_no_ssc = 1;
171 phy_cap.gen1_no_ssc = 1; 171 phy_cap.gen1_no_ssc = 1;
172 if (scic->oem_parameters.sds1.controller.do_enable_ssc == true) { 172 if (ihost->oem_parameters.sds1.controller.do_enable_ssc == true) {
173 phy_cap.gen3_ssc = 1; 173 phy_cap.gen3_ssc = 1;
174 phy_cap.gen2_ssc = 1; 174 phy_cap.gen2_ssc = 1;
175 phy_cap.gen1_ssc = 1; 175 phy_cap.gen1_ssc = 1;
@@ -216,7 +216,7 @@ scic_sds_phy_link_layer_initialization(struct isci_phy *iphy,
216 &iphy->link_layer_registers->afe_lookup_table_control); 216 &iphy->link_layer_registers->afe_lookup_table_control);
217 217
218 llctl = SCU_SAS_LLCTL_GEN_VAL(NO_OUTBOUND_TASK_TIMEOUT, 218 llctl = SCU_SAS_LLCTL_GEN_VAL(NO_OUTBOUND_TASK_TIMEOUT,
219 (u8)scic->user_parameters.sds1.no_outbound_task_timeout); 219 (u8)ihost->user_parameters.sds1.no_outbound_task_timeout);
220 220
221 switch(phy_user->max_speed_generation) { 221 switch(phy_user->max_speed_generation) {
222 case SCIC_SDS_PARM_GEN3_SPEED: 222 case SCIC_SDS_PARM_GEN3_SPEED:
@@ -255,7 +255,7 @@ static void phy_sata_timeout(unsigned long data)
255{ 255{
256 struct sci_timer *tmr = (struct sci_timer *)data; 256 struct sci_timer *tmr = (struct sci_timer *)data;
257 struct isci_phy *iphy = container_of(tmr, typeof(*iphy), sata_timer); 257 struct isci_phy *iphy = container_of(tmr, typeof(*iphy), sata_timer);
258 struct isci_host *ihost = scic_to_ihost(iphy->owning_port->owning_controller); 258 struct isci_host *ihost = iphy->owning_port->owning_controller;
259 unsigned long flags; 259 unsigned long flags;
260 260
261 spin_lock_irqsave(&ihost->scic_lock, flags); 261 spin_lock_irqsave(&ihost->scic_lock, flags);
@@ -890,7 +890,7 @@ enum sci_status scic_sds_phy_frame_handler(struct isci_phy *iphy,
890 u32 frame_index) 890 u32 frame_index)
891{ 891{
892 enum scic_sds_phy_states state = iphy->sm.current_state_id; 892 enum scic_sds_phy_states state = iphy->sm.current_state_id;
893 struct scic_sds_controller *scic = iphy->owning_port->owning_controller; 893 struct isci_host *ihost = iphy->owning_port->owning_controller;
894 enum sci_status result; 894 enum sci_status result;
895 unsigned long flags; 895 unsigned long flags;
896 896
@@ -899,7 +899,7 @@ enum sci_status scic_sds_phy_frame_handler(struct isci_phy *iphy,
899 u32 *frame_words; 899 u32 *frame_words;
900 struct sas_identify_frame iaf; 900 struct sas_identify_frame iaf;
901 901
902 result = scic_sds_unsolicited_frame_control_get_header(&scic->uf_control, 902 result = scic_sds_unsolicited_frame_control_get_header(&ihost->uf_control,
903 frame_index, 903 frame_index,
904 (void **)&frame_words); 904 (void **)&frame_words);
905 905
@@ -933,7 +933,7 @@ enum sci_status scic_sds_phy_frame_handler(struct isci_phy *iphy,
933 "unexpected frame id %x\n", 933 "unexpected frame id %x\n",
934 __func__, frame_index); 934 __func__, frame_index);
935 935
936 scic_sds_controller_release_frame(scic, frame_index); 936 scic_sds_controller_release_frame(ihost, frame_index);
937 return result; 937 return result;
938 } 938 }
939 case SCI_PHY_SUB_AWAIT_SIG_FIS_UF: { 939 case SCI_PHY_SUB_AWAIT_SIG_FIS_UF: {
@@ -950,7 +950,7 @@ enum sci_status scic_sds_phy_frame_handler(struct isci_phy *iphy,
950 950
951 if ((frame_header->fis_type == FIS_REGD2H) && 951 if ((frame_header->fis_type == FIS_REGD2H) &&
952 !(frame_header->status & ATA_BUSY)) { 952 !(frame_header->status & ATA_BUSY)) {
953 scic_sds_unsolicited_frame_control_get_buffer(&scic->uf_control, 953 scic_sds_unsolicited_frame_control_get_buffer(&ihost->uf_control,
954 frame_index, 954 frame_index,
955 (void **)&fis_frame_data); 955 (void **)&fis_frame_data);
956 956
@@ -971,7 +971,7 @@ enum sci_status scic_sds_phy_frame_handler(struct isci_phy *iphy,
971 __func__, frame_index); 971 __func__, frame_index);
972 972
973 /* Regardless of the result we are done with this frame with it */ 973 /* Regardless of the result we are done with this frame with it */
974 scic_sds_controller_release_frame(scic, frame_index); 974 scic_sds_controller_release_frame(ihost, frame_index);
975 975
976 return result; 976 return result;
977 } 977 }
@@ -994,33 +994,33 @@ static void scic_sds_phy_starting_initial_substate_enter(struct sci_base_state_m
994static void scic_sds_phy_starting_await_sas_power_substate_enter(struct sci_base_state_machine *sm) 994static void scic_sds_phy_starting_await_sas_power_substate_enter(struct sci_base_state_machine *sm)
995{ 995{
996 struct isci_phy *iphy = container_of(sm, typeof(*iphy), sm); 996 struct isci_phy *iphy = container_of(sm, typeof(*iphy), sm);
997 struct scic_sds_controller *scic = iphy->owning_port->owning_controller; 997 struct isci_host *ihost = iphy->owning_port->owning_controller;
998 998
999 scic_sds_controller_power_control_queue_insert(scic, iphy); 999 scic_sds_controller_power_control_queue_insert(ihost, iphy);
1000} 1000}
1001 1001
1002static void scic_sds_phy_starting_await_sas_power_substate_exit(struct sci_base_state_machine *sm) 1002static void scic_sds_phy_starting_await_sas_power_substate_exit(struct sci_base_state_machine *sm)
1003{ 1003{
1004 struct isci_phy *iphy = container_of(sm, typeof(*iphy), sm); 1004 struct isci_phy *iphy = container_of(sm, typeof(*iphy), sm);
1005 struct scic_sds_controller *scic = iphy->owning_port->owning_controller; 1005 struct isci_host *ihost = iphy->owning_port->owning_controller;
1006 1006
1007 scic_sds_controller_power_control_queue_remove(scic, iphy); 1007 scic_sds_controller_power_control_queue_remove(ihost, iphy);
1008} 1008}
1009 1009
1010static void scic_sds_phy_starting_await_sata_power_substate_enter(struct sci_base_state_machine *sm) 1010static void scic_sds_phy_starting_await_sata_power_substate_enter(struct sci_base_state_machine *sm)
1011{ 1011{
1012 struct isci_phy *iphy = container_of(sm, typeof(*iphy), sm); 1012 struct isci_phy *iphy = container_of(sm, typeof(*iphy), sm);
1013 struct scic_sds_controller *scic = iphy->owning_port->owning_controller; 1013 struct isci_host *ihost = iphy->owning_port->owning_controller;
1014 1014
1015 scic_sds_controller_power_control_queue_insert(scic, iphy); 1015 scic_sds_controller_power_control_queue_insert(ihost, iphy);
1016} 1016}
1017 1017
1018static void scic_sds_phy_starting_await_sata_power_substate_exit(struct sci_base_state_machine *sm) 1018static void scic_sds_phy_starting_await_sata_power_substate_exit(struct sci_base_state_machine *sm)
1019{ 1019{
1020 struct isci_phy *iphy = container_of(sm, typeof(*iphy), sm); 1020 struct isci_phy *iphy = container_of(sm, typeof(*iphy), sm);
1021 struct scic_sds_controller *scic = iphy->owning_port->owning_controller; 1021 struct isci_host *ihost = iphy->owning_port->owning_controller;
1022 1022
1023 scic_sds_controller_power_control_queue_remove(scic, iphy); 1023 scic_sds_controller_power_control_queue_remove(ihost, iphy);
1024} 1024}
1025 1025
1026static void scic_sds_phy_starting_await_sata_phy_substate_enter(struct sci_base_state_machine *sm) 1026static void scic_sds_phy_starting_await_sata_phy_substate_enter(struct sci_base_state_machine *sm)
@@ -1313,7 +1313,7 @@ void isci_phy_init(struct isci_phy *iphy, struct isci_host *ihost, int index)
1313 u64 sci_sas_addr; 1313 u64 sci_sas_addr;
1314 __be64 sas_addr; 1314 __be64 sas_addr;
1315 1315
1316 scic_oem_parameters_get(&ihost->sci, &oem); 1316 scic_oem_parameters_get(ihost, &oem);
1317 sci_sas_addr = oem.sds1.phys[index].sas_address.high; 1317 sci_sas_addr = oem.sds1.phys[index].sas_address.high;
1318 sci_sas_addr <<= 32; 1318 sci_sas_addr <<= 32;
1319 sci_sas_addr |= oem.sds1.phys[index].sas_address.low; 1319 sci_sas_addr |= oem.sds1.phys[index].sas_address.low;
diff --git a/drivers/scsi/isci/port.c b/drivers/scsi/isci/port.c
index df37b1bf7d11..c434d5a0effa 100644
--- a/drivers/scsi/isci/port.c
+++ b/drivers/scsi/isci/port.c
@@ -365,11 +365,11 @@ static void isci_port_not_ready(struct isci_host *isci_host, struct isci_port *i
365 "%s: isci_port = %p\n", __func__, isci_port); 365 "%s: isci_port = %p\n", __func__, isci_port);
366} 366}
367 367
368static void isci_port_stop_complete(struct scic_sds_controller *scic, 368static void isci_port_stop_complete(struct isci_host *ihost,
369 struct isci_port *iport, 369 struct isci_port *iport,
370 enum sci_status completion_status) 370 enum sci_status completion_status)
371{ 371{
372 dev_dbg(&scic_to_ihost(scic)->pdev->dev, "Port stop complete\n"); 372 dev_dbg(&ihost->pdev->dev, "Port stop complete\n");
373} 373}
374 374
375/** 375/**
@@ -541,8 +541,7 @@ static enum sci_status scic_sds_port_clear_phy(struct isci_port *iport,
541 /* Make sure that this phy is part of this port */ 541 /* Make sure that this phy is part of this port */
542 if (iport->phy_table[iphy->phy_index] == iphy && 542 if (iport->phy_table[iphy->phy_index] == iphy &&
543 phy_get_non_dummy_port(iphy) == iport) { 543 phy_get_non_dummy_port(iphy) == iport) {
544 struct scic_sds_controller *scic = iport->owning_controller; 544 struct isci_host *ihost = iport->owning_controller;
545 struct isci_host *ihost = scic_to_ihost(scic);
546 545
547 /* Yep it is assigned to this port so remove it */ 546 /* Yep it is assigned to this port so remove it */
548 scic_sds_phy_set_port(iphy, &ihost->ports[SCI_MAX_PORTS]); 547 scic_sds_phy_set_port(iphy, &ihost->ports[SCI_MAX_PORTS]);
@@ -654,10 +653,10 @@ static void scic_sds_port_construct_dummy_rnc(struct isci_port *iport, u16 rni)
654 */ 653 */
655static void scic_sds_port_construct_dummy_task(struct isci_port *iport, u16 tag) 654static void scic_sds_port_construct_dummy_task(struct isci_port *iport, u16 tag)
656{ 655{
657 struct scic_sds_controller *scic = iport->owning_controller; 656 struct isci_host *ihost = iport->owning_controller;
658 struct scu_task_context *task_context; 657 struct scu_task_context *task_context;
659 658
660 task_context = &scic->task_context_table[ISCI_TAG_TCI(tag)]; 659 task_context = &ihost->task_context_table[ISCI_TAG_TCI(tag)];
661 memset(task_context, 0, sizeof(struct scu_task_context)); 660 memset(task_context, 0, sizeof(struct scu_task_context));
662 661
663 task_context->initiator_request = 1; 662 task_context->initiator_request = 1;
@@ -674,13 +673,13 @@ static void scic_sds_port_construct_dummy_task(struct isci_port *iport, u16 tag)
674 673
675static void scic_sds_port_destroy_dummy_resources(struct isci_port *iport) 674static void scic_sds_port_destroy_dummy_resources(struct isci_port *iport)
676{ 675{
677 struct scic_sds_controller *scic = iport->owning_controller; 676 struct isci_host *ihost = iport->owning_controller;
678 677
679 if (iport->reserved_tag != SCI_CONTROLLER_INVALID_IO_TAG) 678 if (iport->reserved_tag != SCI_CONTROLLER_INVALID_IO_TAG)
680 isci_free_tag(scic_to_ihost(scic), iport->reserved_tag); 679 isci_free_tag(ihost, iport->reserved_tag);
681 680
682 if (iport->reserved_rni != SCU_DUMMY_INDEX) 681 if (iport->reserved_rni != SCU_DUMMY_INDEX)
683 scic_sds_remote_node_table_release_remote_node_index(&scic->available_remote_nodes, 682 scic_sds_remote_node_table_release_remote_node_index(&ihost->available_remote_nodes,
684 1, iport->reserved_rni); 683 1, iport->reserved_rni);
685 684
686 iport->reserved_rni = SCU_DUMMY_INDEX; 685 iport->reserved_rni = SCU_DUMMY_INDEX;
@@ -749,15 +748,14 @@ static void scic_sds_port_activate_phy(struct isci_port *iport,
749 struct isci_phy *iphy, 748 struct isci_phy *iphy,
750 bool do_notify_user) 749 bool do_notify_user)
751{ 750{
752 struct scic_sds_controller *scic = iport->owning_controller; 751 struct isci_host *ihost = iport->owning_controller;
753 struct isci_host *ihost = scic_to_ihost(scic);
754 752
755 if (iphy->protocol != SCIC_SDS_PHY_PROTOCOL_SATA) 753 if (iphy->protocol != SCIC_SDS_PHY_PROTOCOL_SATA)
756 scic_sds_phy_resume(iphy); 754 scic_sds_phy_resume(iphy);
757 755
758 iport->active_phy_mask |= 1 << iphy->phy_index; 756 iport->active_phy_mask |= 1 << iphy->phy_index;
759 757
760 scic_sds_controller_clear_invalid_phy(scic, iphy); 758 scic_sds_controller_clear_invalid_phy(ihost, iphy);
761 759
762 if (do_notify_user == true) 760 if (do_notify_user == true)
763 isci_port_link_up(ihost, iport, iphy); 761 isci_port_link_up(ihost, iport, iphy);
@@ -767,8 +765,7 @@ void scic_sds_port_deactivate_phy(struct isci_port *iport,
767 struct isci_phy *iphy, 765 struct isci_phy *iphy,
768 bool do_notify_user) 766 bool do_notify_user)
769{ 767{
770 struct scic_sds_controller *scic = scic_sds_port_get_controller(iport); 768 struct isci_host *ihost = scic_sds_port_get_controller(iport);
771 struct isci_host *ihost = scic_to_ihost(scic);
772 769
773 iport->active_phy_mask &= ~(1 << iphy->phy_index); 770 iport->active_phy_mask &= ~(1 << iphy->phy_index);
774 771
@@ -793,16 +790,16 @@ void scic_sds_port_deactivate_phy(struct isci_port *iport,
793static void scic_sds_port_invalid_link_up(struct isci_port *iport, 790static void scic_sds_port_invalid_link_up(struct isci_port *iport,
794 struct isci_phy *iphy) 791 struct isci_phy *iphy)
795{ 792{
796 struct scic_sds_controller *scic = iport->owning_controller; 793 struct isci_host *ihost = iport->owning_controller;
797 794
798 /* 795 /*
799 * Check to see if we have alreay reported this link as bad and if 796 * Check to see if we have alreay reported this link as bad and if
800 * not go ahead and tell the SCI_USER that we have discovered an 797 * not go ahead and tell the SCI_USER that we have discovered an
801 * invalid link. 798 * invalid link.
802 */ 799 */
803 if ((scic->invalid_phy_mask & (1 << iphy->phy_index)) == 0) { 800 if ((ihost->invalid_phy_mask & (1 << iphy->phy_index)) == 0) {
804 scic_sds_controller_set_invalid_phy(scic, iphy); 801 scic_sds_controller_set_invalid_phy(ihost, iphy);
805 dev_warn(&scic_to_ihost(scic)->pdev->dev, "Invalid link up!\n"); 802 dev_warn(&ihost->pdev->dev, "Invalid link up!\n");
806 } 803 }
807} 804}
808 805
@@ -931,7 +928,7 @@ static void port_timeout(unsigned long data)
931{ 928{
932 struct sci_timer *tmr = (struct sci_timer *)data; 929 struct sci_timer *tmr = (struct sci_timer *)data;
933 struct isci_port *iport = container_of(tmr, typeof(*iport), timer); 930 struct isci_port *iport = container_of(tmr, typeof(*iport), timer);
934 struct isci_host *ihost = scic_to_ihost(iport->owning_controller); 931 struct isci_host *ihost = iport->owning_controller;
935 unsigned long flags; 932 unsigned long flags;
936 u32 current_state; 933 u32 current_state;
937 934
@@ -1041,19 +1038,19 @@ static void scic_sds_port_suspend_port_task_scheduler(struct isci_port *iport)
1041 */ 1038 */
1042static void scic_sds_port_post_dummy_request(struct isci_port *iport) 1039static void scic_sds_port_post_dummy_request(struct isci_port *iport)
1043{ 1040{
1044 struct scic_sds_controller *scic = iport->owning_controller; 1041 struct isci_host *ihost = iport->owning_controller;
1045 u16 tag = iport->reserved_tag; 1042 u16 tag = iport->reserved_tag;
1046 struct scu_task_context *tc; 1043 struct scu_task_context *tc;
1047 u32 command; 1044 u32 command;
1048 1045
1049 tc = &scic->task_context_table[ISCI_TAG_TCI(tag)]; 1046 tc = &ihost->task_context_table[ISCI_TAG_TCI(tag)];
1050 tc->abort = 0; 1047 tc->abort = 0;
1051 1048
1052 command = SCU_CONTEXT_COMMAND_REQUEST_TYPE_POST_TC | 1049 command = SCU_CONTEXT_COMMAND_REQUEST_TYPE_POST_TC |
1053 iport->physical_port_index << SCU_CONTEXT_COMMAND_LOGICAL_PORT_SHIFT | 1050 iport->physical_port_index << SCU_CONTEXT_COMMAND_LOGICAL_PORT_SHIFT |
1054 ISCI_TAG_TCI(tag); 1051 ISCI_TAG_TCI(tag);
1055 1052
1056 scic_sds_controller_post_request(scic, command); 1053 scic_sds_controller_post_request(ihost, command);
1057} 1054}
1058 1055
1059/** 1056/**
@@ -1065,19 +1062,19 @@ static void scic_sds_port_post_dummy_request(struct isci_port *iport)
1065 */ 1062 */
1066static void scic_sds_port_abort_dummy_request(struct isci_port *iport) 1063static void scic_sds_port_abort_dummy_request(struct isci_port *iport)
1067{ 1064{
1068 struct scic_sds_controller *scic = iport->owning_controller; 1065 struct isci_host *ihost = iport->owning_controller;
1069 u16 tag = iport->reserved_tag; 1066 u16 tag = iport->reserved_tag;
1070 struct scu_task_context *tc; 1067 struct scu_task_context *tc;
1071 u32 command; 1068 u32 command;
1072 1069
1073 tc = &scic->task_context_table[ISCI_TAG_TCI(tag)]; 1070 tc = &ihost->task_context_table[ISCI_TAG_TCI(tag)];
1074 tc->abort = 1; 1071 tc->abort = 1;
1075 1072
1076 command = SCU_CONTEXT_COMMAND_REQUEST_POST_TC_ABORT | 1073 command = SCU_CONTEXT_COMMAND_REQUEST_POST_TC_ABORT |
1077 iport->physical_port_index << SCU_CONTEXT_COMMAND_LOGICAL_PORT_SHIFT | 1074 iport->physical_port_index << SCU_CONTEXT_COMMAND_LOGICAL_PORT_SHIFT |
1078 ISCI_TAG_TCI(tag); 1075 ISCI_TAG_TCI(tag);
1079 1076
1080 scic_sds_controller_post_request(scic, command); 1077 scic_sds_controller_post_request(ihost, command);
1081} 1078}
1082 1079
1083/** 1080/**
@@ -1115,8 +1112,7 @@ static void scic_sds_port_ready_substate_operational_enter(struct sci_base_state
1115{ 1112{
1116 u32 index; 1113 u32 index;
1117 struct isci_port *iport = container_of(sm, typeof(*iport), sm); 1114 struct isci_port *iport = container_of(sm, typeof(*iport), sm);
1118 struct scic_sds_controller *scic = iport->owning_controller; 1115 struct isci_host *ihost = iport->owning_controller;
1119 struct isci_host *ihost = scic_to_ihost(scic);
1120 1116
1121 isci_port_ready(ihost, iport); 1117 isci_port_ready(ihost, iport);
1122 1118
@@ -1141,13 +1137,13 @@ static void scic_sds_port_ready_substate_operational_enter(struct sci_base_state
1141 1137
1142static void scic_sds_port_invalidate_dummy_remote_node(struct isci_port *iport) 1138static void scic_sds_port_invalidate_dummy_remote_node(struct isci_port *iport)
1143{ 1139{
1144 struct scic_sds_controller *scic = iport->owning_controller; 1140 struct isci_host *ihost = iport->owning_controller;
1145 u8 phys_index = iport->physical_port_index; 1141 u8 phys_index = iport->physical_port_index;
1146 union scu_remote_node_context *rnc; 1142 union scu_remote_node_context *rnc;
1147 u16 rni = iport->reserved_rni; 1143 u16 rni = iport->reserved_rni;
1148 u32 command; 1144 u32 command;
1149 1145
1150 rnc = &scic->remote_node_context_table[rni]; 1146 rnc = &ihost->remote_node_context_table[rni];
1151 1147
1152 rnc->ssp.is_valid = false; 1148 rnc->ssp.is_valid = false;
1153 1149
@@ -1155,13 +1151,13 @@ static void scic_sds_port_invalidate_dummy_remote_node(struct isci_port *iport)
1155 * controller and give it ample time to act before posting the rnc 1151 * controller and give it ample time to act before posting the rnc
1156 * invalidate 1152 * invalidate
1157 */ 1153 */
1158 readl(&scic->smu_registers->interrupt_status); /* flush */ 1154 readl(&ihost->smu_registers->interrupt_status); /* flush */
1159 udelay(10); 1155 udelay(10);
1160 1156
1161 command = SCU_CONTEXT_COMMAND_POST_RNC_INVALIDATE | 1157 command = SCU_CONTEXT_COMMAND_POST_RNC_INVALIDATE |
1162 phys_index << SCU_CONTEXT_COMMAND_LOGICAL_PORT_SHIFT | rni; 1158 phys_index << SCU_CONTEXT_COMMAND_LOGICAL_PORT_SHIFT | rni;
1163 1159
1164 scic_sds_controller_post_request(scic, command); 1160 scic_sds_controller_post_request(ihost, command);
1165} 1161}
1166 1162
1167/** 1163/**
@@ -1175,8 +1171,7 @@ static void scic_sds_port_invalidate_dummy_remote_node(struct isci_port *iport)
1175static void scic_sds_port_ready_substate_operational_exit(struct sci_base_state_machine *sm) 1171static void scic_sds_port_ready_substate_operational_exit(struct sci_base_state_machine *sm)
1176{ 1172{
1177 struct isci_port *iport = container_of(sm, typeof(*iport), sm); 1173 struct isci_port *iport = container_of(sm, typeof(*iport), sm);
1178 struct scic_sds_controller *scic = iport->owning_controller; 1174 struct isci_host *ihost = iport->owning_controller;
1179 struct isci_host *ihost = scic_to_ihost(scic);
1180 1175
1181 /* 1176 /*
1182 * Kill the dummy task for this port if it has not yet posted 1177 * Kill the dummy task for this port if it has not yet posted
@@ -1194,8 +1189,7 @@ static void scic_sds_port_ready_substate_operational_exit(struct sci_base_state_
1194static void scic_sds_port_ready_substate_configuring_enter(struct sci_base_state_machine *sm) 1189static void scic_sds_port_ready_substate_configuring_enter(struct sci_base_state_machine *sm)
1195{ 1190{
1196 struct isci_port *iport = container_of(sm, typeof(*iport), sm); 1191 struct isci_port *iport = container_of(sm, typeof(*iport), sm);
1197 struct scic_sds_controller *scic = iport->owning_controller; 1192 struct isci_host *ihost = iport->owning_controller;
1198 struct isci_host *ihost = scic_to_ihost(scic);
1199 1193
1200 if (iport->active_phy_mask == 0) { 1194 if (iport->active_phy_mask == 0) {
1201 isci_port_not_ready(ihost, iport); 1195 isci_port_not_ready(ihost, iport);
@@ -1218,7 +1212,7 @@ static void scic_sds_port_ready_substate_configuring_exit(struct sci_base_state_
1218 1212
1219enum sci_status scic_sds_port_start(struct isci_port *iport) 1213enum sci_status scic_sds_port_start(struct isci_port *iport)
1220{ 1214{
1221 struct scic_sds_controller *scic = iport->owning_controller; 1215 struct isci_host *ihost = iport->owning_controller;
1222 enum sci_status status = SCI_SUCCESS; 1216 enum sci_status status = SCI_SUCCESS;
1223 enum scic_sds_port_states state; 1217 enum scic_sds_port_states state;
1224 u32 phy_mask; 1218 u32 phy_mask;
@@ -1241,7 +1235,7 @@ enum sci_status scic_sds_port_start(struct isci_port *iport)
1241 1235
1242 if (iport->reserved_rni == SCU_DUMMY_INDEX) { 1236 if (iport->reserved_rni == SCU_DUMMY_INDEX) {
1243 u16 rni = scic_sds_remote_node_table_allocate_remote_node( 1237 u16 rni = scic_sds_remote_node_table_allocate_remote_node(
1244 &scic->available_remote_nodes, 1); 1238 &ihost->available_remote_nodes, 1);
1245 1239
1246 if (rni != SCU_DUMMY_INDEX) 1240 if (rni != SCU_DUMMY_INDEX)
1247 scic_sds_port_construct_dummy_rnc(iport, rni); 1241 scic_sds_port_construct_dummy_rnc(iport, rni);
@@ -1251,7 +1245,6 @@ enum sci_status scic_sds_port_start(struct isci_port *iport)
1251 } 1245 }
1252 1246
1253 if (iport->reserved_tag == SCI_CONTROLLER_INVALID_IO_TAG) { 1247 if (iport->reserved_tag == SCI_CONTROLLER_INVALID_IO_TAG) {
1254 struct isci_host *ihost = scic_to_ihost(scic);
1255 u16 tag; 1248 u16 tag;
1256 1249
1257 tag = isci_alloc_tag(ihost); 1250 tag = isci_alloc_tag(ihost);
@@ -1634,30 +1627,30 @@ scic_sds_port_disable_port_task_scheduler(struct isci_port *iport)
1634 1627
1635static void scic_sds_port_post_dummy_remote_node(struct isci_port *iport) 1628static void scic_sds_port_post_dummy_remote_node(struct isci_port *iport)
1636{ 1629{
1637 struct scic_sds_controller *scic = iport->owning_controller; 1630 struct isci_host *ihost = iport->owning_controller;
1638 u8 phys_index = iport->physical_port_index; 1631 u8 phys_index = iport->physical_port_index;
1639 union scu_remote_node_context *rnc; 1632 union scu_remote_node_context *rnc;
1640 u16 rni = iport->reserved_rni; 1633 u16 rni = iport->reserved_rni;
1641 u32 command; 1634 u32 command;
1642 1635
1643 rnc = &scic->remote_node_context_table[rni]; 1636 rnc = &ihost->remote_node_context_table[rni];
1644 rnc->ssp.is_valid = true; 1637 rnc->ssp.is_valid = true;
1645 1638
1646 command = SCU_CONTEXT_COMMAND_POST_RNC_32 | 1639 command = SCU_CONTEXT_COMMAND_POST_RNC_32 |
1647 phys_index << SCU_CONTEXT_COMMAND_LOGICAL_PORT_SHIFT | rni; 1640 phys_index << SCU_CONTEXT_COMMAND_LOGICAL_PORT_SHIFT | rni;
1648 1641
1649 scic_sds_controller_post_request(scic, command); 1642 scic_sds_controller_post_request(ihost, command);
1650 1643
1651 /* ensure hardware has seen the post rnc command and give it 1644 /* ensure hardware has seen the post rnc command and give it
1652 * ample time to act before sending the suspend 1645 * ample time to act before sending the suspend
1653 */ 1646 */
1654 readl(&scic->smu_registers->interrupt_status); /* flush */ 1647 readl(&ihost->smu_registers->interrupt_status); /* flush */
1655 udelay(10); 1648 udelay(10);
1656 1649
1657 command = SCU_CONTEXT_COMMAND_POST_RNC_SUSPEND_TX_RX | 1650 command = SCU_CONTEXT_COMMAND_POST_RNC_SUSPEND_TX_RX |
1658 phys_index << SCU_CONTEXT_COMMAND_LOGICAL_PORT_SHIFT | rni; 1651 phys_index << SCU_CONTEXT_COMMAND_LOGICAL_PORT_SHIFT | rni;
1659 1652
1660 scic_sds_controller_post_request(scic, command); 1653 scic_sds_controller_post_request(ihost, command);
1661} 1654}
1662 1655
1663static void scic_sds_port_stopped_state_enter(struct sci_base_state_machine *sm) 1656static void scic_sds_port_stopped_state_enter(struct sci_base_state_machine *sm)
@@ -1684,8 +1677,7 @@ static void scic_sds_port_stopped_state_exit(struct sci_base_state_machine *sm)
1684static void scic_sds_port_ready_state_enter(struct sci_base_state_machine *sm) 1677static void scic_sds_port_ready_state_enter(struct sci_base_state_machine *sm)
1685{ 1678{
1686 struct isci_port *iport = container_of(sm, typeof(*iport), sm); 1679 struct isci_port *iport = container_of(sm, typeof(*iport), sm);
1687 struct scic_sds_controller *scic = iport->owning_controller; 1680 struct isci_host *ihost = iport->owning_controller;
1688 struct isci_host *ihost = scic_to_ihost(scic);
1689 u32 prev_state; 1681 u32 prev_state;
1690 1682
1691 prev_state = iport->sm.previous_state_id; 1683 prev_state = iport->sm.previous_state_id;
@@ -1758,7 +1750,7 @@ static const struct sci_base_state scic_sds_port_state_table[] = {
1758}; 1750};
1759 1751
1760void scic_sds_port_construct(struct isci_port *iport, u8 index, 1752void scic_sds_port_construct(struct isci_port *iport, u8 index,
1761 struct scic_sds_controller *scic) 1753 struct isci_host *ihost)
1762{ 1754{
1763 sci_init_sm(&iport->sm, scic_sds_port_state_table, SCI_PORT_STOPPED); 1755 sci_init_sm(&iport->sm, scic_sds_port_state_table, SCI_PORT_STOPPED);
1764 1756
@@ -1767,7 +1759,7 @@ void scic_sds_port_construct(struct isci_port *iport, u8 index,
1767 iport->active_phy_mask = 0; 1759 iport->active_phy_mask = 0;
1768 iport->ready_exit = false; 1760 iport->ready_exit = false;
1769 1761
1770 iport->owning_controller = scic; 1762 iport->owning_controller = ihost;
1771 1763
1772 iport->started_request_count = 0; 1764 iport->started_request_count = 0;
1773 iport->assigned_device_count = 0; 1765 iport->assigned_device_count = 0;
@@ -1810,8 +1802,7 @@ void scic_sds_port_broadcast_change_received(
1810 struct isci_port *iport, 1802 struct isci_port *iport,
1811 struct isci_phy *iphy) 1803 struct isci_phy *iphy)
1812{ 1804{
1813 struct scic_sds_controller *scic = iport->owning_controller; 1805 struct isci_host *ihost = iport->owning_controller;
1814 struct isci_host *ihost = scic_to_ihost(scic);
1815 1806
1816 /* notify the user. */ 1807 /* notify the user. */
1817 isci_port_bc_change_received(ihost, iport, iphy); 1808 isci_port_bc_change_received(ihost, iport, iphy);
diff --git a/drivers/scsi/isci/port.h b/drivers/scsi/isci/port.h
index b9bc89bf6519..9a9be7b47b4a 100644
--- a/drivers/scsi/isci/port.h
+++ b/drivers/scsi/isci/port.h
@@ -115,7 +115,7 @@ struct isci_port {
115 u32 assigned_device_count; 115 u32 assigned_device_count;
116 u32 not_ready_reason; 116 u32 not_ready_reason;
117 struct isci_phy *phy_table[SCI_MAX_PHYS]; 117 struct isci_phy *phy_table[SCI_MAX_PHYS];
118 struct scic_sds_controller *owning_controller; 118 struct isci_host *owning_controller;
119 struct sci_timer timer; 119 struct sci_timer timer;
120 struct scu_port_task_scheduler_registers __iomem *port_task_scheduler_registers; 120 struct scu_port_task_scheduler_registers __iomem *port_task_scheduler_registers;
121 /* XXX rework: only one register, no need to replicate per-port */ 121 /* XXX rework: only one register, no need to replicate per-port */
@@ -243,7 +243,7 @@ static inline void scic_sds_port_decrement_request_count(struct isci_port *iport
243void scic_sds_port_construct( 243void scic_sds_port_construct(
244 struct isci_port *iport, 244 struct isci_port *iport,
245 u8 port_index, 245 u8 port_index,
246 struct scic_sds_controller *scic); 246 struct isci_host *ihost);
247 247
248enum sci_status scic_sds_port_initialize( 248enum sci_status scic_sds_port_initialize(
249 struct isci_port *iport, 249 struct isci_port *iport,
diff --git a/drivers/scsi/isci/port_config.c b/drivers/scsi/isci/port_config.c
index bb62d2a25217..a0a135d54e95 100644
--- a/drivers/scsi/isci/port_config.c
+++ b/drivers/scsi/isci/port_config.c
@@ -113,7 +113,7 @@ static s32 sci_sas_address_compare(
113 * NULL if there is no matching port for the phy. 113 * NULL if there is no matching port for the phy.
114 */ 114 */
115static struct isci_port *scic_sds_port_configuration_agent_find_port( 115static struct isci_port *scic_sds_port_configuration_agent_find_port(
116 struct scic_sds_controller *scic, 116 struct isci_host *ihost,
117 struct isci_phy *iphy) 117 struct isci_phy *iphy)
118{ 118{
119 u8 i; 119 u8 i;
@@ -130,8 +130,7 @@ static struct isci_port *scic_sds_port_configuration_agent_find_port(
130 scic_sds_phy_get_sas_address(iphy, &phy_sas_address); 130 scic_sds_phy_get_sas_address(iphy, &phy_sas_address);
131 scic_sds_phy_get_attached_sas_address(iphy, &phy_attached_device_address); 131 scic_sds_phy_get_attached_sas_address(iphy, &phy_attached_device_address);
132 132
133 for (i = 0; i < scic->logical_port_entries; i++) { 133 for (i = 0; i < ihost->logical_port_entries; i++) {
134 struct isci_host *ihost = scic_to_ihost(scic);
135 struct isci_port *iport = &ihost->ports[i]; 134 struct isci_port *iport = &ihost->ports[i];
136 135
137 scic_sds_port_get_sas_address(iport, &port_sas_address); 136 scic_sds_port_get_sas_address(iport, &port_sas_address);
@@ -158,10 +157,9 @@ static struct isci_port *scic_sds_port_configuration_agent_find_port(
158 * the port configuration is not valid for this port configuration agent. 157 * the port configuration is not valid for this port configuration agent.
159 */ 158 */
160static enum sci_status scic_sds_port_configuration_agent_validate_ports( 159static enum sci_status scic_sds_port_configuration_agent_validate_ports(
161 struct scic_sds_controller *controller, 160 struct isci_host *ihost,
162 struct scic_sds_port_configuration_agent *port_agent) 161 struct scic_sds_port_configuration_agent *port_agent)
163{ 162{
164 struct isci_host *ihost = scic_to_ihost(controller);
165 struct sci_sas_address first_address; 163 struct sci_sas_address first_address;
166 struct sci_sas_address second_address; 164 struct sci_sas_address second_address;
167 165
@@ -239,17 +237,11 @@ static enum sci_status scic_sds_port_configuration_agent_validate_ports(
239 * Manual port configuration agent routines 237 * Manual port configuration agent routines
240 * ****************************************************************************** */ 238 * ****************************************************************************** */
241 239
242/** 240/* verify all of the phys in the same port are using the same SAS address */
243 * 241static enum sci_status
244 * 242scic_sds_mpc_agent_validate_phy_configuration(struct isci_host *ihost,
245 * This routine will verify that all of the phys in the same port are using the 243 struct scic_sds_port_configuration_agent *port_agent)
246 * same SAS address.
247 */
248static enum sci_status scic_sds_mpc_agent_validate_phy_configuration(
249 struct scic_sds_controller *controller,
250 struct scic_sds_port_configuration_agent *port_agent)
251{ 244{
252 struct isci_host *ihost = scic_to_ihost(controller);
253 u32 phy_mask; 245 u32 phy_mask;
254 u32 assigned_phy_mask; 246 u32 assigned_phy_mask;
255 struct sci_sas_address sas_address; 247 struct sci_sas_address sas_address;
@@ -262,7 +254,7 @@ static enum sci_status scic_sds_mpc_agent_validate_phy_configuration(
262 sas_address.low = 0; 254 sas_address.low = 0;
263 255
264 for (port_index = 0; port_index < SCI_MAX_PORTS; port_index++) { 256 for (port_index = 0; port_index < SCI_MAX_PORTS; port_index++) {
265 phy_mask = controller->oem_parameters.sds1.ports[port_index].phy_mask; 257 phy_mask = ihost->oem_parameters.sds1.ports[port_index].phy_mask;
266 258
267 if (!phy_mask) 259 if (!phy_mask)
268 continue; 260 continue;
@@ -324,7 +316,7 @@ static enum sci_status scic_sds_mpc_agent_validate_phy_configuration(
324 phy_index++; 316 phy_index++;
325 } 317 }
326 318
327 return scic_sds_port_configuration_agent_validate_ports(controller, port_agent); 319 return scic_sds_port_configuration_agent_validate_ports(ihost, port_agent);
328} 320}
329 321
330static void mpc_agent_timeout(unsigned long data) 322static void mpc_agent_timeout(unsigned long data)
@@ -332,14 +324,12 @@ static void mpc_agent_timeout(unsigned long data)
332 u8 index; 324 u8 index;
333 struct sci_timer *tmr = (struct sci_timer *)data; 325 struct sci_timer *tmr = (struct sci_timer *)data;
334 struct scic_sds_port_configuration_agent *port_agent; 326 struct scic_sds_port_configuration_agent *port_agent;
335 struct scic_sds_controller *scic;
336 struct isci_host *ihost; 327 struct isci_host *ihost;
337 unsigned long flags; 328 unsigned long flags;
338 u16 configure_phy_mask; 329 u16 configure_phy_mask;
339 330
340 port_agent = container_of(tmr, typeof(*port_agent), timer); 331 port_agent = container_of(tmr, typeof(*port_agent), timer);
341 scic = container_of(port_agent, typeof(*scic), port_agent); 332 ihost = container_of(port_agent, typeof(*ihost), port_agent);
342 ihost = scic_to_ihost(scic);
343 333
344 spin_lock_irqsave(&ihost->scic_lock, flags); 334 spin_lock_irqsave(&ihost->scic_lock, flags);
345 335
@@ -355,7 +345,7 @@ static void mpc_agent_timeout(unsigned long data)
355 struct isci_phy *iphy = &ihost->phys[index]; 345 struct isci_phy *iphy = &ihost->phys[index];
356 346
357 if (configure_phy_mask & (1 << index)) { 347 if (configure_phy_mask & (1 << index)) {
358 port_agent->link_up_handler(scic, port_agent, 348 port_agent->link_up_handler(ihost, port_agent,
359 phy_get_non_dummy_port(iphy), 349 phy_get_non_dummy_port(iphy),
360 iphy); 350 iphy);
361 } 351 }
@@ -365,7 +355,7 @@ done:
365 spin_unlock_irqrestore(&ihost->scic_lock, flags); 355 spin_unlock_irqrestore(&ihost->scic_lock, flags);
366} 356}
367 357
368static void scic_sds_mpc_agent_link_up(struct scic_sds_controller *controller, 358static void scic_sds_mpc_agent_link_up(struct isci_host *ihost,
369 struct scic_sds_port_configuration_agent *port_agent, 359 struct scic_sds_port_configuration_agent *port_agent,
370 struct isci_port *iport, 360 struct isci_port *iport,
371 struct isci_phy *iphy) 361 struct isci_phy *iphy)
@@ -401,7 +391,7 @@ static void scic_sds_mpc_agent_link_up(struct scic_sds_controller *controller,
401 * link down notification from a phy that has no assocoated port? 391 * link down notification from a phy that has no assocoated port?
402 */ 392 */
403static void scic_sds_mpc_agent_link_down( 393static void scic_sds_mpc_agent_link_down(
404 struct scic_sds_controller *scic, 394 struct isci_host *ihost,
405 struct scic_sds_port_configuration_agent *port_agent, 395 struct scic_sds_port_configuration_agent *port_agent,
406 struct isci_port *iport, 396 struct isci_port *iport,
407 struct isci_phy *iphy) 397 struct isci_phy *iphy)
@@ -438,26 +428,17 @@ static void scic_sds_mpc_agent_link_down(
438 } 428 }
439} 429}
440 430
441/* 431/* verify phys are assigned a valid SAS address for automatic port
442 * ****************************************************************************** 432 * configuration mode.
443 * Automatic port configuration agent routines
444 * ****************************************************************************** */
445
446/**
447 *
448 *
449 * This routine will verify that the phys are assigned a valid SAS address for
450 * automatic port configuration mode.
451 */ 433 */
452static enum sci_status scic_sds_apc_agent_validate_phy_configuration( 434static enum sci_status
453 struct scic_sds_controller *controller, 435scic_sds_apc_agent_validate_phy_configuration(struct isci_host *ihost,
454 struct scic_sds_port_configuration_agent *port_agent) 436 struct scic_sds_port_configuration_agent *port_agent)
455{ 437{
456 u8 phy_index; 438 u8 phy_index;
457 u8 port_index; 439 u8 port_index;
458 struct sci_sas_address sas_address; 440 struct sci_sas_address sas_address;
459 struct sci_sas_address phy_assigned_address; 441 struct sci_sas_address phy_assigned_address;
460 struct isci_host *ihost = scic_to_ihost(controller);
461 442
462 phy_index = 0; 443 phy_index = 0;
463 444
@@ -484,10 +465,10 @@ static enum sci_status scic_sds_apc_agent_validate_phy_configuration(
484 } 465 }
485 } 466 }
486 467
487 return scic_sds_port_configuration_agent_validate_ports(controller, port_agent); 468 return scic_sds_port_configuration_agent_validate_ports(ihost, port_agent);
488} 469}
489 470
490static void scic_sds_apc_agent_configure_ports(struct scic_sds_controller *controller, 471static void scic_sds_apc_agent_configure_ports(struct isci_host *ihost,
491 struct scic_sds_port_configuration_agent *port_agent, 472 struct scic_sds_port_configuration_agent *port_agent,
492 struct isci_phy *iphy, 473 struct isci_phy *iphy,
493 bool start_timer) 474 bool start_timer)
@@ -496,9 +477,8 @@ static void scic_sds_apc_agent_configure_ports(struct scic_sds_controller *contr
496 enum sci_status status; 477 enum sci_status status;
497 struct isci_port *iport; 478 struct isci_port *iport;
498 enum SCIC_SDS_APC_ACTIVITY apc_activity = SCIC_SDS_APC_SKIP_PHY; 479 enum SCIC_SDS_APC_ACTIVITY apc_activity = SCIC_SDS_APC_SKIP_PHY;
499 struct isci_host *ihost = scic_to_ihost(controller);
500 480
501 iport = scic_sds_port_configuration_agent_find_port(controller, iphy); 481 iport = scic_sds_port_configuration_agent_find_port(ihost, iphy);
502 482
503 if (iport) { 483 if (iport) {
504 if (scic_sds_port_is_valid_phy_assignment(iport, iphy->phy_index)) 484 if (scic_sds_port_is_valid_phy_assignment(iport, iphy->phy_index))
@@ -619,7 +599,7 @@ static void scic_sds_apc_agent_configure_ports(struct scic_sds_controller *contr
619 * notifications. Is it possible to get a link down notification from a phy 599 * notifications. Is it possible to get a link down notification from a phy
620 * that has no assocoated port? 600 * that has no assocoated port?
621 */ 601 */
622static void scic_sds_apc_agent_link_up(struct scic_sds_controller *scic, 602static void scic_sds_apc_agent_link_up(struct isci_host *ihost,
623 struct scic_sds_port_configuration_agent *port_agent, 603 struct scic_sds_port_configuration_agent *port_agent,
624 struct isci_port *iport, 604 struct isci_port *iport,
625 struct isci_phy *iphy) 605 struct isci_phy *iphy)
@@ -629,7 +609,7 @@ static void scic_sds_apc_agent_link_up(struct scic_sds_controller *scic,
629 if (!iport) { 609 if (!iport) {
630 /* the phy is not the part of this port */ 610 /* the phy is not the part of this port */
631 port_agent->phy_ready_mask |= 1 << phy_index; 611 port_agent->phy_ready_mask |= 1 << phy_index;
632 scic_sds_apc_agent_configure_ports(scic, port_agent, iphy, true); 612 scic_sds_apc_agent_configure_ports(ihost, port_agent, iphy, true);
633 } else { 613 } else {
634 /* the phy is already the part of the port */ 614 /* the phy is already the part of the port */
635 u32 port_state = iport->sm.current_state_id; 615 u32 port_state = iport->sm.current_state_id;
@@ -658,7 +638,7 @@ static void scic_sds_apc_agent_link_up(struct scic_sds_controller *scic,
658 * port? 638 * port?
659 */ 639 */
660static void scic_sds_apc_agent_link_down( 640static void scic_sds_apc_agent_link_down(
661 struct scic_sds_controller *controller, 641 struct isci_host *ihost,
662 struct scic_sds_port_configuration_agent *port_agent, 642 struct scic_sds_port_configuration_agent *port_agent,
663 struct isci_port *iport, 643 struct isci_port *iport,
664 struct isci_phy *iphy) 644 struct isci_phy *iphy)
@@ -683,14 +663,12 @@ static void apc_agent_timeout(unsigned long data)
683 u32 index; 663 u32 index;
684 struct sci_timer *tmr = (struct sci_timer *)data; 664 struct sci_timer *tmr = (struct sci_timer *)data;
685 struct scic_sds_port_configuration_agent *port_agent; 665 struct scic_sds_port_configuration_agent *port_agent;
686 struct scic_sds_controller *scic;
687 struct isci_host *ihost; 666 struct isci_host *ihost;
688 unsigned long flags; 667 unsigned long flags;
689 u16 configure_phy_mask; 668 u16 configure_phy_mask;
690 669
691 port_agent = container_of(tmr, typeof(*port_agent), timer); 670 port_agent = container_of(tmr, typeof(*port_agent), timer);
692 scic = container_of(port_agent, typeof(*scic), port_agent); 671 ihost = container_of(port_agent, typeof(*ihost), port_agent);
693 ihost = scic_to_ihost(scic);
694 672
695 spin_lock_irqsave(&ihost->scic_lock, flags); 673 spin_lock_irqsave(&ihost->scic_lock, flags);
696 674
@@ -708,7 +686,7 @@ static void apc_agent_timeout(unsigned long data)
708 if ((configure_phy_mask & (1 << index)) == 0) 686 if ((configure_phy_mask & (1 << index)) == 0)
709 continue; 687 continue;
710 688
711 scic_sds_apc_agent_configure_ports(scic, port_agent, 689 scic_sds_apc_agent_configure_ports(ihost, port_agent,
712 &ihost->phys[index], false); 690 &ihost->phys[index], false);
713 } 691 }
714 692
@@ -748,17 +726,17 @@ void scic_sds_port_configuration_agent_construct(
748} 726}
749 727
750enum sci_status scic_sds_port_configuration_agent_initialize( 728enum sci_status scic_sds_port_configuration_agent_initialize(
751 struct scic_sds_controller *scic, 729 struct isci_host *ihost,
752 struct scic_sds_port_configuration_agent *port_agent) 730 struct scic_sds_port_configuration_agent *port_agent)
753{ 731{
754 enum sci_status status; 732 enum sci_status status;
755 enum scic_port_configuration_mode mode; 733 enum scic_port_configuration_mode mode;
756 734
757 mode = scic->oem_parameters.sds1.controller.mode_type; 735 mode = ihost->oem_parameters.sds1.controller.mode_type;
758 736
759 if (mode == SCIC_PORT_MANUAL_CONFIGURATION_MODE) { 737 if (mode == SCIC_PORT_MANUAL_CONFIGURATION_MODE) {
760 status = scic_sds_mpc_agent_validate_phy_configuration( 738 status = scic_sds_mpc_agent_validate_phy_configuration(
761 scic, port_agent); 739 ihost, port_agent);
762 740
763 port_agent->link_up_handler = scic_sds_mpc_agent_link_up; 741 port_agent->link_up_handler = scic_sds_mpc_agent_link_up;
764 port_agent->link_down_handler = scic_sds_mpc_agent_link_down; 742 port_agent->link_down_handler = scic_sds_mpc_agent_link_down;
@@ -766,7 +744,7 @@ enum sci_status scic_sds_port_configuration_agent_initialize(
766 sci_init_timer(&port_agent->timer, mpc_agent_timeout); 744 sci_init_timer(&port_agent->timer, mpc_agent_timeout);
767 } else { 745 } else {
768 status = scic_sds_apc_agent_validate_phy_configuration( 746 status = scic_sds_apc_agent_validate_phy_configuration(
769 scic, port_agent); 747 ihost, port_agent);
770 748
771 port_agent->link_up_handler = scic_sds_apc_agent_link_up; 749 port_agent->link_up_handler = scic_sds_apc_agent_link_up;
772 port_agent->link_down_handler = scic_sds_apc_agent_link_down; 750 port_agent->link_down_handler = scic_sds_apc_agent_link_down;
diff --git a/drivers/scsi/isci/probe_roms.h b/drivers/scsi/isci/probe_roms.h
index 95c8d91aab8d..e40cb5f6eba5 100644
--- a/drivers/scsi/isci/probe_roms.h
+++ b/drivers/scsi/isci/probe_roms.h
@@ -165,7 +165,7 @@ struct scic_sds_oem_params;
165int scic_oem_parameters_validate(struct scic_sds_oem_params *oem); 165int scic_oem_parameters_validate(struct scic_sds_oem_params *oem);
166 166
167union scic_oem_parameters; 167union scic_oem_parameters;
168void scic_oem_parameters_get(struct scic_sds_controller *scic, 168void scic_oem_parameters_get(struct isci_host *ihost,
169 union scic_oem_parameters *oem); 169 union scic_oem_parameters *oem);
170 170
171struct isci_orom; 171struct isci_orom;
diff --git a/drivers/scsi/isci/remote_device.c b/drivers/scsi/isci/remote_device.c
index 3b0234049a3d..9043b458c999 100644
--- a/drivers/scsi/isci/remote_device.c
+++ b/drivers/scsi/isci/remote_device.c
@@ -62,7 +62,7 @@
62#include "task.h" 62#include "task.h"
63 63
64/** 64/**
65 * isci_remote_device_not_ready() - This function is called by the scic when 65 * isci_remote_device_not_ready() - This function is called by the ihost when
66 * the remote device is not ready. We mark the isci device as ready (not 66 * the remote device is not ready. We mark the isci device as ready (not
67 * "ready_for_io") and signal the waiting proccess. 67 * "ready_for_io") and signal the waiting proccess.
68 * @isci_host: This parameter specifies the isci host object. 68 * @isci_host: This parameter specifies the isci host object.
@@ -92,7 +92,7 @@ static void isci_remote_device_not_ready(struct isci_host *ihost,
92 "%s: isci_device = %p request = %p\n", 92 "%s: isci_device = %p request = %p\n",
93 __func__, idev, ireq); 93 __func__, idev, ireq);
94 94
95 scic_controller_terminate_request(&ihost->sci, 95 scic_controller_terminate_request(ihost,
96 idev, 96 idev,
97 ireq); 97 ireq);
98 } 98 }
@@ -104,7 +104,7 @@ static void isci_remote_device_not_ready(struct isci_host *ihost,
104} 104}
105 105
106/** 106/**
107 * isci_remote_device_ready() - This function is called by the scic when the 107 * isci_remote_device_ready() - This function is called by the ihost when the
108 * remote device is ready. We mark the isci device as ready and signal the 108 * remote device is ready. We mark the isci device as ready and signal the
109 * waiting proccess. 109 * waiting proccess.
110 * @ihost: our valid isci_host 110 * @ihost: our valid isci_host
@@ -135,8 +135,7 @@ static void rnc_destruct_done(void *_dev)
135 135
136static enum sci_status scic_sds_remote_device_terminate_requests(struct isci_remote_device *idev) 136static enum sci_status scic_sds_remote_device_terminate_requests(struct isci_remote_device *idev)
137{ 137{
138 struct scic_sds_controller *scic = idev->owning_port->owning_controller; 138 struct isci_host *ihost = idev->owning_port->owning_controller;
139 struct isci_host *ihost = scic_to_ihost(scic);
140 enum sci_status status = SCI_SUCCESS; 139 enum sci_status status = SCI_SUCCESS;
141 u32 i; 140 u32 i;
142 141
@@ -148,7 +147,7 @@ static enum sci_status scic_sds_remote_device_terminate_requests(struct isci_rem
148 ireq->target_device != idev) 147 ireq->target_device != idev)
149 continue; 148 continue;
150 149
151 s = scic_controller_terminate_request(scic, idev, ireq); 150 s = scic_controller_terminate_request(ihost, idev, ireq);
152 if (s != SCI_SUCCESS) 151 if (s != SCI_SUCCESS)
153 status = s; 152 status = s;
154 } 153 }
@@ -276,7 +275,7 @@ enum sci_status scic_sds_remote_device_frame_handler(struct isci_remote_device *
276{ 275{
277 struct sci_base_state_machine *sm = &idev->sm; 276 struct sci_base_state_machine *sm = &idev->sm;
278 enum scic_sds_remote_device_states state = sm->current_state_id; 277 enum scic_sds_remote_device_states state = sm->current_state_id;
279 struct scic_sds_controller *scic = idev->owning_port->owning_controller; 278 struct isci_host *ihost = idev->owning_port->owning_controller;
280 enum sci_status status; 279 enum sci_status status;
281 280
282 switch (state) { 281 switch (state) {
@@ -290,7 +289,7 @@ enum sci_status scic_sds_remote_device_frame_handler(struct isci_remote_device *
290 dev_warn(scirdev_to_dev(idev), "%s: in wrong state: %d\n", 289 dev_warn(scirdev_to_dev(idev), "%s: in wrong state: %d\n",
291 __func__, state); 290 __func__, state);
292 /* Return the frame back to the controller */ 291 /* Return the frame back to the controller */
293 scic_sds_controller_release_frame(scic, frame_index); 292 scic_sds_controller_release_frame(ihost, frame_index);
294 return SCI_FAILURE_INVALID_STATE; 293 return SCI_FAILURE_INVALID_STATE;
295 case SCI_DEV_READY: 294 case SCI_DEV_READY:
296 case SCI_STP_DEV_NCQ_ERROR: 295 case SCI_STP_DEV_NCQ_ERROR:
@@ -303,7 +302,7 @@ enum sci_status scic_sds_remote_device_frame_handler(struct isci_remote_device *
303 void *frame_header; 302 void *frame_header;
304 ssize_t word_cnt; 303 ssize_t word_cnt;
305 304
306 status = scic_sds_unsolicited_frame_control_get_header(&scic->uf_control, 305 status = scic_sds_unsolicited_frame_control_get_header(&ihost->uf_control,
307 frame_index, 306 frame_index,
308 &frame_header); 307 &frame_header);
309 if (status != SCI_SUCCESS) 308 if (status != SCI_SUCCESS)
@@ -312,7 +311,7 @@ enum sci_status scic_sds_remote_device_frame_handler(struct isci_remote_device *
312 word_cnt = sizeof(hdr) / sizeof(u32); 311 word_cnt = sizeof(hdr) / sizeof(u32);
313 sci_swab32_cpy(&hdr, frame_header, word_cnt); 312 sci_swab32_cpy(&hdr, frame_header, word_cnt);
314 313
315 ireq = scic_request_by_tag(scic, be16_to_cpu(hdr.tag)); 314 ireq = scic_request_by_tag(ihost, be16_to_cpu(hdr.tag));
316 if (ireq && ireq->target_device == idev) { 315 if (ireq && ireq->target_device == idev) {
317 /* The IO request is now in charge of releasing the frame */ 316 /* The IO request is now in charge of releasing the frame */
318 status = scic_sds_io_request_frame_handler(ireq, frame_index); 317 status = scic_sds_io_request_frame_handler(ireq, frame_index);
@@ -320,14 +319,14 @@ enum sci_status scic_sds_remote_device_frame_handler(struct isci_remote_device *
320 /* We could not map this tag to a valid IO 319 /* We could not map this tag to a valid IO
321 * request Just toss the frame and continue 320 * request Just toss the frame and continue
322 */ 321 */
323 scic_sds_controller_release_frame(scic, frame_index); 322 scic_sds_controller_release_frame(ihost, frame_index);
324 } 323 }
325 break; 324 break;
326 } 325 }
327 case SCI_STP_DEV_NCQ: { 326 case SCI_STP_DEV_NCQ: {
328 struct dev_to_host_fis *hdr; 327 struct dev_to_host_fis *hdr;
329 328
330 status = scic_sds_unsolicited_frame_control_get_header(&scic->uf_control, 329 status = scic_sds_unsolicited_frame_control_get_header(&ihost->uf_control,
331 frame_index, 330 frame_index,
332 (void **)&hdr); 331 (void **)&hdr);
333 if (status != SCI_SUCCESS) 332 if (status != SCI_SUCCESS)
@@ -350,7 +349,7 @@ enum sci_status scic_sds_remote_device_frame_handler(struct isci_remote_device *
350 } else 349 } else
351 status = SCI_FAILURE; 350 status = SCI_FAILURE;
352 351
353 scic_sds_controller_release_frame(scic, frame_index); 352 scic_sds_controller_release_frame(ihost, frame_index);
354 break; 353 break;
355 } 354 }
356 case SCI_STP_DEV_CMD: 355 case SCI_STP_DEV_CMD:
@@ -461,7 +460,7 @@ static void scic_sds_remote_device_start_request(struct isci_remote_device *idev
461 } 460 }
462} 461}
463 462
464enum sci_status scic_sds_remote_device_start_io(struct scic_sds_controller *scic, 463enum sci_status scic_sds_remote_device_start_io(struct isci_host *ihost,
465 struct isci_remote_device *idev, 464 struct isci_remote_device *idev,
466 struct isci_request *ireq) 465 struct isci_request *ireq)
467{ 466{
@@ -597,7 +596,7 @@ static enum sci_status common_complete_io(struct isci_port *iport,
597 return status; 596 return status;
598} 597}
599 598
600enum sci_status scic_sds_remote_device_complete_io(struct scic_sds_controller *scic, 599enum sci_status scic_sds_remote_device_complete_io(struct isci_host *ihost,
601 struct isci_remote_device *idev, 600 struct isci_remote_device *idev,
602 struct isci_request *ireq) 601 struct isci_request *ireq)
603{ 602{
@@ -678,7 +677,7 @@ static void scic_sds_remote_device_continue_request(void *dev)
678 scic_controller_continue_io(idev->working_request); 677 scic_controller_continue_io(idev->working_request);
679} 678}
680 679
681enum sci_status scic_sds_remote_device_start_task(struct scic_sds_controller *scic, 680enum sci_status scic_sds_remote_device_start_task(struct isci_host *ihost,
682 struct isci_remote_device *idev, 681 struct isci_remote_device *idev,
683 struct isci_request *ireq) 682 struct isci_request *ireq)
684{ 683{
@@ -802,13 +801,13 @@ static void remote_device_resume_done(void *_dev)
802static void scic_sds_stp_remote_device_ready_idle_substate_resume_complete_handler(void *_dev) 801static void scic_sds_stp_remote_device_ready_idle_substate_resume_complete_handler(void *_dev)
803{ 802{
804 struct isci_remote_device *idev = _dev; 803 struct isci_remote_device *idev = _dev;
805 struct scic_sds_controller *scic = idev->owning_port->owning_controller; 804 struct isci_host *ihost = idev->owning_port->owning_controller;
806 805
807 /* For NCQ operation we do not issue a isci_remote_device_not_ready(). 806 /* For NCQ operation we do not issue a isci_remote_device_not_ready().
808 * As a result, avoid sending the ready notification. 807 * As a result, avoid sending the ready notification.
809 */ 808 */
810 if (idev->sm.previous_state_id != SCI_STP_DEV_NCQ) 809 if (idev->sm.previous_state_id != SCI_STP_DEV_NCQ)
811 isci_remote_device_ready(scic_to_ihost(scic), idev); 810 isci_remote_device_ready(ihost, idev);
812} 811}
813 812
814static void scic_sds_remote_device_initial_state_enter(struct sci_base_state_machine *sm) 813static void scic_sds_remote_device_initial_state_enter(struct sci_base_state_machine *sm)
@@ -836,7 +835,7 @@ static enum sci_status scic_remote_device_destruct(struct isci_remote_device *id
836{ 835{
837 struct sci_base_state_machine *sm = &idev->sm; 836 struct sci_base_state_machine *sm = &idev->sm;
838 enum scic_sds_remote_device_states state = sm->current_state_id; 837 enum scic_sds_remote_device_states state = sm->current_state_id;
839 struct scic_sds_controller *scic; 838 struct isci_host *ihost;
840 839
841 if (state != SCI_DEV_STOPPED) { 840 if (state != SCI_DEV_STOPPED) {
842 dev_warn(scirdev_to_dev(idev), "%s: in wrong state: %d\n", 841 dev_warn(scirdev_to_dev(idev), "%s: in wrong state: %d\n",
@@ -844,8 +843,8 @@ static enum sci_status scic_remote_device_destruct(struct isci_remote_device *id
844 return SCI_FAILURE_INVALID_STATE; 843 return SCI_FAILURE_INVALID_STATE;
845 } 844 }
846 845
847 scic = idev->owning_port->owning_controller; 846 ihost = idev->owning_port->owning_controller;
848 scic_sds_controller_free_remote_node_context(scic, idev, 847 scic_sds_controller_free_remote_node_context(ihost, idev,
849 idev->rnc.remote_node_index); 848 idev->rnc.remote_node_index);
850 idev->rnc.remote_node_index = SCIC_SDS_REMOTE_NODE_CONTEXT_INVALID_INDEX; 849 idev->rnc.remote_node_index = SCIC_SDS_REMOTE_NODE_CONTEXT_INVALID_INDEX;
851 sci_change_state(sm, SCI_DEV_FINAL); 850 sci_change_state(sm, SCI_DEV_FINAL);
@@ -878,7 +877,7 @@ static void isci_remote_device_deconstruct(struct isci_host *ihost, struct isci_
878static void scic_sds_remote_device_stopped_state_enter(struct sci_base_state_machine *sm) 877static void scic_sds_remote_device_stopped_state_enter(struct sci_base_state_machine *sm)
879{ 878{
880 struct isci_remote_device *idev = container_of(sm, typeof(*idev), sm); 879 struct isci_remote_device *idev = container_of(sm, typeof(*idev), sm);
881 struct scic_sds_controller *scic = idev->owning_port->owning_controller; 880 struct isci_host *ihost = idev->owning_port->owning_controller;
882 u32 prev_state; 881 u32 prev_state;
883 882
884 /* If we are entering from the stopping state let the SCI User know that 883 /* If we are entering from the stopping state let the SCI User know that
@@ -886,16 +885,15 @@ static void scic_sds_remote_device_stopped_state_enter(struct sci_base_state_mac
886 */ 885 */
887 prev_state = idev->sm.previous_state_id; 886 prev_state = idev->sm.previous_state_id;
888 if (prev_state == SCI_DEV_STOPPING) 887 if (prev_state == SCI_DEV_STOPPING)
889 isci_remote_device_deconstruct(scic_to_ihost(scic), idev); 888 isci_remote_device_deconstruct(ihost, idev);
890 889
891 scic_sds_controller_remote_device_stopped(scic, idev); 890 scic_sds_controller_remote_device_stopped(ihost, idev);
892} 891}
893 892
894static void scic_sds_remote_device_starting_state_enter(struct sci_base_state_machine *sm) 893static void scic_sds_remote_device_starting_state_enter(struct sci_base_state_machine *sm)
895{ 894{
896 struct isci_remote_device *idev = container_of(sm, typeof(*idev), sm); 895 struct isci_remote_device *idev = container_of(sm, typeof(*idev), sm);
897 struct scic_sds_controller *scic = scic_sds_remote_device_get_controller(idev); 896 struct isci_host *ihost = scic_sds_remote_device_get_controller(idev);
898 struct isci_host *ihost = scic_to_ihost(scic);
899 897
900 isci_remote_device_not_ready(ihost, idev, 898 isci_remote_device_not_ready(ihost, idev,
901 SCIC_REMOTE_DEVICE_NOT_READY_START_REQUESTED); 899 SCIC_REMOTE_DEVICE_NOT_READY_START_REQUESTED);
@@ -904,7 +902,7 @@ static void scic_sds_remote_device_starting_state_enter(struct sci_base_state_ma
904static void scic_sds_remote_device_ready_state_enter(struct sci_base_state_machine *sm) 902static void scic_sds_remote_device_ready_state_enter(struct sci_base_state_machine *sm)
905{ 903{
906 struct isci_remote_device *idev = container_of(sm, typeof(*idev), sm); 904 struct isci_remote_device *idev = container_of(sm, typeof(*idev), sm);
907 struct scic_sds_controller *scic = idev->owning_port->owning_controller; 905 struct isci_host *ihost = idev->owning_port->owning_controller;
908 struct domain_device *dev = idev->domain_dev; 906 struct domain_device *dev = idev->domain_dev;
909 907
910 if (dev->dev_type == SATA_DEV || (dev->tproto & SAS_PROTOCOL_SATA)) { 908 if (dev->dev_type == SATA_DEV || (dev->tproto & SAS_PROTOCOL_SATA)) {
@@ -912,7 +910,7 @@ static void scic_sds_remote_device_ready_state_enter(struct sci_base_state_machi
912 } else if (dev_is_expander(dev)) { 910 } else if (dev_is_expander(dev)) {
913 sci_change_state(&idev->sm, SCI_SMP_DEV_IDLE); 911 sci_change_state(&idev->sm, SCI_SMP_DEV_IDLE);
914 } else 912 } else
915 isci_remote_device_ready(scic_to_ihost(scic), idev); 913 isci_remote_device_ready(ihost, idev);
916} 914}
917 915
918static void scic_sds_remote_device_ready_state_exit(struct sci_base_state_machine *sm) 916static void scic_sds_remote_device_ready_state_exit(struct sci_base_state_machine *sm)
@@ -921,9 +919,9 @@ static void scic_sds_remote_device_ready_state_exit(struct sci_base_state_machin
921 struct domain_device *dev = idev->domain_dev; 919 struct domain_device *dev = idev->domain_dev;
922 920
923 if (dev->dev_type == SAS_END_DEV) { 921 if (dev->dev_type == SAS_END_DEV) {
924 struct scic_sds_controller *scic = idev->owning_port->owning_controller; 922 struct isci_host *ihost = idev->owning_port->owning_controller;
925 923
926 isci_remote_device_not_ready(scic_to_ihost(scic), idev, 924 isci_remote_device_not_ready(ihost, idev,
927 SCIC_REMOTE_DEVICE_NOT_READY_STOP_REQUESTED); 925 SCIC_REMOTE_DEVICE_NOT_READY_STOP_REQUESTED);
928 } 926 }
929} 927}
@@ -963,40 +961,40 @@ static void scic_sds_stp_remote_device_ready_idle_substate_enter(struct sci_base
963static void scic_sds_stp_remote_device_ready_cmd_substate_enter(struct sci_base_state_machine *sm) 961static void scic_sds_stp_remote_device_ready_cmd_substate_enter(struct sci_base_state_machine *sm)
964{ 962{
965 struct isci_remote_device *idev = container_of(sm, typeof(*idev), sm); 963 struct isci_remote_device *idev = container_of(sm, typeof(*idev), sm);
966 struct scic_sds_controller *scic = scic_sds_remote_device_get_controller(idev); 964 struct isci_host *ihost = scic_sds_remote_device_get_controller(idev);
967 965
968 BUG_ON(idev->working_request == NULL); 966 BUG_ON(idev->working_request == NULL);
969 967
970 isci_remote_device_not_ready(scic_to_ihost(scic), idev, 968 isci_remote_device_not_ready(ihost, idev,
971 SCIC_REMOTE_DEVICE_NOT_READY_SATA_REQUEST_STARTED); 969 SCIC_REMOTE_DEVICE_NOT_READY_SATA_REQUEST_STARTED);
972} 970}
973 971
974static void scic_sds_stp_remote_device_ready_ncq_error_substate_enter(struct sci_base_state_machine *sm) 972static void scic_sds_stp_remote_device_ready_ncq_error_substate_enter(struct sci_base_state_machine *sm)
975{ 973{
976 struct isci_remote_device *idev = container_of(sm, typeof(*idev), sm); 974 struct isci_remote_device *idev = container_of(sm, typeof(*idev), sm);
977 struct scic_sds_controller *scic = scic_sds_remote_device_get_controller(idev); 975 struct isci_host *ihost = scic_sds_remote_device_get_controller(idev);
978 976
979 if (idev->not_ready_reason == SCIC_REMOTE_DEVICE_NOT_READY_SATA_SDB_ERROR_FIS_RECEIVED) 977 if (idev->not_ready_reason == SCIC_REMOTE_DEVICE_NOT_READY_SATA_SDB_ERROR_FIS_RECEIVED)
980 isci_remote_device_not_ready(scic_to_ihost(scic), idev, 978 isci_remote_device_not_ready(ihost, idev,
981 idev->not_ready_reason); 979 idev->not_ready_reason);
982} 980}
983 981
984static void scic_sds_smp_remote_device_ready_idle_substate_enter(struct sci_base_state_machine *sm) 982static void scic_sds_smp_remote_device_ready_idle_substate_enter(struct sci_base_state_machine *sm)
985{ 983{
986 struct isci_remote_device *idev = container_of(sm, typeof(*idev), sm); 984 struct isci_remote_device *idev = container_of(sm, typeof(*idev), sm);
987 struct scic_sds_controller *scic = scic_sds_remote_device_get_controller(idev); 985 struct isci_host *ihost = scic_sds_remote_device_get_controller(idev);
988 986
989 isci_remote_device_ready(scic_to_ihost(scic), idev); 987 isci_remote_device_ready(ihost, idev);
990} 988}
991 989
992static void scic_sds_smp_remote_device_ready_cmd_substate_enter(struct sci_base_state_machine *sm) 990static void scic_sds_smp_remote_device_ready_cmd_substate_enter(struct sci_base_state_machine *sm)
993{ 991{
994 struct isci_remote_device *idev = container_of(sm, typeof(*idev), sm); 992 struct isci_remote_device *idev = container_of(sm, typeof(*idev), sm);
995 struct scic_sds_controller *scic = scic_sds_remote_device_get_controller(idev); 993 struct isci_host *ihost = scic_sds_remote_device_get_controller(idev);
996 994
997 BUG_ON(idev->working_request == NULL); 995 BUG_ON(idev->working_request == NULL);
998 996
999 isci_remote_device_not_ready(scic_to_ihost(scic), idev, 997 isci_remote_device_not_ready(ihost, idev,
1000 SCIC_REMOTE_DEVICE_NOT_READY_SMP_REQUEST_STARTED); 998 SCIC_REMOTE_DEVICE_NOT_READY_SMP_REQUEST_STARTED);
1001} 999}
1002 1000
@@ -1303,7 +1301,7 @@ void isci_remote_device_release(struct kref *kref)
1303 * @isci_host: This parameter specifies the isci host object. 1301 * @isci_host: This parameter specifies the isci host object.
1304 * @isci_device: This parameter specifies the remote device. 1302 * @isci_device: This parameter specifies the remote device.
1305 * 1303 *
1306 * The status of the scic request to stop. 1304 * The status of the ihost request to stop.
1307 */ 1305 */
1308enum sci_status isci_remote_device_stop(struct isci_host *ihost, struct isci_remote_device *idev) 1306enum sci_status isci_remote_device_stop(struct isci_host *ihost, struct isci_remote_device *idev)
1309{ 1307{
diff --git a/drivers/scsi/isci/remote_device.h b/drivers/scsi/isci/remote_device.h
index 45798582fc14..bc4da20a13fa 100644
--- a/drivers/scsi/isci/remote_device.h
+++ b/drivers/scsi/isci/remote_device.h
@@ -402,17 +402,17 @@ enum sci_status scic_sds_remote_device_event_handler(
402 u32 event_code); 402 u32 event_code);
403 403
404enum sci_status scic_sds_remote_device_start_io( 404enum sci_status scic_sds_remote_device_start_io(
405 struct scic_sds_controller *controller, 405 struct isci_host *ihost,
406 struct isci_remote_device *idev, 406 struct isci_remote_device *idev,
407 struct isci_request *ireq); 407 struct isci_request *ireq);
408 408
409enum sci_status scic_sds_remote_device_start_task( 409enum sci_status scic_sds_remote_device_start_task(
410 struct scic_sds_controller *controller, 410 struct isci_host *ihost,
411 struct isci_remote_device *idev, 411 struct isci_remote_device *idev,
412 struct isci_request *ireq); 412 struct isci_request *ireq);
413 413
414enum sci_status scic_sds_remote_device_complete_io( 414enum sci_status scic_sds_remote_device_complete_io(
415 struct scic_sds_controller *controller, 415 struct isci_host *ihost,
416 struct isci_remote_device *idev, 416 struct isci_remote_device *idev,
417 struct isci_request *ireq); 417 struct isci_request *ireq);
418 418
diff --git a/drivers/scsi/isci/remote_node_context.c b/drivers/scsi/isci/remote_node_context.c
index e485744e1263..8a5203b6eb09 100644
--- a/drivers/scsi/isci/remote_node_context.c
+++ b/drivers/scsi/isci/remote_node_context.c
@@ -107,11 +107,11 @@ static void scic_sds_remote_node_context_construct_buffer(
107 struct domain_device *dev = idev->domain_dev; 107 struct domain_device *dev = idev->domain_dev;
108 int rni = sci_rnc->remote_node_index; 108 int rni = sci_rnc->remote_node_index;
109 union scu_remote_node_context *rnc; 109 union scu_remote_node_context *rnc;
110 struct scic_sds_controller *scic; 110 struct isci_host *ihost;
111 __le64 sas_addr; 111 __le64 sas_addr;
112 112
113 scic = scic_sds_remote_device_get_controller(idev); 113 ihost = scic_sds_remote_device_get_controller(idev);
114 rnc = scic_sds_controller_get_remote_node_context_buffer(scic, rni); 114 rnc = scic_sds_controller_get_remote_node_context_buffer(ihost, rni);
115 115
116 memset(rnc, 0, sizeof(union scu_remote_node_context) 116 memset(rnc, 0, sizeof(union scu_remote_node_context)
117 * scic_sds_remote_device_node_count(idev)); 117 * scic_sds_remote_device_node_count(idev));
@@ -135,14 +135,14 @@ static void scic_sds_remote_node_context_construct_buffer(
135 135
136 if (dev->dev_type == SATA_DEV || (dev->tproto & SAS_PROTOCOL_STP)) { 136 if (dev->dev_type == SATA_DEV || (dev->tproto & SAS_PROTOCOL_STP)) {
137 rnc->ssp.connection_occupancy_timeout = 137 rnc->ssp.connection_occupancy_timeout =
138 scic->user_parameters.sds1.stp_max_occupancy_timeout; 138 ihost->user_parameters.sds1.stp_max_occupancy_timeout;
139 rnc->ssp.connection_inactivity_timeout = 139 rnc->ssp.connection_inactivity_timeout =
140 scic->user_parameters.sds1.stp_inactivity_timeout; 140 ihost->user_parameters.sds1.stp_inactivity_timeout;
141 } else { 141 } else {
142 rnc->ssp.connection_occupancy_timeout = 142 rnc->ssp.connection_occupancy_timeout =
143 scic->user_parameters.sds1.ssp_max_occupancy_timeout; 143 ihost->user_parameters.sds1.ssp_max_occupancy_timeout;
144 rnc->ssp.connection_inactivity_timeout = 144 rnc->ssp.connection_inactivity_timeout =
145 scic->user_parameters.sds1.ssp_inactivity_timeout; 145 ihost->user_parameters.sds1.ssp_inactivity_timeout;
146 } 146 }
147 147
148 rnc->ssp.initial_arbitration_wait_time = 0; 148 rnc->ssp.initial_arbitration_wait_time = 0;
diff --git a/drivers/scsi/isci/request.c b/drivers/scsi/isci/request.c
index 90ead662828d..36e674896bc5 100644
--- a/drivers/scsi/isci/request.c
+++ b/drivers/scsi/isci/request.c
@@ -74,19 +74,19 @@ static struct scu_sgl_element_pair *to_sgl_element_pair(struct isci_request *ire
74 return &ireq->sg_table[idx - 2]; 74 return &ireq->sg_table[idx - 2];
75} 75}
76 76
77static dma_addr_t to_sgl_element_pair_dma(struct scic_sds_controller *scic, 77static dma_addr_t to_sgl_element_pair_dma(struct isci_host *ihost,
78 struct isci_request *ireq, u32 idx) 78 struct isci_request *ireq, u32 idx)
79{ 79{
80 u32 offset; 80 u32 offset;
81 81
82 if (idx == 0) { 82 if (idx == 0) {
83 offset = (void *) &ireq->tc->sgl_pair_ab - 83 offset = (void *) &ireq->tc->sgl_pair_ab -
84 (void *) &scic->task_context_table[0]; 84 (void *) &ihost->task_context_table[0];
85 return scic->task_context_dma + offset; 85 return ihost->task_context_dma + offset;
86 } else if (idx == 1) { 86 } else if (idx == 1) {
87 offset = (void *) &ireq->tc->sgl_pair_cd - 87 offset = (void *) &ireq->tc->sgl_pair_cd -
88 (void *) &scic->task_context_table[0]; 88 (void *) &ihost->task_context_table[0];
89 return scic->task_context_dma + offset; 89 return ihost->task_context_dma + offset;
90 } 90 }
91 91
92 return scic_io_request_get_dma_addr(ireq, &ireq->sg_table[idx - 2]); 92 return scic_io_request_get_dma_addr(ireq, &ireq->sg_table[idx - 2]);
@@ -102,8 +102,7 @@ static void init_sgl_element(struct scu_sgl_element *e, struct scatterlist *sg)
102 102
103static void scic_sds_request_build_sgl(struct isci_request *ireq) 103static void scic_sds_request_build_sgl(struct isci_request *ireq)
104{ 104{
105 struct isci_host *isci_host = ireq->isci_host; 105 struct isci_host *ihost = ireq->isci_host;
106 struct scic_sds_controller *scic = &isci_host->sci;
107 struct sas_task *task = isci_request_access_task(ireq); 106 struct sas_task *task = isci_request_access_task(ireq);
108 struct scatterlist *sg = NULL; 107 struct scatterlist *sg = NULL;
109 dma_addr_t dma_addr; 108 dma_addr_t dma_addr;
@@ -125,7 +124,7 @@ static void scic_sds_request_build_sgl(struct isci_request *ireq)
125 memset(&scu_sg->B, 0, sizeof(scu_sg->B)); 124 memset(&scu_sg->B, 0, sizeof(scu_sg->B));
126 125
127 if (prev_sg) { 126 if (prev_sg) {
128 dma_addr = to_sgl_element_pair_dma(scic, 127 dma_addr = to_sgl_element_pair_dma(ihost,
129 ireq, 128 ireq,
130 sg_idx); 129 sg_idx);
131 130
@@ -141,7 +140,7 @@ static void scic_sds_request_build_sgl(struct isci_request *ireq)
141 } else { /* handle when no sg */ 140 } else { /* handle when no sg */
142 scu_sg = to_sgl_element_pair(ireq, sg_idx); 141 scu_sg = to_sgl_element_pair(ireq, sg_idx);
143 142
144 dma_addr = dma_map_single(&isci_host->pdev->dev, 143 dma_addr = dma_map_single(&ihost->pdev->dev,
145 task->scatter, 144 task->scatter,
146 task->total_xfer_len, 145 task->total_xfer_len,
147 task->data_dir); 146 task->data_dir);
@@ -508,7 +507,7 @@ scic_io_request_construct_sata(struct isci_request *ireq,
508 scu_stp_raw_request_construct_task_context(ireq); 507 scu_stp_raw_request_construct_task_context(ireq);
509 return SCI_SUCCESS; 508 return SCI_SUCCESS;
510 } else { 509 } else {
511 dev_err(scic_to_dev(ireq->owning_controller), 510 dev_err(&ireq->owning_controller->pdev->dev,
512 "%s: Request 0x%p received un-handled SAT " 511 "%s: Request 0x%p received un-handled SAT "
513 "management protocol 0x%x.\n", 512 "management protocol 0x%x.\n",
514 __func__, ireq, tmf->tmf_code); 513 __func__, ireq, tmf->tmf_code);
@@ -518,7 +517,7 @@ scic_io_request_construct_sata(struct isci_request *ireq,
518 } 517 }
519 518
520 if (!sas_protocol_ata(task->task_proto)) { 519 if (!sas_protocol_ata(task->task_proto)) {
521 dev_err(scic_to_dev(ireq->owning_controller), 520 dev_err(&ireq->owning_controller->pdev->dev,
522 "%s: Non-ATA protocol in SATA path: 0x%x\n", 521 "%s: Non-ATA protocol in SATA path: 0x%x\n",
523 __func__, 522 __func__,
524 task->task_proto); 523 task->task_proto);
@@ -616,7 +615,7 @@ enum sci_status scic_task_request_construct_sata(struct isci_request *ireq)
616 tmf->tmf_code == isci_tmf_sata_srst_low) { 615 tmf->tmf_code == isci_tmf_sata_srst_low) {
617 scu_stp_raw_request_construct_task_context(ireq); 616 scu_stp_raw_request_construct_task_context(ireq);
618 } else { 617 } else {
619 dev_err(scic_to_dev(ireq->owning_controller), 618 dev_err(&ireq->owning_controller->pdev->dev,
620 "%s: Request 0x%p received un-handled SAT " 619 "%s: Request 0x%p received un-handled SAT "
621 "Protocol 0x%x.\n", 620 "Protocol 0x%x.\n",
622 __func__, ireq, tmf->tmf_code); 621 __func__, ireq, tmf->tmf_code);
@@ -639,11 +638,11 @@ enum sci_status scic_task_request_construct_sata(struct isci_request *ireq)
639#define SCU_TASK_CONTEXT_SRAM 0x200000 638#define SCU_TASK_CONTEXT_SRAM 0x200000
640static u32 sci_req_tx_bytes(struct isci_request *ireq) 639static u32 sci_req_tx_bytes(struct isci_request *ireq)
641{ 640{
642 struct scic_sds_controller *scic = ireq->owning_controller; 641 struct isci_host *ihost = ireq->owning_controller;
643 u32 ret_val = 0; 642 u32 ret_val = 0;
644 643
645 if (readl(&scic->smu_registers->address_modifier) == 0) { 644 if (readl(&ihost->smu_registers->address_modifier) == 0) {
646 void __iomem *scu_reg_base = scic->scu_registers; 645 void __iomem *scu_reg_base = ihost->scu_registers;
647 646
648 /* get the bytes of data from the Address == BAR1 + 20002Ch + (256*TCi) where 647 /* get the bytes of data from the Address == BAR1 + 20002Ch + (256*TCi) where
649 * BAR1 is the scu_registers 648 * BAR1 is the scu_registers
@@ -663,11 +662,11 @@ enum sci_status scic_sds_request_start(struct isci_request *ireq)
663{ 662{
664 enum sci_base_request_states state; 663 enum sci_base_request_states state;
665 struct scu_task_context *tc = ireq->tc; 664 struct scu_task_context *tc = ireq->tc;
666 struct scic_sds_controller *scic = ireq->owning_controller; 665 struct isci_host *ihost = ireq->owning_controller;
667 666
668 state = ireq->sm.current_state_id; 667 state = ireq->sm.current_state_id;
669 if (state != SCI_REQ_CONSTRUCTED) { 668 if (state != SCI_REQ_CONSTRUCTED) {
670 dev_warn(scic_to_dev(scic), 669 dev_warn(&ihost->pdev->dev,
671 "%s: SCIC IO Request requested to start while in wrong " 670 "%s: SCIC IO Request requested to start while in wrong "
672 "state %d\n", __func__, state); 671 "state %d\n", __func__, state);
673 return SCI_FAILURE_INVALID_STATE; 672 return SCI_FAILURE_INVALID_STATE;
@@ -749,7 +748,7 @@ scic_sds_io_request_terminate(struct isci_request *ireq)
749 return SCI_SUCCESS; 748 return SCI_SUCCESS;
750 case SCI_REQ_COMPLETED: 749 case SCI_REQ_COMPLETED:
751 default: 750 default:
752 dev_warn(scic_to_dev(ireq->owning_controller), 751 dev_warn(&ireq->owning_controller->pdev->dev,
753 "%s: SCIC IO Request requested to abort while in wrong " 752 "%s: SCIC IO Request requested to abort while in wrong "
754 "state %d\n", 753 "state %d\n",
755 __func__, 754 __func__,
@@ -763,7 +762,7 @@ scic_sds_io_request_terminate(struct isci_request *ireq)
763enum sci_status scic_sds_request_complete(struct isci_request *ireq) 762enum sci_status scic_sds_request_complete(struct isci_request *ireq)
764{ 763{
765 enum sci_base_request_states state; 764 enum sci_base_request_states state;
766 struct scic_sds_controller *scic = ireq->owning_controller; 765 struct isci_host *ihost = ireq->owning_controller;
767 766
768 state = ireq->sm.current_state_id; 767 state = ireq->sm.current_state_id;
769 if (WARN_ONCE(state != SCI_REQ_COMPLETED, 768 if (WARN_ONCE(state != SCI_REQ_COMPLETED,
@@ -771,7 +770,7 @@ enum sci_status scic_sds_request_complete(struct isci_request *ireq)
771 return SCI_FAILURE_INVALID_STATE; 770 return SCI_FAILURE_INVALID_STATE;
772 771
773 if (ireq->saved_rx_frame_index != SCU_INVALID_FRAME_INDEX) 772 if (ireq->saved_rx_frame_index != SCU_INVALID_FRAME_INDEX)
774 scic_sds_controller_release_frame(scic, 773 scic_sds_controller_release_frame(ihost,
775 ireq->saved_rx_frame_index); 774 ireq->saved_rx_frame_index);
776 775
777 /* XXX can we just stop the machine and remove the 'final' state? */ 776 /* XXX can we just stop the machine and remove the 'final' state? */
@@ -783,12 +782,12 @@ enum sci_status scic_sds_io_request_event_handler(struct isci_request *ireq,
783 u32 event_code) 782 u32 event_code)
784{ 783{
785 enum sci_base_request_states state; 784 enum sci_base_request_states state;
786 struct scic_sds_controller *scic = ireq->owning_controller; 785 struct isci_host *ihost = ireq->owning_controller;
787 786
788 state = ireq->sm.current_state_id; 787 state = ireq->sm.current_state_id;
789 788
790 if (state != SCI_REQ_STP_PIO_DATA_IN) { 789 if (state != SCI_REQ_STP_PIO_DATA_IN) {
791 dev_warn(scic_to_dev(scic), "%s: (%x) in wrong state %d\n", 790 dev_warn(&ihost->pdev->dev, "%s: (%x) in wrong state %d\n",
792 __func__, event_code, state); 791 __func__, event_code, state);
793 792
794 return SCI_FAILURE_INVALID_STATE; 793 return SCI_FAILURE_INVALID_STATE;
@@ -802,7 +801,7 @@ enum sci_status scic_sds_io_request_event_handler(struct isci_request *ireq,
802 sci_change_state(&ireq->sm, SCI_REQ_STP_PIO_WAIT_FRAME); 801 sci_change_state(&ireq->sm, SCI_REQ_STP_PIO_WAIT_FRAME);
803 return SCI_SUCCESS; 802 return SCI_SUCCESS;
804 default: 803 default:
805 dev_err(scic_to_dev(scic), 804 dev_err(&ihost->pdev->dev,
806 "%s: pio request unexpected event %#x\n", 805 "%s: pio request unexpected event %#x\n",
807 __func__, event_code); 806 __func__, event_code);
808 807
@@ -1024,7 +1023,7 @@ static enum sci_status ssp_task_request_await_tc_event(struct isci_request *ireq
1024 * There is a potential for receiving multiple task responses if 1023 * There is a potential for receiving multiple task responses if
1025 * we decide to send the task IU again. 1024 * we decide to send the task IU again.
1026 */ 1025 */
1027 dev_warn(scic_to_dev(ireq->owning_controller), 1026 dev_warn(&ireq->owning_controller->pdev->dev,
1028 "%s: TaskRequest:0x%p CompletionCode:%x - " 1027 "%s: TaskRequest:0x%p CompletionCode:%x - "
1029 "ACK/NAK timeout\n", __func__, ireq, 1028 "ACK/NAK timeout\n", __func__, ireq,
1030 completion_code); 1029 completion_code);
@@ -1073,7 +1072,7 @@ smp_request_await_response_tc_event(struct isci_request *ireq,
1073 * response within 2 ms. This causes our hardware break 1072 * response within 2 ms. This causes our hardware break
1074 * the connection and set TC completion with one of 1073 * the connection and set TC completion with one of
1075 * these SMP_XXX_XX_ERR status. For these type of error, 1074 * these SMP_XXX_XX_ERR status. For these type of error,
1076 * we ask scic user to retry the request. 1075 * we ask ihost user to retry the request.
1077 */ 1076 */
1078 scic_sds_request_set_status(ireq, SCU_TASK_DONE_SMP_RESP_TO_ERR, 1077 scic_sds_request_set_status(ireq, SCU_TASK_DONE_SMP_RESP_TO_ERR,
1079 SCI_FAILURE_RETRY_REQUIRED); 1078 SCI_FAILURE_RETRY_REQUIRED);
@@ -1451,18 +1450,18 @@ static void scic_sds_stp_request_udma_complete_request(
1451static enum sci_status scic_sds_stp_request_udma_general_frame_handler(struct isci_request *ireq, 1450static enum sci_status scic_sds_stp_request_udma_general_frame_handler(struct isci_request *ireq,
1452 u32 frame_index) 1451 u32 frame_index)
1453{ 1452{
1454 struct scic_sds_controller *scic = ireq->owning_controller; 1453 struct isci_host *ihost = ireq->owning_controller;
1455 struct dev_to_host_fis *frame_header; 1454 struct dev_to_host_fis *frame_header;
1456 enum sci_status status; 1455 enum sci_status status;
1457 u32 *frame_buffer; 1456 u32 *frame_buffer;
1458 1457
1459 status = scic_sds_unsolicited_frame_control_get_header(&scic->uf_control, 1458 status = scic_sds_unsolicited_frame_control_get_header(&ihost->uf_control,
1460 frame_index, 1459 frame_index,
1461 (void **)&frame_header); 1460 (void **)&frame_header);
1462 1461
1463 if ((status == SCI_SUCCESS) && 1462 if ((status == SCI_SUCCESS) &&
1464 (frame_header->fis_type == FIS_REGD2H)) { 1463 (frame_header->fis_type == FIS_REGD2H)) {
1465 scic_sds_unsolicited_frame_control_get_buffer(&scic->uf_control, 1464 scic_sds_unsolicited_frame_control_get_buffer(&ihost->uf_control,
1466 frame_index, 1465 frame_index,
1467 (void **)&frame_buffer); 1466 (void **)&frame_buffer);
1468 1467
@@ -1471,7 +1470,7 @@ static enum sci_status scic_sds_stp_request_udma_general_frame_handler(struct is
1471 frame_buffer); 1470 frame_buffer);
1472 } 1471 }
1473 1472
1474 scic_sds_controller_release_frame(scic, frame_index); 1473 scic_sds_controller_release_frame(ihost, frame_index);
1475 1474
1476 return status; 1475 return status;
1477} 1476}
@@ -1480,7 +1479,7 @@ enum sci_status
1480scic_sds_io_request_frame_handler(struct isci_request *ireq, 1479scic_sds_io_request_frame_handler(struct isci_request *ireq,
1481 u32 frame_index) 1480 u32 frame_index)
1482{ 1481{
1483 struct scic_sds_controller *scic = ireq->owning_controller; 1482 struct isci_host *ihost = ireq->owning_controller;
1484 struct isci_stp_request *stp_req = &ireq->stp.req; 1483 struct isci_stp_request *stp_req = &ireq->stp.req;
1485 enum sci_base_request_states state; 1484 enum sci_base_request_states state;
1486 enum sci_status status; 1485 enum sci_status status;
@@ -1492,7 +1491,7 @@ scic_sds_io_request_frame_handler(struct isci_request *ireq,
1492 struct ssp_frame_hdr ssp_hdr; 1491 struct ssp_frame_hdr ssp_hdr;
1493 void *frame_header; 1492 void *frame_header;
1494 1493
1495 scic_sds_unsolicited_frame_control_get_header(&scic->uf_control, 1494 scic_sds_unsolicited_frame_control_get_header(&ihost->uf_control,
1496 frame_index, 1495 frame_index,
1497 &frame_header); 1496 &frame_header);
1498 1497
@@ -1503,7 +1502,7 @@ scic_sds_io_request_frame_handler(struct isci_request *ireq,
1503 struct ssp_response_iu *resp_iu; 1502 struct ssp_response_iu *resp_iu;
1504 ssize_t word_cnt = SSP_RESP_IU_MAX_SIZE / sizeof(u32); 1503 ssize_t word_cnt = SSP_RESP_IU_MAX_SIZE / sizeof(u32);
1505 1504
1506 scic_sds_unsolicited_frame_control_get_buffer(&scic->uf_control, 1505 scic_sds_unsolicited_frame_control_get_buffer(&ihost->uf_control,
1507 frame_index, 1506 frame_index,
1508 (void **)&resp_iu); 1507 (void **)&resp_iu);
1509 1508
@@ -1522,7 +1521,7 @@ scic_sds_io_request_frame_handler(struct isci_request *ireq,
1522 SCI_SUCCESS); 1521 SCI_SUCCESS);
1523 } else { 1522 } else {
1524 /* not a response frame, why did it get forwarded? */ 1523 /* not a response frame, why did it get forwarded? */
1525 dev_err(scic_to_dev(scic), 1524 dev_err(&ihost->pdev->dev,
1526 "%s: SCIC IO Request 0x%p received unexpected " 1525 "%s: SCIC IO Request 0x%p received unexpected "
1527 "frame %d type 0x%02x\n", __func__, ireq, 1526 "frame %d type 0x%02x\n", __func__, ireq,
1528 frame_index, ssp_hdr.frame_type); 1527 frame_index, ssp_hdr.frame_type);
@@ -1532,7 +1531,7 @@ scic_sds_io_request_frame_handler(struct isci_request *ireq,
1532 * In any case we are done with this frame buffer return it to 1531 * In any case we are done with this frame buffer return it to
1533 * the controller 1532 * the controller
1534 */ 1533 */
1535 scic_sds_controller_release_frame(scic, frame_index); 1534 scic_sds_controller_release_frame(ihost, frame_index);
1536 1535
1537 return SCI_SUCCESS; 1536 return SCI_SUCCESS;
1538 } 1537 }
@@ -1540,14 +1539,14 @@ scic_sds_io_request_frame_handler(struct isci_request *ireq,
1540 case SCI_REQ_TASK_WAIT_TC_RESP: 1539 case SCI_REQ_TASK_WAIT_TC_RESP:
1541 scic_sds_io_request_copy_response(ireq); 1540 scic_sds_io_request_copy_response(ireq);
1542 sci_change_state(&ireq->sm, SCI_REQ_COMPLETED); 1541 sci_change_state(&ireq->sm, SCI_REQ_COMPLETED);
1543 scic_sds_controller_release_frame(scic,frame_index); 1542 scic_sds_controller_release_frame(ihost,frame_index);
1544 return SCI_SUCCESS; 1543 return SCI_SUCCESS;
1545 1544
1546 case SCI_REQ_SMP_WAIT_RESP: { 1545 case SCI_REQ_SMP_WAIT_RESP: {
1547 struct smp_resp *rsp_hdr = &ireq->smp.rsp; 1546 struct smp_resp *rsp_hdr = &ireq->smp.rsp;
1548 void *frame_header; 1547 void *frame_header;
1549 1548
1550 scic_sds_unsolicited_frame_control_get_header(&scic->uf_control, 1549 scic_sds_unsolicited_frame_control_get_header(&ihost->uf_control,
1551 frame_index, 1550 frame_index,
1552 &frame_header); 1551 &frame_header);
1553 1552
@@ -1558,7 +1557,7 @@ scic_sds_io_request_frame_handler(struct isci_request *ireq,
1558 if (rsp_hdr->frame_type == SMP_RESPONSE) { 1557 if (rsp_hdr->frame_type == SMP_RESPONSE) {
1559 void *smp_resp; 1558 void *smp_resp;
1560 1559
1561 scic_sds_unsolicited_frame_control_get_buffer(&scic->uf_control, 1560 scic_sds_unsolicited_frame_control_get_buffer(&ihost->uf_control,
1562 frame_index, 1561 frame_index,
1563 &smp_resp); 1562 &smp_resp);
1564 1563
@@ -1577,7 +1576,7 @@ scic_sds_io_request_frame_handler(struct isci_request *ireq,
1577 * This was not a response frame why did it get 1576 * This was not a response frame why did it get
1578 * forwarded? 1577 * forwarded?
1579 */ 1578 */
1580 dev_err(scic_to_dev(scic), 1579 dev_err(&ihost->pdev->dev,
1581 "%s: SCIC SMP Request 0x%p received unexpected " 1580 "%s: SCIC SMP Request 0x%p received unexpected "
1582 "frame %d type 0x%02x\n", 1581 "frame %d type 0x%02x\n",
1583 __func__, 1582 __func__,
@@ -1592,7 +1591,7 @@ scic_sds_io_request_frame_handler(struct isci_request *ireq,
1592 sci_change_state(&ireq->sm, SCI_REQ_COMPLETED); 1591 sci_change_state(&ireq->sm, SCI_REQ_COMPLETED);
1593 } 1592 }
1594 1593
1595 scic_sds_controller_release_frame(scic, frame_index); 1594 scic_sds_controller_release_frame(ihost, frame_index);
1596 1595
1597 return SCI_SUCCESS; 1596 return SCI_SUCCESS;
1598 } 1597 }
@@ -1619,12 +1618,12 @@ scic_sds_io_request_frame_handler(struct isci_request *ireq,
1619 struct dev_to_host_fis *frame_header; 1618 struct dev_to_host_fis *frame_header;
1620 u32 *frame_buffer; 1619 u32 *frame_buffer;
1621 1620
1622 status = scic_sds_unsolicited_frame_control_get_header(&scic->uf_control, 1621 status = scic_sds_unsolicited_frame_control_get_header(&ihost->uf_control,
1623 frame_index, 1622 frame_index,
1624 (void **)&frame_header); 1623 (void **)&frame_header);
1625 1624
1626 if (status != SCI_SUCCESS) { 1625 if (status != SCI_SUCCESS) {
1627 dev_err(scic_to_dev(scic), 1626 dev_err(&ihost->pdev->dev,
1628 "%s: SCIC IO Request 0x%p could not get frame " 1627 "%s: SCIC IO Request 0x%p could not get frame "
1629 "header for frame index %d, status %x\n", 1628 "header for frame index %d, status %x\n",
1630 __func__, 1629 __func__,
@@ -1637,7 +1636,7 @@ scic_sds_io_request_frame_handler(struct isci_request *ireq,
1637 1636
1638 switch (frame_header->fis_type) { 1637 switch (frame_header->fis_type) {
1639 case FIS_REGD2H: 1638 case FIS_REGD2H:
1640 scic_sds_unsolicited_frame_control_get_buffer(&scic->uf_control, 1639 scic_sds_unsolicited_frame_control_get_buffer(&ihost->uf_control,
1641 frame_index, 1640 frame_index,
1642 (void **)&frame_buffer); 1641 (void **)&frame_buffer);
1643 1642
@@ -1651,7 +1650,7 @@ scic_sds_io_request_frame_handler(struct isci_request *ireq,
1651 break; 1650 break;
1652 1651
1653 default: 1652 default:
1654 dev_warn(scic_to_dev(scic), 1653 dev_warn(&ihost->pdev->dev,
1655 "%s: IO Request:0x%p Frame Id:%d protocol " 1654 "%s: IO Request:0x%p Frame Id:%d protocol "
1656 "violation occurred\n", __func__, stp_req, 1655 "violation occurred\n", __func__, stp_req,
1657 frame_index); 1656 frame_index);
@@ -1664,7 +1663,7 @@ scic_sds_io_request_frame_handler(struct isci_request *ireq,
1664 sci_change_state(&ireq->sm, SCI_REQ_COMPLETED); 1663 sci_change_state(&ireq->sm, SCI_REQ_COMPLETED);
1665 1664
1666 /* Frame has been decoded return it to the controller */ 1665 /* Frame has been decoded return it to the controller */
1667 scic_sds_controller_release_frame(scic, frame_index); 1666 scic_sds_controller_release_frame(ihost, frame_index);
1668 1667
1669 return status; 1668 return status;
1670 } 1669 }
@@ -1674,12 +1673,12 @@ scic_sds_io_request_frame_handler(struct isci_request *ireq,
1674 struct dev_to_host_fis *frame_header; 1673 struct dev_to_host_fis *frame_header;
1675 u32 *frame_buffer; 1674 u32 *frame_buffer;
1676 1675
1677 status = scic_sds_unsolicited_frame_control_get_header(&scic->uf_control, 1676 status = scic_sds_unsolicited_frame_control_get_header(&ihost->uf_control,
1678 frame_index, 1677 frame_index,
1679 (void **)&frame_header); 1678 (void **)&frame_header);
1680 1679
1681 if (status != SCI_SUCCESS) { 1680 if (status != SCI_SUCCESS) {
1682 dev_err(scic_to_dev(scic), 1681 dev_err(&ihost->pdev->dev,
1683 "%s: SCIC IO Request 0x%p could not get frame " 1682 "%s: SCIC IO Request 0x%p could not get frame "
1684 "header for frame index %d, status %x\n", 1683 "header for frame index %d, status %x\n",
1685 __func__, stp_req, frame_index, status); 1684 __func__, stp_req, frame_index, status);
@@ -1689,7 +1688,7 @@ scic_sds_io_request_frame_handler(struct isci_request *ireq,
1689 switch (frame_header->fis_type) { 1688 switch (frame_header->fis_type) {
1690 case FIS_PIO_SETUP: 1689 case FIS_PIO_SETUP:
1691 /* Get from the frame buffer the PIO Setup Data */ 1690 /* Get from the frame buffer the PIO Setup Data */
1692 scic_sds_unsolicited_frame_control_get_buffer(&scic->uf_control, 1691 scic_sds_unsolicited_frame_control_get_buffer(&ihost->uf_control,
1693 frame_index, 1692 frame_index,
1694 (void **)&frame_buffer); 1693 (void **)&frame_buffer);
1695 1694
@@ -1736,7 +1735,7 @@ scic_sds_io_request_frame_handler(struct isci_request *ireq,
1736 * FIS when it is still busy? Do nothing since 1735 * FIS when it is still busy? Do nothing since
1737 * we are still in the right state. 1736 * we are still in the right state.
1738 */ 1737 */
1739 dev_dbg(scic_to_dev(scic), 1738 dev_dbg(&ihost->pdev->dev,
1740 "%s: SCIC PIO Request 0x%p received " 1739 "%s: SCIC PIO Request 0x%p received "
1741 "D2H Register FIS with BSY status " 1740 "D2H Register FIS with BSY status "
1742 "0x%x\n", 1741 "0x%x\n",
@@ -1746,7 +1745,7 @@ scic_sds_io_request_frame_handler(struct isci_request *ireq,
1746 break; 1745 break;
1747 } 1746 }
1748 1747
1749 scic_sds_unsolicited_frame_control_get_buffer(&scic->uf_control, 1748 scic_sds_unsolicited_frame_control_get_buffer(&ihost->uf_control,
1750 frame_index, 1749 frame_index,
1751 (void **)&frame_buffer); 1750 (void **)&frame_buffer);
1752 1751
@@ -1767,7 +1766,7 @@ scic_sds_io_request_frame_handler(struct isci_request *ireq,
1767 } 1766 }
1768 1767
1769 /* Frame is decoded return it to the controller */ 1768 /* Frame is decoded return it to the controller */
1770 scic_sds_controller_release_frame(scic, frame_index); 1769 scic_sds_controller_release_frame(ihost, frame_index);
1771 1770
1772 return status; 1771 return status;
1773 } 1772 }
@@ -1776,12 +1775,12 @@ scic_sds_io_request_frame_handler(struct isci_request *ireq,
1776 struct dev_to_host_fis *frame_header; 1775 struct dev_to_host_fis *frame_header;
1777 struct sata_fis_data *frame_buffer; 1776 struct sata_fis_data *frame_buffer;
1778 1777
1779 status = scic_sds_unsolicited_frame_control_get_header(&scic->uf_control, 1778 status = scic_sds_unsolicited_frame_control_get_header(&ihost->uf_control,
1780 frame_index, 1779 frame_index,
1781 (void **)&frame_header); 1780 (void **)&frame_header);
1782 1781
1783 if (status != SCI_SUCCESS) { 1782 if (status != SCI_SUCCESS) {
1784 dev_err(scic_to_dev(scic), 1783 dev_err(&ihost->pdev->dev,
1785 "%s: SCIC IO Request 0x%p could not get frame " 1784 "%s: SCIC IO Request 0x%p could not get frame "
1786 "header for frame index %d, status %x\n", 1785 "header for frame index %d, status %x\n",
1787 __func__, 1786 __func__,
@@ -1792,7 +1791,7 @@ scic_sds_io_request_frame_handler(struct isci_request *ireq,
1792 } 1791 }
1793 1792
1794 if (frame_header->fis_type != FIS_DATA) { 1793 if (frame_header->fis_type != FIS_DATA) {
1795 dev_err(scic_to_dev(scic), 1794 dev_err(&ihost->pdev->dev,
1796 "%s: SCIC PIO Request 0x%p received frame %d " 1795 "%s: SCIC PIO Request 0x%p received frame %d "
1797 "with fis type 0x%02x when expecting a data " 1796 "with fis type 0x%02x when expecting a data "
1798 "fis.\n", 1797 "fis.\n",
@@ -1808,7 +1807,7 @@ scic_sds_io_request_frame_handler(struct isci_request *ireq,
1808 sci_change_state(&ireq->sm, SCI_REQ_COMPLETED); 1807 sci_change_state(&ireq->sm, SCI_REQ_COMPLETED);
1809 1808
1810 /* Frame is decoded return it to the controller */ 1809 /* Frame is decoded return it to the controller */
1811 scic_sds_controller_release_frame(scic, frame_index); 1810 scic_sds_controller_release_frame(ihost, frame_index);
1812 return status; 1811 return status;
1813 } 1812 }
1814 1813
@@ -1816,7 +1815,7 @@ scic_sds_io_request_frame_handler(struct isci_request *ireq,
1816 ireq->saved_rx_frame_index = frame_index; 1815 ireq->saved_rx_frame_index = frame_index;
1817 stp_req->pio_len = 0; 1816 stp_req->pio_len = 0;
1818 } else { 1817 } else {
1819 scic_sds_unsolicited_frame_control_get_buffer(&scic->uf_control, 1818 scic_sds_unsolicited_frame_control_get_buffer(&ihost->uf_control,
1820 frame_index, 1819 frame_index,
1821 (void **)&frame_buffer); 1820 (void **)&frame_buffer);
1822 1821
@@ -1824,7 +1823,7 @@ scic_sds_io_request_frame_handler(struct isci_request *ireq,
1824 (u8 *)frame_buffer); 1823 (u8 *)frame_buffer);
1825 1824
1826 /* Frame is decoded return it to the controller */ 1825 /* Frame is decoded return it to the controller */
1827 scic_sds_controller_release_frame(scic, frame_index); 1826 scic_sds_controller_release_frame(ihost, frame_index);
1828 } 1827 }
1829 1828
1830 /* Check for the end of the transfer, are there more 1829 /* Check for the end of the transfer, are there more
@@ -1849,11 +1848,11 @@ scic_sds_io_request_frame_handler(struct isci_request *ireq,
1849 struct dev_to_host_fis *frame_header; 1848 struct dev_to_host_fis *frame_header;
1850 u32 *frame_buffer; 1849 u32 *frame_buffer;
1851 1850
1852 status = scic_sds_unsolicited_frame_control_get_header(&scic->uf_control, 1851 status = scic_sds_unsolicited_frame_control_get_header(&ihost->uf_control,
1853 frame_index, 1852 frame_index,
1854 (void **)&frame_header); 1853 (void **)&frame_header);
1855 if (status != SCI_SUCCESS) { 1854 if (status != SCI_SUCCESS) {
1856 dev_err(scic_to_dev(scic), 1855 dev_err(&ihost->pdev->dev,
1857 "%s: SCIC IO Request 0x%p could not get frame " 1856 "%s: SCIC IO Request 0x%p could not get frame "
1858 "header for frame index %d, status %x\n", 1857 "header for frame index %d, status %x\n",
1859 __func__, 1858 __func__,
@@ -1865,7 +1864,7 @@ scic_sds_io_request_frame_handler(struct isci_request *ireq,
1865 1864
1866 switch (frame_header->fis_type) { 1865 switch (frame_header->fis_type) {
1867 case FIS_REGD2H: 1866 case FIS_REGD2H:
1868 scic_sds_unsolicited_frame_control_get_buffer(&scic->uf_control, 1867 scic_sds_unsolicited_frame_control_get_buffer(&ihost->uf_control,
1869 frame_index, 1868 frame_index,
1870 (void **)&frame_buffer); 1869 (void **)&frame_buffer);
1871 1870
@@ -1880,7 +1879,7 @@ scic_sds_io_request_frame_handler(struct isci_request *ireq,
1880 break; 1879 break;
1881 1880
1882 default: 1881 default:
1883 dev_warn(scic_to_dev(scic), 1882 dev_warn(&ihost->pdev->dev,
1884 "%s: IO Request:0x%p Frame Id:%d protocol " 1883 "%s: IO Request:0x%p Frame Id:%d protocol "
1885 "violation occurred\n", 1884 "violation occurred\n",
1886 __func__, 1885 __func__,
@@ -1896,7 +1895,7 @@ scic_sds_io_request_frame_handler(struct isci_request *ireq,
1896 sci_change_state(&ireq->sm, SCI_REQ_COMPLETED); 1895 sci_change_state(&ireq->sm, SCI_REQ_COMPLETED);
1897 1896
1898 /* Frame has been decoded return it to the controller */ 1897 /* Frame has been decoded return it to the controller */
1899 scic_sds_controller_release_frame(scic, frame_index); 1898 scic_sds_controller_release_frame(ihost, frame_index);
1900 1899
1901 return status; 1900 return status;
1902 } 1901 }
@@ -1905,18 +1904,18 @@ scic_sds_io_request_frame_handler(struct isci_request *ireq,
1905 * TODO: Is it even possible to get an unsolicited frame in the 1904 * TODO: Is it even possible to get an unsolicited frame in the
1906 * aborting state? 1905 * aborting state?
1907 */ 1906 */
1908 scic_sds_controller_release_frame(scic, frame_index); 1907 scic_sds_controller_release_frame(ihost, frame_index);
1909 return SCI_SUCCESS; 1908 return SCI_SUCCESS;
1910 1909
1911 default: 1910 default:
1912 dev_warn(scic_to_dev(scic), 1911 dev_warn(&ihost->pdev->dev,
1913 "%s: SCIC IO Request given unexpected frame %x while " 1912 "%s: SCIC IO Request given unexpected frame %x while "
1914 "in state %d\n", 1913 "in state %d\n",
1915 __func__, 1914 __func__,
1916 frame_index, 1915 frame_index,
1917 state); 1916 state);
1918 1917
1919 scic_sds_controller_release_frame(scic, frame_index); 1918 scic_sds_controller_release_frame(ihost, frame_index);
1920 return SCI_FAILURE_INVALID_STATE; 1919 return SCI_FAILURE_INVALID_STATE;
1921 } 1920 }
1922} 1921}
@@ -2042,7 +2041,7 @@ scic_sds_io_request_tc_completion(struct isci_request *ireq,
2042 u32 completion_code) 2041 u32 completion_code)
2043{ 2042{
2044 enum sci_base_request_states state; 2043 enum sci_base_request_states state;
2045 struct scic_sds_controller *scic = ireq->owning_controller; 2044 struct isci_host *ihost = ireq->owning_controller;
2046 2045
2047 state = ireq->sm.current_state_id; 2046 state = ireq->sm.current_state_id;
2048 2047
@@ -2089,7 +2088,7 @@ scic_sds_io_request_tc_completion(struct isci_request *ireq,
2089 completion_code); 2088 completion_code);
2090 2089
2091 default: 2090 default:
2092 dev_warn(scic_to_dev(scic), 2091 dev_warn(&ihost->pdev->dev,
2093 "%s: SCIC IO Request given task completion " 2092 "%s: SCIC IO Request given task completion "
2094 "notification %x while in wrong state %d\n", 2093 "notification %x while in wrong state %d\n",
2095 __func__, 2094 __func__,
@@ -2480,7 +2479,7 @@ static void isci_task_save_for_upper_layer_completion(
2480 } 2479 }
2481} 2480}
2482 2481
2483static void isci_request_io_request_complete(struct isci_host *isci_host, 2482static void isci_request_io_request_complete(struct isci_host *ihost,
2484 struct isci_request *request, 2483 struct isci_request *request,
2485 enum sci_io_status completion_status) 2484 enum sci_io_status completion_status)
2486{ 2485{
@@ -2495,7 +2494,7 @@ static void isci_request_io_request_complete(struct isci_host *isci_host,
2495 enum isci_completion_selection complete_to_host 2494 enum isci_completion_selection complete_to_host
2496 = isci_perform_normal_io_completion; 2495 = isci_perform_normal_io_completion;
2497 2496
2498 dev_dbg(&isci_host->pdev->dev, 2497 dev_dbg(&ihost->pdev->dev,
2499 "%s: request = %p, task = %p,\n" 2498 "%s: request = %p, task = %p,\n"
2500 "task->data_dir = %d completion_status = 0x%x\n", 2499 "task->data_dir = %d completion_status = 0x%x\n",
2501 __func__, 2500 __func__,
@@ -2616,7 +2615,7 @@ static void isci_request_io_request_complete(struct isci_host *isci_host,
2616 switch (completion_status) { 2615 switch (completion_status) {
2617 2616
2618 case SCI_IO_FAILURE_RESPONSE_VALID: 2617 case SCI_IO_FAILURE_RESPONSE_VALID:
2619 dev_dbg(&isci_host->pdev->dev, 2618 dev_dbg(&ihost->pdev->dev,
2620 "%s: SCI_IO_FAILURE_RESPONSE_VALID (%p/%p)\n", 2619 "%s: SCI_IO_FAILURE_RESPONSE_VALID (%p/%p)\n",
2621 __func__, 2620 __func__,
2622 request, 2621 request,
@@ -2631,17 +2630,17 @@ static void isci_request_io_request_complete(struct isci_host *isci_host,
2631 /* crack the iu response buffer. */ 2630 /* crack the iu response buffer. */
2632 resp_iu = &request->ssp.rsp; 2631 resp_iu = &request->ssp.rsp;
2633 isci_request_process_response_iu(task, resp_iu, 2632 isci_request_process_response_iu(task, resp_iu,
2634 &isci_host->pdev->dev); 2633 &ihost->pdev->dev);
2635 2634
2636 } else if (SAS_PROTOCOL_SMP == task->task_proto) { 2635 } else if (SAS_PROTOCOL_SMP == task->task_proto) {
2637 2636
2638 dev_err(&isci_host->pdev->dev, 2637 dev_err(&ihost->pdev->dev,
2639 "%s: SCI_IO_FAILURE_RESPONSE_VALID: " 2638 "%s: SCI_IO_FAILURE_RESPONSE_VALID: "
2640 "SAS_PROTOCOL_SMP protocol\n", 2639 "SAS_PROTOCOL_SMP protocol\n",
2641 __func__); 2640 __func__);
2642 2641
2643 } else 2642 } else
2644 dev_err(&isci_host->pdev->dev, 2643 dev_err(&ihost->pdev->dev,
2645 "%s: unknown protocol\n", __func__); 2644 "%s: unknown protocol\n", __func__);
2646 2645
2647 /* use the task status set in the task struct by the 2646 /* use the task status set in the task struct by the
@@ -2662,7 +2661,7 @@ static void isci_request_io_request_complete(struct isci_host *isci_host,
2662 if (task->task_proto == SAS_PROTOCOL_SMP) { 2661 if (task->task_proto == SAS_PROTOCOL_SMP) {
2663 void *rsp = &request->smp.rsp; 2662 void *rsp = &request->smp.rsp;
2664 2663
2665 dev_dbg(&isci_host->pdev->dev, 2664 dev_dbg(&ihost->pdev->dev,
2666 "%s: SMP protocol completion\n", 2665 "%s: SMP protocol completion\n",
2667 __func__); 2666 __func__);
2668 2667
@@ -2687,20 +2686,20 @@ static void isci_request_io_request_complete(struct isci_host *isci_host,
2687 if (task->task_status.residual != 0) 2686 if (task->task_status.residual != 0)
2688 status = SAS_DATA_UNDERRUN; 2687 status = SAS_DATA_UNDERRUN;
2689 2688
2690 dev_dbg(&isci_host->pdev->dev, 2689 dev_dbg(&ihost->pdev->dev,
2691 "%s: SCI_IO_SUCCESS_IO_DONE_EARLY %d\n", 2690 "%s: SCI_IO_SUCCESS_IO_DONE_EARLY %d\n",
2692 __func__, 2691 __func__,
2693 status); 2692 status);
2694 2693
2695 } else 2694 } else
2696 dev_dbg(&isci_host->pdev->dev, 2695 dev_dbg(&ihost->pdev->dev,
2697 "%s: SCI_IO_SUCCESS\n", 2696 "%s: SCI_IO_SUCCESS\n",
2698 __func__); 2697 __func__);
2699 2698
2700 break; 2699 break;
2701 2700
2702 case SCI_IO_FAILURE_TERMINATED: 2701 case SCI_IO_FAILURE_TERMINATED:
2703 dev_dbg(&isci_host->pdev->dev, 2702 dev_dbg(&ihost->pdev->dev,
2704 "%s: SCI_IO_FAILURE_TERMINATED (%p/%p)\n", 2703 "%s: SCI_IO_FAILURE_TERMINATED (%p/%p)\n",
2705 __func__, 2704 __func__,
2706 request, 2705 request,
@@ -2768,7 +2767,7 @@ static void isci_request_io_request_complete(struct isci_host *isci_host,
2768 2767
2769 default: 2768 default:
2770 /* Catch any otherwise unhandled error codes here. */ 2769 /* Catch any otherwise unhandled error codes here. */
2771 dev_warn(&isci_host->pdev->dev, 2770 dev_warn(&ihost->pdev->dev,
2772 "%s: invalid completion code: 0x%x - " 2771 "%s: invalid completion code: 0x%x - "
2773 "isci_request = %p\n", 2772 "isci_request = %p\n",
2774 __func__, completion_status, request); 2773 __func__, completion_status, request);
@@ -2802,11 +2801,11 @@ static void isci_request_io_request_complete(struct isci_host *isci_host,
2802 break; 2801 break;
2803 if (task->num_scatter == 0) 2802 if (task->num_scatter == 0)
2804 /* 0 indicates a single dma address */ 2803 /* 0 indicates a single dma address */
2805 dma_unmap_single(&isci_host->pdev->dev, 2804 dma_unmap_single(&ihost->pdev->dev,
2806 request->zero_scatter_daddr, 2805 request->zero_scatter_daddr,
2807 task->total_xfer_len, task->data_dir); 2806 task->total_xfer_len, task->data_dir);
2808 else /* unmap the sgl dma addresses */ 2807 else /* unmap the sgl dma addresses */
2809 dma_unmap_sg(&isci_host->pdev->dev, task->scatter, 2808 dma_unmap_sg(&ihost->pdev->dev, task->scatter,
2810 request->num_sg_entries, task->data_dir); 2809 request->num_sg_entries, task->data_dir);
2811 break; 2810 break;
2812 case SAS_PROTOCOL_SMP: { 2811 case SAS_PROTOCOL_SMP: {
@@ -2814,7 +2813,7 @@ static void isci_request_io_request_complete(struct isci_host *isci_host,
2814 struct smp_req *smp_req; 2813 struct smp_req *smp_req;
2815 void *kaddr; 2814 void *kaddr;
2816 2815
2817 dma_unmap_sg(&isci_host->pdev->dev, sg, 1, DMA_TO_DEVICE); 2816 dma_unmap_sg(&ihost->pdev->dev, sg, 1, DMA_TO_DEVICE);
2818 2817
2819 /* need to swab it back in case the command buffer is re-used */ 2818 /* need to swab it back in case the command buffer is re-used */
2820 kaddr = kmap_atomic(sg_page(sg), KM_IRQ0); 2819 kaddr = kmap_atomic(sg_page(sg), KM_IRQ0);
@@ -2828,14 +2827,12 @@ static void isci_request_io_request_complete(struct isci_host *isci_host,
2828 } 2827 }
2829 2828
2830 /* Put the completed request on the correct list */ 2829 /* Put the completed request on the correct list */
2831 isci_task_save_for_upper_layer_completion(isci_host, request, response, 2830 isci_task_save_for_upper_layer_completion(ihost, request, response,
2832 status, complete_to_host 2831 status, complete_to_host
2833 ); 2832 );
2834 2833
2835 /* complete the io request to the core. */ 2834 /* complete the io request to the core. */
2836 scic_controller_complete_io(&isci_host->sci, 2835 scic_controller_complete_io(ihost, request->target_device, request);
2837 request->target_device,
2838 request);
2839 isci_put_device(idev); 2836 isci_put_device(idev);
2840 2837
2841 /* set terminated handle so it cannot be completed or 2838 /* set terminated handle so it cannot be completed or
@@ -2885,8 +2882,7 @@ static void scic_sds_request_started_state_enter(struct sci_base_state_machine *
2885static void scic_sds_request_completed_state_enter(struct sci_base_state_machine *sm) 2882static void scic_sds_request_completed_state_enter(struct sci_base_state_machine *sm)
2886{ 2883{
2887 struct isci_request *ireq = container_of(sm, typeof(*ireq), sm); 2884 struct isci_request *ireq = container_of(sm, typeof(*ireq), sm);
2888 struct scic_sds_controller *scic = ireq->owning_controller; 2885 struct isci_host *ihost = ireq->owning_controller;
2889 struct isci_host *ihost = scic_to_ihost(scic);
2890 2886
2891 /* Tell the SCI_USER that the IO request is complete */ 2887 /* Tell the SCI_USER that the IO request is complete */
2892 if (!test_bit(IREQ_TMF, &ireq->flags)) 2888 if (!test_bit(IREQ_TMF, &ireq->flags))
@@ -2985,7 +2981,7 @@ static const struct sci_base_state scic_sds_request_state_table[] = {
2985}; 2981};
2986 2982
2987static void 2983static void
2988scic_sds_general_request_construct(struct scic_sds_controller *scic, 2984scic_sds_general_request_construct(struct isci_host *ihost,
2989 struct isci_remote_device *idev, 2985 struct isci_remote_device *idev,
2990 struct isci_request *ireq) 2986 struct isci_request *ireq)
2991{ 2987{
@@ -3001,7 +2997,7 @@ scic_sds_general_request_construct(struct scic_sds_controller *scic,
3001} 2997}
3002 2998
3003static enum sci_status 2999static enum sci_status
3004scic_io_request_construct(struct scic_sds_controller *scic, 3000scic_io_request_construct(struct isci_host *ihost,
3005 struct isci_remote_device *idev, 3001 struct isci_remote_device *idev,
3006 struct isci_request *ireq) 3002 struct isci_request *ireq)
3007{ 3003{
@@ -3009,7 +3005,7 @@ scic_io_request_construct(struct scic_sds_controller *scic,
3009 enum sci_status status = SCI_SUCCESS; 3005 enum sci_status status = SCI_SUCCESS;
3010 3006
3011 /* Build the common part of the request */ 3007 /* Build the common part of the request */
3012 scic_sds_general_request_construct(scic, idev, ireq); 3008 scic_sds_general_request_construct(ihost, idev, ireq);
3013 3009
3014 if (idev->rnc.remote_node_index == SCIC_SDS_REMOTE_NODE_CONTEXT_INVALID_INDEX) 3010 if (idev->rnc.remote_node_index == SCIC_SDS_REMOTE_NODE_CONTEXT_INVALID_INDEX)
3015 return SCI_FAILURE_INVALID_REMOTE_DEVICE; 3011 return SCI_FAILURE_INVALID_REMOTE_DEVICE;
@@ -3028,7 +3024,7 @@ scic_io_request_construct(struct scic_sds_controller *scic,
3028 return status; 3024 return status;
3029} 3025}
3030 3026
3031enum sci_status scic_task_request_construct(struct scic_sds_controller *scic, 3027enum sci_status scic_task_request_construct(struct isci_host *ihost,
3032 struct isci_remote_device *idev, 3028 struct isci_remote_device *idev,
3033 u16 io_tag, struct isci_request *ireq) 3029 u16 io_tag, struct isci_request *ireq)
3034{ 3030{
@@ -3036,7 +3032,7 @@ enum sci_status scic_task_request_construct(struct scic_sds_controller *scic,
3036 enum sci_status status = SCI_SUCCESS; 3032 enum sci_status status = SCI_SUCCESS;
3037 3033
3038 /* Build the common part of the request */ 3034 /* Build the common part of the request */
3039 scic_sds_general_request_construct(scic, idev, ireq); 3035 scic_sds_general_request_construct(ihost, idev, ireq);
3040 3036
3041 if (dev->dev_type == SAS_END_DEV || 3037 if (dev->dev_type == SAS_END_DEV ||
3042 dev->dev_type == SATA_DEV || (dev->tproto & SAS_PROTOCOL_STP)) { 3038 dev->dev_type == SATA_DEV || (dev->tproto & SAS_PROTOCOL_STP)) {
@@ -3156,7 +3152,7 @@ scic_io_request_construct_smp(struct device *dev,
3156 task_context->initiator_request = 1; 3152 task_context->initiator_request = 1;
3157 task_context->connection_rate = idev->connection_rate; 3153 task_context->connection_rate = idev->connection_rate;
3158 task_context->protocol_engine_index = 3154 task_context->protocol_engine_index =
3159 scic_sds_controller_get_protocol_engine_group(scic); 3155 scic_sds_controller_get_protocol_engine_group(ihost);
3160 task_context->logical_port_index = scic_sds_port_get_index(iport); 3156 task_context->logical_port_index = scic_sds_port_get_index(iport);
3161 task_context->protocol_type = SCU_TASK_CONTEXT_PROTOCOL_SMP; 3157 task_context->protocol_type = SCU_TASK_CONTEXT_PROTOCOL_SMP;
3162 task_context->abort = 0; 3158 task_context->abort = 0;
@@ -3199,7 +3195,7 @@ scic_io_request_construct_smp(struct device *dev,
3199 task_context->task_phase = 0; 3195 task_context->task_phase = 0;
3200 3196
3201 ireq->post_context = (SCU_CONTEXT_COMMAND_REQUEST_TYPE_POST_TC | 3197 ireq->post_context = (SCU_CONTEXT_COMMAND_REQUEST_TYPE_POST_TC |
3202 (scic_sds_controller_get_protocol_engine_group(scic) << 3198 (scic_sds_controller_get_protocol_engine_group(ihost) <<
3203 SCU_CONTEXT_COMMAND_PROTOCOL_ENGINE_GROUP_SHIFT) | 3199 SCU_CONTEXT_COMMAND_PROTOCOL_ENGINE_GROUP_SHIFT) |
3204 (scic_sds_port_get_index(iport) << 3200 (scic_sds_port_get_index(iport) <<
3205 SCU_CONTEXT_COMMAND_LOGICAL_PORT_SHIFT) | 3201 SCU_CONTEXT_COMMAND_LOGICAL_PORT_SHIFT) |
@@ -3245,7 +3241,7 @@ static enum sci_status isci_smp_request_build(struct isci_request *ireq)
3245 3241
3246/** 3242/**
3247 * isci_io_request_build() - This function builds the io request object. 3243 * isci_io_request_build() - This function builds the io request object.
3248 * @isci_host: This parameter specifies the ISCI host object 3244 * @ihost: This parameter specifies the ISCI host object
3249 * @request: This parameter points to the isci_request object allocated in the 3245 * @request: This parameter points to the isci_request object allocated in the
3250 * request construct function. 3246 * request construct function.
3251 * @sci_device: This parameter is the handle for the sci core's remote device 3247 * @sci_device: This parameter is the handle for the sci core's remote device
@@ -3253,14 +3249,14 @@ static enum sci_status isci_smp_request_build(struct isci_request *ireq)
3253 * 3249 *
3254 * SCI_SUCCESS on successfull completion, or specific failure code. 3250 * SCI_SUCCESS on successfull completion, or specific failure code.
3255 */ 3251 */
3256static enum sci_status isci_io_request_build(struct isci_host *isci_host, 3252static enum sci_status isci_io_request_build(struct isci_host *ihost,
3257 struct isci_request *request, 3253 struct isci_request *request,
3258 struct isci_remote_device *idev) 3254 struct isci_remote_device *idev)
3259{ 3255{
3260 enum sci_status status = SCI_SUCCESS; 3256 enum sci_status status = SCI_SUCCESS;
3261 struct sas_task *task = isci_request_access_task(request); 3257 struct sas_task *task = isci_request_access_task(request);
3262 3258
3263 dev_dbg(&isci_host->pdev->dev, 3259 dev_dbg(&ihost->pdev->dev,
3264 "%s: idev = 0x%p; request = %p, " 3260 "%s: idev = 0x%p; request = %p, "
3265 "num_scatter = %d\n", 3261 "num_scatter = %d\n",
3266 __func__, 3262 __func__,
@@ -3277,7 +3273,7 @@ static enum sci_status isci_io_request_build(struct isci_host *isci_host,
3277 !(SAS_PROTOCOL_SMP & task->task_proto)) { 3273 !(SAS_PROTOCOL_SMP & task->task_proto)) {
3278 3274
3279 request->num_sg_entries = dma_map_sg( 3275 request->num_sg_entries = dma_map_sg(
3280 &isci_host->pdev->dev, 3276 &ihost->pdev->dev,
3281 task->scatter, 3277 task->scatter,
3282 task->num_scatter, 3278 task->num_scatter,
3283 task->data_dir 3279 task->data_dir
@@ -3287,10 +3283,10 @@ static enum sci_status isci_io_request_build(struct isci_host *isci_host,
3287 return SCI_FAILURE_INSUFFICIENT_RESOURCES; 3283 return SCI_FAILURE_INSUFFICIENT_RESOURCES;
3288 } 3284 }
3289 3285
3290 status = scic_io_request_construct(&isci_host->sci, idev, request); 3286 status = scic_io_request_construct(ihost, idev, request);
3291 3287
3292 if (status != SCI_SUCCESS) { 3288 if (status != SCI_SUCCESS) {
3293 dev_warn(&isci_host->pdev->dev, 3289 dev_warn(&ihost->pdev->dev,
3294 "%s: failed request construct\n", 3290 "%s: failed request construct\n",
3295 __func__); 3291 __func__);
3296 return SCI_FAILURE; 3292 return SCI_FAILURE;
@@ -3309,7 +3305,7 @@ static enum sci_status isci_io_request_build(struct isci_host *isci_host,
3309 status = isci_request_stp_request_construct(request); 3305 status = isci_request_stp_request_construct(request);
3310 break; 3306 break;
3311 default: 3307 default:
3312 dev_warn(&isci_host->pdev->dev, 3308 dev_warn(&ihost->pdev->dev,
3313 "%s: unknown protocol\n", __func__); 3309 "%s: unknown protocol\n", __func__);
3314 return SCI_FAILURE; 3310 return SCI_FAILURE;
3315 } 3311 }
@@ -3392,7 +3388,7 @@ int isci_request_execute(struct isci_host *ihost, struct isci_remote_device *ide
3392 * request was built that way (ie. 3388 * request was built that way (ie.
3393 * ireq->is_task_management_request is false). 3389 * ireq->is_task_management_request is false).
3394 */ 3390 */
3395 status = scic_controller_start_task(&ihost->sci, 3391 status = scic_controller_start_task(ihost,
3396 idev, 3392 idev,
3397 ireq); 3393 ireq);
3398 } else { 3394 } else {
@@ -3400,7 +3396,7 @@ int isci_request_execute(struct isci_host *ihost, struct isci_remote_device *ide
3400 } 3396 }
3401 } else { 3397 } else {
3402 /* send the request, let the core assign the IO TAG. */ 3398 /* send the request, let the core assign the IO TAG. */
3403 status = scic_controller_start_io(&ihost->sci, idev, 3399 status = scic_controller_start_io(ihost, idev,
3404 ireq); 3400 ireq);
3405 } 3401 }
3406 3402
diff --git a/drivers/scsi/isci/request.h b/drivers/scsi/isci/request.h
index ca64ea207ac8..0cafcead7a01 100644
--- a/drivers/scsi/isci/request.h
+++ b/drivers/scsi/isci/request.h
@@ -145,7 +145,7 @@ struct isci_request {
145 */ 145 */
146 struct completion *io_request_completion; 146 struct completion *io_request_completion;
147 struct sci_base_state_machine sm; 147 struct sci_base_state_machine sm;
148 struct scic_sds_controller *owning_controller; 148 struct isci_host *owning_controller;
149 struct isci_remote_device *target_device; 149 struct isci_remote_device *target_device;
150 u16 io_tag; 150 u16 io_tag;
151 enum sci_request_protocol protocol; 151 enum sci_request_protocol protocol;
@@ -500,7 +500,7 @@ int isci_request_execute(struct isci_host *ihost, struct isci_remote_device *ide
500void isci_terminate_pending_requests(struct isci_host *ihost, 500void isci_terminate_pending_requests(struct isci_host *ihost,
501 struct isci_remote_device *idev); 501 struct isci_remote_device *idev);
502enum sci_status 502enum sci_status
503scic_task_request_construct(struct scic_sds_controller *scic, 503scic_task_request_construct(struct isci_host *ihost,
504 struct isci_remote_device *idev, 504 struct isci_remote_device *idev,
505 u16 io_tag, 505 u16 io_tag,
506 struct isci_request *ireq); 506 struct isci_request *ireq);
diff --git a/drivers/scsi/isci/task.c b/drivers/scsi/isci/task.c
index 89b01eef44b1..3a1fc55a7557 100644
--- a/drivers/scsi/isci/task.c
+++ b/drivers/scsi/isci/task.c
@@ -257,7 +257,7 @@ static struct isci_request *isci_task_request_build(struct isci_host *ihost,
257 return NULL; 257 return NULL;
258 258
259 /* let the core do it's construct. */ 259 /* let the core do it's construct. */
260 status = scic_task_request_construct(&ihost->sci, idev, tag, 260 status = scic_task_request_construct(ihost, idev, tag,
261 ireq); 261 ireq);
262 262
263 if (status != SCI_SUCCESS) { 263 if (status != SCI_SUCCESS) {
@@ -332,7 +332,7 @@ int isci_task_execute_tmf(struct isci_host *ihost,
332 spin_lock_irqsave(&ihost->scic_lock, flags); 332 spin_lock_irqsave(&ihost->scic_lock, flags);
333 333
334 /* start the TMF io. */ 334 /* start the TMF io. */
335 status = scic_controller_start_task(&ihost->sci, idev, ireq); 335 status = scic_controller_start_task(ihost, idev, ireq);
336 336
337 if (status != SCI_TASK_SUCCESS) { 337 if (status != SCI_TASK_SUCCESS) {
338 dev_warn(&ihost->pdev->dev, 338 dev_warn(&ihost->pdev->dev,
@@ -364,7 +364,7 @@ int isci_task_execute_tmf(struct isci_host *ihost,
364 if (tmf->cb_state_func != NULL) 364 if (tmf->cb_state_func != NULL)
365 tmf->cb_state_func(isci_tmf_timed_out, tmf, tmf->cb_data); 365 tmf->cb_state_func(isci_tmf_timed_out, tmf, tmf->cb_data);
366 366
367 scic_controller_terminate_request(&ihost->sci, 367 scic_controller_terminate_request(ihost,
368 idev, 368 idev,
369 ireq); 369 ireq);
370 370
@@ -514,15 +514,14 @@ static void isci_request_cleanup_completed_loiterer(
514 * request, and wait for it to complete. This function must only be called 514 * request, and wait for it to complete. This function must only be called
515 * from a thread that can wait. Note that the request is terminated and 515 * from a thread that can wait. Note that the request is terminated and
516 * completed (back to the host, if started there). 516 * completed (back to the host, if started there).
517 * @isci_host: This SCU. 517 * @ihost: This SCU.
518 * @idev: The target. 518 * @idev: The target.
519 * @isci_request: The I/O request to be terminated. 519 * @isci_request: The I/O request to be terminated.
520 * 520 *
521 */ 521 */
522static void isci_terminate_request_core( 522static void isci_terminate_request_core(struct isci_host *ihost,
523 struct isci_host *isci_host, 523 struct isci_remote_device *idev,
524 struct isci_remote_device *idev, 524 struct isci_request *isci_request)
525 struct isci_request *isci_request)
526{ 525{
527 enum sci_status status = SCI_SUCCESS; 526 enum sci_status status = SCI_SUCCESS;
528 bool was_terminated = false; 527 bool was_terminated = false;
@@ -533,11 +532,11 @@ static void isci_terminate_request_core(
533 struct completion *io_request_completion; 532 struct completion *io_request_completion;
534 struct sas_task *task; 533 struct sas_task *task;
535 534
536 dev_dbg(&isci_host->pdev->dev, 535 dev_dbg(&ihost->pdev->dev,
537 "%s: device = %p; request = %p\n", 536 "%s: device = %p; request = %p\n",
538 __func__, idev, isci_request); 537 __func__, idev, isci_request);
539 538
540 spin_lock_irqsave(&isci_host->scic_lock, flags); 539 spin_lock_irqsave(&ihost->scic_lock, flags);
541 540
542 io_request_completion = isci_request->io_request_completion; 541 io_request_completion = isci_request->io_request_completion;
543 542
@@ -557,12 +556,11 @@ static void isci_terminate_request_core(
557 if (!test_bit(IREQ_TERMINATED, &isci_request->flags)) { 556 if (!test_bit(IREQ_TERMINATED, &isci_request->flags)) {
558 was_terminated = true; 557 was_terminated = true;
559 needs_cleanup_handling = true; 558 needs_cleanup_handling = true;
560 status = scic_controller_terminate_request( 559 status = scic_controller_terminate_request(ihost,
561 &isci_host->sci, 560 idev,
562 idev, 561 isci_request);
563 isci_request);
564 } 562 }
565 spin_unlock_irqrestore(&isci_host->scic_lock, flags); 563 spin_unlock_irqrestore(&ihost->scic_lock, flags);
566 564
567 /* 565 /*
568 * The only time the request to terminate will 566 * The only time the request to terminate will
@@ -570,7 +568,7 @@ static void isci_terminate_request_core(
570 * being aborted. 568 * being aborted.
571 */ 569 */
572 if (status != SCI_SUCCESS) { 570 if (status != SCI_SUCCESS) {
573 dev_err(&isci_host->pdev->dev, 571 dev_err(&ihost->pdev->dev,
574 "%s: scic_controller_terminate_request" 572 "%s: scic_controller_terminate_request"
575 " returned = 0x%x\n", 573 " returned = 0x%x\n",
576 __func__, status); 574 __func__, status);
@@ -579,7 +577,7 @@ static void isci_terminate_request_core(
579 577
580 } else { 578 } else {
581 if (was_terminated) { 579 if (was_terminated) {
582 dev_dbg(&isci_host->pdev->dev, 580 dev_dbg(&ihost->pdev->dev,
583 "%s: before completion wait (%p/%p)\n", 581 "%s: before completion wait (%p/%p)\n",
584 __func__, isci_request, io_request_completion); 582 __func__, isci_request, io_request_completion);
585 583
@@ -593,7 +591,7 @@ static void isci_terminate_request_core(
593 if (!termination_completed) { 591 if (!termination_completed) {
594 592
595 /* The request to terminate has timed out. */ 593 /* The request to terminate has timed out. */
596 spin_lock_irqsave(&isci_host->scic_lock, 594 spin_lock_irqsave(&ihost->scic_lock,
597 flags); 595 flags);
598 596
599 /* Check for state changes. */ 597 /* Check for state changes. */
@@ -623,12 +621,12 @@ static void isci_terminate_request_core(
623 } else 621 } else
624 termination_completed = 1; 622 termination_completed = 1;
625 623
626 spin_unlock_irqrestore(&isci_host->scic_lock, 624 spin_unlock_irqrestore(&ihost->scic_lock,
627 flags); 625 flags);
628 626
629 if (!termination_completed) { 627 if (!termination_completed) {
630 628
631 dev_err(&isci_host->pdev->dev, 629 dev_err(&ihost->pdev->dev,
632 "%s: *** Timeout waiting for " 630 "%s: *** Timeout waiting for "
633 "termination(%p/%p)\n", 631 "termination(%p/%p)\n",
634 __func__, io_request_completion, 632 __func__, io_request_completion,
@@ -642,7 +640,7 @@ static void isci_terminate_request_core(
642 } 640 }
643 } 641 }
644 if (termination_completed) 642 if (termination_completed)
645 dev_dbg(&isci_host->pdev->dev, 643 dev_dbg(&ihost->pdev->dev,
646 "%s: after completion wait (%p/%p)\n", 644 "%s: after completion wait (%p/%p)\n",
647 __func__, isci_request, io_request_completion); 645 __func__, isci_request, io_request_completion);
648 } 646 }
@@ -678,7 +676,7 @@ static void isci_terminate_request_core(
678 } 676 }
679 if (needs_cleanup_handling) 677 if (needs_cleanup_handling)
680 isci_request_cleanup_completed_loiterer( 678 isci_request_cleanup_completed_loiterer(
681 isci_host, idev, isci_request, task); 679 ihost, idev, isci_request, task);
682 } 680 }
683} 681}
684 682
@@ -1253,7 +1251,7 @@ isci_task_request_complete(struct isci_host *ihost,
1253 /* PRINT_TMF( ((struct isci_tmf *)request->task)); */ 1251 /* PRINT_TMF( ((struct isci_tmf *)request->task)); */
1254 tmf_complete = tmf->complete; 1252 tmf_complete = tmf->complete;
1255 1253
1256 scic_controller_complete_io(&ihost->sci, ireq->target_device, ireq); 1254 scic_controller_complete_io(ihost, ireq->target_device, ireq);
1257 /* set the 'terminated' flag handle to make sure it cannot be terminated 1255 /* set the 'terminated' flag handle to make sure it cannot be terminated
1258 * or completed again. 1256 * or completed again.
1259 */ 1257 */
diff --git a/drivers/scsi/isci/unsolicited_frame_control.c b/drivers/scsi/isci/unsolicited_frame_control.c
index 680582d8cde5..a0e6f89fc6a1 100644
--- a/drivers/scsi/isci/unsolicited_frame_control.c
+++ b/drivers/scsi/isci/unsolicited_frame_control.c
@@ -57,9 +57,9 @@
57#include "unsolicited_frame_control.h" 57#include "unsolicited_frame_control.h"
58#include "registers.h" 58#include "registers.h"
59 59
60int scic_sds_unsolicited_frame_control_construct(struct scic_sds_controller *scic) 60int scic_sds_unsolicited_frame_control_construct(struct isci_host *ihost)
61{ 61{
62 struct scic_sds_unsolicited_frame_control *uf_control = &scic->uf_control; 62 struct scic_sds_unsolicited_frame_control *uf_control = &ihost->uf_control;
63 struct scic_sds_unsolicited_frame *uf; 63 struct scic_sds_unsolicited_frame *uf;
64 u32 buf_len, header_len, i; 64 u32 buf_len, header_len, i;
65 dma_addr_t dma; 65 dma_addr_t dma;
@@ -79,7 +79,7 @@ int scic_sds_unsolicited_frame_control_construct(struct scic_sds_controller *sci
79 * memory descriptor entry. The headers and address table will be 79 * memory descriptor entry. The headers and address table will be
80 * placed after the buffers. 80 * placed after the buffers.
81 */ 81 */
82 virt = dmam_alloc_coherent(scic_to_dev(scic), size, &dma, GFP_KERNEL); 82 virt = dmam_alloc_coherent(&ihost->pdev->dev, size, &dma, GFP_KERNEL);
83 if (!virt) 83 if (!virt)
84 return -ENOMEM; 84 return -ENOMEM;
85 85
diff --git a/drivers/scsi/isci/unsolicited_frame_control.h b/drivers/scsi/isci/unsolicited_frame_control.h
index 2954904f025a..c0285a3db562 100644
--- a/drivers/scsi/isci/unsolicited_frame_control.h
+++ b/drivers/scsi/isci/unsolicited_frame_control.h
@@ -214,9 +214,9 @@ struct scic_sds_unsolicited_frame_control {
214 214
215}; 215};
216 216
217struct scic_sds_controller; 217struct isci_host;
218 218
219int scic_sds_unsolicited_frame_control_construct(struct scic_sds_controller *scic); 219int scic_sds_unsolicited_frame_control_construct(struct isci_host *ihost);
220 220
221enum sci_status scic_sds_unsolicited_frame_control_get_header( 221enum sci_status scic_sds_unsolicited_frame_control_get_header(
222 struct scic_sds_unsolicited_frame_control *uf_control, 222 struct scic_sds_unsolicited_frame_control *uf_control,