aboutsummaryrefslogtreecommitdiffstats
path: root/drivers/scsi/ipr.c
diff options
context:
space:
mode:
Diffstat (limited to 'drivers/scsi/ipr.c')
-rw-r--r--drivers/scsi/ipr.c201
1 files changed, 192 insertions, 9 deletions
diff --git a/drivers/scsi/ipr.c b/drivers/scsi/ipr.c
index 8817ea06adbf..7149aada3f3c 100644
--- a/drivers/scsi/ipr.c
+++ b/drivers/scsi/ipr.c
@@ -91,6 +91,7 @@ static unsigned int ipr_max_speed = 1;
91static int ipr_testmode = 0; 91static int ipr_testmode = 0;
92static unsigned int ipr_fastfail = 0; 92static unsigned int ipr_fastfail = 0;
93static unsigned int ipr_transop_timeout = IPR_OPERATIONAL_TIMEOUT; 93static unsigned int ipr_transop_timeout = IPR_OPERATIONAL_TIMEOUT;
94static unsigned int ipr_enable_cache = 1;
94static DEFINE_SPINLOCK(ipr_driver_lock); 95static DEFINE_SPINLOCK(ipr_driver_lock);
95 96
96/* This table describes the differences between DMA controller chips */ 97/* This table describes the differences between DMA controller chips */
@@ -150,6 +151,8 @@ module_param_named(fastfail, ipr_fastfail, int, 0);
150MODULE_PARM_DESC(fastfail, "Reduce timeouts and retries"); 151MODULE_PARM_DESC(fastfail, "Reduce timeouts and retries");
151module_param_named(transop_timeout, ipr_transop_timeout, int, 0); 152module_param_named(transop_timeout, ipr_transop_timeout, int, 0);
152MODULE_PARM_DESC(transop_timeout, "Time in seconds to wait for adapter to come operational (default: 300)"); 153MODULE_PARM_DESC(transop_timeout, "Time in seconds to wait for adapter to come operational (default: 300)");
154module_param_named(enable_cache, ipr_enable_cache, int, 0);
155MODULE_PARM_DESC(enable_cache, "Enable adapter's non-volatile write cache (default: 1)");
153MODULE_LICENSE("GPL"); 156MODULE_LICENSE("GPL");
154MODULE_VERSION(IPR_DRIVER_VERSION); 157MODULE_VERSION(IPR_DRIVER_VERSION);
155 158
@@ -1937,6 +1940,103 @@ static struct bin_attribute ipr_trace_attr = {
1937}; 1940};
1938#endif 1941#endif
1939 1942
1943static const struct {
1944 enum ipr_cache_state state;
1945 char *name;
1946} cache_state [] = {
1947 { CACHE_NONE, "none" },
1948 { CACHE_DISABLED, "disabled" },
1949 { CACHE_ENABLED, "enabled" }
1950};
1951
1952/**
1953 * ipr_show_write_caching - Show the write caching attribute
1954 * @class_dev: class device struct
1955 * @buf: buffer
1956 *
1957 * Return value:
1958 * number of bytes printed to buffer
1959 **/
1960static ssize_t ipr_show_write_caching(struct class_device *class_dev, char *buf)
1961{
1962 struct Scsi_Host *shost = class_to_shost(class_dev);
1963 struct ipr_ioa_cfg *ioa_cfg = (struct ipr_ioa_cfg *)shost->hostdata;
1964 unsigned long lock_flags = 0;
1965 int i, len = 0;
1966
1967 spin_lock_irqsave(ioa_cfg->host->host_lock, lock_flags);
1968 for (i = 0; i < ARRAY_SIZE(cache_state); i++) {
1969 if (cache_state[i].state == ioa_cfg->cache_state) {
1970 len = snprintf(buf, PAGE_SIZE, "%s\n", cache_state[i].name);
1971 break;
1972 }
1973 }
1974 spin_unlock_irqrestore(ioa_cfg->host->host_lock, lock_flags);
1975 return len;
1976}
1977
1978
1979/**
1980 * ipr_store_write_caching - Enable/disable adapter write cache
1981 * @class_dev: class_device struct
1982 * @buf: buffer
1983 * @count: buffer size
1984 *
1985 * This function will enable/disable adapter write cache.
1986 *
1987 * Return value:
1988 * count on success / other on failure
1989 **/
1990static ssize_t ipr_store_write_caching(struct class_device *class_dev,
1991 const char *buf, size_t count)
1992{
1993 struct Scsi_Host *shost = class_to_shost(class_dev);
1994 struct ipr_ioa_cfg *ioa_cfg = (struct ipr_ioa_cfg *)shost->hostdata;
1995 unsigned long lock_flags = 0;
1996 enum ipr_cache_state new_state = CACHE_INVALID;
1997 int i;
1998
1999 if (!capable(CAP_SYS_ADMIN))
2000 return -EACCES;
2001 if (ioa_cfg->cache_state == CACHE_NONE)
2002 return -EINVAL;
2003
2004 for (i = 0; i < ARRAY_SIZE(cache_state); i++) {
2005 if (!strncmp(cache_state[i].name, buf, strlen(cache_state[i].name))) {
2006 new_state = cache_state[i].state;
2007 break;
2008 }
2009 }
2010
2011 if (new_state != CACHE_DISABLED && new_state != CACHE_ENABLED)
2012 return -EINVAL;
2013
2014 spin_lock_irqsave(ioa_cfg->host->host_lock, lock_flags);
2015 if (ioa_cfg->cache_state == new_state) {
2016 spin_unlock_irqrestore(ioa_cfg->host->host_lock, lock_flags);
2017 return count;
2018 }
2019
2020 ioa_cfg->cache_state = new_state;
2021 dev_info(&ioa_cfg->pdev->dev, "%s adapter write cache.\n",
2022 new_state == CACHE_ENABLED ? "Enabling" : "Disabling");
2023 if (!ioa_cfg->in_reset_reload)
2024 ipr_initiate_ioa_reset(ioa_cfg, IPR_SHUTDOWN_NORMAL);
2025 spin_unlock_irqrestore(ioa_cfg->host->host_lock, lock_flags);
2026 wait_event(ioa_cfg->reset_wait_q, !ioa_cfg->in_reset_reload);
2027
2028 return count;
2029}
2030
2031static struct class_device_attribute ipr_ioa_cache_attr = {
2032 .attr = {
2033 .name = "write_cache",
2034 .mode = S_IRUGO | S_IWUSR,
2035 },
2036 .show = ipr_show_write_caching,
2037 .store = ipr_store_write_caching
2038};
2039
1940/** 2040/**
1941 * ipr_show_fw_version - Show the firmware version 2041 * ipr_show_fw_version - Show the firmware version
1942 * @class_dev: class device struct 2042 * @class_dev: class device struct
@@ -2406,6 +2506,7 @@ static struct class_device_attribute *ipr_ioa_attrs[] = {
2406 &ipr_diagnostics_attr, 2506 &ipr_diagnostics_attr,
2407 &ipr_ioa_reset_attr, 2507 &ipr_ioa_reset_attr,
2408 &ipr_update_fw_attr, 2508 &ipr_update_fw_attr,
2509 &ipr_ioa_cache_attr,
2409 NULL, 2510 NULL,
2410}; 2511};
2411 2512
@@ -4148,6 +4249,36 @@ static int ipr_set_supported_devs(struct ipr_cmnd *ipr_cmd)
4148} 4249}
4149 4250
4150/** 4251/**
4252 * ipr_setup_write_cache - Disable write cache if needed
4253 * @ipr_cmd: ipr command struct
4254 *
4255 * This function sets up adapters write cache to desired setting
4256 *
4257 * Return value:
4258 * IPR_RC_JOB_CONTINUE / IPR_RC_JOB_RETURN
4259 **/
4260static int ipr_setup_write_cache(struct ipr_cmnd *ipr_cmd)
4261{
4262 struct ipr_ioa_cfg *ioa_cfg = ipr_cmd->ioa_cfg;
4263
4264 ipr_cmd->job_step = ipr_set_supported_devs;
4265 ipr_cmd->u.res = list_entry(ioa_cfg->used_res_q.next,
4266 struct ipr_resource_entry, queue);
4267
4268 if (ioa_cfg->cache_state != CACHE_DISABLED)
4269 return IPR_RC_JOB_CONTINUE;
4270
4271 ipr_cmd->ioarcb.res_handle = cpu_to_be32(IPR_IOA_RES_HANDLE);
4272 ipr_cmd->ioarcb.cmd_pkt.request_type = IPR_RQTYPE_IOACMD;
4273 ipr_cmd->ioarcb.cmd_pkt.cdb[0] = IPR_IOA_SHUTDOWN;
4274 ipr_cmd->ioarcb.cmd_pkt.cdb[1] = IPR_SHUTDOWN_PREPARE_FOR_NORMAL;
4275
4276 ipr_do_req(ipr_cmd, ipr_reset_ioa_job, ipr_timeout, IPR_INTERNAL_TIMEOUT);
4277
4278 return IPR_RC_JOB_RETURN;
4279}
4280
4281/**
4151 * ipr_get_mode_page - Locate specified mode page 4282 * ipr_get_mode_page - Locate specified mode page
4152 * @mode_pages: mode page buffer 4283 * @mode_pages: mode page buffer
4153 * @page_code: page code to find 4284 * @page_code: page code to find
@@ -4358,10 +4489,7 @@ static int ipr_ioafp_mode_select_page28(struct ipr_cmnd *ipr_cmd)
4358 ioa_cfg->vpd_cbs_dma + offsetof(struct ipr_misc_cbs, mode_pages), 4489 ioa_cfg->vpd_cbs_dma + offsetof(struct ipr_misc_cbs, mode_pages),
4359 length); 4490 length);
4360 4491
4361 ipr_cmd->job_step = ipr_set_supported_devs; 4492 ipr_cmd->job_step = ipr_setup_write_cache;
4362 ipr_cmd->u.res = list_entry(ioa_cfg->used_res_q.next,
4363 struct ipr_resource_entry, queue);
4364
4365 ipr_do_req(ipr_cmd, ipr_reset_ioa_job, ipr_timeout, IPR_INTERNAL_TIMEOUT); 4493 ipr_do_req(ipr_cmd, ipr_reset_ioa_job, ipr_timeout, IPR_INTERNAL_TIMEOUT);
4366 4494
4367 LEAVE; 4495 LEAVE;
@@ -4581,6 +4709,27 @@ static void ipr_ioafp_inquiry(struct ipr_cmnd *ipr_cmd, u8 flags, u8 page,
4581} 4709}
4582 4710
4583/** 4711/**
4712 * ipr_inquiry_page_supported - Is the given inquiry page supported
4713 * @page0: inquiry page 0 buffer
4714 * @page: page code.
4715 *
4716 * This function determines if the specified inquiry page is supported.
4717 *
4718 * Return value:
4719 * 1 if page is supported / 0 if not
4720 **/
4721static int ipr_inquiry_page_supported(struct ipr_inquiry_page0 *page0, u8 page)
4722{
4723 int i;
4724
4725 for (i = 0; i < min_t(u8, page0->len, IPR_INQUIRY_PAGE0_ENTRIES); i++)
4726 if (page0->page[i] == page)
4727 return 1;
4728
4729 return 0;
4730}
4731
4732/**
4584 * ipr_ioafp_page3_inquiry - Send a Page 3 Inquiry to the adapter. 4733 * ipr_ioafp_page3_inquiry - Send a Page 3 Inquiry to the adapter.
4585 * @ipr_cmd: ipr command struct 4734 * @ipr_cmd: ipr command struct
4586 * 4735 *
@@ -4593,6 +4742,36 @@ static void ipr_ioafp_inquiry(struct ipr_cmnd *ipr_cmd, u8 flags, u8 page,
4593static int ipr_ioafp_page3_inquiry(struct ipr_cmnd *ipr_cmd) 4742static int ipr_ioafp_page3_inquiry(struct ipr_cmnd *ipr_cmd)
4594{ 4743{
4595 struct ipr_ioa_cfg *ioa_cfg = ipr_cmd->ioa_cfg; 4744 struct ipr_ioa_cfg *ioa_cfg = ipr_cmd->ioa_cfg;
4745 struct ipr_inquiry_page0 *page0 = &ioa_cfg->vpd_cbs->page0_data;
4746
4747 ENTER;
4748
4749 if (!ipr_inquiry_page_supported(page0, 1))
4750 ioa_cfg->cache_state = CACHE_NONE;
4751
4752 ipr_cmd->job_step = ipr_ioafp_query_ioa_cfg;
4753
4754 ipr_ioafp_inquiry(ipr_cmd, 1, 3,
4755 ioa_cfg->vpd_cbs_dma + offsetof(struct ipr_misc_cbs, page3_data),
4756 sizeof(struct ipr_inquiry_page3));
4757
4758 LEAVE;
4759 return IPR_RC_JOB_RETURN;
4760}
4761
4762/**
4763 * ipr_ioafp_page0_inquiry - Send a Page 0 Inquiry to the adapter.
4764 * @ipr_cmd: ipr command struct
4765 *
4766 * This function sends a Page 0 inquiry to the adapter
4767 * to retrieve supported inquiry pages.
4768 *
4769 * Return value:
4770 * IPR_RC_JOB_CONTINUE / IPR_RC_JOB_RETURN
4771 **/
4772static int ipr_ioafp_page0_inquiry(struct ipr_cmnd *ipr_cmd)
4773{
4774 struct ipr_ioa_cfg *ioa_cfg = ipr_cmd->ioa_cfg;
4596 char type[5]; 4775 char type[5];
4597 4776
4598 ENTER; 4777 ENTER;
@@ -4602,11 +4781,11 @@ static int ipr_ioafp_page3_inquiry(struct ipr_cmnd *ipr_cmd)
4602 type[4] = '\0'; 4781 type[4] = '\0';
4603 ioa_cfg->type = simple_strtoul((char *)type, NULL, 16); 4782 ioa_cfg->type = simple_strtoul((char *)type, NULL, 16);
4604 4783
4605 ipr_cmd->job_step = ipr_ioafp_query_ioa_cfg; 4784 ipr_cmd->job_step = ipr_ioafp_page3_inquiry;
4606 4785
4607 ipr_ioafp_inquiry(ipr_cmd, 1, 3, 4786 ipr_ioafp_inquiry(ipr_cmd, 1, 0,
4608 ioa_cfg->vpd_cbs_dma + offsetof(struct ipr_misc_cbs, page3_data), 4787 ioa_cfg->vpd_cbs_dma + offsetof(struct ipr_misc_cbs, page0_data),
4609 sizeof(struct ipr_inquiry_page3)); 4788 sizeof(struct ipr_inquiry_page0));
4610 4789
4611 LEAVE; 4790 LEAVE;
4612 return IPR_RC_JOB_RETURN; 4791 return IPR_RC_JOB_RETURN;
@@ -4626,7 +4805,7 @@ static int ipr_ioafp_std_inquiry(struct ipr_cmnd *ipr_cmd)
4626 struct ipr_ioa_cfg *ioa_cfg = ipr_cmd->ioa_cfg; 4805 struct ipr_ioa_cfg *ioa_cfg = ipr_cmd->ioa_cfg;
4627 4806
4628 ENTER; 4807 ENTER;
4629 ipr_cmd->job_step = ipr_ioafp_page3_inquiry; 4808 ipr_cmd->job_step = ipr_ioafp_page0_inquiry;
4630 4809
4631 ipr_ioafp_inquiry(ipr_cmd, 0, 0, 4810 ipr_ioafp_inquiry(ipr_cmd, 0, 0,
4632 ioa_cfg->vpd_cbs_dma + offsetof(struct ipr_misc_cbs, ioa_vpd), 4811 ioa_cfg->vpd_cbs_dma + offsetof(struct ipr_misc_cbs, ioa_vpd),
@@ -5629,6 +5808,10 @@ static void __devinit ipr_init_ioa_cfg(struct ipr_ioa_cfg *ioa_cfg,
5629 INIT_WORK(&ioa_cfg->work_q, ipr_worker_thread, ioa_cfg); 5808 INIT_WORK(&ioa_cfg->work_q, ipr_worker_thread, ioa_cfg);
5630 init_waitqueue_head(&ioa_cfg->reset_wait_q); 5809 init_waitqueue_head(&ioa_cfg->reset_wait_q);
5631 ioa_cfg->sdt_state = INACTIVE; 5810 ioa_cfg->sdt_state = INACTIVE;
5811 if (ipr_enable_cache)
5812 ioa_cfg->cache_state = CACHE_ENABLED;
5813 else
5814 ioa_cfg->cache_state = CACHE_DISABLED;
5632 5815
5633 ipr_initialize_bus_attr(ioa_cfg); 5816 ipr_initialize_bus_attr(ioa_cfg);
5634 5817