aboutsummaryrefslogtreecommitdiffstats
path: root/drivers/scsi/sg.c
diff options
context:
space:
mode:
Diffstat (limited to 'drivers/scsi/sg.c')
-rw-r--r--drivers/scsi/sg.c397
1 files changed, 105 insertions, 292 deletions
diff --git a/drivers/scsi/sg.c b/drivers/scsi/sg.c
index cb6de0752ee1..d6391666502c 100644
--- a/drivers/scsi/sg.c
+++ b/drivers/scsi/sg.c
@@ -47,7 +47,6 @@ static int sg_version_num = 30534; /* 2 digits for each component */
47#include <linux/seq_file.h> 47#include <linux/seq_file.h>
48#include <linux/blkdev.h> 48#include <linux/blkdev.h>
49#include <linux/delay.h> 49#include <linux/delay.h>
50#include <linux/scatterlist.h>
51#include <linux/blktrace_api.h> 50#include <linux/blktrace_api.h>
52#include <linux/smp_lock.h> 51#include <linux/smp_lock.h>
53 52
@@ -119,7 +118,8 @@ typedef struct sg_scatter_hold { /* holding area for scsi scatter gather info */
119 unsigned sglist_len; /* size of malloc'd scatter-gather list ++ */ 118 unsigned sglist_len; /* size of malloc'd scatter-gather list ++ */
120 unsigned bufflen; /* Size of (aggregate) data buffer */ 119 unsigned bufflen; /* Size of (aggregate) data buffer */
121 unsigned b_malloc_len; /* actual len malloc'ed in buffer */ 120 unsigned b_malloc_len; /* actual len malloc'ed in buffer */
122 struct scatterlist *buffer;/* scatter list */ 121 struct page **pages;
122 int page_order;
123 char dio_in_use; /* 0->indirect IO (or mmap), 1->dio */ 123 char dio_in_use; /* 0->indirect IO (or mmap), 1->dio */
124 unsigned char cmd_opcode; /* first byte of command */ 124 unsigned char cmd_opcode; /* first byte of command */
125} Sg_scatter_hold; 125} Sg_scatter_hold;
@@ -190,8 +190,6 @@ static ssize_t sg_new_write(Sg_fd *sfp, struct file *file,
190 int read_only, Sg_request **o_srp); 190 int read_only, Sg_request **o_srp);
191static int sg_common_write(Sg_fd * sfp, Sg_request * srp, 191static int sg_common_write(Sg_fd * sfp, Sg_request * srp,
192 unsigned char *cmnd, int timeout, int blocking); 192 unsigned char *cmnd, int timeout, int blocking);
193static int sg_u_iovec(sg_io_hdr_t * hp, int sg_num, int ind,
194 int wr_xf, int *countp, unsigned char __user **up);
195static int sg_write_xfer(Sg_request * srp); 193static int sg_write_xfer(Sg_request * srp);
196static int sg_read_xfer(Sg_request * srp); 194static int sg_read_xfer(Sg_request * srp);
197static int sg_read_oxfer(Sg_request * srp, char __user *outp, int num_read_xfer); 195static int sg_read_oxfer(Sg_request * srp, char __user *outp, int num_read_xfer);
@@ -199,8 +197,6 @@ static void sg_remove_scat(Sg_scatter_hold * schp);
199static void sg_build_reserve(Sg_fd * sfp, int req_size); 197static void sg_build_reserve(Sg_fd * sfp, int req_size);
200static void sg_link_reserve(Sg_fd * sfp, Sg_request * srp, int size); 198static void sg_link_reserve(Sg_fd * sfp, Sg_request * srp, int size);
201static void sg_unlink_reserve(Sg_fd * sfp, Sg_request * srp); 199static void sg_unlink_reserve(Sg_fd * sfp, Sg_request * srp);
202static struct page *sg_page_malloc(int rqSz, int lowDma, int *retSzp);
203static void sg_page_free(struct page *page, int size);
204static Sg_fd *sg_add_sfp(Sg_device * sdp, int dev); 200static Sg_fd *sg_add_sfp(Sg_device * sdp, int dev);
205static int sg_remove_sfp(Sg_device * sdp, Sg_fd * sfp); 201static int sg_remove_sfp(Sg_device * sdp, Sg_fd * sfp);
206static void __sg_remove_sfp(Sg_device * sdp, Sg_fd * sfp); 202static void __sg_remove_sfp(Sg_device * sdp, Sg_fd * sfp);
@@ -771,26 +767,11 @@ sg_common_write(Sg_fd * sfp, Sg_request * srp,
771 break; 767 break;
772 } 768 }
773 hp->duration = jiffies_to_msecs(jiffies); 769 hp->duration = jiffies_to_msecs(jiffies);
774/* Now send everything of to mid-level. The next time we hear about this 770
775 packet is when sg_cmd_done() is called (i.e. a callback). */ 771 srp->rq->timeout = timeout;
776 if (srp->rq) { 772 blk_execute_rq_nowait(sdp->device->request_queue, sdp->disk,
777 srp->rq->timeout = timeout; 773 srp->rq, 1, sg_rq_end_io);
778 blk_execute_rq_nowait(sdp->device->request_queue, sdp->disk, 774 return 0;
779 srp->rq, 1, sg_rq_end_io);
780 return 0;
781 }
782 if (scsi_execute_async(sdp->device, cmnd, hp->cmd_len, data_dir, srp->data.buffer,
783 hp->dxfer_len, srp->data.k_use_sg, timeout,
784 SG_DEFAULT_RETRIES, srp, sg_cmd_done,
785 GFP_ATOMIC)) {
786 SCSI_LOG_TIMEOUT(1, printk("sg_common_write: scsi_execute_async failed\n"));
787 /*
788 * most likely out of mem, but could also be a bad map
789 */
790 sg_finish_rem_req(srp);
791 return -ENOMEM;
792 } else
793 return 0;
794} 775}
795 776
796static int 777static int
@@ -1206,8 +1187,7 @@ sg_vma_fault(struct vm_area_struct *vma, struct vm_fault *vmf)
1206 Sg_fd *sfp; 1187 Sg_fd *sfp;
1207 unsigned long offset, len, sa; 1188 unsigned long offset, len, sa;
1208 Sg_scatter_hold *rsv_schp; 1189 Sg_scatter_hold *rsv_schp;
1209 struct scatterlist *sg; 1190 int k, length;
1210 int k;
1211 1191
1212 if ((NULL == vma) || (!(sfp = (Sg_fd *) vma->vm_private_data))) 1192 if ((NULL == vma) || (!(sfp = (Sg_fd *) vma->vm_private_data)))
1213 return VM_FAULT_SIGBUS; 1193 return VM_FAULT_SIGBUS;
@@ -1217,15 +1197,14 @@ sg_vma_fault(struct vm_area_struct *vma, struct vm_fault *vmf)
1217 return VM_FAULT_SIGBUS; 1197 return VM_FAULT_SIGBUS;
1218 SCSI_LOG_TIMEOUT(3, printk("sg_vma_fault: offset=%lu, scatg=%d\n", 1198 SCSI_LOG_TIMEOUT(3, printk("sg_vma_fault: offset=%lu, scatg=%d\n",
1219 offset, rsv_schp->k_use_sg)); 1199 offset, rsv_schp->k_use_sg));
1220 sg = rsv_schp->buffer;
1221 sa = vma->vm_start; 1200 sa = vma->vm_start;
1222 for (k = 0; (k < rsv_schp->k_use_sg) && (sa < vma->vm_end); 1201 length = 1 << (PAGE_SHIFT + rsv_schp->page_order);
1223 ++k, sg = sg_next(sg)) { 1202 for (k = 0; k < rsv_schp->k_use_sg && sa < vma->vm_end; k++) {
1224 len = vma->vm_end - sa; 1203 len = vma->vm_end - sa;
1225 len = (len < sg->length) ? len : sg->length; 1204 len = (len < length) ? len : length;
1226 if (offset < len) { 1205 if (offset < len) {
1227 struct page *page; 1206 struct page *page = nth_page(rsv_schp->pages[k],
1228 page = virt_to_page(page_address(sg_page(sg)) + offset); 1207 offset >> PAGE_SHIFT);
1229 get_page(page); /* increment page count */ 1208 get_page(page); /* increment page count */
1230 vmf->page = page; 1209 vmf->page = page;
1231 return 0; /* success */ 1210 return 0; /* success */
@@ -1247,8 +1226,7 @@ sg_mmap(struct file *filp, struct vm_area_struct *vma)
1247 Sg_fd *sfp; 1226 Sg_fd *sfp;
1248 unsigned long req_sz, len, sa; 1227 unsigned long req_sz, len, sa;
1249 Sg_scatter_hold *rsv_schp; 1228 Sg_scatter_hold *rsv_schp;
1250 int k; 1229 int k, length;
1251 struct scatterlist *sg;
1252 1230
1253 if ((!filp) || (!vma) || (!(sfp = (Sg_fd *) filp->private_data))) 1231 if ((!filp) || (!vma) || (!(sfp = (Sg_fd *) filp->private_data)))
1254 return -ENXIO; 1232 return -ENXIO;
@@ -1262,11 +1240,10 @@ sg_mmap(struct file *filp, struct vm_area_struct *vma)
1262 return -ENOMEM; /* cannot map more than reserved buffer */ 1240 return -ENOMEM; /* cannot map more than reserved buffer */
1263 1241
1264 sa = vma->vm_start; 1242 sa = vma->vm_start;
1265 sg = rsv_schp->buffer; 1243 length = 1 << (PAGE_SHIFT + rsv_schp->page_order);
1266 for (k = 0; (k < rsv_schp->k_use_sg) && (sa < vma->vm_end); 1244 for (k = 0; k < rsv_schp->k_use_sg && sa < vma->vm_end; k++) {
1267 ++k, sg = sg_next(sg)) {
1268 len = vma->vm_end - sa; 1245 len = vma->vm_end - sa;
1269 len = (len < sg->length) ? len : sg->length; 1246 len = (len < length) ? len : length;
1270 sa += len; 1247 sa += len;
1271 } 1248 }
1272 1249
@@ -1310,7 +1287,6 @@ sg_cmd_done(void *data, char *sense, int result, int resid)
1310 if (0 != result) { 1287 if (0 != result) {
1311 struct scsi_sense_hdr sshdr; 1288 struct scsi_sense_hdr sshdr;
1312 1289
1313 memcpy(srp->sense_b, sense, sizeof (srp->sense_b));
1314 srp->header.status = 0xff & result; 1290 srp->header.status = 0xff & result;
1315 srp->header.masked_status = status_byte(result); 1291 srp->header.masked_status = status_byte(result);
1316 srp->header.msg_status = msg_byte(result); 1292 srp->header.msg_status = msg_byte(result);
@@ -1685,34 +1661,51 @@ static int sg_start_req(Sg_request *srp, unsigned char *cmd)
1685 Sg_scatter_hold *rsv_schp = &sfp->reserve; 1661 Sg_scatter_hold *rsv_schp = &sfp->reserve;
1686 struct request_queue *q = sfp->parentdp->device->request_queue; 1662 struct request_queue *q = sfp->parentdp->device->request_queue;
1687 unsigned long alignment = queue_dma_alignment(q) | q->dma_pad_mask; 1663 unsigned long alignment = queue_dma_alignment(q) | q->dma_pad_mask;
1664 struct rq_map_data map_data;
1688 1665
1689 SCSI_LOG_TIMEOUT(4, printk("sg_start_req: dxfer_len=%d\n", dxfer_len)); 1666 SCSI_LOG_TIMEOUT(4, printk("sg_start_req: dxfer_len=%d\n", dxfer_len));
1690 1667
1668 res = __sg_start_req(srp, hp, cmd);
1669 if (res)
1670 return res;
1671
1691 if ((dxfer_len <= 0) || (dxfer_dir == SG_DXFER_NONE)) 1672 if ((dxfer_len <= 0) || (dxfer_dir == SG_DXFER_NONE))
1692 return __sg_start_req(srp, hp, cmd); 1673 return 0;
1693 1674
1694#ifdef SG_ALLOW_DIO_CODE 1675#ifdef SG_ALLOW_DIO_CODE
1695 if (sg_allow_dio && (hp->flags & SG_FLAG_DIRECT_IO) && 1676 if (sg_allow_dio && (hp->flags & SG_FLAG_DIRECT_IO) &&
1696 (dxfer_dir != SG_DXFER_UNKNOWN) && (0 == hp->iovec_count) && 1677 (dxfer_dir != SG_DXFER_UNKNOWN) && (0 == hp->iovec_count) &&
1697 (!sfp->parentdp->device->host->unchecked_isa_dma) && 1678 (!sfp->parentdp->device->host->unchecked_isa_dma) &&
1698 !(uaddr & alignment) && !(dxfer_len & alignment)) { 1679 !(uaddr & alignment) && !(dxfer_len & alignment))
1699 res = __sg_start_req(srp, hp, cmd); 1680 return sg_build_direct(srp, sfp, dxfer_len);
1700 if (!res)
1701 res = sg_build_direct(srp, sfp, dxfer_len);
1702
1703 return res;
1704 }
1705#endif 1681#endif
1706 if ((!sg_res_in_use(sfp)) && (dxfer_len <= rsv_schp->bufflen)) 1682 if ((!sg_res_in_use(sfp)) && (dxfer_len <= rsv_schp->bufflen))
1707 sg_link_reserve(sfp, srp, dxfer_len); 1683 sg_link_reserve(sfp, srp, dxfer_len);
1708 else { 1684 else
1709 res = sg_build_indirect(req_schp, sfp, dxfer_len); 1685 res = sg_build_indirect(req_schp, sfp, dxfer_len);
1710 if (res) { 1686
1711 sg_remove_scat(req_schp); 1687 if (!res) {
1712 return res; 1688 struct request *rq = srp->rq;
1713 } 1689 Sg_scatter_hold *schp = &srp->data;
1690 int iovec_count = (int) hp->iovec_count;
1691
1692 map_data.pages = schp->pages;
1693 map_data.page_order = schp->page_order;
1694 map_data.nr_entries = schp->k_use_sg;
1695
1696 if (iovec_count)
1697 res = blk_rq_map_user_iov(q, rq, &map_data, hp->dxferp,
1698 iovec_count,
1699 hp->dxfer_len, GFP_ATOMIC);
1700 else
1701 res = blk_rq_map_user(q, rq, &map_data, hp->dxferp,
1702 hp->dxfer_len, GFP_ATOMIC);
1703
1704 if (!res)
1705 srp->bio = rq->bio;
1714 } 1706 }
1715 return 0; 1707
1708 return res;
1716} 1709}
1717 1710
1718static void 1711static void
@@ -1730,6 +1723,7 @@ sg_finish_rem_req(Sg_request * srp)
1730 if (srp->rq) { 1723 if (srp->rq) {
1731 if (srp->bio) 1724 if (srp->bio)
1732 blk_rq_unmap_user(srp->bio); 1725 blk_rq_unmap_user(srp->bio);
1726
1733 blk_put_request(srp->rq); 1727 blk_put_request(srp->rq);
1734 } 1728 }
1735 1729
@@ -1739,21 +1733,12 @@ sg_finish_rem_req(Sg_request * srp)
1739static int 1733static int
1740sg_build_sgat(Sg_scatter_hold * schp, const Sg_fd * sfp, int tablesize) 1734sg_build_sgat(Sg_scatter_hold * schp, const Sg_fd * sfp, int tablesize)
1741{ 1735{
1742 int sg_bufflen = tablesize * sizeof(struct scatterlist); 1736 int sg_bufflen = tablesize * sizeof(struct page *);
1743 gfp_t gfp_flags = GFP_ATOMIC | __GFP_NOWARN; 1737 gfp_t gfp_flags = GFP_ATOMIC | __GFP_NOWARN;
1744 1738
1745 /* 1739 schp->pages = kzalloc(sg_bufflen, gfp_flags);
1746 * TODO: test without low_dma, we should not need it since 1740 if (!schp->pages)
1747 * the block layer will bounce the buffer for us
1748 *
1749 * XXX(hch): we shouldn't need GFP_DMA for the actual S/G list.
1750 */
1751 if (sfp->low_dma)
1752 gfp_flags |= GFP_DMA;
1753 schp->buffer = kzalloc(sg_bufflen, gfp_flags);
1754 if (!schp->buffer)
1755 return -ENOMEM; 1741 return -ENOMEM;
1756 sg_init_table(schp->buffer, tablesize);
1757 schp->sglist_len = sg_bufflen; 1742 schp->sglist_len = sg_bufflen;
1758 return tablesize; /* number of scat_gath elements allocated */ 1743 return tablesize; /* number of scat_gath elements allocated */
1759} 1744}
@@ -1780,11 +1765,10 @@ sg_build_direct(Sg_request * srp, Sg_fd * sfp, int dxfer_len)
1780static int 1765static int
1781sg_build_indirect(Sg_scatter_hold * schp, Sg_fd * sfp, int buff_size) 1766sg_build_indirect(Sg_scatter_hold * schp, Sg_fd * sfp, int buff_size)
1782{ 1767{
1783 struct scatterlist *sg; 1768 int ret_sz = 0, i, k, rem_sz, num, mx_sc_elems;
1784 int ret_sz = 0, k, rem_sz, num, mx_sc_elems;
1785 int sg_tablesize = sfp->parentdp->sg_tablesize; 1769 int sg_tablesize = sfp->parentdp->sg_tablesize;
1786 int blk_size = buff_size; 1770 int blk_size = buff_size, order;
1787 struct page *p = NULL; 1771 gfp_t gfp_mask = GFP_ATOMIC | __GFP_COMP | __GFP_NOWARN;
1788 1772
1789 if (blk_size < 0) 1773 if (blk_size < 0)
1790 return -EFAULT; 1774 return -EFAULT;
@@ -1808,15 +1792,26 @@ sg_build_indirect(Sg_scatter_hold * schp, Sg_fd * sfp, int buff_size)
1808 } else 1792 } else
1809 scatter_elem_sz_prev = num; 1793 scatter_elem_sz_prev = num;
1810 } 1794 }
1811 for (k = 0, sg = schp->buffer, rem_sz = blk_size; 1795
1812 (rem_sz > 0) && (k < mx_sc_elems); 1796 if (sfp->low_dma)
1813 ++k, rem_sz -= ret_sz, sg = sg_next(sg)) { 1797 gfp_mask |= GFP_DMA;
1814 1798
1799 if (!capable(CAP_SYS_ADMIN) || !capable(CAP_SYS_RAWIO))
1800 gfp_mask |= __GFP_ZERO;
1801
1802 order = get_order(num);
1803retry:
1804 ret_sz = 1 << (PAGE_SHIFT + order);
1805
1806 for (k = 0, rem_sz = blk_size; rem_sz > 0 && k < mx_sc_elems;
1807 k++, rem_sz -= ret_sz) {
1808
1815 num = (rem_sz > scatter_elem_sz_prev) ? 1809 num = (rem_sz > scatter_elem_sz_prev) ?
1816 scatter_elem_sz_prev : rem_sz; 1810 scatter_elem_sz_prev : rem_sz;
1817 p = sg_page_malloc(num, sfp->low_dma, &ret_sz); 1811
1818 if (!p) 1812 schp->pages[k] = alloc_pages(gfp_mask, order);
1819 return -ENOMEM; 1813 if (!schp->pages[k])
1814 goto out;
1820 1815
1821 if (num == scatter_elem_sz_prev) { 1816 if (num == scatter_elem_sz_prev) {
1822 if (unlikely(ret_sz > scatter_elem_sz_prev)) { 1817 if (unlikely(ret_sz > scatter_elem_sz_prev)) {
@@ -1824,12 +1819,12 @@ sg_build_indirect(Sg_scatter_hold * schp, Sg_fd * sfp, int buff_size)
1824 scatter_elem_sz_prev = ret_sz; 1819 scatter_elem_sz_prev = ret_sz;
1825 } 1820 }
1826 } 1821 }
1827 sg_set_page(sg, p, (ret_sz > num) ? num : ret_sz, 0);
1828 1822
1829 SCSI_LOG_TIMEOUT(5, printk("sg_build_indirect: k=%d, num=%d, " 1823 SCSI_LOG_TIMEOUT(5, printk("sg_build_indirect: k=%d, num=%d, "
1830 "ret_sz=%d\n", k, num, ret_sz)); 1824 "ret_sz=%d\n", k, num, ret_sz));
1831 } /* end of for loop */ 1825 } /* end of for loop */
1832 1826
1827 schp->page_order = order;
1833 schp->k_use_sg = k; 1828 schp->k_use_sg = k;
1834 SCSI_LOG_TIMEOUT(5, printk("sg_build_indirect: k_use_sg=%d, " 1829 SCSI_LOG_TIMEOUT(5, printk("sg_build_indirect: k_use_sg=%d, "
1835 "rem_sz=%d\n", k, rem_sz)); 1830 "rem_sz=%d\n", k, rem_sz));
@@ -1837,8 +1832,15 @@ sg_build_indirect(Sg_scatter_hold * schp, Sg_fd * sfp, int buff_size)
1837 schp->bufflen = blk_size; 1832 schp->bufflen = blk_size;
1838 if (rem_sz > 0) /* must have failed */ 1833 if (rem_sz > 0) /* must have failed */
1839 return -ENOMEM; 1834 return -ENOMEM;
1840
1841 return 0; 1835 return 0;
1836out:
1837 for (i = 0; i < k; i++)
1838 __free_pages(schp->pages[k], order);
1839
1840 if (--order >= 0)
1841 goto retry;
1842
1843 return -ENOMEM;
1842} 1844}
1843 1845
1844static int 1846static int
@@ -1846,13 +1848,8 @@ sg_write_xfer(Sg_request * srp)
1846{ 1848{
1847 sg_io_hdr_t *hp = &srp->header; 1849 sg_io_hdr_t *hp = &srp->header;
1848 Sg_scatter_hold *schp = &srp->data; 1850 Sg_scatter_hold *schp = &srp->data;
1849 struct scatterlist *sg = schp->buffer;
1850 int num_xfer = 0; 1851 int num_xfer = 0;
1851 int j, k, onum, usglen, ksglen, res;
1852 int iovec_count = (int) hp->iovec_count;
1853 int dxfer_dir = hp->dxfer_direction; 1852 int dxfer_dir = hp->dxfer_direction;
1854 unsigned char *p;
1855 unsigned char __user *up;
1856 int new_interface = ('\0' == hp->interface_id) ? 0 : 1; 1853 int new_interface = ('\0' == hp->interface_id) ? 0 : 1;
1857 1854
1858 if ((SG_DXFER_UNKNOWN == dxfer_dir) || (SG_DXFER_TO_DEV == dxfer_dir) || 1855 if ((SG_DXFER_UNKNOWN == dxfer_dir) || (SG_DXFER_TO_DEV == dxfer_dir) ||
@@ -1866,83 +1863,9 @@ sg_write_xfer(Sg_request * srp)
1866 && ((SG_FLAG_NO_DXFER | SG_FLAG_MMAP_IO) & hp->flags))) 1863 && ((SG_FLAG_NO_DXFER | SG_FLAG_MMAP_IO) & hp->flags)))
1867 return 0; 1864 return 0;
1868 1865
1869 SCSI_LOG_TIMEOUT(4, printk("sg_write_xfer: num_xfer=%d, iovec_count=%d, k_use_sg=%d\n", 1866 SCSI_LOG_TIMEOUT(4, printk("sg_write_xfer: num_xfer=%d, k_use_sg=%d\n",
1870 num_xfer, iovec_count, schp->k_use_sg)); 1867 num_xfer, schp->k_use_sg));
1871 if (iovec_count) {
1872 onum = iovec_count;
1873 if (!access_ok(VERIFY_READ, hp->dxferp, SZ_SG_IOVEC * onum))
1874 return -EFAULT;
1875 } else
1876 onum = 1;
1877
1878 ksglen = sg->length;
1879 p = page_address(sg_page(sg));
1880 for (j = 0, k = 0; j < onum; ++j) {
1881 res = sg_u_iovec(hp, iovec_count, j, 1, &usglen, &up);
1882 if (res)
1883 return res;
1884
1885 for (; p; sg = sg_next(sg), ksglen = sg->length,
1886 p = page_address(sg_page(sg))) {
1887 if (usglen <= 0)
1888 break;
1889 if (ksglen > usglen) {
1890 if (usglen >= num_xfer) {
1891 if (__copy_from_user(p, up, num_xfer))
1892 return -EFAULT;
1893 return 0;
1894 }
1895 if (__copy_from_user(p, up, usglen))
1896 return -EFAULT;
1897 p += usglen;
1898 ksglen -= usglen;
1899 break;
1900 } else {
1901 if (ksglen >= num_xfer) {
1902 if (__copy_from_user(p, up, num_xfer))
1903 return -EFAULT;
1904 return 0;
1905 }
1906 if (__copy_from_user(p, up, ksglen))
1907 return -EFAULT;
1908 up += ksglen;
1909 usglen -= ksglen;
1910 }
1911 ++k;
1912 if (k >= schp->k_use_sg)
1913 return 0;
1914 }
1915 }
1916
1917 return 0;
1918}
1919 1868
1920static int
1921sg_u_iovec(sg_io_hdr_t * hp, int sg_num, int ind,
1922 int wr_xf, int *countp, unsigned char __user **up)
1923{
1924 int num_xfer = (int) hp->dxfer_len;
1925 unsigned char __user *p = hp->dxferp;
1926 int count;
1927
1928 if (0 == sg_num) {
1929 if (wr_xf && ('\0' == hp->interface_id))
1930 count = (int) hp->flags; /* holds "old" input_size */
1931 else
1932 count = num_xfer;
1933 } else {
1934 sg_iovec_t iovec;
1935 if (__copy_from_user(&iovec, p + ind*SZ_SG_IOVEC, SZ_SG_IOVEC))
1936 return -EFAULT;
1937 p = iovec.iov_base;
1938 count = (int) iovec.iov_len;
1939 }
1940 if (!access_ok(wr_xf ? VERIFY_READ : VERIFY_WRITE, p, count))
1941 return -EFAULT;
1942 if (up)
1943 *up = p;
1944 if (countp)
1945 *countp = count;
1946 return 0; 1869 return 0;
1947} 1870}
1948 1871
@@ -1950,21 +1873,18 @@ static void
1950sg_remove_scat(Sg_scatter_hold * schp) 1873sg_remove_scat(Sg_scatter_hold * schp)
1951{ 1874{
1952 SCSI_LOG_TIMEOUT(4, printk("sg_remove_scat: k_use_sg=%d\n", schp->k_use_sg)); 1875 SCSI_LOG_TIMEOUT(4, printk("sg_remove_scat: k_use_sg=%d\n", schp->k_use_sg));
1953 if (schp->buffer && (schp->sglist_len > 0)) { 1876 if (schp->pages && schp->sglist_len > 0) {
1954 struct scatterlist *sg = schp->buffer;
1955
1956 if (!schp->dio_in_use) { 1877 if (!schp->dio_in_use) {
1957 int k; 1878 int k;
1958 1879
1959 for (k = 0; (k < schp->k_use_sg) && sg_page(sg); 1880 for (k = 0; k < schp->k_use_sg && schp->pages[k]; k++) {
1960 ++k, sg = sg_next(sg)) {
1961 SCSI_LOG_TIMEOUT(5, printk( 1881 SCSI_LOG_TIMEOUT(5, printk(
1962 "sg_remove_scat: k=%d, pg=0x%p, len=%d\n", 1882 "sg_remove_scat: k=%d, pg=0x%p\n",
1963 k, sg_page(sg), sg->length)); 1883 k, schp->pages[k]));
1964 sg_page_free(sg_page(sg), sg->length); 1884 __free_pages(schp->pages[k], schp->page_order);
1965 } 1885 }
1966 1886
1967 kfree(schp->buffer); 1887 kfree(schp->pages);
1968 } 1888 }
1969 } 1889 }
1970 memset(schp, 0, sizeof (*schp)); 1890 memset(schp, 0, sizeof (*schp));
@@ -1975,13 +1895,8 @@ sg_read_xfer(Sg_request * srp)
1975{ 1895{
1976 sg_io_hdr_t *hp = &srp->header; 1896 sg_io_hdr_t *hp = &srp->header;
1977 Sg_scatter_hold *schp = &srp->data; 1897 Sg_scatter_hold *schp = &srp->data;
1978 struct scatterlist *sg = schp->buffer;
1979 int num_xfer = 0; 1898 int num_xfer = 0;
1980 int j, k, onum, usglen, ksglen, res;
1981 int iovec_count = (int) hp->iovec_count;
1982 int dxfer_dir = hp->dxfer_direction; 1899 int dxfer_dir = hp->dxfer_direction;
1983 unsigned char *p;
1984 unsigned char __user *up;
1985 int new_interface = ('\0' == hp->interface_id) ? 0 : 1; 1900 int new_interface = ('\0' == hp->interface_id) ? 0 : 1;
1986 1901
1987 if ((SG_DXFER_UNKNOWN == dxfer_dir) || (SG_DXFER_FROM_DEV == dxfer_dir) 1902 if ((SG_DXFER_UNKNOWN == dxfer_dir) || (SG_DXFER_FROM_DEV == dxfer_dir)
@@ -1996,53 +1911,7 @@ sg_read_xfer(Sg_request * srp)
1996 return 0; 1911 return 0;
1997 1912
1998 SCSI_LOG_TIMEOUT(4, printk("sg_read_xfer: num_xfer=%d, iovec_count=%d, k_use_sg=%d\n", 1913 SCSI_LOG_TIMEOUT(4, printk("sg_read_xfer: num_xfer=%d, iovec_count=%d, k_use_sg=%d\n",
1999 num_xfer, iovec_count, schp->k_use_sg)); 1914 num_xfer, (int)hp->iovec_count, schp->k_use_sg));
2000 if (iovec_count) {
2001 onum = iovec_count;
2002 if (!access_ok(VERIFY_READ, hp->dxferp, SZ_SG_IOVEC * onum))
2003 return -EFAULT;
2004 } else
2005 onum = 1;
2006
2007 p = page_address(sg_page(sg));
2008 ksglen = sg->length;
2009 for (j = 0, k = 0; j < onum; ++j) {
2010 res = sg_u_iovec(hp, iovec_count, j, 0, &usglen, &up);
2011 if (res)
2012 return res;
2013
2014 for (; p; sg = sg_next(sg), ksglen = sg->length,
2015 p = page_address(sg_page(sg))) {
2016 if (usglen <= 0)
2017 break;
2018 if (ksglen > usglen) {
2019 if (usglen >= num_xfer) {
2020 if (__copy_to_user(up, p, num_xfer))
2021 return -EFAULT;
2022 return 0;
2023 }
2024 if (__copy_to_user(up, p, usglen))
2025 return -EFAULT;
2026 p += usglen;
2027 ksglen -= usglen;
2028 break;
2029 } else {
2030 if (ksglen >= num_xfer) {
2031 if (__copy_to_user(up, p, num_xfer))
2032 return -EFAULT;
2033 return 0;
2034 }
2035 if (__copy_to_user(up, p, ksglen))
2036 return -EFAULT;
2037 up += ksglen;
2038 usglen -= ksglen;
2039 }
2040 ++k;
2041 if (k >= schp->k_use_sg)
2042 return 0;
2043 }
2044 }
2045
2046 return 0; 1915 return 0;
2047} 1916}
2048 1917
@@ -2050,7 +1919,6 @@ static int
2050sg_read_oxfer(Sg_request * srp, char __user *outp, int num_read_xfer) 1919sg_read_oxfer(Sg_request * srp, char __user *outp, int num_read_xfer)
2051{ 1920{
2052 Sg_scatter_hold *schp = &srp->data; 1921 Sg_scatter_hold *schp = &srp->data;
2053 struct scatterlist *sg = schp->buffer;
2054 int k, num; 1922 int k, num;
2055 1923
2056 SCSI_LOG_TIMEOUT(4, printk("sg_read_oxfer: num_read_xfer=%d\n", 1924 SCSI_LOG_TIMEOUT(4, printk("sg_read_oxfer: num_read_xfer=%d\n",
@@ -2058,15 +1926,18 @@ sg_read_oxfer(Sg_request * srp, char __user *outp, int num_read_xfer)
2058 if ((!outp) || (num_read_xfer <= 0)) 1926 if ((!outp) || (num_read_xfer <= 0))
2059 return 0; 1927 return 0;
2060 1928
2061 for (k = 0; (k < schp->k_use_sg) && sg_page(sg); ++k, sg = sg_next(sg)) { 1929 blk_rq_unmap_user(srp->bio);
2062 num = sg->length; 1930 srp->bio = NULL;
1931
1932 num = 1 << (PAGE_SHIFT + schp->page_order);
1933 for (k = 0; k < schp->k_use_sg && schp->pages[k]; k++) {
2063 if (num > num_read_xfer) { 1934 if (num > num_read_xfer) {
2064 if (__copy_to_user(outp, page_address(sg_page(sg)), 1935 if (__copy_to_user(outp, page_address(schp->pages[k]),
2065 num_read_xfer)) 1936 num_read_xfer))
2066 return -EFAULT; 1937 return -EFAULT;
2067 break; 1938 break;
2068 } else { 1939 } else {
2069 if (__copy_to_user(outp, page_address(sg_page(sg)), 1940 if (__copy_to_user(outp, page_address(schp->pages[k]),
2070 num)) 1941 num))
2071 return -EFAULT; 1942 return -EFAULT;
2072 num_read_xfer -= num; 1943 num_read_xfer -= num;
@@ -2101,24 +1972,22 @@ sg_link_reserve(Sg_fd * sfp, Sg_request * srp, int size)
2101{ 1972{
2102 Sg_scatter_hold *req_schp = &srp->data; 1973 Sg_scatter_hold *req_schp = &srp->data;
2103 Sg_scatter_hold *rsv_schp = &sfp->reserve; 1974 Sg_scatter_hold *rsv_schp = &sfp->reserve;
2104 struct scatterlist *sg = rsv_schp->buffer;
2105 int k, num, rem; 1975 int k, num, rem;
2106 1976
2107 srp->res_used = 1; 1977 srp->res_used = 1;
2108 SCSI_LOG_TIMEOUT(4, printk("sg_link_reserve: size=%d\n", size)); 1978 SCSI_LOG_TIMEOUT(4, printk("sg_link_reserve: size=%d\n", size));
2109 rem = size; 1979 rem = size;
2110 1980
2111 for (k = 0; k < rsv_schp->k_use_sg; ++k, sg = sg_next(sg)) { 1981 num = 1 << (PAGE_SHIFT + rsv_schp->page_order);
2112 num = sg->length; 1982 for (k = 0; k < rsv_schp->k_use_sg; k++) {
2113 if (rem <= num) { 1983 if (rem <= num) {
2114 sfp->save_scat_len = num;
2115 sg->length = rem;
2116 req_schp->k_use_sg = k + 1; 1984 req_schp->k_use_sg = k + 1;
2117 req_schp->sglist_len = rsv_schp->sglist_len; 1985 req_schp->sglist_len = rsv_schp->sglist_len;
2118 req_schp->buffer = rsv_schp->buffer; 1986 req_schp->pages = rsv_schp->pages;
2119 1987
2120 req_schp->bufflen = size; 1988 req_schp->bufflen = size;
2121 req_schp->b_malloc_len = rsv_schp->b_malloc_len; 1989 req_schp->b_malloc_len = rsv_schp->b_malloc_len;
1990 req_schp->page_order = rsv_schp->page_order;
2122 break; 1991 break;
2123 } else 1992 } else
2124 rem -= num; 1993 rem -= num;
@@ -2132,22 +2001,13 @@ static void
2132sg_unlink_reserve(Sg_fd * sfp, Sg_request * srp) 2001sg_unlink_reserve(Sg_fd * sfp, Sg_request * srp)
2133{ 2002{
2134 Sg_scatter_hold *req_schp = &srp->data; 2003 Sg_scatter_hold *req_schp = &srp->data;
2135 Sg_scatter_hold *rsv_schp = &sfp->reserve;
2136 2004
2137 SCSI_LOG_TIMEOUT(4, printk("sg_unlink_reserve: req->k_use_sg=%d\n", 2005 SCSI_LOG_TIMEOUT(4, printk("sg_unlink_reserve: req->k_use_sg=%d\n",
2138 (int) req_schp->k_use_sg)); 2006 (int) req_schp->k_use_sg));
2139 if ((rsv_schp->k_use_sg > 0) && (req_schp->k_use_sg > 0)) {
2140 struct scatterlist *sg = rsv_schp->buffer;
2141
2142 if (sfp->save_scat_len > 0)
2143 (sg + (req_schp->k_use_sg - 1))->length =
2144 (unsigned) sfp->save_scat_len;
2145 else
2146 SCSI_LOG_TIMEOUT(1, printk ("sg_unlink_reserve: BAD save_scat_len\n"));
2147 }
2148 req_schp->k_use_sg = 0; 2007 req_schp->k_use_sg = 0;
2149 req_schp->bufflen = 0; 2008 req_schp->bufflen = 0;
2150 req_schp->buffer = NULL; 2009 req_schp->pages = NULL;
2010 req_schp->page_order = 0;
2151 req_schp->sglist_len = 0; 2011 req_schp->sglist_len = 0;
2152 sfp->save_scat_len = 0; 2012 sfp->save_scat_len = 0;
2153 srp->res_used = 0; 2013 srp->res_used = 0;
@@ -2405,53 +2265,6 @@ sg_res_in_use(Sg_fd * sfp)
2405 return srp ? 1 : 0; 2265 return srp ? 1 : 0;
2406} 2266}
2407 2267
2408/* The size fetched (value output via retSzp) set when non-NULL return */
2409static struct page *
2410sg_page_malloc(int rqSz, int lowDma, int *retSzp)
2411{
2412 struct page *resp = NULL;
2413 gfp_t page_mask;
2414 int order, a_size;
2415 int resSz;
2416
2417 if ((rqSz <= 0) || (NULL == retSzp))
2418 return resp;
2419
2420 if (lowDma)
2421 page_mask = GFP_ATOMIC | GFP_DMA | __GFP_COMP | __GFP_NOWARN;
2422 else
2423 page_mask = GFP_ATOMIC | __GFP_COMP | __GFP_NOWARN;
2424
2425 for (order = 0, a_size = PAGE_SIZE; a_size < rqSz;
2426 order++, a_size <<= 1) ;
2427 resSz = a_size; /* rounded up if necessary */
2428 resp = alloc_pages(page_mask, order);
2429 while ((!resp) && order) {
2430 --order;
2431 a_size >>= 1; /* divide by 2, until PAGE_SIZE */
2432 resp = alloc_pages(page_mask, order); /* try half */
2433 resSz = a_size;
2434 }
2435 if (resp) {
2436 if (!capable(CAP_SYS_ADMIN) || !capable(CAP_SYS_RAWIO))
2437 memset(page_address(resp), 0, resSz);
2438 *retSzp = resSz;
2439 }
2440 return resp;
2441}
2442
2443static void
2444sg_page_free(struct page *page, int size)
2445{
2446 int order, a_size;
2447
2448 if (!page)
2449 return;
2450 for (order = 0, a_size = PAGE_SIZE; a_size < size;
2451 order++, a_size <<= 1) ;
2452 __free_pages(page, order);
2453}
2454
2455#ifdef CONFIG_SCSI_PROC_FS 2268#ifdef CONFIG_SCSI_PROC_FS
2456static int 2269static int
2457sg_idr_max_id(int id, void *p, void *data) 2270sg_idr_max_id(int id, void *p, void *data)