diff options
-rw-r--r-- | include/net/iucv/af_iucv.h | 7 | ||||
-rw-r--r-- | net/iucv/af_iucv.c | 211 |
2 files changed, 132 insertions, 86 deletions
diff --git a/include/net/iucv/af_iucv.h b/include/net/iucv/af_iucv.h index 2ce0c901eadf..85f80eadfa35 100644 --- a/include/net/iucv/af_iucv.h +++ b/include/net/iucv/af_iucv.h | |||
@@ -50,6 +50,12 @@ struct sockaddr_iucv { | |||
50 | 50 | ||
51 | 51 | ||
52 | /* Common socket structures and functions */ | 52 | /* Common socket structures and functions */ |
53 | struct sock_msg_q { | ||
54 | struct iucv_path *path; | ||
55 | struct iucv_message msg; | ||
56 | struct list_head list; | ||
57 | spinlock_t lock; | ||
58 | }; | ||
53 | 59 | ||
54 | #define iucv_sk(__sk) ((struct iucv_sock *) __sk) | 60 | #define iucv_sk(__sk) ((struct iucv_sock *) __sk) |
55 | 61 | ||
@@ -65,6 +71,7 @@ struct iucv_sock { | |||
65 | struct iucv_path *path; | 71 | struct iucv_path *path; |
66 | struct sk_buff_head send_skb_q; | 72 | struct sk_buff_head send_skb_q; |
67 | struct sk_buff_head backlog_skb_q; | 73 | struct sk_buff_head backlog_skb_q; |
74 | struct sock_msg_q message_q; | ||
68 | unsigned int send_tag; | 75 | unsigned int send_tag; |
69 | }; | 76 | }; |
70 | 77 | ||
diff --git a/net/iucv/af_iucv.c b/net/iucv/af_iucv.c index 65358722c397..43e01c8d382b 100644 --- a/net/iucv/af_iucv.c +++ b/net/iucv/af_iucv.c | |||
@@ -224,6 +224,8 @@ static struct sock *iucv_sock_alloc(struct socket *sock, int proto, gfp_t prio) | |||
224 | INIT_LIST_HEAD(&iucv_sk(sk)->accept_q); | 224 | INIT_LIST_HEAD(&iucv_sk(sk)->accept_q); |
225 | spin_lock_init(&iucv_sk(sk)->accept_q_lock); | 225 | spin_lock_init(&iucv_sk(sk)->accept_q_lock); |
226 | skb_queue_head_init(&iucv_sk(sk)->send_skb_q); | 226 | skb_queue_head_init(&iucv_sk(sk)->send_skb_q); |
227 | INIT_LIST_HEAD(&iucv_sk(sk)->message_q.list); | ||
228 | spin_lock_init(&iucv_sk(sk)->message_q.lock); | ||
227 | skb_queue_head_init(&iucv_sk(sk)->backlog_skb_q); | 229 | skb_queue_head_init(&iucv_sk(sk)->backlog_skb_q); |
228 | iucv_sk(sk)->send_tag = 0; | 230 | iucv_sk(sk)->send_tag = 0; |
229 | 231 | ||
@@ -673,6 +675,90 @@ out: | |||
673 | return err; | 675 | return err; |
674 | } | 676 | } |
675 | 677 | ||
678 | static int iucv_fragment_skb(struct sock *sk, struct sk_buff *skb, int len) | ||
679 | { | ||
680 | int dataleft, size, copied = 0; | ||
681 | struct sk_buff *nskb; | ||
682 | |||
683 | dataleft = len; | ||
684 | while (dataleft) { | ||
685 | if (dataleft >= sk->sk_rcvbuf / 4) | ||
686 | size = sk->sk_rcvbuf / 4; | ||
687 | else | ||
688 | size = dataleft; | ||
689 | |||
690 | nskb = alloc_skb(size, GFP_ATOMIC | GFP_DMA); | ||
691 | if (!nskb) | ||
692 | return -ENOMEM; | ||
693 | |||
694 | memcpy(nskb->data, skb->data + copied, size); | ||
695 | copied += size; | ||
696 | dataleft -= size; | ||
697 | |||
698 | skb_reset_transport_header(nskb); | ||
699 | skb_reset_network_header(nskb); | ||
700 | nskb->len = size; | ||
701 | |||
702 | skb_queue_tail(&iucv_sk(sk)->backlog_skb_q, nskb); | ||
703 | } | ||
704 | |||
705 | return 0; | ||
706 | } | ||
707 | |||
708 | static void iucv_process_message(struct sock *sk, struct sk_buff *skb, | ||
709 | struct iucv_path *path, | ||
710 | struct iucv_message *msg) | ||
711 | { | ||
712 | int rc; | ||
713 | |||
714 | if (msg->flags & IPRMDATA) { | ||
715 | skb->data = NULL; | ||
716 | skb->len = 0; | ||
717 | } else { | ||
718 | rc = iucv_message_receive(path, msg, 0, skb->data, | ||
719 | msg->length, NULL); | ||
720 | if (rc) { | ||
721 | kfree_skb(skb); | ||
722 | return; | ||
723 | } | ||
724 | if (skb->truesize >= sk->sk_rcvbuf / 4) { | ||
725 | rc = iucv_fragment_skb(sk, skb, msg->length); | ||
726 | kfree_skb(skb); | ||
727 | skb = NULL; | ||
728 | if (rc) { | ||
729 | iucv_path_sever(path, NULL); | ||
730 | return; | ||
731 | } | ||
732 | skb = skb_dequeue(&iucv_sk(sk)->backlog_skb_q); | ||
733 | } else { | ||
734 | skb_reset_transport_header(skb); | ||
735 | skb_reset_network_header(skb); | ||
736 | skb->len = msg->length; | ||
737 | } | ||
738 | } | ||
739 | |||
740 | if (sock_queue_rcv_skb(sk, skb)) | ||
741 | skb_queue_head(&iucv_sk(sk)->backlog_skb_q, skb); | ||
742 | } | ||
743 | |||
744 | static void iucv_process_message_q(struct sock *sk) | ||
745 | { | ||
746 | struct iucv_sock *iucv = iucv_sk(sk); | ||
747 | struct sk_buff *skb; | ||
748 | struct sock_msg_q *p, *n; | ||
749 | |||
750 | list_for_each_entry_safe(p, n, &iucv->message_q.list, list) { | ||
751 | skb = alloc_skb(p->msg.length, GFP_ATOMIC | GFP_DMA); | ||
752 | if (!skb) | ||
753 | break; | ||
754 | iucv_process_message(sk, skb, p->path, &p->msg); | ||
755 | list_del(&p->list); | ||
756 | kfree(p); | ||
757 | if (!skb_queue_empty(&iucv->backlog_skb_q)) | ||
758 | break; | ||
759 | } | ||
760 | } | ||
761 | |||
676 | static int iucv_sock_recvmsg(struct kiocb *iocb, struct socket *sock, | 762 | static int iucv_sock_recvmsg(struct kiocb *iocb, struct socket *sock, |
677 | struct msghdr *msg, size_t len, int flags) | 763 | struct msghdr *msg, size_t len, int flags) |
678 | { | 764 | { |
@@ -684,8 +770,9 @@ static int iucv_sock_recvmsg(struct kiocb *iocb, struct socket *sock, | |||
684 | int err = 0; | 770 | int err = 0; |
685 | 771 | ||
686 | if ((sk->sk_state == IUCV_DISCONN || sk->sk_state == IUCV_SEVERED) && | 772 | if ((sk->sk_state == IUCV_DISCONN || sk->sk_state == IUCV_SEVERED) && |
687 | skb_queue_empty(&iucv->backlog_skb_q) && | 773 | skb_queue_empty(&iucv->backlog_skb_q) && |
688 | skb_queue_empty(&sk->sk_receive_queue)) | 774 | skb_queue_empty(&sk->sk_receive_queue) && |
775 | list_empty(&iucv->message_q.list)) | ||
689 | return 0; | 776 | return 0; |
690 | 777 | ||
691 | if (flags & (MSG_OOB)) | 778 | if (flags & (MSG_OOB)) |
@@ -724,16 +811,23 @@ static int iucv_sock_recvmsg(struct kiocb *iocb, struct socket *sock, | |||
724 | kfree_skb(skb); | 811 | kfree_skb(skb); |
725 | 812 | ||
726 | /* Queue backlog skbs */ | 813 | /* Queue backlog skbs */ |
727 | rskb = skb_dequeue(&iucv_sk(sk)->backlog_skb_q); | 814 | rskb = skb_dequeue(&iucv->backlog_skb_q); |
728 | while (rskb) { | 815 | while (rskb) { |
729 | if (sock_queue_rcv_skb(sk, rskb)) { | 816 | if (sock_queue_rcv_skb(sk, rskb)) { |
730 | skb_queue_head(&iucv_sk(sk)->backlog_skb_q, | 817 | skb_queue_head(&iucv->backlog_skb_q, |
731 | rskb); | 818 | rskb); |
732 | break; | 819 | break; |
733 | } else { | 820 | } else { |
734 | rskb = skb_dequeue(&iucv_sk(sk)->backlog_skb_q); | 821 | rskb = skb_dequeue(&iucv->backlog_skb_q); |
735 | } | 822 | } |
736 | } | 823 | } |
824 | if (skb_queue_empty(&iucv->backlog_skb_q)) { | ||
825 | spin_lock_bh(&iucv->message_q.lock); | ||
826 | if (!list_empty(&iucv->message_q.list)) | ||
827 | iucv_process_message_q(sk); | ||
828 | spin_unlock_bh(&iucv->message_q.lock); | ||
829 | } | ||
830 | |||
737 | } else | 831 | } else |
738 | skb_queue_head(&sk->sk_receive_queue, skb); | 832 | skb_queue_head(&sk->sk_receive_queue, skb); |
739 | 833 | ||
@@ -975,99 +1069,44 @@ static void iucv_callback_connack(struct iucv_path *path, u8 ipuser[16]) | |||
975 | sk->sk_state_change(sk); | 1069 | sk->sk_state_change(sk); |
976 | } | 1070 | } |
977 | 1071 | ||
978 | static int iucv_fragment_skb(struct sock *sk, struct sk_buff *skb, int len, | ||
979 | struct sk_buff_head *fragmented_skb_q) | ||
980 | { | ||
981 | int dataleft, size, copied = 0; | ||
982 | struct sk_buff *nskb; | ||
983 | |||
984 | dataleft = len; | ||
985 | while (dataleft) { | ||
986 | if (dataleft >= sk->sk_rcvbuf / 4) | ||
987 | size = sk->sk_rcvbuf / 4; | ||
988 | else | ||
989 | size = dataleft; | ||
990 | |||
991 | nskb = alloc_skb(size, GFP_ATOMIC | GFP_DMA); | ||
992 | if (!nskb) | ||
993 | return -ENOMEM; | ||
994 | |||
995 | memcpy(nskb->data, skb->data + copied, size); | ||
996 | copied += size; | ||
997 | dataleft -= size; | ||
998 | |||
999 | skb_reset_transport_header(nskb); | ||
1000 | skb_reset_network_header(nskb); | ||
1001 | nskb->len = size; | ||
1002 | |||
1003 | skb_queue_tail(fragmented_skb_q, nskb); | ||
1004 | } | ||
1005 | |||
1006 | return 0; | ||
1007 | } | ||
1008 | |||
1009 | static void iucv_callback_rx(struct iucv_path *path, struct iucv_message *msg) | 1072 | static void iucv_callback_rx(struct iucv_path *path, struct iucv_message *msg) |
1010 | { | 1073 | { |
1011 | struct sock *sk = path->private; | 1074 | struct sock *sk = path->private; |
1012 | struct iucv_sock *iucv = iucv_sk(sk); | 1075 | struct iucv_sock *iucv = iucv_sk(sk); |
1013 | struct sk_buff *skb, *fskb; | 1076 | struct sk_buff *skb; |
1014 | struct sk_buff_head fragmented_skb_q; | 1077 | struct sock_msg_q *save_msg; |
1015 | int rc; | 1078 | int len; |
1016 | |||
1017 | skb_queue_head_init(&fragmented_skb_q); | ||
1018 | 1079 | ||
1019 | if (sk->sk_shutdown & RCV_SHUTDOWN) | 1080 | if (sk->sk_shutdown & RCV_SHUTDOWN) |
1020 | return; | 1081 | return; |
1021 | 1082 | ||
1083 | if (!list_empty(&iucv->message_q.list) || | ||
1084 | !skb_queue_empty(&iucv->backlog_skb_q)) | ||
1085 | goto save_message; | ||
1086 | |||
1087 | len = atomic_read(&sk->sk_rmem_alloc); | ||
1088 | len += msg->length + sizeof(struct sk_buff); | ||
1089 | if (len > sk->sk_rcvbuf) | ||
1090 | goto save_message; | ||
1091 | |||
1022 | skb = alloc_skb(msg->length, GFP_ATOMIC | GFP_DMA); | 1092 | skb = alloc_skb(msg->length, GFP_ATOMIC | GFP_DMA); |
1023 | if (!skb) { | 1093 | if (!skb) |
1024 | iucv_path_sever(path, NULL); | 1094 | goto save_message; |
1025 | return; | ||
1026 | } | ||
1027 | 1095 | ||
1028 | if (msg->flags & IPRMDATA) { | 1096 | spin_lock(&iucv->message_q.lock); |
1029 | skb->data = NULL; | 1097 | iucv_process_message(sk, skb, path, msg); |
1030 | skb->len = 0; | 1098 | spin_unlock(&iucv->message_q.lock); |
1031 | } else { | ||
1032 | rc = iucv_message_receive(path, msg, 0, skb->data, | ||
1033 | msg->length, NULL); | ||
1034 | if (rc) { | ||
1035 | kfree_skb(skb); | ||
1036 | return; | ||
1037 | } | ||
1038 | if (skb->truesize >= sk->sk_rcvbuf / 4) { | ||
1039 | rc = iucv_fragment_skb(sk, skb, msg->length, | ||
1040 | &fragmented_skb_q); | ||
1041 | kfree_skb(skb); | ||
1042 | skb = NULL; | ||
1043 | if (rc) { | ||
1044 | iucv_path_sever(path, NULL); | ||
1045 | return; | ||
1046 | } | ||
1047 | } else { | ||
1048 | skb_reset_transport_header(skb); | ||
1049 | skb_reset_network_header(skb); | ||
1050 | skb->len = msg->length; | ||
1051 | } | ||
1052 | } | ||
1053 | /* Queue the fragmented skb */ | ||
1054 | fskb = skb_dequeue(&fragmented_skb_q); | ||
1055 | while (fskb) { | ||
1056 | if (!skb_queue_empty(&iucv->backlog_skb_q)) | ||
1057 | skb_queue_tail(&iucv->backlog_skb_q, fskb); | ||
1058 | else if (sock_queue_rcv_skb(sk, fskb)) | ||
1059 | skb_queue_tail(&iucv_sk(sk)->backlog_skb_q, fskb); | ||
1060 | fskb = skb_dequeue(&fragmented_skb_q); | ||
1061 | } | ||
1062 | 1099 | ||
1063 | /* Queue the original skb if it exists (was not fragmented) */ | 1100 | return; |
1064 | if (skb) { | 1101 | |
1065 | if (!skb_queue_empty(&iucv->backlog_skb_q)) | 1102 | save_message: |
1066 | skb_queue_tail(&iucv_sk(sk)->backlog_skb_q, skb); | 1103 | save_msg = kzalloc(sizeof(struct sock_msg_q), GFP_ATOMIC | GFP_DMA); |
1067 | else if (sock_queue_rcv_skb(sk, skb)) | 1104 | save_msg->path = path; |
1068 | skb_queue_tail(&iucv_sk(sk)->backlog_skb_q, skb); | 1105 | save_msg->msg = *msg; |
1069 | } | ||
1070 | 1106 | ||
1107 | spin_lock(&iucv->message_q.lock); | ||
1108 | list_add_tail(&save_msg->list, &iucv->message_q.list); | ||
1109 | spin_unlock(&iucv->message_q.lock); | ||
1071 | } | 1110 | } |
1072 | 1111 | ||
1073 | static void iucv_callback_txdone(struct iucv_path *path, | 1112 | static void iucv_callback_txdone(struct iucv_path *path, |