diff options
Diffstat (limited to 'net/unix/af_unix.c')
-rw-r--r-- | net/unix/af_unix.c | 236 |
1 files changed, 150 insertions, 86 deletions
diff --git a/net/unix/af_unix.c b/net/unix/af_unix.c index f25511903115..dd419d286204 100644 --- a/net/unix/af_unix.c +++ b/net/unix/af_unix.c | |||
@@ -117,7 +117,7 @@ | |||
117 | 117 | ||
118 | static struct hlist_head unix_socket_table[UNIX_HASH_SIZE + 1]; | 118 | static struct hlist_head unix_socket_table[UNIX_HASH_SIZE + 1]; |
119 | static DEFINE_SPINLOCK(unix_table_lock); | 119 | static DEFINE_SPINLOCK(unix_table_lock); |
120 | static atomic_t unix_nr_socks = ATOMIC_INIT(0); | 120 | static atomic_long_t unix_nr_socks; |
121 | 121 | ||
122 | #define unix_sockets_unbound (&unix_socket_table[UNIX_HASH_SIZE]) | 122 | #define unix_sockets_unbound (&unix_socket_table[UNIX_HASH_SIZE]) |
123 | 123 | ||
@@ -144,7 +144,7 @@ static inline void unix_set_secdata(struct scm_cookie *scm, struct sk_buff *skb) | |||
144 | /* | 144 | /* |
145 | * SMP locking strategy: | 145 | * SMP locking strategy: |
146 | * hash table is protected with spinlock unix_table_lock | 146 | * hash table is protected with spinlock unix_table_lock |
147 | * each socket state is protected by separate rwlock. | 147 | * each socket state is protected by separate spin lock. |
148 | */ | 148 | */ |
149 | 149 | ||
150 | static inline unsigned unix_hash_fold(__wsum n) | 150 | static inline unsigned unix_hash_fold(__wsum n) |
@@ -282,7 +282,7 @@ static inline struct sock *unix_find_socket_byname(struct net *net, | |||
282 | return s; | 282 | return s; |
283 | } | 283 | } |
284 | 284 | ||
285 | static struct sock *unix_find_socket_byinode(struct net *net, struct inode *i) | 285 | static struct sock *unix_find_socket_byinode(struct inode *i) |
286 | { | 286 | { |
287 | struct sock *s; | 287 | struct sock *s; |
288 | struct hlist_node *node; | 288 | struct hlist_node *node; |
@@ -292,9 +292,6 @@ static struct sock *unix_find_socket_byinode(struct net *net, struct inode *i) | |||
292 | &unix_socket_table[i->i_ino & (UNIX_HASH_SIZE - 1)]) { | 292 | &unix_socket_table[i->i_ino & (UNIX_HASH_SIZE - 1)]) { |
293 | struct dentry *dentry = unix_sk(s)->dentry; | 293 | struct dentry *dentry = unix_sk(s)->dentry; |
294 | 294 | ||
295 | if (!net_eq(sock_net(s), net)) | ||
296 | continue; | ||
297 | |||
298 | if (dentry && dentry->d_inode == i) { | 295 | if (dentry && dentry->d_inode == i) { |
299 | sock_hold(s); | 296 | sock_hold(s); |
300 | goto found; | 297 | goto found; |
@@ -313,13 +310,17 @@ static inline int unix_writable(struct sock *sk) | |||
313 | 310 | ||
314 | static void unix_write_space(struct sock *sk) | 311 | static void unix_write_space(struct sock *sk) |
315 | { | 312 | { |
316 | read_lock(&sk->sk_callback_lock); | 313 | struct socket_wq *wq; |
314 | |||
315 | rcu_read_lock(); | ||
317 | if (unix_writable(sk)) { | 316 | if (unix_writable(sk)) { |
318 | if (sk_has_sleeper(sk)) | 317 | wq = rcu_dereference(sk->sk_wq); |
319 | wake_up_interruptible_sync(sk->sk_sleep); | 318 | if (wq_has_sleeper(wq)) |
319 | wake_up_interruptible_sync_poll(&wq->wait, | ||
320 | POLLOUT | POLLWRNORM | POLLWRBAND); | ||
320 | sk_wake_async(sk, SOCK_WAKE_SPACE, POLL_OUT); | 321 | sk_wake_async(sk, SOCK_WAKE_SPACE, POLL_OUT); |
321 | } | 322 | } |
322 | read_unlock(&sk->sk_callback_lock); | 323 | rcu_read_unlock(); |
323 | } | 324 | } |
324 | 325 | ||
325 | /* When dgram socket disconnects (or changes its peer), we clear its receive | 326 | /* When dgram socket disconnects (or changes its peer), we clear its receive |
@@ -360,13 +361,13 @@ static void unix_sock_destructor(struct sock *sk) | |||
360 | if (u->addr) | 361 | if (u->addr) |
361 | unix_release_addr(u->addr); | 362 | unix_release_addr(u->addr); |
362 | 363 | ||
363 | atomic_dec(&unix_nr_socks); | 364 | atomic_long_dec(&unix_nr_socks); |
364 | local_bh_disable(); | 365 | local_bh_disable(); |
365 | sock_prot_inuse_add(sock_net(sk), sk->sk_prot, -1); | 366 | sock_prot_inuse_add(sock_net(sk), sk->sk_prot, -1); |
366 | local_bh_enable(); | 367 | local_bh_enable(); |
367 | #ifdef UNIX_REFCNT_DEBUG | 368 | #ifdef UNIX_REFCNT_DEBUG |
368 | printk(KERN_DEBUG "UNIX %p is destroyed, %d are still alive.\n", sk, | 369 | printk(KERN_DEBUG "UNIX %p is destroyed, %ld are still alive.\n", sk, |
369 | atomic_read(&unix_nr_socks)); | 370 | atomic_long_read(&unix_nr_socks)); |
370 | #endif | 371 | #endif |
371 | } | 372 | } |
372 | 373 | ||
@@ -406,9 +407,7 @@ static int unix_release_sock(struct sock *sk, int embrion) | |||
406 | skpair->sk_err = ECONNRESET; | 407 | skpair->sk_err = ECONNRESET; |
407 | unix_state_unlock(skpair); | 408 | unix_state_unlock(skpair); |
408 | skpair->sk_state_change(skpair); | 409 | skpair->sk_state_change(skpair); |
409 | read_lock(&skpair->sk_callback_lock); | ||
410 | sk_wake_async(skpair, SOCK_WAKE_WAITD, POLL_HUP); | 410 | sk_wake_async(skpair, SOCK_WAKE_WAITD, POLL_HUP); |
411 | read_unlock(&skpair->sk_callback_lock); | ||
412 | } | 411 | } |
413 | sock_put(skpair); /* It may now die */ | 412 | sock_put(skpair); /* It may now die */ |
414 | unix_peer(sk) = NULL; | 413 | unix_peer(sk) = NULL; |
@@ -449,11 +448,31 @@ static int unix_release_sock(struct sock *sk, int embrion) | |||
449 | return 0; | 448 | return 0; |
450 | } | 449 | } |
451 | 450 | ||
451 | static void init_peercred(struct sock *sk) | ||
452 | { | ||
453 | put_pid(sk->sk_peer_pid); | ||
454 | if (sk->sk_peer_cred) | ||
455 | put_cred(sk->sk_peer_cred); | ||
456 | sk->sk_peer_pid = get_pid(task_tgid(current)); | ||
457 | sk->sk_peer_cred = get_current_cred(); | ||
458 | } | ||
459 | |||
460 | static void copy_peercred(struct sock *sk, struct sock *peersk) | ||
461 | { | ||
462 | put_pid(sk->sk_peer_pid); | ||
463 | if (sk->sk_peer_cred) | ||
464 | put_cred(sk->sk_peer_cred); | ||
465 | sk->sk_peer_pid = get_pid(peersk->sk_peer_pid); | ||
466 | sk->sk_peer_cred = get_cred(peersk->sk_peer_cred); | ||
467 | } | ||
468 | |||
452 | static int unix_listen(struct socket *sock, int backlog) | 469 | static int unix_listen(struct socket *sock, int backlog) |
453 | { | 470 | { |
454 | int err; | 471 | int err; |
455 | struct sock *sk = sock->sk; | 472 | struct sock *sk = sock->sk; |
456 | struct unix_sock *u = unix_sk(sk); | 473 | struct unix_sock *u = unix_sk(sk); |
474 | struct pid *old_pid = NULL; | ||
475 | const struct cred *old_cred = NULL; | ||
457 | 476 | ||
458 | err = -EOPNOTSUPP; | 477 | err = -EOPNOTSUPP; |
459 | if (sock->type != SOCK_STREAM && sock->type != SOCK_SEQPACKET) | 478 | if (sock->type != SOCK_STREAM && sock->type != SOCK_SEQPACKET) |
@@ -469,12 +488,14 @@ static int unix_listen(struct socket *sock, int backlog) | |||
469 | sk->sk_max_ack_backlog = backlog; | 488 | sk->sk_max_ack_backlog = backlog; |
470 | sk->sk_state = TCP_LISTEN; | 489 | sk->sk_state = TCP_LISTEN; |
471 | /* set credentials so connect can copy them */ | 490 | /* set credentials so connect can copy them */ |
472 | sk->sk_peercred.pid = task_tgid_vnr(current); | 491 | init_peercred(sk); |
473 | current_euid_egid(&sk->sk_peercred.uid, &sk->sk_peercred.gid); | ||
474 | err = 0; | 492 | err = 0; |
475 | 493 | ||
476 | out_unlock: | 494 | out_unlock: |
477 | unix_state_unlock(sk); | 495 | unix_state_unlock(sk); |
496 | put_pid(old_pid); | ||
497 | if (old_cred) | ||
498 | put_cred(old_cred); | ||
478 | out: | 499 | out: |
479 | return err; | 500 | return err; |
480 | } | 501 | } |
@@ -586,8 +607,8 @@ static struct sock *unix_create1(struct net *net, struct socket *sock) | |||
586 | struct sock *sk = NULL; | 607 | struct sock *sk = NULL; |
587 | struct unix_sock *u; | 608 | struct unix_sock *u; |
588 | 609 | ||
589 | atomic_inc(&unix_nr_socks); | 610 | atomic_long_inc(&unix_nr_socks); |
590 | if (atomic_read(&unix_nr_socks) > 2 * get_max_files()) | 611 | if (atomic_long_read(&unix_nr_socks) > 2 * get_max_files()) |
591 | goto out; | 612 | goto out; |
592 | 613 | ||
593 | sk = sk_alloc(net, PF_UNIX, GFP_KERNEL, &unix_proto); | 614 | sk = sk_alloc(net, PF_UNIX, GFP_KERNEL, &unix_proto); |
@@ -612,7 +633,7 @@ static struct sock *unix_create1(struct net *net, struct socket *sock) | |||
612 | unix_insert_socket(unix_sockets_unbound, sk); | 633 | unix_insert_socket(unix_sockets_unbound, sk); |
613 | out: | 634 | out: |
614 | if (sk == NULL) | 635 | if (sk == NULL) |
615 | atomic_dec(&unix_nr_socks); | 636 | atomic_long_dec(&unix_nr_socks); |
616 | else { | 637 | else { |
617 | local_bh_disable(); | 638 | local_bh_disable(); |
618 | sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1); | 639 | sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1); |
@@ -672,6 +693,7 @@ static int unix_autobind(struct socket *sock) | |||
672 | static u32 ordernum = 1; | 693 | static u32 ordernum = 1; |
673 | struct unix_address *addr; | 694 | struct unix_address *addr; |
674 | int err; | 695 | int err; |
696 | unsigned int retries = 0; | ||
675 | 697 | ||
676 | mutex_lock(&u->readlock); | 698 | mutex_lock(&u->readlock); |
677 | 699 | ||
@@ -697,9 +719,17 @@ retry: | |||
697 | if (__unix_find_socket_byname(net, addr->name, addr->len, sock->type, | 719 | if (__unix_find_socket_byname(net, addr->name, addr->len, sock->type, |
698 | addr->hash)) { | 720 | addr->hash)) { |
699 | spin_unlock(&unix_table_lock); | 721 | spin_unlock(&unix_table_lock); |
700 | /* Sanity yield. It is unusual case, but yet... */ | 722 | /* |
701 | if (!(ordernum&0xFF)) | 723 | * __unix_find_socket_byname() may take long time if many names |
702 | yield(); | 724 | * are already in use. |
725 | */ | ||
726 | cond_resched(); | ||
727 | /* Give up if all names seems to be in use. */ | ||
728 | if (retries++ == 0xFFFFF) { | ||
729 | err = -ENOSPC; | ||
730 | kfree(addr); | ||
731 | goto out; | ||
732 | } | ||
703 | goto retry; | 733 | goto retry; |
704 | } | 734 | } |
705 | addr->hash ^= sk->sk_type; | 735 | addr->hash ^= sk->sk_type; |
@@ -735,7 +765,7 @@ static struct sock *unix_find_other(struct net *net, | |||
735 | err = -ECONNREFUSED; | 765 | err = -ECONNREFUSED; |
736 | if (!S_ISSOCK(inode->i_mode)) | 766 | if (!S_ISSOCK(inode->i_mode)) |
737 | goto put_fail; | 767 | goto put_fail; |
738 | u = unix_find_socket_byinode(net, inode); | 768 | u = unix_find_socket_byinode(inode); |
739 | if (!u) | 769 | if (!u) |
740 | goto put_fail; | 770 | goto put_fail; |
741 | 771 | ||
@@ -1127,7 +1157,7 @@ restart: | |||
1127 | goto restart; | 1157 | goto restart; |
1128 | } | 1158 | } |
1129 | 1159 | ||
1130 | err = security_unix_stream_connect(sock, other->sk_socket, newsk); | 1160 | err = security_unix_stream_connect(sk, other, newsk); |
1131 | if (err) { | 1161 | if (err) { |
1132 | unix_state_unlock(sk); | 1162 | unix_state_unlock(sk); |
1133 | goto out_unlock; | 1163 | goto out_unlock; |
@@ -1139,10 +1169,9 @@ restart: | |||
1139 | unix_peer(newsk) = sk; | 1169 | unix_peer(newsk) = sk; |
1140 | newsk->sk_state = TCP_ESTABLISHED; | 1170 | newsk->sk_state = TCP_ESTABLISHED; |
1141 | newsk->sk_type = sk->sk_type; | 1171 | newsk->sk_type = sk->sk_type; |
1142 | newsk->sk_peercred.pid = task_tgid_vnr(current); | 1172 | init_peercred(newsk); |
1143 | current_euid_egid(&newsk->sk_peercred.uid, &newsk->sk_peercred.gid); | ||
1144 | newu = unix_sk(newsk); | 1173 | newu = unix_sk(newsk); |
1145 | newsk->sk_sleep = &newu->peer_wait; | 1174 | newsk->sk_wq = &newu->peer_wq; |
1146 | otheru = unix_sk(other); | 1175 | otheru = unix_sk(other); |
1147 | 1176 | ||
1148 | /* copy address information from listening to new sock*/ | 1177 | /* copy address information from listening to new sock*/ |
@@ -1156,7 +1185,7 @@ restart: | |||
1156 | } | 1185 | } |
1157 | 1186 | ||
1158 | /* Set credentials */ | 1187 | /* Set credentials */ |
1159 | sk->sk_peercred = other->sk_peercred; | 1188 | copy_peercred(sk, other); |
1160 | 1189 | ||
1161 | sock->state = SS_CONNECTED; | 1190 | sock->state = SS_CONNECTED; |
1162 | sk->sk_state = TCP_ESTABLISHED; | 1191 | sk->sk_state = TCP_ESTABLISHED; |
@@ -1198,10 +1227,8 @@ static int unix_socketpair(struct socket *socka, struct socket *sockb) | |||
1198 | sock_hold(skb); | 1227 | sock_hold(skb); |
1199 | unix_peer(ska) = skb; | 1228 | unix_peer(ska) = skb; |
1200 | unix_peer(skb) = ska; | 1229 | unix_peer(skb) = ska; |
1201 | ska->sk_peercred.pid = skb->sk_peercred.pid = task_tgid_vnr(current); | 1230 | init_peercred(ska); |
1202 | current_euid_egid(&skb->sk_peercred.uid, &skb->sk_peercred.gid); | 1231 | init_peercred(skb); |
1203 | ska->sk_peercred.uid = skb->sk_peercred.uid; | ||
1204 | ska->sk_peercred.gid = skb->sk_peercred.gid; | ||
1205 | 1232 | ||
1206 | if (ska->sk_type != SOCK_DGRAM) { | 1233 | if (ska->sk_type != SOCK_DGRAM) { |
1207 | ska->sk_state = TCP_ESTABLISHED; | 1234 | ska->sk_state = TCP_ESTABLISHED; |
@@ -1296,18 +1323,20 @@ static void unix_detach_fds(struct scm_cookie *scm, struct sk_buff *skb) | |||
1296 | int i; | 1323 | int i; |
1297 | 1324 | ||
1298 | scm->fp = UNIXCB(skb).fp; | 1325 | scm->fp = UNIXCB(skb).fp; |
1299 | skb->destructor = sock_wfree; | ||
1300 | UNIXCB(skb).fp = NULL; | 1326 | UNIXCB(skb).fp = NULL; |
1301 | 1327 | ||
1302 | for (i = scm->fp->count-1; i >= 0; i--) | 1328 | for (i = scm->fp->count-1; i >= 0; i--) |
1303 | unix_notinflight(scm->fp->fp[i]); | 1329 | unix_notinflight(scm->fp->fp[i]); |
1304 | } | 1330 | } |
1305 | 1331 | ||
1306 | static void unix_destruct_fds(struct sk_buff *skb) | 1332 | static void unix_destruct_scm(struct sk_buff *skb) |
1307 | { | 1333 | { |
1308 | struct scm_cookie scm; | 1334 | struct scm_cookie scm; |
1309 | memset(&scm, 0, sizeof(scm)); | 1335 | memset(&scm, 0, sizeof(scm)); |
1310 | unix_detach_fds(&scm, skb); | 1336 | scm.pid = UNIXCB(skb).pid; |
1337 | scm.cred = UNIXCB(skb).cred; | ||
1338 | if (UNIXCB(skb).fp) | ||
1339 | unix_detach_fds(&scm, skb); | ||
1311 | 1340 | ||
1312 | /* Alas, it calls VFS */ | 1341 | /* Alas, it calls VFS */ |
1313 | /* So fscking what? fput() had been SMP-safe since the last Summer */ | 1342 | /* So fscking what? fput() had been SMP-safe since the last Summer */ |
@@ -1315,9 +1344,25 @@ static void unix_destruct_fds(struct sk_buff *skb) | |||
1315 | sock_wfree(skb); | 1344 | sock_wfree(skb); |
1316 | } | 1345 | } |
1317 | 1346 | ||
1347 | #define MAX_RECURSION_LEVEL 4 | ||
1348 | |||
1318 | static int unix_attach_fds(struct scm_cookie *scm, struct sk_buff *skb) | 1349 | static int unix_attach_fds(struct scm_cookie *scm, struct sk_buff *skb) |
1319 | { | 1350 | { |
1320 | int i; | 1351 | int i; |
1352 | unsigned char max_level = 0; | ||
1353 | int unix_sock_count = 0; | ||
1354 | |||
1355 | for (i = scm->fp->count - 1; i >= 0; i--) { | ||
1356 | struct sock *sk = unix_get_socket(scm->fp->fp[i]); | ||
1357 | |||
1358 | if (sk) { | ||
1359 | unix_sock_count++; | ||
1360 | max_level = max(max_level, | ||
1361 | unix_sk(sk)->recursion_level); | ||
1362 | } | ||
1363 | } | ||
1364 | if (unlikely(max_level > MAX_RECURSION_LEVEL)) | ||
1365 | return -ETOOMANYREFS; | ||
1321 | 1366 | ||
1322 | /* | 1367 | /* |
1323 | * Need to duplicate file references for the sake of garbage | 1368 | * Need to duplicate file references for the sake of garbage |
@@ -1328,10 +1373,24 @@ static int unix_attach_fds(struct scm_cookie *scm, struct sk_buff *skb) | |||
1328 | if (!UNIXCB(skb).fp) | 1373 | if (!UNIXCB(skb).fp) |
1329 | return -ENOMEM; | 1374 | return -ENOMEM; |
1330 | 1375 | ||
1331 | for (i = scm->fp->count-1; i >= 0; i--) | 1376 | if (unix_sock_count) { |
1332 | unix_inflight(scm->fp->fp[i]); | 1377 | for (i = scm->fp->count - 1; i >= 0; i--) |
1333 | skb->destructor = unix_destruct_fds; | 1378 | unix_inflight(scm->fp->fp[i]); |
1334 | return 0; | 1379 | } |
1380 | return max_level; | ||
1381 | } | ||
1382 | |||
1383 | static int unix_scm_to_skb(struct scm_cookie *scm, struct sk_buff *skb, bool send_fds) | ||
1384 | { | ||
1385 | int err = 0; | ||
1386 | UNIXCB(skb).pid = get_pid(scm->pid); | ||
1387 | UNIXCB(skb).cred = get_cred(scm->cred); | ||
1388 | UNIXCB(skb).fp = NULL; | ||
1389 | if (scm->fp && send_fds) | ||
1390 | err = unix_attach_fds(scm, skb); | ||
1391 | |||
1392 | skb->destructor = unix_destruct_scm; | ||
1393 | return err; | ||
1335 | } | 1394 | } |
1336 | 1395 | ||
1337 | /* | 1396 | /* |
@@ -1353,6 +1412,7 @@ static int unix_dgram_sendmsg(struct kiocb *kiocb, struct socket *sock, | |||
1353 | struct sk_buff *skb; | 1412 | struct sk_buff *skb; |
1354 | long timeo; | 1413 | long timeo; |
1355 | struct scm_cookie tmp_scm; | 1414 | struct scm_cookie tmp_scm; |
1415 | int max_level; | ||
1356 | 1416 | ||
1357 | if (NULL == siocb->scm) | 1417 | if (NULL == siocb->scm) |
1358 | siocb->scm = &tmp_scm; | 1418 | siocb->scm = &tmp_scm; |
@@ -1390,12 +1450,10 @@ static int unix_dgram_sendmsg(struct kiocb *kiocb, struct socket *sock, | |||
1390 | if (skb == NULL) | 1450 | if (skb == NULL) |
1391 | goto out; | 1451 | goto out; |
1392 | 1452 | ||
1393 | memcpy(UNIXCREDS(skb), &siocb->scm->creds, sizeof(struct ucred)); | 1453 | err = unix_scm_to_skb(siocb->scm, skb, true); |
1394 | if (siocb->scm->fp) { | 1454 | if (err < 0) |
1395 | err = unix_attach_fds(siocb->scm, skb); | 1455 | goto out_free; |
1396 | if (err) | 1456 | max_level = err + 1; |
1397 | goto out_free; | ||
1398 | } | ||
1399 | unix_get_secdata(siocb->scm, skb); | 1457 | unix_get_secdata(siocb->scm, skb); |
1400 | 1458 | ||
1401 | skb_reset_transport_header(skb); | 1459 | skb_reset_transport_header(skb); |
@@ -1474,7 +1532,11 @@ restart: | |||
1474 | goto restart; | 1532 | goto restart; |
1475 | } | 1533 | } |
1476 | 1534 | ||
1535 | if (sock_flag(other, SOCK_RCVTSTAMP)) | ||
1536 | __net_timestamp(skb); | ||
1477 | skb_queue_tail(&other->sk_receive_queue, skb); | 1537 | skb_queue_tail(&other->sk_receive_queue, skb); |
1538 | if (max_level > unix_sk(other)->recursion_level) | ||
1539 | unix_sk(other)->recursion_level = max_level; | ||
1478 | unix_state_unlock(other); | 1540 | unix_state_unlock(other); |
1479 | other->sk_data_ready(other, len); | 1541 | other->sk_data_ready(other, len); |
1480 | sock_put(other); | 1542 | sock_put(other); |
@@ -1505,6 +1567,7 @@ static int unix_stream_sendmsg(struct kiocb *kiocb, struct socket *sock, | |||
1505 | int sent = 0; | 1567 | int sent = 0; |
1506 | struct scm_cookie tmp_scm; | 1568 | struct scm_cookie tmp_scm; |
1507 | bool fds_sent = false; | 1569 | bool fds_sent = false; |
1570 | int max_level; | ||
1508 | 1571 | ||
1509 | if (NULL == siocb->scm) | 1572 | if (NULL == siocb->scm) |
1510 | siocb->scm = &tmp_scm; | 1573 | siocb->scm = &tmp_scm; |
@@ -1565,16 +1628,15 @@ static int unix_stream_sendmsg(struct kiocb *kiocb, struct socket *sock, | |||
1565 | */ | 1628 | */ |
1566 | size = min_t(int, size, skb_tailroom(skb)); | 1629 | size = min_t(int, size, skb_tailroom(skb)); |
1567 | 1630 | ||
1568 | memcpy(UNIXCREDS(skb), &siocb->scm->creds, sizeof(struct ucred)); | 1631 | |
1569 | /* Only send the fds in the first buffer */ | 1632 | /* Only send the fds in the first buffer */ |
1570 | if (siocb->scm->fp && !fds_sent) { | 1633 | err = unix_scm_to_skb(siocb->scm, skb, !fds_sent); |
1571 | err = unix_attach_fds(siocb->scm, skb); | 1634 | if (err < 0) { |
1572 | if (err) { | 1635 | kfree_skb(skb); |
1573 | kfree_skb(skb); | 1636 | goto out_err; |
1574 | goto out_err; | ||
1575 | } | ||
1576 | fds_sent = true; | ||
1577 | } | 1637 | } |
1638 | max_level = err + 1; | ||
1639 | fds_sent = true; | ||
1578 | 1640 | ||
1579 | err = memcpy_fromiovec(skb_put(skb, size), msg->msg_iov, size); | 1641 | err = memcpy_fromiovec(skb_put(skb, size), msg->msg_iov, size); |
1580 | if (err) { | 1642 | if (err) { |
@@ -1589,6 +1651,8 @@ static int unix_stream_sendmsg(struct kiocb *kiocb, struct socket *sock, | |||
1589 | goto pipe_err_free; | 1651 | goto pipe_err_free; |
1590 | 1652 | ||
1591 | skb_queue_tail(&other->sk_receive_queue, skb); | 1653 | skb_queue_tail(&other->sk_receive_queue, skb); |
1654 | if (max_level > unix_sk(other)->recursion_level) | ||
1655 | unix_sk(other)->recursion_level = max_level; | ||
1592 | unix_state_unlock(other); | 1656 | unix_state_unlock(other); |
1593 | other->sk_data_ready(other, size); | 1657 | other->sk_data_ready(other, size); |
1594 | sent += size; | 1658 | sent += size; |
@@ -1673,7 +1737,8 @@ static int unix_dgram_recvmsg(struct kiocb *iocb, struct socket *sock, | |||
1673 | goto out_unlock; | 1737 | goto out_unlock; |
1674 | } | 1738 | } |
1675 | 1739 | ||
1676 | wake_up_interruptible_sync(&u->peer_wait); | 1740 | wake_up_interruptible_sync_poll(&u->peer_wait, |
1741 | POLLOUT | POLLWRNORM | POLLWRBAND); | ||
1677 | 1742 | ||
1678 | if (msg->msg_name) | 1743 | if (msg->msg_name) |
1679 | unix_copy_addr(msg, skb->sk); | 1744 | unix_copy_addr(msg, skb->sk); |
@@ -1687,11 +1752,14 @@ static int unix_dgram_recvmsg(struct kiocb *iocb, struct socket *sock, | |||
1687 | if (err) | 1752 | if (err) |
1688 | goto out_free; | 1753 | goto out_free; |
1689 | 1754 | ||
1755 | if (sock_flag(sk, SOCK_RCVTSTAMP)) | ||
1756 | __sock_recv_timestamp(msg, sk, skb); | ||
1757 | |||
1690 | if (!siocb->scm) { | 1758 | if (!siocb->scm) { |
1691 | siocb->scm = &tmp_scm; | 1759 | siocb->scm = &tmp_scm; |
1692 | memset(&tmp_scm, 0, sizeof(tmp_scm)); | 1760 | memset(&tmp_scm, 0, sizeof(tmp_scm)); |
1693 | } | 1761 | } |
1694 | siocb->scm->creds = *UNIXCREDS(skb); | 1762 | scm_set_cred(siocb->scm, UNIXCB(skb).pid, UNIXCB(skb).cred); |
1695 | unix_set_secdata(siocb->scm, skb); | 1763 | unix_set_secdata(siocb->scm, skb); |
1696 | 1764 | ||
1697 | if (!(flags & MSG_PEEK)) { | 1765 | if (!(flags & MSG_PEEK)) { |
@@ -1736,7 +1804,7 @@ static long unix_stream_data_wait(struct sock *sk, long timeo) | |||
1736 | unix_state_lock(sk); | 1804 | unix_state_lock(sk); |
1737 | 1805 | ||
1738 | for (;;) { | 1806 | for (;;) { |
1739 | prepare_to_wait(sk->sk_sleep, &wait, TASK_INTERRUPTIBLE); | 1807 | prepare_to_wait(sk_sleep(sk), &wait, TASK_INTERRUPTIBLE); |
1740 | 1808 | ||
1741 | if (!skb_queue_empty(&sk->sk_receive_queue) || | 1809 | if (!skb_queue_empty(&sk->sk_receive_queue) || |
1742 | sk->sk_err || | 1810 | sk->sk_err || |
@@ -1752,7 +1820,7 @@ static long unix_stream_data_wait(struct sock *sk, long timeo) | |||
1752 | clear_bit(SOCK_ASYNC_WAITDATA, &sk->sk_socket->flags); | 1820 | clear_bit(SOCK_ASYNC_WAITDATA, &sk->sk_socket->flags); |
1753 | } | 1821 | } |
1754 | 1822 | ||
1755 | finish_wait(sk->sk_sleep, &wait); | 1823 | finish_wait(sk_sleep(sk), &wait); |
1756 | unix_state_unlock(sk); | 1824 | unix_state_unlock(sk); |
1757 | return timeo; | 1825 | return timeo; |
1758 | } | 1826 | } |
@@ -1805,6 +1873,7 @@ static int unix_stream_recvmsg(struct kiocb *iocb, struct socket *sock, | |||
1805 | unix_state_lock(sk); | 1873 | unix_state_lock(sk); |
1806 | skb = skb_dequeue(&sk->sk_receive_queue); | 1874 | skb = skb_dequeue(&sk->sk_receive_queue); |
1807 | if (skb == NULL) { | 1875 | if (skb == NULL) { |
1876 | unix_sk(sk)->recursion_level = 0; | ||
1808 | if (copied >= target) | 1877 | if (copied >= target) |
1809 | goto unlock; | 1878 | goto unlock; |
1810 | 1879 | ||
@@ -1840,14 +1909,14 @@ static int unix_stream_recvmsg(struct kiocb *iocb, struct socket *sock, | |||
1840 | 1909 | ||
1841 | if (check_creds) { | 1910 | if (check_creds) { |
1842 | /* Never glue messages from different writers */ | 1911 | /* Never glue messages from different writers */ |
1843 | if (memcmp(UNIXCREDS(skb), &siocb->scm->creds, | 1912 | if ((UNIXCB(skb).pid != siocb->scm->pid) || |
1844 | sizeof(siocb->scm->creds)) != 0) { | 1913 | (UNIXCB(skb).cred != siocb->scm->cred)) { |
1845 | skb_queue_head(&sk->sk_receive_queue, skb); | 1914 | skb_queue_head(&sk->sk_receive_queue, skb); |
1846 | break; | 1915 | break; |
1847 | } | 1916 | } |
1848 | } else { | 1917 | } else { |
1849 | /* Copy credentials */ | 1918 | /* Copy credentials */ |
1850 | siocb->scm->creds = *UNIXCREDS(skb); | 1919 | scm_set_cred(siocb->scm, UNIXCB(skb).pid, UNIXCB(skb).cred); |
1851 | check_creds = 1; | 1920 | check_creds = 1; |
1852 | } | 1921 | } |
1853 | 1922 | ||
@@ -1880,7 +1949,7 @@ static int unix_stream_recvmsg(struct kiocb *iocb, struct socket *sock, | |||
1880 | break; | 1949 | break; |
1881 | } | 1950 | } |
1882 | 1951 | ||
1883 | kfree_skb(skb); | 1952 | consume_skb(skb); |
1884 | 1953 | ||
1885 | if (siocb->scm->fp) | 1954 | if (siocb->scm->fp) |
1886 | break; | 1955 | break; |
@@ -1931,12 +2000,10 @@ static int unix_shutdown(struct socket *sock, int mode) | |||
1931 | other->sk_shutdown |= peer_mode; | 2000 | other->sk_shutdown |= peer_mode; |
1932 | unix_state_unlock(other); | 2001 | unix_state_unlock(other); |
1933 | other->sk_state_change(other); | 2002 | other->sk_state_change(other); |
1934 | read_lock(&other->sk_callback_lock); | ||
1935 | if (peer_mode == SHUTDOWN_MASK) | 2003 | if (peer_mode == SHUTDOWN_MASK) |
1936 | sk_wake_async(other, SOCK_WAKE_WAITD, POLL_HUP); | 2004 | sk_wake_async(other, SOCK_WAKE_WAITD, POLL_HUP); |
1937 | else if (peer_mode & RCV_SHUTDOWN) | 2005 | else if (peer_mode & RCV_SHUTDOWN) |
1938 | sk_wake_async(other, SOCK_WAKE_WAITD, POLL_IN); | 2006 | sk_wake_async(other, SOCK_WAKE_WAITD, POLL_IN); |
1939 | read_unlock(&other->sk_callback_lock); | ||
1940 | } | 2007 | } |
1941 | if (other) | 2008 | if (other) |
1942 | sock_put(other); | 2009 | sock_put(other); |
@@ -1991,7 +2058,7 @@ static unsigned int unix_poll(struct file *file, struct socket *sock, poll_table | |||
1991 | struct sock *sk = sock->sk; | 2058 | struct sock *sk = sock->sk; |
1992 | unsigned int mask; | 2059 | unsigned int mask; |
1993 | 2060 | ||
1994 | sock_poll_wait(file, sk->sk_sleep, wait); | 2061 | sock_poll_wait(file, sk_sleep(sk), wait); |
1995 | mask = 0; | 2062 | mask = 0; |
1996 | 2063 | ||
1997 | /* exceptional events? */ | 2064 | /* exceptional events? */ |
@@ -2000,11 +2067,10 @@ static unsigned int unix_poll(struct file *file, struct socket *sock, poll_table | |||
2000 | if (sk->sk_shutdown == SHUTDOWN_MASK) | 2067 | if (sk->sk_shutdown == SHUTDOWN_MASK) |
2001 | mask |= POLLHUP; | 2068 | mask |= POLLHUP; |
2002 | if (sk->sk_shutdown & RCV_SHUTDOWN) | 2069 | if (sk->sk_shutdown & RCV_SHUTDOWN) |
2003 | mask |= POLLRDHUP; | 2070 | mask |= POLLRDHUP | POLLIN | POLLRDNORM; |
2004 | 2071 | ||
2005 | /* readable? */ | 2072 | /* readable? */ |
2006 | if (!skb_queue_empty(&sk->sk_receive_queue) || | 2073 | if (!skb_queue_empty(&sk->sk_receive_queue)) |
2007 | (sk->sk_shutdown & RCV_SHUTDOWN)) | ||
2008 | mask |= POLLIN | POLLRDNORM; | 2074 | mask |= POLLIN | POLLRDNORM; |
2009 | 2075 | ||
2010 | /* Connection-based need to check for termination and startup */ | 2076 | /* Connection-based need to check for termination and startup */ |
@@ -2028,20 +2094,19 @@ static unsigned int unix_dgram_poll(struct file *file, struct socket *sock, | |||
2028 | struct sock *sk = sock->sk, *other; | 2094 | struct sock *sk = sock->sk, *other; |
2029 | unsigned int mask, writable; | 2095 | unsigned int mask, writable; |
2030 | 2096 | ||
2031 | sock_poll_wait(file, sk->sk_sleep, wait); | 2097 | sock_poll_wait(file, sk_sleep(sk), wait); |
2032 | mask = 0; | 2098 | mask = 0; |
2033 | 2099 | ||
2034 | /* exceptional events? */ | 2100 | /* exceptional events? */ |
2035 | if (sk->sk_err || !skb_queue_empty(&sk->sk_error_queue)) | 2101 | if (sk->sk_err || !skb_queue_empty(&sk->sk_error_queue)) |
2036 | mask |= POLLERR; | 2102 | mask |= POLLERR; |
2037 | if (sk->sk_shutdown & RCV_SHUTDOWN) | 2103 | if (sk->sk_shutdown & RCV_SHUTDOWN) |
2038 | mask |= POLLRDHUP; | 2104 | mask |= POLLRDHUP | POLLIN | POLLRDNORM; |
2039 | if (sk->sk_shutdown == SHUTDOWN_MASK) | 2105 | if (sk->sk_shutdown == SHUTDOWN_MASK) |
2040 | mask |= POLLHUP; | 2106 | mask |= POLLHUP; |
2041 | 2107 | ||
2042 | /* readable? */ | 2108 | /* readable? */ |
2043 | if (!skb_queue_empty(&sk->sk_receive_queue) || | 2109 | if (!skb_queue_empty(&sk->sk_receive_queue)) |
2044 | (sk->sk_shutdown & RCV_SHUTDOWN)) | ||
2045 | mask |= POLLIN | POLLRDNORM; | 2110 | mask |= POLLIN | POLLRDNORM; |
2046 | 2111 | ||
2047 | /* Connection-based need to check for termination and startup */ | 2112 | /* Connection-based need to check for termination and startup */ |
@@ -2053,20 +2118,19 @@ static unsigned int unix_dgram_poll(struct file *file, struct socket *sock, | |||
2053 | return mask; | 2118 | return mask; |
2054 | } | 2119 | } |
2055 | 2120 | ||
2056 | /* writable? */ | 2121 | /* No write status requested, avoid expensive OUT tests. */ |
2057 | writable = unix_writable(sk); | 2122 | if (wait && !(wait->key & (POLLWRBAND | POLLWRNORM | POLLOUT))) |
2058 | if (writable) { | 2123 | return mask; |
2059 | other = unix_peer_get(sk); | ||
2060 | if (other) { | ||
2061 | if (unix_peer(other) != sk) { | ||
2062 | sock_poll_wait(file, &unix_sk(other)->peer_wait, | ||
2063 | wait); | ||
2064 | if (unix_recvq_full(other)) | ||
2065 | writable = 0; | ||
2066 | } | ||
2067 | 2124 | ||
2068 | sock_put(other); | 2125 | writable = unix_writable(sk); |
2126 | other = unix_peer_get(sk); | ||
2127 | if (other) { | ||
2128 | if (unix_peer(other) != sk) { | ||
2129 | sock_poll_wait(file, &unix_sk(other)->peer_wait, wait); | ||
2130 | if (unix_recvq_full(other)) | ||
2131 | writable = 0; | ||
2069 | } | 2132 | } |
2133 | sock_put(other); | ||
2070 | } | 2134 | } |
2071 | 2135 | ||
2072 | if (writable) | 2136 | if (writable) |
@@ -2224,7 +2288,7 @@ static const struct net_proto_family unix_family_ops = { | |||
2224 | }; | 2288 | }; |
2225 | 2289 | ||
2226 | 2290 | ||
2227 | static int unix_net_init(struct net *net) | 2291 | static int __net_init unix_net_init(struct net *net) |
2228 | { | 2292 | { |
2229 | int error = -ENOMEM; | 2293 | int error = -ENOMEM; |
2230 | 2294 | ||
@@ -2243,7 +2307,7 @@ out: | |||
2243 | return error; | 2307 | return error; |
2244 | } | 2308 | } |
2245 | 2309 | ||
2246 | static void unix_net_exit(struct net *net) | 2310 | static void __net_exit unix_net_exit(struct net *net) |
2247 | { | 2311 | { |
2248 | unix_sysctl_unregister(net); | 2312 | unix_sysctl_unregister(net); |
2249 | proc_net_remove(net, "unix"); | 2313 | proc_net_remove(net, "unix"); |