Lines Matching refs:nsk
801 struct sock *sk = sock->sk, *nsk; in iucv_sock_accept() local
816 while (!(nsk = iucv_accept_dequeue(sk, newsock))) { in iucv_sock_accept()
1592 struct sock *sk, *nsk; in iucv_callback_connreq() local
1637 nsk = iucv_sock_alloc(NULL, sk->sk_protocol, GFP_ATOMIC, 0); in iucv_callback_connreq()
1638 if (!nsk) { in iucv_callback_connreq()
1644 niucv = iucv_sk(nsk); in iucv_callback_connreq()
1645 iucv_sock_init(nsk, sk); in iucv_callback_connreq()
1647 nsk->sk_allocation |= GFP_DMA; in iucv_callback_connreq()
1665 err = pr_iucv->path_accept(path, &af_iucv_handler, nuser_data, nsk); in iucv_callback_connreq()
1667 iucv_sever_path(nsk, 1); in iucv_callback_connreq()
1668 iucv_sock_kill(nsk); in iucv_callback_connreq()
1672 iucv_accept_enqueue(sk, nsk); in iucv_callback_connreq()
1675 nsk->sk_state = IUCV_CONNECTED; in iucv_callback_connreq()
1846 struct sock *nsk; in afiucv_hs_callback_syn() local
1859 nsk = iucv_sock_alloc(NULL, sk->sk_protocol, GFP_ATOMIC, 0); in afiucv_hs_callback_syn()
1863 !nsk) { in afiucv_hs_callback_syn()
1868 iucv_sock_kill(nsk); in afiucv_hs_callback_syn()
1873 niucv = iucv_sk(nsk); in afiucv_hs_callback_syn()
1874 iucv_sock_init(nsk, sk); in afiucv_hs_callback_syn()
1885 nsk->sk_bound_dev_if = sk->sk_bound_dev_if; in afiucv_hs_callback_syn()
1894 iucv_accept_enqueue(sk, nsk); in afiucv_hs_callback_syn()
1895 nsk->sk_state = IUCV_CONNECTED; in afiucv_hs_callback_syn()
1898 iucv_sock_kill(nsk); in afiucv_hs_callback_syn()