1 /* SPDX-License-Identifier: GPL-2.0-only */ 2 /* 3 * VMware vSockets Driver 4 * 5 * Copyright (C) 2007-2013 VMware, Inc. All rights reserved. 6 */ 7 8 #ifndef __AF_VSOCK_H__ 9 #define __AF_VSOCK_H__ 10 11 #include <linux/kernel.h> 12 #include <linux/workqueue.h> 13 #include <uapi/linux/vm_sockets.h> 14 15 #include "vsock_addr.h" 16 17 #define LAST_RESERVED_PORT 1023 18 19 #define VSOCK_HASH_SIZE 251 20 extern struct list_head vsock_bind_table[VSOCK_HASH_SIZE + 1]; 21 extern struct list_head vsock_connected_table[VSOCK_HASH_SIZE]; 22 extern spinlock_t vsock_table_lock; 23 24 #define vsock_sk(__sk) ((struct vsock_sock *)__sk) 25 #define sk_vsock(__vsk) (&(__vsk)->sk) 26 27 struct vsock_sock { 28 /* sk must be the first member. */ 29 struct sock sk; 30 const struct vsock_transport *transport; 31 struct sockaddr_vm local_addr; 32 struct sockaddr_vm remote_addr; 33 /* Links for the global tables of bound and connected sockets. */ 34 struct list_head bound_table; 35 struct list_head connected_table; 36 /* Accessed without the socket lock held. This means it can never be 37 * modified outsided of socket create or destruct. 38 */ 39 bool trusted; 40 bool cached_peer_allow_dgram; /* Dgram communication allowed to 41 * cached peer? 42 */ 43 u32 cached_peer; /* Context ID of last dgram destination check. */ 44 const struct cred *owner; 45 /* Rest are SOCK_STREAM only. */ 46 long connect_timeout; 47 /* Listening socket that this came from. */ 48 struct sock *listener; 49 /* Used for pending list and accept queue during connection handshake. 50 * The listening socket is the head for both lists. Sockets created 51 * for connection requests are placed in the pending list until they 52 * are connected, at which point they are put in the accept queue list 53 * so they can be accepted in accept(). If accept() cannot accept the 54 * connection, it is marked as rejected so the cleanup function knows 55 * to clean up the socket. 56 */ 57 struct list_head pending_links; 58 struct list_head accept_queue; 59 bool rejected; 60 struct delayed_work connect_work; 61 struct delayed_work pending_work; 62 struct delayed_work close_work; 63 bool close_work_scheduled; 64 u32 peer_shutdown; 65 bool sent_request; 66 bool ignore_connecting_rst; 67 68 /* Protected by lock_sock(sk) */ 69 u64 buffer_size; 70 u64 buffer_min_size; 71 u64 buffer_max_size; 72 73 /* Private to transport. */ 74 void *trans; 75 }; 76 77 s64 vsock_stream_has_data(struct vsock_sock *vsk); 78 s64 vsock_stream_has_space(struct vsock_sock *vsk); 79 struct sock *vsock_create_connected(struct sock *parent); 80 81 /**** TRANSPORT ****/ 82 83 struct vsock_transport_recv_notify_data { 84 u64 data1; /* Transport-defined. */ 85 u64 data2; /* Transport-defined. */ 86 bool notify_on_block; 87 }; 88 89 struct vsock_transport_send_notify_data { 90 u64 data1; /* Transport-defined. */ 91 u64 data2; /* Transport-defined. */ 92 }; 93 94 struct vsock_transport { 95 /* Initialize/tear-down socket. */ 96 int (*init)(struct vsock_sock *, struct vsock_sock *); 97 void (*destruct)(struct vsock_sock *); 98 void (*release)(struct vsock_sock *); 99 100 /* Cancel all pending packets sent on vsock. */ 101 int (*cancel_pkt)(struct vsock_sock *vsk); 102 103 /* Connections. */ 104 int (*connect)(struct vsock_sock *); 105 106 /* DGRAM. */ 107 int (*dgram_bind)(struct vsock_sock *, struct sockaddr_vm *); 108 int (*dgram_dequeue)(struct vsock_sock *vsk, struct msghdr *msg, 109 size_t len, int flags); 110 int (*dgram_enqueue)(struct vsock_sock *, struct sockaddr_vm *, 111 struct msghdr *, size_t len); 112 bool (*dgram_allow)(u32 cid, u32 port); 113 114 /* STREAM. */ 115 /* TODO: stream_bind() */ 116 ssize_t (*stream_dequeue)(struct vsock_sock *, struct msghdr *, 117 size_t len, int flags); 118 ssize_t (*stream_enqueue)(struct vsock_sock *, struct msghdr *, 119 size_t len); 120 s64 (*stream_has_data)(struct vsock_sock *); 121 s64 (*stream_has_space)(struct vsock_sock *); 122 u64 (*stream_rcvhiwat)(struct vsock_sock *); 123 bool (*stream_is_active)(struct vsock_sock *); 124 bool (*stream_allow)(u32 cid, u32 port); 125 126 /* Notification. */ 127 int (*notify_poll_in)(struct vsock_sock *, size_t, bool *); 128 int (*notify_poll_out)(struct vsock_sock *, size_t, bool *); 129 int (*notify_recv_init)(struct vsock_sock *, size_t, 130 struct vsock_transport_recv_notify_data *); 131 int (*notify_recv_pre_block)(struct vsock_sock *, size_t, 132 struct vsock_transport_recv_notify_data *); 133 int (*notify_recv_pre_dequeue)(struct vsock_sock *, size_t, 134 struct vsock_transport_recv_notify_data *); 135 int (*notify_recv_post_dequeue)(struct vsock_sock *, size_t, 136 ssize_t, bool, struct vsock_transport_recv_notify_data *); 137 int (*notify_send_init)(struct vsock_sock *, 138 struct vsock_transport_send_notify_data *); 139 int (*notify_send_pre_block)(struct vsock_sock *, 140 struct vsock_transport_send_notify_data *); 141 int (*notify_send_pre_enqueue)(struct vsock_sock *, 142 struct vsock_transport_send_notify_data *); 143 int (*notify_send_post_enqueue)(struct vsock_sock *, ssize_t, 144 struct vsock_transport_send_notify_data *); 145 /* sk_lock held by the caller */ 146 void (*notify_buffer_size)(struct vsock_sock *, u64 *); 147 148 /* Shutdown. */ 149 int (*shutdown)(struct vsock_sock *, int); 150 151 /* Addressing. */ 152 u32 (*get_local_cid)(void); 153 }; 154 155 /**** CORE ****/ 156 157 int __vsock_core_init(const struct vsock_transport *t, struct module *owner); 158 static inline int vsock_core_init(const struct vsock_transport *t) 159 { 160 return __vsock_core_init(t, THIS_MODULE); 161 } 162 void vsock_core_exit(void); 163 164 /* The transport may downcast this to access transport-specific functions */ 165 const struct vsock_transport *vsock_core_get_transport(struct vsock_sock *vsk); 166 167 /**** UTILS ****/ 168 169 /* vsock_table_lock must be held */ 170 static inline bool __vsock_in_bound_table(struct vsock_sock *vsk) 171 { 172 return !list_empty(&vsk->bound_table); 173 } 174 175 /* vsock_table_lock must be held */ 176 static inline bool __vsock_in_connected_table(struct vsock_sock *vsk) 177 { 178 return !list_empty(&vsk->connected_table); 179 } 180 181 void vsock_release_pending(struct sock *pending); 182 void vsock_add_pending(struct sock *listener, struct sock *pending); 183 void vsock_remove_pending(struct sock *listener, struct sock *pending); 184 void vsock_enqueue_accept(struct sock *listener, struct sock *connected); 185 void vsock_insert_connected(struct vsock_sock *vsk); 186 void vsock_remove_bound(struct vsock_sock *vsk); 187 void vsock_remove_connected(struct vsock_sock *vsk); 188 struct sock *vsock_find_bound_socket(struct sockaddr_vm *addr); 189 struct sock *vsock_find_connected_socket(struct sockaddr_vm *src, 190 struct sockaddr_vm *dst); 191 void vsock_remove_sock(struct vsock_sock *vsk); 192 void vsock_for_each_connected_socket(void (*fn)(struct sock *sk)); 193 194 /**** TAP ****/ 195 196 struct vsock_tap { 197 struct net_device *dev; 198 struct module *module; 199 struct list_head list; 200 }; 201 202 int vsock_init_tap(void); 203 int vsock_add_tap(struct vsock_tap *vt); 204 int vsock_remove_tap(struct vsock_tap *vt); 205 void vsock_deliver_tap(struct sk_buff *build_skb(void *opaque), void *opaque); 206 207 #endif /* __AF_VSOCK_H__ */ 208