267 lines
		
	
	
		
			7.4 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			267 lines
		
	
	
		
			7.4 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /* SPDX-License-Identifier: GPL-2.0 */
 | |
| #ifndef _LINUX_VIRTIO_VSOCK_H
 | |
| #define _LINUX_VIRTIO_VSOCK_H
 | |
| 
 | |
| #include <uapi/linux/virtio_vsock.h>
 | |
| #include <linux/socket.h>
 | |
| #include <net/sock.h>
 | |
| #include <net/af_vsock.h>
 | |
| 
 | |
| #define VIRTIO_VSOCK_SKB_HEADROOM (sizeof(struct virtio_vsock_hdr))
 | |
| 
 | |
| struct virtio_vsock_skb_cb {
 | |
| 	bool reply;
 | |
| 	bool tap_delivered;
 | |
| 	u32 offset;
 | |
| };
 | |
| 
 | |
| #define VIRTIO_VSOCK_SKB_CB(skb) ((struct virtio_vsock_skb_cb *)((skb)->cb))
 | |
| 
 | |
| static inline struct virtio_vsock_hdr *virtio_vsock_hdr(struct sk_buff *skb)
 | |
| {
 | |
| 	return (struct virtio_vsock_hdr *)skb->head;
 | |
| }
 | |
| 
 | |
| static inline bool virtio_vsock_skb_reply(struct sk_buff *skb)
 | |
| {
 | |
| 	return VIRTIO_VSOCK_SKB_CB(skb)->reply;
 | |
| }
 | |
| 
 | |
| static inline void virtio_vsock_skb_set_reply(struct sk_buff *skb)
 | |
| {
 | |
| 	VIRTIO_VSOCK_SKB_CB(skb)->reply = true;
 | |
| }
 | |
| 
 | |
| static inline bool virtio_vsock_skb_tap_delivered(struct sk_buff *skb)
 | |
| {
 | |
| 	return VIRTIO_VSOCK_SKB_CB(skb)->tap_delivered;
 | |
| }
 | |
| 
 | |
| static inline void virtio_vsock_skb_set_tap_delivered(struct sk_buff *skb)
 | |
| {
 | |
| 	VIRTIO_VSOCK_SKB_CB(skb)->tap_delivered = true;
 | |
| }
 | |
| 
 | |
| static inline void virtio_vsock_skb_clear_tap_delivered(struct sk_buff *skb)
 | |
| {
 | |
| 	VIRTIO_VSOCK_SKB_CB(skb)->tap_delivered = false;
 | |
| }
 | |
| 
 | |
| static inline void virtio_vsock_skb_rx_put(struct sk_buff *skb)
 | |
| {
 | |
| 	u32 len;
 | |
| 
 | |
| 	len = le32_to_cpu(virtio_vsock_hdr(skb)->len);
 | |
| 
 | |
| 	if (len > 0)
 | |
| 		skb_put(skb, len);
 | |
| }
 | |
| 
 | |
| static inline struct sk_buff *virtio_vsock_alloc_skb(unsigned int size, gfp_t mask)
 | |
| {
 | |
| 	struct sk_buff *skb;
 | |
| 
 | |
| 	if (size < VIRTIO_VSOCK_SKB_HEADROOM)
 | |
| 		return NULL;
 | |
| 
 | |
| 	skb = alloc_skb(size, mask);
 | |
| 	if (!skb)
 | |
| 		return NULL;
 | |
| 
 | |
| 	skb_reserve(skb, VIRTIO_VSOCK_SKB_HEADROOM);
 | |
| 	return skb;
 | |
| }
 | |
| 
 | |
| static inline void
 | |
| virtio_vsock_skb_queue_head(struct sk_buff_head *list, struct sk_buff *skb)
 | |
| {
 | |
| 	spin_lock_bh(&list->lock);
 | |
| 	__skb_queue_head(list, skb);
 | |
| 	spin_unlock_bh(&list->lock);
 | |
| }
 | |
| 
 | |
| static inline void
 | |
| virtio_vsock_skb_queue_tail(struct sk_buff_head *list, struct sk_buff *skb)
 | |
| {
 | |
| 	spin_lock_bh(&list->lock);
 | |
| 	__skb_queue_tail(list, skb);
 | |
| 	spin_unlock_bh(&list->lock);
 | |
| }
 | |
| 
 | |
| static inline struct sk_buff *virtio_vsock_skb_dequeue(struct sk_buff_head *list)
 | |
| {
 | |
| 	struct sk_buff *skb;
 | |
| 
 | |
| 	spin_lock_bh(&list->lock);
 | |
| 	skb = __skb_dequeue(list);
 | |
| 	spin_unlock_bh(&list->lock);
 | |
| 
 | |
| 	return skb;
 | |
| }
 | |
| 
 | |
| static inline void virtio_vsock_skb_queue_purge(struct sk_buff_head *list)
 | |
| {
 | |
| 	spin_lock_bh(&list->lock);
 | |
| 	__skb_queue_purge(list);
 | |
| 	spin_unlock_bh(&list->lock);
 | |
| }
 | |
| 
 | |
| static inline size_t virtio_vsock_skb_len(struct sk_buff *skb)
 | |
| {
 | |
| 	return (size_t)(skb_end_pointer(skb) - skb->head);
 | |
| }
 | |
| 
 | |
| #define VIRTIO_VSOCK_DEFAULT_RX_BUF_SIZE	(1024 * 4)
 | |
| #define VIRTIO_VSOCK_MAX_BUF_SIZE		0xFFFFFFFFUL
 | |
| #define VIRTIO_VSOCK_MAX_PKT_BUF_SIZE		(1024 * 64)
 | |
| 
 | |
| enum {
 | |
| 	VSOCK_VQ_RX     = 0, /* for host to guest data */
 | |
| 	VSOCK_VQ_TX     = 1, /* for guest to host data */
 | |
| 	VSOCK_VQ_EVENT  = 2,
 | |
| 	VSOCK_VQ_MAX    = 3,
 | |
| };
 | |
| 
 | |
| /* Per-socket state (accessed via vsk->trans) */
 | |
| struct virtio_vsock_sock {
 | |
| 	struct vsock_sock *vsk;
 | |
| 
 | |
| 	spinlock_t tx_lock;
 | |
| 	spinlock_t rx_lock;
 | |
| 
 | |
| 	/* Protected by tx_lock */
 | |
| 	u32 tx_cnt;
 | |
| 	u32 peer_fwd_cnt;
 | |
| 	u32 peer_buf_alloc;
 | |
| 	size_t bytes_unsent;
 | |
| 
 | |
| 	/* Protected by rx_lock */
 | |
| 	u32 fwd_cnt;
 | |
| 	u32 last_fwd_cnt;
 | |
| 	u32 rx_bytes;
 | |
| 	u32 buf_alloc;
 | |
| 	struct sk_buff_head rx_queue;
 | |
| 	u32 msg_count;
 | |
| };
 | |
| 
 | |
| struct virtio_vsock_pkt_info {
 | |
| 	u32 remote_cid, remote_port;
 | |
| 	struct vsock_sock *vsk;
 | |
| 	struct msghdr *msg;
 | |
| 	u32 pkt_len;
 | |
| 	u16 type;
 | |
| 	u16 op;
 | |
| 	u32 flags;
 | |
| 	bool reply;
 | |
| };
 | |
| 
 | |
| struct virtio_transport {
 | |
| 	/* This must be the first field */
 | |
| 	struct vsock_transport transport;
 | |
| 
 | |
| 	/* Takes ownership of the packet */
 | |
| 	int (*send_pkt)(struct sk_buff *skb);
 | |
| 
 | |
| 	/* Used in MSG_ZEROCOPY mode. Checks, that provided data
 | |
| 	 * (number of buffers) could be transmitted with zerocopy
 | |
| 	 * mode. If this callback is not implemented for the current
 | |
| 	 * transport - this means that this transport doesn't need
 | |
| 	 * extra checks and can perform zerocopy transmission by
 | |
| 	 * default.
 | |
| 	 */
 | |
| 	bool (*can_msgzerocopy)(int bufs_num);
 | |
| };
 | |
| 
 | |
| ssize_t
 | |
| virtio_transport_stream_dequeue(struct vsock_sock *vsk,
 | |
| 				struct msghdr *msg,
 | |
| 				size_t len,
 | |
| 				int type);
 | |
| int
 | |
| virtio_transport_dgram_dequeue(struct vsock_sock *vsk,
 | |
| 			       struct msghdr *msg,
 | |
| 			       size_t len, int flags);
 | |
| 
 | |
| int
 | |
| virtio_transport_seqpacket_enqueue(struct vsock_sock *vsk,
 | |
| 				   struct msghdr *msg,
 | |
| 				   size_t len);
 | |
| ssize_t
 | |
| virtio_transport_seqpacket_dequeue(struct vsock_sock *vsk,
 | |
| 				   struct msghdr *msg,
 | |
| 				   int flags);
 | |
| s64 virtio_transport_stream_has_data(struct vsock_sock *vsk);
 | |
| s64 virtio_transport_stream_has_space(struct vsock_sock *vsk);
 | |
| u32 virtio_transport_seqpacket_has_data(struct vsock_sock *vsk);
 | |
| 
 | |
| ssize_t virtio_transport_unsent_bytes(struct vsock_sock *vsk);
 | |
| 
 | |
| void virtio_transport_consume_skb_sent(struct sk_buff *skb,
 | |
| 				       bool consume);
 | |
| 
 | |
| int virtio_transport_do_socket_init(struct vsock_sock *vsk,
 | |
| 				 struct vsock_sock *psk);
 | |
| int
 | |
| virtio_transport_notify_poll_in(struct vsock_sock *vsk,
 | |
| 				size_t target,
 | |
| 				bool *data_ready_now);
 | |
| int
 | |
| virtio_transport_notify_poll_out(struct vsock_sock *vsk,
 | |
| 				 size_t target,
 | |
| 				 bool *space_available_now);
 | |
| 
 | |
| int virtio_transport_notify_recv_init(struct vsock_sock *vsk,
 | |
| 	size_t target, struct vsock_transport_recv_notify_data *data);
 | |
| int virtio_transport_notify_recv_pre_block(struct vsock_sock *vsk,
 | |
| 	size_t target, struct vsock_transport_recv_notify_data *data);
 | |
| int virtio_transport_notify_recv_pre_dequeue(struct vsock_sock *vsk,
 | |
| 	size_t target, struct vsock_transport_recv_notify_data *data);
 | |
| int virtio_transport_notify_recv_post_dequeue(struct vsock_sock *vsk,
 | |
| 	size_t target, ssize_t copied, bool data_read,
 | |
| 	struct vsock_transport_recv_notify_data *data);
 | |
| int virtio_transport_notify_send_init(struct vsock_sock *vsk,
 | |
| 	struct vsock_transport_send_notify_data *data);
 | |
| int virtio_transport_notify_send_pre_block(struct vsock_sock *vsk,
 | |
| 	struct vsock_transport_send_notify_data *data);
 | |
| int virtio_transport_notify_send_pre_enqueue(struct vsock_sock *vsk,
 | |
| 	struct vsock_transport_send_notify_data *data);
 | |
| int virtio_transport_notify_send_post_enqueue(struct vsock_sock *vsk,
 | |
| 	ssize_t written, struct vsock_transport_send_notify_data *data);
 | |
| void virtio_transport_notify_buffer_size(struct vsock_sock *vsk, u64 *val);
 | |
| 
 | |
| u64 virtio_transport_stream_rcvhiwat(struct vsock_sock *vsk);
 | |
| bool virtio_transport_stream_is_active(struct vsock_sock *vsk);
 | |
| bool virtio_transport_stream_allow(u32 cid, u32 port);
 | |
| int virtio_transport_dgram_bind(struct vsock_sock *vsk,
 | |
| 				struct sockaddr_vm *addr);
 | |
| bool virtio_transport_dgram_allow(u32 cid, u32 port);
 | |
| 
 | |
| int virtio_transport_connect(struct vsock_sock *vsk);
 | |
| 
 | |
| int virtio_transport_shutdown(struct vsock_sock *vsk, int mode);
 | |
| 
 | |
| void virtio_transport_release(struct vsock_sock *vsk);
 | |
| 
 | |
| ssize_t
 | |
| virtio_transport_stream_enqueue(struct vsock_sock *vsk,
 | |
| 				struct msghdr *msg,
 | |
| 				size_t len);
 | |
| int
 | |
| virtio_transport_dgram_enqueue(struct vsock_sock *vsk,
 | |
| 			       struct sockaddr_vm *remote_addr,
 | |
| 			       struct msghdr *msg,
 | |
| 			       size_t len);
 | |
| 
 | |
| void virtio_transport_destruct(struct vsock_sock *vsk);
 | |
| 
 | |
| void virtio_transport_recv_pkt(struct virtio_transport *t,
 | |
| 			       struct sk_buff *skb);
 | |
| void virtio_transport_inc_tx_pkt(struct virtio_vsock_sock *vvs, struct sk_buff *skb);
 | |
| u32 virtio_transport_get_credit(struct virtio_vsock_sock *vvs, u32 wanted);
 | |
| void virtio_transport_put_credit(struct virtio_vsock_sock *vvs, u32 credit);
 | |
| void virtio_transport_deliver_tap_pkt(struct sk_buff *skb);
 | |
| int virtio_transport_purge_skbs(void *vsk, struct sk_buff_head *list);
 | |
| int virtio_transport_read_skb(struct vsock_sock *vsk, skb_read_actor_t read_actor);
 | |
| int virtio_transport_notify_set_rcvlowat(struct vsock_sock *vsk, int val);
 | |
| #endif /* _LINUX_VIRTIO_VSOCK_H */
 |