Add the rt linux 4.1.3-rt3 as base
[kvmfornfv.git] / kernel / net / vmw_vsock / vmci_transport_notify_qstate.c
diff --git a/kernel/net/vmw_vsock/vmci_transport_notify_qstate.c b/kernel/net/vmw_vsock/vmci_transport_notify_qstate.c
new file mode 100644 (file)
index 0000000..dc9c792
--- /dev/null
@@ -0,0 +1,438 @@
+/*
+ * VMware vSockets Driver
+ *
+ * Copyright (C) 2009-2013 VMware, Inc. All rights reserved.
+ *
+ * This program is free software; you can redistribute it and/or modify it
+ * under the terms of the GNU General Public License as published by the Free
+ * Software Foundation version 2 and no later version.
+ *
+ * This program is distributed in the hope that it will be useful, but WITHOUT
+ * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
+ * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
+ * more details.
+ */
+
+#include <linux/types.h>
+#include <linux/socket.h>
+#include <linux/stddef.h>
+#include <net/sock.h>
+
+#include "vmci_transport_notify.h"
+
+#define PKT_FIELD(vsk, field_name) \
+       (vmci_trans(vsk)->notify.pkt_q_state.field_name)
+
+static bool vmci_transport_notify_waiting_write(struct vsock_sock *vsk)
+{
+       bool retval;
+       u64 notify_limit;
+
+       if (!PKT_FIELD(vsk, peer_waiting_write))
+               return false;
+
+       /* When the sender blocks, we take that as a sign that the sender is
+        * faster than the receiver. To reduce the transmit rate of the sender,
+        * we delay the sending of the read notification by decreasing the
+        * write_notify_window. The notification is delayed until the number of
+        * bytes used in the queue drops below the write_notify_window.
+        */
+
+       if (!PKT_FIELD(vsk, peer_waiting_write_detected)) {
+               PKT_FIELD(vsk, peer_waiting_write_detected) = true;
+               if (PKT_FIELD(vsk, write_notify_window) < PAGE_SIZE) {
+                       PKT_FIELD(vsk, write_notify_window) =
+                           PKT_FIELD(vsk, write_notify_min_window);
+               } else {
+                       PKT_FIELD(vsk, write_notify_window) -= PAGE_SIZE;
+                       if (PKT_FIELD(vsk, write_notify_window) <
+                           PKT_FIELD(vsk, write_notify_min_window))
+                               PKT_FIELD(vsk, write_notify_window) =
+                                   PKT_FIELD(vsk, write_notify_min_window);
+
+               }
+       }
+       notify_limit = vmci_trans(vsk)->consume_size -
+               PKT_FIELD(vsk, write_notify_window);
+
+       /* The notify_limit is used to delay notifications in the case where
+        * flow control is enabled. Below the test is expressed in terms of
+        * free space in the queue: if free_space > ConsumeSize -
+        * write_notify_window then notify An alternate way of expressing this
+        * is to rewrite the expression to use the data ready in the receive
+        * queue: if write_notify_window > bufferReady then notify as
+        * free_space == ConsumeSize - bufferReady.
+        */
+
+       retval = vmci_qpair_consume_free_space(vmci_trans(vsk)->qpair) >
+               notify_limit;
+
+       if (retval) {
+               /* Once we notify the peer, we reset the detected flag so the
+                * next wait will again cause a decrease in the window size.
+                */
+
+               PKT_FIELD(vsk, peer_waiting_write_detected) = false;
+       }
+       return retval;
+}
+
+static void
+vmci_transport_handle_read(struct sock *sk,
+                          struct vmci_transport_packet *pkt,
+                          bool bottom_half,
+                          struct sockaddr_vm *dst, struct sockaddr_vm *src)
+{
+       sk->sk_write_space(sk);
+}
+
+static void
+vmci_transport_handle_wrote(struct sock *sk,
+                           struct vmci_transport_packet *pkt,
+                           bool bottom_half,
+                           struct sockaddr_vm *dst, struct sockaddr_vm *src)
+{
+       sk->sk_data_ready(sk);
+}
+
+static void vsock_block_update_write_window(struct sock *sk)
+{
+       struct vsock_sock *vsk = vsock_sk(sk);
+
+       if (PKT_FIELD(vsk, write_notify_window) < vmci_trans(vsk)->consume_size)
+               PKT_FIELD(vsk, write_notify_window) =
+                   min(PKT_FIELD(vsk, write_notify_window) + PAGE_SIZE,
+                       vmci_trans(vsk)->consume_size);
+}
+
+static int vmci_transport_send_read_notification(struct sock *sk)
+{
+       struct vsock_sock *vsk;
+       bool sent_read;
+       unsigned int retries;
+       int err;
+
+       vsk = vsock_sk(sk);
+       sent_read = false;
+       retries = 0;
+       err = 0;
+
+       if (vmci_transport_notify_waiting_write(vsk)) {
+               /* Notify the peer that we have read, retrying the send on
+                * failure up to our maximum value.  XXX For now we just log
+                * the failure, but later we should schedule a work item to
+                * handle the resend until it succeeds.  That would require
+                * keeping track of work items in the vsk and cleaning them up
+                * upon socket close.
+                */
+               while (!(vsk->peer_shutdown & RCV_SHUTDOWN) &&
+                      !sent_read &&
+                      retries < VMCI_TRANSPORT_MAX_DGRAM_RESENDS) {
+                       err = vmci_transport_send_read(sk);
+                       if (err >= 0)
+                               sent_read = true;
+
+                       retries++;
+               }
+
+               if (retries >= VMCI_TRANSPORT_MAX_DGRAM_RESENDS && !sent_read)
+                       pr_err("%p unable to send read notification to peer\n",
+                              sk);
+               else
+                       PKT_FIELD(vsk, peer_waiting_write) = false;
+
+       }
+       return err;
+}
+
+static void vmci_transport_notify_pkt_socket_init(struct sock *sk)
+{
+       struct vsock_sock *vsk = vsock_sk(sk);
+
+       PKT_FIELD(vsk, write_notify_window) = PAGE_SIZE;
+       PKT_FIELD(vsk, write_notify_min_window) = PAGE_SIZE;
+       PKT_FIELD(vsk, peer_waiting_write) = false;
+       PKT_FIELD(vsk, peer_waiting_write_detected) = false;
+}
+
+static void vmci_transport_notify_pkt_socket_destruct(struct vsock_sock *vsk)
+{
+       PKT_FIELD(vsk, write_notify_window) = PAGE_SIZE;
+       PKT_FIELD(vsk, write_notify_min_window) = PAGE_SIZE;
+       PKT_FIELD(vsk, peer_waiting_write) = false;
+       PKT_FIELD(vsk, peer_waiting_write_detected) = false;
+}
+
+static int
+vmci_transport_notify_pkt_poll_in(struct sock *sk,
+                                 size_t target, bool *data_ready_now)
+{
+       struct vsock_sock *vsk = vsock_sk(sk);
+
+       if (vsock_stream_has_data(vsk)) {
+               *data_ready_now = true;
+       } else {
+               /* We can't read right now because there is nothing in the
+                * queue. Ask for notifications when there is something to
+                * read.
+                */
+               if (sk->sk_state == SS_CONNECTED)
+                       vsock_block_update_write_window(sk);
+               *data_ready_now = false;
+       }
+
+       return 0;
+}
+
+static int
+vmci_transport_notify_pkt_poll_out(struct sock *sk,
+                                  size_t target, bool *space_avail_now)
+{
+       s64 produce_q_free_space;
+       struct vsock_sock *vsk = vsock_sk(sk);
+
+       produce_q_free_space = vsock_stream_has_space(vsk);
+       if (produce_q_free_space > 0) {
+               *space_avail_now = true;
+               return 0;
+       } else if (produce_q_free_space == 0) {
+               /* This is a connected socket but we can't currently send data.
+                * Nothing else to do.
+                */
+               *space_avail_now = false;
+       }
+
+       return 0;
+}
+
+static int
+vmci_transport_notify_pkt_recv_init(
+                               struct sock *sk,
+                               size_t target,
+                               struct vmci_transport_recv_notify_data *data)
+{
+       struct vsock_sock *vsk = vsock_sk(sk);
+
+       data->consume_head = 0;
+       data->produce_tail = 0;
+       data->notify_on_block = false;
+
+       if (PKT_FIELD(vsk, write_notify_min_window) < target + 1) {
+               PKT_FIELD(vsk, write_notify_min_window) = target + 1;
+               if (PKT_FIELD(vsk, write_notify_window) <
+                   PKT_FIELD(vsk, write_notify_min_window)) {
+                       /* If the current window is smaller than the new
+                        * minimal window size, we need to reevaluate whether
+                        * we need to notify the sender. If the number of ready
+                        * bytes are smaller than the new window, we need to
+                        * send a notification to the sender before we block.
+                        */
+
+                       PKT_FIELD(vsk, write_notify_window) =
+                           PKT_FIELD(vsk, write_notify_min_window);
+                       data->notify_on_block = true;
+               }
+       }
+
+       return 0;
+}
+
+static int
+vmci_transport_notify_pkt_recv_pre_block(
+                               struct sock *sk,
+                               size_t target,
+                               struct vmci_transport_recv_notify_data *data)
+{
+       int err = 0;
+
+       vsock_block_update_write_window(sk);
+
+       if (data->notify_on_block) {
+               err = vmci_transport_send_read_notification(sk);
+               if (err < 0)
+                       return err;
+               data->notify_on_block = false;
+       }
+
+       return err;
+}
+
+static int
+vmci_transport_notify_pkt_recv_post_dequeue(
+                               struct sock *sk,
+                               size_t target,
+                               ssize_t copied,
+                               bool data_read,
+                               struct vmci_transport_recv_notify_data *data)
+{
+       struct vsock_sock *vsk;
+       int err;
+       bool was_full = false;
+       u64 free_space;
+
+       vsk = vsock_sk(sk);
+       err = 0;
+
+       if (data_read) {
+               smp_mb();
+
+               free_space =
+                       vmci_qpair_consume_free_space(vmci_trans(vsk)->qpair);
+               was_full = free_space == copied;
+
+               if (was_full)
+                       PKT_FIELD(vsk, peer_waiting_write) = true;
+
+               err = vmci_transport_send_read_notification(sk);
+               if (err < 0)
+                       return err;
+
+               /* See the comment in
+                * vmci_transport_notify_pkt_send_post_enqueue().
+                */
+               sk->sk_data_ready(sk);
+       }
+
+       return err;
+}
+
+static int
+vmci_transport_notify_pkt_send_init(
+                               struct sock *sk,
+                               struct vmci_transport_send_notify_data *data)
+{
+       data->consume_head = 0;
+       data->produce_tail = 0;
+
+       return 0;
+}
+
+static int
+vmci_transport_notify_pkt_send_post_enqueue(
+                               struct sock *sk,
+                               ssize_t written,
+                               struct vmci_transport_send_notify_data *data)
+{
+       int err = 0;
+       struct vsock_sock *vsk;
+       bool sent_wrote = false;
+       bool was_empty;
+       int retries = 0;
+
+       vsk = vsock_sk(sk);
+
+       smp_mb();
+
+       was_empty =
+               vmci_qpair_produce_buf_ready(vmci_trans(vsk)->qpair) == written;
+       if (was_empty) {
+               while (!(vsk->peer_shutdown & RCV_SHUTDOWN) &&
+                      !sent_wrote &&
+                      retries < VMCI_TRANSPORT_MAX_DGRAM_RESENDS) {
+                       err = vmci_transport_send_wrote(sk);
+                       if (err >= 0)
+                               sent_wrote = true;
+
+                       retries++;
+               }
+       }
+
+       if (retries >= VMCI_TRANSPORT_MAX_DGRAM_RESENDS && !sent_wrote) {
+               pr_err("%p unable to send wrote notification to peer\n",
+                      sk);
+               return err;
+       }
+
+       return err;
+}
+
+static void
+vmci_transport_notify_pkt_handle_pkt(
+                               struct sock *sk,
+                               struct vmci_transport_packet *pkt,
+                               bool bottom_half,
+                               struct sockaddr_vm *dst,
+                               struct sockaddr_vm *src, bool *pkt_processed)
+{
+       bool processed = false;
+
+       switch (pkt->type) {
+       case VMCI_TRANSPORT_PACKET_TYPE_WROTE:
+               vmci_transport_handle_wrote(sk, pkt, bottom_half, dst, src);
+               processed = true;
+               break;
+       case VMCI_TRANSPORT_PACKET_TYPE_READ:
+               vmci_transport_handle_read(sk, pkt, bottom_half, dst, src);
+               processed = true;
+               break;
+       }
+
+       if (pkt_processed)
+               *pkt_processed = processed;
+}
+
+static void vmci_transport_notify_pkt_process_request(struct sock *sk)
+{
+       struct vsock_sock *vsk = vsock_sk(sk);
+
+       PKT_FIELD(vsk, write_notify_window) = vmci_trans(vsk)->consume_size;
+       if (vmci_trans(vsk)->consume_size <
+               PKT_FIELD(vsk, write_notify_min_window))
+               PKT_FIELD(vsk, write_notify_min_window) =
+                       vmci_trans(vsk)->consume_size;
+}
+
+static void vmci_transport_notify_pkt_process_negotiate(struct sock *sk)
+{
+       struct vsock_sock *vsk = vsock_sk(sk);
+
+       PKT_FIELD(vsk, write_notify_window) = vmci_trans(vsk)->consume_size;
+       if (vmci_trans(vsk)->consume_size <
+               PKT_FIELD(vsk, write_notify_min_window))
+               PKT_FIELD(vsk, write_notify_min_window) =
+                       vmci_trans(vsk)->consume_size;
+}
+
+static int
+vmci_transport_notify_pkt_recv_pre_dequeue(
+                               struct sock *sk,
+                               size_t target,
+                               struct vmci_transport_recv_notify_data *data)
+{
+       return 0; /* NOP for QState. */
+}
+
+static int
+vmci_transport_notify_pkt_send_pre_block(
+                               struct sock *sk,
+                               struct vmci_transport_send_notify_data *data)
+{
+       return 0; /* NOP for QState. */
+}
+
+static int
+vmci_transport_notify_pkt_send_pre_enqueue(
+                               struct sock *sk,
+                               struct vmci_transport_send_notify_data *data)
+{
+       return 0; /* NOP for QState. */
+}
+
+/* Socket always on control packet based operations. */
+struct vmci_transport_notify_ops vmci_transport_notify_pkt_q_state_ops = {
+       vmci_transport_notify_pkt_socket_init,
+       vmci_transport_notify_pkt_socket_destruct,
+       vmci_transport_notify_pkt_poll_in,
+       vmci_transport_notify_pkt_poll_out,
+       vmci_transport_notify_pkt_handle_pkt,
+       vmci_transport_notify_pkt_recv_init,
+       vmci_transport_notify_pkt_recv_pre_block,
+       vmci_transport_notify_pkt_recv_pre_dequeue,
+       vmci_transport_notify_pkt_recv_post_dequeue,
+       vmci_transport_notify_pkt_send_init,
+       vmci_transport_notify_pkt_send_pre_block,
+       vmci_transport_notify_pkt_send_pre_enqueue,
+       vmci_transport_notify_pkt_send_post_enqueue,
+       vmci_transport_notify_pkt_process_request,
+       vmci_transport_notify_pkt_process_negotiate,
+};