tls: Fix write space handling
authorBoris Pismenny <borisp@mellanox.com>
Wed, 27 Feb 2019 15:38:04 +0000 (17:38 +0200)
committerDavid S. Miller <davem@davemloft.net>
Mon, 4 Mar 2019 06:10:16 +0000 (22:10 -0800)
TLS device cannot use the sw context. This patch returns the original
tls device write space handler and moves the sw/device specific portions
to the relevant files.

Also, we remove the write_space call for the tls_sw flow, because it
handles partial records in its delayed tx work handler.

Fixes: a42055e8d2c3 ("net/tls: Add support for async encryption of records for performance")
Signed-off-by: Boris Pismenny <borisp@mellanox.com>
Reviewed-by: Eran Ben Elisha <eranbe@mellanox.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
include/net/tls.h
net/tls/tls_device.c
net/tls/tls_main.c
net/tls/tls_sw.c

index a528a082da73eddeedeaa6d154ca4559af72c2df..a5a938583295c0789df287c737b7a1c87556c9f1 100644 (file)
@@ -519,6 +519,9 @@ static inline bool tls_sw_has_ctx_tx(const struct sock *sk)
        return !!tls_sw_ctx_tx(ctx);
 }
 
+void tls_sw_write_space(struct sock *sk, struct tls_context *ctx);
+void tls_device_write_space(struct sock *sk, struct tls_context *ctx);
+
 static inline struct tls_offload_context_rx *
 tls_offload_ctx_rx(const struct tls_context *tls_ctx)
 {
index 3e5e8e021a878acca89b4a5c89d52c7a257701c2..4a1da837a733d0ec583b7b6f33f3c73466812645 100644 (file)
@@ -546,6 +546,23 @@ static int tls_device_push_pending_record(struct sock *sk, int flags)
        return tls_push_data(sk, &msg_iter, 0, flags, TLS_RECORD_TYPE_DATA);
 }
 
+void tls_device_write_space(struct sock *sk, struct tls_context *ctx)
+{
+       int rc = 0;
+
+       if (!sk->sk_write_pending && tls_is_partially_sent_record(ctx)) {
+               gfp_t sk_allocation = sk->sk_allocation;
+
+               sk->sk_allocation = GFP_ATOMIC;
+               rc = tls_push_partial_record(sk, ctx,
+                                            MSG_DONTWAIT | MSG_NOSIGNAL);
+               sk->sk_allocation = sk_allocation;
+       }
+
+       if (!rc)
+               ctx->sk_write_space(sk);
+}
+
 void handle_device_resync(struct sock *sk, u32 seq, u64 rcd_sn)
 {
        struct tls_context *tls_ctx = tls_get_ctx(sk);
index 7e05af75536d1c355d3dc95813ddc34c06af9660..17e8667917aa3c14fbe0111f426072c559db053e 100644 (file)
@@ -212,7 +212,6 @@ int tls_push_partial_record(struct sock *sk, struct tls_context *ctx,
 static void tls_write_space(struct sock *sk)
 {
        struct tls_context *ctx = tls_get_ctx(sk);
-       struct tls_sw_context_tx *tx_ctx = tls_sw_ctx_tx(ctx);
 
        /* If in_tcp_sendpages call lower protocol write space handler
         * to ensure we wake up any waiting operations there. For example
@@ -223,14 +222,12 @@ static void tls_write_space(struct sock *sk)
                return;
        }
 
-       /* Schedule the transmission if tx list is ready */
-       if (is_tx_ready(tx_ctx) && !sk->sk_write_pending) {
-               /* Schedule the transmission */
-               if (!test_and_set_bit(BIT_TX_SCHEDULED, &tx_ctx->tx_bitmask))
-                       schedule_delayed_work(&tx_ctx->tx_work.work, 0);
-       }
-
-       ctx->sk_write_space(sk);
+#ifdef CONFIG_TLS_DEVICE
+       if (ctx->tx_conf == TLS_HW)
+               tls_device_write_space(sk, ctx);
+       else
+#endif
+               tls_sw_write_space(sk, ctx);
 }
 
 static void tls_ctx_free(struct tls_context *ctx)
index 1cc830582fa8af2e0ff978005eba5c5a609ae51b..917caacd4d31fed4cf52fde7c95b4e55ec2da235 100644 (file)
@@ -2126,6 +2126,19 @@ static void tx_work_handler(struct work_struct *work)
        release_sock(sk);
 }
 
+void tls_sw_write_space(struct sock *sk, struct tls_context *ctx)
+{
+       struct tls_sw_context_tx *tx_ctx = tls_sw_ctx_tx(ctx);
+
+       /* Schedule the transmission if tx list is ready */
+       if (is_tx_ready(tx_ctx) && !sk->sk_write_pending) {
+               /* Schedule the transmission */
+               if (!test_and_set_bit(BIT_TX_SCHEDULED,
+                                     &tx_ctx->tx_bitmask))
+                       schedule_delayed_work(&tx_ctx->tx_work.work, 0);
+       }
+}
+
 int tls_set_sw_offload(struct sock *sk, struct tls_context *ctx, int tx)
 {
        struct tls_context *tls_ctx = tls_get_ctx(sk);