chcr_ktls: use tls_offload_context_tx and driver_state like other drivers
authorSabrina Dubroca <sd@queasysnail.net>
Mon, 9 Oct 2023 20:50:53 +0000 (22:50 +0200)
committerDavid S. Miller <davem@davemloft.net>
Fri, 13 Oct 2023 10:26:10 +0000 (11:26 +0100)
chcr_ktls uses the space reserved in driver_state by
tls_set_device_offload, but makes up into own wrapper around
tls_offload_context_tx instead of accessing driver_state via the
__tls_driver_ctx helper.

In this driver, driver_state is only used to store a pointer to a
larger context struct allocated by the driver.

Signed-off-by: Sabrina Dubroca <sd@queasysnail.net>
Signed-off-by: David S. Miller <davem@davemloft.net>
drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/chcr_ktls.c
drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/chcr_ktls.h

index bcdc7fc2f4276ace76c6fd3a35a93fd918fed792..6482728794dd4f0e7e63d51d8fe4842210cfba4a 100644 (file)
@@ -361,9 +361,7 @@ static void chcr_ktls_dev_del(struct net_device *netdev,
                              struct tls_context *tls_ctx,
                              enum tls_offload_ctx_dir direction)
 {
-       struct chcr_ktls_ofld_ctx_tx *tx_ctx =
-                               chcr_get_ktls_tx_context(tls_ctx);
-       struct chcr_ktls_info *tx_info = tx_ctx->chcr_info;
+       struct chcr_ktls_info *tx_info = chcr_get_ktls_tx_info(tls_ctx);
        struct ch_ktls_port_stats_debug *port_stats;
        struct chcr_ktls_uld_ctx *u_ctx;
 
@@ -396,7 +394,7 @@ static void chcr_ktls_dev_del(struct net_device *netdev,
        port_stats = &tx_info->adap->ch_ktls_stats.ktls_port[tx_info->port_id];
        atomic64_inc(&port_stats->ktls_tx_connection_close);
        kvfree(tx_info);
-       tx_ctx->chcr_info = NULL;
+       chcr_set_ktls_tx_info(tls_ctx, NULL);
        /* release module refcount */
        module_put(THIS_MODULE);
 }
@@ -417,7 +415,6 @@ static int chcr_ktls_dev_add(struct net_device *netdev, struct sock *sk,
 {
        struct tls_context *tls_ctx = tls_get_ctx(sk);
        struct ch_ktls_port_stats_debug *port_stats;
-       struct chcr_ktls_ofld_ctx_tx *tx_ctx;
        struct chcr_ktls_uld_ctx *u_ctx;
        struct chcr_ktls_info *tx_info;
        struct dst_entry *dst;
@@ -427,8 +424,6 @@ static int chcr_ktls_dev_add(struct net_device *netdev, struct sock *sk,
        u8 daaddr[16];
        int ret = -1;
 
-       tx_ctx = chcr_get_ktls_tx_context(tls_ctx);
-
        pi = netdev_priv(netdev);
        adap = pi->adapter;
        port_stats = &adap->ch_ktls_stats.ktls_port[pi->port_id];
@@ -440,7 +435,7 @@ static int chcr_ktls_dev_add(struct net_device *netdev, struct sock *sk,
                goto out;
        }
 
-       if (tx_ctx->chcr_info)
+       if (chcr_get_ktls_tx_info(tls_ctx))
                goto out;
 
        if (u_ctx && u_ctx->detach)
@@ -566,7 +561,7 @@ static int chcr_ktls_dev_add(struct net_device *netdev, struct sock *sk,
                goto free_tid;
 
        atomic64_inc(&port_stats->ktls_tx_ctx);
-       tx_ctx->chcr_info = tx_info;
+       chcr_set_ktls_tx_info(tls_ctx, tx_info);
 
        return 0;
 
@@ -647,7 +642,7 @@ static int chcr_ktls_cpl_act_open_rpl(struct adapter *adap,
 {
        const struct cpl_act_open_rpl *p = (void *)input;
        struct chcr_ktls_info *tx_info = NULL;
-       struct chcr_ktls_ofld_ctx_tx *tx_ctx;
+       struct tls_offload_context_tx *tx_ctx;
        struct chcr_ktls_uld_ctx *u_ctx;
        unsigned int atid, tid, status;
        struct tls_context *tls_ctx;
@@ -686,7 +681,7 @@ static int chcr_ktls_cpl_act_open_rpl(struct adapter *adap,
                cxgb4_insert_tid(t, tx_info, tx_info->tid, tx_info->ip_family);
                /* Adding tid */
                tls_ctx = tls_get_ctx(tx_info->sk);
-               tx_ctx = chcr_get_ktls_tx_context(tls_ctx);
+               tx_ctx = tls_offload_ctx_tx(tls_ctx);
                u_ctx = adap->uld[CXGB4_ULD_KTLS].handle;
                if (u_ctx) {
                        ret = xa_insert_bh(&u_ctx->tid_list, tid, tx_ctx,
@@ -1924,7 +1919,7 @@ static int chcr_ktls_xmit(struct sk_buff *skb, struct net_device *dev)
 {
        u32 tls_end_offset, tcp_seq, skb_data_len, skb_offset;
        struct ch_ktls_port_stats_debug *port_stats;
-       struct chcr_ktls_ofld_ctx_tx *tx_ctx;
+       struct tls_offload_context_tx *tx_ctx;
        struct ch_ktls_stats_debug *stats;
        struct tcphdr *th = tcp_hdr(skb);
        int data_len, qidx, ret = 0, mss;
@@ -1944,6 +1939,7 @@ static int chcr_ktls_xmit(struct sk_buff *skb, struct net_device *dev)
        mss = skb_is_gso(skb) ? skb_shinfo(skb)->gso_size : data_len;
 
        tls_ctx = tls_get_ctx(skb->sk);
+       tx_ctx = tls_offload_ctx_tx(tls_ctx);
        tls_netdev = rcu_dereference_bh(tls_ctx->netdev);
        /* Don't quit on NULL: if tls_device_down is running in parallel,
         * netdev might become NULL, even if tls_is_skb_tx_device_offloaded was
@@ -1952,8 +1948,7 @@ static int chcr_ktls_xmit(struct sk_buff *skb, struct net_device *dev)
        if (unlikely(tls_netdev && tls_netdev != dev))
                goto out;
 
-       tx_ctx = chcr_get_ktls_tx_context(tls_ctx);
-       tx_info = tx_ctx->chcr_info;
+       tx_info = chcr_get_ktls_tx_info(tls_ctx);
 
        if (unlikely(!tx_info))
                goto out;
@@ -1979,19 +1974,19 @@ static int chcr_ktls_xmit(struct sk_buff *skb, struct net_device *dev)
         * we will send the complete record again.
         */
 
-       spin_lock_irqsave(&tx_ctx->base.lock, flags);
+       spin_lock_irqsave(&tx_ctx->lock, flags);
 
        do {
 
                cxgb4_reclaim_completed_tx(adap, &q->q, true);
                /* fetch the tls record */
-               record = tls_get_record(&tx_ctx->base, tcp_seq,
+               record = tls_get_record(tx_ctx, tcp_seq,
                                        &tx_info->record_no);
                /* By the time packet reached to us, ACK is received, and record
                 * won't be found in that case, handle it gracefully.
                 */
                if (unlikely(!record)) {
-                       spin_unlock_irqrestore(&tx_ctx->base.lock, flags);
+                       spin_unlock_irqrestore(&tx_ctx->lock, flags);
                        atomic64_inc(&port_stats->ktls_tx_drop_no_sync_data);
                        goto out;
                }
@@ -2015,7 +2010,7 @@ static int chcr_ktls_xmit(struct sk_buff *skb, struct net_device *dev)
                                                      tls_end_offset !=
                                                      record->len);
                        if (ret) {
-                               spin_unlock_irqrestore(&tx_ctx->base.lock,
+                               spin_unlock_irqrestore(&tx_ctx->lock,
                                                       flags);
                                goto out;
                        }
@@ -2046,7 +2041,7 @@ static int chcr_ktls_xmit(struct sk_buff *skb, struct net_device *dev)
                                /* free the refcount taken earlier */
                                if (tls_end_offset < data_len)
                                        dev_kfree_skb_any(skb);
-                               spin_unlock_irqrestore(&tx_ctx->base.lock, flags);
+                               spin_unlock_irqrestore(&tx_ctx->lock, flags);
                                goto out;
                        }
 
@@ -2082,7 +2077,7 @@ static int chcr_ktls_xmit(struct sk_buff *skb, struct net_device *dev)
 
                /* if any failure, come out from the loop. */
                if (ret) {
-                       spin_unlock_irqrestore(&tx_ctx->base.lock, flags);
+                       spin_unlock_irqrestore(&tx_ctx->lock, flags);
                        if (th->fin)
                                dev_kfree_skb_any(skb);
 
@@ -2097,7 +2092,7 @@ static int chcr_ktls_xmit(struct sk_buff *skb, struct net_device *dev)
 
        } while (data_len > 0);
 
-       spin_unlock_irqrestore(&tx_ctx->base.lock, flags);
+       spin_unlock_irqrestore(&tx_ctx->lock, flags);
        atomic64_inc(&port_stats->ktls_tx_encrypted_packets);
        atomic64_add(skb_data_len, &port_stats->ktls_tx_encrypted_bytes);
 
@@ -2185,17 +2180,17 @@ static void clear_conn_resources(struct chcr_ktls_info *tx_info)
 static void ch_ktls_reset_all_conn(struct chcr_ktls_uld_ctx *u_ctx)
 {
        struct ch_ktls_port_stats_debug *port_stats;
-       struct chcr_ktls_ofld_ctx_tx *tx_ctx;
+       struct tls_offload_context_tx *tx_ctx;
        struct chcr_ktls_info *tx_info;
        unsigned long index;
 
        xa_for_each(&u_ctx->tid_list, index, tx_ctx) {
-               tx_info = tx_ctx->chcr_info;
+               tx_info = __chcr_get_ktls_tx_info(tx_ctx);
                clear_conn_resources(tx_info);
                port_stats = &tx_info->adap->ch_ktls_stats.ktls_port[tx_info->port_id];
                atomic64_inc(&port_stats->ktls_tx_connection_close);
                kvfree(tx_info);
-               tx_ctx->chcr_info = NULL;
+               memset(tx_ctx->driver_state, 0, TLS_DRIVER_STATE_SIZE_TX);
                /* release module refcount */
                module_put(THIS_MODULE);
        }
index 10572dc55365a9e418eb9f02587ea3841a5e78a3..dbbba92bf54097239374864ca6c0fdcde0b8095f 100644 (file)
@@ -67,8 +67,7 @@ struct chcr_ktls_info {
        bool pending_close;
 };
 
-struct chcr_ktls_ofld_ctx_tx {
-       struct tls_offload_context_tx base;
+struct chcr_ktls_ctx_tx {
        struct chcr_ktls_info *chcr_info;
 };
 
@@ -79,14 +78,33 @@ struct chcr_ktls_uld_ctx {
        bool detach;
 };
 
-static inline struct chcr_ktls_ofld_ctx_tx *
-chcr_get_ktls_tx_context(struct tls_context *tls_ctx)
+static inline struct chcr_ktls_info *
+__chcr_get_ktls_tx_info(struct tls_offload_context_tx *octx)
 {
-       BUILD_BUG_ON(sizeof(struct chcr_ktls_ofld_ctx_tx) >
-                    TLS_OFFLOAD_CONTEXT_SIZE_TX);
-       return container_of(tls_offload_ctx_tx(tls_ctx),
-                           struct chcr_ktls_ofld_ctx_tx,
-                           base);
+       struct chcr_ktls_ctx_tx *priv_ctx;
+
+       BUILD_BUG_ON(sizeof(struct chcr_ktls_ctx_tx) > TLS_DRIVER_STATE_SIZE_TX);
+       priv_ctx = (struct chcr_ktls_ctx_tx *)octx->driver_state;
+       return priv_ctx->chcr_info;
+}
+
+static inline struct chcr_ktls_info *
+chcr_get_ktls_tx_info(struct tls_context *tls_ctx)
+{
+       struct chcr_ktls_ctx_tx *priv_ctx;
+
+       BUILD_BUG_ON(sizeof(struct chcr_ktls_ctx_tx) > TLS_DRIVER_STATE_SIZE_TX);
+       priv_ctx = (struct chcr_ktls_ctx_tx *)__tls_driver_ctx(tls_ctx, TLS_OFFLOAD_CTX_DIR_TX);
+       return priv_ctx->chcr_info;
+}
+
+static inline void
+chcr_set_ktls_tx_info(struct tls_context *tls_ctx, struct chcr_ktls_info *chcr_info)
+{
+       struct chcr_ktls_ctx_tx *priv_ctx;
+
+       priv_ctx = __tls_driver_ctx(tls_ctx, TLS_OFFLOAD_CTX_DIR_TX);
+       priv_ctx->chcr_info = chcr_info;
 }
 
 static inline int chcr_get_first_rx_qid(struct adapter *adap)