| 1 | // SPDX-License-Identifier: GPL-2.0-only |
| 2 | /* |
| 3 | * Copyright 2007-2012 Siemens AG |
| 4 | * |
| 5 | * Written by: |
| 6 | * Dmitry Eremin-Solenikov <dbaryshkov@gmail.com> |
| 7 | * Sergey Lapin <slapin@ossfans.org> |
| 8 | * Maxim Gorbachyov <maxim.gorbachev@siemens.com> |
| 9 | * Alexander Smirnov <alex.bluesman.smirnov@gmail.com> |
| 10 | */ |
| 11 | |
| 12 | #include <linux/netdevice.h> |
| 13 | #include <linux/if_arp.h> |
| 14 | #include <linux/crc-ccitt.h> |
| 15 | #include <asm/unaligned.h> |
| 16 | |
| 17 | #include <net/rtnetlink.h> |
| 18 | #include <net/ieee802154_netdev.h> |
| 19 | #include <net/mac802154.h> |
| 20 | #include <net/cfg802154.h> |
| 21 | |
| 22 | #include "ieee802154_i.h" |
| 23 | #include "driver-ops.h" |
| 24 | |
| 25 | void ieee802154_xmit_sync_worker(struct work_struct *work) |
| 26 | { |
| 27 | struct ieee802154_local *local = |
| 28 | container_of(work, struct ieee802154_local, sync_tx_work); |
| 29 | struct sk_buff *skb = local->tx_skb; |
| 30 | struct net_device *dev = skb->dev; |
| 31 | int res; |
| 32 | |
| 33 | res = drv_xmit_sync(local, skb); |
| 34 | if (res) |
| 35 | goto err_tx; |
| 36 | |
| 37 | DEV_STATS_INC(dev, tx_packets); |
| 38 | DEV_STATS_ADD(dev, tx_bytes, skb->len); |
| 39 | |
| 40 | ieee802154_xmit_complete(&local->hw, skb, false); |
| 41 | |
| 42 | return; |
| 43 | |
| 44 | err_tx: |
| 45 | /* Restart the netif queue on each sub_if_data object. */ |
| 46 | ieee802154_release_queue(local); |
| 47 | if (atomic_dec_and_test(&local->phy->ongoing_txs)) |
| 48 | wake_up(&local->phy->sync_txq); |
| 49 | kfree_skb(skb); |
| 50 | netdev_dbg(dev, "transmission failed\n"); |
| 51 | } |
| 52 | |
| 53 | static netdev_tx_t |
| 54 | ieee802154_tx(struct ieee802154_local *local, struct sk_buff *skb) |
| 55 | { |
| 56 | struct net_device *dev = skb->dev; |
| 57 | int ret; |
| 58 | |
| 59 | if (!(local->hw.flags & IEEE802154_HW_TX_OMIT_CKSUM)) { |
| 60 | struct sk_buff *nskb; |
| 61 | u16 crc; |
| 62 | |
| 63 | if (unlikely(skb_tailroom(skb) < IEEE802154_FCS_LEN)) { |
| 64 | nskb = skb_copy_expand(skb, 0, IEEE802154_FCS_LEN, |
| 65 | GFP_ATOMIC); |
| 66 | if (likely(nskb)) { |
| 67 | consume_skb(skb); |
| 68 | skb = nskb; |
| 69 | } else { |
| 70 | goto err_free_skb; |
| 71 | } |
| 72 | } |
| 73 | |
| 74 | crc = crc_ccitt(0, skb->data, skb->len); |
| 75 | put_unaligned_le16(crc, skb_put(skb, 2)); |
| 76 | } |
| 77 | |
| 78 | /* Stop the netif queue on each sub_if_data object. */ |
| 79 | ieee802154_hold_queue(local); |
| 80 | atomic_inc(&local->phy->ongoing_txs); |
| 81 | |
| 82 | /* Drivers should preferably implement the async callback. In some rare |
| 83 | * cases they only provide a sync callback which we will use as a |
| 84 | * fallback. |
| 85 | */ |
| 86 | if (local->ops->xmit_async) { |
| 87 | unsigned int len = skb->len; |
| 88 | |
| 89 | ret = drv_xmit_async(local, skb); |
| 90 | if (ret) |
| 91 | goto err_wake_netif_queue; |
| 92 | |
| 93 | DEV_STATS_INC(dev, tx_packets); |
| 94 | DEV_STATS_ADD(dev, tx_bytes, len); |
| 95 | } else { |
| 96 | local->tx_skb = skb; |
| 97 | queue_work(local->workqueue, &local->sync_tx_work); |
| 98 | } |
| 99 | |
| 100 | return NETDEV_TX_OK; |
| 101 | |
| 102 | err_wake_netif_queue: |
| 103 | ieee802154_release_queue(local); |
| 104 | if (atomic_dec_and_test(&local->phy->ongoing_txs)) |
| 105 | wake_up(&local->phy->sync_txq); |
| 106 | err_free_skb: |
| 107 | kfree_skb(skb); |
| 108 | return NETDEV_TX_OK; |
| 109 | } |
| 110 | |
| 111 | static int ieee802154_sync_queue(struct ieee802154_local *local) |
| 112 | { |
| 113 | int ret; |
| 114 | |
| 115 | ieee802154_hold_queue(local); |
| 116 | ieee802154_disable_queue(local); |
| 117 | wait_event(local->phy->sync_txq, !atomic_read(&local->phy->ongoing_txs)); |
| 118 | ret = local->tx_result; |
| 119 | ieee802154_release_queue(local); |
| 120 | |
| 121 | return ret; |
| 122 | } |
| 123 | |
| 124 | int ieee802154_sync_and_hold_queue(struct ieee802154_local *local) |
| 125 | { |
| 126 | int ret; |
| 127 | |
| 128 | ieee802154_hold_queue(local); |
| 129 | ret = ieee802154_sync_queue(local); |
| 130 | set_bit(WPAN_PHY_FLAG_STATE_QUEUE_STOPPED, &local->phy->flags); |
| 131 | |
| 132 | return ret; |
| 133 | } |
| 134 | |
| 135 | int ieee802154_mlme_op_pre(struct ieee802154_local *local) |
| 136 | { |
| 137 | return ieee802154_sync_and_hold_queue(local); |
| 138 | } |
| 139 | |
| 140 | int ieee802154_mlme_tx_locked(struct ieee802154_local *local, |
| 141 | struct ieee802154_sub_if_data *sdata, |
| 142 | struct sk_buff *skb) |
| 143 | { |
| 144 | /* Avoid possible calls to ->ndo_stop() when we asynchronously perform |
| 145 | * MLME transmissions. |
| 146 | */ |
| 147 | ASSERT_RTNL(); |
| 148 | |
| 149 | /* Ensure the device was not stopped, otherwise error out */ |
| 150 | if (!local->open_count) |
| 151 | return -ENETDOWN; |
| 152 | |
| 153 | /* Warn if the ieee802154 core thinks MLME frames can be sent while the |
| 154 | * net interface expects this cannot happen. |
| 155 | */ |
| 156 | if (WARN_ON_ONCE(!netif_running(sdata->dev))) |
| 157 | return -ENETDOWN; |
| 158 | |
| 159 | ieee802154_tx(local, skb); |
| 160 | return ieee802154_sync_queue(local); |
| 161 | } |
| 162 | |
| 163 | int ieee802154_mlme_tx(struct ieee802154_local *local, |
| 164 | struct ieee802154_sub_if_data *sdata, |
| 165 | struct sk_buff *skb) |
| 166 | { |
| 167 | int ret; |
| 168 | |
| 169 | rtnl_lock(); |
| 170 | ret = ieee802154_mlme_tx_locked(local, sdata, skb); |
| 171 | rtnl_unlock(); |
| 172 | |
| 173 | return ret; |
| 174 | } |
| 175 | |
| 176 | void ieee802154_mlme_op_post(struct ieee802154_local *local) |
| 177 | { |
| 178 | ieee802154_release_queue(local); |
| 179 | } |
| 180 | |
| 181 | int ieee802154_mlme_tx_one(struct ieee802154_local *local, |
| 182 | struct ieee802154_sub_if_data *sdata, |
| 183 | struct sk_buff *skb) |
| 184 | { |
| 185 | int ret; |
| 186 | |
| 187 | ieee802154_mlme_op_pre(local); |
| 188 | ret = ieee802154_mlme_tx(local, sdata, skb); |
| 189 | ieee802154_mlme_op_post(local); |
| 190 | |
| 191 | return ret; |
| 192 | } |
| 193 | |
| 194 | int ieee802154_mlme_tx_one_locked(struct ieee802154_local *local, |
| 195 | struct ieee802154_sub_if_data *sdata, |
| 196 | struct sk_buff *skb) |
| 197 | { |
| 198 | int ret; |
| 199 | |
| 200 | ieee802154_mlme_op_pre(local); |
| 201 | ret = ieee802154_mlme_tx_locked(local, sdata, skb); |
| 202 | ieee802154_mlme_op_post(local); |
| 203 | |
| 204 | return ret; |
| 205 | } |
| 206 | |
| 207 | static bool ieee802154_queue_is_stopped(struct ieee802154_local *local) |
| 208 | { |
| 209 | return test_bit(WPAN_PHY_FLAG_STATE_QUEUE_STOPPED, &local->phy->flags); |
| 210 | } |
| 211 | |
| 212 | static netdev_tx_t |
| 213 | ieee802154_hot_tx(struct ieee802154_local *local, struct sk_buff *skb) |
| 214 | { |
| 215 | /* Warn if the net interface tries to transmit frames while the |
| 216 | * ieee802154 core assumes the queue is stopped. |
| 217 | */ |
| 218 | WARN_ON_ONCE(ieee802154_queue_is_stopped(local)); |
| 219 | |
| 220 | return ieee802154_tx(local, skb); |
| 221 | } |
| 222 | |
| 223 | netdev_tx_t |
| 224 | ieee802154_monitor_start_xmit(struct sk_buff *skb, struct net_device *dev) |
| 225 | { |
| 226 | struct ieee802154_sub_if_data *sdata = IEEE802154_DEV_TO_SUB_IF(dev); |
| 227 | |
| 228 | skb->skb_iif = dev->ifindex; |
| 229 | |
| 230 | return ieee802154_hot_tx(sdata->local, skb); |
| 231 | } |
| 232 | |
| 233 | netdev_tx_t |
| 234 | ieee802154_subif_start_xmit(struct sk_buff *skb, struct net_device *dev) |
| 235 | { |
| 236 | struct ieee802154_sub_if_data *sdata = IEEE802154_DEV_TO_SUB_IF(dev); |
| 237 | int rc; |
| 238 | |
| 239 | /* TODO we should move it to wpan_dev_hard_header and dev_hard_header |
| 240 | * functions. The reason is wireshark will show a mac header which is |
| 241 | * with security fields but the payload is not encrypted. |
| 242 | */ |
| 243 | rc = mac802154_llsec_encrypt(&sdata->sec, skb); |
| 244 | if (rc) { |
| 245 | netdev_warn(dev, "encryption failed: %i\n", rc); |
| 246 | kfree_skb(skb); |
| 247 | return NETDEV_TX_OK; |
| 248 | } |
| 249 | |
| 250 | skb->skb_iif = dev->ifindex; |
| 251 | |
| 252 | return ieee802154_hot_tx(sdata->local, skb); |
| 253 | } |