tx_rings[num].stats64, packets), \
STAT_GENET_SOFT_MIB64("txq" __stringify(num) "_bytes", \
tx_rings[num].stats64, bytes), \
+ STAT_GENET_SOFT_MIB64("txq" __stringify(num) "_errors", \
+ tx_rings[num].stats64, errors), \
+ STAT_GENET_SOFT_MIB64("txq" __stringify(num) "_dropped", \
+ tx_rings[num].stats64, dropped), \
STAT_GENET_SOFT_MIB64("rxq" __stringify(num) "_bytes", \
rx_rings[num].stats64, bytes), \
STAT_GENET_SOFT_MIB64("rxq" __stringify(num) "_packets", \
STAT_GENET_SOFT_MIB64("rxq" __stringify(num) "_errors", \
rx_rings[num].stats64, errors), \
STAT_GENET_SOFT_MIB64("rxq" __stringify(num) "_dropped", \
- rx_rings[num].stats64, dropped)
+ rx_rings[num].stats64, dropped), \
+ STAT_GENET_SOFT_MIB64("rxq" __stringify(num) "_multicast", \
+ rx_rings[num].stats64, multicast), \
+ STAT_GENET_SOFT_MIB64("rxq" __stringify(num) "_missed", \
+ rx_rings[num].stats64, missed), \
+ STAT_GENET_SOFT_MIB64("rxq" __stringify(num) "_length_errors", \
+ rx_rings[num].stats64, length_errors), \
+ STAT_GENET_SOFT_MIB64("rxq" __stringify(num) "_over_errors", \
+ rx_rings[num].stats64, over_errors), \
+ STAT_GENET_SOFT_MIB64("rxq" __stringify(num) "_crc_errors", \
+ rx_rings[num].stats64, crc_errors), \
+ STAT_GENET_SOFT_MIB64("rxq" __stringify(num) "_frame_errors", \
+ rx_rings[num].stats64, frame_errors), \
+ STAT_GENET_SOFT_MIB64("rxq" __stringify(num) "_fragmented_errors", \
+ rx_rings[num].stats64, fragmented_errors), \
+ STAT_GENET_SOFT_MIB64("rxq" __stringify(num) "_broadcast", \
+ rx_rings[num].stats64, broadcast)
/* There is a 0xC gap between the end of RX and beginning of TX stats and then
* between the end of TX stats and the beginning of the RX RUNT
STAT_RTNL(rx_dropped),
STAT_RTNL(tx_dropped),
STAT_RTNL(multicast),
+ STAT_RTNL(rx_missed_errors),
+ STAT_RTNL(rx_length_errors),
+ STAT_RTNL(rx_over_errors),
+ STAT_RTNL(rx_crc_errors),
+ STAT_RTNL(rx_frame_errors),
/* UniMAC RSV counters */
STAT_GENET_MIB_RX("rx_64_octets", mib.rx.pkt_cnt.cnt_64),
STAT_GENET_MIB_RX("rx_65_127_oct", mib.rx.pkt_cnt.cnt_127),
if (unlikely(!(dma_flag & DMA_EOP) || !(dma_flag & DMA_SOP))) {
netif_err(priv, rx_status, dev,
"dropping fragmented packet!\n");
- BCMGENET_STATS64_INC(stats, errors);
+ BCMGENET_STATS64_INC(stats, fragmented_errors);
dev_kfree_skb_any(skb);
goto next;
}
u64_stats_add(&stats->bytes, len);
if (dma_flag & DMA_RX_MULT)
u64_stats_inc(&stats->multicast);
+ else if (dma_flag & DMA_RX_BRDCAST)
+ u64_stats_inc(&stats->broadcast);
u64_stats_update_end(&stats->syncp);
/* Notify kernel */
struct bcmgenet_tx_stats64 *tx_stats;
struct bcmgenet_rx_stats64 *rx_stats;
u64 rx_length_errors, rx_over_errors;
+ u64 rx_missed, rx_fragmented_errors;
u64 rx_crc_errors, rx_frame_errors;
u64 tx_errors, tx_dropped;
u64 rx_errors, rx_dropped;
unsigned int start;
unsigned int q;
u64 multicast;
- u64 rx_missed;
for (q = 0; q <= priv->hw_params->tx_queues; q++) {
tx_stats = &priv->tx_rings[q].stats64;
rx_over_errors = u64_stats_read(&rx_stats->over_errors);
rx_crc_errors = u64_stats_read(&rx_stats->crc_errors);
rx_frame_errors = u64_stats_read(&rx_stats->frame_errors);
+ rx_fragmented_errors = u64_stats_read(&rx_stats->fragmented_errors);
multicast = u64_stats_read(&rx_stats->multicast);
} while (u64_stats_fetch_retry(&rx_stats->syncp, start));
rx_errors += rx_length_errors;
rx_errors += rx_crc_errors;
rx_errors += rx_frame_errors;
+ rx_errors += rx_fragmented_errors;
stats->rx_bytes += rx_bytes;
stats->rx_packets += rx_packets;