1 From bf3f81f3773cc9f6b273d769aca96512780c6189 Mon Sep 17 00:00:00 2001
2 From: Po Liu <Po.Liu@nxp.com>
3 Date: Tue, 3 Dec 2019 16:52:57 +0800
4 Subject: [PATCH] enetc: add support tsn capabilities qbv/qci/qbu/cbs
6 Support Qbv/Qci/Qbu/Credit Base Shaper etc.
7 This patch using the generic netlink adapt layer driver net/tsn/*
8 and include/net/tsn.h interface load by user space. The user space
9 refer the include/uapi/linux/tsn.h.
11 Signed-off-by: Po Liu <Po.Liu@nxp.com>
13 drivers/net/ethernet/freescale/enetc/Kconfig | 10 +
14 drivers/net/ethernet/freescale/enetc/Makefile | 1 +
15 drivers/net/ethernet/freescale/enetc/enetc.c | 13 +-
16 drivers/net/ethernet/freescale/enetc/enetc.h | 38 +
17 .../net/ethernet/freescale/enetc/enetc_ethtool.c | 59 +
18 drivers/net/ethernet/freescale/enetc/enetc_hw.h | 438 ++++-
19 drivers/net/ethernet/freescale/enetc/enetc_pf.c | 15 +-
20 drivers/net/ethernet/freescale/enetc/enetc_tsn.c | 2049 ++++++++++++++++++++
21 8 files changed, 2614 insertions(+), 9 deletions(-)
22 create mode 100644 drivers/net/ethernet/freescale/enetc/enetc_tsn.c
24 --- a/drivers/net/ethernet/freescale/enetc/Kconfig
25 +++ b/drivers/net/ethernet/freescale/enetc/Kconfig
26 @@ -60,3 +60,13 @@ config FSL_ENETC_QOS
27 enable/disable from user space via Qos commands(tc). In the kernel
28 side, it can be loaded by Qos driver. Currently, it is only support
29 taprio(802.1Qbv) and Credit Based Shaper(802.1Qbu).
32 + bool "TSN Support for NXP ENETC driver"
34 + depends on TSN && FSL_ENETC
36 + This driver supports TSN on Freescale ENETC driver. Provide
37 + interface to config the tsn capabilities of ENETC. The interface link
38 + to the /net/tsn/* and include/net/tsn.h. User space refer the
39 + include/uapi/linux/tsn.h.
40 --- a/drivers/net/ethernet/freescale/enetc/Makefile
41 +++ b/drivers/net/ethernet/freescale/enetc/Makefile
42 @@ -6,6 +6,7 @@ obj-$(CONFIG_FSL_ENETC) += fsl-enetc.o
43 fsl-enetc-y := enetc_pf.o enetc_mdio.o $(common-objs)
44 fsl-enetc-$(CONFIG_PCI_IOV) += enetc_msg.o
45 fsl-enetc-$(CONFIG_FSL_ENETC_QOS) += enetc_qos.o
46 +fsl-enetc-$(CONFIG_ENETC_TSN) += enetc_tsn.o
48 obj-$(CONFIG_FSL_ENETC_VF) += fsl-enetc-vf.o
49 fsl-enetc-vf-y := enetc_vf.o $(common-objs)
50 --- a/drivers/net/ethernet/freescale/enetc/enetc.c
51 +++ b/drivers/net/ethernet/freescale/enetc/enetc.c
52 @@ -145,7 +145,8 @@ static int enetc_map_tx_buffs(struct ene
53 do_tstamp = (active_offloads & ENETC_F_TX_TSTAMP) &&
54 (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP);
55 tx_swbd->do_tstamp = do_tstamp;
56 - tx_swbd->check_wb = tx_swbd->do_tstamp;
57 + tx_swbd->qbv_en = !!(active_offloads & ENETC_F_QBV);
58 + tx_swbd->check_wb = tx_swbd->do_tstamp || tx_swbd->qbv_en;
60 if (do_vlan || do_tstamp)
61 flags |= ENETC_TXBD_FLAGS_EX;
62 @@ -342,7 +343,7 @@ static void enetc_tstamp_tx(struct sk_bu
63 static bool enetc_clean_tx_ring(struct enetc_bdr *tx_ring, int napi_budget)
65 struct net_device *ndev = tx_ring->ndev;
66 - int tx_frm_cnt = 0, tx_byte_cnt = 0;
67 + int tx_frm_cnt = 0, tx_byte_cnt = 0, tx_win_drop = 0;
68 struct enetc_tx_swbd *tx_swbd;
71 @@ -372,6 +373,10 @@ static bool enetc_clean_tx_ring(struct e
76 + if (tx_swbd->qbv_en &&
77 + txbd->wb.status & ENETC_TXBD_STATS_WIN)
81 if (likely(tx_swbd->dma))
82 @@ -415,6 +420,7 @@ static bool enetc_clean_tx_ring(struct e
83 tx_ring->next_to_clean = i;
84 tx_ring->stats.packets += tx_frm_cnt;
85 tx_ring->stats.bytes += tx_byte_cnt;
86 + tx_ring->stats.win_drop += tx_win_drop;
88 if (unlikely(tx_frm_cnt && netif_carrier_ok(ndev) &&
89 __netif_subqueue_stopped(ndev, tx_ring->index) &&
90 @@ -778,6 +784,9 @@ void enetc_get_si_caps(struct enetc_si *
92 if (val & ENETC_SIPCAPR0_QBV)
93 si->hw_features |= ENETC_SI_F_QBV;
95 + if (val & ENETC_SIPCAPR0_QBU)
96 + si->hw_features |= ENETC_SI_F_QBU;
99 static int enetc_dma_alloc_bdr(struct enetc_bdr *r, size_t bd_size)
100 --- a/drivers/net/ethernet/freescale/enetc/enetc.h
101 +++ b/drivers/net/ethernet/freescale/enetc/enetc.h
103 #include <linux/ethtool.h>
104 #include <linux/if_vlan.h>
105 #include <linux/phy.h>
106 +#include <net/tsn.h>
108 #include "enetc_hw.h"
110 @@ -24,6 +25,7 @@ struct enetc_tx_swbd {
117 #define ENETC_RX_MAXFRM_SIZE ENETC_MAC_MAXFRM_SIZE
118 @@ -42,6 +44,7 @@ struct enetc_ring_stats {
119 unsigned int packets;
121 unsigned int rx_alloc_errs;
122 + unsigned int win_drop;
125 #define ENETC_BDR_DEFAULT_SIZE 1024
126 @@ -111,6 +114,28 @@ struct enetc_msg_swbd {
130 +#ifdef CONFIG_ENETC_TSN
131 +/* Credit-Based Shaper parameters */
140 + u32 tc_max_sized_frame;
141 + u32 max_interfrence_size;
145 + u32 port_transmit_rate;
146 + u32 port_max_size_frame;
152 #define ENETC_REV1 0x1
154 ENETC_ERR_TXCSUM = BIT(0),
155 @@ -119,6 +144,7 @@ enum enetc_errata {
158 #define ENETC_SI_F_QBV BIT(0)
159 +#define ENETC_SI_F_QBU BIT(1)
161 /* PCI IEP device data */
163 @@ -136,6 +162,10 @@ struct enetc_si {
164 int num_rss; /* number of RSS buckets */
167 +#ifdef CONFIG_ENETC_TSN
168 + struct enetc_cbs *ecbs;
173 #define ENETC_SI_ALIGN 32
174 @@ -177,6 +207,7 @@ enum enetc_active_offloads {
175 ENETC_F_RX_TSTAMP = BIT(0),
176 ENETC_F_TX_TSTAMP = BIT(1),
177 ENETC_F_QBV = BIT(2),
178 + ENETC_F_QBU = BIT(3),
181 struct enetc_ndev_priv {
182 @@ -266,3 +297,10 @@ int enetc_setup_tc_cbs(struct net_device
183 #define enetc_sched_speed_set(ndev) (void)0
184 #define enetc_setup_tc_cbs(ndev, type_data) -EOPNOTSUPP
186 +#ifdef CONFIG_ENETC_TSN
187 +void enetc_tsn_pf_init(struct net_device *netdev, struct pci_dev *pdev);
188 +void enetc_tsn_pf_deinit(struct net_device *netdev);
190 +#define enetc_tsn_pf_init(netdev, pdev) (void)0
191 +#define enetc_tsn_pf_deinit(netdev) (void)0
193 --- a/drivers/net/ethernet/freescale/enetc/enetc_ethtool.c
194 +++ b/drivers/net/ethernet/freescale/enetc/enetc_ethtool.c
195 @@ -187,6 +187,21 @@ static const struct {
196 { ENETC_PICDR(3), "ICM DR3 discarded frames" },
199 +static const struct {
201 + char name[ETH_GSTRING_LEN];
202 +} enetc_pmac_counters[] = {
203 + { ENETC_PM1_RFRM, "PMAC rx frames" },
204 + { ENETC_PM1_RPKT, "PMAC rx packets" },
205 + { ENETC_PM1_RDRP, "PMAC rx dropped packets" },
206 + { ENETC_PM1_RFRG, "PMAC rx fragment packets" },
207 + { ENETC_PM1_TFRM, "PMAC tx frames" },
208 + { ENETC_PM1_TERR, "PMAC tx error frames" },
209 + { ENETC_PM1_TPKT, "PMAC tx packets" },
210 + { ENETC_MAC_MERGE_MMFCRXR, "MAC merge fragment rx counter" },
211 + { ENETC_MAC_MERGE_MMFCTXR, "MAC merge fragment tx counter"},
214 static const char rx_ring_stats[][ETH_GSTRING_LEN] = {
215 "Rx ring %2d frames",
216 "Rx ring %2d alloc errors",
217 @@ -196,6 +211,10 @@ static const char tx_ring_stats[][ETH_GS
218 "Tx ring %2d frames",
221 +static const char tx_windrop_stats[][ETH_GSTRING_LEN] = {
222 + "Tx window drop %2d frames",
225 static int enetc_get_sset_count(struct net_device *ndev, int sset)
227 struct enetc_ndev_priv *priv = netdev_priv(ndev);
228 @@ -213,6 +232,12 @@ static int enetc_get_sset_count(struct n
230 len += ARRAY_SIZE(enetc_port_counters);
232 + if (priv->active_offloads & ENETC_F_QBU)
233 + len += ARRAY_SIZE(enetc_pmac_counters);
235 + if (priv->active_offloads & ENETC_F_QBV)
236 + len += ARRAY_SIZE(tx_windrop_stats) * priv->num_tx_rings;
241 @@ -251,6 +276,28 @@ static void enetc_get_strings(struct net
243 p += ETH_GSTRING_LEN;
246 + if (!(priv->active_offloads & ENETC_F_QBU))
249 + for (i = 0; i < ARRAY_SIZE(enetc_pmac_counters); i++) {
250 + strlcpy(p, enetc_pmac_counters[i].name,
252 + p += ETH_GSTRING_LEN;
255 + if (!((priv->active_offloads & ENETC_F_QBV)))
258 + for (i = 0; i < priv->num_tx_rings; i++) {
259 + for (j = 0; j < ARRAY_SIZE(tx_windrop_stats); j++) {
260 + snprintf(p, ETH_GSTRING_LEN,
261 + tx_windrop_stats[j],
263 + p += ETH_GSTRING_LEN;
270 @@ -278,6 +325,18 @@ static void enetc_get_ethtool_stats(stru
272 for (i = 0; i < ARRAY_SIZE(enetc_port_counters); i++)
273 data[o++] = enetc_port_rd(hw, enetc_port_counters[i].reg);
275 + if (!(priv->active_offloads & ENETC_F_QBU))
278 + for (i = 0; i < ARRAY_SIZE(enetc_pmac_counters); i++)
279 + data[o++] = enetc_port_rd(hw, enetc_pmac_counters[i].reg);
281 + if (!((priv->active_offloads & ENETC_F_QBV)))
284 + for (i = 0; i < priv->num_tx_rings; i++)
285 + data[o++] = priv->tx_ring[i]->stats.win_drop;
288 #define ENETC_RSSHASH_L3 (RXH_L2DA | RXH_VLAN | RXH_L3_PROTO | RXH_IP_SRC | \
289 --- a/drivers/net/ethernet/freescale/enetc/enetc_hw.h
290 +++ b/drivers/net/ethernet/freescale/enetc/enetc_hw.h
292 #define ENETC_SICTR1 0x1c
293 #define ENETC_SIPCAPR0 0x20
294 #define ENETC_SIPCAPR0_QBV BIT(4)
295 +#define ENETC_SIPCAPR0_QBU BIT(3)
296 #define ENETC_SIPCAPR0_RSS BIT(8)
297 #define ENETC_SIPCAPR1 0x24
298 #define ENETC_SITGTGR 0x30
299 @@ -243,10 +244,20 @@ enum enetc_bdr_type {TX, RX};
300 #define ENETC_PCS_IF_MODE_SGMII_AN 0x0003
302 #define ENETC_PM0_IF_MODE 0x8300
303 +#define ENETC_PM1_IF_MODE 0x9300
304 #define ENETC_PMO_IFM_RG BIT(2)
305 #define ENETC_PM0_IFM_RLP (BIT(5) | BIT(11))
306 #define ENETC_PM0_IFM_RGAUTO (BIT(15) | ENETC_PMO_IFM_RG | BIT(1))
307 #define ENETC_PM0_IFM_XGMII BIT(12)
308 +#define ENETC_PSIDCAPR 0x1b08
309 +#define ENETC_PSIDCAPR_MSK GENMASK(15, 0)
310 +#define ENETC_PSFCAPR 0x1b18
311 +#define ENETC_PSFCAPR_MSK GENMASK(15, 0)
312 +#define ENETC_PSGCAPR 0x1b28
313 +#define ENETC_PSGCAPR_GCL_MSK GENMASK(18, 16)
314 +#define ENETC_PSGCAPR_SGIT_MSK GENMASK(15, 0)
315 +#define ENETC_PFMCAPR 0x1b38
316 +#define ENETC_PFMCAPR_MSK GENMASK(15, 0)
319 #define ENETC_PM0_REOCT 0x8100
320 @@ -300,6 +311,15 @@ enum enetc_bdr_type {TX, RX};
321 #define ENETC_PM0_TSCOL 0x82E0
322 #define ENETC_PM0_TLCOL 0x82E8
323 #define ENETC_PM0_TECOL 0x82F0
324 +#define ENETC_PM1_RFRM 0x9120
325 +#define ENETC_PM1_RDRP 0x9158
326 +#define ENETC_PM1_RPKT 0x9160
327 +#define ENETC_PM1_RFRG 0x91B8
328 +#define ENETC_PM1_TFRM 0x9220
329 +#define ENETC_PM1_TERR 0x9238
330 +#define ENETC_PM1_TPKT 0x9260
331 +#define ENETC_MAC_MERGE_MMFCRXR 0x1f14
332 +#define ENETC_MAC_MERGE_MMFCTXR 0x1f18
335 #define ENETC_PICDR(n) (0x0700 + (n) * 8) /* n = [0..3] */
336 @@ -458,6 +478,7 @@ union enetc_tx_bd {
337 #define ENETC_TXBD_FLAGS_CSUM BIT(3)
338 #define ENETC_TXBD_FLAGS_EX BIT(6)
339 #define ENETC_TXBD_FLAGS_F BIT(7)
340 +#define ENETC_TXBD_STATS_WIN BIT(7)
342 static inline void enetc_clear_tx_bd(union enetc_tx_bd *txbd)
344 @@ -485,6 +506,8 @@ static inline __le16 enetc_txbd_l3_csoff
345 #define ENETC_TXBD_L4_UDP BIT(5)
346 #define ENETC_TXBD_L4_TCP BIT(6)
348 +#define enetc_tsn_is_enabled() IS_ENABLED(CONFIG_ENETC_TSN)
353 @@ -631,21 +654,307 @@ enum bdcr_cmd_class {
356 BDCR_CMD_RECV_CLASSIFIER,
357 + BDCR_CMD_STREAM_IDENTIFY,
358 + BDCR_CMD_STREAM_FILTER,
359 + BDCR_CMD_STREAM_GCL,
360 + BDCR_CMD_FLOW_METER,
362 BDCR_CMD_MAX_LEN = __BDCR_CMD_MAX_LEN - 1,
365 +/* class 7, command 0, Stream Identity Entry Configuration */
366 +struct streamid_conf {
367 + __le32 stream_handle; /* init gate value */
375 +#define ENETC_CBDR_SID_VID_MASK 0xfff
376 +#define ENETC_CBDR_SID_VIDM BIT(12)
377 +#define ENETC_CBDR_SID_TG_MASK 0xc000
378 +/* streamid_conf address point to this data space */
379 +struct null_streamid_data {
384 +struct smac_streamid_data {
389 +/* class 7, command 1, query config , long format */
390 +/* No need structure define */
392 +#define ENETC_CDBR_SID_ENABLE BIT(7)
393 +/* Stream ID Query Response Data Buffer */
394 +struct streamid_query_resp {
405 +/* class 7, command 2, qeury status count, Stream ID query long format */
406 +struct streamid_stat_query {
408 + __le32 input_ports;
411 +/* Stream Identity Statistics Query */
412 +struct streamid_stat_query_resp {
418 +#define ENETC_CBDR_SFI_PRI_MASK 0x7
419 +#define ENETC_CBDR_SFI_PRIM BIT(3)
420 +#define ENETC_CBDR_SFI_BLOV BIT(4)
421 +#define ENETC_CBDR_SFI_BLEN BIT(5)
422 +#define ENETC_CBDR_SFI_MSDUEN BIT(6)
423 +#define ENETC_CBDR_SFI_FMITEN BIT(7)
424 +#define ENETC_CBDR_SFI_ENABLE BIT(7)
425 +/* class 8, command 0, Stream Filter Instance, Short Format */
427 + __le32 stream_handle;
431 + /* Max Service Data Unit or Flow Meter Instance Table index.
432 + * Depending on the value of FLT this represents either Max
433 + * Service Data Unit (max frame size) allowed by the filter
434 + * entry or is an index into the Flow Meter Instance table
435 + * index identifying the policer which will be used to police
438 + __le16 fm_inst_table_index;
440 + __le16 sg_inst_table_index;
442 + __le32 input_ports;
447 +/* class 8, command 1, Stream Filter Instance, write back, short Format */
453 + u16 fm_inst_table_index;
455 + u16 sg_inst_table_index;
462 +/* class 8, command 2 stream Filter Instance status query short format
463 + * command no need structure define
464 + * Stream Filter Instance Query Statistics Response data
466 +struct sfi_counter_data {
471 + u32 stream_gate_dropl;
472 + u32 stream_gate_droph;
473 + u32 flow_meter_dropl;
474 + u32 flow_meter_droph;
477 +#define ENETC_CBDR_SGI_OIPV_MASK 0x7
478 +#define ENETC_CBDR_SGI_OIPV_EN BIT(3)
479 +#define ENETC_CBDR_SGI_CGTST BIT(6)
480 +#define ENETC_CBDR_SGI_OGTST BIT(7)
481 +#define ENETC_CBDR_SGI_CFG_CHG BIT(1)
482 +#define ENETC_CBDR_SGI_CFG_PND BIT(2)
483 +#define ENETC_CBDR_SGI_OEX BIT(4)
484 +#define ENETC_CBDR_SGI_OEXEN BIT(5)
485 +#define ENETC_CBDR_SGI_IRX BIT(6)
486 +#define ENETC_CBDR_SGI_IRXEN BIT(7)
487 +#define ENETC_CBDR_SGI_ACLLEN_MASK 0x3
488 +#define ENETC_CBDR_SGI_OCLLEN_MASK 0xc
489 +#define ENETC_CBDR_SGI_EN BIT(7)
490 +/* class 9, command 0, Stream Gate Instance Table, Short Format
491 + * class 9, command 2, Stream Gate Instance Table entry query write back
506 +#define ENETC_CBDR_SGI_AIPV_MASK 0x7
507 +#define ENETC_CBDR_SGI_AIPV_EN BIT(3)
508 +#define ENETC_CBDR_SGI_AGTST BIT(7)
510 +/* class 9, command 1, Stream Gate Control List, Long Format */
522 + u8 cct[8]; /* Config change time */
526 +/* stream control list class 9 , cmd 1 data buffer */
532 + /*struct sgce *sgcl;*/
535 +/* class 9, command 2, stream gate instant table enery query, short format
536 + * write back see struct sgi_table. Do not need define.
537 + * class 9, command 3 Stream Gate Control List Query Descriptor - Long Format
538 + * ocl_len or acl_len to be 0, oper or admin would not show in the data space
539 + * true len will be write back in the space.
547 +/* define for 'stat' */
548 +#define ENETC_CBDR_SGIQ_AIPV_MASK 0x7
549 +#define ENETC_CBDR_SGIQ_AIPV_EN BIT(3)
550 +#define ENETC_CBDR_SGIQ_AGTST BIT(4)
551 +#define ENETC_CBDR_SGIQ_ACL_LEN_MASK 0x60
552 +#define ENETC_CBDR_SGIQ_OIPV_MASK 0x380
553 +#define ENETC_CBDR_SGIQ_OIPV_EN BIT(10)
554 +#define ENETC_CBDR_SGIQ_OGTST BIT(11)
555 +#define ENETC_CBDR_SGIQ_OCL_LEN_MASK 0x3000
556 +/* class 9, command 3 data space */
557 +struct sgcl_query_resp {
572 +/* class 9, command 4 Stream Gate Instance Table Query Statistics Response
573 + * short command, write back, no command define
575 +struct sgi_query_stat_resp {
584 +#define ENETC_CBDR_FMI_MR BIT(0)
585 +#define ENETC_CBDR_FMI_MREN BIT(1)
586 +#define ENETC_CBDR_FMI_DOY BIT(2)
587 +#define ENETC_CBDR_FMI_CM BIT(3)
588 +#define ENETC_CBDR_FMI_CF BIT(4)
589 +#define ENETC_CBDR_FMI_NDOR BIT(5)
590 +#define ENETC_CBDR_FMI_OALEN BIT(6)
591 +#define ENETC_CBDR_FMI_IRFPP_MASK 0x1f
592 +/* class 10: command 0/1, Flow Meter Instance Set, short Format */
605 +/* class:10, command:2, Flow Meter Instance Statistics Query Response */
606 +struct fmi_query_stat_resp {
630 /* class 5, command 0 */
631 struct tgs_gcl_conf {
632 u8 atc; /* init gate value */
651 +#define ENETC_CBDR_SGL_IOMEN BIT(0)
652 +#define ENETC_CBDR_SGL_IPVEN BIT(3)
653 +#define ENETC_CBDR_SGL_GTST BIT(4)
654 +#define ENETC_CBDR_SGL_IPV_MASK 0xe
655 +/* Stream Gate Control List Entry */
662 /* gate control list entry */
665 @@ -662,13 +971,55 @@ struct tgs_gcl_data {
669 +/* class 5, command 1 */
670 +struct tgs_gcl_query {
674 + __le16 acl_len; /* admin list length */
675 + __le16 ocl_len; /* operation list length */
678 + u16 admin_list_len;
685 +/* tgs_gcl_query command response data format */
686 +struct tgs_gcl_resp {
687 + u32 abtl; /* base time */
689 + u32 act; /* cycle time */
690 + u32 acte; /* cycle time extend */
691 + u32 cctl; /* config change time */
693 + u32 obtl; /* operation base time */
695 + u32 oct; /* operation cycle time */
696 + u32 octe; /* operation cycle time extend */
697 + u32 ccel; /* config change error */
699 + /*struct gce *gcl;*/
704 + struct sfi_conf sfi_conf;
705 + struct sgi_table sgi_table;
706 + struct sgi_query_stat_resp sgi_query_stat_resp;
707 + struct fmi_conf fmi_conf;
712 - struct tgs_gcl_conf gcl_conf;
713 + struct tgs_gcl_conf gcl_conf;
714 + struct tgs_gcl_query gcl_query;
715 + struct streamid_conf sid_set;
716 + struct streamid_stat_query sid_stat;
717 + struct sgcl_conf sgcl_conf;
718 + struct sgcl_query sgcl_query;
722 @@ -683,11 +1034,88 @@ struct enetc_cbd {
724 #define ENETC_CLK 400000000ULL
726 +#define ENETC_PTCFPR(n) (0x1910 + (n) * 4) /* n = [0 ..7] */
727 +#define ENETC_FPE BIT(31)
729 +/* Port capability register 0 */
730 +#define ENETC_PCAPR0_PSFPM BIT(10)
731 +#define ENETC_PCAPR0_PSFP BIT(9)
732 +#define ENETC_PCAPR0_TSN BIT(4)
733 +#define ENETC_PCAPR0_QBU BIT(3)
735 /* port time gating control register */
736 #define ENETC_QBV_PTGCR_OFFSET 0x11a00
737 #define ENETC_QBV_TGE BIT(31)
738 #define ENETC_QBV_TGPE BIT(30)
739 +#define ENETC_QBV_TGDROP_DISABLE BIT(29)
741 /* Port time gating capability register */
742 #define ENETC_QBV_PTGCAPR_OFFSET 0x11a08
743 #define ENETC_QBV_MAX_GCL_LEN_MASK GENMASK(15, 0)
745 +/* Port time gating tick granularity register */
746 +#define ENETC_QBV_PTGTGR_OFFSET 0x11a0c
747 +#define ENETC_QBV_TICK_GRAN_MASK 0xffffffff
749 +/* Port time gating admin gate list status register */
750 +#define ENETC_QBV_PTGAGLSR_OFFSET 0x11a10
752 +#define ENETC_QBV_CFG_PEND_MASK 0x00000002
754 +/* Port time gating admin gate list length register */
755 +#define ENETC_QBV_PTGAGLLR_OFFSET 0x11a14
756 +#define ENETC_QBV_ADMIN_GATE_LIST_LENGTH_MASK 0xffff
758 +/* Port time gating operational gate list status register */
759 +#define ENETC_QBV_PTGOGLSR_OFFSET 0x11a18
760 +#define ENETC_QBV_HTA_POS_MASK 0xffff0000
762 +#define ENETC_QBV_CURR_POS_MASK 0x0000ffff
764 +/* Port time gating operational gate list length register */
765 +#define ENETC_QBV_PTGOGLLR_OFFSET 0x11a1c
766 +#define ENETC_QBV_OPER_GATE_LIST_LENGTH_MASK 0xffff
768 +/* Port time gating current time register */
769 +#define ENETC_QBV_PTGCTR_OFFSET 0x11a20
770 +#define ENETC_QBV_CURR_TIME_MASK 0xffffffffffffffff
772 +/* Port traffic class a time gating control register */
773 +#define ENETC_QBV_PTC0TGCR_OFFSET 0x11a40
774 +#define ENETC_QBV_PTC1TGCR_OFFSET 0x11a50
775 +#define ENETC_QBV_PTC2TGCR_OFFSET 0x11a60
776 +#define ENETC_QBV_PTC3TGCR_OFFSET 0x11a70
777 +#define ENETC_QBV_PTC4TGCR_OFFSET 0x11a80
778 +#define ENETC_QBV_PTC5TGCR_OFFSET 0x11a90
779 +#define ENETC_QBV_PTC6TGCR_OFFSET 0x11aa0
780 +#define ENETC_QBV_PTC7TGCR_OFFSET 0x11ab0
782 +/* Maximum Service Data Unit. */
783 +#define ENETC_PTC0MSDUR 0x12020
784 +#define ENETC_PTC1MSDUR 0x12024
785 +#define ENETC_PTC2MSDUR 0x12028
786 +#define ENETC_PTC3MSDUR 0x1202c
787 +#define ENETC_PTC4MSDUR 0x12030
788 +#define ENETC_PTC5MSDUR 0x12034
789 +#define ENETC_PTC6MSDUR 0x12038
790 +#define ENETC_PTC7MSDUR 0x1203c
792 +#define ENETC_QBV_MAXSDU_MASK 0xffff
794 +/* Port traffic class a time gating status register */
795 +#define ENETC_QBV_PTC0TGSR_OFFSET 0x11a44
796 +#define ENETC_QBV_HTA_STATE_MASK 0x10000
797 +#define ENETC_QBV_CURR_STATE_MASK 0x1
799 +/* Port traffic class a time gating transmission overrun counter register*/
800 +#define ENETC_QBV_PTC0TGTOCR_OFFSET 0x11a48
801 +#define ENETC_QBV_TX_OVERRUN_MASK 0xffffffffffffffff
802 +#define ENETC_TGLSTR 0xa200
803 +#define ENETC_TGS_MIN_DIS_MASK 0x80000000
804 +#define ENETC_MIN_LOOKAHEAD_MASK 0xffff
806 +#define ENETC_PPSFPMR 0x11b00
807 +#define ENETC_PPSFPMR_PSFPEN BIT(0)
808 +#define ENETC_PPSFPMR_VS BIT(1)
809 +#define ENETC_PPSFPMR_PVC BIT(2)
810 +#define ENETC_PPSFPMR_PVZC BIT(3)
811 --- a/drivers/net/ethernet/freescale/enetc/enetc_pf.c
812 +++ b/drivers/net/ethernet/freescale/enetc/enetc_pf.c
813 @@ -525,12 +525,16 @@ static void enetc_configure_port_mac(str
814 ENETC_PM0_TX_EN | ENETC_PM0_RX_EN);
815 /* set auto-speed for RGMII */
816 if (enetc_port_rd(hw, ENETC_PM0_IF_MODE) & ENETC_PMO_IFM_RG ||
817 - phy_mode == PHY_INTERFACE_MODE_RGMII)
818 + phy_mode == PHY_INTERFACE_MODE_RGMII) {
819 enetc_port_wr(hw, ENETC_PM0_IF_MODE, ENETC_PM0_IFM_RGAUTO);
820 + enetc_port_wr(hw, ENETC_PM1_IF_MODE, ENETC_PM0_IFM_RGAUTO);
823 if (phy_mode == PHY_INTERFACE_MODE_XGMII ||
824 - phy_mode == PHY_INTERFACE_MODE_USXGMII)
825 + phy_mode == PHY_INTERFACE_MODE_USXGMII) {
826 enetc_port_wr(hw, ENETC_PM0_IF_MODE, ENETC_PM0_IFM_XGMII);
827 + enetc_port_wr(hw, ENETC_PM1_IF_MODE, ENETC_PM0_IFM_XGMII);
831 static void enetc_configure_port_pmac(struct enetc_hw *hw)
832 @@ -749,6 +753,9 @@ static void enetc_pf_netdev_setup(struct
833 if (si->hw_features & ENETC_SI_F_QBV)
834 priv->active_offloads |= ENETC_F_QBV;
836 + if (enetc_tsn_is_enabled() && (si->hw_features & ENETC_SI_F_QBU))
837 + priv->active_offloads |= ENETC_F_QBU;
839 /* pick up primary MAC address from SI */
840 enetc_get_primary_mac_addr(&si->hw, ndev->dev_addr);
842 @@ -1027,6 +1034,8 @@ static int enetc_pf_probe(struct pci_dev
843 netif_info(priv, probe, ndev, "%s v%s\n",
844 enetc_drv_name, enetc_drv_ver);
846 + enetc_tsn_pf_init(ndev, pdev);
851 @@ -1063,6 +1072,8 @@ static void enetc_pf_remove(struct pci_d
852 netif_info(priv, drv, si->ndev, "%s v%s remove\n",
853 enetc_drv_name, enetc_drv_ver);
855 + enetc_tsn_pf_deinit(si->ndev);
857 unregister_netdev(si->ndev);
859 enetc_mdio_remove(pf);
861 +++ b/drivers/net/ethernet/freescale/enetc/enetc_tsn.c
863 +// SPDX-License-Identifier: (GPL-2.0+ OR BSD-3-Clause)
864 +/* Copyright 2017-2019 NXP */
866 +#ifdef CONFIG_ENETC_TSN
869 +#include <net/tsn.h>
870 +#include <linux/module.h>
871 +#include <linux/irqflags.h>
872 +#include <linux/preempt.h>
874 +static u32 get_ndev_speed(struct net_device *netdev);
876 +static int alloc_cbdr(struct enetc_si *si, struct enetc_cbd **curr_cbd)
878 + struct enetc_cbdr *ring = &si->cbd_ring;
881 + i = ring->next_to_use;
882 + *curr_cbd = ENETC_CBD(*ring, i);
884 + memset(*curr_cbd, 0, sizeof(struct enetc_cbd));
888 +/* Transmit the BD control ring by writing the pir register.
889 + * Update the counters maintained by software.
891 +static int xmit_cbdr(struct enetc_si *si, int i)
893 + struct enetc_cbdr *ring = &si->cbd_ring;
894 + struct enetc_cbd *dest_cbd;
897 + i = (i + 1) % ring->bd_count;
899 + ring->next_to_use = i;
900 + /* let H/W know BD ring has been updated */
901 + enetc_wr_reg(ring->pir, i);
903 + timeout = ENETC_CBDR_TIMEOUT;
906 + if (enetc_rd_reg(ring->cir) == i)
908 + usleep_range(10, 20);
915 + nc = ring->next_to_clean;
917 + while (enetc_rd_reg(ring->cir) != nc) {
918 + dest_cbd = ENETC_CBD(*ring, nc);
919 + if (dest_cbd->status_flags & ENETC_CBD_STATUS_MASK)
922 + nc = (nc + 1) % ring->bd_count;
925 + ring->next_to_clean = nc;
930 +static inline u64 get_current_time(struct enetc_si *si)
934 + tmp = (u64)enetc_rd(&si->hw, ENETC_SICTR0);
935 + return ((u64)enetc_rd(&si->hw, ENETC_SICTR1) << 32) + tmp;
938 +/* Class 10: Flow Meter Instance Statistics Query Descriptor - Long Format */
939 +int enetc_qci_fmi_counters_get(struct net_device *ndev, u32 index,
940 + struct fmi_query_stat_resp *counters)
942 + struct enetc_cbd *cbdr;
943 + struct fmi_query_stat_resp *fmi_data;
945 + u16 data_size, dma_size;
947 + struct enetc_ndev_priv *priv = netdev_priv(ndev);
949 + curr_cbd = alloc_cbdr(priv->si, &cbdr);
951 + cbdr->index = cpu_to_le16((u16)index);
953 + cbdr->cls = BDCR_CMD_FLOW_METER;
954 + cbdr->status_flags = 0;
956 + data_size = sizeof(struct fmi_query_stat_resp);
958 + fmi_data = kzalloc(data_size, __GFP_DMA | GFP_KERNEL);
962 + dma_size = cpu_to_le16(data_size);
963 + cbdr->length = dma_size;
965 + dma = dma_map_single(&priv->si->pdev->dev, fmi_data,
966 + data_size, DMA_FROM_DEVICE);
967 + if (dma_mapping_error(&priv->si->pdev->dev, dma)) {
968 + netdev_err(priv->si->ndev, "DMA mapping failed!\n");
972 + cbdr->addr[0] = lower_32_bits(dma);
973 + cbdr->addr[1] = upper_32_bits(dma);
975 + xmit_cbdr(priv->si, curr_cbd);
977 + memcpy(counters, fmi_data, sizeof(struct fmi_query_stat_resp));
979 + memset(cbdr, 0, sizeof(*cbdr));
984 +u16 enetc_get_max_gcl_len(struct enetc_hw *hw)
986 + return (enetc_rd(hw, ENETC_QBV_PTGCAPR_OFFSET)
987 + & ENETC_QBV_MAX_GCL_LEN_MASK);
990 +void enetc_pspeed_set(struct net_device *ndev)
994 + struct enetc_ndev_priv *priv = netdev_priv(ndev);
996 + speed = get_ndev_speed(ndev);
997 + pspeed = enetc_port_rd(&priv->si->hw, ENETC_PMR)
998 + & ENETC_PMR_PSPEED_MASK;
1001 + if (pspeed != ENETC_PMR_PSPEED_1000M) {
1003 + pspeed = ENETC_PMR_PSPEED_1000M;
1007 + if (pspeed != ENETC_PMR_PSPEED_2500M) {
1009 + pspeed = ENETC_PMR_PSPEED_2500M;
1014 + if (pspeed != ENETC_PMR_PSPEED_100M) {
1016 + pspeed = ENETC_PMR_PSPEED_100M;
1020 + if (pspeed != ENETC_PMR_PSPEED_10M) {
1022 + pspeed = ENETC_PMR_PSPEED_10M;
1026 + netdev_err(ndev, "not support speed\n");
1030 + enetc_port_wr(&priv->si->hw, ENETC_PMR,
1031 + (enetc_port_rd(&priv->si->hw, ENETC_PMR)
1032 + & (~ENETC_PMR_PSPEED_MASK))
1037 +/* CBD Class 5: Time Gated Scheduling Gate Control List configuration
1038 + * Descriptor - Long Format
1040 +int enetc_qbv_set(struct net_device *ndev, struct tsn_qbv_conf *admin_conf)
1042 + struct enetc_cbd *cbdr;
1043 + struct tgs_gcl_data *gcl_data;
1044 + struct tgs_gcl_conf *gcl_config;
1051 + struct tsn_qbv_basic *admin_basic = &admin_conf->admin;
1052 + struct enetc_ndev_priv *priv = netdev_priv(ndev);
1055 + struct tsn_port *port;
1057 + port = tsn_get_port(ndev);
1059 + netdev_err(priv->si->ndev, "TSN device not registered!\n");
1063 + enetc_pspeed_set(ndev);
1065 + gcl_len = admin_basic->control_list_length;
1066 + if (gcl_len > enetc_get_max_gcl_len(&priv->si->hw))
1069 + temp = enetc_rd(&priv->si->hw, ENETC_QBV_PTGCR_OFFSET);
1070 + if (admin_conf->gate_enabled && !(temp & ENETC_QBV_TGE)) {
1071 + enetc_wr(&priv->si->hw, ENETC_QBV_PTGCR_OFFSET,
1072 + temp & (~ENETC_QBV_TGE));
1073 + usleep_range(10, 20);
1074 + enetc_wr(&priv->si->hw, ENETC_QBV_PTGCR_OFFSET,
1075 + temp | ENETC_QBV_TGE);
1076 + } else if (!admin_conf->gate_enabled) {
1077 + enetc_wr(&priv->si->hw, ENETC_QBV_PTGCR_OFFSET,
1078 + temp & (~ENETC_QBV_TGE));
1079 + memcpy(&port->nd.ntdata, admin_conf, sizeof(*admin_conf));
1080 + call_tsn_notifiers(TSN_QBV_CONFIGCHANGETIME_ARRIVE,
1085 + /* Set the maximum frame size for each traffic class index
1086 + * PTCaMSDUR[MAXSDU]. The maximum frame size cannot exceed
1087 + * 9,600 bytes (0x2580). Frames that exceed the limit are
1090 + if (admin_conf->maxsdu) {
1091 + enetc_wr(&priv->si->hw, ENETC_PTC0MSDUR, admin_conf->maxsdu);
1092 + enetc_wr(&priv->si->hw, ENETC_PTC1MSDUR, admin_conf->maxsdu);
1093 + enetc_wr(&priv->si->hw, ENETC_PTC2MSDUR, admin_conf->maxsdu);
1094 + enetc_wr(&priv->si->hw, ENETC_PTC3MSDUR, admin_conf->maxsdu);
1095 + enetc_wr(&priv->si->hw, ENETC_PTC4MSDUR, admin_conf->maxsdu);
1096 + enetc_wr(&priv->si->hw, ENETC_PTC5MSDUR, admin_conf->maxsdu);
1097 + enetc_wr(&priv->si->hw, ENETC_PTC6MSDUR, admin_conf->maxsdu);
1098 + enetc_wr(&priv->si->hw, ENETC_PTC7MSDUR, admin_conf->maxsdu);
1101 + /* Configure the (administrative) gate control list using the
1102 + * control BD descriptor.
1104 + curr_cbd = alloc_cbdr(priv->si, &cbdr);
1106 + gcl_config = &cbdr->gcl_conf;
1108 + data_size = struct_size(gcl_data, entry, gcl_len);
1110 + gcl_data = kzalloc(data_size, __GFP_DMA | GFP_KERNEL);
1114 + gce = &gcl_data->entry[0];
1116 + gcl_config->atc = admin_basic->gate_states;
1117 + gcl_config->acl_len = cpu_to_le16(gcl_len);
1119 + if (!admin_basic->base_time) {
1121 + cpu_to_le32(enetc_rd(&priv->si->hw, ENETC_SICTR0));
1123 + cpu_to_le32(enetc_rd(&priv->si->hw, ENETC_SICTR1));
1126 + cpu_to_le32(lower_32_bits(admin_basic->base_time));
1128 + cpu_to_le32(upper_32_bits(admin_basic->base_time));
1131 + gcl_data->ct = cpu_to_le32(admin_basic->cycle_time);
1132 + gcl_data->cte = cpu_to_le32(admin_basic->cycle_time_extension);
1134 + for (i = 0; i < gcl_len; i++) {
1135 + struct gce *temp_gce = gce + i;
1136 + struct tsn_qbv_entry *temp_entry;
1138 + temp_entry = admin_basic->control_list + i;
1140 + temp_gce->gate = temp_entry->gate_state;
1141 + temp_gce->period = cpu_to_le32(temp_entry->time_interval);
1144 + cbdr->length = cpu_to_le16(data_size);
1145 + cbdr->status_flags = 0;
1147 + dma = dma_map_single(&priv->si->pdev->dev, gcl_data,
1148 + data_size, DMA_TO_DEVICE);
1149 + if (dma_mapping_error(&priv->si->pdev->dev, dma)) {
1150 + netdev_err(priv->si->ndev, "DMA mapping failed!\n");
1155 + cbdr->addr[0] = lower_32_bits(dma);
1156 + cbdr->addr[1] = upper_32_bits(dma);
1158 + cbdr->cls = BDCR_CMD_PORT_GCL;
1160 + /* Updated by ENETC on completion of the configuration
1161 + * command. A zero value indicates success.
1163 + cbdr->status_flags = 0;
1165 + xmit_cbdr(priv->si, curr_cbd);
1167 + memcpy(&port->nd.ntdata, admin_conf, sizeof(*admin_conf));
1169 + tempclock = ((u64)le32_to_cpu(gcl_config->ccth)) << 32;
1170 + port->nd.ntdata.qbv_notify.admin.base_time =
1171 + le32_to_cpu(gcl_config->cctl) + tempclock;
1173 + memset(cbdr, 0, sizeof(struct enetc_cbd));
1174 + dma_unmap_single(&priv->si->pdev->dev, dma, data_size, DMA_TO_DEVICE);
1177 + call_tsn_notifiers(TSN_QBV_CONFIGCHANGETIME_ARRIVE,
1183 +/* CBD Class 5: Time Gated Scheduling Gate Control List query
1184 + * Descriptor - Long Format
1186 +int enetc_qbv_get(struct net_device *ndev, struct tsn_qbv_conf *admin_conf)
1188 + struct enetc_cbd *cbdr;
1189 + struct tgs_gcl_resp *gcl_data;
1190 + struct tgs_gcl_query *gcl_query;
1192 + struct tsn_qbv_basic *admin_basic = &admin_conf->admin;
1193 + struct enetc_ndev_priv *priv = netdev_priv(ndev);
1197 + u16 data_size, dma_size;
1203 + if (enetc_rd(&priv->si->hw, ENETC_QBV_PTGCR_OFFSET) & ENETC_QBV_TGE) {
1204 + admin_conf->gate_enabled = true;
1206 + admin_conf->gate_enabled = false;
1210 + curr_cbd = alloc_cbdr(priv->si, &cbdr);
1212 + gcl_query = &cbdr->gcl_query;
1214 + maxlen = enetc_get_max_gcl_len(&priv->si->hw);
1216 + data_size = sizeof(struct tgs_gcl_resp)
1217 + + sizeof(struct gce) * 2 * maxlen;
1219 + gcl_data = kzalloc(data_size, __GFP_DMA | GFP_KERNEL);
1223 + gce = (struct gce *)(gcl_data + 1);
1225 + gcl_query->acl_len = cpu_to_le16(maxlen);
1227 + dma_size = cpu_to_le16(data_size);
1228 + cbdr->length = dma_size;
1229 + cbdr->status_flags = 0;
1231 + dma = dma_map_single(&priv->si->pdev->dev, gcl_data,
1232 + data_size, DMA_FROM_DEVICE);
1233 + if (dma_mapping_error(&priv->si->pdev->dev, dma)) {
1234 + netdev_err(priv->si->ndev, "DMA mapping failed!\n");
1239 + cbdr->addr[0] = lower_32_bits(dma);
1240 + cbdr->addr[1] = upper_32_bits(dma);
1242 + cbdr->cls = BDCR_CMD_PORT_GCL;
1243 + xmit_cbdr(priv->si, curr_cbd);
1244 + dma_unmap_single(&priv->si->pdev->dev, dma, data_size, DMA_FROM_DEVICE);
1246 + /* since cbdr already passed to free, below could be get wrong */
1247 + admin_len = le16_to_cpu(gcl_query->admin_list_len);
1248 + oper_len = le16_to_cpu(gcl_query->oper_list_len);
1250 + admin_basic->control_list_length = admin_len;
1252 + temp = ((u64)le32_to_cpu(gcl_data->abth)) << 32;
1253 + admin_basic->base_time = le32_to_cpu(gcl_data->abtl) + temp;
1255 + admin_basic->cycle_time = le32_to_cpu(gcl_data->act);
1256 + admin_basic->cycle_time_extension = le32_to_cpu(gcl_data->acte);
1258 + admin_basic->control_list = kcalloc(admin_len,
1259 + sizeof(admin_basic->control_list),
1261 + if (!admin_basic->control_list) {
1262 + memset(cbdr, 0, sizeof(*cbdr));
1267 + for (i = 0; i < admin_len; i++) {
1268 + struct gce *temp_gce = gce + i;
1269 + struct tsn_qbv_entry *temp_entry;
1271 + temp_entry = admin_basic->control_list + i;
1273 + temp_entry->gate_state = temp_gce->gate;
1274 + temp_entry->time_interval = le32_to_cpu(temp_gce->period);
1277 + /* Updated by ENETC on completion of the configuration
1278 + * command. A zero value indicates success.
1280 + admin_conf->config_change = true;
1282 + memset(cbdr, 0, sizeof(*cbdr));
1288 +int enetc_qbv_get_status(struct net_device *ndev,
1289 + struct tsn_qbv_status *status)
1291 + struct enetc_cbd *cbdr;
1292 + struct tgs_gcl_resp *gcl_data;
1293 + struct tgs_gcl_query *gcl_query;
1295 + struct tsn_qbv_basic *oper_basic;
1296 + struct enetc_ndev_priv *priv;
1300 + u16 data_size, dma_size;
1312 + oper_basic = &status->oper;
1313 + priv = netdev_priv(ndev);
1315 + if (!(enetc_rd(&priv->si->hw, ENETC_QBV_PTGCR_OFFSET) & ENETC_QBV_TGE))
1318 + curr_cbd = alloc_cbdr(priv->si, &cbdr);
1320 + gcl_query = &cbdr->gcl_query;
1322 + maxlen = enetc_get_max_gcl_len(&priv->si->hw);
1324 + data_size = sizeof(struct tgs_gcl_resp) +
1325 + sizeof(struct gce) * 2 * maxlen;
1327 + gcl_data = kzalloc(data_size, __GFP_DMA | GFP_KERNEL);
1331 + gce = (struct gce *)(gcl_data + 1);
1333 + gcl_query->acl_len = cpu_to_le16(maxlen);
1334 + gcl_query->ocl_len = cpu_to_le16(maxlen);
1336 + dma_size = cpu_to_le16(data_size);
1337 + cbdr->length = dma_size;
1338 + cbdr->status_flags = 0; /* long format command no ie */
1340 + dma = dma_map_single(&priv->si->pdev->dev, gcl_data,
1341 + data_size, DMA_FROM_DEVICE);
1342 + if (dma_mapping_error(&priv->si->pdev->dev, dma)) {
1343 + netdev_err(priv->si->ndev, "DMA mapping failed!\n");
1348 + cbdr->addr[0] = lower_32_bits(dma);
1349 + cbdr->addr[1] = upper_32_bits(dma);
1351 + cbdr->cls = BDCR_CMD_PORT_GCL;
1352 + xmit_cbdr(priv->si, curr_cbd);
1353 + dma_unmap_single(&priv->si->pdev->dev, dma, data_size, DMA_FROM_DEVICE);
1355 + /* since cbdr already passed to free, below could be get wrong */
1356 + admin_len = le16_to_cpu(gcl_query->admin_list_len);
1357 + oper_len = le16_to_cpu(gcl_query->oper_list_len);
1359 + if (enetc_rd(&priv->si->hw, ENETC_QBV_PTGAGLSR_OFFSET) &
1360 + ENETC_QBV_CFG_PEND_MASK) {
1361 + status->config_pending = true;
1365 + /* The Oper and Admin timing fields exist in the response buffer even
1366 + * if no valid corresponding lists exists. These fields are considered
1367 + * invalid if the corresponding list does not exist.
1369 + status->config_pending = false;
1370 + temp = ((u64)le32_to_cpu(gcl_data->ccth)) << 32;
1371 + status->config_change_time = le32_to_cpu(gcl_data->cctl) + temp;
1373 + temp = ((u64)le32_to_cpu(gcl_data->cceh)) << 32;
1374 + status->config_change_error = le32_to_cpu(gcl_data->ccel) + temp;
1376 + /* changed to SITGTGR */
1377 + status->tick_granularity = enetc_rd(&priv->si->hw, ENETC_SITGTGR);
1379 + /* current time */
1380 + status->current_time = get_current_time(priv->si);
1382 + status->supported_list_max = maxlen;
1384 + /* status->oper.gate_states , no init oper/admin gate state */
1385 + status->oper.control_list_length = oper_len;
1386 + temp = ((u64)le32_to_cpu(gcl_data->obth)) << 32;
1387 + status->oper.base_time = le32_to_cpu(gcl_data->obtl) + temp;
1388 + status->oper.cycle_time = le32_to_cpu(gcl_data->oct);
1389 + status->oper.cycle_time_extension = le32_to_cpu(gcl_data->octe);
1391 + oper_basic->control_list =
1392 + kcalloc(oper_len, sizeof(oper_basic->control_list), GFP_KERNEL);
1393 + if (!oper_basic->control_list) {
1394 + memset(cbdr, 0, sizeof(*cbdr));
1399 + for (i = 0; i < oper_len; i++) {
1400 + struct gce *temp_gce = gce + maxlen + i;
1401 + struct tsn_qbv_entry *temp_entry = oper_basic->control_list + i;
1403 + temp_entry->gate_state = temp_gce->gate;
1404 + temp_entry->time_interval = le32_to_cpu(temp_gce->period);
1408 + memset(cbdr, 0, sizeof(*cbdr));
1413 +/* CBD Class 7: Stream Identity Entry Set Descriptor - Long Format */
1414 +int enetc_cb_streamid_set(struct net_device *ndev, u32 index,
1415 + bool en, struct tsn_cb_streamid *streamid)
1417 + struct enetc_cbd *cbdr;
1419 + struct null_streamid_data *si_data1;
1420 + struct smac_streamid_data *si_data2;
1421 + struct streamid_conf *si_conf;
1422 + struct enetc_ndev_priv *priv;
1424 + u16 data_size, dma_size;
1430 + priv = netdev_priv(ndev);
1432 + curr_cbd = alloc_cbdr(priv->si, &cbdr);
1434 + cbdr->index = cpu_to_le16((u16)index);
1436 + cbdr->cls = BDCR_CMD_STREAM_IDENTIFY;
1437 + cbdr->status_flags = 0;
1439 + data_size = sizeof(struct null_streamid_data);
1440 + si_data = kzalloc(data_size, __GFP_DMA | GFP_KERNEL);
1441 + cbdr->length = cpu_to_le16(data_size);
1443 + dma = dma_map_single(&priv->si->pdev->dev, si_data,
1444 + data_size, DMA_FROM_DEVICE);
1445 + if (dma_mapping_error(&priv->si->pdev->dev, dma)) {
1446 + netdev_err(priv->si->ndev, "DMA mapping failed!\n");
1451 + cbdr->addr[0] = lower_32_bits(dma);
1452 + cbdr->addr[1] = upper_32_bits(dma);
1453 + si_data1 = (struct null_streamid_data *)si_data;
1454 + si_data1->dmac[0] = 0xFF;
1455 + si_data1->dmac[1] = 0xFF;
1456 + si_data1->dmac[2] = 0xFF;
1457 + si_data1->dmac[3] = 0xFF;
1458 + si_data1->dmac[4] = 0xFF;
1459 + si_data1->dmac[5] = 0xFF;
1460 + si_data1->vid_vidm_tg =
1461 + cpu_to_le16(ENETC_CBDR_SID_VID_MASK
1462 + + ((0x3 << 14) | ENETC_CBDR_SID_VIDM));
1464 + si_conf = &cbdr->sid_set;
1465 + /* Only one port supported for one entry, set itself */
1466 + si_conf->iports = 1 << (priv->si->pdev->devfn & 0x7);
1467 + si_conf->id_type = 1;
1468 + si_conf->oui[2] = 0x0;
1469 + si_conf->oui[1] = 0x80;
1470 + si_conf->oui[0] = 0xC2;
1472 + xmit_cbdr(priv->si, curr_cbd);
1474 + memset(cbdr, 0, sizeof(*cbdr));
1480 + curr_cbd = alloc_cbdr(priv->si, &cbdr);
1482 + cbdr->index = cpu_to_le16((u16)index);
1484 + cbdr->cls = BDCR_CMD_STREAM_IDENTIFY;
1485 + cbdr->status_flags = 0;
1487 + si_conf = &cbdr->sid_set;
1488 + si_conf->en = 0x80;
1489 + si_conf->stream_handle = cpu_to_le32(streamid->handle);
1490 + si_conf->iports = 1 << (priv->si->pdev->devfn & 0x7);
1491 + si_conf->id_type = streamid->type;
1492 + si_conf->oui[2] = 0x0;
1493 + si_conf->oui[1] = 0x80;
1494 + si_conf->oui[0] = 0xC2;
1496 + if (si_conf->id_type == 1) {
1497 + data_size = sizeof(struct null_streamid_data);
1498 + si_data = kzalloc(data_size, __GFP_DMA | GFP_KERNEL);
1499 + } else if (si_conf->id_type == 2) {
1500 + data_size = sizeof(struct smac_streamid_data);
1501 + si_data = kzalloc(data_size, __GFP_DMA | GFP_KERNEL);
1509 + dma_size = cpu_to_le16(data_size);
1510 + cbdr->length = dma_size;
1511 + cbdr->status_flags = 0;
1513 + dma = dma_map_single(&priv->si->pdev->dev, si_data,
1514 + data_size, DMA_FROM_DEVICE);
1515 + if (dma_mapping_error(&priv->si->pdev->dev, dma)) {
1516 + netdev_err(priv->si->ndev, "DMA mapping failed!\n");
1517 + memset(cbdr, 0, sizeof(*cbdr));
1521 + cbdr->addr[0] = lower_32_bits(dma);
1522 + cbdr->addr[1] = upper_32_bits(dma);
1524 + /* VIDM default to be 1.
1525 + * VID Match. If set (b1) then the VID must match, otherwise
1526 + * any VID is considered a match. VIDM setting is only used
1527 + * when TG is set to b01.
1529 + if (si_conf->id_type == 1) {
1530 + si_data1 = (struct null_streamid_data *)si_data;
1531 + si_data1->dmac[0] = streamid->para.nid.dmac & 0xFF;
1532 + si_data1->dmac[1] = (streamid->para.nid.dmac >> 8) & 0xFF;
1533 + si_data1->dmac[2] = (streamid->para.nid.dmac >> 16) & 0xFF;
1534 + si_data1->dmac[3] = (streamid->para.nid.dmac >> 24) & 0xFF;
1535 + si_data1->dmac[4] = (streamid->para.nid.dmac >> 32) & 0xFF;
1536 + si_data1->dmac[5] = (streamid->para.nid.dmac >> 40) & 0xFF;
1537 + si_data1->vid_vidm_tg =
1538 + cpu_to_le16((streamid->para.nid.vid & ENETC_CBDR_SID_VID_MASK) +
1539 + ((((u16)(streamid->para.nid.tagged) & 0x3) << 14)
1540 + | ENETC_CBDR_SID_VIDM));
1541 + } else if (si_conf->id_type == 2) {
1542 + si_data2 = (struct smac_streamid_data *)si_data;
1543 + si_data2->smac[0] = streamid->para.sid.smac & 0xFF;
1544 + si_data2->smac[1] = (streamid->para.sid.smac >> 8) & 0xFF;
1545 + si_data2->smac[2] = (streamid->para.sid.smac >> 16) & 0xFF;
1546 + si_data2->smac[3] = (streamid->para.sid.smac >> 24) & 0xFF;
1547 + si_data2->smac[4] = (streamid->para.sid.smac >> 32) & 0xFF;
1548 + si_data2->smac[5] = (streamid->para.sid.smac >> 40) & 0xFF;
1549 + si_data2->vid_vidm_tg =
1550 + cpu_to_le16((streamid->para.sid.vid & ENETC_CBDR_SID_VID_MASK) +
1551 + ((((u16)(streamid->para.sid.tagged) & 0x3) << 14)
1552 + | ENETC_CBDR_SID_VIDM));
1555 + xmit_cbdr(priv->si, curr_cbd);
1557 + memset(cbdr, 0, sizeof(*cbdr));
1563 +/* CBD Class 7: Stream Identity Entry Query Descriptor - Long Format */
1564 +int enetc_cb_streamid_get(struct net_device *ndev, u32 index,
1565 + struct tsn_cb_streamid *streamid)
1567 + struct enetc_cbd *cbdr;
1568 + struct streamid_query_resp *si_data;
1569 + struct enetc_ndev_priv *priv;
1571 + u16 data_size, dma_size;
1578 + priv = netdev_priv(ndev);
1580 + curr_cbd = alloc_cbdr(priv->si, &cbdr);
1582 + cbdr->index = cpu_to_le32(index);
1584 + cbdr->cls = BDCR_CMD_STREAM_IDENTIFY;
1585 + cbdr->status_flags = 0;
1587 + data_size = sizeof(struct streamid_query_resp);
1588 + si_data = kzalloc(data_size, __GFP_DMA | GFP_KERNEL);
1592 + dma_size = cpu_to_le16(data_size);
1593 + cbdr->length = dma_size;
1594 + cbdr->status_flags = 0; /* long format command no ie */
1596 + dma = dma_map_single(&priv->si->pdev->dev, si_data,
1597 + data_size, DMA_FROM_DEVICE);
1598 + if (dma_mapping_error(&priv->si->pdev->dev, dma)) {
1599 + netdev_err(priv->si->ndev, "DMA mapping failed!\n");
1603 + cbdr->addr[0] = lower_32_bits(dma);
1604 + cbdr->addr[1] = upper_32_bits(dma);
1606 + xmit_cbdr(priv->si, curr_cbd);
1608 + streamid->type = si_data->id_type;
1610 + if (streamid->type == 1) {
1611 + streamid->para.nid.dmac = si_data->mac[0]
1612 + + ((u64)si_data->mac[1] << 8)
1613 + + ((u64)si_data->mac[2] << 16)
1614 + + ((u64)si_data->mac[3] << 24)
1615 + + ((u64)si_data->mac[4] << 32)
1616 + + ((u64)si_data->mac[5] << 40);
1617 + /* VID Match. If set (b1) then the VID must match, otherwise
1618 + * any VID is considered a match.
1620 + streamid->para.nid.vid =
1621 + le16_to_cpu(si_data->vid_vidm_tg
1622 + & ENETC_CBDR_SID_VID_MASK);
1623 + streamid->para.nid.tagged =
1624 + le16_to_cpu(si_data->vid_vidm_tg >> 14 & 0x3);
1625 + } else if (streamid->type == 2) {
1626 + streamid->para.sid.smac = si_data->mac[0]
1627 + + ((u64)si_data->mac[1] << 8)
1628 + + ((u64)si_data->mac[2] << 16)
1629 + + ((u64)si_data->mac[3] << 24)
1630 + + ((u64)si_data->mac[4] << 32)
1631 + + ((u64)si_data->mac[5] << 40);
1632 + /* VID Match. If set (b1) then the VID must match, otherwise
1633 + * any VID is considered a match.
1635 + streamid->para.sid.vid =
1636 + le16_to_cpu(si_data->vid_vidm_tg
1637 + & ENETC_CBDR_SID_VID_MASK);
1638 + streamid->para.sid.tagged =
1639 + le16_to_cpu(si_data->vid_vidm_tg >> 14 & 0x3);
1642 + streamid->handle = le32_to_cpu(si_data->stream_handle);
1643 + streamid->ifac_iport = le32_to_cpu(si_data->input_ports);
1644 + valid = si_data->en ? 1 : 0;
1646 + memset(cbdr, 0, sizeof(*cbdr));
1652 +/* CBD Class 7: Stream Identity Statistics Query Descriptor - Long Format */
1653 +int enetc_cb_streamid_counters_get(struct net_device *ndev, u32 index,
1654 + struct tsn_cb_streamid_counters *counters)
1659 +void enetc_qci_enable(struct enetc_hw *hw)
1661 + enetc_wr(hw, ENETC_PPSFPMR, enetc_rd(hw, ENETC_PPSFPMR)
1662 + | ENETC_PPSFPMR_PSFPEN | ENETC_PPSFPMR_VS
1663 + | ENETC_PPSFPMR_PVC | ENETC_PPSFPMR_PVZC);
1666 +void enetc_qci_disable(struct enetc_hw *hw)
1668 + enetc_wr(hw, ENETC_PPSFPMR, enetc_rd(hw, ENETC_PPSFPMR)
1669 + & ~ENETC_PPSFPMR_PSFPEN & ~ENETC_PPSFPMR_VS
1670 + & ~ENETC_PPSFPMR_PVC & ~ENETC_PPSFPMR_PVZC);
1673 +/* CBD Class 8: Stream Filter Instance Set Descriptor - Short Format */
1674 +int enetc_qci_sfi_set(struct net_device *ndev, u32 index, bool en,
1675 + struct tsn_qci_psfp_sfi_conf *tsn_qci_sfi)
1677 + struct enetc_cbd *cbdr;
1678 + struct sfi_conf *sfi_config;
1680 + struct enetc_ndev_priv *priv = netdev_priv(ndev);
1683 + curr_cbd = alloc_cbdr(priv->si, &cbdr);
1685 + cbdr->index = cpu_to_le16(index);
1687 + cbdr->cls = BDCR_CMD_STREAM_FILTER;
1688 + cbdr->status_flags = 0x80;
1689 + cbdr->length = cpu_to_le16(1);
1691 + sfi_config = &cbdr->sfi_conf;
1693 + sfi_config->en = 0x80;
1695 + if (tsn_qci_sfi->stream_handle_spec >= 0) {
1696 + sfi_config->stream_handle =
1697 + cpu_to_le32(tsn_qci_sfi->stream_handle_spec);
1698 + sfi_config->sthm |= 0x80;
1701 + sfi_config->sg_inst_table_index =
1702 + cpu_to_le16(tsn_qci_sfi->stream_gate_instance_id);
1703 + sfi_config->input_ports = 1 << (priv->si->pdev->devfn & 0x7);
1705 + /* The priority value which may be matched against the
1706 + * frame’s priority value to determine a match for this entry.
1708 + if (tsn_qci_sfi->priority_spec >= 0)
1709 + sfi_config->multi |= (tsn_qci_sfi->priority_spec & 0x7) | 0x8;
1711 + /* Filter Type. Identifies the contents of the MSDU/FM_INST_INDEX
1712 + * field as being either an MSDU value or an index into the Flow
1713 + * Meter Instance table.
1715 + if (tsn_qci_sfi->stream_filter.maximum_sdu_size != 0) {
1716 + sfi_config->msdu =
1717 + cpu_to_le16(tsn_qci_sfi->stream_filter.maximum_sdu_size);
1718 + sfi_config->multi |= 0x40;
1721 + if (tsn_qci_sfi->stream_filter.flow_meter_instance_id >= 0) {
1722 + sfi_config->fm_inst_table_index =
1723 + cpu_to_le16(tsn_qci_sfi->stream_filter.flow_meter_instance_id);
1724 + sfi_config->multi |= 0x80;
1727 + /* Stream blocked due to oversized frame enable. TRUE or FALSE */
1728 + if (tsn_qci_sfi->block_oversize_enable)
1729 + sfi_config->multi |= 0x20;
1731 + /* Stream blocked due to oversized frame. TRUE or FALSE */
1732 + if (tsn_qci_sfi->block_oversize)
1733 + sfi_config->multi |= 0x10;
1735 + xmit_cbdr(priv->si, curr_cbd);
1737 + memset(cbdr, 0, sizeof(*cbdr));
1741 +/* CBD Class 8: Stream Filter Instance Query Descriptor - Short Format */
1742 +int enetc_qci_sfi_get(struct net_device *ndev, u32 index,
1743 + struct tsn_qci_psfp_sfi_conf *tsn_qci_sfi)
1745 + struct enetc_cbd *cbdr;
1746 + struct sfi_conf *sfi_config;
1747 + struct enetc_ndev_priv *priv = netdev_priv(ndev);
1750 + curr_cbd = alloc_cbdr(priv->si, &cbdr);
1752 + cbdr->index = cpu_to_le16(index);
1754 + cbdr->cls = BDCR_CMD_STREAM_FILTER;
1755 + cbdr->status_flags = 0x80;
1757 + xmit_cbdr(priv->si, curr_cbd);
1759 + sfi_config = &cbdr->sfi_conf;
1760 + if (sfi_config->sthm & 0x80)
1761 + tsn_qci_sfi->stream_handle_spec =
1762 + le32_to_cpu(sfi_config->stream_handle);
1764 + tsn_qci_sfi->stream_handle_spec = -1;
1766 + tsn_qci_sfi->stream_gate_instance_id =
1767 + le16_to_cpu(sfi_config->sg_inst_table_index);
1769 + if (sfi_config->multi & 0x8)
1770 + tsn_qci_sfi->priority_spec =
1771 + le16_to_cpu(sfi_config->multi & 0x7);
1773 + tsn_qci_sfi->priority_spec = -1;
1775 + /* Filter Type. Identifies the contents of the MSDU/FM_INST_INDEX
1776 + * field as being either an MSDU value or an index into the Flow
1777 + * Meter Instance table.
1779 + if (sfi_config->multi & 0x80)
1780 + tsn_qci_sfi->stream_filter.flow_meter_instance_id =
1781 + le16_to_cpu(sfi_config->fm_inst_table_index);
1783 + tsn_qci_sfi->stream_filter.flow_meter_instance_id = -1;
1785 + if (sfi_config->multi & 0x40)
1786 + tsn_qci_sfi->stream_filter.maximum_sdu_size =
1787 + le16_to_cpu(sfi_config->msdu);
1789 + /* Stream blocked due to oversized frame enable. TRUE or FALSE */
1790 + if (sfi_config->multi & 0x20)
1791 + tsn_qci_sfi->block_oversize_enable = true;
1792 + /* Stream blocked due to oversized frame. TRUE or FALSE */
1793 + if (sfi_config->multi & 0x10)
1794 + tsn_qci_sfi->block_oversize = true;
1796 + if (sfi_config->en & 0x80) {
1797 + memset(cbdr, 0, sizeof(*cbdr));
1801 + memset(cbdr, 0, sizeof(*cbdr));
1805 +/* CBD Class 8: Stream Filter Instance Query Statistics
1806 + * Descriptor - Long Format
1808 +int enetc_qci_sfi_counters_get(struct net_device *ndev, u32 index,
1809 + struct tsn_qci_psfp_sfi_counters *counters)
1811 + struct enetc_cbd *cbdr;
1812 + struct enetc_ndev_priv *priv = netdev_priv(ndev);
1814 + struct sfi_counter_data *sfi_counter_data;
1816 + u16 data_size, dma_size;
1818 + curr_cbd = alloc_cbdr(priv->si, &cbdr);
1820 + cbdr->index = cpu_to_le16((u16)index);
1822 + cbdr->cls = BDCR_CMD_STREAM_FILTER;
1823 + cbdr->status_flags = 0;
1825 + data_size = sizeof(struct sfi_counter_data);
1826 + sfi_counter_data = kzalloc(data_size, __GFP_DMA | GFP_KERNEL);
1827 + if (!sfi_counter_data)
1830 + dma = dma_map_single(&priv->si->pdev->dev, sfi_counter_data,
1831 + data_size, DMA_FROM_DEVICE);
1832 + if (dma_mapping_error(&priv->si->pdev->dev, dma)) {
1833 + netdev_err(priv->si->ndev, "DMA mapping failed!\n");
1834 + kfree(sfi_counter_data);
1837 + cbdr->addr[0] = lower_32_bits(dma);
1838 + cbdr->addr[1] = upper_32_bits(dma);
1840 + dma_size = cpu_to_le16(data_size);
1841 + cbdr->length = dma_size;
1843 + xmit_cbdr(priv->si, curr_cbd);
1845 + counters->matching_frames_count =
1846 + ((u64)le32_to_cpu(sfi_counter_data->matchh) << 32)
1847 + + sfi_counter_data->matchl;
1849 + counters->not_passing_sdu_count =
1850 + ((u64)le32_to_cpu(sfi_counter_data->msdu_droph) << 32)
1851 + + sfi_counter_data->msdu_dropl;
1853 + counters->passing_sdu_count = counters->matching_frames_count
1854 + - counters->not_passing_sdu_count;
1856 + counters->not_passing_frames_count =
1857 + ((u64)le32_to_cpu(sfi_counter_data->stream_gate_droph) << 32)
1858 + + le32_to_cpu(sfi_counter_data->stream_gate_dropl);
1860 + counters->passing_frames_count = counters->matching_frames_count
1861 + - counters->not_passing_sdu_count
1862 + - counters->not_passing_frames_count;
1864 + counters->red_frames_count =
1865 + ((u64)le32_to_cpu(sfi_counter_data->flow_meter_droph) << 32)
1866 + + le32_to_cpu(sfi_counter_data->flow_meter_dropl);
1868 + memset(cbdr, 0, sizeof(*cbdr));
1872 +/* CBD Class 9: Stream Gate Instance Table Entry Set
1873 + * Descriptor - Short Format
1875 +int enetc_qci_sgi_set(struct net_device *ndev, u32 index,
1876 + struct tsn_qci_psfp_sgi_conf *tsn_qci_sgi)
1878 + struct enetc_cbd *cbdr, *cbdr_sgcl;
1879 + struct sgi_table *sgi_config;
1880 + struct sgcl_conf *sgcl_config;
1881 + struct sgcl_data *sgcl_data;
1882 + struct sgce *sgce;
1883 + struct enetc_ndev_priv *priv = netdev_priv(ndev);
1886 + u16 data_size, dma_size;
1889 + /* disable first */
1890 + curr_cbd = alloc_cbdr(priv->si, &cbdr);
1891 + memset(cbdr, 0, sizeof(*cbdr));
1893 + cbdr->index = cpu_to_le16(index);
1895 + cbdr->cls = BDCR_CMD_STREAM_GCL;
1896 + cbdr->status_flags = 0x80;
1898 + xmit_cbdr(priv->si, curr_cbd);
1900 + if (!tsn_qci_sgi->gate_enabled) {
1901 + memset(cbdr, 0, sizeof(*cbdr));
1906 + curr_cbd = alloc_cbdr(priv->si, &cbdr);
1907 + memset(cbdr, 0, sizeof(*cbdr));
1909 + cbdr->index = cpu_to_le16(index);
1911 + cbdr->cls = BDCR_CMD_STREAM_GCL;
1912 + cbdr->status_flags = 0x80;
1914 + sgi_config = &cbdr->sgi_table;
1916 + sgi_config->ocgtst = tsn_qci_sgi->admin.control_list_length ?
1917 + 0x80 : (tsn_qci_sgi->admin.gate_states ? 0x80 : 0x0);
1919 + sgi_config->oipv =
1920 + tsn_qci_sgi->admin.control_list_length ?
1921 + 0x0 : ((tsn_qci_sgi->admin.init_ipv < 0) ?
1922 + 0x0 : ((tsn_qci_sgi->admin.init_ipv & 0x7) | 0x8));
1924 + sgi_config->en = 0x80;
1926 + if (tsn_qci_sgi->block_invalid_rx_enable)
1927 + sgi_config->gset |= 0x80;
1928 + if (tsn_qci_sgi->block_invalid_rx)
1929 + sgi_config->gset |= 0x40;
1930 + if (tsn_qci_sgi->block_octets_exceeded)
1931 + sgi_config->gset |= 0x10;
1932 + if (tsn_qci_sgi->block_octets_exceeded_enable)
1933 + sgi_config->gset |= 0x20;
1935 + xmit_cbdr(priv->si, curr_cbd);
1937 + if (tsn_qci_sgi->admin.control_list_length == 0)
1940 + curr_cbd = alloc_cbdr(priv->si, &cbdr_sgcl);
1941 + memset(cbdr, 0, sizeof(*cbdr));
1943 + cbdr_sgcl->index = cpu_to_le16(index);
1944 + cbdr_sgcl->cmd = 1;
1945 + cbdr_sgcl->cls = BDCR_CMD_STREAM_GCL;
1946 + cbdr_sgcl->status_flags = 0;
1948 + sgcl_config = &cbdr_sgcl->sgcl_conf;
1950 + /* tsn_qci_sgi->admin.control_list_length is not zero now */
1951 + if (tsn_qci_sgi->admin.control_list_length > 4)
1954 + sgcl_config->acl_len =
1955 + (tsn_qci_sgi->admin.control_list_length - 1) & 0x3;
1957 + data_size = sizeof(struct sgcl_data) +
1958 + (sgcl_config->acl_len + 1) * sizeof(struct sgce);
1960 + sgcl_data = kzalloc(data_size, __GFP_DMA | GFP_KERNEL);
1964 + dma_size = cpu_to_le16(data_size);
1965 + cbdr_sgcl->length = dma_size;
1967 + dma = dma_map_single(&priv->si->pdev->dev,
1968 + sgcl_data, data_size,
1970 + if (dma_mapping_error(&priv->si->pdev->dev, dma)) {
1971 + netdev_err(priv->si->ndev, "DMA mapping failed!\n");
1972 + memset(cbdr, 0, sizeof(*cbdr));
1973 + memset(cbdr_sgcl, 0, sizeof(*cbdr_sgcl));
1977 + cbdr_sgcl->addr[0] = lower_32_bits(dma);
1978 + cbdr_sgcl->addr[1] = upper_32_bits(dma);
1980 + sgce = (struct sgce *)(sgcl_data + 1);
1982 + if (tsn_qci_sgi->admin.gate_states)
1983 + sgcl_config->agtst = 0x80;
1985 + sgcl_data->ct = cpu_to_le32(tsn_qci_sgi->admin.cycle_time);
1986 + sgcl_data->cte = cpu_to_le32(tsn_qci_sgi->admin.cycle_time_extension);
1988 + if (tsn_qci_sgi->admin.init_ipv >= 0)
1989 + sgcl_config->aipv = (tsn_qci_sgi->admin.init_ipv & 0x7) | 0x8;
1991 + for (i = 0; i < tsn_qci_sgi->admin.control_list_length; i++) {
1992 + struct tsn_qci_psfp_gcl *temp_sgcl = tsn_qci_sgi->admin.gcl + i;
1993 + struct sgce *temp_entry = (struct sgce *)(sgce + i);
1995 + if (temp_sgcl->gate_state)
1996 + temp_entry->multi |= 0x10;
1998 + if (temp_sgcl->ipv >= 0)
1999 + temp_entry->multi |= ((temp_sgcl->ipv & 0x7) << 5)
2002 + if (temp_sgcl->octet_max)
2003 + temp_entry->multi |= 0x01;
2005 + temp_entry->interval = cpu_to_le32(temp_sgcl->time_interval);
2006 + temp_entry->msdu[0] = temp_sgcl->octet_max & 0xFF;
2007 + temp_entry->msdu[1] = (temp_sgcl->octet_max >> 8) & 0xFF;
2008 + temp_entry->msdu[2] = (temp_sgcl->octet_max >> 16) & 0xFF;
2011 + if (!tsn_qci_sgi->admin.base_time) {
2013 + cpu_to_le32(enetc_rd(&priv->si->hw, ENETC_SICTR0));
2015 + cpu_to_le32(enetc_rd(&priv->si->hw, ENETC_SICTR1));
2019 + tempu = upper_32_bits(tsn_qci_sgi->admin.base_time);
2020 + templ = lower_32_bits(tsn_qci_sgi->admin.base_time);
2021 + sgcl_data->bth = cpu_to_le32(tempu);
2022 + sgcl_data->btl = cpu_to_le32(templ);
2025 + xmit_cbdr(priv->si, curr_cbd);
2027 + memset(cbdr_sgcl, 0, sizeof(*cbdr_sgcl));
2031 + memset(cbdr, 0, sizeof(*cbdr));
2035 +/* CBD Class 9: Stream Gate Instance Table Entry Query
2036 + * Descriptor - Short Format
2038 +int enetc_qci_sgi_get(struct net_device *ndev, u32 index,
2039 + struct tsn_qci_psfp_sgi_conf *tsn_qci_sgi)
2041 + struct enetc_cbd *cbdr, *cbdr_sgcl;
2042 + struct sgi_table *sgi_config;
2043 + struct sgcl_query *sgcl_query;
2044 + struct sgcl_query_resp *sgcl_data;
2045 + struct sgce *sgce;
2046 + struct enetc_ndev_priv *priv = netdev_priv(ndev);
2048 + u16 data_size, dma_size, gcl_data_stat = 0;
2052 + curr_cbd = alloc_cbdr(priv->si, &cbdr);
2054 + cbdr->index = cpu_to_le16(index);
2056 + cbdr->cls = BDCR_CMD_STREAM_GCL;
2057 + cbdr->status_flags = 0x80;
2059 + xmit_cbdr(priv->si, curr_cbd);
2061 + sgi_config = &cbdr->sgi_table;
2063 + tsn_qci_sgi->admin.gate_states = (sgi_config->ocgtst & 0x80) ?
2065 + if (sgi_config->oipv & 0x08)
2066 + tsn_qci_sgi->admin.init_ipv = sgi_config->oipv & 0x7;
2068 + tsn_qci_sgi->admin.init_ipv = -1;
2070 + if (sgi_config->en & 0x80)
2071 + tsn_qci_sgi->gate_enabled = true;
2072 + if (sgi_config->gset & 0x80)
2073 + tsn_qci_sgi->block_invalid_rx_enable = true;
2074 + if (sgi_config->gset & 0x40)
2075 + tsn_qci_sgi->block_invalid_rx = true;
2076 + if (sgi_config->gset & 0x20)
2077 + tsn_qci_sgi->block_octets_exceeded_enable = true;
2078 + if (sgi_config->gset & 0x10)
2079 + tsn_qci_sgi->block_octets_exceeded = true;
2081 + /* Check gate list length is zero? */
2082 + if (!(sgi_config->oacl_len & 0x30)) {
2083 + tsn_qci_sgi->admin.control_list_length = 0;
2087 + curr_cbd = alloc_cbdr(priv->si, &cbdr_sgcl);
2089 + cbdr_sgcl->index = cpu_to_le16(index);
2090 + cbdr_sgcl->cmd = 3;
2091 + cbdr_sgcl->cls = BDCR_CMD_STREAM_GCL;
2092 + cbdr_sgcl->status_flags = 0;
2094 + data_size = sizeof(struct sgcl_query_resp) + 4 * sizeof(struct sgce);
2096 + sgcl_data = kzalloc(data_size, __GFP_DMA | GFP_KERNEL);
2100 + dma_size = cpu_to_le16(data_size);
2101 + cbdr_sgcl->length = dma_size;
2102 + cbdr_sgcl->status_flags = 0;
2104 + sgcl_query = &cbdr_sgcl->sgcl_query;
2106 + sgcl_query->oacl_len = 0x10;
2108 + dma = dma_map_single(&priv->si->pdev->dev, sgcl_data,
2109 + data_size, DMA_FROM_DEVICE);
2110 + if (dma_mapping_error(&priv->si->pdev->dev, dma)) {
2111 + netdev_err(priv->si->ndev, "DMA mapping failed!\n");
2112 + memset(cbdr, 0, sizeof(*cbdr));
2113 + memset(cbdr_sgcl, 0, sizeof(*cbdr_sgcl));
2117 + cbdr_sgcl->addr[0] = lower_32_bits(dma);
2118 + cbdr_sgcl->addr[1] = upper_32_bits(dma);
2120 + xmit_cbdr(priv->si, curr_cbd);
2122 + sgce = (struct sgce *)(sgcl_data + 1);
2124 + gcl_data_stat = le16_to_cpu(sgcl_data->stat);
2125 + if (gcl_data_stat & 0x10)
2126 + tsn_qci_sgi->admin.gate_states = true;
2128 + if (gcl_data_stat & 0x80)
2129 + tsn_qci_sgi->admin.init_ipv = gcl_data_stat & 0x7;
2131 + tsn_qci_sgi->admin.init_ipv = -1;
2133 + /* admin_len can also get from gcl_data_stat bit 5,6
2134 + * OR sgi_config->oacl_len
2136 + admin_len = (sgcl_query->oacl_len & 0x3) + 1;
2137 + tsn_qci_sgi->admin.control_list_length = admin_len;
2138 + tsn_qci_sgi->admin.cycle_time = le32_to_cpu(sgcl_data->act);
2139 + tsn_qci_sgi->admin.cycle_time_extension = le32_to_cpu(sgcl_data->acte);
2140 + tsn_qci_sgi->admin.base_time = ((u64)(le32_to_cpu(sgcl_data->abth))
2142 + + le32_to_cpu(sgcl_data->abtl);
2144 + tsn_qci_sgi->admin.gcl = kcalloc(admin_len,
2145 + sizeof(struct tsn_qci_psfp_gcl),
2147 + if (!tsn_qci_sgi->admin.gcl) {
2152 + for (i = 0; i < admin_len; i++) {
2153 + struct tsn_qci_psfp_gcl *temp_sgcl = tsn_qci_sgi->admin.gcl + i;
2154 + struct sgce *temp_entry = (struct sgce *)(sgce + i);
2156 + if (temp_entry->multi & 0x10)
2157 + temp_sgcl->gate_state = true;
2159 + if (temp_entry->multi & 0x08)
2160 + temp_sgcl->ipv = temp_entry->multi >> 5;
2162 + temp_sgcl->ipv = -1;
2164 + temp_sgcl->time_interval = le32_to_cpu(temp_entry->interval);
2166 + if (temp_entry->multi & 0x01)
2167 + temp_sgcl->octet_max = (temp_entry->msdu[0] & 0xff)
2168 + | (((u32)temp_entry->msdu[1] << 8) & 0xff00)
2169 + | (((u32)temp_entry->msdu[1] << 16) & 0xff0000);
2171 + temp_sgcl->octet_max = 0;
2174 + memset(cbdr_sgcl, 0, sizeof(*cbdr_sgcl));
2178 + memset(cbdr, 0, sizeof(*cbdr));
2182 +/* CBD Class 9: Stream Gate Instance Table Entry Query Descriptor
2183 + * CBD Class 9: Stream Gate Control List Query Descriptor
2185 +int enetc_qci_sgi_status_get(struct net_device *ndev, u16 index,
2186 + struct tsn_psfp_sgi_status *status)
2188 + struct enetc_cbd *cbdr_sgi, *cbdr_sgcl;
2189 + struct sgi_table *sgi_config;
2190 + struct sgcl_query *sgcl_query;
2191 + struct sgcl_query_resp *sgcl_data;
2192 + struct sgce *sgce;
2193 + struct enetc_ndev_priv *priv = netdev_priv(ndev);
2195 + u16 data_size, dma_size, gcl_data_stat = 0;
2199 + curr_cbd = alloc_cbdr(priv->si, &cbdr_sgi);
2201 + cbdr_sgi->index = cpu_to_le16(index);
2202 + cbdr_sgi->cmd = 2;
2203 + cbdr_sgi->cls = BDCR_CMD_STREAM_GCL;
2204 + cbdr_sgi->status_flags = 0x80;
2206 + sgi_config = &cbdr_sgi->sgi_table;
2208 + if (sgi_config->gset & 0x4)
2209 + status->config_pending = true;
2211 + status->oper.gate_states = ((sgi_config->ocgtst & 0x80) ? true : false);
2213 + /* Check gate list length is zero */
2214 + if (!(sgi_config->oacl_len & 0x30)) {
2215 + status->oper.control_list_length = 0;
2219 + xmit_cbdr(priv->si, curr_cbd);
2221 + curr_cbd = alloc_cbdr(priv->si, &cbdr_sgcl);
2223 + cbdr_sgcl->index = cpu_to_le16(index);
2224 + cbdr_sgcl->cmd = 3;
2225 + cbdr_sgcl->cls = BDCR_CMD_STREAM_GCL;
2226 + cbdr_sgcl->status_flags = 0;
2229 + data_size = sizeof(struct sgcl_query_resp) + 4 * sizeof(struct sgce);
2231 + sgcl_data = kzalloc(data_size, __GFP_DMA | GFP_KERNEL);
2235 + dma_size = cpu_to_le16(data_size);
2236 + cbdr_sgcl->length = dma_size;
2237 + cbdr_sgcl->status_flags = 0;
2239 + sgcl_query = &cbdr_sgcl->sgcl_query;
2241 + sgcl_query->oacl_len = 0x20;
2243 + dma = dma_map_single(&priv->si->pdev->dev, sgcl_data,
2244 + data_size, DMA_FROM_DEVICE);
2245 + if (dma_mapping_error(&priv->si->pdev->dev, dma)) {
2246 + netdev_err(priv->si->ndev, "DMA mapping failed!\n");
2247 + memset(cbdr_sgi, 0, sizeof(*cbdr_sgi));
2248 + memset(cbdr_sgcl, 0, sizeof(*cbdr_sgcl));
2252 + cbdr_sgcl->addr[0] = lower_32_bits(dma);
2253 + cbdr_sgcl->addr[1] = upper_32_bits(dma);
2255 + xmit_cbdr(priv->si, curr_cbd);
2257 + sgce = (struct sgce *)(sgcl_data + 1);
2259 + /* oper_len can also get from gcl_data_stat bit 5,6
2260 + * OR sgi_config->oacl_len
2262 + oper_len = ((sgcl_query->oacl_len & 0x0c) >> 2) + 1;
2264 + /* Get Stream Gate Control List */
2265 + status->oper.cycle_time = le32_to_cpu(sgcl_data->oct);
2266 + status->oper.cycle_time_extension = le32_to_cpu(sgcl_data->octe);
2267 + status->oper.base_time = le32_to_cpu(sgcl_data->obtl)
2268 + + ((u64)le32_to_cpu(sgcl_data->obth) << 32);
2269 + status->oper.control_list_length = oper_len;
2271 + gcl_data_stat = le16_to_cpu(sgcl_data->stat);
2272 + if (gcl_data_stat & 0x400)
2273 + status->oper.init_ipv = gcl_data_stat & 0x38 >> 7;
2275 + status->oper.init_ipv = -1;
2277 + if (gcl_data_stat & 0x800)
2278 + status->oper.gate_states = true;
2280 + status->oper.gcl = kcalloc(oper_len,
2281 + sizeof(struct tsn_qci_psfp_gcl),
2283 + if (!status->oper.gcl) {
2284 + memset(cbdr_sgi, 0, sizeof(*cbdr_sgi));
2285 + memset(cbdr_sgcl, 0, sizeof(*cbdr_sgcl));
2290 + for (i = 0; i < oper_len; i++) {
2291 + struct tsn_qci_psfp_gcl *temp_sgcl = status->oper.gcl + i;
2292 + struct sgce *temp_entry = (struct sgce *)(sgce + i);
2294 + if (temp_entry->multi & 0x10)
2295 + temp_sgcl->gate_state = true;
2297 + if (temp_entry->multi & 0x08)
2298 + temp_sgcl->ipv = temp_entry->multi >> 5;
2300 + temp_sgcl->ipv = -1;
2302 + temp_sgcl->time_interval = le32_to_cpu(temp_entry->interval);
2304 + if (temp_entry->multi & 0x01)
2305 + temp_sgcl->octet_max = temp_entry->msdu[0]
2306 + | ((((u32)temp_entry->msdu[1]) << 8)
2308 + | ((((u32)temp_entry->msdu[2]) << 16)
2311 + temp_sgcl->octet_max = 0;
2314 + status->config_change_time = le32_to_cpu(sgcl_data->cctl)
2315 + + ((u64)le32_to_cpu(sgcl_data->ccth) << 32);
2317 + memset(cbdr_sgcl, 0, sizeof(*cbdr_sgcl));
2321 + /* changed to SITGTGR */
2322 + status->tick_granularity = enetc_rd(&priv->si->hw, ENETC_SITGTGR);
2324 + /* current time */
2325 + status->current_time = get_current_time(priv->si);
2327 + memset(cbdr_sgi, 0, sizeof(*cbdr_sgi));
2332 +/* CBD Class 10: Flow Meter Instance Set Descriptor - Short Format */
2333 +int enetc_qci_fmi_set(struct net_device *ndev, u32 index, bool enable,
2334 + struct tsn_qci_psfp_fmi *tsn_qci_fmi)
2336 + struct enetc_cbd *cbdr;
2337 + struct fmi_conf *fmi_config;
2339 + struct enetc_ndev_priv *priv = netdev_priv(ndev);
2343 + curr_cbd = alloc_cbdr(priv->si, &cbdr);
2345 + cbdr->index = cpu_to_le16((u16)index);
2347 + cbdr->cls = BDCR_CMD_FLOW_METER;
2348 + cbdr->status_flags = 0x80;
2350 + xmit_cbdr(priv->si, curr_cbd);
2353 + memset(cbdr, 0, sizeof(*cbdr));
2358 + curr_cbd = alloc_cbdr(priv->si, &cbdr);
2359 + memset(cbdr, 0, sizeof(*cbdr));
2360 + cbdr->index = cpu_to_le16((u16)index);
2362 + cbdr->cls = BDCR_CMD_FLOW_METER;
2363 + cbdr->status_flags = 0x80;
2365 + fmi_config = &cbdr->fmi_conf;
2366 + fmi_config->en = 0x80;
2367 + if (tsn_qci_fmi->cir) {
2368 + temp = (u64)1000 * tsn_qci_fmi->cir;
2369 + temp = temp / 3725;
2371 + fmi_config->cir = cpu_to_le32((u32)temp);
2372 + fmi_config->cbs = cpu_to_le32(tsn_qci_fmi->cbs);
2374 + if (tsn_qci_fmi->eir) {
2375 + temp = (u64)1000 * tsn_qci_fmi->eir;
2376 + temp = temp / 3725;
2378 + fmi_config->eir = cpu_to_le32((u32)temp);
2379 + fmi_config->ebs = cpu_to_le32(tsn_qci_fmi->ebs);
2381 + if (tsn_qci_fmi->mark_red)
2382 + fmi_config->conf |= 0x1;
2384 + if (tsn_qci_fmi->mark_red_enable)
2385 + fmi_config->conf |= 0x2;
2387 + if (tsn_qci_fmi->drop_on_yellow)
2388 + fmi_config->conf |= 0x4;
2390 + if (tsn_qci_fmi->cm)
2391 + fmi_config->conf |= 0x8;
2393 + if (tsn_qci_fmi->cf)
2394 + fmi_config->conf |= 0x10;
2396 + xmit_cbdr(priv->si, curr_cbd);
2398 + memset(cbdr, 0, sizeof(*cbdr));
2402 +/* CBD Class 10: Flow Meter Instance Query Descriptor - Short Format */
2403 +int enetc_qci_fmi_get(struct net_device *ndev, u32 index,
2404 + struct tsn_qci_psfp_fmi *tsn_qci_fmi,
2405 + struct tsn_qci_psfp_fmi_counters *counters)
2407 + struct enetc_cbd *cbdr;
2408 + struct fmi_conf *fmi_config;
2409 + struct enetc_ndev_priv *priv = netdev_priv(ndev);
2411 + u16 data_size, dma_size;
2413 + struct fmi_query_stat_resp *fmi_counter_data;
2416 + curr_cbd = alloc_cbdr(priv->si, &cbdr);
2418 + cbdr->index = cpu_to_le16(index);
2420 + cbdr->cls = BDCR_CMD_FLOW_METER;
2421 + cbdr->status_flags = 0x80;
2423 + xmit_cbdr(priv->si, curr_cbd);
2425 + fmi_config = &cbdr->fmi_conf;
2426 + if (fmi_config->cir) {
2427 + temp = (u64)3725 * fmi_config->cir;
2428 + temp = temp / 1000;
2430 + tsn_qci_fmi->cir = le32_to_cpu((u32)temp);
2431 + tsn_qci_fmi->cbs = le32_to_cpu(fmi_config->cbs);
2433 + if (fmi_config->eir) {
2434 + temp = (u64)3725 * fmi_config->eir;
2435 + temp = temp / 1000;
2437 + tsn_qci_fmi->eir = le32_to_cpu((u32)temp);
2438 + tsn_qci_fmi->ebs = le32_to_cpu(fmi_config->ebs);
2440 + if (fmi_config->conf & 0x1)
2441 + tsn_qci_fmi->mark_red = true;
2443 + if (fmi_config->conf & 0x2)
2444 + tsn_qci_fmi->mark_red_enable = true;
2446 + if (fmi_config->conf & 0x4)
2447 + tsn_qci_fmi->drop_on_yellow = true;
2449 + if (fmi_config->conf & 0x8)
2450 + tsn_qci_fmi->cm = true;
2452 + if (fmi_config->conf & 0x10)
2453 + tsn_qci_fmi->cf = true;
2455 + memset(cbdr, 0, sizeof(*cbdr));
2457 + /* Get counters */
2458 + curr_cbd = alloc_cbdr(priv->si, &cbdr);
2460 + cbdr->index = cpu_to_le16(index);
2462 + cbdr->cls = BDCR_CMD_FLOW_METER;
2463 + cbdr->status_flags = 0x0;
2465 + data_size = sizeof(struct fmi_query_stat_resp);
2466 + fmi_counter_data = kzalloc(data_size, __GFP_DMA | GFP_KERNEL);
2467 + if (!fmi_counter_data)
2470 + dma = dma_map_single(&priv->si->pdev->dev, fmi_counter_data,
2471 + data_size, DMA_FROM_DEVICE);
2472 + if (dma_mapping_error(&priv->si->pdev->dev, dma)) {
2473 + netdev_err(priv->si->ndev, "DMA mapping failed!\n");
2474 + kfree(fmi_counter_data);
2477 + cbdr->addr[0] = lower_32_bits(dma);
2478 + cbdr->addr[1] = upper_32_bits(dma);
2480 + dma_size = cpu_to_le16(data_size);
2481 + cbdr->length = dma_size;
2483 + xmit_cbdr(priv->si, curr_cbd);
2485 + memcpy(counters, fmi_counter_data, sizeof(*counters));
2490 +int enetc_qbu_set(struct net_device *ndev, u8 ptvector)
2494 + struct enetc_ndev_priv *priv = netdev_priv(ndev);
2496 + temp = enetc_rd(&priv->si->hw, ENETC_QBV_PTGCR_OFFSET);
2497 + if (temp & ENETC_QBV_TGE)
2498 + enetc_wr(&priv->si->hw, ENETC_QBV_PTGCR_OFFSET,
2499 + temp & (~ENETC_QBV_TGPE));
2501 + for (i = 0; i < 8; i++) {
2502 + /* 1 Enabled. Traffic is transmitted on the preemptive MAC. */
2503 + temp = enetc_port_rd(&priv->si->hw, ENETC_PTCFPR(i));
2505 + if ((ptvector >> i) & 0x1)
2506 + enetc_port_wr(&priv->si->hw,
2508 + temp | ENETC_FPE);
2510 + enetc_port_wr(&priv->si->hw,
2512 + temp & ~ENETC_FPE);
2518 +int enetc_qbu_get(struct net_device *ndev,
2519 + struct tsn_preempt_status *preemptstat)
2522 + struct enetc_ndev_priv *priv = netdev_priv(ndev);
2524 + if (enetc_port_rd(&priv->si->hw, ENETC_PFPMR) & ENETC_PFPMR_PMACE) {
2525 + preemptstat->preemption_active = true;
2526 + if (enetc_rd(&priv->si->hw, ENETC_QBV_PTGCR_OFFSET)
2528 + preemptstat->hold_request = 1;
2530 + preemptstat->hold_request = 2;
2532 + preemptstat->preemption_active = false;
2536 + for (i = 0; i < 8; i++)
2537 + if (enetc_port_rd(&priv->si->hw, ENETC_PTCFPR(i)) & 0x80000000)
2538 + preemptstat->admin_state |= 1 << i;
2540 + preemptstat->hold_advance =
2541 + enetc_rd(&priv->si->hw, ENETC_QBV_PTGCR_OFFSET) & 0xFFFF;
2542 + preemptstat->release_advance =
2543 + enetc_rd(&priv->si->hw, ENETC_QBV_PTGCR_OFFSET) & 0xFFFF;
2548 +u32 __enetc_tsn_get_cap(struct enetc_si *si)
2553 + reg = enetc_port_rd(&si->hw, ENETC_PCAPR0);
2555 + if (reg & ENETC_PCAPR0_PSFP)
2556 + cap |= TSN_CAP_QCI;
2558 + if (reg & ENETC_PCAPR0_TSN)
2559 + cap |= TSN_CAP_QBV;
2561 + if (reg & ENETC_PCAPR0_QBU)
2562 + cap |= TSN_CAP_QBU;
2564 + cap |= TSN_CAP_CBS;
2565 + cap |= TSN_CAP_TBS;
2570 +u32 enetc_tsn_get_capability(struct net_device *ndev)
2572 + struct enetc_ndev_priv *priv = netdev_priv(ndev);
2574 + return __enetc_tsn_get_cap(priv->si);
2577 +static int __enetc_get_max_cap(struct enetc_si *si,
2578 + struct tsn_qci_psfp_stream_param *stream_para)
2582 + /* Port stream filter capability */
2583 + reg = enetc_port_rd(&si->hw, ENETC_PSFCAPR);
2584 + stream_para->max_sf_instance = reg & ENETC_PSFCAPR_MSK;
2585 + /* Port stream filter capability */
2586 + reg = enetc_port_rd(&si->hw, ENETC_PSGCAPR);
2587 + stream_para->max_sg_instance = (reg & ENETC_PSGCAPR_SGIT_MSK);
2588 + stream_para->supported_list_max = (reg & ENETC_PSGCAPR_GCL_MSK) >> 16;
2589 + /* Port flow meter capability */
2590 + reg = enetc_port_rd(&si->hw, ENETC_PFMCAPR);
2591 + stream_para->max_fm_instance = reg & ENETC_PFMCAPR_MSK;
2596 +int enetc_get_max_cap(struct net_device *ndev,
2597 + struct tsn_qci_psfp_stream_param *stream_para)
2599 + struct enetc_ndev_priv *priv = netdev_priv(ndev);
2601 + return __enetc_get_max_cap(priv->si, stream_para);
2604 +static int enetc_set_cbs(struct net_device *ndev, u8 tc, u8 bw)
2606 + struct enetc_ndev_priv *priv = netdev_priv(ndev);
2607 + struct enetc_si *si = priv->si;
2608 + struct enetc_cbs *ecbs = si->ecbs;
2612 + u32 port_transmit_rate;
2613 + u32 port_frame_max_size;
2617 + u32 max_interfrence_size;
2624 + port_transmit_rate = get_ndev_speed(si->ndev);
2625 + if (port_transmit_rate != ecbs->port_transmit_rate)
2626 + ecbs->port_transmit_rate = port_transmit_rate;
2627 + port_frame_max_size = ecbs->port_max_size_frame;
2628 + tc_nums = ecbs->tc_nums;
2631 + if (tc >= tc_nums) {
2632 + dev_err(&ndev->dev, "Make sure the TC less than %d\n", tc_nums);
2637 + if (cbs[tc].enable) {
2638 + /* Make sure the other TC that are numerically
2639 + * lower than this TC have been disabled.
2641 + for (i = 0; i < tc; i++) {
2642 + if (cbs[i].enable)
2646 + dev_err(&ndev->dev,
2647 + "TC%d has been disabled first\n", i);
2650 + memset(&cbs[tc], 0, sizeof(*cbs));
2651 + cbs[tc].enable = false;
2652 + enetc_port_wr(&si->hw, ENETC_PTCCBSR1(tc), 0);
2653 + enetc_port_wr(&si->hw, ENETC_PTCCBSR0(tc), 0);
2658 + /* Make sure the other TC that are numerically
2659 + * higher than this TC have been enabled.
2661 + for (i = tc_nums - 1; i > tc; i--) {
2662 + if (!cbs[i].enable) {
2663 + dev_err(&ndev->dev,
2664 + "TC%d has been enabled first\n", i);
2667 + bw_sum += cbs[i].bw;
2670 + if (bw_sum + bw >= 100) {
2671 + dev_err(&ndev->dev,
2672 + "The sum of all CBS Bandwidth cann't exceed 100\n");
2677 + cbs[tc].tc_max_sized_frame = enetc_port_rd(&si->hw, ENETC_PTCMSDUR(tc));
2678 + cbs[tc].idle_slope = port_transmit_rate / 100 * bw;
2679 + cbs[tc].send_slope = port_transmit_rate - cbs[tc].idle_slope;
2681 + /* For TC7, the max_interfrence_size is ENETC_MAC_MAXFRM_SIZE.
2682 + * For TC6, the max_interfrence_size is calculated as below:
2684 + * max_interfrence_size = (M0 + Ma + Ra * M0 / (R0 - Ra))
2686 + * For other traffic class, for example SR class Q:
2688 + * R0 * (M0 + Ma + ... + Mp)
2689 + * max_interfrence_size = ------------------------------
2690 + * (R0 - Ra) + ... + (R0 - Rp)
2694 + if (tc == tc_nums - 1) {
2695 + cbs[tc].max_interfrence_size = port_frame_max_size * 8;
2697 + } else if (tc == tc_nums - 2) {
2698 + cbs[tc].max_interfrence_size = (port_frame_max_size
2699 + + cbs[tc + 1].tc_max_sized_frame
2700 + + port_frame_max_size * (cbs[tc + 1].idle_slope
2701 + / cbs[tc + 1].send_slope)) * 8;
2703 + max_interfrence_size = port_frame_max_size;
2705 + for (i = tc + 1; i < tc_nums; i++) {
2706 + send_slope += cbs[i].send_slope;
2707 + max_interfrence_size += cbs[i].tc_max_sized_frame;
2709 + max_interfrence_size = ((u64)port_transmit_rate
2710 + * max_interfrence_size) / send_slope;
2711 + cbs[tc].max_interfrence_size = max_interfrence_size * 8;
2714 + cbs[tc].hi_credit = cbs[tc].max_interfrence_size * cbs[tc].bw / 100;
2715 + cbs[tc].lo_credit = cbs[tc].tc_max_sized_frame * (cbs[tc].send_slope
2716 + / port_transmit_rate);
2719 + hi_credit = (ENETC_CLK * 100L) * (u64)cbs[tc].hi_credit
2720 + / port_transmit_rate;
2721 + enetc_port_wr(&si->hw, ENETC_PTCCBSR1(tc), hi_credit);
2723 + /* Set bw register and enable this traffic class*/
2724 + enetc_port_wr(&si->hw, ENETC_PTCCBSR0(tc),
2725 + (cbs[tc].bw & 0x7F) | (1 << 31));
2726 + cbs[tc].enable = true;
2731 +static int enetc_get_cbs(struct net_device *ndev, u8 tc)
2733 + struct enetc_ndev_priv *priv = netdev_priv(ndev);
2734 + struct enetc_si *si = priv->si;
2735 + struct enetc_cbs *ecbs = si->ecbs;
2741 + if (tc >= ecbs->tc_nums) {
2742 + dev_err(&ndev->dev, "The maximum of TC is %d\n", ecbs->tc_nums);
2746 + return cbs[tc].bw;
2749 +static int enetc_set_tsd(struct net_device *ndev, struct tsn_tsd *ttsd)
2754 +static int enetc_get_tsd(struct net_device *ndev, struct tsn_tsd_status *tts)
2759 +static u32 get_ndev_speed(struct net_device *netdev)
2761 + struct ethtool_link_ksettings ksettings;
2764 + if (netdev->ethtool_ops->get_link_ksettings) {
2765 + if (netdev->ethtool_ops->begin) {
2766 + rc = netdev->ethtool_ops->begin(netdev);
2771 + memset(&ksettings, 0, sizeof(ksettings));
2773 + if (!netdev->ethtool_ops->get_link_ksettings)
2776 + rc = netdev->ethtool_ops->get_link_ksettings(netdev,
2779 + if (netdev->ethtool_ops->complete)
2780 + netdev->ethtool_ops->complete(netdev);
2783 + return (rc < 0) ? 0 : ksettings.base.speed;
2786 +static void enetc_cbs_init(struct enetc_si *si)
2788 + struct enetc_ndev_priv *priv = netdev_priv(si->ndev);
2791 + tc_nums = priv->num_tx_rings;
2792 + si->ecbs = kzalloc(sizeof(*si->ecbs) +
2793 + sizeof(struct cbs) * tc_nums, GFP_KERNEL);
2797 + si->ecbs->port_max_size_frame = si->ndev->mtu + ETH_HLEN
2798 + + VLAN_HLEN + ETH_FCS_LEN;
2799 + si->ecbs->tc_nums = tc_nums;
2800 + si->ecbs->port_transmit_rate = get_ndev_speed(si->ndev);
2802 + /*This trick is used only for CFP*/
2803 + if (!si->ecbs->port_transmit_rate)
2804 + si->ecbs->port_transmit_rate = 1000000000;
2806 + if (!si->ecbs->port_transmit_rate) {
2807 + dev_err(&si->pdev->dev, "Failure to get port speed for CBS\n");
2813 +static void enetc_qbv_init(struct enetc_hw *hw)
2815 + /* Set PSPEED to be 1Gbps */
2816 + enetc_port_wr(hw, ENETC_PMR,
2817 + (enetc_port_rd(hw, ENETC_PMR)
2818 + & (~ENETC_PMR_PSPEED_MASK))
2819 + | ENETC_PMR_PSPEED_1000M);
2822 +void enetc_tsn_init(struct net_device *ndev)
2824 + struct enetc_ndev_priv *priv = netdev_priv(ndev);
2825 + struct enetc_si *si = priv->si;
2826 + u32 capability = 0;
2828 + capability = __enetc_tsn_get_cap(si);
2830 + if (capability & TSN_CAP_CBS)
2831 + enetc_cbs_init(si);
2833 + if (capability & TSN_CAP_QBV)
2834 + enetc_qbv_init(&si->hw);
2836 + if (capability & TSN_CAP_QCI)
2837 + enetc_qci_enable(&si->hw);
2839 + dev_info(&si->pdev->dev, "%s: setup done\n", __func__);
2842 +void enetc_tsn_deinit(struct net_device *ndev)
2844 + struct enetc_ndev_priv *priv = netdev_priv(ndev);
2845 + struct enetc_si *si = priv->si;
2847 + dev_info(&si->pdev->dev, "%s: release\n", __func__);
2850 +static struct tsn_ops enetc_tsn_ops_full = {
2851 + .device_init = enetc_tsn_init,
2852 + .device_deinit = enetc_tsn_deinit,
2853 + .get_capability = enetc_tsn_get_capability,
2854 + .qbv_set = enetc_qbv_set,
2855 + .qbv_get = enetc_qbv_get,
2856 + .qbv_get_status = enetc_qbv_get_status,
2857 + .cb_streamid_set = enetc_cb_streamid_set,
2858 + .cb_streamid_get = enetc_cb_streamid_get,
2859 + .cb_streamid_counters_get = enetc_cb_streamid_counters_get,
2860 + .qci_get_maxcap = enetc_get_max_cap,
2861 + .qci_sfi_set = enetc_qci_sfi_set,
2862 + .qci_sfi_get = enetc_qci_sfi_get,
2863 + .qci_sfi_counters_get = enetc_qci_sfi_counters_get,
2864 + .qci_sgi_set = enetc_qci_sgi_set,
2865 + .qci_sgi_get = enetc_qci_sgi_get,
2866 + .qci_sgi_status_get = enetc_qci_sgi_status_get,
2867 + .qci_fmi_set = enetc_qci_fmi_set,
2868 + .qci_fmi_get = enetc_qci_fmi_get,
2869 + .qbu_set = enetc_qbu_set,
2870 + .qbu_get = enetc_qbu_get,
2871 + .cbs_set = enetc_set_cbs,
2872 + .cbs_get = enetc_get_cbs,
2873 + .tsd_set = enetc_set_tsd,
2874 + .tsd_get = enetc_get_tsd,
2877 +static struct tsn_ops enetc_tsn_ops_part = {
2878 + .device_init = enetc_tsn_init,
2879 + .device_deinit = enetc_tsn_deinit,
2880 + .get_capability = enetc_tsn_get_capability,
2881 + .cb_streamid_set = enetc_cb_streamid_set,
2882 + .cb_streamid_get = enetc_cb_streamid_get,
2883 + .cb_streamid_counters_get = enetc_cb_streamid_counters_get,
2884 + .qci_get_maxcap = enetc_get_max_cap,
2885 + .qci_sfi_set = enetc_qci_sfi_set,
2886 + .qci_sfi_get = enetc_qci_sfi_get,
2887 + .qci_sfi_counters_get = enetc_qci_sfi_counters_get,
2888 + .qci_sgi_set = enetc_qci_sgi_set,
2889 + .qci_sgi_get = enetc_qci_sgi_get,
2890 + .qci_sgi_status_get = enetc_qci_sgi_status_get,
2891 + .qci_fmi_set = enetc_qci_fmi_set,
2892 + .qci_fmi_get = enetc_qci_fmi_get,
2895 +void enetc_tsn_pf_init(struct net_device *netdev, struct pci_dev *pdev)
2897 + int port = pdev->devfn & 0x7;
2899 + if (port == 1 || port == 3)
2900 + tsn_port_register(netdev, &enetc_tsn_ops_part,
2901 + (u16)pdev->bus->number);
2903 + tsn_port_register(netdev, &enetc_tsn_ops_full,
2904 + (u16)pdev->bus->number);
2907 +void enetc_tsn_pf_deinit(struct net_device *netdev)
2909 + tsn_port_unregister(netdev);
2911 +#endif /* #if IS_ENABLED(CONFIG_ENETC_TSN) */