xref: /dpdk/drivers/common/sfc_efx/base/medford2_nic.c (revision 950fe1ed0818e4e3582f62f02030c834decf09c0)
15e111ed8SAndrew Rybchenko /* SPDX-License-Identifier: BSD-3-Clause
25e111ed8SAndrew Rybchenko  *
3672386c1SAndrew Rybchenko  * Copyright(c) 2019-2021 Xilinx, Inc.
45e111ed8SAndrew Rybchenko  * Copyright(c) 2015-2019 Solarflare Communications Inc.
55e111ed8SAndrew Rybchenko  */
65e111ed8SAndrew Rybchenko 
75e111ed8SAndrew Rybchenko #include "efx.h"
85e111ed8SAndrew Rybchenko #include "efx_impl.h"
95e111ed8SAndrew Rybchenko 
105e111ed8SAndrew Rybchenko 
115e111ed8SAndrew Rybchenko #if EFSYS_OPT_MEDFORD2
125e111ed8SAndrew Rybchenko 
135e111ed8SAndrew Rybchenko static	__checkReturn	efx_rc_t
medford2_nic_get_required_pcie_bandwidth(__in efx_nic_t * enp,__out uint32_t * bandwidth_mbpsp)145e111ed8SAndrew Rybchenko medford2_nic_get_required_pcie_bandwidth(
155e111ed8SAndrew Rybchenko 	__in		efx_nic_t *enp,
165e111ed8SAndrew Rybchenko 	__out		uint32_t *bandwidth_mbpsp)
175e111ed8SAndrew Rybchenko {
185e111ed8SAndrew Rybchenko 	uint32_t bandwidth;
195e111ed8SAndrew Rybchenko 	efx_rc_t rc;
205e111ed8SAndrew Rybchenko 
215e111ed8SAndrew Rybchenko 	/* FIXME: support new Medford2 dynamic port modes */
225e111ed8SAndrew Rybchenko 
235e111ed8SAndrew Rybchenko 	if ((rc = ef10_nic_get_port_mode_bandwidth(enp,
245e111ed8SAndrew Rybchenko 						    &bandwidth)) != 0)
255e111ed8SAndrew Rybchenko 		goto fail1;
265e111ed8SAndrew Rybchenko 
275e111ed8SAndrew Rybchenko 	*bandwidth_mbpsp = bandwidth;
285e111ed8SAndrew Rybchenko 
295e111ed8SAndrew Rybchenko 	return (0);
305e111ed8SAndrew Rybchenko 
315e111ed8SAndrew Rybchenko fail1:
325e111ed8SAndrew Rybchenko 	EFSYS_PROBE1(fail1, efx_rc_t, rc);
335e111ed8SAndrew Rybchenko 
345e111ed8SAndrew Rybchenko 	return (rc);
355e111ed8SAndrew Rybchenko }
365e111ed8SAndrew Rybchenko 
375e111ed8SAndrew Rybchenko 	__checkReturn	efx_rc_t
medford2_board_cfg(__in efx_nic_t * enp)385e111ed8SAndrew Rybchenko medford2_board_cfg(
395e111ed8SAndrew Rybchenko 	__in		efx_nic_t *enp)
405e111ed8SAndrew Rybchenko {
415e111ed8SAndrew Rybchenko 	efx_nic_cfg_t *encp = &(enp->en_nic_cfg);
425e111ed8SAndrew Rybchenko 	uint32_t sysclk, dpcpu_clk;
435e111ed8SAndrew Rybchenko 	uint32_t end_padding;
445e111ed8SAndrew Rybchenko 	uint32_t bandwidth;
455e111ed8SAndrew Rybchenko 	efx_rc_t rc;
465e111ed8SAndrew Rybchenko 
475e111ed8SAndrew Rybchenko 	/*
4892fedcd3SIvan Malov 	 * Event queue creation is complete when an
4992fedcd3SIvan Malov 	 * EVQ_INIT_DONE_EV event is received.
5092fedcd3SIvan Malov 	 */
5192fedcd3SIvan Malov 	encp->enc_evq_init_done_ev_supported = B_TRUE;
5292fedcd3SIvan Malov 
5392fedcd3SIvan Malov 	/*
545e111ed8SAndrew Rybchenko 	 * Enable firmware workarounds for hardware errata.
555e111ed8SAndrew Rybchenko 	 * Expected responses are:
565e111ed8SAndrew Rybchenko 	 *  - 0 (zero):
575e111ed8SAndrew Rybchenko 	 *	Success: workaround enabled or disabled as requested.
585e111ed8SAndrew Rybchenko 	 *  - MC_CMD_ERR_ENOSYS (reported as ENOTSUP):
595e111ed8SAndrew Rybchenko 	 *	Firmware does not support the MC_CMD_WORKAROUND request.
605e111ed8SAndrew Rybchenko 	 *	(assume that the workaround is not supported).
615e111ed8SAndrew Rybchenko 	 *  - MC_CMD_ERR_ENOENT (reported as ENOENT):
625e111ed8SAndrew Rybchenko 	 *	Firmware does not support the requested workaround.
635e111ed8SAndrew Rybchenko 	 *  - MC_CMD_ERR_EPERM  (reported as EACCES):
645e111ed8SAndrew Rybchenko 	 *	Unprivileged function cannot enable/disable workarounds.
655e111ed8SAndrew Rybchenko 	 *
665e111ed8SAndrew Rybchenko 	 * See efx_mcdi_request_errcode() for MCDI error translations.
675e111ed8SAndrew Rybchenko 	 */
685e111ed8SAndrew Rybchenko 
695e111ed8SAndrew Rybchenko 
705e111ed8SAndrew Rybchenko 	if (EFX_PCI_FUNCTION_IS_VF(encp)) {
715e111ed8SAndrew Rybchenko 		/*
725e111ed8SAndrew Rybchenko 		 * Interrupt testing does not work for VFs on Medford2.
735e111ed8SAndrew Rybchenko 		 * See bug50084 and bug71432 comment 21.
745e111ed8SAndrew Rybchenko 		 */
755e111ed8SAndrew Rybchenko 		encp->enc_bug41750_workaround = B_TRUE;
765e111ed8SAndrew Rybchenko 	}
775e111ed8SAndrew Rybchenko 
785e111ed8SAndrew Rybchenko 	/*
795e111ed8SAndrew Rybchenko 	 * If the bug61265 workaround is enabled, then interrupt holdoff timers
805e111ed8SAndrew Rybchenko 	 * cannot be controlled by timer table writes, so MCDI must be used
815e111ed8SAndrew Rybchenko 	 * (timer table writes can still be used for wakeup timers).
825e111ed8SAndrew Rybchenko 	 */
835e111ed8SAndrew Rybchenko 	rc = efx_mcdi_set_workaround(enp, MC_CMD_WORKAROUND_BUG61265, B_TRUE,
845e111ed8SAndrew Rybchenko 	    NULL);
855e111ed8SAndrew Rybchenko 	if ((rc == 0) || (rc == EACCES))
865e111ed8SAndrew Rybchenko 		encp->enc_bug61265_workaround = B_TRUE;
875e111ed8SAndrew Rybchenko 	else if ((rc == ENOTSUP) || (rc == ENOENT))
885e111ed8SAndrew Rybchenko 		encp->enc_bug61265_workaround = B_FALSE;
895e111ed8SAndrew Rybchenko 	else
905e111ed8SAndrew Rybchenko 		goto fail1;
915e111ed8SAndrew Rybchenko 
925e111ed8SAndrew Rybchenko 	/* Checksums for TSO sends should always be correct on Medford2. */
935e111ed8SAndrew Rybchenko 	encp->enc_bug61297_workaround = B_FALSE;
945e111ed8SAndrew Rybchenko 
955e111ed8SAndrew Rybchenko 	/* Get clock frequencies (in MHz). */
965e111ed8SAndrew Rybchenko 	if ((rc = efx_mcdi_get_clock(enp, &sysclk, &dpcpu_clk)) != 0)
975e111ed8SAndrew Rybchenko 		goto fail2;
985e111ed8SAndrew Rybchenko 
995e111ed8SAndrew Rybchenko 	/*
1005e111ed8SAndrew Rybchenko 	 * The Medford2 timer quantum is 1536 dpcpu_clk cycles, documented for
1015e111ed8SAndrew Rybchenko 	 * the EV_TMR_VAL field of EV_TIMER_TBL. Scale for MHz and ns units.
1025e111ed8SAndrew Rybchenko 	 */
1035e111ed8SAndrew Rybchenko 	encp->enc_evq_timer_quantum_ns = 1536000UL / dpcpu_clk; /* 1536 cycles */
1045e111ed8SAndrew Rybchenko 	encp->enc_evq_timer_max_us = (encp->enc_evq_timer_quantum_ns <<
1055e111ed8SAndrew Rybchenko 		    FRF_CZ_TC_TIMER_VAL_WIDTH) / 1000;
1065e111ed8SAndrew Rybchenko 
1075e111ed8SAndrew Rybchenko 	encp->enc_ev_desc_size = EF10_EVQ_DESC_SIZE;
1085e111ed8SAndrew Rybchenko 	encp->enc_rx_desc_size = EF10_RXQ_DESC_SIZE;
1095e111ed8SAndrew Rybchenko 	encp->enc_tx_desc_size = EF10_TXQ_DESC_SIZE;
1105e111ed8SAndrew Rybchenko 
1115e111ed8SAndrew Rybchenko 	/* Alignment for receive packet DMA buffers */
1125e111ed8SAndrew Rybchenko 	encp->enc_rx_buf_align_start = 1;
1135e111ed8SAndrew Rybchenko 
1145e111ed8SAndrew Rybchenko 	/* Get the RX DMA end padding alignment configuration */
1155e111ed8SAndrew Rybchenko 	if ((rc = efx_mcdi_get_rxdp_config(enp, &end_padding)) != 0) {
1165e111ed8SAndrew Rybchenko 		if (rc != EACCES)
1175e111ed8SAndrew Rybchenko 			goto fail3;
1185e111ed8SAndrew Rybchenko 
1195e111ed8SAndrew Rybchenko 		/* Assume largest tail padding size supported by hardware */
1205e111ed8SAndrew Rybchenko 		end_padding = 256;
1215e111ed8SAndrew Rybchenko 	}
1225e111ed8SAndrew Rybchenko 	encp->enc_rx_buf_align_end = end_padding;
1235e111ed8SAndrew Rybchenko 
1245e111ed8SAndrew Rybchenko 	encp->enc_evq_max_nevs = EF10_EVQ_MAXNEVS;
1255e111ed8SAndrew Rybchenko 	encp->enc_evq_min_nevs = EF10_EVQ_MINNEVS;
1265e111ed8SAndrew Rybchenko 
1275e111ed8SAndrew Rybchenko 	encp->enc_rxq_max_ndescs = EF10_RXQ_MAXNDESCS;
1285e111ed8SAndrew Rybchenko 	encp->enc_rxq_min_ndescs = EF10_RXQ_MINNDESCS;
1295e111ed8SAndrew Rybchenko 
1305e111ed8SAndrew Rybchenko 	/*
1315e111ed8SAndrew Rybchenko 	 * The maximum supported transmit queue size is 2048. TXQs with 4096
1325e111ed8SAndrew Rybchenko 	 * descriptors are not supported as the top bit is used for vfifo
1335e111ed8SAndrew Rybchenko 	 * stuffing.
1345e111ed8SAndrew Rybchenko 	 */
1355e111ed8SAndrew Rybchenko 	encp->enc_txq_max_ndescs = MEDFORD2_TXQ_MAXNDESCS;
1365e111ed8SAndrew Rybchenko 	encp->enc_txq_min_ndescs = EF10_TXQ_MINNDESCS;
1375e111ed8SAndrew Rybchenko 
1385e111ed8SAndrew Rybchenko 	EFX_STATIC_ASSERT(MEDFORD2_PIOBUF_NBUFS <= EF10_MAX_PIOBUF_NBUFS);
1395e111ed8SAndrew Rybchenko 	encp->enc_piobuf_limit = MEDFORD2_PIOBUF_NBUFS;
1405e111ed8SAndrew Rybchenko 	encp->enc_piobuf_size = MEDFORD2_PIOBUF_SIZE;
1415e111ed8SAndrew Rybchenko 	encp->enc_piobuf_min_alloc_size = MEDFORD2_MIN_PIO_ALLOC_SIZE;
1425e111ed8SAndrew Rybchenko 
1435e111ed8SAndrew Rybchenko 	/*
1445e111ed8SAndrew Rybchenko 	 * Medford2 stores a single global copy of VPD, not per-PF as on
1455e111ed8SAndrew Rybchenko 	 * Huntington.
1465e111ed8SAndrew Rybchenko 	 */
1475e111ed8SAndrew Rybchenko 	encp->enc_vpd_is_global = B_TRUE;
1485e111ed8SAndrew Rybchenko 
1495e111ed8SAndrew Rybchenko 	rc = medford2_nic_get_required_pcie_bandwidth(enp, &bandwidth);
1505e111ed8SAndrew Rybchenko 	if (rc != 0)
1515e111ed8SAndrew Rybchenko 		goto fail4;
1525e111ed8SAndrew Rybchenko 	encp->enc_required_pcie_bandwidth_mbps = bandwidth;
1535e111ed8SAndrew Rybchenko 	encp->enc_max_pcie_link_gen = EFX_PCIE_LINK_SPEED_GEN3;
1545e111ed8SAndrew Rybchenko 
155*950fe1edSDenis Pryazhennikov 	encp->enc_table_api_supported = B_FALSE;
156*950fe1edSDenis Pryazhennikov 
1575e111ed8SAndrew Rybchenko 	return (0);
1585e111ed8SAndrew Rybchenko 
1595e111ed8SAndrew Rybchenko fail4:
1605e111ed8SAndrew Rybchenko 	EFSYS_PROBE(fail4);
1615e111ed8SAndrew Rybchenko fail3:
1625e111ed8SAndrew Rybchenko 	EFSYS_PROBE(fail3);
1635e111ed8SAndrew Rybchenko fail2:
1645e111ed8SAndrew Rybchenko 	EFSYS_PROBE(fail2);
1655e111ed8SAndrew Rybchenko fail1:
1665e111ed8SAndrew Rybchenko 	EFSYS_PROBE1(fail1, efx_rc_t, rc);
1675e111ed8SAndrew Rybchenko 
1685e111ed8SAndrew Rybchenko 	return (rc);
1695e111ed8SAndrew Rybchenko }
1705e111ed8SAndrew Rybchenko 
1715e111ed8SAndrew Rybchenko #endif	/* EFSYS_OPT_MEDFORD2 */
172