DPDK logo

Elixir Cross Referencer

/* SPDX-License-Identifier: BSD-3-Clause
 *
 * Copyright(c) 2019-2021 Xilinx, Inc.
 * Copyright(c) 2016-2019 Solarflare Communications Inc.
 *
 * This software was jointly developed between OKTET Labs (under contract
 * for Solarflare) and Solarflare Communications, Inc.
 */

#ifndef _SFC_EV_H_
#define _SFC_EV_H_

#include <ethdev_driver.h>

#include "efx.h"

#include "sfc.h"

#ifdef __cplusplus
extern "C" {
#endif

struct sfc_adapter;
struct sfc_dp_rxq;
struct sfc_dp_txq;

enum sfc_evq_state {
	SFC_EVQ_UNINITIALIZED = 0,
	SFC_EVQ_INITIALIZED,
	SFC_EVQ_STARTING,
	SFC_EVQ_STARTED,

	SFC_EVQ_NSTATES
};

enum sfc_evq_type {
	SFC_EVQ_TYPE_MGMT = 0,
	SFC_EVQ_TYPE_RX,
	SFC_EVQ_TYPE_TX,

	SFC_EVQ_NTYPES
};

struct sfc_evq {
	/* Used on datapath */
	efx_evq_t			*common;
	const efx_ev_callbacks_t	*callbacks;
	unsigned int			read_ptr;
	unsigned int			read_ptr_primed;
	boolean_t			exception;
	efsys_mem_t			mem;
	struct sfc_dp_rxq		*dp_rxq;
	struct sfc_dp_txq		*dp_txq;

	/* Not used on datapath */
	struct sfc_adapter		*sa;
	unsigned int			evq_index;
	enum sfc_evq_state		init_state;
	enum sfc_evq_type		type;
	unsigned int			entries;
};

static inline sfc_sw_index_t
sfc_mgmt_evq_sw_index(__rte_unused const struct sfc_adapter_shared *sas)
{
	return 0;
}

/* Return the number of Rx queues reserved for driver's internal use */
static inline unsigned int
sfc_nb_reserved_rxq(const struct sfc_adapter_shared *sas)
{
	return sfc_nb_counter_rxq(sas);
}

static inline unsigned int
sfc_nb_reserved_evq(const struct sfc_adapter_shared *sas)
{
	/* An EvQ is required for each reserved RxQ */
	return 1 + sfc_nb_reserved_rxq(sas);
}

/*
 * The mapping functions that return SW index of a specific reserved
 * queue rely on the relative order of reserved queues. Some reserved
 * queues are optional, and if they are disabled or not supported, then
 * the function for that specific reserved queue will return previous
 * valid index of a reserved queue in the dependency chain or
 * SFC_SW_INDEX_INVALID if it is the first reserved queue in the chain.
 * If at least one of the reserved queues in the chain is enabled, then
 * the corresponding function will give valid SW index, even if previous
 * functions in the chain returned SFC_SW_INDEX_INVALID, since this value
 * is one less than the first valid SW index.
 *
 * The dependency mechanism is utilized to avoid regid defines for SW indices
 * for reserved queues and to allow these indices to shrink and make space
 * for ethdev queue indices when some of the reserved queues are disabled.
 */

static inline sfc_sw_index_t
sfc_counters_rxq_sw_index(const struct sfc_adapter_shared *sas)
{
	return sas->counters_rxq_allocated ? 0 : SFC_SW_INDEX_INVALID;
}

/*
 * Functions below define event queue to transmit/receive queue and vice
 * versa mapping.
 * SFC_ETHDEV_QID_INVALID is returned when sw_index is converted to
 * ethdev_qid, but sw_index represents a reserved queue for driver's
 * internal use.
 * Own event queue is allocated for management, each Rx and each Tx queue.
 * Zero event queue is used for management events.
 * When counters are supported, one Rx event queue is reserved.
 * Rx event queues follow reserved event queues.
 * Tx event queues follow Rx event queues.
 */

static inline sfc_ethdev_qid_t
sfc_ethdev_rx_qid_by_rxq_sw_index(struct sfc_adapter_shared *sas,
				  sfc_sw_index_t rxq_sw_index)
{
	if (rxq_sw_index < sfc_nb_reserved_rxq(sas))
		return SFC_ETHDEV_QID_INVALID;

	return rxq_sw_index - sfc_nb_reserved_rxq(sas);
}

static inline sfc_sw_index_t
sfc_rxq_sw_index_by_ethdev_rx_qid(struct sfc_adapter_shared *sas,
				  sfc_ethdev_qid_t ethdev_qid)
{
	return sfc_nb_reserved_rxq(sas) + ethdev_qid;
}

static inline sfc_sw_index_t
sfc_evq_sw_index_by_rxq_sw_index(struct sfc_adapter *sa,
				 sfc_sw_index_t rxq_sw_index)
{
	struct sfc_adapter_shared *sas = sfc_sa2shared(sa);
	sfc_ethdev_qid_t ethdev_qid;

	ethdev_qid = sfc_ethdev_rx_qid_by_rxq_sw_index(sas, rxq_sw_index);
	if (ethdev_qid == SFC_ETHDEV_QID_INVALID) {
		/* One EvQ is reserved for management */
		return 1 + rxq_sw_index;
	}

	return sfc_nb_reserved_evq(sas) + ethdev_qid;
}

static inline sfc_ethdev_qid_t
sfc_ethdev_tx_qid_by_txq_sw_index(__rte_unused struct sfc_adapter_shared *sas,
				  sfc_sw_index_t txq_sw_index)
{
	/* Only ethdev queues are present for now */
	return txq_sw_index;
}

static inline sfc_sw_index_t
sfc_txq_sw_index_by_ethdev_tx_qid(__rte_unused struct sfc_adapter_shared *sas,
				  sfc_ethdev_qid_t ethdev_qid)
{
	/* Only ethdev queues are present for now */
	return ethdev_qid;
}

static inline sfc_sw_index_t
sfc_evq_sw_index_by_txq_sw_index(struct sfc_adapter *sa,
				 sfc_sw_index_t txq_sw_index)
{
	return sfc_nb_reserved_evq(sfc_sa2shared(sa)) +
		sa->eth_dev->data->nb_rx_queues + txq_sw_index;
}

int sfc_ev_attach(struct sfc_adapter *sa);
void sfc_ev_detach(struct sfc_adapter *sa);
int sfc_ev_start(struct sfc_adapter *sa);
void sfc_ev_stop(struct sfc_adapter *sa);

int sfc_ev_qinit(struct sfc_adapter *sa,
		 enum sfc_evq_type type, unsigned int type_index,
		 unsigned int entries, int socket_id, struct sfc_evq **evqp);
void sfc_ev_qfini(struct sfc_evq *evq);
int sfc_ev_qstart(struct sfc_evq *evq, unsigned int hw_index);
void sfc_ev_qstop(struct sfc_evq *evq);

int sfc_ev_qprime(struct sfc_evq *evq);
void sfc_ev_qpoll(struct sfc_evq *evq);

void sfc_ev_mgmt_qpoll(struct sfc_adapter *sa);

#ifdef __cplusplus
}
#endif
#endif /* _SFC_EV_H_ */