1/* SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB */
2/* Copyright (c) 2019, Mellanox Technologies */
3
4#ifndef	_DR_TYPES_
5#define	_DR_TYPES_
6
7#include <linux/mlx5/vport.h>
8#include <linux/refcount.h>
9#include "fs_core.h"
10#include "wq.h"
11#include "lib/mlx5.h"
12#include "mlx5_ifc_dr.h"
13#include "mlx5dr.h"
14#include "dr_dbg.h"
15
16#define DR_RULE_MAX_STES 18
17#define DR_ACTION_MAX_STES 5
18#define DR_STE_SVLAN 0x1
19#define DR_STE_CVLAN 0x2
20#define DR_SZ_MATCH_PARAM (MLX5_ST_SZ_DW_MATCH_PARAM * 4)
21#define DR_NUM_OF_FLEX_PARSERS 8
22#define DR_STE_MAX_FLEX_0_ID 3
23#define DR_STE_MAX_FLEX_1_ID 7
24#define DR_ACTION_CACHE_LINE_SIZE 64
25
26#define mlx5dr_err(dmn, arg...) mlx5_core_err((dmn)->mdev, ##arg)
27#define mlx5dr_info(dmn, arg...) mlx5_core_info((dmn)->mdev, ##arg)
28#define mlx5dr_dbg(dmn, arg...) mlx5_core_dbg((dmn)->mdev, ##arg)
29
30struct mlx5dr_ptrn_mgr;
31struct mlx5dr_arg_mgr;
32struct mlx5dr_arg_obj;
33
34static inline bool dr_is_flex_parser_0_id(u8 parser_id)
35{
36	return parser_id <= DR_STE_MAX_FLEX_0_ID;
37}
38
39static inline bool dr_is_flex_parser_1_id(u8 parser_id)
40{
41	return parser_id > DR_STE_MAX_FLEX_0_ID;
42}
43
44enum mlx5dr_icm_chunk_size {
45	DR_CHUNK_SIZE_1,
46	DR_CHUNK_SIZE_MIN = DR_CHUNK_SIZE_1, /* keep updated when changing */
47	DR_CHUNK_SIZE_2,
48	DR_CHUNK_SIZE_4,
49	DR_CHUNK_SIZE_8,
50	DR_CHUNK_SIZE_16,
51	DR_CHUNK_SIZE_32,
52	DR_CHUNK_SIZE_64,
53	DR_CHUNK_SIZE_128,
54	DR_CHUNK_SIZE_256,
55	DR_CHUNK_SIZE_512,
56	DR_CHUNK_SIZE_1K,
57	DR_CHUNK_SIZE_2K,
58	DR_CHUNK_SIZE_4K,
59	DR_CHUNK_SIZE_8K,
60	DR_CHUNK_SIZE_16K,
61	DR_CHUNK_SIZE_32K,
62	DR_CHUNK_SIZE_64K,
63	DR_CHUNK_SIZE_128K,
64	DR_CHUNK_SIZE_256K,
65	DR_CHUNK_SIZE_512K,
66	DR_CHUNK_SIZE_1024K,
67	DR_CHUNK_SIZE_2048K,
68	DR_CHUNK_SIZE_MAX,
69};
70
71enum mlx5dr_icm_type {
72	DR_ICM_TYPE_STE,
73	DR_ICM_TYPE_MODIFY_ACTION,
74	DR_ICM_TYPE_MODIFY_HDR_PTRN,
75	DR_ICM_TYPE_MAX,
76};
77
78static inline enum mlx5dr_icm_chunk_size
79mlx5dr_icm_next_higher_chunk(enum mlx5dr_icm_chunk_size chunk)
80{
81	chunk += 2;
82	if (chunk < DR_CHUNK_SIZE_MAX)
83		return chunk;
84
85	return DR_CHUNK_SIZE_MAX;
86}
87
88enum {
89	DR_STE_SIZE = 64,
90	DR_STE_SIZE_CTRL = 32,
91	DR_STE_SIZE_MATCH_TAG = 32,
92	DR_STE_SIZE_TAG = 16,
93	DR_STE_SIZE_MASK = 16,
94	DR_STE_SIZE_REDUCED = DR_STE_SIZE - DR_STE_SIZE_MASK,
95};
96
97enum mlx5dr_ste_ctx_action_cap {
98	DR_STE_CTX_ACTION_CAP_NONE = 0,
99	DR_STE_CTX_ACTION_CAP_TX_POP   = 1 << 0,
100	DR_STE_CTX_ACTION_CAP_RX_PUSH  = 1 << 1,
101	DR_STE_CTX_ACTION_CAP_RX_ENCAP = 1 << 2,
102	DR_STE_CTX_ACTION_CAP_POP_MDFY = 1 << 3,
103};
104
105enum {
106	DR_MODIFY_ACTION_SIZE = 8,
107};
108
109enum mlx5dr_matcher_criteria {
110	DR_MATCHER_CRITERIA_EMPTY = 0,
111	DR_MATCHER_CRITERIA_OUTER = 1 << 0,
112	DR_MATCHER_CRITERIA_MISC = 1 << 1,
113	DR_MATCHER_CRITERIA_INNER = 1 << 2,
114	DR_MATCHER_CRITERIA_MISC2 = 1 << 3,
115	DR_MATCHER_CRITERIA_MISC3 = 1 << 4,
116	DR_MATCHER_CRITERIA_MISC4 = 1 << 5,
117	DR_MATCHER_CRITERIA_MISC5 = 1 << 6,
118	DR_MATCHER_CRITERIA_MAX = 1 << 7,
119};
120
121enum mlx5dr_action_type {
122	DR_ACTION_TYP_TNL_L2_TO_L2,
123	DR_ACTION_TYP_L2_TO_TNL_L2,
124	DR_ACTION_TYP_TNL_L3_TO_L2,
125	DR_ACTION_TYP_L2_TO_TNL_L3,
126	DR_ACTION_TYP_DROP,
127	DR_ACTION_TYP_QP,
128	DR_ACTION_TYP_FT,
129	DR_ACTION_TYP_CTR,
130	DR_ACTION_TYP_TAG,
131	DR_ACTION_TYP_MODIFY_HDR,
132	DR_ACTION_TYP_VPORT,
133	DR_ACTION_TYP_POP_VLAN,
134	DR_ACTION_TYP_PUSH_VLAN,
135	DR_ACTION_TYP_INSERT_HDR,
136	DR_ACTION_TYP_REMOVE_HDR,
137	DR_ACTION_TYP_SAMPLER,
138	DR_ACTION_TYP_ASO_FLOW_METER,
139	DR_ACTION_TYP_RANGE,
140	DR_ACTION_TYP_MAX,
141};
142
143enum mlx5dr_ipv {
144	DR_RULE_IPV4,
145	DR_RULE_IPV6,
146	DR_RULE_IPV_MAX,
147};
148
149struct mlx5dr_icm_pool;
150struct mlx5dr_icm_chunk;
151struct mlx5dr_icm_buddy_mem;
152struct mlx5dr_ste_htbl;
153struct mlx5dr_match_param;
154struct mlx5dr_cmd_caps;
155struct mlx5dr_rule_rx_tx;
156struct mlx5dr_matcher_rx_tx;
157struct mlx5dr_ste_ctx;
158struct mlx5dr_send_info_pool;
159struct mlx5dr_icm_hot_chunk;
160
161struct mlx5dr_ste {
162	/* refcount: indicates the num of rules that using this ste */
163	u32 refcount;
164
165	/* this ste is part of a rule, located in ste's chain */
166	u8 ste_chain_location;
167
168	/* attached to the miss_list head at each htbl entry */
169	struct list_head miss_list_node;
170
171	/* this ste is member of htbl */
172	struct mlx5dr_ste_htbl *htbl;
173
174	struct mlx5dr_ste_htbl *next_htbl;
175
176	/* The rule this STE belongs to */
177	struct mlx5dr_rule_rx_tx *rule_rx_tx;
178};
179
180struct mlx5dr_ste_htbl_ctrl {
181	/* total number of valid entries belonging to this hash table. This
182	 * includes the non collision and collision entries
183	 */
184	unsigned int num_of_valid_entries;
185
186	/* total number of collisions entries attached to this table */
187	unsigned int num_of_collisions;
188};
189
190struct mlx5dr_ste_htbl {
191	u16 lu_type;
192	u16 byte_mask;
193	u32 refcount;
194	struct mlx5dr_icm_chunk *chunk;
195	struct mlx5dr_ste *pointing_ste;
196	struct mlx5dr_ste_htbl_ctrl ctrl;
197};
198
199struct mlx5dr_ste_send_info {
200	struct mlx5dr_ste *ste;
201	struct list_head send_list;
202	u16 size;
203	u16 offset;
204	u8 data_cont[DR_STE_SIZE];
205	u8 *data;
206};
207
208void mlx5dr_send_fill_and_append_ste_send_info(struct mlx5dr_ste *ste, u16 size,
209					       u16 offset, u8 *data,
210					       struct mlx5dr_ste_send_info *ste_info,
211					       struct list_head *send_list,
212					       bool copy_data);
213
214struct mlx5dr_ste_build {
215	u8 inner:1;
216	u8 rx:1;
217	u8 vhca_id_valid:1;
218	struct mlx5dr_domain *dmn;
219	struct mlx5dr_cmd_caps *caps;
220	u16 lu_type;
221	u16 byte_mask;
222	u8 bit_mask[DR_STE_SIZE_MASK];
223	int (*ste_build_tag_func)(struct mlx5dr_match_param *spec,
224				  struct mlx5dr_ste_build *sb,
225				  u8 *tag);
226};
227
228struct mlx5dr_ste_htbl *
229mlx5dr_ste_htbl_alloc(struct mlx5dr_icm_pool *pool,
230		      enum mlx5dr_icm_chunk_size chunk_size,
231		      u16 lu_type, u16 byte_mask);
232
233int mlx5dr_ste_htbl_free(struct mlx5dr_ste_htbl *htbl);
234
235static inline void mlx5dr_htbl_put(struct mlx5dr_ste_htbl *htbl)
236{
237	htbl->refcount--;
238	if (!htbl->refcount)
239		mlx5dr_ste_htbl_free(htbl);
240}
241
242static inline void mlx5dr_htbl_get(struct mlx5dr_ste_htbl *htbl)
243{
244	htbl->refcount++;
245}
246
247/* STE utils */
248u32 mlx5dr_ste_calc_hash_index(u8 *hw_ste_p, struct mlx5dr_ste_htbl *htbl);
249bool mlx5dr_ste_is_miss_addr_set(struct mlx5dr_ste_ctx *ste_ctx, u8 *hw_ste_p);
250void mlx5dr_ste_set_miss_addr(struct mlx5dr_ste_ctx *ste_ctx,
251			      u8 *hw_ste, u64 miss_addr);
252void mlx5dr_ste_set_hit_addr(struct mlx5dr_ste_ctx *ste_ctx,
253			     u8 *hw_ste, u64 icm_addr, u32 ht_size);
254void mlx5dr_ste_set_hit_addr_by_next_htbl(struct mlx5dr_ste_ctx *ste_ctx,
255					  u8 *hw_ste,
256					  struct mlx5dr_ste_htbl *next_htbl);
257void mlx5dr_ste_set_bit_mask(u8 *hw_ste_p, u8 *bit_mask);
258bool mlx5dr_ste_is_last_in_rule(struct mlx5dr_matcher_rx_tx *nic_matcher,
259				u8 ste_location);
260u64 mlx5dr_ste_get_icm_addr(struct mlx5dr_ste *ste);
261u64 mlx5dr_ste_get_mr_addr(struct mlx5dr_ste *ste);
262struct list_head *mlx5dr_ste_get_miss_list(struct mlx5dr_ste *ste);
263
264#define MLX5DR_MAX_VLANS 2
265#define MLX5DR_INVALID_PATTERN_INDEX 0xffffffff
266
267struct mlx5dr_ste_actions_attr {
268	u32	modify_index;
269	u32	modify_pat_idx;
270	u16	modify_actions;
271	u8	*single_modify_action;
272	u32	decap_index;
273	u32	decap_pat_idx;
274	u16	decap_actions;
275	u8	decap_with_vlan:1;
276	u64	final_icm_addr;
277	u32	flow_tag;
278	u32	ctr_id;
279	u16	gvmi;
280	u16	hit_gvmi;
281	struct {
282		u32	id;
283		u32	size;
284		u8	param_0;
285		u8	param_1;
286	} reformat;
287	struct {
288		int	count;
289		u32	headers[MLX5DR_MAX_VLANS];
290	} vlans;
291
292	struct {
293		u32 obj_id;
294		u32 offset;
295		u8 dest_reg_id;
296		u8 init_color;
297	} aso_flow_meter;
298
299	struct {
300		u64	miss_icm_addr;
301		u32	definer_id;
302		u32	min;
303		u32	max;
304	} range;
305};
306
307void mlx5dr_ste_set_actions_rx(struct mlx5dr_ste_ctx *ste_ctx,
308			       struct mlx5dr_domain *dmn,
309			       u8 *action_type_set,
310			       u8 *last_ste,
311			       struct mlx5dr_ste_actions_attr *attr,
312			       u32 *added_stes);
313void mlx5dr_ste_set_actions_tx(struct mlx5dr_ste_ctx *ste_ctx,
314			       struct mlx5dr_domain *dmn,
315			       u8 *action_type_set,
316			       u8 *last_ste,
317			       struct mlx5dr_ste_actions_attr *attr,
318			       u32 *added_stes);
319
320void mlx5dr_ste_set_action_set(struct mlx5dr_ste_ctx *ste_ctx,
321			       __be64 *hw_action,
322			       u8 hw_field,
323			       u8 shifter,
324			       u8 length,
325			       u32 data);
326void mlx5dr_ste_set_action_add(struct mlx5dr_ste_ctx *ste_ctx,
327			       __be64 *hw_action,
328			       u8 hw_field,
329			       u8 shifter,
330			       u8 length,
331			       u32 data);
332void mlx5dr_ste_set_action_copy(struct mlx5dr_ste_ctx *ste_ctx,
333				__be64 *hw_action,
334				u8 dst_hw_field,
335				u8 dst_shifter,
336				u8 dst_len,
337				u8 src_hw_field,
338				u8 src_shifter);
339int mlx5dr_ste_set_action_decap_l3_list(struct mlx5dr_ste_ctx *ste_ctx,
340					void *data,
341					u32 data_sz,
342					u8 *hw_action,
343					u32 hw_action_sz,
344					u16 *used_hw_action_num);
345int mlx5dr_ste_alloc_modify_hdr(struct mlx5dr_action *action);
346void mlx5dr_ste_free_modify_hdr(struct mlx5dr_action *action);
347
348const struct mlx5dr_ste_action_modify_field *
349mlx5dr_ste_conv_modify_hdr_sw_field(struct mlx5dr_ste_ctx *ste_ctx, u16 sw_field);
350
351struct mlx5dr_ste_ctx *mlx5dr_ste_get_ctx(u8 version);
352void mlx5dr_ste_free(struct mlx5dr_ste *ste,
353		     struct mlx5dr_matcher *matcher,
354		     struct mlx5dr_matcher_rx_tx *nic_matcher);
355static inline void mlx5dr_ste_put(struct mlx5dr_ste *ste,
356				  struct mlx5dr_matcher *matcher,
357				  struct mlx5dr_matcher_rx_tx *nic_matcher)
358{
359	ste->refcount--;
360	if (!ste->refcount)
361		mlx5dr_ste_free(ste, matcher, nic_matcher);
362}
363
364/* initial as 0, increased only when ste appears in a new rule */
365static inline void mlx5dr_ste_get(struct mlx5dr_ste *ste)
366{
367	ste->refcount++;
368}
369
370static inline bool mlx5dr_ste_is_not_used(struct mlx5dr_ste *ste)
371{
372	return !ste->refcount;
373}
374
375bool mlx5dr_ste_equal_tag(void *src, void *dst);
376int mlx5dr_ste_create_next_htbl(struct mlx5dr_matcher *matcher,
377				struct mlx5dr_matcher_rx_tx *nic_matcher,
378				struct mlx5dr_ste *ste,
379				u8 *cur_hw_ste,
380				enum mlx5dr_icm_chunk_size log_table_size);
381
382/* STE build functions */
383int mlx5dr_ste_build_pre_check(struct mlx5dr_domain *dmn,
384			       u8 match_criteria,
385			       struct mlx5dr_match_param *mask,
386			       struct mlx5dr_match_param *value);
387int mlx5dr_ste_build_ste_arr(struct mlx5dr_matcher *matcher,
388			     struct mlx5dr_matcher_rx_tx *nic_matcher,
389			     struct mlx5dr_match_param *value,
390			     u8 *ste_arr);
391void mlx5dr_ste_build_eth_l2_src_dst(struct mlx5dr_ste_ctx *ste_ctx,
392				     struct mlx5dr_ste_build *builder,
393				     struct mlx5dr_match_param *mask,
394				     bool inner, bool rx);
395void mlx5dr_ste_build_eth_l3_ipv4_5_tuple(struct mlx5dr_ste_ctx *ste_ctx,
396					  struct mlx5dr_ste_build *sb,
397					  struct mlx5dr_match_param *mask,
398					  bool inner, bool rx);
399void mlx5dr_ste_build_eth_l3_ipv4_misc(struct mlx5dr_ste_ctx *ste_ctx,
400				       struct mlx5dr_ste_build *sb,
401				       struct mlx5dr_match_param *mask,
402				       bool inner, bool rx);
403void mlx5dr_ste_build_eth_l3_ipv6_dst(struct mlx5dr_ste_ctx *ste_ctx,
404				      struct mlx5dr_ste_build *sb,
405				      struct mlx5dr_match_param *mask,
406				      bool inner, bool rx);
407void mlx5dr_ste_build_eth_l3_ipv6_src(struct mlx5dr_ste_ctx *ste_ctx,
408				      struct mlx5dr_ste_build *sb,
409				      struct mlx5dr_match_param *mask,
410				      bool inner, bool rx);
411void mlx5dr_ste_build_eth_l2_src(struct mlx5dr_ste_ctx *ste_ctx,
412				 struct mlx5dr_ste_build *sb,
413				 struct mlx5dr_match_param *mask,
414				 bool inner, bool rx);
415void mlx5dr_ste_build_eth_l2_dst(struct mlx5dr_ste_ctx *ste_ctx,
416				 struct mlx5dr_ste_build *sb,
417				 struct mlx5dr_match_param *mask,
418				 bool inner, bool rx);
419void mlx5dr_ste_build_eth_l2_tnl(struct mlx5dr_ste_ctx *ste_ctx,
420				 struct mlx5dr_ste_build *sb,
421				 struct mlx5dr_match_param *mask,
422				 bool inner, bool rx);
423void mlx5dr_ste_build_eth_ipv6_l3_l4(struct mlx5dr_ste_ctx *ste_ctx,
424				     struct mlx5dr_ste_build *sb,
425				     struct mlx5dr_match_param *mask,
426				     bool inner, bool rx);
427void mlx5dr_ste_build_eth_l4_misc(struct mlx5dr_ste_ctx *ste_ctx,
428				  struct mlx5dr_ste_build *sb,
429				  struct mlx5dr_match_param *mask,
430				  bool inner, bool rx);
431void mlx5dr_ste_build_tnl_gre(struct mlx5dr_ste_ctx *ste_ctx,
432			      struct mlx5dr_ste_build *sb,
433			      struct mlx5dr_match_param *mask,
434			      bool inner, bool rx);
435void mlx5dr_ste_build_mpls(struct mlx5dr_ste_ctx *ste_ctx,
436			   struct mlx5dr_ste_build *sb,
437			   struct mlx5dr_match_param *mask,
438			   bool inner, bool rx);
439void mlx5dr_ste_build_tnl_mpls_over_gre(struct mlx5dr_ste_ctx *ste_ctx,
440					struct mlx5dr_ste_build *sb,
441					struct mlx5dr_match_param *mask,
442					struct mlx5dr_cmd_caps *caps,
443					bool inner, bool rx);
444void mlx5dr_ste_build_tnl_mpls_over_udp(struct mlx5dr_ste_ctx *ste_ctx,
445					struct mlx5dr_ste_build *sb,
446					struct mlx5dr_match_param *mask,
447					struct mlx5dr_cmd_caps *caps,
448					bool inner, bool rx);
449void mlx5dr_ste_build_icmp(struct mlx5dr_ste_ctx *ste_ctx,
450			   struct mlx5dr_ste_build *sb,
451			   struct mlx5dr_match_param *mask,
452			   struct mlx5dr_cmd_caps *caps,
453			   bool inner, bool rx);
454void mlx5dr_ste_build_tnl_vxlan_gpe(struct mlx5dr_ste_ctx *ste_ctx,
455				    struct mlx5dr_ste_build *sb,
456				    struct mlx5dr_match_param *mask,
457				    bool inner, bool rx);
458void mlx5dr_ste_build_tnl_geneve(struct mlx5dr_ste_ctx *ste_ctx,
459				 struct mlx5dr_ste_build *sb,
460				 struct mlx5dr_match_param *mask,
461				 bool inner, bool rx);
462void mlx5dr_ste_build_tnl_geneve_tlv_opt(struct mlx5dr_ste_ctx *ste_ctx,
463					 struct mlx5dr_ste_build *sb,
464					 struct mlx5dr_match_param *mask,
465					 struct mlx5dr_cmd_caps *caps,
466					 bool inner, bool rx);
467void mlx5dr_ste_build_tnl_geneve_tlv_opt_exist(struct mlx5dr_ste_ctx *ste_ctx,
468					       struct mlx5dr_ste_build *sb,
469					       struct mlx5dr_match_param *mask,
470					       struct mlx5dr_cmd_caps *caps,
471					       bool inner, bool rx);
472void mlx5dr_ste_build_tnl_gtpu(struct mlx5dr_ste_ctx *ste_ctx,
473			       struct mlx5dr_ste_build *sb,
474			       struct mlx5dr_match_param *mask,
475			       bool inner, bool rx);
476void mlx5dr_ste_build_tnl_gtpu_flex_parser_0(struct mlx5dr_ste_ctx *ste_ctx,
477					     struct mlx5dr_ste_build *sb,
478					     struct mlx5dr_match_param *mask,
479					     struct mlx5dr_cmd_caps *caps,
480					     bool inner, bool rx);
481void mlx5dr_ste_build_tnl_gtpu_flex_parser_1(struct mlx5dr_ste_ctx *ste_ctx,
482					     struct mlx5dr_ste_build *sb,
483					     struct mlx5dr_match_param *mask,
484					     struct mlx5dr_cmd_caps *caps,
485					     bool inner, bool rx);
486void mlx5dr_ste_build_tnl_header_0_1(struct mlx5dr_ste_ctx *ste_ctx,
487				     struct mlx5dr_ste_build *sb,
488				     struct mlx5dr_match_param *mask,
489				     bool inner, bool rx);
490void mlx5dr_ste_build_general_purpose(struct mlx5dr_ste_ctx *ste_ctx,
491				      struct mlx5dr_ste_build *sb,
492				      struct mlx5dr_match_param *mask,
493				      bool inner, bool rx);
494void mlx5dr_ste_build_register_0(struct mlx5dr_ste_ctx *ste_ctx,
495				 struct mlx5dr_ste_build *sb,
496				 struct mlx5dr_match_param *mask,
497				 bool inner, bool rx);
498void mlx5dr_ste_build_register_1(struct mlx5dr_ste_ctx *ste_ctx,
499				 struct mlx5dr_ste_build *sb,
500				 struct mlx5dr_match_param *mask,
501				 bool inner, bool rx);
502void mlx5dr_ste_build_src_gvmi_qpn(struct mlx5dr_ste_ctx *ste_ctx,
503				   struct mlx5dr_ste_build *sb,
504				   struct mlx5dr_match_param *mask,
505				   struct mlx5dr_domain *dmn,
506				   bool inner, bool rx);
507void mlx5dr_ste_build_flex_parser_0(struct mlx5dr_ste_ctx *ste_ctx,
508				    struct mlx5dr_ste_build *sb,
509				    struct mlx5dr_match_param *mask,
510				    bool inner, bool rx);
511void mlx5dr_ste_build_flex_parser_1(struct mlx5dr_ste_ctx *ste_ctx,
512				    struct mlx5dr_ste_build *sb,
513				    struct mlx5dr_match_param *mask,
514				    bool inner, bool rx);
515void mlx5dr_ste_build_empty_always_hit(struct mlx5dr_ste_build *sb, bool rx);
516
517/* Actions utils */
518int mlx5dr_actions_build_ste_arr(struct mlx5dr_matcher *matcher,
519				 struct mlx5dr_matcher_rx_tx *nic_matcher,
520				 struct mlx5dr_action *actions[],
521				 u32 num_actions,
522				 u8 *ste_arr,
523				 u32 *new_hw_ste_arr_sz);
524
525struct mlx5dr_match_spec {
526	u32 smac_47_16;		/* Source MAC address of incoming packet */
527	/* Incoming packet Ethertype - this is the Ethertype
528	 * following the last VLAN tag of the packet
529	 */
530	u32 smac_15_0:16;	/* Source MAC address of incoming packet */
531	u32 ethertype:16;
532
533	u32 dmac_47_16;		/* Destination MAC address of incoming packet */
534
535	u32 dmac_15_0:16;	/* Destination MAC address of incoming packet */
536	/* Priority of first VLAN tag in the incoming packet.
537	 * Valid only when cvlan_tag==1 or svlan_tag==1
538	 */
539	u32 first_prio:3;
540	/* CFI bit of first VLAN tag in the incoming packet.
541	 * Valid only when cvlan_tag==1 or svlan_tag==1
542	 */
543	u32 first_cfi:1;
544	/* VLAN ID of first VLAN tag in the incoming packet.
545	 * Valid only when cvlan_tag==1 or svlan_tag==1
546	 */
547	u32 first_vid:12;
548
549	u32 ip_protocol:8;	/* IP protocol */
550	/* Differentiated Services Code Point derived from
551	 * Traffic Class/TOS field of IPv6/v4
552	 */
553	u32 ip_dscp:6;
554	/* Explicit Congestion Notification derived from
555	 * Traffic Class/TOS field of IPv6/v4
556	 */
557	u32 ip_ecn:2;
558	/* The first vlan in the packet is c-vlan (0x8100).
559	 * cvlan_tag and svlan_tag cannot be set together
560	 */
561	u32 cvlan_tag:1;
562	/* The first vlan in the packet is s-vlan (0x8a88).
563	 * cvlan_tag and svlan_tag cannot be set together
564	 */
565	u32 svlan_tag:1;
566	u32 frag:1;		/* Packet is an IP fragment */
567	u32 ip_version:4;	/* IP version */
568	/* TCP flags. ;Bit 0: FIN;Bit 1: SYN;Bit 2: RST;Bit 3: PSH;Bit 4: ACK;
569	 *             Bit 5: URG;Bit 6: ECE;Bit 7: CWR;Bit 8: NS
570	 */
571	u32 tcp_flags:9;
572
573	/* TCP source port.;tcp and udp sport/dport are mutually exclusive */
574	u32 tcp_sport:16;
575	/* TCP destination port.
576	 * tcp and udp sport/dport are mutually exclusive
577	 */
578	u32 tcp_dport:16;
579
580	u32 reserved_auto1:16;
581	u32 ipv4_ihl:4;
582	u32 reserved_auto2:4;
583	u32 ttl_hoplimit:8;
584
585	/* UDP source port.;tcp and udp sport/dport are mutually exclusive */
586	u32 udp_sport:16;
587	/* UDP destination port.;tcp and udp sport/dport are mutually exclusive */
588	u32 udp_dport:16;
589
590	/* IPv6 source address of incoming packets
591	 * For IPv4 address use bits 31:0 (rest of the bits are reserved)
592	 * This field should be qualified by an appropriate ethertype
593	 */
594	u32 src_ip_127_96;
595	/* IPv6 source address of incoming packets
596	 * For IPv4 address use bits 31:0 (rest of the bits are reserved)
597	 * This field should be qualified by an appropriate ethertype
598	 */
599	u32 src_ip_95_64;
600	/* IPv6 source address of incoming packets
601	 * For IPv4 address use bits 31:0 (rest of the bits are reserved)
602	 * This field should be qualified by an appropriate ethertype
603	 */
604	u32 src_ip_63_32;
605	/* IPv6 source address of incoming packets
606	 * For IPv4 address use bits 31:0 (rest of the bits are reserved)
607	 * This field should be qualified by an appropriate ethertype
608	 */
609	u32 src_ip_31_0;
610	/* IPv6 destination address of incoming packets
611	 * For IPv4 address use bits 31:0 (rest of the bits are reserved)
612	 * This field should be qualified by an appropriate ethertype
613	 */
614	u32 dst_ip_127_96;
615	/* IPv6 destination address of incoming packets
616	 * For IPv4 address use bits 31:0 (rest of the bits are reserved)
617	 * This field should be qualified by an appropriate ethertype
618	 */
619	u32 dst_ip_95_64;
620	/* IPv6 destination address of incoming packets
621	 * For IPv4 address use bits 31:0 (rest of the bits are reserved)
622	 * This field should be qualified by an appropriate ethertype
623	 */
624	u32 dst_ip_63_32;
625	/* IPv6 destination address of incoming packets
626	 * For IPv4 address use bits 31:0 (rest of the bits are reserved)
627	 * This field should be qualified by an appropriate ethertype
628	 */
629	u32 dst_ip_31_0;
630};
631
632struct mlx5dr_match_misc {
633	/* used with GRE, checksum exist when gre_c_present == 1 */
634	u32 gre_c_present:1;
635	u32 reserved_auto1:1;
636	/* used with GRE, key exist when gre_k_present == 1 */
637	u32 gre_k_present:1;
638	/* used with GRE, sequence number exist when gre_s_present == 1 */
639	u32 gre_s_present:1;
640	u32 source_vhca_port:4;
641	u32 source_sqn:24;		/* Source SQN */
642
643	u32 source_eswitch_owner_vhca_id:16;
644	/* Source port.;0xffff determines wire port */
645	u32 source_port:16;
646
647	/* Priority of second VLAN tag in the outer header of the incoming packet.
648	 * Valid only when outer_second_cvlan_tag ==1 or outer_second_svlan_tag ==1
649	 */
650	u32 outer_second_prio:3;
651	/* CFI bit of first VLAN tag in the outer header of the incoming packet.
652	 * Valid only when outer_second_cvlan_tag ==1 or outer_second_svlan_tag ==1
653	 */
654	u32 outer_second_cfi:1;
655	/* VLAN ID of first VLAN tag the outer header of the incoming packet.
656	 * Valid only when outer_second_cvlan_tag ==1 or outer_second_svlan_tag ==1
657	 */
658	u32 outer_second_vid:12;
659	/* Priority of second VLAN tag in the inner header of the incoming packet.
660	 * Valid only when inner_second_cvlan_tag ==1 or inner_second_svlan_tag ==1
661	 */
662	u32 inner_second_prio:3;
663	/* CFI bit of first VLAN tag in the inner header of the incoming packet.
664	 * Valid only when inner_second_cvlan_tag ==1 or inner_second_svlan_tag ==1
665	 */
666	u32 inner_second_cfi:1;
667	/* VLAN ID of first VLAN tag the inner header of the incoming packet.
668	 * Valid only when inner_second_cvlan_tag ==1 or inner_second_svlan_tag ==1
669	 */
670	u32 inner_second_vid:12;
671
672	u32 outer_second_cvlan_tag:1;
673	u32 inner_second_cvlan_tag:1;
674	/* The second vlan in the outer header of the packet is c-vlan (0x8100).
675	 * outer_second_cvlan_tag and outer_second_svlan_tag cannot be set together
676	 */
677	u32 outer_second_svlan_tag:1;
678	/* The second vlan in the inner header of the packet is c-vlan (0x8100).
679	 * inner_second_cvlan_tag and inner_second_svlan_tag cannot be set together
680	 */
681	u32 inner_second_svlan_tag:1;
682	/* The second vlan in the outer header of the packet is s-vlan (0x8a88).
683	 * outer_second_cvlan_tag and outer_second_svlan_tag cannot be set together
684	 */
685	u32 reserved_auto2:12;
686	/* The second vlan in the inner header of the packet is s-vlan (0x8a88).
687	 * inner_second_cvlan_tag and inner_second_svlan_tag cannot be set together
688	 */
689	u32 gre_protocol:16;		/* GRE Protocol (outer) */
690
691	u32 gre_key_h:24;		/* GRE Key[31:8] (outer) */
692	u32 gre_key_l:8;		/* GRE Key [7:0] (outer) */
693
694	u32 vxlan_vni:24;		/* VXLAN VNI (outer) */
695	u32 reserved_auto3:8;
696
697	u32 geneve_vni:24;		/* GENEVE VNI field (outer) */
698	u32 reserved_auto4:6;
699	u32 geneve_tlv_option_0_exist:1;
700	u32 geneve_oam:1;		/* GENEVE OAM field (outer) */
701
702	u32 reserved_auto5:12;
703	u32 outer_ipv6_flow_label:20;	/* Flow label of incoming IPv6 packet (outer) */
704
705	u32 reserved_auto6:12;
706	u32 inner_ipv6_flow_label:20;	/* Flow label of incoming IPv6 packet (inner) */
707
708	u32 reserved_auto7:10;
709	u32 geneve_opt_len:6;		/* GENEVE OptLen (outer) */
710	u32 geneve_protocol_type:16;	/* GENEVE protocol type (outer) */
711
712	u32 reserved_auto8:8;
713	u32 bth_dst_qp:24;		/* Destination QP in BTH header */
714
715	u32 reserved_auto9;
716	u32 outer_esp_spi;
717	u32 reserved_auto10[3];
718};
719
720struct mlx5dr_match_misc2 {
721	u32 outer_first_mpls_label:20;		/* First MPLS LABEL (outer) */
722	u32 outer_first_mpls_exp:3;		/* First MPLS EXP (outer) */
723	u32 outer_first_mpls_s_bos:1;		/* First MPLS S_BOS (outer) */
724	u32 outer_first_mpls_ttl:8;		/* First MPLS TTL (outer) */
725
726	u32 inner_first_mpls_label:20;		/* First MPLS LABEL (inner) */
727	u32 inner_first_mpls_exp:3;		/* First MPLS EXP (inner) */
728	u32 inner_first_mpls_s_bos:1;		/* First MPLS S_BOS (inner) */
729	u32 inner_first_mpls_ttl:8;		/* First MPLS TTL (inner) */
730
731	u32 outer_first_mpls_over_gre_label:20;	/* last MPLS LABEL (outer) */
732	u32 outer_first_mpls_over_gre_exp:3;	/* last MPLS EXP (outer) */
733	u32 outer_first_mpls_over_gre_s_bos:1;	/* last MPLS S_BOS (outer) */
734	u32 outer_first_mpls_over_gre_ttl:8;	/* last MPLS TTL (outer) */
735
736	u32 outer_first_mpls_over_udp_label:20;	/* last MPLS LABEL (outer) */
737	u32 outer_first_mpls_over_udp_exp:3;	/* last MPLS EXP (outer) */
738	u32 outer_first_mpls_over_udp_s_bos:1;	/* last MPLS S_BOS (outer) */
739	u32 outer_first_mpls_over_udp_ttl:8;	/* last MPLS TTL (outer) */
740
741	u32 metadata_reg_c_7;			/* metadata_reg_c_7 */
742	u32 metadata_reg_c_6;			/* metadata_reg_c_6 */
743	u32 metadata_reg_c_5;			/* metadata_reg_c_5 */
744	u32 metadata_reg_c_4;			/* metadata_reg_c_4 */
745	u32 metadata_reg_c_3;			/* metadata_reg_c_3 */
746	u32 metadata_reg_c_2;			/* metadata_reg_c_2 */
747	u32 metadata_reg_c_1;			/* metadata_reg_c_1 */
748	u32 metadata_reg_c_0;			/* metadata_reg_c_0 */
749	u32 metadata_reg_a;			/* metadata_reg_a */
750	u32 reserved_auto1[3];
751};
752
753struct mlx5dr_match_misc3 {
754	u32 inner_tcp_seq_num;
755	u32 outer_tcp_seq_num;
756	u32 inner_tcp_ack_num;
757	u32 outer_tcp_ack_num;
758
759	u32 reserved_auto1:8;
760	u32 outer_vxlan_gpe_vni:24;
761
762	u32 outer_vxlan_gpe_next_protocol:8;
763	u32 outer_vxlan_gpe_flags:8;
764	u32 reserved_auto2:16;
765
766	u32 icmpv4_header_data;
767	u32 icmpv6_header_data;
768
769	u8 icmpv4_type;
770	u8 icmpv4_code;
771	u8 icmpv6_type;
772	u8 icmpv6_code;
773
774	u32 geneve_tlv_option_0_data;
775
776	u32 gtpu_teid;
777
778	u8 gtpu_msg_type;
779	u8 gtpu_msg_flags;
780	u32 reserved_auto3:16;
781
782	u32 gtpu_dw_2;
783	u32 gtpu_first_ext_dw_0;
784	u32 gtpu_dw_0;
785	u32 reserved_auto4;
786};
787
788struct mlx5dr_match_misc4 {
789	u32 prog_sample_field_value_0;
790	u32 prog_sample_field_id_0;
791	u32 prog_sample_field_value_1;
792	u32 prog_sample_field_id_1;
793	u32 prog_sample_field_value_2;
794	u32 prog_sample_field_id_2;
795	u32 prog_sample_field_value_3;
796	u32 prog_sample_field_id_3;
797	u32 reserved_auto1[8];
798};
799
800struct mlx5dr_match_misc5 {
801	u32 macsec_tag_0;
802	u32 macsec_tag_1;
803	u32 macsec_tag_2;
804	u32 macsec_tag_3;
805	u32 tunnel_header_0;
806	u32 tunnel_header_1;
807	u32 tunnel_header_2;
808	u32 tunnel_header_3;
809};
810
811struct mlx5dr_match_param {
812	struct mlx5dr_match_spec outer;
813	struct mlx5dr_match_misc misc;
814	struct mlx5dr_match_spec inner;
815	struct mlx5dr_match_misc2 misc2;
816	struct mlx5dr_match_misc3 misc3;
817	struct mlx5dr_match_misc4 misc4;
818	struct mlx5dr_match_misc5 misc5;
819};
820
821#define DR_MASK_IS_ICMPV4_SET(_misc3) ((_misc3)->icmpv4_type || \
822				       (_misc3)->icmpv4_code || \
823				       (_misc3)->icmpv4_header_data)
824
825#define DR_MASK_IS_SRC_IP_SET(_spec) ((_spec)->src_ip_127_96 || \
826				      (_spec)->src_ip_95_64  || \
827				      (_spec)->src_ip_63_32  || \
828				      (_spec)->src_ip_31_0)
829
830#define DR_MASK_IS_DST_IP_SET(_spec) ((_spec)->dst_ip_127_96 || \
831				      (_spec)->dst_ip_95_64  || \
832				      (_spec)->dst_ip_63_32  || \
833				      (_spec)->dst_ip_31_0)
834
835struct mlx5dr_esw_caps {
836	u64 drop_icm_address_rx;
837	u64 drop_icm_address_tx;
838	u64 uplink_icm_address_rx;
839	u64 uplink_icm_address_tx;
840	u8 sw_owner:1;
841	u8 sw_owner_v2:1;
842};
843
844struct mlx5dr_cmd_vport_cap {
845	u16 vport_gvmi;
846	u16 vhca_gvmi;
847	u16 num;
848	u64 icm_address_rx;
849	u64 icm_address_tx;
850};
851
852struct mlx5dr_roce_cap {
853	u8 roce_en:1;
854	u8 fl_rc_qp_when_roce_disabled:1;
855	u8 fl_rc_qp_when_roce_enabled:1;
856};
857
858struct mlx5dr_vports {
859	struct mlx5dr_cmd_vport_cap esw_manager_caps;
860	struct mlx5dr_cmd_vport_cap uplink_caps;
861	struct xarray vports_caps_xa;
862};
863
864struct mlx5dr_cmd_caps {
865	u16 gvmi;
866	u64 nic_rx_drop_address;
867	u64 nic_tx_drop_address;
868	u64 nic_tx_allow_address;
869	u64 esw_rx_drop_address;
870	u64 esw_tx_drop_address;
871	u32 log_icm_size;
872	u64 hdr_modify_icm_addr;
873	u32 log_modify_pattern_icm_size;
874	u64 hdr_modify_pattern_icm_addr;
875	u32 flex_protocols;
876	u8 flex_parser_id_icmp_dw0;
877	u8 flex_parser_id_icmp_dw1;
878	u8 flex_parser_id_icmpv6_dw0;
879	u8 flex_parser_id_icmpv6_dw1;
880	u8 flex_parser_id_geneve_tlv_option_0;
881	u8 flex_parser_id_mpls_over_gre;
882	u8 flex_parser_id_mpls_over_udp;
883	u8 flex_parser_id_gtpu_dw_0;
884	u8 flex_parser_id_gtpu_teid;
885	u8 flex_parser_id_gtpu_dw_2;
886	u8 flex_parser_id_gtpu_first_ext_dw_0;
887	u8 flex_parser_ok_bits_supp;
888	u8 max_ft_level;
889	u16 roce_min_src_udp;
890	u8 sw_format_ver;
891	bool eswitch_manager;
892	bool rx_sw_owner;
893	bool tx_sw_owner;
894	bool fdb_sw_owner;
895	u8 rx_sw_owner_v2:1;
896	u8 tx_sw_owner_v2:1;
897	u8 fdb_sw_owner_v2:1;
898	struct mlx5dr_esw_caps esw_caps;
899	struct mlx5dr_vports vports;
900	bool prio_tag_required;
901	struct mlx5dr_roce_cap roce_caps;
902	u16 log_header_modify_argument_granularity;
903	u16 log_header_modify_argument_max_alloc;
904	bool support_modify_argument;
905	u8 is_ecpf:1;
906	u8 isolate_vl_tc:1;
907};
908
909enum mlx5dr_domain_nic_type {
910	DR_DOMAIN_NIC_TYPE_RX,
911	DR_DOMAIN_NIC_TYPE_TX,
912};
913
914struct mlx5dr_domain_rx_tx {
915	u64 drop_icm_addr;
916	u64 default_icm_addr;
917	enum mlx5dr_domain_nic_type type;
918	struct mutex mutex; /* protect rx/tx domain */
919};
920
921struct mlx5dr_domain_info {
922	bool supp_sw_steering;
923	u32 max_inline_size;
924	u32 max_send_wr;
925	u32 max_log_sw_icm_sz;
926	u32 max_log_action_icm_sz;
927	u32 max_log_modify_hdr_pattern_icm_sz;
928	struct mlx5dr_domain_rx_tx rx;
929	struct mlx5dr_domain_rx_tx tx;
930	struct mlx5dr_cmd_caps caps;
931};
932
933struct mlx5dr_domain {
934	struct mlx5_core_dev *mdev;
935	u32 pdn;
936	struct mlx5_uars_page *uar;
937	enum mlx5dr_domain_type type;
938	refcount_t refcount;
939	struct mlx5dr_icm_pool *ste_icm_pool;
940	struct mlx5dr_icm_pool *action_icm_pool;
941	struct mlx5dr_send_info_pool *send_info_pool_rx;
942	struct mlx5dr_send_info_pool *send_info_pool_tx;
943	struct kmem_cache *chunks_kmem_cache;
944	struct kmem_cache *htbls_kmem_cache;
945	struct mlx5dr_ptrn_mgr *ptrn_mgr;
946	struct mlx5dr_arg_mgr *arg_mgr;
947	struct mlx5dr_send_ring *send_ring;
948	struct mlx5dr_domain_info info;
949	struct xarray csum_fts_xa;
950	struct mlx5dr_ste_ctx *ste_ctx;
951	struct list_head dbg_tbl_list;
952	struct mlx5dr_dbg_dump_info dump_info;
953	struct xarray definers_xa;
954	struct xarray peer_dmn_xa;
955	/* memory management statistics */
956	u32 num_buddies[DR_ICM_TYPE_MAX];
957};
958
959struct mlx5dr_table_rx_tx {
960	struct mlx5dr_ste_htbl *s_anchor;
961	struct mlx5dr_domain_rx_tx *nic_dmn;
962	u64 default_icm_addr;
963	struct list_head nic_matcher_list;
964};
965
966struct mlx5dr_table {
967	struct mlx5dr_domain *dmn;
968	struct mlx5dr_table_rx_tx rx;
969	struct mlx5dr_table_rx_tx tx;
970	u32 level;
971	u32 table_type;
972	u32 table_id;
973	u32 flags;
974	struct list_head matcher_list;
975	struct mlx5dr_action *miss_action;
976	refcount_t refcount;
977	struct list_head dbg_node;
978};
979
980struct mlx5dr_matcher_rx_tx {
981	struct mlx5dr_ste_htbl *s_htbl;
982	struct mlx5dr_ste_htbl *e_anchor;
983	struct mlx5dr_ste_build *ste_builder;
984	struct mlx5dr_ste_build ste_builder_arr[DR_RULE_IPV_MAX]
985					       [DR_RULE_IPV_MAX]
986					       [DR_RULE_MAX_STES];
987	u8 num_of_builders;
988	u8 num_of_builders_arr[DR_RULE_IPV_MAX][DR_RULE_IPV_MAX];
989	u64 default_icm_addr;
990	struct mlx5dr_table_rx_tx *nic_tbl;
991	u32 prio;
992	struct list_head list_node;
993	u32 rules;
994};
995
996struct mlx5dr_matcher {
997	struct mlx5dr_table *tbl;
998	struct mlx5dr_matcher_rx_tx rx;
999	struct mlx5dr_matcher_rx_tx tx;
1000	struct list_head list_node; /* Used for both matchers and dbg managing */
1001	u32 prio;
1002	struct mlx5dr_match_param mask;
1003	u8 match_criteria;
1004	refcount_t refcount;
1005	struct list_head dbg_rule_list;
1006};
1007
1008struct mlx5dr_ste_action_modify_field {
1009	u16 hw_field;
1010	u8 start;
1011	u8 end;
1012	u8 l3_type;
1013	u8 l4_type;
1014};
1015
1016struct mlx5dr_ptrn_obj {
1017	struct mlx5dr_icm_chunk *chunk;
1018	u8 *data;
1019	u16 num_of_actions;
1020	u32 index;
1021	refcount_t refcount;
1022	struct list_head list;
1023};
1024
1025struct mlx5dr_arg_obj {
1026	u32 obj_id;
1027	u32 obj_offset;
1028	struct list_head list_node;
1029	u32 log_chunk_size;
1030};
1031
1032struct mlx5dr_action_rewrite {
1033	struct mlx5dr_domain *dmn;
1034	struct mlx5dr_icm_chunk *chunk;
1035	u8 *data;
1036	u16 num_of_actions;
1037	u32 index;
1038	u8 single_action_opt:1;
1039	u8 allow_rx:1;
1040	u8 allow_tx:1;
1041	u8 modify_ttl:1;
1042	struct mlx5dr_ptrn_obj *ptrn;
1043	struct mlx5dr_arg_obj *arg;
1044};
1045
1046struct mlx5dr_action_reformat {
1047	struct mlx5dr_domain *dmn;
1048	u32 id;
1049	u32 size;
1050	u8 param_0;
1051	u8 param_1;
1052};
1053
1054struct mlx5dr_action_sampler {
1055	struct mlx5dr_domain *dmn;
1056	u64 rx_icm_addr;
1057	u64 tx_icm_addr;
1058	u32 sampler_id;
1059};
1060
1061struct mlx5dr_action_dest_tbl {
1062	u8 is_fw_tbl:1;
1063	u8 is_wire_ft:1;
1064	union {
1065		struct mlx5dr_table *tbl;
1066		struct {
1067			struct mlx5dr_domain *dmn;
1068			u32 id;
1069			u32 group_id;
1070			enum fs_flow_table_type type;
1071			u64 rx_icm_addr;
1072			u64 tx_icm_addr;
1073			struct mlx5dr_action **ref_actions;
1074			u32 num_of_ref_actions;
1075		} fw_tbl;
1076	};
1077};
1078
1079struct mlx5dr_action_range {
1080	struct mlx5dr_domain *dmn;
1081	struct mlx5dr_action *hit_tbl_action;
1082	struct mlx5dr_action *miss_tbl_action;
1083	u32 definer_id;
1084	u32 min;
1085	u32 max;
1086};
1087
1088struct mlx5dr_action_ctr {
1089	u32 ctr_id;
1090	u32 offset;
1091};
1092
1093struct mlx5dr_action_vport {
1094	struct mlx5dr_domain *dmn;
1095	struct mlx5dr_cmd_vport_cap *caps;
1096};
1097
1098struct mlx5dr_action_push_vlan {
1099	u32 vlan_hdr; /* tpid_pcp_dei_vid */
1100};
1101
1102struct mlx5dr_action_flow_tag {
1103	u32 flow_tag;
1104};
1105
1106struct mlx5dr_rule_action_member {
1107	struct mlx5dr_action *action;
1108	struct list_head list;
1109};
1110
1111struct mlx5dr_action_aso_flow_meter {
1112	struct mlx5dr_domain *dmn;
1113	u32 obj_id;
1114	u32 offset;
1115	u8 dest_reg_id;
1116	u8 init_color;
1117};
1118
1119struct mlx5dr_action {
1120	enum mlx5dr_action_type action_type;
1121	refcount_t refcount;
1122
1123	union {
1124		void *data;
1125		struct mlx5dr_action_rewrite *rewrite;
1126		struct mlx5dr_action_reformat *reformat;
1127		struct mlx5dr_action_sampler *sampler;
1128		struct mlx5dr_action_dest_tbl *dest_tbl;
1129		struct mlx5dr_action_ctr *ctr;
1130		struct mlx5dr_action_vport *vport;
1131		struct mlx5dr_action_push_vlan *push_vlan;
1132		struct mlx5dr_action_flow_tag *flow_tag;
1133		struct mlx5dr_action_aso_flow_meter *aso;
1134		struct mlx5dr_action_range *range;
1135	};
1136};
1137
1138enum mlx5dr_connect_type {
1139	CONNECT_HIT	= 1,
1140	CONNECT_MISS	= 2,
1141};
1142
1143struct mlx5dr_htbl_connect_info {
1144	enum mlx5dr_connect_type type;
1145	union {
1146		struct mlx5dr_ste_htbl *hit_next_htbl;
1147		u64 miss_icm_addr;
1148	};
1149};
1150
1151struct mlx5dr_rule_rx_tx {
1152	struct mlx5dr_matcher_rx_tx *nic_matcher;
1153	struct mlx5dr_ste *last_rule_ste;
1154};
1155
1156struct mlx5dr_rule {
1157	struct mlx5dr_matcher *matcher;
1158	struct mlx5dr_rule_rx_tx rx;
1159	struct mlx5dr_rule_rx_tx tx;
1160	struct list_head rule_actions_list;
1161	struct list_head dbg_node;
1162	u32 flow_source;
1163};
1164
1165void mlx5dr_rule_set_last_member(struct mlx5dr_rule_rx_tx *nic_rule,
1166				 struct mlx5dr_ste *ste,
1167				 bool force);
1168int mlx5dr_rule_get_reverse_rule_members(struct mlx5dr_ste **ste_arr,
1169					 struct mlx5dr_ste *curr_ste,
1170					 int *num_of_stes);
1171
1172struct mlx5dr_icm_chunk {
1173	struct mlx5dr_icm_buddy_mem *buddy_mem;
1174
1175	/* indicates the index of this chunk in the whole memory,
1176	 * used for deleting the chunk from the buddy
1177	 */
1178	unsigned int seg;
1179	enum mlx5dr_icm_chunk_size size;
1180
1181	/* Memory optimisation */
1182	struct mlx5dr_ste *ste_arr;
1183	u8 *hw_ste_arr;
1184	struct list_head *miss_list;
1185};
1186
1187static inline void mlx5dr_domain_nic_lock(struct mlx5dr_domain_rx_tx *nic_dmn)
1188{
1189	mutex_lock(&nic_dmn->mutex);
1190}
1191
1192static inline void mlx5dr_domain_nic_unlock(struct mlx5dr_domain_rx_tx *nic_dmn)
1193{
1194	mutex_unlock(&nic_dmn->mutex);
1195}
1196
1197static inline void mlx5dr_domain_lock(struct mlx5dr_domain *dmn)
1198{
1199	mlx5dr_domain_nic_lock(&dmn->info.rx);
1200	mlx5dr_domain_nic_lock(&dmn->info.tx);
1201}
1202
1203static inline void mlx5dr_domain_unlock(struct mlx5dr_domain *dmn)
1204{
1205	mlx5dr_domain_nic_unlock(&dmn->info.tx);
1206	mlx5dr_domain_nic_unlock(&dmn->info.rx);
1207}
1208
1209int mlx5dr_matcher_add_to_tbl_nic(struct mlx5dr_domain *dmn,
1210				  struct mlx5dr_matcher_rx_tx *nic_matcher);
1211int mlx5dr_matcher_remove_from_tbl_nic(struct mlx5dr_domain *dmn,
1212				       struct mlx5dr_matcher_rx_tx *nic_matcher);
1213
1214int mlx5dr_matcher_select_builders(struct mlx5dr_matcher *matcher,
1215				   struct mlx5dr_matcher_rx_tx *nic_matcher,
1216				   enum mlx5dr_ipv outer_ipv,
1217				   enum mlx5dr_ipv inner_ipv);
1218
1219u64 mlx5dr_icm_pool_get_chunk_mr_addr(struct mlx5dr_icm_chunk *chunk);
1220u32 mlx5dr_icm_pool_get_chunk_rkey(struct mlx5dr_icm_chunk *chunk);
1221u64 mlx5dr_icm_pool_get_chunk_icm_addr(struct mlx5dr_icm_chunk *chunk);
1222u32 mlx5dr_icm_pool_get_chunk_num_of_entries(struct mlx5dr_icm_chunk *chunk);
1223u32 mlx5dr_icm_pool_get_chunk_byte_size(struct mlx5dr_icm_chunk *chunk);
1224u8 *mlx5dr_ste_get_hw_ste(struct mlx5dr_ste *ste);
1225
1226struct mlx5dr_ste_htbl *mlx5dr_icm_pool_alloc_htbl(struct mlx5dr_icm_pool *pool);
1227void mlx5dr_icm_pool_free_htbl(struct mlx5dr_icm_pool *pool, struct mlx5dr_ste_htbl *htbl);
1228
1229static inline int
1230mlx5dr_icm_pool_dm_type_to_entry_size(enum mlx5dr_icm_type icm_type)
1231{
1232	if (icm_type == DR_ICM_TYPE_STE)
1233		return DR_STE_SIZE;
1234
1235	return DR_MODIFY_ACTION_SIZE;
1236}
1237
1238static inline u32
1239mlx5dr_icm_pool_chunk_size_to_entries(enum mlx5dr_icm_chunk_size chunk_size)
1240{
1241	return 1 << chunk_size;
1242}
1243
1244static inline int
1245mlx5dr_icm_pool_chunk_size_to_byte(enum mlx5dr_icm_chunk_size chunk_size,
1246				   enum mlx5dr_icm_type icm_type)
1247{
1248	int num_of_entries;
1249	int entry_size;
1250
1251	entry_size = mlx5dr_icm_pool_dm_type_to_entry_size(icm_type);
1252	num_of_entries = mlx5dr_icm_pool_chunk_size_to_entries(chunk_size);
1253
1254	return entry_size * num_of_entries;
1255}
1256
1257static inline int
1258mlx5dr_ste_htbl_increase_threshold(struct mlx5dr_ste_htbl *htbl)
1259{
1260	int num_of_entries =
1261		mlx5dr_icm_pool_chunk_size_to_entries(htbl->chunk->size);
1262
1263	/* Threshold is 50%, one is added to table of size 1 */
1264	return (num_of_entries + 1) / 2;
1265}
1266
1267static inline bool
1268mlx5dr_ste_htbl_may_grow(struct mlx5dr_ste_htbl *htbl)
1269{
1270	if (htbl->chunk->size == DR_CHUNK_SIZE_MAX - 1 || !htbl->byte_mask)
1271		return false;
1272
1273	return true;
1274}
1275
1276struct mlx5dr_cmd_vport_cap *
1277mlx5dr_domain_get_vport_cap(struct mlx5dr_domain *dmn, u16 vport);
1278
1279struct mlx5dr_cmd_query_flow_table_details {
1280	u8 status;
1281	u8 level;
1282	u64 sw_owner_icm_root_1;
1283	u64 sw_owner_icm_root_0;
1284};
1285
1286struct mlx5dr_cmd_create_flow_table_attr {
1287	u32 table_type;
1288	u16 uid;
1289	u64 icm_addr_rx;
1290	u64 icm_addr_tx;
1291	u8 level;
1292	bool sw_owner;
1293	bool term_tbl;
1294	bool decap_en;
1295	bool reformat_en;
1296};
1297
1298/* internal API functions */
1299int mlx5dr_cmd_query_device(struct mlx5_core_dev *mdev,
1300			    struct mlx5dr_cmd_caps *caps);
1301int mlx5dr_cmd_query_esw_vport_context(struct mlx5_core_dev *mdev,
1302				       bool other_vport, u16 vport_number,
1303				       u64 *icm_address_rx,
1304				       u64 *icm_address_tx);
1305int mlx5dr_cmd_query_gvmi(struct mlx5_core_dev *mdev,
1306			  bool other_vport, u16 vport_number, u16 *gvmi);
1307int mlx5dr_cmd_query_esw_caps(struct mlx5_core_dev *mdev,
1308			      struct mlx5dr_esw_caps *caps);
1309int mlx5dr_cmd_query_flow_sampler(struct mlx5_core_dev *dev,
1310				  u32 sampler_id,
1311				  u64 *rx_icm_addr,
1312				  u64 *tx_icm_addr);
1313int mlx5dr_cmd_sync_steering(struct mlx5_core_dev *mdev);
1314int mlx5dr_cmd_set_fte_modify_and_vport(struct mlx5_core_dev *mdev,
1315					u32 table_type,
1316					u32 table_id,
1317					u32 group_id,
1318					u32 modify_header_id,
1319					u16 vport_id);
1320int mlx5dr_cmd_del_flow_table_entry(struct mlx5_core_dev *mdev,
1321				    u32 table_type,
1322				    u32 table_id);
1323int mlx5dr_cmd_alloc_modify_header(struct mlx5_core_dev *mdev,
1324				   u32 table_type,
1325				   u8 num_of_actions,
1326				   u64 *actions,
1327				   u32 *modify_header_id);
1328int mlx5dr_cmd_dealloc_modify_header(struct mlx5_core_dev *mdev,
1329				     u32 modify_header_id);
1330int mlx5dr_cmd_create_empty_flow_group(struct mlx5_core_dev *mdev,
1331				       u32 table_type,
1332				       u32 table_id,
1333				       u32 *group_id);
1334int mlx5dr_cmd_destroy_flow_group(struct mlx5_core_dev *mdev,
1335				  u32 table_type,
1336				  u32 table_id,
1337				  u32 group_id);
1338int mlx5dr_cmd_create_flow_table(struct mlx5_core_dev *mdev,
1339				 struct mlx5dr_cmd_create_flow_table_attr *attr,
1340				 u64 *fdb_rx_icm_addr,
1341				 u32 *table_id);
1342int mlx5dr_cmd_destroy_flow_table(struct mlx5_core_dev *mdev,
1343				  u32 table_id,
1344				  u32 table_type);
1345int mlx5dr_cmd_query_flow_table(struct mlx5_core_dev *dev,
1346				enum fs_flow_table_type type,
1347				u32 table_id,
1348				struct mlx5dr_cmd_query_flow_table_details *output);
1349int mlx5dr_cmd_create_reformat_ctx(struct mlx5_core_dev *mdev,
1350				   enum mlx5_reformat_ctx_type rt,
1351				   u8 reformat_param_0,
1352				   u8 reformat_param_1,
1353				   size_t reformat_size,
1354				   void *reformat_data,
1355				   u32 *reformat_id);
1356void mlx5dr_cmd_destroy_reformat_ctx(struct mlx5_core_dev *mdev,
1357				     u32 reformat_id);
1358int mlx5dr_cmd_create_definer(struct mlx5_core_dev *mdev,
1359			      u16 format_id,
1360			      u8 *dw_selectors,
1361			      u8 *byte_selectors,
1362			      u8 *match_mask,
1363			      u32 *definer_id);
1364void mlx5dr_cmd_destroy_definer(struct mlx5_core_dev *mdev,
1365				u32 definer_id);
1366
1367struct mlx5dr_cmd_gid_attr {
1368	u8 gid[16];
1369	u8 mac[6];
1370	u32 roce_ver;
1371};
1372
1373int mlx5dr_cmd_query_gid(struct mlx5_core_dev *mdev, u8 vhca_port_num,
1374			 u16 index, struct mlx5dr_cmd_gid_attr *attr);
1375
1376int mlx5dr_cmd_create_modify_header_arg(struct mlx5_core_dev *dev,
1377					u16 log_obj_range, u32 pd,
1378					u32 *obj_id);
1379void mlx5dr_cmd_destroy_modify_header_arg(struct mlx5_core_dev *dev,
1380					  u32 obj_id);
1381
1382struct mlx5dr_icm_pool *mlx5dr_icm_pool_create(struct mlx5dr_domain *dmn,
1383					       enum mlx5dr_icm_type icm_type);
1384void mlx5dr_icm_pool_destroy(struct mlx5dr_icm_pool *pool);
1385
1386struct mlx5dr_icm_chunk *
1387mlx5dr_icm_alloc_chunk(struct mlx5dr_icm_pool *pool,
1388		       enum mlx5dr_icm_chunk_size chunk_size);
1389void mlx5dr_icm_free_chunk(struct mlx5dr_icm_chunk *chunk);
1390
1391void mlx5dr_ste_prepare_for_postsend(struct mlx5dr_ste_ctx *ste_ctx,
1392				     u8 *hw_ste_p, u32 ste_size);
1393int mlx5dr_ste_htbl_init_and_postsend(struct mlx5dr_domain *dmn,
1394				      struct mlx5dr_domain_rx_tx *nic_dmn,
1395				      struct mlx5dr_ste_htbl *htbl,
1396				      struct mlx5dr_htbl_connect_info *connect_info,
1397				      bool update_hw_ste);
1398void mlx5dr_ste_set_formatted_ste(struct mlx5dr_ste_ctx *ste_ctx,
1399				  u16 gvmi,
1400				  enum mlx5dr_domain_nic_type nic_type,
1401				  struct mlx5dr_ste_htbl *htbl,
1402				  u8 *formatted_ste,
1403				  struct mlx5dr_htbl_connect_info *connect_info);
1404void mlx5dr_ste_copy_param(u8 match_criteria,
1405			   struct mlx5dr_match_param *set_param,
1406			   struct mlx5dr_match_parameters *mask,
1407			   bool clear);
1408
1409struct mlx5dr_qp {
1410	struct mlx5_core_dev *mdev;
1411	struct mlx5_wq_qp wq;
1412	struct mlx5_uars_page *uar;
1413	struct mlx5_wq_ctrl wq_ctrl;
1414	u32 qpn;
1415	struct {
1416		unsigned int head;
1417		unsigned int pc;
1418		unsigned int cc;
1419		unsigned int size;
1420		unsigned int *wqe_head;
1421		unsigned int wqe_cnt;
1422	} sq;
1423	struct {
1424		unsigned int pc;
1425		unsigned int cc;
1426		unsigned int size;
1427		unsigned int wqe_cnt;
1428	} rq;
1429	int max_inline_data;
1430};
1431
1432struct mlx5dr_cq {
1433	struct mlx5_core_dev *mdev;
1434	struct mlx5_cqwq wq;
1435	struct mlx5_wq_ctrl wq_ctrl;
1436	struct mlx5_core_cq mcq;
1437	struct mlx5dr_qp *qp;
1438};
1439
1440struct mlx5dr_mr {
1441	struct mlx5_core_dev *mdev;
1442	u32 mkey;
1443	dma_addr_t dma_addr;
1444	void *addr;
1445	size_t size;
1446};
1447
1448struct mlx5dr_send_ring {
1449	struct mlx5dr_cq *cq;
1450	struct mlx5dr_qp *qp;
1451	struct mlx5dr_mr *mr;
1452	/* How much wqes are waiting for completion */
1453	u32 pending_wqe;
1454	/* Signal request per this trash hold value */
1455	u16 signal_th;
1456	/* Each post_send_size less than max_post_send_size */
1457	u32 max_post_send_size;
1458	/* manage the send queue */
1459	u32 tx_head;
1460	void *buf;
1461	u32 buf_size;
1462	u8 *sync_buff;
1463	struct mlx5dr_mr *sync_mr;
1464	spinlock_t lock; /* Protect the data path of the send ring */
1465	bool err_state; /* send_ring is not usable in err state */
1466};
1467
1468int mlx5dr_send_ring_alloc(struct mlx5dr_domain *dmn);
1469void mlx5dr_send_ring_free(struct mlx5dr_domain *dmn,
1470			   struct mlx5dr_send_ring *send_ring);
1471int mlx5dr_send_ring_force_drain(struct mlx5dr_domain *dmn);
1472int mlx5dr_send_postsend_ste(struct mlx5dr_domain *dmn,
1473			     struct mlx5dr_ste *ste,
1474			     u8 *data,
1475			     u16 size,
1476			     u16 offset);
1477int mlx5dr_send_postsend_htbl(struct mlx5dr_domain *dmn,
1478			      struct mlx5dr_ste_htbl *htbl,
1479			      u8 *formatted_ste, u8 *mask);
1480int mlx5dr_send_postsend_formatted_htbl(struct mlx5dr_domain *dmn,
1481					struct mlx5dr_ste_htbl *htbl,
1482					u8 *ste_init_data,
1483					bool update_hw_ste);
1484int mlx5dr_send_postsend_action(struct mlx5dr_domain *dmn,
1485				struct mlx5dr_action *action);
1486int mlx5dr_send_postsend_pattern(struct mlx5dr_domain *dmn,
1487				 struct mlx5dr_icm_chunk *chunk,
1488				 u16 num_of_actions,
1489				 u8 *data);
1490int mlx5dr_send_postsend_args(struct mlx5dr_domain *dmn, u64 arg_id,
1491			      u16 num_of_actions, u8 *actions_data);
1492
1493int mlx5dr_send_info_pool_create(struct mlx5dr_domain *dmn);
1494void mlx5dr_send_info_pool_destroy(struct mlx5dr_domain *dmn);
1495struct mlx5dr_ste_send_info *mlx5dr_send_info_alloc(struct mlx5dr_domain *dmn,
1496						    enum mlx5dr_domain_nic_type nic_type);
1497void mlx5dr_send_info_free(struct mlx5dr_ste_send_info *ste_send_info);
1498
1499struct mlx5dr_cmd_ft_info {
1500	u32 id;
1501	u16 vport;
1502	enum fs_flow_table_type type;
1503};
1504
1505struct mlx5dr_cmd_flow_destination_hw_info {
1506	enum mlx5_flow_destination_type type;
1507	union {
1508		u32 tir_num;
1509		u32 ft_num;
1510		u32 ft_id;
1511		u32 counter_id;
1512		u32 sampler_id;
1513		struct {
1514			u16 num;
1515			u16 vhca_id;
1516			u32 reformat_id;
1517			u8 flags;
1518		} vport;
1519	};
1520};
1521
1522struct mlx5dr_cmd_fte_info {
1523	u32 dests_size;
1524	u32 index;
1525	struct mlx5_flow_context flow_context;
1526	u32 *val;
1527	struct mlx5_flow_act action;
1528	struct mlx5dr_cmd_flow_destination_hw_info *dest_arr;
1529	bool ignore_flow_level;
1530};
1531
1532int mlx5dr_cmd_set_fte(struct mlx5_core_dev *dev,
1533		       int opmod, int modify_mask,
1534		       struct mlx5dr_cmd_ft_info *ft,
1535		       u32 group_id,
1536		       struct mlx5dr_cmd_fte_info *fte);
1537
1538bool mlx5dr_ste_supp_ttl_cs_recalc(struct mlx5dr_cmd_caps *caps);
1539
1540struct mlx5dr_fw_recalc_cs_ft {
1541	u64 rx_icm_addr;
1542	u32 table_id;
1543	u32 group_id;
1544	u32 modify_hdr_id;
1545};
1546
1547struct mlx5dr_fw_recalc_cs_ft *
1548mlx5dr_fw_create_recalc_cs_ft(struct mlx5dr_domain *dmn, u16 vport_num);
1549void mlx5dr_fw_destroy_recalc_cs_ft(struct mlx5dr_domain *dmn,
1550				    struct mlx5dr_fw_recalc_cs_ft *recalc_cs_ft);
1551int mlx5dr_domain_get_recalc_cs_ft_addr(struct mlx5dr_domain *dmn,
1552					u16 vport_num,
1553					u64 *rx_icm_addr);
1554int mlx5dr_fw_create_md_tbl(struct mlx5dr_domain *dmn,
1555			    struct mlx5dr_cmd_flow_destination_hw_info *dest,
1556			    int num_dest,
1557			    bool reformat_req,
1558			    u32 *tbl_id,
1559			    u32 *group_id,
1560			    bool ignore_flow_level,
1561			    u32 flow_source);
1562void mlx5dr_fw_destroy_md_tbl(struct mlx5dr_domain *dmn, u32 tbl_id,
1563			      u32 group_id);
1564
1565static inline bool mlx5dr_is_fw_table(struct mlx5_flow_table *ft)
1566{
1567	return !ft->fs_dr_table.dr_table;
1568}
1569
1570static inline bool mlx5dr_supp_match_ranges(struct mlx5_core_dev *dev)
1571{
1572	return (MLX5_CAP_GEN(dev, steering_format_version) >=
1573		MLX5_STEERING_FORMAT_CONNECTX_6DX) &&
1574	       (MLX5_CAP_GEN_64(dev, match_definer_format_supported) &
1575			(1ULL << MLX5_IFC_DEFINER_FORMAT_ID_SELECT));
1576}
1577
1578bool mlx5dr_domain_is_support_ptrn_arg(struct mlx5dr_domain *dmn);
1579struct mlx5dr_ptrn_mgr *mlx5dr_ptrn_mgr_create(struct mlx5dr_domain *dmn);
1580void mlx5dr_ptrn_mgr_destroy(struct mlx5dr_ptrn_mgr *mgr);
1581struct mlx5dr_ptrn_obj *mlx5dr_ptrn_cache_get_pattern(struct mlx5dr_ptrn_mgr *mgr,
1582						      u16 num_of_actions, u8 *data);
1583void mlx5dr_ptrn_cache_put_pattern(struct mlx5dr_ptrn_mgr *mgr,
1584				   struct mlx5dr_ptrn_obj *pattern);
1585struct mlx5dr_arg_mgr *mlx5dr_arg_mgr_create(struct mlx5dr_domain *dmn);
1586void mlx5dr_arg_mgr_destroy(struct mlx5dr_arg_mgr *mgr);
1587struct mlx5dr_arg_obj *mlx5dr_arg_get_obj(struct mlx5dr_arg_mgr *mgr,
1588					  u16 num_of_actions,
1589					  u8 *data);
1590void mlx5dr_arg_put_obj(struct mlx5dr_arg_mgr *mgr,
1591			struct mlx5dr_arg_obj *arg_obj);
1592u32 mlx5dr_arg_get_obj_id(struct mlx5dr_arg_obj *arg_obj);
1593
1594#endif  /* _DR_TYPES_H_ */
1595