Lines Matching refs:prof
630 ice_find_prot_off(struct ice_hw *hw, enum ice_block blk, u8 prof, u16 fv_idx, in ice_find_prot_off() argument
635 if (prof >= hw->blk[blk].es.count) in ice_find_prot_off()
641 fv_ext = hw->blk[blk].es.t + (prof * hw->blk[blk].es.fvw); in ice_find_prot_off()
1160 ice_prof_has_mask_idx(struct ice_hw *hw, enum ice_block blk, u8 prof, u16 idx, in ice_prof_has_mask_idx() argument
1175 if (hw->blk[blk].es.mask_ena[prof] & BIT(i)) in ice_prof_has_mask_idx()
1203 ice_prof_has_mask(struct ice_hw *hw, enum ice_block blk, u8 prof, u16 *masks) in ice_prof_has_mask() argument
1209 if (!ice_prof_has_mask_idx(hw, blk, prof, i, masks[i])) in ice_prof_has_mask()
1939 sizeof(*hw->blk[block_id].prof.t); in ice_fill_tbl()
1940 dst = (u8 *)hw->blk[block_id].prof.t; in ice_fill_tbl()
1941 dst_len = hw->blk[block_id].prof.count * in ice_fill_tbl()
1942 sizeof(*hw->blk[block_id].prof.t); in ice_fill_tbl()
2013 ice_fill_tbl(hw, blk_id, hw->blk[blk_id].prof.sid); in ice_fill_blk_tbls()
2115 devm_kfree(ice_hw_to_dev(hw), hw->blk[i].prof.t); in ice_free_hw_tbls()
2153 struct ice_prof_tcam *prof = &hw->blk[i].prof; in ice_clear_hw_tbls() local
2175 memset(prof->t, 0, prof->count * sizeof(*prof->t)); in ice_clear_hw_tbls()
2199 struct ice_prof_tcam *prof = &hw->blk[i].prof; in ice_init_hw_tbls() local
2260 prof->sid = ice_blk_sids[i][ICE_SID_PR_OFF]; in ice_init_hw_tbls()
2261 prof->count = blk_sizes[i].prof_tcam; in ice_init_hw_tbls()
2262 prof->max_prof_id = blk_sizes[i].prof_id; in ice_init_hw_tbls()
2263 prof->cdid_bits = blk_sizes[i].prof_cdid_bits; in ice_init_hw_tbls()
2264 prof->t = devm_kcalloc(ice_hw_to_dev(hw), prof->count, in ice_init_hw_tbls()
2265 sizeof(*prof->t), GFP_KERNEL); in ice_init_hw_tbls()
2267 if (!prof->t) in ice_init_hw_tbls()
2337 switch (hw->blk[blk].prof.cdid_bits) { in ice_prof_gen_key()
2392 dc_msk, nm_msk, hw->blk[blk].prof.t[idx].key); in ice_tcam_write_entry()
2394 hw->blk[blk].prof.t[idx].addr = cpu_to_le16(idx); in ice_tcam_write_entry()
2395 hw->blk[blk].prof.t[idx].prof_id = prof_id; in ice_tcam_write_entry()
2519 &hw->blk[blk].prof.t[tmp->tcam_idx].key, in ice_prof_bld_tcam()
2520 sizeof(hw->blk[blk].prof.t->key)); in ice_prof_bld_tcam()
2932 ice_add_prof_attrib(struct ice_prof_map *prof, u8 ptg, u16 ptype, in ice_add_prof_attrib() argument
2942 prof->ptg[prof->ptg_cnt] = ptg; in ice_add_prof_attrib()
2944 &prof->attr[prof->ptg_cnt]); in ice_add_prof_attrib()
2946 if (++prof->ptg_cnt >= ICE_MAX_PTG_PER_PROFILE) in ice_add_prof_attrib()
2979 struct ice_prof_map *prof; in ice_add_prof() local
3018 prof = devm_kzalloc(ice_hw_to_dev(hw), sizeof(*prof), GFP_KERNEL); in ice_add_prof()
3019 if (!prof) { in ice_add_prof()
3024 prof->profile_cookie = id; in ice_add_prof()
3025 prof->prof_id = prof_id; in ice_add_prof()
3026 prof->ptg_cnt = 0; in ice_add_prof()
3027 prof->context = 0; in ice_add_prof()
3030 while (bytes && prof->ptg_cnt < ICE_MAX_PTG_PER_PROFILE) { in ice_add_prof()
3061 status = ice_add_prof_attrib(prof, ptg, ptype, in ice_add_prof()
3069 prof->ptg[prof->ptg_cnt] = ptg; in ice_add_prof()
3070 prof->attr[prof->ptg_cnt].flags = 0; in ice_add_prof()
3071 prof->attr[prof->ptg_cnt].mask = 0; in ice_add_prof()
3073 if (++prof->ptg_cnt >= in ice_add_prof()
3083 list_add(&prof->list, &hw->blk[blk].es.prof_map); in ice_add_prof()
3168 struct ice_vsig_prof *prof) in ice_rem_prof_id() argument
3173 for (i = 0; i < prof->tcam_count; i++) in ice_rem_prof_id()
3174 if (prof->tcam[i].in_use) { in ice_rem_prof_id()
3175 prof->tcam[i].in_use = false; in ice_rem_prof_id()
3177 prof->tcam[i].tcam_idx); in ice_rem_prof_id()