// SPDX-License-Identifier: BSD-3-Clause-Clear /* * Copyright (c) 2018-2019 The Linux Foundation. All rights reserved. * Copyright (c) 2021-2023 Qualcomm Innovation Center, Inc. All rights reserved. */ #include "core.h" #include "peer.h" #include "debug.h" static struct ath11k_peer *ath11k_peer_find_list_by_id(struct ath11k_base *ab, int peer_id) { struct ath11k_peer *peer; lockdep_assert_held(&ab->base_lock); list_for_each_entry(peer, &ab->peers, list) { if (peer->peer_id != peer_id) continue; return peer; } return NULL; } struct ath11k_peer *ath11k_peer_find(struct ath11k_base *ab, int vdev_id, const u8 *addr) { struct ath11k_peer *peer; lockdep_assert_held(&ab->base_lock); list_for_each_entry(peer, &ab->peers, list) { if (peer->vdev_id != vdev_id) continue; if (!ether_addr_equal(peer->addr, addr)) continue; return peer; } return NULL; } struct ath11k_peer *ath11k_peer_find_by_addr(struct ath11k_base *ab, const u8 *addr) { struct ath11k_peer *peer; lockdep_assert_held(&ab->base_lock); if (!ab->rhead_peer_addr) return NULL; peer = rhashtable_lookup_fast(ab->rhead_peer_addr, addr, ab->rhash_peer_addr_param); return peer; } struct ath11k_peer *ath11k_peer_find_by_id(struct ath11k_base *ab, int peer_id) { struct ath11k_peer *peer; lockdep_assert_held(&ab->base_lock); if (!ab->rhead_peer_id) return NULL; peer = rhashtable_lookup_fast(ab->rhead_peer_id, &peer_id, ab->rhash_peer_id_param); return peer; } struct ath11k_peer *ath11k_peer_find_by_vdev_id(struct ath11k_base *ab, int vdev_id) { struct ath11k_peer *peer; spin_lock_bh(&ab->base_lock); list_for_each_entry(peer, &ab->peers, list) { if (vdev_id == peer->vdev_id) { spin_unlock_bh(&ab->base_lock); return peer; } } spin_unlock_bh(&ab->base_lock); return NULL; } void ath11k_peer_unmap_event(struct ath11k_base *ab, u16 peer_id) { struct ath11k_peer *peer; spin_lock_bh(&ab->base_lock); peer = ath11k_peer_find_list_by_id(ab, peer_id); if (!peer) { ath11k_warn(ab, "peer-unmap-event: unknown peer id %d\n", peer_id); goto exit; } ath11k_dbg(ab, ATH11K_DBG_DP_HTT, "peer unmap vdev %d peer %pM id %d\n", peer->vdev_id, peer->addr, peer_id); list_del(&peer->list); kfree(peer); wake_up(&ab->peer_mapping_wq); exit: spin_unlock_bh(&ab->base_lock); } void ath11k_peer_map_event(struct ath11k_base *ab, u8 vdev_id, u16 peer_id, u8 *mac_addr, u16 ast_hash, u16 hw_peer_id) { struct ath11k_peer *peer; spin_lock_bh(&ab->base_lock); peer = ath11k_peer_find(ab, vdev_id, mac_addr); if (!peer) { peer = kzalloc(sizeof(*peer), GFP_ATOMIC); if (!peer) goto exit; peer->vdev_id = vdev_id; peer->peer_id = peer_id; peer->ast_hash = ast_hash; peer->hw_peer_id = hw_peer_id; ether_addr_copy(peer->addr, mac_addr); list_add(&peer->list, &ab->peers); wake_up(&ab->peer_mapping_wq); } ath11k_dbg(ab, ATH11K_DBG_DP_HTT, "peer map vdev %d peer %pM id %d\n", vdev_id, mac_addr, peer_id); exit: spin_unlock_bh(&ab->base_lock); } static int ath11k_wait_for_peer_common(struct ath11k_base *ab, int vdev_id, const u8 *addr, bool expect_mapped) { int ret; ret = wait_event_timeout(ab->peer_mapping_wq, ({ bool mapped; spin_lock_bh(&ab->base_lock); mapped = !!ath11k_peer_find(ab, vdev_id, addr); spin_unlock_bh(&ab->base_lock); (mapped == expect_mapped || test_bit(ATH11K_FLAG_CRASH_FLUSH, &ab->dev_flags)); }), 3 * HZ); if (ret <= 0) return -ETIMEDOUT; return 0; } static inline int ath11k_peer_rhash_insert(struct ath11k_base *ab, struct rhashtable *rtbl, struct rhash_head *rhead, struct rhashtable_params *params, void *key) { struct ath11k_peer *tmp; lockdep_assert_held(&ab->tbl_mtx_lock); tmp = rhashtable_lookup_get_insert_fast(rtbl, rhead, *params); if (!tmp) return 0; else if (IS_ERR(tmp)) return PTR_ERR(tmp); else return -EEXIST; } static inline int ath11k_peer_rhash_remove(struct ath11k_base *ab, struct rhashtable *rtbl, struct rhash_head *rhead, struct rhashtable_params *params) { int ret; lockdep_assert_held(&ab->tbl_mtx_lock); ret = rhashtable_remove_fast(rtbl, rhead, *params); if (ret && ret != -ENOENT) return ret; return 0; } static int ath11k_peer_rhash_add(struct ath11k_base *ab, struct ath11k_peer *peer) { int ret; lockdep_assert_held(&ab->base_lock); lockdep_assert_held(&ab->tbl_mtx_lock); if (!ab->rhead_peer_id || !ab->rhead_peer_addr) return -EPERM; ret = ath11k_peer_rhash_insert(ab, ab->rhead_peer_id, &peer->rhash_id, &ab->rhash_peer_id_param, &peer->peer_id); if (ret) { ath11k_warn(ab, "failed to add peer %pM with id %d in rhash_id ret %d\n", peer->addr, peer->peer_id, ret); return ret; } ret = ath11k_peer_rhash_insert(ab, ab->rhead_peer_addr, &peer->rhash_addr, &ab->rhash_peer_addr_param, &peer->addr); if (ret) { ath11k_warn(ab, "failed to add peer %pM with id %d in rhash_addr ret %d\n", peer->addr, peer->peer_id, ret); goto err_clean; } return 0; err_clean: ath11k_peer_rhash_remove(ab, ab->rhead_peer_id, &peer->rhash_id, &ab->rhash_peer_id_param); return ret; } void ath11k_peer_cleanup(struct ath11k *ar, u32 vdev_id) { struct ath11k_peer *peer, *tmp; struct ath11k_base *ab = ar->ab; lockdep_assert_held(&ar->conf_mutex); mutex_lock(&ab->tbl_mtx_lock); spin_lock_bh(&ab->base_lock); list_for_each_entry_safe(peer, tmp, &ab->peers, list) { if (peer->vdev_id != vdev_id) continue; ath11k_warn(ab, "removing stale peer %pM from vdev_id %d\n", peer->addr, vdev_id); ath11k_peer_rhash_delete(ab, peer); list_del(&peer->list); kfree(peer); ar->num_peers--; } spin_unlock_bh(&ab->base_lock); mutex_unlock(&ab->tbl_mtx_lock); } static int ath11k_wait_for_peer_deleted(struct ath11k *ar, int vdev_id, const u8 *addr) { return ath11k_wait_for_peer_common(ar->ab, vdev_id, addr, false); } int ath11k_wait_for_peer_delete_done(struct ath11k *ar, u32 vdev_id, const u8 *addr) { int ret; unsigned long time_left; ret = ath11k_wait_for_peer_deleted(ar, vdev_id, addr); if (ret) { ath11k_warn(ar->ab, "failed wait for peer deleted"); return ret; } time_left = wait_for_completion_timeout(&ar->peer_delete_done, 3 * HZ); if (time_left == 0) { ath11k_warn(ar->ab, "Timeout in receiving peer delete response\n"); return -ETIMEDOUT; } return 0; } static int __ath11k_peer_delete(struct ath11k *ar, u32 vdev_id, const u8 *addr) { int ret; struct ath11k_peer *peer; struct ath11k_base *ab = ar->ab; lockdep_assert_held(&ar->conf_mutex); mutex_lock(&ab->tbl_mtx_lock); spin_lock_bh(&ab->base_lock); peer = ath11k_peer_find_by_addr(ab, addr); /* Check if the found peer is what we want to remove. * While the sta is transitioning to another band we may * have 2 peer with the same addr assigned to different * vdev_id. Make sure we are deleting the correct peer. */ if (peer && peer->vdev_id == vdev_id) ath11k_peer_rhash_delete(ab, peer); /* Fallback to peer list search if the correct peer can't be found. * Skip the deletion of the peer from the rhash since it has already * been deleted in peer add. */ if (!peer) peer = ath11k_peer_find(ab, vdev_id, addr); if (!peer) { spin_unlock_bh(&ab->base_lock); mutex_unlock(&ab->tbl_mtx_lock); ath11k_warn(ab, "failed to find peer vdev_id %d addr %pM in delete\n", vdev_id, addr); return -EINVAL; } spin_unlock_bh(&ab->base_lock); mutex_unlock(&ab->tbl_mtx_lock); reinit_completion(&ar->peer_delete_done); ret = ath11k_wmi_send_peer_delete_cmd(ar, addr, vdev_id); if (ret) { ath11k_warn(ab, "failed to delete peer vdev_id %d addr %pM ret %d\n", vdev_id, addr, ret); return ret; } ret = ath11k_wait_for_peer_delete_done(ar, vdev_id, addr); if (ret) return ret; return 0; } int ath11k_peer_delete(struct ath11k *ar, u32 vdev_id, u8 *addr) { int ret; lockdep_assert_held(&ar->conf_mutex); ret = __ath11k_peer_delete(ar, vdev_id, addr); if (ret) return ret; ar->num_peers--; return 0; } static int ath11k_wait_for_peer_created(struct ath11k *ar, int vdev_id, const u8 *addr) { return ath11k_wait_for_peer_common(ar->ab, vdev_id, addr, true); } int ath11k_peer_create(struct ath11k *ar, struct ath11k_vif *arvif, struct ieee80211_sta *sta, struct peer_create_params *param) { struct ath11k_peer *peer; struct ath11k_sta *arsta; int ret, fbret; lockdep_assert_held(&ar->conf_mutex); if (ar->num_peers > (ar->max_num_peers - 1)) { ath11k_warn(ar->ab, "failed to create peer due to insufficient peer entry resource in firmware\n"); return -ENOBUFS; } mutex_lock(&ar->ab->tbl_mtx_lock); spin_lock_bh(&ar->ab->base_lock); peer = ath11k_peer_find_by_addr(ar->ab, param->peer_addr); if (peer) { if (peer->vdev_id == param->vdev_id) { spin_unlock_bh(&ar->ab->base_lock); mutex_unlock(&ar->ab->tbl_mtx_lock); return -EINVAL; } /* Assume sta is transitioning to another band. * Remove here the peer from rhash. */ ath11k_peer_rhash_delete(ar->ab, peer); } spin_unlock_bh(&ar->ab->base_lock); mutex_unlock(&ar->ab->tbl_mtx_lock); ret = ath11k_wmi_send_peer_create_cmd(ar, param); if (ret) { ath11k_warn(ar->ab, "failed to send peer create vdev_id %d ret %d\n", param->vdev_id, ret); return ret; } ret = ath11k_wait_for_peer_created(ar, param->vdev_id, param->peer_addr); if (ret) return ret; mutex_lock(&ar->ab->tbl_mtx_lock); spin_lock_bh(&ar->ab->base_lock); peer = ath11k_peer_find(ar->ab, param->vdev_id, param->peer_addr); if (!peer) { spin_unlock_bh(&ar->ab->base_lock); mutex_unlock(&ar->ab->tbl_mtx_lock); ath11k_warn(ar->ab, "failed to find peer %pM on vdev %i after creation\n", param->peer_addr, param->vdev_id); ret = -ENOENT; goto cleanup; } ret = ath11k_peer_rhash_add(ar->ab, peer); if (ret) { spin_unlock_bh(&ar->ab->base_lock); mutex_unlock(&ar->ab->tbl_mtx_lock); goto cleanup; } peer->pdev_idx = ar->pdev_idx; peer->sta = sta; if (arvif->vif->type == NL80211_IFTYPE_STATION) { arvif->ast_hash = peer->ast_hash; arvif->ast_idx = peer->hw_peer_id; } peer->sec_type = HAL_ENCRYPT_TYPE_OPEN; peer->sec_type_grp = HAL_ENCRYPT_TYPE_OPEN; if (sta) { arsta = (struct ath11k_sta *)sta->drv_priv; arsta->tcl_metadata |= FIELD_PREP(HTT_TCL_META_DATA_TYPE, 0) | FIELD_PREP(HTT_TCL_META_DATA_PEER_ID, peer->peer_id); /* set HTT extension valid bit to 0 by default */ arsta->tcl_metadata &= ~HTT_TCL_META_DATA_VALID_HTT; } ar->num_peers++; spin_unlock_bh(&ar->ab->base_lock); mutex_unlock(&ar->ab->tbl_mtx_lock); return 0; cleanup: fbret = __ath11k_peer_delete(ar, param->vdev_id, param->peer_addr); if (fbret) ath11k_warn(ar->ab, "failed peer %pM delete vdev_id %d fallback ret %d\n", param->peer_addr, param->vdev_id, fbret); return ret; } int ath11k_peer_rhash_delete(struct ath11k_base *ab, struct ath11k_peer *peer) { int ret; lockdep_assert_held(&ab->base_lock); lockdep_assert_held(&ab->tbl_mtx_lock); if (!ab->rhead_peer_id || !ab->rhead_peer_addr) return -EPERM; ret = ath11k_peer_rhash_remove(ab, ab->rhead_peer_addr, &peer->rhash_addr, &ab->rhash_peer_addr_param); if (ret) { ath11k_warn(ab, "failed to remove peer %pM id %d in rhash_addr ret %d\n", peer->addr, peer->peer_id, ret); return ret; } ret = ath11k_peer_rhash_remove(ab, ab->rhead_peer_id, &peer->rhash_id, &ab->rhash_peer_id_param); if (ret) { ath11k_warn(ab, "failed to remove peer %pM id %d in rhash_id ret %d\n", peer->addr, peer->peer_id, ret); return ret; } return 0; } static int ath11k_peer_rhash_id_tbl_init(struct ath11k_base *ab) { struct rhashtable_params *param; struct rhashtable *rhash_id_tbl; int ret; size_t size; lockdep_assert_held(&ab->tbl_mtx_lock); if (ab->rhead_peer_id) return 0; size = sizeof(*ab->rhead_peer_id); rhash_id_tbl = kzalloc(size, GFP_KERNEL); if (!rhash_id_tbl) { ath11k_warn(ab, "failed to init rhash id table due to no mem (size %zu)\n", size); return -ENOMEM; } param = &ab->rhash_peer_id_param; param->key_offset = offsetof(struct ath11k_peer, peer_id); param->head_offset = offsetof(struct ath11k_peer, rhash_id); param->key_len = sizeof_field(struct ath11k_peer, peer_id); param->automatic_shrinking = true; param->nelem_hint = ab->num_radios * TARGET_NUM_PEERS_PDEV(ab); ret = rhashtable_init(rhash_id_tbl, param); if (ret) { ath11k_warn(ab, "failed to init peer id rhash table %d\n", ret); goto err_free; } spin_lock_bh(&ab->base_lock); if (!ab->rhead_peer_id) { ab->rhead_peer_id = rhash_id_tbl; } else { spin_unlock_bh(&ab->base_lock); goto cleanup_tbl; } spin_unlock_bh(&ab->base_lock); return 0; cleanup_tbl: rhashtable_destroy(rhash_id_tbl); err_free: kfree(rhash_id_tbl); return ret; } static int ath11k_peer_rhash_addr_tbl_init(struct ath11k_base *ab) { struct rhashtable_params *param; struct rhashtable *rhash_addr_tbl; int ret; size_t size; lockdep_assert_held(&ab->tbl_mtx_lock); if (ab->rhead_peer_addr) return 0; size = sizeof(*ab->rhead_peer_addr); rhash_addr_tbl = kzalloc(size, GFP_KERNEL); if (!rhash_addr_tbl) { ath11k_warn(ab, "failed to init rhash addr table due to no mem (size %zu)\n", size); return -ENOMEM; } param = &ab->rhash_peer_addr_param; param->key_offset = offsetof(struct ath11k_peer, addr); param->head_offset = offsetof(struct ath11k_peer, rhash_addr); param->key_len = sizeof_field(struct ath11k_peer, addr); param->automatic_shrinking = true; param->nelem_hint = ab->num_radios * TARGET_NUM_PEERS_PDEV(ab); ret = rhashtable_init(rhash_addr_tbl, param); if (ret) { ath11k_warn(ab, "failed to init peer addr rhash table %d\n", ret); goto err_free; } spin_lock_bh(&ab->base_lock); if (!ab->rhead_peer_addr) { ab->rhead_peer_addr = rhash_addr_tbl; } else { spin_unlock_bh(&ab->base_lock); goto cleanup_tbl; } spin_unlock_bh(&ab->base_lock); return 0; cleanup_tbl: rhashtable_destroy(rhash_addr_tbl); err_free: kfree(rhash_addr_tbl); return ret; } static inline void ath11k_peer_rhash_id_tbl_destroy(struct ath11k_base *ab) { lockdep_assert_held(&ab->tbl_mtx_lock); if (!ab->rhead_peer_id) return; rhashtable_destroy(ab->rhead_peer_id); kfree(ab->rhead_peer_id); ab->rhead_peer_id = NULL; } static inline void ath11k_peer_rhash_addr_tbl_destroy(struct ath11k_base *ab) { lockdep_assert_held(&ab->tbl_mtx_lock); if (!ab->rhead_peer_addr) return; rhashtable_destroy(ab->rhead_peer_addr); kfree(ab->rhead_peer_addr); ab->rhead_peer_addr = NULL; } int ath11k_peer_rhash_tbl_init(struct ath11k_base *ab) { int ret; mutex_lock(&ab->tbl_mtx_lock); ret = ath11k_peer_rhash_id_tbl_init(ab); if (ret) goto out; ret = ath11k_peer_rhash_addr_tbl_init(ab); if (ret) goto cleanup_tbl; mutex_unlock(&ab->tbl_mtx_lock); return 0; cleanup_tbl: ath11k_peer_rhash_id_tbl_destroy(ab); out: mutex_unlock(&ab->tbl_mtx_lock); return ret; } void ath11k_peer_rhash_tbl_destroy(struct ath11k_base *ab) { mutex_lock(&ab->tbl_mtx_lock); ath11k_peer_rhash_addr_tbl_destroy(ab); ath11k_peer_rhash_id_tbl_destroy(ab); mutex_unlock(&ab->tbl_mtx_lock); }