1 /* 2 * Copyright (c) 2016-2020 The Linux Foundation. All rights reserved. 3 * 4 * Permission to use, copy, modify, and/or distribute this software for 5 * any purpose with or without fee is hereby granted, provided that the 6 * above copyright notice and this permission notice appear in all 7 * copies. 8 * 9 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL 10 * WARRANTIES WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED 11 * WARRANTIES OF MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE 12 * AUTHOR BE LIABLE FOR ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL 13 * DAMAGES OR ANY DAMAGES WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR 14 * PROFITS, WHETHER IN AN ACTION OF CONTRACT, NEGLIGENCE OR OTHER 15 * TORTIOUS ACTION, ARISING OUT OF OR IN CONNECTION WITH THE USE OR 16 * PERFORMANCE OF THIS SOFTWARE. 17 */ 18 #ifndef _DP_PEER_H_ 19 #define _DP_PEER_H_ 20 21 #include <qdf_types.h> 22 #include <qdf_lock.h> 23 #include "dp_types.h" 24 25 #define DP_INVALID_PEER_ID 0xffff 26 27 #define DP_FW_PEER_STATS_CMP_TIMEOUT_MSEC 5000 28 /** 29 * __dp_peer_find_by_id() - Returns peer object given the peer id 30 * 31 * @soc : core DP soc context 32 * @peer_id : peer id from peer object can be retrieved 33 * 34 * Return: struct dp_peer*: Pointer to DP peer object 35 */ 36 static inline struct dp_peer * 37 __dp_peer_find_by_id(struct dp_soc *soc, 38 uint16_t peer_id) 39 { 40 struct dp_peer *peer; 41 42 /* TODO: Hold lock */ 43 peer = (peer_id >= soc->max_peers) ? NULL : 44 soc->peer_id_to_obj_map[peer_id]; 45 46 return peer; 47 } 48 49 #ifdef PEER_PROTECTED_ACCESS 50 /** 51 * dp_peer_find_by_id() - Returns peer object given the peer id 52 * if delete_in_progress in not set for peer 53 * 54 * @soc : core DP soc context 55 * @peer_id : peer id from peer object can be retrieved 56 * 57 * Return: struct dp_peer*: Pointer to DP peer object 58 */ 59 static inline 60 struct dp_peer *dp_peer_find_by_id(struct dp_soc *soc, 61 uint16_t peer_id) 62 { 63 struct dp_peer *peer; 64 65 qdf_spin_lock_bh(&soc->peer_ref_mutex); 66 peer = __dp_peer_find_by_id(soc, peer_id); 67 if (!peer || (peer && peer->delete_in_progress)) { 68 qdf_spin_unlock_bh(&soc->peer_ref_mutex); 69 return NULL; 70 } 71 qdf_atomic_inc(&peer->ref_cnt); 72 qdf_spin_unlock_bh(&soc->peer_ref_mutex); 73 74 return peer; 75 } 76 #else 77 static inline struct dp_peer * 78 dp_peer_find_by_id(struct dp_soc *soc, 79 uint16_t peer_id) 80 { 81 struct dp_peer *peer; 82 83 peer = __dp_peer_find_by_id (soc, peer_id); 84 if (peer && peer->delete_in_progress) { 85 return NULL; 86 } 87 88 return peer; 89 } 90 #endif /* PEER_LOCK_REF_PROTECT */ 91 92 void dp_print_ast_stats(struct dp_soc *soc); 93 void dp_rx_peer_map_handler(struct dp_soc *soc, uint16_t peer_id, 94 uint16_t hw_peer_id, uint8_t vdev_id, 95 uint8_t *peer_mac_addr, uint16_t ast_hash, 96 uint8_t is_wds); 97 void dp_rx_peer_unmap_handler(struct dp_soc *soc, uint16_t peer_id, 98 uint8_t vdev_id, uint8_t *peer_mac_addr, 99 uint8_t is_wds); 100 void dp_rx_sec_ind_handler(struct dp_soc *soc, uint16_t peer_id, 101 enum cdp_sec_type sec_type, int is_unicast, 102 u_int32_t *michael_key, u_int32_t *rx_pn); 103 uint8_t dp_get_peer_mac_addr_frm_id(struct cdp_soc_t *soc_handle, 104 uint16_t peer_id, uint8_t *peer_mac); 105 106 int dp_peer_add_ast(struct dp_soc *soc, struct dp_peer *peer, 107 uint8_t *mac_addr, enum cdp_txrx_ast_entry_type type, 108 uint32_t flags); 109 110 void dp_peer_del_ast(struct dp_soc *soc, struct dp_ast_entry *ast_entry); 111 112 void dp_peer_ast_unmap_handler(struct dp_soc *soc, 113 struct dp_ast_entry *ast_entry); 114 115 int dp_peer_update_ast(struct dp_soc *soc, struct dp_peer *peer, 116 struct dp_ast_entry *ast_entry, uint32_t flags); 117 118 struct dp_ast_entry *dp_peer_ast_hash_find_by_pdevid(struct dp_soc *soc, 119 uint8_t *ast_mac_addr, 120 uint8_t pdev_id); 121 122 struct dp_ast_entry *dp_peer_ast_hash_find_soc(struct dp_soc *soc, 123 uint8_t *ast_mac_addr); 124 125 struct dp_ast_entry *dp_peer_ast_list_find(struct dp_soc *soc, 126 struct dp_peer *peer, 127 uint8_t *ast_mac_addr); 128 129 uint8_t dp_peer_ast_get_pdev_id(struct dp_soc *soc, 130 struct dp_ast_entry *ast_entry); 131 132 133 uint8_t dp_peer_ast_get_next_hop(struct dp_soc *soc, 134 struct dp_ast_entry *ast_entry); 135 136 void dp_peer_ast_set_type(struct dp_soc *soc, 137 struct dp_ast_entry *ast_entry, 138 enum cdp_txrx_ast_entry_type type); 139 140 void dp_peer_ast_send_wds_del(struct dp_soc *soc, 141 struct dp_ast_entry *ast_entry); 142 143 void dp_peer_free_hmwds_cb(struct cdp_ctrl_objmgr_psoc *ctrl_psoc, 144 struct cdp_soc *dp_soc, 145 void *cookie, 146 enum cdp_ast_free_status status); 147 148 void dp_peer_ast_hash_remove(struct dp_soc *soc, 149 struct dp_ast_entry *ase); 150 151 void dp_peer_free_ast_entry(struct dp_soc *soc, 152 struct dp_ast_entry *ast_entry); 153 154 void dp_peer_unlink_ast_entry(struct dp_soc *soc, 155 struct dp_ast_entry *ast_entry); 156 157 /* 158 * dp_peer_find_by_id_exist - check if peer exists for given id 159 * @soc: core DP soc context 160 * @peer_id: peer id from peer object can be retrieved 161 * 162 * Return: true if peer exists of false otherwise 163 */ 164 bool dp_peer_find_by_id_valid(struct dp_soc *soc, uint16_t peer_id); 165 166 #define DP_AST_ASSERT(_condition) \ 167 do { \ 168 if (!(_condition)) { \ 169 dp_print_ast_stats(soc);\ 170 QDF_BUG(_condition); \ 171 } \ 172 } while (0) 173 174 /** 175 * dp_peer_update_inactive_time - Update inactive time for peer 176 * @pdev: pdev object 177 * @tag_type: htt_tlv_tag type 178 * #tag_buf: buf message 179 */ 180 void 181 dp_peer_update_inactive_time(struct dp_pdev *pdev, uint32_t tag_type, 182 uint32_t *tag_buf); 183 184 #ifndef QCA_MULTIPASS_SUPPORT 185 /** 186 * dp_peer_set_vlan_id: set vlan_id for this peer 187 * @cdp_soc: soc handle 188 * @vdev_id: id of vdev object 189 * @peer_mac: mac address 190 * @vlan_id: vlan id for peer 191 * 192 * return: void 193 */ 194 static inline 195 void dp_peer_set_vlan_id(struct cdp_soc_t *cdp_soc, 196 uint8_t vdev_id, uint8_t *peer_mac, 197 uint16_t vlan_id) 198 { 199 } 200 201 /** 202 * dp_set_vlan_groupkey: set vlan map for vdev 203 * @soc: pointer to soc 204 * @vdev_id: id of vdev handle 205 * @vlan_id: vlan_id 206 * @group_key: group key for vlan 207 * 208 * return: set success/failure 209 */ 210 static inline 211 QDF_STATUS dp_set_vlan_groupkey(struct cdp_soc_t *soc, uint8_t vdev_id, 212 uint16_t vlan_id, uint16_t group_key) 213 { 214 return QDF_STATUS_SUCCESS; 215 } 216 217 /** 218 * dp_peer_multipass_list_init: initialize multipass peer list 219 * @vdev: pointer to vdev 220 * 221 * return: void 222 */ 223 static inline 224 void dp_peer_multipass_list_init(struct dp_vdev *vdev) 225 { 226 } 227 228 /** 229 * dp_peer_multipass_list_remove: remove peer from special peer list 230 * @peer: peer handle 231 * 232 * return: void 233 */ 234 static inline 235 void dp_peer_multipass_list_remove(struct dp_peer *peer) 236 { 237 } 238 #else 239 void dp_peer_set_vlan_id(struct cdp_soc_t *cdp_soc, 240 uint8_t vdev_id, uint8_t *peer_mac, 241 uint16_t vlan_id); 242 QDF_STATUS dp_set_vlan_groupkey(struct cdp_soc_t *soc, uint8_t vdev_id, 243 uint16_t vlan_id, uint16_t group_key); 244 void dp_peer_multipass_list_init(struct dp_vdev *vdev); 245 void dp_peer_multipass_list_remove(struct dp_peer *peer); 246 #endif 247 248 249 #ifndef QCA_PEER_MULTIQ_SUPPORT 250 /** 251 * dp_peer_reset_flowq_map() - reset peer flowq map table 252 * @peer - dp peer handle 253 * 254 * Return: none 255 */ 256 static inline 257 void dp_peer_reset_flowq_map(struct dp_peer *peer) 258 { 259 } 260 261 /** 262 * dp_peer_ast_index_flow_queue_map_create() - create ast index flow queue map 263 * @soc - genereic soc handle 264 * @is_wds - flag to indicate if peer is wds 265 * @peer_id - peer_id from htt peer map message 266 * @peer_mac_addr - mac address of the peer 267 * @ast_info - ast flow override information from peer map 268 * 269 * Return: none 270 */ 271 static inline 272 void dp_peer_ast_index_flow_queue_map_create(void *soc_hdl, 273 bool is_wds, uint16_t peer_id, uint8_t *peer_mac_addr, 274 struct dp_ast_flow_override_info *ast_info) 275 { 276 } 277 #else 278 void dp_peer_reset_flowq_map(struct dp_peer *peer); 279 void dp_peer_ast_index_flow_queue_map_create(void *soc_hdl, 280 bool is_wds, uint16_t peer_id, uint8_t *peer_mac_addr, 281 struct dp_ast_flow_override_info *ast_info); 282 #endif 283 284 /** 285 * dp_peer_update_pkt_capture_params: Set Rx & Tx Capture flags for a peer 286 * @soc: DP SOC handle 287 * @pdev_id: id of DP pdev handle 288 * @is_rx_pkt_cap_enable: enable/disable Rx packet capture in monitor mode 289 * @is_tx_pkt_cap_enable: enable/disable Tx packet capture in monitor mode 290 * @peer_mac: MAC address for which the above need to be enabled/disabled 291 * 292 * Return: Success if Rx & Tx capture is enabled for peer, false otherwise 293 */ 294 QDF_STATUS 295 dp_peer_update_pkt_capture_params(ol_txrx_soc_handle soc, 296 uint8_t pdev_id, 297 bool is_rx_pkt_cap_enable, 298 bool is_tx_pkt_cap_enable, 299 uint8_t *peer_mac); 300 301 /* 302 * dp_rx_tid_delete_cb() - Callback to flush reo descriptor HW cache 303 * after deleting the entries (ie., setting valid=0) 304 * 305 * @soc: DP SOC handle 306 * @cb_ctxt: Callback context 307 * @reo_status: REO command status 308 */ 309 void dp_rx_tid_delete_cb(struct dp_soc *soc, 310 void *cb_ctxt, 311 union hal_reo_status *reo_status); 312 #endif /* _DP_PEER_H_ */ 313