Lines Matching refs:rxf
25 #define bna_rxf_vlan_cfg_soft_reset(rxf) \ argument
27 (rxf)->vlan_pending_bitmask = (u8)BFI_VLAN_BMASK_ALL; \
28 (rxf)->vlan_strip_pending = true; \
31 #define bna_rxf_rss_cfg_soft_reset(rxf) \ argument
33 if ((rxf)->rss_status == BNA_STATUS_T_ENABLED) \
34 (rxf)->rss_pending = (BNA_RSS_F_RIT_PENDING | \
39 static int bna_rxf_cfg_apply(struct bna_rxf *rxf);
40 static void bna_rxf_cfg_reset(struct bna_rxf *rxf);
41 static int bna_rxf_ucast_cfg_apply(struct bna_rxf *rxf);
42 static int bna_rxf_promisc_cfg_apply(struct bna_rxf *rxf);
43 static int bna_rxf_allmulti_cfg_apply(struct bna_rxf *rxf);
44 static int bna_rxf_vlan_strip_cfg_apply(struct bna_rxf *rxf);
45 static int bna_rxf_ucast_cfg_reset(struct bna_rxf *rxf,
47 static int bna_rxf_promisc_cfg_reset(struct bna_rxf *rxf,
49 static int bna_rxf_allmulti_cfg_reset(struct bna_rxf *rxf,
62 bna_rxf_sm_stopped_entry(struct bna_rxf *rxf) in bna_rxf_sm_stopped_entry() argument
64 call_rxf_stop_cbfn(rxf); in bna_rxf_sm_stopped_entry()
68 bna_rxf_sm_stopped(struct bna_rxf *rxf, enum bna_rxf_event event) in bna_rxf_sm_stopped() argument
72 bfa_fsm_set_state(rxf, bna_rxf_sm_cfg_wait); in bna_rxf_sm_stopped()
76 call_rxf_stop_cbfn(rxf); in bna_rxf_sm_stopped()
84 call_rxf_cam_fltr_cbfn(rxf); in bna_rxf_sm_stopped()
93 bna_rxf_sm_cfg_wait_entry(struct bna_rxf *rxf) in bna_rxf_sm_cfg_wait_entry() argument
95 if (!bna_rxf_cfg_apply(rxf)) { in bna_rxf_sm_cfg_wait_entry()
97 bfa_fsm_set_state(rxf, bna_rxf_sm_started); in bna_rxf_sm_cfg_wait_entry()
102 bna_rxf_sm_cfg_wait(struct bna_rxf *rxf, enum bna_rxf_event event) in bna_rxf_sm_cfg_wait() argument
106 bfa_fsm_set_state(rxf, bna_rxf_sm_last_resp_wait); in bna_rxf_sm_cfg_wait()
110 bna_rxf_cfg_reset(rxf); in bna_rxf_sm_cfg_wait()
111 call_rxf_start_cbfn(rxf); in bna_rxf_sm_cfg_wait()
112 call_rxf_cam_fltr_cbfn(rxf); in bna_rxf_sm_cfg_wait()
113 bfa_fsm_set_state(rxf, bna_rxf_sm_stopped); in bna_rxf_sm_cfg_wait()
121 if (!bna_rxf_cfg_apply(rxf)) { in bna_rxf_sm_cfg_wait()
123 bfa_fsm_set_state(rxf, bna_rxf_sm_started); in bna_rxf_sm_cfg_wait()
133 bna_rxf_sm_started_entry(struct bna_rxf *rxf) in bna_rxf_sm_started_entry() argument
135 call_rxf_start_cbfn(rxf); in bna_rxf_sm_started_entry()
136 call_rxf_cam_fltr_cbfn(rxf); in bna_rxf_sm_started_entry()
140 bna_rxf_sm_started(struct bna_rxf *rxf, enum bna_rxf_event event) in bna_rxf_sm_started() argument
145 bna_rxf_cfg_reset(rxf); in bna_rxf_sm_started()
146 bfa_fsm_set_state(rxf, bna_rxf_sm_stopped); in bna_rxf_sm_started()
150 bfa_fsm_set_state(rxf, bna_rxf_sm_cfg_wait); in bna_rxf_sm_started()
159 bna_rxf_sm_last_resp_wait_entry(struct bna_rxf *rxf) in bna_rxf_sm_last_resp_wait_entry() argument
164 bna_rxf_sm_last_resp_wait(struct bna_rxf *rxf, enum bna_rxf_event event) in bna_rxf_sm_last_resp_wait() argument
169 bna_rxf_cfg_reset(rxf); in bna_rxf_sm_last_resp_wait()
170 bfa_fsm_set_state(rxf, bna_rxf_sm_stopped); in bna_rxf_sm_last_resp_wait()
179 bna_bfi_ucast_req(struct bna_rxf *rxf, struct bna_mac *mac, in bna_bfi_ucast_req() argument
182 struct bfi_enet_ucast_req *req = &rxf->bfi_enet_cmd.ucast_req; in bna_bfi_ucast_req()
184 bfi_msgq_mhdr_set(req->mh, BFI_MC_ENET, req_type, 0, rxf->rx->rid); in bna_bfi_ucast_req()
188 bfa_msgq_cmd_set(&rxf->msgq_cmd, NULL, NULL, in bna_bfi_ucast_req()
190 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_ucast_req()
194 bna_bfi_mcast_add_req(struct bna_rxf *rxf, struct bna_mac *mac) in bna_bfi_mcast_add_req() argument
197 &rxf->bfi_enet_cmd.mcast_add_req; in bna_bfi_mcast_add_req()
200 0, rxf->rx->rid); in bna_bfi_mcast_add_req()
204 bfa_msgq_cmd_set(&rxf->msgq_cmd, NULL, NULL, in bna_bfi_mcast_add_req()
206 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_mcast_add_req()
210 bna_bfi_mcast_del_req(struct bna_rxf *rxf, u16 handle) in bna_bfi_mcast_del_req() argument
213 &rxf->bfi_enet_cmd.mcast_del_req; in bna_bfi_mcast_del_req()
216 0, rxf->rx->rid); in bna_bfi_mcast_del_req()
220 bfa_msgq_cmd_set(&rxf->msgq_cmd, NULL, NULL, in bna_bfi_mcast_del_req()
222 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_mcast_del_req()
226 bna_bfi_mcast_filter_req(struct bna_rxf *rxf, enum bna_status status) in bna_bfi_mcast_filter_req() argument
228 struct bfi_enet_enable_req *req = &rxf->bfi_enet_cmd.req; in bna_bfi_mcast_filter_req()
231 BFI_ENET_H2I_MAC_MCAST_FILTER_REQ, 0, rxf->rx->rid); in bna_bfi_mcast_filter_req()
235 bfa_msgq_cmd_set(&rxf->msgq_cmd, NULL, NULL, in bna_bfi_mcast_filter_req()
237 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_mcast_filter_req()
241 bna_bfi_rx_promisc_req(struct bna_rxf *rxf, enum bna_status status) in bna_bfi_rx_promisc_req() argument
243 struct bfi_enet_enable_req *req = &rxf->bfi_enet_cmd.req; in bna_bfi_rx_promisc_req()
246 BFI_ENET_H2I_RX_PROMISCUOUS_REQ, 0, rxf->rx->rid); in bna_bfi_rx_promisc_req()
250 bfa_msgq_cmd_set(&rxf->msgq_cmd, NULL, NULL, in bna_bfi_rx_promisc_req()
252 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_rx_promisc_req()
256 bna_bfi_rx_vlan_filter_set(struct bna_rxf *rxf, u8 block_idx) in bna_bfi_rx_vlan_filter_set() argument
258 struct bfi_enet_rx_vlan_req *req = &rxf->bfi_enet_cmd.vlan_req; in bna_bfi_rx_vlan_filter_set()
263 BFI_ENET_H2I_RX_VLAN_SET_REQ, 0, rxf->rx->rid); in bna_bfi_rx_vlan_filter_set()
269 if (rxf->vlan_filter_status == BNA_STATUS_T_ENABLED) in bna_bfi_rx_vlan_filter_set()
271 htonl(rxf->vlan_filter_table[j]); in bna_bfi_rx_vlan_filter_set()
275 bfa_msgq_cmd_set(&rxf->msgq_cmd, NULL, NULL, in bna_bfi_rx_vlan_filter_set()
277 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_rx_vlan_filter_set()
281 bna_bfi_vlan_strip_enable(struct bna_rxf *rxf) in bna_bfi_vlan_strip_enable() argument
283 struct bfi_enet_enable_req *req = &rxf->bfi_enet_cmd.req; in bna_bfi_vlan_strip_enable()
286 BFI_ENET_H2I_RX_VLAN_STRIP_ENABLE_REQ, 0, rxf->rx->rid); in bna_bfi_vlan_strip_enable()
289 req->enable = rxf->vlan_strip_status; in bna_bfi_vlan_strip_enable()
290 bfa_msgq_cmd_set(&rxf->msgq_cmd, NULL, NULL, in bna_bfi_vlan_strip_enable()
292 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_vlan_strip_enable()
296 bna_bfi_rit_cfg(struct bna_rxf *rxf) in bna_bfi_rit_cfg() argument
298 struct bfi_enet_rit_req *req = &rxf->bfi_enet_cmd.rit_req; in bna_bfi_rit_cfg()
301 BFI_ENET_H2I_RIT_CFG_REQ, 0, rxf->rx->rid); in bna_bfi_rit_cfg()
304 req->size = htons(rxf->rit_size); in bna_bfi_rit_cfg()
305 memcpy(&req->table[0], rxf->rit, rxf->rit_size); in bna_bfi_rit_cfg()
306 bfa_msgq_cmd_set(&rxf->msgq_cmd, NULL, NULL, in bna_bfi_rit_cfg()
308 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_rit_cfg()
312 bna_bfi_rss_cfg(struct bna_rxf *rxf) in bna_bfi_rss_cfg() argument
314 struct bfi_enet_rss_cfg_req *req = &rxf->bfi_enet_cmd.rss_req; in bna_bfi_rss_cfg()
318 BFI_ENET_H2I_RSS_CFG_REQ, 0, rxf->rx->rid); in bna_bfi_rss_cfg()
321 req->cfg.type = rxf->rss_cfg.hash_type; in bna_bfi_rss_cfg()
322 req->cfg.mask = rxf->rss_cfg.hash_mask; in bna_bfi_rss_cfg()
325 htonl(rxf->rss_cfg.toeplitz_hash_key[i]); in bna_bfi_rss_cfg()
326 bfa_msgq_cmd_set(&rxf->msgq_cmd, NULL, NULL, in bna_bfi_rss_cfg()
328 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_rss_cfg()
332 bna_bfi_rss_enable(struct bna_rxf *rxf) in bna_bfi_rss_enable() argument
334 struct bfi_enet_enable_req *req = &rxf->bfi_enet_cmd.req; in bna_bfi_rss_enable()
337 BFI_ENET_H2I_RSS_ENABLE_REQ, 0, rxf->rx->rid); in bna_bfi_rss_enable()
340 req->enable = rxf->rss_status; in bna_bfi_rss_enable()
341 bfa_msgq_cmd_set(&rxf->msgq_cmd, NULL, NULL, in bna_bfi_rss_enable()
343 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_rss_enable()
348 bna_rxf_mcmac_get(struct bna_rxf *rxf, const u8 *mac_addr) in bna_rxf_mcmac_get() argument
352 list_for_each_entry(mac, &rxf->mcast_active_q, qe) in bna_rxf_mcmac_get()
356 list_for_each_entry(mac, &rxf->mcast_pending_del_q, qe) in bna_rxf_mcmac_get()
364 bna_rxf_mchandle_get(struct bna_rxf *rxf, int handle) in bna_rxf_mchandle_get() argument
368 list_for_each_entry(mchandle, &rxf->mcast_handle_q, qe) in bna_rxf_mchandle_get()
376 bna_rxf_mchandle_attach(struct bna_rxf *rxf, u8 *mac_addr, int handle) in bna_rxf_mchandle_attach() argument
381 mcmac = bna_rxf_mcmac_get(rxf, mac_addr); in bna_rxf_mchandle_attach()
382 mchandle = bna_rxf_mchandle_get(rxf, handle); in bna_rxf_mchandle_attach()
384 mchandle = bna_mcam_mod_handle_get(&rxf->rx->bna->mcam_mod); in bna_rxf_mchandle_attach()
387 list_add_tail(&mchandle->qe, &rxf->mcast_handle_q); in bna_rxf_mchandle_attach()
394 bna_rxf_mcast_del(struct bna_rxf *rxf, struct bna_mac *mac, in bna_rxf_mcast_del() argument
407 bna_bfi_mcast_del_req(rxf, mchandle->handle); in bna_rxf_mcast_del()
411 bna_mcam_mod_handle_put(&rxf->rx->bna->mcam_mod, mchandle); in bna_rxf_mcast_del()
419 bna_rxf_mcast_cfg_apply(struct bna_rxf *rxf) in bna_rxf_mcast_cfg_apply() argument
425 while (!list_empty(&rxf->mcast_pending_del_q)) { in bna_rxf_mcast_cfg_apply()
426 mac = list_first_entry(&rxf->mcast_pending_del_q, in bna_rxf_mcast_cfg_apply()
428 ret = bna_rxf_mcast_del(rxf, mac, BNA_HARD_CLEANUP); in bna_rxf_mcast_cfg_apply()
429 list_move_tail(&mac->qe, bna_mcam_mod_del_q(rxf->rx->bna)); in bna_rxf_mcast_cfg_apply()
435 if (!list_empty(&rxf->mcast_pending_add_q)) { in bna_rxf_mcast_cfg_apply()
436 mac = list_first_entry(&rxf->mcast_pending_add_q, in bna_rxf_mcast_cfg_apply()
438 list_move_tail(&mac->qe, &rxf->mcast_active_q); in bna_rxf_mcast_cfg_apply()
439 bna_bfi_mcast_add_req(rxf, mac); in bna_rxf_mcast_cfg_apply()
447 bna_rxf_vlan_cfg_apply(struct bna_rxf *rxf) in bna_rxf_vlan_cfg_apply() argument
452 if (rxf->vlan_pending_bitmask) { in bna_rxf_vlan_cfg_apply()
453 vlan_pending_bitmask = rxf->vlan_pending_bitmask; in bna_rxf_vlan_cfg_apply()
458 rxf->vlan_pending_bitmask &= ~BIT(block_idx); in bna_rxf_vlan_cfg_apply()
459 bna_bfi_rx_vlan_filter_set(rxf, block_idx); in bna_rxf_vlan_cfg_apply()
467 bna_rxf_mcast_cfg_reset(struct bna_rxf *rxf, enum bna_cleanup_type cleanup) in bna_rxf_mcast_cfg_reset() argument
473 while (!list_empty(&rxf->mcast_pending_del_q)) { in bna_rxf_mcast_cfg_reset()
474 mac = list_first_entry(&rxf->mcast_pending_del_q, in bna_rxf_mcast_cfg_reset()
476 ret = bna_rxf_mcast_del(rxf, mac, cleanup); in bna_rxf_mcast_cfg_reset()
477 list_move_tail(&mac->qe, bna_mcam_mod_del_q(rxf->rx->bna)); in bna_rxf_mcast_cfg_reset()
483 while (!list_empty(&rxf->mcast_active_q)) { in bna_rxf_mcast_cfg_reset()
484 mac = list_first_entry(&rxf->mcast_active_q, in bna_rxf_mcast_cfg_reset()
486 list_move_tail(&mac->qe, &rxf->mcast_pending_add_q); in bna_rxf_mcast_cfg_reset()
487 if (bna_rxf_mcast_del(rxf, mac, cleanup)) in bna_rxf_mcast_cfg_reset()
495 bna_rxf_rss_cfg_apply(struct bna_rxf *rxf) in bna_rxf_rss_cfg_apply() argument
497 if (rxf->rss_pending) { in bna_rxf_rss_cfg_apply()
498 if (rxf->rss_pending & BNA_RSS_F_RIT_PENDING) { in bna_rxf_rss_cfg_apply()
499 rxf->rss_pending &= ~BNA_RSS_F_RIT_PENDING; in bna_rxf_rss_cfg_apply()
500 bna_bfi_rit_cfg(rxf); in bna_rxf_rss_cfg_apply()
504 if (rxf->rss_pending & BNA_RSS_F_CFG_PENDING) { in bna_rxf_rss_cfg_apply()
505 rxf->rss_pending &= ~BNA_RSS_F_CFG_PENDING; in bna_rxf_rss_cfg_apply()
506 bna_bfi_rss_cfg(rxf); in bna_rxf_rss_cfg_apply()
510 if (rxf->rss_pending & BNA_RSS_F_STATUS_PENDING) { in bna_rxf_rss_cfg_apply()
511 rxf->rss_pending &= ~BNA_RSS_F_STATUS_PENDING; in bna_rxf_rss_cfg_apply()
512 bna_bfi_rss_enable(rxf); in bna_rxf_rss_cfg_apply()
521 bna_rxf_cfg_apply(struct bna_rxf *rxf) in bna_rxf_cfg_apply() argument
523 if (bna_rxf_ucast_cfg_apply(rxf)) in bna_rxf_cfg_apply()
526 if (bna_rxf_mcast_cfg_apply(rxf)) in bna_rxf_cfg_apply()
529 if (bna_rxf_promisc_cfg_apply(rxf)) in bna_rxf_cfg_apply()
532 if (bna_rxf_allmulti_cfg_apply(rxf)) in bna_rxf_cfg_apply()
535 if (bna_rxf_vlan_cfg_apply(rxf)) in bna_rxf_cfg_apply()
538 if (bna_rxf_vlan_strip_cfg_apply(rxf)) in bna_rxf_cfg_apply()
541 if (bna_rxf_rss_cfg_apply(rxf)) in bna_rxf_cfg_apply()
548 bna_rxf_cfg_reset(struct bna_rxf *rxf) in bna_rxf_cfg_reset() argument
550 bna_rxf_ucast_cfg_reset(rxf, BNA_SOFT_CLEANUP); in bna_rxf_cfg_reset()
551 bna_rxf_mcast_cfg_reset(rxf, BNA_SOFT_CLEANUP); in bna_rxf_cfg_reset()
552 bna_rxf_promisc_cfg_reset(rxf, BNA_SOFT_CLEANUP); in bna_rxf_cfg_reset()
553 bna_rxf_allmulti_cfg_reset(rxf, BNA_SOFT_CLEANUP); in bna_rxf_cfg_reset()
554 bna_rxf_vlan_cfg_soft_reset(rxf); in bna_rxf_cfg_reset()
555 bna_rxf_rss_cfg_soft_reset(rxf); in bna_rxf_cfg_reset()
559 bna_rit_init(struct bna_rxf *rxf, int rit_size) in bna_rit_init() argument
561 struct bna_rx *rx = rxf->rx; in bna_rit_init()
565 rxf->rit_size = rit_size; in bna_rit_init()
567 rxf->rit[offset] = rxp->cq.ccb->id; in bna_rit_init()
573 bna_bfi_rxf_cfg_rsp(struct bna_rxf *rxf, struct bfi_msgq_mhdr *msghdr) in bna_bfi_rxf_cfg_rsp() argument
575 bfa_fsm_send_event(rxf, RXF_E_FW_RESP); in bna_bfi_rxf_cfg_rsp()
579 bna_bfi_rxf_ucast_set_rsp(struct bna_rxf *rxf, in bna_bfi_rxf_ucast_set_rsp() argument
587 rxf->ucast_active_set = 0; in bna_bfi_rxf_ucast_set_rsp()
590 bfa_fsm_send_event(rxf, RXF_E_FW_RESP); in bna_bfi_rxf_ucast_set_rsp()
594 bna_bfi_rxf_mcast_add_rsp(struct bna_rxf *rxf, in bna_bfi_rxf_mcast_add_rsp() argument
598 &rxf->bfi_enet_cmd.mcast_add_req; in bna_bfi_rxf_mcast_add_rsp()
602 bna_rxf_mchandle_attach(rxf, (u8 *)&req->mac_addr, in bna_bfi_rxf_mcast_add_rsp()
604 bfa_fsm_send_event(rxf, RXF_E_FW_RESP); in bna_bfi_rxf_mcast_add_rsp()
608 bna_rxf_init(struct bna_rxf *rxf, in bna_rxf_init() argument
613 rxf->rx = rx; in bna_rxf_init()
615 INIT_LIST_HEAD(&rxf->ucast_pending_add_q); in bna_rxf_init()
616 INIT_LIST_HEAD(&rxf->ucast_pending_del_q); in bna_rxf_init()
617 rxf->ucast_pending_set = 0; in bna_rxf_init()
618 rxf->ucast_active_set = 0; in bna_rxf_init()
619 INIT_LIST_HEAD(&rxf->ucast_active_q); in bna_rxf_init()
620 rxf->ucast_pending_mac = NULL; in bna_rxf_init()
622 INIT_LIST_HEAD(&rxf->mcast_pending_add_q); in bna_rxf_init()
623 INIT_LIST_HEAD(&rxf->mcast_pending_del_q); in bna_rxf_init()
624 INIT_LIST_HEAD(&rxf->mcast_active_q); in bna_rxf_init()
625 INIT_LIST_HEAD(&rxf->mcast_handle_q); in bna_rxf_init()
627 rxf->rit = (u8 *) in bna_rxf_init()
629 bna_rit_init(rxf, q_config->num_paths); in bna_rxf_init()
631 rxf->rss_status = q_config->rss_status; in bna_rxf_init()
632 if (rxf->rss_status == BNA_STATUS_T_ENABLED) { in bna_rxf_init()
633 rxf->rss_cfg = q_config->rss_config; in bna_rxf_init()
634 rxf->rss_pending |= BNA_RSS_F_CFG_PENDING; in bna_rxf_init()
635 rxf->rss_pending |= BNA_RSS_F_RIT_PENDING; in bna_rxf_init()
636 rxf->rss_pending |= BNA_RSS_F_STATUS_PENDING; in bna_rxf_init()
639 rxf->vlan_filter_status = BNA_STATUS_T_DISABLED; in bna_rxf_init()
640 memset(rxf->vlan_filter_table, 0, in bna_rxf_init()
642 rxf->vlan_filter_table[0] |= 1; /* for pure priority tagged frames */ in bna_rxf_init()
643 rxf->vlan_pending_bitmask = (u8)BFI_VLAN_BMASK_ALL; in bna_rxf_init()
645 rxf->vlan_strip_status = q_config->vlan_strip_status; in bna_rxf_init()
647 bfa_fsm_set_state(rxf, bna_rxf_sm_stopped); in bna_rxf_init()
651 bna_rxf_uninit(struct bna_rxf *rxf) in bna_rxf_uninit() argument
655 rxf->ucast_pending_set = 0; in bna_rxf_uninit()
656 rxf->ucast_active_set = 0; in bna_rxf_uninit()
658 while (!list_empty(&rxf->ucast_pending_add_q)) { in bna_rxf_uninit()
659 mac = list_first_entry(&rxf->ucast_pending_add_q, in bna_rxf_uninit()
661 list_move_tail(&mac->qe, bna_ucam_mod_free_q(rxf->rx->bna)); in bna_rxf_uninit()
664 if (rxf->ucast_pending_mac) { in bna_rxf_uninit()
665 list_add_tail(&rxf->ucast_pending_mac->qe, in bna_rxf_uninit()
666 bna_ucam_mod_free_q(rxf->rx->bna)); in bna_rxf_uninit()
667 rxf->ucast_pending_mac = NULL; in bna_rxf_uninit()
670 while (!list_empty(&rxf->mcast_pending_add_q)) { in bna_rxf_uninit()
671 mac = list_first_entry(&rxf->mcast_pending_add_q, in bna_rxf_uninit()
673 list_move_tail(&mac->qe, bna_mcam_mod_free_q(rxf->rx->bna)); in bna_rxf_uninit()
676 rxf->rxmode_pending = 0; in bna_rxf_uninit()
677 rxf->rxmode_pending_bitmask = 0; in bna_rxf_uninit()
678 if (rxf->rx->bna->promisc_rid == rxf->rx->rid) in bna_rxf_uninit()
679 rxf->rx->bna->promisc_rid = BFI_INVALID_RID; in bna_rxf_uninit()
680 if (rxf->rx->bna->default_mode_rid == rxf->rx->rid) in bna_rxf_uninit()
681 rxf->rx->bna->default_mode_rid = BFI_INVALID_RID; in bna_rxf_uninit()
683 rxf->rss_pending = 0; in bna_rxf_uninit()
684 rxf->vlan_strip_pending = false; in bna_rxf_uninit()
686 rxf->rx = NULL; in bna_rxf_uninit()
696 bna_rxf_start(struct bna_rxf *rxf) in bna_rxf_start() argument
698 rxf->start_cbfn = bna_rx_cb_rxf_started; in bna_rxf_start()
699 rxf->start_cbarg = rxf->rx; in bna_rxf_start()
700 bfa_fsm_send_event(rxf, RXF_E_START); in bna_rxf_start()
710 bna_rxf_stop(struct bna_rxf *rxf) in bna_rxf_stop() argument
712 rxf->stop_cbfn = bna_rx_cb_rxf_stopped; in bna_rxf_stop()
713 rxf->stop_cbarg = rxf->rx; in bna_rxf_stop()
714 bfa_fsm_send_event(rxf, RXF_E_STOP); in bna_rxf_stop()
718 bna_rxf_fail(struct bna_rxf *rxf) in bna_rxf_fail() argument
720 bfa_fsm_send_event(rxf, RXF_E_FAIL); in bna_rxf_fail()
726 struct bna_rxf *rxf = &rx->rxf; in bna_rx_ucast_set() local
728 if (rxf->ucast_pending_mac == NULL) { in bna_rx_ucast_set()
729 rxf->ucast_pending_mac = in bna_rx_ucast_set()
730 bna_cam_mod_mac_get(bna_ucam_mod_free_q(rxf->rx->bna)); in bna_rx_ucast_set()
731 if (rxf->ucast_pending_mac == NULL) in bna_rx_ucast_set()
735 ether_addr_copy(rxf->ucast_pending_mac->addr, ucmac); in bna_rx_ucast_set()
736 rxf->ucast_pending_set = 1; in bna_rx_ucast_set()
737 rxf->cam_fltr_cbfn = NULL; in bna_rx_ucast_set()
738 rxf->cam_fltr_cbarg = rx->bna->bnad; in bna_rx_ucast_set()
740 bfa_fsm_send_event(rxf, RXF_E_CONFIG); in bna_rx_ucast_set()
749 struct bna_rxf *rxf = &rx->rxf; in bna_rx_mcast_add() local
753 if (bna_mac_find(&rxf->mcast_active_q, addr) || in bna_rx_mcast_add()
754 bna_mac_find(&rxf->mcast_pending_add_q, addr)) { in bna_rx_mcast_add()
760 mac = bna_cam_mod_mac_get(bna_mcam_mod_free_q(rxf->rx->bna)); in bna_rx_mcast_add()
764 list_add_tail(&mac->qe, &rxf->mcast_pending_add_q); in bna_rx_mcast_add()
766 rxf->cam_fltr_cbfn = cbfn; in bna_rx_mcast_add()
767 rxf->cam_fltr_cbarg = rx->bna->bnad; in bna_rx_mcast_add()
769 bfa_fsm_send_event(rxf, RXF_E_CONFIG); in bna_rx_mcast_add()
778 struct bna_rxf *rxf = &rx->rxf; in bna_rx_ucast_listset() local
785 while (!list_empty(&rxf->ucast_pending_add_q)) { in bna_rx_ucast_listset()
786 mac = list_first_entry(&rxf->ucast_pending_add_q, in bna_rx_ucast_listset()
792 while (!list_empty(&rxf->ucast_active_q)) { in bna_rx_ucast_listset()
793 mac = list_first_entry(&rxf->ucast_active_q, in bna_rx_ucast_listset()
798 list_add_tail(&del_mac->qe, &rxf->ucast_pending_del_q); in bna_rx_ucast_listset()
816 list_move_tail(&mac->qe, &rxf->ucast_pending_add_q); in bna_rx_ucast_listset()
819 bfa_fsm_send_event(rxf, RXF_E_CONFIG); in bna_rx_ucast_listset()
836 struct bna_rxf *rxf = &rx->rxf; in bna_rx_mcast_listset() local
843 while (!list_empty(&rxf->mcast_pending_add_q)) { in bna_rx_mcast_listset()
844 mac = list_first_entry(&rxf->mcast_pending_add_q, in bna_rx_mcast_listset()
850 while (!list_empty(&rxf->mcast_active_q)) { in bna_rx_mcast_listset()
851 mac = list_first_entry(&rxf->mcast_active_q, in bna_rx_mcast_listset()
856 list_add_tail(&del_mac->qe, &rxf->mcast_pending_del_q); in bna_rx_mcast_listset()
876 list_move_tail(&mac->qe, &rxf->mcast_pending_add_q); in bna_rx_mcast_listset()
879 bfa_fsm_send_event(rxf, RXF_E_CONFIG); in bna_rx_mcast_listset()
895 struct bna_rxf *rxf = &rx->rxf; in bna_rx_mcast_delall() local
900 while (!list_empty(&rxf->mcast_pending_add_q)) { in bna_rx_mcast_delall()
901 mac = list_first_entry(&rxf->mcast_pending_add_q, in bna_rx_mcast_delall()
903 list_move_tail(&mac->qe, bna_mcam_mod_free_q(rxf->rx->bna)); in bna_rx_mcast_delall()
907 while (!list_empty(&rxf->mcast_active_q)) { in bna_rx_mcast_delall()
908 mac = list_first_entry(&rxf->mcast_active_q, in bna_rx_mcast_delall()
911 del_mac = bna_cam_mod_mac_get(bna_mcam_mod_del_q(rxf->rx->bna)); in bna_rx_mcast_delall()
913 list_add_tail(&del_mac->qe, &rxf->mcast_pending_del_q); in bna_rx_mcast_delall()
915 list_add_tail(&mac->qe, bna_mcam_mod_free_q(rxf->rx->bna)); in bna_rx_mcast_delall()
920 bfa_fsm_send_event(rxf, RXF_E_CONFIG); in bna_rx_mcast_delall()
926 struct bna_rxf *rxf = &rx->rxf; in bna_rx_vlan_add() local
931 rxf->vlan_filter_table[index] |= bit; in bna_rx_vlan_add()
932 if (rxf->vlan_filter_status == BNA_STATUS_T_ENABLED) { in bna_rx_vlan_add()
933 rxf->vlan_pending_bitmask |= BIT(group_id); in bna_rx_vlan_add()
934 bfa_fsm_send_event(rxf, RXF_E_CONFIG); in bna_rx_vlan_add()
941 struct bna_rxf *rxf = &rx->rxf; in bna_rx_vlan_del() local
946 rxf->vlan_filter_table[index] &= ~bit; in bna_rx_vlan_del()
947 if (rxf->vlan_filter_status == BNA_STATUS_T_ENABLED) { in bna_rx_vlan_del()
948 rxf->vlan_pending_bitmask |= BIT(group_id); in bna_rx_vlan_del()
949 bfa_fsm_send_event(rxf, RXF_E_CONFIG); in bna_rx_vlan_del()
954 bna_rxf_ucast_cfg_apply(struct bna_rxf *rxf) in bna_rxf_ucast_cfg_apply() argument
959 if (!list_empty(&rxf->ucast_pending_del_q)) { in bna_rxf_ucast_cfg_apply()
960 mac = list_first_entry(&rxf->ucast_pending_del_q, in bna_rxf_ucast_cfg_apply()
962 bna_bfi_ucast_req(rxf, mac, BFI_ENET_H2I_MAC_UCAST_DEL_REQ); in bna_rxf_ucast_cfg_apply()
963 list_move_tail(&mac->qe, bna_ucam_mod_del_q(rxf->rx->bna)); in bna_rxf_ucast_cfg_apply()
968 if (rxf->ucast_pending_set) { in bna_rxf_ucast_cfg_apply()
969 rxf->ucast_pending_set = 0; in bna_rxf_ucast_cfg_apply()
970 ether_addr_copy(rxf->ucast_active_mac.addr, in bna_rxf_ucast_cfg_apply()
971 rxf->ucast_pending_mac->addr); in bna_rxf_ucast_cfg_apply()
972 rxf->ucast_active_set = 1; in bna_rxf_ucast_cfg_apply()
973 bna_bfi_ucast_req(rxf, &rxf->ucast_active_mac, in bna_rxf_ucast_cfg_apply()
979 if (!list_empty(&rxf->ucast_pending_add_q)) { in bna_rxf_ucast_cfg_apply()
980 mac = list_first_entry(&rxf->ucast_pending_add_q, in bna_rxf_ucast_cfg_apply()
982 list_move_tail(&mac->qe, &rxf->ucast_active_q); in bna_rxf_ucast_cfg_apply()
983 bna_bfi_ucast_req(rxf, mac, BFI_ENET_H2I_MAC_UCAST_ADD_REQ); in bna_rxf_ucast_cfg_apply()
991 bna_rxf_ucast_cfg_reset(struct bna_rxf *rxf, enum bna_cleanup_type cleanup) in bna_rxf_ucast_cfg_reset() argument
996 while (!list_empty(&rxf->ucast_pending_del_q)) { in bna_rxf_ucast_cfg_reset()
997 mac = list_first_entry(&rxf->ucast_pending_del_q, in bna_rxf_ucast_cfg_reset()
1001 bna_ucam_mod_del_q(rxf->rx->bna)); in bna_rxf_ucast_cfg_reset()
1003 bna_bfi_ucast_req(rxf, mac, in bna_rxf_ucast_cfg_reset()
1006 bna_ucam_mod_del_q(rxf->rx->bna)); in bna_rxf_ucast_cfg_reset()
1012 while (!list_empty(&rxf->ucast_active_q)) { in bna_rxf_ucast_cfg_reset()
1013 mac = list_first_entry(&rxf->ucast_active_q, in bna_rxf_ucast_cfg_reset()
1015 list_move_tail(&mac->qe, &rxf->ucast_pending_add_q); in bna_rxf_ucast_cfg_reset()
1017 bna_bfi_ucast_req(rxf, mac, in bna_rxf_ucast_cfg_reset()
1023 if (rxf->ucast_active_set) { in bna_rxf_ucast_cfg_reset()
1024 rxf->ucast_pending_set = 1; in bna_rxf_ucast_cfg_reset()
1025 rxf->ucast_active_set = 0; in bna_rxf_ucast_cfg_reset()
1027 bna_bfi_ucast_req(rxf, &rxf->ucast_active_mac, in bna_rxf_ucast_cfg_reset()
1037 bna_rxf_promisc_cfg_apply(struct bna_rxf *rxf) in bna_rxf_promisc_cfg_apply() argument
1039 struct bna *bna = rxf->rx->bna; in bna_rxf_promisc_cfg_apply()
1042 if (is_promisc_enable(rxf->rxmode_pending, in bna_rxf_promisc_cfg_apply()
1043 rxf->rxmode_pending_bitmask)) { in bna_rxf_promisc_cfg_apply()
1045 promisc_inactive(rxf->rxmode_pending, in bna_rxf_promisc_cfg_apply()
1046 rxf->rxmode_pending_bitmask); in bna_rxf_promisc_cfg_apply()
1047 rxf->rxmode_active |= BNA_RXMODE_PROMISC; in bna_rxf_promisc_cfg_apply()
1048 bna_bfi_rx_promisc_req(rxf, BNA_STATUS_T_ENABLED); in bna_rxf_promisc_cfg_apply()
1050 } else if (is_promisc_disable(rxf->rxmode_pending, in bna_rxf_promisc_cfg_apply()
1051 rxf->rxmode_pending_bitmask)) { in bna_rxf_promisc_cfg_apply()
1053 promisc_inactive(rxf->rxmode_pending, in bna_rxf_promisc_cfg_apply()
1054 rxf->rxmode_pending_bitmask); in bna_rxf_promisc_cfg_apply()
1055 rxf->rxmode_active &= ~BNA_RXMODE_PROMISC; in bna_rxf_promisc_cfg_apply()
1057 bna_bfi_rx_promisc_req(rxf, BNA_STATUS_T_DISABLED); in bna_rxf_promisc_cfg_apply()
1065 bna_rxf_promisc_cfg_reset(struct bna_rxf *rxf, enum bna_cleanup_type cleanup) in bna_rxf_promisc_cfg_reset() argument
1067 struct bna *bna = rxf->rx->bna; in bna_rxf_promisc_cfg_reset()
1070 if (is_promisc_disable(rxf->rxmode_pending, in bna_rxf_promisc_cfg_reset()
1071 rxf->rxmode_pending_bitmask)) { in bna_rxf_promisc_cfg_reset()
1072 promisc_inactive(rxf->rxmode_pending, in bna_rxf_promisc_cfg_reset()
1073 rxf->rxmode_pending_bitmask); in bna_rxf_promisc_cfg_reset()
1074 rxf->rxmode_active &= ~BNA_RXMODE_PROMISC; in bna_rxf_promisc_cfg_reset()
1077 bna_bfi_rx_promisc_req(rxf, BNA_STATUS_T_DISABLED); in bna_rxf_promisc_cfg_reset()
1083 if (rxf->rxmode_active & BNA_RXMODE_PROMISC) { in bna_rxf_promisc_cfg_reset()
1084 promisc_enable(rxf->rxmode_pending, in bna_rxf_promisc_cfg_reset()
1085 rxf->rxmode_pending_bitmask); in bna_rxf_promisc_cfg_reset()
1086 rxf->rxmode_active &= ~BNA_RXMODE_PROMISC; in bna_rxf_promisc_cfg_reset()
1088 bna_bfi_rx_promisc_req(rxf, BNA_STATUS_T_DISABLED); in bna_rxf_promisc_cfg_reset()
1097 bna_rxf_allmulti_cfg_apply(struct bna_rxf *rxf) in bna_rxf_allmulti_cfg_apply() argument
1100 if (is_allmulti_enable(rxf->rxmode_pending, in bna_rxf_allmulti_cfg_apply()
1101 rxf->rxmode_pending_bitmask)) { in bna_rxf_allmulti_cfg_apply()
1103 allmulti_inactive(rxf->rxmode_pending, in bna_rxf_allmulti_cfg_apply()
1104 rxf->rxmode_pending_bitmask); in bna_rxf_allmulti_cfg_apply()
1105 rxf->rxmode_active |= BNA_RXMODE_ALLMULTI; in bna_rxf_allmulti_cfg_apply()
1106 bna_bfi_mcast_filter_req(rxf, BNA_STATUS_T_DISABLED); in bna_rxf_allmulti_cfg_apply()
1108 } else if (is_allmulti_disable(rxf->rxmode_pending, in bna_rxf_allmulti_cfg_apply()
1109 rxf->rxmode_pending_bitmask)) { in bna_rxf_allmulti_cfg_apply()
1111 allmulti_inactive(rxf->rxmode_pending, in bna_rxf_allmulti_cfg_apply()
1112 rxf->rxmode_pending_bitmask); in bna_rxf_allmulti_cfg_apply()
1113 rxf->rxmode_active &= ~BNA_RXMODE_ALLMULTI; in bna_rxf_allmulti_cfg_apply()
1114 bna_bfi_mcast_filter_req(rxf, BNA_STATUS_T_ENABLED); in bna_rxf_allmulti_cfg_apply()
1122 bna_rxf_allmulti_cfg_reset(struct bna_rxf *rxf, enum bna_cleanup_type cleanup) in bna_rxf_allmulti_cfg_reset() argument
1125 if (is_allmulti_disable(rxf->rxmode_pending, in bna_rxf_allmulti_cfg_reset()
1126 rxf->rxmode_pending_bitmask)) { in bna_rxf_allmulti_cfg_reset()
1127 allmulti_inactive(rxf->rxmode_pending, in bna_rxf_allmulti_cfg_reset()
1128 rxf->rxmode_pending_bitmask); in bna_rxf_allmulti_cfg_reset()
1129 rxf->rxmode_active &= ~BNA_RXMODE_ALLMULTI; in bna_rxf_allmulti_cfg_reset()
1131 bna_bfi_mcast_filter_req(rxf, BNA_STATUS_T_ENABLED); in bna_rxf_allmulti_cfg_reset()
1137 if (rxf->rxmode_active & BNA_RXMODE_ALLMULTI) { in bna_rxf_allmulti_cfg_reset()
1138 allmulti_enable(rxf->rxmode_pending, in bna_rxf_allmulti_cfg_reset()
1139 rxf->rxmode_pending_bitmask); in bna_rxf_allmulti_cfg_reset()
1140 rxf->rxmode_active &= ~BNA_RXMODE_ALLMULTI; in bna_rxf_allmulti_cfg_reset()
1142 bna_bfi_mcast_filter_req(rxf, BNA_STATUS_T_ENABLED); in bna_rxf_allmulti_cfg_reset()
1151 bna_rxf_promisc_enable(struct bna_rxf *rxf) in bna_rxf_promisc_enable() argument
1153 struct bna *bna = rxf->rx->bna; in bna_rxf_promisc_enable()
1156 if (is_promisc_enable(rxf->rxmode_pending, in bna_rxf_promisc_enable()
1157 rxf->rxmode_pending_bitmask) || in bna_rxf_promisc_enable()
1158 (rxf->rxmode_active & BNA_RXMODE_PROMISC)) { in bna_rxf_promisc_enable()
1160 } else if (is_promisc_disable(rxf->rxmode_pending, in bna_rxf_promisc_enable()
1161 rxf->rxmode_pending_bitmask)) { in bna_rxf_promisc_enable()
1163 promisc_inactive(rxf->rxmode_pending, in bna_rxf_promisc_enable()
1164 rxf->rxmode_pending_bitmask); in bna_rxf_promisc_enable()
1167 promisc_enable(rxf->rxmode_pending, in bna_rxf_promisc_enable()
1168 rxf->rxmode_pending_bitmask); in bna_rxf_promisc_enable()
1169 bna->promisc_rid = rxf->rx->rid; in bna_rxf_promisc_enable()
1177 bna_rxf_promisc_disable(struct bna_rxf *rxf) in bna_rxf_promisc_disable() argument
1179 struct bna *bna = rxf->rx->bna; in bna_rxf_promisc_disable()
1182 if (is_promisc_disable(rxf->rxmode_pending, in bna_rxf_promisc_disable()
1183 rxf->rxmode_pending_bitmask) || in bna_rxf_promisc_disable()
1184 (!(rxf->rxmode_active & BNA_RXMODE_PROMISC))) { in bna_rxf_promisc_disable()
1186 } else if (is_promisc_enable(rxf->rxmode_pending, in bna_rxf_promisc_disable()
1187 rxf->rxmode_pending_bitmask)) { in bna_rxf_promisc_disable()
1189 promisc_inactive(rxf->rxmode_pending, in bna_rxf_promisc_disable()
1190 rxf->rxmode_pending_bitmask); in bna_rxf_promisc_disable()
1192 } else if (rxf->rxmode_active & BNA_RXMODE_PROMISC) { in bna_rxf_promisc_disable()
1194 promisc_disable(rxf->rxmode_pending, in bna_rxf_promisc_disable()
1195 rxf->rxmode_pending_bitmask); in bna_rxf_promisc_disable()
1203 bna_rxf_allmulti_enable(struct bna_rxf *rxf) in bna_rxf_allmulti_enable() argument
1207 if (is_allmulti_enable(rxf->rxmode_pending, in bna_rxf_allmulti_enable()
1208 rxf->rxmode_pending_bitmask) || in bna_rxf_allmulti_enable()
1209 (rxf->rxmode_active & BNA_RXMODE_ALLMULTI)) { in bna_rxf_allmulti_enable()
1211 } else if (is_allmulti_disable(rxf->rxmode_pending, in bna_rxf_allmulti_enable()
1212 rxf->rxmode_pending_bitmask)) { in bna_rxf_allmulti_enable()
1214 allmulti_inactive(rxf->rxmode_pending, in bna_rxf_allmulti_enable()
1215 rxf->rxmode_pending_bitmask); in bna_rxf_allmulti_enable()
1218 allmulti_enable(rxf->rxmode_pending, in bna_rxf_allmulti_enable()
1219 rxf->rxmode_pending_bitmask); in bna_rxf_allmulti_enable()
1227 bna_rxf_allmulti_disable(struct bna_rxf *rxf) in bna_rxf_allmulti_disable() argument
1231 if (is_allmulti_disable(rxf->rxmode_pending, in bna_rxf_allmulti_disable()
1232 rxf->rxmode_pending_bitmask) || in bna_rxf_allmulti_disable()
1233 (!(rxf->rxmode_active & BNA_RXMODE_ALLMULTI))) { in bna_rxf_allmulti_disable()
1235 } else if (is_allmulti_enable(rxf->rxmode_pending, in bna_rxf_allmulti_disable()
1236 rxf->rxmode_pending_bitmask)) { in bna_rxf_allmulti_disable()
1238 allmulti_inactive(rxf->rxmode_pending, in bna_rxf_allmulti_disable()
1239 rxf->rxmode_pending_bitmask); in bna_rxf_allmulti_disable()
1240 } else if (rxf->rxmode_active & BNA_RXMODE_ALLMULTI) { in bna_rxf_allmulti_disable()
1242 allmulti_disable(rxf->rxmode_pending, in bna_rxf_allmulti_disable()
1243 rxf->rxmode_pending_bitmask); in bna_rxf_allmulti_disable()
1251 bna_rxf_vlan_strip_cfg_apply(struct bna_rxf *rxf) in bna_rxf_vlan_strip_cfg_apply() argument
1253 if (rxf->vlan_strip_pending) { in bna_rxf_vlan_strip_cfg_apply()
1254 rxf->vlan_strip_pending = false; in bna_rxf_vlan_strip_cfg_apply()
1255 bna_bfi_vlan_strip_enable(rxf); in bna_rxf_vlan_strip_cfg_apply()
1408 bna_rxf_start(&rx->rxf); in bna_rx_sm_rxf_start_wait_entry()
1422 bna_rxf_fail(&rx->rxf); in bna_rx_sm_rxf_stop_wait()
1428 bna_rxf_stop(&rx->rxf); in bna_rx_sm_rxf_stop_wait()
1487 bna_rxf_stop(&rx->rxf); in bna_rx_sm_started()
1493 bna_rxf_fail(&rx->rxf); in bna_rx_sm_started()
1514 bna_rxf_fail(&rx->rxf); in bna_rx_sm_rxf_start_wait()
1707 cfg_req->rx_cfg.strip_vlan = rx->rxf.vlan_strip_status; in bna_bfi_rx_enet_start()
2473 bna_rxf_init(&rx->rxf, rx, rx_cfg, res_info); in bna_rx_create()
2491 bna_rxf_uninit(&rx->rxf); in bna_rx_destroy()
2572 struct bna_rxf *rxf = &rx->rxf; in bna_rx_vlan_strip_enable() local
2574 if (rxf->vlan_strip_status == BNA_STATUS_T_DISABLED) { in bna_rx_vlan_strip_enable()
2575 rxf->vlan_strip_status = BNA_STATUS_T_ENABLED; in bna_rx_vlan_strip_enable()
2576 rxf->vlan_strip_pending = true; in bna_rx_vlan_strip_enable()
2577 bfa_fsm_send_event(rxf, RXF_E_CONFIG); in bna_rx_vlan_strip_enable()
2584 struct bna_rxf *rxf = &rx->rxf; in bna_rx_vlan_strip_disable() local
2586 if (rxf->vlan_strip_status != BNA_STATUS_T_DISABLED) { in bna_rx_vlan_strip_disable()
2587 rxf->vlan_strip_status = BNA_STATUS_T_DISABLED; in bna_rx_vlan_strip_disable()
2588 rxf->vlan_strip_pending = true; in bna_rx_vlan_strip_disable()
2589 bfa_fsm_send_event(rxf, RXF_E_CONFIG); in bna_rx_vlan_strip_disable()
2597 struct bna_rxf *rxf = &rx->rxf; in bna_rx_mode_set() local
2605 (rx->bna->promisc_rid != rxf->rx->rid)) in bna_rx_mode_set()
2620 (rx->bna->default_mode_rid != rxf->rx->rid)) { in bna_rx_mode_set()
2632 if (bna_rxf_promisc_enable(rxf)) in bna_rx_mode_set()
2635 if (bna_rxf_promisc_disable(rxf)) in bna_rx_mode_set()
2640 if (bna_rxf_allmulti_enable(rxf)) in bna_rx_mode_set()
2643 if (bna_rxf_allmulti_disable(rxf)) in bna_rx_mode_set()
2650 rxf->cam_fltr_cbfn = NULL; in bna_rx_mode_set()
2651 rxf->cam_fltr_cbarg = rx->bna->bnad; in bna_rx_mode_set()
2652 bfa_fsm_send_event(rxf, RXF_E_CONFIG); in bna_rx_mode_set()
2664 struct bna_rxf *rxf = &rx->rxf; in bna_rx_vlanfilter_enable() local
2666 if (rxf->vlan_filter_status == BNA_STATUS_T_DISABLED) { in bna_rx_vlanfilter_enable()
2667 rxf->vlan_filter_status = BNA_STATUS_T_ENABLED; in bna_rx_vlanfilter_enable()
2668 rxf->vlan_pending_bitmask = (u8)BFI_VLAN_BMASK_ALL; in bna_rx_vlanfilter_enable()
2669 bfa_fsm_send_event(rxf, RXF_E_CONFIG); in bna_rx_vlanfilter_enable()