1 /* 2 * Copyright (c) 2015-2021 The Linux Foundation. All rights reserved. 3 * Copyright (c) 2021-2024 Qualcomm Innovation Center, Inc. All rights reserved. 4 * 5 * Permission to use, copy, modify, and/or distribute this software for 6 * any purpose with or without fee is hereby granted, provided that the 7 * above copyright notice and this permission notice appear in all 8 * copies. 9 * 10 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL 11 * WARRANTIES WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED 12 * WARRANTIES OF MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE 13 * AUTHOR BE LIABLE FOR ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL 14 * DAMAGES OR ANY DAMAGES WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR 15 * PROFITS, WHETHER IN AN ACTION OF CONTRACT, NEGLIGENCE OR OTHER 16 * TORTIOUS ACTION, ARISING OUT OF OR IN CONNECTION WITH THE USE OR 17 * PERFORMANCE OF THIS SOFTWARE. 18 */ 19 20 #include "targcfg.h" 21 #include "qdf_lock.h" 22 #include "qdf_status.h" 23 #include "qdf_status.h" 24 #include <qdf_atomic.h> /* qdf_atomic_read */ 25 #include <targaddrs.h> 26 #include "hif_io32.h" 27 #include <hif.h> 28 #include <target_type.h> 29 #include "regtable.h" 30 #define ATH_MODULE_NAME hif 31 #include <a_debug.h> 32 #include "hif_main.h" 33 #include "hif_hw_version.h" 34 #if (defined(HIF_PCI) || defined(HIF_SNOC) || defined(HIF_AHB) || \ 35 defined(HIF_IPCI)) 36 #include "ce_tasklet.h" 37 #include "ce_api.h" 38 #endif 39 #include "qdf_trace.h" 40 #include "qdf_status.h" 41 #include "hif_debug.h" 42 #include "mp_dev.h" 43 #if defined(QCA_WIFI_QCA8074) || defined(QCA_WIFI_QCA6018) || \ 44 defined(QCA_WIFI_QCA5018) || defined(QCA_WIFI_QCA9574) || \ 45 defined(QCA_WIFI_QCA5332) 46 #include "hal_api.h" 47 #endif 48 #include "hif_napi.h" 49 #include "hif_unit_test_suspend_i.h" 50 #include "qdf_module.h" 51 #ifdef HIF_CE_LOG_INFO 52 #include <qdf_notifier.h> 53 #include <qdf_hang_event_notifier.h> 54 #endif 55 #include <linux/cpumask.h> 56 57 #include <pld_common.h> 58 #include "ce_internal.h" 59 #include <qdf_tracepoint.h> 60 #include "qdf_ssr_driver_dump.h" 61 hif_dump(struct hif_opaque_softc * hif_ctx,uint8_t cmd_id,bool start)62 void hif_dump(struct hif_opaque_softc *hif_ctx, uint8_t cmd_id, bool start) 63 { 64 hif_trigger_dump(hif_ctx, cmd_id, start); 65 } 66 67 /** 68 * hif_get_target_id(): hif_get_target_id 69 * @scn: scn 70 * 71 * Return the virtual memory base address to the caller 72 * 73 * @scn: hif_softc 74 * 75 * Return: A_target_id_t 76 */ hif_get_target_id(struct hif_softc * scn)77 A_target_id_t hif_get_target_id(struct hif_softc *scn) 78 { 79 return scn->mem; 80 } 81 82 /** 83 * hif_get_targetdef(): hif_get_targetdef 84 * @hif_ctx: hif context 85 * 86 * Return: void * 87 */ hif_get_targetdef(struct hif_opaque_softc * hif_ctx)88 void *hif_get_targetdef(struct hif_opaque_softc *hif_ctx) 89 { 90 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 91 92 return scn->targetdef; 93 } 94 95 #ifdef FORCE_WAKE 96 #ifndef QCA_WIFI_WCN6450 hif_srng_init_phase(struct hif_opaque_softc * hif_ctx,bool init_phase)97 void hif_srng_init_phase(struct hif_opaque_softc *hif_ctx, 98 bool init_phase) 99 { 100 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 101 102 if (ce_srng_based(scn)) 103 hal_set_init_phase(scn->hal_soc, init_phase); 104 } 105 #else hif_srng_init_phase(struct hif_opaque_softc * hif_ctx,bool init_phase)106 void hif_srng_init_phase(struct hif_opaque_softc *hif_ctx, 107 bool init_phase) 108 { 109 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 110 111 hal_set_init_phase(scn->hal_soc, init_phase); 112 } 113 #endif 114 #endif /* FORCE_WAKE */ 115 116 #ifdef HIF_IPCI hif_shutdown_notifier_cb(void * hif_ctx)117 void hif_shutdown_notifier_cb(void *hif_ctx) 118 { 119 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 120 121 scn->recovery = true; 122 } 123 #endif 124 125 /** 126 * hif_vote_link_down(): unvote for link up 127 * @hif_ctx: hif context 128 * 129 * Call hif_vote_link_down to release a previous request made using 130 * hif_vote_link_up. A hif_vote_link_down call should only be made 131 * after a corresponding hif_vote_link_up, otherwise you could be 132 * negating a vote from another source. When no votes are present 133 * hif will not guarantee the linkstate after hif_bus_suspend. 134 * 135 * SYNCHRONIZE WITH hif_vote_link_up by only calling in MC thread 136 * and initialization deinitialization sequencences. 137 * 138 * Return: n/a 139 */ hif_vote_link_down(struct hif_opaque_softc * hif_ctx)140 void hif_vote_link_down(struct hif_opaque_softc *hif_ctx) 141 { 142 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 143 144 QDF_BUG(scn); 145 if (scn->linkstate_vote == 0) 146 QDF_DEBUG_PANIC("linkstate_vote(%d) has already been 0", 147 scn->linkstate_vote); 148 149 scn->linkstate_vote--; 150 hif_info("Down_linkstate_vote %d", scn->linkstate_vote); 151 if (scn->linkstate_vote == 0) 152 hif_bus_prevent_linkdown(scn, false); 153 } 154 155 /** 156 * hif_vote_link_up(): vote to prevent bus from suspending 157 * @hif_ctx: hif context 158 * 159 * Makes hif guarantee that fw can message the host normally 160 * during suspend. 161 * 162 * SYNCHRONIZE WITH hif_vote_link_up by only calling in MC thread 163 * and initialization deinitialization sequencences. 164 * 165 * Return: n/a 166 */ hif_vote_link_up(struct hif_opaque_softc * hif_ctx)167 void hif_vote_link_up(struct hif_opaque_softc *hif_ctx) 168 { 169 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 170 171 QDF_BUG(scn); 172 scn->linkstate_vote++; 173 hif_info("Up_linkstate_vote %d", scn->linkstate_vote); 174 if (scn->linkstate_vote == 1) 175 hif_bus_prevent_linkdown(scn, true); 176 } 177 178 /** 179 * hif_can_suspend_link(): query if hif is permitted to suspend the link 180 * @hif_ctx: hif context 181 * 182 * Hif will ensure that the link won't be suspended if the upperlayers 183 * don't want it to. 184 * 185 * SYNCHRONIZATION: MC thread is stopped before bus suspend thus 186 * we don't need extra locking to ensure votes dont change while 187 * we are in the process of suspending or resuming. 188 * 189 * Return: false if hif will guarantee link up during suspend. 190 */ hif_can_suspend_link(struct hif_opaque_softc * hif_ctx)191 bool hif_can_suspend_link(struct hif_opaque_softc *hif_ctx) 192 { 193 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 194 195 QDF_BUG(scn); 196 return scn->linkstate_vote == 0; 197 } 198 199 /** 200 * hif_hia_item_address(): hif_hia_item_address 201 * @target_type: target_type 202 * @item_offset: item_offset 203 * 204 * Return: n/a 205 */ hif_hia_item_address(uint32_t target_type,uint32_t item_offset)206 uint32_t hif_hia_item_address(uint32_t target_type, uint32_t item_offset) 207 { 208 switch (target_type) { 209 case TARGET_TYPE_AR6002: 210 return AR6002_HOST_INTEREST_ADDRESS + item_offset; 211 case TARGET_TYPE_AR6003: 212 return AR6003_HOST_INTEREST_ADDRESS + item_offset; 213 case TARGET_TYPE_AR6004: 214 return AR6004_HOST_INTEREST_ADDRESS + item_offset; 215 case TARGET_TYPE_AR6006: 216 return AR6006_HOST_INTEREST_ADDRESS + item_offset; 217 case TARGET_TYPE_AR9888: 218 return AR9888_HOST_INTEREST_ADDRESS + item_offset; 219 case TARGET_TYPE_AR6320: 220 case TARGET_TYPE_AR6320V2: 221 return AR6320_HOST_INTEREST_ADDRESS + item_offset; 222 case TARGET_TYPE_ADRASTEA: 223 /* ADRASTEA doesn't have a host interest address */ 224 ASSERT(0); 225 return 0; 226 case TARGET_TYPE_AR900B: 227 return AR900B_HOST_INTEREST_ADDRESS + item_offset; 228 case TARGET_TYPE_QCA9984: 229 return QCA9984_HOST_INTEREST_ADDRESS + item_offset; 230 case TARGET_TYPE_QCA9888: 231 return QCA9888_HOST_INTEREST_ADDRESS + item_offset; 232 233 default: 234 ASSERT(0); 235 return 0; 236 } 237 } 238 239 /** 240 * hif_max_num_receives_reached() - check max receive is reached 241 * @scn: HIF Context 242 * @count: unsigned int. 243 * 244 * Output check status as bool 245 * 246 * Return: bool 247 */ hif_max_num_receives_reached(struct hif_softc * scn,unsigned int count)248 bool hif_max_num_receives_reached(struct hif_softc *scn, unsigned int count) 249 { 250 if (QDF_IS_EPPING_ENABLED(hif_get_conparam(scn))) 251 return count > 120; 252 else 253 return count > MAX_NUM_OF_RECEIVES; 254 } 255 256 /** 257 * init_buffer_count() - initial buffer count 258 * @maxSize: qdf_size_t 259 * 260 * routine to modify the initial buffer count to be allocated on an os 261 * platform basis. Platform owner will need to modify this as needed 262 * 263 * Return: qdf_size_t 264 */ init_buffer_count(qdf_size_t maxSize)265 qdf_size_t init_buffer_count(qdf_size_t maxSize) 266 { 267 return maxSize; 268 } 269 270 /** 271 * hif_save_htc_htt_config_endpoint() - save htt_tx_endpoint 272 * @hif_ctx: hif context 273 * @htc_htt_tx_endpoint: htt_tx_endpoint 274 * 275 * Return: void 276 */ hif_save_htc_htt_config_endpoint(struct hif_opaque_softc * hif_ctx,int htc_htt_tx_endpoint)277 void hif_save_htc_htt_config_endpoint(struct hif_opaque_softc *hif_ctx, 278 int htc_htt_tx_endpoint) 279 { 280 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 281 282 if (!scn) { 283 hif_err("scn or scn->hif_sc is NULL!"); 284 return; 285 } 286 287 scn->htc_htt_tx_endpoint = htc_htt_tx_endpoint; 288 } 289 qdf_export_symbol(hif_save_htc_htt_config_endpoint); 290 291 static const struct qwlan_hw qwlan_hw_list[] = { 292 { 293 .id = AR6320_REV1_VERSION, 294 .subid = 0, 295 .name = "QCA6174_REV1", 296 }, 297 { 298 .id = AR6320_REV1_1_VERSION, 299 .subid = 0x1, 300 .name = "QCA6174_REV1_1", 301 }, 302 { 303 .id = AR6320_REV1_3_VERSION, 304 .subid = 0x2, 305 .name = "QCA6174_REV1_3", 306 }, 307 { 308 .id = AR6320_REV2_1_VERSION, 309 .subid = 0x4, 310 .name = "QCA6174_REV2_1", 311 }, 312 { 313 .id = AR6320_REV2_1_VERSION, 314 .subid = 0x5, 315 .name = "QCA6174_REV2_2", 316 }, 317 { 318 .id = AR6320_REV3_VERSION, 319 .subid = 0x6, 320 .name = "QCA6174_REV2.3", 321 }, 322 { 323 .id = AR6320_REV3_VERSION, 324 .subid = 0x8, 325 .name = "QCA6174_REV3", 326 }, 327 { 328 .id = AR6320_REV3_VERSION, 329 .subid = 0x9, 330 .name = "QCA6174_REV3_1", 331 }, 332 { 333 .id = AR6320_REV3_2_VERSION, 334 .subid = 0xA, 335 .name = "AR6320_REV3_2_VERSION", 336 }, 337 { 338 .id = QCA6390_V1, 339 .subid = 0x0, 340 .name = "QCA6390_V1", 341 }, 342 { 343 .id = QCA6490_V1, 344 .subid = 0x0, 345 .name = "QCA6490_V1", 346 }, 347 { 348 .id = WCN3990_v1, 349 .subid = 0x0, 350 .name = "WCN3990_V1", 351 }, 352 { 353 .id = WCN3990_v2, 354 .subid = 0x0, 355 .name = "WCN3990_V2", 356 }, 357 { 358 .id = WCN3990_v2_1, 359 .subid = 0x0, 360 .name = "WCN3990_V2.1", 361 }, 362 { 363 .id = WCN3998, 364 .subid = 0x0, 365 .name = "WCN3998", 366 }, 367 { 368 .id = QCA9379_REV1_VERSION, 369 .subid = 0xC, 370 .name = "QCA9379_REV1", 371 }, 372 { 373 .id = QCA9379_REV1_VERSION, 374 .subid = 0xD, 375 .name = "QCA9379_REV1_1", 376 }, 377 { 378 .id = MANGO_V1, 379 .subid = 0xF, 380 .name = "MANGO_V1", 381 }, 382 { 383 .id = PEACH_V1, 384 .subid = 0, 385 .name = "PEACH_V1", 386 }, 387 388 { 389 .id = KIWI_V1, 390 .subid = 0, 391 .name = "KIWI_V1", 392 }, 393 { 394 .id = KIWI_V2, 395 .subid = 0, 396 .name = "KIWI_V2", 397 }, 398 { 399 .id = WCN6750_V1, 400 .subid = 0, 401 .name = "WCN6750_V1", 402 }, 403 { 404 .id = WCN6750_V2, 405 .subid = 0, 406 .name = "WCN6750_V2", 407 }, 408 { 409 .id = WCN6450_V1, 410 .subid = 0, 411 .name = "WCN6450_V1", 412 }, 413 { 414 .id = QCA6490_v2_1, 415 .subid = 0, 416 .name = "QCA6490", 417 }, 418 { 419 .id = QCA6490_v2, 420 .subid = 0, 421 .name = "QCA6490", 422 }, 423 { 424 .id = WCN3990_TALOS, 425 .subid = 0, 426 .name = "WCN3990", 427 }, 428 { 429 .id = WCN3990_MOOREA, 430 .subid = 0, 431 .name = "WCN3990", 432 }, 433 { 434 .id = WCN3990_SAIPAN, 435 .subid = 0, 436 .name = "WCN3990", 437 }, 438 { 439 .id = WCN3990_RENNELL, 440 .subid = 0, 441 .name = "WCN3990", 442 }, 443 { 444 .id = WCN3990_BITRA, 445 .subid = 0, 446 .name = "WCN3990", 447 }, 448 { 449 .id = WCN3990_DIVAR, 450 .subid = 0, 451 .name = "WCN3990", 452 }, 453 { 454 .id = WCN3990_ATHERTON, 455 .subid = 0, 456 .name = "WCN3990", 457 }, 458 { 459 .id = WCN3990_STRAIT, 460 .subid = 0, 461 .name = "WCN3990", 462 }, 463 { 464 .id = WCN3990_NETRANI, 465 .subid = 0, 466 .name = "WCN3990", 467 }, 468 { 469 .id = WCN3990_CLARENCE, 470 .subid = 0, 471 .name = "WCN3990", 472 } 473 }; 474 475 /** 476 * hif_get_hw_name(): get a human readable name for the hardware 477 * @info: Target Info 478 * 479 * Return: human readable name for the underlying wifi hardware. 480 */ hif_get_hw_name(struct hif_target_info * info)481 static const char *hif_get_hw_name(struct hif_target_info *info) 482 { 483 int i; 484 485 hif_debug("target version = %d, target revision = %d", 486 info->target_version, 487 info->target_revision); 488 489 if (info->hw_name) 490 return info->hw_name; 491 492 for (i = 0; i < ARRAY_SIZE(qwlan_hw_list); i++) { 493 if (info->target_version == qwlan_hw_list[i].id && 494 info->target_revision == qwlan_hw_list[i].subid) { 495 return qwlan_hw_list[i].name; 496 } 497 } 498 499 info->hw_name = qdf_mem_malloc(64); 500 if (!info->hw_name) 501 return "Unknown Device (nomem)"; 502 503 i = qdf_snprint(info->hw_name, 64, "HW_VERSION=%x.", 504 info->target_version); 505 if (i < 0) 506 return "Unknown Device (snprintf failure)"; 507 else 508 return info->hw_name; 509 } 510 511 /** 512 * hif_get_hw_info(): hif_get_hw_info 513 * @scn: scn 514 * @version: version 515 * @revision: revision 516 * @target_name: target name 517 * 518 * Return: n/a 519 */ hif_get_hw_info(struct hif_opaque_softc * scn,u32 * version,u32 * revision,const char ** target_name)520 void hif_get_hw_info(struct hif_opaque_softc *scn, u32 *version, u32 *revision, 521 const char **target_name) 522 { 523 struct hif_target_info *info = hif_get_target_info_handle(scn); 524 struct hif_softc *sc = HIF_GET_SOFTC(scn); 525 526 if (sc->bus_type == QDF_BUS_TYPE_USB) 527 hif_usb_get_hw_info(sc); 528 529 *version = info->target_version; 530 *revision = info->target_revision; 531 *target_name = hif_get_hw_name(info); 532 } 533 534 /** 535 * hif_get_dev_ba(): API to get device base address. 536 * @hif_handle: hif handle 537 * 538 * Return: device base address 539 */ hif_get_dev_ba(struct hif_opaque_softc * hif_handle)540 void *hif_get_dev_ba(struct hif_opaque_softc *hif_handle) 541 { 542 struct hif_softc *scn = (struct hif_softc *)hif_handle; 543 544 return scn->mem; 545 } 546 qdf_export_symbol(hif_get_dev_ba); 547 548 /** 549 * hif_get_dev_ba_ce(): API to get device ce base address. 550 * @hif_handle: hif handle 551 * 552 * Return: dev mem base address for CE 553 */ hif_get_dev_ba_ce(struct hif_opaque_softc * hif_handle)554 void *hif_get_dev_ba_ce(struct hif_opaque_softc *hif_handle) 555 { 556 struct hif_softc *scn = (struct hif_softc *)hif_handle; 557 558 return scn->mem_ce; 559 } 560 561 qdf_export_symbol(hif_get_dev_ba_ce); 562 563 /** 564 * hif_get_dev_ba_pmm(): API to get device pmm base address. 565 * @hif_handle: scn 566 * 567 * Return: dev mem base address for PMM 568 */ 569 hif_get_dev_ba_pmm(struct hif_opaque_softc * hif_handle)570 void *hif_get_dev_ba_pmm(struct hif_opaque_softc *hif_handle) 571 { 572 struct hif_softc *scn = (struct hif_softc *)hif_handle; 573 574 return scn->mem_pmm_base; 575 } 576 577 qdf_export_symbol(hif_get_dev_ba_pmm); 578 hif_get_soc_version(struct hif_opaque_softc * hif_handle)579 uint32_t hif_get_soc_version(struct hif_opaque_softc *hif_handle) 580 { 581 struct hif_softc *scn = (struct hif_softc *)hif_handle; 582 583 return scn->target_info.soc_version; 584 } 585 586 qdf_export_symbol(hif_get_soc_version); 587 588 /** 589 * hif_get_dev_ba_cmem(): API to get device ce base address. 590 * @hif_handle: hif handle 591 * 592 * Return: dev mem base address for CMEM 593 */ hif_get_dev_ba_cmem(struct hif_opaque_softc * hif_handle)594 void *hif_get_dev_ba_cmem(struct hif_opaque_softc *hif_handle) 595 { 596 struct hif_softc *scn = (struct hif_softc *)hif_handle; 597 598 return scn->mem_cmem; 599 } 600 601 qdf_export_symbol(hif_get_dev_ba_cmem); 602 603 #ifdef FEATURE_RUNTIME_PM hif_runtime_prevent_linkdown(struct hif_softc * scn,bool is_get)604 void hif_runtime_prevent_linkdown(struct hif_softc *scn, bool is_get) 605 { 606 if (is_get) 607 qdf_runtime_pm_prevent_suspend(&scn->prevent_linkdown_lock); 608 else 609 qdf_runtime_pm_allow_suspend(&scn->prevent_linkdown_lock); 610 } 611 612 static inline hif_rtpm_lock_init(struct hif_softc * scn)613 void hif_rtpm_lock_init(struct hif_softc *scn) 614 { 615 qdf_runtime_lock_init(&scn->prevent_linkdown_lock); 616 } 617 618 static inline hif_rtpm_lock_deinit(struct hif_softc * scn)619 void hif_rtpm_lock_deinit(struct hif_softc *scn) 620 { 621 qdf_runtime_lock_deinit(&scn->prevent_linkdown_lock); 622 } 623 #else 624 static inline hif_rtpm_lock_init(struct hif_softc * scn)625 void hif_rtpm_lock_init(struct hif_softc *scn) 626 { 627 } 628 629 static inline hif_rtpm_lock_deinit(struct hif_softc * scn)630 void hif_rtpm_lock_deinit(struct hif_softc *scn) 631 { 632 } 633 #endif 634 635 #ifdef WLAN_CE_INTERRUPT_THRESHOLD_CONFIG 636 /** 637 * hif_get_interrupt_threshold_cfg_from_psoc() - Retrieve ini cfg from psoc 638 * @scn: hif context 639 * @psoc: psoc objmgr handle 640 * 641 * Return: None 642 */ 643 static inline hif_get_interrupt_threshold_cfg_from_psoc(struct hif_softc * scn,struct wlan_objmgr_psoc * psoc)644 void hif_get_interrupt_threshold_cfg_from_psoc(struct hif_softc *scn, 645 struct wlan_objmgr_psoc *psoc) 646 { 647 if (psoc) { 648 scn->ini_cfg.ce_status_ring_timer_threshold = 649 cfg_get(psoc, 650 CFG_CE_STATUS_RING_TIMER_THRESHOLD); 651 scn->ini_cfg.ce_status_ring_batch_count_threshold = 652 cfg_get(psoc, 653 CFG_CE_STATUS_RING_BATCH_COUNT_THRESHOLD); 654 } 655 } 656 #else 657 static inline hif_get_interrupt_threshold_cfg_from_psoc(struct hif_softc * scn,struct wlan_objmgr_psoc * psoc)658 void hif_get_interrupt_threshold_cfg_from_psoc(struct hif_softc *scn, 659 struct wlan_objmgr_psoc *psoc) 660 { 661 } 662 #endif /* WLAN_CE_INTERRUPT_THRESHOLD_CONFIG */ 663 664 /** 665 * hif_get_cfg_from_psoc() - Retrieve ini cfg from psoc 666 * @scn: hif context 667 * @psoc: psoc objmgr handle 668 * 669 * Return: None 670 */ 671 static inline hif_get_cfg_from_psoc(struct hif_softc * scn,struct wlan_objmgr_psoc * psoc)672 void hif_get_cfg_from_psoc(struct hif_softc *scn, 673 struct wlan_objmgr_psoc *psoc) 674 { 675 if (psoc) { 676 scn->ini_cfg.disable_wake_irq = 677 cfg_get(psoc, CFG_DISABLE_WAKE_IRQ); 678 /** 679 * Wake IRQ can't share the same IRQ with the copy engines 680 * In one MSI mode, we don't know whether wake IRQ is triggered 681 * or not in wake IRQ handler. known issue CR 2055359 682 * If you want to support Wake IRQ. Please allocate at least 683 * 2 MSI vector. The first is for wake IRQ while the others 684 * share the second vector 685 */ 686 if (pld_is_one_msi(scn->qdf_dev->dev)) { 687 hif_debug("Disable wake IRQ once it is one MSI mode"); 688 scn->ini_cfg.disable_wake_irq = true; 689 } 690 hif_get_interrupt_threshold_cfg_from_psoc(scn, psoc); 691 } 692 } 693 694 #if defined(HIF_CE_LOG_INFO) || defined(HIF_BUS_LOG_INFO) 695 /** 696 * hif_recovery_notifier_cb - Recovery notifier callback to log 697 * hang event data 698 * @block: notifier block 699 * @state: state 700 * @data: notifier data 701 * 702 * Return: status 703 */ 704 static hif_recovery_notifier_cb(struct notifier_block * block,unsigned long state,void * data)705 int hif_recovery_notifier_cb(struct notifier_block *block, unsigned long state, 706 void *data) 707 { 708 struct qdf_notifer_data *notif_data = data; 709 qdf_notif_block *notif_block; 710 struct hif_softc *hif_handle; 711 bool bus_id_invalid; 712 713 if (!data || !block) 714 return -EINVAL; 715 716 notif_block = qdf_container_of(block, qdf_notif_block, notif_block); 717 718 hif_handle = notif_block->priv_data; 719 if (!hif_handle) 720 return -EINVAL; 721 722 bus_id_invalid = hif_log_bus_info(hif_handle, notif_data->hang_data, 723 ¬if_data->offset); 724 if (bus_id_invalid) 725 return NOTIFY_STOP_MASK; 726 727 hif_log_ce_info(hif_handle, notif_data->hang_data, 728 ¬if_data->offset); 729 730 return 0; 731 } 732 733 /** 734 * hif_register_recovery_notifier - Register hif recovery notifier 735 * @hif_handle: hif handle 736 * 737 * Return: status 738 */ 739 static hif_register_recovery_notifier(struct hif_softc * hif_handle)740 QDF_STATUS hif_register_recovery_notifier(struct hif_softc *hif_handle) 741 { 742 qdf_notif_block *hif_notifier; 743 744 if (!hif_handle) 745 return QDF_STATUS_E_FAILURE; 746 747 hif_notifier = &hif_handle->hif_recovery_notifier; 748 749 hif_notifier->notif_block.notifier_call = hif_recovery_notifier_cb; 750 hif_notifier->priv_data = hif_handle; 751 return qdf_hang_event_register_notifier(hif_notifier); 752 } 753 754 /** 755 * hif_unregister_recovery_notifier - Un-register hif recovery notifier 756 * @hif_handle: hif handle 757 * 758 * Return: status 759 */ 760 static hif_unregister_recovery_notifier(struct hif_softc * hif_handle)761 QDF_STATUS hif_unregister_recovery_notifier(struct hif_softc *hif_handle) 762 { 763 qdf_notif_block *hif_notifier = &hif_handle->hif_recovery_notifier; 764 765 return qdf_hang_event_unregister_notifier(hif_notifier); 766 } 767 #else 768 static inline hif_register_recovery_notifier(struct hif_softc * hif_handle)769 QDF_STATUS hif_register_recovery_notifier(struct hif_softc *hif_handle) 770 { 771 return QDF_STATUS_SUCCESS; 772 } 773 774 static inline hif_unregister_recovery_notifier(struct hif_softc * hif_handle)775 QDF_STATUS hif_unregister_recovery_notifier(struct hif_softc *hif_handle) 776 { 777 return QDF_STATUS_SUCCESS; 778 } 779 #endif 780 781 #if defined(HIF_CPU_PERF_AFFINE_MASK) || \ 782 defined(FEATURE_ENABLE_CE_DP_IRQ_AFFINE) 783 /** 784 * __hif_cpu_hotplug_notify() - CPU hotplug event handler 785 * @context: HIF context 786 * @cpu: CPU Id of the CPU generating the event 787 * @cpu_up: true if the CPU is online 788 * 789 * Return: None 790 */ __hif_cpu_hotplug_notify(void * context,uint32_t cpu,bool cpu_up)791 static void __hif_cpu_hotplug_notify(void *context, 792 uint32_t cpu, bool cpu_up) 793 { 794 struct hif_softc *scn = context; 795 796 if (!scn) 797 return; 798 if (hif_is_driver_unloading(scn) || hif_is_recovery_in_progress(scn)) 799 return; 800 801 if (cpu_up) { 802 hif_config_irq_set_perf_affinity_hint(GET_HIF_OPAQUE_HDL(scn)); 803 hif_debug("Setting affinity for online CPU: %d", cpu); 804 } else { 805 hif_debug("Skip setting affinity for offline CPU: %d", cpu); 806 } 807 } 808 809 /** 810 * hif_cpu_hotplug_notify - cpu core up/down notification 811 * handler 812 * @context: HIF context 813 * @cpu: CPU generating the event 814 * @cpu_up: true if the CPU is online 815 * 816 * Return: None 817 */ hif_cpu_hotplug_notify(void * context,uint32_t cpu,bool cpu_up)818 static void hif_cpu_hotplug_notify(void *context, uint32_t cpu, bool cpu_up) 819 { 820 struct qdf_op_sync *op_sync; 821 822 if (qdf_op_protect(&op_sync)) 823 return; 824 825 __hif_cpu_hotplug_notify(context, cpu, cpu_up); 826 827 qdf_op_unprotect(op_sync); 828 } 829 hif_cpu_online_cb(void * context,uint32_t cpu)830 static void hif_cpu_online_cb(void *context, uint32_t cpu) 831 { 832 hif_cpu_hotplug_notify(context, cpu, true); 833 } 834 hif_cpu_before_offline_cb(void * context,uint32_t cpu)835 static void hif_cpu_before_offline_cb(void *context, uint32_t cpu) 836 { 837 hif_cpu_hotplug_notify(context, cpu, false); 838 } 839 hif_cpuhp_register(struct hif_softc * scn)840 static void hif_cpuhp_register(struct hif_softc *scn) 841 { 842 if (!scn) { 843 hif_info_high("cannot register hotplug notifiers"); 844 return; 845 } 846 qdf_cpuhp_register(&scn->cpuhp_event_handle, 847 scn, 848 hif_cpu_online_cb, 849 hif_cpu_before_offline_cb); 850 } 851 hif_cpuhp_unregister(struct hif_softc * scn)852 static void hif_cpuhp_unregister(struct hif_softc *scn) 853 { 854 if (!scn) { 855 hif_info_high("cannot unregister hotplug notifiers"); 856 return; 857 } 858 qdf_cpuhp_unregister(&scn->cpuhp_event_handle); 859 } 860 861 #else hif_cpuhp_register(struct hif_softc * scn)862 static void hif_cpuhp_register(struct hif_softc *scn) 863 { 864 } 865 hif_cpuhp_unregister(struct hif_softc * scn)866 static void hif_cpuhp_unregister(struct hif_softc *scn) 867 { 868 } 869 #endif /* ifdef HIF_CPU_PERF_AFFINE_MASK */ 870 871 #ifdef HIF_DETECTION_LATENCY_ENABLE 872 /* 873 * Bitmask to control enablement of latency detection for the tasklets, 874 * bit-X represents for tasklet of WLAN_CE_X. 875 */ 876 #ifndef DETECTION_LATENCY_TASKLET_MASK 877 #define DETECTION_LATENCY_TASKLET_MASK (BIT(2) | BIT(7)) 878 #endif 879 880 static inline int __hif_tasklet_latency(struct hif_softc * scn,bool from_timer,int idx)881 __hif_tasklet_latency(struct hif_softc *scn, bool from_timer, int idx) 882 { 883 qdf_time_t sched_time = 884 scn->latency_detect.tasklet_info[idx].sched_time; 885 qdf_time_t exec_time = 886 scn->latency_detect.tasklet_info[idx].exec_time; 887 qdf_time_t curr_time = qdf_system_ticks(); 888 uint32_t threshold = scn->latency_detect.threshold; 889 qdf_time_t expect_exec_time = 890 sched_time + qdf_system_msecs_to_ticks(threshold); 891 892 /* 2 kinds of check here. 893 * from_timer==true: check if tasklet stall 894 * from_timer==false: check tasklet execute comes late 895 */ 896 if (from_timer ? 897 (qdf_system_time_after(sched_time, exec_time) && 898 qdf_system_time_after(curr_time, expect_exec_time)) : 899 qdf_system_time_after(exec_time, expect_exec_time)) { 900 hif_err("tasklet[%d] latency detected: from_timer %d, curr_time %lu, sched_time %lu, exec_time %lu, threshold %ums, timeout %ums, cpu_id %d, called: %ps", 901 idx, from_timer, curr_time, sched_time, 902 exec_time, threshold, 903 scn->latency_detect.timeout, 904 qdf_get_cpu(), (void *)_RET_IP_); 905 qdf_trigger_self_recovery(NULL, 906 QDF_TASKLET_CREDIT_LATENCY_DETECT); 907 return -ETIMEDOUT; 908 } 909 910 return 0; 911 } 912 913 /** 914 * hif_tasklet_latency_detect_enabled() - check whether latency detect 915 * is enabled for the tasklet which is specified by idx 916 * @scn: HIF opaque context 917 * @idx: CE id 918 * 919 * Return: true if latency detect is enabled for the specified tasklet, 920 * false otherwise. 921 */ 922 static inline bool hif_tasklet_latency_detect_enabled(struct hif_softc * scn,int idx)923 hif_tasklet_latency_detect_enabled(struct hif_softc *scn, int idx) 924 { 925 if (QDF_GLOBAL_MISSION_MODE != hif_get_conparam(scn)) 926 return false; 927 928 if (!scn->latency_detect.enable_detection) 929 return false; 930 931 if (idx < 0 || idx >= HIF_TASKLET_IN_MONITOR || 932 !qdf_test_bit(idx, scn->latency_detect.tasklet_bmap)) 933 return false; 934 935 return true; 936 } 937 hif_tasklet_latency_record_exec(struct hif_softc * scn,int idx)938 void hif_tasklet_latency_record_exec(struct hif_softc *scn, int idx) 939 { 940 if (!hif_tasklet_latency_detect_enabled(scn, idx)) 941 return; 942 943 /* 944 * hif_set_enable_detection(true) might come between 945 * hif_tasklet_latency_record_sched() and 946 * hif_tasklet_latency_record_exec() during wlan startup, then the 947 * sched_time is 0 but exec_time is not, and hit the timeout case in 948 * __hif_tasklet_latency(). 949 * To avoid such issue, skip exec_time recording if sched_time has not 950 * been recorded. 951 */ 952 if (!scn->latency_detect.tasklet_info[idx].sched_time) 953 return; 954 955 scn->latency_detect.tasklet_info[idx].exec_time = qdf_system_ticks(); 956 __hif_tasklet_latency(scn, false, idx); 957 } 958 hif_tasklet_latency_record_sched(struct hif_softc * scn,int idx)959 void hif_tasklet_latency_record_sched(struct hif_softc *scn, int idx) 960 { 961 if (!hif_tasklet_latency_detect_enabled(scn, idx)) 962 return; 963 964 scn->latency_detect.tasklet_info[idx].sched_cpuid = qdf_get_cpu(); 965 scn->latency_detect.tasklet_info[idx].sched_time = qdf_system_ticks(); 966 } 967 hif_credit_latency(struct hif_softc * scn,bool from_timer)968 static inline void hif_credit_latency(struct hif_softc *scn, bool from_timer) 969 { 970 qdf_time_t credit_request_time = 971 scn->latency_detect.credit_request_time; 972 qdf_time_t credit_report_time = scn->latency_detect.credit_report_time; 973 qdf_time_t curr_jiffies = qdf_system_ticks(); 974 uint32_t threshold = scn->latency_detect.threshold; 975 int cpu_id = qdf_get_cpu(); 976 977 /* 2 kinds of check here. 978 * from_timer==true: check if credit report stall 979 * from_timer==false: check credit report comes late 980 */ 981 982 if ((from_timer ? 983 qdf_system_time_after(credit_request_time, credit_report_time) : 984 qdf_system_time_after(credit_report_time, credit_request_time)) && 985 qdf_system_time_after(curr_jiffies, 986 credit_request_time + 987 qdf_system_msecs_to_ticks(threshold))) { 988 hif_err("credit report latency: from timer %d, curr_jiffies %lu, credit_request_time %lu, credit_report_time %lu, threshold %ums, timeout %ums, cpu_id %d, called: %ps", 989 from_timer, curr_jiffies, credit_request_time, 990 credit_report_time, threshold, 991 scn->latency_detect.timeout, 992 cpu_id, (void *)_RET_IP_); 993 goto latency; 994 } 995 return; 996 997 latency: 998 qdf_trigger_self_recovery(NULL, QDF_TASKLET_CREDIT_LATENCY_DETECT); 999 } 1000 hif_tasklet_latency(struct hif_softc * scn,bool from_timer)1001 static inline void hif_tasklet_latency(struct hif_softc *scn, bool from_timer) 1002 { 1003 int i, ret; 1004 1005 for (i = 0; i < HIF_TASKLET_IN_MONITOR; i++) { 1006 if (!qdf_test_bit(i, scn->latency_detect.tasklet_bmap)) 1007 continue; 1008 1009 ret = __hif_tasklet_latency(scn, from_timer, i); 1010 if (ret) 1011 return; 1012 } 1013 } 1014 1015 /** 1016 * hif_check_detection_latency(): to check if latency for tasklet/credit 1017 * 1018 * @scn: hif context 1019 * @from_timer: if called from timer handler 1020 * @bitmap_type: indicate if check tasklet or credit 1021 * 1022 * Return: none 1023 */ hif_check_detection_latency(struct hif_softc * scn,bool from_timer,uint32_t bitmap_type)1024 void hif_check_detection_latency(struct hif_softc *scn, 1025 bool from_timer, 1026 uint32_t bitmap_type) 1027 { 1028 if (QDF_GLOBAL_MISSION_MODE != hif_get_conparam(scn)) 1029 return; 1030 1031 if (!scn->latency_detect.enable_detection) 1032 return; 1033 1034 if (bitmap_type & BIT(HIF_DETECT_TASKLET)) 1035 hif_tasklet_latency(scn, from_timer); 1036 1037 if (bitmap_type & BIT(HIF_DETECT_CREDIT)) 1038 hif_credit_latency(scn, from_timer); 1039 } 1040 hif_latency_detect_timeout_handler(void * arg)1041 static void hif_latency_detect_timeout_handler(void *arg) 1042 { 1043 struct hif_softc *scn = (struct hif_softc *)arg; 1044 int next_cpu, i; 1045 qdf_cpu_mask cpu_mask = {0}; 1046 struct hif_latency_detect *detect = &scn->latency_detect; 1047 1048 hif_check_detection_latency(scn, true, 1049 BIT(HIF_DETECT_TASKLET) | 1050 BIT(HIF_DETECT_CREDIT)); 1051 1052 /* it need to make sure timer start on a different cpu, 1053 * so it can detect the tasklet schedule stall, but there 1054 * is still chance that, after timer has been started, then 1055 * irq/tasklet happens on the same cpu, then tasklet will 1056 * execute before softirq timer, if this tasklet stall, the 1057 * timer can't detect it, we can accept this as a limitation, 1058 * if tasklet stall, anyway other place will detect it, just 1059 * a little later. 1060 */ 1061 qdf_cpumask_copy(&cpu_mask, (const qdf_cpu_mask *)cpu_active_mask); 1062 for (i = 0; i < HIF_TASKLET_IN_MONITOR; i++) { 1063 if (!qdf_test_bit(i, detect->tasklet_bmap)) 1064 continue; 1065 1066 qdf_cpumask_clear_cpu(detect->tasklet_info[i].sched_cpuid, 1067 &cpu_mask); 1068 } 1069 1070 next_cpu = cpumask_first(&cpu_mask); 1071 if (qdf_unlikely(next_cpu >= nr_cpu_ids)) { 1072 hif_debug("start timer on local"); 1073 /* it doesn't found a available cpu, start on local cpu*/ 1074 qdf_timer_mod(&detect->timer, detect->timeout); 1075 } else { 1076 qdf_timer_start_on(&detect->timer, detect->timeout, next_cpu); 1077 } 1078 } 1079 hif_latency_detect_timer_init(struct hif_softc * scn)1080 static void hif_latency_detect_timer_init(struct hif_softc *scn) 1081 { 1082 scn->latency_detect.timeout = 1083 DETECTION_TIMER_TIMEOUT; 1084 scn->latency_detect.threshold = 1085 DETECTION_LATENCY_THRESHOLD; 1086 1087 hif_info("timer timeout %u, latency threshold %u", 1088 scn->latency_detect.timeout, 1089 scn->latency_detect.threshold); 1090 1091 scn->latency_detect.is_timer_started = false; 1092 1093 qdf_timer_init(NULL, 1094 &scn->latency_detect.timer, 1095 &hif_latency_detect_timeout_handler, 1096 scn, 1097 QDF_TIMER_TYPE_SW_SPIN); 1098 } 1099 hif_latency_detect_timer_deinit(struct hif_softc * scn)1100 static void hif_latency_detect_timer_deinit(struct hif_softc *scn) 1101 { 1102 hif_info("deinit timer"); 1103 qdf_timer_free(&scn->latency_detect.timer); 1104 } 1105 hif_latency_detect_init(struct hif_softc * scn)1106 static void hif_latency_detect_init(struct hif_softc *scn) 1107 { 1108 uint32_t tasklet_mask; 1109 int i; 1110 1111 if (QDF_GLOBAL_MISSION_MODE != hif_get_conparam(scn)) 1112 return; 1113 1114 tasklet_mask = DETECTION_LATENCY_TASKLET_MASK; 1115 hif_info("tasklet mask is 0x%x", tasklet_mask); 1116 for (i = 0; i < HIF_TASKLET_IN_MONITOR; i++) { 1117 if (BIT(i) & tasklet_mask) 1118 qdf_set_bit(i, scn->latency_detect.tasklet_bmap); 1119 } 1120 1121 hif_latency_detect_timer_init(scn); 1122 } 1123 hif_latency_detect_deinit(struct hif_softc * scn)1124 static void hif_latency_detect_deinit(struct hif_softc *scn) 1125 { 1126 int i; 1127 1128 if (QDF_GLOBAL_MISSION_MODE != hif_get_conparam(scn)) 1129 return; 1130 1131 hif_latency_detect_timer_deinit(scn); 1132 for (i = 0; i < HIF_TASKLET_IN_MONITOR; i++) 1133 qdf_clear_bit(i, scn->latency_detect.tasklet_bmap); 1134 } 1135 hif_latency_detect_timer_start(struct hif_opaque_softc * hif_ctx)1136 void hif_latency_detect_timer_start(struct hif_opaque_softc *hif_ctx) 1137 { 1138 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 1139 1140 if (QDF_GLOBAL_MISSION_MODE != hif_get_conparam(scn)) 1141 return; 1142 1143 hif_debug_rl("start timer"); 1144 if (scn->latency_detect.is_timer_started) { 1145 hif_info("timer has been started"); 1146 return; 1147 } 1148 1149 qdf_timer_start(&scn->latency_detect.timer, 1150 scn->latency_detect.timeout); 1151 scn->latency_detect.is_timer_started = true; 1152 } 1153 hif_latency_detect_timer_stop(struct hif_opaque_softc * hif_ctx)1154 void hif_latency_detect_timer_stop(struct hif_opaque_softc *hif_ctx) 1155 { 1156 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 1157 1158 if (QDF_GLOBAL_MISSION_MODE != hif_get_conparam(scn)) 1159 return; 1160 1161 hif_debug_rl("stop timer"); 1162 1163 qdf_timer_sync_cancel(&scn->latency_detect.timer); 1164 scn->latency_detect.is_timer_started = false; 1165 } 1166 hif_latency_detect_credit_record_time(enum hif_credit_exchange_type type,struct hif_opaque_softc * hif_ctx)1167 void hif_latency_detect_credit_record_time( 1168 enum hif_credit_exchange_type type, 1169 struct hif_opaque_softc *hif_ctx) 1170 { 1171 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 1172 1173 if (!scn) { 1174 hif_err("Could not do runtime put, scn is null"); 1175 return; 1176 } 1177 1178 if (QDF_GLOBAL_MISSION_MODE != hif_get_conparam(scn)) 1179 return; 1180 1181 if (HIF_REQUEST_CREDIT == type) 1182 scn->latency_detect.credit_request_time = qdf_system_ticks(); 1183 else if (HIF_PROCESS_CREDIT_REPORT == type) 1184 scn->latency_detect.credit_report_time = qdf_system_ticks(); 1185 1186 hif_check_detection_latency(scn, false, BIT(HIF_DETECT_CREDIT)); 1187 } 1188 hif_set_enable_detection(struct hif_opaque_softc * hif_ctx,bool value)1189 void hif_set_enable_detection(struct hif_opaque_softc *hif_ctx, bool value) 1190 { 1191 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 1192 1193 if (!scn) { 1194 hif_err("Could not do runtime put, scn is null"); 1195 return; 1196 } 1197 1198 if (QDF_GLOBAL_MISSION_MODE != hif_get_conparam(scn)) 1199 return; 1200 1201 scn->latency_detect.enable_detection = value; 1202 } 1203 #else hif_latency_detect_init(struct hif_softc * scn)1204 static inline void hif_latency_detect_init(struct hif_softc *scn) 1205 {} 1206 hif_latency_detect_deinit(struct hif_softc * scn)1207 static inline void hif_latency_detect_deinit(struct hif_softc *scn) 1208 {} 1209 #endif 1210 1211 #ifdef WLAN_FEATURE_AFFINITY_MGR 1212 #define AFFINITY_THRESHOLD 5000000 1213 static inline void hif_affinity_mgr_init(struct hif_softc * scn,struct wlan_objmgr_psoc * psoc)1214 hif_affinity_mgr_init(struct hif_softc *scn, struct wlan_objmgr_psoc *psoc) 1215 { 1216 unsigned int cpus; 1217 qdf_cpu_mask allowed_mask = {0}; 1218 1219 scn->affinity_mgr_supported = 1220 (cfg_get(psoc, CFG_IRQ_AFFINE_AUDIO_USE_CASE) && 1221 qdf_walt_get_cpus_taken_supported()); 1222 1223 hif_info("Affinity Manager supported: %d", scn->affinity_mgr_supported); 1224 1225 if (!scn->affinity_mgr_supported) 1226 return; 1227 1228 scn->time_threshold = AFFINITY_THRESHOLD; 1229 qdf_for_each_possible_cpu(cpus) 1230 if (qdf_topology_physical_package_id(cpus) == 1231 CPU_CLUSTER_TYPE_LITTLE) 1232 qdf_cpumask_set_cpu(cpus, &allowed_mask); 1233 qdf_cpumask_copy(&scn->allowed_mask, &allowed_mask); 1234 } 1235 #else 1236 static inline void hif_affinity_mgr_init(struct hif_softc * scn,struct wlan_objmgr_psoc * psoc)1237 hif_affinity_mgr_init(struct hif_softc *scn, struct wlan_objmgr_psoc *psoc) 1238 { 1239 } 1240 #endif 1241 1242 #ifdef FEATURE_DIRECT_LINK 1243 /** 1244 * hif_init_direct_link_rcv_pipe_num(): Initialize the direct link receive 1245 * pipe number 1246 * @scn: hif context 1247 * 1248 * Return: None 1249 */ 1250 static inline hif_init_direct_link_rcv_pipe_num(struct hif_softc * scn)1251 void hif_init_direct_link_rcv_pipe_num(struct hif_softc *scn) 1252 { 1253 scn->dl_recv_pipe_num = INVALID_PIPE_NO; 1254 } 1255 #else 1256 static inline hif_init_direct_link_rcv_pipe_num(struct hif_softc * scn)1257 void hif_init_direct_link_rcv_pipe_num(struct hif_softc *scn) 1258 { 1259 } 1260 #endif 1261 hif_open(qdf_device_t qdf_ctx,uint32_t mode,enum qdf_bus_type bus_type,struct hif_driver_state_callbacks * cbk,struct wlan_objmgr_psoc * psoc)1262 struct hif_opaque_softc *hif_open(qdf_device_t qdf_ctx, 1263 uint32_t mode, 1264 enum qdf_bus_type bus_type, 1265 struct hif_driver_state_callbacks *cbk, 1266 struct wlan_objmgr_psoc *psoc) 1267 { 1268 struct hif_softc *scn; 1269 QDF_STATUS status = QDF_STATUS_SUCCESS; 1270 int bus_context_size = hif_bus_get_context_size(bus_type); 1271 1272 if (bus_context_size == 0) { 1273 hif_err("context size 0 not allowed"); 1274 return NULL; 1275 } 1276 1277 scn = (struct hif_softc *)qdf_mem_malloc(bus_context_size); 1278 if (!scn) 1279 return GET_HIF_OPAQUE_HDL(scn); 1280 1281 scn->qdf_dev = qdf_ctx; 1282 scn->hif_con_param = mode; 1283 qdf_atomic_init(&scn->active_tasklet_cnt); 1284 qdf_atomic_init(&scn->active_oom_work_cnt); 1285 1286 qdf_atomic_init(&scn->active_grp_tasklet_cnt); 1287 qdf_atomic_init(&scn->link_suspended); 1288 qdf_atomic_init(&scn->tasklet_from_intr); 1289 hif_system_pm_set_state_on(GET_HIF_OPAQUE_HDL(scn)); 1290 qdf_mem_copy(&scn->callbacks, cbk, 1291 sizeof(struct hif_driver_state_callbacks)); 1292 scn->bus_type = bus_type; 1293 1294 hif_allow_ep_vote_access(GET_HIF_OPAQUE_HDL(scn)); 1295 hif_get_cfg_from_psoc(scn, psoc); 1296 1297 hif_set_event_hist_mask(GET_HIF_OPAQUE_HDL(scn)); 1298 status = hif_bus_open(scn, bus_type); 1299 if (status != QDF_STATUS_SUCCESS) { 1300 hif_err("hif_bus_open error = %d, bus_type = %d", 1301 status, bus_type); 1302 qdf_mem_free(scn); 1303 scn = NULL; 1304 goto out; 1305 } 1306 1307 hif_rtpm_lock_init(scn); 1308 1309 hif_cpuhp_register(scn); 1310 hif_latency_detect_init(scn); 1311 hif_affinity_mgr_init(scn, psoc); 1312 hif_init_direct_link_rcv_pipe_num(scn); 1313 hif_ce_desc_history_log_register(scn); 1314 hif_desc_history_log_register(); 1315 qdf_ssr_driver_dump_register_region("hif", scn, sizeof(*scn)); 1316 1317 out: 1318 return GET_HIF_OPAQUE_HDL(scn); 1319 } 1320 1321 #ifdef ADRASTEA_RRI_ON_DDR 1322 /** 1323 * hif_uninit_rri_on_ddr(): free consistent memory allocated for rri 1324 * @scn: hif context 1325 * 1326 * Return: none 1327 */ hif_uninit_rri_on_ddr(struct hif_softc * scn)1328 void hif_uninit_rri_on_ddr(struct hif_softc *scn) 1329 { 1330 if (scn->vaddr_rri_on_ddr) 1331 qdf_mem_free_consistent(scn->qdf_dev, scn->qdf_dev->dev, 1332 RRI_ON_DDR_MEM_SIZE, 1333 scn->vaddr_rri_on_ddr, 1334 scn->paddr_rri_on_ddr, 0); 1335 scn->vaddr_rri_on_ddr = NULL; 1336 } 1337 #endif 1338 1339 /** 1340 * hif_close(): hif_close 1341 * @hif_ctx: hif_ctx 1342 * 1343 * Return: n/a 1344 */ hif_close(struct hif_opaque_softc * hif_ctx)1345 void hif_close(struct hif_opaque_softc *hif_ctx) 1346 { 1347 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 1348 1349 if (!scn) { 1350 hif_err("hif_opaque_softc is NULL"); 1351 return; 1352 } 1353 1354 qdf_ssr_driver_dump_unregister_region("hif"); 1355 hif_desc_history_log_unregister(); 1356 hif_ce_desc_history_log_unregister(); 1357 hif_latency_detect_deinit(scn); 1358 1359 if (scn->athdiag_procfs_inited) { 1360 athdiag_procfs_remove(); 1361 scn->athdiag_procfs_inited = false; 1362 } 1363 1364 if (scn->target_info.hw_name) { 1365 char *hw_name = scn->target_info.hw_name; 1366 1367 scn->target_info.hw_name = "ErrUnloading"; 1368 qdf_mem_free(hw_name); 1369 } 1370 1371 hif_uninit_rri_on_ddr(scn); 1372 hif_cleanup_static_buf_to_target(scn); 1373 hif_cpuhp_unregister(scn); 1374 hif_rtpm_lock_deinit(scn); 1375 1376 hif_bus_close(scn); 1377 1378 qdf_mem_free(scn); 1379 } 1380 1381 /** 1382 * hif_get_num_active_grp_tasklets() - get the number of active 1383 * datapath group tasklets pending to be completed. 1384 * @scn: HIF context 1385 * 1386 * Returns: the number of datapath group tasklets which are active 1387 */ hif_get_num_active_grp_tasklets(struct hif_softc * scn)1388 static inline int hif_get_num_active_grp_tasklets(struct hif_softc *scn) 1389 { 1390 return qdf_atomic_read(&scn->active_grp_tasklet_cnt); 1391 } 1392 1393 #if (defined(QCA_WIFI_QCA8074) || defined(QCA_WIFI_QCA6018) || \ 1394 defined(QCA_WIFI_QCA6290) || defined(QCA_WIFI_QCA6390) || \ 1395 defined(QCA_WIFI_QCN9000) || defined(QCA_WIFI_QCA6490) || \ 1396 defined(QCA_WIFI_QCA6750) || defined(QCA_WIFI_QCA5018) || \ 1397 defined(QCA_WIFI_KIWI) || defined(QCA_WIFI_QCN9224) || \ 1398 defined(QCA_WIFI_QCN6432) || \ 1399 defined(QCA_WIFI_QCA9574)) || defined(QCA_WIFI_QCA5332) 1400 /** 1401 * hif_get_num_pending_work() - get the number of entries in 1402 * the workqueue pending to be completed. 1403 * @scn: HIF context 1404 * 1405 * Returns: the number of tasklets which are active 1406 */ hif_get_num_pending_work(struct hif_softc * scn)1407 static inline int hif_get_num_pending_work(struct hif_softc *scn) 1408 { 1409 return hal_get_reg_write_pending_work(scn->hal_soc); 1410 } 1411 #elif defined(FEATURE_HIF_DELAYED_REG_WRITE) hif_get_num_pending_work(struct hif_softc * scn)1412 static inline int hif_get_num_pending_work(struct hif_softc *scn) 1413 { 1414 return qdf_atomic_read(&scn->active_work_cnt); 1415 } 1416 #else 1417 hif_get_num_pending_work(struct hif_softc * scn)1418 static inline int hif_get_num_pending_work(struct hif_softc *scn) 1419 { 1420 return 0; 1421 } 1422 #endif 1423 hif_try_complete_tasks(struct hif_softc * scn)1424 QDF_STATUS hif_try_complete_tasks(struct hif_softc *scn) 1425 { 1426 uint32_t task_drain_wait_cnt = 0; 1427 int tasklet = 0, grp_tasklet = 0, work = 0, oom_work = 0; 1428 1429 while ((tasklet = hif_get_num_active_tasklets(scn)) || 1430 (grp_tasklet = hif_get_num_active_grp_tasklets(scn)) || 1431 (work = hif_get_num_pending_work(scn)) || 1432 (oom_work = hif_get_num_active_oom_work(scn))) { 1433 if (++task_drain_wait_cnt > HIF_TASK_DRAIN_WAIT_CNT) { 1434 hif_err("pending tasklets %d grp tasklets %d work %d oom work %d", 1435 tasklet, grp_tasklet, work, oom_work); 1436 /* 1437 * There is chance of OOM thread getting scheduled 1438 * continuously or execution get delayed during low 1439 * memory state. So avoid panic and prevent suspend 1440 * if OOM thread is unable to complete pending 1441 * work. 1442 */ 1443 if (oom_work) 1444 hif_err("OOM thread is still pending %d tasklets %d grp tasklets %d work %d", 1445 oom_work, tasklet, grp_tasklet, work); 1446 else 1447 QDF_DEBUG_PANIC("Complete tasks takes more than %u ms: tasklets %d grp tasklets %d work %d oom_work %d", 1448 HIF_TASK_DRAIN_WAIT_CNT * 10, 1449 tasklet, grp_tasklet, work, 1450 oom_work); 1451 return QDF_STATUS_E_FAULT; 1452 } 1453 hif_info("waiting for tasklets %d grp tasklets %d work %d oom_work %d", 1454 tasklet, grp_tasklet, work, oom_work); 1455 msleep(10); 1456 } 1457 1458 return QDF_STATUS_SUCCESS; 1459 } 1460 hif_try_complete_dp_tasks(struct hif_opaque_softc * hif_ctx)1461 QDF_STATUS hif_try_complete_dp_tasks(struct hif_opaque_softc *hif_ctx) 1462 { 1463 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 1464 uint32_t task_drain_wait_cnt = 0; 1465 int grp_tasklet = 0, work = 0; 1466 1467 while ((grp_tasklet = hif_get_num_active_grp_tasklets(scn)) || 1468 (work = hif_get_num_pending_work(scn))) { 1469 if (++task_drain_wait_cnt > HIF_TASK_DRAIN_WAIT_CNT) { 1470 hif_err("pending grp tasklets %d work %d", 1471 grp_tasklet, work); 1472 QDF_DEBUG_PANIC("Complete tasks takes more than %u ms: grp tasklets %d work %d", 1473 HIF_TASK_DRAIN_WAIT_CNT * 10, 1474 grp_tasklet, work); 1475 return QDF_STATUS_E_FAULT; 1476 } 1477 hif_info("waiting for grp tasklets %d work %d", 1478 grp_tasklet, work); 1479 msleep(10); 1480 } 1481 1482 return QDF_STATUS_SUCCESS; 1483 } 1484 1485 #ifdef HIF_HAL_REG_ACCESS_SUPPORT hif_reg_window_write(struct hif_softc * scn,uint32_t offset,uint32_t value)1486 void hif_reg_window_write(struct hif_softc *scn, uint32_t offset, 1487 uint32_t value) 1488 { 1489 hal_write32_mb(scn->hal_soc, offset, value); 1490 } 1491 hif_reg_window_read(struct hif_softc * scn,uint32_t offset)1492 uint32_t hif_reg_window_read(struct hif_softc *scn, uint32_t offset) 1493 { 1494 return hal_read32_mb(scn->hal_soc, offset); 1495 } 1496 #endif 1497 1498 #if defined(HIF_IPCI) && defined(FEATURE_HAL_DELAYED_REG_WRITE) hif_try_prevent_ep_vote_access(struct hif_opaque_softc * hif_ctx)1499 QDF_STATUS hif_try_prevent_ep_vote_access(struct hif_opaque_softc *hif_ctx) 1500 { 1501 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 1502 uint32_t work_drain_wait_cnt = 0; 1503 uint32_t wait_cnt = 0; 1504 int work = 0; 1505 1506 qdf_atomic_set(&scn->dp_ep_vote_access, 1507 HIF_EP_VOTE_ACCESS_DISABLE); 1508 qdf_atomic_set(&scn->ep_vote_access, 1509 HIF_EP_VOTE_ACCESS_DISABLE); 1510 1511 while ((work = hif_get_num_pending_work(scn))) { 1512 if (++work_drain_wait_cnt > HIF_WORK_DRAIN_WAIT_CNT) { 1513 qdf_atomic_set(&scn->dp_ep_vote_access, 1514 HIF_EP_VOTE_ACCESS_ENABLE); 1515 qdf_atomic_set(&scn->ep_vote_access, 1516 HIF_EP_VOTE_ACCESS_ENABLE); 1517 hif_err("timeout wait for pending work %d ", work); 1518 return QDF_STATUS_E_FAULT; 1519 } 1520 qdf_sleep(10); 1521 } 1522 1523 if (pld_is_pci_ep_awake(scn->qdf_dev->dev) == -ENOTSUPP) 1524 return QDF_STATUS_SUCCESS; 1525 1526 while (pld_is_pci_ep_awake(scn->qdf_dev->dev)) { 1527 if (++wait_cnt > HIF_EP_WAKE_RESET_WAIT_CNT) { 1528 hif_err("Release EP vote is not proceed by Fw"); 1529 return QDF_STATUS_E_FAULT; 1530 } 1531 qdf_sleep(5); 1532 } 1533 1534 return QDF_STATUS_SUCCESS; 1535 } 1536 hif_set_ep_intermediate_vote_access(struct hif_opaque_softc * hif_ctx)1537 void hif_set_ep_intermediate_vote_access(struct hif_opaque_softc *hif_ctx) 1538 { 1539 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 1540 uint8_t vote_access; 1541 1542 vote_access = qdf_atomic_read(&scn->ep_vote_access); 1543 1544 if (vote_access != HIF_EP_VOTE_ACCESS_DISABLE) 1545 hif_info("EP vote changed from:%u to intermediate state", 1546 vote_access); 1547 1548 if (QDF_IS_STATUS_ERROR(hif_try_prevent_ep_vote_access(hif_ctx))) 1549 QDF_BUG(0); 1550 1551 qdf_atomic_set(&scn->ep_vote_access, 1552 HIF_EP_VOTE_INTERMEDIATE_ACCESS); 1553 } 1554 hif_allow_ep_vote_access(struct hif_opaque_softc * hif_ctx)1555 void hif_allow_ep_vote_access(struct hif_opaque_softc *hif_ctx) 1556 { 1557 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 1558 1559 qdf_atomic_set(&scn->dp_ep_vote_access, 1560 HIF_EP_VOTE_ACCESS_ENABLE); 1561 qdf_atomic_set(&scn->ep_vote_access, 1562 HIF_EP_VOTE_ACCESS_ENABLE); 1563 } 1564 hif_set_ep_vote_access(struct hif_opaque_softc * hif_ctx,uint8_t type,uint8_t access)1565 void hif_set_ep_vote_access(struct hif_opaque_softc *hif_ctx, 1566 uint8_t type, uint8_t access) 1567 { 1568 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 1569 1570 if (type == HIF_EP_VOTE_DP_ACCESS) 1571 qdf_atomic_set(&scn->dp_ep_vote_access, access); 1572 else 1573 qdf_atomic_set(&scn->ep_vote_access, access); 1574 } 1575 hif_get_ep_vote_access(struct hif_opaque_softc * hif_ctx,uint8_t type)1576 uint8_t hif_get_ep_vote_access(struct hif_opaque_softc *hif_ctx, 1577 uint8_t type) 1578 { 1579 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 1580 1581 if (type == HIF_EP_VOTE_DP_ACCESS) 1582 return qdf_atomic_read(&scn->dp_ep_vote_access); 1583 else 1584 return qdf_atomic_read(&scn->ep_vote_access); 1585 } 1586 #endif 1587 1588 #ifdef FEATURE_HIF_DELAYED_REG_WRITE 1589 #ifdef MEMORY_DEBUG 1590 #define HIF_REG_WRITE_QUEUE_LEN 128 1591 #else 1592 #define HIF_REG_WRITE_QUEUE_LEN 32 1593 #endif 1594 1595 /** 1596 * hif_print_reg_write_stats() - Print hif delayed reg write stats 1597 * @hif_ctx: hif opaque handle 1598 * 1599 * Return: None 1600 */ hif_print_reg_write_stats(struct hif_opaque_softc * hif_ctx)1601 void hif_print_reg_write_stats(struct hif_opaque_softc *hif_ctx) 1602 { 1603 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 1604 struct CE_state *ce_state; 1605 uint32_t *hist; 1606 int i; 1607 1608 hist = scn->wstats.sched_delay; 1609 hif_debug("wstats: enq %u deq %u coal %u direct %u q_depth %u max_q %u sched-delay hist %u %u %u %u", 1610 qdf_atomic_read(&scn->wstats.enqueues), 1611 scn->wstats.dequeues, 1612 qdf_atomic_read(&scn->wstats.coalesces), 1613 qdf_atomic_read(&scn->wstats.direct), 1614 qdf_atomic_read(&scn->wstats.q_depth), 1615 scn->wstats.max_q_depth, 1616 hist[HIF_REG_WRITE_SCHED_DELAY_SUB_100us], 1617 hist[HIF_REG_WRITE_SCHED_DELAY_SUB_1000us], 1618 hist[HIF_REG_WRITE_SCHED_DELAY_SUB_5000us], 1619 hist[HIF_REG_WRITE_SCHED_DELAY_GT_5000us]); 1620 1621 for (i = 0; i < scn->ce_count; i++) { 1622 ce_state = scn->ce_id_to_state[i]; 1623 if (!ce_state) 1624 continue; 1625 1626 hif_debug("ce%d: enq %u deq %u coal %u direct %u", 1627 i, ce_state->wstats.enqueues, 1628 ce_state->wstats.dequeues, 1629 ce_state->wstats.coalesces, 1630 ce_state->wstats.direct); 1631 } 1632 } 1633 1634 /** 1635 * hif_is_reg_write_tput_level_high() - throughput level for delayed reg writes 1636 * @scn: hif_softc pointer 1637 * 1638 * Return: true if throughput is high, else false. 1639 */ hif_is_reg_write_tput_level_high(struct hif_softc * scn)1640 static inline bool hif_is_reg_write_tput_level_high(struct hif_softc *scn) 1641 { 1642 int bw_level = hif_get_bandwidth_level(GET_HIF_OPAQUE_HDL(scn)); 1643 1644 return (bw_level >= PLD_BUS_WIDTH_MEDIUM) ? true : false; 1645 } 1646 1647 /** 1648 * hif_reg_write_fill_sched_delay_hist() - fill reg write delay histogram 1649 * @scn: hif_softc pointer 1650 * @delay_us: delay in us 1651 * 1652 * Return: None 1653 */ hif_reg_write_fill_sched_delay_hist(struct hif_softc * scn,uint64_t delay_us)1654 static inline void hif_reg_write_fill_sched_delay_hist(struct hif_softc *scn, 1655 uint64_t delay_us) 1656 { 1657 uint32_t *hist; 1658 1659 hist = scn->wstats.sched_delay; 1660 1661 if (delay_us < 100) 1662 hist[HIF_REG_WRITE_SCHED_DELAY_SUB_100us]++; 1663 else if (delay_us < 1000) 1664 hist[HIF_REG_WRITE_SCHED_DELAY_SUB_1000us]++; 1665 else if (delay_us < 5000) 1666 hist[HIF_REG_WRITE_SCHED_DELAY_SUB_5000us]++; 1667 else 1668 hist[HIF_REG_WRITE_SCHED_DELAY_GT_5000us]++; 1669 } 1670 1671 /** 1672 * hif_process_reg_write_q_elem() - process a register write queue element 1673 * @scn: hif_softc pointer 1674 * @q_elem: pointer to hal register write queue element 1675 * 1676 * Return: The value which was written to the address 1677 */ 1678 static int32_t hif_process_reg_write_q_elem(struct hif_softc * scn,struct hif_reg_write_q_elem * q_elem)1679 hif_process_reg_write_q_elem(struct hif_softc *scn, 1680 struct hif_reg_write_q_elem *q_elem) 1681 { 1682 struct CE_state *ce_state = q_elem->ce_state; 1683 uint32_t write_val = -1; 1684 1685 qdf_spin_lock_bh(&ce_state->ce_index_lock); 1686 1687 ce_state->reg_write_in_progress = false; 1688 ce_state->wstats.dequeues++; 1689 1690 if (ce_state->src_ring) { 1691 q_elem->dequeue_val = ce_state->src_ring->write_index; 1692 hal_write32_mb(scn->hal_soc, ce_state->ce_wrt_idx_offset, 1693 ce_state->src_ring->write_index); 1694 write_val = ce_state->src_ring->write_index; 1695 } else if (ce_state->dest_ring) { 1696 q_elem->dequeue_val = ce_state->dest_ring->write_index; 1697 hal_write32_mb(scn->hal_soc, ce_state->ce_wrt_idx_offset, 1698 ce_state->dest_ring->write_index); 1699 write_val = ce_state->dest_ring->write_index; 1700 } else { 1701 hif_debug("invalid reg write received"); 1702 qdf_assert(0); 1703 } 1704 1705 q_elem->valid = 0; 1706 ce_state->last_dequeue_time = q_elem->dequeue_time; 1707 1708 qdf_spin_unlock_bh(&ce_state->ce_index_lock); 1709 1710 return write_val; 1711 } 1712 1713 /** 1714 * hif_reg_write_work() - Worker to process delayed writes 1715 * @arg: hif_softc pointer 1716 * 1717 * Return: None 1718 */ hif_reg_write_work(void * arg)1719 static void hif_reg_write_work(void *arg) 1720 { 1721 struct hif_softc *scn = arg; 1722 struct hif_reg_write_q_elem *q_elem; 1723 uint32_t offset; 1724 uint64_t delta_us; 1725 int32_t q_depth, write_val; 1726 uint32_t num_processed = 0; 1727 int32_t ring_id; 1728 1729 q_elem = &scn->reg_write_queue[scn->read_idx]; 1730 q_elem->work_scheduled_time = qdf_get_log_timestamp(); 1731 q_elem->cpu_id = qdf_get_cpu(); 1732 1733 /* Make sure q_elem consistent in the memory for multi-cores */ 1734 qdf_rmb(); 1735 if (!q_elem->valid) 1736 return; 1737 1738 q_depth = qdf_atomic_read(&scn->wstats.q_depth); 1739 if (q_depth > scn->wstats.max_q_depth) 1740 scn->wstats.max_q_depth = q_depth; 1741 1742 if (hif_prevent_link_low_power_states(GET_HIF_OPAQUE_HDL(scn))) { 1743 scn->wstats.prevent_l1_fails++; 1744 return; 1745 } 1746 1747 while (true) { 1748 qdf_rmb(); 1749 if (!q_elem->valid) 1750 break; 1751 1752 qdf_rmb(); 1753 q_elem->dequeue_time = qdf_get_log_timestamp(); 1754 ring_id = q_elem->ce_state->id; 1755 offset = q_elem->offset; 1756 delta_us = qdf_log_timestamp_to_usecs(q_elem->dequeue_time - 1757 q_elem->enqueue_time); 1758 hif_reg_write_fill_sched_delay_hist(scn, delta_us); 1759 1760 scn->wstats.dequeues++; 1761 qdf_atomic_dec(&scn->wstats.q_depth); 1762 1763 write_val = hif_process_reg_write_q_elem(scn, q_elem); 1764 hif_debug("read_idx %u ce_id %d offset 0x%x dequeue_val %d", 1765 scn->read_idx, ring_id, offset, write_val); 1766 1767 qdf_trace_dp_del_reg_write(ring_id, q_elem->enqueue_val, 1768 q_elem->dequeue_val, 1769 q_elem->enqueue_time, 1770 q_elem->dequeue_time); 1771 num_processed++; 1772 scn->read_idx = (scn->read_idx + 1) & 1773 (HIF_REG_WRITE_QUEUE_LEN - 1); 1774 q_elem = &scn->reg_write_queue[scn->read_idx]; 1775 } 1776 1777 hif_allow_link_low_power_states(GET_HIF_OPAQUE_HDL(scn)); 1778 1779 /* 1780 * Decrement active_work_cnt by the number of elements dequeued after 1781 * hif_allow_link_low_power_states. 1782 * This makes sure that hif_try_complete_tasks will wait till we make 1783 * the bus access in hif_allow_link_low_power_states. This will avoid 1784 * race condition between delayed register worker and bus suspend 1785 * (system suspend or runtime suspend). 1786 * 1787 * The following decrement should be done at the end! 1788 */ 1789 qdf_atomic_sub(num_processed, &scn->active_work_cnt); 1790 } 1791 1792 /** 1793 * hif_delayed_reg_write_deinit() - De-Initialize delayed reg write processing 1794 * @scn: hif_softc pointer 1795 * 1796 * De-initialize main data structures to process register writes in a delayed 1797 * workqueue. 1798 * 1799 * Return: None 1800 */ hif_delayed_reg_write_deinit(struct hif_softc * scn)1801 static void hif_delayed_reg_write_deinit(struct hif_softc *scn) 1802 { 1803 qdf_flush_work(&scn->reg_write_work); 1804 qdf_disable_work(&scn->reg_write_work); 1805 qdf_flush_workqueue(0, scn->reg_write_wq); 1806 qdf_destroy_workqueue(0, scn->reg_write_wq); 1807 qdf_mem_free(scn->reg_write_queue); 1808 } 1809 1810 /** 1811 * hif_delayed_reg_write_init() - Initialization function for delayed reg writes 1812 * @scn: hif_softc pointer 1813 * 1814 * Initialize main data structures to process register writes in a delayed 1815 * workqueue. 1816 */ 1817 hif_delayed_reg_write_init(struct hif_softc * scn)1818 static QDF_STATUS hif_delayed_reg_write_init(struct hif_softc *scn) 1819 { 1820 qdf_atomic_init(&scn->active_work_cnt); 1821 scn->reg_write_wq = 1822 qdf_alloc_high_prior_ordered_workqueue("hif_register_write_wq"); 1823 qdf_create_work(0, &scn->reg_write_work, hif_reg_write_work, scn); 1824 scn->reg_write_queue = qdf_mem_malloc(HIF_REG_WRITE_QUEUE_LEN * 1825 sizeof(*scn->reg_write_queue)); 1826 if (!scn->reg_write_queue) { 1827 hif_err("unable to allocate memory for delayed reg write"); 1828 QDF_BUG(0); 1829 return QDF_STATUS_E_NOMEM; 1830 } 1831 1832 /* Initial value of indices */ 1833 scn->read_idx = 0; 1834 qdf_atomic_set(&scn->write_idx, -1); 1835 1836 return QDF_STATUS_SUCCESS; 1837 } 1838 hif_reg_write_enqueue(struct hif_softc * scn,struct CE_state * ce_state,uint32_t value)1839 static void hif_reg_write_enqueue(struct hif_softc *scn, 1840 struct CE_state *ce_state, 1841 uint32_t value) 1842 { 1843 struct hif_reg_write_q_elem *q_elem; 1844 uint32_t write_idx; 1845 1846 if (ce_state->reg_write_in_progress) { 1847 hif_debug("Already in progress ce_id %d offset 0x%x value %u", 1848 ce_state->id, ce_state->ce_wrt_idx_offset, value); 1849 qdf_atomic_inc(&scn->wstats.coalesces); 1850 ce_state->wstats.coalesces++; 1851 return; 1852 } 1853 1854 write_idx = qdf_atomic_inc_return(&scn->write_idx); 1855 write_idx = write_idx & (HIF_REG_WRITE_QUEUE_LEN - 1); 1856 1857 q_elem = &scn->reg_write_queue[write_idx]; 1858 if (q_elem->valid) { 1859 hif_err("queue full"); 1860 QDF_BUG(0); 1861 return; 1862 } 1863 1864 qdf_atomic_inc(&scn->wstats.enqueues); 1865 ce_state->wstats.enqueues++; 1866 1867 qdf_atomic_inc(&scn->wstats.q_depth); 1868 1869 q_elem->ce_state = ce_state; 1870 q_elem->offset = ce_state->ce_wrt_idx_offset; 1871 q_elem->enqueue_val = value; 1872 q_elem->enqueue_time = qdf_get_log_timestamp(); 1873 1874 /* 1875 * Before the valid flag is set to true, all the other 1876 * fields in the q_elem needs to be updated in memory. 1877 * Else there is a chance that the dequeuing worker thread 1878 * might read stale entries and process incorrect srng. 1879 */ 1880 qdf_wmb(); 1881 q_elem->valid = true; 1882 1883 /* 1884 * After all other fields in the q_elem has been updated 1885 * in memory successfully, the valid flag needs to be updated 1886 * in memory in time too. 1887 * Else there is a chance that the dequeuing worker thread 1888 * might read stale valid flag and the work will be bypassed 1889 * for this round. And if there is no other work scheduled 1890 * later, this hal register writing won't be updated any more. 1891 */ 1892 qdf_wmb(); 1893 1894 ce_state->reg_write_in_progress = true; 1895 qdf_atomic_inc(&scn->active_work_cnt); 1896 1897 hif_debug("write_idx %u ce_id %d offset 0x%x value %u", 1898 write_idx, ce_state->id, ce_state->ce_wrt_idx_offset, value); 1899 1900 qdf_queue_work(scn->qdf_dev, scn->reg_write_wq, 1901 &scn->reg_write_work); 1902 } 1903 hif_delayed_reg_write(struct hif_softc * scn,uint32_t ctrl_addr,uint32_t val)1904 void hif_delayed_reg_write(struct hif_softc *scn, uint32_t ctrl_addr, 1905 uint32_t val) 1906 { 1907 struct CE_state *ce_state; 1908 int ce_id = COPY_ENGINE_ID(ctrl_addr); 1909 1910 ce_state = scn->ce_id_to_state[ce_id]; 1911 1912 if (!ce_state->htt_tx_data && !ce_state->htt_rx_data) { 1913 hif_reg_write_enqueue(scn, ce_state, val); 1914 return; 1915 } 1916 1917 if (hif_is_reg_write_tput_level_high(scn) || 1918 (PLD_MHI_STATE_L0 == pld_get_mhi_state(scn->qdf_dev->dev))) { 1919 hal_write32_mb(scn->hal_soc, ce_state->ce_wrt_idx_offset, val); 1920 qdf_atomic_inc(&scn->wstats.direct); 1921 ce_state->wstats.direct++; 1922 } else { 1923 hif_reg_write_enqueue(scn, ce_state, val); 1924 } 1925 } 1926 #else hif_delayed_reg_write_init(struct hif_softc * scn)1927 static inline QDF_STATUS hif_delayed_reg_write_init(struct hif_softc *scn) 1928 { 1929 return QDF_STATUS_SUCCESS; 1930 } 1931 hif_delayed_reg_write_deinit(struct hif_softc * scn)1932 static inline void hif_delayed_reg_write_deinit(struct hif_softc *scn) 1933 { 1934 } 1935 #endif 1936 1937 #if defined(QCA_WIFI_WCN6450) hif_hal_attach(struct hif_softc * scn)1938 static QDF_STATUS hif_hal_attach(struct hif_softc *scn) 1939 { 1940 scn->hal_soc = hal_attach(hif_softc_to_hif_opaque_softc(scn), 1941 scn->qdf_dev); 1942 if (!scn->hal_soc) 1943 return QDF_STATUS_E_FAILURE; 1944 1945 return QDF_STATUS_SUCCESS; 1946 } 1947 hif_hal_detach(struct hif_softc * scn)1948 static QDF_STATUS hif_hal_detach(struct hif_softc *scn) 1949 { 1950 hal_detach(scn->hal_soc); 1951 scn->hal_soc = NULL; 1952 1953 return QDF_STATUS_SUCCESS; 1954 } 1955 #elif (defined(QCA_WIFI_QCA8074) || defined(QCA_WIFI_QCA6018) || \ 1956 defined(QCA_WIFI_QCA6290) || defined(QCA_WIFI_QCA6390) || \ 1957 defined(QCA_WIFI_QCN9000) || defined(QCA_WIFI_QCA6490) || \ 1958 defined(QCA_WIFI_QCA6750) || defined(QCA_WIFI_QCA5018) || \ 1959 defined(QCA_WIFI_KIWI) || defined(QCA_WIFI_QCN9224) || \ 1960 defined(QCA_WIFI_QCA9574)) || defined(QCA_WIFI_QCA5332) hif_hal_attach(struct hif_softc * scn)1961 static QDF_STATUS hif_hal_attach(struct hif_softc *scn) 1962 { 1963 if (ce_srng_based(scn)) { 1964 scn->hal_soc = hal_attach( 1965 hif_softc_to_hif_opaque_softc(scn), 1966 scn->qdf_dev); 1967 if (!scn->hal_soc) 1968 return QDF_STATUS_E_FAILURE; 1969 } 1970 1971 return QDF_STATUS_SUCCESS; 1972 } 1973 hif_hal_detach(struct hif_softc * scn)1974 static QDF_STATUS hif_hal_detach(struct hif_softc *scn) 1975 { 1976 if (ce_srng_based(scn)) { 1977 hal_detach(scn->hal_soc); 1978 scn->hal_soc = NULL; 1979 } 1980 1981 return QDF_STATUS_SUCCESS; 1982 } 1983 #else hif_hal_attach(struct hif_softc * scn)1984 static QDF_STATUS hif_hal_attach(struct hif_softc *scn) 1985 { 1986 return QDF_STATUS_SUCCESS; 1987 } 1988 hif_hal_detach(struct hif_softc * scn)1989 static QDF_STATUS hif_hal_detach(struct hif_softc *scn) 1990 { 1991 return QDF_STATUS_SUCCESS; 1992 } 1993 #endif 1994 hif_init_dma_mask(struct device * dev,enum qdf_bus_type bus_type)1995 int hif_init_dma_mask(struct device *dev, enum qdf_bus_type bus_type) 1996 { 1997 int ret; 1998 1999 switch (bus_type) { 2000 case QDF_BUS_TYPE_IPCI: 2001 ret = qdf_set_dma_coherent_mask(dev, 2002 DMA_COHERENT_MASK_DEFAULT); 2003 if (ret) { 2004 hif_err("Failed to set dma mask error = %d", ret); 2005 return ret; 2006 } 2007 2008 break; 2009 default: 2010 /* Follow the existing sequence for other targets */ 2011 break; 2012 } 2013 2014 return 0; 2015 } 2016 2017 /** 2018 * hif_enable(): hif_enable 2019 * @hif_ctx: hif_ctx 2020 * @dev: dev 2021 * @bdev: bus dev 2022 * @bid: bus ID 2023 * @bus_type: bus type 2024 * @type: enable type 2025 * 2026 * Return: QDF_STATUS 2027 */ hif_enable(struct hif_opaque_softc * hif_ctx,struct device * dev,void * bdev,const struct hif_bus_id * bid,enum qdf_bus_type bus_type,enum hif_enable_type type)2028 QDF_STATUS hif_enable(struct hif_opaque_softc *hif_ctx, struct device *dev, 2029 void *bdev, 2030 const struct hif_bus_id *bid, 2031 enum qdf_bus_type bus_type, 2032 enum hif_enable_type type) 2033 { 2034 QDF_STATUS status; 2035 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 2036 2037 if (!scn) { 2038 hif_err("hif_ctx = NULL"); 2039 return QDF_STATUS_E_NULL_VALUE; 2040 } 2041 2042 status = hif_enable_bus(scn, dev, bdev, bid, type); 2043 if (status != QDF_STATUS_SUCCESS) { 2044 hif_err("hif_enable_bus error = %d", status); 2045 return status; 2046 } 2047 2048 status = hif_hal_attach(scn); 2049 if (status != QDF_STATUS_SUCCESS) { 2050 hif_err("hal attach failed"); 2051 goto disable_bus; 2052 } 2053 2054 if (hif_delayed_reg_write_init(scn) != QDF_STATUS_SUCCESS) { 2055 hif_err("unable to initialize delayed reg write"); 2056 goto hal_detach; 2057 } 2058 2059 if (hif_bus_configure(scn)) { 2060 hif_err("Target probe failed"); 2061 status = QDF_STATUS_E_FAILURE; 2062 goto hal_detach; 2063 } 2064 2065 hif_ut_suspend_init(scn); 2066 hif_register_recovery_notifier(scn); 2067 hif_latency_detect_timer_start(hif_ctx); 2068 2069 /* 2070 * Flag to avoid potential unallocated memory access from MSI 2071 * interrupt handler which could get scheduled as soon as MSI 2072 * is enabled, i.e to take care of the race due to the order 2073 * in where MSI is enabled before the memory, that will be 2074 * in interrupt handlers, is allocated. 2075 */ 2076 2077 scn->hif_init_done = true; 2078 2079 hif_debug("OK"); 2080 2081 return QDF_STATUS_SUCCESS; 2082 2083 hal_detach: 2084 hif_hal_detach(scn); 2085 disable_bus: 2086 hif_disable_bus(scn); 2087 return status; 2088 } 2089 hif_disable(struct hif_opaque_softc * hif_ctx,enum hif_disable_type type)2090 void hif_disable(struct hif_opaque_softc *hif_ctx, enum hif_disable_type type) 2091 { 2092 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 2093 2094 if (!scn) 2095 return; 2096 2097 hif_delayed_reg_write_deinit(scn); 2098 hif_set_enable_detection(hif_ctx, false); 2099 hif_latency_detect_timer_stop(hif_ctx); 2100 2101 hif_unregister_recovery_notifier(scn); 2102 2103 hif_nointrs(scn); 2104 if (scn->hif_init_done == false) 2105 hif_shutdown_device(hif_ctx); 2106 else 2107 hif_stop(hif_ctx); 2108 2109 hif_hal_detach(scn); 2110 2111 hif_disable_bus(scn); 2112 2113 hif_wlan_disable(scn); 2114 2115 scn->notice_send = false; 2116 2117 hif_debug("X"); 2118 } 2119 2120 #ifdef CE_TASKLET_DEBUG_ENABLE hif_enable_ce_latency_stats(struct hif_opaque_softc * hif_ctx,uint8_t val)2121 void hif_enable_ce_latency_stats(struct hif_opaque_softc *hif_ctx, uint8_t val) 2122 { 2123 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 2124 2125 if (!scn) 2126 return; 2127 2128 scn->ce_latency_stats = val; 2129 } 2130 #endif 2131 hif_display_stats(struct hif_opaque_softc * hif_ctx)2132 void hif_display_stats(struct hif_opaque_softc *hif_ctx) 2133 { 2134 hif_display_bus_stats(hif_ctx); 2135 } 2136 2137 qdf_export_symbol(hif_display_stats); 2138 hif_clear_stats(struct hif_opaque_softc * hif_ctx)2139 void hif_clear_stats(struct hif_opaque_softc *hif_ctx) 2140 { 2141 hif_clear_bus_stats(hif_ctx); 2142 } 2143 2144 /** 2145 * hif_crash_shutdown_dump_bus_register() - dump bus registers 2146 * @hif_ctx: hif_ctx 2147 * 2148 * Return: n/a 2149 */ 2150 #if defined(TARGET_RAMDUMP_AFTER_KERNEL_PANIC) && defined(WLAN_FEATURE_BMI) 2151 hif_crash_shutdown_dump_bus_register(void * hif_ctx)2152 static void hif_crash_shutdown_dump_bus_register(void *hif_ctx) 2153 { 2154 struct hif_opaque_softc *scn = hif_ctx; 2155 2156 if (hif_check_soc_status(scn)) 2157 return; 2158 2159 if (hif_dump_registers(scn)) 2160 hif_err("Failed to dump bus registers!"); 2161 } 2162 2163 /** 2164 * hif_crash_shutdown(): hif_crash_shutdown 2165 * 2166 * This function is called by the platform driver to dump CE registers 2167 * 2168 * @hif_ctx: hif_ctx 2169 * 2170 * Return: n/a 2171 */ hif_crash_shutdown(struct hif_opaque_softc * hif_ctx)2172 void hif_crash_shutdown(struct hif_opaque_softc *hif_ctx) 2173 { 2174 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 2175 2176 if (!hif_ctx) 2177 return; 2178 2179 if (scn->bus_type == QDF_BUS_TYPE_SNOC) { 2180 hif_warn("RAM dump disabled for bustype %d", scn->bus_type); 2181 return; 2182 } 2183 2184 if (TARGET_STATUS_RESET == scn->target_status) { 2185 hif_warn("Target is already asserted, ignore!"); 2186 return; 2187 } 2188 2189 if (hif_is_load_or_unload_in_progress(scn)) { 2190 hif_err("Load/unload is in progress, ignore!"); 2191 return; 2192 } 2193 2194 hif_crash_shutdown_dump_bus_register(hif_ctx); 2195 hif_set_target_status(hif_ctx, TARGET_STATUS_RESET); 2196 2197 if (ol_copy_ramdump(hif_ctx)) 2198 goto out; 2199 2200 hif_info("RAM dump collecting completed!"); 2201 2202 out: 2203 return; 2204 } 2205 #else hif_crash_shutdown(struct hif_opaque_softc * hif_ctx)2206 void hif_crash_shutdown(struct hif_opaque_softc *hif_ctx) 2207 { 2208 hif_debug("Collecting target RAM dump disabled"); 2209 } 2210 #endif /* TARGET_RAMDUMP_AFTER_KERNEL_PANIC */ 2211 2212 #ifdef QCA_WIFI_3_0 2213 /** 2214 * hif_check_fw_reg(): hif_check_fw_reg 2215 * @scn: scn 2216 * 2217 * Return: int 2218 */ hif_check_fw_reg(struct hif_opaque_softc * scn)2219 int hif_check_fw_reg(struct hif_opaque_softc *scn) 2220 { 2221 return 0; 2222 } 2223 #endif 2224 2225 /** 2226 * hif_read_phy_mem_base(): hif_read_phy_mem_base 2227 * @scn: scn 2228 * @phy_mem_base: physical mem base 2229 * 2230 * Return: n/a 2231 */ hif_read_phy_mem_base(struct hif_softc * scn,qdf_dma_addr_t * phy_mem_base)2232 void hif_read_phy_mem_base(struct hif_softc *scn, qdf_dma_addr_t *phy_mem_base) 2233 { 2234 *phy_mem_base = scn->mem_pa; 2235 } 2236 qdf_export_symbol(hif_read_phy_mem_base); 2237 2238 /** 2239 * hif_get_device_type(): hif_get_device_type 2240 * @device_id: device_id 2241 * @revision_id: revision_id 2242 * @hif_type: returned hif_type 2243 * @target_type: returned target_type 2244 * 2245 * Return: int 2246 */ hif_get_device_type(uint32_t device_id,uint32_t revision_id,uint32_t * hif_type,uint32_t * target_type)2247 int hif_get_device_type(uint32_t device_id, 2248 uint32_t revision_id, 2249 uint32_t *hif_type, uint32_t *target_type) 2250 { 2251 int ret = 0; 2252 2253 switch (device_id) { 2254 case ADRASTEA_DEVICE_ID_P2_E12: 2255 2256 *hif_type = HIF_TYPE_ADRASTEA; 2257 *target_type = TARGET_TYPE_ADRASTEA; 2258 break; 2259 2260 case AR9888_DEVICE_ID: 2261 *hif_type = HIF_TYPE_AR9888; 2262 *target_type = TARGET_TYPE_AR9888; 2263 break; 2264 2265 case AR6320_DEVICE_ID: 2266 switch (revision_id) { 2267 case AR6320_FW_1_1: 2268 case AR6320_FW_1_3: 2269 *hif_type = HIF_TYPE_AR6320; 2270 *target_type = TARGET_TYPE_AR6320; 2271 break; 2272 2273 case AR6320_FW_2_0: 2274 case AR6320_FW_3_0: 2275 case AR6320_FW_3_2: 2276 *hif_type = HIF_TYPE_AR6320V2; 2277 *target_type = TARGET_TYPE_AR6320V2; 2278 break; 2279 2280 default: 2281 hif_err("dev_id = 0x%x, rev_id = 0x%x", 2282 device_id, revision_id); 2283 ret = -ENODEV; 2284 goto end; 2285 } 2286 break; 2287 2288 case AR9887_DEVICE_ID: 2289 *hif_type = HIF_TYPE_AR9888; 2290 *target_type = TARGET_TYPE_AR9888; 2291 hif_info(" *********** AR9887 **************"); 2292 break; 2293 2294 case QCA9984_DEVICE_ID: 2295 *hif_type = HIF_TYPE_QCA9984; 2296 *target_type = TARGET_TYPE_QCA9984; 2297 hif_info(" *********** QCA9984 *************"); 2298 break; 2299 2300 case QCA9888_DEVICE_ID: 2301 *hif_type = HIF_TYPE_QCA9888; 2302 *target_type = TARGET_TYPE_QCA9888; 2303 hif_info(" *********** QCA9888 *************"); 2304 break; 2305 2306 case AR900B_DEVICE_ID: 2307 *hif_type = HIF_TYPE_AR900B; 2308 *target_type = TARGET_TYPE_AR900B; 2309 hif_info(" *********** AR900B *************"); 2310 break; 2311 2312 case QCA8074_DEVICE_ID: 2313 *hif_type = HIF_TYPE_QCA8074; 2314 *target_type = TARGET_TYPE_QCA8074; 2315 hif_info(" *********** QCA8074 *************"); 2316 break; 2317 2318 case QCA6290_EMULATION_DEVICE_ID: 2319 case QCA6290_DEVICE_ID: 2320 *hif_type = HIF_TYPE_QCA6290; 2321 *target_type = TARGET_TYPE_QCA6290; 2322 hif_info(" *********** QCA6290EMU *************"); 2323 break; 2324 2325 case QCN9000_DEVICE_ID: 2326 *hif_type = HIF_TYPE_QCN9000; 2327 *target_type = TARGET_TYPE_QCN9000; 2328 hif_info(" *********** QCN9000 *************"); 2329 break; 2330 2331 case QCN9224_DEVICE_ID: 2332 *hif_type = HIF_TYPE_QCN9224; 2333 *target_type = TARGET_TYPE_QCN9224; 2334 hif_info(" *********** QCN9224 *************"); 2335 break; 2336 2337 case QCN6122_DEVICE_ID: 2338 *hif_type = HIF_TYPE_QCN6122; 2339 *target_type = TARGET_TYPE_QCN6122; 2340 hif_info(" *********** QCN6122 *************"); 2341 break; 2342 2343 case QCN9160_DEVICE_ID: 2344 *hif_type = HIF_TYPE_QCN9160; 2345 *target_type = TARGET_TYPE_QCN9160; 2346 hif_info(" *********** QCN9160 *************"); 2347 break; 2348 2349 case QCN6432_DEVICE_ID: 2350 *hif_type = HIF_TYPE_QCN6432; 2351 *target_type = TARGET_TYPE_QCN6432; 2352 hif_info(" *********** QCN6432 *************"); 2353 break; 2354 2355 case QCN7605_DEVICE_ID: 2356 case QCN7605_COMPOSITE: 2357 case QCN7605_STANDALONE: 2358 case QCN7605_STANDALONE_V2: 2359 case QCN7605_COMPOSITE_V2: 2360 *hif_type = HIF_TYPE_QCN7605; 2361 *target_type = TARGET_TYPE_QCN7605; 2362 hif_info(" *********** QCN7605 *************"); 2363 break; 2364 2365 case QCA6390_DEVICE_ID: 2366 case QCA6390_EMULATION_DEVICE_ID: 2367 *hif_type = HIF_TYPE_QCA6390; 2368 *target_type = TARGET_TYPE_QCA6390; 2369 hif_info(" *********** QCA6390 *************"); 2370 break; 2371 2372 case QCA6490_DEVICE_ID: 2373 case QCA6490_EMULATION_DEVICE_ID: 2374 *hif_type = HIF_TYPE_QCA6490; 2375 *target_type = TARGET_TYPE_QCA6490; 2376 hif_info(" *********** QCA6490 *************"); 2377 break; 2378 2379 case QCA6750_DEVICE_ID: 2380 case QCA6750_EMULATION_DEVICE_ID: 2381 *hif_type = HIF_TYPE_QCA6750; 2382 *target_type = TARGET_TYPE_QCA6750; 2383 hif_info(" *********** QCA6750 *************"); 2384 break; 2385 2386 case KIWI_DEVICE_ID: 2387 *hif_type = HIF_TYPE_KIWI; 2388 *target_type = TARGET_TYPE_KIWI; 2389 hif_info(" *********** KIWI *************"); 2390 break; 2391 2392 case MANGO_DEVICE_ID: 2393 *hif_type = HIF_TYPE_MANGO; 2394 *target_type = TARGET_TYPE_MANGO; 2395 hif_info(" *********** MANGO *************"); 2396 break; 2397 2398 case PEACH_DEVICE_ID: 2399 *hif_type = HIF_TYPE_PEACH; 2400 *target_type = TARGET_TYPE_PEACH; 2401 hif_info(" *********** PEACH *************"); 2402 break; 2403 2404 case QCA8074V2_DEVICE_ID: 2405 *hif_type = HIF_TYPE_QCA8074V2; 2406 *target_type = TARGET_TYPE_QCA8074V2; 2407 hif_info(" *********** QCA8074V2 *************"); 2408 break; 2409 2410 case QCA6018_DEVICE_ID: 2411 case RUMIM2M_DEVICE_ID_NODE0: 2412 case RUMIM2M_DEVICE_ID_NODE1: 2413 case RUMIM2M_DEVICE_ID_NODE2: 2414 case RUMIM2M_DEVICE_ID_NODE3: 2415 case RUMIM2M_DEVICE_ID_NODE4: 2416 case RUMIM2M_DEVICE_ID_NODE5: 2417 *hif_type = HIF_TYPE_QCA6018; 2418 *target_type = TARGET_TYPE_QCA6018; 2419 hif_info(" *********** QCA6018 *************"); 2420 break; 2421 2422 case QCA5018_DEVICE_ID: 2423 *hif_type = HIF_TYPE_QCA5018; 2424 *target_type = TARGET_TYPE_QCA5018; 2425 hif_info(" *********** qca5018 *************"); 2426 break; 2427 2428 case QCA5332_DEVICE_ID: 2429 *hif_type = HIF_TYPE_QCA5332; 2430 *target_type = TARGET_TYPE_QCA5332; 2431 hif_info(" *********** QCA5332 *************"); 2432 break; 2433 2434 case QCA9574_DEVICE_ID: 2435 *hif_type = HIF_TYPE_QCA9574; 2436 *target_type = TARGET_TYPE_QCA9574; 2437 hif_info(" *********** QCA9574 *************"); 2438 break; 2439 2440 case WCN6450_DEVICE_ID: 2441 *hif_type = HIF_TYPE_WCN6450; 2442 *target_type = TARGET_TYPE_WCN6450; 2443 hif_info(" *********** WCN6450 *************"); 2444 break; 2445 2446 default: 2447 hif_err("Unsupported device ID = 0x%x!", device_id); 2448 ret = -ENODEV; 2449 break; 2450 } 2451 2452 if (*target_type == TARGET_TYPE_UNKNOWN) { 2453 hif_err("Unsupported target_type!"); 2454 ret = -ENODEV; 2455 } 2456 end: 2457 return ret; 2458 } 2459 2460 /** 2461 * hif_get_bus_type() - return the bus type 2462 * @hif_hdl: HIF Context 2463 * 2464 * Return: enum qdf_bus_type 2465 */ hif_get_bus_type(struct hif_opaque_softc * hif_hdl)2466 enum qdf_bus_type hif_get_bus_type(struct hif_opaque_softc *hif_hdl) 2467 { 2468 struct hif_softc *scn = HIF_GET_SOFTC(hif_hdl); 2469 2470 return scn->bus_type; 2471 } 2472 2473 /* 2474 * Target info and ini parameters are global to the driver 2475 * Hence these structures are exposed to all the modules in 2476 * the driver and they don't need to maintains multiple copies 2477 * of the same info, instead get the handle from hif and 2478 * modify them in hif 2479 */ 2480 2481 /** 2482 * hif_get_ini_handle() - API to get hif_config_param handle 2483 * @hif_ctx: HIF Context 2484 * 2485 * Return: pointer to hif_config_info 2486 */ hif_get_ini_handle(struct hif_opaque_softc * hif_ctx)2487 struct hif_config_info *hif_get_ini_handle(struct hif_opaque_softc *hif_ctx) 2488 { 2489 struct hif_softc *sc = HIF_GET_SOFTC(hif_ctx); 2490 2491 return &sc->hif_config; 2492 } 2493 2494 /** 2495 * hif_get_target_info_handle() - API to get hif_target_info handle 2496 * @hif_ctx: HIF context 2497 * 2498 * Return: Pointer to hif_target_info 2499 */ hif_get_target_info_handle(struct hif_opaque_softc * hif_ctx)2500 struct hif_target_info *hif_get_target_info_handle( 2501 struct hif_opaque_softc *hif_ctx) 2502 { 2503 struct hif_softc *sc = HIF_GET_SOFTC(hif_ctx); 2504 2505 return &sc->target_info; 2506 2507 } 2508 qdf_export_symbol(hif_get_target_info_handle); 2509 2510 #ifdef RECEIVE_OFFLOAD hif_offld_flush_cb_register(struct hif_opaque_softc * scn,void (offld_flush_handler)(void *))2511 void hif_offld_flush_cb_register(struct hif_opaque_softc *scn, 2512 void (offld_flush_handler)(void *)) 2513 { 2514 if (hif_napi_enabled(scn, -1)) 2515 hif_napi_rx_offld_flush_cb_register(scn, offld_flush_handler); 2516 else 2517 hif_err("NAPI not enabled"); 2518 } 2519 qdf_export_symbol(hif_offld_flush_cb_register); 2520 hif_offld_flush_cb_deregister(struct hif_opaque_softc * scn)2521 void hif_offld_flush_cb_deregister(struct hif_opaque_softc *scn) 2522 { 2523 if (hif_napi_enabled(scn, -1)) 2524 hif_napi_rx_offld_flush_cb_deregister(scn); 2525 else 2526 hif_err("NAPI not enabled"); 2527 } 2528 qdf_export_symbol(hif_offld_flush_cb_deregister); 2529 hif_get_rx_ctx_id(int ctx_id,struct hif_opaque_softc * hif_hdl)2530 int hif_get_rx_ctx_id(int ctx_id, struct hif_opaque_softc *hif_hdl) 2531 { 2532 if (hif_napi_enabled(hif_hdl, -1)) 2533 return NAPI_PIPE2ID(ctx_id); 2534 else 2535 return ctx_id; 2536 } 2537 #else /* RECEIVE_OFFLOAD */ hif_get_rx_ctx_id(int ctx_id,struct hif_opaque_softc * hif_hdl)2538 int hif_get_rx_ctx_id(int ctx_id, struct hif_opaque_softc *hif_hdl) 2539 { 2540 return 0; 2541 } 2542 qdf_export_symbol(hif_get_rx_ctx_id); 2543 #endif /* RECEIVE_OFFLOAD */ 2544 2545 #if defined(FEATURE_LRO) 2546 2547 /** 2548 * hif_get_lro_info - Returns LRO instance for instance ID 2549 * @ctx_id: LRO instance ID 2550 * @hif_hdl: HIF Context 2551 * 2552 * Return: Pointer to LRO instance. 2553 */ hif_get_lro_info(int ctx_id,struct hif_opaque_softc * hif_hdl)2554 void *hif_get_lro_info(int ctx_id, struct hif_opaque_softc *hif_hdl) 2555 { 2556 void *data; 2557 2558 if (hif_napi_enabled(hif_hdl, -1)) 2559 data = hif_napi_get_lro_info(hif_hdl, ctx_id); 2560 else 2561 data = hif_ce_get_lro_ctx(hif_hdl, ctx_id); 2562 2563 return data; 2564 } 2565 #endif 2566 2567 /** 2568 * hif_get_target_status - API to get target status 2569 * @hif_ctx: HIF Context 2570 * 2571 * Return: enum hif_target_status 2572 */ hif_get_target_status(struct hif_opaque_softc * hif_ctx)2573 enum hif_target_status hif_get_target_status(struct hif_opaque_softc *hif_ctx) 2574 { 2575 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 2576 2577 return scn->target_status; 2578 } 2579 qdf_export_symbol(hif_get_target_status); 2580 2581 /** 2582 * hif_set_target_status() - API to set target status 2583 * @hif_ctx: HIF Context 2584 * @status: Target Status 2585 * 2586 * Return: void 2587 */ hif_set_target_status(struct hif_opaque_softc * hif_ctx,enum hif_target_status status)2588 void hif_set_target_status(struct hif_opaque_softc *hif_ctx, enum 2589 hif_target_status status) 2590 { 2591 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 2592 2593 scn->target_status = status; 2594 } 2595 2596 /** 2597 * hif_init_ini_config() - API to initialize HIF configuration parameters 2598 * @hif_ctx: HIF Context 2599 * @cfg: HIF Configuration 2600 * 2601 * Return: void 2602 */ hif_init_ini_config(struct hif_opaque_softc * hif_ctx,struct hif_config_info * cfg)2603 void hif_init_ini_config(struct hif_opaque_softc *hif_ctx, 2604 struct hif_config_info *cfg) 2605 { 2606 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 2607 2608 qdf_mem_copy(&scn->hif_config, cfg, sizeof(struct hif_config_info)); 2609 } 2610 2611 /** 2612 * hif_get_conparam() - API to get driver mode in HIF 2613 * @scn: HIF Context 2614 * 2615 * Return: driver mode of operation 2616 */ hif_get_conparam(struct hif_softc * scn)2617 uint32_t hif_get_conparam(struct hif_softc *scn) 2618 { 2619 if (!scn) 2620 return 0; 2621 2622 return scn->hif_con_param; 2623 } 2624 2625 /** 2626 * hif_get_callbacks_handle() - API to get callbacks Handle 2627 * @scn: HIF Context 2628 * 2629 * Return: pointer to HIF Callbacks 2630 */ hif_get_callbacks_handle(struct hif_softc * scn)2631 struct hif_driver_state_callbacks *hif_get_callbacks_handle( 2632 struct hif_softc *scn) 2633 { 2634 return &scn->callbacks; 2635 } 2636 2637 /** 2638 * hif_is_driver_unloading() - API to query upper layers if driver is unloading 2639 * @scn: HIF Context 2640 * 2641 * Return: True/False 2642 */ hif_is_driver_unloading(struct hif_softc * scn)2643 bool hif_is_driver_unloading(struct hif_softc *scn) 2644 { 2645 struct hif_driver_state_callbacks *cbk = hif_get_callbacks_handle(scn); 2646 2647 if (cbk && cbk->is_driver_unloading) 2648 return cbk->is_driver_unloading(cbk->context); 2649 2650 return false; 2651 } 2652 2653 /** 2654 * hif_is_load_or_unload_in_progress() - API to query upper layers if 2655 * load/unload in progress 2656 * @scn: HIF Context 2657 * 2658 * Return: True/False 2659 */ hif_is_load_or_unload_in_progress(struct hif_softc * scn)2660 bool hif_is_load_or_unload_in_progress(struct hif_softc *scn) 2661 { 2662 struct hif_driver_state_callbacks *cbk = hif_get_callbacks_handle(scn); 2663 2664 if (cbk && cbk->is_load_unload_in_progress) 2665 return cbk->is_load_unload_in_progress(cbk->context); 2666 2667 return false; 2668 } 2669 2670 /** 2671 * hif_is_recovery_in_progress() - API to query upper layers if recovery in 2672 * progress 2673 * @scn: HIF Context 2674 * 2675 * Return: True/False 2676 */ hif_is_recovery_in_progress(struct hif_softc * scn)2677 bool hif_is_recovery_in_progress(struct hif_softc *scn) 2678 { 2679 struct hif_driver_state_callbacks *cbk = hif_get_callbacks_handle(scn); 2680 2681 if (cbk && cbk->is_recovery_in_progress) 2682 return cbk->is_recovery_in_progress(cbk->context); 2683 2684 return false; 2685 } 2686 2687 #if defined(HIF_PCI) || defined(HIF_SNOC) || defined(HIF_AHB) || \ 2688 defined(HIF_IPCI) 2689 2690 /** 2691 * hif_update_pipe_callback() - API to register pipe specific callbacks 2692 * @osc: Opaque softc 2693 * @pipeid: pipe id 2694 * @callbacks: callbacks to register 2695 * 2696 * Return: void 2697 */ 2698 hif_update_pipe_callback(struct hif_opaque_softc * osc,u_int8_t pipeid,struct hif_msg_callbacks * callbacks)2699 void hif_update_pipe_callback(struct hif_opaque_softc *osc, 2700 u_int8_t pipeid, 2701 struct hif_msg_callbacks *callbacks) 2702 { 2703 struct hif_softc *scn = HIF_GET_SOFTC(osc); 2704 struct HIF_CE_state *hif_state = HIF_GET_CE_STATE(scn); 2705 struct HIF_CE_pipe_info *pipe_info; 2706 2707 QDF_BUG(pipeid < CE_COUNT_MAX); 2708 2709 hif_debug("pipeid: %d", pipeid); 2710 2711 pipe_info = &hif_state->pipe_info[pipeid]; 2712 2713 qdf_mem_copy(&pipe_info->pipe_callbacks, 2714 callbacks, sizeof(pipe_info->pipe_callbacks)); 2715 } 2716 qdf_export_symbol(hif_update_pipe_callback); 2717 2718 /** 2719 * hif_is_target_ready() - API to query if target is in ready state 2720 * progress 2721 * @scn: HIF Context 2722 * 2723 * Return: True/False 2724 */ hif_is_target_ready(struct hif_softc * scn)2725 bool hif_is_target_ready(struct hif_softc *scn) 2726 { 2727 struct hif_driver_state_callbacks *cbk = hif_get_callbacks_handle(scn); 2728 2729 if (cbk && cbk->is_target_ready) 2730 return cbk->is_target_ready(cbk->context); 2731 /* 2732 * if callback is not registered then there is no way to determine 2733 * if target is ready. In-such case return true to indicate that 2734 * target is ready. 2735 */ 2736 return true; 2737 } 2738 qdf_export_symbol(hif_is_target_ready); 2739 hif_get_bandwidth_level(struct hif_opaque_softc * hif_handle)2740 int hif_get_bandwidth_level(struct hif_opaque_softc *hif_handle) 2741 { 2742 struct hif_softc *scn = HIF_GET_SOFTC(hif_handle); 2743 struct hif_driver_state_callbacks *cbk = hif_get_callbacks_handle(scn); 2744 2745 if (cbk && cbk->get_bandwidth_level) 2746 return cbk->get_bandwidth_level(cbk->context); 2747 2748 return 0; 2749 } 2750 2751 qdf_export_symbol(hif_get_bandwidth_level); 2752 2753 #ifdef DP_MEM_PRE_ALLOC hif_mem_alloc_consistent_unaligned(struct hif_softc * scn,qdf_size_t size,qdf_dma_addr_t * paddr,uint32_t ring_type,uint8_t * is_mem_prealloc)2754 void *hif_mem_alloc_consistent_unaligned(struct hif_softc *scn, 2755 qdf_size_t size, 2756 qdf_dma_addr_t *paddr, 2757 uint32_t ring_type, 2758 uint8_t *is_mem_prealloc) 2759 { 2760 void *vaddr = NULL; 2761 struct hif_driver_state_callbacks *cbk = 2762 hif_get_callbacks_handle(scn); 2763 2764 *is_mem_prealloc = false; 2765 if (cbk && cbk->prealloc_get_consistent_mem_unaligned) { 2766 vaddr = cbk->prealloc_get_consistent_mem_unaligned(size, 2767 paddr, 2768 ring_type); 2769 if (vaddr) { 2770 *is_mem_prealloc = true; 2771 goto end; 2772 } 2773 } 2774 2775 vaddr = qdf_mem_alloc_consistent(scn->qdf_dev, 2776 scn->qdf_dev->dev, 2777 size, 2778 paddr); 2779 end: 2780 dp_info("%s va_unaligned %pK pa_unaligned %pK size %d ring_type %d", 2781 *is_mem_prealloc ? "pre-alloc" : "dynamic-alloc", vaddr, 2782 (void *)*paddr, (int)size, ring_type); 2783 2784 return vaddr; 2785 } 2786 hif_mem_free_consistent_unaligned(struct hif_softc * scn,qdf_size_t size,void * vaddr,qdf_dma_addr_t paddr,qdf_dma_context_t memctx,uint8_t is_mem_prealloc)2787 void hif_mem_free_consistent_unaligned(struct hif_softc *scn, 2788 qdf_size_t size, 2789 void *vaddr, 2790 qdf_dma_addr_t paddr, 2791 qdf_dma_context_t memctx, 2792 uint8_t is_mem_prealloc) 2793 { 2794 struct hif_driver_state_callbacks *cbk = 2795 hif_get_callbacks_handle(scn); 2796 2797 if (is_mem_prealloc) { 2798 if (cbk && cbk->prealloc_put_consistent_mem_unaligned) { 2799 cbk->prealloc_put_consistent_mem_unaligned(vaddr); 2800 } else { 2801 dp_warn("dp_prealloc_put_consistent_unligned NULL"); 2802 QDF_BUG(0); 2803 } 2804 } else { 2805 qdf_mem_free_consistent(scn->qdf_dev, scn->qdf_dev->dev, 2806 size, vaddr, paddr, memctx); 2807 } 2808 } 2809 hif_prealloc_get_multi_pages(struct hif_softc * scn,uint32_t desc_type,qdf_size_t elem_size,uint16_t elem_num,struct qdf_mem_multi_page_t * pages,bool cacheable)2810 void hif_prealloc_get_multi_pages(struct hif_softc *scn, uint32_t desc_type, 2811 qdf_size_t elem_size, uint16_t elem_num, 2812 struct qdf_mem_multi_page_t *pages, 2813 bool cacheable) 2814 { 2815 struct hif_driver_state_callbacks *cbk = 2816 hif_get_callbacks_handle(scn); 2817 2818 if (cbk && cbk->prealloc_get_multi_pages) 2819 cbk->prealloc_get_multi_pages(desc_type, elem_size, elem_num, 2820 pages, cacheable); 2821 2822 if (!pages->num_pages) 2823 qdf_mem_multi_pages_alloc(scn->qdf_dev, pages, 2824 elem_size, elem_num, 0, cacheable); 2825 } 2826 hif_prealloc_put_multi_pages(struct hif_softc * scn,uint32_t desc_type,struct qdf_mem_multi_page_t * pages,bool cacheable)2827 void hif_prealloc_put_multi_pages(struct hif_softc *scn, uint32_t desc_type, 2828 struct qdf_mem_multi_page_t *pages, 2829 bool cacheable) 2830 { 2831 struct hif_driver_state_callbacks *cbk = 2832 hif_get_callbacks_handle(scn); 2833 2834 if (cbk && cbk->prealloc_put_multi_pages && 2835 pages->is_mem_prealloc) 2836 cbk->prealloc_put_multi_pages(desc_type, pages); 2837 2838 if (!pages->is_mem_prealloc) 2839 qdf_mem_multi_pages_free(scn->qdf_dev, pages, 0, 2840 cacheable); 2841 } 2842 #endif 2843 2844 /** 2845 * hif_batch_send() - API to access hif specific function 2846 * ce_batch_send. 2847 * @osc: HIF Context 2848 * @msdu: list of msdus to be sent 2849 * @transfer_id: transfer id 2850 * @len: downloaded length 2851 * @sendhead: 2852 * 2853 * Return: list of msds not sent 2854 */ hif_batch_send(struct hif_opaque_softc * osc,qdf_nbuf_t msdu,uint32_t transfer_id,u_int32_t len,uint32_t sendhead)2855 qdf_nbuf_t hif_batch_send(struct hif_opaque_softc *osc, qdf_nbuf_t msdu, 2856 uint32_t transfer_id, u_int32_t len, uint32_t sendhead) 2857 { 2858 void *ce_tx_hdl = hif_get_ce_handle(osc, CE_HTT_TX_CE); 2859 2860 if (!ce_tx_hdl) 2861 return NULL; 2862 2863 return ce_batch_send((struct CE_handle *)ce_tx_hdl, msdu, transfer_id, 2864 len, sendhead); 2865 } 2866 qdf_export_symbol(hif_batch_send); 2867 2868 /** 2869 * hif_update_tx_ring() - API to access hif specific function 2870 * ce_update_tx_ring. 2871 * @osc: HIF Context 2872 * @num_htt_cmpls: number of htt compl received. 2873 * 2874 * Return: void 2875 */ hif_update_tx_ring(struct hif_opaque_softc * osc,u_int32_t num_htt_cmpls)2876 void hif_update_tx_ring(struct hif_opaque_softc *osc, u_int32_t num_htt_cmpls) 2877 { 2878 void *ce_tx_hdl = hif_get_ce_handle(osc, CE_HTT_TX_CE); 2879 2880 ce_update_tx_ring(ce_tx_hdl, num_htt_cmpls); 2881 } 2882 qdf_export_symbol(hif_update_tx_ring); 2883 2884 2885 /** 2886 * hif_send_single() - API to access hif specific function 2887 * ce_send_single. 2888 * @osc: HIF Context 2889 * @msdu : msdu to be sent 2890 * @transfer_id: transfer id 2891 * @len : downloaded length 2892 * 2893 * Return: msdu sent status 2894 */ hif_send_single(struct hif_opaque_softc * osc,qdf_nbuf_t msdu,uint32_t transfer_id,u_int32_t len)2895 QDF_STATUS hif_send_single(struct hif_opaque_softc *osc, qdf_nbuf_t msdu, 2896 uint32_t transfer_id, u_int32_t len) 2897 { 2898 void *ce_tx_hdl = hif_get_ce_handle(osc, CE_HTT_TX_CE); 2899 2900 if (!ce_tx_hdl) 2901 return QDF_STATUS_E_NULL_VALUE; 2902 2903 return ce_send_single((struct CE_handle *)ce_tx_hdl, msdu, transfer_id, 2904 len); 2905 } 2906 qdf_export_symbol(hif_send_single); 2907 #endif 2908 2909 /** 2910 * hif_reg_write() - API to access hif specific function 2911 * hif_write32_mb. 2912 * @hif_ctx : HIF Context 2913 * @offset : offset on which value has to be written 2914 * @value : value to be written 2915 * 2916 * Return: None 2917 */ hif_reg_write(struct hif_opaque_softc * hif_ctx,uint32_t offset,uint32_t value)2918 void hif_reg_write(struct hif_opaque_softc *hif_ctx, uint32_t offset, 2919 uint32_t value) 2920 { 2921 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 2922 2923 hif_write32_mb(scn, scn->mem + offset, value); 2924 2925 } 2926 qdf_export_symbol(hif_reg_write); 2927 2928 /** 2929 * hif_reg_read() - API to access hif specific function 2930 * hif_read32_mb. 2931 * @hif_ctx : HIF Context 2932 * @offset : offset from which value has to be read 2933 * 2934 * Return: Read value 2935 */ hif_reg_read(struct hif_opaque_softc * hif_ctx,uint32_t offset)2936 uint32_t hif_reg_read(struct hif_opaque_softc *hif_ctx, uint32_t offset) 2937 { 2938 2939 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 2940 2941 return hif_read32_mb(scn, scn->mem + offset); 2942 } 2943 qdf_export_symbol(hif_reg_read); 2944 2945 /** 2946 * hif_ramdump_handler(): generic ramdump handler 2947 * @scn: struct hif_opaque_softc 2948 * 2949 * Return: None 2950 */ hif_ramdump_handler(struct hif_opaque_softc * scn)2951 void hif_ramdump_handler(struct hif_opaque_softc *scn) 2952 { 2953 if (hif_get_bus_type(scn) == QDF_BUS_TYPE_USB) 2954 hif_usb_ramdump_handler(scn); 2955 } 2956 hif_pm_get_wake_irq_type(struct hif_opaque_softc * hif_ctx)2957 hif_pm_wake_irq_type hif_pm_get_wake_irq_type(struct hif_opaque_softc *hif_ctx) 2958 { 2959 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 2960 2961 return scn->wake_irq_type; 2962 } 2963 hif_wake_interrupt_handler(int irq,void * context)2964 irqreturn_t hif_wake_interrupt_handler(int irq, void *context) 2965 { 2966 struct hif_softc *scn = context; 2967 2968 hif_info("wake interrupt received on irq %d", irq); 2969 2970 hif_rtpm_set_monitor_wake_intr(0); 2971 hif_rtpm_request_resume(); 2972 2973 if (scn->initial_wakeup_cb) 2974 scn->initial_wakeup_cb(scn->initial_wakeup_priv); 2975 2976 if (hif_is_ut_suspended(scn)) 2977 hif_ut_fw_resume(scn); 2978 2979 qdf_pm_system_wakeup(); 2980 2981 return IRQ_HANDLED; 2982 } 2983 hif_set_initial_wakeup_cb(struct hif_opaque_softc * hif_ctx,void (* callback)(void *),void * priv)2984 void hif_set_initial_wakeup_cb(struct hif_opaque_softc *hif_ctx, 2985 void (*callback)(void *), 2986 void *priv) 2987 { 2988 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 2989 2990 scn->initial_wakeup_cb = callback; 2991 scn->initial_wakeup_priv = priv; 2992 } 2993 hif_set_ce_service_max_yield_time(struct hif_opaque_softc * hif,uint32_t ce_service_max_yield_time)2994 void hif_set_ce_service_max_yield_time(struct hif_opaque_softc *hif, 2995 uint32_t ce_service_max_yield_time) 2996 { 2997 struct hif_softc *hif_ctx = HIF_GET_SOFTC(hif); 2998 2999 hif_ctx->ce_service_max_yield_time = 3000 ce_service_max_yield_time * 1000; 3001 } 3002 3003 unsigned long long hif_get_ce_service_max_yield_time(struct hif_opaque_softc * hif)3004 hif_get_ce_service_max_yield_time(struct hif_opaque_softc *hif) 3005 { 3006 struct hif_softc *hif_ctx = HIF_GET_SOFTC(hif); 3007 3008 return hif_ctx->ce_service_max_yield_time; 3009 } 3010 hif_set_ce_service_max_rx_ind_flush(struct hif_opaque_softc * hif,uint8_t ce_service_max_rx_ind_flush)3011 void hif_set_ce_service_max_rx_ind_flush(struct hif_opaque_softc *hif, 3012 uint8_t ce_service_max_rx_ind_flush) 3013 { 3014 struct hif_softc *hif_ctx = HIF_GET_SOFTC(hif); 3015 3016 if (ce_service_max_rx_ind_flush == 0 || 3017 ce_service_max_rx_ind_flush > MSG_FLUSH_NUM) 3018 hif_ctx->ce_service_max_rx_ind_flush = MSG_FLUSH_NUM; 3019 else 3020 hif_ctx->ce_service_max_rx_ind_flush = 3021 ce_service_max_rx_ind_flush; 3022 } 3023 3024 #ifdef SYSTEM_PM_CHECK __hif_system_pm_set_state(struct hif_opaque_softc * hif,enum hif_system_pm_state state)3025 void __hif_system_pm_set_state(struct hif_opaque_softc *hif, 3026 enum hif_system_pm_state state) 3027 { 3028 struct hif_softc *hif_ctx = HIF_GET_SOFTC(hif); 3029 3030 qdf_atomic_set(&hif_ctx->sys_pm_state, state); 3031 } 3032 hif_system_pm_get_state(struct hif_opaque_softc * hif)3033 int32_t hif_system_pm_get_state(struct hif_opaque_softc *hif) 3034 { 3035 struct hif_softc *hif_ctx = HIF_GET_SOFTC(hif); 3036 3037 return qdf_atomic_read(&hif_ctx->sys_pm_state); 3038 } 3039 hif_system_pm_state_check(struct hif_opaque_softc * hif)3040 int hif_system_pm_state_check(struct hif_opaque_softc *hif) 3041 { 3042 struct hif_softc *hif_ctx = HIF_GET_SOFTC(hif); 3043 int32_t sys_pm_state; 3044 3045 if (!hif_ctx) { 3046 hif_err("hif context is null"); 3047 return -EFAULT; 3048 } 3049 3050 sys_pm_state = qdf_atomic_read(&hif_ctx->sys_pm_state); 3051 if (sys_pm_state == HIF_SYSTEM_PM_STATE_BUS_SUSPENDING || 3052 sys_pm_state == HIF_SYSTEM_PM_STATE_BUS_SUSPENDED) { 3053 hif_info("Triggering system wakeup"); 3054 qdf_pm_system_wakeup(); 3055 return -EAGAIN; 3056 } 3057 3058 return 0; 3059 } 3060 #endif 3061 #ifdef WLAN_FEATURE_AFFINITY_MGR 3062 /* 3063 * hif_audio_cpu_affinity_allowed() - Check if audio cpu affinity allowed 3064 * 3065 * @scn: hif handle 3066 * @cfg: hif affinity manager configuration for IRQ 3067 * @audio_taken_cpu: Current CPUs which are taken by audio. 3068 * @current_time: Current system time. 3069 * 3070 * This API checks for 2 conditions 3071 * 1) Last audio taken mask and current taken mask are different 3072 * 2) Last time when IRQ was affined away due to audio taken CPUs is 3073 * more than time threshold (5 Seconds in current case). 3074 * If both condition satisfies then only return true. 3075 * 3076 * Return: bool: true if it is allowed to affine away audio taken cpus. 3077 */ 3078 static inline bool hif_audio_cpu_affinity_allowed(struct hif_softc * scn,struct hif_cpu_affinity * cfg,qdf_cpu_mask audio_taken_cpu,uint64_t current_time)3079 hif_audio_cpu_affinity_allowed(struct hif_softc *scn, 3080 struct hif_cpu_affinity *cfg, 3081 qdf_cpu_mask audio_taken_cpu, 3082 uint64_t current_time) 3083 { 3084 if (!qdf_cpumask_equal(&audio_taken_cpu, &cfg->walt_taken_mask) && 3085 (qdf_log_timestamp_to_usecs(current_time - 3086 cfg->last_affined_away) 3087 < scn->time_threshold)) 3088 return false; 3089 return true; 3090 } 3091 3092 /* 3093 * hif_affinity_mgr_check_update_mask() - Check if cpu mask need to be updated 3094 * 3095 * @scn: hif handle 3096 * @cfg: hif affinity manager configuration for IRQ 3097 * @audio_taken_cpu: Current CPUs which are taken by audio. 3098 * @cpu_mask: CPU mask which need to be updated. 3099 * @current_time: Current system time. 3100 * 3101 * This API checks if Pro audio use case is running and if cpu_mask need 3102 * to be updated 3103 * 3104 * Return: QDF_STATUS 3105 */ 3106 static inline QDF_STATUS hif_affinity_mgr_check_update_mask(struct hif_softc * scn,struct hif_cpu_affinity * cfg,qdf_cpu_mask audio_taken_cpu,qdf_cpu_mask * cpu_mask,uint64_t current_time)3107 hif_affinity_mgr_check_update_mask(struct hif_softc *scn, 3108 struct hif_cpu_affinity *cfg, 3109 qdf_cpu_mask audio_taken_cpu, 3110 qdf_cpu_mask *cpu_mask, 3111 uint64_t current_time) 3112 { 3113 qdf_cpu_mask allowed_mask; 3114 3115 /* 3116 * Case 1: audio_taken_mask is empty 3117 * Check if passed cpu_mask and wlan_requested_mask is same or not. 3118 * If both mask are different copy wlan_requested_mask(IRQ affinity 3119 * mask requested by WLAN) to cpu_mask. 3120 * 3121 * Case 2: audio_taken_mask is not empty 3122 * 1. Only allow update if last time when IRQ was affined away due to 3123 * audio taken CPUs is more than 5 seconds or update is requested 3124 * by WLAN 3125 * 2. Only allow silver cores to be affined away. 3126 * 3. Check if any allowed CPUs for audio use case is set in cpu_mask. 3127 * i. If any CPU mask is set, mask out that CPU from the cpu_mask 3128 * ii. If after masking out audio taken cpu(Silver cores) cpu_mask 3129 * is empty, set mask to all cpu except cpus taken by audio. 3130 * Example: 3131 *| Audio mask | mask allowed | cpu_mask | WLAN req mask | new cpu_mask| 3132 *| 0x00 | 0x00 | 0x0C | 0x0C | 0x0C | 3133 *| 0x00 | 0x00 | 0x03 | 0x03 | 0x03 | 3134 *| 0x00 | 0x00 | 0xFC | 0x03 | 0x03 | 3135 *| 0x00 | 0x00 | 0x03 | 0x0C | 0x0C | 3136 *| 0x0F | 0x03 | 0x0C | 0x0C | 0x0C | 3137 *| 0x0F | 0x03 | 0x03 | 0x03 | 0xFC | 3138 *| 0x03 | 0x03 | 0x0C | 0x0C | 0x0C | 3139 *| 0x03 | 0x03 | 0x03 | 0x03 | 0xFC | 3140 *| 0x03 | 0x03 | 0xFC | 0x03 | 0xFC | 3141 *| 0xF0 | 0x00 | 0x0C | 0x0C | 0x0C | 3142 *| 0xF0 | 0x00 | 0x03 | 0x03 | 0x03 | 3143 */ 3144 3145 /* Check if audio taken mask is empty*/ 3146 if (qdf_likely(qdf_cpumask_empty(&audio_taken_cpu))) { 3147 /* If CPU mask requested by WLAN for the IRQ and 3148 * cpu_mask passed CPU mask set for IRQ is different 3149 * Copy requested mask into cpu_mask and return 3150 */ 3151 if (qdf_unlikely(!qdf_cpumask_equal(cpu_mask, 3152 &cfg->wlan_requested_mask))) { 3153 qdf_cpumask_copy(cpu_mask, &cfg->wlan_requested_mask); 3154 return QDF_STATUS_SUCCESS; 3155 } 3156 return QDF_STATUS_E_ALREADY; 3157 } 3158 3159 if (!(hif_audio_cpu_affinity_allowed(scn, cfg, audio_taken_cpu, 3160 current_time) || 3161 cfg->update_requested)) 3162 return QDF_STATUS_E_AGAIN; 3163 3164 /* Only allow Silver cores to be affine away */ 3165 qdf_cpumask_and(&allowed_mask, &scn->allowed_mask, &audio_taken_cpu); 3166 if (qdf_cpumask_intersects(cpu_mask, &allowed_mask)) { 3167 /* If any of taken CPU(Silver cores) mask is set in cpu_mask, 3168 * mask out the audio taken CPUs from the cpu_mask. 3169 */ 3170 qdf_cpumask_andnot(cpu_mask, &cfg->wlan_requested_mask, 3171 &allowed_mask); 3172 /* If cpu_mask is empty set it to all CPUs 3173 * except taken by audio(Silver cores) 3174 */ 3175 if (qdf_unlikely(qdf_cpumask_empty(cpu_mask))) 3176 qdf_cpumask_complement(cpu_mask, &allowed_mask); 3177 return QDF_STATUS_SUCCESS; 3178 } 3179 3180 return QDF_STATUS_E_ALREADY; 3181 } 3182 3183 static inline QDF_STATUS hif_check_and_affine_irq(struct hif_softc * scn,struct hif_cpu_affinity * cfg,qdf_cpu_mask audio_taken_cpu,qdf_cpu_mask cpu_mask,uint64_t current_time)3184 hif_check_and_affine_irq(struct hif_softc *scn, struct hif_cpu_affinity *cfg, 3185 qdf_cpu_mask audio_taken_cpu, qdf_cpu_mask cpu_mask, 3186 uint64_t current_time) 3187 { 3188 QDF_STATUS status; 3189 3190 status = hif_affinity_mgr_check_update_mask(scn, cfg, 3191 audio_taken_cpu, 3192 &cpu_mask, 3193 current_time); 3194 /* Set IRQ affinity if CPU mask was updated */ 3195 if (QDF_IS_STATUS_SUCCESS(status)) { 3196 status = hif_irq_set_affinity_hint(cfg->irq, 3197 &cpu_mask); 3198 if (QDF_IS_STATUS_SUCCESS(status)) { 3199 /* Store audio taken CPU mask */ 3200 qdf_cpumask_copy(&cfg->walt_taken_mask, 3201 &audio_taken_cpu); 3202 /* Store CPU mask which was set for IRQ*/ 3203 qdf_cpumask_copy(&cfg->current_irq_mask, 3204 &cpu_mask); 3205 /* Set time when IRQ affinity was updated */ 3206 cfg->last_updated = current_time; 3207 if (hif_audio_cpu_affinity_allowed(scn, cfg, 3208 audio_taken_cpu, 3209 current_time)) 3210 /* If CPU mask was updated due to CPU 3211 * taken by audio, update 3212 * last_affined_away time 3213 */ 3214 cfg->last_affined_away = current_time; 3215 } 3216 } 3217 3218 return status; 3219 } 3220 hif_affinity_mgr_affine_irq(struct hif_softc * scn)3221 void hif_affinity_mgr_affine_irq(struct hif_softc *scn) 3222 { 3223 bool audio_affinity_allowed = false; 3224 int i, j, ce_id; 3225 uint64_t current_time; 3226 char cpu_str[10]; 3227 QDF_STATUS status; 3228 qdf_cpu_mask cpu_mask, audio_taken_cpu; 3229 struct HIF_CE_state *hif_state; 3230 struct hif_exec_context *hif_ext_group; 3231 struct CE_attr *host_ce_conf; 3232 struct HIF_CE_state *ce_sc; 3233 struct hif_cpu_affinity *cfg; 3234 3235 if (!scn->affinity_mgr_supported) 3236 return; 3237 3238 current_time = hif_get_log_timestamp(); 3239 /* Get CPU mask for audio taken CPUs */ 3240 audio_taken_cpu = qdf_walt_get_cpus_taken(); 3241 3242 ce_sc = HIF_GET_CE_STATE(scn); 3243 host_ce_conf = ce_sc->host_ce_config; 3244 for (ce_id = 0; ce_id < scn->ce_count; ce_id++) { 3245 if (host_ce_conf[ce_id].flags & CE_ATTR_DISABLE_INTR) 3246 continue; 3247 cfg = &scn->ce_irq_cpu_mask[ce_id]; 3248 qdf_cpumask_copy(&cpu_mask, &cfg->current_irq_mask); 3249 status = 3250 hif_check_and_affine_irq(scn, cfg, audio_taken_cpu, 3251 cpu_mask, current_time); 3252 if (QDF_IS_STATUS_SUCCESS(status)) 3253 audio_affinity_allowed = true; 3254 } 3255 3256 hif_state = HIF_GET_CE_STATE(scn); 3257 for (i = 0; i < hif_state->hif_num_extgroup; i++) { 3258 hif_ext_group = hif_state->hif_ext_group[i]; 3259 for (j = 0; j < hif_ext_group->numirq; j++) { 3260 cfg = &scn->irq_cpu_mask[hif_ext_group->grp_id][j]; 3261 qdf_cpumask_copy(&cpu_mask, &cfg->current_irq_mask); 3262 status = 3263 hif_check_and_affine_irq(scn, cfg, audio_taken_cpu, 3264 cpu_mask, current_time); 3265 if (QDF_IS_STATUS_SUCCESS(status)) { 3266 qdf_atomic_set(&hif_ext_group->force_napi_complete, -1); 3267 audio_affinity_allowed = true; 3268 } 3269 } 3270 } 3271 if (audio_affinity_allowed) { 3272 qdf_thread_cpumap_print_to_pagebuf(false, cpu_str, 3273 &audio_taken_cpu); 3274 hif_info("Audio taken CPU mask: %s", cpu_str); 3275 } 3276 } 3277 3278 static inline QDF_STATUS hif_affinity_mgr_set_irq_affinity(struct hif_softc * scn,uint32_t irq,struct hif_cpu_affinity * cfg,qdf_cpu_mask * cpu_mask)3279 hif_affinity_mgr_set_irq_affinity(struct hif_softc *scn, uint32_t irq, 3280 struct hif_cpu_affinity *cfg, 3281 qdf_cpu_mask *cpu_mask) 3282 { 3283 uint64_t current_time; 3284 char cpu_str[10]; 3285 QDF_STATUS status, mask_updated; 3286 qdf_cpu_mask audio_taken_cpu = qdf_walt_get_cpus_taken(); 3287 3288 current_time = hif_get_log_timestamp(); 3289 qdf_cpumask_copy(&cfg->wlan_requested_mask, cpu_mask); 3290 cfg->update_requested = true; 3291 mask_updated = hif_affinity_mgr_check_update_mask(scn, cfg, 3292 audio_taken_cpu, 3293 cpu_mask, 3294 current_time); 3295 status = hif_irq_set_affinity_hint(irq, cpu_mask); 3296 if (QDF_IS_STATUS_SUCCESS(status)) { 3297 qdf_cpumask_copy(&cfg->walt_taken_mask, &audio_taken_cpu); 3298 qdf_cpumask_copy(&cfg->current_irq_mask, cpu_mask); 3299 if (QDF_IS_STATUS_SUCCESS(mask_updated)) { 3300 cfg->last_updated = current_time; 3301 if (hif_audio_cpu_affinity_allowed(scn, cfg, 3302 audio_taken_cpu, 3303 current_time)) { 3304 cfg->last_affined_away = current_time; 3305 qdf_thread_cpumap_print_to_pagebuf(false, 3306 cpu_str, 3307 &audio_taken_cpu); 3308 hif_info_rl("Audio taken CPU mask: %s", 3309 cpu_str); 3310 } 3311 } 3312 } 3313 cfg->update_requested = false; 3314 return status; 3315 } 3316 3317 QDF_STATUS hif_affinity_mgr_set_qrg_irq_affinity(struct hif_softc * scn,uint32_t irq,uint32_t grp_id,uint32_t irq_index,qdf_cpu_mask * cpu_mask)3318 hif_affinity_mgr_set_qrg_irq_affinity(struct hif_softc *scn, uint32_t irq, 3319 uint32_t grp_id, uint32_t irq_index, 3320 qdf_cpu_mask *cpu_mask) 3321 { 3322 struct hif_cpu_affinity *cfg; 3323 3324 if (!scn->affinity_mgr_supported) 3325 return hif_irq_set_affinity_hint(irq, cpu_mask); 3326 3327 cfg = &scn->irq_cpu_mask[grp_id][irq_index]; 3328 return hif_affinity_mgr_set_irq_affinity(scn, irq, cfg, cpu_mask); 3329 } 3330 3331 QDF_STATUS hif_affinity_mgr_set_ce_irq_affinity(struct hif_softc * scn,uint32_t irq,uint32_t ce_id,qdf_cpu_mask * cpu_mask)3332 hif_affinity_mgr_set_ce_irq_affinity(struct hif_softc *scn, uint32_t irq, 3333 uint32_t ce_id, qdf_cpu_mask *cpu_mask) 3334 { 3335 struct hif_cpu_affinity *cfg; 3336 3337 if (!scn->affinity_mgr_supported) 3338 return hif_irq_set_affinity_hint(irq, cpu_mask); 3339 3340 cfg = &scn->ce_irq_cpu_mask[ce_id]; 3341 return hif_affinity_mgr_set_irq_affinity(scn, irq, cfg, cpu_mask); 3342 } 3343 3344 void hif_affinity_mgr_init_ce_irq(struct hif_softc * scn,int id,int irq)3345 hif_affinity_mgr_init_ce_irq(struct hif_softc *scn, int id, int irq) 3346 { 3347 unsigned int cpus; 3348 qdf_cpu_mask cpu_mask = {0}; 3349 struct hif_cpu_affinity *cfg = NULL; 3350 3351 if (!scn->affinity_mgr_supported) 3352 return; 3353 3354 /* Set CPU Mask to Silver core */ 3355 qdf_for_each_possible_cpu(cpus) 3356 if (qdf_topology_physical_package_id(cpus) == 3357 CPU_CLUSTER_TYPE_LITTLE) 3358 qdf_cpumask_set_cpu(cpus, &cpu_mask); 3359 3360 cfg = &scn->ce_irq_cpu_mask[id]; 3361 qdf_cpumask_copy(&cfg->current_irq_mask, &cpu_mask); 3362 qdf_cpumask_copy(&cfg->wlan_requested_mask, &cpu_mask); 3363 cfg->irq = irq; 3364 cfg->last_updated = 0; 3365 cfg->last_affined_away = 0; 3366 cfg->update_requested = false; 3367 } 3368 3369 void hif_affinity_mgr_init_grp_irq(struct hif_softc * scn,int grp_id,int irq_num,int irq)3370 hif_affinity_mgr_init_grp_irq(struct hif_softc *scn, int grp_id, 3371 int irq_num, int irq) 3372 { 3373 unsigned int cpus; 3374 qdf_cpu_mask cpu_mask = {0}; 3375 struct hif_cpu_affinity *cfg = NULL; 3376 3377 if (!scn->affinity_mgr_supported) 3378 return; 3379 3380 /* Set CPU Mask to Silver core */ 3381 qdf_for_each_possible_cpu(cpus) 3382 if (qdf_topology_physical_package_id(cpus) == 3383 CPU_CLUSTER_TYPE_LITTLE) 3384 qdf_cpumask_set_cpu(cpus, &cpu_mask); 3385 3386 cfg = &scn->irq_cpu_mask[grp_id][irq_num]; 3387 qdf_cpumask_copy(&cfg->current_irq_mask, &cpu_mask); 3388 qdf_cpumask_copy(&cfg->wlan_requested_mask, &cpu_mask); 3389 cfg->irq = irq; 3390 cfg->last_updated = 0; 3391 cfg->last_affined_away = 0; 3392 cfg->update_requested = false; 3393 } 3394 #endif 3395 3396 #if defined(HIF_CPU_PERF_AFFINE_MASK) || \ 3397 defined(FEATURE_ENABLE_CE_DP_IRQ_AFFINE) hif_config_irq_set_perf_affinity_hint(struct hif_opaque_softc * hif_ctx)3398 void hif_config_irq_set_perf_affinity_hint( 3399 struct hif_opaque_softc *hif_ctx) 3400 { 3401 struct hif_softc *scn = HIF_GET_SOFTC(hif_ctx); 3402 3403 hif_config_irq_affinity(scn); 3404 } 3405 3406 qdf_export_symbol(hif_config_irq_set_perf_affinity_hint); 3407 #endif 3408