/linux-6.12.1/drivers/ufs/core/ |
D | ufshcd.c | 113 static bool is_mcq_supported(struct ufs_hba *hba) in is_mcq_supported() argument 115 return hba->mcq_sup && use_mcq_mode; in is_mcq_supported() 155 int ufshcd_dump_regs(struct ufs_hba *hba, size_t offset, size_t len, in ufshcd_dump_regs() argument 173 regs[pos / 4] = ufshcd_readl(hba, offset + pos); in ufshcd_dump_regs() 261 static bool ufshcd_is_ufs_dev_busy(struct ufs_hba *hba) in ufshcd_is_ufs_dev_busy() argument 263 return (hba->clk_gating.active_reqs || hba->outstanding_reqs || hba->outstanding_tasks || in ufshcd_is_ufs_dev_busy() 264 hba->active_uic_cmd || hba->uic_async_done); in ufshcd_is_ufs_dev_busy() 295 static irqreturn_t ufshcd_tmc_handler(struct ufs_hba *hba); 297 static int ufshcd_reset_and_restore(struct ufs_hba *hba); 299 static int ufshcd_clear_tm_cmd(struct ufs_hba *hba, int tag); [all …]
|
D | ufshcd-priv.h | 9 static inline bool ufshcd_is_user_access_allowed(struct ufs_hba *hba) in ufshcd_is_user_access_allowed() argument 11 return !hba->shutting_down; in ufshcd_is_user_access_allowed() 14 void ufshcd_schedule_eh_work(struct ufs_hba *hba); 17 struct ufs_hba *hba) in ufshcd_keep_autobkops_enabled_except_suspend() argument 19 return hba->caps & UFSHCD_CAP_KEEP_AUTO_BKOPS_ENABLED_EXCEPT_SUSPEND; in ufshcd_keep_autobkops_enabled_except_suspend() 22 static inline u8 ufshcd_wb_get_query_index(struct ufs_hba *hba) in ufshcd_wb_get_query_index() argument 24 if (hba->dev_info.wb_buffer_type == WB_BUF_MODE_LU_DEDICATED) in ufshcd_wb_get_query_index() 25 return hba->dev_info.wb_dedicated_lu; in ufshcd_wb_get_query_index() 29 static inline bool ufshcd_is_wb_buf_flush_allowed(struct ufs_hba *hba) in ufshcd_is_wb_buf_flush_allowed() argument 31 return ufshcd_is_wb_allowed(hba) && in ufshcd_is_wb_buf_flush_allowed() [all …]
|
D | ufs-mcq.c | 91 void ufshcd_mcq_config_mac(struct ufs_hba *hba, u32 max_active_cmds) in ufshcd_mcq_config_mac() argument 95 val = ufshcd_readl(hba, REG_UFS_MCQ_CFG); in ufshcd_mcq_config_mac() 98 ufshcd_writel(hba, val, REG_UFS_MCQ_CFG); in ufshcd_mcq_config_mac() 111 struct ufs_hw_queue *ufshcd_mcq_req_to_hwq(struct ufs_hba *hba, in ufshcd_mcq_req_to_hwq() argument 116 return hctx ? &hba->uhq[hctx->queue_num] : NULL; in ufshcd_mcq_req_to_hwq() 126 unsigned int ufshcd_mcq_queue_cfg_addr(struct ufs_hba *hba) in ufshcd_mcq_queue_cfg_addr() argument 128 return FIELD_GET(QCFGPTR, hba->mcq_capabilities) * 0x200; in ufshcd_mcq_queue_cfg_addr() 143 int ufshcd_mcq_decide_queue_depth(struct ufs_hba *hba) in ufshcd_mcq_decide_queue_depth() argument 147 if (!hba->vops || !hba->vops->get_hba_mac) { in ufshcd_mcq_decide_queue_depth() 153 hba->capabilities = in ufshcd_mcq_decide_queue_depth() [all …]
|
D | ufshcd-crypto.c | 20 static int ufshcd_program_key(struct ufs_hba *hba, in ufshcd_program_key() argument 24 u32 slot_offset = hba->crypto_cfg_register + slot * sizeof(*cfg); in ufshcd_program_key() 27 ufshcd_hold(hba); in ufshcd_program_key() 29 if (hba->vops && hba->vops->program_key) { in ufshcd_program_key() 30 err = hba->vops->program_key(hba, cfg, slot); in ufshcd_program_key() 35 ufshcd_writel(hba, 0, slot_offset + 16 * sizeof(cfg->reg_val[0])); in ufshcd_program_key() 37 ufshcd_writel(hba, le32_to_cpu(cfg->reg_val[i]), in ufshcd_program_key() 41 ufshcd_writel(hba, le32_to_cpu(cfg->reg_val[17]), in ufshcd_program_key() 44 ufshcd_writel(hba, le32_to_cpu(cfg->reg_val[16]), in ufshcd_program_key() 47 ufshcd_release(hba); in ufshcd_program_key() [all …]
|
D | ufs-debugfs.c | 36 struct ufs_hba *hba = hba_from_file(s->file); in ufs_debugfs_stats_show() local 37 struct ufs_event_hist *e = hba->ufs_stats.event; in ufs_debugfs_stats_show() 62 struct ufs_hba *hba = data; in ee_usr_mask_get() local 64 *val = hba->ee_usr_mask; in ee_usr_mask_get() 68 static int ufs_debugfs_get_user_access(struct ufs_hba *hba) in ufs_debugfs_get_user_access() argument 69 __acquires(&hba->host_sem) in ufs_debugfs_get_user_access() 71 down(&hba->host_sem); in ufs_debugfs_get_user_access() 72 if (!ufshcd_is_user_access_allowed(hba)) { in ufs_debugfs_get_user_access() 73 up(&hba->host_sem); in ufs_debugfs_get_user_access() 76 ufshcd_rpm_get_sync(hba); in ufs_debugfs_get_user_access() [all …]
|
D | ufs-hwmon.c | 14 struct ufs_hba *hba; member 18 static int ufs_read_temp_enable(struct ufs_hba *hba, u8 mask, long *val) in ufs_read_temp_enable() argument 23 err = ufshcd_query_attr(hba, UPIU_QUERY_OPCODE_READ_ATTR, QUERY_ATTR_IDN_EE_CONTROL, 0, 0, in ufs_read_temp_enable() 33 static int ufs_get_temp(struct ufs_hba *hba, enum attr_idn idn, long *val) in ufs_get_temp() argument 38 err = ufshcd_query_attr(hba, UPIU_QUERY_OPCODE_READ_ATTR, idn, 0, 0, &value); in ufs_get_temp() 54 struct ufs_hba *hba = data->hba; in ufs_hwmon_read() local 57 down(&hba->host_sem); in ufs_hwmon_read() 59 if (!ufshcd_is_user_access_allowed(hba)) { in ufs_hwmon_read() 60 up(&hba->host_sem); in ufs_hwmon_read() 64 ufshcd_rpm_get_sync(hba); in ufs_hwmon_read() [all …]
|
D | ufs-sysfs.c | 89 struct ufs_hba *hba = dev_get_drvdata(dev); in ufs_sysfs_pm_lvl_store() local 90 struct ufs_dev_info *dev_info = &hba->dev_info; in ufs_sysfs_pm_lvl_store() 100 (!(hba->caps & UFSHCD_CAP_DEEPSLEEP) || in ufs_sysfs_pm_lvl_store() 104 spin_lock_irqsave(hba->host->host_lock, flags); in ufs_sysfs_pm_lvl_store() 106 hba->rpm_lvl = value; in ufs_sysfs_pm_lvl_store() 108 hba->spm_lvl = value; in ufs_sysfs_pm_lvl_store() 109 spin_unlock_irqrestore(hba->host->host_lock, flags); in ufs_sysfs_pm_lvl_store() 116 struct ufs_hba *hba = dev_get_drvdata(dev); in rpm_lvl_show() local 118 return sysfs_emit(buf, "%d\n", hba->rpm_lvl); in rpm_lvl_show() 130 struct ufs_hba *hba = dev_get_drvdata(dev); in rpm_target_dev_state_show() local [all …]
|
D | ufshcd-crypto.h | 40 static inline int ufshcd_crypto_fill_prdt(struct ufs_hba *hba, in ufshcd_crypto_fill_prdt() argument 46 if (crypt_ctx && hba->vops && hba->vops->fill_crypto_prdt) in ufshcd_crypto_fill_prdt() 47 return hba->vops->fill_crypto_prdt(hba, crypt_ctx, in ufshcd_crypto_fill_prdt() 53 static inline void ufshcd_crypto_clear_prdt(struct ufs_hba *hba, in ufshcd_crypto_clear_prdt() argument 56 if (!(hba->quirks & UFSHCD_QUIRK_KEYS_IN_PRDT)) in ufshcd_crypto_clear_prdt() 64 ufshcd_sg_entry_size(hba) * scsi_sg_count(lrbp->cmd)); in ufshcd_crypto_clear_prdt() 67 bool ufshcd_crypto_enable(struct ufs_hba *hba); 69 int ufshcd_hba_init_crypto_capabilities(struct ufs_hba *hba); 71 void ufshcd_init_crypto(struct ufs_hba *hba); 73 void ufshcd_crypto_register(struct ufs_hba *hba, struct request_queue *q); [all …]
|
D | ufs_bsg.c | 16 static int ufs_bsg_get_query_desc_size(struct ufs_hba *hba, int *desc_len, in ufs_bsg_get_query_desc_size() argument 29 static int ufs_bsg_alloc_desc_buffer(struct ufs_hba *hba, struct bsg_job *job, in ufs_bsg_alloc_desc_buffer() argument 42 if (ufs_bsg_get_query_desc_size(hba, desc_len, qr)) { in ufs_bsg_alloc_desc_buffer() 43 dev_err(hba->dev, "Illegal desc size\n"); in ufs_bsg_alloc_desc_buffer() 48 dev_err(hba->dev, "Illegal desc size\n"); in ufs_bsg_alloc_desc_buffer() 67 static int ufs_bsg_exec_advanced_rpmb_req(struct ufs_hba *hba, struct bsg_job *job) in ufs_bsg_exec_advanced_rpmb_req() argument 79 if (hba->ufs_version < ufshci_version(4, 0) || !hba->dev_info.b_advanced_rpmb_en) in ufs_bsg_exec_advanced_rpmb_req() 111 sg_cnt = dma_map_sg(hba->host->dma_dev, payload->sg_list, payload->sg_cnt, dir); in ufs_bsg_exec_advanced_rpmb_req() 118 ret = ufshcd_advanced_rpmb_req_handler(hba, &rpmb_request->bsg_request.upiu_req, in ufs_bsg_exec_advanced_rpmb_req() 123 dma_unmap_sg(hba->host->dma_dev, payload->sg_list, payload->sg_cnt, dir); in ufs_bsg_exec_advanced_rpmb_req() [all …]
|
/linux-6.12.1/drivers/ufs/host/ |
D | ufs-mediatek.c | 31 static int ufs_mtk_config_mcq(struct ufs_hba *hba, bool irq); 95 static bool ufs_mtk_is_boost_crypt_enabled(struct ufs_hba *hba) in ufs_mtk_is_boost_crypt_enabled() argument 97 struct ufs_mtk_host *host = ufshcd_get_variant(hba); in ufs_mtk_is_boost_crypt_enabled() 102 static bool ufs_mtk_is_va09_supported(struct ufs_hba *hba) in ufs_mtk_is_va09_supported() argument 104 struct ufs_mtk_host *host = ufshcd_get_variant(hba); in ufs_mtk_is_va09_supported() 109 static bool ufs_mtk_is_broken_vcc(struct ufs_hba *hba) in ufs_mtk_is_broken_vcc() argument 111 struct ufs_mtk_host *host = ufshcd_get_variant(hba); in ufs_mtk_is_broken_vcc() 116 static bool ufs_mtk_is_pmc_via_fastauto(struct ufs_hba *hba) in ufs_mtk_is_pmc_via_fastauto() argument 118 struct ufs_mtk_host *host = ufshcd_get_variant(hba); in ufs_mtk_is_pmc_via_fastauto() 123 static bool ufs_mtk_is_tx_skew_fix(struct ufs_hba *hba) in ufs_mtk_is_tx_skew_fix() argument [all …]
|
D | ufs-hisi.c | 25 static int ufs_hisi_check_hibern8(struct ufs_hba *hba) in ufs_hisi_check_hibern8() argument 33 err = ufshcd_dme_get(hba, UIC_ARG_MIB_SEL(MPHY_TX_FSM_STATE, 0), in ufs_hisi_check_hibern8() 35 err |= ufshcd_dme_get(hba, in ufs_hisi_check_hibern8() 50 err = ufshcd_dme_get(hba, UIC_ARG_MIB_SEL(MPHY_TX_FSM_STATE, 0), in ufs_hisi_check_hibern8() 52 err |= ufshcd_dme_get(hba, in ufs_hisi_check_hibern8() 57 dev_err(hba->dev, "%s: unable to get TX_FSM_STATE, err %d\n", in ufs_hisi_check_hibern8() 62 dev_err(hba->dev, "%s: invalid TX_FSM_STATE, lane0 = %d, lane1 = %d\n", in ufs_hisi_check_hibern8() 69 static void ufs_hisi_clk_init(struct ufs_hba *hba) in ufs_hisi_clk_init() argument 71 struct ufs_hisi_host *host = ufshcd_get_variant(hba); in ufs_hisi_clk_init() 83 static void ufs_hisi_soc_init(struct ufs_hba *hba) in ufs_hisi_soc_init() argument [all …]
|
D | ufshcd-pci.c | 26 void (*late_init)(struct ufs_hba *hba); 100 static int ufs_intel_hce_enable_notify(struct ufs_hba *hba, in ufs_intel_hce_enable_notify() argument 104 if (status == POST_CHANGE && hba->caps & UFSHCD_CAP_CRYPTO) { in ufs_intel_hce_enable_notify() 105 u32 hce = ufshcd_readl(hba, REG_CONTROLLER_ENABLE); in ufs_intel_hce_enable_notify() 108 ufshcd_writel(hba, hce, REG_CONTROLLER_ENABLE); in ufs_intel_hce_enable_notify() 114 static int ufs_intel_disable_lcc(struct ufs_hba *hba) in ufs_intel_disable_lcc() argument 119 ufshcd_dme_get(hba, attr, &lcc_enable); in ufs_intel_disable_lcc() 121 ufshcd_disable_host_tx_lcc(hba); in ufs_intel_disable_lcc() 126 static int ufs_intel_link_startup_notify(struct ufs_hba *hba, in ufs_intel_link_startup_notify() argument 133 err = ufs_intel_disable_lcc(hba); in ufs_intel_link_startup_notify() [all …]
|
D | ufs-qcom.c | 100 static int ufs_qcom_set_core_clk_ctrl(struct ufs_hba *hba, bool is_scale_up); 111 if (host->hba->caps & UFSHCD_CAP_CRYPTO) in ufs_qcom_ice_enable() 117 struct ufs_hba *hba = host->hba; in ufs_qcom_ice_init() local 118 struct device *dev = hba->dev; in ufs_qcom_ice_init() 131 hba->caps |= UFSHCD_CAP_CRYPTO; in ufs_qcom_ice_init() 138 if (host->hba->caps & UFSHCD_CAP_CRYPTO) in ufs_qcom_ice_resume() 146 if (host->hba->caps & UFSHCD_CAP_CRYPTO) in ufs_qcom_ice_suspend() 152 static int ufs_qcom_ice_program_key(struct ufs_hba *hba, in ufs_qcom_ice_program_key() argument 156 struct ufs_qcom_host *host = ufshcd_get_variant(hba); in ufs_qcom_ice_program_key() 162 cap = hba->crypto_cap_array[cfg->crypto_cap_idx]; in ufs_qcom_ice_program_key() [all …]
|
D | ufs-exynos.c | 224 struct ufs_hba *hba = ufs->hba; in exynosauto_ufs_post_hce_enable() local 227 ufshcd_rmwl(hba, MHCTRL_EN_VH_MASK, MHCTRL_EN_VH(1), MHCTRL); in exynosauto_ufs_post_hce_enable() 238 struct ufs_hba *hba = ufs->hba; in exynosauto_ufs_pre_link() local 245 ufshcd_dme_set(hba, UIC_ARG_MIB(0x200), 0x40); in exynosauto_ufs_pre_link() 247 ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(VND_RX_CLK_PRD, i), in exynosauto_ufs_pre_link() 249 ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(VND_RX_CLK_PRD_EN, i), 0x0); in exynosauto_ufs_pre_link() 251 ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(VND_RX_LINERESET_VALUE2, i), in exynosauto_ufs_pre_link() 253 ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(VND_RX_LINERESET_VALUE1, i), in exynosauto_ufs_pre_link() 255 ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(VND_RX_LINERESET_VALUE0, i), in exynosauto_ufs_pre_link() 258 ufshcd_dme_set(hba, UIC_ARG_MIB_SEL(0x2f, i), 0x79); in exynosauto_ufs_pre_link() [all …]
|
D | cdns-pltfrm.c | 37 static void cdns_ufs_get_l4_attr(struct ufs_hba *hba) in cdns_ufs_get_l4_attr() argument 39 struct cdns_ufs_host *host = ufshcd_get_variant(hba); in cdns_ufs_get_l4_attr() 41 ufshcd_dme_get(hba, UIC_ARG_MIB(T_PEERDEVICEID), in cdns_ufs_get_l4_attr() 43 ufshcd_dme_get(hba, UIC_ARG_MIB(T_PEERCPORTID), in cdns_ufs_get_l4_attr() 45 ufshcd_dme_get(hba, UIC_ARG_MIB(T_TRAFFICCLASS), in cdns_ufs_get_l4_attr() 47 ufshcd_dme_get(hba, UIC_ARG_MIB(T_PROTOCOLID), in cdns_ufs_get_l4_attr() 49 ufshcd_dme_get(hba, UIC_ARG_MIB(T_CPORTFLAGS), in cdns_ufs_get_l4_attr() 51 ufshcd_dme_get(hba, UIC_ARG_MIB(T_TXTOKENVALUE), in cdns_ufs_get_l4_attr() 53 ufshcd_dme_get(hba, UIC_ARG_MIB(T_RXTOKENVALUE), in cdns_ufs_get_l4_attr() 55 ufshcd_dme_get(hba, UIC_ARG_MIB(T_LOCALBUFFERSPACE), in cdns_ufs_get_l4_attr() [all …]
|
D | ufs-sprd.c | 23 static struct ufs_sprd_priv *ufs_sprd_get_priv_data(struct ufs_hba *hba) in ufs_sprd_get_priv_data() argument 25 struct ufs_sprd_host *host = ufshcd_get_variant(hba); in ufs_sprd_get_priv_data() 43 static void ufs_sprd_get_unipro_ver(struct ufs_hba *hba) in ufs_sprd_get_unipro_ver() argument 45 struct ufs_sprd_host *host = ufshcd_get_variant(hba); in ufs_sprd_get_unipro_ver() 47 if (ufshcd_dme_get(hba, UIC_ARG_MIB(PA_LOCALVERINFO), &host->unipro_ver)) in ufs_sprd_get_unipro_ver() 51 static void ufs_sprd_ctrl_uic_compl(struct ufs_hba *hba, bool enable) in ufs_sprd_ctrl_uic_compl() argument 53 u32 set = ufshcd_readl(hba, REG_INTERRUPT_ENABLE); in ufs_sprd_ctrl_uic_compl() 59 ufshcd_writel(hba, set, REG_INTERRUPT_ENABLE); in ufs_sprd_ctrl_uic_compl() 95 static int ufs_sprd_parse_dt(struct device *dev, struct ufs_hba *hba, struct ufs_sprd_host *host) in ufs_sprd_parse_dt() argument 132 static int ufs_sprd_common_init(struct ufs_hba *hba) in ufs_sprd_common_init() argument [all …]
|
/linux-6.12.1/drivers/scsi/ |
D | hptiop.c | 39 static int iop_send_sync_msg(struct hptiop_hba *hba, u32 msg, u32 millisec); 40 static void hptiop_finish_scsi_req(struct hptiop_hba *hba, u32 tag, 42 static void hptiop_host_request_callback_itl(struct hptiop_hba *hba, u32 tag); 43 static void hptiop_iop_request_callback_itl(struct hptiop_hba *hba, u32 tag); 44 static void hptiop_message_callback(struct hptiop_hba *hba, u32 msg); 46 static int iop_wait_ready_itl(struct hptiop_hba *hba, u32 millisec) in iop_wait_ready_itl() argument 52 req = readl(&hba->u.itl.iop->inbound_queue); in iop_wait_ready_itl() 59 writel(req, &hba->u.itl.iop->outbound_queue); in iop_wait_ready_itl() 60 readl(&hba->u.itl.iop->outbound_intstatus); in iop_wait_ready_itl() 67 static int iop_wait_ready_mv(struct hptiop_hba *hba, u32 millisec) in iop_wait_ready_mv() argument [all …]
|
D | stex.c | 391 static struct status_msg *stex_get_status(struct st_hba *hba) in stex_get_status() argument 393 struct status_msg *status = hba->status_buffer + hba->status_tail; in stex_get_status() 395 ++hba->status_tail; in stex_get_status() 396 hba->status_tail %= hba->sts_count+1; in stex_get_status() 409 static struct req_msg *stex_alloc_req(struct st_hba *hba) in stex_alloc_req() argument 411 struct req_msg *req = hba->dma_mem + hba->req_head * hba->rq_size; in stex_alloc_req() 413 ++hba->req_head; in stex_alloc_req() 414 hba->req_head %= hba->rq_count+1; in stex_alloc_req() 419 static struct req_msg *stex_ss_alloc_req(struct st_hba *hba) in stex_ss_alloc_req() argument 421 return (struct req_msg *)(hba->dma_mem + in stex_ss_alloc_req() [all …]
|
/linux-6.12.1/drivers/scsi/bnx2i/ |
D | bnx2i_init.c | 82 void bnx2i_identify_device(struct bnx2i_hba *hba, struct cnic_dev *dev) in bnx2i_identify_device() argument 84 hba->cnic_dev_type = 0; in bnx2i_identify_device() 86 if (hba->pci_did == PCI_DEVICE_ID_NX2_5706 || in bnx2i_identify_device() 87 hba->pci_did == PCI_DEVICE_ID_NX2_5706S) { in bnx2i_identify_device() 88 set_bit(BNX2I_NX2_DEV_5706, &hba->cnic_dev_type); in bnx2i_identify_device() 89 } else if (hba->pci_did == PCI_DEVICE_ID_NX2_5708 || in bnx2i_identify_device() 90 hba->pci_did == PCI_DEVICE_ID_NX2_5708S) { in bnx2i_identify_device() 91 set_bit(BNX2I_NX2_DEV_5708, &hba->cnic_dev_type); in bnx2i_identify_device() 92 } else if (hba->pci_did == PCI_DEVICE_ID_NX2_5709 || in bnx2i_identify_device() 93 hba->pci_did == PCI_DEVICE_ID_NX2_5709S) { in bnx2i_identify_device() [all …]
|
D | bnx2i_iscsi.c | 34 static int bnx2i_adapter_ready(struct bnx2i_hba *hba) in bnx2i_adapter_ready() argument 38 if (!hba || !test_bit(ADAPTER_STATE_UP, &hba->adapter_state) || in bnx2i_adapter_ready() 39 test_bit(ADAPTER_STATE_GOING_DOWN, &hba->adapter_state) || in bnx2i_adapter_ready() 40 test_bit(ADAPTER_STATE_LINK_DOWN, &hba->adapter_state)) in bnx2i_adapter_ready() 143 static int bnx2i_map_scsi_sg(struct bnx2i_hba *hba, struct bnx2i_cmd *cmd) in bnx2i_map_scsi_sg() argument 190 bd_count = bnx2i_map_scsi_sg(cmd->conn->hba, cmd); in bnx2i_iscsi_map_sg_list() 238 static int bnx2i_bind_conn_to_iscsi_cid(struct bnx2i_hba *hba, in bnx2i_bind_conn_to_iscsi_cid() argument 242 if (hba && hba->cid_que.conn_cid_tbl[iscsi_cid]) { in bnx2i_bind_conn_to_iscsi_cid() 248 hba->cid_que.conn_cid_tbl[iscsi_cid] = bnx2i_conn; in bnx2i_bind_conn_to_iscsi_cid() 258 struct bnx2i_conn *bnx2i_get_conn_from_id(struct bnx2i_hba *hba, in bnx2i_get_conn_from_id() argument [all …]
|
/linux-6.12.1/include/ufs/ |
D | ufshcd.h | 100 #define ufshcd_is_link_off(hba) ((hba)->uic_link_state == UIC_LINK_OFF_STATE) argument 101 #define ufshcd_is_link_active(hba) ((hba)->uic_link_state == \ argument 103 #define ufshcd_is_link_hibern8(hba) ((hba)->uic_link_state == \ argument 105 #define ufshcd_is_link_broken(hba) ((hba)->uic_link_state == \ argument 107 #define ufshcd_set_link_off(hba) ((hba)->uic_link_state = UIC_LINK_OFF_STATE) argument 108 #define ufshcd_set_link_active(hba) ((hba)->uic_link_state = \ argument 110 #define ufshcd_set_link_hibern8(hba) ((hba)->uic_link_state = \ argument 112 #define ufshcd_set_link_broken(hba) ((hba)->uic_link_state = \ argument 356 void (*setup_xfer_req)(struct ufs_hba *hba, int tag, 361 int (*apply_dev_quirks)(struct ufs_hba *hba); [all …]
|
/linux-6.12.1/drivers/parisc/ |
D | lba_pci.c | 111 static inline struct lba_device *LBA_DEV(struct pci_hba_data *hba) in LBA_DEV() argument 113 return container_of(hba, struct lba_device, hba); in LBA_DEV() 191 u8 first_bus = d->hba.hba_bus->busn_res.start; in lba_device_present() 192 u8 last_sub_bus = d->hba.hba_bus->busn_res.end; in lba_device_present() 207 error_config = READ_REG32(d->hba.base_addr + LBA_ERROR_CONFIG); \ 210 status_control = READ_REG32(d->hba.base_addr + LBA_STAT_CTL); \ 216 arb_mask = READ_REG32(d->hba.base_addr + LBA_ARB_MASK); \ 222 WRITE_REG32(0x1, d->hba.base_addr + LBA_ARB_MASK); \ 228 WRITE_REG32(error_config | LBA_SMART_MODE, d->hba.base_addr + LBA_ERROR_CONFIG); \ 237 WRITE_REG32(tok | PCI_VENDOR_ID, (d)->hba.base_addr + LBA_PCI_CFG_ADDR);\ [all …]
|
/linux-6.12.1/drivers/scsi/bnx2fc/ |
D | bnx2fc_fcoe.c | 78 static int bnx2fc_em_config(struct fc_lport *lport, struct bnx2fc_hba *hba); 79 static int bnx2fc_bind_adapter_devices(struct bnx2fc_hba *hba); 80 static void bnx2fc_unbind_adapter_devices(struct bnx2fc_hba *hba); 81 static int bnx2fc_bind_pcidev(struct bnx2fc_hba *hba); 82 static void bnx2fc_unbind_pcidev(struct bnx2fc_hba *hba); 93 static int bnx2fc_fw_init(struct bnx2fc_hba *hba); 94 static void bnx2fc_fw_destroy(struct bnx2fc_hba *hba); 193 struct bnx2fc_hba *hba = interface->hba; in bnx2fc_cleanup() local 198 mutex_lock(&hba->hba_mutex); in bnx2fc_cleanup() 199 spin_lock_bh(&hba->hba_lock); in bnx2fc_cleanup() [all …]
|
D | bnx2fc_hwi.c | 20 static void bnx2fc_fastpath_notification(struct bnx2fc_hba *hba, 22 static void bnx2fc_process_ofld_cmpl(struct bnx2fc_hba *hba, 24 static void bnx2fc_process_enable_conn_cmpl(struct bnx2fc_hba *hba, 26 static void bnx2fc_init_failure(struct bnx2fc_hba *hba, u32 err_code); 27 static void bnx2fc_process_conn_destroy_cmpl(struct bnx2fc_hba *hba, 30 int bnx2fc_send_stat_req(struct bnx2fc_hba *hba) in bnx2fc_send_stat_req() argument 42 stat_req.stat_params_addr_lo = (u32) hba->stats_buf_dma; in bnx2fc_send_stat_req() 43 stat_req.stat_params_addr_hi = (u32) ((u64)hba->stats_buf_dma >> 32); in bnx2fc_send_stat_req() 47 if (hba->cnic && hba->cnic->submit_kwqes) in bnx2fc_send_stat_req() 48 rc = hba->cnic->submit_kwqes(hba->cnic, kwqe_arr, num_kwqes); in bnx2fc_send_stat_req() [all …]
|
/linux-6.12.1/drivers/target/ |
D | target_core_hba.c | 111 struct se_hba *hba; in core_alloc_hba() local 114 hba = kzalloc(sizeof(*hba), GFP_KERNEL); in core_alloc_hba() 115 if (!hba) { in core_alloc_hba() 120 spin_lock_init(&hba->device_lock); in core_alloc_hba() 121 mutex_init(&hba->hba_access_mutex); in core_alloc_hba() 123 hba->hba_index = scsi_get_new_index(SCSI_INST_INDEX); in core_alloc_hba() 124 hba->hba_flags |= hba_flags; in core_alloc_hba() 126 hba->backend = core_get_backend(plugin_name); in core_alloc_hba() 127 if (!hba->backend) { in core_alloc_hba() 132 ret = hba->backend->ops->attach_hba(hba, plugin_dep_id); in core_alloc_hba() [all …]
|