/linux-6.12.1/drivers/crypto/intel/qat/qat_common/ |
D | adf_gen4_ras.c | 10 static void enable_errsou_reporting(void __iomem *csr) in enable_errsou_reporting() argument 13 ADF_CSR_WR(csr, ADF_GEN4_ERRMSK0, 0); in enable_errsou_reporting() 16 ADF_CSR_WR(csr, ADF_GEN4_ERRMSK1, 0); in enable_errsou_reporting() 22 ADF_CSR_WR(csr, ADF_GEN4_ERRMSK2, in enable_errsou_reporting() 30 ADF_CSR_WR(csr, ADF_GEN4_ERRMSK3, in enable_errsou_reporting() 35 static void disable_errsou_reporting(void __iomem *csr) in disable_errsou_reporting() argument 40 ADF_CSR_WR(csr, ADF_GEN4_ERRMSK0, ADF_GEN4_ERRSOU0_BIT); in disable_errsou_reporting() 43 ADF_CSR_WR(csr, ADF_GEN4_ERRMSK1, ADF_GEN4_ERRSOU1_BITMASK); in disable_errsou_reporting() 46 val = ADF_CSR_RD(csr, ADF_GEN4_ERRMSK2); in disable_errsou_reporting() 48 ADF_CSR_WR(csr, ADF_GEN4_ERRMSK2, val); in disable_errsou_reporting() [all …]
|
D | icp_qat_hw_20_comp.h | 23 ICP_QAT_FW_COMP_20_BUILD_CONFIG_LOWER(struct icp_qat_hw_comp_20_config_csr_lower csr) in ICP_QAT_FW_COMP_20_BUILD_CONFIG_LOWER() argument 27 QAT_FIELD_SET(val32, csr.algo, in ICP_QAT_FW_COMP_20_BUILD_CONFIG_LOWER() 30 QAT_FIELD_SET(val32, csr.sd, in ICP_QAT_FW_COMP_20_BUILD_CONFIG_LOWER() 33 QAT_FIELD_SET(val32, csr.edmm, in ICP_QAT_FW_COMP_20_BUILD_CONFIG_LOWER() 36 QAT_FIELD_SET(val32, csr.hbs, in ICP_QAT_FW_COMP_20_BUILD_CONFIG_LOWER() 39 QAT_FIELD_SET(val32, csr.lllbd, in ICP_QAT_FW_COMP_20_BUILD_CONFIG_LOWER() 42 QAT_FIELD_SET(val32, csr.mmctrl, in ICP_QAT_FW_COMP_20_BUILD_CONFIG_LOWER() 45 QAT_FIELD_SET(val32, csr.hash_col, in ICP_QAT_FW_COMP_20_BUILD_CONFIG_LOWER() 48 QAT_FIELD_SET(val32, csr.hash_update, in ICP_QAT_FW_COMP_20_BUILD_CONFIG_LOWER() 51 QAT_FIELD_SET(val32, csr.skip_ctrl, in ICP_QAT_FW_COMP_20_BUILD_CONFIG_LOWER() [all …]
|
/linux-6.12.1/arch/loongarch/kvm/ |
D | vcpu.c | 68 struct loongarch_csrs *csr = vcpu->arch.csr; in kvm_save_guest_pmu() local 70 kvm_save_hw_gcsr(csr, LOONGARCH_CSR_PERFCNTR0); in kvm_save_guest_pmu() 71 kvm_save_hw_gcsr(csr, LOONGARCH_CSR_PERFCNTR1); in kvm_save_guest_pmu() 72 kvm_save_hw_gcsr(csr, LOONGARCH_CSR_PERFCNTR2); in kvm_save_guest_pmu() 73 kvm_save_hw_gcsr(csr, LOONGARCH_CSR_PERFCNTR3); in kvm_save_guest_pmu() 74 kvm_read_clear_hw_gcsr(csr, LOONGARCH_CSR_PERFCTRL0); in kvm_save_guest_pmu() 75 kvm_read_clear_hw_gcsr(csr, LOONGARCH_CSR_PERFCTRL1); in kvm_save_guest_pmu() 76 kvm_read_clear_hw_gcsr(csr, LOONGARCH_CSR_PERFCTRL2); in kvm_save_guest_pmu() 77 kvm_read_clear_hw_gcsr(csr, LOONGARCH_CSR_PERFCTRL3); in kvm_save_guest_pmu() 82 struct loongarch_csrs *csr = vcpu->arch.csr; in kvm_restore_guest_pmu() local [all …]
|
D | timer.c | 46 kvm_write_sw_gcsr(vcpu->arch.csr, LOONGARCH_CSR_TVAL, 0); in kvm_init_timer() 57 struct loongarch_csrs *csr = vcpu->arch.csr; in kvm_restore_timer() local 60 * Set guest stable timer cfg csr in kvm_restore_timer() 61 * Disable timer before restore estat CSR register, avoid to in kvm_restore_timer() 64 cfg = kvm_read_sw_gcsr(csr, LOONGARCH_CSR_TCFG); in kvm_restore_timer() 67 kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_ESTAT); in kvm_restore_timer() 68 kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_TCFG); in kvm_restore_timer() 71 kvm_restore_hw_gcsr(csr, LOONGARCH_CSR_TVAL); in kvm_restore_timer() 83 * If oneshot timer is fired, CSR TVAL will be -1, there are two in kvm_restore_timer() 90 ticks = kvm_read_sw_gcsr(csr, LOONGARCH_CSR_TVAL); in kvm_restore_timer() [all …]
|
/linux-6.12.1/arch/sparc/kernel/ |
D | ebus.c | 74 u32 csr = 0; in ebus_dma_irq() local 77 csr = readl(p->regs + EBDMA_CSR); in ebus_dma_irq() 78 writel(csr, p->regs + EBDMA_CSR); in ebus_dma_irq() 81 if (csr & EBDMA_CSR_ERR_PEND) { in ebus_dma_irq() 85 } else if (csr & EBDMA_CSR_INT_PEND) { in ebus_dma_irq() 87 (csr & EBDMA_CSR_TC) ? in ebus_dma_irq() 99 u32 csr; in ebus_dma_register() local 113 csr = EBDMA_CSR_BURST_SZ_16 | EBDMA_CSR_EN_CNT; in ebus_dma_register() 116 csr |= EBDMA_CSR_TCI_DIS; in ebus_dma_register() 118 writel(csr, p->regs + EBDMA_CSR); in ebus_dma_register() [all …]
|
/linux-6.12.1/arch/alpha/kernel/ |
D | core_tsunami.c | 181 volatile unsigned long *csr; in tsunami_pci_tbi() local 186 csr = &pchip->tlbia.csr; in tsunami_pci_tbi() 188 csr = &pchip->tlbiv.csr; in tsunami_pci_tbi() 194 *csr = value; in tsunami_pci_tbi() 196 *csr; in tsunami_pci_tbi() 227 TSUNAMI_cchip->misc.csr |= (1L << 28); /* clear NXM... */ in tsunami_probe_write() 231 if (TSUNAMI_cchip->misc.csr & (1L << 28)) { in tsunami_probe_write() 232 int source = (TSUNAMI_cchip->misc.csr >> 29) & 7; in tsunami_probe_write() 233 TSUNAMI_cchip->misc.csr |= (1L << 28); /* ...and unlock NXS. */ in tsunami_probe_write() 251 if (tsunami_probe_read(&pchip->pctl.csr) == 0) in tsunami_init_one_pchip() [all …]
|
D | core_wildfire.c | 121 pci->pci_window[0].wbase.csr = hose->sg_isa->dma_base | 3; in wildfire_init_hose() 122 pci->pci_window[0].wmask.csr = (hose->sg_isa->size - 1) & 0xfff00000; in wildfire_init_hose() 123 pci->pci_window[0].tbase.csr = virt_to_phys(hose->sg_isa->ptes); in wildfire_init_hose() 125 pci->pci_window[1].wbase.csr = 0x40000000 | 1; in wildfire_init_hose() 126 pci->pci_window[1].wmask.csr = (0x40000000 -1) & 0xfff00000; in wildfire_init_hose() 127 pci->pci_window[1].tbase.csr = 0; in wildfire_init_hose() 129 pci->pci_window[2].wbase.csr = 0x80000000 | 1; in wildfire_init_hose() 130 pci->pci_window[2].wmask.csr = (0x40000000 -1) & 0xfff00000; in wildfire_init_hose() 131 pci->pci_window[2].tbase.csr = 0x40000000; in wildfire_init_hose() 133 pci->pci_window[3].wbase.csr = hose->sg_pci->dma_base | 3; in wildfire_init_hose() [all …]
|
D | core_titan.c | 207 volatile unsigned long *csr; in titan_pci_tbi() local 220 csr = &port->port_specific.g.gtlbia.csr; in titan_pci_tbi() 222 csr = &port->port_specific.g.gtlbiv.csr; in titan_pci_tbi() 229 *csr = value; in titan_pci_tbi() 231 *csr; in titan_pci_tbi() 240 pctl.pctl_q_whole = port->pctl.csr; in titan_query_agp() 293 saved_config[index].wsba[0] = port->wsba[0].csr; in titan_init_one_pachip_port() 294 saved_config[index].wsm[0] = port->wsm[0].csr; in titan_init_one_pachip_port() 295 saved_config[index].tba[0] = port->tba[0].csr; in titan_init_one_pachip_port() 297 saved_config[index].wsba[1] = port->wsba[1].csr; in titan_init_one_pachip_port() [all …]
|
D | sys_marvel.c | 96 ctl = &io7->csrs->PO7_LSI_CTL[irq & 0xff].csr; /* assume LSI */ in io7_get_irq_ctl() 98 ctl = &io7->csrs->PO7_MSI_CTL[((irq - 0x80) >> 5) & 0x0f].csr; in io7_get_irq_ctl() 174 volatile unsigned long *csr, in io7_redirect_irq() argument 179 val = *csr; in io7_redirect_irq() 183 *csr = val; in io7_redirect_irq() 185 *csr; in io7_redirect_irq() 196 val = io7->csrs->PO7_LSI_CTL[which].csr; in io7_redirect_one_lsi() 200 io7->csrs->PO7_LSI_CTL[which].csr = val; in io7_redirect_one_lsi() 202 io7->csrs->PO7_LSI_CTL[which].csr; in io7_redirect_one_lsi() 213 val = io7->csrs->PO7_MSI_CTL[which].csr; in io7_redirect_one_msi() [all …]
|
D | err_ev7.c | 194 "DM CSR PH", "DM CSR PH", "DM CSR PH", 195 "DM CSR PH", "reserved", 198 "DM CSR PH", "DM CSR PH", "DM CSR PH", 199 "DM CSR PH", "reserved", 202 "DM CSR PH", "DM CSR PH", "DM CSR PH", 203 "DM CSR PH", "reserved", 206 "DM CSR PH", "DM CSR PH", "DM CSR PH", 207 "DM CSR PH", "reserved",
|
/linux-6.12.1/drivers/crypto/starfive/ |
D | jh7110-rsa.c | 82 rctx->csr.pka.v = 0; in starfive_rsa_montgomery_form() 84 writel(rctx->csr.pka.v, cryp->base + STARFIVE_PKA_CACR_OFFSET); in starfive_rsa_montgomery_form() 90 rctx->csr.pka.v = 0; in starfive_rsa_montgomery_form() 91 rctx->csr.pka.cln_done = 1; in starfive_rsa_montgomery_form() 92 rctx->csr.pka.opsize = opsize; in starfive_rsa_montgomery_form() 93 rctx->csr.pka.exposize = opsize; in starfive_rsa_montgomery_form() 94 rctx->csr.pka.cmd = CRYPTO_CMD_PRE; in starfive_rsa_montgomery_form() 95 rctx->csr.pka.start = 1; in starfive_rsa_montgomery_form() 96 rctx->csr.pka.not_r2 = 1; in starfive_rsa_montgomery_form() 97 rctx->csr.pka.ie = 1; in starfive_rsa_montgomery_form() [all …]
|
/linux-6.12.1/drivers/usb/musb/ |
D | musb_gadget.c | 229 u16 fifo_count = 0, csr; in txstate() local 248 csr = musb_readw(epio, MUSB_TXCSR); in txstate() 254 if (csr & MUSB_TXCSR_TXPKTRDY) { in txstate() 256 musb_ep->end_point.name, csr); in txstate() 260 if (csr & MUSB_TXCSR_P_SENDSTALL) { in txstate() 262 musb_ep->end_point.name, csr); in txstate() 268 csr); in txstate() 275 /* setup DMA, then program endpoint CSR */ in txstate() 301 csr &= ~(MUSB_TXCSR_AUTOSET in txstate() 303 musb_writew(epio, MUSB_TXCSR, csr in txstate() [all …]
|
D | musb_gadget_ep0.c | 243 u16 csr; in service_zero_data_request() local 266 csr = musb_readw(regs, MUSB_TXCSR); in service_zero_data_request() 267 csr |= MUSB_TXCSR_CLRDATATOG | in service_zero_data_request() 269 csr &= ~(MUSB_TXCSR_P_SENDSTALL | in service_zero_data_request() 272 musb_writew(regs, MUSB_TXCSR, csr); in service_zero_data_request() 274 csr = musb_readw(regs, MUSB_RXCSR); in service_zero_data_request() 275 csr |= MUSB_RXCSR_CLRDATATOG | in service_zero_data_request() 277 csr &= ~(MUSB_RXCSR_P_SENDSTALL | in service_zero_data_request() 279 musb_writew(regs, MUSB_RXCSR, csr); in service_zero_data_request() 403 u16 csr; in service_zero_data_request() local [all …]
|
D | musb_cppi41.c | 56 u16 csr; in save_rx_toggle() local 64 csr = musb_readw(cppi41_channel->hw_ep->regs, MUSB_RXCSR); in save_rx_toggle() 65 toggle = csr & MUSB_RXCSR_H_DATATOGGLE ? 1 : 0; in save_rx_toggle() 74 u16 csr; in update_rx_toggle() local 83 csr = musb_readw(hw_ep->regs, MUSB_RXCSR); in update_rx_toggle() 84 toggle = csr & MUSB_RXCSR_H_DATATOGGLE ? 1 : 0; in update_rx_toggle() 92 csr |= MUSB_RXCSR_H_DATATOGGLE | MUSB_RXCSR_H_WR_DATATOGGLE; in update_rx_toggle() 93 musb_writew(cppi41_channel->hw_ep->regs, MUSB_RXCSR, csr); in update_rx_toggle() 105 u16 csr; in musb_is_tx_fifo_empty() local 108 csr = musb_readw(epio, MUSB_TXCSR); in musb_is_tx_fifo_empty() [all …]
|
D | musb_host.c | 90 u16 csr; in musb_h_tx_flush_fifo() local 93 csr = musb_readw(epio, MUSB_TXCSR); in musb_h_tx_flush_fifo() 94 while (csr & MUSB_TXCSR_FIFONOTEMPTY) { in musb_h_tx_flush_fifo() 95 csr |= MUSB_TXCSR_FLUSHFIFO | MUSB_TXCSR_TXPKTRDY; in musb_h_tx_flush_fifo() 96 musb_writew(epio, MUSB_TXCSR, csr); in musb_h_tx_flush_fifo() 97 csr = musb_readw(epio, MUSB_TXCSR); in musb_h_tx_flush_fifo() 114 "Could not flush host TX%d fifo: csr: %04x\n", in musb_h_tx_flush_fifo() 115 ep->epnum, csr)) in musb_h_tx_flush_fifo() 124 u16 csr; in musb_h_ep0_flush_fifo() local 129 csr = musb_readw(epio, MUSB_TXCSR); in musb_h_ep0_flush_fifo() [all …]
|
/linux-6.12.1/drivers/scsi/ |
D | sun3_scsi.c | 73 unsigned short csr; /* control/status reg */ member 117 /* bits in csr reg */ 190 // safe bits for the CSR 196 unsigned short csr = dregs->csr; in scsi_sun3_intr() local 200 dregs->csr &= ~CSR_DMA_ENABLE; in scsi_sun3_intr() 203 if(csr & ~CSR_GOOD) { in scsi_sun3_intr() 204 if (csr & CSR_DMA_BUSERR) in scsi_sun3_intr() 206 if (csr & CSR_DMA_CONFLICT) in scsi_sun3_intr() 211 if(csr & (CSR_SDB_INT | CSR_DMA_INT)) { in scsi_sun3_intr() 242 dregs->csr &= ~CSR_FIFO; in sun3scsi_dma_setup() [all …]
|
/linux-6.12.1/arch/loongarch/include/asm/ |
D | kvm_csr.h | 14 #define gcsr_read(csr) \ argument 20 : [reg] "i" (csr) \ 25 #define gcsr_write(v, csr) \ argument 31 : [reg] "i" (csr) \ 36 #define gcsr_xchg(v, m, csr) \ argument 42 : [mask] "r" (m), [reg] "i" (csr) \ 182 #define kvm_save_hw_gcsr(csr, gid) (csr->csrs[gid] = gcsr_read(gid)) argument 183 #define kvm_restore_hw_gcsr(csr, gid) (gcsr_write(csr->csrs[gid], gid)) argument 185 #define kvm_read_clear_hw_gcsr(csr, gid) (csr->csrs[gid] = gcsr_write(0, gid)) argument 189 static __always_inline unsigned long kvm_read_sw_gcsr(struct loongarch_csrs *csr, int gid) in kvm_read_sw_gcsr() argument [all …]
|
/linux-6.12.1/drivers/usb/gadget/udc/ |
D | at91_udc.c | 111 u32 csr; in proc_ep_show() local 118 csr = __raw_readl(ep->creg); in proc_ep_show() 131 seq_printf(s, "csr %08x rxbytes=%d %s %s %s" EIGHTBITS "\n", in proc_ep_show() 132 csr, in proc_ep_show() 133 (csr & 0x07ff0000) >> 16, in proc_ep_show() 134 (csr & (1 << 15)) ? "enabled" : "disabled", in proc_ep_show() 135 (csr & (1 << 11)) ? "DATA1" : "DATA0", in proc_ep_show() 136 types[(csr & 0x700) >> 8], in proc_ep_show() 139 (!(csr & 0x700)) in proc_ep_show() 140 ? ((csr & (1 << 7)) ? " IN" : " OUT") in proc_ep_show() [all …]
|
/linux-6.12.1/arch/riscv/kvm/ |
D | vcpu.c | 20 #include <asm/csr.h> 51 struct kvm_vcpu_csr *csr = &vcpu->arch.guest_csr; in kvm_riscv_reset_vcpu() local 69 memcpy(csr, reset_csr, sizeof(*csr)); in kvm_riscv_reset_vcpu() 342 struct kvm_vcpu_csr *csr = &vcpu->arch.guest_csr; in kvm_riscv_vcpu_flush_interrupts() local 349 csr->hvip &= ~mask; in kvm_riscv_vcpu_flush_interrupts() 350 csr->hvip |= val; in kvm_riscv_vcpu_flush_interrupts() 361 struct kvm_vcpu_csr *csr = &vcpu->arch.guest_csr; in kvm_riscv_vcpu_sync_interrupts() local 364 csr->vsie = csr_read(CSR_VSIE); in kvm_riscv_vcpu_sync_interrupts() 368 if ((csr->hvip ^ hvip) & (1UL << IRQ_VS_SOFT)) { in kvm_riscv_vcpu_sync_interrupts() 381 if ((csr->hvip ^ hvip) & (1UL << IRQ_PMU_OVF)) { in kvm_riscv_vcpu_sync_interrupts() [all …]
|
/linux-6.12.1/arch/sh/kernel/cpu/ |
D | adc.c | 16 unsigned char csr; in adc_single() local 22 csr = __raw_readb(ADCSR); in adc_single() 23 csr = channel | ADCSR_ADST | ADCSR_CKS; in adc_single() 24 __raw_writeb(csr, ADCSR); in adc_single() 27 csr = __raw_readb(ADCSR); in adc_single() 28 } while ((csr & ADCSR_ADF) == 0); in adc_single() 30 csr &= ~(ADCSR_ADF | ADCSR_ADST); in adc_single() 31 __raw_writeb(csr, ADCSR); in adc_single()
|
/linux-6.12.1/arch/alpha/include/asm/ |
D | core_t2.h | 206 unsigned long elcmc_bcc; /* CSR 0 */ 207 unsigned long elcmc_bcce; /* CSR 1 */ 208 unsigned long elcmc_bccea; /* CSR 2 */ 209 unsigned long elcmc_bcue; /* CSR 3 */ 210 unsigned long elcmc_bcuea; /* CSR 4 */ 211 unsigned long elcmc_dter; /* CSR 5 */ 212 unsigned long elcmc_cbctl; /* CSR 6 */ 213 unsigned long elcmc_cbe; /* CSR 7 */ 214 unsigned long elcmc_cbeal; /* CSR 8 */ 215 unsigned long elcmc_cbeah; /* CSR 9 */ [all …]
|
/linux-6.12.1/drivers/watchdog/ |
D | shwdt.c | 85 u8 csr; in sh_wdt_start() local 95 csr = sh_wdt_read_csr(); in sh_wdt_start() 96 csr |= WTCSR_WT | clock_division_ratio; in sh_wdt_start() 97 sh_wdt_write_csr(csr); in sh_wdt_start() 109 csr = sh_wdt_read_csr(); in sh_wdt_start() 110 csr |= WTCSR_TME; in sh_wdt_start() 111 csr &= ~WTCSR_RSTS; in sh_wdt_start() 112 sh_wdt_write_csr(csr); in sh_wdt_start() 115 csr = sh_wdt_read_rstcsr(); in sh_wdt_start() 116 csr &= ~RSTCSR_RSTS; in sh_wdt_start() [all …]
|
/linux-6.12.1/Documentation/devicetree/bindings/clock/ |
D | xgene.txt | 36 - reg : shall be a list of address and length pairs describing the CSR 40 may include "csr-reg" and/or "div-reg". If this property 42 only "csr-reg". 49 - csr-offset : Offset to the CSR reset register from the reset address base. 51 - csr-mask : CSR reset mask bit. Default is 0xF. 54 - enable-mask : CSR enable mask bit. Default is 0xF. 55 - divider-offset : Offset to the divider CSR register from the divider base. 96 reg-name = "csr-reg"; 120 reg-names = "csr-reg", "div-reg"; 121 csr-offset = <0x0>; [all …]
|
/linux-6.12.1/include/linux/ |
D | rio_regs.h | 103 #define RIO_MBOX_CSR 0x40 /* [II, <= 1.2] Mailbox CSR */ 129 #define RIO_WRITE_PORT_CSR 0x44 /* [I, <= 1.2] Write Port CSR */ 130 #define RIO_DOORBELL_CSR 0x44 /* [II, <= 1.2] Doorbell CSR */ 146 #define RIO_PELL_CTRL_CSR 0x4c /* [I] PE Logical Layer Control CSR */ 156 #define RIO_DID_CSR 0x60 /* [III] Base Device ID CSR */ 160 #define RIO_HOST_DID_LOCK_CSR 0x68 /* [III] Host Base Device ID Lock CSR */ 161 #define RIO_COMPONENT_TAG_CSR 0x6c /* [III] Component Tag CSR */ 301 #define RIO_EM_LTL_ERR_DETECT 0x008 /* Logical/Transport Layer Error Detect CSR */ 302 #define RIO_EM_LTL_ERR_EN 0x00c /* Logical/Transport Layer Error Enable CSR */ 307 #define RIO_EM_LTL_HIADDR_CAP 0x010 /* Logical/Transport Layer High Address Capture CSR */ [all …]
|
/linux-6.12.1/drivers/net/pcs/ |
D | pcs-xpcs-plat.c | 41 static u16 xpcs_mmio_addr_page(ptrdiff_t csr) in xpcs_mmio_addr_page() argument 43 return FIELD_GET(0x1fff00, csr); in xpcs_mmio_addr_page() 46 static ptrdiff_t xpcs_mmio_addr_offset(ptrdiff_t csr) in xpcs_mmio_addr_offset() argument 48 return FIELD_GET(0xff, csr); in xpcs_mmio_addr_offset() 54 ptrdiff_t csr, ofs; in xpcs_mmio_read_reg_indirect() local 58 csr = xpcs_mmio_addr_format(dev, reg); in xpcs_mmio_read_reg_indirect() 59 page = xpcs_mmio_addr_page(csr); in xpcs_mmio_read_reg_indirect() 60 ofs = xpcs_mmio_addr_offset(csr); in xpcs_mmio_read_reg_indirect() 85 ptrdiff_t csr, ofs; in xpcs_mmio_write_reg_indirect() local 89 csr = xpcs_mmio_addr_format(dev, reg); in xpcs_mmio_write_reg_indirect() [all …]
|