/linux-6.12.1/drivers/macintosh/ |
D | rack-meter.c | 51 struct rackmeter *rm; member 95 static void rackmeter_setup_i2s(struct rackmeter *rm) in rackmeter_setup_i2s() argument 97 struct macio_chip *macio = rm->mdev->bus->chip; in rackmeter_setup_i2s() 106 pmac_call_feature(PMAC_FTR_SOUND_CHIP_ENABLE, rm->i2s, 0, 1); in rackmeter_setup_i2s() 121 out_le32(rm->i2s_regs + 0x10, 0x01fa0000); in rackmeter_setup_i2s() 122 (void)in_le32(rm->i2s_regs + 0x10); in rackmeter_setup_i2s() 132 static void rackmeter_set_default_pattern(struct rackmeter *rm) in rackmeter_set_default_pattern() argument 138 rm->ubuf[i] = (i & 1) * 255; in rackmeter_set_default_pattern() 140 rm->ubuf[i] = ((~i) & 1) * 255; in rackmeter_set_default_pattern() 144 static void rackmeter_do_pause(struct rackmeter *rm, int pause) in rackmeter_do_pause() argument [all …]
|
/linux-6.12.1/arch/arm/net/ |
D | bpf_jit_32.h | 165 #define _AL3_R(op, rd, rn, rm) ((op ## _R) | (rd) << 12 | (rn) << 16 | (rm)) argument 171 #define ARM_ADD_R(rd, rn, rm) _AL3_R(ARM_INST_ADD, rd, rn, rm) argument 172 #define ARM_ADDS_R(rd, rn, rm) _AL3_R(ARM_INST_ADDS, rd, rn, rm) argument 175 #define ARM_ADC_R(rd, rn, rm) _AL3_R(ARM_INST_ADC, rd, rn, rm) argument 178 #define ARM_AND_R(rd, rn, rm) _AL3_R(ARM_INST_AND, rd, rn, rm) argument 179 #define ARM_ANDS_R(rd, rn, rm) _AL3_R(ARM_INST_ANDS, rd, rn, rm) argument 182 #define ARM_BIC_R(rd, rn, rm) _AL3_R(ARM_INST_BIC, rd, rn, rm) argument 186 #define ARM_BX(rm) (ARM_INST_BX | (rm)) argument 187 #define ARM_BLX_R(rm) (ARM_INST_BLX_R | (rm)) argument 189 #define ARM_CMP_R(rn, rm) _AL3_R(ARM_INST_CMP, 0, rn, rm) argument [all …]
|
/linux-6.12.1/net/rds/ |
D | message.c | 51 void rds_message_addref(struct rds_message *rm) in rds_message_addref() argument 53 rdsdebug("addref rm %p ref %d\n", rm, refcount_read(&rm->m_refcount)); in rds_message_addref() 54 refcount_inc(&rm->m_refcount); in rds_message_addref() 130 static void rds_message_purge(struct rds_message *rm) in rds_message_purge() argument 135 if (unlikely(test_bit(RDS_MSG_PAGEVEC, &rm->m_flags))) in rds_message_purge() 138 spin_lock_irqsave(&rm->m_rs_lock, flags); in rds_message_purge() 139 if (rm->m_rs) { in rds_message_purge() 140 struct rds_sock *rs = rm->m_rs; in rds_message_purge() 142 if (rm->data.op_mmp_znotifier) { in rds_message_purge() 144 rds_rm_zerocopy_callback(rs, rm->data.op_mmp_znotifier); in rds_message_purge() [all …]
|
D | send.c | 67 struct rds_message *rm, *tmp; in rds_send_path_reset() local 71 rm = cp->cp_xmit_rm; in rds_send_path_reset() 77 rds_message_unmapped(rm); in rds_send_path_reset() 78 rds_message_put(rm); in rds_send_path_reset() 95 list_for_each_entry_safe(rm, tmp, &cp->cp_retrans, m_conn_item) { in rds_send_path_reset() 96 set_bit(RDS_MSG_ACK_REQUIRED, &rm->m_flags); in rds_send_path_reset() 97 set_bit(RDS_MSG_RETRANSMITTED, &rm->m_flags); in rds_send_path_reset() 139 struct rds_message *rm; in rds_send_xmit() local 201 rm = cp->cp_xmit_rm; in rds_send_xmit() 203 if (!rm) { in rds_send_xmit() [all …]
|
D | tcp_send.c | 70 int rds_tcp_xmit(struct rds_connection *conn, struct rds_message *rm, in rds_tcp_xmit() argument 73 struct rds_conn_path *cp = rm->m_inc.i_conn_path; in rds_tcp_xmit() 86 rm->m_ack_seq = tc->t_last_sent_nxt + in rds_tcp_xmit() 88 be32_to_cpu(rm->m_inc.i_hdr.h_len) - 1; in rds_tcp_xmit() 90 set_bit(RDS_MSG_HAS_ACK_SEQ, &rm->m_flags); in rds_tcp_xmit() 91 tc->t_last_expected_una = rm->m_ack_seq + 1; in rds_tcp_xmit() 93 if (test_bit(RDS_MSG_RETRANSMITTED, &rm->m_flags)) in rds_tcp_xmit() 94 rm->m_inc.i_hdr.h_flags |= RDS_FLAG_RETRANSMITTED; in rds_tcp_xmit() 97 rm, rds_tcp_write_seq(tc), in rds_tcp_xmit() 98 (unsigned long long)rm->m_ack_seq); in rds_tcp_xmit() [all …]
|
D | rdma.c | 611 int rds_cmsg_rdma_args(struct rds_sock *rs, struct rds_message *rm, in rds_cmsg_rdma_args() argument 616 struct rm_rdma_op *op = &rm->rdma; in rds_cmsg_rdma_args() 626 || rm->rdma.op_active) in rds_cmsg_rdma_args() 672 op->op_sg = rds_message_alloc_sgs(rm, nr_pages); in rds_cmsg_rdma_args() 811 int rds_cmsg_rdma_dest(struct rds_sock *rs, struct rds_message *rm, in rds_cmsg_rdma_dest() argument 820 rm->m_rdma_cookie != 0) in rds_cmsg_rdma_dest() 823 memcpy(&rm->m_rdma_cookie, CMSG_DATA(cmsg), sizeof(rm->m_rdma_cookie)); in rds_cmsg_rdma_dest() 830 r_key = rds_rdma_cookie_key(rm->m_rdma_cookie); in rds_cmsg_rdma_dest() 843 rm->rdma.op_rdma_mr = mr; in rds_cmsg_rdma_dest() 854 int rds_cmsg_rdma_map(struct rds_sock *rs, struct rds_message *rm, in rds_cmsg_rdma_map() argument [all …]
|
D | ib_send.c | 48 static void rds_ib_send_complete(struct rds_message *rm, in rds_ib_send_complete() argument 50 void (*complete)(struct rds_message *rm, int status)) in rds_ib_send_complete() argument 70 complete(rm, notify_status); in rds_ib_send_complete() 154 struct rds_message *rm = NULL; in rds_ib_send_unmap_op() local 160 rm = container_of(send->s_op, struct rds_message, data); in rds_ib_send_unmap_op() 167 rm = container_of(send->s_op, struct rds_message, rdma); in rds_ib_send_unmap_op() 174 rm = container_of(send->s_op, struct rds_message, atomic); in rds_ib_send_unmap_op() 187 return rm; in rds_ib_send_unmap_op() 245 struct rds_message *rm = NULL; in rds_ib_send_cqe_handler() local 276 rm = rds_ib_send_unmap_op(ic, send, wc->status); in rds_ib_send_cqe_handler() [all …]
|
/linux-6.12.1/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
D | r535.c | 36 &vmm->rm.client, &vmm->rm.device); in r535_mmu_promote_vmm() 40 args = nvkm_gsp_rm_alloc_get(&vmm->rm.device.object, 0x90f10000, FERMI_VASPACE_A, in r535_mmu_promote_vmm() 41 sizeof(*args), &vmm->rm.object); in r535_mmu_promote_vmm() 47 ret = nvkm_gsp_rm_alloc_wr(&vmm->rm.object, args); in r535_mmu_promote_vmm() 56 &vmm->rm.rsvd); in r535_mmu_promote_vmm() 61 ctrl = nvkm_gsp_rm_ctrl_get(&vmm->rm.object, in r535_mmu_promote_vmm() 68 ctrl->virtAddrLo = vmm->rm.rsvd->addr; in r535_mmu_promote_vmm() 69 ctrl->virtAddrHi = vmm->rm.rsvd->addr + vmm->rm.rsvd->size - 1; in r535_mmu_promote_vmm() 86 ret = nvkm_gsp_rm_ctrl_wr(&vmm->rm.object, ctrl); in r535_mmu_promote_vmm() 103 struct nvkm_mmu_func *rm; in r535_mmu_new() local [all …]
|
/linux-6.12.1/drivers/gpu/drm/msm/disp/dpu1/ |
D | dpu_rm.c | 38 struct dpu_rm *rm, in dpu_rm_init() argument 45 if (!rm || !cat || !mmio) { in dpu_rm_init() 51 memset(rm, 0, sizeof(*rm)); in dpu_rm_init() 64 rm->mixer_blks[lm->id - LM_0] = &hw->base; in dpu_rm_init() 78 rm->merge_3d_blks[merge_3d->id - MERGE_3D_0] = &hw->base; in dpu_rm_init() 93 hw->merge_3d = to_dpu_hw_merge_3d(rm->merge_3d_blks[pp->merge_3d - MERGE_3D_0]); in dpu_rm_init() 94 rm->pingpong_blks[pp->id - PINGPONG_0] = &hw->base; in dpu_rm_init() 107 rm->hw_intf[intf->id - INTF_0] = hw; in dpu_rm_init() 120 rm->hw_wb[wb->id - WB_0] = hw; in dpu_rm_init() 133 rm->ctl_blks[ctl->id - CTL_0] = &hw->base; in dpu_rm_init() [all …]
|
D | dpu_rm.h | 51 struct dpu_rm *rm, 69 int dpu_rm_reserve(struct dpu_rm *rm, 88 int dpu_rm_get_assigned_resources(struct dpu_rm *rm, 105 static inline struct dpu_hw_intf *dpu_rm_get_intf(struct dpu_rm *rm, enum dpu_intf intf_idx) in dpu_rm_get_intf() argument 107 return rm->hw_intf[intf_idx - INTF_0]; in dpu_rm_get_intf() 115 static inline struct dpu_hw_wb *dpu_rm_get_wb(struct dpu_rm *rm, enum dpu_wb wb_idx) in dpu_rm_get_wb() argument 117 return rm->hw_wb[wb_idx - WB_0]; in dpu_rm_get_wb() 125 static inline struct dpu_hw_sspp *dpu_rm_get_sspp(struct dpu_rm *rm, enum dpu_sspp sspp_idx) in dpu_rm_get_sspp() argument 127 return rm->hw_sspp[sspp_idx - SSPP_NONE]; in dpu_rm_get_sspp()
|
/linux-6.12.1/drivers/gpu/drm/nouveau/nvkm/engine/nvjpg/ |
D | r535.c | 33 struct nvkm_gsp_object rm; member 41 nvkm_gsp_rm_free(&obj->rm); in r535_nvjpg_obj_dtor() 64 args = nvkm_gsp_rm_alloc_get(&chan->rm.object, oclass->handle, oclass->base.oclass, in r535_nvjpg_obj_ctor() 65 sizeof(*args), &obj->rm); in r535_nvjpg_obj_ctor() 72 return nvkm_gsp_rm_alloc_wr(&obj->rm, args); in r535_nvjpg_obj_ctor() 86 struct nvkm_engine_func *rm; in r535_nvjpg_new() local 91 if (!(rm = kzalloc(sizeof(*rm) + (nclass + 1) * sizeof(rm->sclass[0]), GFP_KERNEL))) in r535_nvjpg_new() 94 rm->dtor = r535_nvjpg_dtor; in r535_nvjpg_new() 96 rm->sclass[i].minver = hw->sclass[i].minver; in r535_nvjpg_new() 97 rm->sclass[i].maxver = hw->sclass[i].maxver; in r535_nvjpg_new() [all …]
|
/linux-6.12.1/drivers/gpu/drm/nouveau/nvkm/engine/ce/ |
D | r535.c | 34 struct nvkm_gsp_object rm; member 42 nvkm_gsp_rm_free(&obj->rm); in r535_ce_obj_dtor() 65 args = nvkm_gsp_rm_alloc_get(&chan->rm.object, oclass->handle, oclass->base.oclass, in r535_ce_obj_ctor() 66 sizeof(*args), &obj->rm); in r535_ce_obj_ctor() 73 return nvkm_gsp_rm_alloc_wr(&obj->rm, args); in r535_ce_obj_ctor() 87 struct nvkm_engine_func *rm; in r535_ce_new() local 92 if (!(rm = kzalloc(sizeof(*rm) + (nclass + 1) * sizeof(rm->sclass[0]), GFP_KERNEL))) in r535_ce_new() 95 rm->dtor = r535_ce_dtor; in r535_ce_new() 97 rm->sclass[i].minver = hw->sclass[i].minver; in r535_ce_new() 98 rm->sclass[i].maxver = hw->sclass[i].maxver; in r535_ce_new() [all …]
|
/linux-6.12.1/drivers/gpu/drm/nouveau/nvkm/engine/ofa/ |
D | r535.c | 34 struct nvkm_gsp_object rm; member 42 nvkm_gsp_rm_free(&obj->rm); in r535_ofa_obj_dtor() 65 args = nvkm_gsp_rm_alloc_get(&chan->rm.object, oclass->handle, oclass->base.oclass, in r535_ofa_obj_ctor() 66 sizeof(*args), &obj->rm); in r535_ofa_obj_ctor() 72 return nvkm_gsp_rm_alloc_wr(&obj->rm, args); in r535_ofa_obj_ctor() 86 struct nvkm_engine_func *rm; in r535_ofa_new() local 91 if (!(rm = kzalloc(sizeof(*rm) + (nclass + 1) * sizeof(rm->sclass[0]), GFP_KERNEL))) in r535_ofa_new() 94 rm->dtor = r535_ofa_dtor; in r535_ofa_new() 96 rm->sclass[i].minver = hw->sclass[i].minver; in r535_ofa_new() 97 rm->sclass[i].maxver = hw->sclass[i].maxver; in r535_ofa_new() [all …]
|
/linux-6.12.1/drivers/gpu/drm/nouveau/nvkm/engine/disp/ |
D | r535.c | 77 nvkm_gsp_rm_free(&chan->rm.object); in r535_chan_fini() 131 args = nvkm_gsp_rm_alloc_get(&chan->disp->rm.object, in r535_curs_init() 133 chan->object.oclass, sizeof(*args), &chan->rm.object); in r535_curs_init() 139 return nvkm_gsp_rm_alloc_wr(&chan->rm.object, args); in r535_curs_init() 161 (chan->disp->rm.client.object.handle & 0x3fff)); in r535_dmac_bind() 184 args = nvkm_gsp_rm_alloc_get(&chan->disp->rm.object, in r535_dmac_init() 186 chan->object.oclass, sizeof(*args), &chan->rm.object); in r535_dmac_init() 193 return nvkm_gsp_rm_alloc_wr(&chan->rm.object, args); in r535_dmac_init() 268 ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom, in r535_sor_bl_set() 277 return nvkm_gsp_rm_ctrl_wr(&disp->rm.objcom, ctrl); in r535_sor_bl_set() [all …]
|
/linux-6.12.1/tools/perf/arch/x86/util/ |
D | perf_regs.c | 207 regmatch_t rm[6]; in arch_sdt_arg_parse_op() local 226 regexec(&sdt_op_regex, old_op, 6, rm, 0) || in arch_sdt_arg_parse_op() 227 rm[4].rm_eo - rm[4].rm_so > SDT_REG_NAME_SIZE) { in arch_sdt_arg_parse_op() 244 if (rm[3].rm_so != rm[3].rm_eo) { in arch_sdt_arg_parse_op() 245 if (rm[1].rm_so != rm[1].rm_eo) in arch_sdt_arg_parse_op() 246 prefix[0] = *(old_op + rm[1].rm_so); in arch_sdt_arg_parse_op() 247 else if (rm[2].rm_so != rm[2].rm_eo) in arch_sdt_arg_parse_op() 254 sdt_rename_register(old_op + rm[4].rm_so, rm[4].rm_eo - rm[4].rm_so, in arch_sdt_arg_parse_op() 259 (rm[2].rm_eo - rm[2].rm_so) + in arch_sdt_arg_parse_op() 260 (rm[3].rm_eo - rm[3].rm_so) + in arch_sdt_arg_parse_op() [all …]
|
/linux-6.12.1/drivers/gpu/drm/nouveau/nvkm/engine/nvdec/ |
D | r535.c | 33 struct nvkm_gsp_object rm; member 41 nvkm_gsp_rm_free(&obj->rm); in r535_nvdec_obj_dtor() 64 args = nvkm_gsp_rm_alloc_get(&chan->rm.object, oclass->handle, oclass->base.oclass, in r535_nvdec_obj_ctor() 65 sizeof(*args), &obj->rm); in r535_nvdec_obj_ctor() 72 return nvkm_gsp_rm_alloc_wr(&obj->rm, args); in r535_nvdec_obj_ctor() 88 struct nvkm_engine_func *rm; in r535_nvdec_new() local 93 if (!(rm = kzalloc(sizeof(*rm) + (nclass + 1) * sizeof(rm->sclass[0]), GFP_KERNEL))) in r535_nvdec_new() 96 rm->dtor = r535_nvdec_dtor; in r535_nvdec_new() 98 rm->sclass[i].minver = hw->sclass[i].minver; in r535_nvdec_new() 99 rm->sclass[i].maxver = hw->sclass[i].maxver; in r535_nvdec_new() [all …]
|
/linux-6.12.1/drivers/gpu/drm/nouveau/nvkm/engine/nvenc/ |
D | r535.c | 33 struct nvkm_gsp_object rm; member 41 nvkm_gsp_rm_free(&obj->rm); in r535_nvenc_obj_dtor() 64 args = nvkm_gsp_rm_alloc_get(&chan->rm.object, oclass->handle, oclass->base.oclass, in r535_nvenc_obj_ctor() 65 sizeof(*args), &obj->rm); in r535_nvenc_obj_ctor() 72 return nvkm_gsp_rm_alloc_wr(&obj->rm, args); in r535_nvenc_obj_ctor() 88 struct nvkm_engine_func *rm; in r535_nvenc_new() local 93 if (!(rm = kzalloc(sizeof(*rm) + (nclass + 1) * sizeof(rm->sclass[0]), GFP_KERNEL))) in r535_nvenc_new() 96 rm->dtor = r535_nvenc_dtor; in r535_nvenc_new() 98 rm->sclass[i].minver = hw->sclass[i].minver; in r535_nvenc_new() 99 rm->sclass[i].maxver = hw->sclass[i].maxver; in r535_nvenc_new() [all …]
|
/linux-6.12.1/drivers/gpu/drm/nouveau/nvkm/subdev/bar/ |
D | r535.c | 81 vmm->rm.bar2_pdb = gsp->bar.rm_bar2_pdb; in r535_bar_bar2_init() 151 struct nvkm_bar_func *rm; in r535_bar_new_() local 155 if (!(rm = kzalloc(sizeof(*rm), GFP_KERNEL))) in r535_bar_new_() 158 rm->dtor = r535_bar_dtor; in r535_bar_new_() 159 rm->oneinit = hw->oneinit; in r535_bar_new_() 160 rm->bar1.init = r535_bar_bar1_init; in r535_bar_new_() 161 rm->bar1.fini = r535_bar_bar1_fini; in r535_bar_new_() 162 rm->bar1.wait = r535_bar_bar1_wait; in r535_bar_new_() 163 rm->bar1.vmm = hw->bar1.vmm; in r535_bar_new_() 164 rm->bar2.init = r535_bar_bar2_init; in r535_bar_new_() [all …]
|
/linux-6.12.1/tools/perf/arch/powerpc/util/ |
D | perf_regs.c | 138 regmatch_t rm[5]; in arch_sdt_arg_parse_op() local 151 if (!regexec(&sdt_op_regex1, old_op, 3, rm, 0)) { in arch_sdt_arg_parse_op() 155 new_len += (int)(rm[2].rm_eo - rm[2].rm_so); in arch_sdt_arg_parse_op() 162 (int)(rm[2].rm_eo - rm[2].rm_so), old_op + rm[2].rm_so); in arch_sdt_arg_parse_op() 163 } else if (!regexec(&sdt_op_regex2, old_op, 5, rm, 0)) { in arch_sdt_arg_parse_op() 168 prefix = (rm[1].rm_so == -1) ? '+' : '-'; in arch_sdt_arg_parse_op() 171 new_len += (int)(rm[2].rm_eo - rm[2].rm_so); in arch_sdt_arg_parse_op() 172 new_len += (int)(rm[4].rm_eo - rm[4].rm_so); in arch_sdt_arg_parse_op() 179 (int)(rm[2].rm_eo - rm[2].rm_so), old_op + rm[2].rm_so, in arch_sdt_arg_parse_op() 180 (int)(rm[4].rm_eo - rm[4].rm_so), old_op + rm[4].rm_so); in arch_sdt_arg_parse_op()
|
/linux-6.12.1/drivers/gpu/drm/nouveau/nvkm/engine/fifo/ |
D | r535.c | 69 nvkm_gsp_rm_free(&chan->rm.object); in r535_chan_ramfc_clear() 71 dma_free_coherent(fifo->engine.subdev.device->dev, fifo->rm.mthdbuf_size, in r535_chan_ramfc_clear() 72 chan->rm.mthdbuf.ptr, chan->rm.mthdbuf.addr); in r535_chan_ramfc_clear() 74 nvkm_cgrp_vctx_put(chan->cgrp, &chan->rm.grctx); in r535_chan_ramfc_clear() 105 chan->rm.mthdbuf.ptr = dma_alloc_coherent(fifo->engine.subdev.device->dev, in r535_chan_ramfc_write() 106 fifo->rm.mthdbuf_size, in r535_chan_ramfc_write() 107 &chan->rm.mthdbuf.addr, GFP_KERNEL); in r535_chan_ramfc_write() 108 if (!chan->rm.mthdbuf.ptr) in r535_chan_ramfc_write() 111 args = nvkm_gsp_rm_alloc_get(&chan->vmm->rm.device.object, 0xf1f00000 | chan->id, in r535_chan_ramfc_write() 113 &chan->rm.object); in r535_chan_ramfc_write() [all …]
|
/linux-6.12.1/tools/perf/arch/arm64/util/ |
D | perf_regs.c | 98 regmatch_t rm[5]; in arch_sdt_arg_parse_op() local 104 if (!regexec(&sdt_op_regex1, old_op, 3, rm, 0)) { in arch_sdt_arg_parse_op() 107 new_len += (int)(rm[1].rm_eo - rm[1].rm_so); in arch_sdt_arg_parse_op() 114 (int)(rm[1].rm_eo - rm[1].rm_so), old_op + rm[1].rm_so); in arch_sdt_arg_parse_op() 115 } else if (!regexec(&sdt_op_regex2, old_op, 5, rm, 0)) { in arch_sdt_arg_parse_op() 120 if (rm[2].rm_so == -1) in arch_sdt_arg_parse_op() 123 new_len += (int)(rm[2].rm_eo - rm[2].rm_so); in arch_sdt_arg_parse_op() 129 if (rm[2].rm_so == -1) in arch_sdt_arg_parse_op() 133 (int)(rm[2].rm_eo - rm[2].rm_so), in arch_sdt_arg_parse_op() 134 old_op + rm[2].rm_so); in arch_sdt_arg_parse_op()
|
/linux-6.12.1/arch/x86/math-emu/ |
D | reg_divide.c | 28 int FPU_div(int flags, int rm, int control_w) in FPU_div() argument 37 deststnr = rm; in FPU_div() 46 a = (FPU_REG *) rm; in FPU_div() 49 a = &st(rm); in FPU_div() 51 taga = FPU_gettagi(rm); in FPU_div() 58 b = (FPU_REG *) rm; in FPU_div() 61 b = &st(rm); in FPU_div() 63 tagb = FPU_gettagi(rm); in FPU_div() 133 return real_2op_NaN((FPU_REG *) rm, flags & 0x0f, 0, in FPU_div() 141 return real_2op_NaN(st0_ptr, tag, rm, in FPU_div() [all …]
|
/linux-6.12.1/drivers/gpu/drm/nouveau/nvkm/subdev/devinit/ |
D | r535.c | 36 struct nvkm_devinit_func *rm; in r535_devinit_new() local 39 if (!(rm = kzalloc(sizeof(*rm), GFP_KERNEL))) in r535_devinit_new() 42 rm->dtor = r535_devinit_dtor; in r535_devinit_new() 43 rm->post = hw->post; in r535_devinit_new() 44 rm->disable = hw->disable; in r535_devinit_new() 46 ret = nv50_devinit_new_(rm, device, type, inst, pdevinit); in r535_devinit_new() 48 kfree(rm); in r535_devinit_new()
|
/linux-6.12.1/drivers/gpu/drm/nouveau/nvkm/subdev/vfn/ |
D | r535.c | 35 struct nvkm_vfn_func *rm; in r535_vfn_new() local 38 if (!(rm = kzalloc(sizeof(*rm), GFP_KERNEL))) in r535_vfn_new() 41 rm->dtor = r535_vfn_dtor; in r535_vfn_new() 42 rm->intr = hw->intr; in r535_vfn_new() 43 rm->user = hw->user; in r535_vfn_new() 45 ret = nvkm_vfn_new_(rm, device, type, inst, addr, pvfn); in r535_vfn_new() 47 kfree(rm); in r535_vfn_new()
|
/linux-6.12.1/sound/soc/codecs/ |
D | tas5805m.c | 174 static void set_dsp_scale(struct regmap *rm, int offset, int vol) in set_dsp_scale() argument 185 regmap_bulk_write(rm, offset, v, ARRAY_SIZE(v)); in set_dsp_scale() 190 struct regmap *rm = tas5805m->regmap; in tas5805m_refresh() local 195 regmap_write(rm, REG_PAGE, 0x00); in tas5805m_refresh() 196 regmap_write(rm, REG_BOOK, 0x8c); in tas5805m_refresh() 197 regmap_write(rm, REG_PAGE, 0x2a); in tas5805m_refresh() 203 set_dsp_scale(rm, 0x24, tas5805m->vol[0]); in tas5805m_refresh() 204 set_dsp_scale(rm, 0x28, tas5805m->vol[1]); in tas5805m_refresh() 206 regmap_write(rm, REG_PAGE, 0x00); in tas5805m_refresh() 207 regmap_write(rm, REG_BOOK, 0x00); in tas5805m_refresh() [all …]
|