/linux-6.12.1/drivers/gpu/drm/i915/gvt/ |
D | sched_policy.c | 245 mutex_lock(&gvt->sched_lock); in intel_gvt_schedule() 261 mutex_unlock(&gvt->sched_lock); in intel_gvt_schedule() 384 mutex_lock(&gvt->sched_lock); in intel_gvt_init_sched_policy() 387 mutex_unlock(&gvt->sched_lock); in intel_gvt_init_sched_policy() 394 mutex_lock(&gvt->sched_lock); in intel_gvt_clean_sched_policy() 396 mutex_unlock(&gvt->sched_lock); in intel_gvt_clean_sched_policy() 410 mutex_lock(&vgpu->gvt->sched_lock); in intel_vgpu_init_sched_policy() 412 mutex_unlock(&vgpu->gvt->sched_lock); in intel_vgpu_init_sched_policy() 419 mutex_lock(&vgpu->gvt->sched_lock); in intel_vgpu_clean_sched_policy() 421 mutex_unlock(&vgpu->gvt->sched_lock); in intel_vgpu_clean_sched_policy() [all …]
|
D | scheduler.c | 853 mutex_lock(&gvt->sched_lock); in pick_next_workload() 902 mutex_unlock(&gvt->sched_lock); in pick_next_workload() 1077 mutex_lock(&gvt->sched_lock); in complete_current_workload() 1146 mutex_unlock(&gvt->sched_lock); in complete_current_workload()
|
D | gvt.h | 327 struct mutex sched_lock; member
|
/linux-6.12.1/net/netfilter/ipvs/ |
D | ip_vs_rr.c | 37 spin_lock_bh(&svc->sched_lock); in ip_vs_rr_del_dest() 44 spin_unlock_bh(&svc->sched_lock); in ip_vs_rr_del_dest() 62 spin_lock_bh(&svc->sched_lock); in ip_vs_rr_schedule() 84 spin_unlock_bh(&svc->sched_lock); in ip_vs_rr_schedule() 90 spin_unlock_bh(&svc->sched_lock); in ip_vs_rr_schedule()
|
D | ip_vs_lblcr.c | 401 spin_lock_bh(&svc->sched_lock); in ip_vs_lblcr_flush() 408 spin_unlock_bh(&svc->sched_lock); in ip_vs_lblcr_flush() 431 spin_lock(&svc->sched_lock); in ip_vs_lblcr_full_check() 440 spin_unlock(&svc->sched_lock); in ip_vs_lblcr_full_check() 486 spin_lock(&svc->sched_lock); in ip_vs_lblcr_check_expire() 495 spin_unlock(&svc->sched_lock); in ip_vs_lblcr_check_expire() 665 spin_lock_bh(&svc->sched_lock); in ip_vs_lblcr_schedule() 673 spin_unlock_bh(&svc->sched_lock); in ip_vs_lblcr_schedule() 688 spin_lock_bh(&svc->sched_lock); in ip_vs_lblcr_schedule() 691 spin_unlock_bh(&svc->sched_lock); in ip_vs_lblcr_schedule() [all …]
|
D | ip_vs_lblc.c | 235 spin_lock_bh(&svc->sched_lock); in ip_vs_lblc_flush() 243 spin_unlock_bh(&svc->sched_lock); in ip_vs_lblc_flush() 266 spin_lock(&svc->sched_lock); in ip_vs_lblc_full_check() 276 spin_unlock(&svc->sched_lock); in ip_vs_lblc_full_check() 322 spin_lock(&svc->sched_lock); in ip_vs_lblc_check_expire() 331 spin_unlock(&svc->sched_lock); in ip_vs_lblc_check_expire() 518 spin_lock_bh(&svc->sched_lock); in ip_vs_lblc_schedule() 521 spin_unlock_bh(&svc->sched_lock); in ip_vs_lblc_schedule()
|
D | ip_vs_wrr.c | 143 spin_lock_bh(&svc->sched_lock); in ip_vs_wrr_dest_changed() 151 spin_unlock_bh(&svc->sched_lock); in ip_vs_wrr_dest_changed() 169 spin_lock_bh(&svc->sched_lock); in ip_vs_wrr_schedule() 221 spin_unlock_bh(&svc->sched_lock); in ip_vs_wrr_schedule()
|
/linux-6.12.1/drivers/net/ethernet/intel/ice/devlink/ |
D | devlink.c | 810 mutex_lock(&pi->sched_lock); in ice_devlink_rate_init_tx_topology() 813 mutex_unlock(&pi->sched_lock); in ice_devlink_rate_init_tx_topology() 837 mutex_lock(&pi->sched_lock); in ice_devlink_rate_clear_tx_topology() 839 mutex_unlock(&pi->sched_lock); in ice_devlink_rate_clear_tx_topology() 856 mutex_lock(&pi->sched_lock); in ice_set_object_tx_share() 860 mutex_unlock(&pi->sched_lock); in ice_set_object_tx_share() 882 mutex_lock(&pi->sched_lock); in ice_set_object_tx_max() 886 mutex_unlock(&pi->sched_lock); in ice_set_object_tx_max() 913 mutex_lock(&pi->sched_lock); in ice_set_object_tx_priority() 916 mutex_unlock(&pi->sched_lock); in ice_set_object_tx_priority() [all …]
|
/linux-6.12.1/drivers/gpu/drm/etnaviv/ |
D | etnaviv_sched.c | 108 mutex_lock(&gpu->sched_lock); in etnaviv_sched_push_job() 127 mutex_unlock(&gpu->sched_lock); in etnaviv_sched_push_job()
|
D | etnaviv_gpu.h | 116 struct mutex sched_lock; member
|
/linux-6.12.1/drivers/gpu/drm/v3d/ |
D | v3d_submit.c | 993 mutex_lock(&v3d->sched_lock); in v3d_submit_cl_ioctl() 1019 mutex_unlock(&v3d->sched_lock); in v3d_submit_cl_ioctl() 1035 mutex_unlock(&v3d->sched_lock); in v3d_submit_cl_ioctl() 1126 mutex_lock(&v3d->sched_lock); in v3d_submit_tfu_ioctl() 1128 mutex_unlock(&v3d->sched_lock); in v3d_submit_tfu_ioctl() 1207 mutex_lock(&v3d->sched_lock); in v3d_submit_csd_ioctl() 1216 mutex_unlock(&v3d->sched_lock); in v3d_submit_csd_ioctl() 1231 mutex_unlock(&v3d->sched_lock); in v3d_submit_csd_ioctl() 1328 mutex_lock(&v3d->sched_lock); in v3d_submit_cpu_ioctl() 1351 mutex_unlock(&v3d->sched_lock); in v3d_submit_cpu_ioctl() [all …]
|
D | v3d_gem.c | 265 ret = drmm_mutex_init(dev, &v3d->sched_lock); in v3d_gem_init()
|
D | v3d_drv.h | 170 struct mutex sched_lock; member
|
/linux-6.12.1/drivers/net/ethernet/intel/ice/ |
D | ice_sched.c | 841 mutex_lock(&pi->sched_lock); in ice_sched_clear_port() 843 mutex_unlock(&pi->sched_lock); in ice_sched_clear_port() 844 mutex_destroy(&pi->sched_lock); in ice_sched_clear_port() 1303 mutex_init(&pi->sched_lock); in ice_sched_init_port() 2034 mutex_lock(&pi->sched_lock); in ice_sched_rm_vsi_cfg() 2084 mutex_unlock(&pi->sched_lock); in ice_sched_rm_vsi_cfg() 2683 mutex_lock(&pi->sched_lock); in ice_cfg_agg() 2687 mutex_unlock(&pi->sched_lock); in ice_cfg_agg() 2955 mutex_lock(&pi->sched_lock); in ice_move_vsi_to_agg() 2961 mutex_unlock(&pi->sched_lock); in ice_move_vsi_to_agg() [all …]
|
D | ice_common.c | 4788 mutex_lock(&pi->sched_lock); in ice_ena_vsi_txq() 4852 mutex_unlock(&pi->sched_lock); in ice_ena_vsi_txq() 4899 mutex_lock(&pi->sched_lock); in ice_dis_vsi_txq() 4930 mutex_unlock(&pi->sched_lock); in ice_dis_vsi_txq() 4957 mutex_lock(&pi->sched_lock); in ice_cfg_vsi_qs() 4970 mutex_unlock(&pi->sched_lock); in ice_cfg_vsi_qs() 5041 mutex_lock(&pi->sched_lock); in ice_ena_vsi_rdma_qset() 5083 mutex_unlock(&pi->sched_lock); in ice_ena_vsi_rdma_qset() 5110 mutex_lock(&pi->sched_lock); in ice_dis_vsi_rdma_qset() 5133 mutex_unlock(&pi->sched_lock); in ice_dis_vsi_rdma_qset()
|
D | ice_dcb.c | 1635 mutex_lock(&pi->sched_lock); in ice_query_port_ets() 1639 mutex_unlock(&pi->sched_lock); in ice_query_port_ets()
|
D | ice_type.h | 751 struct mutex sched_lock; /* protect access to TXSched tree */ member
|
/linux-6.12.1/drivers/gpu/drm/panfrost/ |
D | panfrost_device.h | 135 struct mutex sched_lock; member
|
D | panfrost_job.c | 303 mutex_lock(&pfdev->sched_lock); in panfrost_job_push() 311 mutex_unlock(&pfdev->sched_lock); in panfrost_job_push() 319 mutex_unlock(&pfdev->sched_lock); in panfrost_job_push()
|
D | panfrost_device.c | 204 mutex_init(&pfdev->sched_lock); in panfrost_device_init()
|
/linux-6.12.1/drivers/hv/ |
D | channel.c | 896 spin_lock_irqsave(&channel->sched_lock, flags); in vmbus_reset_channel_cb() 898 spin_unlock_irqrestore(&channel->sched_lock, flags); in vmbus_reset_channel_cb()
|
D | vmbus_drv.c | 1250 spin_lock(&channel->sched_lock); in vmbus_chan_sched() 1273 spin_unlock(&channel->sched_lock); in vmbus_chan_sched()
|
D | channel_mgmt.c | 360 spin_lock_init(&channel->sched_lock); in alloc_channel()
|
/linux-6.12.1/drivers/pci/controller/ |
D | pci-hyperv.c | 1992 spin_lock_irqsave(&channel->sched_lock, flags); in hv_compose_msi_msg() 1994 spin_unlock_irqrestore(&channel->sched_lock, flags); in hv_compose_msi_msg() 1998 spin_unlock_irqrestore(&channel->sched_lock, flags); in hv_compose_msi_msg()
|
/linux-6.12.1/include/linux/ |
D | hyperv.h | 899 spinlock_t sched_lock; member
|