/linux-6.12.1/drivers/gpu/drm/amd/display/amdgpu_dm/ |
D | amdgpu_dm_crtc.c | 194 struct idle_workqueue *idle_work; in amdgpu_dm_idle_worker() local 196 idle_work = container_of(work, struct idle_workqueue, work); in amdgpu_dm_idle_worker() 197 idle_work->dm->idle_workqueue->running = true; in amdgpu_dm_idle_worker() 199 while (idle_work->enable) { in amdgpu_dm_idle_worker() 201 mutex_lock(&idle_work->dm->dc_lock); in amdgpu_dm_idle_worker() 202 if (!idle_work->dm->dc->idle_optimizations_allowed) { in amdgpu_dm_idle_worker() 203 mutex_unlock(&idle_work->dm->dc_lock); in amdgpu_dm_idle_worker() 206 dc_allow_idle_optimizations(idle_work->dm->dc, false); in amdgpu_dm_idle_worker() 208 mutex_unlock(&idle_work->dm->dc_lock); in amdgpu_dm_idle_worker() 210 mutex_lock(&idle_work->dm->dc_lock); in amdgpu_dm_idle_worker() [all …]
|
/linux-6.12.1/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_jpeg.c | 41 INIT_DELAYED_WORK(&adev->jpeg.idle_work, amdgpu_jpeg_idle_work_handler); in amdgpu_jpeg_sw_init() 95 cancel_delayed_work_sync(&adev->jpeg.idle_work); in amdgpu_jpeg_suspend() 108 container_of(work, struct amdgpu_device, jpeg.idle_work.work); in amdgpu_jpeg_idle_work_handler() 124 schedule_delayed_work(&adev->jpeg.idle_work, JPEG_IDLE_TIMEOUT); in amdgpu_jpeg_idle_work_handler() 132 cancel_delayed_work_sync(&adev->jpeg.idle_work); in amdgpu_jpeg_ring_begin_use() 143 schedule_delayed_work(&ring->adev->jpeg.idle_work, JPEG_IDLE_TIMEOUT); in amdgpu_jpeg_ring_end_use()
|
D | amdgpu_vce.c | 195 INIT_DELAYED_WORK(&adev->vce.idle_work, amdgpu_vce_idle_work_handler); in amdgpu_vce_sw_init() 264 cancel_delayed_work_sync(&adev->vce.idle_work); in amdgpu_vce_suspend() 335 container_of(work, struct amdgpu_device, vce.idle_work.work); in amdgpu_vce_idle_work_handler() 352 schedule_delayed_work(&adev->vce.idle_work, VCE_IDLE_TIMEOUT); in amdgpu_vce_idle_work_handler() 372 set_clocks = !cancel_delayed_work_sync(&adev->vce.idle_work); in amdgpu_vce_ring_begin_use() 398 schedule_delayed_work(&ring->adev->vce.idle_work, VCE_IDLE_TIMEOUT); in amdgpu_vce_ring_end_use()
|
D | amdgpu_vpe.c | 327 container_of(work, struct amdgpu_device, vpe.idle_work.work); in vpe_idle_work_handler() 335 schedule_delayed_work(&adev->vpe.idle_work, VPE_IDLE_TIMEOUT); in vpe_idle_work_handler() 354 INIT_DELAYED_WORK(&adev->vpe.idle_work, vpe_idle_work_handler); in vpe_common_init() 441 cancel_delayed_work_sync(&adev->vpe.idle_work); in vpe_suspend() 843 cancel_delayed_work_sync(&adev->vpe.idle_work); in vpe_ring_begin_use() 867 schedule_delayed_work(&adev->vpe.idle_work, VPE_IDLE_TIMEOUT); in vpe_ring_end_use()
|
D | amdgpu_vce.h | 45 struct delayed_work idle_work; member
|
D | amdgpu_uvd.h | 66 struct delayed_work idle_work; member
|
D | amdgpu_uvd.c | 192 INIT_DELAYED_WORK(&adev->uvd.idle_work, amdgpu_uvd_idle_work_handler); in amdgpu_uvd_sw_init() 428 cancel_delayed_work_sync(&adev->uvd.idle_work); in amdgpu_uvd_prepare_suspend() 1262 container_of(work, struct amdgpu_device, uvd.idle_work.work); in amdgpu_uvd_idle_work_handler() 1285 schedule_delayed_work(&adev->uvd.idle_work, UVD_IDLE_TIMEOUT); in amdgpu_uvd_idle_work_handler() 1297 set_clocks = !cancel_delayed_work_sync(&adev->uvd.idle_work); in amdgpu_uvd_ring_begin_use() 1314 schedule_delayed_work(&ring->adev->uvd.idle_work, UVD_IDLE_TIMEOUT); in amdgpu_uvd_ring_end_use()
|
D | amdgpu_vpe.h | 77 struct delayed_work idle_work; member
|
D | amdgpu_vcn.c | 121 INIT_DELAYED_WORK(&adev->vcn.idle_work, amdgpu_vcn_idle_work_handler); in amdgpu_vcn_sw_init() 305 cancel_delayed_work_sync(&adev->vcn.idle_work); in amdgpu_vcn_suspend() 380 container_of(work, struct amdgpu_device, vcn.idle_work.work); in amdgpu_vcn_idle_work_handler() 418 schedule_delayed_work(&adev->vcn.idle_work, VCN_IDLE_TIMEOUT); in amdgpu_vcn_idle_work_handler() 429 if (!cancel_delayed_work_sync(&adev->vcn.idle_work)) { in amdgpu_vcn_ring_begin_use() 478 schedule_delayed_work(&ring->adev->vcn.idle_work, VCN_IDLE_TIMEOUT); in amdgpu_vcn_ring_end_use()
|
D | amdgpu_jpeg.h | 121 struct delayed_work idle_work; member
|
D | vce_v2_0.c | 482 cancel_delayed_work_sync(&adev->vce.idle_work); in vce_v2_0_hw_fini() 504 cancel_delayed_work_sync(&adev->vce.idle_work); in vce_v2_0_suspend()
|
D | vcn_v1_0.c | 152 adev->vcn.idle_work.work.func = vcn_v1_0_idle_work_handler; in vcn_v1_0_sw_init() 279 cancel_delayed_work_sync(&adev->vcn.idle_work); in vcn_v1_0_hw_fini() 303 idle_work_unexecuted = cancel_delayed_work_sync(&adev->vcn.idle_work); in vcn_v1_0_suspend() 1832 container_of(work, struct amdgpu_device, vcn.idle_work.work); in vcn_v1_0_idle_work_handler() 1865 schedule_delayed_work(&adev->vcn.idle_work, VCN_IDLE_TIMEOUT); in vcn_v1_0_idle_work_handler() 1872 bool set_clocks = !cancel_delayed_work_sync(&adev->vcn.idle_work); in vcn_v1_0_ring_begin_use() 1924 schedule_delayed_work(&ring->adev->vcn.idle_work, VCN_IDLE_TIMEOUT); in vcn_v1_0_ring_end_use()
|
D | uvd_v3_1.c | 699 cancel_delayed_work_sync(&adev->uvd.idle_work); in uvd_v3_1_hw_fini() 730 cancel_delayed_work_sync(&adev->uvd.idle_work); in uvd_v3_1_suspend()
|
D | uvd_v4_2.c | 213 cancel_delayed_work_sync(&adev->uvd.idle_work); in uvd_v4_2_hw_fini() 244 cancel_delayed_work_sync(&adev->uvd.idle_work); in uvd_v4_2_suspend()
|
D | uvd_v5_0.c | 211 cancel_delayed_work_sync(&adev->uvd.idle_work); in uvd_v5_0_hw_fini() 242 cancel_delayed_work_sync(&adev->uvd.idle_work); in uvd_v5_0_suspend()
|
D | vce_v3_0.c | 493 cancel_delayed_work_sync(&adev->vce.idle_work); in vce_v3_0_hw_fini() 519 cancel_delayed_work_sync(&adev->vce.idle_work); in vce_v3_0_suspend()
|
D | jpeg_v3_0.c | 167 cancel_delayed_work_sync(&adev->vcn.idle_work); in jpeg_v3_0_hw_fini()
|
D | amdgpu_vcn.h | 308 struct delayed_work idle_work; member
|
/linux-6.12.1/drivers/gpu/drm/msm/ |
D | msm_gpu_devfreq.c | 195 msm_hrtimer_work_init(&df->idle_work, gpu->worker, msm_devfreq_idle_work, in msm_devfreq_init() 201 hrtimer_cancel(&df->idle_work.timer); in cancel_idle_work() 202 kthread_cancel_work_sync(&df->idle_work.work); in cancel_idle_work() 338 struct msm_gpu_devfreq, idle_work.work); in msm_devfreq_idle_work() 367 msm_hrtimer_queue_work(&df->idle_work, ms_to_ktime(1), in msm_devfreq_idle()
|
D | msm_gpu.h | 152 struct msm_hrtimer_work idle_work; member
|
/linux-6.12.1/drivers/gpu/drm/radeon/ |
D | radeon_vce.c | 65 INIT_DELAYED_WORK(&rdev->vce.idle_work, radeon_vce_idle_work_handler); in radeon_vce_init() 264 container_of(work, struct radeon_device, vce.idle_work.work); in radeon_vce_idle_work_handler() 274 schedule_delayed_work(&rdev->vce.idle_work, in radeon_vce_idle_work_handler() 289 bool set_clocks = !cancel_delayed_work_sync(&rdev->vce.idle_work); in radeon_vce_note_usage() 290 set_clocks &= schedule_delayed_work(&rdev->vce.idle_work, in radeon_vce_note_usage()
|
D | radeon_uvd.c | 72 INIT_DELAYED_WORK(&rdev->uvd.idle_work, radeon_uvd_idle_work_handler); in radeon_uvd_init() 862 container_of(work, struct radeon_device, uvd.idle_work.work); in radeon_uvd_idle_work_handler() 873 schedule_delayed_work(&rdev->uvd.idle_work, in radeon_uvd_idle_work_handler() 881 bool set_clocks = !cancel_delayed_work_sync(&rdev->uvd.idle_work); in radeon_uvd_note_usage() 882 set_clocks &= schedule_delayed_work(&rdev->uvd.idle_work, in radeon_uvd_note_usage()
|
/linux-6.12.1/drivers/net/wireless/zydas/zd1211rw/ |
D | zd_usb.h | 175 struct delayed_work idle_work; member
|
D | zd_usb.c | 848 cancel_delayed_work_sync(&rx->idle_work); in zd_usb_disable_rx() 1130 container_of(work, struct zd_usb, rx.idle_work.work); in zd_rx_idle_timer_handler() 1153 mod_delayed_work(zd_workqueue, &rx->idle_work, ZD_RX_IDLE_INTERVAL); in zd_usb_reset_rx_idle_timer() 1179 INIT_DELAYED_WORK(&rx->idle_work, zd_rx_idle_timer_handler); in init_usb_rx()
|
/linux-6.12.1/drivers/bus/ |
D | ti-sysc.c | 160 struct delayed_work idle_work; member 2914 ddata = container_of(work, struct sysc, idle_work.work); in ti_sysc_idle() 3273 INIT_DELAYED_WORK(&ddata->idle_work, ti_sysc_idle); in sysc_probe() 3279 schedule_delayed_work(&ddata->idle_work, 3000); in sysc_probe() 3304 if (cancel_delayed_work_sync(&ddata->idle_work)) in sysc_remove() 3305 ti_sysc_idle(&ddata->idle_work.work); in sysc_remove()
|