1 // SPDX-License-Identifier: (GPL-2.0 OR BSD-3-Clause)
2 /*
3  * Wave5 series multi-standard codec IP - wave5 backend logic
4  *
5  * Copyright (C) 2021-2023 CHIPS&MEDIA INC
6  */
7 
8 #include <linux/iopoll.h>
9 #include <linux/bitfield.h>
10 #include "wave5-vpu.h"
11 #include "wave5.h"
12 #include "wave5-regdefine.h"
13 
14 #define FIO_TIMEOUT			10000000
15 #define FIO_CTRL_READY			BIT(31)
16 #define FIO_CTRL_WRITE			BIT(16)
17 #define VPU_BUSY_CHECK_TIMEOUT		10000000
18 #define QUEUE_REPORT_MASK		0xffff
19 
20 /* Encoder support fields */
21 #define W521_FEATURE_HEVC10BIT_ENC	BIT(3)
22 #define W521_FEATURE_AVC10BIT_ENC	BIT(11)
23 #define W521_FEATURE_AVC_ENCODER	BIT(1)
24 #define W521_FEATURE_HEVC_ENCODER	BIT(0)
25 
26 /* Decoder support fields */
27 #define W521_FEATURE_AVC_DECODER	BIT(3)
28 #define W521_FEATURE_HEVC_DECODER	BIT(2)
29 #define W515_FEATURE_HEVC10BIT_DEC	BIT(1)
30 #define W515_FEATURE_HEVC_DECODER	BIT(0)
31 
32 #define W521_FEATURE_BACKBONE		BIT(16)
33 #define W521_FEATURE_VCORE_BACKBONE	BIT(22)
34 #define W521_FEATURE_VCPU_BACKBONE	BIT(28)
35 
36 #define REMAP_CTRL_MAX_SIZE_BITS	((W5_REMAP_MAX_SIZE >> 12) & 0x1ff)
37 #define REMAP_CTRL_REGISTER_VALUE(index)	(			\
38 	(BIT(31) | (index << 12) | BIT(11) | REMAP_CTRL_MAX_SIZE_BITS)	\
39 )
40 
41 #define FASTIO_ADDRESS_MASK		GENMASK(15, 0)
42 #define SEQ_PARAM_PROFILE_MASK		GENMASK(30, 24)
43 
44 static void _wave5_print_reg_err(struct vpu_device *vpu_dev, u32 reg_fail_reason,
45 				 const char *func);
46 #define PRINT_REG_ERR(dev, reason)	_wave5_print_reg_err((dev), (reason), __func__)
47 
cmd_to_str(int cmd,bool is_dec)48 static inline const char *cmd_to_str(int cmd, bool is_dec)
49 {
50 	switch (cmd) {
51 	case W5_INIT_VPU:
52 		return "W5_INIT_VPU";
53 	case W5_WAKEUP_VPU:
54 		return "W5_WAKEUP_VPU";
55 	case W5_SLEEP_VPU:
56 		return "W5_SLEEP_VPU";
57 	case W5_CREATE_INSTANCE:
58 		return "W5_CREATE_INSTANCE";
59 	case W5_FLUSH_INSTANCE:
60 		return "W5_FLUSH_INSTANCE";
61 	case W5_DESTROY_INSTANCE:
62 		return "W5_DESTROY_INSTANCE";
63 	case W5_INIT_SEQ:
64 		return "W5_INIT_SEQ";
65 	case W5_SET_FB:
66 		return "W5_SET_FB";
67 	case W5_DEC_ENC_PIC:
68 		if (is_dec)
69 			return "W5_DEC_PIC";
70 		return "W5_ENC_PIC";
71 	case W5_ENC_SET_PARAM:
72 		return "W5_ENC_SET_PARAM";
73 	case W5_QUERY:
74 		return "W5_QUERY";
75 	case W5_UPDATE_BS:
76 		return "W5_UPDATE_BS";
77 	case W5_MAX_VPU_COMD:
78 		return "W5_MAX_VPU_COMD";
79 	default:
80 		return "UNKNOWN";
81 	}
82 }
83 
_wave5_print_reg_err(struct vpu_device * vpu_dev,u32 reg_fail_reason,const char * func)84 static void _wave5_print_reg_err(struct vpu_device *vpu_dev, u32 reg_fail_reason,
85 				 const char *func)
86 {
87 	struct device *dev = vpu_dev->dev;
88 	u32 reg_val;
89 
90 	switch (reg_fail_reason) {
91 	case WAVE5_SYSERR_QUEUEING_FAIL:
92 		reg_val = vpu_read_reg(vpu_dev, W5_RET_QUEUE_FAIL_REASON);
93 		dev_dbg(dev, "%s: queueing failure: 0x%x\n", func, reg_val);
94 		break;
95 	case WAVE5_SYSERR_RESULT_NOT_READY:
96 		dev_err(dev, "%s: result not ready: 0x%x\n", func, reg_fail_reason);
97 		break;
98 	case WAVE5_SYSERR_ACCESS_VIOLATION_HW:
99 		dev_err(dev, "%s: access violation: 0x%x\n", func, reg_fail_reason);
100 		break;
101 	case WAVE5_SYSERR_WATCHDOG_TIMEOUT:
102 		dev_err(dev, "%s: watchdog timeout: 0x%x\n", func, reg_fail_reason);
103 		break;
104 	case WAVE5_SYSERR_BUS_ERROR:
105 		dev_err(dev, "%s: bus error: 0x%x\n", func, reg_fail_reason);
106 		break;
107 	case WAVE5_SYSERR_DOUBLE_FAULT:
108 		dev_err(dev, "%s: double fault: 0x%x\n", func, reg_fail_reason);
109 		break;
110 	case WAVE5_SYSERR_VPU_STILL_RUNNING:
111 		dev_err(dev, "%s: still running: 0x%x\n", func, reg_fail_reason);
112 		break;
113 	case WAVE5_SYSERR_VLC_BUF_FULL:
114 		dev_err(dev, "%s: vlc buf full: 0x%x\n", func, reg_fail_reason);
115 		break;
116 	default:
117 		dev_err(dev, "%s: failure:: 0x%x\n", func, reg_fail_reason);
118 		break;
119 	}
120 }
121 
wave5_wait_fio_readl(struct vpu_device * vpu_dev,u32 addr,u32 val)122 static int wave5_wait_fio_readl(struct vpu_device *vpu_dev, u32 addr, u32 val)
123 {
124 	u32 ctrl;
125 	int ret;
126 
127 	ctrl = addr & 0xffff;
128 	wave5_vdi_write_register(vpu_dev, W5_VPU_FIO_CTRL_ADDR, ctrl);
129 	ret = read_poll_timeout(wave5_vdi_read_register, ctrl, ctrl & FIO_CTRL_READY,
130 				0, FIO_TIMEOUT, false, vpu_dev, W5_VPU_FIO_CTRL_ADDR);
131 	if (ret)
132 		return ret;
133 
134 	if (wave5_vdi_read_register(vpu_dev, W5_VPU_FIO_DATA) != val)
135 		return -ETIMEDOUT;
136 
137 	return 0;
138 }
139 
wave5_fio_writel(struct vpu_device * vpu_dev,unsigned int addr,unsigned int data)140 static void wave5_fio_writel(struct vpu_device *vpu_dev, unsigned int addr, unsigned int data)
141 {
142 	int ret;
143 	unsigned int ctrl;
144 
145 	wave5_vdi_write_register(vpu_dev, W5_VPU_FIO_DATA, data);
146 	ctrl = FIELD_GET(FASTIO_ADDRESS_MASK, addr);
147 	ctrl |= FIO_CTRL_WRITE;
148 	wave5_vdi_write_register(vpu_dev, W5_VPU_FIO_CTRL_ADDR, ctrl);
149 	ret = read_poll_timeout(wave5_vdi_read_register, ctrl, ctrl & FIO_CTRL_READY, 0,
150 				FIO_TIMEOUT, false, vpu_dev, W5_VPU_FIO_CTRL_ADDR);
151 	if (ret)
152 		dev_dbg_ratelimited(vpu_dev->dev, "FIO write timeout: addr=0x%x data=%x\n",
153 				    ctrl, data);
154 }
155 
wave5_wait_bus_busy(struct vpu_device * vpu_dev,unsigned int addr)156 static int wave5_wait_bus_busy(struct vpu_device *vpu_dev, unsigned int addr)
157 {
158 	u32 gdi_status_check_value = 0x3f;
159 
160 	if (vpu_dev->product_code == WAVE515_CODE)
161 		gdi_status_check_value = 0x0738;
162 	if (vpu_dev->product_code == WAVE521C_CODE ||
163 	    vpu_dev->product_code == WAVE521_CODE ||
164 	    vpu_dev->product_code == WAVE521E1_CODE)
165 		gdi_status_check_value = 0x00ff1f3f;
166 
167 	return wave5_wait_fio_readl(vpu_dev, addr, gdi_status_check_value);
168 }
169 
wave5_wait_vpu_busy(struct vpu_device * vpu_dev,unsigned int addr)170 static int wave5_wait_vpu_busy(struct vpu_device *vpu_dev, unsigned int addr)
171 {
172 	u32 data;
173 
174 	return read_poll_timeout(wave5_vdi_read_register, data, data == 0,
175 				 0, VPU_BUSY_CHECK_TIMEOUT, false, vpu_dev, addr);
176 }
177 
wave5_wait_vcpu_bus_busy(struct vpu_device * vpu_dev,unsigned int addr)178 static int wave5_wait_vcpu_bus_busy(struct vpu_device *vpu_dev, unsigned int addr)
179 {
180 	return wave5_wait_fio_readl(vpu_dev, addr, 0);
181 }
182 
wave5_vpu_is_init(struct vpu_device * vpu_dev)183 bool wave5_vpu_is_init(struct vpu_device *vpu_dev)
184 {
185 	return vpu_read_reg(vpu_dev, W5_VCPU_CUR_PC) != 0;
186 }
187 
wave5_vpu_get_product_id(struct vpu_device * vpu_dev)188 unsigned int wave5_vpu_get_product_id(struct vpu_device *vpu_dev)
189 {
190 	u32 val = vpu_read_reg(vpu_dev, W5_PRODUCT_NUMBER);
191 
192 	switch (val) {
193 	case WAVE515_CODE:
194 		return PRODUCT_ID_515;
195 	case WAVE521C_CODE:
196 		return PRODUCT_ID_521;
197 	case WAVE521_CODE:
198 	case WAVE521C_DUAL_CODE:
199 	case WAVE521E1_CODE:
200 	case WAVE511_CODE:
201 	case WAVE517_CODE:
202 	case WAVE537_CODE:
203 		dev_err(vpu_dev->dev, "Unsupported product id (%x)\n", val);
204 		break;
205 	default:
206 		dev_err(vpu_dev->dev, "Invalid product id (%x)\n", val);
207 		break;
208 	}
209 
210 	return PRODUCT_ID_NONE;
211 }
212 
wave5_bit_issue_command(struct vpu_device * vpu_dev,struct vpu_instance * inst,u32 cmd)213 static void wave5_bit_issue_command(struct vpu_device *vpu_dev, struct vpu_instance *inst, u32 cmd)
214 {
215 	u32 instance_index;
216 	u32 codec_mode;
217 
218 	if (inst) {
219 		instance_index = inst->id;
220 		codec_mode = inst->std;
221 
222 		vpu_write_reg(vpu_dev, W5_CMD_INSTANCE_INFO, (codec_mode << 16) |
223 			      (instance_index & 0xffff));
224 		vpu_write_reg(vpu_dev, W5_VPU_BUSY_STATUS, 1);
225 	}
226 
227 	vpu_write_reg(vpu_dev, W5_COMMAND, cmd);
228 
229 	if (inst) {
230 		dev_dbg(vpu_dev->dev, "%s: cmd=0x%x (%s)\n", __func__, cmd,
231 			cmd_to_str(cmd, inst->type == VPU_INST_TYPE_DEC));
232 	} else {
233 		dev_dbg(vpu_dev->dev, "%s: cmd=0x%x\n", __func__, cmd);
234 	}
235 
236 	vpu_write_reg(vpu_dev, W5_VPU_HOST_INT_REQ, 1);
237 }
238 
wave5_vpu_firmware_command_queue_error_check(struct vpu_device * dev,u32 * fail_res)239 static int wave5_vpu_firmware_command_queue_error_check(struct vpu_device *dev, u32 *fail_res)
240 {
241 	u32 reason = 0;
242 
243 	/* Check if we were able to add a command into the VCPU QUEUE */
244 	if (!vpu_read_reg(dev, W5_RET_SUCCESS)) {
245 		reason = vpu_read_reg(dev, W5_RET_FAIL_REASON);
246 		PRINT_REG_ERR(dev, reason);
247 
248 		/*
249 		 * The fail_res argument will be either NULL or 0.
250 		 * If the fail_res argument is NULL, then just return -EIO.
251 		 * Otherwise, assign the reason to fail_res, so that the
252 		 * calling function can use it.
253 		 */
254 		if (fail_res)
255 			*fail_res = reason;
256 		else
257 			return -EIO;
258 
259 		if (reason == WAVE5_SYSERR_VPU_STILL_RUNNING)
260 			return -EBUSY;
261 	}
262 	return 0;
263 }
264 
send_firmware_command(struct vpu_instance * inst,u32 cmd,bool check_success,u32 * queue_status,u32 * fail_result)265 static int send_firmware_command(struct vpu_instance *inst, u32 cmd, bool check_success,
266 				 u32 *queue_status, u32 *fail_result)
267 {
268 	int ret;
269 
270 	wave5_bit_issue_command(inst->dev, inst, cmd);
271 	ret = wave5_wait_vpu_busy(inst->dev, W5_VPU_BUSY_STATUS);
272 	if (ret) {
273 		dev_warn(inst->dev->dev, "%s: command: '%s', timed out\n", __func__,
274 			 cmd_to_str(cmd, inst->type == VPU_INST_TYPE_DEC));
275 		return -ETIMEDOUT;
276 	}
277 
278 	if (queue_status)
279 		*queue_status = vpu_read_reg(inst->dev, W5_RET_QUEUE_STATUS);
280 
281 	/* In some cases we want to send multiple commands before checking
282 	 * whether they are queued properly
283 	 */
284 	if (!check_success)
285 		return 0;
286 
287 	return wave5_vpu_firmware_command_queue_error_check(inst->dev, fail_result);
288 }
289 
wave5_send_query(struct vpu_device * vpu_dev,struct vpu_instance * inst,enum query_opt query_opt)290 static int wave5_send_query(struct vpu_device *vpu_dev, struct vpu_instance *inst,
291 			    enum query_opt query_opt)
292 {
293 	int ret;
294 
295 	vpu_write_reg(vpu_dev, W5_QUERY_OPTION, query_opt);
296 	vpu_write_reg(vpu_dev, W5_VPU_BUSY_STATUS, 1);
297 	wave5_bit_issue_command(vpu_dev, inst, W5_QUERY);
298 
299 	ret = wave5_wait_vpu_busy(vpu_dev, W5_VPU_BUSY_STATUS);
300 	if (ret) {
301 		dev_warn(vpu_dev->dev, "command: 'W5_QUERY', timed out opt=0x%x\n", query_opt);
302 		return ret;
303 	}
304 
305 	return wave5_vpu_firmware_command_queue_error_check(vpu_dev, NULL);
306 }
307 
setup_wave5_interrupts(struct vpu_device * vpu_dev)308 static void setup_wave5_interrupts(struct vpu_device *vpu_dev)
309 {
310 	u32 reg_val = 0;
311 
312 	if (vpu_dev->attr.support_encoders) {
313 		/* Encoder interrupt */
314 		reg_val |= BIT(INT_WAVE5_ENC_SET_PARAM);
315 		reg_val |= BIT(INT_WAVE5_ENC_PIC);
316 		reg_val |= BIT(INT_WAVE5_BSBUF_FULL);
317 	}
318 
319 	if (vpu_dev->attr.support_decoders) {
320 		/* Decoder interrupt */
321 		reg_val |= BIT(INT_WAVE5_INIT_SEQ);
322 		reg_val |= BIT(INT_WAVE5_DEC_PIC);
323 		reg_val |= BIT(INT_WAVE5_BSBUF_EMPTY);
324 	}
325 
326 	return vpu_write_reg(vpu_dev, W5_VPU_VINT_ENABLE, reg_val);
327 }
328 
setup_wave5_properties(struct device * dev)329 static int setup_wave5_properties(struct device *dev)
330 {
331 	struct vpu_device *vpu_dev = dev_get_drvdata(dev);
332 	struct vpu_attr *p_attr = &vpu_dev->attr;
333 	u32 reg_val;
334 	u8 *str;
335 	int ret;
336 	u32 hw_config_def0, hw_config_def1, hw_config_feature;
337 
338 	ret = wave5_send_query(vpu_dev, NULL, GET_VPU_INFO);
339 	if (ret)
340 		return ret;
341 
342 	reg_val = vpu_read_reg(vpu_dev, W5_RET_PRODUCT_NAME);
343 	str = (u8 *)&reg_val;
344 	p_attr->product_name[0] = str[3];
345 	p_attr->product_name[1] = str[2];
346 	p_attr->product_name[2] = str[1];
347 	p_attr->product_name[3] = str[0];
348 	p_attr->product_name[4] = 0;
349 
350 	p_attr->product_id = wave5_vpu_get_product_id(vpu_dev);
351 	p_attr->product_version = vpu_read_reg(vpu_dev, W5_RET_PRODUCT_VERSION);
352 	p_attr->fw_version = vpu_read_reg(vpu_dev, W5_RET_FW_VERSION);
353 	p_attr->customer_id = vpu_read_reg(vpu_dev, W5_RET_CUSTOMER_ID);
354 	hw_config_def0 = vpu_read_reg(vpu_dev, W5_RET_STD_DEF0);
355 	hw_config_def1 = vpu_read_reg(vpu_dev, W5_RET_STD_DEF1);
356 	hw_config_feature = vpu_read_reg(vpu_dev, W5_RET_CONF_FEATURE);
357 
358 	if (vpu_dev->product_code == WAVE515_CODE) {
359 		p_attr->support_hevc10bit_dec = FIELD_GET(W515_FEATURE_HEVC10BIT_DEC,
360 							  hw_config_feature);
361 		p_attr->support_decoders = FIELD_GET(W515_FEATURE_HEVC_DECODER,
362 						     hw_config_def1) << STD_HEVC;
363 	} else {
364 		p_attr->support_hevc10bit_enc = FIELD_GET(W521_FEATURE_HEVC10BIT_ENC,
365 							  hw_config_feature);
366 		p_attr->support_avc10bit_enc = FIELD_GET(W521_FEATURE_AVC10BIT_ENC,
367 							 hw_config_feature);
368 
369 		p_attr->support_decoders = FIELD_GET(W521_FEATURE_AVC_DECODER,
370 						     hw_config_def1) << STD_AVC;
371 		p_attr->support_decoders |= FIELD_GET(W521_FEATURE_HEVC_DECODER,
372 						      hw_config_def1) << STD_HEVC;
373 		p_attr->support_encoders = FIELD_GET(W521_FEATURE_AVC_ENCODER,
374 						     hw_config_def1) << STD_AVC;
375 		p_attr->support_encoders |= FIELD_GET(W521_FEATURE_HEVC_ENCODER,
376 						      hw_config_def1) << STD_HEVC;
377 
378 		p_attr->support_backbone = FIELD_GET(W521_FEATURE_BACKBONE,
379 						     hw_config_def0);
380 		p_attr->support_vcpu_backbone = FIELD_GET(W521_FEATURE_VCPU_BACKBONE,
381 							  hw_config_def0);
382 		p_attr->support_vcore_backbone = FIELD_GET(W521_FEATURE_VCORE_BACKBONE,
383 							   hw_config_def0);
384 	}
385 
386 	setup_wave5_interrupts(vpu_dev);
387 
388 	return 0;
389 }
390 
wave5_vpu_get_version(struct vpu_device * vpu_dev,u32 * revision)391 int wave5_vpu_get_version(struct vpu_device *vpu_dev, u32 *revision)
392 {
393 	u32 reg_val;
394 	int ret;
395 
396 	ret = wave5_send_query(vpu_dev, NULL, GET_VPU_INFO);
397 	if (ret)
398 		return ret;
399 
400 	reg_val = vpu_read_reg(vpu_dev, W5_RET_FW_VERSION);
401 	if (revision) {
402 		*revision = reg_val;
403 		return 0;
404 	}
405 
406 	return -EINVAL;
407 }
408 
remap_page(struct vpu_device * vpu_dev,dma_addr_t code_base,u32 index)409 static void remap_page(struct vpu_device *vpu_dev, dma_addr_t code_base, u32 index)
410 {
411 	vpu_write_reg(vpu_dev, W5_VPU_REMAP_CTRL, REMAP_CTRL_REGISTER_VALUE(index));
412 	vpu_write_reg(vpu_dev, W5_VPU_REMAP_VADDR, index * W5_REMAP_MAX_SIZE);
413 	vpu_write_reg(vpu_dev, W5_VPU_REMAP_PADDR, code_base + index * W5_REMAP_MAX_SIZE);
414 }
415 
wave5_vpu_init(struct device * dev,u8 * fw,size_t size)416 int wave5_vpu_init(struct device *dev, u8 *fw, size_t size)
417 {
418 	struct vpu_buf *common_vb;
419 	dma_addr_t code_base, temp_base;
420 	u32 code_size, temp_size;
421 	u32 i, reg_val, reason_code;
422 	int ret;
423 	struct vpu_device *vpu_dev = dev_get_drvdata(dev);
424 
425 	common_vb = &vpu_dev->common_mem;
426 
427 	code_base = common_vb->daddr;
428 
429 	if (vpu_dev->product_code == WAVE515_CODE)
430 		code_size = WAVE515_MAX_CODE_BUF_SIZE;
431 	else
432 		code_size = WAVE521_MAX_CODE_BUF_SIZE;
433 
434 	/* ALIGN TO 4KB */
435 	code_size &= ~0xfff;
436 	if (code_size < size * 2)
437 		return -EINVAL;
438 
439 	temp_base = code_base + code_size;
440 	temp_size = WAVE5_TEMPBUF_SIZE;
441 
442 	ret = wave5_vdi_write_memory(vpu_dev, common_vb, 0, fw, size);
443 	if (ret < 0) {
444 		dev_err(vpu_dev->dev, "VPU init, Writing firmware to common buffer, fail: %d\n",
445 			ret);
446 		return ret;
447 	}
448 
449 	vpu_write_reg(vpu_dev, W5_PO_CONF, 0);
450 
451 	/* clear registers */
452 
453 	for (i = W5_CMD_REG_BASE; i < W5_CMD_REG_END; i += 4)
454 		vpu_write_reg(vpu_dev, i, 0x00);
455 
456 	remap_page(vpu_dev, code_base, W5_REMAP_INDEX0);
457 	remap_page(vpu_dev, code_base, W5_REMAP_INDEX1);
458 
459 	vpu_write_reg(vpu_dev, W5_ADDR_CODE_BASE, code_base);
460 	vpu_write_reg(vpu_dev, W5_CODE_SIZE, code_size);
461 	vpu_write_reg(vpu_dev, W5_CODE_PARAM, (WAVE5_UPPER_PROC_AXI_ID << 4) | 0);
462 	vpu_write_reg(vpu_dev, W5_ADDR_TEMP_BASE, temp_base);
463 	vpu_write_reg(vpu_dev, W5_TEMP_SIZE, temp_size);
464 
465 	/* These register must be reset explicitly */
466 	vpu_write_reg(vpu_dev, W5_HW_OPTION, 0);
467 
468 	if (vpu_dev->product_code != WAVE515_CODE) {
469 		wave5_fio_writel(vpu_dev, W5_BACKBONE_PROC_EXT_ADDR, 0);
470 		wave5_fio_writel(vpu_dev, W5_BACKBONE_AXI_PARAM, 0);
471 		vpu_write_reg(vpu_dev, W5_SEC_AXI_PARAM, 0);
472 	}
473 
474 	reg_val = vpu_read_reg(vpu_dev, W5_VPU_RET_VPU_CONFIG0);
475 	if (FIELD_GET(W521_FEATURE_BACKBONE, reg_val)) {
476 		reg_val = ((WAVE5_PROC_AXI_ID << 28) |
477 			   (WAVE5_PRP_AXI_ID << 24) |
478 			   (WAVE5_FBD_Y_AXI_ID << 20) |
479 			   (WAVE5_FBC_Y_AXI_ID << 16) |
480 			   (WAVE5_FBD_C_AXI_ID << 12) |
481 			   (WAVE5_FBC_C_AXI_ID << 8) |
482 			   (WAVE5_PRI_AXI_ID << 4) |
483 			   WAVE5_SEC_AXI_ID);
484 		wave5_fio_writel(vpu_dev, W5_BACKBONE_PROG_AXI_ID, reg_val);
485 	}
486 
487 	if (vpu_dev->product_code == WAVE515_CODE) {
488 		dma_addr_t task_buf_base;
489 
490 		vpu_write_reg(vpu_dev, W5_CMD_INIT_NUM_TASK_BUF, WAVE515_COMMAND_QUEUE_DEPTH);
491 		vpu_write_reg(vpu_dev, W5_CMD_INIT_TASK_BUF_SIZE, WAVE515_ONE_TASKBUF_SIZE);
492 
493 		for (i = 0; i < WAVE515_COMMAND_QUEUE_DEPTH; i++) {
494 			task_buf_base = temp_base + temp_size +
495 					(i * WAVE515_ONE_TASKBUF_SIZE);
496 			vpu_write_reg(vpu_dev,
497 				      W5_CMD_INIT_ADDR_TASK_BUF0 + (i * 4),
498 				      task_buf_base);
499 		}
500 
501 		vpu_write_reg(vpu_dev, W515_CMD_ADDR_SEC_AXI, vpu_dev->sram_buf.daddr);
502 		vpu_write_reg(vpu_dev, W515_CMD_SEC_AXI_SIZE, vpu_dev->sram_buf.size);
503 	}
504 
505 	vpu_write_reg(vpu_dev, W5_VPU_BUSY_STATUS, 1);
506 	vpu_write_reg(vpu_dev, W5_COMMAND, W5_INIT_VPU);
507 	vpu_write_reg(vpu_dev, W5_VPU_REMAP_CORE_START, 1);
508 	ret = wave5_wait_vpu_busy(vpu_dev, W5_VPU_BUSY_STATUS);
509 	if (ret) {
510 		dev_err(vpu_dev->dev, "VPU init(W5_VPU_REMAP_CORE_START) timeout\n");
511 		return ret;
512 	}
513 
514 	ret = wave5_vpu_firmware_command_queue_error_check(vpu_dev, &reason_code);
515 	if (ret)
516 		return ret;
517 
518 	return setup_wave5_properties(dev);
519 }
520 
wave5_vpu_build_up_dec_param(struct vpu_instance * inst,struct dec_open_param * param)521 int wave5_vpu_build_up_dec_param(struct vpu_instance *inst,
522 				 struct dec_open_param *param)
523 {
524 	int ret;
525 	struct dec_info *p_dec_info = &inst->codec_info->dec_info;
526 	struct vpu_device *vpu_dev = inst->dev;
527 
528 	p_dec_info->cycle_per_tick = 256;
529 	if (vpu_dev->sram_buf.size) {
530 		p_dec_info->sec_axi_info.use_bit_enable = 1;
531 		p_dec_info->sec_axi_info.use_ip_enable = 1;
532 		p_dec_info->sec_axi_info.use_lf_row_enable = 1;
533 	}
534 	switch (inst->std) {
535 	case W_HEVC_DEC:
536 		p_dec_info->seq_change_mask = SEQ_CHANGE_ENABLE_ALL_HEVC;
537 		break;
538 	case W_AVC_DEC:
539 		p_dec_info->seq_change_mask = SEQ_CHANGE_ENABLE_ALL_AVC;
540 		break;
541 	default:
542 		return -EINVAL;
543 	}
544 
545 	if (vpu_dev->product == PRODUCT_ID_515)
546 		p_dec_info->vb_work.size = WAVE515DEC_WORKBUF_SIZE;
547 	else
548 		p_dec_info->vb_work.size = WAVE521DEC_WORKBUF_SIZE;
549 
550 	ret = wave5_vdi_allocate_dma_memory(inst->dev, &p_dec_info->vb_work);
551 	if (ret)
552 		return ret;
553 
554 	if (inst->dev->product_code != WAVE515_CODE)
555 		vpu_write_reg(inst->dev, W5_CMD_DEC_VCORE_INFO, 1);
556 
557 	wave5_vdi_clear_memory(inst->dev, &p_dec_info->vb_work);
558 
559 	vpu_write_reg(inst->dev, W5_ADDR_WORK_BASE, p_dec_info->vb_work.daddr);
560 	vpu_write_reg(inst->dev, W5_WORK_SIZE, p_dec_info->vb_work.size);
561 
562 	if (inst->dev->product_code != WAVE515_CODE) {
563 		vpu_write_reg(inst->dev, W5_CMD_ADDR_SEC_AXI, vpu_dev->sram_buf.daddr);
564 		vpu_write_reg(inst->dev, W5_CMD_SEC_AXI_SIZE, vpu_dev->sram_buf.size);
565 	}
566 
567 	vpu_write_reg(inst->dev, W5_CMD_DEC_BS_START_ADDR, p_dec_info->stream_buf_start_addr);
568 	vpu_write_reg(inst->dev, W5_CMD_DEC_BS_SIZE, p_dec_info->stream_buf_size);
569 
570 	/* NOTE: SDMA reads MSB first */
571 	vpu_write_reg(inst->dev, W5_CMD_BS_PARAM, BITSTREAM_ENDIANNESS_BIG_ENDIAN);
572 
573 	if (inst->dev->product_code != WAVE515_CODE) {
574 		/* This register must be reset explicitly */
575 		vpu_write_reg(inst->dev, W5_CMD_EXT_ADDR, 0);
576 		vpu_write_reg(inst->dev, W5_CMD_NUM_CQ_DEPTH_M1,
577 			      WAVE521_COMMAND_QUEUE_DEPTH - 1);
578 	}
579 
580 	ret = send_firmware_command(inst, W5_CREATE_INSTANCE, true, NULL, NULL);
581 	if (ret) {
582 		wave5_vdi_free_dma_memory(vpu_dev, &p_dec_info->vb_work);
583 		return ret;
584 	}
585 
586 	p_dec_info->product_code = vpu_read_reg(inst->dev, W5_PRODUCT_NUMBER);
587 
588 	return 0;
589 }
590 
wave5_vpu_hw_flush_instance(struct vpu_instance * inst)591 int wave5_vpu_hw_flush_instance(struct vpu_instance *inst)
592 {
593 	struct dec_info *p_dec_info = &inst->codec_info->dec_info;
594 	u32 instance_queue_count, report_queue_count;
595 	u32 reg_val = 0;
596 	u32 fail_res = 0;
597 	int ret;
598 
599 	ret = send_firmware_command(inst, W5_FLUSH_INSTANCE, true, &reg_val, &fail_res);
600 	if (ret)
601 		return ret;
602 
603 	instance_queue_count = (reg_val >> 16) & 0xff;
604 	report_queue_count = (reg_val & QUEUE_REPORT_MASK);
605 	if (instance_queue_count != 0 || report_queue_count != 0) {
606 		dev_warn(inst->dev->dev,
607 			 "FLUSH_INSTANCE cmd didn't reset the amount of queued commands & reports");
608 	}
609 
610 	/* reset our local copy of the counts */
611 	p_dec_info->instance_queue_count = 0;
612 	p_dec_info->report_queue_count = 0;
613 
614 	return 0;
615 }
616 
get_bitstream_options(struct dec_info * info)617 static u32 get_bitstream_options(struct dec_info *info)
618 {
619 	u32 bs_option = BSOPTION_ENABLE_EXPLICIT_END;
620 
621 	if (info->stream_endflag)
622 		bs_option |= BSOPTION_HIGHLIGHT_STREAM_END;
623 	return bs_option;
624 }
625 
wave5_vpu_dec_init_seq(struct vpu_instance * inst)626 int wave5_vpu_dec_init_seq(struct vpu_instance *inst)
627 {
628 	struct dec_info *p_dec_info = &inst->codec_info->dec_info;
629 	u32 bs_option, cmd_option = INIT_SEQ_NORMAL;
630 	u32 reg_val, fail_res;
631 	int ret;
632 
633 	if (!inst->codec_info)
634 		return -EINVAL;
635 
636 	vpu_write_reg(inst->dev, W5_BS_RD_PTR, p_dec_info->stream_rd_ptr);
637 	vpu_write_reg(inst->dev, W5_BS_WR_PTR, p_dec_info->stream_wr_ptr);
638 
639 	bs_option = get_bitstream_options(p_dec_info);
640 
641 	/* Without RD_PTR_VALID_FLAG Wave515 ignores RD_PTR value */
642 	if (inst->dev->product_code == WAVE515_CODE)
643 		bs_option |= BSOPTION_RD_PTR_VALID_FLAG;
644 
645 	vpu_write_reg(inst->dev, W5_BS_OPTION, bs_option);
646 
647 	vpu_write_reg(inst->dev, W5_COMMAND_OPTION, cmd_option);
648 	vpu_write_reg(inst->dev, W5_CMD_DEC_USER_MASK, p_dec_info->user_data_enable);
649 
650 	ret = send_firmware_command(inst, W5_INIT_SEQ, true, &reg_val, &fail_res);
651 	if (ret)
652 		return ret;
653 
654 	p_dec_info->instance_queue_count = (reg_val >> 16) & 0xff;
655 	p_dec_info->report_queue_count = (reg_val & QUEUE_REPORT_MASK);
656 
657 	dev_dbg(inst->dev->dev, "%s: init seq sent (queue %u : %u)\n", __func__,
658 		p_dec_info->instance_queue_count, p_dec_info->report_queue_count);
659 
660 	return 0;
661 }
662 
wave5_get_dec_seq_result(struct vpu_instance * inst,struct dec_initial_info * info)663 static void wave5_get_dec_seq_result(struct vpu_instance *inst, struct dec_initial_info *info)
664 {
665 	u32 reg_val;
666 	u32 profile_compatibility_flag;
667 	struct dec_info *p_dec_info = &inst->codec_info->dec_info;
668 
669 	p_dec_info->stream_rd_ptr = wave5_dec_get_rd_ptr(inst);
670 	info->rd_ptr = p_dec_info->stream_rd_ptr;
671 
672 	p_dec_info->frame_display_flag = vpu_read_reg(inst->dev, W5_RET_DEC_DISP_IDC);
673 
674 	reg_val = vpu_read_reg(inst->dev, W5_RET_DEC_PIC_SIZE);
675 	info->pic_width = ((reg_val >> 16) & 0xffff);
676 	info->pic_height = (reg_val & 0xffff);
677 	info->min_frame_buffer_count = vpu_read_reg(inst->dev, W5_RET_DEC_NUM_REQUIRED_FB);
678 
679 	reg_val = vpu_read_reg(inst->dev, W5_RET_DEC_CROP_LEFT_RIGHT);
680 	info->pic_crop_rect.left = (reg_val >> 16) & 0xffff;
681 	info->pic_crop_rect.right = reg_val & 0xffff;
682 	reg_val = vpu_read_reg(inst->dev, W5_RET_DEC_CROP_TOP_BOTTOM);
683 	info->pic_crop_rect.top = (reg_val >> 16) & 0xffff;
684 	info->pic_crop_rect.bottom = reg_val & 0xffff;
685 
686 	reg_val = vpu_read_reg(inst->dev, W5_RET_DEC_COLOR_SAMPLE_INFO);
687 	info->luma_bitdepth = reg_val & 0xf;
688 	info->chroma_bitdepth = (reg_val >> 4) & 0xf;
689 
690 	reg_val = vpu_read_reg(inst->dev, W5_RET_DEC_SEQ_PARAM);
691 	profile_compatibility_flag = (reg_val >> 12) & 0xff;
692 	info->profile = (reg_val >> 24) & 0x1f;
693 
694 	if (inst->std == W_HEVC_DEC) {
695 		/* guessing profile */
696 		if (!info->profile) {
697 			if ((profile_compatibility_flag & 0x06) == 0x06)
698 				info->profile = HEVC_PROFILE_MAIN; /* main profile */
699 			else if (profile_compatibility_flag & 0x04)
700 				info->profile = HEVC_PROFILE_MAIN10; /* main10 profile */
701 			else if (profile_compatibility_flag & 0x08)
702 				/* main still picture profile */
703 				info->profile = HEVC_PROFILE_STILLPICTURE;
704 			else
705 				info->profile = HEVC_PROFILE_MAIN; /* for old version HM */
706 		}
707 	} else if (inst->std == W_AVC_DEC) {
708 		info->profile = FIELD_GET(SEQ_PARAM_PROFILE_MASK, reg_val);
709 	}
710 
711 	if (inst->dev->product_code != WAVE515_CODE) {
712 		info->vlc_buf_size = vpu_read_reg(inst->dev, W5_RET_VLC_BUF_SIZE);
713 		info->param_buf_size = vpu_read_reg(inst->dev, W5_RET_PARAM_BUF_SIZE);
714 		p_dec_info->vlc_buf_size = info->vlc_buf_size;
715 		p_dec_info->param_buf_size = info->param_buf_size;
716 	}
717 }
718 
wave5_vpu_dec_get_seq_info(struct vpu_instance * inst,struct dec_initial_info * info)719 int wave5_vpu_dec_get_seq_info(struct vpu_instance *inst, struct dec_initial_info *info)
720 {
721 	int ret;
722 	u32 reg_val;
723 	struct dec_info *p_dec_info = &inst->codec_info->dec_info;
724 
725 	vpu_write_reg(inst->dev, W5_CMD_DEC_ADDR_REPORT_BASE, p_dec_info->user_data_buf_addr);
726 	vpu_write_reg(inst->dev, W5_CMD_DEC_REPORT_SIZE, p_dec_info->user_data_buf_size);
727 	vpu_write_reg(inst->dev, W5_CMD_DEC_REPORT_PARAM, REPORT_PARAM_ENDIANNESS_BIG_ENDIAN);
728 
729 	/* send QUERY cmd */
730 	ret = wave5_send_query(inst->dev, inst, GET_RESULT);
731 	if (ret)
732 		return ret;
733 
734 	reg_val = vpu_read_reg(inst->dev, W5_RET_QUEUE_STATUS);
735 
736 	p_dec_info->instance_queue_count = (reg_val >> 16) & 0xff;
737 	p_dec_info->report_queue_count = (reg_val & QUEUE_REPORT_MASK);
738 
739 	dev_dbg(inst->dev->dev, "%s: init seq complete (queue %u : %u)\n", __func__,
740 		p_dec_info->instance_queue_count, p_dec_info->report_queue_count);
741 
742 	/* this is not a fatal error, set ret to -EIO but don't return immediately */
743 	if (vpu_read_reg(inst->dev, W5_RET_DEC_DECODING_SUCCESS) != 1) {
744 		info->seq_init_err_reason = vpu_read_reg(inst->dev, W5_RET_DEC_ERR_INFO);
745 		ret = -EIO;
746 	}
747 
748 	wave5_get_dec_seq_result(inst, info);
749 
750 	return ret;
751 }
752 
wave5_vpu_dec_register_framebuffer(struct vpu_instance * inst,struct frame_buffer * fb_arr,enum tiled_map_type map_type,unsigned int count)753 int wave5_vpu_dec_register_framebuffer(struct vpu_instance *inst, struct frame_buffer *fb_arr,
754 				       enum tiled_map_type map_type, unsigned int count)
755 {
756 	int ret;
757 	struct dec_info *p_dec_info = &inst->codec_info->dec_info;
758 	struct dec_initial_info *init_info = &p_dec_info->initial_info;
759 	size_t remain, idx, j, i, cnt_8_chunk, size;
760 	u32 start_no, end_no;
761 	u32 reg_val, cbcr_interleave, nv21, pic_size;
762 	u32 addr_y, addr_cb, addr_cr;
763 	u32 mv_col_size, frame_width, frame_height, fbc_y_tbl_size, fbc_c_tbl_size;
764 	struct vpu_buf vb_buf;
765 	bool justified = WTL_RIGHT_JUSTIFIED;
766 	u32 format_no = WTL_PIXEL_8BIT;
767 	u32 color_format = 0;
768 	u32 pixel_order = 1;
769 	u32 bwb_flag = (map_type == LINEAR_FRAME_MAP) ? 1 : 0;
770 
771 	cbcr_interleave = inst->cbcr_interleave;
772 	nv21 = inst->nv21;
773 	mv_col_size = 0;
774 	fbc_y_tbl_size = 0;
775 	fbc_c_tbl_size = 0;
776 
777 	if (map_type >= COMPRESSED_FRAME_MAP) {
778 		cbcr_interleave = 0;
779 		nv21 = 0;
780 
781 		switch (inst->std) {
782 		case W_HEVC_DEC:
783 			mv_col_size = WAVE5_DEC_HEVC_BUF_SIZE(init_info->pic_width,
784 							      init_info->pic_height);
785 			break;
786 		case W_AVC_DEC:
787 			mv_col_size = WAVE5_DEC_AVC_BUF_SIZE(init_info->pic_width,
788 							     init_info->pic_height);
789 			break;
790 		default:
791 			return -EINVAL;
792 		}
793 
794 		if (inst->std == W_HEVC_DEC || inst->std == W_AVC_DEC) {
795 			size = ALIGN(ALIGN(mv_col_size, 16), BUFFER_MARGIN) + BUFFER_MARGIN;
796 			ret = wave5_vdi_allocate_array(inst->dev, p_dec_info->vb_mv, count, size);
797 			if (ret)
798 				goto free_mv_buffers;
799 		}
800 
801 		frame_width = init_info->pic_width;
802 		frame_height = init_info->pic_height;
803 		fbc_y_tbl_size = ALIGN(WAVE5_FBC_LUMA_TABLE_SIZE(frame_width, frame_height), 16);
804 		fbc_c_tbl_size = ALIGN(WAVE5_FBC_CHROMA_TABLE_SIZE(frame_width, frame_height), 16);
805 
806 		size = ALIGN(fbc_y_tbl_size, BUFFER_MARGIN) + BUFFER_MARGIN;
807 		ret = wave5_vdi_allocate_array(inst->dev, p_dec_info->vb_fbc_y_tbl, count, size);
808 		if (ret)
809 			goto free_fbc_y_tbl_buffers;
810 
811 		size = ALIGN(fbc_c_tbl_size, BUFFER_MARGIN) + BUFFER_MARGIN;
812 		ret = wave5_vdi_allocate_array(inst->dev, p_dec_info->vb_fbc_c_tbl, count, size);
813 		if (ret)
814 			goto free_fbc_c_tbl_buffers;
815 
816 		pic_size = (init_info->pic_width << 16) | (init_info->pic_height);
817 
818 		if (inst->dev->product_code != WAVE515_CODE) {
819 			vb_buf.size = (p_dec_info->vlc_buf_size * VLC_BUF_NUM) +
820 				(p_dec_info->param_buf_size * WAVE521_COMMAND_QUEUE_DEPTH);
821 			vb_buf.daddr = 0;
822 
823 			if (vb_buf.size != p_dec_info->vb_task.size) {
824 				wave5_vdi_free_dma_memory(inst->dev,
825 							  &p_dec_info->vb_task);
826 				ret = wave5_vdi_allocate_dma_memory(inst->dev,
827 								    &vb_buf);
828 				if (ret)
829 					goto free_fbc_c_tbl_buffers;
830 
831 				p_dec_info->vb_task = vb_buf;
832 			}
833 
834 			vpu_write_reg(inst->dev, W5_CMD_SET_FB_ADDR_TASK_BUF,
835 				      p_dec_info->vb_task.daddr);
836 			vpu_write_reg(inst->dev, W5_CMD_SET_FB_TASK_BUF_SIZE,
837 				      vb_buf.size);
838 		}
839 	} else {
840 		pic_size = (init_info->pic_width << 16) | (init_info->pic_height);
841 
842 		if (inst->output_format == FORMAT_422)
843 			color_format = 1;
844 	}
845 	vpu_write_reg(inst->dev, W5_PIC_SIZE, pic_size);
846 
847 	reg_val = (bwb_flag << 28) |
848 		  (pixel_order << 23) |
849 		  (justified << 22) |
850 		  (format_no << 20) |
851 		  (color_format << 19) |
852 		  (nv21 << 17) |
853 		  (cbcr_interleave << 16) |
854 		  (fb_arr[0].stride);
855 	vpu_write_reg(inst->dev, W5_COMMON_PIC_INFO, reg_val);
856 
857 	remain = count;
858 	cnt_8_chunk = DIV_ROUND_UP(count, 8);
859 	idx = 0;
860 	for (j = 0; j < cnt_8_chunk; j++) {
861 		reg_val = (j == cnt_8_chunk - 1) << 4 | ((j == 0) << 3);
862 		vpu_write_reg(inst->dev, W5_SFB_OPTION, reg_val);
863 		start_no = j * 8;
864 		end_no = start_no + ((remain >= 8) ? 8 : remain) - 1;
865 
866 		vpu_write_reg(inst->dev, W5_SET_FB_NUM, (start_no << 8) | end_no);
867 
868 		for (i = 0; i < 8 && i < remain; i++) {
869 			addr_y = fb_arr[i + start_no].buf_y;
870 			addr_cb = fb_arr[i + start_no].buf_cb;
871 			addr_cr = fb_arr[i + start_no].buf_cr;
872 			vpu_write_reg(inst->dev, W5_ADDR_LUMA_BASE0 + (i << 4), addr_y);
873 			vpu_write_reg(inst->dev, W5_ADDR_CB_BASE0 + (i << 4), addr_cb);
874 			if (map_type >= COMPRESSED_FRAME_MAP) {
875 				/* luma FBC offset table */
876 				vpu_write_reg(inst->dev, W5_ADDR_FBC_Y_OFFSET0 + (i << 4),
877 					      p_dec_info->vb_fbc_y_tbl[idx].daddr);
878 				/* chroma FBC offset table */
879 				vpu_write_reg(inst->dev, W5_ADDR_FBC_C_OFFSET0 + (i << 4),
880 					      p_dec_info->vb_fbc_c_tbl[idx].daddr);
881 				vpu_write_reg(inst->dev, W5_ADDR_MV_COL0 + (i << 2),
882 					      p_dec_info->vb_mv[idx].daddr);
883 			} else {
884 				vpu_write_reg(inst->dev, W5_ADDR_CR_BASE0 + (i << 4), addr_cr);
885 				vpu_write_reg(inst->dev, W5_ADDR_FBC_C_OFFSET0 + (i << 4), 0);
886 				vpu_write_reg(inst->dev, W5_ADDR_MV_COL0 + (i << 2), 0);
887 			}
888 			idx++;
889 		}
890 		remain -= i;
891 
892 		ret = send_firmware_command(inst, W5_SET_FB, false, NULL, NULL);
893 		if (ret)
894 			goto free_buffers;
895 	}
896 
897 	reg_val = vpu_read_reg(inst->dev, W5_RET_SUCCESS);
898 	if (!reg_val) {
899 		ret = -EIO;
900 		goto free_buffers;
901 	}
902 
903 	return 0;
904 
905 free_buffers:
906 	wave5_vdi_free_dma_memory(inst->dev, &p_dec_info->vb_task);
907 free_fbc_c_tbl_buffers:
908 	for (i = 0; i < count; i++)
909 		wave5_vdi_free_dma_memory(inst->dev, &p_dec_info->vb_fbc_c_tbl[i]);
910 free_fbc_y_tbl_buffers:
911 	for (i = 0; i < count; i++)
912 		wave5_vdi_free_dma_memory(inst->dev, &p_dec_info->vb_fbc_y_tbl[i]);
913 free_mv_buffers:
914 	for (i = 0; i < count; i++)
915 		wave5_vdi_free_dma_memory(inst->dev, &p_dec_info->vb_mv[i]);
916 	return ret;
917 }
918 
wave5_vpu_dec_validate_sec_axi(struct vpu_instance * inst)919 static u32 wave5_vpu_dec_validate_sec_axi(struct vpu_instance *inst)
920 {
921 	u32 bitdepth = inst->codec_info->dec_info.initial_info.luma_bitdepth;
922 	struct dec_info *p_dec_info = &inst->codec_info->dec_info;
923 	u32 bit_size = 0, ip_size = 0, lf_size = 0, ret = 0;
924 	u32 sram_size = inst->dev->sram_size;
925 	u32 width = inst->src_fmt.width;
926 
927 	if (!sram_size)
928 		return 0;
929 
930 	/*
931 	 * TODO: calculate bit_size, ip_size, lf_size from width and bitdepth
932 	 * for Wave521.
933 	 */
934 	if (inst->dev->product_code == WAVE515_CODE) {
935 		bit_size = DIV_ROUND_UP(width, 16) * 5 * 8;
936 		ip_size = ALIGN(width, 16) * 2 * bitdepth / 8;
937 		lf_size = ALIGN(width, 16) * 10 * bitdepth / 8;
938 	}
939 
940 	if (p_dec_info->sec_axi_info.use_bit_enable && sram_size >= bit_size) {
941 		ret |= BIT(0);
942 		sram_size -= bit_size;
943 	}
944 
945 	if (p_dec_info->sec_axi_info.use_ip_enable && sram_size >= ip_size) {
946 		ret |= BIT(9);
947 		sram_size -= ip_size;
948 	}
949 
950 	if (p_dec_info->sec_axi_info.use_lf_row_enable && sram_size >= lf_size)
951 		ret |= BIT(15);
952 
953 	return ret;
954 }
955 
wave5_vpu_decode(struct vpu_instance * inst,u32 * fail_res)956 int wave5_vpu_decode(struct vpu_instance *inst, u32 *fail_res)
957 {
958 	u32 reg_val;
959 	struct dec_info *p_dec_info = &inst->codec_info->dec_info;
960 	int ret;
961 
962 	vpu_write_reg(inst->dev, W5_BS_RD_PTR, p_dec_info->stream_rd_ptr);
963 	vpu_write_reg(inst->dev, W5_BS_WR_PTR, p_dec_info->stream_wr_ptr);
964 
965 	vpu_write_reg(inst->dev, W5_BS_OPTION, get_bitstream_options(p_dec_info));
966 
967 	/* secondary AXI */
968 	reg_val = wave5_vpu_dec_validate_sec_axi(inst);
969 	vpu_write_reg(inst->dev, W5_USE_SEC_AXI, reg_val);
970 
971 	/* set attributes of user buffer */
972 	vpu_write_reg(inst->dev, W5_CMD_DEC_USER_MASK, p_dec_info->user_data_enable);
973 
974 	vpu_write_reg(inst->dev, W5_COMMAND_OPTION, DEC_PIC_NORMAL);
975 	vpu_write_reg(inst->dev, W5_CMD_DEC_TEMPORAL_ID_PLUS1,
976 		      (p_dec_info->target_spatial_id << 9) |
977 		      (p_dec_info->temp_id_select_mode << 8) | p_dec_info->target_temp_id);
978 	vpu_write_reg(inst->dev, W5_CMD_SEQ_CHANGE_ENABLE_FLAG, p_dec_info->seq_change_mask);
979 	/* When reordering is disabled we force the latency of the framebuffers */
980 	vpu_write_reg(inst->dev, W5_CMD_DEC_FORCE_FB_LATENCY_PLUS1, !p_dec_info->reorder_enable);
981 
982 	ret = send_firmware_command(inst, W5_DEC_ENC_PIC, true, &reg_val, fail_res);
983 	if (ret == -ETIMEDOUT)
984 		return ret;
985 
986 	p_dec_info->instance_queue_count = (reg_val >> 16) & 0xff;
987 	p_dec_info->report_queue_count = (reg_val & QUEUE_REPORT_MASK);
988 
989 	dev_dbg(inst->dev->dev, "%s: dec pic sent (queue %u : %u)\n", __func__,
990 		p_dec_info->instance_queue_count, p_dec_info->report_queue_count);
991 
992 	if (ret)
993 		return ret;
994 
995 	return 0;
996 }
997 
wave5_vpu_dec_get_result(struct vpu_instance * inst,struct dec_output_info * result)998 int wave5_vpu_dec_get_result(struct vpu_instance *inst, struct dec_output_info *result)
999 {
1000 	int ret;
1001 	u32 index, nal_unit_type, reg_val, sub_layer_info;
1002 	struct dec_info *p_dec_info = &inst->codec_info->dec_info;
1003 	struct vpu_device *vpu_dev = inst->dev;
1004 
1005 	vpu_write_reg(inst->dev, W5_CMD_DEC_ADDR_REPORT_BASE, p_dec_info->user_data_buf_addr);
1006 	vpu_write_reg(inst->dev, W5_CMD_DEC_REPORT_SIZE, p_dec_info->user_data_buf_size);
1007 	vpu_write_reg(inst->dev, W5_CMD_DEC_REPORT_PARAM, REPORT_PARAM_ENDIANNESS_BIG_ENDIAN);
1008 
1009 	/* send QUERY cmd */
1010 	ret = wave5_send_query(vpu_dev, inst, GET_RESULT);
1011 	if (ret)
1012 		return ret;
1013 
1014 	reg_val = vpu_read_reg(inst->dev, W5_RET_QUEUE_STATUS);
1015 
1016 	p_dec_info->instance_queue_count = (reg_val >> 16) & 0xff;
1017 	p_dec_info->report_queue_count = (reg_val & QUEUE_REPORT_MASK);
1018 
1019 	dev_dbg(inst->dev->dev, "%s: dec pic complete (queue %u : %u)\n", __func__,
1020 		p_dec_info->instance_queue_count, p_dec_info->report_queue_count);
1021 
1022 	reg_val = vpu_read_reg(inst->dev, W5_RET_DEC_PIC_TYPE);
1023 
1024 	nal_unit_type = (reg_val >> 4) & 0x3f;
1025 
1026 	if (inst->std == W_HEVC_DEC) {
1027 		if (reg_val & 0x04)
1028 			result->pic_type = PIC_TYPE_B;
1029 		else if (reg_val & 0x02)
1030 			result->pic_type = PIC_TYPE_P;
1031 		else if (reg_val & 0x01)
1032 			result->pic_type = PIC_TYPE_I;
1033 		else
1034 			result->pic_type = PIC_TYPE_MAX;
1035 		if ((nal_unit_type == 19 || nal_unit_type == 20) && result->pic_type == PIC_TYPE_I)
1036 			/* IDR_W_RADL, IDR_N_LP */
1037 			result->pic_type = PIC_TYPE_IDR;
1038 	} else if (inst->std == W_AVC_DEC) {
1039 		if (reg_val & 0x04)
1040 			result->pic_type = PIC_TYPE_B;
1041 		else if (reg_val & 0x02)
1042 			result->pic_type = PIC_TYPE_P;
1043 		else if (reg_val & 0x01)
1044 			result->pic_type = PIC_TYPE_I;
1045 		else
1046 			result->pic_type = PIC_TYPE_MAX;
1047 		if (nal_unit_type == 5 && result->pic_type == PIC_TYPE_I)
1048 			result->pic_type = PIC_TYPE_IDR;
1049 	}
1050 	index = vpu_read_reg(inst->dev, W5_RET_DEC_DISPLAY_INDEX);
1051 	result->index_frame_display = index;
1052 	index = vpu_read_reg(inst->dev, W5_RET_DEC_DECODED_INDEX);
1053 	result->index_frame_decoded = index;
1054 	result->index_frame_decoded_for_tiled = index;
1055 
1056 	sub_layer_info = vpu_read_reg(inst->dev, W5_RET_DEC_SUB_LAYER_INFO);
1057 	result->temporal_id = sub_layer_info & 0x7;
1058 
1059 	if (inst->std == W_HEVC_DEC || inst->std == W_AVC_DEC) {
1060 		result->decoded_poc = -1;
1061 		if (result->index_frame_decoded >= 0 ||
1062 		    result->index_frame_decoded == DECODED_IDX_FLAG_SKIP)
1063 			result->decoded_poc = vpu_read_reg(inst->dev, W5_RET_DEC_PIC_POC);
1064 	}
1065 
1066 	result->sequence_changed = vpu_read_reg(inst->dev, W5_RET_DEC_NOTIFICATION);
1067 	reg_val = vpu_read_reg(inst->dev, W5_RET_DEC_PIC_SIZE);
1068 	result->dec_pic_width = reg_val >> 16;
1069 	result->dec_pic_height = reg_val & 0xffff;
1070 
1071 	if (result->sequence_changed) {
1072 		memcpy((void *)&p_dec_info->new_seq_info, (void *)&p_dec_info->initial_info,
1073 		       sizeof(struct dec_initial_info));
1074 		wave5_get_dec_seq_result(inst, &p_dec_info->new_seq_info);
1075 	}
1076 
1077 	result->dec_host_cmd_tick = vpu_read_reg(inst->dev, W5_RET_DEC_HOST_CMD_TICK);
1078 	result->dec_decode_end_tick = vpu_read_reg(inst->dev, W5_RET_DEC_DECODING_ENC_TICK);
1079 
1080 	if (!p_dec_info->first_cycle_check) {
1081 		result->frame_cycle =
1082 			(result->dec_decode_end_tick - result->dec_host_cmd_tick) *
1083 			p_dec_info->cycle_per_tick;
1084 		vpu_dev->last_performance_cycles = result->dec_decode_end_tick;
1085 		p_dec_info->first_cycle_check = true;
1086 	} else if (result->index_frame_decoded_for_tiled != -1) {
1087 		result->frame_cycle =
1088 			(result->dec_decode_end_tick - vpu_dev->last_performance_cycles) *
1089 			p_dec_info->cycle_per_tick;
1090 		vpu_dev->last_performance_cycles = result->dec_decode_end_tick;
1091 		if (vpu_dev->last_performance_cycles < result->dec_host_cmd_tick)
1092 			result->frame_cycle =
1093 				(result->dec_decode_end_tick - result->dec_host_cmd_tick) *
1094 				p_dec_info->cycle_per_tick;
1095 	}
1096 
1097 	/* no remaining command. reset frame cycle. */
1098 	if (p_dec_info->instance_queue_count == 0 && p_dec_info->report_queue_count == 0)
1099 		p_dec_info->first_cycle_check = false;
1100 
1101 	return 0;
1102 }
1103 
wave5_vpu_re_init(struct device * dev,u8 * fw,size_t size)1104 int wave5_vpu_re_init(struct device *dev, u8 *fw, size_t size)
1105 {
1106 	struct vpu_buf *common_vb;
1107 	dma_addr_t code_base, temp_base;
1108 	dma_addr_t old_code_base, temp_size;
1109 	u32 code_size, reason_code;
1110 	u32 reg_val;
1111 	struct vpu_device *vpu_dev = dev_get_drvdata(dev);
1112 
1113 	common_vb = &vpu_dev->common_mem;
1114 
1115 	code_base = common_vb->daddr;
1116 
1117 	if (vpu_dev->product_code == WAVE515_CODE)
1118 		code_size = WAVE515_MAX_CODE_BUF_SIZE;
1119 	else
1120 		code_size = WAVE521_MAX_CODE_BUF_SIZE;
1121 
1122 	/* ALIGN TO 4KB */
1123 	code_size &= ~0xfff;
1124 	if (code_size < size * 2)
1125 		return -EINVAL;
1126 
1127 	temp_base = code_base + code_size;
1128 	temp_size = WAVE5_TEMPBUF_SIZE;
1129 
1130 	old_code_base = vpu_read_reg(vpu_dev, W5_VPU_REMAP_PADDR);
1131 
1132 	if (old_code_base != code_base + W5_REMAP_INDEX1 * W5_REMAP_MAX_SIZE) {
1133 		int ret;
1134 
1135 		ret = wave5_vdi_write_memory(vpu_dev, common_vb, 0, fw, size);
1136 		if (ret < 0) {
1137 			dev_err(vpu_dev->dev,
1138 				"VPU init, Writing firmware to common buffer, fail: %d\n", ret);
1139 			return ret;
1140 		}
1141 
1142 		vpu_write_reg(vpu_dev, W5_PO_CONF, 0);
1143 
1144 		ret = wave5_vpu_reset(dev, SW_RESET_ON_BOOT);
1145 		if (ret < 0) {
1146 			dev_err(vpu_dev->dev, "VPU init, Resetting the VPU, fail: %d\n", ret);
1147 			return ret;
1148 		}
1149 
1150 		remap_page(vpu_dev, code_base, W5_REMAP_INDEX0);
1151 		remap_page(vpu_dev, code_base, W5_REMAP_INDEX1);
1152 
1153 		vpu_write_reg(vpu_dev, W5_ADDR_CODE_BASE, code_base);
1154 		vpu_write_reg(vpu_dev, W5_CODE_SIZE, code_size);
1155 		vpu_write_reg(vpu_dev, W5_CODE_PARAM, (WAVE5_UPPER_PROC_AXI_ID << 4) | 0);
1156 		vpu_write_reg(vpu_dev, W5_ADDR_TEMP_BASE, temp_base);
1157 		vpu_write_reg(vpu_dev, W5_TEMP_SIZE, temp_size);
1158 
1159 		/* These register must be reset explicitly */
1160 		vpu_write_reg(vpu_dev, W5_HW_OPTION, 0);
1161 
1162 		if (vpu_dev->product_code != WAVE515_CODE) {
1163 			wave5_fio_writel(vpu_dev, W5_BACKBONE_PROC_EXT_ADDR, 0);
1164 			wave5_fio_writel(vpu_dev, W5_BACKBONE_AXI_PARAM, 0);
1165 			vpu_write_reg(vpu_dev, W5_SEC_AXI_PARAM, 0);
1166 		}
1167 
1168 		reg_val = vpu_read_reg(vpu_dev, W5_VPU_RET_VPU_CONFIG0);
1169 		if (FIELD_GET(W521_FEATURE_BACKBONE, reg_val)) {
1170 			reg_val = ((WAVE5_PROC_AXI_ID << 28) |
1171 					(WAVE5_PRP_AXI_ID << 24) |
1172 					(WAVE5_FBD_Y_AXI_ID << 20) |
1173 					(WAVE5_FBC_Y_AXI_ID << 16) |
1174 					(WAVE5_FBD_C_AXI_ID << 12) |
1175 					(WAVE5_FBC_C_AXI_ID << 8) |
1176 					(WAVE5_PRI_AXI_ID << 4) |
1177 					WAVE5_SEC_AXI_ID);
1178 			wave5_fio_writel(vpu_dev, W5_BACKBONE_PROG_AXI_ID, reg_val);
1179 		}
1180 
1181 		if (vpu_dev->product_code == WAVE515_CODE) {
1182 			dma_addr_t task_buf_base;
1183 			u32 i;
1184 
1185 			vpu_write_reg(vpu_dev, W5_CMD_INIT_NUM_TASK_BUF,
1186 				      WAVE515_COMMAND_QUEUE_DEPTH);
1187 			vpu_write_reg(vpu_dev, W5_CMD_INIT_TASK_BUF_SIZE,
1188 				      WAVE515_ONE_TASKBUF_SIZE);
1189 
1190 			for (i = 0; i < WAVE515_COMMAND_QUEUE_DEPTH; i++) {
1191 				task_buf_base = temp_base + temp_size +
1192 						(i * WAVE515_ONE_TASKBUF_SIZE);
1193 				vpu_write_reg(vpu_dev,
1194 					      W5_CMD_INIT_ADDR_TASK_BUF0 + (i * 4),
1195 					      task_buf_base);
1196 			}
1197 
1198 			vpu_write_reg(vpu_dev, W515_CMD_ADDR_SEC_AXI,
1199 				      vpu_dev->sram_buf.daddr);
1200 			vpu_write_reg(vpu_dev, W515_CMD_SEC_AXI_SIZE,
1201 				      vpu_dev->sram_buf.size);
1202 		}
1203 
1204 		vpu_write_reg(vpu_dev, W5_VPU_BUSY_STATUS, 1);
1205 		vpu_write_reg(vpu_dev, W5_COMMAND, W5_INIT_VPU);
1206 		vpu_write_reg(vpu_dev, W5_VPU_REMAP_CORE_START, 1);
1207 
1208 		ret = wave5_wait_vpu_busy(vpu_dev, W5_VPU_BUSY_STATUS);
1209 		if (ret) {
1210 			dev_err(vpu_dev->dev, "VPU reinit(W5_VPU_REMAP_CORE_START) timeout\n");
1211 			return ret;
1212 		}
1213 
1214 		ret = wave5_vpu_firmware_command_queue_error_check(vpu_dev, &reason_code);
1215 		if (ret)
1216 			return ret;
1217 	}
1218 
1219 	return setup_wave5_properties(dev);
1220 }
1221 
wave5_vpu_sleep_wake(struct device * dev,bool i_sleep_wake,const uint16_t * code,size_t size)1222 static int wave5_vpu_sleep_wake(struct device *dev, bool i_sleep_wake, const uint16_t *code,
1223 				size_t size)
1224 {
1225 	u32 reg_val;
1226 	struct vpu_buf *common_vb;
1227 	dma_addr_t code_base, temp_base;
1228 	u32 code_size, temp_size, reason_code;
1229 	struct vpu_device *vpu_dev = dev_get_drvdata(dev);
1230 	int ret;
1231 
1232 	if (i_sleep_wake) {
1233 		ret = wave5_wait_vpu_busy(vpu_dev, W5_VPU_BUSY_STATUS);
1234 		if (ret)
1235 			return ret;
1236 
1237 		/*
1238 		 * Declare who has ownership for the host interface access
1239 		 * 1 = VPU
1240 		 * 0 = Host processor
1241 		 */
1242 		vpu_write_reg(vpu_dev, W5_VPU_BUSY_STATUS, 1);
1243 		vpu_write_reg(vpu_dev, W5_COMMAND, W5_SLEEP_VPU);
1244 		/* Send an interrupt named HOST to the VPU */
1245 		vpu_write_reg(vpu_dev, W5_VPU_HOST_INT_REQ, 1);
1246 
1247 		ret = wave5_wait_vpu_busy(vpu_dev, W5_VPU_BUSY_STATUS);
1248 		if (ret)
1249 			return ret;
1250 
1251 		ret = wave5_vpu_firmware_command_queue_error_check(vpu_dev, &reason_code);
1252 		if (ret)
1253 			return ret;
1254 	} else { /* restore */
1255 		common_vb = &vpu_dev->common_mem;
1256 
1257 		code_base = common_vb->daddr;
1258 
1259 		if (vpu_dev->product_code == WAVE515_CODE)
1260 			code_size = WAVE515_MAX_CODE_BUF_SIZE;
1261 		else
1262 			code_size = WAVE521_MAX_CODE_BUF_SIZE;
1263 
1264 		/* ALIGN TO 4KB */
1265 		code_size &= ~0xfff;
1266 		if (code_size < size * 2) {
1267 			dev_err(dev, "size too small\n");
1268 			return -EINVAL;
1269 		}
1270 
1271 		temp_base = code_base + code_size;
1272 		temp_size = WAVE5_TEMPBUF_SIZE;
1273 
1274 		/* Power on without DEBUG mode */
1275 		vpu_write_reg(vpu_dev, W5_PO_CONF, 0);
1276 
1277 		remap_page(vpu_dev, code_base, W5_REMAP_INDEX0);
1278 		remap_page(vpu_dev, code_base, W5_REMAP_INDEX1);
1279 
1280 		vpu_write_reg(vpu_dev, W5_ADDR_CODE_BASE, code_base);
1281 		vpu_write_reg(vpu_dev, W5_CODE_SIZE, code_size);
1282 		vpu_write_reg(vpu_dev, W5_CODE_PARAM, (WAVE5_UPPER_PROC_AXI_ID << 4) | 0);
1283 
1284 		/* These register must be reset explicitly */
1285 		vpu_write_reg(vpu_dev, W5_HW_OPTION, 0);
1286 
1287 		if (vpu_dev->product_code != WAVE515_CODE) {
1288 			wave5_fio_writel(vpu_dev, W5_BACKBONE_PROC_EXT_ADDR, 0);
1289 			wave5_fio_writel(vpu_dev, W5_BACKBONE_AXI_PARAM, 0);
1290 			vpu_write_reg(vpu_dev, W5_SEC_AXI_PARAM, 0);
1291 		}
1292 
1293 		setup_wave5_interrupts(vpu_dev);
1294 
1295 		reg_val = vpu_read_reg(vpu_dev, W5_VPU_RET_VPU_CONFIG0);
1296 		if (FIELD_GET(W521_FEATURE_BACKBONE, reg_val)) {
1297 			reg_val = ((WAVE5_PROC_AXI_ID << 28) |
1298 					(WAVE5_PRP_AXI_ID << 24) |
1299 					(WAVE5_FBD_Y_AXI_ID << 20) |
1300 					(WAVE5_FBC_Y_AXI_ID << 16) |
1301 					(WAVE5_FBD_C_AXI_ID << 12) |
1302 					(WAVE5_FBC_C_AXI_ID << 8) |
1303 					(WAVE5_PRI_AXI_ID << 4) |
1304 					WAVE5_SEC_AXI_ID);
1305 			wave5_fio_writel(vpu_dev, W5_BACKBONE_PROG_AXI_ID, reg_val);
1306 		}
1307 
1308 		if (vpu_dev->product_code == WAVE515_CODE) {
1309 			dma_addr_t task_buf_base;
1310 			u32 i;
1311 
1312 			vpu_write_reg(vpu_dev, W5_CMD_INIT_NUM_TASK_BUF,
1313 				      WAVE515_COMMAND_QUEUE_DEPTH);
1314 			vpu_write_reg(vpu_dev, W5_CMD_INIT_TASK_BUF_SIZE,
1315 				      WAVE515_ONE_TASKBUF_SIZE);
1316 
1317 			for (i = 0; i < WAVE515_COMMAND_QUEUE_DEPTH; i++) {
1318 				task_buf_base = temp_base + temp_size +
1319 						(i * WAVE515_ONE_TASKBUF_SIZE);
1320 				vpu_write_reg(vpu_dev,
1321 					      W5_CMD_INIT_ADDR_TASK_BUF0 + (i * 4),
1322 					      task_buf_base);
1323 			}
1324 
1325 			vpu_write_reg(vpu_dev, W515_CMD_ADDR_SEC_AXI,
1326 				      vpu_dev->sram_buf.daddr);
1327 			vpu_write_reg(vpu_dev, W515_CMD_SEC_AXI_SIZE,
1328 				      vpu_dev->sram_buf.size);
1329 		}
1330 
1331 		vpu_write_reg(vpu_dev, W5_VPU_BUSY_STATUS, 1);
1332 		vpu_write_reg(vpu_dev, W5_COMMAND, W5_WAKEUP_VPU);
1333 		/* Start VPU after settings */
1334 		vpu_write_reg(vpu_dev, W5_VPU_REMAP_CORE_START, 1);
1335 
1336 		ret = wave5_wait_vpu_busy(vpu_dev, W5_VPU_BUSY_STATUS);
1337 		if (ret) {
1338 			dev_err(vpu_dev->dev, "VPU wakeup(W5_VPU_REMAP_CORE_START) timeout\n");
1339 			return ret;
1340 		}
1341 
1342 		return wave5_vpu_firmware_command_queue_error_check(vpu_dev, &reason_code);
1343 	}
1344 
1345 	return 0;
1346 }
1347 
wave5_vpu_reset(struct device * dev,enum sw_reset_mode reset_mode)1348 int wave5_vpu_reset(struct device *dev, enum sw_reset_mode reset_mode)
1349 {
1350 	u32 val = 0;
1351 	int ret = 0;
1352 	struct vpu_device *vpu_dev = dev_get_drvdata(dev);
1353 	struct vpu_attr *p_attr = &vpu_dev->attr;
1354 	/* VPU doesn't send response. force to set BUSY flag to 0. */
1355 	vpu_write_reg(vpu_dev, W5_VPU_BUSY_STATUS, 0);
1356 
1357 	if (reset_mode == SW_RESET_SAFETY) {
1358 		ret = wave5_vpu_sleep_wake(dev, true, NULL, 0);
1359 		if (ret)
1360 			return ret;
1361 	}
1362 
1363 	val = vpu_read_reg(vpu_dev, W5_VPU_RET_VPU_CONFIG0);
1364 	if ((val >> 16) & 0x1)
1365 		p_attr->support_backbone = true;
1366 	if ((val >> 22) & 0x1)
1367 		p_attr->support_vcore_backbone = true;
1368 	if ((val >> 28) & 0x1)
1369 		p_attr->support_vcpu_backbone = true;
1370 
1371 	/* waiting for completion of bus transaction */
1372 	if (p_attr->support_backbone) {
1373 		dev_dbg(dev, "%s: backbone supported\n", __func__);
1374 
1375 		if (p_attr->support_vcore_backbone) {
1376 			if (p_attr->support_vcpu_backbone) {
1377 				/* step1 : disable request */
1378 				wave5_fio_writel(vpu_dev, W5_BACKBONE_BUS_CTRL_VCPU, 0xFF);
1379 
1380 				/* step2 : waiting for completion of bus transaction */
1381 				ret = wave5_wait_vcpu_bus_busy(vpu_dev,
1382 							       W5_BACKBONE_BUS_STATUS_VCPU);
1383 				if (ret) {
1384 					wave5_fio_writel(vpu_dev, W5_BACKBONE_BUS_CTRL_VCPU, 0x00);
1385 					return ret;
1386 				}
1387 			}
1388 			/* step1 : disable request */
1389 			wave5_fio_writel(vpu_dev, W5_BACKBONE_BUS_CTRL_VCORE0, 0x7);
1390 
1391 			/* step2 : waiting for completion of bus transaction */
1392 			if (wave5_wait_bus_busy(vpu_dev, W5_BACKBONE_BUS_STATUS_VCORE0)) {
1393 				wave5_fio_writel(vpu_dev, W5_BACKBONE_BUS_CTRL_VCORE0, 0x00);
1394 				return -EBUSY;
1395 			}
1396 		} else {
1397 			/* step1 : disable request */
1398 			wave5_fio_writel(vpu_dev, W5_COMBINED_BACKBONE_BUS_CTRL, 0x7);
1399 
1400 			/* step2 : waiting for completion of bus transaction */
1401 			if (wave5_wait_bus_busy(vpu_dev, W5_COMBINED_BACKBONE_BUS_STATUS)) {
1402 				wave5_fio_writel(vpu_dev, W5_COMBINED_BACKBONE_BUS_CTRL, 0x00);
1403 				return -EBUSY;
1404 			}
1405 		}
1406 	} else {
1407 		dev_dbg(dev, "%s: backbone NOT supported\n", __func__);
1408 		/* step1 : disable request */
1409 		wave5_fio_writel(vpu_dev, W5_GDI_BUS_CTRL, 0x100);
1410 
1411 		/* step2 : waiting for completion of bus transaction */
1412 		ret = wave5_wait_bus_busy(vpu_dev, W5_GDI_BUS_STATUS);
1413 		if (ret) {
1414 			wave5_fio_writel(vpu_dev, W5_GDI_BUS_CTRL, 0x00);
1415 			return ret;
1416 		}
1417 	}
1418 
1419 	switch (reset_mode) {
1420 	case SW_RESET_ON_BOOT:
1421 	case SW_RESET_FORCE:
1422 	case SW_RESET_SAFETY:
1423 		val = W5_RST_BLOCK_ALL;
1424 		break;
1425 	default:
1426 		return -EINVAL;
1427 	}
1428 
1429 	if (val) {
1430 		vpu_write_reg(vpu_dev, W5_VPU_RESET_REQ, val);
1431 
1432 		ret = wave5_wait_vpu_busy(vpu_dev, W5_VPU_RESET_STATUS);
1433 		if (ret) {
1434 			vpu_write_reg(vpu_dev, W5_VPU_RESET_REQ, 0);
1435 			return ret;
1436 		}
1437 		vpu_write_reg(vpu_dev, W5_VPU_RESET_REQ, 0);
1438 	}
1439 	/* step3 : must clear GDI_BUS_CTRL after done SW_RESET */
1440 	if (p_attr->support_backbone) {
1441 		if (p_attr->support_vcore_backbone) {
1442 			if (p_attr->support_vcpu_backbone)
1443 				wave5_fio_writel(vpu_dev, W5_BACKBONE_BUS_CTRL_VCPU, 0x00);
1444 			wave5_fio_writel(vpu_dev, W5_BACKBONE_BUS_CTRL_VCORE0, 0x00);
1445 		} else {
1446 			wave5_fio_writel(vpu_dev, W5_COMBINED_BACKBONE_BUS_CTRL, 0x00);
1447 		}
1448 	} else {
1449 		wave5_fio_writel(vpu_dev, W5_GDI_BUS_CTRL, 0x00);
1450 	}
1451 	if (reset_mode == SW_RESET_SAFETY || reset_mode == SW_RESET_FORCE)
1452 		ret = wave5_vpu_sleep_wake(dev, false, NULL, 0);
1453 
1454 	return ret;
1455 }
1456 
wave5_vpu_dec_finish_seq(struct vpu_instance * inst,u32 * fail_res)1457 int wave5_vpu_dec_finish_seq(struct vpu_instance *inst, u32 *fail_res)
1458 {
1459 	return send_firmware_command(inst, W5_DESTROY_INSTANCE, true, NULL, fail_res);
1460 }
1461 
wave5_vpu_dec_set_bitstream_flag(struct vpu_instance * inst,bool eos)1462 int wave5_vpu_dec_set_bitstream_flag(struct vpu_instance *inst, bool eos)
1463 {
1464 	struct dec_info *p_dec_info = &inst->codec_info->dec_info;
1465 
1466 	p_dec_info->stream_endflag = eos ? 1 : 0;
1467 	vpu_write_reg(inst->dev, W5_BS_OPTION, get_bitstream_options(p_dec_info));
1468 	vpu_write_reg(inst->dev, W5_BS_WR_PTR, p_dec_info->stream_wr_ptr);
1469 
1470 	return send_firmware_command(inst, W5_UPDATE_BS, true, NULL, NULL);
1471 }
1472 
wave5_dec_clr_disp_flag(struct vpu_instance * inst,unsigned int index)1473 int wave5_dec_clr_disp_flag(struct vpu_instance *inst, unsigned int index)
1474 {
1475 	struct dec_info *p_dec_info = &inst->codec_info->dec_info;
1476 	int ret;
1477 
1478 	vpu_write_reg(inst->dev, W5_CMD_DEC_CLR_DISP_IDC, BIT(index));
1479 	vpu_write_reg(inst->dev, W5_CMD_DEC_SET_DISP_IDC, 0);
1480 
1481 	ret = wave5_send_query(inst->dev, inst, UPDATE_DISP_FLAG);
1482 	if (ret)
1483 		return ret;
1484 
1485 	p_dec_info->frame_display_flag = vpu_read_reg(inst->dev, W5_RET_DEC_DISP_IDC);
1486 
1487 	return 0;
1488 }
1489 
wave5_dec_set_disp_flag(struct vpu_instance * inst,unsigned int index)1490 int wave5_dec_set_disp_flag(struct vpu_instance *inst, unsigned int index)
1491 {
1492 	int ret;
1493 
1494 	vpu_write_reg(inst->dev, W5_CMD_DEC_CLR_DISP_IDC, 0);
1495 	vpu_write_reg(inst->dev, W5_CMD_DEC_SET_DISP_IDC, BIT(index));
1496 
1497 	ret = wave5_send_query(inst->dev, inst, UPDATE_DISP_FLAG);
1498 	if (ret)
1499 		return ret;
1500 
1501 	return 0;
1502 }
1503 
wave5_vpu_clear_interrupt(struct vpu_instance * inst,u32 flags)1504 int wave5_vpu_clear_interrupt(struct vpu_instance *inst, u32 flags)
1505 {
1506 	u32 interrupt_reason;
1507 
1508 	interrupt_reason = vpu_read_reg(inst->dev, W5_VPU_VINT_REASON_USR);
1509 	interrupt_reason &= ~flags;
1510 	vpu_write_reg(inst->dev, W5_VPU_VINT_REASON_USR, interrupt_reason);
1511 
1512 	return 0;
1513 }
1514 
wave5_dec_get_rd_ptr(struct vpu_instance * inst)1515 dma_addr_t wave5_dec_get_rd_ptr(struct vpu_instance *inst)
1516 {
1517 	int ret;
1518 
1519 	ret = wave5_send_query(inst->dev, inst, GET_BS_RD_PTR);
1520 	if (ret)
1521 		return inst->codec_info->dec_info.stream_rd_ptr;
1522 
1523 	return vpu_read_reg(inst->dev, W5_RET_QUERY_DEC_BS_RD_PTR);
1524 }
1525 
wave5_dec_set_rd_ptr(struct vpu_instance * inst,dma_addr_t addr)1526 int wave5_dec_set_rd_ptr(struct vpu_instance *inst, dma_addr_t addr)
1527 {
1528 	int ret;
1529 
1530 	vpu_write_reg(inst->dev, W5_RET_QUERY_DEC_SET_BS_RD_PTR, addr);
1531 
1532 	ret = wave5_send_query(inst->dev, inst, SET_BS_RD_PTR);
1533 
1534 	return ret;
1535 }
1536 
1537 /************************************************************************/
1538 /* ENCODER functions */
1539 /************************************************************************/
1540 
wave5_vpu_build_up_enc_param(struct device * dev,struct vpu_instance * inst,struct enc_open_param * open_param)1541 int wave5_vpu_build_up_enc_param(struct device *dev, struct vpu_instance *inst,
1542 				 struct enc_open_param *open_param)
1543 {
1544 	int ret;
1545 	struct enc_info *p_enc_info = &inst->codec_info->enc_info;
1546 	u32 reg_val;
1547 	struct vpu_device *vpu_dev = dev_get_drvdata(dev);
1548 	dma_addr_t buffer_addr;
1549 	size_t buffer_size;
1550 
1551 	p_enc_info->cycle_per_tick = 256;
1552 	if (vpu_dev->sram_buf.size) {
1553 		p_enc_info->sec_axi_info.use_enc_rdo_enable = 1;
1554 		p_enc_info->sec_axi_info.use_enc_lf_enable = 1;
1555 	}
1556 
1557 	p_enc_info->vb_work.size = WAVE521ENC_WORKBUF_SIZE;
1558 	ret = wave5_vdi_allocate_dma_memory(vpu_dev, &p_enc_info->vb_work);
1559 	if (ret) {
1560 		memset(&p_enc_info->vb_work, 0, sizeof(p_enc_info->vb_work));
1561 		return ret;
1562 	}
1563 
1564 	wave5_vdi_clear_memory(vpu_dev, &p_enc_info->vb_work);
1565 
1566 	vpu_write_reg(inst->dev, W5_ADDR_WORK_BASE, p_enc_info->vb_work.daddr);
1567 	vpu_write_reg(inst->dev, W5_WORK_SIZE, p_enc_info->vb_work.size);
1568 
1569 	vpu_write_reg(inst->dev, W5_CMD_ADDR_SEC_AXI, vpu_dev->sram_buf.daddr);
1570 	vpu_write_reg(inst->dev, W5_CMD_SEC_AXI_SIZE, vpu_dev->sram_buf.size);
1571 
1572 	reg_val = (open_param->line_buf_int_en << 6) | BITSTREAM_ENDIANNESS_BIG_ENDIAN;
1573 	vpu_write_reg(inst->dev, W5_CMD_BS_PARAM, reg_val);
1574 	vpu_write_reg(inst->dev, W5_CMD_EXT_ADDR, 0);
1575 	vpu_write_reg(inst->dev, W5_CMD_NUM_CQ_DEPTH_M1, WAVE521_COMMAND_QUEUE_DEPTH - 1);
1576 
1577 	/* This register must be reset explicitly */
1578 	vpu_write_reg(inst->dev, W5_CMD_ENC_SRC_OPTIONS, 0);
1579 	vpu_write_reg(inst->dev, W5_CMD_ENC_VCORE_INFO, 1);
1580 
1581 	ret = send_firmware_command(inst, W5_CREATE_INSTANCE, true, NULL, NULL);
1582 	if (ret)
1583 		goto free_vb_work;
1584 
1585 	buffer_addr = open_param->bitstream_buffer;
1586 	buffer_size = open_param->bitstream_buffer_size;
1587 	p_enc_info->stream_rd_ptr = buffer_addr;
1588 	p_enc_info->stream_wr_ptr = buffer_addr;
1589 	p_enc_info->line_buf_int_en = open_param->line_buf_int_en;
1590 	p_enc_info->stream_buf_start_addr = buffer_addr;
1591 	p_enc_info->stream_buf_size = buffer_size;
1592 	p_enc_info->stream_buf_end_addr = buffer_addr + buffer_size;
1593 	p_enc_info->stride = 0;
1594 	p_enc_info->initial_info_obtained = false;
1595 	p_enc_info->product_code = vpu_read_reg(inst->dev, W5_PRODUCT_NUMBER);
1596 
1597 	return 0;
1598 free_vb_work:
1599 	if (wave5_vdi_free_dma_memory(vpu_dev, &p_enc_info->vb_work))
1600 		memset(&p_enc_info->vb_work, 0, sizeof(p_enc_info->vb_work));
1601 	return ret;
1602 }
1603 
wave5_set_enc_crop_info(u32 codec,struct enc_wave_param * param,int rot_mode,int src_width,int src_height)1604 static void wave5_set_enc_crop_info(u32 codec, struct enc_wave_param *param, int rot_mode,
1605 				    int src_width, int src_height)
1606 {
1607 	int aligned_width = (codec == W_HEVC_ENC) ? ALIGN(src_width, 32) : ALIGN(src_width, 16);
1608 	int aligned_height = (codec == W_HEVC_ENC) ? ALIGN(src_height, 32) : ALIGN(src_height, 16);
1609 	int pad_right, pad_bot;
1610 	int crop_right, crop_left, crop_top, crop_bot;
1611 	int prp_mode = rot_mode >> 1; /* remove prp_enable bit */
1612 
1613 	if (codec == W_HEVC_ENC &&
1614 	    (!rot_mode || prp_mode == 14)) /* prp_mode 14 : hor_mir && ver_mir && rot_180 */
1615 		return;
1616 
1617 	pad_right = aligned_width - src_width;
1618 	pad_bot = aligned_height - src_height;
1619 
1620 	if (param->conf_win_right > 0)
1621 		crop_right = param->conf_win_right + pad_right;
1622 	else
1623 		crop_right = pad_right;
1624 
1625 	if (param->conf_win_bot > 0)
1626 		crop_bot = param->conf_win_bot + pad_bot;
1627 	else
1628 		crop_bot = pad_bot;
1629 
1630 	crop_top = param->conf_win_top;
1631 	crop_left = param->conf_win_left;
1632 
1633 	param->conf_win_top = crop_top;
1634 	param->conf_win_left = crop_left;
1635 	param->conf_win_bot = crop_bot;
1636 	param->conf_win_right = crop_right;
1637 
1638 	switch (prp_mode) {
1639 	case 0:
1640 		return;
1641 	case 1:
1642 	case 15:
1643 		param->conf_win_top = crop_right;
1644 		param->conf_win_left = crop_top;
1645 		param->conf_win_bot = crop_left;
1646 		param->conf_win_right = crop_bot;
1647 		break;
1648 	case 2:
1649 	case 12:
1650 		param->conf_win_top = crop_bot;
1651 		param->conf_win_left = crop_right;
1652 		param->conf_win_bot = crop_top;
1653 		param->conf_win_right = crop_left;
1654 		break;
1655 	case 3:
1656 	case 13:
1657 		param->conf_win_top = crop_left;
1658 		param->conf_win_left = crop_bot;
1659 		param->conf_win_bot = crop_right;
1660 		param->conf_win_right = crop_top;
1661 		break;
1662 	case 4:
1663 	case 10:
1664 		param->conf_win_top = crop_bot;
1665 		param->conf_win_bot = crop_top;
1666 		break;
1667 	case 8:
1668 	case 6:
1669 		param->conf_win_left = crop_right;
1670 		param->conf_win_right = crop_left;
1671 		break;
1672 	case 5:
1673 	case 11:
1674 		param->conf_win_top = crop_left;
1675 		param->conf_win_left = crop_top;
1676 		param->conf_win_bot = crop_right;
1677 		param->conf_win_right = crop_bot;
1678 		break;
1679 	case 7:
1680 	case 9:
1681 		param->conf_win_top = crop_right;
1682 		param->conf_win_left = crop_bot;
1683 		param->conf_win_bot = crop_left;
1684 		param->conf_win_right = crop_top;
1685 		break;
1686 	default:
1687 		WARN(1, "Invalid prp_mode: %d, must be in range of 1 - 15\n", prp_mode);
1688 	}
1689 }
1690 
wave5_vpu_enc_init_seq(struct vpu_instance * inst)1691 int wave5_vpu_enc_init_seq(struct vpu_instance *inst)
1692 {
1693 	u32 reg_val = 0, rot_mir_mode, fixed_cu_size_mode = 0x7;
1694 	struct enc_info *p_enc_info = &inst->codec_info->enc_info;
1695 	struct enc_open_param *p_open_param = &p_enc_info->open_param;
1696 	struct enc_wave_param *p_param = &p_open_param->wave_param;
1697 
1698 	/*
1699 	 * OPT_COMMON:
1700 	 *	the last SET_PARAM command should be called with OPT_COMMON
1701 	 */
1702 	rot_mir_mode = 0;
1703 	if (p_enc_info->rotation_enable) {
1704 		switch (p_enc_info->rotation_angle) {
1705 		case 0:
1706 			rot_mir_mode |= NONE_ROTATE;
1707 			break;
1708 		case 90:
1709 			rot_mir_mode |= ROT_CLOCKWISE_90;
1710 			break;
1711 		case 180:
1712 			rot_mir_mode |= ROT_CLOCKWISE_180;
1713 			break;
1714 		case 270:
1715 			rot_mir_mode |= ROT_CLOCKWISE_270;
1716 			break;
1717 		}
1718 	}
1719 
1720 	if (p_enc_info->mirror_enable) {
1721 		switch (p_enc_info->mirror_direction) {
1722 		case MIRDIR_NONE:
1723 			rot_mir_mode |= NONE_ROTATE;
1724 			break;
1725 		case MIRDIR_VER:
1726 			rot_mir_mode |= MIR_VER_FLIP;
1727 			break;
1728 		case MIRDIR_HOR:
1729 			rot_mir_mode |= MIR_HOR_FLIP;
1730 			break;
1731 		case MIRDIR_HOR_VER:
1732 			rot_mir_mode |= MIR_HOR_VER_FLIP;
1733 			break;
1734 		}
1735 	}
1736 
1737 	wave5_set_enc_crop_info(inst->std, p_param, rot_mir_mode, p_open_param->pic_width,
1738 				p_open_param->pic_height);
1739 
1740 	/* SET_PARAM + COMMON */
1741 	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_SET_PARAM_OPTION, OPT_COMMON);
1742 	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_SRC_SIZE, p_open_param->pic_height << 16
1743 			| p_open_param->pic_width);
1744 	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CUSTOM_MAP_ENDIAN, VDI_LITTLE_ENDIAN);
1745 
1746 	reg_val = p_param->profile |
1747 		(p_param->level << 3) |
1748 		(p_param->internal_bit_depth << 14);
1749 	if (inst->std == W_HEVC_ENC)
1750 		reg_val |= (p_param->tier << 12) |
1751 			(p_param->tmvp_enable << 23) |
1752 			(p_param->sao_enable << 24) |
1753 			(p_param->skip_intra_trans << 25) |
1754 			(p_param->strong_intra_smooth_enable << 27) |
1755 			(p_param->en_still_picture << 30);
1756 	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_SPS_PARAM, reg_val);
1757 
1758 	reg_val = (p_param->lossless_enable) |
1759 		(p_param->const_intra_pred_flag << 1) |
1760 		(p_param->lf_cross_slice_boundary_enable << 2) |
1761 		(p_param->wpp_enable << 4) |
1762 		(p_param->disable_deblk << 5) |
1763 		((p_param->beta_offset_div2 & 0xF) << 6) |
1764 		((p_param->tc_offset_div2 & 0xF) << 10) |
1765 		((p_param->chroma_cb_qp_offset & 0x1F) << 14) |
1766 		((p_param->chroma_cr_qp_offset & 0x1F) << 19) |
1767 		(p_param->transform8x8_enable << 29) |
1768 		(p_param->entropy_coding_mode << 30);
1769 	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_PPS_PARAM, reg_val);
1770 
1771 	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_GOP_PARAM, p_param->gop_preset_idx);
1772 
1773 	if (inst->std == W_AVC_ENC)
1774 		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_INTRA_PARAM, p_param->intra_qp |
1775 				((p_param->intra_period & 0x7ff) << 6) |
1776 				((p_param->avc_idr_period & 0x7ff) << 17));
1777 	else if (inst->std == W_HEVC_ENC)
1778 		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_INTRA_PARAM,
1779 			      p_param->decoding_refresh_type | (p_param->intra_qp << 3) |
1780 				(p_param->intra_period << 16));
1781 
1782 	reg_val = (p_param->rdo_skip << 2) |
1783 		(p_param->lambda_scaling_enable << 3) |
1784 		(fixed_cu_size_mode << 5) |
1785 		(p_param->intra_nx_n_enable << 8) |
1786 		(p_param->max_num_merge << 18);
1787 
1788 	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_RDO_PARAM, reg_val);
1789 
1790 	if (inst->std == W_AVC_ENC)
1791 		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_INTRA_REFRESH,
1792 			      p_param->intra_mb_refresh_arg << 16 | p_param->intra_mb_refresh_mode);
1793 	else if (inst->std == W_HEVC_ENC)
1794 		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_INTRA_REFRESH,
1795 			      p_param->intra_refresh_arg << 16 | p_param->intra_refresh_mode);
1796 
1797 	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_RC_FRAME_RATE, p_open_param->frame_rate_info);
1798 	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_RC_TARGET_RATE, p_open_param->bit_rate);
1799 
1800 	reg_val = p_open_param->rc_enable |
1801 		(p_param->hvs_qp_enable << 2) |
1802 		(p_param->hvs_qp_scale << 4) |
1803 		((p_param->initial_rc_qp & 0x3F) << 14) |
1804 		(p_open_param->vbv_buffer_size << 20);
1805 	if (inst->std == W_AVC_ENC)
1806 		reg_val |= (p_param->mb_level_rc_enable << 1);
1807 	else if (inst->std == W_HEVC_ENC)
1808 		reg_val |= (p_param->cu_level_rc_enable << 1);
1809 	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_RC_PARAM, reg_val);
1810 
1811 	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_RC_WEIGHT_PARAM,
1812 		      p_param->rc_weight_buf << 8 | p_param->rc_weight_param);
1813 
1814 	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_RC_MIN_MAX_QP, p_param->min_qp_i |
1815 		      (p_param->max_qp_i << 6) | (p_param->hvs_max_delta_qp << 12));
1816 
1817 	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_RC_INTER_MIN_MAX_QP, p_param->min_qp_p |
1818 		      (p_param->max_qp_p << 6) | (p_param->min_qp_b << 12) |
1819 		      (p_param->max_qp_b << 18));
1820 
1821 	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_RC_BIT_RATIO_LAYER_0_3, 0);
1822 	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_RC_BIT_RATIO_LAYER_4_7, 0);
1823 	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_ROT_PARAM, rot_mir_mode);
1824 
1825 	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_BG_PARAM, 0);
1826 	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CUSTOM_LAMBDA_ADDR, 0);
1827 	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CONF_WIN_TOP_BOT,
1828 		      p_param->conf_win_bot << 16 | p_param->conf_win_top);
1829 	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CONF_WIN_LEFT_RIGHT,
1830 		      p_param->conf_win_right << 16 | p_param->conf_win_left);
1831 
1832 	if (inst->std == W_AVC_ENC)
1833 		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_INDEPENDENT_SLICE,
1834 			      p_param->avc_slice_arg << 16 | p_param->avc_slice_mode);
1835 	else if (inst->std == W_HEVC_ENC)
1836 		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_INDEPENDENT_SLICE,
1837 			      p_param->independ_slice_mode_arg << 16 |
1838 			      p_param->independ_slice_mode);
1839 
1840 	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_USER_SCALING_LIST_ADDR, 0);
1841 	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_NUM_UNITS_IN_TICK, 0);
1842 	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_TIME_SCALE, 0);
1843 	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_NUM_TICKS_POC_DIFF_ONE, 0);
1844 
1845 	if (inst->std == W_HEVC_ENC) {
1846 		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CUSTOM_MD_PU04, 0);
1847 		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CUSTOM_MD_PU08, 0);
1848 		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CUSTOM_MD_PU16, 0);
1849 		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CUSTOM_MD_PU32, 0);
1850 		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CUSTOM_MD_CU08, 0);
1851 		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CUSTOM_MD_CU16, 0);
1852 		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_CUSTOM_MD_CU32, 0);
1853 		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_DEPENDENT_SLICE,
1854 			      p_param->depend_slice_mode_arg << 16 | p_param->depend_slice_mode);
1855 
1856 		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_NR_PARAM, 0);
1857 
1858 		vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_NR_WEIGHT,
1859 			      p_param->nr_intra_weight_y |
1860 			      (p_param->nr_intra_weight_cb << 5) |
1861 			      (p_param->nr_intra_weight_cr << 10) |
1862 			      (p_param->nr_inter_weight_y << 15) |
1863 			      (p_param->nr_inter_weight_cb << 20) |
1864 			      (p_param->nr_inter_weight_cr << 25));
1865 	}
1866 	vpu_write_reg(inst->dev, W5_CMD_ENC_SEQ_VUI_HRD_PARAM, 0);
1867 
1868 	return send_firmware_command(inst, W5_ENC_SET_PARAM, true, NULL, NULL);
1869 }
1870 
wave5_vpu_enc_get_seq_info(struct vpu_instance * inst,struct enc_initial_info * info)1871 int wave5_vpu_enc_get_seq_info(struct vpu_instance *inst, struct enc_initial_info *info)
1872 {
1873 	int ret;
1874 	u32 reg_val;
1875 	struct enc_info *p_enc_info = &inst->codec_info->enc_info;
1876 
1877 	/* send QUERY cmd */
1878 	ret = wave5_send_query(inst->dev, inst, GET_RESULT);
1879 	if (ret)
1880 		return ret;
1881 
1882 	dev_dbg(inst->dev->dev, "%s: init seq\n", __func__);
1883 
1884 	reg_val = vpu_read_reg(inst->dev, W5_RET_QUEUE_STATUS);
1885 
1886 	p_enc_info->instance_queue_count = (reg_val >> 16) & 0xff;
1887 	p_enc_info->report_queue_count = (reg_val & QUEUE_REPORT_MASK);
1888 
1889 	if (vpu_read_reg(inst->dev, W5_RET_ENC_ENCODING_SUCCESS) != 1) {
1890 		info->seq_init_err_reason = vpu_read_reg(inst->dev, W5_RET_ENC_ERR_INFO);
1891 		ret = -EIO;
1892 	} else {
1893 		info->warn_info = vpu_read_reg(inst->dev, W5_RET_ENC_WARN_INFO);
1894 	}
1895 
1896 	info->min_frame_buffer_count = vpu_read_reg(inst->dev, W5_RET_ENC_NUM_REQUIRED_FB);
1897 	info->min_src_frame_count = vpu_read_reg(inst->dev, W5_RET_ENC_MIN_SRC_BUF_NUM);
1898 	info->vlc_buf_size = vpu_read_reg(inst->dev, W5_RET_VLC_BUF_SIZE);
1899 	info->param_buf_size = vpu_read_reg(inst->dev, W5_RET_PARAM_BUF_SIZE);
1900 	p_enc_info->vlc_buf_size = info->vlc_buf_size;
1901 	p_enc_info->param_buf_size = info->param_buf_size;
1902 
1903 	return ret;
1904 }
1905 
calculate_luma_stride(u32 width,u32 bit_depth)1906 static u32 calculate_luma_stride(u32 width, u32 bit_depth)
1907 {
1908 	return ALIGN(ALIGN(width, 16) * ((bit_depth > 8) ? 5 : 4), 32);
1909 }
1910 
calculate_chroma_stride(u32 width,u32 bit_depth)1911 static u32 calculate_chroma_stride(u32 width, u32 bit_depth)
1912 {
1913 	return ALIGN(ALIGN(width / 2, 16) * ((bit_depth > 8) ? 5 : 4), 32);
1914 }
1915 
wave5_vpu_enc_register_framebuffer(struct device * dev,struct vpu_instance * inst,struct frame_buffer * fb_arr,enum tiled_map_type map_type,unsigned int count)1916 int wave5_vpu_enc_register_framebuffer(struct device *dev, struct vpu_instance *inst,
1917 				       struct frame_buffer *fb_arr, enum tiled_map_type map_type,
1918 				       unsigned int count)
1919 {
1920 	struct vpu_device *vpu_dev = dev_get_drvdata(dev);
1921 	int ret = 0;
1922 	u32 stride;
1923 	u32 start_no, end_no;
1924 	size_t remain, idx, j, i, cnt_8_chunk;
1925 	u32 reg_val = 0, pic_size = 0, mv_col_size, fbc_y_tbl_size, fbc_c_tbl_size;
1926 	u32 sub_sampled_size = 0;
1927 	u32 luma_stride, chroma_stride;
1928 	u32 buf_height = 0, buf_width = 0;
1929 	u32 bit_depth;
1930 	bool avc_encoding = (inst->std == W_AVC_ENC);
1931 	struct vpu_buf vb_mv = {0};
1932 	struct vpu_buf vb_fbc_y_tbl = {0};
1933 	struct vpu_buf vb_fbc_c_tbl = {0};
1934 	struct vpu_buf vb_sub_sam_buf = {0};
1935 	struct vpu_buf vb_task = {0};
1936 	struct enc_open_param *p_open_param;
1937 	struct enc_info *p_enc_info = &inst->codec_info->enc_info;
1938 
1939 	p_open_param = &p_enc_info->open_param;
1940 	mv_col_size = 0;
1941 	fbc_y_tbl_size = 0;
1942 	fbc_c_tbl_size = 0;
1943 	stride = p_enc_info->stride;
1944 	bit_depth = p_open_param->wave_param.internal_bit_depth;
1945 
1946 	if (avc_encoding) {
1947 		buf_width = ALIGN(p_open_param->pic_width, 16);
1948 		buf_height = ALIGN(p_open_param->pic_height, 16);
1949 
1950 		if ((p_enc_info->rotation_angle || p_enc_info->mirror_direction) &&
1951 		    !(p_enc_info->rotation_angle == 180 &&
1952 					p_enc_info->mirror_direction == MIRDIR_HOR_VER)) {
1953 			buf_width = ALIGN(p_open_param->pic_width, 16);
1954 			buf_height = ALIGN(p_open_param->pic_height, 16);
1955 		}
1956 
1957 		if (p_enc_info->rotation_angle == 90 || p_enc_info->rotation_angle == 270) {
1958 			buf_width = ALIGN(p_open_param->pic_height, 16);
1959 			buf_height = ALIGN(p_open_param->pic_width, 16);
1960 		}
1961 	} else {
1962 		buf_width = ALIGN(p_open_param->pic_width, 8);
1963 		buf_height = ALIGN(p_open_param->pic_height, 8);
1964 
1965 		if ((p_enc_info->rotation_angle || p_enc_info->mirror_direction) &&
1966 		    !(p_enc_info->rotation_angle == 180 &&
1967 					p_enc_info->mirror_direction == MIRDIR_HOR_VER)) {
1968 			buf_width = ALIGN(p_open_param->pic_width, 32);
1969 			buf_height = ALIGN(p_open_param->pic_height, 32);
1970 		}
1971 
1972 		if (p_enc_info->rotation_angle == 90 || p_enc_info->rotation_angle == 270) {
1973 			buf_width = ALIGN(p_open_param->pic_height, 32);
1974 			buf_height = ALIGN(p_open_param->pic_width, 32);
1975 		}
1976 	}
1977 
1978 	pic_size = (buf_width << 16) | buf_height;
1979 
1980 	if (avc_encoding) {
1981 		mv_col_size = WAVE5_ENC_AVC_BUF_SIZE(buf_width, buf_height);
1982 		vb_mv.daddr = 0;
1983 		vb_mv.size = ALIGN(mv_col_size * count, BUFFER_MARGIN) + BUFFER_MARGIN;
1984 	} else {
1985 		mv_col_size = WAVE5_ENC_HEVC_BUF_SIZE(buf_width, buf_height);
1986 		mv_col_size = ALIGN(mv_col_size, 16);
1987 		vb_mv.daddr = 0;
1988 		vb_mv.size = ALIGN(mv_col_size * count, BUFFER_MARGIN) + BUFFER_MARGIN;
1989 	}
1990 
1991 	ret = wave5_vdi_allocate_dma_memory(vpu_dev, &vb_mv);
1992 	if (ret)
1993 		return ret;
1994 
1995 	p_enc_info->vb_mv = vb_mv;
1996 
1997 	fbc_y_tbl_size = ALIGN(WAVE5_FBC_LUMA_TABLE_SIZE(buf_width, buf_height), 16);
1998 	fbc_c_tbl_size = ALIGN(WAVE5_FBC_CHROMA_TABLE_SIZE(buf_width, buf_height), 16);
1999 
2000 	vb_fbc_y_tbl.daddr = 0;
2001 	vb_fbc_y_tbl.size = ALIGN(fbc_y_tbl_size * count, BUFFER_MARGIN) + BUFFER_MARGIN;
2002 	ret = wave5_vdi_allocate_dma_memory(vpu_dev, &vb_fbc_y_tbl);
2003 	if (ret)
2004 		goto free_vb_fbc_y_tbl;
2005 
2006 	p_enc_info->vb_fbc_y_tbl = vb_fbc_y_tbl;
2007 
2008 	vb_fbc_c_tbl.daddr = 0;
2009 	vb_fbc_c_tbl.size = ALIGN(fbc_c_tbl_size * count, BUFFER_MARGIN) + BUFFER_MARGIN;
2010 	ret = wave5_vdi_allocate_dma_memory(vpu_dev, &vb_fbc_c_tbl);
2011 	if (ret)
2012 		goto free_vb_fbc_c_tbl;
2013 
2014 	p_enc_info->vb_fbc_c_tbl = vb_fbc_c_tbl;
2015 
2016 	if (avc_encoding)
2017 		sub_sampled_size = WAVE5_SUBSAMPLED_ONE_SIZE_AVC(buf_width, buf_height);
2018 	else
2019 		sub_sampled_size = WAVE5_SUBSAMPLED_ONE_SIZE(buf_width, buf_height);
2020 	vb_sub_sam_buf.size = ALIGN(sub_sampled_size * count, BUFFER_MARGIN) + BUFFER_MARGIN;
2021 	vb_sub_sam_buf.daddr = 0;
2022 	ret = wave5_vdi_allocate_dma_memory(vpu_dev, &vb_sub_sam_buf);
2023 	if (ret)
2024 		goto free_vb_sam_buf;
2025 
2026 	p_enc_info->vb_sub_sam_buf = vb_sub_sam_buf;
2027 
2028 	vb_task.size = (p_enc_info->vlc_buf_size * VLC_BUF_NUM) +
2029 			(p_enc_info->param_buf_size * WAVE521_COMMAND_QUEUE_DEPTH);
2030 	vb_task.daddr = 0;
2031 	if (p_enc_info->vb_task.size == 0) {
2032 		ret = wave5_vdi_allocate_dma_memory(vpu_dev, &vb_task);
2033 		if (ret)
2034 			goto free_vb_task;
2035 
2036 		p_enc_info->vb_task = vb_task;
2037 
2038 		vpu_write_reg(inst->dev, W5_CMD_SET_FB_ADDR_TASK_BUF,
2039 			      p_enc_info->vb_task.daddr);
2040 		vpu_write_reg(inst->dev, W5_CMD_SET_FB_TASK_BUF_SIZE, vb_task.size);
2041 	}
2042 
2043 	/* set sub-sampled buffer base addr */
2044 	vpu_write_reg(inst->dev, W5_ADDR_SUB_SAMPLED_FB_BASE, vb_sub_sam_buf.daddr);
2045 	/* set sub-sampled buffer size for one frame */
2046 	vpu_write_reg(inst->dev, W5_SUB_SAMPLED_ONE_FB_SIZE, sub_sampled_size);
2047 
2048 	vpu_write_reg(inst->dev, W5_PIC_SIZE, pic_size);
2049 
2050 	/* set stride of luma/chroma for compressed buffer */
2051 	if ((p_enc_info->rotation_angle || p_enc_info->mirror_direction) &&
2052 	    !(p_enc_info->rotation_angle == 180 &&
2053 	    p_enc_info->mirror_direction == MIRDIR_HOR_VER)) {
2054 		luma_stride = calculate_luma_stride(buf_width, bit_depth);
2055 		chroma_stride = calculate_chroma_stride(buf_width / 2, bit_depth);
2056 	} else {
2057 		luma_stride = calculate_luma_stride(p_open_param->pic_width, bit_depth);
2058 		chroma_stride = calculate_chroma_stride(p_open_param->pic_width / 2, bit_depth);
2059 	}
2060 
2061 	vpu_write_reg(inst->dev, W5_FBC_STRIDE, luma_stride << 16 | chroma_stride);
2062 	vpu_write_reg(inst->dev, W5_COMMON_PIC_INFO, stride);
2063 
2064 	remain = count;
2065 	cnt_8_chunk = DIV_ROUND_UP(count, 8);
2066 	idx = 0;
2067 	for (j = 0; j < cnt_8_chunk; j++) {
2068 		reg_val = (j == cnt_8_chunk - 1) << 4 | ((j == 0) << 3);
2069 		vpu_write_reg(inst->dev, W5_SFB_OPTION, reg_val);
2070 		start_no = j * 8;
2071 		end_no = start_no + ((remain >= 8) ? 8 : remain) - 1;
2072 
2073 		vpu_write_reg(inst->dev, W5_SET_FB_NUM, (start_no << 8) | end_no);
2074 
2075 		for (i = 0; i < 8 && i < remain; i++) {
2076 			vpu_write_reg(inst->dev, W5_ADDR_LUMA_BASE0 + (i << 4), fb_arr[i +
2077 					start_no].buf_y);
2078 			vpu_write_reg(inst->dev, W5_ADDR_CB_BASE0 + (i << 4),
2079 				      fb_arr[i + start_no].buf_cb);
2080 			/* luma FBC offset table */
2081 			vpu_write_reg(inst->dev, W5_ADDR_FBC_Y_OFFSET0 + (i << 4),
2082 				      vb_fbc_y_tbl.daddr + idx * fbc_y_tbl_size);
2083 			/* chroma FBC offset table */
2084 			vpu_write_reg(inst->dev, W5_ADDR_FBC_C_OFFSET0 + (i << 4),
2085 				      vb_fbc_c_tbl.daddr + idx * fbc_c_tbl_size);
2086 
2087 			vpu_write_reg(inst->dev, W5_ADDR_MV_COL0 + (i << 2),
2088 				      vb_mv.daddr + idx * mv_col_size);
2089 			idx++;
2090 		}
2091 		remain -= i;
2092 
2093 		ret = send_firmware_command(inst, W5_SET_FB, false, NULL, NULL);
2094 		if (ret)
2095 			goto free_vb_mem;
2096 	}
2097 
2098 	ret = wave5_vpu_firmware_command_queue_error_check(vpu_dev, NULL);
2099 	if (ret)
2100 		goto free_vb_mem;
2101 
2102 	return ret;
2103 
2104 free_vb_mem:
2105 	wave5_vdi_free_dma_memory(vpu_dev, &vb_task);
2106 free_vb_task:
2107 	wave5_vdi_free_dma_memory(vpu_dev, &vb_sub_sam_buf);
2108 free_vb_sam_buf:
2109 	wave5_vdi_free_dma_memory(vpu_dev, &vb_fbc_c_tbl);
2110 free_vb_fbc_c_tbl:
2111 	wave5_vdi_free_dma_memory(vpu_dev, &vb_fbc_y_tbl);
2112 free_vb_fbc_y_tbl:
2113 	wave5_vdi_free_dma_memory(vpu_dev, &vb_mv);
2114 	return ret;
2115 }
2116 
wave5_vpu_enc_validate_sec_axi(struct vpu_instance * inst)2117 static u32 wave5_vpu_enc_validate_sec_axi(struct vpu_instance *inst)
2118 {
2119 	struct enc_info *p_enc_info = &inst->codec_info->enc_info;
2120 	u32 rdo_size = 0, lf_size = 0, ret = 0;
2121 	u32 sram_size = inst->dev->sram_size;
2122 
2123 	if (!sram_size)
2124 		return 0;
2125 
2126 	/*
2127 	 * TODO: calculate rdo_size and lf_size from inst->src_fmt.width and
2128 	 * inst->codec_info->enc_info.open_param.wave_param.internal_bit_depth
2129 	 */
2130 
2131 	if (p_enc_info->sec_axi_info.use_enc_rdo_enable && sram_size >= rdo_size) {
2132 		ret |= BIT(11);
2133 		sram_size -= rdo_size;
2134 	}
2135 
2136 	if (p_enc_info->sec_axi_info.use_enc_lf_enable && sram_size >= lf_size)
2137 		ret |= BIT(15);
2138 
2139 	return ret;
2140 }
2141 
wave5_vpu_encode(struct vpu_instance * inst,struct enc_param * option,u32 * fail_res)2142 int wave5_vpu_encode(struct vpu_instance *inst, struct enc_param *option, u32 *fail_res)
2143 {
2144 	u32 src_frame_format;
2145 	u32 reg_val = 0;
2146 	u32 src_stride_c = 0;
2147 	struct enc_info *p_enc_info = &inst->codec_info->enc_info;
2148 	struct frame_buffer *p_src_frame = option->source_frame;
2149 	struct enc_open_param *p_open_param = &p_enc_info->open_param;
2150 	bool justified = WTL_RIGHT_JUSTIFIED;
2151 	u32 format_no = WTL_PIXEL_8BIT;
2152 	int ret;
2153 
2154 	vpu_write_reg(inst->dev, W5_CMD_ENC_BS_START_ADDR, option->pic_stream_buffer_addr);
2155 	vpu_write_reg(inst->dev, W5_CMD_ENC_BS_SIZE, option->pic_stream_buffer_size);
2156 	p_enc_info->stream_buf_start_addr = option->pic_stream_buffer_addr;
2157 	p_enc_info->stream_buf_size = option->pic_stream_buffer_size;
2158 	p_enc_info->stream_buf_end_addr =
2159 		option->pic_stream_buffer_addr + option->pic_stream_buffer_size;
2160 
2161 	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SRC_AXI_SEL, DEFAULT_SRC_AXI);
2162 	/* secondary AXI */
2163 	reg_val = wave5_vpu_enc_validate_sec_axi(inst);
2164 	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_USE_SEC_AXI, reg_val);
2165 
2166 	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_REPORT_PARAM, 0);
2167 
2168 	/*
2169 	 * CODEOPT_ENC_VCL is used to implicitly encode header/headers to generate bitstream.
2170 	 * (use ENC_PUT_VIDEO_HEADER for give_command to encode only a header)
2171 	 */
2172 	if (option->code_option.implicit_header_encode)
2173 		vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_CODE_OPTION,
2174 			      CODEOPT_ENC_HEADER_IMPLICIT | CODEOPT_ENC_VCL |
2175 			      (option->code_option.encode_aud << 5) |
2176 			      (option->code_option.encode_eos << 6) |
2177 			      (option->code_option.encode_eob << 7));
2178 	else
2179 		vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_CODE_OPTION,
2180 			      option->code_option.implicit_header_encode |
2181 			      (option->code_option.encode_vcl << 1) |
2182 			      (option->code_option.encode_vps << 2) |
2183 			      (option->code_option.encode_sps << 3) |
2184 			      (option->code_option.encode_pps << 4) |
2185 			      (option->code_option.encode_aud << 5) |
2186 			      (option->code_option.encode_eos << 6) |
2187 			      (option->code_option.encode_eob << 7));
2188 
2189 	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_PIC_PARAM, 0);
2190 
2191 	if (option->src_end_flag)
2192 		/* no more source images. */
2193 		vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SRC_PIC_IDX, 0xFFFFFFFF);
2194 	else
2195 		vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SRC_PIC_IDX, option->src_idx);
2196 
2197 	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SRC_ADDR_Y, p_src_frame->buf_y);
2198 	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SRC_ADDR_U, p_src_frame->buf_cb);
2199 	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SRC_ADDR_V, p_src_frame->buf_cr);
2200 
2201 	switch (p_open_param->src_format) {
2202 	case FORMAT_420:
2203 	case FORMAT_422:
2204 	case FORMAT_YUYV:
2205 	case FORMAT_YVYU:
2206 	case FORMAT_UYVY:
2207 	case FORMAT_VYUY:
2208 		justified = WTL_LEFT_JUSTIFIED;
2209 		format_no = WTL_PIXEL_8BIT;
2210 		src_stride_c = inst->cbcr_interleave ? p_src_frame->stride :
2211 			(p_src_frame->stride / 2);
2212 		src_stride_c = (p_open_param->src_format == FORMAT_422) ? src_stride_c * 2 :
2213 			src_stride_c;
2214 		break;
2215 	case FORMAT_420_P10_16BIT_MSB:
2216 	case FORMAT_422_P10_16BIT_MSB:
2217 	case FORMAT_YUYV_P10_16BIT_MSB:
2218 	case FORMAT_YVYU_P10_16BIT_MSB:
2219 	case FORMAT_UYVY_P10_16BIT_MSB:
2220 	case FORMAT_VYUY_P10_16BIT_MSB:
2221 		justified = WTL_RIGHT_JUSTIFIED;
2222 		format_no = WTL_PIXEL_16BIT;
2223 		src_stride_c = inst->cbcr_interleave ? p_src_frame->stride :
2224 			(p_src_frame->stride / 2);
2225 		src_stride_c = (p_open_param->src_format ==
2226 				FORMAT_422_P10_16BIT_MSB) ? src_stride_c * 2 : src_stride_c;
2227 		break;
2228 	case FORMAT_420_P10_16BIT_LSB:
2229 	case FORMAT_422_P10_16BIT_LSB:
2230 	case FORMAT_YUYV_P10_16BIT_LSB:
2231 	case FORMAT_YVYU_P10_16BIT_LSB:
2232 	case FORMAT_UYVY_P10_16BIT_LSB:
2233 	case FORMAT_VYUY_P10_16BIT_LSB:
2234 		justified = WTL_LEFT_JUSTIFIED;
2235 		format_no = WTL_PIXEL_16BIT;
2236 		src_stride_c = inst->cbcr_interleave ? p_src_frame->stride :
2237 			(p_src_frame->stride / 2);
2238 		src_stride_c = (p_open_param->src_format ==
2239 				FORMAT_422_P10_16BIT_LSB) ? src_stride_c * 2 : src_stride_c;
2240 		break;
2241 	case FORMAT_420_P10_32BIT_MSB:
2242 	case FORMAT_422_P10_32BIT_MSB:
2243 	case FORMAT_YUYV_P10_32BIT_MSB:
2244 	case FORMAT_YVYU_P10_32BIT_MSB:
2245 	case FORMAT_UYVY_P10_32BIT_MSB:
2246 	case FORMAT_VYUY_P10_32BIT_MSB:
2247 		justified = WTL_RIGHT_JUSTIFIED;
2248 		format_no = WTL_PIXEL_32BIT;
2249 		src_stride_c = inst->cbcr_interleave ? p_src_frame->stride :
2250 			ALIGN(p_src_frame->stride / 2, 16) * BIT(inst->cbcr_interleave);
2251 		src_stride_c = (p_open_param->src_format ==
2252 				FORMAT_422_P10_32BIT_MSB) ? src_stride_c * 2 : src_stride_c;
2253 		break;
2254 	case FORMAT_420_P10_32BIT_LSB:
2255 	case FORMAT_422_P10_32BIT_LSB:
2256 	case FORMAT_YUYV_P10_32BIT_LSB:
2257 	case FORMAT_YVYU_P10_32BIT_LSB:
2258 	case FORMAT_UYVY_P10_32BIT_LSB:
2259 	case FORMAT_VYUY_P10_32BIT_LSB:
2260 		justified = WTL_LEFT_JUSTIFIED;
2261 		format_no = WTL_PIXEL_32BIT;
2262 		src_stride_c = inst->cbcr_interleave ? p_src_frame->stride :
2263 			ALIGN(p_src_frame->stride / 2, 16) * BIT(inst->cbcr_interleave);
2264 		src_stride_c = (p_open_param->src_format ==
2265 				FORMAT_422_P10_32BIT_LSB) ? src_stride_c * 2 : src_stride_c;
2266 		break;
2267 	default:
2268 		return -EINVAL;
2269 	}
2270 
2271 	src_frame_format = (inst->cbcr_interleave << 1) | (inst->nv21);
2272 	switch (p_open_param->packed_format) {
2273 	case PACKED_YUYV:
2274 		src_frame_format = 4;
2275 		break;
2276 	case PACKED_YVYU:
2277 		src_frame_format = 5;
2278 		break;
2279 	case PACKED_UYVY:
2280 		src_frame_format = 6;
2281 		break;
2282 	case PACKED_VYUY:
2283 		src_frame_format = 7;
2284 		break;
2285 	default:
2286 		break;
2287 	}
2288 
2289 	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SRC_STRIDE,
2290 		      (p_src_frame->stride << 16) | src_stride_c);
2291 	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SRC_FORMAT, src_frame_format |
2292 		      (format_no << 3) | (justified << 5) | (PIC_SRC_ENDIANNESS_BIG_ENDIAN << 6));
2293 
2294 	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_CUSTOM_MAP_OPTION_ADDR, 0);
2295 	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_CUSTOM_MAP_OPTION_PARAM, 0);
2296 	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_LONGTERM_PIC, 0);
2297 	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_WP_PIXEL_SIGMA_Y, 0);
2298 	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_WP_PIXEL_SIGMA_C, 0);
2299 	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_WP_PIXEL_MEAN_Y, 0);
2300 	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_WP_PIXEL_MEAN_C, 0);
2301 	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_PREFIX_SEI_INFO, 0);
2302 	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_PREFIX_SEI_NAL_ADDR, 0);
2303 	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SUFFIX_SEI_INFO, 0);
2304 	vpu_write_reg(inst->dev, W5_CMD_ENC_PIC_SUFFIX_SEI_NAL_ADDR, 0);
2305 
2306 	ret = send_firmware_command(inst, W5_DEC_ENC_PIC, true, &reg_val, fail_res);
2307 	if (ret == -ETIMEDOUT)
2308 		return ret;
2309 
2310 	p_enc_info->instance_queue_count = (reg_val >> 16) & 0xff;
2311 	p_enc_info->report_queue_count = (reg_val & QUEUE_REPORT_MASK);
2312 
2313 	if (ret)
2314 		return ret;
2315 
2316 	return 0;
2317 }
2318 
wave5_vpu_enc_get_result(struct vpu_instance * inst,struct enc_output_info * result)2319 int wave5_vpu_enc_get_result(struct vpu_instance *inst, struct enc_output_info *result)
2320 {
2321 	int ret;
2322 	u32 encoding_success;
2323 	u32 reg_val;
2324 	struct enc_info *p_enc_info = &inst->codec_info->enc_info;
2325 	struct vpu_device *vpu_dev = inst->dev;
2326 
2327 	ret = wave5_send_query(inst->dev, inst, GET_RESULT);
2328 	if (ret)
2329 		return ret;
2330 
2331 	dev_dbg(inst->dev->dev, "%s: enc pic complete\n", __func__);
2332 
2333 	reg_val = vpu_read_reg(inst->dev, W5_RET_QUEUE_STATUS);
2334 
2335 	p_enc_info->instance_queue_count = (reg_val >> 16) & 0xff;
2336 	p_enc_info->report_queue_count = (reg_val & QUEUE_REPORT_MASK);
2337 
2338 	encoding_success = vpu_read_reg(inst->dev, W5_RET_ENC_ENCODING_SUCCESS);
2339 	if (!encoding_success) {
2340 		result->error_reason = vpu_read_reg(inst->dev, W5_RET_ENC_ERR_INFO);
2341 		return -EIO;
2342 	}
2343 
2344 	result->warn_info = vpu_read_reg(inst->dev, W5_RET_ENC_WARN_INFO);
2345 
2346 	reg_val = vpu_read_reg(inst->dev, W5_RET_ENC_PIC_TYPE);
2347 	result->pic_type = reg_val & 0xFFFF;
2348 
2349 	result->enc_vcl_nut = vpu_read_reg(inst->dev, W5_RET_ENC_VCL_NUT);
2350 	/*
2351 	 * To get the reconstructed frame use the following index on
2352 	 * inst->frame_buf
2353 	 */
2354 	result->recon_frame_index = vpu_read_reg(inst->dev, W5_RET_ENC_PIC_IDX);
2355 	result->enc_pic_byte = vpu_read_reg(inst->dev, W5_RET_ENC_PIC_BYTE);
2356 	result->enc_src_idx = vpu_read_reg(inst->dev, W5_RET_ENC_USED_SRC_IDX);
2357 	p_enc_info->stream_wr_ptr = vpu_read_reg(inst->dev, W5_RET_ENC_WR_PTR);
2358 	p_enc_info->stream_rd_ptr = vpu_read_reg(inst->dev, W5_RET_ENC_RD_PTR);
2359 
2360 	result->bitstream_buffer = vpu_read_reg(inst->dev, W5_RET_ENC_RD_PTR);
2361 	result->rd_ptr = p_enc_info->stream_rd_ptr;
2362 	result->wr_ptr = p_enc_info->stream_wr_ptr;
2363 
2364 	/*result for header only(no vcl) encoding */
2365 	if (result->recon_frame_index == RECON_IDX_FLAG_HEADER_ONLY)
2366 		result->bitstream_size = result->enc_pic_byte;
2367 	else if (result->recon_frame_index < 0)
2368 		result->bitstream_size = 0;
2369 	else
2370 		result->bitstream_size = result->enc_pic_byte;
2371 
2372 	result->enc_host_cmd_tick = vpu_read_reg(inst->dev, W5_RET_ENC_HOST_CMD_TICK);
2373 	result->enc_encode_end_tick = vpu_read_reg(inst->dev, W5_RET_ENC_ENCODING_END_TICK);
2374 
2375 	if (!p_enc_info->first_cycle_check) {
2376 		result->frame_cycle = (result->enc_encode_end_tick - result->enc_host_cmd_tick) *
2377 			p_enc_info->cycle_per_tick;
2378 		p_enc_info->first_cycle_check = true;
2379 	} else {
2380 		result->frame_cycle =
2381 			(result->enc_encode_end_tick - vpu_dev->last_performance_cycles) *
2382 			p_enc_info->cycle_per_tick;
2383 		if (vpu_dev->last_performance_cycles < result->enc_host_cmd_tick)
2384 			result->frame_cycle = (result->enc_encode_end_tick -
2385 					result->enc_host_cmd_tick) * p_enc_info->cycle_per_tick;
2386 	}
2387 	vpu_dev->last_performance_cycles = result->enc_encode_end_tick;
2388 
2389 	return 0;
2390 }
2391 
wave5_vpu_enc_finish_seq(struct vpu_instance * inst,u32 * fail_res)2392 int wave5_vpu_enc_finish_seq(struct vpu_instance *inst, u32 *fail_res)
2393 {
2394 	return send_firmware_command(inst, W5_DESTROY_INSTANCE, true, NULL, fail_res);
2395 }
2396 
wave5_vpu_enc_check_common_param_valid(struct vpu_instance * inst,struct enc_open_param * open_param)2397 static bool wave5_vpu_enc_check_common_param_valid(struct vpu_instance *inst,
2398 						   struct enc_open_param *open_param)
2399 {
2400 	bool low_delay = true;
2401 	struct enc_wave_param *param = &open_param->wave_param;
2402 	struct vpu_device *vpu_dev = inst->dev;
2403 	struct device *dev = vpu_dev->dev;
2404 	u32 num_ctu_row = (open_param->pic_height + 64 - 1) / 64;
2405 	u32 num_ctu_col = (open_param->pic_width + 64 - 1) / 64;
2406 	u32 ctu_sz = num_ctu_col * num_ctu_row;
2407 
2408 	if (inst->std == W_HEVC_ENC && low_delay &&
2409 	    param->decoding_refresh_type == DEC_REFRESH_TYPE_CRA) {
2410 		dev_warn(dev,
2411 			 "dec_refresh_type(CRA) shouldn't be used together with low delay GOP\n");
2412 		dev_warn(dev, "Suggested configuration parameter: decoding refresh type (IDR)\n");
2413 		param->decoding_refresh_type = 2;
2414 	}
2415 
2416 	if (param->wpp_enable && param->independ_slice_mode) {
2417 		unsigned int num_ctb_in_width = ALIGN(open_param->pic_width, 64) >> 6;
2418 
2419 		if (param->independ_slice_mode_arg % num_ctb_in_width) {
2420 			dev_err(dev, "independ_slice_mode_arg %u must be a multiple of %u\n",
2421 				param->independ_slice_mode_arg, num_ctb_in_width);
2422 			return false;
2423 		}
2424 	}
2425 
2426 	/* multi-slice & wpp */
2427 	if (param->wpp_enable && param->depend_slice_mode) {
2428 		dev_err(dev, "wpp_enable && depend_slice_mode cannot be used simultaneously\n");
2429 		return false;
2430 	}
2431 
2432 	if (!param->independ_slice_mode && param->depend_slice_mode) {
2433 		dev_err(dev, "depend_slice_mode requires independ_slice_mode\n");
2434 		return false;
2435 	} else if (param->independ_slice_mode &&
2436 		   param->depend_slice_mode == DEPEND_SLICE_MODE_RECOMMENDED &&
2437 		   param->independ_slice_mode_arg < param->depend_slice_mode_arg) {
2438 		dev_err(dev, "independ_slice_mode_arg: %u must be smaller than %u\n",
2439 			param->independ_slice_mode_arg, param->depend_slice_mode_arg);
2440 		return false;
2441 	}
2442 
2443 	if (param->independ_slice_mode && param->independ_slice_mode_arg > 65535) {
2444 		dev_err(dev, "independ_slice_mode_arg: %u must be smaller than 65535\n",
2445 			param->independ_slice_mode_arg);
2446 		return false;
2447 	}
2448 
2449 	if (param->depend_slice_mode && param->depend_slice_mode_arg > 65535) {
2450 		dev_err(dev, "depend_slice_mode_arg: %u must be smaller than 65535\n",
2451 			param->depend_slice_mode_arg);
2452 		return false;
2453 	}
2454 
2455 	if (param->conf_win_top % 2) {
2456 		dev_err(dev, "conf_win_top: %u, must be a multiple of 2\n", param->conf_win_top);
2457 		return false;
2458 	}
2459 
2460 	if (param->conf_win_bot % 2) {
2461 		dev_err(dev, "conf_win_bot: %u, must be a multiple of 2\n", param->conf_win_bot);
2462 		return false;
2463 	}
2464 
2465 	if (param->conf_win_left % 2) {
2466 		dev_err(dev, "conf_win_left: %u, must be a multiple of 2\n", param->conf_win_left);
2467 		return false;
2468 	}
2469 
2470 	if (param->conf_win_right % 2) {
2471 		dev_err(dev, "conf_win_right: %u, Must be a multiple of 2\n",
2472 			param->conf_win_right);
2473 		return false;
2474 	}
2475 
2476 	if (param->lossless_enable && open_param->rc_enable) {
2477 		dev_err(dev, "option rate_control cannot be used with lossless_coding\n");
2478 		return false;
2479 	}
2480 
2481 	if (param->lossless_enable && !param->skip_intra_trans) {
2482 		dev_err(dev, "option intra_trans_skip must be enabled with lossless_coding\n");
2483 		return false;
2484 	}
2485 
2486 	/* intra refresh */
2487 	if (param->intra_refresh_mode && param->intra_refresh_arg == 0) {
2488 		dev_err(dev, "Invalid refresh argument, mode: %u, refresh: %u must be > 0\n",
2489 			param->intra_refresh_mode, param->intra_refresh_arg);
2490 		return false;
2491 	}
2492 	switch (param->intra_refresh_mode) {
2493 	case REFRESH_MODE_CTU_ROWS:
2494 		if (param->intra_mb_refresh_arg > num_ctu_row)
2495 			goto invalid_refresh_argument;
2496 		break;
2497 	case REFRESH_MODE_CTU_COLUMNS:
2498 		if (param->intra_refresh_arg > num_ctu_col)
2499 			goto invalid_refresh_argument;
2500 		break;
2501 	case REFRESH_MODE_CTU_STEP_SIZE:
2502 		if (param->intra_refresh_arg > ctu_sz)
2503 			goto invalid_refresh_argument;
2504 		break;
2505 	case REFRESH_MODE_CTUS:
2506 		if (param->intra_refresh_arg > ctu_sz)
2507 			goto invalid_refresh_argument;
2508 		if (param->lossless_enable) {
2509 			dev_err(dev, "mode: %u cannot be used lossless_enable",
2510 				param->intra_refresh_mode);
2511 			return false;
2512 		}
2513 	}
2514 	return true;
2515 
2516 invalid_refresh_argument:
2517 	dev_err(dev, "Invalid refresh argument, mode: %u, refresh: %u > W(%u)xH(%u)\n",
2518 		param->intra_refresh_mode, param->intra_refresh_arg,
2519 		num_ctu_row, num_ctu_col);
2520 	return false;
2521 }
2522 
wave5_vpu_enc_check_param_valid(struct vpu_device * vpu_dev,struct enc_open_param * open_param)2523 static bool wave5_vpu_enc_check_param_valid(struct vpu_device *vpu_dev,
2524 					    struct enc_open_param *open_param)
2525 {
2526 	struct enc_wave_param *param = &open_param->wave_param;
2527 
2528 	if (open_param->rc_enable) {
2529 		if (param->min_qp_i > param->max_qp_i || param->min_qp_p > param->max_qp_p ||
2530 		    param->min_qp_b > param->max_qp_b) {
2531 			dev_err(vpu_dev->dev, "Configuration failed because min_qp is greater than max_qp\n");
2532 			dev_err(vpu_dev->dev, "Suggested configuration parameters: min_qp = max_qp\n");
2533 			return false;
2534 		}
2535 
2536 		if (open_param->bit_rate <= (int)open_param->frame_rate_info) {
2537 			dev_err(vpu_dev->dev,
2538 				"enc_bit_rate: %u must be greater than the frame_rate: %u\n",
2539 				open_param->bit_rate, (int)open_param->frame_rate_info);
2540 			return false;
2541 		}
2542 	}
2543 
2544 	return true;
2545 }
2546 
wave5_vpu_enc_check_open_param(struct vpu_instance * inst,struct enc_open_param * open_param)2547 int wave5_vpu_enc_check_open_param(struct vpu_instance *inst, struct enc_open_param *open_param)
2548 {
2549 	u32 pic_width;
2550 	u32 pic_height;
2551 	s32 product_id = inst->dev->product;
2552 	struct vpu_attr *p_attr = &inst->dev->attr;
2553 	struct enc_wave_param *param;
2554 
2555 	if (!open_param)
2556 		return -EINVAL;
2557 
2558 	param = &open_param->wave_param;
2559 	pic_width = open_param->pic_width;
2560 	pic_height = open_param->pic_height;
2561 
2562 	if (inst->id >= MAX_NUM_INSTANCE) {
2563 		dev_err(inst->dev->dev, "Too many simultaneous instances: %d (max: %u)\n",
2564 			inst->id, MAX_NUM_INSTANCE);
2565 		return -EOPNOTSUPP;
2566 	}
2567 
2568 	if (inst->std != W_HEVC_ENC &&
2569 	    !(inst->std == W_AVC_ENC && product_id == PRODUCT_ID_521)) {
2570 		dev_err(inst->dev->dev, "Unsupported encoder-codec & product combination\n");
2571 		return -EOPNOTSUPP;
2572 	}
2573 
2574 	if (param->internal_bit_depth == 10) {
2575 		if (inst->std == W_HEVC_ENC && !p_attr->support_hevc10bit_enc) {
2576 			dev_err(inst->dev->dev,
2577 				"Flag support_hevc10bit_enc must be set to encode 10bit HEVC\n");
2578 			return -EOPNOTSUPP;
2579 		} else if (inst->std == W_AVC_ENC && !p_attr->support_avc10bit_enc) {
2580 			dev_err(inst->dev->dev,
2581 				"Flag support_avc10bit_enc must be set to encode 10bit AVC\n");
2582 			return -EOPNOTSUPP;
2583 		}
2584 	}
2585 
2586 	if (!open_param->frame_rate_info) {
2587 		dev_err(inst->dev->dev, "No frame rate information.\n");
2588 		return -EINVAL;
2589 	}
2590 
2591 	if (open_param->bit_rate > MAX_BIT_RATE) {
2592 		dev_err(inst->dev->dev, "Invalid encoding bit-rate: %u (valid: 0-%u)\n",
2593 			open_param->bit_rate, MAX_BIT_RATE);
2594 		return -EINVAL;
2595 	}
2596 
2597 	if (pic_width < W5_MIN_ENC_PIC_WIDTH || pic_width > W5_MAX_ENC_PIC_WIDTH ||
2598 	    pic_height < W5_MIN_ENC_PIC_HEIGHT || pic_height > W5_MAX_ENC_PIC_HEIGHT) {
2599 		dev_err(inst->dev->dev, "Invalid encoding dimension: %ux%u\n",
2600 			pic_width, pic_height);
2601 		return -EINVAL;
2602 	}
2603 
2604 	if (param->profile) {
2605 		if (inst->std == W_HEVC_ENC) {
2606 			if ((param->profile != HEVC_PROFILE_MAIN ||
2607 			     (param->profile == HEVC_PROFILE_MAIN &&
2608 			      param->internal_bit_depth > 8)) &&
2609 			    (param->profile != HEVC_PROFILE_MAIN10 ||
2610 			     (param->profile == HEVC_PROFILE_MAIN10 &&
2611 			      param->internal_bit_depth < 10)) &&
2612 			    param->profile != HEVC_PROFILE_STILLPICTURE) {
2613 				dev_err(inst->dev->dev,
2614 					"Invalid HEVC encoding profile: %u (bit-depth: %u)\n",
2615 					param->profile, param->internal_bit_depth);
2616 				return -EINVAL;
2617 			}
2618 		} else if (inst->std == W_AVC_ENC) {
2619 			if ((param->internal_bit_depth > 8 &&
2620 			     param->profile != H264_PROFILE_HIGH10)) {
2621 				dev_err(inst->dev->dev,
2622 					"Invalid AVC encoding profile: %u (bit-depth: %u)\n",
2623 					param->profile, param->internal_bit_depth);
2624 				return -EINVAL;
2625 			}
2626 		}
2627 	}
2628 
2629 	if (param->decoding_refresh_type > DEC_REFRESH_TYPE_IDR) {
2630 		dev_err(inst->dev->dev, "Invalid decoding refresh type: %u (valid: 0-2)\n",
2631 			param->decoding_refresh_type);
2632 		return -EINVAL;
2633 	}
2634 
2635 	if (param->intra_refresh_mode > REFRESH_MODE_CTUS) {
2636 		dev_err(inst->dev->dev, "Invalid intra refresh mode: %d (valid: 0-4)\n",
2637 			param->intra_refresh_mode);
2638 		return -EINVAL;
2639 	}
2640 
2641 	if (inst->std == W_HEVC_ENC && param->independ_slice_mode &&
2642 	    param->depend_slice_mode > DEPEND_SLICE_MODE_BOOST) {
2643 		dev_err(inst->dev->dev,
2644 			"Can't combine slice modes: independent and fast dependent for HEVC\n");
2645 		return -EINVAL;
2646 	}
2647 
2648 	if (!param->disable_deblk) {
2649 		if (param->beta_offset_div2 < -6 || param->beta_offset_div2 > 6) {
2650 			dev_err(inst->dev->dev, "Invalid beta offset: %d (valid: -6-6)\n",
2651 				param->beta_offset_div2);
2652 			return -EINVAL;
2653 		}
2654 
2655 		if (param->tc_offset_div2 < -6 || param->tc_offset_div2 > 6) {
2656 			dev_err(inst->dev->dev, "Invalid tc offset: %d (valid: -6-6)\n",
2657 				param->tc_offset_div2);
2658 			return -EINVAL;
2659 		}
2660 	}
2661 
2662 	if (param->intra_qp > MAX_INTRA_QP) {
2663 		dev_err(inst->dev->dev,
2664 			"Invalid intra quantization parameter: %u (valid: 0-%u)\n",
2665 			param->intra_qp, MAX_INTRA_QP);
2666 		return -EINVAL;
2667 	}
2668 
2669 	if (open_param->rc_enable) {
2670 		if (param->min_qp_i > MAX_INTRA_QP || param->max_qp_i > MAX_INTRA_QP ||
2671 		    param->min_qp_p > MAX_INTRA_QP || param->max_qp_p > MAX_INTRA_QP ||
2672 		    param->min_qp_b > MAX_INTRA_QP || param->max_qp_b > MAX_INTRA_QP) {
2673 			dev_err(inst->dev->dev,
2674 				"Invalid quantization parameter min/max values: "
2675 				"I: %u-%u, P: %u-%u, B: %u-%u (valid for each: 0-%u)\n",
2676 				param->min_qp_i, param->max_qp_i, param->min_qp_p, param->max_qp_p,
2677 				param->min_qp_b, param->max_qp_b, MAX_INTRA_QP);
2678 			return -EINVAL;
2679 		}
2680 
2681 		if (param->hvs_qp_enable && param->hvs_max_delta_qp > MAX_HVS_MAX_DELTA_QP) {
2682 			dev_err(inst->dev->dev,
2683 				"Invalid HVS max delta quantization parameter: %u (valid: 0-%u)\n",
2684 				param->hvs_max_delta_qp, MAX_HVS_MAX_DELTA_QP);
2685 			return -EINVAL;
2686 		}
2687 
2688 		if (open_param->vbv_buffer_size < MIN_VBV_BUFFER_SIZE ||
2689 		    open_param->vbv_buffer_size > MAX_VBV_BUFFER_SIZE) {
2690 			dev_err(inst->dev->dev, "VBV buffer size: %u (valid: %u-%u)\n",
2691 				open_param->vbv_buffer_size, MIN_VBV_BUFFER_SIZE,
2692 				MAX_VBV_BUFFER_SIZE);
2693 			return -EINVAL;
2694 		}
2695 	}
2696 
2697 	if (!wave5_vpu_enc_check_common_param_valid(inst, open_param))
2698 		return -EINVAL;
2699 
2700 	if (!wave5_vpu_enc_check_param_valid(inst->dev, open_param))
2701 		return -EINVAL;
2702 
2703 	if (param->chroma_cb_qp_offset < -12 || param->chroma_cb_qp_offset > 12) {
2704 		dev_err(inst->dev->dev,
2705 			"Invalid chroma Cb quantization parameter offset: %d (valid: -12-12)\n",
2706 			param->chroma_cb_qp_offset);
2707 		return -EINVAL;
2708 	}
2709 
2710 	if (param->chroma_cr_qp_offset < -12 || param->chroma_cr_qp_offset > 12) {
2711 		dev_err(inst->dev->dev,
2712 			"Invalid chroma Cr quantization parameter offset: %d (valid: -12-12)\n",
2713 			param->chroma_cr_qp_offset);
2714 		return -EINVAL;
2715 	}
2716 
2717 	if (param->intra_refresh_mode == REFRESH_MODE_CTU_STEP_SIZE && !param->intra_refresh_arg) {
2718 		dev_err(inst->dev->dev,
2719 			"Intra refresh mode CTU step-size requires an argument\n");
2720 		return -EINVAL;
2721 	}
2722 
2723 	if (inst->std == W_HEVC_ENC) {
2724 		if (param->nr_intra_weight_y > MAX_INTRA_WEIGHT ||
2725 		    param->nr_intra_weight_cb > MAX_INTRA_WEIGHT ||
2726 		    param->nr_intra_weight_cr > MAX_INTRA_WEIGHT) {
2727 			dev_err(inst->dev->dev,
2728 				"Invalid intra weight Y(%u) Cb(%u) Cr(%u) (valid: %u)\n",
2729 				param->nr_intra_weight_y, param->nr_intra_weight_cb,
2730 				param->nr_intra_weight_cr, MAX_INTRA_WEIGHT);
2731 			return -EINVAL;
2732 		}
2733 
2734 		if (param->nr_inter_weight_y > MAX_INTER_WEIGHT ||
2735 		    param->nr_inter_weight_cb > MAX_INTER_WEIGHT ||
2736 		    param->nr_inter_weight_cr > MAX_INTER_WEIGHT) {
2737 			dev_err(inst->dev->dev,
2738 				"Invalid inter weight Y(%u) Cb(%u) Cr(%u) (valid: %u)\n",
2739 				param->nr_inter_weight_y, param->nr_inter_weight_cb,
2740 				param->nr_inter_weight_cr, MAX_INTER_WEIGHT);
2741 			return -EINVAL;
2742 		}
2743 	}
2744 
2745 	return 0;
2746 }
2747