1  // SPDX-License-Identifier: ISC
2  /*
3   * Copyright (c) 2005-2011 Atheros Communications Inc.
4   * Copyright (c) 2011-2017 Qualcomm Atheros, Inc.
5   * Copyright (c) 2022 Qualcomm Innovation Center, Inc. All rights reserved.
6   */
7  
8  #include "core.h"
9  #include "hif.h"
10  #include "debug.h"
11  
12  /********/
13  /* Send */
14  /********/
15  
ath10k_htc_control_tx_complete(struct ath10k * ar,struct sk_buff * skb)16  static void ath10k_htc_control_tx_complete(struct ath10k *ar,
17  					   struct sk_buff *skb)
18  {
19  	kfree_skb(skb);
20  }
21  
ath10k_htc_build_tx_ctrl_skb(void * ar)22  static struct sk_buff *ath10k_htc_build_tx_ctrl_skb(void *ar)
23  {
24  	struct sk_buff *skb;
25  	struct ath10k_skb_cb *skb_cb;
26  
27  	skb = dev_alloc_skb(ATH10K_HTC_CONTROL_BUFFER_SIZE);
28  	if (!skb)
29  		return NULL;
30  
31  	skb_reserve(skb, 20); /* FIXME: why 20 bytes? */
32  	WARN_ONCE((unsigned long)skb->data & 3, "unaligned skb");
33  
34  	skb_cb = ATH10K_SKB_CB(skb);
35  	memset(skb_cb, 0, sizeof(*skb_cb));
36  
37  	ath10k_dbg(ar, ATH10K_DBG_HTC, "%s: skb %pK\n", __func__, skb);
38  	return skb;
39  }
40  
ath10k_htc_restore_tx_skb(struct ath10k_htc * htc,struct sk_buff * skb)41  static inline void ath10k_htc_restore_tx_skb(struct ath10k_htc *htc,
42  					     struct sk_buff *skb)
43  {
44  	struct ath10k_skb_cb *skb_cb = ATH10K_SKB_CB(skb);
45  
46  	if (htc->ar->bus_param.dev_type != ATH10K_DEV_TYPE_HL)
47  		dma_unmap_single(htc->ar->dev, skb_cb->paddr, skb->len, DMA_TO_DEVICE);
48  	skb_pull(skb, sizeof(struct ath10k_htc_hdr));
49  }
50  
ath10k_htc_notify_tx_completion(struct ath10k_htc_ep * ep,struct sk_buff * skb)51  void ath10k_htc_notify_tx_completion(struct ath10k_htc_ep *ep,
52  				     struct sk_buff *skb)
53  {
54  	struct ath10k *ar = ep->htc->ar;
55  	struct ath10k_htc_hdr *hdr;
56  
57  	ath10k_dbg(ar, ATH10K_DBG_HTC, "%s: ep %d skb %pK\n", __func__,
58  		   ep->eid, skb);
59  
60  	/* A corner case where the copy completion is reaching to host but still
61  	 * copy engine is processing it due to which host unmaps corresponding
62  	 * memory and causes SMMU fault, hence as workaround adding delay
63  	 * the unmapping memory to avoid SMMU faults.
64  	 */
65  	if (ar->hw_params.delay_unmap_buffer &&
66  	    ep->ul_pipe_id == 3)
67  		mdelay(2);
68  
69  	hdr = (struct ath10k_htc_hdr *)skb->data;
70  	ath10k_htc_restore_tx_skb(ep->htc, skb);
71  
72  	if (!ep->ep_ops.ep_tx_complete) {
73  		ath10k_warn(ar, "no tx handler for eid %d\n", ep->eid);
74  		dev_kfree_skb_any(skb);
75  		return;
76  	}
77  
78  	if (hdr->flags & ATH10K_HTC_FLAG_SEND_BUNDLE) {
79  		dev_kfree_skb_any(skb);
80  		return;
81  	}
82  
83  	ep->ep_ops.ep_tx_complete(ep->htc->ar, skb);
84  }
85  EXPORT_SYMBOL(ath10k_htc_notify_tx_completion);
86  
ath10k_htc_prepare_tx_skb(struct ath10k_htc_ep * ep,struct sk_buff * skb)87  static void ath10k_htc_prepare_tx_skb(struct ath10k_htc_ep *ep,
88  				      struct sk_buff *skb)
89  {
90  	struct ath10k_htc_hdr *hdr;
91  
92  	hdr = (struct ath10k_htc_hdr *)skb->data;
93  	memset(hdr, 0, sizeof(struct ath10k_htc_hdr));
94  
95  	hdr->eid = ep->eid;
96  	hdr->len = __cpu_to_le16(skb->len - sizeof(*hdr));
97  	hdr->flags = 0;
98  	if (ep->tx_credit_flow_enabled && !ep->bundle_tx)
99  		hdr->flags |= ATH10K_HTC_FLAG_NEED_CREDIT_UPDATE;
100  
101  	spin_lock_bh(&ep->htc->tx_lock);
102  	hdr->seq_no = ep->seq_no++;
103  	spin_unlock_bh(&ep->htc->tx_lock);
104  }
105  
ath10k_htc_consume_credit(struct ath10k_htc_ep * ep,unsigned int len,bool consume)106  static int ath10k_htc_consume_credit(struct ath10k_htc_ep *ep,
107  				     unsigned int len,
108  				     bool consume)
109  {
110  	struct ath10k_htc *htc = ep->htc;
111  	struct ath10k *ar = htc->ar;
112  	enum ath10k_htc_ep_id eid = ep->eid;
113  	int credits, ret = 0;
114  
115  	if (!ep->tx_credit_flow_enabled)
116  		return 0;
117  
118  	credits = DIV_ROUND_UP(len, ep->tx_credit_size);
119  	spin_lock_bh(&htc->tx_lock);
120  
121  	if (ep->tx_credits < credits) {
122  		ath10k_dbg(ar, ATH10K_DBG_HTC,
123  			   "htc insufficient credits ep %d required %d available %d consume %d\n",
124  			   eid, credits, ep->tx_credits, consume);
125  		ret = -EAGAIN;
126  		goto unlock;
127  	}
128  
129  	if (consume) {
130  		ep->tx_credits -= credits;
131  		ath10k_dbg(ar, ATH10K_DBG_HTC,
132  			   "htc ep %d consumed %d credits total %d\n",
133  			   eid, credits, ep->tx_credits);
134  	}
135  
136  unlock:
137  	spin_unlock_bh(&htc->tx_lock);
138  	return ret;
139  }
140  
ath10k_htc_release_credit(struct ath10k_htc_ep * ep,unsigned int len)141  static void ath10k_htc_release_credit(struct ath10k_htc_ep *ep, unsigned int len)
142  {
143  	struct ath10k_htc *htc = ep->htc;
144  	struct ath10k *ar = htc->ar;
145  	enum ath10k_htc_ep_id eid = ep->eid;
146  	int credits;
147  
148  	if (!ep->tx_credit_flow_enabled)
149  		return;
150  
151  	credits = DIV_ROUND_UP(len, ep->tx_credit_size);
152  	spin_lock_bh(&htc->tx_lock);
153  	ep->tx_credits += credits;
154  	ath10k_dbg(ar, ATH10K_DBG_HTC,
155  		   "htc ep %d reverted %d credits back total %d\n",
156  		   eid, credits, ep->tx_credits);
157  	spin_unlock_bh(&htc->tx_lock);
158  
159  	if (ep->ep_ops.ep_tx_credits)
160  		ep->ep_ops.ep_tx_credits(htc->ar);
161  }
162  
ath10k_htc_send(struct ath10k_htc * htc,enum ath10k_htc_ep_id eid,struct sk_buff * skb)163  int ath10k_htc_send(struct ath10k_htc *htc,
164  		    enum ath10k_htc_ep_id eid,
165  		    struct sk_buff *skb)
166  {
167  	struct ath10k *ar = htc->ar;
168  	struct ath10k_htc_ep *ep = &htc->endpoint[eid];
169  	struct ath10k_skb_cb *skb_cb = ATH10K_SKB_CB(skb);
170  	struct ath10k_hif_sg_item sg_item;
171  	struct device *dev = htc->ar->dev;
172  	int ret;
173  	unsigned int skb_len;
174  
175  	if (htc->ar->state == ATH10K_STATE_WEDGED)
176  		return -ECOMM;
177  
178  	if (eid >= ATH10K_HTC_EP_COUNT) {
179  		ath10k_warn(ar, "Invalid endpoint id: %d\n", eid);
180  		return -ENOENT;
181  	}
182  
183  	skb_push(skb, sizeof(struct ath10k_htc_hdr));
184  
185  	skb_len = skb->len;
186  	ret = ath10k_htc_consume_credit(ep, skb_len, true);
187  	if (ret)
188  		goto err_pull;
189  
190  	ath10k_htc_prepare_tx_skb(ep, skb);
191  
192  	skb_cb->eid = eid;
193  	if (ar->bus_param.dev_type != ATH10K_DEV_TYPE_HL) {
194  		skb_cb->paddr = dma_map_single(dev, skb->data, skb->len,
195  					       DMA_TO_DEVICE);
196  		ret = dma_mapping_error(dev, skb_cb->paddr);
197  		if (ret) {
198  			ret = -EIO;
199  			goto err_credits;
200  		}
201  	}
202  
203  	sg_item.transfer_id = ep->eid;
204  	sg_item.transfer_context = skb;
205  	sg_item.vaddr = skb->data;
206  	sg_item.paddr = skb_cb->paddr;
207  	sg_item.len = skb->len;
208  
209  	ret = ath10k_hif_tx_sg(htc->ar, ep->ul_pipe_id, &sg_item, 1);
210  	if (ret)
211  		goto err_unmap;
212  
213  	return 0;
214  
215  err_unmap:
216  	if (ar->bus_param.dev_type != ATH10K_DEV_TYPE_HL)
217  		dma_unmap_single(dev, skb_cb->paddr, skb->len, DMA_TO_DEVICE);
218  err_credits:
219  	ath10k_htc_release_credit(ep, skb_len);
220  err_pull:
221  	skb_pull(skb, sizeof(struct ath10k_htc_hdr));
222  	return ret;
223  }
224  
ath10k_htc_tx_completion_handler(struct ath10k * ar,struct sk_buff * skb)225  void ath10k_htc_tx_completion_handler(struct ath10k *ar, struct sk_buff *skb)
226  {
227  	struct ath10k_htc *htc = &ar->htc;
228  	struct ath10k_skb_cb *skb_cb;
229  	struct ath10k_htc_ep *ep;
230  
231  	if (WARN_ON_ONCE(!skb))
232  		return;
233  
234  	skb_cb = ATH10K_SKB_CB(skb);
235  	ep = &htc->endpoint[skb_cb->eid];
236  
237  	ath10k_htc_notify_tx_completion(ep, skb);
238  	/* the skb now belongs to the completion handler */
239  }
240  EXPORT_SYMBOL(ath10k_htc_tx_completion_handler);
241  
242  /***********/
243  /* Receive */
244  /***********/
245  
246  static void
ath10k_htc_process_credit_report(struct ath10k_htc * htc,const struct ath10k_htc_credit_report * report,int len,enum ath10k_htc_ep_id eid)247  ath10k_htc_process_credit_report(struct ath10k_htc *htc,
248  				 const struct ath10k_htc_credit_report *report,
249  				 int len,
250  				 enum ath10k_htc_ep_id eid)
251  {
252  	struct ath10k *ar = htc->ar;
253  	struct ath10k_htc_ep *ep;
254  	int i, n_reports;
255  
256  	if (len % sizeof(*report))
257  		ath10k_warn(ar, "Uneven credit report len %d", len);
258  
259  	n_reports = len / sizeof(*report);
260  
261  	spin_lock_bh(&htc->tx_lock);
262  	for (i = 0; i < n_reports; i++, report++) {
263  		if (report->eid >= ATH10K_HTC_EP_COUNT)
264  			break;
265  
266  		ep = &htc->endpoint[report->eid];
267  		ep->tx_credits += report->credits;
268  
269  		ath10k_dbg(ar, ATH10K_DBG_HTC, "htc ep %d got %d credits (total %d)\n",
270  			   report->eid, report->credits, ep->tx_credits);
271  
272  		if (ep->ep_ops.ep_tx_credits) {
273  			spin_unlock_bh(&htc->tx_lock);
274  			ep->ep_ops.ep_tx_credits(htc->ar);
275  			spin_lock_bh(&htc->tx_lock);
276  		}
277  	}
278  	spin_unlock_bh(&htc->tx_lock);
279  }
280  
281  static int
ath10k_htc_process_lookahead(struct ath10k_htc * htc,const struct ath10k_htc_lookahead_report * report,int len,enum ath10k_htc_ep_id eid,void * next_lookaheads,int * next_lookaheads_len)282  ath10k_htc_process_lookahead(struct ath10k_htc *htc,
283  			     const struct ath10k_htc_lookahead_report *report,
284  			     int len,
285  			     enum ath10k_htc_ep_id eid,
286  			     void *next_lookaheads,
287  			     int *next_lookaheads_len)
288  {
289  	struct ath10k *ar = htc->ar;
290  
291  	/* Invalid lookahead flags are actually transmitted by
292  	 * the target in the HTC control message.
293  	 * Since this will happen at every boot we silently ignore
294  	 * the lookahead in this case
295  	 */
296  	if (report->pre_valid != ((~report->post_valid) & 0xFF))
297  		return 0;
298  
299  	if (next_lookaheads && next_lookaheads_len) {
300  		ath10k_dbg(ar, ATH10K_DBG_HTC,
301  			   "htc rx lookahead found pre_valid 0x%x post_valid 0x%x\n",
302  			   report->pre_valid, report->post_valid);
303  
304  		/* look ahead bytes are valid, copy them over */
305  		memcpy((u8 *)next_lookaheads, report->lookahead, 4);
306  
307  		*next_lookaheads_len = 1;
308  	}
309  
310  	return 0;
311  }
312  
313  static int
ath10k_htc_process_lookahead_bundle(struct ath10k_htc * htc,const struct ath10k_htc_lookahead_bundle * report,int len,enum ath10k_htc_ep_id eid,void * next_lookaheads,int * next_lookaheads_len)314  ath10k_htc_process_lookahead_bundle(struct ath10k_htc *htc,
315  				    const struct ath10k_htc_lookahead_bundle *report,
316  				    int len,
317  				    enum ath10k_htc_ep_id eid,
318  				    void *next_lookaheads,
319  				    int *next_lookaheads_len)
320  {
321  	struct ath10k *ar = htc->ar;
322  	int bundle_cnt = len / sizeof(*report);
323  
324  	if (!bundle_cnt || (bundle_cnt > htc->max_msgs_per_htc_bundle)) {
325  		ath10k_warn(ar, "Invalid lookahead bundle count: %d\n",
326  			    bundle_cnt);
327  		return -EINVAL;
328  	}
329  
330  	if (next_lookaheads && next_lookaheads_len) {
331  		int i;
332  
333  		for (i = 0; i < bundle_cnt; i++) {
334  			memcpy(((u8 *)next_lookaheads) + 4 * i,
335  			       report->lookahead, 4);
336  			report++;
337  		}
338  
339  		*next_lookaheads_len = bundle_cnt;
340  	}
341  
342  	return 0;
343  }
344  
ath10k_htc_process_trailer(struct ath10k_htc * htc,u8 * buffer,int length,enum ath10k_htc_ep_id src_eid,void * next_lookaheads,int * next_lookaheads_len)345  int ath10k_htc_process_trailer(struct ath10k_htc *htc,
346  			       u8 *buffer,
347  			       int length,
348  			       enum ath10k_htc_ep_id src_eid,
349  			       void *next_lookaheads,
350  			       int *next_lookaheads_len)
351  {
352  	struct ath10k_htc_lookahead_bundle *bundle;
353  	struct ath10k *ar = htc->ar;
354  	int status = 0;
355  	struct ath10k_htc_record *record;
356  	u8 *orig_buffer;
357  	int orig_length;
358  	size_t len;
359  
360  	orig_buffer = buffer;
361  	orig_length = length;
362  
363  	while (length > 0) {
364  		record = (struct ath10k_htc_record *)buffer;
365  
366  		if (length < sizeof(record->hdr)) {
367  			status = -EINVAL;
368  			break;
369  		}
370  
371  		if (record->hdr.len > length) {
372  			/* no room left in buffer for record */
373  			ath10k_warn(ar, "Invalid record length: %d\n",
374  				    record->hdr.len);
375  			status = -EINVAL;
376  			break;
377  		}
378  
379  		switch (record->hdr.id) {
380  		case ATH10K_HTC_RECORD_CREDITS:
381  			len = sizeof(struct ath10k_htc_credit_report);
382  			if (record->hdr.len < len) {
383  				ath10k_warn(ar, "Credit report too long\n");
384  				status = -EINVAL;
385  				break;
386  			}
387  			ath10k_htc_process_credit_report(htc,
388  							 record->credit_report,
389  							 record->hdr.len,
390  							 src_eid);
391  			break;
392  		case ATH10K_HTC_RECORD_LOOKAHEAD:
393  			len = sizeof(struct ath10k_htc_lookahead_report);
394  			if (record->hdr.len < len) {
395  				ath10k_warn(ar, "Lookahead report too long\n");
396  				status = -EINVAL;
397  				break;
398  			}
399  			status = ath10k_htc_process_lookahead(htc,
400  							      record->lookahead_report,
401  							      record->hdr.len,
402  							      src_eid,
403  							      next_lookaheads,
404  							      next_lookaheads_len);
405  			break;
406  		case ATH10K_HTC_RECORD_LOOKAHEAD_BUNDLE:
407  			bundle = record->lookahead_bundle;
408  			status = ath10k_htc_process_lookahead_bundle(htc,
409  								     bundle,
410  								     record->hdr.len,
411  								     src_eid,
412  								     next_lookaheads,
413  								     next_lookaheads_len);
414  			break;
415  		default:
416  			ath10k_warn(ar, "Unhandled record: id:%d length:%d\n",
417  				    record->hdr.id, record->hdr.len);
418  			break;
419  		}
420  
421  		if (status)
422  			break;
423  
424  		/* multiple records may be present in a trailer */
425  		buffer += sizeof(record->hdr) + record->hdr.len;
426  		length -= sizeof(record->hdr) + record->hdr.len;
427  	}
428  
429  	if (status)
430  		ath10k_dbg_dump(ar, ATH10K_DBG_HTC, "htc rx bad trailer", "",
431  				orig_buffer, orig_length);
432  
433  	return status;
434  }
435  EXPORT_SYMBOL(ath10k_htc_process_trailer);
436  
ath10k_htc_rx_completion_handler(struct ath10k * ar,struct sk_buff * skb)437  void ath10k_htc_rx_completion_handler(struct ath10k *ar, struct sk_buff *skb)
438  {
439  	int status = 0;
440  	struct ath10k_htc *htc = &ar->htc;
441  	struct ath10k_htc_hdr *hdr;
442  	struct ath10k_htc_ep *ep;
443  	u16 payload_len;
444  	u32 trailer_len = 0;
445  	size_t min_len;
446  	u8 eid;
447  	bool trailer_present;
448  
449  	hdr = (struct ath10k_htc_hdr *)skb->data;
450  	skb_pull(skb, sizeof(*hdr));
451  
452  	eid = hdr->eid;
453  
454  	if (eid >= ATH10K_HTC_EP_COUNT) {
455  		ath10k_warn(ar, "HTC Rx: invalid eid %d\n", eid);
456  		ath10k_dbg_dump(ar, ATH10K_DBG_HTC, "htc bad header", "",
457  				hdr, sizeof(*hdr));
458  		goto out;
459  	}
460  
461  	ep = &htc->endpoint[eid];
462  	if (ep->service_id == ATH10K_HTC_SVC_ID_UNUSED) {
463  		ath10k_warn(ar, "htc rx endpoint %d is not connected\n", eid);
464  		goto out;
465  	}
466  
467  	payload_len = __le16_to_cpu(hdr->len);
468  
469  	if (payload_len + sizeof(*hdr) > ATH10K_HTC_MAX_LEN) {
470  		ath10k_warn(ar, "HTC rx frame too long, len: %zu\n",
471  			    payload_len + sizeof(*hdr));
472  		ath10k_dbg_dump(ar, ATH10K_DBG_HTC, "htc bad rx pkt len", "",
473  				hdr, sizeof(*hdr));
474  		goto out;
475  	}
476  
477  	if (skb->len < payload_len) {
478  		ath10k_dbg(ar, ATH10K_DBG_HTC,
479  			   "HTC Rx: insufficient length, got %d, expected %d\n",
480  			   skb->len, payload_len);
481  		ath10k_dbg_dump(ar, ATH10K_DBG_HTC, "htc bad rx pkt len",
482  				"", hdr, sizeof(*hdr));
483  		goto out;
484  	}
485  
486  	/* get flags to check for trailer */
487  	trailer_present = hdr->flags & ATH10K_HTC_FLAG_TRAILER_PRESENT;
488  	if (trailer_present) {
489  		u8 *trailer;
490  
491  		trailer_len = hdr->trailer_len;
492  		min_len = sizeof(struct ath10k_ath10k_htc_record_hdr);
493  
494  		if ((trailer_len < min_len) ||
495  		    (trailer_len > payload_len)) {
496  			ath10k_warn(ar, "Invalid trailer length: %d\n",
497  				    trailer_len);
498  			goto out;
499  		}
500  
501  		trailer = (u8 *)hdr;
502  		trailer += sizeof(*hdr);
503  		trailer += payload_len;
504  		trailer -= trailer_len;
505  		status = ath10k_htc_process_trailer(htc, trailer,
506  						    trailer_len, hdr->eid,
507  						    NULL, NULL);
508  		if (status)
509  			goto out;
510  
511  		skb_trim(skb, skb->len - trailer_len);
512  	}
513  
514  	if (((int)payload_len - (int)trailer_len) <= 0)
515  		/* zero length packet with trailer data, just drop these */
516  		goto out;
517  
518  	ath10k_dbg(ar, ATH10K_DBG_HTC, "htc rx completion ep %d skb %pK\n",
519  		   eid, skb);
520  	ep->ep_ops.ep_rx_complete(ar, skb);
521  
522  	/* skb is now owned by the rx completion handler */
523  	skb = NULL;
524  out:
525  	kfree_skb(skb);
526  }
527  EXPORT_SYMBOL(ath10k_htc_rx_completion_handler);
528  
ath10k_htc_control_rx_complete(struct ath10k * ar,struct sk_buff * skb)529  static void ath10k_htc_control_rx_complete(struct ath10k *ar,
530  					   struct sk_buff *skb)
531  {
532  	struct ath10k_htc *htc = &ar->htc;
533  	struct ath10k_htc_msg *msg = (struct ath10k_htc_msg *)skb->data;
534  
535  	switch (__le16_to_cpu(msg->hdr.message_id)) {
536  	case ATH10K_HTC_MSG_READY_ID:
537  	case ATH10K_HTC_MSG_CONNECT_SERVICE_RESP_ID:
538  		/* handle HTC control message */
539  		if (completion_done(&htc->ctl_resp)) {
540  			/* this is a fatal error, target should not be
541  			 * sending unsolicited messages on the ep 0
542  			 */
543  			ath10k_warn(ar, "HTC rx ctrl still processing\n");
544  			complete(&htc->ctl_resp);
545  			goto out;
546  		}
547  
548  		htc->control_resp_len =
549  			min_t(int, skb->len,
550  			      ATH10K_HTC_MAX_CTRL_MSG_LEN);
551  
552  		memcpy(htc->control_resp_buffer, skb->data,
553  		       htc->control_resp_len);
554  
555  		complete(&htc->ctl_resp);
556  		break;
557  	case ATH10K_HTC_MSG_SEND_SUSPEND_COMPLETE:
558  		htc->htc_ops.target_send_suspend_complete(ar);
559  		break;
560  	default:
561  		ath10k_warn(ar, "ignoring unsolicited htc ep0 event\n");
562  		break;
563  	}
564  
565  out:
566  	kfree_skb(skb);
567  }
568  
569  /***************/
570  /* Init/Deinit */
571  /***************/
572  
htc_service_name(enum ath10k_htc_svc_id id)573  static const char *htc_service_name(enum ath10k_htc_svc_id id)
574  {
575  	switch (id) {
576  	case ATH10K_HTC_SVC_ID_RESERVED:
577  		return "Reserved";
578  	case ATH10K_HTC_SVC_ID_RSVD_CTRL:
579  		return "Control";
580  	case ATH10K_HTC_SVC_ID_WMI_CONTROL:
581  		return "WMI";
582  	case ATH10K_HTC_SVC_ID_WMI_DATA_BE:
583  		return "DATA BE";
584  	case ATH10K_HTC_SVC_ID_WMI_DATA_BK:
585  		return "DATA BK";
586  	case ATH10K_HTC_SVC_ID_WMI_DATA_VI:
587  		return "DATA VI";
588  	case ATH10K_HTC_SVC_ID_WMI_DATA_VO:
589  		return "DATA VO";
590  	case ATH10K_HTC_SVC_ID_NMI_CONTROL:
591  		return "NMI Control";
592  	case ATH10K_HTC_SVC_ID_NMI_DATA:
593  		return "NMI Data";
594  	case ATH10K_HTC_SVC_ID_HTT_DATA_MSG:
595  		return "HTT Data";
596  	case ATH10K_HTC_SVC_ID_HTT_DATA2_MSG:
597  		return "HTT Data";
598  	case ATH10K_HTC_SVC_ID_HTT_DATA3_MSG:
599  		return "HTT Data";
600  	case ATH10K_HTC_SVC_ID_TEST_RAW_STREAMS:
601  		return "RAW";
602  	case ATH10K_HTC_SVC_ID_HTT_LOG_MSG:
603  		return "PKTLOG";
604  	}
605  
606  	return "Unknown";
607  }
608  
ath10k_htc_reset_endpoint_states(struct ath10k_htc * htc)609  static void ath10k_htc_reset_endpoint_states(struct ath10k_htc *htc)
610  {
611  	struct ath10k_htc_ep *ep;
612  	int i;
613  
614  	for (i = ATH10K_HTC_EP_0; i < ATH10K_HTC_EP_COUNT; i++) {
615  		ep = &htc->endpoint[i];
616  		ep->service_id = ATH10K_HTC_SVC_ID_UNUSED;
617  		ep->max_ep_message_len = 0;
618  		ep->max_tx_queue_depth = 0;
619  		ep->eid = i;
620  		ep->htc = htc;
621  		ep->tx_credit_flow_enabled = true;
622  	}
623  }
624  
ath10k_htc_get_credit_allocation(struct ath10k_htc * htc,u16 service_id)625  static u8 ath10k_htc_get_credit_allocation(struct ath10k_htc *htc,
626  					   u16 service_id)
627  {
628  	u8 allocation = 0;
629  
630  	/* The WMI control service is the only service with flow control.
631  	 * Let it have all transmit credits.
632  	 */
633  	if (service_id == ATH10K_HTC_SVC_ID_WMI_CONTROL)
634  		allocation = htc->total_transmit_credits;
635  
636  	return allocation;
637  }
638  
ath10k_htc_send_bundle(struct ath10k_htc_ep * ep,struct sk_buff * bundle_skb,struct sk_buff_head * tx_save_head)639  static int ath10k_htc_send_bundle(struct ath10k_htc_ep *ep,
640  				  struct sk_buff *bundle_skb,
641  				  struct sk_buff_head *tx_save_head)
642  {
643  	struct ath10k_hif_sg_item sg_item;
644  	struct ath10k_htc *htc = ep->htc;
645  	struct ath10k *ar = htc->ar;
646  	struct sk_buff *skb;
647  	int ret, cn = 0;
648  	unsigned int skb_len;
649  
650  	ath10k_dbg(ar, ATH10K_DBG_HTC, "bundle skb len %d\n", bundle_skb->len);
651  	skb_len = bundle_skb->len;
652  	ret = ath10k_htc_consume_credit(ep, skb_len, true);
653  
654  	if (!ret) {
655  		sg_item.transfer_id = ep->eid;
656  		sg_item.transfer_context = bundle_skb;
657  		sg_item.vaddr = bundle_skb->data;
658  		sg_item.len = bundle_skb->len;
659  
660  		ret = ath10k_hif_tx_sg(htc->ar, ep->ul_pipe_id, &sg_item, 1);
661  		if (ret)
662  			ath10k_htc_release_credit(ep, skb_len);
663  	}
664  
665  	if (ret)
666  		dev_kfree_skb_any(bundle_skb);
667  
668  	for (cn = 0; (skb = skb_dequeue_tail(tx_save_head)); cn++) {
669  		if (ret) {
670  			skb_pull(skb, sizeof(struct ath10k_htc_hdr));
671  			skb_queue_head(&ep->tx_req_head, skb);
672  		} else {
673  			skb_queue_tail(&ep->tx_complete_head, skb);
674  		}
675  	}
676  
677  	if (!ret)
678  		queue_work(ar->workqueue_tx_complete, &ar->tx_complete_work);
679  
680  	ath10k_dbg(ar, ATH10K_DBG_HTC,
681  		   "bundle tx status %d eid %d req count %d count %d len %d\n",
682  		   ret, ep->eid, skb_queue_len(&ep->tx_req_head), cn, skb_len);
683  	return ret;
684  }
685  
ath10k_htc_send_one_skb(struct ath10k_htc_ep * ep,struct sk_buff * skb)686  static void ath10k_htc_send_one_skb(struct ath10k_htc_ep *ep, struct sk_buff *skb)
687  {
688  	struct ath10k_htc *htc = ep->htc;
689  	struct ath10k *ar = htc->ar;
690  	int ret;
691  
692  	ret = ath10k_htc_send(htc, ep->eid, skb);
693  
694  	if (ret)
695  		skb_queue_head(&ep->tx_req_head, skb);
696  
697  	ath10k_dbg(ar, ATH10K_DBG_HTC, "tx one status %d eid %d len %d pending count %d\n",
698  		   ret, ep->eid, skb->len, skb_queue_len(&ep->tx_req_head));
699  }
700  
ath10k_htc_send_bundle_skbs(struct ath10k_htc_ep * ep)701  static int ath10k_htc_send_bundle_skbs(struct ath10k_htc_ep *ep)
702  {
703  	struct ath10k_htc *htc = ep->htc;
704  	struct sk_buff *bundle_skb, *skb;
705  	struct sk_buff_head tx_save_head;
706  	struct ath10k_htc_hdr *hdr;
707  	u8 *bundle_buf;
708  	int ret = 0, credit_pad, credit_remainder, trans_len, bundles_left = 0;
709  
710  	if (htc->ar->state == ATH10K_STATE_WEDGED)
711  		return -ECOMM;
712  
713  	if (ep->tx_credit_flow_enabled &&
714  	    ep->tx_credits < ATH10K_MIN_CREDIT_PER_HTC_TX_BUNDLE)
715  		return 0;
716  
717  	bundles_left = ATH10K_MAX_MSG_PER_HTC_TX_BUNDLE * ep->tx_credit_size;
718  	bundle_skb = dev_alloc_skb(bundles_left);
719  
720  	if (!bundle_skb)
721  		return -ENOMEM;
722  
723  	bundle_buf = bundle_skb->data;
724  	skb_queue_head_init(&tx_save_head);
725  
726  	while (true) {
727  		skb = skb_dequeue(&ep->tx_req_head);
728  		if (!skb)
729  			break;
730  
731  		credit_pad = 0;
732  		trans_len = skb->len + sizeof(*hdr);
733  		credit_remainder = trans_len % ep->tx_credit_size;
734  
735  		if (credit_remainder != 0) {
736  			credit_pad = ep->tx_credit_size - credit_remainder;
737  			trans_len += credit_pad;
738  		}
739  
740  		ret = ath10k_htc_consume_credit(ep,
741  						bundle_buf + trans_len - bundle_skb->data,
742  						false);
743  		if (ret) {
744  			skb_queue_head(&ep->tx_req_head, skb);
745  			break;
746  		}
747  
748  		if (bundles_left < trans_len) {
749  			bundle_skb->len = bundle_buf - bundle_skb->data;
750  			ret = ath10k_htc_send_bundle(ep, bundle_skb, &tx_save_head);
751  
752  			if (ret) {
753  				skb_queue_head(&ep->tx_req_head, skb);
754  				return ret;
755  			}
756  
757  			if (skb_queue_len(&ep->tx_req_head) == 0) {
758  				ath10k_htc_send_one_skb(ep, skb);
759  				return ret;
760  			}
761  
762  			if (ep->tx_credit_flow_enabled &&
763  			    ep->tx_credits < ATH10K_MIN_CREDIT_PER_HTC_TX_BUNDLE) {
764  				skb_queue_head(&ep->tx_req_head, skb);
765  				return 0;
766  			}
767  
768  			bundles_left =
769  				ATH10K_MAX_MSG_PER_HTC_TX_BUNDLE * ep->tx_credit_size;
770  			bundle_skb = dev_alloc_skb(bundles_left);
771  
772  			if (!bundle_skb) {
773  				skb_queue_head(&ep->tx_req_head, skb);
774  				return -ENOMEM;
775  			}
776  			bundle_buf = bundle_skb->data;
777  			skb_queue_head_init(&tx_save_head);
778  		}
779  
780  		skb_push(skb, sizeof(struct ath10k_htc_hdr));
781  		ath10k_htc_prepare_tx_skb(ep, skb);
782  
783  		memcpy(bundle_buf, skb->data, skb->len);
784  		hdr = (struct ath10k_htc_hdr *)bundle_buf;
785  		hdr->flags |= ATH10K_HTC_FLAG_SEND_BUNDLE;
786  		hdr->pad_len = __cpu_to_le16(credit_pad);
787  		bundle_buf += trans_len;
788  		bundles_left -= trans_len;
789  		skb_queue_tail(&tx_save_head, skb);
790  	}
791  
792  	if (bundle_buf != bundle_skb->data) {
793  		bundle_skb->len = bundle_buf - bundle_skb->data;
794  		ret = ath10k_htc_send_bundle(ep, bundle_skb, &tx_save_head);
795  	} else {
796  		dev_kfree_skb_any(bundle_skb);
797  	}
798  
799  	return ret;
800  }
801  
ath10k_htc_bundle_tx_work(struct work_struct * work)802  static void ath10k_htc_bundle_tx_work(struct work_struct *work)
803  {
804  	struct ath10k *ar = container_of(work, struct ath10k, bundle_tx_work);
805  	struct ath10k_htc_ep *ep;
806  	struct sk_buff *skb;
807  	int i;
808  
809  	for (i = 0; i < ARRAY_SIZE(ar->htc.endpoint); i++) {
810  		ep = &ar->htc.endpoint[i];
811  
812  		if (!ep->bundle_tx)
813  			continue;
814  
815  		ath10k_dbg(ar, ATH10K_DBG_HTC, "bundle tx work eid %d count %d\n",
816  			   ep->eid, skb_queue_len(&ep->tx_req_head));
817  
818  		if (skb_queue_len(&ep->tx_req_head) >=
819  		    ATH10K_MIN_MSG_PER_HTC_TX_BUNDLE) {
820  			ath10k_htc_send_bundle_skbs(ep);
821  		} else {
822  			skb = skb_dequeue(&ep->tx_req_head);
823  
824  			if (!skb)
825  				continue;
826  			ath10k_htc_send_one_skb(ep, skb);
827  		}
828  	}
829  }
830  
ath10k_htc_tx_complete_work(struct work_struct * work)831  static void ath10k_htc_tx_complete_work(struct work_struct *work)
832  {
833  	struct ath10k *ar = container_of(work, struct ath10k, tx_complete_work);
834  	struct ath10k_htc_ep *ep;
835  	enum ath10k_htc_ep_id eid;
836  	struct sk_buff *skb;
837  	int i;
838  
839  	for (i = 0; i < ARRAY_SIZE(ar->htc.endpoint); i++) {
840  		ep = &ar->htc.endpoint[i];
841  		eid = ep->eid;
842  		if (ep->bundle_tx && eid == ar->htt.eid) {
843  			ath10k_dbg(ar, ATH10K_DBG_HTC, "bundle tx complete eid %d pending complete count%d\n",
844  				   ep->eid, skb_queue_len(&ep->tx_complete_head));
845  
846  			while (true) {
847  				skb = skb_dequeue(&ep->tx_complete_head);
848  				if (!skb)
849  					break;
850  				ath10k_htc_notify_tx_completion(ep, skb);
851  			}
852  		}
853  	}
854  }
855  
ath10k_htc_send_hl(struct ath10k_htc * htc,enum ath10k_htc_ep_id eid,struct sk_buff * skb)856  int ath10k_htc_send_hl(struct ath10k_htc *htc,
857  		       enum ath10k_htc_ep_id eid,
858  		       struct sk_buff *skb)
859  {
860  	struct ath10k_htc_ep *ep = &htc->endpoint[eid];
861  	struct ath10k *ar = htc->ar;
862  
863  	if (sizeof(struct ath10k_htc_hdr) + skb->len > ep->tx_credit_size) {
864  		ath10k_dbg(ar, ATH10K_DBG_HTC, "tx exceed max len %d\n", skb->len);
865  		return -ENOMEM;
866  	}
867  
868  	ath10k_dbg(ar, ATH10K_DBG_HTC, "htc send hl eid %d bundle %d tx count %d len %d\n",
869  		   eid, ep->bundle_tx, skb_queue_len(&ep->tx_req_head), skb->len);
870  
871  	if (ep->bundle_tx) {
872  		skb_queue_tail(&ep->tx_req_head, skb);
873  		queue_work(ar->workqueue, &ar->bundle_tx_work);
874  		return 0;
875  	} else {
876  		return ath10k_htc_send(htc, eid, skb);
877  	}
878  }
879  
ath10k_htc_setup_tx_req(struct ath10k_htc_ep * ep)880  void ath10k_htc_setup_tx_req(struct ath10k_htc_ep *ep)
881  {
882  	if (ep->htc->max_msgs_per_htc_bundle >= ATH10K_MIN_MSG_PER_HTC_TX_BUNDLE &&
883  	    !ep->bundle_tx) {
884  		ep->bundle_tx = true;
885  		skb_queue_head_init(&ep->tx_req_head);
886  		skb_queue_head_init(&ep->tx_complete_head);
887  	}
888  }
889  
ath10k_htc_stop_hl(struct ath10k * ar)890  void ath10k_htc_stop_hl(struct ath10k *ar)
891  {
892  	struct ath10k_htc_ep *ep;
893  	int i;
894  
895  	cancel_work_sync(&ar->bundle_tx_work);
896  	cancel_work_sync(&ar->tx_complete_work);
897  
898  	for (i = 0; i < ARRAY_SIZE(ar->htc.endpoint); i++) {
899  		ep = &ar->htc.endpoint[i];
900  
901  		if (!ep->bundle_tx)
902  			continue;
903  
904  		ath10k_dbg(ar, ATH10K_DBG_HTC, "stop tx work eid %d count %d\n",
905  			   ep->eid, skb_queue_len(&ep->tx_req_head));
906  
907  		skb_queue_purge(&ep->tx_req_head);
908  	}
909  }
910  
ath10k_htc_wait_target(struct ath10k_htc * htc)911  int ath10k_htc_wait_target(struct ath10k_htc *htc)
912  {
913  	struct ath10k *ar = htc->ar;
914  	int i, status = 0;
915  	unsigned long time_left;
916  	struct ath10k_htc_msg *msg;
917  	u16 message_id;
918  
919  	time_left = wait_for_completion_timeout(&htc->ctl_resp,
920  						ATH10K_HTC_WAIT_TIMEOUT_HZ);
921  	if (!time_left) {
922  		/* Workaround: In some cases the PCI HIF doesn't
923  		 * receive interrupt for the control response message
924  		 * even if the buffer was completed. It is suspected
925  		 * iomap writes unmasking PCI CE irqs aren't propagated
926  		 * properly in KVM PCI-passthrough sometimes.
927  		 */
928  		ath10k_warn(ar, "failed to receive control response completion, polling..\n");
929  
930  		for (i = 0; i < CE_COUNT; i++)
931  			ath10k_hif_send_complete_check(htc->ar, i, 1);
932  
933  		time_left =
934  		wait_for_completion_timeout(&htc->ctl_resp,
935  					    ATH10K_HTC_WAIT_TIMEOUT_HZ);
936  
937  		if (!time_left)
938  			status = -ETIMEDOUT;
939  	}
940  
941  	if (status < 0) {
942  		ath10k_err(ar, "ctl_resp never came in (%d)\n", status);
943  		return status;
944  	}
945  
946  	if (htc->control_resp_len < sizeof(msg->hdr) + sizeof(msg->ready)) {
947  		ath10k_err(ar, "Invalid HTC ready msg len:%d\n",
948  			   htc->control_resp_len);
949  		return -ECOMM;
950  	}
951  
952  	msg = (struct ath10k_htc_msg *)htc->control_resp_buffer;
953  	message_id   = __le16_to_cpu(msg->hdr.message_id);
954  
955  	if (message_id != ATH10K_HTC_MSG_READY_ID) {
956  		ath10k_err(ar, "Invalid HTC ready msg: 0x%x\n", message_id);
957  		return -ECOMM;
958  	}
959  
960  	if (ar->hw_params.use_fw_tx_credits)
961  		htc->total_transmit_credits = __le16_to_cpu(msg->ready.credit_count);
962  	else
963  		htc->total_transmit_credits = 1;
964  
965  	htc->target_credit_size = __le16_to_cpu(msg->ready.credit_size);
966  
967  	ath10k_dbg(ar, ATH10K_DBG_HTC,
968  		   "Target ready! transmit resources: %d size:%d actual credits:%d\n",
969  		   htc->total_transmit_credits,
970  		   htc->target_credit_size,
971  		   msg->ready.credit_count);
972  
973  	if ((htc->total_transmit_credits == 0) ||
974  	    (htc->target_credit_size == 0)) {
975  		ath10k_err(ar, "Invalid credit size received\n");
976  		return -ECOMM;
977  	}
978  
979  	/* The only way to determine if the ready message is an extended
980  	 * message is from the size.
981  	 */
982  	if (htc->control_resp_len >=
983  	    sizeof(msg->hdr) + sizeof(msg->ready_ext)) {
984  		htc->alt_data_credit_size =
985  			__le16_to_cpu(msg->ready_ext.reserved) &
986  			ATH10K_HTC_MSG_READY_EXT_ALT_DATA_MASK;
987  		htc->max_msgs_per_htc_bundle =
988  			min_t(u8, msg->ready_ext.max_msgs_per_htc_bundle,
989  			      HTC_HOST_MAX_MSG_PER_RX_BUNDLE);
990  		ath10k_dbg(ar, ATH10K_DBG_HTC,
991  			   "Extended ready message RX bundle size %d alt size %d\n",
992  			   htc->max_msgs_per_htc_bundle,
993  			   htc->alt_data_credit_size);
994  	}
995  
996  	INIT_WORK(&ar->bundle_tx_work, ath10k_htc_bundle_tx_work);
997  	INIT_WORK(&ar->tx_complete_work, ath10k_htc_tx_complete_work);
998  
999  	return 0;
1000  }
1001  
ath10k_htc_change_tx_credit_flow(struct ath10k_htc * htc,enum ath10k_htc_ep_id eid,bool enable)1002  void ath10k_htc_change_tx_credit_flow(struct ath10k_htc *htc,
1003  				      enum ath10k_htc_ep_id eid,
1004  				      bool enable)
1005  {
1006  	struct ath10k *ar = htc->ar;
1007  	struct ath10k_htc_ep *ep = &ar->htc.endpoint[eid];
1008  
1009  	ep->tx_credit_flow_enabled = enable;
1010  }
1011  
ath10k_htc_connect_service(struct ath10k_htc * htc,struct ath10k_htc_svc_conn_req * conn_req,struct ath10k_htc_svc_conn_resp * conn_resp)1012  int ath10k_htc_connect_service(struct ath10k_htc *htc,
1013  			       struct ath10k_htc_svc_conn_req *conn_req,
1014  			       struct ath10k_htc_svc_conn_resp *conn_resp)
1015  {
1016  	struct ath10k *ar = htc->ar;
1017  	struct ath10k_htc_msg *msg;
1018  	struct ath10k_htc_conn_svc *req_msg;
1019  	struct ath10k_htc_conn_svc_response resp_msg_dummy;
1020  	struct ath10k_htc_conn_svc_response *resp_msg = &resp_msg_dummy;
1021  	enum ath10k_htc_ep_id assigned_eid = ATH10K_HTC_EP_COUNT;
1022  	struct ath10k_htc_ep *ep;
1023  	struct sk_buff *skb;
1024  	unsigned int max_msg_size = 0;
1025  	int length, status;
1026  	unsigned long time_left;
1027  	bool disable_credit_flow_ctrl = false;
1028  	u16 message_id, service_id, flags = 0;
1029  	u8 tx_alloc = 0;
1030  
1031  	/* special case for HTC pseudo control service */
1032  	if (conn_req->service_id == ATH10K_HTC_SVC_ID_RSVD_CTRL) {
1033  		disable_credit_flow_ctrl = true;
1034  		assigned_eid = ATH10K_HTC_EP_0;
1035  		max_msg_size = ATH10K_HTC_MAX_CTRL_MSG_LEN;
1036  		memset(&resp_msg_dummy, 0, sizeof(resp_msg_dummy));
1037  		goto setup;
1038  	}
1039  
1040  	tx_alloc = ath10k_htc_get_credit_allocation(htc,
1041  						    conn_req->service_id);
1042  	if (!tx_alloc)
1043  		ath10k_dbg(ar, ATH10K_DBG_BOOT,
1044  			   "boot htc service %s does not allocate target credits\n",
1045  			   htc_service_name(conn_req->service_id));
1046  
1047  	skb = ath10k_htc_build_tx_ctrl_skb(htc->ar);
1048  	if (!skb) {
1049  		ath10k_err(ar, "Failed to allocate HTC packet\n");
1050  		return -ENOMEM;
1051  	}
1052  
1053  	length = sizeof(msg->hdr) + sizeof(msg->connect_service);
1054  	skb_put(skb, length);
1055  	memset(skb->data, 0, length);
1056  
1057  	msg = (struct ath10k_htc_msg *)skb->data;
1058  	msg->hdr.message_id =
1059  		__cpu_to_le16(ATH10K_HTC_MSG_CONNECT_SERVICE_ID);
1060  
1061  	flags |= SM(tx_alloc, ATH10K_HTC_CONN_FLAGS_RECV_ALLOC);
1062  
1063  	/* Only enable credit flow control for WMI ctrl service */
1064  	if (conn_req->service_id != ATH10K_HTC_SVC_ID_WMI_CONTROL) {
1065  		flags |= ATH10K_HTC_CONN_FLAGS_DISABLE_CREDIT_FLOW_CTRL;
1066  		disable_credit_flow_ctrl = true;
1067  	}
1068  
1069  	req_msg = &msg->connect_service;
1070  	req_msg->flags = __cpu_to_le16(flags);
1071  	req_msg->service_id = __cpu_to_le16(conn_req->service_id);
1072  
1073  	reinit_completion(&htc->ctl_resp);
1074  
1075  	status = ath10k_htc_send(htc, ATH10K_HTC_EP_0, skb);
1076  	if (status) {
1077  		kfree_skb(skb);
1078  		return status;
1079  	}
1080  
1081  	/* wait for response */
1082  	time_left = wait_for_completion_timeout(&htc->ctl_resp,
1083  						ATH10K_HTC_CONN_SVC_TIMEOUT_HZ);
1084  	if (!time_left) {
1085  		ath10k_err(ar, "Service connect timeout\n");
1086  		return -ETIMEDOUT;
1087  	}
1088  
1089  	/* we controlled the buffer creation, it's aligned */
1090  	msg = (struct ath10k_htc_msg *)htc->control_resp_buffer;
1091  	resp_msg = &msg->connect_service_response;
1092  	message_id = __le16_to_cpu(msg->hdr.message_id);
1093  	service_id = __le16_to_cpu(resp_msg->service_id);
1094  
1095  	if ((message_id != ATH10K_HTC_MSG_CONNECT_SERVICE_RESP_ID) ||
1096  	    (htc->control_resp_len < sizeof(msg->hdr) +
1097  	     sizeof(msg->connect_service_response))) {
1098  		ath10k_err(ar, "Invalid resp message ID 0x%x", message_id);
1099  		return -EPROTO;
1100  	}
1101  
1102  	ath10k_dbg(ar, ATH10K_DBG_HTC,
1103  		   "HTC Service %s connect response: status: 0x%x, assigned ep: 0x%x\n",
1104  		   htc_service_name(service_id),
1105  		   resp_msg->status, resp_msg->eid);
1106  
1107  	conn_resp->connect_resp_code = resp_msg->status;
1108  
1109  	/* check response status */
1110  	if (resp_msg->status != ATH10K_HTC_CONN_SVC_STATUS_SUCCESS) {
1111  		ath10k_err(ar, "HTC Service %s connect request failed: 0x%x)\n",
1112  			   htc_service_name(service_id),
1113  			   resp_msg->status);
1114  		return -EPROTO;
1115  	}
1116  
1117  	assigned_eid = (enum ath10k_htc_ep_id)resp_msg->eid;
1118  	max_msg_size = __le16_to_cpu(resp_msg->max_msg_size);
1119  
1120  setup:
1121  
1122  	if (assigned_eid >= ATH10K_HTC_EP_COUNT)
1123  		return -EPROTO;
1124  
1125  	if (max_msg_size == 0)
1126  		return -EPROTO;
1127  
1128  	ep = &htc->endpoint[assigned_eid];
1129  	ep->eid = assigned_eid;
1130  
1131  	if (ep->service_id != ATH10K_HTC_SVC_ID_UNUSED)
1132  		return -EPROTO;
1133  
1134  	/* return assigned endpoint to caller */
1135  	conn_resp->eid = assigned_eid;
1136  	conn_resp->max_msg_len = __le16_to_cpu(resp_msg->max_msg_size);
1137  
1138  	/* setup the endpoint */
1139  	ep->service_id = conn_req->service_id;
1140  	ep->max_tx_queue_depth = conn_req->max_send_queue_depth;
1141  	ep->max_ep_message_len = __le16_to_cpu(resp_msg->max_msg_size);
1142  	ep->tx_credits = tx_alloc;
1143  	ep->tx_credit_size = htc->target_credit_size;
1144  
1145  	if (conn_req->service_id == ATH10K_HTC_SVC_ID_HTT_DATA_MSG &&
1146  	    htc->alt_data_credit_size != 0)
1147  		ep->tx_credit_size = htc->alt_data_credit_size;
1148  
1149  	/* copy all the callbacks */
1150  	ep->ep_ops = conn_req->ep_ops;
1151  
1152  	status = ath10k_hif_map_service_to_pipe(htc->ar,
1153  						ep->service_id,
1154  						&ep->ul_pipe_id,
1155  						&ep->dl_pipe_id);
1156  	if (status) {
1157  		ath10k_dbg(ar, ATH10K_DBG_BOOT, "unsupported HTC service id: %d\n",
1158  			   ep->service_id);
1159  		return status;
1160  	}
1161  
1162  	ath10k_dbg(ar, ATH10K_DBG_BOOT,
1163  		   "boot htc service '%s' ul pipe %d dl pipe %d eid %d ready\n",
1164  		   htc_service_name(ep->service_id), ep->ul_pipe_id,
1165  		   ep->dl_pipe_id, ep->eid);
1166  
1167  	if (disable_credit_flow_ctrl && ep->tx_credit_flow_enabled) {
1168  		ep->tx_credit_flow_enabled = false;
1169  		ath10k_dbg(ar, ATH10K_DBG_BOOT,
1170  			   "boot htc service '%s' eid %d TX flow control disabled\n",
1171  			   htc_service_name(ep->service_id), assigned_eid);
1172  	}
1173  
1174  	return status;
1175  }
1176  
ath10k_htc_alloc_skb(struct ath10k * ar,int size)1177  struct sk_buff *ath10k_htc_alloc_skb(struct ath10k *ar, int size)
1178  {
1179  	struct sk_buff *skb;
1180  
1181  	skb = dev_alloc_skb(size + sizeof(struct ath10k_htc_hdr));
1182  	if (!skb)
1183  		return NULL;
1184  
1185  	skb_reserve(skb, sizeof(struct ath10k_htc_hdr));
1186  
1187  	/* FW/HTC requires 4-byte aligned streams */
1188  	if (!IS_ALIGNED((unsigned long)skb->data, 4))
1189  		ath10k_warn(ar, "Unaligned HTC tx skb\n");
1190  
1191  	return skb;
1192  }
1193  
ath10k_htc_pktlog_process_rx(struct ath10k * ar,struct sk_buff * skb)1194  static void ath10k_htc_pktlog_process_rx(struct ath10k *ar, struct sk_buff *skb)
1195  {
1196  	trace_ath10k_htt_pktlog(ar, skb->data, skb->len);
1197  	dev_kfree_skb_any(skb);
1198  }
1199  
ath10k_htc_pktlog_connect(struct ath10k * ar)1200  static int ath10k_htc_pktlog_connect(struct ath10k *ar)
1201  {
1202  	struct ath10k_htc_svc_conn_resp conn_resp;
1203  	struct ath10k_htc_svc_conn_req conn_req;
1204  	int status;
1205  
1206  	memset(&conn_req, 0, sizeof(conn_req));
1207  	memset(&conn_resp, 0, sizeof(conn_resp));
1208  
1209  	conn_req.ep_ops.ep_tx_complete = NULL;
1210  	conn_req.ep_ops.ep_rx_complete = ath10k_htc_pktlog_process_rx;
1211  	conn_req.ep_ops.ep_tx_credits = NULL;
1212  
1213  	/* connect to control service */
1214  	conn_req.service_id = ATH10K_HTC_SVC_ID_HTT_LOG_MSG;
1215  	status = ath10k_htc_connect_service(&ar->htc, &conn_req, &conn_resp);
1216  	if (status) {
1217  		ath10k_warn(ar, "failed to connect to PKTLOG service: %d\n",
1218  			    status);
1219  		return status;
1220  	}
1221  
1222  	return 0;
1223  }
1224  
ath10k_htc_pktlog_svc_supported(struct ath10k * ar)1225  static bool ath10k_htc_pktlog_svc_supported(struct ath10k *ar)
1226  {
1227  	u8 ul_pipe_id;
1228  	u8 dl_pipe_id;
1229  	int status;
1230  
1231  	status = ath10k_hif_map_service_to_pipe(ar, ATH10K_HTC_SVC_ID_HTT_LOG_MSG,
1232  						&ul_pipe_id,
1233  						&dl_pipe_id);
1234  	if (status) {
1235  		ath10k_dbg(ar, ATH10K_DBG_BOOT, "unsupported HTC pktlog service id: %d\n",
1236  			   ATH10K_HTC_SVC_ID_HTT_LOG_MSG);
1237  
1238  		return false;
1239  	}
1240  
1241  	return true;
1242  }
1243  
ath10k_htc_start(struct ath10k_htc * htc)1244  int ath10k_htc_start(struct ath10k_htc *htc)
1245  {
1246  	struct ath10k *ar = htc->ar;
1247  	struct sk_buff *skb;
1248  	int status = 0;
1249  	struct ath10k_htc_msg *msg;
1250  
1251  	skb = ath10k_htc_build_tx_ctrl_skb(htc->ar);
1252  	if (!skb)
1253  		return -ENOMEM;
1254  
1255  	skb_put(skb, sizeof(msg->hdr) + sizeof(msg->setup_complete_ext));
1256  	memset(skb->data, 0, skb->len);
1257  
1258  	msg = (struct ath10k_htc_msg *)skb->data;
1259  	msg->hdr.message_id =
1260  		__cpu_to_le16(ATH10K_HTC_MSG_SETUP_COMPLETE_EX_ID);
1261  
1262  	if (ar->hif.bus == ATH10K_BUS_SDIO) {
1263  		/* Extra setup params used by SDIO */
1264  		msg->setup_complete_ext.flags =
1265  			__cpu_to_le32(ATH10K_HTC_SETUP_COMPLETE_FLAGS_RX_BNDL_EN);
1266  		msg->setup_complete_ext.max_msgs_per_bundled_recv =
1267  			htc->max_msgs_per_htc_bundle;
1268  	}
1269  	ath10k_dbg(ar, ATH10K_DBG_HTC, "HTC is using TX credit flow control\n");
1270  
1271  	status = ath10k_htc_send(htc, ATH10K_HTC_EP_0, skb);
1272  	if (status) {
1273  		kfree_skb(skb);
1274  		return status;
1275  	}
1276  
1277  	if (ath10k_htc_pktlog_svc_supported(ar)) {
1278  		status = ath10k_htc_pktlog_connect(ar);
1279  		if (status) {
1280  			ath10k_err(ar, "failed to connect to pktlog: %d\n", status);
1281  			return status;
1282  		}
1283  	}
1284  
1285  	return 0;
1286  }
1287  
1288  /* registered target arrival callback from the HIF layer */
ath10k_htc_init(struct ath10k * ar)1289  int ath10k_htc_init(struct ath10k *ar)
1290  {
1291  	int status;
1292  	struct ath10k_htc *htc = &ar->htc;
1293  	struct ath10k_htc_svc_conn_req conn_req;
1294  	struct ath10k_htc_svc_conn_resp conn_resp;
1295  
1296  	spin_lock_init(&htc->tx_lock);
1297  
1298  	ath10k_htc_reset_endpoint_states(htc);
1299  
1300  	htc->ar = ar;
1301  
1302  	/* setup our pseudo HTC control endpoint connection */
1303  	memset(&conn_req, 0, sizeof(conn_req));
1304  	memset(&conn_resp, 0, sizeof(conn_resp));
1305  	conn_req.ep_ops.ep_tx_complete = ath10k_htc_control_tx_complete;
1306  	conn_req.ep_ops.ep_rx_complete = ath10k_htc_control_rx_complete;
1307  	conn_req.max_send_queue_depth = ATH10K_NUM_CONTROL_TX_BUFFERS;
1308  	conn_req.service_id = ATH10K_HTC_SVC_ID_RSVD_CTRL;
1309  
1310  	/* connect fake service */
1311  	status = ath10k_htc_connect_service(htc, &conn_req, &conn_resp);
1312  	if (status) {
1313  		ath10k_err(ar, "could not connect to htc service (%d)\n",
1314  			   status);
1315  		return status;
1316  	}
1317  
1318  	init_completion(&htc->ctl_resp);
1319  
1320  	return 0;
1321  }
1322