1 /* 2 * Copyright (c) 2016-2018 The Linux Foundation. All rights reserved. 3 * 4 * Permission to use, copy, modify, and/or distribute this software for 5 * any purpose with or without fee is hereby granted, provided that the 6 * above copyright notice and this permission notice appear in all 7 * copies. 8 * 9 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL 10 * WARRANTIES WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED 11 * WARRANTIES OF MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE 12 * AUTHOR BE LIABLE FOR ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL 13 * DAMAGES OR ANY DAMAGES WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR 14 * PROFITS, WHETHER IN AN ACTION OF CONTRACT, NEGLIGENCE OR OTHER 15 * TORTIOUS ACTION, ARISING OUT OF OR IN CONNECTION WITH THE USE OR 16 * PERFORMANCE OF THIS SOFTWARE. 17 */ 18 19 #include "dp_types.h" 20 #include "hal_reo.h" 21 #include "dp_internal.h" 22 23 QDF_STATUS dp_reo_send_cmd(struct dp_soc *soc, enum hal_reo_cmd_type type, 24 struct hal_reo_cmd_params *params, 25 void (*callback_fn), void *data) 26 { 27 struct dp_reo_cmd_info *reo_cmd; 28 int num; 29 30 switch (type) { 31 case CMD_GET_QUEUE_STATS: 32 num = hal_reo_cmd_queue_stats(soc->reo_cmd_ring.hal_srng, 33 soc->hal_soc, params); 34 break; 35 case CMD_FLUSH_QUEUE: 36 num = hal_reo_cmd_flush_queue(soc->reo_cmd_ring.hal_srng, 37 soc->hal_soc, params); 38 break; 39 case CMD_FLUSH_CACHE: 40 num = hal_reo_cmd_flush_cache(soc->reo_cmd_ring.hal_srng, 41 soc->hal_soc, params); 42 break; 43 case CMD_UNBLOCK_CACHE: 44 num = hal_reo_cmd_unblock_cache(soc->reo_cmd_ring.hal_srng, 45 soc->hal_soc, params); 46 break; 47 case CMD_FLUSH_TIMEOUT_LIST: 48 num = hal_reo_cmd_flush_timeout_list(soc->reo_cmd_ring.hal_srng, 49 soc->hal_soc, params); 50 break; 51 case CMD_UPDATE_RX_REO_QUEUE: 52 num = hal_reo_cmd_update_rx_queue(soc->reo_cmd_ring.hal_srng, 53 soc->hal_soc, params); 54 break; 55 default: 56 QDF_TRACE(QDF_MODULE_ID_TXRX, QDF_TRACE_LEVEL_ERROR, 57 "%s: Invalid REO command type\n", __func__); 58 return QDF_STATUS_E_FAILURE; 59 }; 60 61 if (num < 0) { 62 qdf_print("%s: Error with sending REO command\n", __func__); 63 return QDF_STATUS_E_FAILURE; 64 } 65 66 if (callback_fn) { 67 reo_cmd = qdf_mem_malloc(sizeof(*reo_cmd)); 68 if (!reo_cmd) { 69 QDF_TRACE(QDF_MODULE_ID_TXRX, QDF_TRACE_LEVEL_ERROR, 70 "%s: alloc failed for REO cmd:%d!!\n", 71 __func__, type); 72 return QDF_STATUS_E_NOMEM; 73 } 74 75 reo_cmd->cmd = num; 76 reo_cmd->cmd_type = type; 77 reo_cmd->handler = callback_fn; 78 reo_cmd->data = data; 79 qdf_spin_lock_bh(&soc->rx.reo_cmd_lock); 80 TAILQ_INSERT_TAIL(&soc->rx.reo_cmd_list, reo_cmd, 81 reo_cmd_list_elem); 82 qdf_spin_unlock_bh(&soc->rx.reo_cmd_lock); 83 } 84 85 return QDF_STATUS_SUCCESS; 86 } 87 88 void dp_reo_status_ring_handler(struct dp_soc *soc) 89 { 90 uint32_t *reo_desc; 91 struct dp_reo_cmd_info *reo_cmd = NULL; 92 union hal_reo_status reo_status; 93 int num; 94 95 if (hal_srng_access_start(soc->hal_soc, 96 soc->reo_status_ring.hal_srng)) { 97 return; 98 } 99 reo_desc = hal_srng_dst_get_next(soc->hal_soc, 100 soc->reo_status_ring.hal_srng); 101 102 while (reo_desc) { 103 uint16_t tlv = HAL_GET_TLV(reo_desc); 104 105 switch (tlv) { 106 case HAL_REO_QUEUE_STATS_STATUS_TLV: 107 hal_reo_queue_stats_status(reo_desc, 108 &reo_status.queue_status); 109 num = reo_status.queue_status.header.cmd_num; 110 break; 111 case HAL_REO_FLUSH_QUEUE_STATUS_TLV: 112 hal_reo_flush_queue_status(reo_desc, 113 &reo_status.fl_queue_status); 114 num = reo_status.fl_queue_status.header.cmd_num; 115 break; 116 case HAL_REO_FLUSH_CACHE_STATUS_TLV: 117 hal_reo_flush_cache_status(reo_desc, soc->hal_soc, 118 &reo_status.fl_cache_status); 119 num = reo_status.fl_cache_status.header.cmd_num; 120 break; 121 case HAL_REO_UNBLK_CACHE_STATUS_TLV: 122 hal_reo_unblock_cache_status(reo_desc, soc->hal_soc, 123 &reo_status.unblk_cache_status); 124 num = reo_status.unblk_cache_status.header.cmd_num; 125 break; 126 case HAL_REO_TIMOUT_LIST_STATUS_TLV: 127 hal_reo_flush_timeout_list_status(reo_desc, 128 &reo_status.fl_timeout_status); 129 num = reo_status.fl_timeout_status.header.cmd_num; 130 break; 131 case HAL_REO_DESC_THRES_STATUS_TLV: 132 hal_reo_desc_thres_reached_status(reo_desc, 133 &reo_status.thres_status); 134 num = reo_status.thres_status.header.cmd_num; 135 break; 136 case HAL_REO_UPDATE_RX_QUEUE_STATUS_TLV: 137 hal_reo_rx_update_queue_status(reo_desc, 138 &reo_status.rx_queue_status); 139 num = reo_status.rx_queue_status.header.cmd_num; 140 break; 141 default: 142 QDF_TRACE(QDF_MODULE_ID_DP, QDF_TRACE_LEVEL_WARN, 143 "%s, no handler for TLV:%d\n", __func__, tlv); 144 goto next; 145 } /* switch */ 146 147 qdf_spin_lock_bh(&soc->rx.reo_cmd_lock); 148 TAILQ_FOREACH(reo_cmd, &soc->rx.reo_cmd_list, 149 reo_cmd_list_elem) { 150 if (reo_cmd->cmd == num) { 151 TAILQ_REMOVE(&soc->rx.reo_cmd_list, reo_cmd, 152 reo_cmd_list_elem); 153 break; 154 } 155 } 156 qdf_spin_unlock_bh(&soc->rx.reo_cmd_lock); 157 158 if (reo_cmd) { 159 reo_cmd->handler(soc, reo_cmd->data, 160 &reo_status); 161 qdf_mem_free(reo_cmd); 162 } 163 164 next: 165 reo_desc = hal_srng_dst_get_next(soc, 166 soc->reo_status_ring.hal_srng); 167 } /* while */ 168 169 hal_srng_access_end(soc->hal_soc, soc->reo_status_ring.hal_srng); 170 } 171 172 /** 173 * dp_reo_cmdlist_destroy - Free REO commands in the queue 174 * @soc: DP SoC hanle 175 * 176 */ 177 void dp_reo_cmdlist_destroy(struct dp_soc *soc) 178 { 179 struct dp_reo_cmd_info *reo_cmd = NULL; 180 struct dp_reo_cmd_info *tmp_cmd = NULL; 181 union hal_reo_status reo_status; 182 183 reo_status.queue_status.header.status = 184 HAL_REO_CMD_DRAIN; 185 186 qdf_spin_lock_bh(&soc->rx.reo_cmd_lock); 187 TAILQ_FOREACH_SAFE(reo_cmd, &soc->rx.reo_cmd_list, 188 reo_cmd_list_elem, tmp_cmd) { 189 TAILQ_REMOVE(&soc->rx.reo_cmd_list, reo_cmd, 190 reo_cmd_list_elem); 191 reo_cmd->handler(soc, reo_cmd->data, &reo_status); 192 qdf_mem_free(reo_cmd); 193 } 194 qdf_spin_unlock_bh(&soc->rx.reo_cmd_lock); 195 } 196