1 /* 2 * Copyright (c) 2016-2018 The Linux Foundation. All rights reserved. 3 * 4 * Permission to use, copy, modify, and/or distribute this software for 5 * any purpose with or without fee is hereby granted, provided that the 6 * above copyright notice and this permission notice appear in all 7 * copies. 8 * 9 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL 10 * WARRANTIES WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED 11 * WARRANTIES OF MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE 12 * AUTHOR BE LIABLE FOR ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL 13 * DAMAGES OR ANY DAMAGES WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR 14 * PROFITS, WHETHER IN AN ACTION OF CONTRACT, NEGLIGENCE OR OTHER 15 * TORTIOUS ACTION, ARISING OUT OF OR IN CONNECTION WITH THE USE OR 16 * PERFORMANCE OF THIS SOFTWARE. 17 */ 18 19 #include "dp_types.h" 20 #include "hal_reo.h" 21 #include "dp_internal.h" 22 23 QDF_STATUS dp_reo_send_cmd(struct dp_soc *soc, enum hal_reo_cmd_type type, 24 struct hal_reo_cmd_params *params, 25 void (*callback_fn), void *data) 26 { 27 struct dp_reo_cmd_info *reo_cmd; 28 int num; 29 30 switch (type) { 31 case CMD_GET_QUEUE_STATS: 32 num = hal_reo_cmd_queue_stats(soc->reo_cmd_ring.hal_srng, 33 soc->hal_soc, params); 34 break; 35 case CMD_FLUSH_QUEUE: 36 num = hal_reo_cmd_flush_queue(soc->reo_cmd_ring.hal_srng, 37 soc->hal_soc, params); 38 break; 39 case CMD_FLUSH_CACHE: 40 num = hal_reo_cmd_flush_cache(soc->reo_cmd_ring.hal_srng, 41 soc->hal_soc, params); 42 break; 43 case CMD_UNBLOCK_CACHE: 44 num = hal_reo_cmd_unblock_cache(soc->reo_cmd_ring.hal_srng, 45 soc->hal_soc, params); 46 break; 47 case CMD_FLUSH_TIMEOUT_LIST: 48 num = hal_reo_cmd_flush_timeout_list(soc->reo_cmd_ring.hal_srng, 49 soc->hal_soc, params); 50 break; 51 case CMD_UPDATE_RX_REO_QUEUE: 52 num = hal_reo_cmd_update_rx_queue(soc->reo_cmd_ring.hal_srng, 53 soc->hal_soc, params); 54 break; 55 default: 56 QDF_TRACE(QDF_MODULE_ID_TXRX, QDF_TRACE_LEVEL_ERROR, 57 "%s: Invalid REO command type", __func__); 58 return QDF_STATUS_E_FAILURE; 59 }; 60 61 if (num < 0) { 62 qdf_print("%s: Error with sending REO command", __func__); 63 return QDF_STATUS_E_FAILURE; 64 } 65 66 if (callback_fn) { 67 reo_cmd = qdf_mem_malloc(sizeof(*reo_cmd)); 68 if (!reo_cmd) { 69 QDF_TRACE(QDF_MODULE_ID_TXRX, QDF_TRACE_LEVEL_ERROR, 70 "%s: alloc failed for REO cmd:%d!!", 71 __func__, type); 72 return QDF_STATUS_E_NOMEM; 73 } 74 75 reo_cmd->cmd = num; 76 reo_cmd->cmd_type = type; 77 reo_cmd->handler = callback_fn; 78 reo_cmd->data = data; 79 qdf_spin_lock_bh(&soc->rx.reo_cmd_lock); 80 TAILQ_INSERT_TAIL(&soc->rx.reo_cmd_list, reo_cmd, 81 reo_cmd_list_elem); 82 qdf_spin_unlock_bh(&soc->rx.reo_cmd_lock); 83 } 84 85 return QDF_STATUS_SUCCESS; 86 } 87 88 void dp_reo_status_ring_handler(struct dp_soc *soc) 89 { 90 uint32_t *reo_desc; 91 struct dp_reo_cmd_info *reo_cmd = NULL; 92 union hal_reo_status reo_status; 93 int num; 94 95 if (hal_srng_access_start(soc->hal_soc, 96 soc->reo_status_ring.hal_srng)) { 97 return; 98 } 99 reo_desc = hal_srng_dst_get_next(soc->hal_soc, 100 soc->reo_status_ring.hal_srng); 101 102 while (reo_desc) { 103 uint16_t tlv = HAL_GET_TLV(reo_desc); 104 105 switch (tlv) { 106 case HAL_REO_QUEUE_STATS_STATUS_TLV: 107 hal_reo_queue_stats_status(reo_desc, 108 &reo_status.queue_status, 109 soc->hal_soc); 110 num = reo_status.queue_status.header.cmd_num; 111 break; 112 case HAL_REO_FLUSH_QUEUE_STATUS_TLV: 113 hal_reo_flush_queue_status(reo_desc, 114 &reo_status.fl_queue_status, 115 soc->hal_soc); 116 num = reo_status.fl_queue_status.header.cmd_num; 117 break; 118 case HAL_REO_FLUSH_CACHE_STATUS_TLV: 119 hal_reo_flush_cache_status(reo_desc, soc->hal_soc, 120 &reo_status.fl_cache_status, 121 soc->hal_soc); 122 num = reo_status.fl_cache_status.header.cmd_num; 123 break; 124 case HAL_REO_UNBLK_CACHE_STATUS_TLV: 125 hal_reo_unblock_cache_status(reo_desc, soc->hal_soc, 126 &reo_status.unblk_cache_status); 127 num = reo_status.unblk_cache_status.header.cmd_num; 128 break; 129 case HAL_REO_TIMOUT_LIST_STATUS_TLV: 130 hal_reo_flush_timeout_list_status(reo_desc, 131 &reo_status.fl_timeout_status, 132 soc->hal_soc); 133 num = reo_status.fl_timeout_status.header.cmd_num; 134 break; 135 case HAL_REO_DESC_THRES_STATUS_TLV: 136 hal_reo_desc_thres_reached_status(reo_desc, 137 &reo_status.thres_status, 138 soc->hal_soc); 139 num = reo_status.thres_status.header.cmd_num; 140 break; 141 case HAL_REO_UPDATE_RX_QUEUE_STATUS_TLV: 142 hal_reo_rx_update_queue_status(reo_desc, 143 &reo_status.rx_queue_status, 144 soc->hal_soc); 145 num = reo_status.rx_queue_status.header.cmd_num; 146 break; 147 default: 148 QDF_TRACE(QDF_MODULE_ID_DP, QDF_TRACE_LEVEL_WARN, 149 "%s, no handler for TLV:%d", __func__, tlv); 150 goto next; 151 } /* switch */ 152 153 qdf_spin_lock_bh(&soc->rx.reo_cmd_lock); 154 TAILQ_FOREACH(reo_cmd, &soc->rx.reo_cmd_list, 155 reo_cmd_list_elem) { 156 if (reo_cmd->cmd == num) { 157 TAILQ_REMOVE(&soc->rx.reo_cmd_list, reo_cmd, 158 reo_cmd_list_elem); 159 break; 160 } 161 } 162 qdf_spin_unlock_bh(&soc->rx.reo_cmd_lock); 163 164 if (reo_cmd) { 165 reo_cmd->handler(soc, reo_cmd->data, 166 &reo_status); 167 qdf_mem_free(reo_cmd); 168 } 169 170 next: 171 reo_desc = hal_srng_dst_get_next(soc, 172 soc->reo_status_ring.hal_srng); 173 } /* while */ 174 175 hal_srng_access_end(soc->hal_soc, soc->reo_status_ring.hal_srng); 176 } 177 178 /** 179 * dp_reo_cmdlist_destroy - Free REO commands in the queue 180 * @soc: DP SoC hanle 181 * 182 */ 183 void dp_reo_cmdlist_destroy(struct dp_soc *soc) 184 { 185 struct dp_reo_cmd_info *reo_cmd = NULL; 186 struct dp_reo_cmd_info *tmp_cmd = NULL; 187 union hal_reo_status reo_status; 188 189 reo_status.queue_status.header.status = 190 HAL_REO_CMD_DRAIN; 191 192 qdf_spin_lock_bh(&soc->rx.reo_cmd_lock); 193 TAILQ_FOREACH_SAFE(reo_cmd, &soc->rx.reo_cmd_list, 194 reo_cmd_list_elem, tmp_cmd) { 195 TAILQ_REMOVE(&soc->rx.reo_cmd_list, reo_cmd, 196 reo_cmd_list_elem); 197 reo_cmd->handler(soc, reo_cmd->data, &reo_status); 198 qdf_mem_free(reo_cmd); 199 } 200 qdf_spin_unlock_bh(&soc->rx.reo_cmd_lock); 201 } 202