1 /* 2 * Copyright (c) 2014-2021 The Linux Foundation. All rights reserved. 3 * Copyright (c) 2021-2023 Qualcomm Innovation Center, Inc. All rights reserved. 4 * 5 * Permission to use, copy, modify, and/or distribute this software for 6 * any purpose with or without fee is hereby granted, provided that the 7 * above copyright notice and this permission notice appear in all 8 * copies. 9 * 10 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL 11 * WARRANTIES WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED 12 * WARRANTIES OF MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE 13 * AUTHOR BE LIABLE FOR ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL 14 * DAMAGES OR ANY DAMAGES WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR 15 * PROFITS, WHETHER IN AN ACTION OF CONTRACT, NEGLIGENCE OR OTHER 16 * TORTIOUS ACTION, ARISING OUT OF OR IN CONNECTION WITH THE USE OR 17 * PERFORMANCE OF THIS SOFTWARE. 18 */ 19 20 /** 21 * DOC: i_qdf_util.h 22 * This file provides OS dependent API's. 23 */ 24 25 #ifndef _I_QDF_UTIL_H 26 #define _I_QDF_UTIL_H 27 28 #include <linux/compiler.h> 29 #include <linux/kernel.h> 30 #include <linux/types.h> 31 #include <linux/mm.h> 32 #include <linux/errno.h> 33 #include <linux/average.h> 34 35 #include <linux/random.h> 36 #include <linux/io.h> 37 38 #include <qdf_types.h> 39 #include <asm/byteorder.h> 40 41 #if LINUX_VERSION_CODE <= KERNEL_VERSION(3, 3, 8) 42 #include <asm/system.h> 43 #else 44 #if defined(__LINUX_MIPS32_ARCH__) || defined(__LINUX_MIPS64_ARCH__) 45 #include <asm/dec/system.h> 46 #else 47 #endif 48 #endif 49 50 #include <qdf_types.h> 51 #include <linux/io.h> 52 #include <asm/byteorder.h> 53 54 #ifdef QCA_PARTNER_PLATFORM 55 #include "ath_carr_pltfrm.h" 56 #else 57 #include <linux/byteorder/generic.h> 58 #endif 59 60 #include <linux/rcupdate.h> 61 62 typedef wait_queue_head_t __qdf_wait_queue_head_t; 63 64 /* Generic compiler-dependent macros if defined by the OS */ 65 #define __qdf_wait_queue_interruptible(wait_queue, condition) \ 66 wait_event_interruptible(wait_queue, condition) 67 68 #define __qdf_wait_queue_timeout(wait_queue, condition, timeout) \ 69 wait_event_timeout(wait_queue, condition, timeout) 70 71 72 #define __qdf_init_waitqueue_head(_q) init_waitqueue_head(_q) 73 74 #define __qdf_wake_up_interruptible(_q) wake_up_interruptible(_q) 75 76 #define __qdf_wake_up(_q) wake_up(_q) 77 78 #define __qdf_wake_up_completion(_q) wake_up_completion(_q) 79 80 #define __qdf_unlikely(_expr) unlikely(_expr) 81 #define __qdf_likely(_expr) likely(_expr) 82 83 #define __qdf_bitmap(name, bits) DECLARE_BITMAP(name, bits) 84 85 /** 86 * __qdf_set_bit() - set bit in address 87 * @nr: bit number to be set 88 * @addr: address buffer pointer 89 * 90 * Return: none 91 */ 92 static inline void __qdf_set_bit(unsigned int nr, unsigned long *addr) 93 { 94 __set_bit(nr, addr); 95 } 96 97 static inline void __qdf_clear_bit(unsigned int nr, unsigned long *addr) 98 { 99 __clear_bit(nr, addr); 100 } 101 102 static inline bool __qdf_test_bit(unsigned int nr, unsigned long *addr) 103 { 104 return test_bit(nr, addr); 105 } 106 107 static inline bool __qdf_test_and_clear_bit(unsigned int nr, 108 unsigned long *addr) 109 { 110 return __test_and_clear_bit(nr, addr); 111 } 112 113 static inline unsigned long __qdf_find_first_bit(unsigned long *addr, 114 unsigned long nbits) 115 { 116 return find_first_bit(addr, nbits); 117 } 118 119 static inline bool __qdf_bitmap_empty(unsigned long *addr, 120 unsigned long nbits) 121 { 122 return bitmap_empty(addr, nbits); 123 } 124 125 static inline int __qdf_bitmap_and(unsigned long *dst, unsigned long *src1, 126 unsigned long *src2, unsigned long nbits) 127 { 128 return bitmap_and(dst, src1, src2, nbits); 129 } 130 131 /** 132 * __qdf_set_macaddr_broadcast() - set a QDF MacAddress to the 'broadcast' 133 * @mac_addr: pointer to the qdf MacAddress to set to broadcast 134 * 135 * This function sets a QDF MacAddress to the 'broadcast' MacAddress. Broadcast 136 * MacAddress contains all 0xFF bytes. 137 * 138 * Return: none 139 */ 140 static inline void __qdf_set_macaddr_broadcast(struct qdf_mac_addr *mac_addr) 141 { 142 memset(mac_addr, 0xff, QDF_MAC_ADDR_SIZE); 143 } 144 145 /** 146 * __qdf_zero_macaddr() - zero out a MacAddress 147 * @mac_addr: pointer to the struct qdf_mac_addr to zero. 148 * 149 * This function zeros out a QDF MacAddress type. 150 * 151 * Return: none 152 */ 153 static inline void __qdf_zero_macaddr(struct qdf_mac_addr *mac_addr) 154 { 155 memset(mac_addr, 0, QDF_MAC_ADDR_SIZE); 156 } 157 158 /** 159 * __qdf_is_macaddr_equal() - compare two QDF MacAddress 160 * @mac_addr1: Pointer to one qdf MacAddress to compare 161 * @mac_addr2: Pointer to the other qdf MacAddress to compare 162 * 163 * This function returns a bool that tells if a two QDF MacAddress' 164 * are equivalent. 165 * 166 * Return: true if the MacAddress's are equal 167 * not true if the MacAddress's are not equal 168 */ 169 static inline bool __qdf_is_macaddr_equal(const struct qdf_mac_addr *mac_addr1, 170 const struct qdf_mac_addr *mac_addr2) 171 { 172 return 0 == memcmp(mac_addr1, mac_addr2, QDF_MAC_ADDR_SIZE); 173 } 174 175 #define __qdf_in_interrupt in_interrupt 176 177 #define __qdf_min(_a, _b) min(_a, _b) 178 #define __qdf_max(_a, _b) max(_a, _b) 179 180 /* 181 * Setting it to blank as feature is not intended to be supported 182 * on linux version less than 4.3 183 */ 184 #if LINUX_VERSION_CODE < KERNEL_VERSION(4, 3, 0) 185 #define __QDF_DECLARE_EWMA(name, _factor, _weight) 186 187 #define __qdf_ewma_tx_lag int 188 #define __qdf_ewma_rx_rssi int 189 #else 190 #if LINUX_VERSION_CODE >= KERNEL_VERSION(4, 11, 0) 191 #define __QDF_DECLARE_EWMA(name, _factor, _weight) \ 192 DECLARE_EWMA(name, ilog2(_factor), _weight) 193 #else 194 #define __QDF_DECLARE_EWMA(name, _factor, _weight) \ 195 DECLARE_EWMA(name, _factor, _weight) 196 #endif 197 198 #define __qdf_ewma_tx_lag struct ewma_tx_lag 199 #define __qdf_ewma_rx_rssi struct ewma_rx_rssi 200 #endif 201 202 #define __qdf_ffz(mask) (~(mask) == 0 ? -1 : ffz(mask)) 203 204 #define MEMINFO_KB(x) ((x) << (PAGE_SHIFT - 10)) /* In kilobytes */ 205 206 #define __qdf_assert(expr) do { \ 207 if (unlikely(!(expr))) { \ 208 pr_err("Assertion failed! %s:%s %s:%d\n", \ 209 # expr, __func__, __FILE__, __LINE__); \ 210 dump_stack(); \ 211 QDF_BUG_ON_ASSERT(0); \ 212 } \ 213 } while (0) 214 215 #define __qdf_target_assert(expr) do { \ 216 if (unlikely(!(expr))) { \ 217 qdf_err("Assertion failed! %s:%s %s:%d", \ 218 #expr, __FUNCTION__, __FILE__, __LINE__); \ 219 dump_stack(); \ 220 QDF_DEBUG_PANIC("Take care of the TARGET ASSERT first\n"); \ 221 } \ 222 } while (0) 223 224 #define QDF_COMPILE_TIME_ASSERT(assertion_name, predicate) \ 225 typedef char assertion_name[(predicate) ? 1 : -1] 226 227 #define __qdf_container_of(ptr, type, member) container_of(ptr, type, member) 228 229 #define __qdf_ntohs ntohs 230 #define __qdf_ntohl ntohl 231 232 #define __qdf_htons htons 233 #define __qdf_htonl htonl 234 235 #define __qdf_cpu_to_le16 cpu_to_le16 236 #define __qdf_cpu_to_le32 cpu_to_le32 237 #define __qdf_cpu_to_le64 cpu_to_le64 238 239 #define __qdf_le16_to_cpu le16_to_cpu 240 #define __qdf_le32_to_cpu le32_to_cpu 241 #define __qdf_le64_to_cpu le64_to_cpu 242 243 #define __qdf_cpu_to_be16 cpu_to_be16 244 #define __qdf_cpu_to_be32 cpu_to_be32 245 #define __qdf_cpu_to_be64 cpu_to_be64 246 247 #define __qdf_be16_to_cpu be16_to_cpu 248 #define __qdf_be32_to_cpu be32_to_cpu 249 #define __qdf_be64_to_cpu be64_to_cpu 250 251 #define __qdf_wmb() wmb() 252 #define __qdf_rmb() rmb() 253 #define __qdf_mb() mb() 254 #define __qdf_ioread32(offset) ioread32(offset) 255 #define __qdf_iowrite32(offset, value) iowrite32(value, offset) 256 257 #define __qdf_roundup(x, y) roundup(x, y) 258 #define __qdf_ceil(x, y) DIV_ROUND_UP(x, y) 259 #define __qdf_abs(x) abs(x) 260 261 #if LINUX_VERSION_CODE < KERNEL_VERSION(4, 3, 0) 262 #define __qdf_ewma_tx_lag_init(tx_lag) 263 #define __qdf_ewma_tx_lag_add(tx_lag, value) 264 #define __qdf_ewma_tx_lag_read(tx_lag) 265 266 #define __qdf_ewma_rx_rssi_init(rx_rssi) 267 #define __qdf_ewma_rx_rssi_add(rx_rssi, value) 268 #define __qdf_ewma_rx_rssi_read(rx_rssi) 269 #else 270 #define __qdf_ewma_tx_lag_init(tx_lag) \ 271 ewma_tx_lag_init(tx_lag) 272 273 #define __qdf_ewma_tx_lag_add(tx_lag, value) \ 274 ewma_tx_lag_add(tx_lag, value) 275 276 #define __qdf_ewma_tx_lag_read(tx_lag) \ 277 ewma_tx_lag_read(tx_lag) 278 279 #define __qdf_ewma_rx_rssi_init(rx_rssi) \ 280 ewma_rx_rssi_init(rx_rssi) 281 282 #define __qdf_ewma_rx_rssi_add(rx_rssi, value) \ 283 ewma_rx_rssi_add(rx_rssi, value) 284 285 #define __qdf_ewma_rx_rssi_read(rx_rssi) \ 286 ewma_rx_rssi_read(rx_rssi) 287 #endif 288 289 #define __qdf_prefetch(x) prefetch(x) 290 291 #ifdef QCA_CONFIG_SMP 292 /** 293 * __qdf_get_cpu() - get cpu_index 294 * 295 * Return: cpu_index 296 */ 297 static inline 298 int __qdf_get_cpu(void) 299 { 300 int cpu_index = get_cpu(); 301 302 put_cpu(); 303 return cpu_index; 304 } 305 #else 306 static inline 307 int __qdf_get_cpu(void) 308 { 309 return 0; 310 } 311 #endif 312 313 static inline int __qdf_device_init_wakeup(__qdf_device_t qdf_dev, bool enable) 314 { 315 return device_init_wakeup(qdf_dev->dev, enable); 316 } 317 318 /** 319 * __qdf_get_totalramsize() - Get total ram size in Kb 320 * 321 * Return: Total ram size in Kb 322 */ 323 static inline uint64_t 324 __qdf_get_totalramsize(void) 325 { 326 struct sysinfo meminfo; 327 328 si_meminfo(&meminfo); 329 return MEMINFO_KB(meminfo.totalram); 330 } 331 332 /** 333 * __qdf_get_lower_32_bits() - get lower 32 bits from an address. 334 * @addr: address 335 * 336 * This api returns the lower 32 bits of an address. 337 * 338 * Return: lower 32 bits. 339 */ 340 static inline 341 uint32_t __qdf_get_lower_32_bits(__qdf_dma_addr_t addr) 342 { 343 return lower_32_bits(addr); 344 } 345 346 /** 347 * __qdf_get_upper_32_bits() - get upper 32 bits from an address. 348 * @addr: address 349 * 350 * This api returns the upper 32 bits of an address. 351 * 352 * Return: upper 32 bits. 353 */ 354 static inline 355 uint32_t __qdf_get_upper_32_bits(__qdf_dma_addr_t addr) 356 { 357 return upper_32_bits(addr); 358 } 359 360 /** 361 * __qdf_rounddown_pow_of_two() - Round down to nearest power of two 362 * @n: number to be tested 363 * 364 * Test if the input number is power of two, and return the nearest power of two 365 * 366 * Return: number rounded down to the nearest power of two 367 */ 368 static inline 369 unsigned long __qdf_rounddown_pow_of_two(unsigned long n) 370 { 371 if (is_power_of_2(n)) 372 return n; /* already a power of 2 */ 373 374 return __rounddown_pow_of_two(n); 375 } 376 377 #if LINUX_VERSION_CODE >= KERNEL_VERSION(3, 13, 0) 378 379 /** 380 * __qdf_set_dma_coherent_mask() - set max number of bits allowed in dma addr 381 * @dev: device pointer 382 * @addr_bits: max number of bits allowed in dma address 383 * 384 * This API sets the maximum allowed number of bits in the dma address. 385 * 386 * Return: 0 - success, non zero - failure 387 */ 388 static inline 389 int __qdf_set_dma_coherent_mask(struct device *dev, uint8_t addr_bits) 390 { 391 return dma_set_mask_and_coherent(dev, DMA_BIT_MASK(addr_bits)); 392 } 393 394 #else 395 396 /** 397 * __qdf_set_dma_coherent_mask() - set max number of bits allowed in dma addr 398 * @dev: device pointer 399 * @addr_bits: max number of bits allowed in dma address 400 * 401 * This API sets the maximum allowed number of bits in the dma address. 402 * 403 * Return: 0 - success, non zero - failure 404 */ 405 static inline 406 int __qdf_set_dma_coherent_mask(struct device *dev, uint8_t addr_bits) 407 { 408 return dma_set_coherent_mask(dev, DMA_BIT_MASK(addr_bits)); 409 } 410 #endif 411 /** 412 * __qdf_get_random_bytes() - returns nbytes bytes of random data 413 * @buf: buffer to fill 414 * @nbytes: number of bytes to fill 415 * 416 * Return: void 417 */ 418 static inline 419 void __qdf_get_random_bytes(void *buf, int nbytes) 420 { 421 return get_random_bytes(buf, nbytes); 422 } 423 424 /** 425 * __qdf_do_div() - wrapper function for kernel macro(do_div). 426 * @dividend: Dividend value 427 * @divisor : Divisor value 428 * 429 * Return: Quotient 430 */ 431 static inline 432 uint64_t __qdf_do_div(uint64_t dividend, uint32_t divisor) 433 { 434 do_div(dividend, divisor); 435 /*do_div macro updates dividend with Quotient of dividend/divisor */ 436 return dividend; 437 } 438 439 /** 440 * __qdf_do_div_rem() - wrapper function for kernel macro(do_div) 441 * to get remainder. 442 * @dividend: Dividend value 443 * @divisor : Divisor value 444 * 445 * Return: remainder 446 */ 447 static inline 448 uint64_t __qdf_do_div_rem(uint64_t dividend, uint32_t divisor) 449 { 450 return do_div(dividend, divisor); 451 } 452 453 /** 454 * __qdf_hex_to_bin() - Wrapper function to kernel API to get unsigned 455 * integer from hexa decimal ASCII character. 456 * @ch: hexa decimal ASCII character 457 * 458 * Return: For hexa decimal ASCII char return actual decimal value 459 * else -1 for bad input. 460 */ 461 static inline 462 int __qdf_hex_to_bin(char ch) 463 { 464 return hex_to_bin(ch); 465 } 466 467 /** 468 * __qdf_hex_str_to_binary() - Wrapper function to get array of unsigned 469 * integers from string of hexa decimal ASCII characters. 470 * @dst: output array to hold converted values 471 * @src: input string of hexa decimal ASCII characters 472 * @count: size of dst string 473 * 474 * Return: For a string of hexa decimal ASCII characters return 0 475 * else -1 for bad input. 476 */ 477 static inline 478 int __qdf_hex_str_to_binary(u8 *dst, const char *src, size_t count) 479 { 480 return hex2bin(dst, src, count); 481 } 482 483 /** 484 * __qdf_fls() - find last set bit in a given 32 bit input 485 * @x: 32 bit mask 486 * 487 * Return: zero if the input is zero, otherwise returns the bit 488 * position of the last set bit, where the LSB is 1 and MSB is 32. 489 */ 490 static inline 491 int __qdf_fls(uint32_t x) 492 { 493 return fls(x); 494 } 495 496 /** 497 * __qdf_get_smp_processor_id() - Get the current CPU id 498 * 499 * Return: current CPU id 500 */ 501 static inline int __qdf_get_smp_processor_id(void) 502 { 503 return smp_processor_id(); 504 } 505 506 /** 507 * __qdf_in_atomic: Check whether current thread running in atomic context 508 * 509 * Return: true if current thread is running in the atomic context 510 * else it will be return false. 511 */ 512 static inline bool __qdf_in_atomic(void) 513 { 514 if (in_interrupt() || !preemptible() || rcu_preempt_depth()) 515 return true; 516 517 return false; 518 } 519 520 #endif /*_I_QDF_UTIL_H*/ 521