123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657 |
- /* SPDX-License-Identifier: GPL-2.0-only
- * Copyright (C) 2020 Marvell.
- */
- #ifndef __SOC_OTX2_ASM_H
- #define __SOC_OTX2_ASM_H
- #include <linux/types.h>
- #if defined(CONFIG_ARM64)
- /*
- * otx2_lmt_flush is used for LMT store operation.
- * On octeontx2 platform CPT instruction enqueue and
- * NIX packet send are only possible via LMTST
- * operations and it uses LDEOR instruction targeting
- * the coprocessor address.
- */
- #define otx2_lmt_flush(ioaddr) \
- ({ \
- u64 result = 0; \
- __asm__ volatile(".cpu generic+lse\n" \
- "ldeor xzr, %x[rf], [%[rs]]" \
- : [rf]"=r" (result) \
- : [rs]"r" (ioaddr)); \
- (result); \
- })
- /*
- * STEORL store to memory with release semantics.
- * This will avoid using DMB barrier after each LMTST
- * operation.
- */
- #define cn10k_lmt_flush(val, addr) \
- ({ \
- __asm__ volatile(".cpu generic+lse\n" \
- "steorl %x[rf],[%[rs]]" \
- : [rf] "+r"(val) \
- : [rs] "r"(addr)); \
- })
- static inline u64 otx2_atomic64_fetch_add(u64 incr, u64 *ptr)
- {
- u64 result;
- asm volatile (".cpu generic+lse\n"
- "ldadda %x[i], %x[r], [%[b]]"
- : [r] "=r" (result), "+m" (*ptr)
- : [i] "r" (incr), [b] "r" (ptr)
- : "memory");
- return result;
- }
- #else
- #define otx2_lmt_flush(ioaddr) ({ 0; })
- #define cn10k_lmt_flush(val, addr) ({ addr = val; })
- #define otx2_atomic64_fetch_add(incr, ptr) ({ incr; })
- #endif
- #endif /* __SOC_OTX2_ASM_H */
|