Quellcodebibliothek Statistik Leitseite products/sources/formale Sprachen/C/Linux/arch/loongarch/include/asm/   (Open Source Betriebssystem Version 6.17.9©)  Datei vom 24.10.2025 mit Größe 10 kB image not shown  

Quelle  atomic.h   Sprache: C

 
/* SPDX-License-Identifier: GPL-2.0 */
/*
 * Atomic operations.
 *
 * Copyright (C) 2020-2022 Loongson Technology Corporation Limited
 */

#ifndef _ASM_ATOMIC_H
#define _ASM_ATOMIC_H

#include <linux/types.h>
#include <asm/barrier.h>
#include <asm/cmpxchg.h>

#if __SIZEOF_LONG__ == 4
#define __LL  "ll.w "
#define __SC  "sc.w "
#define __AMADD  "amadd.w "
#define __AMOR  "amor.w "
#define __AMAND_DB "amand_db.w "
#define __AMOR_DB "amor_db.w "
#define __AMXOR_DB "amxor_db.w "
#elif __SIZEOF_LONG__ == 8
#define __LL  "ll.d "
#define __SC  "sc.d "
#define __AMADD  "amadd.d "
#define __AMOR  "amor.d "
#define __AMAND_DB "amand_db.d "
#define __AMOR_DB "amor_db.d "
#define __AMXOR_DB "amxor_db.d "
#endif

#define ATOMIC_INIT(i)   { (i) }

#define arch_atomic_read(v) READ_ONCE((v)->counter)
#define arch_atomic_set(v, i) WRITE_ONCE((v)->counter, (i))

#define ATOMIC_OP(op, I, asm_op)     \
static inline void arch_atomic_##op(int i, atomic_t *v)   \
{         \
 __asm__ __volatile__(      \
 "am"#asm_op".w" " $zero, %1, %0 \n"    \
 : "+ZB" (v->counter)      \
 : "r" (I)       \
 : "memory");       \
}

#define ATOMIC_OP_RETURN(op, I, asm_op, c_op, mb, suffix)  \
static inline int arch_atomic_##op##_return##suffix(int i, atomic_t *v) \
{         \
 int result;       \
         \
 __asm__ __volatile__(      \
 "am"#asm_op#mb".w" " %1, %2, %0 \n"   \
 : "+ZB" (v->counter), "=&r" (result)    \
 : "r" (I)       \
 : "memory");       \
         \
 return result c_op I;      \
}

#define ATOMIC_FETCH_OP(op, I, asm_op, mb, suffix)   \
static inline int arch_atomic_fetch_##op##suffix(int i, atomic_t *v) \
{         \
 int result;       \
         \
 __asm__ __volatile__(      \
 "am"#asm_op#mb".w" " %1, %2, %0 \n"   \
 : "+ZB" (v->counter), "=&r" (result)    \
 : "r" (I)       \
 : "memory");       \
         \
 return result;       \
}

#define ATOMIC_OPS(op, I, asm_op, c_op)     \
 ATOMIC_OP(op, I, asm_op)     \
 ATOMIC_OP_RETURN(op, I, asm_op, c_op, _db,         )  \
 ATOMIC_OP_RETURN(op, I, asm_op, c_op,    , _relaxed)  \
 ATOMIC_FETCH_OP(op, I, asm_op, _db,         )   \
 ATOMIC_FETCH_OP(op, I, asm_op,    , _relaxed)

ATOMIC_OPS(add, i, add, +)
ATOMIC_OPS(sub, -i, add, +)

#define arch_atomic_add_return  arch_atomic_add_return
#define arch_atomic_add_return_acquire arch_atomic_add_return
#define arch_atomic_add_return_release arch_atomic_add_return
#define arch_atomic_add_return_relaxed arch_atomic_add_return_relaxed
#define arch_atomic_sub_return  arch_atomic_sub_return
#define arch_atomic_sub_return_acquire arch_atomic_sub_return
#define arch_atomic_sub_return_release arch_atomic_sub_return
#define arch_atomic_sub_return_relaxed arch_atomic_sub_return_relaxed
#define arch_atomic_fetch_add  arch_atomic_fetch_add
#define arch_atomic_fetch_add_acquire arch_atomic_fetch_add
#define arch_atomic_fetch_add_release arch_atomic_fetch_add
#define arch_atomic_fetch_add_relaxed arch_atomic_fetch_add_relaxed
#define arch_atomic_fetch_sub  arch_atomic_fetch_sub
#define arch_atomic_fetch_sub_acquire arch_atomic_fetch_sub
#define arch_atomic_fetch_sub_release arch_atomic_fetch_sub
#define arch_atomic_fetch_sub_relaxed arch_atomic_fetch_sub_relaxed

#undef ATOMIC_OPS

#define ATOMIC_OPS(op, I, asm_op)     \
 ATOMIC_OP(op, I, asm_op)     \
 ATOMIC_FETCH_OP(op, I, asm_op, _db,         )   \
 ATOMIC_FETCH_OP(op, I, asm_op,    , _relaxed)

ATOMIC_OPS(and, i, and)
ATOMIC_OPS(or, i, or)
ATOMIC_OPS(xor, i, xor)

#define arch_atomic_fetch_and  arch_atomic_fetch_and
#define arch_atomic_fetch_and_acquire arch_atomic_fetch_and
#define arch_atomic_fetch_and_release arch_atomic_fetch_and
#define arch_atomic_fetch_and_relaxed arch_atomic_fetch_and_relaxed
#define arch_atomic_fetch_or  arch_atomic_fetch_or
#define arch_atomic_fetch_or_acquire arch_atomic_fetch_or
#define arch_atomic_fetch_or_release arch_atomic_fetch_or
#define arch_atomic_fetch_or_relaxed arch_atomic_fetch_or_relaxed
#define arch_atomic_fetch_xor  arch_atomic_fetch_xor
#define arch_atomic_fetch_xor_acquire arch_atomic_fetch_xor
#define arch_atomic_fetch_xor_release arch_atomic_fetch_xor
#define arch_atomic_fetch_xor_relaxed arch_atomic_fetch_xor_relaxed

#undef ATOMIC_OPS
#undef ATOMIC_FETCH_OP
#undef ATOMIC_OP_RETURN
#undef ATOMIC_OP

static inline int arch_atomic_fetch_add_unless(atomic_t *v, int a, int u)
{
       int prev, rc;

 __asm__ __volatile__ (
  "0: ll.w %[p], %[c]\n"
  " beq %[p], %[u], 1f\n"
  " add.w %[rc], %[p], %[a]\n"
  " sc.w %[rc], %[c]\n"
  " beqz %[rc], 0b\n"
  " b 2f\n"
  "1:\n"
  __WEAK_LLSC_MB
  "2:\n"
  : [p]"=&r" (prev), [rc]"=&r" (rc),
    [c]"=ZB" (v->counter)
  : [a]"r" (a), [u]"r" (u)
  : "memory");

 return prev;
}
#define arch_atomic_fetch_add_unless arch_atomic_fetch_add_unless

static inline int arch_atomic_sub_if_positive(int i, atomic_t *v)
{
 int result;
 int temp;

 if (__builtin_constant_p(i)) {
  __asm__ __volatile__(
  "1: ll.w %1, %2 # atomic_sub_if_positive\n"
  " addi.w %0, %1, %3 \n"
  " move %1, %0 \n"
  " bltz %0, 2f \n"
  " sc.w %1, %2 \n"
  " beqz %1, 1b \n"
  "2: \n"
  __WEAK_LLSC_MB
  : "=&r" (result), "=&r" (temp), "+ZC" (v->counter)
  : "I" (-i));
 } else {
  __asm__ __volatile__(
  "1: ll.w %1, %2 # atomic_sub_if_positive\n"
  " sub.w %0, %1, %3 \n"
  " move %1, %0 \n"
  " bltz %0, 2f \n"
  " sc.w %1, %2 \n"
  " beqz %1, 1b \n"
  "2: \n"
  __WEAK_LLSC_MB
  : "=&r" (result), "=&r" (temp), "+ZC" (v->counter)
  : "r" (i));
 }

 return result;
}

#define arch_atomic_dec_if_positive(v) arch_atomic_sub_if_positive(1, v)

#ifdef CONFIG_64BIT

#define ATOMIC64_INIT(i)    { (i) }

#define arch_atomic64_read(v) READ_ONCE((v)->counter)
#define arch_atomic64_set(v, i) WRITE_ONCE((v)->counter, (i))

#define ATOMIC64_OP(op, I, asm_op)     \
static inline void arch_atomic64_##op(long i, atomic64_t *v)  \
{         \
 __asm__ __volatile__(      \
 "am"#asm_op".d " " $zero, %1, %0 \n"   \
 : "+ZB" (v->counter)      \
 : "r" (I)       \
 : "memory");       \
}

#define ATOMIC64_OP_RETURN(op, I, asm_op, c_op, mb, suffix)   \
static inline long arch_atomic64_##op##_return##suffix(long i, atomic64_t *v) \
{          \
 long result;        \
 __asm__ __volatile__(       \
 "am"#asm_op#mb".d " " %1, %2, %0 \n"   \
 : "+ZB" (v->counter), "=&r" (result)     \
 : "r" (I)        \
 : "memory");        \
          \
 return result c_op I;       \
}

#define ATOMIC64_FETCH_OP(op, I, asm_op, mb, suffix)    \
static inline long arch_atomic64_fetch_##op##suffix(long i, atomic64_t *v) \
{          \
 long result;        \
          \
 __asm__ __volatile__(       \
 "am"#asm_op#mb".d " " %1, %2, %0 \n"   \
 : "+ZB" (v->counter), "=&r" (result)     \
 : "r" (I)        \
 : "memory");        \
          \
 return result;        \
}

#define ATOMIC64_OPS(op, I, asm_op, c_op)          \
 ATOMIC64_OP(op, I, asm_op)           \
 ATOMIC64_OP_RETURN(op, I, asm_op, c_op, _db,         )        \
 ATOMIC64_OP_RETURN(op, I, asm_op, c_op,    , _relaxed)        \
 ATOMIC64_FETCH_OP(op, I, asm_op, _db,         )         \
 ATOMIC64_FETCH_OP(op, I, asm_op,    , _relaxed)

ATOMIC64_OPS(add, i, add, +)
ATOMIC64_OPS(sub, -i, add, +)

#define arch_atomic64_add_return  arch_atomic64_add_return
#define arch_atomic64_add_return_acquire arch_atomic64_add_return
#define arch_atomic64_add_return_release arch_atomic64_add_return
#define arch_atomic64_add_return_relaxed arch_atomic64_add_return_relaxed
#define arch_atomic64_sub_return  arch_atomic64_sub_return
#define arch_atomic64_sub_return_acquire arch_atomic64_sub_return
#define arch_atomic64_sub_return_release arch_atomic64_sub_return
#define arch_atomic64_sub_return_relaxed arch_atomic64_sub_return_relaxed
#define arch_atomic64_fetch_add   arch_atomic64_fetch_add
#define arch_atomic64_fetch_add_acquire  arch_atomic64_fetch_add
#define arch_atomic64_fetch_add_release  arch_atomic64_fetch_add
#define arch_atomic64_fetch_add_relaxed  arch_atomic64_fetch_add_relaxed
#define arch_atomic64_fetch_sub   arch_atomic64_fetch_sub
#define arch_atomic64_fetch_sub_acquire  arch_atomic64_fetch_sub
#define arch_atomic64_fetch_sub_release  arch_atomic64_fetch_sub
#define arch_atomic64_fetch_sub_relaxed  arch_atomic64_fetch_sub_relaxed

#undef ATOMIC64_OPS

#define ATOMIC64_OPS(op, I, asm_op)           \
 ATOMIC64_OP(op, I, asm_op)           \
 ATOMIC64_FETCH_OP(op, I, asm_op, _db,         )         \
 ATOMIC64_FETCH_OP(op, I, asm_op,    , _relaxed)

ATOMIC64_OPS(and, i, and)
ATOMIC64_OPS(or, i, or)
ATOMIC64_OPS(xor, i, xor)

#define arch_atomic64_fetch_and  arch_atomic64_fetch_and
#define arch_atomic64_fetch_and_acquire arch_atomic64_fetch_and
#define arch_atomic64_fetch_and_release arch_atomic64_fetch_and
#define arch_atomic64_fetch_and_relaxed arch_atomic64_fetch_and_relaxed
#define arch_atomic64_fetch_or  arch_atomic64_fetch_or
#define arch_atomic64_fetch_or_acquire arch_atomic64_fetch_or
#define arch_atomic64_fetch_or_release arch_atomic64_fetch_or
#define arch_atomic64_fetch_or_relaxed arch_atomic64_fetch_or_relaxed
#define arch_atomic64_fetch_xor  arch_atomic64_fetch_xor
#define arch_atomic64_fetch_xor_acquire arch_atomic64_fetch_xor
#define arch_atomic64_fetch_xor_release arch_atomic64_fetch_xor
#define arch_atomic64_fetch_xor_relaxed arch_atomic64_fetch_xor_relaxed

#undef ATOMIC64_OPS
#undef ATOMIC64_FETCH_OP
#undef ATOMIC64_OP_RETURN
#undef ATOMIC64_OP

static inline long arch_atomic64_fetch_add_unless(atomic64_t *v, long a, long u)
{
       long prev, rc;

 __asm__ __volatile__ (
  "0: ll.d %[p], %[c]\n"
  " beq %[p], %[u], 1f\n"
  " add.d %[rc], %[p], %[a]\n"
  " sc.d %[rc], %[c]\n"
  " beqz %[rc], 0b\n"
  " b 2f\n"
  "1:\n"
  __WEAK_LLSC_MB
  "2:\n"
  : [p]"=&r" (prev), [rc]"=&r" (rc),
    [c] "=ZB" (v->counter)
  : [a]"r" (a), [u]"r" (u)
  : "memory");

 return prev;
}
#define arch_atomic64_fetch_add_unless arch_atomic64_fetch_add_unless

static inline long arch_atomic64_sub_if_positive(long i, atomic64_t *v)
{
 long result;
 long temp;

 if (__builtin_constant_p(i)) {
  __asm__ __volatile__(
  "1: ll.d %1, %2 # atomic64_sub_if_positive \n"
  " addi.d %0, %1, %3 \n"
  " move %1, %0 \n"
  " bltz %0, 2f \n"
  " sc.d %1, %2 \n"
  " beqz %1, 1b \n"
  "2: \n"
  __WEAK_LLSC_MB
  : "=&r" (result), "=&r" (temp), "+ZC" (v->counter)
  : "I" (-i));
 } else {
  __asm__ __volatile__(
  "1: ll.d %1, %2 # atomic64_sub_if_positive \n"
  " sub.d %0, %1, %3 \n"
  " move %1, %0 \n"
  " bltz %0, 2f \n"
  " sc.d %1, %2 \n"
  " beqz %1, 1b \n"
  "2: \n"
  __WEAK_LLSC_MB
  : "=&r" (result), "=&r" (temp), "+ZC" (v->counter)
  : "r" (i));
 }

 return result;
}

#define arch_atomic64_dec_if_positive(v) arch_atomic64_sub_if_positive(1, v)

#endif /* CONFIG_64BIT */

#endif /* _ASM_ATOMIC_H */

Messung V0.5
C=99 H=95 G=96

¤ Dauer der Verarbeitung: 0.15 Sekunden  (vorverarbeitet)  ¤

*© Formatika GbR, Deutschland






Wurzel

Suchen

Beweissystem der NASA

Beweissystem Isabelle

NIST Cobol Testsuite

Cephes Mathematical Library

Wiener Entwicklungsmethode

Haftungshinweis

Die Informationen auf dieser Webseite wurden nach bestem Wissen sorgfältig zusammengestellt. Es wird jedoch weder Vollständigkeit, noch Richtigkeit, noch Qualität der bereit gestellten Informationen zugesichert.

Bemerkung:

Die farbliche Syntaxdarstellung und die Messung sind noch experimentell.