lib: sbi: Replace test-and-set locks by ticket locks
Replace the test-and-set spinlock implementation with ticket locks in order to get fairness (in form of FIFO order). The implementation uses a 32-bit wide struct, which consists of two 16-bit counters (owner and next). This is inspired by similar spinlock implementations on other architectures. This allows that the code works for both, RV32 and RV64. Signed-off-by: Christoph Muellner <cmuellner@linux.com> Reviewed-by: Anup Patel <anup.patel@wdc.com> Reviewed-by: Xiang W <wxjstz@126.com>
This commit is contained in:
parent
d0e406fa44
commit
4d8e2f135d
|
@ -2,26 +2,37 @@
|
||||||
* SPDX-License-Identifier: BSD-2-Clause
|
* SPDX-License-Identifier: BSD-2-Clause
|
||||||
*
|
*
|
||||||
* Copyright (c) 2019 Western Digital Corporation or its affiliates.
|
* Copyright (c) 2019 Western Digital Corporation or its affiliates.
|
||||||
*
|
* Copyright (c) 2021 Christoph Müllner <cmuellner@linux.com>
|
||||||
* Authors:
|
|
||||||
* Anup Patel <anup.patel@wdc.com>
|
|
||||||
*/
|
*/
|
||||||
|
|
||||||
#ifndef __RISCV_LOCKS_H__
|
#ifndef __RISCV_LOCKS_H__
|
||||||
#define __RISCV_LOCKS_H__
|
#define __RISCV_LOCKS_H__
|
||||||
|
|
||||||
|
#include <sbi/sbi_types.h>
|
||||||
|
|
||||||
|
#define TICKET_SHIFT 16
|
||||||
|
|
||||||
typedef struct {
|
typedef struct {
|
||||||
volatile long lock;
|
#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
|
||||||
} spinlock_t;
|
u16 next;
|
||||||
|
u16 owner;
|
||||||
|
#else
|
||||||
|
u16 owner;
|
||||||
|
u16 next;
|
||||||
|
#endif
|
||||||
|
} __aligned(4) spinlock_t;
|
||||||
|
|
||||||
#define __RISCV_SPIN_UNLOCKED 0
|
#define __SPIN_LOCK_UNLOCKED \
|
||||||
|
(spinlock_t) { 0, 0 }
|
||||||
|
|
||||||
#define SPIN_LOCK_INIT(x) (x).lock = __RISCV_SPIN_UNLOCKED
|
#define SPIN_LOCK_INIT(x) \
|
||||||
|
x = __SPIN_LOCK_UNLOCKED
|
||||||
|
|
||||||
#define SPIN_LOCK_INITIALIZER \
|
#define SPIN_LOCK_INITIALIZER \
|
||||||
{ \
|
__SPIN_LOCK_UNLOCKED
|
||||||
.lock = __RISCV_SPIN_UNLOCKED, \
|
|
||||||
}
|
#define DEFINE_SPIN_LOCK(x) \
|
||||||
|
spinlock_t SPIN_LOCK_INIT(x)
|
||||||
|
|
||||||
int spin_lock_check(spinlock_t *lock);
|
int spin_lock_check(spinlock_t *lock);
|
||||||
|
|
||||||
|
|
|
@ -2,44 +2,76 @@
|
||||||
* SPDX-License-Identifier: BSD-2-Clause
|
* SPDX-License-Identifier: BSD-2-Clause
|
||||||
*
|
*
|
||||||
* Copyright (c) 2019 Western Digital Corporation or its affiliates.
|
* Copyright (c) 2019 Western Digital Corporation or its affiliates.
|
||||||
*
|
* Copyright (c) 2021 Christoph Müllner <cmuellner@linux.com>
|
||||||
* Authors:
|
|
||||||
* Anup Patel <anup.patel@wdc.com>
|
|
||||||
*/
|
*/
|
||||||
|
|
||||||
#include <sbi/riscv_barrier.h>
|
#include <sbi/riscv_barrier.h>
|
||||||
#include <sbi/riscv_locks.h>
|
#include <sbi/riscv_locks.h>
|
||||||
|
|
||||||
int spin_lock_check(spinlock_t *lock)
|
static inline int spin_lock_unlocked(spinlock_t lock)
|
||||||
{
|
{
|
||||||
return (lock->lock == __RISCV_SPIN_UNLOCKED) ? 0 : 1;
|
return lock.owner == lock.next;
|
||||||
|
}
|
||||||
|
|
||||||
|
bool spin_lock_check(spinlock_t *lock)
|
||||||
|
{
|
||||||
|
RISCV_FENCE(r, rw);
|
||||||
|
return !spin_lock_unlocked(*lock);
|
||||||
}
|
}
|
||||||
|
|
||||||
int spin_trylock(spinlock_t *lock)
|
int spin_trylock(spinlock_t *lock)
|
||||||
{
|
{
|
||||||
int tmp = 1, busy;
|
unsigned long inc = 1u << TICKET_SHIFT;
|
||||||
|
unsigned long mask = 0xffffu << TICKET_SHIFT;
|
||||||
|
u32 l0, tmp1, tmp2;
|
||||||
|
|
||||||
__asm__ __volatile__(
|
__asm__ __volatile__(
|
||||||
" amoswap.w %0, %2, %1\n" RISCV_ACQUIRE_BARRIER
|
/* Get the current lock counters. */
|
||||||
: "=r"(busy), "+A"(lock->lock)
|
"1: lr.w.aq %0, %3\n"
|
||||||
: "r"(tmp)
|
" slli %2, %0, %6\n"
|
||||||
|
" and %2, %2, %5\n"
|
||||||
|
" and %1, %0, %5\n"
|
||||||
|
/* Is the lock free right now? */
|
||||||
|
" bne %1, %2, 2f\n"
|
||||||
|
" add %0, %0, %4\n"
|
||||||
|
/* Acquire the lock. */
|
||||||
|
" sc.w.rl %0, %0, %3\n"
|
||||||
|
" bnez %0, 1b\n"
|
||||||
|
"2:"
|
||||||
|
: "=&r"(l0), "=&r"(tmp1), "=&r"(tmp2), "+A"(*lock)
|
||||||
|
: "r"(inc), "r"(mask), "I"(TICKET_SHIFT)
|
||||||
: "memory");
|
: "memory");
|
||||||
|
|
||||||
return !busy;
|
return !l0;
|
||||||
}
|
}
|
||||||
|
|
||||||
void spin_lock(spinlock_t *lock)
|
void spin_lock(spinlock_t *lock)
|
||||||
{
|
{
|
||||||
while (1) {
|
unsigned long inc = 1u << TICKET_SHIFT;
|
||||||
if (spin_lock_check(lock))
|
unsigned long mask = 0xffffu;
|
||||||
continue;
|
u32 l0, tmp1, tmp2;
|
||||||
|
|
||||||
if (spin_trylock(lock))
|
__asm__ __volatile__(
|
||||||
break;
|
/* Atomically increment the next ticket. */
|
||||||
}
|
" amoadd.w.aqrl %0, %4, %3\n"
|
||||||
|
|
||||||
|
/* Did we get the lock? */
|
||||||
|
" srli %1, %0, %6\n"
|
||||||
|
" and %1, %1, %5\n"
|
||||||
|
"1: and %2, %0, %5\n"
|
||||||
|
" beq %1, %2, 2f\n"
|
||||||
|
|
||||||
|
/* If not, then spin on the lock. */
|
||||||
|
" lw %0, %3\n"
|
||||||
|
RISCV_ACQUIRE_BARRIER
|
||||||
|
" j 1b\n"
|
||||||
|
"2:"
|
||||||
|
: "=&r"(l0), "=&r"(tmp1), "=&r"(tmp2), "+A"(*lock)
|
||||||
|
: "r"(inc), "r"(mask), "I"(TICKET_SHIFT)
|
||||||
|
: "memory");
|
||||||
}
|
}
|
||||||
|
|
||||||
void spin_unlock(spinlock_t *lock)
|
void spin_unlock(spinlock_t *lock)
|
||||||
{
|
{
|
||||||
__smp_store_release(&lock->lock, __RISCV_SPIN_UNLOCKED);
|
__smp_store_release(&lock->owner, lock->owner + 1);
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in New Issue