linux/kernel/locking/semaphore.c
<<
>>
Prefs
   1// SPDX-License-Identifier: GPL-2.0-only
   2/*
   3 * Copyright (c) 2008 Intel Corporation
   4 * Author: Matthew Wilcox <willy@linux.intel.com>
   5 *
   6 * This file implements counting semaphores.
   7 * A counting semaphore may be acquired 'n' times before sleeping.
   8 * See mutex.c for single-acquisition sleeping locks which enforce
   9 * rules which allow code to be debugged more easily.
  10 */
  11
  12/*
  13 * Some notes on the implementation:
  14 *
  15 * The spinlock controls access to the other members of the semaphore.
  16 * down_trylock() and up() can be called from interrupt context, so we
  17 * have to disable interrupts when taking the lock.  It turns out various
  18 * parts of the kernel expect to be able to use down() on a semaphore in
  19 * interrupt context when they know it will succeed, so we have to use
  20 * irqsave variants for down(), down_interruptible() and down_killable()
  21 * too.
  22 *
  23 * The ->count variable represents how many more tasks can acquire this
  24 * semaphore.  If it's zero, there may be tasks waiting on the wait_list.
  25 */
  26
  27#include <linux/compiler.h>
  28#include <linux/kernel.h>
  29#include <linux/export.h>
  30#include <linux/sched.h>
  31#include <linux/sched/debug.h>
  32#include <linux/semaphore.h>
  33#include <linux/spinlock.h>
  34#include <linux/ftrace.h>
  35
  36static noinline void __down(struct semaphore *sem);
  37static noinline int __down_interruptible(struct semaphore *sem);
  38static noinline int __down_killable(struct semaphore *sem);
  39static noinline int __down_timeout(struct semaphore *sem, long timeout);
  40static noinline void __up(struct semaphore *sem);
  41
  42/**
  43 * down - acquire the semaphore
  44 * @sem: the semaphore to be acquired
  45 *
  46 * Acquires the semaphore.  If no more tasks are allowed to acquire the
  47 * semaphore, calling this function will put the task to sleep until the
  48 * semaphore is released.
  49 *
  50 * Use of this function is deprecated, please use down_interruptible() or
  51 * down_killable() instead.
  52 */
  53void down(struct semaphore *sem)
  54{
  55        unsigned long flags;
  56
  57        might_sleep();
  58        raw_spin_lock_irqsave(&sem->lock, flags);
  59        if (likely(sem->count > 0))
  60                sem->count--;
  61        else
  62                __down(sem);
  63        raw_spin_unlock_irqrestore(&sem->lock, flags);
  64}
  65EXPORT_SYMBOL(down);
  66
  67/**
  68 * down_interruptible - acquire the semaphore unless interrupted
  69 * @sem: the semaphore to be acquired
  70 *
  71 * Attempts to acquire the semaphore.  If no more tasks are allowed to
  72 * acquire the semaphore, calling this function will put the task to sleep.
  73 * If the sleep is interrupted by a signal, this function will return -EINTR.
  74 * If the semaphore is successfully acquired, this function returns 0.
  75 */
  76int down_interruptible(struct semaphore *sem)
  77{
  78        unsigned long flags;
  79        int result = 0;
  80
  81        might_sleep();
  82        raw_spin_lock_irqsave(&sem->lock, flags);
  83        if (likely(sem->count > 0))
  84                sem->count--;
  85        else
  86                result = __down_interruptible(sem);
  87        raw_spin_unlock_irqrestore(&sem->lock, flags);
  88
  89        return result;
  90}
  91EXPORT_SYMBOL(down_interruptible);
  92
  93/**
  94 * down_killable - acquire the semaphore unless killed
  95 * @sem: the semaphore to be acquired
  96 *
  97 * Attempts to acquire the semaphore.  If no more tasks are allowed to
  98 * acquire the semaphore, calling this function will put the task to sleep.
  99 * If the sleep is interrupted by a fatal signal, this function will return
 100 * -EINTR.  If the semaphore is successfully acquired, this function returns
 101 * 0.
 102 */
 103int down_killable(struct semaphore *sem)
 104{
 105        unsigned long flags;
 106        int result = 0;
 107
 108        might_sleep();
 109        raw_spin_lock_irqsave(&sem->lock, flags);
 110        if (likely(sem->count > 0))
 111                sem->count--;
 112        else
 113                result = __down_killable(sem);
 114        raw_spin_unlock_irqrestore(&sem->lock, flags);
 115
 116        return result;
 117}
 118EXPORT_SYMBOL(down_killable);
 119
 120/**
 121 * down_trylock - try to acquire the semaphore, without waiting
 122 * @sem: the semaphore to be acquired
 123 *
 124 * Try to acquire the semaphore atomically.  Returns 0 if the semaphore has
 125 * been acquired successfully or 1 if it cannot be acquired.
 126 *
 127 * NOTE: This return value is inverted from both spin_trylock and
 128 * mutex_trylock!  Be careful about this when converting code.
 129 *
 130 * Unlike mutex_trylock, this function can be used from interrupt context,
 131 * and the semaphore can be released by any task or interrupt.
 132 */
 133int down_trylock(struct semaphore *sem)
 134{
 135        unsigned long flags;
 136        int count;
 137
 138        raw_spin_lock_irqsave(&sem->lock, flags);
 139        count = sem->count - 1;
 140        if (likely(count >= 0))
 141                sem->count = count;
 142        raw_spin_unlock_irqrestore(&sem->lock, flags);
 143
 144        return (count < 0);
 145}
 146EXPORT_SYMBOL(down_trylock);
 147
 148/**
 149 * down_timeout - acquire the semaphore within a specified time
 150 * @sem: the semaphore to be acquired
 151 * @timeout: how long to wait before failing
 152 *
 153 * Attempts to acquire the semaphore.  If no more tasks are allowed to
 154 * acquire the semaphore, calling this function will put the task to sleep.
 155 * If the semaphore is not released within the specified number of jiffies,
 156 * this function returns -ETIME.  It returns 0 if the semaphore was acquired.
 157 */
 158int down_timeout(struct semaphore *sem, long timeout)
 159{
 160        unsigned long flags;
 161        int result = 0;
 162
 163        might_sleep();
 164        raw_spin_lock_irqsave(&sem->lock, flags);
 165        if (likely(sem->count > 0))
 166                sem->count--;
 167        else
 168                result = __down_timeout(sem, timeout);
 169        raw_spin_unlock_irqrestore(&sem->lock, flags);
 170
 171        return result;
 172}
 173EXPORT_SYMBOL(down_timeout);
 174
 175/**
 176 * up - release the semaphore
 177 * @sem: the semaphore to release
 178 *
 179 * Release the semaphore.  Unlike mutexes, up() may be called from any
 180 * context and even by tasks which have never called down().
 181 */
 182void up(struct semaphore *sem)
 183{
 184        unsigned long flags;
 185
 186        raw_spin_lock_irqsave(&sem->lock, flags);
 187        if (likely(list_empty(&sem->wait_list)))
 188                sem->count++;
 189        else
 190                __up(sem);
 191        raw_spin_unlock_irqrestore(&sem->lock, flags);
 192}
 193EXPORT_SYMBOL(up);
 194
 195/* Functions for the contended case */
 196
 197struct semaphore_waiter {
 198        struct list_head list;
 199        struct task_struct *task;
 200        bool up;
 201};
 202
 203/*
 204 * Because this function is inlined, the 'state' parameter will be
 205 * constant, and thus optimised away by the compiler.  Likewise the
 206 * 'timeout' parameter for the cases without timeouts.
 207 */
 208static inline int __sched __down_common(struct semaphore *sem, long state,
 209                                                                long timeout)
 210{
 211        struct semaphore_waiter waiter;
 212
 213        list_add_tail(&waiter.list, &sem->wait_list);
 214        waiter.task = current;
 215        waiter.up = false;
 216
 217        for (;;) {
 218                if (signal_pending_state(state, current))
 219                        goto interrupted;
 220                if (unlikely(timeout <= 0))
 221                        goto timed_out;
 222                __set_current_state(state);
 223                raw_spin_unlock_irq(&sem->lock);
 224                timeout = schedule_timeout(timeout);
 225                raw_spin_lock_irq(&sem->lock);
 226                if (waiter.up)
 227                        return 0;
 228        }
 229
 230 timed_out:
 231        list_del(&waiter.list);
 232        return -ETIME;
 233
 234 interrupted:
 235        list_del(&waiter.list);
 236        return -EINTR;
 237}
 238
 239static noinline void __sched __down(struct semaphore *sem)
 240{
 241        __down_common(sem, TASK_UNINTERRUPTIBLE, MAX_SCHEDULE_TIMEOUT);
 242}
 243
 244static noinline int __sched __down_interruptible(struct semaphore *sem)
 245{
 246        return __down_common(sem, TASK_INTERRUPTIBLE, MAX_SCHEDULE_TIMEOUT);
 247}
 248
 249static noinline int __sched __down_killable(struct semaphore *sem)
 250{
 251        return __down_common(sem, TASK_KILLABLE, MAX_SCHEDULE_TIMEOUT);
 252}
 253
 254static noinline int __sched __down_timeout(struct semaphore *sem, long timeout)
 255{
 256        return __down_common(sem, TASK_UNINTERRUPTIBLE, timeout);
 257}
 258
 259static noinline void __sched __up(struct semaphore *sem)
 260{
 261        struct semaphore_waiter *waiter = list_first_entry(&sem->wait_list,
 262                                                struct semaphore_waiter, list);
 263        list_del(&waiter->list);
 264        waiter->up = true;
 265        wake_up_process(waiter->task);
 266}
 267