mirror of
https://github.com/armbian/linux-cix.git
synced 2026-01-06 12:30:45 -08:00
sched: Change task_struct::state
Change the type and name of task_struct::state. Drop the volatile and shrink it to an 'unsigned int'. Rename it in order to find all uses such that we can use READ_ONCE/WRITE_ONCE as appropriate. Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org> Reviewed-by: Daniel Bristot de Oliveira <bristot@redhat.com> Acked-by: Will Deacon <will@kernel.org> Acked-by: Daniel Thompson <daniel.thompson@linaro.org> Link: https://lore.kernel.org/r/20210611082838.550736351@infradead.org
This commit is contained in:
@@ -1788,7 +1788,7 @@ format_mca_init_stack(void *mca_data, unsigned long offset,
|
||||
ti->task = p;
|
||||
ti->cpu = cpu;
|
||||
p->stack = ti;
|
||||
p->state = TASK_UNINTERRUPTIBLE;
|
||||
p->__state = TASK_UNINTERRUPTIBLE;
|
||||
cpumask_set_cpu(cpu, &p->cpus_mask);
|
||||
INIT_LIST_HEAD(&p->tasks);
|
||||
p->parent = p->real_parent = p->group_leader = p;
|
||||
|
||||
@@ -641,11 +641,11 @@ ptrace_attach_sync_user_rbs (struct task_struct *child)
|
||||
read_lock(&tasklist_lock);
|
||||
if (child->sighand) {
|
||||
spin_lock_irq(&child->sighand->siglock);
|
||||
if (child->state == TASK_STOPPED &&
|
||||
if (READ_ONCE(child->__state) == TASK_STOPPED &&
|
||||
!test_and_set_tsk_thread_flag(child, TIF_RESTORE_RSE)) {
|
||||
set_notify_resume(child);
|
||||
|
||||
child->state = TASK_TRACED;
|
||||
WRITE_ONCE(child->__state, TASK_TRACED);
|
||||
stopped = 1;
|
||||
}
|
||||
spin_unlock_irq(&child->sighand->siglock);
|
||||
@@ -665,9 +665,9 @@ ptrace_attach_sync_user_rbs (struct task_struct *child)
|
||||
read_lock(&tasklist_lock);
|
||||
if (child->sighand) {
|
||||
spin_lock_irq(&child->sighand->siglock);
|
||||
if (child->state == TASK_TRACED &&
|
||||
if (READ_ONCE(child->__state) == TASK_TRACED &&
|
||||
(child->signal->flags & SIGNAL_STOP_STOPPED)) {
|
||||
child->state = TASK_STOPPED;
|
||||
WRITE_ONCE(child->__state, TASK_STOPPED);
|
||||
}
|
||||
spin_unlock_irq(&child->sighand->siglock);
|
||||
}
|
||||
|
||||
@@ -3162,6 +3162,7 @@ memzcan(void)
|
||||
|
||||
static void show_task(struct task_struct *tsk)
|
||||
{
|
||||
unsigned int p_state = READ_ONCE(tsk->__state);
|
||||
char state;
|
||||
|
||||
/*
|
||||
@@ -3169,14 +3170,14 @@ static void show_task(struct task_struct *tsk)
|
||||
* appropriate for calling from xmon. This could be moved
|
||||
* to a common, generic, routine used by both.
|
||||
*/
|
||||
state = (tsk->state == 0) ? 'R' :
|
||||
(tsk->state < 0) ? 'U' :
|
||||
(tsk->state & TASK_UNINTERRUPTIBLE) ? 'D' :
|
||||
(tsk->state & TASK_STOPPED) ? 'T' :
|
||||
(tsk->state & TASK_TRACED) ? 'C' :
|
||||
state = (p_state == 0) ? 'R' :
|
||||
(p_state < 0) ? 'U' :
|
||||
(p_state & TASK_UNINTERRUPTIBLE) ? 'D' :
|
||||
(p_state & TASK_STOPPED) ? 'T' :
|
||||
(p_state & TASK_TRACED) ? 'C' :
|
||||
(tsk->exit_state & EXIT_ZOMBIE) ? 'Z' :
|
||||
(tsk->exit_state & EXIT_DEAD) ? 'E' :
|
||||
(tsk->state & TASK_INTERRUPTIBLE) ? 'S' : '?';
|
||||
(p_state & TASK_INTERRUPTIBLE) ? 'S' : '?';
|
||||
|
||||
printf("%16px %16lx %16px %6d %6d %c %2d %s\n", tsk,
|
||||
tsk->thread.ksp, tsk->thread.regs,
|
||||
|
||||
@@ -3886,7 +3886,7 @@ static bool blk_mq_poll_hybrid(struct request_queue *q,
|
||||
int blk_poll(struct request_queue *q, blk_qc_t cookie, bool spin)
|
||||
{
|
||||
struct blk_mq_hw_ctx *hctx;
|
||||
long state;
|
||||
unsigned int state;
|
||||
|
||||
if (!blk_qc_t_valid(cookie) ||
|
||||
!test_bit(QUEUE_FLAG_POLL, &q->queue_flags))
|
||||
|
||||
@@ -2328,7 +2328,7 @@ static bool md_in_flight_bios(struct mapped_device *md)
|
||||
return sum != 0;
|
||||
}
|
||||
|
||||
static int dm_wait_for_bios_completion(struct mapped_device *md, long task_state)
|
||||
static int dm_wait_for_bios_completion(struct mapped_device *md, unsigned int task_state)
|
||||
{
|
||||
int r = 0;
|
||||
DEFINE_WAIT(wait);
|
||||
@@ -2351,7 +2351,7 @@ static int dm_wait_for_bios_completion(struct mapped_device *md, long task_state
|
||||
return r;
|
||||
}
|
||||
|
||||
static int dm_wait_for_completion(struct mapped_device *md, long task_state)
|
||||
static int dm_wait_for_completion(struct mapped_device *md, unsigned int task_state)
|
||||
{
|
||||
int r = 0;
|
||||
|
||||
@@ -2478,7 +2478,7 @@ static void unlock_fs(struct mapped_device *md)
|
||||
* are being added to md->deferred list.
|
||||
*/
|
||||
static int __dm_suspend(struct mapped_device *md, struct dm_table *map,
|
||||
unsigned suspend_flags, long task_state,
|
||||
unsigned suspend_flags, unsigned int task_state,
|
||||
int dmf_suspended_flag)
|
||||
{
|
||||
bool do_lockfs = suspend_flags & DM_SUSPEND_LOCKFS_FLAG;
|
||||
|
||||
@@ -1537,7 +1537,8 @@ static int fill_psinfo(struct elf_prpsinfo *psinfo, struct task_struct *p,
|
||||
{
|
||||
const struct cred *cred;
|
||||
unsigned int i, len;
|
||||
|
||||
unsigned int state;
|
||||
|
||||
/* first copy the parameters from user space */
|
||||
memset(psinfo, 0, sizeof(struct elf_prpsinfo));
|
||||
|
||||
@@ -1559,7 +1560,8 @@ static int fill_psinfo(struct elf_prpsinfo *psinfo, struct task_struct *p,
|
||||
psinfo->pr_pgrp = task_pgrp_vnr(p);
|
||||
psinfo->pr_sid = task_session_vnr(p);
|
||||
|
||||
i = p->state ? ffz(~p->state) + 1 : 0;
|
||||
state = READ_ONCE(p->__state);
|
||||
i = state ? ffz(~state) + 1 : 0;
|
||||
psinfo->pr_state = i;
|
||||
psinfo->pr_sname = (i > 5) ? '.' : "RSDTZW"[i];
|
||||
psinfo->pr_zomb = psinfo->pr_sname == 'Z';
|
||||
@@ -1571,7 +1573,7 @@ static int fill_psinfo(struct elf_prpsinfo *psinfo, struct task_struct *p,
|
||||
SET_GID(psinfo->pr_gid, from_kgid_munged(cred->user_ns, cred->gid));
|
||||
rcu_read_unlock();
|
||||
strncpy(psinfo->pr_fname, p->comm, sizeof(psinfo->pr_fname));
|
||||
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
||||
@@ -1331,6 +1331,7 @@ static int fill_psinfo(struct elf_prpsinfo *psinfo, struct task_struct *p,
|
||||
{
|
||||
const struct cred *cred;
|
||||
unsigned int i, len;
|
||||
unsigned int state;
|
||||
|
||||
/* first copy the parameters from user space */
|
||||
memset(psinfo, 0, sizeof(struct elf_prpsinfo));
|
||||
@@ -1353,7 +1354,8 @@ static int fill_psinfo(struct elf_prpsinfo *psinfo, struct task_struct *p,
|
||||
psinfo->pr_pgrp = task_pgrp_vnr(p);
|
||||
psinfo->pr_sid = task_session_vnr(p);
|
||||
|
||||
i = p->state ? ffz(~p->state) + 1 : 0;
|
||||
state = READ_ONCE(p->__state);
|
||||
i = state ? ffz(~state) + 1 : 0;
|
||||
psinfo->pr_state = i;
|
||||
psinfo->pr_sname = (i > 5) ? '.' : "RSDTZW"[i];
|
||||
psinfo->pr_zomb = psinfo->pr_sname == 'Z';
|
||||
|
||||
@@ -337,7 +337,7 @@ out:
|
||||
return ret;
|
||||
}
|
||||
|
||||
static inline long userfaultfd_get_blocking_state(unsigned int flags)
|
||||
static inline unsigned int userfaultfd_get_blocking_state(unsigned int flags)
|
||||
{
|
||||
if (flags & FAULT_FLAG_INTERRUPTIBLE)
|
||||
return TASK_INTERRUPTIBLE;
|
||||
@@ -370,7 +370,7 @@ vm_fault_t handle_userfault(struct vm_fault *vmf, unsigned long reason)
|
||||
struct userfaultfd_wait_queue uwq;
|
||||
vm_fault_t ret = VM_FAULT_SIGBUS;
|
||||
bool must_wait;
|
||||
long blocking_state;
|
||||
unsigned int blocking_state;
|
||||
|
||||
/*
|
||||
* We don't do userfault handling for the final child pid update.
|
||||
|
||||
@@ -113,13 +113,13 @@ struct task_group;
|
||||
__TASK_TRACED | EXIT_DEAD | EXIT_ZOMBIE | \
|
||||
TASK_PARKED)
|
||||
|
||||
#define task_is_running(task) (READ_ONCE((task)->state) == TASK_RUNNING)
|
||||
#define task_is_running(task) (READ_ONCE((task)->__state) == TASK_RUNNING)
|
||||
|
||||
#define task_is_traced(task) ((task->state & __TASK_TRACED) != 0)
|
||||
#define task_is_traced(task) ((READ_ONCE(task->__state) & __TASK_TRACED) != 0)
|
||||
|
||||
#define task_is_stopped(task) ((task->state & __TASK_STOPPED) != 0)
|
||||
#define task_is_stopped(task) ((READ_ONCE(task->__state) & __TASK_STOPPED) != 0)
|
||||
|
||||
#define task_is_stopped_or_traced(task) ((task->state & (__TASK_STOPPED | __TASK_TRACED)) != 0)
|
||||
#define task_is_stopped_or_traced(task) ((READ_ONCE(task->__state) & (__TASK_STOPPED | __TASK_TRACED)) != 0)
|
||||
|
||||
#ifdef CONFIG_DEBUG_ATOMIC_SLEEP
|
||||
|
||||
@@ -134,14 +134,14 @@ struct task_group;
|
||||
do { \
|
||||
WARN_ON_ONCE(is_special_task_state(state_value));\
|
||||
current->task_state_change = _THIS_IP_; \
|
||||
current->state = (state_value); \
|
||||
WRITE_ONCE(current->__state, (state_value)); \
|
||||
} while (0)
|
||||
|
||||
#define set_current_state(state_value) \
|
||||
do { \
|
||||
WARN_ON_ONCE(is_special_task_state(state_value));\
|
||||
current->task_state_change = _THIS_IP_; \
|
||||
smp_store_mb(current->state, (state_value)); \
|
||||
smp_store_mb(current->__state, (state_value)); \
|
||||
} while (0)
|
||||
|
||||
#define set_special_state(state_value) \
|
||||
@@ -150,7 +150,7 @@ struct task_group;
|
||||
WARN_ON_ONCE(!is_special_task_state(state_value)); \
|
||||
raw_spin_lock_irqsave(¤t->pi_lock, flags); \
|
||||
current->task_state_change = _THIS_IP_; \
|
||||
current->state = (state_value); \
|
||||
WRITE_ONCE(current->__state, (state_value)); \
|
||||
raw_spin_unlock_irqrestore(¤t->pi_lock, flags); \
|
||||
} while (0)
|
||||
#else
|
||||
@@ -192,10 +192,10 @@ struct task_group;
|
||||
* Also see the comments of try_to_wake_up().
|
||||
*/
|
||||
#define __set_current_state(state_value) \
|
||||
current->state = (state_value)
|
||||
WRITE_ONCE(current->__state, (state_value))
|
||||
|
||||
#define set_current_state(state_value) \
|
||||
smp_store_mb(current->state, (state_value))
|
||||
smp_store_mb(current->__state, (state_value))
|
||||
|
||||
/*
|
||||
* set_special_state() should be used for those states when the blocking task
|
||||
@@ -207,13 +207,13 @@ struct task_group;
|
||||
do { \
|
||||
unsigned long flags; /* may shadow */ \
|
||||
raw_spin_lock_irqsave(¤t->pi_lock, flags); \
|
||||
current->state = (state_value); \
|
||||
WRITE_ONCE(current->__state, (state_value)); \
|
||||
raw_spin_unlock_irqrestore(¤t->pi_lock, flags); \
|
||||
} while (0)
|
||||
|
||||
#endif
|
||||
|
||||
#define get_current_state() READ_ONCE(current->state)
|
||||
#define get_current_state() READ_ONCE(current->__state)
|
||||
|
||||
/* Task command name length: */
|
||||
#define TASK_COMM_LEN 16
|
||||
@@ -666,8 +666,7 @@ struct task_struct {
|
||||
*/
|
||||
struct thread_info thread_info;
|
||||
#endif
|
||||
/* -1 unrunnable, 0 runnable, >0 stopped: */
|
||||
volatile long state;
|
||||
unsigned int __state;
|
||||
|
||||
/*
|
||||
* This begins the randomizable portion of task_struct. Only
|
||||
@@ -1532,7 +1531,7 @@ static inline pid_t task_pgrp_nr(struct task_struct *tsk)
|
||||
|
||||
static inline unsigned int task_state_index(struct task_struct *tsk)
|
||||
{
|
||||
unsigned int tsk_state = READ_ONCE(tsk->state);
|
||||
unsigned int tsk_state = READ_ONCE(tsk->__state);
|
||||
unsigned int state = (tsk_state | tsk->exit_state) & TASK_REPORT;
|
||||
|
||||
BUILD_BUG_ON_NOT_POWER_OF_2(TASK_REPORT_MAX);
|
||||
@@ -1840,10 +1839,10 @@ static __always_inline void scheduler_ipi(void)
|
||||
*/
|
||||
preempt_fold_need_resched();
|
||||
}
|
||||
extern unsigned long wait_task_inactive(struct task_struct *, long match_state);
|
||||
extern unsigned long wait_task_inactive(struct task_struct *, unsigned int match_state);
|
||||
#else
|
||||
static inline void scheduler_ipi(void) { }
|
||||
static inline unsigned long wait_task_inactive(struct task_struct *p, long match_state)
|
||||
static inline unsigned long wait_task_inactive(struct task_struct *p, unsigned int match_state)
|
||||
{
|
||||
return 1;
|
||||
}
|
||||
|
||||
@@ -14,7 +14,7 @@ extern void dump_cpu_task(int cpu);
|
||||
/*
|
||||
* Only dump TASK_* tasks. (0 for all tasks)
|
||||
*/
|
||||
extern void show_state_filter(unsigned long state_filter);
|
||||
extern void show_state_filter(unsigned int state_filter);
|
||||
|
||||
static inline void show_state(void)
|
||||
{
|
||||
|
||||
@@ -382,7 +382,7 @@ static inline int fatal_signal_pending(struct task_struct *p)
|
||||
return task_sigpending(p) && __fatal_signal_pending(p);
|
||||
}
|
||||
|
||||
static inline int signal_pending_state(long state, struct task_struct *p)
|
||||
static inline int signal_pending_state(unsigned int state, struct task_struct *p)
|
||||
{
|
||||
if (!(state & (TASK_INTERRUPTIBLE | TASK_WAKEKILL)))
|
||||
return 0;
|
||||
|
||||
@@ -71,7 +71,7 @@ struct task_struct init_task
|
||||
.thread_info = INIT_THREAD_INFO(init_task),
|
||||
.stack_refcount = REFCOUNT_INIT(1),
|
||||
#endif
|
||||
.state = 0,
|
||||
.__state = 0,
|
||||
.stack = init_stack,
|
||||
.usage = REFCOUNT_INIT(2),
|
||||
.flags = PF_KTHREAD,
|
||||
|
||||
@@ -713,7 +713,7 @@ int cgroupstats_build(struct cgroupstats *stats, struct dentry *dentry)
|
||||
|
||||
css_task_iter_start(&cgrp->self, 0, &it);
|
||||
while ((tsk = css_task_iter_next(&it))) {
|
||||
switch (tsk->state) {
|
||||
switch (READ_ONCE(tsk->__state)) {
|
||||
case TASK_RUNNING:
|
||||
stats->nr_running++;
|
||||
break;
|
||||
|
||||
@@ -609,23 +609,25 @@ unsigned long kdb_task_state_string(const char *s)
|
||||
*/
|
||||
char kdb_task_state_char (const struct task_struct *p)
|
||||
{
|
||||
int cpu;
|
||||
char state;
|
||||
unsigned int p_state;
|
||||
unsigned long tmp;
|
||||
char state;
|
||||
int cpu;
|
||||
|
||||
if (!p ||
|
||||
copy_from_kernel_nofault(&tmp, (char *)p, sizeof(unsigned long)))
|
||||
return 'E';
|
||||
|
||||
cpu = kdb_process_cpu(p);
|
||||
state = (p->state == 0) ? 'R' :
|
||||
(p->state < 0) ? 'U' :
|
||||
(p->state & TASK_UNINTERRUPTIBLE) ? 'D' :
|
||||
(p->state & TASK_STOPPED) ? 'T' :
|
||||
(p->state & TASK_TRACED) ? 'C' :
|
||||
p_state = READ_ONCE(p->__state);
|
||||
state = (p_state == 0) ? 'R' :
|
||||
(p_state < 0) ? 'U' :
|
||||
(p_state & TASK_UNINTERRUPTIBLE) ? 'D' :
|
||||
(p_state & TASK_STOPPED) ? 'T' :
|
||||
(p_state & TASK_TRACED) ? 'C' :
|
||||
(p->exit_state & EXIT_ZOMBIE) ? 'Z' :
|
||||
(p->exit_state & EXIT_DEAD) ? 'E' :
|
||||
(p->state & TASK_INTERRUPTIBLE) ? 'S' : '?';
|
||||
(p_state & TASK_INTERRUPTIBLE) ? 'S' : '?';
|
||||
if (is_idle_task(p)) {
|
||||
/* Idle task. Is it really idle, apart from the kdb
|
||||
* interrupt? */
|
||||
|
||||
@@ -425,7 +425,7 @@ static int memcg_charge_kernel_stack(struct task_struct *tsk)
|
||||
|
||||
static void release_task_stack(struct task_struct *tsk)
|
||||
{
|
||||
if (WARN_ON(tsk->state != TASK_DEAD))
|
||||
if (WARN_ON(READ_ONCE(tsk->__state) != TASK_DEAD))
|
||||
return; /* Better to leak the stack than to free prematurely */
|
||||
|
||||
account_kernel_stack(tsk, -1);
|
||||
@@ -2392,7 +2392,7 @@ bad_fork_cleanup_count:
|
||||
atomic_dec(&p->cred->user->processes);
|
||||
exit_creds(p);
|
||||
bad_fork_free:
|
||||
p->state = TASK_DEAD;
|
||||
WRITE_ONCE(p->__state, TASK_DEAD);
|
||||
put_task_stack(p);
|
||||
delayed_free_task(p);
|
||||
fork_out:
|
||||
|
||||
@@ -196,7 +196,7 @@ static void check_hung_uninterruptible_tasks(unsigned long timeout)
|
||||
last_break = jiffies;
|
||||
}
|
||||
/* use "==" to skip the TASK_KILLABLE tasks waiting on NFS */
|
||||
if (t->state == TASK_UNINTERRUPTIBLE)
|
||||
if (READ_ONCE(t->__state) == TASK_UNINTERRUPTIBLE)
|
||||
check_hung_task(t, timeout);
|
||||
}
|
||||
unlock:
|
||||
|
||||
@@ -457,7 +457,7 @@ struct task_struct *kthread_create_on_node(int (*threadfn)(void *data),
|
||||
}
|
||||
EXPORT_SYMBOL(kthread_create_on_node);
|
||||
|
||||
static void __kthread_bind_mask(struct task_struct *p, const struct cpumask *mask, long state)
|
||||
static void __kthread_bind_mask(struct task_struct *p, const struct cpumask *mask, unsigned int state)
|
||||
{
|
||||
unsigned long flags;
|
||||
|
||||
@@ -473,7 +473,7 @@ static void __kthread_bind_mask(struct task_struct *p, const struct cpumask *mas
|
||||
raw_spin_unlock_irqrestore(&p->pi_lock, flags);
|
||||
}
|
||||
|
||||
static void __kthread_bind(struct task_struct *p, unsigned int cpu, long state)
|
||||
static void __kthread_bind(struct task_struct *p, unsigned int cpu, unsigned int state)
|
||||
{
|
||||
__kthread_bind_mask(p, cpumask_of(cpu), state);
|
||||
}
|
||||
|
||||
@@ -923,7 +923,7 @@ __ww_mutex_add_waiter(struct mutex_waiter *waiter,
|
||||
* Lock a mutex (possibly interruptible), slowpath:
|
||||
*/
|
||||
static __always_inline int __sched
|
||||
__mutex_lock_common(struct mutex *lock, long state, unsigned int subclass,
|
||||
__mutex_lock_common(struct mutex *lock, unsigned int state, unsigned int subclass,
|
||||
struct lockdep_map *nest_lock, unsigned long ip,
|
||||
struct ww_acquire_ctx *ww_ctx, const bool use_ww_ctx)
|
||||
{
|
||||
@@ -1098,14 +1098,14 @@ err_early_kill:
|
||||
}
|
||||
|
||||
static int __sched
|
||||
__mutex_lock(struct mutex *lock, long state, unsigned int subclass,
|
||||
__mutex_lock(struct mutex *lock, unsigned int state, unsigned int subclass,
|
||||
struct lockdep_map *nest_lock, unsigned long ip)
|
||||
{
|
||||
return __mutex_lock_common(lock, state, subclass, nest_lock, ip, NULL, false);
|
||||
}
|
||||
|
||||
static int __sched
|
||||
__ww_mutex_lock(struct mutex *lock, long state, unsigned int subclass,
|
||||
__ww_mutex_lock(struct mutex *lock, unsigned int state, unsigned int subclass,
|
||||
struct lockdep_map *nest_lock, unsigned long ip,
|
||||
struct ww_acquire_ctx *ww_ctx)
|
||||
{
|
||||
|
||||
@@ -1135,7 +1135,7 @@ void __sched rt_mutex_init_waiter(struct rt_mutex_waiter *waiter)
|
||||
*
|
||||
* Must be called with lock->wait_lock held and interrupts disabled
|
||||
*/
|
||||
static int __sched __rt_mutex_slowlock(struct rt_mutex *lock, int state,
|
||||
static int __sched __rt_mutex_slowlock(struct rt_mutex *lock, unsigned int state,
|
||||
struct hrtimer_sleeper *timeout,
|
||||
struct rt_mutex_waiter *waiter)
|
||||
{
|
||||
@@ -1190,7 +1190,7 @@ static void __sched rt_mutex_handle_deadlock(int res, int detect_deadlock,
|
||||
/*
|
||||
* Slow path lock function:
|
||||
*/
|
||||
static int __sched rt_mutex_slowlock(struct rt_mutex *lock, int state,
|
||||
static int __sched rt_mutex_slowlock(struct rt_mutex *lock, unsigned int state,
|
||||
struct hrtimer_sleeper *timeout,
|
||||
enum rtmutex_chainwalk chwalk)
|
||||
{
|
||||
|
||||
@@ -889,7 +889,7 @@ rwsem_spin_on_owner(struct rw_semaphore *sem)
|
||||
* Wait for the read lock to be granted
|
||||
*/
|
||||
static struct rw_semaphore __sched *
|
||||
rwsem_down_read_slowpath(struct rw_semaphore *sem, long count, int state)
|
||||
rwsem_down_read_slowpath(struct rw_semaphore *sem, long count, unsigned int state)
|
||||
{
|
||||
long adjustment = -RWSEM_READER_BIAS;
|
||||
long rcnt = (count >> RWSEM_READER_SHIFT);
|
||||
|
||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user