You've already forked linux-apfs
mirror of
https://github.com/linux-apfs/linux-apfs.git
synced 2026-05-01 15:00:59 -07:00
[PATCH] sched: cleanup, remove task_t, convert to struct task_struct
cleanup: remove task_t and convert all the uses to struct task_struct. I introduced it for the scheduler anno and it was a mistake. Conversion was mostly scripted, the result was reviewed and all secondary whitespace and style impact (if any) was fixed up by hand. Signed-off-by: Ingo Molnar <mingo@elte.hu> Signed-off-by: Andrew Morton <akpm@osdl.org> Signed-off-by: Linus Torvalds <torvalds@osdl.org>
This commit is contained in:
committed by
Linus Torvalds
parent
48f24c4da1
commit
36c8b58689
@@ -474,7 +474,7 @@ out:
|
||||
*/
|
||||
|
||||
unsigned long
|
||||
thread_saved_pc(task_t *t)
|
||||
thread_saved_pc(struct task_struct *t)
|
||||
{
|
||||
unsigned long base = (unsigned long)task_stack_page(t);
|
||||
unsigned long fp, sp = task_thread_info(t)->pcb.ksp;
|
||||
|
||||
@@ -678,7 +678,7 @@ copy_reg(const u64 *fr, u64 fnat, u64 *tr, u64 *tnat)
|
||||
*/
|
||||
|
||||
static void
|
||||
ia64_mca_modify_comm(const task_t *previous_current)
|
||||
ia64_mca_modify_comm(const struct task_struct *previous_current)
|
||||
{
|
||||
char *p, comm[sizeof(current->comm)];
|
||||
if (previous_current->pid)
|
||||
@@ -709,7 +709,7 @@ ia64_mca_modify_comm(const task_t *previous_current)
|
||||
* that we can do backtrace on the MCA/INIT handler code itself.
|
||||
*/
|
||||
|
||||
static task_t *
|
||||
static struct task_struct *
|
||||
ia64_mca_modify_original_stack(struct pt_regs *regs,
|
||||
const struct switch_stack *sw,
|
||||
struct ia64_sal_os_state *sos,
|
||||
@@ -719,7 +719,7 @@ ia64_mca_modify_original_stack(struct pt_regs *regs,
|
||||
ia64_va va;
|
||||
extern char ia64_leave_kernel[]; /* Need asm address, not function descriptor */
|
||||
const pal_min_state_area_t *ms = sos->pal_min_state;
|
||||
task_t *previous_current;
|
||||
struct task_struct *previous_current;
|
||||
struct pt_regs *old_regs;
|
||||
struct switch_stack *old_sw;
|
||||
unsigned size = sizeof(struct pt_regs) +
|
||||
@@ -1023,7 +1023,7 @@ ia64_mca_handler(struct pt_regs *regs, struct switch_stack *sw,
|
||||
pal_processor_state_info_t *psp = (pal_processor_state_info_t *)
|
||||
&sos->proc_state_param;
|
||||
int recover, cpu = smp_processor_id();
|
||||
task_t *previous_current;
|
||||
struct task_struct *previous_current;
|
||||
struct ia64_mca_notify_die nd =
|
||||
{ .sos = sos, .monarch_cpu = &monarch_cpu };
|
||||
|
||||
@@ -1352,7 +1352,7 @@ ia64_init_handler(struct pt_regs *regs, struct switch_stack *sw,
|
||||
{
|
||||
static atomic_t slaves;
|
||||
static atomic_t monarchs;
|
||||
task_t *previous_current;
|
||||
struct task_struct *previous_current;
|
||||
int cpu = smp_processor_id();
|
||||
struct ia64_mca_notify_die nd =
|
||||
{ .sos = sos, .monarch_cpu = &monarch_cpu };
|
||||
|
||||
@@ -124,7 +124,7 @@ extern void __devinit calibrate_delay (void);
|
||||
extern void start_ap (void);
|
||||
extern unsigned long ia64_iobase;
|
||||
|
||||
task_t *task_for_booting_cpu;
|
||||
struct task_struct *task_for_booting_cpu;
|
||||
|
||||
/*
|
||||
* State for each CPU
|
||||
|
||||
@@ -65,7 +65,7 @@ need_resched:
|
||||
#endif
|
||||
|
||||
FEXPORT(ret_from_fork)
|
||||
jal schedule_tail # a0 = task_t *prev
|
||||
jal schedule_tail # a0 = struct task_struct *prev
|
||||
|
||||
FEXPORT(syscall_exit)
|
||||
local_irq_disable # make sure need_resched and
|
||||
|
||||
@@ -47,7 +47,7 @@ unsigned long mt_fpemul_threshold = 0;
|
||||
* used in sys_sched_set/getaffinity() in kernel/sched.c, so
|
||||
* cloned here.
|
||||
*/
|
||||
static inline task_t *find_process_by_pid(pid_t pid)
|
||||
static inline struct task_struct *find_process_by_pid(pid_t pid)
|
||||
{
|
||||
return pid ? find_task_by_pid(pid) : current;
|
||||
}
|
||||
@@ -62,7 +62,7 @@ asmlinkage long mipsmt_sys_sched_setaffinity(pid_t pid, unsigned int len,
|
||||
cpumask_t new_mask;
|
||||
cpumask_t effective_mask;
|
||||
int retval;
|
||||
task_t *p;
|
||||
struct task_struct *p;
|
||||
|
||||
if (len < sizeof(new_mask))
|
||||
return -EINVAL;
|
||||
@@ -127,7 +127,7 @@ asmlinkage long mipsmt_sys_sched_getaffinity(pid_t pid, unsigned int len,
|
||||
unsigned int real_len;
|
||||
cpumask_t mask;
|
||||
int retval;
|
||||
task_t *p;
|
||||
struct task_struct *p;
|
||||
|
||||
real_len = sizeof(mask);
|
||||
if (len < real_len)
|
||||
|
||||
@@ -119,7 +119,7 @@ void suspend_new_thread(int fd)
|
||||
panic("read failed in suspend_new_thread, err = %d", -err);
|
||||
}
|
||||
|
||||
void schedule_tail(task_t *prev);
|
||||
void schedule_tail(struct task_struct *prev);
|
||||
|
||||
static void new_thread_handler(int sig)
|
||||
{
|
||||
|
||||
@@ -2336,7 +2336,7 @@ static int fionbio(struct file *file, int __user *p)
|
||||
|
||||
static int tiocsctty(struct tty_struct *tty, int arg)
|
||||
{
|
||||
task_t *p;
|
||||
struct task_struct *p;
|
||||
|
||||
if (current->signal->leader &&
|
||||
(current->signal->session == tty->session))
|
||||
|
||||
+2
-2
@@ -120,7 +120,7 @@ struct epoll_filefd {
|
||||
*/
|
||||
struct wake_task_node {
|
||||
struct list_head llink;
|
||||
task_t *task;
|
||||
struct task_struct *task;
|
||||
wait_queue_head_t *wq;
|
||||
};
|
||||
|
||||
@@ -413,7 +413,7 @@ static void ep_poll_safewake(struct poll_safewake *psw, wait_queue_head_t *wq)
|
||||
{
|
||||
int wake_nests = 0;
|
||||
unsigned long flags;
|
||||
task_t *this_task = current;
|
||||
struct task_struct *this_task = current;
|
||||
struct list_head *lsthead = &psw->wake_task_list, *lnk;
|
||||
struct wake_task_node *tncur;
|
||||
struct wake_task_node tnode;
|
||||
|
||||
@@ -68,7 +68,7 @@ struct thread_info {
|
||||
#define end_of_stack(p) (unsigned long *)((void *)(p) + IA64_RBS_OFFSET)
|
||||
|
||||
#define __HAVE_ARCH_TASK_STRUCT_ALLOCATOR
|
||||
#define alloc_task_struct() ((task_t *)__get_free_pages(GFP_KERNEL | __GFP_COMP, KERNEL_STACK_SIZE_ORDER))
|
||||
#define alloc_task_struct() ((struct task_struct *)__get_free_pages(GFP_KERNEL | __GFP_COMP, KERNEL_STACK_SIZE_ORDER))
|
||||
#define free_task_struct(tsk) free_pages((unsigned long) (tsk), KERNEL_STACK_SIZE_ORDER)
|
||||
|
||||
#endif /* !__ASSEMBLY */
|
||||
|
||||
@@ -18,7 +18,7 @@
|
||||
* switch_to(prev, next) should switch from task `prev' to `next'
|
||||
* `prev' will never be the same as `next'.
|
||||
*
|
||||
* `next' and `prev' should be task_t, but it isn't always defined
|
||||
* `next' and `prev' should be struct task_struct, but it isn't always defined
|
||||
*/
|
||||
|
||||
#define switch_to(prev, next, last) do { \
|
||||
|
||||
@@ -12,7 +12,7 @@
|
||||
*/
|
||||
|
||||
#define switch_to(prev, next, last) do { \
|
||||
task_t *__last; \
|
||||
struct task_struct *__last; \
|
||||
register unsigned long *__ts1 __asm__ ("r1") = &prev->thread.sp; \
|
||||
register unsigned long *__ts2 __asm__ ("r2") = &prev->thread.pc; \
|
||||
register unsigned long *__ts4 __asm__ ("r4") = (unsigned long *)prev; \
|
||||
|
||||
+28
-27
@@ -184,11 +184,11 @@ extern unsigned long weighted_cpuload(const int cpu);
|
||||
extern rwlock_t tasklist_lock;
|
||||
extern spinlock_t mmlist_lock;
|
||||
|
||||
typedef struct task_struct task_t;
|
||||
struct task_struct;
|
||||
|
||||
extern void sched_init(void);
|
||||
extern void sched_init_smp(void);
|
||||
extern void init_idle(task_t *idle, int cpu);
|
||||
extern void init_idle(struct task_struct *idle, int cpu);
|
||||
|
||||
extern cpumask_t nohz_cpu_mask;
|
||||
|
||||
@@ -383,7 +383,7 @@ struct signal_struct {
|
||||
wait_queue_head_t wait_chldexit; /* for wait4() */
|
||||
|
||||
/* current thread group signal load-balancing target: */
|
||||
task_t *curr_target;
|
||||
struct task_struct *curr_target;
|
||||
|
||||
/* shared signal handling: */
|
||||
struct sigpending shared_pending;
|
||||
@@ -699,7 +699,7 @@ extern int groups_search(struct group_info *group_info, gid_t grp);
|
||||
((gi)->blocks[(i)/NGROUPS_PER_BLOCK][(i)%NGROUPS_PER_BLOCK])
|
||||
|
||||
#ifdef ARCH_HAS_PREFETCH_SWITCH_STACK
|
||||
extern void prefetch_stack(struct task_struct*);
|
||||
extern void prefetch_stack(struct task_struct *t);
|
||||
#else
|
||||
static inline void prefetch_stack(struct task_struct *t) { }
|
||||
#endif
|
||||
@@ -1031,9 +1031,9 @@ static inline void put_task_struct(struct task_struct *t)
|
||||
#define used_math() tsk_used_math(current)
|
||||
|
||||
#ifdef CONFIG_SMP
|
||||
extern int set_cpus_allowed(task_t *p, cpumask_t new_mask);
|
||||
extern int set_cpus_allowed(struct task_struct *p, cpumask_t new_mask);
|
||||
#else
|
||||
static inline int set_cpus_allowed(task_t *p, cpumask_t new_mask)
|
||||
static inline int set_cpus_allowed(struct task_struct *p, cpumask_t new_mask)
|
||||
{
|
||||
if (!cpu_isset(0, new_mask))
|
||||
return -EINVAL;
|
||||
@@ -1042,7 +1042,8 @@ static inline int set_cpus_allowed(task_t *p, cpumask_t new_mask)
|
||||
#endif
|
||||
|
||||
extern unsigned long long sched_clock(void);
|
||||
extern unsigned long long current_sched_time(const task_t *current_task);
|
||||
extern unsigned long long
|
||||
current_sched_time(const struct task_struct *current_task);
|
||||
|
||||
/* sched_exec is called by processes performing an exec */
|
||||
#ifdef CONFIG_SMP
|
||||
@@ -1060,27 +1061,27 @@ static inline void idle_task_exit(void) {}
|
||||
extern void sched_idle_next(void);
|
||||
|
||||
#ifdef CONFIG_RT_MUTEXES
|
||||
extern int rt_mutex_getprio(task_t *p);
|
||||
extern void rt_mutex_setprio(task_t *p, int prio);
|
||||
extern void rt_mutex_adjust_pi(task_t *p);
|
||||
extern int rt_mutex_getprio(struct task_struct *p);
|
||||
extern void rt_mutex_setprio(struct task_struct *p, int prio);
|
||||
extern void rt_mutex_adjust_pi(struct task_struct *p);
|
||||
#else
|
||||
static inline int rt_mutex_getprio(task_t *p)
|
||||
static inline int rt_mutex_getprio(struct task_struct *p)
|
||||
{
|
||||
return p->normal_prio;
|
||||
}
|
||||
# define rt_mutex_adjust_pi(p) do { } while (0)
|
||||
#endif
|
||||
|
||||
extern void set_user_nice(task_t *p, long nice);
|
||||
extern int task_prio(const task_t *p);
|
||||
extern int task_nice(const task_t *p);
|
||||
extern int can_nice(const task_t *p, const int nice);
|
||||
extern int task_curr(const task_t *p);
|
||||
extern void set_user_nice(struct task_struct *p, long nice);
|
||||
extern int task_prio(const struct task_struct *p);
|
||||
extern int task_nice(const struct task_struct *p);
|
||||
extern int can_nice(const struct task_struct *p, const int nice);
|
||||
extern int task_curr(const struct task_struct *p);
|
||||
extern int idle_cpu(int cpu);
|
||||
extern int sched_setscheduler(struct task_struct *, int, struct sched_param *);
|
||||
extern task_t *idle_task(int cpu);
|
||||
extern task_t *curr_task(int cpu);
|
||||
extern void set_curr_task(int cpu, task_t *p);
|
||||
extern struct task_struct *idle_task(int cpu);
|
||||
extern struct task_struct *curr_task(int cpu);
|
||||
extern void set_curr_task(int cpu, struct task_struct *p);
|
||||
|
||||
void yield(void);
|
||||
|
||||
@@ -1137,8 +1138,8 @@ extern void FASTCALL(wake_up_new_task(struct task_struct * tsk,
|
||||
#else
|
||||
static inline void kick_process(struct task_struct *tsk) { }
|
||||
#endif
|
||||
extern void FASTCALL(sched_fork(task_t * p, int clone_flags));
|
||||
extern void FASTCALL(sched_exit(task_t * p));
|
||||
extern void FASTCALL(sched_fork(struct task_struct * p, int clone_flags));
|
||||
extern void FASTCALL(sched_exit(struct task_struct * p));
|
||||
|
||||
extern int in_group_p(gid_t);
|
||||
extern int in_egroup_p(gid_t);
|
||||
@@ -1243,17 +1244,17 @@ extern NORET_TYPE void do_group_exit(int);
|
||||
extern void daemonize(const char *, ...);
|
||||
extern int allow_signal(int);
|
||||
extern int disallow_signal(int);
|
||||
extern task_t *child_reaper;
|
||||
extern struct task_struct *child_reaper;
|
||||
|
||||
extern int do_execve(char *, char __user * __user *, char __user * __user *, struct pt_regs *);
|
||||
extern long do_fork(unsigned long, unsigned long, struct pt_regs *, unsigned long, int __user *, int __user *);
|
||||
task_t *fork_idle(int);
|
||||
struct task_struct *fork_idle(int);
|
||||
|
||||
extern void set_task_comm(struct task_struct *tsk, char *from);
|
||||
extern void get_task_comm(char *to, struct task_struct *tsk);
|
||||
|
||||
#ifdef CONFIG_SMP
|
||||
extern void wait_task_inactive(task_t * p);
|
||||
extern void wait_task_inactive(struct task_struct * p);
|
||||
#else
|
||||
#define wait_task_inactive(p) do { } while (0)
|
||||
#endif
|
||||
@@ -1279,13 +1280,13 @@ extern void wait_task_inactive(task_t * p);
|
||||
/* de_thread depends on thread_group_leader not being a pid based check */
|
||||
#define thread_group_leader(p) (p == p->group_leader)
|
||||
|
||||
static inline task_t *next_thread(const task_t *p)
|
||||
static inline struct task_struct *next_thread(const struct task_struct *p)
|
||||
{
|
||||
return list_entry(rcu_dereference(p->thread_group.next),
|
||||
task_t, thread_group);
|
||||
struct task_struct, thread_group);
|
||||
}
|
||||
|
||||
static inline int thread_group_empty(task_t *p)
|
||||
static inline int thread_group_empty(struct task_struct *p)
|
||||
{
|
||||
return list_empty(&p->thread_group);
|
||||
}
|
||||
|
||||
+4
-4
@@ -46,7 +46,7 @@ asmlinkage long sys_capget(cap_user_header_t header, cap_user_data_t dataptr)
|
||||
int ret = 0;
|
||||
pid_t pid;
|
||||
__u32 version;
|
||||
task_t *target;
|
||||
struct task_struct *target;
|
||||
struct __user_cap_data_struct data;
|
||||
|
||||
if (get_user(version, &header->version))
|
||||
@@ -96,7 +96,7 @@ static inline int cap_set_pg(int pgrp, kernel_cap_t *effective,
|
||||
kernel_cap_t *inheritable,
|
||||
kernel_cap_t *permitted)
|
||||
{
|
||||
task_t *g, *target;
|
||||
struct task_struct *g, *target;
|
||||
int ret = -EPERM;
|
||||
int found = 0;
|
||||
|
||||
@@ -128,7 +128,7 @@ static inline int cap_set_all(kernel_cap_t *effective,
|
||||
kernel_cap_t *inheritable,
|
||||
kernel_cap_t *permitted)
|
||||
{
|
||||
task_t *g, *target;
|
||||
struct task_struct *g, *target;
|
||||
int ret = -EPERM;
|
||||
int found = 0;
|
||||
|
||||
@@ -172,7 +172,7 @@ asmlinkage long sys_capset(cap_user_header_t header, const cap_user_data_t data)
|
||||
{
|
||||
kernel_cap_t inheritable, permitted, effective;
|
||||
__u32 version;
|
||||
task_t *target;
|
||||
struct task_struct *target;
|
||||
int ret;
|
||||
pid_t pid;
|
||||
|
||||
|
||||
+19
-16
@@ -134,8 +134,8 @@ static void delayed_put_task_struct(struct rcu_head *rhp)
|
||||
|
||||
void release_task(struct task_struct * p)
|
||||
{
|
||||
struct task_struct *leader;
|
||||
int zap_leader;
|
||||
task_t *leader;
|
||||
repeat:
|
||||
atomic_dec(&p->user->processes);
|
||||
write_lock_irq(&tasklist_lock);
|
||||
@@ -209,7 +209,7 @@ out:
|
||||
*
|
||||
* "I ask you, have you ever known what it is to be an orphan?"
|
||||
*/
|
||||
static int will_become_orphaned_pgrp(int pgrp, task_t *ignored_task)
|
||||
static int will_become_orphaned_pgrp(int pgrp, struct task_struct *ignored_task)
|
||||
{
|
||||
struct task_struct *p;
|
||||
int ret = 1;
|
||||
@@ -582,7 +582,8 @@ static void exit_mm(struct task_struct * tsk)
|
||||
mmput(mm);
|
||||
}
|
||||
|
||||
static inline void choose_new_parent(task_t *p, task_t *reaper)
|
||||
static inline void
|
||||
choose_new_parent(struct task_struct *p, struct task_struct *reaper)
|
||||
{
|
||||
/*
|
||||
* Make sure we're not reparenting to ourselves and that
|
||||
@@ -592,7 +593,8 @@ static inline void choose_new_parent(task_t *p, task_t *reaper)
|
||||
p->real_parent = reaper;
|
||||
}
|
||||
|
||||
static void reparent_thread(task_t *p, task_t *father, int traced)
|
||||
static void
|
||||
reparent_thread(struct task_struct *p, struct task_struct *father, int traced)
|
||||
{
|
||||
/* We don't want people slaying init. */
|
||||
if (p->exit_signal != -1)
|
||||
@@ -656,8 +658,8 @@ static void reparent_thread(task_t *p, task_t *father, int traced)
|
||||
* group, and if no such member exists, give it to
|
||||
* the global child reaper process (ie "init")
|
||||
*/
|
||||
static void forget_original_parent(struct task_struct * father,
|
||||
struct list_head *to_release)
|
||||
static void
|
||||
forget_original_parent(struct task_struct *father, struct list_head *to_release)
|
||||
{
|
||||
struct task_struct *p, *reaper = father;
|
||||
struct list_head *_p, *_n;
|
||||
@@ -680,7 +682,7 @@ static void forget_original_parent(struct task_struct * father,
|
||||
*/
|
||||
list_for_each_safe(_p, _n, &father->children) {
|
||||
int ptrace;
|
||||
p = list_entry(_p,struct task_struct,sibling);
|
||||
p = list_entry(_p, struct task_struct, sibling);
|
||||
|
||||
ptrace = p->ptrace;
|
||||
|
||||
@@ -709,7 +711,7 @@ static void forget_original_parent(struct task_struct * father,
|
||||
list_add(&p->ptrace_list, to_release);
|
||||
}
|
||||
list_for_each_safe(_p, _n, &father->ptrace_children) {
|
||||
p = list_entry(_p,struct task_struct,ptrace_list);
|
||||
p = list_entry(_p, struct task_struct, ptrace_list);
|
||||
choose_new_parent(p, reaper);
|
||||
reparent_thread(p, father, 1);
|
||||
}
|
||||
@@ -829,7 +831,7 @@ static void exit_notify(struct task_struct *tsk)
|
||||
|
||||
list_for_each_safe(_p, _n, &ptrace_dead) {
|
||||
list_del_init(_p);
|
||||
t = list_entry(_p,struct task_struct,ptrace_list);
|
||||
t = list_entry(_p, struct task_struct, ptrace_list);
|
||||
release_task(t);
|
||||
}
|
||||
|
||||
@@ -1010,7 +1012,7 @@ asmlinkage void sys_exit_group(int error_code)
|
||||
do_group_exit((error_code & 0xff) << 8);
|
||||
}
|
||||
|
||||
static int eligible_child(pid_t pid, int options, task_t *p)
|
||||
static int eligible_child(pid_t pid, int options, struct task_struct *p)
|
||||
{
|
||||
if (pid > 0) {
|
||||
if (p->pid != pid)
|
||||
@@ -1051,12 +1053,13 @@ static int eligible_child(pid_t pid, int options, task_t *p)
|
||||
return 1;
|
||||
}
|
||||
|
||||
static int wait_noreap_copyout(task_t *p, pid_t pid, uid_t uid,
|
||||
static int wait_noreap_copyout(struct task_struct *p, pid_t pid, uid_t uid,
|
||||
int why, int status,
|
||||
struct siginfo __user *infop,
|
||||
struct rusage __user *rusagep)
|
||||
{
|
||||
int retval = rusagep ? getrusage(p, RUSAGE_BOTH, rusagep) : 0;
|
||||
|
||||
put_task_struct(p);
|
||||
if (!retval)
|
||||
retval = put_user(SIGCHLD, &infop->si_signo);
|
||||
@@ -1081,7 +1084,7 @@ static int wait_noreap_copyout(task_t *p, pid_t pid, uid_t uid,
|
||||
* the lock and this task is uninteresting. If we return nonzero, we have
|
||||
* released the lock and the system call should return.
|
||||
*/
|
||||
static int wait_task_zombie(task_t *p, int noreap,
|
||||
static int wait_task_zombie(struct task_struct *p, int noreap,
|
||||
struct siginfo __user *infop,
|
||||
int __user *stat_addr, struct rusage __user *ru)
|
||||
{
|
||||
@@ -1243,8 +1246,8 @@ static int wait_task_zombie(task_t *p, int noreap,
|
||||
* the lock and this task is uninteresting. If we return nonzero, we have
|
||||
* released the lock and the system call should return.
|
||||
*/
|
||||
static int wait_task_stopped(task_t *p, int delayed_group_leader, int noreap,
|
||||
struct siginfo __user *infop,
|
||||
static int wait_task_stopped(struct task_struct *p, int delayed_group_leader,
|
||||
int noreap, struct siginfo __user *infop,
|
||||
int __user *stat_addr, struct rusage __user *ru)
|
||||
{
|
||||
int retval, exit_code;
|
||||
@@ -1358,7 +1361,7 @@ bail_ref:
|
||||
* the lock and this task is uninteresting. If we return nonzero, we have
|
||||
* released the lock and the system call should return.
|
||||
*/
|
||||
static int wait_task_continued(task_t *p, int noreap,
|
||||
static int wait_task_continued(struct task_struct *p, int noreap,
|
||||
struct siginfo __user *infop,
|
||||
int __user *stat_addr, struct rusage __user *ru)
|
||||
{
|
||||
@@ -1444,7 +1447,7 @@ repeat:
|
||||
int ret;
|
||||
|
||||
list_for_each(_p,&tsk->children) {
|
||||
p = list_entry(_p,struct task_struct,sibling);
|
||||
p = list_entry(_p, struct task_struct, sibling);
|
||||
|
||||
ret = eligible_child(pid, options, p);
|
||||
if (!ret)
|
||||
|
||||
+9
-9
@@ -933,13 +933,13 @@ static inline void rt_mutex_init_task(struct task_struct *p)
|
||||
* parts of the process environment (as per the clone
|
||||
* flags). The actual kick-off is left to the caller.
|
||||
*/
|
||||
static task_t *copy_process(unsigned long clone_flags,
|
||||
unsigned long stack_start,
|
||||
struct pt_regs *regs,
|
||||
unsigned long stack_size,
|
||||
int __user *parent_tidptr,
|
||||
int __user *child_tidptr,
|
||||
int pid)
|
||||
static struct task_struct *copy_process(unsigned long clone_flags,
|
||||
unsigned long stack_start,
|
||||
struct pt_regs *regs,
|
||||
unsigned long stack_size,
|
||||
int __user *parent_tidptr,
|
||||
int __user *child_tidptr,
|
||||
int pid)
|
||||
{
|
||||
int retval;
|
||||
struct task_struct *p = NULL;
|
||||
@@ -1294,9 +1294,9 @@ struct pt_regs * __devinit __attribute__((weak)) idle_regs(struct pt_regs *regs)
|
||||
return regs;
|
||||
}
|
||||
|
||||
task_t * __devinit fork_idle(int cpu)
|
||||
struct task_struct * __devinit fork_idle(int cpu)
|
||||
{
|
||||
task_t *task;
|
||||
struct task_struct *task;
|
||||
struct pt_regs regs;
|
||||
|
||||
task = copy_process(CLONE_VM, 0, idle_regs(®s), 0, NULL, NULL, 0);
|
||||
|
||||
+1
-1
@@ -669,7 +669,7 @@ static int hrtimer_wakeup(struct hrtimer *timer)
|
||||
return HRTIMER_NORESTART;
|
||||
}
|
||||
|
||||
void hrtimer_init_sleeper(struct hrtimer_sleeper *sl, task_t *task)
|
||||
void hrtimer_init_sleeper(struct hrtimer_sleeper *sl, struct task_struct *task)
|
||||
{
|
||||
sl->timer.function = hrtimer_wakeup;
|
||||
sl->task = task;
|
||||
|
||||
+3
-3
@@ -218,7 +218,7 @@ struct pid * fastcall find_pid(int nr)
|
||||
return NULL;
|
||||
}
|
||||
|
||||
int fastcall attach_pid(task_t *task, enum pid_type type, int nr)
|
||||
int fastcall attach_pid(struct task_struct *task, enum pid_type type, int nr)
|
||||
{
|
||||
struct pid_link *link;
|
||||
struct pid *pid;
|
||||
@@ -233,7 +233,7 @@ int fastcall attach_pid(task_t *task, enum pid_type type, int nr)
|
||||
return 0;
|
||||
}
|
||||
|
||||
void fastcall detach_pid(task_t *task, enum pid_type type)
|
||||
void fastcall detach_pid(struct task_struct *task, enum pid_type type)
|
||||
{
|
||||
struct pid_link *link;
|
||||
struct pid *pid;
|
||||
@@ -267,7 +267,7 @@ struct task_struct * fastcall pid_task(struct pid *pid, enum pid_type type)
|
||||
/*
|
||||
* Must be called under rcu_read_lock() or with tasklist_lock read-held.
|
||||
*/
|
||||
task_t *find_task_by_pid_type(int type, int nr)
|
||||
struct task_struct *find_task_by_pid_type(int type, int nr)
|
||||
{
|
||||
return pid_task(find_pid(nr), type);
|
||||
}
|
||||
|
||||
+3
-3
@@ -28,7 +28,7 @@
|
||||
*
|
||||
* Must be called with the tasklist lock write-held.
|
||||
*/
|
||||
void __ptrace_link(task_t *child, task_t *new_parent)
|
||||
void __ptrace_link(struct task_struct *child, struct task_struct *new_parent)
|
||||
{
|
||||
BUG_ON(!list_empty(&child->ptrace_list));
|
||||
if (child->parent == new_parent)
|
||||
@@ -46,7 +46,7 @@ void __ptrace_link(task_t *child, task_t *new_parent)
|
||||
* TASK_TRACED, resume it now.
|
||||
* Requires that irqs be disabled.
|
||||
*/
|
||||
void ptrace_untrace(task_t *child)
|
||||
void ptrace_untrace(struct task_struct *child)
|
||||
{
|
||||
spin_lock(&child->sighand->siglock);
|
||||
if (child->state == TASK_TRACED) {
|
||||
@@ -65,7 +65,7 @@ void ptrace_untrace(task_t *child)
|
||||
*
|
||||
* Must be called with the tasklist lock write-held.
|
||||
*/
|
||||
void __ptrace_unlink(task_t *child)
|
||||
void __ptrace_unlink(struct task_struct *child)
|
||||
{
|
||||
BUG_ON(!child->ptrace);
|
||||
|
||||
|
||||
@@ -96,7 +96,7 @@ void deadlock_trace_off(void)
|
||||
rt_trace_on = 0;
|
||||
}
|
||||
|
||||
static void printk_task(task_t *p)
|
||||
static void printk_task(struct task_struct *p)
|
||||
{
|
||||
if (p)
|
||||
printk("%16s:%5d [%p, %3d]", p->comm, p->pid, p, p->prio);
|
||||
@@ -231,7 +231,8 @@ void debug_rt_mutex_init(struct rt_mutex *lock, const char *name)
|
||||
lock->name = name;
|
||||
}
|
||||
|
||||
void rt_mutex_deadlock_account_lock(struct rt_mutex *lock, task_t *task)
|
||||
void
|
||||
rt_mutex_deadlock_account_lock(struct rt_mutex *lock, struct task_struct *task)
|
||||
{
|
||||
}
|
||||
|
||||
|
||||
@@ -33,7 +33,7 @@ struct test_thread_data {
|
||||
};
|
||||
|
||||
static struct test_thread_data thread_data[MAX_RT_TEST_THREADS];
|
||||
static task_t *threads[MAX_RT_TEST_THREADS];
|
||||
static struct task_struct *threads[MAX_RT_TEST_THREADS];
|
||||
static struct rt_mutex mutexes[MAX_RT_TEST_MUTEXES];
|
||||
|
||||
enum test_opcodes {
|
||||
@@ -361,8 +361,8 @@ static ssize_t sysfs_test_command(struct sys_device *dev, const char *buf,
|
||||
static ssize_t sysfs_test_status(struct sys_device *dev, char *buf)
|
||||
{
|
||||
struct test_thread_data *td;
|
||||
struct task_struct *tsk;
|
||||
char *curr = buf;
|
||||
task_t *tsk;
|
||||
int i;
|
||||
|
||||
td = container_of(dev, struct test_thread_data, sysdev);
|
||||
|
||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user