You've already forked linux-apfs
mirror of
https://github.com/linux-apfs/linux-apfs.git
synced 2026-05-01 15:00:59 -07:00
Merge git://git.kernel.org/pub/scm/linux/kernel/git/kyle/parisc-2.6
* git://git.kernel.org/pub/scm/linux/kernel/git/kyle/parisc-2.6: (23 commits) parisc: move dereference_function_descriptor to process.c parisc: Move kernel Elf_Fdesc define to <asm/elf.h> parisc: fix build when ARCH_HAS_KMAP parisc: fix "make tar-pkg" parisc: drivers: fix warnings parisc: select BUG always parisc: asm/pdc.h should include asm/page.h parisc: led: remove proc_dir_entry::owner parisc: fix macro expansion in atomic.h parisc: iosapic: fix build breakage parisc: oops_enter()/oops_exit() in die() parisc: document light weight syscall ABI parisc: blink all or loadavg LEDs on oops parisc: add ftrace (function and graph tracer) functionality parisc: simplify sys_clone() parisc: add LATENCYTOP_SUPPORT and CONFIG_STACKTRACE_SUPPORT parisc: allow to build with 16k default kernel page size parisc: expose 32/64-bit capabilities in cpuinfo parisc: use constants instead of numbers in assembly parisc: fix usage of 32bit PTE page table entries on 32bit kernels ...
This commit is contained in:
@@ -9,9 +9,13 @@ config PARISC
|
||||
def_bool y
|
||||
select HAVE_IDE
|
||||
select HAVE_OPROFILE
|
||||
select HAVE_FUNCTION_TRACER if 64BIT
|
||||
select HAVE_FUNCTION_GRAPH_TRACER if 64BIT
|
||||
select HAVE_FUNCTION_TRACE_MCOUNT_TEST if 64BIT
|
||||
select RTC_CLASS
|
||||
select RTC_DRV_GENERIC
|
||||
select INIT_ALL_POSSIBLE
|
||||
select BUG
|
||||
help
|
||||
The PA-RISC microprocessor is designed by Hewlett-Packard and used
|
||||
in many of their workstations & servers (HP9000 700 and 800 series,
|
||||
@@ -75,6 +79,9 @@ config GENERIC_HARDIRQS
|
||||
config GENERIC_IRQ_PROBE
|
||||
def_bool y
|
||||
|
||||
config HAVE_LATENCYTOP_SUPPORT
|
||||
def_bool y
|
||||
|
||||
config IRQ_PER_CPU
|
||||
bool
|
||||
default y
|
||||
@@ -83,6 +90,9 @@ config IRQ_PER_CPU
|
||||
config PM
|
||||
bool
|
||||
|
||||
config STACKTRACE_SUPPORT
|
||||
def_bool y
|
||||
|
||||
config ISA_DMA_API
|
||||
bool
|
||||
|
||||
|
||||
@@ -56,7 +56,9 @@ cflags-y += -mdisable-fpregs
|
||||
|
||||
# Without this, "ld -r" results in .text sections that are too big
|
||||
# (> 0x40000) for branches to reach stubs.
|
||||
cflags-y += -ffunction-sections
|
||||
ifndef CONFIG_FUNCTION_TRACER
|
||||
cflags-y += -ffunction-sections
|
||||
endif
|
||||
|
||||
# select which processor to optimise for
|
||||
cflags-$(CONFIG_PA7100) += -march=1.1 -mschedule=7100
|
||||
|
||||
@@ -25,7 +25,7 @@
|
||||
* Since "a" is usually an address, use one spinlock per cacheline.
|
||||
*/
|
||||
# define ATOMIC_HASH_SIZE 4
|
||||
# define ATOMIC_HASH(a) (&(__atomic_hash[ (((unsigned long) a)/L1_CACHE_BYTES) & (ATOMIC_HASH_SIZE-1) ]))
|
||||
# define ATOMIC_HASH(a) (&(__atomic_hash[ (((unsigned long) (a))/L1_CACHE_BYTES) & (ATOMIC_HASH_SIZE-1) ]))
|
||||
|
||||
extern raw_spinlock_t __atomic_hash[ATOMIC_HASH_SIZE] __lock_aligned;
|
||||
|
||||
@@ -222,13 +222,13 @@ static __inline__ int atomic_add_unless(atomic_t *v, int a, int u)
|
||||
|
||||
#define atomic_inc_not_zero(v) atomic_add_unless((v), 1, 0)
|
||||
|
||||
#define atomic_add(i,v) ((void)(__atomic_add_return( ((int)i),(v))))
|
||||
#define atomic_sub(i,v) ((void)(__atomic_add_return(-((int)i),(v))))
|
||||
#define atomic_add(i,v) ((void)(__atomic_add_return( ((int)(i)),(v))))
|
||||
#define atomic_sub(i,v) ((void)(__atomic_add_return(-((int)(i)),(v))))
|
||||
#define atomic_inc(v) ((void)(__atomic_add_return( 1,(v))))
|
||||
#define atomic_dec(v) ((void)(__atomic_add_return( -1,(v))))
|
||||
|
||||
#define atomic_add_return(i,v) (__atomic_add_return( ((int)i),(v)))
|
||||
#define atomic_sub_return(i,v) (__atomic_add_return(-((int)i),(v)))
|
||||
#define atomic_add_return(i,v) (__atomic_add_return( ((int)(i)),(v)))
|
||||
#define atomic_sub_return(i,v) (__atomic_add_return(-((int)(i)),(v)))
|
||||
#define atomic_inc_return(v) (__atomic_add_return( 1,(v)))
|
||||
#define atomic_dec_return(v) (__atomic_add_return( -1,(v)))
|
||||
|
||||
@@ -289,13 +289,13 @@ atomic64_read(const atomic64_t *v)
|
||||
return v->counter;
|
||||
}
|
||||
|
||||
#define atomic64_add(i,v) ((void)(__atomic64_add_return( ((s64)i),(v))))
|
||||
#define atomic64_sub(i,v) ((void)(__atomic64_add_return(-((s64)i),(v))))
|
||||
#define atomic64_add(i,v) ((void)(__atomic64_add_return( ((s64)(i)),(v))))
|
||||
#define atomic64_sub(i,v) ((void)(__atomic64_add_return(-((s64)(i)),(v))))
|
||||
#define atomic64_inc(v) ((void)(__atomic64_add_return( 1,(v))))
|
||||
#define atomic64_dec(v) ((void)(__atomic64_add_return( -1,(v))))
|
||||
|
||||
#define atomic64_add_return(i,v) (__atomic64_add_return( ((s64)i),(v)))
|
||||
#define atomic64_sub_return(i,v) (__atomic64_add_return(-((s64)i),(v)))
|
||||
#define atomic64_add_return(i,v) (__atomic64_add_return( ((s64)(i)),(v)))
|
||||
#define atomic64_sub_return(i,v) (__atomic64_add_return(-((s64)(i)),(v)))
|
||||
#define atomic64_inc_return(v) (__atomic64_add_return( 1,(v)))
|
||||
#define atomic64_dec_return(v) (__atomic64_add_return( -1,(v)))
|
||||
|
||||
|
||||
@@ -97,6 +97,9 @@ void mark_rodata_ro(void);
|
||||
|
||||
#ifdef CONFIG_PA8X00
|
||||
/* Only pa8800, pa8900 needs this */
|
||||
|
||||
#include <asm/kmap_types.h>
|
||||
|
||||
#define ARCH_HAS_KMAP
|
||||
|
||||
void kunmap_parisc(void *addr);
|
||||
|
||||
@@ -168,6 +168,16 @@ typedef struct elf64_fdesc {
|
||||
__u64 gp;
|
||||
} Elf64_Fdesc;
|
||||
|
||||
#ifdef __KERNEL__
|
||||
|
||||
#ifdef CONFIG_64BIT
|
||||
#define Elf_Fdesc Elf64_Fdesc
|
||||
#else
|
||||
#define Elf_Fdesc Elf32_Fdesc
|
||||
#endif /*CONFIG_64BIT*/
|
||||
|
||||
#endif /*__KERNEL__*/
|
||||
|
||||
/* Legal values for p_type field of Elf32_Phdr/Elf64_Phdr. */
|
||||
|
||||
#define PT_HP_TLS (PT_LOOS + 0x0)
|
||||
|
||||
@@ -0,0 +1,25 @@
|
||||
#ifndef _ASM_PARISC_FTRACE_H
|
||||
#define _ASM_PARISC_FTRACE_H
|
||||
|
||||
#ifndef __ASSEMBLY__
|
||||
extern void mcount(void);
|
||||
|
||||
/*
|
||||
* Stack of return addresses for functions of a thread.
|
||||
* Used in struct thread_info
|
||||
*/
|
||||
struct ftrace_ret_stack {
|
||||
unsigned long ret;
|
||||
unsigned long func;
|
||||
unsigned long long calltime;
|
||||
};
|
||||
|
||||
/*
|
||||
* Primary handler of a function return.
|
||||
* It relays on ftrace_return_to_handler.
|
||||
* Defined in entry.S
|
||||
*/
|
||||
extern void return_to_handler(void);
|
||||
#endif /* __ASSEMBLY__ */
|
||||
|
||||
#endif /* _ASM_PARISC_FTRACE_H */
|
||||
@@ -36,16 +36,7 @@ void clear_user_page(void *page, unsigned long vaddr, struct page *pg);
|
||||
*/
|
||||
#define STRICT_MM_TYPECHECKS
|
||||
#ifdef STRICT_MM_TYPECHECKS
|
||||
typedef struct { unsigned long pte;
|
||||
#if !defined(CONFIG_64BIT)
|
||||
unsigned long future_flags;
|
||||
/* XXX: it's possible to remove future_flags and change BITS_PER_PTE_ENTRY
|
||||
to 2, but then strangely the identical 32bit kernel boots on a
|
||||
c3000(pa20), but not any longer on a 715(pa11).
|
||||
Still investigating... HelgeD.
|
||||
*/
|
||||
#endif
|
||||
} pte_t; /* either 32 or 64bit */
|
||||
typedef struct { unsigned long pte; } pte_t; /* either 32 or 64bit */
|
||||
|
||||
/* NOTE: even on 64 bits, these entries are __u32 because we allocate
|
||||
* the pmd and pgd in ZONE_DMA (i.e. under 4GB) */
|
||||
@@ -111,7 +102,7 @@ extern int npmem_ranges;
|
||||
#define BITS_PER_PMD_ENTRY 2
|
||||
#define BITS_PER_PGD_ENTRY 2
|
||||
#else
|
||||
#define BITS_PER_PTE_ENTRY 3
|
||||
#define BITS_PER_PTE_ENTRY 2
|
||||
#define BITS_PER_PMD_ENTRY 2
|
||||
#define BITS_PER_PGD_ENTRY BITS_PER_PMD_ENTRY
|
||||
#endif
|
||||
|
||||
@@ -49,6 +49,8 @@
|
||||
#define PDC_MODEL_CPU_ID 6 /* returns cpu-id (only newer machines!) */
|
||||
#define PDC_MODEL_CAPABILITIES 7 /* returns OS32/OS64-flags */
|
||||
/* Values for PDC_MODEL_CAPABILITIES non-equivalent virtual aliasing support */
|
||||
#define PDC_MODEL_OS64 (1 << 0)
|
||||
#define PDC_MODEL_OS32 (1 << 1)
|
||||
#define PDC_MODEL_IOPDIR_FDC (1 << 2)
|
||||
#define PDC_MODEL_NVA_MASK (3 << 4)
|
||||
#define PDC_MODEL_NVA_SUPPORTED (0 << 4)
|
||||
@@ -341,6 +343,8 @@
|
||||
|
||||
#ifdef __KERNEL__
|
||||
|
||||
#include <asm/page.h> /* for __PAGE_OFFSET */
|
||||
|
||||
extern int pdc_type;
|
||||
|
||||
/* Values for pdc_type */
|
||||
|
||||
@@ -50,11 +50,7 @@
|
||||
printk("%s:%d: bad pgd %08lx.\n", __FILE__, __LINE__, (unsigned long)pgd_val(e))
|
||||
|
||||
/* This is the size of the initially mapped kernel memory */
|
||||
#ifdef CONFIG_64BIT
|
||||
#define KERNEL_INITIAL_ORDER 24 /* 0 to 1<<24 = 16MB */
|
||||
#else
|
||||
#define KERNEL_INITIAL_ORDER 23 /* 0 to 1<<23 = 8MB */
|
||||
#endif
|
||||
#define KERNEL_INITIAL_SIZE (1 << KERNEL_INITIAL_ORDER)
|
||||
|
||||
#if defined(CONFIG_64BIT) && defined(CONFIG_PARISC_PAGE_SIZE_4KB)
|
||||
@@ -91,16 +87,25 @@
|
||||
|
||||
/* Definitions for 1st level */
|
||||
#define PGDIR_SHIFT (PMD_SHIFT + BITS_PER_PMD)
|
||||
#if (PGDIR_SHIFT + PAGE_SHIFT + PGD_ORDER - BITS_PER_PGD_ENTRY) > BITS_PER_LONG
|
||||
#define BITS_PER_PGD (BITS_PER_LONG - PGDIR_SHIFT)
|
||||
#else
|
||||
#define BITS_PER_PGD (PAGE_SHIFT + PGD_ORDER - BITS_PER_PGD_ENTRY)
|
||||
#endif
|
||||
#define PGDIR_SIZE (1UL << PGDIR_SHIFT)
|
||||
#define PGDIR_MASK (~(PGDIR_SIZE-1))
|
||||
#define PTRS_PER_PGD (1UL << BITS_PER_PGD)
|
||||
#define USER_PTRS_PER_PGD PTRS_PER_PGD
|
||||
|
||||
#ifdef CONFIG_64BIT
|
||||
#define MAX_ADDRBITS (PGDIR_SHIFT + BITS_PER_PGD)
|
||||
#define MAX_ADDRESS (1UL << MAX_ADDRBITS)
|
||||
|
||||
#define SPACEID_SHIFT (MAX_ADDRBITS - 32)
|
||||
#else
|
||||
#define MAX_ADDRBITS (BITS_PER_LONG)
|
||||
#define MAX_ADDRESS (1UL << MAX_ADDRBITS)
|
||||
#define SPACEID_SHIFT 0
|
||||
#endif
|
||||
|
||||
/* This calculates the number of initial pages we need for the initial
|
||||
* page tables */
|
||||
|
||||
@@ -29,7 +29,8 @@ extern void smp_send_reschedule(int cpu);
|
||||
extern void smp_send_all_nop(void);
|
||||
|
||||
extern void arch_send_call_function_single_ipi(int cpu);
|
||||
extern void arch_send_call_function_ipi(cpumask_t mask);
|
||||
extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
|
||||
#define arch_send_call_function_ipi_mask arch_send_call_function_ipi_mask
|
||||
|
||||
#endif /* !ASSEMBLY */
|
||||
|
||||
|
||||
@@ -11,10 +11,25 @@ obj-y := cache.o pacache.o setup.o traps.o time.o irq.o \
|
||||
process.o processor.o pdc_cons.o pdc_chassis.o unwind.o \
|
||||
topology.o
|
||||
|
||||
ifdef CONFIG_FUNCTION_TRACER
|
||||
# Do not profile debug and lowlevel utilities
|
||||
CFLAGS_REMOVE_ftrace.o = -pg
|
||||
CFLAGS_REMOVE_cache.o = -pg
|
||||
CFLAGS_REMOVE_irq.o = -pg
|
||||
CFLAGS_REMOVE_pacache.o = -pg
|
||||
CFLAGS_REMOVE_perf.o = -pg
|
||||
CFLAGS_REMOVE_traps.o = -pg
|
||||
CFLAGS_REMOVE_unaligned.o = -pg
|
||||
CFLAGS_REMOVE_unwind.o = -pg
|
||||
endif
|
||||
|
||||
obj-$(CONFIG_SMP) += smp.o
|
||||
obj-$(CONFIG_PA11) += pci-dma.o
|
||||
obj-$(CONFIG_PCI) += pci.o
|
||||
obj-$(CONFIG_MODULES) += module.o
|
||||
obj-$(CONFIG_64BIT) += binfmt_elf32.o sys_parisc32.o signal32.o
|
||||
obj-$(CONFIG_STACKTRACE)+= stacktrace.o
|
||||
# only supported for PCX-W/U in 64-bit mode at the moment
|
||||
obj-$(CONFIG_64BIT) += perf.o perf_asm.o
|
||||
obj-$(CONFIG_FUNCTION_TRACER) += ftrace.o
|
||||
obj-$(CONFIG_FUNCTION_GRAPH_TRACER) += ftrace.o
|
||||
|
||||
+48
-12
@@ -505,6 +505,18 @@
|
||||
STREG \pte,0(\ptep)
|
||||
.endm
|
||||
|
||||
/* bitshift difference between a PFN (based on kernel's PAGE_SIZE)
|
||||
* to a CPU TLB 4k PFN (4k => 12 bits to shift) */
|
||||
#define PAGE_ADD_SHIFT (PAGE_SHIFT-12)
|
||||
|
||||
/* Drop prot bits and convert to page addr for iitlbt and idtlbt */
|
||||
.macro convert_for_tlb_insert20 pte
|
||||
extrd,u \pte,(63-ASM_PFN_PTE_SHIFT)+(63-58)+PAGE_ADD_SHIFT,\
|
||||
64-PAGE_SHIFT-PAGE_ADD_SHIFT,\pte
|
||||
depdi _PAGE_SIZE_ENCODING_DEFAULT,63,\
|
||||
(63-58)+PAGE_ADD_SHIFT,\pte
|
||||
.endm
|
||||
|
||||
/* Convert the pte and prot to tlb insertion values. How
|
||||
* this happens is quite subtle, read below */
|
||||
.macro make_insert_tlb spc,pte,prot
|
||||
@@ -544,8 +556,7 @@
|
||||
depi 1,12,1,\prot
|
||||
|
||||
/* Drop prot bits and convert to page addr for iitlbt and idtlbt */
|
||||
extrd,u \pte,(63-ASM_PFN_PTE_SHIFT)+(63-58),64-PAGE_SHIFT,\pte
|
||||
depdi _PAGE_SIZE_ENCODING_DEFAULT,63,63-58,\pte
|
||||
convert_for_tlb_insert20 \pte
|
||||
.endm
|
||||
|
||||
/* Identical macro to make_insert_tlb above, except it
|
||||
@@ -563,8 +574,8 @@
|
||||
|
||||
/* Get rid of prot bits and convert to page addr for iitlba */
|
||||
|
||||
depi _PAGE_SIZE_ENCODING_DEFAULT,31,ASM_PFN_PTE_SHIFT,\pte
|
||||
extru \pte,24,25,\pte
|
||||
depi 0,31,ASM_PFN_PTE_SHIFT,\pte
|
||||
SHRREG \pte,(ASM_PFN_PTE_SHIFT-(31-26)),\pte
|
||||
.endm
|
||||
|
||||
/* This is for ILP32 PA2.0 only. The TLB insertion needs
|
||||
@@ -1244,10 +1255,9 @@ nadtlb_check_flush_20w:
|
||||
depdi,z 7,7,3,prot
|
||||
depdi 1,10,1,prot
|
||||
|
||||
/* Get rid of prot bits and convert to page addr for idtlbt */
|
||||
/* Drop prot bits from pte and convert to page addr for idtlbt */
|
||||
convert_for_tlb_insert20 pte
|
||||
|
||||
depdi 0,63,12,pte
|
||||
extrd,u pte,56,52,pte
|
||||
idtlbt pte,prot
|
||||
|
||||
rfir
|
||||
@@ -1337,8 +1347,8 @@ nadtlb_check_flush_11:
|
||||
|
||||
/* Get rid of prot bits and convert to page addr for idtlba */
|
||||
|
||||
depi 0,31,12,pte
|
||||
extru pte,24,25,pte
|
||||
depi 0,31,ASM_PFN_PTE_SHIFT,pte
|
||||
SHRREG pte,(ASM_PFN_PTE_SHIFT-(31-26)),pte
|
||||
|
||||
mfsp %sr1,t0 /* Save sr1 so we can use it in tlb inserts */
|
||||
mtsp spc,%sr1
|
||||
@@ -1403,10 +1413,9 @@ nadtlb_check_flush_20:
|
||||
depdi,z 7,7,3,prot
|
||||
depdi 1,10,1,prot
|
||||
|
||||
/* Get rid of prot bits and convert to page addr for idtlbt */
|
||||
/* Drop prot bits from pte and convert to page addr for idtlbt */
|
||||
convert_for_tlb_insert20 pte
|
||||
|
||||
depdi 0,63,12,pte
|
||||
extrd,u pte,56,32,pte
|
||||
idtlbt pte,prot
|
||||
|
||||
rfir
|
||||
@@ -2176,6 +2185,33 @@ syscall_do_resched:
|
||||
ENDPROC(syscall_exit)
|
||||
|
||||
|
||||
#ifdef CONFIG_FUNCTION_TRACER
|
||||
.import ftrace_function_trampoline,code
|
||||
ENTRY(_mcount)
|
||||
copy %r3, %arg2
|
||||
b ftrace_function_trampoline
|
||||
nop
|
||||
ENDPROC(_mcount)
|
||||
|
||||
ENTRY(return_to_handler)
|
||||
load32 return_trampoline, %rp
|
||||
copy %ret0, %arg0
|
||||
copy %ret1, %arg1
|
||||
b ftrace_return_to_handler
|
||||
nop
|
||||
return_trampoline:
|
||||
copy %ret0, %rp
|
||||
copy %r23, %ret0
|
||||
copy %r24, %ret1
|
||||
|
||||
.globl ftrace_stub
|
||||
ftrace_stub:
|
||||
bv %r0(%rp)
|
||||
nop
|
||||
ENDPROC(return_to_handler)
|
||||
#endif /* CONFIG_FUNCTION_TRACER */
|
||||
|
||||
|
||||
get_register:
|
||||
/*
|
||||
* get_register is used by the non access tlb miss handlers to
|
||||
|
||||
@@ -527,7 +527,11 @@ int pdc_model_capabilities(unsigned long *capabilities)
|
||||
pdc_result[0] = 0; /* preset zero (call may not be implemented!) */
|
||||
retval = mem_pdc_call(PDC_MODEL, PDC_MODEL_CAPABILITIES, __pa(pdc_result), 0);
|
||||
convert_to_wide(pdc_result);
|
||||
*capabilities = pdc_result[0];
|
||||
if (retval == PDC_OK) {
|
||||
*capabilities = pdc_result[0];
|
||||
} else {
|
||||
*capabilities = PDC_MODEL_OS32;
|
||||
}
|
||||
spin_unlock_irqrestore(&pdc_lock, flags);
|
||||
|
||||
return retval;
|
||||
|
||||
@@ -0,0 +1,185 @@
|
||||
/*
|
||||
* Code for tracing calls in Linux kernel.
|
||||
* Copyright (C) 2009 Helge Deller <deller@gmx.de>
|
||||
*
|
||||
* based on code for x86 which is:
|
||||
* Copyright (C) 2007-2008 Steven Rostedt <srostedt@redhat.com>
|
||||
*
|
||||
* future possible enhancements:
|
||||
* - add CONFIG_DYNAMIC_FTRACE
|
||||
* - add CONFIG_STACK_TRACER
|
||||
*/
|
||||
|
||||
#include <linux/init.h>
|
||||
#include <linux/ftrace.h>
|
||||
|
||||
#include <asm/sections.h>
|
||||
#include <asm/ftrace.h>
|
||||
|
||||
|
||||
|
||||
#ifdef CONFIG_FUNCTION_GRAPH_TRACER
|
||||
|
||||
/* Add a function return address to the trace stack on thread info.*/
|
||||
static int push_return_trace(unsigned long ret, unsigned long long time,
|
||||
unsigned long func, int *depth)
|
||||
{
|
||||
int index;
|
||||
|
||||
if (!current->ret_stack)
|
||||
return -EBUSY;
|
||||
|
||||
/* The return trace stack is full */
|
||||
if (current->curr_ret_stack == FTRACE_RETFUNC_DEPTH - 1) {
|
||||
atomic_inc(¤t->trace_overrun);
|
||||
return -EBUSY;
|
||||
}
|
||||
|
||||
index = ++current->curr_ret_stack;
|
||||
barrier();
|
||||
current->ret_stack[index].ret = ret;
|
||||
current->ret_stack[index].func = func;
|
||||
current->ret_stack[index].calltime = time;
|
||||
*depth = index;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
/* Retrieve a function return address to the trace stack on thread info.*/
|
||||
static void pop_return_trace(struct ftrace_graph_ret *trace, unsigned long *ret)
|
||||
{
|
||||
int index;
|
||||
|
||||
index = current->curr_ret_stack;
|
||||
|
||||
if (unlikely(index < 0)) {
|
||||
ftrace_graph_stop();
|
||||
WARN_ON(1);
|
||||
/* Might as well panic, otherwise we have no where to go */
|
||||
*ret = (unsigned long)
|
||||
dereference_function_descriptor(&panic);
|
||||
return;
|
||||
}
|
||||
|
||||
*ret = current->ret_stack[index].ret;
|
||||
trace->func = current->ret_stack[index].func;
|
||||
trace->calltime = current->ret_stack[index].calltime;
|
||||
trace->overrun = atomic_read(¤t->trace_overrun);
|
||||
trace->depth = index;
|
||||
barrier();
|
||||
current->curr_ret_stack--;
|
||||
|
||||
}
|
||||
|
||||
/*
|
||||
* Send the trace to the ring-buffer.
|
||||
* @return the original return address.
|
||||
*/
|
||||
unsigned long ftrace_return_to_handler(unsigned long retval0,
|
||||
unsigned long retval1)
|
||||
{
|
||||
struct ftrace_graph_ret trace;
|
||||
unsigned long ret;
|
||||
|
||||
pop_return_trace(&trace, &ret);
|
||||
trace.rettime = cpu_clock(raw_smp_processor_id());
|
||||
ftrace_graph_return(&trace);
|
||||
|
||||
if (unlikely(!ret)) {
|
||||
ftrace_graph_stop();
|
||||
WARN_ON(1);
|
||||
/* Might as well panic. What else to do? */
|
||||
ret = (unsigned long)
|
||||
dereference_function_descriptor(&panic);
|
||||
}
|
||||
|
||||
/* HACK: we hand over the old functions' return values
|
||||
in %r23 and %r24. Assembly in entry.S will take care
|
||||
and move those to their final registers %ret0 and %ret1 */
|
||||
asm( "copy %0, %%r23 \n\t"
|
||||
"copy %1, %%r24 \n" : : "r" (retval0), "r" (retval1) );
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
/*
|
||||
* Hook the return address and push it in the stack of return addrs
|
||||
* in current thread info.
|
||||
*/
|
||||
void prepare_ftrace_return(unsigned long *parent, unsigned long self_addr)
|
||||
{
|
||||
unsigned long old;
|
||||
unsigned long long calltime;
|
||||
struct ftrace_graph_ent trace;
|
||||
|
||||
if (unlikely(atomic_read(¤t->tracing_graph_pause)))
|
||||
return;
|
||||
|
||||
old = *parent;
|
||||
*parent = (unsigned long)
|
||||
dereference_function_descriptor(&return_to_handler);
|
||||
|
||||
if (unlikely(!__kernel_text_address(old))) {
|
||||
ftrace_graph_stop();
|
||||
*parent = old;
|
||||
WARN_ON(1);
|
||||
return;
|
||||
}
|
||||
|
||||
calltime = cpu_clock(raw_smp_processor_id());
|
||||
|
||||
if (push_return_trace(old, calltime,
|
||||
self_addr, &trace.depth) == -EBUSY) {
|
||||
*parent = old;
|
||||
return;
|
||||
}
|
||||
|
||||
trace.func = self_addr;
|
||||
|
||||
/* Only trace if the calling function expects to */
|
||||
if (!ftrace_graph_entry(&trace)) {
|
||||
current->curr_ret_stack--;
|
||||
*parent = old;
|
||||
}
|
||||
}
|
||||
|
||||
#endif /* CONFIG_FUNCTION_GRAPH_TRACER */
|
||||
|
||||
|
||||
void ftrace_function_trampoline(unsigned long parent,
|
||||
unsigned long self_addr,
|
||||
unsigned long org_sp_gr3)
|
||||
{
|
||||
extern ftrace_func_t ftrace_trace_function;
|
||||
|
||||
if (function_trace_stop)
|
||||
return;
|
||||
|
||||
if (ftrace_trace_function != ftrace_stub) {
|
||||
ftrace_trace_function(parent, self_addr);
|
||||
return;
|
||||
}
|
||||
#ifdef CONFIG_FUNCTION_GRAPH_TRACER
|
||||
if (ftrace_graph_entry && ftrace_graph_return) {
|
||||
unsigned long sp;
|
||||
unsigned long *parent_rp;
|
||||
|
||||
asm volatile ("copy %%r30, %0" : "=r"(sp));
|
||||
/* sanity check: is stack pointer which we got from
|
||||
assembler function in entry.S in a reasonable
|
||||
range compared to current stack pointer? */
|
||||
if ((sp - org_sp_gr3) > 0x400)
|
||||
return;
|
||||
|
||||
/* calculate pointer to %rp in stack */
|
||||
parent_rp = (unsigned long *) org_sp_gr3 - 0x10;
|
||||
/* sanity check: parent_rp should hold parent */
|
||||
if (*parent_rp != parent)
|
||||
return;
|
||||
|
||||
prepare_ftrace_return(parent_rp, self_addr);
|
||||
return;
|
||||
}
|
||||
#endif
|
||||
}
|
||||
|
||||
@@ -311,12 +311,12 @@ unsigned long txn_alloc_addr(unsigned int virt_irq)
|
||||
next_cpu++; /* assign to "next" CPU we want this bugger on */
|
||||
|
||||
/* validate entry */
|
||||
while ((next_cpu < NR_CPUS) &&
|
||||
while ((next_cpu < nr_cpu_ids) &&
|
||||
(!per_cpu(cpu_data, next_cpu).txn_addr ||
|
||||
!cpu_online(next_cpu)))
|
||||
next_cpu++;
|
||||
|
||||
if (next_cpu >= NR_CPUS)
|
||||
if (next_cpu >= nr_cpu_ids)
|
||||
next_cpu = 0; /* nothing else, assign monarch */
|
||||
|
||||
return txn_affinity_addr(virt_irq, next_cpu);
|
||||
|
||||
@@ -61,9 +61,7 @@
|
||||
#include <linux/string.h>
|
||||
#include <linux/kernel.h>
|
||||
#include <linux/bug.h>
|
||||
#include <linux/uaccess.h>
|
||||
|
||||
#include <asm/sections.h>
|
||||
#include <asm/unwind.h>
|
||||
|
||||
#if 0
|
||||
@@ -115,8 +113,6 @@ struct got_entry {
|
||||
Elf32_Addr addr;
|
||||
};
|
||||
|
||||
#define Elf_Fdesc Elf32_Fdesc
|
||||
|
||||
struct stub_entry {
|
||||
Elf32_Word insns[2]; /* each stub entry has two insns */
|
||||
};
|
||||
@@ -125,8 +121,6 @@ struct got_entry {
|
||||
Elf64_Addr addr;
|
||||
};
|
||||
|
||||
#define Elf_Fdesc Elf64_Fdesc
|
||||
|
||||
struct stub_entry {
|
||||
Elf64_Word insns[4]; /* each stub entry has four insns */
|
||||
};
|
||||
@@ -916,15 +910,3 @@ void module_arch_cleanup(struct module *mod)
|
||||
deregister_unwind_table(mod);
|
||||
module_bug_cleanup(mod);
|
||||
}
|
||||
|
||||
#ifdef CONFIG_64BIT
|
||||
void *dereference_function_descriptor(void *ptr)
|
||||
{
|
||||
Elf64_Fdesc *desc = ptr;
|
||||
void *p;
|
||||
|
||||
if (!probe_kernel_address(&desc->addr, p))
|
||||
ptr = p;
|
||||
return ptr;
|
||||
}
|
||||
#endif
|
||||
|
||||
@@ -153,5 +153,10 @@ EXPORT_SYMBOL(node_data);
|
||||
EXPORT_SYMBOL(pfnnid_map);
|
||||
#endif
|
||||
|
||||
#ifdef CONFIG_FUNCTION_TRACER
|
||||
extern void _mcount(void);
|
||||
EXPORT_SYMBOL(_mcount);
|
||||
#endif
|
||||
|
||||
/* from pacache.S -- needed for copy_page */
|
||||
EXPORT_SYMBOL(copy_user_page_asm);
|
||||
|
||||
@@ -46,14 +46,15 @@
|
||||
#include <linux/stddef.h>
|
||||
#include <linux/unistd.h>
|
||||
#include <linux/kallsyms.h>
|
||||
#include <linux/uaccess.h>
|
||||
|
||||
#include <asm/io.h>
|
||||
#include <asm/asm-offsets.h>
|
||||
#include <asm/pdc.h>
|
||||
#include <asm/pdc_chassis.h>
|
||||
#include <asm/pgalloc.h>
|
||||
#include <asm/uaccess.h>
|
||||
#include <asm/unwind.h>
|
||||
#include <asm/sections.h>
|
||||
|
||||
/*
|
||||
* The idle thread. There's no useful work to be
|
||||
@@ -231,8 +232,8 @@ sys_clone(unsigned long clone_flags, unsigned long usp,
|
||||
|
||||
However, these last 3 args are only examined
|
||||
if the proper flags are set. */
|
||||
int __user *child_tidptr;
|
||||
int __user *parent_tidptr;
|
||||
int __user *parent_tidptr = (int __user *)regs->gr[24];
|
||||
int __user *child_tidptr = (int __user *)regs->gr[22];
|
||||
|
||||
/* usp must be word aligned. This also prevents users from
|
||||
* passing in the value 1 (which is the signal for a special
|
||||
@@ -243,16 +244,6 @@ sys_clone(unsigned long clone_flags, unsigned long usp,
|
||||
if (usp == 0)
|
||||
usp = regs->gr[30];
|
||||
|
||||
if (clone_flags & CLONE_PARENT_SETTID)
|
||||
parent_tidptr = (int __user *)regs->gr[24];
|
||||
else
|
||||
parent_tidptr = NULL;
|
||||
|
||||
if (clone_flags & (CLONE_CHILD_SETTID | CLONE_CHILD_CLEARTID))
|
||||
child_tidptr = (int __user *)regs->gr[22];
|
||||
else
|
||||
child_tidptr = NULL;
|
||||
|
||||
return do_fork(clone_flags, usp, regs, 0, parent_tidptr, child_tidptr);
|
||||
}
|
||||
|
||||
@@ -400,3 +391,15 @@ get_wchan(struct task_struct *p)
|
||||
} while (count++ < 16);
|
||||
return 0;
|
||||
}
|
||||
|
||||
#ifdef CONFIG_64BIT
|
||||
void *dereference_function_descriptor(void *ptr)
|
||||
{
|
||||
Elf64_Fdesc *desc = ptr;
|
||||
void *p;
|
||||
|
||||
if (!probe_kernel_address(&desc->addr, p))
|
||||
ptr = p;
|
||||
return ptr;
|
||||
}
|
||||
#endif
|
||||
|
||||
@@ -100,8 +100,8 @@ static int __cpuinit processor_probe(struct parisc_device *dev)
|
||||
struct cpuinfo_parisc *p;
|
||||
|
||||
#ifdef CONFIG_SMP
|
||||
if (num_online_cpus() >= NR_CPUS) {
|
||||
printk(KERN_INFO "num_online_cpus() >= NR_CPUS\n");
|
||||
if (num_online_cpus() >= nr_cpu_ids) {
|
||||
printk(KERN_INFO "num_online_cpus() >= nr_cpu_ids\n");
|
||||
return 1;
|
||||
}
|
||||
#else
|
||||
@@ -214,7 +214,7 @@ static int __cpuinit processor_probe(struct parisc_device *dev)
|
||||
*/
|
||||
#ifdef CONFIG_SMP
|
||||
if (cpuid) {
|
||||
cpu_set(cpuid, cpu_present_map);
|
||||
set_cpu_present(cpuid, true);
|
||||
cpu_up(cpuid);
|
||||
}
|
||||
#endif
|
||||
@@ -364,6 +364,13 @@ show_cpuinfo (struct seq_file *m, void *v)
|
||||
boot_cpu_data.cpu_hz / 1000000,
|
||||
boot_cpu_data.cpu_hz % 1000000 );
|
||||
|
||||
seq_printf(m, "capabilities\t:");
|
||||
if (boot_cpu_data.pdc.capabilities & PDC_MODEL_OS32)
|
||||
seq_printf(m, " os32");
|
||||
if (boot_cpu_data.pdc.capabilities & PDC_MODEL_OS64)
|
||||
seq_printf(m, " os64");
|
||||
seq_printf(m, "\n");
|
||||
|
||||
seq_printf(m, "model\t\t: %s\n"
|
||||
"model name\t: %s\n",
|
||||
boot_cpu_data.pdc.sys_model_name,
|
||||
|
||||
+11
-10
@@ -31,6 +31,7 @@
|
||||
#include <linux/err.h>
|
||||
#include <linux/delay.h>
|
||||
#include <linux/bitops.h>
|
||||
#include <linux/ftrace.h>
|
||||
|
||||
#include <asm/system.h>
|
||||
#include <asm/atomic.h>
|
||||
@@ -113,14 +114,14 @@ halt_processor(void)
|
||||
{
|
||||
/* REVISIT : redirect I/O Interrupts to another CPU? */
|
||||
/* REVISIT : does PM *know* this CPU isn't available? */
|
||||
cpu_clear(smp_processor_id(), cpu_online_map);
|
||||
set_cpu_online(smp_processor_id(), false);
|
||||
local_irq_disable();
|
||||
for (;;)
|
||||
;
|
||||
}
|
||||
|
||||
|
||||
irqreturn_t
|
||||
irqreturn_t __irq_entry
|
||||
ipi_interrupt(int irq, void *dev_id)
|
||||
{
|
||||
int this_cpu = smp_processor_id();
|
||||
@@ -214,11 +215,11 @@ ipi_send(int cpu, enum ipi_message_type op)
|
||||
}
|
||||
|
||||
static void
|
||||
send_IPI_mask(cpumask_t mask, enum ipi_message_type op)
|
||||
send_IPI_mask(const struct cpumask *mask, enum ipi_message_type op)
|
||||
{
|
||||
int cpu;
|
||||
|
||||
for_each_cpu_mask(cpu, mask)
|
||||
for_each_cpu(cpu, mask)
|
||||
ipi_send(cpu, op);
|
||||
}
|
||||
|
||||
@@ -257,7 +258,7 @@ smp_send_all_nop(void)
|
||||
send_IPI_allbutself(IPI_NOP);
|
||||
}
|
||||
|
||||
void arch_send_call_function_ipi(cpumask_t mask)
|
||||
void arch_send_call_function_ipi_mask(const struct cpumask *mask)
|
||||
{
|
||||
send_IPI_mask(mask, IPI_CALL_FUNC);
|
||||
}
|
||||
@@ -296,13 +297,14 @@ smp_cpu_init(int cpunum)
|
||||
mb();
|
||||
|
||||
/* Well, support 2.4 linux scheme as well. */
|
||||
if (cpu_test_and_set(cpunum, cpu_online_map))
|
||||
if (cpu_isset(cpunum, cpu_online_map))
|
||||
{
|
||||
extern void machine_halt(void); /* arch/parisc.../process.c */
|
||||
|
||||
printk(KERN_CRIT "CPU#%d already initialized!\n", cpunum);
|
||||
machine_halt();
|
||||
}
|
||||
set_cpu_online(cpunum, true);
|
||||
|
||||
/* Initialise the idle task for this CPU */
|
||||
atomic_inc(&init_mm.mm_count);
|
||||
@@ -424,8 +426,8 @@ void __init smp_prepare_boot_cpu(void)
|
||||
/* Setup BSP mappings */
|
||||
printk(KERN_INFO "SMP: bootstrap CPU ID is %d\n", bootstrap_processor);
|
||||
|
||||
cpu_set(bootstrap_processor, cpu_online_map);
|
||||
cpu_set(bootstrap_processor, cpu_present_map);
|
||||
set_cpu_online(bootstrap_processor, true);
|
||||
set_cpu_present(bootstrap_processor, true);
|
||||
}
|
||||
|
||||
|
||||
@@ -436,8 +438,7 @@ void __init smp_prepare_boot_cpu(void)
|
||||
*/
|
||||
void __init smp_prepare_cpus(unsigned int max_cpus)
|
||||
{
|
||||
cpus_clear(cpu_present_map);
|
||||
cpu_set(0, cpu_present_map);
|
||||
init_cpu_present(cpumask_of(0));
|
||||
|
||||
parisc_max_cpus = max_cpus;
|
||||
if (!max_cpus)
|
||||
|
||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user