You've already forked linux-apfs
mirror of
https://github.com/linux-apfs/linux-apfs.git
synced 2026-05-01 15:00:59 -07:00
Merge tag 'arc-v3.13-rc1-part1' of git://git.kernel.org/pub/scm/linux/kernel/git/vgupta/arc
Pull ARC changes from Vineet Gupta: - Towards a working SMP setup (ASID allocation, TLB Flush,...) - Support for TRACE_IRQFLAGS, LOCKDEP - cacheflush backend consolidation for I/D - Lots of allmodconfig fixlets from Chen - Other improvements/fixes * tag 'arc-v3.13-rc1-part1' of git://git.kernel.org/pub/scm/linux/kernel/git/vgupta/arc: (25 commits) ARC: [plat-arcfpga] defconfig update smp, ARC: kill SMP single function call interrupt ARC: [SMP] Disallow RTSC ARC: [SMP] Fix build failures for large NR_CPUS ARC: [SMP] enlarge possible NR_CPUS ARC: [SMP] TLB flush ARC: [SMP] ASID allocation arc: export symbol for pm_power_off in reset.c arc: export symbol for save_stack_trace() in stacktrace.c arc: remove '__init' for get_hw_config_num_irq() arc: remove '__init' for first_lines_of_secondary() arc: remove '__init' for setup_processor() and arc_init_IRQ() arc: kgdb: add default implementation for kgdb_roundup_cpus() ARC: Fix bogus gcc warning and micro-optimise TLB iteration loop ARC: Add support for irqflags tracing and lockdep ARC: Reset the value of Interrupt Priority Register ARC: Reduce #ifdef'ery for unaligned access emulation ARC: Change calling convention of do_page_fault() ARC: cacheflush optim - PTAG can be loop invariant if V-P is const ARC: cacheflush refactor #3: Unify the {d,i}cache flush leaf helpers ...
This commit is contained in:
+9
-7
@@ -35,6 +35,12 @@ config ARC
|
||||
select PERF_USE_VMALLOC
|
||||
select HAVE_DEBUG_STACKOVERFLOW
|
||||
|
||||
config TRACE_IRQFLAGS_SUPPORT
|
||||
def_bool y
|
||||
|
||||
config LOCKDEP_SUPPORT
|
||||
def_bool y
|
||||
|
||||
config SCHED_OMIT_FRAME_POINTER
|
||||
def_bool y
|
||||
|
||||
@@ -130,17 +136,14 @@ if SMP
|
||||
config ARC_HAS_COH_CACHES
|
||||
def_bool n
|
||||
|
||||
config ARC_HAS_COH_RTSC
|
||||
def_bool n
|
||||
|
||||
config ARC_HAS_REENTRANT_IRQ_LV2
|
||||
def_bool n
|
||||
|
||||
endif
|
||||
|
||||
config NR_CPUS
|
||||
int "Maximum number of CPUs (2-32)"
|
||||
range 2 32
|
||||
int "Maximum number of CPUs (2-4096)"
|
||||
range 2 4096
|
||||
depends on SMP
|
||||
default "2"
|
||||
|
||||
@@ -326,8 +329,7 @@ config ARC_HAS_RTSC
|
||||
bool "Insn: RTSC (64-bit r/o cycle counter)"
|
||||
default y
|
||||
depends on ARC_CPU_REL_4_10
|
||||
# if SMP, enable RTSC only if counter is coherent across cores
|
||||
depends on !SMP || ARC_HAS_COH_RTSC
|
||||
depends on !SMP
|
||||
|
||||
endmenu # "ARC CPU Configuration"
|
||||
|
||||
|
||||
@@ -2,6 +2,8 @@ CONFIG_CROSS_COMPILE="arc-linux-uclibc-"
|
||||
# CONFIG_LOCALVERSION_AUTO is not set
|
||||
CONFIG_DEFAULT_HOSTNAME="ARCLinux"
|
||||
# CONFIG_SWAP is not set
|
||||
CONFIG_SYSVIPC=y
|
||||
CONFIG_POSIX_MQUEUE=y
|
||||
CONFIG_HIGH_RES_TIMERS=y
|
||||
CONFIG_IKCONFIG=y
|
||||
CONFIG_IKCONFIG_PROC=y
|
||||
@@ -62,4 +64,5 @@ CONFIG_TMPFS=y
|
||||
CONFIG_NFS_FS=y
|
||||
# CONFIG_ENABLE_WARN_DEPRECATED is not set
|
||||
# CONFIG_ENABLE_MUST_CHECK is not set
|
||||
# CONFIG_DEBUG_PREEMPT is not set
|
||||
CONFIG_XZ_DEC=y
|
||||
|
||||
@@ -17,13 +17,7 @@
|
||||
#endif
|
||||
|
||||
#define L1_CACHE_BYTES (1 << L1_CACHE_SHIFT)
|
||||
|
||||
/* For a rare case where customers have differently config I/D */
|
||||
#define ARC_ICACHE_LINE_LEN L1_CACHE_BYTES
|
||||
#define ARC_DCACHE_LINE_LEN L1_CACHE_BYTES
|
||||
|
||||
#define ICACHE_LINE_MASK (~(ARC_ICACHE_LINE_LEN - 1))
|
||||
#define DCACHE_LINE_MASK (~(ARC_DCACHE_LINE_LEN - 1))
|
||||
#define CACHE_LINE_MASK (~(L1_CACHE_BYTES - 1))
|
||||
|
||||
/*
|
||||
* ARC700 doesn't cache any access in top 256M.
|
||||
|
||||
@@ -18,8 +18,8 @@
|
||||
|
||||
#include <asm-generic/irq.h>
|
||||
|
||||
extern void __init arc_init_IRQ(void);
|
||||
extern int __init get_hw_config_num_irq(void);
|
||||
extern void arc_init_IRQ(void);
|
||||
extern int get_hw_config_num_irq(void);
|
||||
|
||||
void arc_local_timer_setup(unsigned int cpu);
|
||||
|
||||
|
||||
@@ -151,16 +151,38 @@ static inline void arch_unmask_irq(unsigned int irq)
|
||||
|
||||
#else
|
||||
|
||||
#ifdef CONFIG_TRACE_IRQFLAGS
|
||||
|
||||
.macro TRACE_ASM_IRQ_DISABLE
|
||||
bl trace_hardirqs_off
|
||||
.endm
|
||||
|
||||
.macro TRACE_ASM_IRQ_ENABLE
|
||||
bl trace_hardirqs_on
|
||||
.endm
|
||||
|
||||
#else
|
||||
|
||||
.macro TRACE_ASM_IRQ_DISABLE
|
||||
.endm
|
||||
|
||||
.macro TRACE_ASM_IRQ_ENABLE
|
||||
.endm
|
||||
|
||||
#endif
|
||||
|
||||
.macro IRQ_DISABLE scratch
|
||||
lr \scratch, [status32]
|
||||
bic \scratch, \scratch, (STATUS_E1_MASK | STATUS_E2_MASK)
|
||||
flag \scratch
|
||||
TRACE_ASM_IRQ_DISABLE
|
||||
.endm
|
||||
|
||||
.macro IRQ_ENABLE scratch
|
||||
lr \scratch, [status32]
|
||||
or \scratch, \scratch, (STATUS_E1_MASK | STATUS_E2_MASK)
|
||||
flag \scratch
|
||||
TRACE_ASM_IRQ_ENABLE
|
||||
.endm
|
||||
|
||||
#endif /* __ASSEMBLY__ */
|
||||
|
||||
@@ -48,7 +48,7 @@
|
||||
#ifndef __ASSEMBLY__
|
||||
|
||||
typedef struct {
|
||||
unsigned long asid; /* 8 bit MMU PID + Generation cycle */
|
||||
unsigned long asid[NR_CPUS]; /* 8 bit MMU PID + Generation cycle */
|
||||
} mm_context_t;
|
||||
|
||||
#ifdef CONFIG_ARC_DBG_TLB_PARANOIA
|
||||
|
||||
@@ -30,13 +30,13 @@
|
||||
* "Fast Context Switch" i.e. no TLB flush on ctxt-switch
|
||||
*
|
||||
* Linux assigns each task a unique ASID. A simple round-robin allocation
|
||||
* of H/w ASID is done using software tracker @asid_cache.
|
||||
* of H/w ASID is done using software tracker @asid_cpu.
|
||||
* When it reaches max 255, the allocation cycle starts afresh by flushing
|
||||
* the entire TLB and wrapping ASID back to zero.
|
||||
*
|
||||
* A new allocation cycle, post rollover, could potentially reassign an ASID
|
||||
* to a different task. Thus the rule is to refresh the ASID in a new cycle.
|
||||
* The 32 bit @asid_cache (and mm->asid) have 8 bits MMU PID and rest 24 bits
|
||||
* The 32 bit @asid_cpu (and mm->asid) have 8 bits MMU PID and rest 24 bits
|
||||
* serve as cycle/generation indicator and natural 32 bit unsigned math
|
||||
* automagically increments the generation when lower 8 bits rollover.
|
||||
*/
|
||||
@@ -47,9 +47,11 @@
|
||||
#define MM_CTXT_FIRST_CYCLE (MM_CTXT_ASID_MASK + 1)
|
||||
#define MM_CTXT_NO_ASID 0UL
|
||||
|
||||
#define hw_pid(mm) (mm->context.asid & MM_CTXT_ASID_MASK)
|
||||
#define asid_mm(mm, cpu) mm->context.asid[cpu]
|
||||
#define hw_pid(mm, cpu) (asid_mm(mm, cpu) & MM_CTXT_ASID_MASK)
|
||||
|
||||
extern unsigned int asid_cache;
|
||||
DECLARE_PER_CPU(unsigned int, asid_cache);
|
||||
#define asid_cpu(cpu) per_cpu(asid_cache, cpu)
|
||||
|
||||
/*
|
||||
* Get a new ASID if task doesn't have a valid one (unalloc or from prev cycle)
|
||||
@@ -57,6 +59,7 @@ extern unsigned int asid_cache;
|
||||
*/
|
||||
static inline void get_new_mmu_context(struct mm_struct *mm)
|
||||
{
|
||||
const unsigned int cpu = smp_processor_id();
|
||||
unsigned long flags;
|
||||
|
||||
local_irq_save(flags);
|
||||
@@ -71,28 +74,28 @@ static inline void get_new_mmu_context(struct mm_struct *mm)
|
||||
* first need to destroy the context, setting it to invalid
|
||||
* value.
|
||||
*/
|
||||
if (!((mm->context.asid ^ asid_cache) & MM_CTXT_CYCLE_MASK))
|
||||
if (!((asid_mm(mm, cpu) ^ asid_cpu(cpu)) & MM_CTXT_CYCLE_MASK))
|
||||
goto set_hw;
|
||||
|
||||
/* move to new ASID and handle rollover */
|
||||
if (unlikely(!(++asid_cache & MM_CTXT_ASID_MASK))) {
|
||||
if (unlikely(!(++asid_cpu(cpu) & MM_CTXT_ASID_MASK))) {
|
||||
|
||||
flush_tlb_all();
|
||||
local_flush_tlb_all();
|
||||
|
||||
/*
|
||||
* Above checke for rollover of 8 bit ASID in 32 bit container.
|
||||
* If the container itself wrapped around, set it to a non zero
|
||||
* "generation" to distinguish from no context
|
||||
*/
|
||||
if (!asid_cache)
|
||||
asid_cache = MM_CTXT_FIRST_CYCLE;
|
||||
if (!asid_cpu(cpu))
|
||||
asid_cpu(cpu) = MM_CTXT_FIRST_CYCLE;
|
||||
}
|
||||
|
||||
/* Assign new ASID to tsk */
|
||||
mm->context.asid = asid_cache;
|
||||
asid_mm(mm, cpu) = asid_cpu(cpu);
|
||||
|
||||
set_hw:
|
||||
write_aux_reg(ARC_REG_PID, hw_pid(mm) | MMU_ENABLE);
|
||||
write_aux_reg(ARC_REG_PID, hw_pid(mm, cpu) | MMU_ENABLE);
|
||||
|
||||
local_irq_restore(flags);
|
||||
}
|
||||
@@ -104,16 +107,45 @@ set_hw:
|
||||
static inline int
|
||||
init_new_context(struct task_struct *tsk, struct mm_struct *mm)
|
||||
{
|
||||
mm->context.asid = MM_CTXT_NO_ASID;
|
||||
int i;
|
||||
|
||||
for_each_possible_cpu(i)
|
||||
asid_mm(mm, i) = MM_CTXT_NO_ASID;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static inline void destroy_context(struct mm_struct *mm)
|
||||
{
|
||||
unsigned long flags;
|
||||
|
||||
/* Needed to elide CONFIG_DEBUG_PREEMPT warning */
|
||||
local_irq_save(flags);
|
||||
asid_mm(mm, smp_processor_id()) = MM_CTXT_NO_ASID;
|
||||
local_irq_restore(flags);
|
||||
}
|
||||
|
||||
/* Prepare the MMU for task: setup PID reg with allocated ASID
|
||||
If task doesn't have an ASID (never alloc or stolen, get a new ASID)
|
||||
*/
|
||||
static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
|
||||
struct task_struct *tsk)
|
||||
{
|
||||
const int cpu = smp_processor_id();
|
||||
|
||||
/*
|
||||
* Note that the mm_cpumask is "aggregating" only, we don't clear it
|
||||
* for the switched-out task, unlike some other arches.
|
||||
* It is used to enlist cpus for sending TLB flush IPIs and not sending
|
||||
* it to CPUs where a task once ran-on, could cause stale TLB entry
|
||||
* re-use, specially for a multi-threaded task.
|
||||
* e.g. T1 runs on C1, migrates to C3. T2 running on C2 munmaps.
|
||||
* For a non-aggregating mm_cpumask, IPI not sent C1, and if T1
|
||||
* were to re-migrate to C1, it could access the unmapped region
|
||||
* via any existing stale TLB entries.
|
||||
*/
|
||||
cpumask_set_cpu(cpu, mm_cpumask(next));
|
||||
|
||||
#ifndef CONFIG_SMP
|
||||
/* PGD cached in MMU reg to avoid 3 mem lookups: task->mm->pgd */
|
||||
write_aux_reg(ARC_REG_SCRATCH_DATA0, next->pgd);
|
||||
@@ -131,11 +163,6 @@ static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
|
||||
*/
|
||||
#define activate_mm(prev, next) switch_mm(prev, next, NULL)
|
||||
|
||||
static inline void destroy_context(struct mm_struct *mm)
|
||||
{
|
||||
mm->context.asid = MM_CTXT_NO_ASID;
|
||||
}
|
||||
|
||||
/* it seemed that deactivate_mm( ) is a reasonable place to do book-keeping
|
||||
* for retiring-mm. However destroy_context( ) still needs to do that because
|
||||
* between mm_release( ) = >deactive_mm( ) and
|
||||
|
||||
@@ -31,7 +31,7 @@ struct cpuinfo_data {
|
||||
extern int root_mountflags, end_mem;
|
||||
extern int running_on_hw;
|
||||
|
||||
void __init setup_processor(void);
|
||||
void setup_processor(void);
|
||||
void __init setup_arch_memory(void);
|
||||
|
||||
#endif /* __ASMARC_SETUP_H */
|
||||
|
||||
@@ -30,7 +30,7 @@ extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
|
||||
* APIs provided by arch SMP code to rest of arch code
|
||||
*/
|
||||
extern void __init smp_init_cpus(void);
|
||||
extern void __init first_lines_of_secondary(void);
|
||||
extern void first_lines_of_secondary(void);
|
||||
extern const char *arc_platform_smp_cpuinfo(void);
|
||||
|
||||
/*
|
||||
|
||||
@@ -18,11 +18,18 @@ void local_flush_tlb_kernel_range(unsigned long start, unsigned long end);
|
||||
void local_flush_tlb_range(struct vm_area_struct *vma,
|
||||
unsigned long start, unsigned long end);
|
||||
|
||||
/* XXX: Revisit for SMP */
|
||||
#ifndef CONFIG_SMP
|
||||
#define flush_tlb_range(vma, s, e) local_flush_tlb_range(vma, s, e)
|
||||
#define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page)
|
||||
#define flush_tlb_kernel_range(s, e) local_flush_tlb_kernel_range(s, e)
|
||||
#define flush_tlb_all() local_flush_tlb_all()
|
||||
#define flush_tlb_mm(mm) local_flush_tlb_mm(mm)
|
||||
|
||||
#else
|
||||
extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start,
|
||||
unsigned long end);
|
||||
extern void flush_tlb_page(struct vm_area_struct *vma, unsigned long page);
|
||||
extern void flush_tlb_kernel_range(unsigned long start, unsigned long end);
|
||||
extern void flush_tlb_all(void);
|
||||
extern void flush_tlb_mm(struct mm_struct *mm);
|
||||
#endif /* CONFIG_SMP */
|
||||
#endif
|
||||
|
||||
@@ -22,7 +22,8 @@ static inline int
|
||||
misaligned_fixup(unsigned long address, struct pt_regs *regs,
|
||||
struct callee_regs *cregs)
|
||||
{
|
||||
return 0;
|
||||
/* Not fixed */
|
||||
return 1;
|
||||
}
|
||||
#endif
|
||||
|
||||
|
||||
@@ -17,6 +17,8 @@
|
||||
#include <asm/asm-offsets.h>
|
||||
#include <linux/sched.h>
|
||||
|
||||
#define KSP_WORD_OFF ((TASK_THREAD + THREAD_KSP) / 4)
|
||||
|
||||
struct task_struct *__sched
|
||||
__switch_to(struct task_struct *prev_task, struct task_struct *next_task)
|
||||
{
|
||||
@@ -45,7 +47,16 @@ __switch_to(struct task_struct *prev_task, struct task_struct *next_task)
|
||||
#endif
|
||||
|
||||
/* set ksp of outgoing task in tsk->thread.ksp */
|
||||
#if KSP_WORD_OFF <= 255
|
||||
"st.as sp, [%3, %1] \n\t"
|
||||
#else
|
||||
/*
|
||||
* Workaround for NR_CPUS=4k
|
||||
* %1 is bigger than 255 (S9 offset for st.as)
|
||||
*/
|
||||
"add2 r24, %3, %1 \n\t"
|
||||
"st sp, [r24] \n\t"
|
||||
#endif
|
||||
|
||||
"sync \n\t"
|
||||
|
||||
@@ -97,7 +108,7 @@ __switch_to(struct task_struct *prev_task, struct task_struct *next_task)
|
||||
/* FP/BLINK restore generated by gcc (standard func epilogue */
|
||||
|
||||
: "=r"(tmp)
|
||||
: "n"((TASK_THREAD + THREAD_KSP) / 4), "r"(next), "r"(prev)
|
||||
: "n"(KSP_WORD_OFF), "r"(next), "r"(prev)
|
||||
: "blink"
|
||||
);
|
||||
|
||||
|
||||
@@ -14,6 +14,8 @@
|
||||
#include <asm/asm-offsets.h>
|
||||
#include <asm/linkage.h>
|
||||
|
||||
#define KSP_WORD_OFF ((TASK_THREAD + THREAD_KSP) / 4)
|
||||
|
||||
;################### Low Level Context Switch ##########################
|
||||
|
||||
.section .sched.text,"ax",@progbits
|
||||
@@ -28,8 +30,13 @@ __switch_to:
|
||||
SAVE_CALLEE_SAVED_KERNEL
|
||||
|
||||
/* Save the now KSP in task->thread.ksp */
|
||||
st.as sp, [r0, (TASK_THREAD + THREAD_KSP)/4]
|
||||
|
||||
#if KSP_WORD_OFF <= 255
|
||||
st.as sp, [r0, KSP_WORD_OFF]
|
||||
#else
|
||||
/* Workaround for NR_CPUS=4k as ST.as can only take s9 offset */
|
||||
add2 r24, r0, KSP_WORD_OFF
|
||||
st sp, [r24]
|
||||
#endif
|
||||
/*
|
||||
* Return last task in r0 (return reg)
|
||||
* On ARC, Return reg = First Arg reg = r0.
|
||||
|
||||
+15
-9
@@ -250,6 +250,14 @@ ARC_ENTRY handle_interrupt_level1
|
||||
lr r0, [icause1]
|
||||
and r0, r0, 0x1f
|
||||
|
||||
#ifdef CONFIG_TRACE_IRQFLAGS
|
||||
; icause1 needs to be read early, before calling tracing, which
|
||||
; can clobber scratch regs, hence use of stack to stash it
|
||||
push r0
|
||||
TRACE_ASM_IRQ_DISABLE
|
||||
pop r0
|
||||
#endif
|
||||
|
||||
bl.d @arch_do_IRQ
|
||||
mov r1, sp
|
||||
|
||||
@@ -337,9 +345,9 @@ ARC_ENTRY EV_TLBProtV
|
||||
; vineetg: Mar 6th: Random Seg Fault issue #1
|
||||
; ecr and efa were not saved in case an Intr sneaks in
|
||||
; after fake rtie
|
||||
;
|
||||
|
||||
lr r2, [ecr]
|
||||
lr r1, [efa] ; Faulting Data address
|
||||
lr r0, [efa] ; Faulting Data address
|
||||
|
||||
; --------(4) Return from CPU Exception Mode ---------
|
||||
; Fake a rtie, but rtie to next label
|
||||
@@ -348,6 +356,8 @@ ARC_ENTRY EV_TLBProtV
|
||||
|
||||
FAKE_RET_FROM_EXCPN r9
|
||||
|
||||
mov r1, sp
|
||||
|
||||
;------ (5) Type of Protection Violation? ----------
|
||||
;
|
||||
; ProtV Hardware Exception is triggered for Access Faults of 2 types
|
||||
@@ -358,16 +368,12 @@ ARC_ENTRY EV_TLBProtV
|
||||
bbit1 r2, ECR_C_BIT_PROTV_MISALIG_DATA, 4f
|
||||
|
||||
;========= (6a) Access Violation Processing ========
|
||||
mov r0, sp ; pt_regs
|
||||
bl do_page_fault
|
||||
b ret_from_exception
|
||||
|
||||
;========== (6b) Non aligned access ============
|
||||
4:
|
||||
mov r0, r1
|
||||
mov r1, sp ; pt_regs
|
||||
|
||||
#ifdef CONFIG_ARC_MISALIGN_ACCESS
|
||||
SAVE_CALLEE_SAVED_USER
|
||||
mov r2, sp ; callee_regs
|
||||
|
||||
@@ -376,9 +382,6 @@ ARC_ENTRY EV_TLBProtV
|
||||
; TBD: optimize - do this only if a callee reg was involved
|
||||
; either a dst of emulated LD/ST or src with address-writeback
|
||||
RESTORE_CALLEE_SAVED_USER
|
||||
#else
|
||||
bl do_misaligned_error
|
||||
#endif
|
||||
|
||||
b ret_from_exception
|
||||
|
||||
@@ -575,6 +578,7 @@ resume_user_mode_begin:
|
||||
; --- (Slow Path #2) pending signal ---
|
||||
mov r0, sp ; pt_regs for arg to do_signal()/do_notify_resume()
|
||||
|
||||
GET_CURR_THR_INFO_FLAGS r9
|
||||
bbit0 r9, TIF_SIGPENDING, .Lchk_notify_resume
|
||||
|
||||
; Normal Trap/IRQ entry only saves Scratch (caller-saved) regs
|
||||
@@ -640,6 +644,8 @@ resume_kernel_mode:
|
||||
|
||||
restore_regs :
|
||||
|
||||
TRACE_ASM_IRQ_ENABLE
|
||||
|
||||
lr r10, [status32]
|
||||
|
||||
; Restore REG File. In case multiple Events outstanding,
|
||||
|
||||
@@ -95,7 +95,7 @@ stext:
|
||||
;----------------------------------------------------------------
|
||||
; First lines of code run by secondary before jumping to 'C'
|
||||
;----------------------------------------------------------------
|
||||
.section .init.text, "ax",@progbits
|
||||
.section .text, "ax",@progbits
|
||||
.type first_lines_of_secondary, @function
|
||||
.globl first_lines_of_secondary
|
||||
|
||||
|
||||
@@ -39,10 +39,14 @@ void arc_init_IRQ(void)
|
||||
level_mask |= IS_ENABLED(CONFIG_ARC_IRQ5_LV2) << 5;
|
||||
level_mask |= IS_ENABLED(CONFIG_ARC_IRQ6_LV2) << 6;
|
||||
|
||||
if (level_mask) {
|
||||
/*
|
||||
* Write to register, even if no LV2 IRQs configured to reset it
|
||||
* in case bootloader had mucked with it
|
||||
*/
|
||||
write_aux_reg(AUX_IRQ_LEV, level_mask);
|
||||
|
||||
if (level_mask)
|
||||
pr_info("Level-2 interrupts bitset %x\n", level_mask);
|
||||
write_aux_reg(AUX_IRQ_LEV, level_mask);
|
||||
}
|
||||
}
|
||||
|
||||
/*
|
||||
@@ -146,7 +150,7 @@ void arch_do_IRQ(unsigned int irq, struct pt_regs *regs)
|
||||
set_irq_regs(old_regs);
|
||||
}
|
||||
|
||||
int __init get_hw_config_num_irq(void)
|
||||
int get_hw_config_num_irq(void)
|
||||
{
|
||||
uint32_t val = read_aux_reg(ARC_REG_VECBASE_BCR);
|
||||
|
||||
|
||||
@@ -196,6 +196,18 @@ void kgdb_arch_set_pc(struct pt_regs *regs, unsigned long ip)
|
||||
instruction_pointer(regs) = ip;
|
||||
}
|
||||
|
||||
static void kgdb_call_nmi_hook(void *ignored)
|
||||
{
|
||||
kgdb_nmicallback(raw_smp_processor_id(), NULL);
|
||||
}
|
||||
|
||||
void kgdb_roundup_cpus(unsigned long flags)
|
||||
{
|
||||
local_irq_enable();
|
||||
smp_call_function(kgdb_call_nmi_hook, NULL, 0);
|
||||
local_irq_disable();
|
||||
}
|
||||
|
||||
struct kgdb_arch arch_kgdb_ops = {
|
||||
/* breakpoint instruction: TRAP_S 0x3 */
|
||||
#ifdef CONFIG_CPU_BIG_ENDIAN
|
||||
|
||||
@@ -87,13 +87,13 @@ static void __kprobes save_previous_kprobe(struct kprobe_ctlblk *kcb)
|
||||
|
||||
static void __kprobes restore_previous_kprobe(struct kprobe_ctlblk *kcb)
|
||||
{
|
||||
__get_cpu_var(current_kprobe) = kcb->prev_kprobe.kp;
|
||||
__this_cpu_write(current_kprobe, kcb->prev_kprobe.kp);
|
||||
kcb->kprobe_status = kcb->prev_kprobe.status;
|
||||
}
|
||||
|
||||
static inline void __kprobes set_current_kprobe(struct kprobe *p)
|
||||
{
|
||||
__get_cpu_var(current_kprobe) = p;
|
||||
__this_cpu_write(current_kprobe, p);
|
||||
}
|
||||
|
||||
static void __kprobes resume_execution(struct kprobe *p, unsigned long addr,
|
||||
@@ -237,7 +237,7 @@ int __kprobes arc_kprobe_handler(unsigned long addr, struct pt_regs *regs)
|
||||
|
||||
return 1;
|
||||
} else if (kprobe_running()) {
|
||||
p = __get_cpu_var(current_kprobe);
|
||||
p = __this_cpu_read(current_kprobe);
|
||||
if (p->break_handler && p->break_handler(p, regs)) {
|
||||
setup_singlestep(p, regs);
|
||||
kcb->kprobe_status = KPROBE_HIT_SS;
|
||||
|
||||
@@ -31,3 +31,4 @@ void machine_power_off(void)
|
||||
}
|
||||
|
||||
void (*pm_power_off) (void) = NULL;
|
||||
EXPORT_SYMBOL(pm_power_off);
|
||||
|
||||
@@ -37,8 +37,7 @@ struct task_struct *_current_task[NR_CPUS]; /* For stack switching */
|
||||
|
||||
struct cpuinfo_arc cpuinfo_arc700[NR_CPUS];
|
||||
|
||||
|
||||
void read_arc_build_cfg_regs(void)
|
||||
static void read_arc_build_cfg_regs(void)
|
||||
{
|
||||
struct bcr_perip uncached_space;
|
||||
struct cpuinfo_arc *cpu = &cpuinfo_arc700[smp_processor_id()];
|
||||
@@ -106,7 +105,7 @@ static const struct cpuinfo_data arc_cpu_tbl[] = {
|
||||
{ {0x00, NULL } }
|
||||
};
|
||||
|
||||
char *arc_cpu_mumbojumbo(int cpu_id, char *buf, int len)
|
||||
static char *arc_cpu_mumbojumbo(int cpu_id, char *buf, int len)
|
||||
{
|
||||
int n = 0;
|
||||
struct cpuinfo_arc *cpu = &cpuinfo_arc700[cpu_id];
|
||||
@@ -171,7 +170,7 @@ static const struct id_to_str mac_mul_nm[] = {
|
||||
{0x6, "Dual 16x16 and 32x16"}
|
||||
};
|
||||
|
||||
char *arc_extn_mumbojumbo(int cpu_id, char *buf, int len)
|
||||
static char *arc_extn_mumbojumbo(int cpu_id, char *buf, int len)
|
||||
{
|
||||
int n = 0;
|
||||
struct cpuinfo_arc *cpu = &cpuinfo_arc700[cpu_id];
|
||||
@@ -234,7 +233,7 @@ char *arc_extn_mumbojumbo(int cpu_id, char *buf, int len)
|
||||
return buf;
|
||||
}
|
||||
|
||||
void arc_chk_ccms(void)
|
||||
static void arc_chk_ccms(void)
|
||||
{
|
||||
#if defined(CONFIG_ARC_HAS_DCCM) || defined(CONFIG_ARC_HAS_ICCM)
|
||||
struct cpuinfo_arc *cpu = &cpuinfo_arc700[smp_processor_id()];
|
||||
@@ -269,7 +268,7 @@ void arc_chk_ccms(void)
|
||||
* hardware has dedicated regs which need to be saved/restored on ctx-sw
|
||||
* (Single Precision uses core regs), thus kernel is kind of oblivious to it
|
||||
*/
|
||||
void arc_chk_fpu(void)
|
||||
static void arc_chk_fpu(void)
|
||||
{
|
||||
struct cpuinfo_arc *cpu = &cpuinfo_arc700[smp_processor_id()];
|
||||
|
||||
|
||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user