You've already forked linux-apfs
mirror of
https://github.com/linux-apfs/linux-apfs.git
synced 2026-05-01 15:00:59 -07:00
x86: Fix misspellings in comments
Signed-off-by: Adam Buchbinder <adam.buchbinder@gmail.com> Cc: Linus Torvalds <torvalds@linux-foundation.org> Cc: Peter Zijlstra <peterz@infradead.org> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: trivial@kernel.org Signed-off-by: Ingo Molnar <mingo@kernel.org>
This commit is contained in:
committed by
Ingo Molnar
parent
1b74dde7c4
commit
6a6256f9e0
@@ -178,7 +178,7 @@ notrace static cycle_t vread_tsc(void)
|
||||
|
||||
/*
|
||||
* GCC likes to generate cmov here, but this branch is extremely
|
||||
* predictable (it's just a funciton of time and the likely is
|
||||
* predictable (it's just a function of time and the likely is
|
||||
* very likely) and there's a data dependence, so force GCC
|
||||
* to generate a branch instead. I don't barrier() because
|
||||
* we don't actually need a barrier, and if this function
|
||||
|
||||
@@ -52,7 +52,7 @@ int ftrace_int3_handler(struct pt_regs *regs);
|
||||
* this screws up the trace output when tracing a ia32 task.
|
||||
* Instead of reporting bogus syscalls, just do not trace them.
|
||||
*
|
||||
* If the user realy wants these, then they should use the
|
||||
* If the user really wants these, then they should use the
|
||||
* raw syscall tracepoints with filtering.
|
||||
*/
|
||||
#define ARCH_TRACE_IGNORE_COMPAT_SYSCALLS 1
|
||||
|
||||
@@ -25,7 +25,7 @@
|
||||
* This should be totally fair - if anything is waiting, a process that wants a
|
||||
* lock will go to the back of the queue. When the currently active lock is
|
||||
* released, if there's a writer at the front of the queue, then that and only
|
||||
* that will be woken up; if there's a bunch of consequtive readers at the
|
||||
* that will be woken up; if there's a bunch of consecutive readers at the
|
||||
* front, then they'll all be woken up, but no other readers will be.
|
||||
*/
|
||||
|
||||
|
||||
@@ -956,7 +956,7 @@ static int __init early_acpi_parse_madt_lapic_addr_ovr(void)
|
||||
|
||||
/*
|
||||
* Note that the LAPIC address is obtained from the MADT (32-bit value)
|
||||
* and (optionally) overriden by a LAPIC_ADDR_OVR entry (64-bit value).
|
||||
* and (optionally) overridden by a LAPIC_ADDR_OVR entry (64-bit value).
|
||||
*/
|
||||
|
||||
count = acpi_table_parse_madt(ACPI_MADT_TYPE_LOCAL_APIC_OVERRIDE,
|
||||
@@ -984,7 +984,7 @@ static int __init acpi_parse_madt_lapic_entries(void)
|
||||
|
||||
/*
|
||||
* Note that the LAPIC address is obtained from the MADT (32-bit value)
|
||||
* and (optionally) overriden by a LAPIC_ADDR_OVR entry (64-bit value).
|
||||
* and (optionally) overridden by a LAPIC_ADDR_OVR entry (64-bit value).
|
||||
*/
|
||||
|
||||
count = acpi_table_parse_madt(ACPI_MADT_TYPE_LOCAL_APIC_OVERRIDE,
|
||||
|
||||
@@ -1611,7 +1611,7 @@ void __init enable_IR_x2apic(void)
|
||||
legacy_pic->mask_all();
|
||||
mask_ioapic_entries();
|
||||
|
||||
/* If irq_remapping_prepare() succeded, try to enable it */
|
||||
/* If irq_remapping_prepare() succeeded, try to enable it */
|
||||
if (ir_stat >= 0)
|
||||
ir_stat = try_to_enable_IR();
|
||||
/* ir_stat contains the remap mode or an error code */
|
||||
|
||||
@@ -860,7 +860,7 @@ int uv_set_vga_state(struct pci_dev *pdev, bool decode,
|
||||
*/
|
||||
void uv_cpu_init(void)
|
||||
{
|
||||
/* CPU 0 initilization will be done via uv_system_init. */
|
||||
/* CPU 0 initialization will be done via uv_system_init. */
|
||||
if (!uv_blade_info)
|
||||
return;
|
||||
|
||||
|
||||
@@ -1088,7 +1088,7 @@ static int apm_get_battery_status(u_short which, u_short *status,
|
||||
* @device: identity of device
|
||||
* @enable: on/off
|
||||
*
|
||||
* Activate or deactive power management on either a specific device
|
||||
* Activate or deactivate power management on either a specific device
|
||||
* or the entire system (%APM_DEVICE_ALL).
|
||||
*/
|
||||
|
||||
|
||||
@@ -82,7 +82,7 @@ static void init_amd_k5(struct cpuinfo_x86 *c)
|
||||
#ifdef CONFIG_X86_32
|
||||
/*
|
||||
* General Systems BIOSen alias the cpu frequency registers
|
||||
* of the Elan at 0x000df000. Unfortuantly, one of the Linux
|
||||
* of the Elan at 0x000df000. Unfortunately, one of the Linux
|
||||
* drivers subsequently pokes it, and changes the CPU speed.
|
||||
* Workaround : Remove the unneeded alias.
|
||||
*/
|
||||
|
||||
@@ -884,7 +884,7 @@ static void identify_cpu(struct cpuinfo_x86 *c)
|
||||
if (this_cpu->c_identify)
|
||||
this_cpu->c_identify(c);
|
||||
|
||||
/* Clear/Set all flags overriden by options, after probe */
|
||||
/* Clear/Set all flags overridden by options, after probe */
|
||||
for (i = 0; i < NCAPINTS; i++) {
|
||||
c->x86_capability[i] &= ~cpu_caps_cleared[i];
|
||||
c->x86_capability[i] |= cpu_caps_set[i];
|
||||
@@ -943,7 +943,7 @@ static void identify_cpu(struct cpuinfo_x86 *c)
|
||||
x86_init_cache_qos(c);
|
||||
|
||||
/*
|
||||
* Clear/Set all flags overriden by options, need do it
|
||||
* Clear/Set all flags overridden by options, need do it
|
||||
* before following smp all cpus cap AND.
|
||||
*/
|
||||
for (i = 0; i < NCAPINTS; i++) {
|
||||
|
||||
@@ -42,7 +42,7 @@ EXPORT_SYMBOL_GPL(mtrr_state);
|
||||
* "BIOS and Kernel Developer's Guide for the AMD Athlon 64 and AMD
|
||||
* Opteron Processors" (26094 Rev. 3.30 February 2006), section
|
||||
* "13.2.1.2 SYSCFG Register": "The MtrrFixDramModEn bit should be set
|
||||
* to 1 during BIOS initalization of the fixed MTRRs, then cleared to
|
||||
* to 1 during BIOS initialization of the fixed MTRRs, then cleared to
|
||||
* 0 for operation."
|
||||
*/
|
||||
static inline void k8_check_syscfg_dram_mod_en(void)
|
||||
|
||||
@@ -272,7 +272,7 @@ struct cpu_hw_events {
|
||||
* events to select for counter rescheduling.
|
||||
*
|
||||
* Care must be taken as the rescheduling algorithm is O(n!) which
|
||||
* will increase scheduling cycles for an over-commited system
|
||||
* will increase scheduling cycles for an over-committed system
|
||||
* dramatically. The number of such EVENT_CONSTRAINT_OVERLAP() macros
|
||||
* and its counter masks must be kept at a minimum.
|
||||
*/
|
||||
|
||||
@@ -649,7 +649,7 @@ int intel_pmu_setup_lbr_filter(struct perf_event *event)
|
||||
|
||||
/*
|
||||
* return the type of control flow change at address "from"
|
||||
* intruction is not necessarily a branch (in case of interrupt).
|
||||
* instruction is not necessarily a branch (in case of interrupt).
|
||||
*
|
||||
* The branch type returned also includes the priv level of the
|
||||
* target of the control flow change (X86_BR_USER, X86_BR_KERNEL).
|
||||
|
||||
@@ -287,7 +287,7 @@ static __init void early_pci_serial_init(char *s)
|
||||
}
|
||||
|
||||
/*
|
||||
* Lastly, initalize the hardware
|
||||
* Lastly, initialize the hardware
|
||||
*/
|
||||
if (*s) {
|
||||
if (strcmp(s, "nocfg") == 0)
|
||||
|
||||
@@ -8,7 +8,7 @@
|
||||
/*
|
||||
* The xstateregs_active() routine is the same as the regset_fpregs_active() routine,
|
||||
* as the "regset->n" for the xstate regset will be updated based on the feature
|
||||
* capabilites supported by the xsave.
|
||||
* capabilities supported by the xsave.
|
||||
*/
|
||||
int regset_fpregs_active(struct task_struct *target, const struct user_regset *regset)
|
||||
{
|
||||
|
||||
@@ -271,7 +271,7 @@ static int bzImage64_probe(const char *buf, unsigned long len)
|
||||
int ret = -ENOEXEC;
|
||||
struct setup_header *header;
|
||||
|
||||
/* kernel should be atleast two sectors long */
|
||||
/* kernel should be at least two sectors long */
|
||||
if (len < 2 * 512) {
|
||||
pr_err("File is too short to be a bzImage\n");
|
||||
return ret;
|
||||
|
||||
@@ -609,9 +609,9 @@ static struct notifier_block kgdb_notifier = {
|
||||
};
|
||||
|
||||
/**
|
||||
* kgdb_arch_init - Perform any architecture specific initalization.
|
||||
* kgdb_arch_init - Perform any architecture specific initialization.
|
||||
*
|
||||
* This function will handle the initalization of any architecture
|
||||
* This function will handle the initialization of any architecture
|
||||
* specific callbacks.
|
||||
*/
|
||||
int kgdb_arch_init(void)
|
||||
|
||||
@@ -226,7 +226,7 @@ static void kvm_setup_secondary_clock(void)
|
||||
* registered memory location. If the guest happens to shutdown, this memory
|
||||
* won't be valid. In cases like kexec, in which you install a new kernel, this
|
||||
* means a random memory location will be kept being written. So before any
|
||||
* kind of shutdown from our side, we unregister the clock by writting anything
|
||||
* kind of shutdown from our side, we unregister the clock by writing anything
|
||||
* that does not have the 'enable' bit set in the msr
|
||||
*/
|
||||
#ifdef CONFIG_KEXEC_CORE
|
||||
|
||||
@@ -876,7 +876,7 @@ void tsc_restore_sched_clock_state(void)
|
||||
local_irq_save(flags);
|
||||
|
||||
/*
|
||||
* We're comming out of suspend, there's no concurrency yet; don't
|
||||
* We're coming out of suspend, there's no concurrency yet; don't
|
||||
* bother being nice about the RCU stuff, just write to both
|
||||
* data fields.
|
||||
*/
|
||||
|
||||
+2
-2
@@ -478,7 +478,7 @@ static bool spte_is_locklessly_modifiable(u64 spte)
|
||||
static bool spte_has_volatile_bits(u64 spte)
|
||||
{
|
||||
/*
|
||||
* Always atomicly update spte if it can be updated
|
||||
* Always atomically update spte if it can be updated
|
||||
* out of mmu-lock, it can ensure dirty bit is not lost,
|
||||
* also, it can help us to get a stable is_writable_pte()
|
||||
* to ensure tlb flush is not missed.
|
||||
@@ -549,7 +549,7 @@ static bool mmu_spte_update(u64 *sptep, u64 new_spte)
|
||||
|
||||
/*
|
||||
* For the spte updated out of mmu-lock is safe, since
|
||||
* we always atomicly update it, see the comments in
|
||||
* we always atomically update it, see the comments in
|
||||
* spte_has_volatile_bits().
|
||||
*/
|
||||
if (spte_is_locklessly_modifiable(old_spte) &&
|
||||
|
||||
+2
-2
@@ -5475,7 +5475,7 @@ static int handle_set_cr4(struct kvm_vcpu *vcpu, unsigned long val)
|
||||
return kvm_set_cr4(vcpu, val);
|
||||
}
|
||||
|
||||
/* called to set cr0 as approriate for clts instruction exit. */
|
||||
/* called to set cr0 as appropriate for clts instruction exit. */
|
||||
static void handle_clts(struct kvm_vcpu *vcpu)
|
||||
{
|
||||
if (is_guest_mode(vcpu)) {
|
||||
@@ -7223,7 +7223,7 @@ static int handle_vmwrite(struct kvm_vcpu *vcpu)
|
||||
/* The value to write might be 32 or 64 bits, depending on L1's long
|
||||
* mode, and eventually we need to write that into a field of several
|
||||
* possible lengths. The code below first zero-extends the value to 64
|
||||
* bit (field_value), and then copies only the approriate number of
|
||||
* bit (field_value), and then copies only the appropriate number of
|
||||
* bits into the vmcs12 field.
|
||||
*/
|
||||
u64 field_value = 0;
|
||||
|
||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user