x86: Fix various typos in comments, take #2
Fix another ~42 single-word typos in arch/x86/ code comments, missed a few in the first pass, in particular in .S files. Signed-off-by: Ingo Molnar <mingo@kernel.org> Cc: Borislav Petkov <bp@alien8.de> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: Peter Zijlstra <peterz@infradead.org> Cc: Paolo Bonzini <pbonzini@redhat.com> Cc: Bjorn Helgaas <bhelgaas@google.com> Cc: linux-kernel@vger.kernel.org
This commit is contained in:
@@ -1,6 +1,6 @@
|
||||
// SPDX-License-Identifier: GPL-2.0-only
|
||||
/*
|
||||
* Shared support code for AMD K8 northbridges and derivates.
|
||||
* Shared support code for AMD K8 northbridges and derivatives.
|
||||
* Copyright 2006 Andi Kleen, SUSE Labs.
|
||||
*/
|
||||
|
||||
|
||||
@@ -1025,7 +1025,7 @@ static int apm_enable_power_management(int enable)
|
||||
* status which gives the rough battery status, and current power
|
||||
* source. The bat value returned give an estimate as a percentage
|
||||
* of life and a status value for the battery. The estimated life
|
||||
* if reported is a lifetime in secodnds/minutes at current power
|
||||
* if reported is a lifetime in seconds/minutes at current power
|
||||
* consumption.
|
||||
*/
|
||||
|
||||
|
||||
@@ -301,7 +301,7 @@ static void early_init_intel(struct cpuinfo_x86 *c)
|
||||
* The operating system must reload CR3 to cause the TLB to be flushed"
|
||||
*
|
||||
* As a result, boot_cpu_has(X86_FEATURE_PGE) in arch/x86/include/asm/tlbflush.h
|
||||
* should be false so that __flush_tlb_all() causes CR3 insted of CR4.PGE
|
||||
* should be false so that __flush_tlb_all() causes CR3 instead of CR4.PGE
|
||||
* to be modified.
|
||||
*/
|
||||
if (c->x86 == 5 && c->x86_model == 9) {
|
||||
|
||||
@@ -142,7 +142,7 @@ static struct severity {
|
||||
MASK(MCI_STATUS_OVER|MCI_UC_SAR, MCI_STATUS_UC|MCI_STATUS_AR)
|
||||
),
|
||||
MCESEV(
|
||||
KEEP, "Non signalled machine check",
|
||||
KEEP, "Non signaled machine check",
|
||||
SER, BITCLR(MCI_STATUS_S)
|
||||
),
|
||||
|
||||
|
||||
@@ -799,7 +799,7 @@ void mtrr_ap_init(void)
|
||||
*
|
||||
* This routine is called in two cases:
|
||||
*
|
||||
* 1. very earily time of software resume, when there absolutely
|
||||
* 1. very early time of software resume, when there absolutely
|
||||
* isn't mtrr entry changes;
|
||||
*
|
||||
* 2. cpu hotadd time. We let mtrr_add/del_page hold cpuhotplug
|
||||
|
||||
@@ -397,7 +397,7 @@ void mon_event_count(void *info)
|
||||
* timer. Having 1s interval makes the calculation of bandwidth simpler.
|
||||
*
|
||||
* Although MBA's goal is to restrict the bandwidth to a maximum, there may
|
||||
* be a need to increase the bandwidth to avoid uncecessarily restricting
|
||||
* be a need to increase the bandwidth to avoid unnecessarily restricting
|
||||
* the L2 <-> L3 traffic.
|
||||
*
|
||||
* Since MBA controls the L2 external bandwidth where as MBM measures the
|
||||
@@ -480,7 +480,7 @@ static void update_mba_bw(struct rdtgroup *rgrp, struct rdt_domain *dom_mbm)
|
||||
|
||||
/*
|
||||
* Delta values are updated dynamically package wise for each
|
||||
* rdtgrp everytime the throttle MSR changes value.
|
||||
* rdtgrp every time the throttle MSR changes value.
|
||||
*
|
||||
* This is because (1)the increase in bandwidth is not perfectly
|
||||
* linear and only "approximately" linear even when the hardware
|
||||
|
||||
@@ -2555,7 +2555,7 @@ static int mkdir_mondata_subdir_alldom(struct kernfs_node *parent_kn,
|
||||
/*
|
||||
* This creates a directory mon_data which contains the monitored data.
|
||||
*
|
||||
* mon_data has one directory for each domain whic are named
|
||||
* mon_data has one directory for each domain which are named
|
||||
* in the format mon_<domain_name>_<domain_id>. For ex: A mon_data
|
||||
* with L3 domain looks as below:
|
||||
* ./mon_data:
|
||||
|
||||
@@ -107,7 +107,7 @@ SYM_CODE_START_LOCAL_NOALIGN(identity_mapped)
|
||||
* - Write protect disabled
|
||||
* - No task switch
|
||||
* - Don't do FP software emulation.
|
||||
* - Proctected mode enabled
|
||||
* - Protected mode enabled
|
||||
*/
|
||||
movl %cr0, %eax
|
||||
andl $~(X86_CR0_PG | X86_CR0_AM | X86_CR0_WP | X86_CR0_TS | X86_CR0_EM), %eax
|
||||
|
||||
@@ -121,7 +121,7 @@ SYM_CODE_START_LOCAL_NOALIGN(identity_mapped)
|
||||
* - Write protect disabled
|
||||
* - No task switch
|
||||
* - Don't do FP software emulation.
|
||||
* - Proctected mode enabled
|
||||
* - Protected mode enabled
|
||||
*/
|
||||
movq %cr0, %rax
|
||||
andq $~(X86_CR0_AM | X86_CR0_WP | X86_CR0_TS | X86_CR0_EM), %rax
|
||||
|
||||
@@ -204,7 +204,7 @@ static void native_stop_other_cpus(int wait)
|
||||
}
|
||||
/*
|
||||
* Don't wait longer than 10 ms if the caller didn't
|
||||
* reqeust it. If wait is true, the machine hangs here if
|
||||
* request it. If wait is true, the machine hangs here if
|
||||
* one or more CPUs do not reach shutdown state.
|
||||
*/
|
||||
timeout = USEC_PER_MSEC * 10;
|
||||
|
||||
@@ -472,7 +472,7 @@ retry:
|
||||
/*
|
||||
* Add the result to the previous adjustment value.
|
||||
*
|
||||
* The adjustement value is slightly off by the overhead of the
|
||||
* The adjustment value is slightly off by the overhead of the
|
||||
* sync mechanism (observed values are ~200 TSC cycles), but this
|
||||
* really depends on CPU, node distance and frequency. So
|
||||
* compensating for this is hard to get right. Experiments show
|
||||
|
||||
@@ -272,7 +272,7 @@ static int emulate_umip_insn(struct insn *insn, int umip_inst,
|
||||
* by whether the operand is a register or a memory location.
|
||||
* If operand is a register, return as many bytes as the operand
|
||||
* size. If operand is memory, return only the two least
|
||||
* siginificant bytes.
|
||||
* significant bytes.
|
||||
*/
|
||||
if (X86_MODRM_MOD(insn->modrm.value) == 3)
|
||||
*data_size = insn->opnd_bytes;
|
||||
|
||||
Reference in New Issue
Block a user