[PATCH 7/9] lib/cpumask: add num_{possible,present,active}_cpus_{eq,gt,le}

From: Yury Norov
Date: Sat Nov 27 2021 - 23:00:12 EST


Add num_{possible,present,active}_cpus_{eq,gt,le} and replace num_*_cpus()
with one of new functions where appropriate. This allows num_*_cpus_*()
to return earlier depending on the condition.

Signed-off-by: Yury Norov <yury.norov@xxxxxxxxx>
---
arch/arc/kernel/smp.c | 2 +-
arch/arm/kernel/machine_kexec.c | 2 +-
arch/arm/mach-exynos/exynos.c | 2 +-
arch/arm/mm/cache-b15-rac.c | 2 +-
arch/arm64/kernel/smp.c | 2 +-
arch/arm64/mm/context.c | 2 +-
arch/csky/mm/asid.c | 2 +-
arch/csky/mm/context.c | 2 +-
arch/ia64/mm/tlb.c | 6 ++---
arch/mips/kernel/i8253.c | 2 +-
arch/mips/kernel/perf_event_mipsxx.c | 4 ++--
arch/mips/kernel/rtlx-cmp.c | 2 +-
arch/mips/kernel/smp.c | 4 ++--
arch/mips/kernel/vpe-cmp.c | 2 +-
.../loongson2ef/common/cs5536/cs5536_mfgpt.c | 2 +-
arch/mips/mm/context.c | 2 +-
arch/mips/mm/tlbex.c | 2 +-
arch/nios2/kernel/cpuinfo.c | 2 +-
arch/powerpc/platforms/85xx/smp.c | 2 +-
arch/powerpc/platforms/pseries/hotplug-cpu.c | 4 ++--
arch/powerpc/sysdev/mpic.c | 2 +-
arch/powerpc/xmon/xmon.c | 6 ++---
arch/riscv/kvm/vmid.c | 2 +-
arch/sparc/kernel/mdesc.c | 6 ++---
arch/x86/events/amd/core.c | 2 +-
arch/x86/kernel/alternative.c | 8 +++----
arch/x86/kernel/apic/apic.c | 4 ++--
arch/x86/kernel/apic/apic_flat_64.c | 2 +-
arch/x86/kernel/apic/probe_32.c | 2 +-
arch/x86/kernel/cpu/mce/dev-mcelog.c | 2 +-
arch/x86/kernel/hpet.c | 2 +-
arch/x86/kernel/i8253.c | 2 +-
arch/x86/kernel/kvm.c | 2 +-
arch/x86/kernel/kvmclock.c | 2 +-
arch/x86/kernel/tsc.c | 2 +-
arch/x86/xen/smp_pv.c | 2 +-
arch/x86/xen/spinlock.c | 2 +-
drivers/clk/samsung/clk-exynos4.c | 2 +-
drivers/clocksource/ingenic-timer.c | 3 +--
drivers/cpufreq/pcc-cpufreq.c | 2 +-
drivers/dma/mv_xor.c | 5 ++--
drivers/gpu/drm/i810/i810_drv.c | 2 +-
drivers/irqchip/irq-gic.c | 2 +-
drivers/net/caif/caif_virtio.c | 2 +-
.../cavium/liquidio/cn23xx_vf_device.c | 2 +-
drivers/net/ethernet/hisilicon/hns/hns_enet.c | 2 +-
.../net/ethernet/marvell/mvpp2/mvpp2_main.c | 2 +-
drivers/net/wireless/ath/ath9k/hw.c | 2 +-
drivers/net/wireless/marvell/mwifiex/main.c | 4 ++--
drivers/net/wireless/st/cw1200/queue.c | 3 +--
drivers/nvdimm/region.c | 2 +-
drivers/nvme/host/pci.c | 2 +-
drivers/perf/arm_pmu.c | 2 +-
.../intel/speed_select_if/isst_if_common.c | 6 ++---
drivers/soc/bcm/brcmstb/biuctrl.c | 2 +-
drivers/soc/fsl/dpio/dpio-service.c | 4 ++--
drivers/spi/spi-dw-bt1.c | 2 +-
drivers/virt/acrn/hsm.c | 2 +-
fs/xfs/xfs_sysfs.c | 2 +-
include/linux/cpumask.h | 23 +++++++++++++++++++
include/linux/kdb.h | 2 +-
kernel/debug/kdb/kdb_bt.c | 2 +-
kernel/printk/printk.c | 2 +-
kernel/reboot.c | 4 ++--
kernel/time/clockevents.c | 2 +-
mm/percpu.c | 6 ++---
mm/slab.c | 2 +-
67 files changed, 110 insertions(+), 90 deletions(-)

diff --git a/arch/arc/kernel/smp.c b/arch/arc/kernel/smp.c
index 78e6d069b1c1..d4f2765755c9 100644
--- a/arch/arc/kernel/smp.c
+++ b/arch/arc/kernel/smp.c
@@ -103,7 +103,7 @@ void __init smp_prepare_cpus(unsigned int max_cpus)
* if platform didn't set the present map already, do it now
* boot cpu is set to present already by init/main.c
*/
- if (num_present_cpus() <= 1)
+ if (num_present_cpus_le(2))
init_cpu_present(cpu_possible_mask);
}

diff --git a/arch/arm/kernel/machine_kexec.c b/arch/arm/kernel/machine_kexec.c
index f567032a09c0..8875e2ee0083 100644
--- a/arch/arm/kernel/machine_kexec.c
+++ b/arch/arm/kernel/machine_kexec.c
@@ -44,7 +44,7 @@ int machine_kexec_prepare(struct kimage *image)
* and implements CPU hotplug for the current HW. If not, we won't be
* able to kexec reliably, so fail the prepare operation.
*/
- if (num_possible_cpus() > 1 && platform_can_secondary_boot() &&
+ if (num_possible_cpus_gt(1) && platform_can_secondary_boot() &&
!platform_can_cpu_hotplug())
return -EINVAL;

diff --git a/arch/arm/mach-exynos/exynos.c b/arch/arm/mach-exynos/exynos.c
index 8b48326be9fd..ba658402ac1e 100644
--- a/arch/arm/mach-exynos/exynos.c
+++ b/arch/arm/mach-exynos/exynos.c
@@ -120,7 +120,7 @@ void exynos_set_delayed_reset_assertion(bool enable)
if (of_machine_is_compatible("samsung,exynos4")) {
unsigned int tmp, core_id;

- for (core_id = 0; core_id < num_possible_cpus(); core_id++) {
+ for (core_id = 0; num_possible_cpus_gt(core_id); core_id++) {
tmp = pmu_raw_readl(EXYNOS_ARM_CORE_OPTION(core_id));
if (enable)
tmp |= S5P_USE_DELAYED_RESET_ASSERTION;
diff --git a/arch/arm/mm/cache-b15-rac.c b/arch/arm/mm/cache-b15-rac.c
index bdc07030997b..202c3a6cf98b 100644
--- a/arch/arm/mm/cache-b15-rac.c
+++ b/arch/arm/mm/cache-b15-rac.c
@@ -296,7 +296,7 @@ static int __init b15_rac_init(void)
if (!dn)
return -ENODEV;

- if (WARN(num_possible_cpus() > 4, "RAC only supports 4 CPUs\n"))
+ if (WARN(num_possible_cpus_gt(4), "RAC only supports 4 CPUs\n"))
goto out;

b15_rac_base = of_iomap(dn, 0);
diff --git a/arch/arm64/kernel/smp.c b/arch/arm64/kernel/smp.c
index 27df5c1e6baa..bd1280e5081b 100644
--- a/arch/arm64/kernel/smp.c
+++ b/arch/arm64/kernel/smp.c
@@ -1099,7 +1099,7 @@ static bool have_cpu_die(void)

bool cpus_are_stuck_in_kernel(void)
{
- bool smp_spin_tables = (num_possible_cpus() > 1 && !have_cpu_die());
+ bool smp_spin_tables = (num_possible_cpus_gt(1) && !have_cpu_die());

return !!cpus_stuck_in_kernel || smp_spin_tables ||
is_protected_kvm_enabled();
diff --git a/arch/arm64/mm/context.c b/arch/arm64/mm/context.c
index cd72576ae2b7..702248dc105e 100644
--- a/arch/arm64/mm/context.c
+++ b/arch/arm64/mm/context.c
@@ -384,7 +384,7 @@ static int asids_update_limit(void)
* Expect allocation after rollover to fail if we don't have at least
* one more ASID than CPUs. ASID #0 is reserved for init_mm.
*/
- WARN_ON(num_available_asids - 1 <= num_possible_cpus());
+ WARN_ON(num_possible_cpus_gt(num_available_asids - 2));
pr_info("ASID allocator initialised with %lu entries\n",
num_available_asids);

diff --git a/arch/csky/mm/asid.c b/arch/csky/mm/asid.c
index b2e914745c1d..4dd6eb62a9e0 100644
--- a/arch/csky/mm/asid.c
+++ b/arch/csky/mm/asid.c
@@ -176,7 +176,7 @@ int asid_allocator_init(struct asid_info *info,
* Expect allocation after rollover to fail if we don't have at least
* one more ASID than CPUs. ASID #0 is always reserved.
*/
- WARN_ON(NUM_CTXT_ASIDS(info) - 1 <= num_possible_cpus());
+ WARN_ON(num_possible_cpus_gt(NUM_CTXT_ASIDS(info) - 2));
atomic64_set(&info->generation, ASID_FIRST_VERSION(info));
info->map = kcalloc(BITS_TO_LONGS(NUM_CTXT_ASIDS(info)),
sizeof(*info->map), GFP_KERNEL);
diff --git a/arch/csky/mm/context.c b/arch/csky/mm/context.c
index 0d95bdd93846..c12312215bde 100644
--- a/arch/csky/mm/context.c
+++ b/arch/csky/mm/context.c
@@ -28,7 +28,7 @@ static void asid_flush_cpu_ctxt(void)

static int asids_init(void)
{
- BUG_ON(((1 << CONFIG_CPU_ASID_BITS) - 1) <= num_possible_cpus());
+ BUG_ON(num_possible_cpus_gt((1 << CONFIG_CPU_ASID_BITS) - 2));

if (asid_allocator_init(&asid_info, CONFIG_CPU_ASID_BITS, 1,
asid_flush_cpu_ctxt))
diff --git a/arch/ia64/mm/tlb.c b/arch/ia64/mm/tlb.c
index a5bce13ab047..44f623f5dc5e 100644
--- a/arch/ia64/mm/tlb.c
+++ b/arch/ia64/mm/tlb.c
@@ -202,7 +202,7 @@ setup_ptcg_sem(int max_purges, int nptcg_from)
goto resetsema;
}
if (kp_override) {
- need_ptcg_sem = num_possible_cpus() > nptcg;
+ need_ptcg_sem = num_possible_cpus_gt(nptcg);
return;
}

@@ -221,7 +221,7 @@ setup_ptcg_sem(int max_purges, int nptcg_from)
}
if (palo_override) {
if (nptcg != PALO_MAX_TLB_PURGES)
- need_ptcg_sem = (num_possible_cpus() > nptcg);
+ need_ptcg_sem = num_possible_cpus_gt(nptcg);
return;
}

@@ -238,7 +238,7 @@ setup_ptcg_sem(int max_purges, int nptcg_from)
need_ptcg_sem = 0;
return;
} else
- need_ptcg_sem = (num_possible_cpus() > nptcg);
+ need_ptcg_sem = num_possible_cpus_gt(nptcg);

resetsema:
spinaphore_init(&ptcg_sem, max_purges);
diff --git a/arch/mips/kernel/i8253.c b/arch/mips/kernel/i8253.c
index ca21210e06b5..89a63538be4a 100644
--- a/arch/mips/kernel/i8253.c
+++ b/arch/mips/kernel/i8253.c
@@ -29,7 +29,7 @@ void __init setup_pit_timer(void)

static int __init init_pit_clocksource(void)
{
- if (num_possible_cpus() > 1 || /* PIT does not scale! */
+ if (num_possible_cpus_gt(1) || /* PIT does not scale! */
!clockevent_state_periodic(&i8253_clockevent))
return 0;

diff --git a/arch/mips/kernel/perf_event_mipsxx.c b/arch/mips/kernel/perf_event_mipsxx.c
index 1641d274fe37..4b6458899b05 100644
--- a/arch/mips/kernel/perf_event_mipsxx.c
+++ b/arch/mips/kernel/perf_event_mipsxx.c
@@ -135,7 +135,7 @@ static DEFINE_RWLOCK(pmuint_rwlock);
/* Copied from op_model_mipsxx.c */
static unsigned int vpe_shift(void)
{
- if (num_possible_cpus() > 1)
+ if (num_possible_cpus_gt(1))
return 1;

return 0;
@@ -704,7 +704,7 @@ static unsigned int mipspmu_perf_event_encode(const struct mips_perf_event *pev)
* event_id.
*/
#ifdef CONFIG_MIPS_MT_SMP
- if (num_possible_cpus() > 1)
+ if (num_possible_cpus_gt(1))
return ((unsigned int)pev->range << 24) |
(pev->cntr_mask & 0xffff00) |
(pev->event_id & 0xff);
diff --git a/arch/mips/kernel/rtlx-cmp.c b/arch/mips/kernel/rtlx-cmp.c
index d26dcc4b46e7..e4bb83bc46c6 100644
--- a/arch/mips/kernel/rtlx-cmp.c
+++ b/arch/mips/kernel/rtlx-cmp.c
@@ -54,7 +54,7 @@ int __init rtlx_module_init(void)
return -ENODEV;
}

- if (num_possible_cpus() - aprp_cpu_index() < 1) {
+ if (num_possible_cpus_le(aprp_cpu_index() + 1)) {
pr_warn("No TCs reserved for AP/SP, not initializing RTLX.\n"
"Pass maxcpus=<n> argument as kernel argument\n");

diff --git a/arch/mips/kernel/smp.c b/arch/mips/kernel/smp.c
index d542fb7af3ba..6a0bbf249528 100644
--- a/arch/mips/kernel/smp.c
+++ b/arch/mips/kernel/smp.c
@@ -248,7 +248,7 @@ int mips_smp_ipi_allocate(const struct cpumask *mask)
* setup, if we're running with only a single CPU.
*/
if (!ipidomain) {
- BUG_ON(num_present_cpus() > 1);
+ BUG_ON(num_present_cpus_gt(1));
return 0;
}

@@ -314,7 +314,7 @@ int mips_smp_ipi_free(const struct cpumask *mask)

static int __init mips_smp_ipi_init(void)
{
- if (num_possible_cpus() == 1)
+ if (num_possible_cpus_eq(1))
return 0;

mips_smp_ipi_allocate(cpu_possible_mask);
diff --git a/arch/mips/kernel/vpe-cmp.c b/arch/mips/kernel/vpe-cmp.c
index e673603e11e5..c1dc00cda666 100644
--- a/arch/mips/kernel/vpe-cmp.c
+++ b/arch/mips/kernel/vpe-cmp.c
@@ -98,7 +98,7 @@ int __init vpe_module_init(void)
return -ENODEV;
}

- if (num_possible_cpus() - aprp_cpu_index() < 1) {
+ if (num_possible_cpus_le(aprp_cpu_index() + 1)) {
pr_warn("No VPEs reserved for AP/SP, not initialize VPE loader\n"
"Pass maxcpus=<n> argument as kernel argument\n");
return -ENODEV;
diff --git a/arch/mips/loongson2ef/common/cs5536/cs5536_mfgpt.c b/arch/mips/loongson2ef/common/cs5536/cs5536_mfgpt.c
index f21a540a1dd2..37166fa866c4 100644
--- a/arch/mips/loongson2ef/common/cs5536/cs5536_mfgpt.c
+++ b/arch/mips/loongson2ef/common/cs5536/cs5536_mfgpt.c
@@ -194,7 +194,7 @@ static struct clocksource clocksource_mfgpt = {

int __init init_mfgpt_clocksource(void)
{
- if (num_possible_cpus() > 1) /* MFGPT does not scale! */
+ if (num_possible_cpus_gt(1)) /* MFGPT does not scale! */
return 0;

return clocksource_register_hz(&clocksource_mfgpt, MFGPT_TICK_RATE);
diff --git a/arch/mips/mm/context.c b/arch/mips/mm/context.c
index b25564090939..bf508e38d30a 100644
--- a/arch/mips/mm/context.c
+++ b/arch/mips/mm/context.c
@@ -274,7 +274,7 @@ static int mmid_init(void)
* one more MMID than CPUs.
*/
num_mmids = asid_first_version(0);
- WARN_ON(num_mmids <= num_possible_cpus());
+ WARN_ON(num_possible_cpus_gt(num_mmids - 1));

atomic64_set(&mmid_version, asid_first_version(0));
mmid_map = kcalloc(BITS_TO_LONGS(num_mmids), sizeof(*mmid_map),
diff --git a/arch/mips/mm/tlbex.c b/arch/mips/mm/tlbex.c
index bede66b072a7..92dae5cfa0a4 100644
--- a/arch/mips/mm/tlbex.c
+++ b/arch/mips/mm/tlbex.c
@@ -363,7 +363,7 @@ static struct work_registers build_get_work_registers(u32 **p)
return r;
}

- if (num_possible_cpus() > 1) {
+ if (num_possible_cpus_gt(1)) {
/* Get smp_processor_id */
UASM_i_CPUID_MFC0(p, K0, SMP_CPUID_REG);
UASM_i_SRL_SAFE(p, K0, K0, SMP_CPUID_REGSHIFT);
diff --git a/arch/nios2/kernel/cpuinfo.c b/arch/nios2/kernel/cpuinfo.c
index 203870c4b86d..7bdc511eba60 100644
--- a/arch/nios2/kernel/cpuinfo.c
+++ b/arch/nios2/kernel/cpuinfo.c
@@ -172,7 +172,7 @@ static void *cpuinfo_start(struct seq_file *m, loff_t *pos)
{
unsigned long i = *pos;

- return i < num_possible_cpus() ? (void *) (i + 1) : NULL;
+ return num_possible_cpus_gt(i) ? (void *) (i + 1) : NULL;
}

static void *cpuinfo_next(struct seq_file *m, void *v, loff_t *pos)
diff --git a/arch/powerpc/platforms/85xx/smp.c b/arch/powerpc/platforms/85xx/smp.c
index 83f4a6389a28..15573310fab4 100644
--- a/arch/powerpc/platforms/85xx/smp.c
+++ b/arch/powerpc/platforms/85xx/smp.c
@@ -280,7 +280,7 @@ static int smp_85xx_kick_cpu(int nr)
int primary = nr;
#endif

- WARN_ON(nr < 0 || nr >= num_possible_cpus());
+ WARN_ON(nr < 0 || num_possible_cpus_le(nr + 1));

pr_debug("kick CPU #%d\n", nr);

diff --git a/arch/powerpc/platforms/pseries/hotplug-cpu.c b/arch/powerpc/platforms/pseries/hotplug-cpu.c
index 5ab44600c8d3..b0d66de92309 100644
--- a/arch/powerpc/platforms/pseries/hotplug-cpu.c
+++ b/arch/powerpc/platforms/pseries/hotplug-cpu.c
@@ -365,7 +365,7 @@ static int dlpar_offline_cpu(struct device_node *dn)
cpu_maps_update_begin();
break;
}
- if (cpu == num_possible_cpus()) {
+ if (num_possible_cpus_eq(cpu)) {
pr_warn("Could not find cpu to offline with physical id 0x%x\n",
thread);
}
@@ -408,7 +408,7 @@ static int dlpar_online_cpu(struct device_node *dn)

break;
}
- if (cpu == num_possible_cpus())
+ if (num_possible_cpus_eq(cpu))
printk(KERN_WARNING "Could not find cpu to online "
"with physical id 0x%x\n", thread);
}
diff --git a/arch/powerpc/sysdev/mpic.c b/arch/powerpc/sysdev/mpic.c
index 995fb2ada507..ded5007f2af9 100644
--- a/arch/powerpc/sysdev/mpic.c
+++ b/arch/powerpc/sysdev/mpic.c
@@ -1440,7 +1440,7 @@ struct mpic * __init mpic_alloc(struct device_node *node,
* The MPIC driver will crash if there are more cores than we
* can initialize, so we may as well catch that problem here.
*/
- BUG_ON(num_possible_cpus() > MPIC_MAX_CPUS);
+ BUG_ON(num_possible_cpus_gt(MPIC_MAX_CPUS));

/* Map the per-CPU registers */
for_each_possible_cpu(i) {
diff --git a/arch/powerpc/xmon/xmon.c b/arch/powerpc/xmon/xmon.c
index 2073be312fe9..938346f9af7d 100644
--- a/arch/powerpc/xmon/xmon.c
+++ b/arch/powerpc/xmon/xmon.c
@@ -2747,7 +2747,7 @@ static void dump_all_pacas(void)
{
int cpu;

- if (num_possible_cpus() == 0) {
+ if (num_possible_cpus_eq(0)) {
printf("No possible cpus, use 'dp #' to dump individual cpus\n");
return;
}
@@ -2809,7 +2809,7 @@ static void dump_all_xives(void)
{
int cpu;

- if (num_possible_cpus() == 0) {
+ if (num_possible_cpus_eq(0)) {
printf("No possible cpus, use 'dx #' to dump individual cpus\n");
return;
}
@@ -3692,7 +3692,7 @@ symbol_lookup(void)
ptr >= (void __percpu *)__per_cpu_start &&
ptr < (void __percpu *)__per_cpu_end)
{
- if (scanhex(&cpu) && cpu < num_possible_cpus()) {
+ if (scanhex(&cpu) && num_possible_cpus_gt(cpu)) {
addr = (unsigned long)per_cpu_ptr(ptr, cpu);
} else {
cpu = raw_smp_processor_id();
diff --git a/arch/riscv/kvm/vmid.c b/arch/riscv/kvm/vmid.c
index 2c6253b293bc..6e176baedf65 100644
--- a/arch/riscv/kvm/vmid.c
+++ b/arch/riscv/kvm/vmid.c
@@ -36,7 +36,7 @@ void kvm_riscv_stage2_vmid_detect(void)
__kvm_riscv_hfence_gvma_all();

/* We don't use VMID bits if they are not sufficient */
- if ((1UL << vmid_bits) < num_possible_cpus())
+ if (num_possible_cpus_gt(1UL << vmid_bits))
vmid_bits = 0;
}

diff --git a/arch/sparc/kernel/mdesc.c b/arch/sparc/kernel/mdesc.c
index 30f171b7b00c..b779c6607ff3 100644
--- a/arch/sparc/kernel/mdesc.c
+++ b/arch/sparc/kernel/mdesc.c
@@ -885,7 +885,7 @@ static void __mark_core_id(struct mdesc_handle *hp, u64 node,
{
const u64 *id = mdesc_get_property(hp, node, "id", NULL);

- if (*id < num_possible_cpus())
+ if (num_possible_cpus_gt(*id))
cpu_data(*id).core_id = core_id;
}

@@ -894,7 +894,7 @@ static void __mark_max_cache_id(struct mdesc_handle *hp, u64 node,
{
const u64 *id = mdesc_get_property(hp, node, "id", NULL);

- if (*id < num_possible_cpus()) {
+ if (num_possible_cpus_gt(*id)) {
cpu_data(*id).max_cache_id = max_cache_id;

/**
@@ -986,7 +986,7 @@ static void set_sock_ids_by_socket(struct mdesc_handle *hp, u64 mp)
continue;

id = mdesc_get_property(hp, t, "id", NULL);
- if (*id < num_possible_cpus())
+ if (num_possible_cpus_gt(*id))
cpu_data(*id).sock_id = idx;
}
idx++;
diff --git a/arch/x86/events/amd/core.c b/arch/x86/events/amd/core.c
index 9687a8aef01c..d69ed09a85b0 100644
--- a/arch/x86/events/amd/core.c
+++ b/arch/x86/events/amd/core.c
@@ -1007,7 +1007,7 @@ __init int amd_pmu_init(void)
if (ret)
return ret;

- if (num_possible_cpus() == 1) {
+ if (num_possible_cpus_eq(1)) {
/*
* No point in allocating data structures to serialize
* against other CPUs, when there is only the one CPU.
diff --git a/arch/x86/kernel/alternative.c b/arch/x86/kernel/alternative.c
index 23fb4d51a5da..55fd70fdb213 100644
--- a/arch/x86/kernel/alternative.c
+++ b/arch/x86/kernel/alternative.c
@@ -574,7 +574,7 @@ void __init_or_module alternatives_smp_module_add(struct module *mod,
if (!uniproc_patched)
goto unlock;

- if (num_possible_cpus() == 1)
+ if (num_possible_cpus_eq(1))
/* Don't bother remembering, we'll never have to undo it. */
goto smp_unlock;

@@ -620,7 +620,7 @@ void alternatives_enable_smp(void)
struct smp_alt_module *mod;

/* Why bother if there are no other CPUs? */
- BUG_ON(num_possible_cpus() == 1);
+ BUG_ON(num_possible_cpus_eq(1));

mutex_lock(&text_mutex);

@@ -833,14 +833,14 @@ void __init alternative_instructions(void)

#ifdef CONFIG_SMP
/* Patch to UP if other cpus not imminent. */
- if (!noreplace_smp && (num_present_cpus() == 1 || setup_max_cpus <= 1)) {
+ if (!noreplace_smp && (num_present_cpus_eq(1) || setup_max_cpus <= 1)) {
uniproc_patched = true;
alternatives_smp_module_add(NULL, "core kernel",
__smp_locks, __smp_locks_end,
_text, _etext);
}

- if (!uniproc_patched || num_possible_cpus() == 1) {
+ if (!uniproc_patched || num_possible_cpus_eq(1)) {
free_init_pages("SMP alternatives",
(unsigned long)__smp_locks,
(unsigned long)__smp_locks_end);
diff --git a/arch/x86/kernel/apic/apic.c b/arch/x86/kernel/apic/apic.c
index b70344bf6600..9a3d0748ca86 100644
--- a/arch/x86/kernel/apic/apic.c
+++ b/arch/x86/kernel/apic/apic.c
@@ -1020,7 +1020,7 @@ void __init setup_boot_APIC_clock(void)
if (disable_apic_timer) {
pr_info("Disabling APIC timer\n");
/* No broadcast on UP ! */
- if (num_possible_cpus() > 1) {
+ if (num_possible_cpus_gt(1)) {
lapic_clockevent.mult = 1;
setup_APIC_timer();
}
@@ -1029,7 +1029,7 @@ void __init setup_boot_APIC_clock(void)

if (calibrate_APIC_clock()) {
/* No broadcast on UP ! */
- if (num_possible_cpus() > 1)
+ if (num_possible_cpus_gt(1))
setup_APIC_timer();
return;
}
diff --git a/arch/x86/kernel/apic/apic_flat_64.c b/arch/x86/kernel/apic/apic_flat_64.c
index 8f72b4351c9f..3dfd4c5d30dc 100644
--- a/arch/x86/kernel/apic/apic_flat_64.c
+++ b/arch/x86/kernel/apic/apic_flat_64.c
@@ -189,7 +189,7 @@ static void physflat_init_apic_ldr(void)

static int physflat_probe(void)
{
- if (apic == &apic_physflat || num_possible_cpus() > 8 ||
+ if (apic == &apic_physflat || num_possible_cpus_gt(8) ||
jailhouse_paravirt())
return 1;

diff --git a/arch/x86/kernel/apic/probe_32.c b/arch/x86/kernel/apic/probe_32.c
index a61f642b1b90..b65c1572aaf5 100644
--- a/arch/x86/kernel/apic/probe_32.c
+++ b/arch/x86/kernel/apic/probe_32.c
@@ -138,7 +138,7 @@ void __init default_setup_apic_routing(void)
{
int version = boot_cpu_apic_version;

- if (num_possible_cpus() > 8) {
+ if (num_possible_cpus_gt(8)) {
switch (boot_cpu_data.x86_vendor) {
case X86_VENDOR_INTEL:
if (!APIC_XAPIC(version)) {
diff --git a/arch/x86/kernel/cpu/mce/dev-mcelog.c b/arch/x86/kernel/cpu/mce/dev-mcelog.c
index 100fbeebdc72..34e44b0d9546 100644
--- a/arch/x86/kernel/cpu/mce/dev-mcelog.c
+++ b/arch/x86/kernel/cpu/mce/dev-mcelog.c
@@ -310,7 +310,7 @@ static ssize_t mce_chrdev_write(struct file *filp, const char __user *ubuf,
if (copy_from_user(&m, ubuf, usize))
return -EFAULT;

- if (m.extcpu >= num_possible_cpus() || !cpu_online(m.extcpu))
+ if (num_possible_cpus_le(m.extcpu + 1) || !cpu_online(m.extcpu))
return -EINVAL;

/*
diff --git a/arch/x86/kernel/hpet.c b/arch/x86/kernel/hpet.c
index 882213df3713..e432e6248599 100644
--- a/arch/x86/kernel/hpet.c
+++ b/arch/x86/kernel/hpet.c
@@ -737,7 +737,7 @@ static void __init hpet_select_clockevents(void)
hc->irq = irq;
hc->mode = HPET_MODE_CLOCKEVT;

- if (++hpet_base.nr_clockevents == num_possible_cpus())
+ if (num_possible_cpus_eq(++hpet_base.nr_clockevents))
break;
}

diff --git a/arch/x86/kernel/i8253.c b/arch/x86/kernel/i8253.c
index 2b7999a1a50a..e6e30a7bc80f 100644
--- a/arch/x86/kernel/i8253.c
+++ b/arch/x86/kernel/i8253.c
@@ -57,7 +57,7 @@ static int __init init_pit_clocksource(void)
* - when HPET is enabled
* - when local APIC timer is active (PIT is switched off)
*/
- if (num_possible_cpus() > 1 || is_hpet_enabled() ||
+ if (num_possible_cpus_gt(1) || is_hpet_enabled() ||
!clockevent_state_periodic(&i8253_clockevent))
return 0;

diff --git a/arch/x86/kernel/kvm.c b/arch/x86/kernel/kvm.c
index 59abbdad7729..375226dcf29e 100644
--- a/arch/x86/kernel/kvm.c
+++ b/arch/x86/kernel/kvm.c
@@ -1057,7 +1057,7 @@ void __init kvm_spinlock_init(void)
goto out;
}

- if (num_possible_cpus() == 1) {
+ if (num_possible_cpus_eq(1)) {
pr_info("PV spinlocks disabled, single CPU\n");
goto out;
}
diff --git a/arch/x86/kernel/kvmclock.c b/arch/x86/kernel/kvmclock.c
index 462dd8e9b03d..12c1fb1dfd07 100644
--- a/arch/x86/kernel/kvmclock.c
+++ b/arch/x86/kernel/kvmclock.c
@@ -205,7 +205,7 @@ static void __init kvmclock_init_mem(void)
struct page *p;
int r;

- if (HVC_BOOT_ARRAY_SIZE >= num_possible_cpus())
+ if (num_possible_cpus_le(HVC_BOOT_ARRAY_SIZE + 1))
return;

ncpus = num_possible_cpus() - HVC_BOOT_ARRAY_SIZE;
diff --git a/arch/x86/kernel/tsc.c b/arch/x86/kernel/tsc.c
index 2e076a459a0c..2245c9721d4a 100644
--- a/arch/x86/kernel/tsc.c
+++ b/arch/x86/kernel/tsc.c
@@ -1223,7 +1223,7 @@ int unsynchronized_tsc(void)
*/
if (boot_cpu_data.x86_vendor != X86_VENDOR_INTEL) {
/* assume multi socket systems are not synchronized: */
- if (num_possible_cpus() > 1)
+ if (num_possible_cpus_gt(1))
return 1;
}

diff --git a/arch/x86/xen/smp_pv.c b/arch/x86/xen/smp_pv.c
index 6a8f3b53ab83..b32ca28292ae 100644
--- a/arch/x86/xen/smp_pv.c
+++ b/arch/x86/xen/smp_pv.c
@@ -254,7 +254,7 @@ static void __init xen_pv_smp_prepare_cpus(unsigned int max_cpus)
cpumask_copy(xen_cpu_initialized_map, cpumask_of(0));

/* Restrict the possible_map according to max_cpus. */
- while ((num_possible_cpus() > 1) && (num_possible_cpus() > max_cpus)) {
+ while (num_possible_cpus_gt(max(1, max_cpus))) {
for (cpu = nr_cpu_ids - 1; !cpu_possible(cpu); cpu--)
continue;
set_cpu_possible(cpu, false);
diff --git a/arch/x86/xen/spinlock.c b/arch/x86/xen/spinlock.c
index 043c73dfd2c9..58caaa9aec3e 100644
--- a/arch/x86/xen/spinlock.c
+++ b/arch/x86/xen/spinlock.c
@@ -125,7 +125,7 @@ PV_CALLEE_SAVE_REGS_THUNK(xen_vcpu_stolen);
void __init xen_init_spinlocks(void)
{
/* Don't need to use pvqspinlock code if there is only 1 vCPU. */
- if (num_possible_cpus() == 1 || nopvspin)
+ if (num_possible_cpus_eq(1) || nopvspin)
xen_pvspin = false;

if (!xen_pvspin) {
diff --git a/drivers/clk/samsung/clk-exynos4.c b/drivers/clk/samsung/clk-exynos4.c
index 22009cb53428..64d7de6b885c 100644
--- a/drivers/clk/samsung/clk-exynos4.c
+++ b/drivers/clk/samsung/clk-exynos4.c
@@ -1178,7 +1178,7 @@ static void __init exynos4x12_core_down_clock(void)
PWR_CTRL1_USE_CORE1_WFE | PWR_CTRL1_USE_CORE0_WFE |
PWR_CTRL1_USE_CORE1_WFI | PWR_CTRL1_USE_CORE0_WFI);
/* On Exynos4412 enable it also on core 2 and 3 */
- if (num_possible_cpus() == 4)
+ if (num_possible_cpus_eq(4))
tmp |= PWR_CTRL1_USE_CORE3_WFE | PWR_CTRL1_USE_CORE2_WFE |
PWR_CTRL1_USE_CORE3_WFI | PWR_CTRL1_USE_CORE2_WFI;
writel_relaxed(tmp, reg_base + PWR_CTRL1);
diff --git a/drivers/clocksource/ingenic-timer.c b/drivers/clocksource/ingenic-timer.c
index 24ed0f1f089b..c4a34d26357c 100644
--- a/drivers/clocksource/ingenic-timer.c
+++ b/drivers/clocksource/ingenic-timer.c
@@ -302,8 +302,7 @@ static int __init ingenic_tcu_init(struct device_node *np)
(u32 *)&tcu->pwm_channels_mask);

/* Verify that we have at least num_possible_cpus() + 1 free channels */
- if (hweight8(tcu->pwm_channels_mask) >
- soc_info->num_channels - num_possible_cpus() + 1) {
+ if (num_possible_cpus_gt(soc_info->num_channels + 1 - hweight8(tcu->pwm_channels_mask))) {
pr_crit("%s: Invalid PWM channel mask: 0x%02lx\n", __func__,
tcu->pwm_channels_mask);
ret = -EINVAL;
diff --git a/drivers/cpufreq/pcc-cpufreq.c b/drivers/cpufreq/pcc-cpufreq.c
index 9f3fc7a073d0..8bf76eaa9e1e 100644
--- a/drivers/cpufreq/pcc-cpufreq.c
+++ b/drivers/cpufreq/pcc-cpufreq.c
@@ -593,7 +593,7 @@ static int __init pcc_cpufreq_init(void)
return ret;
}

- if (num_present_cpus() > 4) {
+ if (num_present_cpus_gt(4)) {
pcc_cpufreq_driver.flags |= CPUFREQ_NO_AUTO_DYNAMIC_SWITCHING;
pr_err("%s: Too many CPUs, dynamic performance scaling disabled\n",
__func__);
diff --git a/drivers/dma/mv_xor.c b/drivers/dma/mv_xor.c
index 23b232b57518..f99177e72158 100644
--- a/drivers/dma/mv_xor.c
+++ b/drivers/dma/mv_xor.c
@@ -1293,7 +1293,7 @@ static int mv_xor_probe(struct platform_device *pdev)
struct mv_xor_device *xordev;
struct mv_xor_platform_data *pdata = dev_get_platdata(&pdev->dev);
struct resource *res;
- unsigned int max_engines, max_channels;
+ unsigned int max_channels;
int i, ret;

dev_notice(&pdev->dev, "Marvell shared XOR driver\n");
@@ -1362,7 +1362,6 @@ static int mv_xor_probe(struct platform_device *pdev)
* separate engines when possible. For dual-CPU Armada 3700
* SoC with single XOR engine allow using its both channels.
*/
- max_engines = num_present_cpus();
if (xordev->xor_type == XOR_ARMADA_37XX)
max_channels = num_present_cpus();
else
@@ -1370,7 +1369,7 @@ static int mv_xor_probe(struct platform_device *pdev)
MV_XOR_MAX_CHANNELS,
DIV_ROUND_UP(num_present_cpus(), 2));

- if (mv_xor_engine_count >= max_engines)
+ if (num_present_cpus_le(mv_xor_engine_count + 1))
return 0;

if (pdev->dev.of_node) {
diff --git a/drivers/gpu/drm/i810/i810_drv.c b/drivers/gpu/drm/i810/i810_drv.c
index 0e53a066d4db..c70745fa4166 100644
--- a/drivers/gpu/drm/i810/i810_drv.c
+++ b/drivers/gpu/drm/i810/i810_drv.c
@@ -80,7 +80,7 @@ static struct pci_driver i810_pci_driver = {

static int __init i810_init(void)
{
- if (num_possible_cpus() > 1) {
+ if (num_possible_cpus_gt(1)) {
pr_err("drm/i810 does not support SMP\n");
return -EINVAL;
}
diff --git a/drivers/irqchip/irq-gic.c b/drivers/irqchip/irq-gic.c
index b8bb46c65a97..4e319e4ba9dc 100644
--- a/drivers/irqchip/irq-gic.c
+++ b/drivers/irqchip/irq-gic.c
@@ -430,7 +430,7 @@ static u8 gic_get_cpumask(struct gic_chip_data *gic)
break;
}

- if (!mask && num_possible_cpus() > 1)
+ if (!mask && num_possible_cpus_gt(1))
pr_crit("GIC CPU mask not found - kernel will fail to boot.\n");

return mask;
diff --git a/drivers/net/caif/caif_virtio.c b/drivers/net/caif/caif_virtio.c
index 91230894692d..c7aa3f6dc635 100644
--- a/drivers/net/caif/caif_virtio.c
+++ b/drivers/net/caif/caif_virtio.c
@@ -537,7 +537,7 @@ static netdev_tx_t cfv_netdev_tx(struct sk_buff *skb, struct net_device *netdev)
*
* Flow-on is triggered when sufficient buffers are freed
*/
- if (unlikely(cfv->vq_tx->num_free <= num_present_cpus())) {
+ if (unlikely(num_present_cpus_gt(cfv->vq_tx->num_free - 1))) {
flow_off = true;
cfv->stats.tx_full_ring++;
}
diff --git a/drivers/net/ethernet/cavium/liquidio/cn23xx_vf_device.c b/drivers/net/ethernet/cavium/liquidio/cn23xx_vf_device.c
index fda49404968c..79d5ded30b65 100644
--- a/drivers/net/ethernet/cavium/liquidio/cn23xx_vf_device.c
+++ b/drivers/net/ethernet/cavium/liquidio/cn23xx_vf_device.c
@@ -649,7 +649,7 @@ int cn23xx_setup_octeon_vf_device(struct octeon_device *oct)
rings_per_vf);
oct->sriov_info.rings_per_vf = rings_per_vf;
} else {
- if (rings_per_vf > num_present_cpus()) {
+ if (num_present_cpus_le(rings_per_vf)) {
dev_warn(&oct->pci_dev->dev,
"PF configured rings_per_vf:%d greater than num_cpu:%d. Using rings_per_vf:%d equal to num cpus\n",
rings_per_vf,
diff --git a/drivers/net/ethernet/hisilicon/hns/hns_enet.c b/drivers/net/ethernet/hisilicon/hns/hns_enet.c
index 22a463e15678..7d97939413d2 100644
--- a/drivers/net/ethernet/hisilicon/hns/hns_enet.c
+++ b/drivers/net/ethernet/hisilicon/hns/hns_enet.c
@@ -1239,7 +1239,7 @@ static int hns_nic_init_affinity_mask(int q_num, int ring_idx,
* The cpu mask set by ring index according to the ring flag
* which indicate the ring is tx or rx.
*/
- if (q_num == num_possible_cpus()) {
+ if (num_possible_cpus_eq(q_num)) {
if (is_tx_ring(ring))
cpu = ring_idx;
else
diff --git a/drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c b/drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
index a48e804c46f2..34ad59fd51d6 100644
--- a/drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
+++ b/drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
@@ -3315,7 +3315,7 @@ static int mvpp2_setup_txqs(struct mvpp2_port *port)
goto err_cleanup;

/* Assign this queue to a CPU */
- if (queue < num_possible_cpus())
+ if (num_possible_cpus_gt(queue))
netif_set_xps_queue(port->dev, cpumask_of(queue), queue);
}

diff --git a/drivers/net/wireless/ath/ath9k/hw.c b/drivers/net/wireless/ath/ath9k/hw.c
index 172081ffe477..33d3cddc6c7b 100644
--- a/drivers/net/wireless/ath/ath9k/hw.c
+++ b/drivers/net/wireless/ath/ath9k/hw.c
@@ -429,7 +429,7 @@ static void ath9k_hw_init_config(struct ath_hw *ah)
* This issue is not present on PCI-Express devices or pre-AR5416
* devices (legacy, 802.11abg).
*/
- if (num_possible_cpus() > 1)
+ if (num_possible_cpus_gt(1))
ah->config.serialize_regmode = SER_REG_MODE_AUTO;

if (NR_CPUS > 1 && ah->config.serialize_regmode == SER_REG_MODE_AUTO) {
diff --git a/drivers/net/wireless/marvell/mwifiex/main.c b/drivers/net/wireless/marvell/mwifiex/main.c
index 19b996c6a260..6ce0236a3203 100644
--- a/drivers/net/wireless/marvell/mwifiex/main.c
+++ b/drivers/net/wireless/marvell/mwifiex/main.c
@@ -1536,7 +1536,7 @@ mwifiex_reinit_sw(struct mwifiex_adapter *adapter)
adapter->cmd_wait_q.status = 0;
adapter->scan_wait_q_woken = false;

- if ((num_possible_cpus() > 1) || adapter->iface_type == MWIFIEX_USB)
+ if (num_possible_cpus_gt(1) || adapter->iface_type == MWIFIEX_USB)
adapter->rx_work_enabled = true;

adapter->workqueue =
@@ -1691,7 +1691,7 @@ mwifiex_add_card(void *card, struct completion *fw_done,
adapter->cmd_wait_q.status = 0;
adapter->scan_wait_q_woken = false;

- if ((num_possible_cpus() > 1) || adapter->iface_type == MWIFIEX_USB)
+ if (num_possible_cpus_gt(1) || adapter->iface_type == MWIFIEX_USB)
adapter->rx_work_enabled = true;

adapter->workqueue =
diff --git a/drivers/net/wireless/st/cw1200/queue.c b/drivers/net/wireless/st/cw1200/queue.c
index 12952b1c29df..4d47a1e26d55 100644
--- a/drivers/net/wireless/st/cw1200/queue.c
+++ b/drivers/net/wireless/st/cw1200/queue.c
@@ -312,8 +312,7 @@ int cw1200_queue_put(struct cw1200_queue *queue,
* Leave extra queue slots so we don't overflow.
*/
if (queue->overfull == false &&
- queue->num_queued >=
- (queue->capacity - (num_present_cpus() - 1))) {
+ num_present_cpus_gt(queue->capacity - queue->num_queued)) {
queue->overfull = true;
__cw1200_queue_lock(queue);
mod_timer(&queue->gc, jiffies);
diff --git a/drivers/nvdimm/region.c b/drivers/nvdimm/region.c
index e0c34120df37..474f1ed5d9b9 100644
--- a/drivers/nvdimm/region.c
+++ b/drivers/nvdimm/region.c
@@ -17,7 +17,7 @@ static int nd_region_probe(struct device *dev)
struct nd_region *nd_region = to_nd_region(dev);

if (nd_region->num_lanes > num_online_cpus()
- && nd_region->num_lanes < num_possible_cpus()
+ && num_possible_cpus_gt(nd_region->num_lanes)
&& !test_and_set_bit(0, &once)) {
dev_dbg(dev, "online cpus (%d) < concurrent i/o lanes (%d) < possible cpus (%d)\n",
num_online_cpus(), nd_region->num_lanes,
diff --git a/drivers/nvme/host/pci.c b/drivers/nvme/host/pci.c
index ca2ee806d74b..34958f775ad8 100644
--- a/drivers/nvme/host/pci.c
+++ b/drivers/nvme/host/pci.c
@@ -79,7 +79,7 @@ static int io_queue_count_set(const char *val, const struct kernel_param *kp)
int ret;

ret = kstrtouint(val, 10, &n);
- if (ret != 0 || n > num_possible_cpus())
+ if (ret != 0 || num_possible_cpus_le(n))
return -EINVAL;
return param_set_uint(val, kp);
}
diff --git a/drivers/perf/arm_pmu.c b/drivers/perf/arm_pmu.c
index a31b302b0ade..5f43a7bde55d 100644
--- a/drivers/perf/arm_pmu.c
+++ b/drivers/perf/arm_pmu.c
@@ -637,7 +637,7 @@ int armpmu_request_irq(int irq, int cpu)

err = irq_force_affinity(irq, cpumask_of(cpu));

- if (err && num_possible_cpus() > 1) {
+ if (err && num_possible_cpus_gt(1)) {
pr_warn("unable to set irq affinity (irq=%d, cpu=%u)\n",
irq, cpu);
goto err_out;
diff --git a/drivers/platform/x86/intel/speed_select_if/isst_if_common.c b/drivers/platform/x86/intel/speed_select_if/isst_if_common.c
index c9a85eb2e860..c25902969475 100644
--- a/drivers/platform/x86/intel/speed_select_if/isst_if_common.c
+++ b/drivers/platform/x86/intel/speed_select_if/isst_if_common.c
@@ -297,7 +297,7 @@ static struct pci_dev *_isst_if_get_pci_dev(int cpu, int bus_no, int dev, int fn
int i, bus_number;

if (bus_no < 0 || bus_no > 1 || cpu < 0 || cpu >= nr_cpu_ids ||
- cpu >= num_possible_cpus())
+ num_possible_cpus_le(cpu + 1))
return NULL;

bus_number = isst_cpu_info[cpu].bus_info[bus_no];
@@ -362,7 +362,7 @@ struct pci_dev *isst_if_get_pci_dev(int cpu, int bus_no, int dev, int fn)
struct pci_dev *pci_dev;

if (bus_no < 0 || bus_no > 1 || cpu < 0 || cpu >= nr_cpu_ids ||
- cpu >= num_possible_cpus())
+ num_possible_cpus_le(cpu + 1))
return NULL;

pci_dev = isst_cpu_info[cpu].pci_dev[bus_no];
@@ -442,7 +442,7 @@ static long isst_if_proc_phyid_req(u8 *cmd_ptr, int *write_only, int resume)

cpu_map = (struct isst_if_cpu_map *)cmd_ptr;
if (cpu_map->logical_cpu >= nr_cpu_ids ||
- cpu_map->logical_cpu >= num_possible_cpus())
+ num_possible_cpus_le(cpu_map->logical_cpu + 1))
return -EINVAL;

*write_only = 0;
diff --git a/drivers/soc/bcm/brcmstb/biuctrl.c b/drivers/soc/bcm/brcmstb/biuctrl.c
index 2c975d79fe8e..6a75cbe836a4 100644
--- a/drivers/soc/bcm/brcmstb/biuctrl.c
+++ b/drivers/soc/bcm/brcmstb/biuctrl.c
@@ -181,7 +181,7 @@ static void __init a72_b53_rac_enable_all(struct device_node *np)
if (IS_ENABLED(CONFIG_CACHE_B15_RAC))
return;

- if (WARN(num_possible_cpus() > 4, "RAC only supports 4 CPUs\n"))
+ if (WARN(num_possible_cpus_gt(4), "RAC only supports 4 CPUs\n"))
return;

pref_dist = cbc_readl(RAC_CONFIG1_REG);
diff --git a/drivers/soc/fsl/dpio/dpio-service.c b/drivers/soc/fsl/dpio/dpio-service.c
index 1d2b27e3ea63..b38c519f2294 100644
--- a/drivers/soc/fsl/dpio/dpio-service.c
+++ b/drivers/soc/fsl/dpio/dpio-service.c
@@ -60,7 +60,7 @@ static inline struct dpaa2_io *service_select_by_cpu(struct dpaa2_io *d,
if (d)
return d;

- if (cpu != DPAA2_IO_ANY_CPU && cpu >= num_possible_cpus())
+ if (cpu != DPAA2_IO_ANY_CPU && num_possible_cpus_le(cpu + 1))
return NULL;

/*
@@ -140,7 +140,7 @@ struct dpaa2_io *dpaa2_io_create(const struct dpaa2_io_desc *desc,
return NULL;

/* check if CPU is out of range (-1 means any cpu) */
- if (desc->cpu != DPAA2_IO_ANY_CPU && desc->cpu >= num_possible_cpus()) {
+ if (desc->cpu != DPAA2_IO_ANY_CPU && num_possible_cpus_le(desc->cpu + 1)) {
kfree(obj);
return NULL;
}
diff --git a/drivers/spi/spi-dw-bt1.c b/drivers/spi/spi-dw-bt1.c
index c06553416123..ab6b6a32a0d6 100644
--- a/drivers/spi/spi-dw-bt1.c
+++ b/drivers/spi/spi-dw-bt1.c
@@ -241,7 +241,7 @@ static int dw_spi_bt1_sys_init(struct platform_device *pdev,
* though, but still tends to be not fast enough at low CPU
* frequencies.
*/
- if (num_possible_cpus() > 1)
+ if (num_possible_cpus_gt(1))
dws->max_mem_freq = 10000000U;
else
dws->max_mem_freq = 20000000U;
diff --git a/drivers/virt/acrn/hsm.c b/drivers/virt/acrn/hsm.c
index 5419794fccf1..50cd69012dcf 100644
--- a/drivers/virt/acrn/hsm.c
+++ b/drivers/virt/acrn/hsm.c
@@ -431,7 +431,7 @@ static ssize_t remove_cpu_store(struct device *dev,
if (kstrtoull(buf, 0, &cpu) < 0)
return -EINVAL;

- if (cpu >= num_possible_cpus() || cpu == 0 || !cpu_is_hotpluggable(cpu))
+ if (num_possible_cpus_le(cpu + 1) || cpu == 0 || !cpu_is_hotpluggable(cpu))
return -EINVAL;

if (cpu_online(cpu))
diff --git a/fs/xfs/xfs_sysfs.c b/fs/xfs/xfs_sysfs.c
index 8608f804388f..5580d60ec962 100644
--- a/fs/xfs/xfs_sysfs.c
+++ b/fs/xfs/xfs_sysfs.c
@@ -211,7 +211,7 @@ pwork_threads_store(
if (ret)
return ret;

- if (val < -1 || val > num_possible_cpus())
+ if (val < -1 || num_possible_cpus_le(val))
return -EINVAL;

xfs_globals.pwork_threads = val;
diff --git a/include/linux/cpumask.h b/include/linux/cpumask.h
index b5e50cf74785..ea0699fa4d4c 100644
--- a/include/linux/cpumask.h
+++ b/include/linux/cpumask.h
@@ -945,8 +945,19 @@ static inline unsigned int num_online_cpus(void)
return atomic_read(&__num_online_cpus);
}
#define num_possible_cpus() cpumask_weight(cpu_possible_mask)
+#define num_possible_cpus_eq(n) cpumask_weight_eq(cpu_possible_mask, (n))
+#define num_possible_cpus_gt(n) cpumask_weight_gt(cpu_possible_mask, (n))
+#define num_possible_cpus_le(n) cpumask_weight_le(cpu_possible_mask, (n))
+
#define num_present_cpus() cpumask_weight(cpu_present_mask)
+#define num_present_cpus_eq(n) cpumask_weight_eq(cpu_present_mask, (n))
+#define num_present_cpus_gt(n) cpumask_weight_gt(cpu_present_mask, (n))
+#define num_present_cpus_le(n) cpumask_weight_le(cpu_present_mask, (n))
+
#define num_active_cpus() cpumask_weight(cpu_active_mask)
+#define num_active_cpus_eq(n) cpumask_weight_eq(cpu_active_mask, (n))
+#define num_active_cpus_gt(n) cpumask_weight_gt(cpu_active_mask, (n))
+#define num_active_cpus_le(n) cpumask_weight_le(cpu_active_mask, (n))

static inline bool cpu_online(unsigned int cpu)
{
@@ -976,9 +987,21 @@ static inline bool cpu_dying(unsigned int cpu)
#else

#define num_online_cpus() 1U
+
#define num_possible_cpus() 1U
+#define num_possible_cpus_eq(n) (1U == (n))
+#define num_possible_cpus_gt(n) (1U > (n))
+#define num_possible_cpus_le(n) (1U < (n))
+
#define num_present_cpus() 1U
+#define num_present_cpus_eq(n) (1U == (n))
+#define num_present_cpus_gt(n) (1U > (n))
+#define num_present_cpus_le(n) (1U < (n))
+
#define num_active_cpus() 1U
+#define num_active_cpus_eq(n) (1U == (n))
+#define num_active_cpus_gt(n) (1U > (n))
+#define num_active_cpus_le(n) (1U < (n))

static inline bool cpu_online(unsigned int cpu)
{
diff --git a/include/linux/kdb.h b/include/linux/kdb.h
index ea0f5e580fac..48269d32b038 100644
--- a/include/linux/kdb.h
+++ b/include/linux/kdb.h
@@ -191,7 +191,7 @@ static inline
int kdb_process_cpu(const struct task_struct *p)
{
unsigned int cpu = task_cpu(p);
- if (cpu > num_possible_cpus())
+ if (num_possible_cpus_le(cpu))
cpu = 0;
return cpu;
}
diff --git a/kernel/debug/kdb/kdb_bt.c b/kernel/debug/kdb/kdb_bt.c
index 10b454554ab0..b6435a41a537 100644
--- a/kernel/debug/kdb/kdb_bt.c
+++ b/kernel/debug/kdb/kdb_bt.c
@@ -108,7 +108,7 @@ kdb_bt_cpu(unsigned long cpu)
{
struct task_struct *kdb_tsk;

- if (cpu >= num_possible_cpus() || !cpu_online(cpu)) {
+ if (num_possible_cpus_le(cpu + 1) || !cpu_online(cpu)) {
kdb_printf("WARNING: no process for cpu %ld\n", cpu);
return;
}
diff --git a/kernel/printk/printk.c b/kernel/printk/printk.c
index cbc35d586afb..08e6df52eb4d 100644
--- a/kernel/printk/printk.c
+++ b/kernel/printk/printk.c
@@ -1005,7 +1005,7 @@ static void __init log_buf_add_cpu(void)
* set_cpu_possible() after setup_arch() but just in
* case lets ensure this is valid.
*/
- if (num_possible_cpus() == 1)
+ if (num_possible_cpus_eq(1))
return;

cpu_extra = (num_possible_cpus() - 1) * __LOG_CPU_MAX_BUF_LEN;
diff --git a/kernel/reboot.c b/kernel/reboot.c
index 6bcc5d6a6572..f21c2c20505d 100644
--- a/kernel/reboot.c
+++ b/kernel/reboot.c
@@ -639,7 +639,7 @@ static int __init reboot_setup(char *str)
if (isdigit(str[0])) {
int cpu = simple_strtoul(str, NULL, 0);

- if (cpu >= num_possible_cpus()) {
+ if (num_possible_cpus_le(cpu + 1)) {
pr_err("Ignoring the CPU number in reboot= option. "
"CPU %d exceeds possible cpu number %d\n",
cpu, num_possible_cpus());
@@ -844,7 +844,7 @@ static ssize_t cpu_store(struct kobject *kobj, struct kobj_attribute *attr,
if (rc)
return rc;

- if (cpunum >= num_possible_cpus())
+ if (num_possible_cpus_le(cpunum + 1))
return -ERANGE;

reboot_default = 0;
diff --git a/kernel/time/clockevents.c b/kernel/time/clockevents.c
index 32d6629a55b2..c1fdfa4084c3 100644
--- a/kernel/time/clockevents.c
+++ b/kernel/time/clockevents.c
@@ -448,7 +448,7 @@ void clockevents_register_device(struct clock_event_device *dev)
clockevent_set_state(dev, CLOCK_EVT_STATE_DETACHED);

if (!dev->cpumask) {
- WARN_ON(num_possible_cpus() > 1);
+ WARN_ON(num_possible_cpus_gt(1));
dev->cpumask = cpumask_of(smp_processor_id());
}

diff --git a/mm/percpu.c b/mm/percpu.c
index 293009cc03ef..76e846b3d48e 100644
--- a/mm/percpu.c
+++ b/mm/percpu.c
@@ -2936,7 +2936,7 @@ static struct pcpu_alloc_info * __init __flatten pcpu_build_alloc_info(
* greater-than comparison ensures upa==1 always
* passes the following check.
*/
- if (wasted > num_possible_cpus() / 3)
+ if (num_possible_cpus_le(wasted * 3))
continue;

/* and then don't consume more memory */
@@ -3193,7 +3193,7 @@ int __init pcpu_page_first_chunk(size_t reserved_size,

/* allocate pages */
j = 0;
- for (unit = 0; unit < num_possible_cpus(); unit++) {
+ for (unit = 0; num_possible_cpus_gt(unit); unit++) {
unsigned int cpu = ai->groups[0].cpu_map[unit];
for (i = 0; i < unit_pages; i++) {
void *ptr;
@@ -3215,7 +3215,7 @@ int __init pcpu_page_first_chunk(size_t reserved_size,
vm.size = num_possible_cpus() * ai->unit_size;
vm_area_register_early(&vm, PAGE_SIZE);

- for (unit = 0; unit < num_possible_cpus(); unit++) {
+ for (unit = 0; num_possible_cpus_gt(unit); unit++) {
unsigned long unit_addr =
(unsigned long)vm.addr + unit * ai->unit_size;

diff --git a/mm/slab.c b/mm/slab.c
index ca4822f6b2b6..3cf2ee629b23 100644
--- a/mm/slab.c
+++ b/mm/slab.c
@@ -3929,7 +3929,7 @@ static int enable_cpucache(struct kmem_cache *cachep, gfp_t gfp)
* to a larger limit. Thus disabled by default.
*/
shared = 0;
- if (cachep->size <= PAGE_SIZE && num_possible_cpus() > 1)
+ if (cachep->size <= PAGE_SIZE && num_possible_cpus_gt(1))
shared = 8;

#if DEBUG
--
2.25.1