// SPDX-License-Identifier: GPL-2.0-only /* * * Copyright (C) 2013 Citrix Systems * * Author: Stefano Stabellini */ #define pr_fmt(fmt) "arm-pv: " fmt #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include struct static_key paravirt_steal_enabled; struct static_key paravirt_steal_rq_enabled; static u64 native_steal_clock(int cpu) { return 0; } DEFINE_STATIC_CALL(pv_steal_clock, native_steal_clock); struct pv_time_stolen_time_region { struct pvclock_vcpu_stolen_time __rcu *kaddr; }; static DEFINE_PER_CPU(struct pv_time_stolen_time_region, stolen_time_region); static bool steal_acc = true; static int __init parse_no_stealacc(char *arg) { steal_acc = false; return 0; } early_param("no-steal-acc", parse_no_stealacc); /* return stolen time in ns by asking the hypervisor */ static u64 para_steal_clock(int cpu) { struct pvclock_vcpu_stolen_time *kaddr = NULL; struct pv_time_stolen_time_region *reg; u64 ret = 0; reg = per_cpu_ptr(&stolen_time_region, cpu); /* * paravirt_steal_clock() may be called before the CPU * online notification callback runs. Until the callback * has run we just return zero. */ rcu_read_lock(); kaddr = rcu_dereference(reg->kaddr); if (!kaddr) { rcu_read_unlock(); return 0; } ret = le64_to_cpu(READ_ONCE(kaddr->stolen_time)); rcu_read_unlock(); return ret; } static int stolen_time_cpu_down_prepare(unsigned int cpu) { struct pvclock_vcpu_stolen_time *kaddr = NULL; struct pv_time_stolen_time_region *reg; reg = this_cpu_ptr(&stolen_time_region); if (!reg->kaddr) return 0; kaddr = rcu_replace_pointer(reg->kaddr, NULL, true); synchronize_rcu(); memunmap(kaddr); return 0; } static int stolen_time_cpu_online(unsigned int cpu) { struct pvclock_vcpu_stolen_time *kaddr = NULL; struct pv_time_stolen_time_region *reg; struct arm_smccc_res res; reg = this_cpu_ptr(&stolen_time_region); arm_smccc_1_1_invoke(ARM_SMCCC_HV_PV_TIME_ST, &res); if (res.a0 == SMCCC_RET_NOT_SUPPORTED) return -EINVAL; kaddr = memremap(res.a0, sizeof(struct pvclock_vcpu_stolen_time), MEMREMAP_WB); rcu_assign_pointer(reg->kaddr, kaddr); if (!reg->kaddr) { pr_warn("Failed to map stolen time data structure\n"); return -ENOMEM; } if (le32_to_cpu(kaddr->revision) != 0 || le32_to_cpu(kaddr->attributes) != 0) { pr_warn_once("Unexpected revision or attributes in stolen time data\n"); return -ENXIO; } return 0; } static int __init pv_time_init_stolen_time(void) { int ret; ret = cpuhp_setup_state(CPUHP_AP_ONLINE_DYN, "hypervisor/arm/pvtime:online", stolen_time_cpu_online, stolen_time_cpu_down_prepare); if (ret < 0) return ret; return 0; } static bool __init has_pv_steal_clock(void) { struct arm_smccc_res res; arm_smccc_1_1_invoke(ARM_SMCCC_ARCH_FEATURES_FUNC_ID, ARM_SMCCC_HV_PV_TIME_FEATURES, &res); if (res.a0 != SMCCC_RET_SUCCESS) return false; arm_smccc_1_1_invoke(ARM_SMCCC_HV_PV_TIME_FEATURES, ARM_SMCCC_HV_PV_TIME_ST, &res); return (res.a0 == SMCCC_RET_SUCCESS); } int __init pv_time_init(void) { int ret; if (!has_pv_steal_clock()) return 0; ret = pv_time_init_stolen_time(); if (ret) return ret; static_call_update(pv_steal_clock, para_steal_clock); static_key_slow_inc(¶virt_steal_enabled); if (steal_acc) static_key_slow_inc(¶virt_steal_rq_enabled); pr_info("using stolen time PV\n"); return 0; } #ifdef CONFIG_PARAVIRT_SCHED DEFINE_PER_CPU(struct pvsched_vcpu_state, pvsched_vcpu_region) __aligned(64); EXPORT_PER_CPU_SYMBOL(pvsched_vcpu_region); static bool kvm_vcpu_is_preempted(int cpu) { struct pvsched_vcpu_state *reg; u32 preempted; reg = &per_cpu(pvsched_vcpu_region, cpu); if (!reg) { pr_warn_once("PV sched enabled but not configured for cpu %d\n", cpu); return false; } preempted = le32_to_cpu(READ_ONCE(reg->preempted)); return !!preempted; } static int pvsched_vcpu_state_dying_cpu(unsigned int cpu) { struct pvsched_vcpu_state *reg; struct arm_smccc_res res; reg = this_cpu_ptr(&pvsched_vcpu_region); if (!reg) return -EFAULT; arm_smccc_1_1_invoke(ARM_SMCCC_HV_PV_SCHED_IPA_RELEASE, &res); memset(reg, 0, sizeof(*reg)); return 0; } static int init_pvsched_vcpu_state(unsigned int cpu) { struct pvsched_vcpu_state *reg; struct arm_smccc_res res; reg = this_cpu_ptr(&pvsched_vcpu_region); if (!reg) return -EFAULT; /* Pass the memory address to host via hypercall */ arm_smccc_1_1_invoke(ARM_SMCCC_HV_PV_SCHED_IPA_INIT, virt_to_phys(reg), &res); return 0; } static int kvm_arm_init_pvsched(void) { int ret; ret = cpuhp_setup_state(CPUHP_AP_ARM_KVM_PVSCHED_STARTING, "hypervisor/arm/pvsched:starting", init_pvsched_vcpu_state, pvsched_vcpu_state_dying_cpu); if (ret < 0) { pr_warn("PV sched init failed\n"); return ret; } return 0; } static bool has_kvm_pvsched(void) { struct arm_smccc_res res; /* To detect the presence of PV sched support we require SMCCC 1.1+ */ if (arm_smccc_1_1_get_conduit() == SMCCC_CONDUIT_NONE) return false; arm_smccc_1_1_invoke(ARM_SMCCC_ARCH_FEATURES_FUNC_ID, ARM_SMCCC_HV_PV_SCHED_FEATURES, &res); return (res.a0 == SMCCC_RET_SUCCESS); } int __init pv_sched_init(void) { int ret; if (is_hyp_mode_available()) return 0; if (!has_kvm_pvsched()) { pr_warn("PV sched is not available\n"); return 0; } ret = kvm_arm_init_pvsched(); if (ret) return ret; static_call_update(pv_vcpu_preempted, kvm_vcpu_is_preempted); pr_info("using PV sched preempted\n"); return 0; } early_initcall(pv_sched_init); #endif /* CONFIG_PARAVIRT_SCHED */