KVM: arm64: nv: Add sanitising to VNCR-backed FGT sysregs
authorMarc Zyngier <maz@kernel.org>
Wed, 14 Feb 2024 13:18:06 +0000 (13:18 +0000)
committerOliver Upton <oliver.upton@linux.dev>
Mon, 19 Feb 2024 17:13:00 +0000 (17:13 +0000)
Fine Grained Traps are controlled by a whole bunch of features.
Each one of them must be checked and the corresponding masks
computed so that we don't let the guest apply traps it shouldn't
be using.

This takes care of HFG[IRW]TR_EL2, HDFG[RW]TR_EL2, and HAFGRTR_EL2.

Signed-off-by: Marc Zyngier <maz@kernel.org>
Reviewed-by: Joey Gouly <joey.gouly@arm.com>
Link: https://lore.kernel.org/r/20240214131827.2856277-6-maz@kernel.org
Signed-off-by: Oliver Upton <oliver.upton@linux.dev>
arch/arm64/kvm/nested.c

index 29ed529681e221b3b1de671797380b8cb305e430..eda05eb16c0a3f34a75f61a5bd479161ee41de74 100644 (file)
@@ -265,6 +265,134 @@ int kvm_init_nv_sysregs(struct kvm *kvm)
                res1 |= HCR_E2H;
        set_sysreg_masks(kvm, HCR_EL2, res0, res1);
 
+       /* HFG[RW]TR_EL2 */
+       res0 = res1 = 0;
+       if (!(__vcpu_has_feature(&kvm->arch, KVM_ARM_VCPU_PTRAUTH_ADDRESS) &&
+             __vcpu_has_feature(&kvm->arch, KVM_ARM_VCPU_PTRAUTH_GENERIC)))
+               res0 |= (HFGxTR_EL2_APDAKey | HFGxTR_EL2_APDBKey |
+                        HFGxTR_EL2_APGAKey | HFGxTR_EL2_APIAKey |
+                        HFGxTR_EL2_APIBKey);
+       if (!kvm_has_feat(kvm, ID_AA64MMFR1_EL1, LO, IMP))
+               res0 |= (HFGxTR_EL2_LORC_EL1 | HFGxTR_EL2_LOREA_EL1 |
+                        HFGxTR_EL2_LORID_EL1 | HFGxTR_EL2_LORN_EL1 |
+                        HFGxTR_EL2_LORSA_EL1);
+       if (!kvm_has_feat(kvm, ID_AA64PFR0_EL1, CSV2, CSV2_2) &&
+           !kvm_has_feat(kvm, ID_AA64PFR1_EL1, CSV2_frac, CSV2_1p2))
+               res0 |= (HFGxTR_EL2_SCXTNUM_EL1 | HFGxTR_EL2_SCXTNUM_EL0);
+       if (!kvm_has_feat(kvm, ID_AA64PFR0_EL1, GIC, IMP))
+               res0 |= HFGxTR_EL2_ICC_IGRPENn_EL1;
+       if (!kvm_has_feat(kvm, ID_AA64PFR0_EL1, RAS, IMP))
+               res0 |= (HFGxTR_EL2_ERRIDR_EL1 | HFGxTR_EL2_ERRSELR_EL1 |
+                        HFGxTR_EL2_ERXFR_EL1 | HFGxTR_EL2_ERXCTLR_EL1 |
+                        HFGxTR_EL2_ERXSTATUS_EL1 | HFGxTR_EL2_ERXMISCn_EL1 |
+                        HFGxTR_EL2_ERXPFGF_EL1 | HFGxTR_EL2_ERXPFGCTL_EL1 |
+                        HFGxTR_EL2_ERXPFGCDN_EL1 | HFGxTR_EL2_ERXADDR_EL1);
+       if (!kvm_has_feat(kvm, ID_AA64ISAR1_EL1, LS64, LS64_ACCDATA))
+               res0 |= HFGxTR_EL2_nACCDATA_EL1;
+       if (!kvm_has_feat(kvm, ID_AA64PFR1_EL1, GCS, IMP))
+               res0 |= (HFGxTR_EL2_nGCS_EL0 | HFGxTR_EL2_nGCS_EL1);
+       if (!kvm_has_feat(kvm, ID_AA64PFR1_EL1, SME, IMP))
+               res0 |= (HFGxTR_EL2_nSMPRI_EL1 | HFGxTR_EL2_nTPIDR2_EL0);
+       if (!kvm_has_feat(kvm, ID_AA64PFR1_EL1, THE, IMP))
+               res0 |= HFGxTR_EL2_nRCWMASK_EL1;
+       if (!kvm_has_feat(kvm, ID_AA64MMFR3_EL1, S1PIE, IMP))
+               res0 |= (HFGxTR_EL2_nPIRE0_EL1 | HFGxTR_EL2_nPIR_EL1);
+       if (!kvm_has_feat(kvm, ID_AA64MMFR3_EL1, S1POE, IMP))
+               res0 |= (HFGxTR_EL2_nPOR_EL0 | HFGxTR_EL2_nPOR_EL1);
+       if (!kvm_has_feat(kvm, ID_AA64MMFR3_EL1, S2POE, IMP))
+               res0 |= HFGxTR_EL2_nS2POR_EL1;
+       if (!kvm_has_feat(kvm, ID_AA64MMFR3_EL1, AIE, IMP))
+               res0 |= (HFGxTR_EL2_nMAIR2_EL1 | HFGxTR_EL2_nAMAIR2_EL1);
+       set_sysreg_masks(kvm, HFGRTR_EL2, res0 | __HFGRTR_EL2_RES0, res1);
+       set_sysreg_masks(kvm, HFGWTR_EL2, res0 | __HFGWTR_EL2_RES0, res1);
+
+       /* HDFG[RW]TR_EL2 */
+       res0 = res1 = 0;
+       if (!kvm_has_feat(kvm, ID_AA64DFR0_EL1, DoubleLock, IMP))
+               res0 |= HDFGRTR_EL2_OSDLR_EL1;
+       if (!kvm_has_feat(kvm, ID_AA64DFR0_EL1, PMUVer, IMP))
+               res0 |= (HDFGRTR_EL2_PMEVCNTRn_EL0 | HDFGRTR_EL2_PMEVTYPERn_EL0 |
+                        HDFGRTR_EL2_PMCCFILTR_EL0 | HDFGRTR_EL2_PMCCNTR_EL0 |
+                        HDFGRTR_EL2_PMCNTEN | HDFGRTR_EL2_PMINTEN |
+                        HDFGRTR_EL2_PMOVS | HDFGRTR_EL2_PMSELR_EL0 |
+                        HDFGRTR_EL2_PMMIR_EL1 | HDFGRTR_EL2_PMUSERENR_EL0 |
+                        HDFGRTR_EL2_PMCEIDn_EL0);
+       if (!kvm_has_feat(kvm, ID_AA64DFR0_EL1, PMSVer, IMP))
+               res0 |= (HDFGRTR_EL2_PMBLIMITR_EL1 | HDFGRTR_EL2_PMBPTR_EL1 |
+                        HDFGRTR_EL2_PMBSR_EL1 | HDFGRTR_EL2_PMSCR_EL1 |
+                        HDFGRTR_EL2_PMSEVFR_EL1 | HDFGRTR_EL2_PMSFCR_EL1 |
+                        HDFGRTR_EL2_PMSICR_EL1 | HDFGRTR_EL2_PMSIDR_EL1 |
+                        HDFGRTR_EL2_PMSIRR_EL1 | HDFGRTR_EL2_PMSLATFR_EL1 |
+                        HDFGRTR_EL2_PMBIDR_EL1);
+       if (!kvm_has_feat(kvm, ID_AA64DFR0_EL1, TraceVer, IMP))
+               res0 |= (HDFGRTR_EL2_TRC | HDFGRTR_EL2_TRCAUTHSTATUS |
+                        HDFGRTR_EL2_TRCAUXCTLR | HDFGRTR_EL2_TRCCLAIM |
+                        HDFGRTR_EL2_TRCCNTVRn | HDFGRTR_EL2_TRCID |
+                        HDFGRTR_EL2_TRCIMSPECn | HDFGRTR_EL2_TRCOSLSR |
+                        HDFGRTR_EL2_TRCPRGCTLR | HDFGRTR_EL2_TRCSEQSTR |
+                        HDFGRTR_EL2_TRCSSCSRn | HDFGRTR_EL2_TRCSTATR |
+                        HDFGRTR_EL2_TRCVICTLR);
+       if (!kvm_has_feat(kvm, ID_AA64DFR0_EL1, TraceBuffer, IMP))
+               res0 |= (HDFGRTR_EL2_TRBBASER_EL1 | HDFGRTR_EL2_TRBIDR_EL1 |
+                        HDFGRTR_EL2_TRBLIMITR_EL1 | HDFGRTR_EL2_TRBMAR_EL1 |
+                        HDFGRTR_EL2_TRBPTR_EL1 | HDFGRTR_EL2_TRBSR_EL1 |
+                        HDFGRTR_EL2_TRBTRG_EL1);
+       if (!kvm_has_feat(kvm, ID_AA64DFR0_EL1, BRBE, IMP))
+               res0 |= (HDFGRTR_EL2_nBRBIDR | HDFGRTR_EL2_nBRBCTL |
+                        HDFGRTR_EL2_nBRBDATA);
+       if (!kvm_has_feat(kvm, ID_AA64DFR0_EL1, PMSVer, V1P2))
+               res0 |= HDFGRTR_EL2_nPMSNEVFR_EL1;
+       set_sysreg_masks(kvm, HDFGRTR_EL2, res0 | HDFGRTR_EL2_RES0, res1);
+
+       /* Reuse the bits from the read-side and add the write-specific stuff */
+       if (!kvm_has_feat(kvm, ID_AA64DFR0_EL1, PMUVer, IMP))
+               res0 |= (HDFGWTR_EL2_PMCR_EL0 | HDFGWTR_EL2_PMSWINC_EL0);
+       if (!kvm_has_feat(kvm, ID_AA64DFR0_EL1, TraceVer, IMP))
+               res0 |= HDFGWTR_EL2_TRCOSLAR;
+       if (!kvm_has_feat(kvm, ID_AA64DFR0_EL1, TraceFilt, IMP))
+               res0 |= HDFGWTR_EL2_TRFCR_EL1;
+       set_sysreg_masks(kvm, HFGWTR_EL2, res0 | HDFGWTR_EL2_RES0, res1);
+
+       /* HFGITR_EL2 */
+       res0 = HFGITR_EL2_RES0;
+       res1 = HFGITR_EL2_RES1;
+       if (!kvm_has_feat(kvm, ID_AA64ISAR1_EL1, DPB, DPB2))
+               res0 |= HFGITR_EL2_DCCVADP;
+       if (!kvm_has_feat(kvm, ID_AA64MMFR1_EL1, PAN, PAN2))
+               res0 |= (HFGITR_EL2_ATS1E1RP | HFGITR_EL2_ATS1E1WP);
+       if (!kvm_has_feat(kvm, ID_AA64ISAR0_EL1, TLB, OS))
+               res0 |= (HFGITR_EL2_TLBIRVAALE1OS | HFGITR_EL2_TLBIRVALE1OS |
+                        HFGITR_EL2_TLBIRVAAE1OS | HFGITR_EL2_TLBIRVAE1OS |
+                        HFGITR_EL2_TLBIVAALE1OS | HFGITR_EL2_TLBIVALE1OS |
+                        HFGITR_EL2_TLBIVAAE1OS | HFGITR_EL2_TLBIASIDE1OS |
+                        HFGITR_EL2_TLBIVAE1OS | HFGITR_EL2_TLBIVMALLE1OS);
+       if (!kvm_has_feat(kvm, ID_AA64ISAR0_EL1, TLB, RANGE))
+               res0 |= (HFGITR_EL2_TLBIRVAALE1 | HFGITR_EL2_TLBIRVALE1 |
+                        HFGITR_EL2_TLBIRVAAE1 | HFGITR_EL2_TLBIRVAE1 |
+                        HFGITR_EL2_TLBIRVAALE1IS | HFGITR_EL2_TLBIRVALE1IS |
+                        HFGITR_EL2_TLBIRVAAE1IS | HFGITR_EL2_TLBIRVAE1IS |
+                        HFGITR_EL2_TLBIRVAALE1OS | HFGITR_EL2_TLBIRVALE1OS |
+                        HFGITR_EL2_TLBIRVAAE1OS | HFGITR_EL2_TLBIRVAE1OS);
+       if (!kvm_has_feat(kvm, ID_AA64ISAR1_EL1, SPECRES, IMP))
+               res0 |= (HFGITR_EL2_CFPRCTX | HFGITR_EL2_DVPRCTX |
+                        HFGITR_EL2_CPPRCTX);
+       if (!kvm_has_feat(kvm, ID_AA64DFR0_EL1, BRBE, IMP))
+               res0 |= (HFGITR_EL2_nBRBINJ | HFGITR_EL2_nBRBIALL);
+       if (!kvm_has_feat(kvm, ID_AA64PFR1_EL1, GCS, IMP))
+               res0 |= (HFGITR_EL2_nGCSPUSHM_EL1 | HFGITR_EL2_nGCSSTR_EL1 |
+                        HFGITR_EL2_nGCSEPP);
+       if (!kvm_has_feat(kvm, ID_AA64ISAR1_EL1, SPECRES, COSP_RCTX))
+               res0 |= HFGITR_EL2_COSPRCTX;
+       if (!kvm_has_feat(kvm, ID_AA64ISAR2_EL1, ATS1A, IMP))
+               res0 |= HFGITR_EL2_ATS1E1A;
+       set_sysreg_masks(kvm, HFGITR_EL2, res0, res1);
+
+       /* HAFGRTR_EL2 - not a lot to see here */
+       res0 = HAFGRTR_EL2_RES0;
+       res1 = HAFGRTR_EL2_RES1;
+       if (!kvm_has_feat(kvm, ID_AA64PFR0_EL1, AMU, V1P1))
+               res0 |= ~(res0 | res1);
+       set_sysreg_masks(kvm, HAFGRTR_EL2, res0, res1);
 out:
        mutex_unlock(&kvm->arch.config_lock);