arm64: alternative: Apply alternatives early in boot process
authorDaniel Thompson <daniel.thompson@linaro.org>
Thu, 31 Jan 2019 14:58:53 +0000 (14:58 +0000)
committerCatalin Marinas <catalin.marinas@arm.com>
Wed, 6 Feb 2019 10:05:20 +0000 (10:05 +0000)
Currently alternatives are applied very late in the boot process (and
a long time after we enable scheduling). Some alternative sequences,
such as those that alter the way CPU context is stored, must be applied
much earlier in the boot sequence.

Introduce apply_boot_alternatives() to allow some alternatives to be
applied immediately after we detect the CPU features of the boot CPU.

Signed-off-by: Daniel Thompson <daniel.thompson@linaro.org>
[julien.thierry@arm.com: rename to fit new cpufeature framework better,
 apply BOOT_SCOPE feature early in boot]
Signed-off-by: Julien Thierry <julien.thierry@arm.com>
Reviewed-by: Suzuki K Poulose <suzuki.poulose@arm.com>
Reviewed-by: Marc Zyngier <marc.zyngier@arm.com>
Cc: Will Deacon <will.deacon@arm.com>
Cc: Christoffer Dall <christoffer.dall@arm.com>
Cc: Suzuki K Poulose <suzuki.poulose@arm.com>
Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
arch/arm64/include/asm/alternative.h
arch/arm64/include/asm/cpufeature.h
arch/arm64/kernel/alternative.c
arch/arm64/kernel/cpufeature.c
arch/arm64/kernel/smp.c

index 9806a2357e70afcfed9c1f6d51c7cecc33469168..b9f8d787eea9fc57b724d7a87dff61172551b156 100644 (file)
@@ -25,6 +25,7 @@ struct alt_instr {
 typedef void (*alternative_cb_t)(struct alt_instr *alt,
                                 __le32 *origptr, __le32 *updptr, int nr_inst);
 
+void __init apply_boot_alternatives(void);
 void __init apply_alternatives_all(void);
 bool alternative_is_applied(u16 cpufeature);
 
index 89c3f318f6be5a627300181af1d30c3a6643447a..e505e1fbd2b933b42af6f717e48e7eb284248659 100644 (file)
@@ -391,6 +391,10 @@ extern DECLARE_BITMAP(cpu_hwcaps, ARM64_NCAPS);
 extern struct static_key_false cpu_hwcap_keys[ARM64_NCAPS];
 extern struct static_key_false arm64_const_caps_ready;
 
+/* ARM64 CAPS + alternative_cb */
+#define ARM64_NPATCHABLE (ARM64_NCAPS + 1)
+extern DECLARE_BITMAP(boot_capabilities, ARM64_NPATCHABLE);
+
 #define for_each_available_cap(cap)            \
        for_each_set_bit(cap, cpu_hwcaps, ARM64_NCAPS)
 
index c947d2246017515b8b430078577ba7c1c27e20cb..a9b4677631532306d33303a362a5c63a6be16785 100644 (file)
@@ -155,7 +155,8 @@ static void clean_dcache_range_nopatch(u64 start, u64 end)
        } while (cur += d_size, cur < end);
 }
 
-static void __apply_alternatives(void *alt_region, bool is_module)
+static void __apply_alternatives(void *alt_region,  bool is_module,
+                                unsigned long *feature_mask)
 {
        struct alt_instr *alt;
        struct alt_region *region = alt_region;
@@ -165,6 +166,9 @@ static void __apply_alternatives(void *alt_region, bool is_module)
        for (alt = region->begin; alt < region->end; alt++) {
                int nr_inst;
 
+               if (!test_bit(alt->cpufeature, feature_mask))
+                       continue;
+
                /* Use ARM64_CB_PATCH as an unconditional patch */
                if (alt->cpufeature < ARM64_CB_PATCH &&
                    !cpus_have_cap(alt->cpufeature))
@@ -203,8 +207,11 @@ static void __apply_alternatives(void *alt_region, bool is_module)
                __flush_icache_all();
                isb();
 
-               /* We applied all that was available */
-               bitmap_copy(applied_alternatives, cpu_hwcaps, ARM64_NCAPS);
+               /* Ignore ARM64_CB bit from feature mask */
+               bitmap_or(applied_alternatives, applied_alternatives,
+                         feature_mask, ARM64_NCAPS);
+               bitmap_and(applied_alternatives, applied_alternatives,
+                          cpu_hwcaps, ARM64_NCAPS);
        }
 }
 
@@ -225,8 +232,13 @@ static int __apply_alternatives_multi_stop(void *unused)
                        cpu_relax();
                isb();
        } else {
+               DECLARE_BITMAP(remaining_capabilities, ARM64_NPATCHABLE);
+
+               bitmap_complement(remaining_capabilities, boot_capabilities,
+                                 ARM64_NPATCHABLE);
+
                BUG_ON(all_alternatives_applied);
-               __apply_alternatives(&region, false);
+               __apply_alternatives(&region, false, remaining_capabilities);
                /* Barriers provided by the cache flushing */
                WRITE_ONCE(all_alternatives_applied, 1);
        }
@@ -240,6 +252,24 @@ void __init apply_alternatives_all(void)
        stop_machine(__apply_alternatives_multi_stop, NULL, cpu_online_mask);
 }
 
+/*
+ * This is called very early in the boot process (directly after we run
+ * a feature detect on the boot CPU). No need to worry about other CPUs
+ * here.
+ */
+void __init apply_boot_alternatives(void)
+{
+       struct alt_region region = {
+               .begin  = (struct alt_instr *)__alt_instructions,
+               .end    = (struct alt_instr *)__alt_instructions_end,
+       };
+
+       /* If called on non-boot cpu things could go wrong */
+       WARN_ON(smp_processor_id() != 0);
+
+       __apply_alternatives(&region, false, &boot_capabilities[0]);
+}
+
 #ifdef CONFIG_MODULES
 void apply_alternatives_module(void *start, size_t length)
 {
@@ -247,7 +277,10 @@ void apply_alternatives_module(void *start, size_t length)
                .begin  = start,
                .end    = start + length,
        };
+       DECLARE_BITMAP(all_capabilities, ARM64_NPATCHABLE);
+
+       bitmap_fill(all_capabilities, ARM64_NPATCHABLE);
 
-       __apply_alternatives(&region, true);
+       __apply_alternatives(&region, true, &all_capabilities[0]);
 }
 #endif
index d607ea33228c3f1ce4bc98f61c2f11fef560388e..b530fb24e6c6f5db4605bfffeb6b2b583f9da820 100644 (file)
@@ -54,6 +54,9 @@ DECLARE_BITMAP(cpu_hwcaps, ARM64_NCAPS);
 EXPORT_SYMBOL(cpu_hwcaps);
 static struct arm64_cpu_capabilities const __ro_after_init *cpu_hwcaps_ptrs[ARM64_NCAPS];
 
+/* Need also bit for ARM64_CB_PATCH */
+DECLARE_BITMAP(boot_capabilities, ARM64_NPATCHABLE);
+
 /*
  * Flag to indicate if we have computed the system wide
  * capabilities based on the boot time active CPUs. This
@@ -1677,6 +1680,9 @@ static void update_cpu_capabilities(u16 scope_mask)
                if (caps->desc)
                        pr_info("detected: %s\n", caps->desc);
                cpus_set_cap(caps->capability);
+
+               if ((scope_mask & SCOPE_BOOT_CPU) && (caps->type & SCOPE_BOOT_CPU))
+                       set_bit(caps->capability, boot_capabilities);
        }
 }
 
index 1598d6f7200a5a2f2a323f66c1b0fa95ad4a77b0..a944edd39d2d12a65e89fd1645aaebf70acfe25a 100644 (file)
@@ -419,6 +419,13 @@ void __init smp_prepare_boot_cpu(void)
         */
        jump_label_init();
        cpuinfo_store_boot_cpu();
+
+       /*
+        * We now know enough about the boot CPU to apply the
+        * alternatives that cannot wait until interrupt handling
+        * and/or scheduling is enabled.
+        */
+       apply_boot_alternatives();
 }
 
 static u64 __init of_get_cpu_mpidr(struct device_node *dn)