It's not entirely complete, doesn't deal with accelerators (reverts to
tlbie), not all the boot code is converted, kernel space invalidations
not converted, and KVM not converted, also radix only to start with. We
can start to add more cases if this will be useful.
Thanks,
Nick
---
arch/powerpc/mm/book3s64/radix_tlb.c | 149 +++++++++++++++++++++++++--
1 file changed, 140 insertions(+), 9 deletions(-)
diff --git a/arch/powerpc/mm/book3s64/radix_tlb.c
b/arch/powerpc/mm/book3s64/radix_tlb.c
index 71f7fede2fa4..56ceecbd3d5c 100644
--- a/arch/powerpc/mm/book3s64/radix_tlb.c
+++ b/arch/powerpc/mm/book3s64/radix_tlb.c
@@ -11,6 +11,7 @@
#include <linux/mmu_context.h>
#include <linux/sched/mm.h>
+#include <asm/debugfs.h>
#include <asm/ppc-opcode.h>
#include <asm/tlb.h>
#include <asm/tlbflush.h>
@@ -285,6 +286,30 @@ static inline void _tlbie_pid(unsigned long pid, unsigned
long ric)
asm volatile("eieio; tlbsync; ptesync": : :"memory");
}
+struct tlbiel_pid {
+ unsigned long pid;
+ unsigned long ric;
+};
+
+static void do_tlbiel_pid(void *info)
+{
+ struct tlbiel_pid *t = info;
+
+ if (t->ric == RIC_FLUSH_TLB)
+ _tlbiel_pid(t->pid, RIC_FLUSH_TLB);
+ else if (t->ric == RIC_FLUSH_PWC)
+ _tlbiel_pid(t->pid, RIC_FLUSH_PWC);
+ else
+ _tlbiel_pid(t->pid, RIC_FLUSH_ALL);
+}
+
+static inline void _tlbiel_pid_broadcast(const struct cpumask *cpus,
+ unsigned long pid, unsigned long ric)
+{
+ struct tlbiel_pid t = { .pid = pid, .ric = ric };
+ on_each_cpu_mask(cpus, do_tlbiel_pid, &t, 1);
+}
+
static inline void _tlbiel_lpid(unsigned long lpid, unsigned long ric)
{
int set;
@@ -420,6 +445,61 @@ static __always_inline void _tlbie_va(unsigned long va,
unsigned long pid,
asm volatile("eieio; tlbsync; ptesync": : :"memory");
}
+struct tlbiel_va {
+ unsigned long pid;
+ unsigned long va;
+ unsigned long psize;
+ unsigned long ric;
+};
+
+static void do_tlbiel_va(void *info)
+{
+ struct tlbiel_va *t = info;
+
+ if (t->ric == RIC_FLUSH_TLB)
+ _tlbiel_va(t->va, t->pid, t->psize, RIC_FLUSH_TLB);
+ else if (t->ric == RIC_FLUSH_PWC)
+ _tlbiel_va(t->va, t->pid, t->psize, RIC_FLUSH_PWC);
+ else
+ _tlbiel_va(t->va, t->pid, t->psize, RIC_FLUSH_ALL);
+}
+
+static inline void _tlbiel_va_broadcast(const struct cpumask *cpus,
+ unsigned long va, unsigned long pid,
+ unsigned long psize, unsigned long ric)
+{
+ struct tlbiel_va t = { .va = va, .pid = pid, .psize = psize, .ric = ric
};
+ on_each_cpu_mask(cpus, do_tlbiel_va, &t, 1);
+}
+
+struct tlbiel_va_range {
+ unsigned long pid;
+ unsigned long start;
+ unsigned long end;
+ unsigned long page_size;
+ unsigned long psize;
+ bool also_pwc;
+};
+
+static void do_tlbiel_va_range(void *info)
+{
+ struct tlbiel_va_range *t = info;
+
+ _tlbiel_va_range(t->start, t->end, t->pid, t->page_size,
+ t->psize, t->also_pwc);
+}
+
+static inline void _tlbiel_va_range_broadcast(const struct cpumask *cpus,
+ unsigned long start, unsigned long end,
+ unsigned long pid, unsigned long page_size,
+ unsigned long psize, bool also_pwc)
+{
+ struct tlbiel_va_range t = { .start = start, .end = end,
+ .pid = pid, .page_size = page_size,
+ .psize = psize, .also_pwc = also_pwc };
+ on_each_cpu_mask(cpus, do_tlbiel_va_range, &t, 1);
+}
+
static __always_inline void _tlbie_lpid_va(unsigned long va, unsigned long
lpid,
unsigned long psize, unsigned long ric)
{
@@ -524,6 +604,12 @@ static bool mm_needs_flush_escalation(struct mm_struct *mm)
return false;
}
+static bool tlbie_enabled = true;
+static bool use_tlbie(void)
+{
+ return tlbie_enabled;
+}
+
#ifdef CONFIG_SMP
static void do_exit_flush_lazy_tlb(void *arg)
{
@@ -582,8 +668,10 @@ void radix__flush_tlb_mm(struct mm_struct *mm)
if (mm_needs_flush_escalation(mm))
_tlbie_pid(pid, RIC_FLUSH_ALL);
- else
+ else if (use_tlbie())
_tlbie_pid(pid, RIC_FLUSH_TLB);
+ else
+ _tlbiel_pid_broadcast(mm_cpumask(mm), pid,
RIC_FLUSH_TLB);
} else {
local:
_tlbiel_pid(pid, RIC_FLUSH_TLB);
@@ -609,7 +697,10 @@ static void __flush_all_mm(struct mm_struct *mm, bool
fullmm)
goto local;
}
}
- _tlbie_pid(pid, RIC_FLUSH_ALL);
+ if (mm_needs_flush_escalation(mm) || use_tlbie())
+ _tlbie_pid(pid, RIC_FLUSH_ALL);
+ else
+ _tlbiel_pid_broadcast(mm_cpumask(mm), pid,
RIC_FLUSH_ALL);
} else {
local:
_tlbiel_pid(pid, RIC_FLUSH_ALL);
@@ -644,7 +735,11 @@ void radix__flush_tlb_page_psize(struct mm_struct *mm,
unsigned long vmaddr,
exit_flush_lazy_tlbs(mm);
goto local;
}
- _tlbie_va(vmaddr, pid, psize, RIC_FLUSH_TLB);
+ if (mm_needs_flush_escalation(mm) || use_tlbie())
+ _tlbie_va(vmaddr, pid, psize, RIC_FLUSH_TLB);
+ else
+ _tlbiel_va_broadcast(mm_cpumask(mm),
+ vmaddr, pid, psize, RIC_FLUSH_TLB);
} else {
local:
_tlbiel_va(vmaddr, pid, psize, RIC_FLUSH_TLB);
@@ -731,8 +826,11 @@ static inline void __radix__flush_tlb_range(struct
mm_struct *mm,
} else {
if (mm_needs_flush_escalation(mm))
_tlbie_pid(pid, RIC_FLUSH_ALL);
- else
+ else if (use_tlbie())
_tlbie_pid(pid, RIC_FLUSH_TLB);
+ else
+ _tlbiel_pid_broadcast(mm_cpumask(mm),
+ pid, RIC_FLUSH_TLB);
}
} else {
bool hflush = flush_all_sizes;
@@ -757,8 +855,8 @@ static inline void __radix__flush_tlb_range(struct
mm_struct *mm,
gflush = false;
}
- asm volatile("ptesync": : :"memory");
if (local) {
+ asm volatile("ptesync": : :"memory");
__tlbiel_va_range(start, end, pid, page_size,
mmu_virtual_psize);
if (hflush)
__tlbiel_va_range(hstart, hend, pid,
@@ -767,7 +865,8 @@ static inline void __radix__flush_tlb_range(struct
mm_struct *mm,
__tlbiel_va_range(gstart, gend, pid,
PUD_SIZE, MMU_PAGE_1G);
asm volatile("ptesync": : :"memory");
- } else {
+ } else if (use_tlbie()) {
+ asm volatile("ptesync": : :"memory");
__tlbie_va_range(start, end, pid, page_size,
mmu_virtual_psize);
if (hflush)
__tlbie_va_range(hstart, hend, pid,
@@ -777,6 +876,15 @@ static inline void __radix__flush_tlb_range(struct
mm_struct *mm,
PUD_SIZE, MMU_PAGE_1G);
fixup_tlbie();
asm volatile("eieio; tlbsync; ptesync": : :"memory");
+ } else {
+ _tlbiel_va_range_broadcast(mm_cpumask(mm),
+ start, end, pid, page_size,
mmu_virtual_psize, false);
+ if (hflush)
+ _tlbiel_va_range_broadcast(mm_cpumask(mm),
+ hstart, hend, pid, PMD_SIZE,
MMU_PAGE_2M, false);
+ if (gflush)
+ _tlbiel_va_range_broadcast(mm_cpumask(mm),
+ gstart, gend, pid, PUD_SIZE,
MMU_PAGE_1G, false);
}
}
preempt_enable();
@@ -969,13 +1077,22 @@ static __always_inline void
__radix__flush_tlb_range_psize(struct mm_struct *mm,
if (mm_needs_flush_escalation(mm))
also_pwc = true;
- _tlbie_pid(pid, also_pwc ? RIC_FLUSH_ALL : RIC_FLUSH_TLB);
+ if (use_tlbie())
+ _tlbie_pid(pid,
+ also_pwc ? RIC_FLUSH_ALL :
RIC_FLUSH_TLB);
+ else
+ _tlbiel_pid_broadcast(mm_cpumask(mm), pid,
+ also_pwc ? RIC_FLUSH_ALL :
RIC_FLUSH_TLB);
+
}
} else {
if (local)
_tlbiel_va_range(start, end, pid, page_size, psize,
also_pwc);
- else
+ else if (mm_needs_flush_escalation(mm) || use_tlbie())
_tlbie_va_range(start, end, pid, page_size, psize,
also_pwc);
+ else
+ _tlbiel_va_range_broadcast(mm_cpumask(mm),
+ start, end, pid, page_size, psize,
also_pwc);
}
preempt_enable();
}
@@ -1017,7 +1134,11 @@ void radix__flush_tlb_collapsed_pmd(struct mm_struct
*mm, unsigned long addr)
exit_flush_lazy_tlbs(mm);
goto local;
}
- _tlbie_va_range(addr, end, pid, PAGE_SIZE, mmu_virtual_psize,
true);
+ if (mm_needs_flush_escalation(mm) || use_tlbie())
+ _tlbie_va_range(addr, end, pid, PAGE_SIZE,
mmu_virtual_psize, true);
+ else
+ _tlbiel_va_range_broadcast(mm_cpumask(mm),
+ addr, end, pid, PAGE_SIZE,
mmu_virtual_psize, true);
} else {
local:
_tlbiel_va_range(addr, end, pid, PAGE_SIZE, mmu_virtual_psize,
true);
@@ -1100,3 +1221,13 @@ extern void radix_kvm_prefetch_workaround(struct
mm_struct *mm)
}
EXPORT_SYMBOL_GPL(radix_kvm_prefetch_workaround);
#endif /* CONFIG_KVM_BOOK3S_HV_POSSIBLE */
+
+static int __init radix_tlb_setup(void)
+{
+ debugfs_create_bool("tlbie_enabled", 0600,
+ powerpc_debugfs_root,
+ &tlbie_enabled);
+
+ return 0;
+}
+arch_initcall(radix_tlb_setup);