// SPDX-License-Identifier: GPL-2.0 /* * Alpha TLB shootdown helpers * * Copyright (C) 2025 Magnus Lindholm * * Alpha-specific TLB flush helpers that cannot be expressed purely * as inline functions. * * These helpers provide combined MM context handling (ASN rollover) * and immediate TLB invalidation for page migration and memory * compaction paths, where lazy shootdowns are insufficient. */ #include #include #include #include #include #include #define asn_locked() (cpu_data[smp_processor_id()].asn_lock) /* * Migration/compaction helper: combine mm context (ASN) handling with an * immediate per-page TLB invalidate and (for exec) an instruction barrier. * * This mirrors the SMP combined IPI handler semantics, but runs locally on UP. */ #ifndef CONFIG_SMP void migrate_flush_tlb_page(struct vm_area_struct *vma, unsigned long addr) { struct mm_struct *mm = vma->vm_mm; int tbi_type = (vma->vm_flags & VM_EXEC) ? 3 : 2; /* * First do the mm-context side: * If we're currently running this mm, reload a fresh context ASN. * Otherwise, mark context invalid. * * On UP, this is mostly about matching the SMP semantics and ensuring * exec/i-cache tagging assumptions hold when compaction migrates pages. */ if (mm == current->active_mm) flush_tlb_current(mm); else flush_tlb_other(mm); /* * Then do the immediate translation kill for this VA. * For exec mappings, order instruction fetch after invalidation. */ tbi(tbi_type, addr); } #else struct tlb_mm_and_addr { struct mm_struct *mm; unsigned long addr; int tbi_type; /* 2 = DTB, 3 = ITB+DTB */ }; static void ipi_flush_mm_and_page(void *x) { struct tlb_mm_and_addr *d = x; /* Part 1: mm context side (Alpha uses ASN/context as a key mechanism). */ if (d->mm == current->active_mm && !asn_locked()) __load_new_mm_context(d->mm); else flush_tlb_other(d->mm); /* Part 2: immediate per-VA invalidation on this CPU. */ tbi(d->tbi_type, d->addr); } void migrate_flush_tlb_page(struct vm_area_struct *vma, unsigned long addr) { struct mm_struct *mm = vma->vm_mm; struct tlb_mm_and_addr d = { .mm = mm, .addr = addr, .tbi_type = (vma->vm_flags & VM_EXEC) ? 3 : 2, }; /* * One synchronous rendezvous: every CPU runs ipi_flush_mm_and_page(). * This is the "combined" version of flush_tlb_mm + per-page invalidate. */ preempt_disable(); on_each_cpu(ipi_flush_mm_and_page, &d, 1); /* * mimic flush_tlb_mm()'s mm_users<=1 optimization. */ if (atomic_read(&mm->mm_users) <= 1) { int cpu, this_cpu; this_cpu = smp_processor_id(); for (cpu = 0; cpu < NR_CPUS; cpu++) { if (!cpu_online(cpu) || cpu == this_cpu) continue; if (READ_ONCE(mm->context[cpu])) WRITE_ONCE(mm->context[cpu], 0); } } preempt_enable(); } #endif