#ifndef _SPARC64_TLB_H #define _SPARC64_TLB_H #include <linux/swap.h> #include <linux/pagemap.h> #include <asm/pgalloc.h> #include <asm/tlbflush.h> #include <asm/mmu_context.h> #define TLB_BATCH_NR 192 /* * For UP we don't need to worry about TLB flush * and page free order so much.. */ #ifdef CONFIG_SMP #define FREE_PTE_NR 506 #define tlb_fast_mode(bp) ((bp)->pages_nr == ~0U) #else #define FREE_PTE_NR 1 #define tlb_fast_mode(bp) 1 #endif struct mmu_gather { struct mm_struct *mm; unsigned int pages_nr; unsigned int need_flush; unsigned int fullmm; unsigned int tlb_nr; unsigned long vaddrs[TLB_BATCH_NR]; struct page *pages[FREE_PTE_NR]; }; DECLARE_PER_CPU(struct mmu_gather, mmu_gathers); #ifdef CONFIG_SMP extern void smp_flush_tlb_pending(struct mm_struct *, unsigned long, unsigned long *); #endif extern void __flush_tlb_pending(unsigned long, unsigned long, unsigned long *); extern void flush_tlb_pending(void); static inline struct mmu_gather *tlb_gather_mmu(struct mm_struct *mm, unsigned int full_mm_flush) { struct mmu_gather *mp = &get_cpu_var(mmu_gathers); BUG_ON(mp->tlb_nr); mp->mm = mm; mp->pages_nr = num_online_cpus() > 1 ? 0U : ~0U; mp->fullmm = full_mm_flush; return mp; } static inline void tlb_flush_mmu(struct mmu_gather *mp) { if (mp->need_flush) { free_pages_and_swap_cache(mp->pages, mp->pages_nr); mp->pages_nr = 0; mp->need_flush = 0; } } #ifdef CONFIG_SMP extern void smp_flush_tlb_mm(struct mm_struct *mm); #define do_flush_tlb_mm(mm) smp_flush_tlb_mm(mm) #else #define do_flush_tlb_mm(mm) __flush_tlb_mm(CTX_HWBITS(mm->context), SECONDARY_CONTEXT) #endif static inline void tlb_finish_mmu(struct mmu_gather *mp, unsigned long start, unsigned long end) { tlb_flush_mmu(mp); if (mp->fullmm) mp->fullmm = 0; else flush_tlb_pending(); /* keep the page table cache within bounds */ check_pgt_cache(); put_cpu_var(mmu_gathers); } static inline void tlb_remove_page(struct mmu_gather *mp, struct page *page) { if (tlb_fast_mode(mp)) { free_page_and_swap_cache(page); return; } mp->need_flush = 1; mp->pages[mp->pages_nr++] = page; if (mp->pages_nr >= FREE_PTE_NR) tlb_flush_mmu(mp); } #define tlb_remove_tlb_entry(mp,ptep,addr) do { } while (0) #define pte_free_tlb(mp, ptepage) pte_free((mp)->mm, ptepage) #define pmd_free_tlb(mp, pmdp) pmd_free((mp)->mm, pmdp) #define pud_free_tlb(tlb,pudp) __pud_free_tlb(tlb,pudp) #define tlb_migrate_finish(mm) do { } while (0) #define tlb_start_vma(tlb, vma) do { } while (0) #define tlb_end_vma(tlb, vma) do { } while (0) #endif /* _SPARC64_TLB_H */