1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
|
/* SPDX-License-Identifier: GPL-2.0 */
#ifndef _ASM_M32R_TLBFLUSH_H
#define _ASM_M32R_TLBFLUSH_H
#include <asm/m32r.h>
/*
* TLB flushing:
*
* - flush_tlb() flushes the current mm struct TLBs
* - flush_tlb_all() flushes all processes TLBs
* - flush_tlb_mm(mm) flushes the specified mm context TLB's
* - flush_tlb_page(vma, vmaddr) flushes one page
* - flush_tlb_range(vma, start, end) flushes a range of pages
* - flush_tlb_kernel_range(start, end) flushes a range of kernel pages
*/
extern void local_flush_tlb_all(void);
extern void local_flush_tlb_mm(struct mm_struct *);
extern void local_flush_tlb_page(struct vm_area_struct *, unsigned long);
extern void local_flush_tlb_range(struct vm_area_struct *, unsigned long,
unsigned long);
#ifndef CONFIG_SMP
#ifdef CONFIG_MMU
#define flush_tlb_all() local_flush_tlb_all()
#define flush_tlb_mm(mm) local_flush_tlb_mm(mm)
#define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page)
#define flush_tlb_range(vma, start, end) \
local_flush_tlb_range(vma, start, end)
#define flush_tlb_kernel_range(start, end) local_flush_tlb_all()
#else /* CONFIG_MMU */
#define flush_tlb_all() do { } while (0)
#define flush_tlb_mm(mm) do { } while (0)
#define flush_tlb_page(vma, vmaddr) do { } while (0)
#define flush_tlb_range(vma, start, end) do { } while (0)
#endif /* CONFIG_MMU */
#else /* CONFIG_SMP */
extern void smp_flush_tlb_all(void);
extern void smp_flush_tlb_mm(struct mm_struct *);
extern void smp_flush_tlb_page(struct vm_area_struct *, unsigned long);
extern void smp_flush_tlb_range(struct vm_area_struct *, unsigned long,
unsigned long);
#define flush_tlb_all() smp_flush_tlb_all()
#define flush_tlb_mm(mm) smp_flush_tlb_mm(mm)
#define flush_tlb_page(vma, page) smp_flush_tlb_page(vma, page)
#define flush_tlb_range(vma, start, end) \
smp_flush_tlb_range(vma, start, end)
#define flush_tlb_kernel_range(start, end) smp_flush_tlb_all()
#endif /* CONFIG_SMP */
static __inline__ void __flush_tlb_page(unsigned long page)
{
unsigned int tmpreg0, tmpreg1, tmpreg2;
__asm__ __volatile__ (
"seth %0, #high(%4) \n\t"
"st %3, @(%5, %0) \n\t"
"ldi %1, #1 \n\t"
"st %1, @(%6, %0) \n\t"
"add3 %1, %0, %7 \n\t"
".fillinsn \n"
"1: \n\t"
"ld %2, @(%6, %0) \n\t"
"bnez %2, 1b \n\t"
"ld %0, @%1+ \n\t"
"ld %1, @%1 \n\t"
"st %2, @+%0 \n\t"
"st %2, @+%1 \n\t"
: "=&r" (tmpreg0), "=&r" (tmpreg1), "=&r" (tmpreg2)
: "r" (page), "i" (MMU_REG_BASE), "i" (MSVA_offset),
"i" (MTOP_offset), "i" (MIDXI_offset)
: "memory"
);
}
static __inline__ void __flush_tlb_all(void)
{
unsigned int tmpreg0, tmpreg1;
__asm__ __volatile__ (
"seth %0, #high(%2) \n\t"
"or3 %0, %0, #low(%2) \n\t"
"ldi %1, #0xc \n\t"
"st %1, @%0 \n\t"
".fillinsn \n"
"1: \n\t"
"ld %1, @%0 \n\t"
"bnez %1, 1b \n\t"
: "=&r" (tmpreg0), "=&r" (tmpreg1)
: "i" (MTOP) : "memory"
);
}
extern void update_mmu_cache(struct vm_area_struct *, unsigned long, pte_t *);
#endif /* _ASM_M32R_TLBFLUSH_H */
|