diff options
author | Fenghua Yu <fenghua.yu@intel.com> | 2008-10-17 15:14:13 -0400 |
---|---|---|
committer | Tony Luck <tony.luck@intel.com> | 2008-10-17 15:14:13 -0400 |
commit | 62fdd7678a26efadd6ac5c2869543caff77d2df0 (patch) | |
tree | 0dd67208590c4540ff6a4476579a55bcac0d1fce /arch/ia64/lib | |
parent | 6bb7a935489dab20802dde6c2cb7d8582f4849bf (diff) |
[IA64] Add Variable Page Size and IA64 Support in Intel IOMMU
The patch contains Intel IOMMU IA64 specific code. It defines new
machvec dig_vtd, hooks for IOMMU, DMAR table detection, cache line flush
function, etc.
For a generic kernel with CONFIG_DMAR=y, if Intel IOMMU is detected,
dig_vtd is used for machinve vector. Otherwise, kernel falls back to
dig machine vector. Kernel parameter "machvec=dig" or "intel_iommu=off"
can be used to force kernel to boot dig machine vector.
Signed-off-by: Fenghua Yu <fenghua.yu@intel.com>
Signed-off-by: Tony Luck <tony.luck@intel.com>
Diffstat (limited to 'arch/ia64/lib')
-rw-r--r-- | arch/ia64/lib/flush.S | 55 |
1 files changed, 55 insertions, 0 deletions
diff --git a/arch/ia64/lib/flush.S b/arch/ia64/lib/flush.S index 2a0d27f2f21b..1d8c88860063 100644 --- a/arch/ia64/lib/flush.S +++ b/arch/ia64/lib/flush.S | |||
@@ -60,3 +60,58 @@ GLOBAL_ENTRY(flush_icache_range) | |||
60 | mov ar.lc=r3 // restore ar.lc | 60 | mov ar.lc=r3 // restore ar.lc |
61 | br.ret.sptk.many rp | 61 | br.ret.sptk.many rp |
62 | END(flush_icache_range) | 62 | END(flush_icache_range) |
63 | |||
64 | /* | ||
65 | * clflush_cache_range(start,size) | ||
66 | * | ||
67 | * Flush cache lines from start to start+size-1. | ||
68 | * | ||
69 | * Must deal with range from start to start+size-1 but nothing else | ||
70 | * (need to be careful not to touch addresses that may be | ||
71 | * unmapped). | ||
72 | * | ||
73 | * Note: "in0" and "in1" are preserved for debugging purposes. | ||
74 | */ | ||
75 | .section .kprobes.text,"ax" | ||
76 | GLOBAL_ENTRY(clflush_cache_range) | ||
77 | |||
78 | .prologue | ||
79 | alloc r2=ar.pfs,2,0,0,0 | ||
80 | movl r3=ia64_cache_stride_shift | ||
81 | mov r21=1 | ||
82 | add r22=in1,in0 | ||
83 | ;; | ||
84 | ld8 r20=[r3] // r20: stride shift | ||
85 | sub r22=r22,r0,1 // last byte address | ||
86 | ;; | ||
87 | shr.u r23=in0,r20 // start / (stride size) | ||
88 | shr.u r22=r22,r20 // (last byte address) / (stride size) | ||
89 | shl r21=r21,r20 // r21: stride size of the i-cache(s) | ||
90 | ;; | ||
91 | sub r8=r22,r23 // number of strides - 1 | ||
92 | shl r24=r23,r20 // r24: addresses for "fc" = | ||
93 | // "start" rounded down to stride | ||
94 | // boundary | ||
95 | .save ar.lc,r3 | ||
96 | mov r3=ar.lc // save ar.lc | ||
97 | ;; | ||
98 | |||
99 | .body | ||
100 | mov ar.lc=r8 | ||
101 | ;; | ||
102 | /* | ||
103 | * 32 byte aligned loop, even number of (actually 2) bundles | ||
104 | */ | ||
105 | .Loop_fc: | ||
106 | fc r24 // issuable on M0 only | ||
107 | add r24=r21,r24 // we flush "stride size" bytes per iteration | ||
108 | nop.i 0 | ||
109 | br.cloop.sptk.few .Loop_fc | ||
110 | ;; | ||
111 | sync.i | ||
112 | ;; | ||
113 | srlz.i | ||
114 | ;; | ||
115 | mov ar.lc=r3 // restore ar.lc | ||
116 | br.ret.sptk.many rp | ||
117 | END(clflush_cache_range) | ||