diff options
author | Joerg Roedel <joerg.roedel@amd.com> | 2008-09-25 06:13:54 -0400 |
---|---|---|
committer | Ingo Molnar <mingo@elte.hu> | 2008-09-27 14:13:46 -0400 |
commit | 3610f2116e961cdcbd3546a3828470f7aa636212 (patch) | |
tree | b3db81f6aba75bb3b0453862b1e3d6aa3045dc1f /arch/x86/kernel/pci-gart_64.c | |
parent | 237a62247c2879331986a300d6ab36ad21264c68 (diff) |
x86/iommu: convert GART need_flush to bool
Signed-off-by: Joerg Roedel <joerg.roedel@amd.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'arch/x86/kernel/pci-gart_64.c')
-rw-r--r-- | arch/x86/kernel/pci-gart_64.c | 10 |
1 files changed, 5 insertions, 5 deletions
diff --git a/arch/x86/kernel/pci-gart_64.c b/arch/x86/kernel/pci-gart_64.c index aa569db73a2..aecea068f58 100644 --- a/arch/x86/kernel/pci-gart_64.c +++ b/arch/x86/kernel/pci-gart_64.c | |||
@@ -80,7 +80,7 @@ AGPEXTERN int agp_memory_reserved; | |||
80 | AGPEXTERN __u32 *agp_gatt_table; | 80 | AGPEXTERN __u32 *agp_gatt_table; |
81 | 81 | ||
82 | static unsigned long next_bit; /* protected by iommu_bitmap_lock */ | 82 | static unsigned long next_bit; /* protected by iommu_bitmap_lock */ |
83 | static int need_flush; /* global flush state. set for each gart wrap */ | 83 | static bool need_flush; /* global flush state. set for each gart wrap */ |
84 | 84 | ||
85 | static unsigned long alloc_iommu(struct device *dev, int size, | 85 | static unsigned long alloc_iommu(struct device *dev, int size, |
86 | unsigned long align_mask) | 86 | unsigned long align_mask) |
@@ -98,7 +98,7 @@ static unsigned long alloc_iommu(struct device *dev, int size, | |||
98 | offset = iommu_area_alloc(iommu_gart_bitmap, iommu_pages, next_bit, | 98 | offset = iommu_area_alloc(iommu_gart_bitmap, iommu_pages, next_bit, |
99 | size, base_index, boundary_size, align_mask); | 99 | size, base_index, boundary_size, align_mask); |
100 | if (offset == -1) { | 100 | if (offset == -1) { |
101 | need_flush = 1; | 101 | need_flush = true; |
102 | offset = iommu_area_alloc(iommu_gart_bitmap, iommu_pages, 0, | 102 | offset = iommu_area_alloc(iommu_gart_bitmap, iommu_pages, 0, |
103 | size, base_index, boundary_size, | 103 | size, base_index, boundary_size, |
104 | align_mask); | 104 | align_mask); |
@@ -107,11 +107,11 @@ static unsigned long alloc_iommu(struct device *dev, int size, | |||
107 | next_bit = offset+size; | 107 | next_bit = offset+size; |
108 | if (next_bit >= iommu_pages) { | 108 | if (next_bit >= iommu_pages) { |
109 | next_bit = 0; | 109 | next_bit = 0; |
110 | need_flush = 1; | 110 | need_flush = true; |
111 | } | 111 | } |
112 | } | 112 | } |
113 | if (iommu_fullflush) | 113 | if (iommu_fullflush) |
114 | need_flush = 1; | 114 | need_flush = true; |
115 | spin_unlock_irqrestore(&iommu_bitmap_lock, flags); | 115 | spin_unlock_irqrestore(&iommu_bitmap_lock, flags); |
116 | 116 | ||
117 | return offset; | 117 | return offset; |
@@ -136,7 +136,7 @@ static void flush_gart(void) | |||
136 | spin_lock_irqsave(&iommu_bitmap_lock, flags); | 136 | spin_lock_irqsave(&iommu_bitmap_lock, flags); |
137 | if (need_flush) { | 137 | if (need_flush) { |
138 | k8_flush_garts(); | 138 | k8_flush_garts(); |
139 | need_flush = 0; | 139 | need_flush = false; |
140 | } | 140 | } |
141 | spin_unlock_irqrestore(&iommu_bitmap_lock, flags); | 141 | spin_unlock_irqrestore(&iommu_bitmap_lock, flags); |
142 | } | 142 | } |