diff options
author | Stephen Hemminger <stephen@networkplumber.org> | 2013-12-29 15:12:29 -0500 |
---|---|---|
committer | Marcelo Tosatti <mtosatti@redhat.com> | 2014-01-08 16:02:58 -0500 |
commit | 7940876e1330671708186ac3386aa521ffb5c182 (patch) | |
tree | abe06165870e77111ba9e3d1b5f900c1cd8c4cef /virt/kvm/kvm_main.c | |
parent | 2f0a6397dd3cac2fb05b46cad08c1d532c04d6b8 (diff) |
kvm: make local functions static
Running 'make namespacecheck' found lots of functions that
should be declared static, since only used in one file.
Signed-off-by: Stephen Hemminger <stephen@networkplumber.org>
Signed-off-by: Marcelo Tosatti <mtosatti@redhat.com>
Diffstat (limited to 'virt/kvm/kvm_main.c')
-rw-r--r-- | virt/kvm/kvm_main.c | 35 |
1 files changed, 18 insertions, 17 deletions
diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c index 3efba97bdce2..e7c6ddd8ecc0 100644 --- a/virt/kvm/kvm_main.c +++ b/virt/kvm/kvm_main.c | |||
@@ -95,6 +95,12 @@ static int hardware_enable_all(void); | |||
95 | static void hardware_disable_all(void); | 95 | static void hardware_disable_all(void); |
96 | 96 | ||
97 | static void kvm_io_bus_destroy(struct kvm_io_bus *bus); | 97 | static void kvm_io_bus_destroy(struct kvm_io_bus *bus); |
98 | static void update_memslots(struct kvm_memslots *slots, | ||
99 | struct kvm_memory_slot *new, u64 last_generation); | ||
100 | |||
101 | static void kvm_release_pfn_dirty(pfn_t pfn); | ||
102 | static void mark_page_dirty_in_slot(struct kvm *kvm, | ||
103 | struct kvm_memory_slot *memslot, gfn_t gfn); | ||
98 | 104 | ||
99 | bool kvm_rebooting; | 105 | bool kvm_rebooting; |
100 | EXPORT_SYMBOL_GPL(kvm_rebooting); | 106 | EXPORT_SYMBOL_GPL(kvm_rebooting); |
@@ -553,7 +559,7 @@ static void kvm_free_physmem_slot(struct kvm *kvm, struct kvm_memory_slot *free, | |||
553 | free->npages = 0; | 559 | free->npages = 0; |
554 | } | 560 | } |
555 | 561 | ||
556 | void kvm_free_physmem(struct kvm *kvm) | 562 | static void kvm_free_physmem(struct kvm *kvm) |
557 | { | 563 | { |
558 | struct kvm_memslots *slots = kvm->memslots; | 564 | struct kvm_memslots *slots = kvm->memslots; |
559 | struct kvm_memory_slot *memslot; | 565 | struct kvm_memory_slot *memslot; |
@@ -675,8 +681,9 @@ static void sort_memslots(struct kvm_memslots *slots) | |||
675 | slots->id_to_index[slots->memslots[i].id] = i; | 681 | slots->id_to_index[slots->memslots[i].id] = i; |
676 | } | 682 | } |
677 | 683 | ||
678 | void update_memslots(struct kvm_memslots *slots, struct kvm_memory_slot *new, | 684 | static void update_memslots(struct kvm_memslots *slots, |
679 | u64 last_generation) | 685 | struct kvm_memory_slot *new, |
686 | u64 last_generation) | ||
680 | { | 687 | { |
681 | if (new) { | 688 | if (new) { |
682 | int id = new->id; | 689 | int id = new->id; |
@@ -924,8 +931,8 @@ int kvm_set_memory_region(struct kvm *kvm, | |||
924 | } | 931 | } |
925 | EXPORT_SYMBOL_GPL(kvm_set_memory_region); | 932 | EXPORT_SYMBOL_GPL(kvm_set_memory_region); |
926 | 933 | ||
927 | int kvm_vm_ioctl_set_memory_region(struct kvm *kvm, | 934 | static int kvm_vm_ioctl_set_memory_region(struct kvm *kvm, |
928 | struct kvm_userspace_memory_region *mem) | 935 | struct kvm_userspace_memory_region *mem) |
929 | { | 936 | { |
930 | if (mem->slot >= KVM_USER_MEM_SLOTS) | 937 | if (mem->slot >= KVM_USER_MEM_SLOTS) |
931 | return -EINVAL; | 938 | return -EINVAL; |
@@ -1047,7 +1054,7 @@ static unsigned long gfn_to_hva_many(struct kvm_memory_slot *slot, gfn_t gfn, | |||
1047 | } | 1054 | } |
1048 | 1055 | ||
1049 | unsigned long gfn_to_hva_memslot(struct kvm_memory_slot *slot, | 1056 | unsigned long gfn_to_hva_memslot(struct kvm_memory_slot *slot, |
1050 | gfn_t gfn) | 1057 | gfn_t gfn) |
1051 | { | 1058 | { |
1052 | return gfn_to_hva_many(slot, gfn, NULL); | 1059 | return gfn_to_hva_many(slot, gfn, NULL); |
1053 | } | 1060 | } |
@@ -1387,18 +1394,11 @@ void kvm_release_page_dirty(struct page *page) | |||
1387 | } | 1394 | } |
1388 | EXPORT_SYMBOL_GPL(kvm_release_page_dirty); | 1395 | EXPORT_SYMBOL_GPL(kvm_release_page_dirty); |
1389 | 1396 | ||
1390 | void kvm_release_pfn_dirty(pfn_t pfn) | 1397 | static void kvm_release_pfn_dirty(pfn_t pfn) |
1391 | { | 1398 | { |
1392 | kvm_set_pfn_dirty(pfn); | 1399 | kvm_set_pfn_dirty(pfn); |
1393 | kvm_release_pfn_clean(pfn); | 1400 | kvm_release_pfn_clean(pfn); |
1394 | } | 1401 | } |
1395 | EXPORT_SYMBOL_GPL(kvm_release_pfn_dirty); | ||
1396 | |||
1397 | void kvm_set_page_dirty(struct page *page) | ||
1398 | { | ||
1399 | kvm_set_pfn_dirty(page_to_pfn(page)); | ||
1400 | } | ||
1401 | EXPORT_SYMBOL_GPL(kvm_set_page_dirty); | ||
1402 | 1402 | ||
1403 | void kvm_set_pfn_dirty(pfn_t pfn) | 1403 | void kvm_set_pfn_dirty(pfn_t pfn) |
1404 | { | 1404 | { |
@@ -1640,8 +1640,9 @@ int kvm_clear_guest(struct kvm *kvm, gpa_t gpa, unsigned long len) | |||
1640 | } | 1640 | } |
1641 | EXPORT_SYMBOL_GPL(kvm_clear_guest); | 1641 | EXPORT_SYMBOL_GPL(kvm_clear_guest); |
1642 | 1642 | ||
1643 | void mark_page_dirty_in_slot(struct kvm *kvm, struct kvm_memory_slot *memslot, | 1643 | static void mark_page_dirty_in_slot(struct kvm *kvm, |
1644 | gfn_t gfn) | 1644 | struct kvm_memory_slot *memslot, |
1645 | gfn_t gfn) | ||
1645 | { | 1646 | { |
1646 | if (memslot && memslot->dirty_bitmap) { | 1647 | if (memslot && memslot->dirty_bitmap) { |
1647 | unsigned long rel_gfn = gfn - memslot->base_gfn; | 1648 | unsigned long rel_gfn = gfn - memslot->base_gfn; |
@@ -1757,7 +1758,7 @@ EXPORT_SYMBOL_GPL(kvm_vcpu_yield_to); | |||
1757 | * locking does not harm. It may result in trying to yield to same VCPU, fail | 1758 | * locking does not harm. It may result in trying to yield to same VCPU, fail |
1758 | * and continue with next VCPU and so on. | 1759 | * and continue with next VCPU and so on. |
1759 | */ | 1760 | */ |
1760 | bool kvm_vcpu_eligible_for_directed_yield(struct kvm_vcpu *vcpu) | 1761 | static bool kvm_vcpu_eligible_for_directed_yield(struct kvm_vcpu *vcpu) |
1761 | { | 1762 | { |
1762 | bool eligible; | 1763 | bool eligible; |
1763 | 1764 | ||