diff options
Diffstat (limited to 'drivers/xen/tmem.c')
-rw-r--r-- | drivers/xen/tmem.c | 87 |
1 files changed, 39 insertions, 48 deletions
diff --git a/drivers/xen/tmem.c b/drivers/xen/tmem.c index e3600be4e7fa..18e8bd8fa947 100644 --- a/drivers/xen/tmem.c +++ b/drivers/xen/tmem.c | |||
@@ -11,11 +11,7 @@ | |||
11 | #include <linux/init.h> | 11 | #include <linux/init.h> |
12 | #include <linux/pagemap.h> | 12 | #include <linux/pagemap.h> |
13 | #include <linux/cleancache.h> | 13 | #include <linux/cleancache.h> |
14 | |||
15 | /* temporary ifdef until include/linux/frontswap.h is upstream */ | ||
16 | #ifdef CONFIG_FRONTSWAP | ||
17 | #include <linux/frontswap.h> | 14 | #include <linux/frontswap.h> |
18 | #endif | ||
19 | 15 | ||
20 | #include <xen/xen.h> | 16 | #include <xen/xen.h> |
21 | #include <xen/interface/xen.h> | 17 | #include <xen/interface/xen.h> |
@@ -24,6 +20,34 @@ | |||
24 | #include <asm/xen/hypervisor.h> | 20 | #include <asm/xen/hypervisor.h> |
25 | #include <xen/tmem.h> | 21 | #include <xen/tmem.h> |
26 | 22 | ||
23 | #ifndef CONFIG_XEN_TMEM_MODULE | ||
24 | bool __read_mostly tmem_enabled = false; | ||
25 | |||
26 | static int __init enable_tmem(char *s) | ||
27 | { | ||
28 | tmem_enabled = true; | ||
29 | return 1; | ||
30 | } | ||
31 | __setup("tmem", enable_tmem); | ||
32 | #endif | ||
33 | |||
34 | #ifdef CONFIG_CLEANCACHE | ||
35 | static bool cleancache __read_mostly = true; | ||
36 | module_param(cleancache, bool, S_IRUGO); | ||
37 | static bool selfballooning __read_mostly = true; | ||
38 | module_param(selfballooning, bool, S_IRUGO); | ||
39 | #endif /* CONFIG_CLEANCACHE */ | ||
40 | |||
41 | #ifdef CONFIG_FRONTSWAP | ||
42 | static bool frontswap __read_mostly = true; | ||
43 | module_param(frontswap, bool, S_IRUGO); | ||
44 | #endif /* CONFIG_FRONTSWAP */ | ||
45 | |||
46 | #ifdef CONFIG_XEN_SELFBALLOONING | ||
47 | static bool selfshrinking __read_mostly = true; | ||
48 | module_param(selfshrinking, bool, S_IRUGO); | ||
49 | #endif /* CONFIG_XEN_SELFBALLOONING */ | ||
50 | |||
27 | #define TMEM_CONTROL 0 | 51 | #define TMEM_CONTROL 0 |
28 | #define TMEM_NEW_POOL 1 | 52 | #define TMEM_NEW_POOL 1 |
29 | #define TMEM_DESTROY_POOL 2 | 53 | #define TMEM_DESTROY_POOL 2 |
@@ -129,16 +153,6 @@ static int xen_tmem_flush_object(u32 pool_id, struct tmem_oid oid) | |||
129 | return xen_tmem_op(TMEM_FLUSH_OBJECT, pool_id, oid, 0, 0, 0, 0, 0); | 153 | return xen_tmem_op(TMEM_FLUSH_OBJECT, pool_id, oid, 0, 0, 0, 0, 0); |
130 | } | 154 | } |
131 | 155 | ||
132 | #ifndef CONFIG_XEN_TMEM_MODULE | ||
133 | bool __read_mostly tmem_enabled = false; | ||
134 | |||
135 | static int __init enable_tmem(char *s) | ||
136 | { | ||
137 | tmem_enabled = true; | ||
138 | return 1; | ||
139 | } | ||
140 | __setup("tmem", enable_tmem); | ||
141 | #endif | ||
142 | 156 | ||
143 | #ifdef CONFIG_CLEANCACHE | 157 | #ifdef CONFIG_CLEANCACHE |
144 | static int xen_tmem_destroy_pool(u32 pool_id) | 158 | static int xen_tmem_destroy_pool(u32 pool_id) |
@@ -230,20 +244,6 @@ static int tmem_cleancache_init_shared_fs(char *uuid, size_t pagesize) | |||
230 | return xen_tmem_new_pool(shared_uuid, TMEM_POOL_SHARED, pagesize); | 244 | return xen_tmem_new_pool(shared_uuid, TMEM_POOL_SHARED, pagesize); |
231 | } | 245 | } |
232 | 246 | ||
233 | static bool disable_cleancache __read_mostly; | ||
234 | static bool disable_selfballooning __read_mostly; | ||
235 | #ifdef CONFIG_XEN_TMEM_MODULE | ||
236 | module_param(disable_cleancache, bool, S_IRUGO); | ||
237 | module_param(disable_selfballooning, bool, S_IRUGO); | ||
238 | #else | ||
239 | static int __init no_cleancache(char *s) | ||
240 | { | ||
241 | disable_cleancache = true; | ||
242 | return 1; | ||
243 | } | ||
244 | __setup("nocleancache", no_cleancache); | ||
245 | #endif | ||
246 | |||
247 | static struct cleancache_ops tmem_cleancache_ops = { | 247 | static struct cleancache_ops tmem_cleancache_ops = { |
248 | .put_page = tmem_cleancache_put_page, | 248 | .put_page = tmem_cleancache_put_page, |
249 | .get_page = tmem_cleancache_get_page, | 249 | .get_page = tmem_cleancache_get_page, |
@@ -361,20 +361,6 @@ static void tmem_frontswap_init(unsigned ignored) | |||
361 | xen_tmem_new_pool(private, TMEM_POOL_PERSIST, PAGE_SIZE); | 361 | xen_tmem_new_pool(private, TMEM_POOL_PERSIST, PAGE_SIZE); |
362 | } | 362 | } |
363 | 363 | ||
364 | static bool disable_frontswap __read_mostly; | ||
365 | static bool disable_frontswap_selfshrinking __read_mostly; | ||
366 | #ifdef CONFIG_XEN_TMEM_MODULE | ||
367 | module_param(disable_frontswap, bool, S_IRUGO); | ||
368 | module_param(disable_frontswap_selfshrinking, bool, S_IRUGO); | ||
369 | #else | ||
370 | static int __init no_frontswap(char *s) | ||
371 | { | ||
372 | disable_frontswap = true; | ||
373 | return 1; | ||
374 | } | ||
375 | __setup("nofrontswap", no_frontswap); | ||
376 | #endif | ||
377 | |||
378 | static struct frontswap_ops tmem_frontswap_ops = { | 364 | static struct frontswap_ops tmem_frontswap_ops = { |
379 | .store = tmem_frontswap_store, | 365 | .store = tmem_frontswap_store, |
380 | .load = tmem_frontswap_load, | 366 | .load = tmem_frontswap_load, |
@@ -382,8 +368,6 @@ static struct frontswap_ops tmem_frontswap_ops = { | |||
382 | .invalidate_area = tmem_frontswap_flush_area, | 368 | .invalidate_area = tmem_frontswap_flush_area, |
383 | .init = tmem_frontswap_init | 369 | .init = tmem_frontswap_init |
384 | }; | 370 | }; |
385 | #else /* CONFIG_FRONTSWAP */ | ||
386 | #define disable_frontswap_selfshrinking 1 | ||
387 | #endif | 371 | #endif |
388 | 372 | ||
389 | static int xen_tmem_init(void) | 373 | static int xen_tmem_init(void) |
@@ -391,7 +375,7 @@ static int xen_tmem_init(void) | |||
391 | if (!xen_domain()) | 375 | if (!xen_domain()) |
392 | return 0; | 376 | return 0; |
393 | #ifdef CONFIG_FRONTSWAP | 377 | #ifdef CONFIG_FRONTSWAP |
394 | if (tmem_enabled && !disable_frontswap) { | 378 | if (tmem_enabled && frontswap) { |
395 | char *s = ""; | 379 | char *s = ""; |
396 | struct frontswap_ops *old_ops = | 380 | struct frontswap_ops *old_ops = |
397 | frontswap_register_ops(&tmem_frontswap_ops); | 381 | frontswap_register_ops(&tmem_frontswap_ops); |
@@ -408,7 +392,7 @@ static int xen_tmem_init(void) | |||
408 | #endif | 392 | #endif |
409 | #ifdef CONFIG_CLEANCACHE | 393 | #ifdef CONFIG_CLEANCACHE |
410 | BUG_ON(sizeof(struct cleancache_filekey) != sizeof(struct tmem_oid)); | 394 | BUG_ON(sizeof(struct cleancache_filekey) != sizeof(struct tmem_oid)); |
411 | if (tmem_enabled && !disable_cleancache) { | 395 | if (tmem_enabled && cleancache) { |
412 | char *s = ""; | 396 | char *s = ""; |
413 | struct cleancache_ops *old_ops = | 397 | struct cleancache_ops *old_ops = |
414 | cleancache_register_ops(&tmem_cleancache_ops); | 398 | cleancache_register_ops(&tmem_cleancache_ops); |
@@ -419,8 +403,15 @@ static int xen_tmem_init(void) | |||
419 | } | 403 | } |
420 | #endif | 404 | #endif |
421 | #ifdef CONFIG_XEN_SELFBALLOONING | 405 | #ifdef CONFIG_XEN_SELFBALLOONING |
422 | xen_selfballoon_init(!disable_selfballooning, | 406 | /* |
423 | !disable_frontswap_selfshrinking); | 407 | * There is no point of driving pages to the swap system if they |
408 | * aren't going anywhere in tmem universe. | ||
409 | */ | ||
410 | if (!frontswap) { | ||
411 | selfshrinking = false; | ||
412 | selfballooning = false; | ||
413 | } | ||
414 | xen_selfballoon_init(selfballooning, selfshrinking); | ||
424 | #endif | 415 | #endif |
425 | return 0; | 416 | return 0; |
426 | } | 417 | } |