diff options
author | Minchan Kim <minchan@kernel.org> | 2016-07-26 18:23:28 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@linux-foundation.org> | 2016-07-26 19:19:19 -0400 |
commit | bfd093f5e7f09c1e41c43e7605893069975cd734 (patch) | |
tree | 0a0fa79efe9fcda15eaae41f4735e7afee01c55c /mm/zsmalloc.c | |
parent | 4aa409cab7c39c90f4b725ff22f52bbf5d2fc4e0 (diff) |
zsmalloc: use freeobj for index
Zsmalloc stores first free object's <PFN, obj_idx> position into freeobj
in each zspage. If we change it with index from first_page instead of
position, it makes page migration simple because we don't need to
correct other entries for linked list if a page is migrated out.
Link: http://lkml.kernel.org/r/1464736881-24886-11-git-send-email-minchan@kernel.org
Signed-off-by: Minchan Kim <minchan@kernel.org>
Cc: Sergey Senozhatsky <sergey.senozhatsky@gmail.com>
Cc: Vlastimil Babka <vbabka@suse.cz>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
Diffstat (limited to 'mm/zsmalloc.c')
-rw-r--r-- | mm/zsmalloc.c | 139 |
1 files changed, 73 insertions, 66 deletions
diff --git a/mm/zsmalloc.c b/mm/zsmalloc.c index dd3708611f65..c6fb543cfb98 100644 --- a/mm/zsmalloc.c +++ b/mm/zsmalloc.c | |||
@@ -71,9 +71,7 @@ | |||
71 | * Object location (<PFN>, <obj_idx>) is encoded as | 71 | * Object location (<PFN>, <obj_idx>) is encoded as |
72 | * as single (unsigned long) handle value. | 72 | * as single (unsigned long) handle value. |
73 | * | 73 | * |
74 | * Note that object index <obj_idx> is relative to system | 74 | * Note that object index <obj_idx> starts from 0. |
75 | * page <PFN> it is stored in, so for each sub-page belonging | ||
76 | * to a zspage, obj_idx starts with 0. | ||
77 | * | 75 | * |
78 | * This is made more complicated by various memory models and PAE. | 76 | * This is made more complicated by various memory models and PAE. |
79 | */ | 77 | */ |
@@ -214,10 +212,10 @@ struct size_class { | |||
214 | struct link_free { | 212 | struct link_free { |
215 | union { | 213 | union { |
216 | /* | 214 | /* |
217 | * Position of next free chunk (encodes <PFN, obj_idx>) | 215 | * Free object index; |
218 | * It's valid for non-allocated object | 216 | * It's valid for non-allocated object |
219 | */ | 217 | */ |
220 | void *next; | 218 | unsigned long next; |
221 | /* | 219 | /* |
222 | * Handle of allocated object. | 220 | * Handle of allocated object. |
223 | */ | 221 | */ |
@@ -261,7 +259,7 @@ struct zspage { | |||
261 | unsigned int class:CLASS_BITS; | 259 | unsigned int class:CLASS_BITS; |
262 | }; | 260 | }; |
263 | unsigned int inuse; | 261 | unsigned int inuse; |
264 | void *freeobj; | 262 | unsigned int freeobj; |
265 | struct page *first_page; | 263 | struct page *first_page; |
266 | struct list_head list; /* fullness list */ | 264 | struct list_head list; /* fullness list */ |
267 | }; | 265 | }; |
@@ -459,14 +457,14 @@ static inline void set_first_obj_offset(struct page *page, int offset) | |||
459 | page->index = offset; | 457 | page->index = offset; |
460 | } | 458 | } |
461 | 459 | ||
462 | static inline unsigned long get_freeobj(struct zspage *zspage) | 460 | static inline unsigned int get_freeobj(struct zspage *zspage) |
463 | { | 461 | { |
464 | return (unsigned long)zspage->freeobj; | 462 | return zspage->freeobj; |
465 | } | 463 | } |
466 | 464 | ||
467 | static inline void set_freeobj(struct zspage *zspage, unsigned long obj) | 465 | static inline void set_freeobj(struct zspage *zspage, unsigned int obj) |
468 | { | 466 | { |
469 | zspage->freeobj = (void *)obj; | 467 | zspage->freeobj = obj; |
470 | } | 468 | } |
471 | 469 | ||
472 | static void get_zspage_mapping(struct zspage *zspage, | 470 | static void get_zspage_mapping(struct zspage *zspage, |
@@ -810,6 +808,10 @@ static int get_pages_per_zspage(int class_size) | |||
810 | return max_usedpc_order; | 808 | return max_usedpc_order; |
811 | } | 809 | } |
812 | 810 | ||
811 | static struct page *get_first_page(struct zspage *zspage) | ||
812 | { | ||
813 | return zspage->first_page; | ||
814 | } | ||
813 | 815 | ||
814 | static struct zspage *get_zspage(struct page *page) | 816 | static struct zspage *get_zspage(struct page *page) |
815 | { | 817 | { |
@@ -821,37 +823,33 @@ static struct page *get_next_page(struct page *page) | |||
821 | return page->next; | 823 | return page->next; |
822 | } | 824 | } |
823 | 825 | ||
824 | /* | 826 | /** |
825 | * Encode <page, obj_idx> as a single handle value. | 827 | * obj_to_location - get (<page>, <obj_idx>) from encoded object value |
826 | * We use the least bit of handle for tagging. | 828 | * @page: page object resides in zspage |
829 | * @obj_idx: object index | ||
827 | */ | 830 | */ |
828 | static void *location_to_obj(struct page *page, unsigned long obj_idx) | 831 | static void obj_to_location(unsigned long obj, struct page **page, |
832 | unsigned int *obj_idx) | ||
829 | { | 833 | { |
830 | unsigned long obj; | 834 | obj >>= OBJ_TAG_BITS; |
835 | *page = pfn_to_page(obj >> OBJ_INDEX_BITS); | ||
836 | *obj_idx = (obj & OBJ_INDEX_MASK); | ||
837 | } | ||
831 | 838 | ||
832 | if (!page) { | 839 | /** |
833 | VM_BUG_ON(obj_idx); | 840 | * location_to_obj - get obj value encoded from (<page>, <obj_idx>) |
834 | return NULL; | 841 | * @page: page object resides in zspage |
835 | } | 842 | * @obj_idx: object index |
843 | */ | ||
844 | static unsigned long location_to_obj(struct page *page, unsigned int obj_idx) | ||
845 | { | ||
846 | unsigned long obj; | ||
836 | 847 | ||
837 | obj = page_to_pfn(page) << OBJ_INDEX_BITS; | 848 | obj = page_to_pfn(page) << OBJ_INDEX_BITS; |
838 | obj |= ((obj_idx) & OBJ_INDEX_MASK); | 849 | obj |= obj_idx & OBJ_INDEX_MASK; |
839 | obj <<= OBJ_TAG_BITS; | 850 | obj <<= OBJ_TAG_BITS; |
840 | 851 | ||
841 | return (void *)obj; | 852 | return obj; |
842 | } | ||
843 | |||
844 | /* | ||
845 | * Decode <page, obj_idx> pair from the given object handle. We adjust the | ||
846 | * decoded obj_idx back to its original value since it was adjusted in | ||
847 | * location_to_obj(). | ||
848 | */ | ||
849 | static void obj_to_location(unsigned long obj, struct page **page, | ||
850 | unsigned long *obj_idx) | ||
851 | { | ||
852 | obj >>= OBJ_TAG_BITS; | ||
853 | *page = pfn_to_page(obj >> OBJ_INDEX_BITS); | ||
854 | *obj_idx = (obj & OBJ_INDEX_MASK); | ||
855 | } | 853 | } |
856 | 854 | ||
857 | static unsigned long handle_to_obj(unsigned long handle) | 855 | static unsigned long handle_to_obj(unsigned long handle) |
@@ -869,16 +867,6 @@ static unsigned long obj_to_head(struct size_class *class, struct page *page, | |||
869 | return *(unsigned long *)obj; | 867 | return *(unsigned long *)obj; |
870 | } | 868 | } |
871 | 869 | ||
872 | static unsigned long obj_idx_to_offset(struct page *page, | ||
873 | unsigned long obj_idx, int class_size) | ||
874 | { | ||
875 | unsigned long off; | ||
876 | |||
877 | off = get_first_obj_offset(page); | ||
878 | |||
879 | return off + obj_idx * class_size; | ||
880 | } | ||
881 | |||
882 | static inline int trypin_tag(unsigned long handle) | 870 | static inline int trypin_tag(unsigned long handle) |
883 | { | 871 | { |
884 | return bit_spin_trylock(HANDLE_PIN_BIT, (unsigned long *)handle); | 872 | return bit_spin_trylock(HANDLE_PIN_BIT, (unsigned long *)handle); |
@@ -922,13 +910,13 @@ static void free_zspage(struct zs_pool *pool, struct zspage *zspage) | |||
922 | /* Initialize a newly allocated zspage */ | 910 | /* Initialize a newly allocated zspage */ |
923 | static void init_zspage(struct size_class *class, struct zspage *zspage) | 911 | static void init_zspage(struct size_class *class, struct zspage *zspage) |
924 | { | 912 | { |
913 | unsigned int freeobj = 1; | ||
925 | unsigned long off = 0; | 914 | unsigned long off = 0; |
926 | struct page *page = zspage->first_page; | 915 | struct page *page = zspage->first_page; |
927 | 916 | ||
928 | while (page) { | 917 | while (page) { |
929 | struct page *next_page; | 918 | struct page *next_page; |
930 | struct link_free *link; | 919 | struct link_free *link; |
931 | unsigned int i = 1; | ||
932 | void *vaddr; | 920 | void *vaddr; |
933 | 921 | ||
934 | set_first_obj_offset(page, off); | 922 | set_first_obj_offset(page, off); |
@@ -937,7 +925,7 @@ static void init_zspage(struct size_class *class, struct zspage *zspage) | |||
937 | link = (struct link_free *)vaddr + off / sizeof(*link); | 925 | link = (struct link_free *)vaddr + off / sizeof(*link); |
938 | 926 | ||
939 | while ((off += class->size) < PAGE_SIZE) { | 927 | while ((off += class->size) < PAGE_SIZE) { |
940 | link->next = location_to_obj(page, i++); | 928 | link->next = freeobj++ << OBJ_ALLOCATED_TAG; |
941 | link += class->size / sizeof(*link); | 929 | link += class->size / sizeof(*link); |
942 | } | 930 | } |
943 | 931 | ||
@@ -947,14 +935,21 @@ static void init_zspage(struct size_class *class, struct zspage *zspage) | |||
947 | * page (if present) | 935 | * page (if present) |
948 | */ | 936 | */ |
949 | next_page = get_next_page(page); | 937 | next_page = get_next_page(page); |
950 | link->next = location_to_obj(next_page, 0); | 938 | if (next_page) { |
939 | link->next = freeobj++ << OBJ_ALLOCATED_TAG; | ||
940 | } else { | ||
941 | /* | ||
942 | * Reset OBJ_ALLOCATED_TAG bit to last link to tell | ||
943 | * whether it's allocated object or not. | ||
944 | */ | ||
945 | link->next = -1 << OBJ_ALLOCATED_TAG; | ||
946 | } | ||
951 | kunmap_atomic(vaddr); | 947 | kunmap_atomic(vaddr); |
952 | page = next_page; | 948 | page = next_page; |
953 | off %= PAGE_SIZE; | 949 | off %= PAGE_SIZE; |
954 | } | 950 | } |
955 | 951 | ||
956 | set_freeobj(zspage, | 952 | set_freeobj(zspage, 0); |
957 | (unsigned long)location_to_obj(zspage->first_page, 0)); | ||
958 | } | 953 | } |
959 | 954 | ||
960 | static void create_page_chain(struct zspage *zspage, struct page *pages[], | 955 | static void create_page_chain(struct zspage *zspage, struct page *pages[], |
@@ -1271,7 +1266,8 @@ void *zs_map_object(struct zs_pool *pool, unsigned long handle, | |||
1271 | { | 1266 | { |
1272 | struct zspage *zspage; | 1267 | struct zspage *zspage; |
1273 | struct page *page; | 1268 | struct page *page; |
1274 | unsigned long obj, obj_idx, off; | 1269 | unsigned long obj, off; |
1270 | unsigned int obj_idx; | ||
1275 | 1271 | ||
1276 | unsigned int class_idx; | 1272 | unsigned int class_idx; |
1277 | enum fullness_group fg; | 1273 | enum fullness_group fg; |
@@ -1295,7 +1291,7 @@ void *zs_map_object(struct zs_pool *pool, unsigned long handle, | |||
1295 | zspage = get_zspage(page); | 1291 | zspage = get_zspage(page); |
1296 | get_zspage_mapping(zspage, &class_idx, &fg); | 1292 | get_zspage_mapping(zspage, &class_idx, &fg); |
1297 | class = pool->size_class[class_idx]; | 1293 | class = pool->size_class[class_idx]; |
1298 | off = obj_idx_to_offset(page, obj_idx, class->size); | 1294 | off = (class->size * obj_idx) & ~PAGE_MASK; |
1299 | 1295 | ||
1300 | area = &get_cpu_var(zs_map_area); | 1296 | area = &get_cpu_var(zs_map_area); |
1301 | area->vm_mm = mm; | 1297 | area->vm_mm = mm; |
@@ -1324,7 +1320,8 @@ void zs_unmap_object(struct zs_pool *pool, unsigned long handle) | |||
1324 | { | 1320 | { |
1325 | struct zspage *zspage; | 1321 | struct zspage *zspage; |
1326 | struct page *page; | 1322 | struct page *page; |
1327 | unsigned long obj, obj_idx, off; | 1323 | unsigned long obj, off; |
1324 | unsigned int obj_idx; | ||
1328 | 1325 | ||
1329 | unsigned int class_idx; | 1326 | unsigned int class_idx; |
1330 | enum fullness_group fg; | 1327 | enum fullness_group fg; |
@@ -1336,7 +1333,7 @@ void zs_unmap_object(struct zs_pool *pool, unsigned long handle) | |||
1336 | zspage = get_zspage(page); | 1333 | zspage = get_zspage(page); |
1337 | get_zspage_mapping(zspage, &class_idx, &fg); | 1334 | get_zspage_mapping(zspage, &class_idx, &fg); |
1338 | class = pool->size_class[class_idx]; | 1335 | class = pool->size_class[class_idx]; |
1339 | off = obj_idx_to_offset(page, obj_idx, class->size); | 1336 | off = (class->size * obj_idx) & ~PAGE_MASK; |
1340 | 1337 | ||
1341 | area = this_cpu_ptr(&zs_map_area); | 1338 | area = this_cpu_ptr(&zs_map_area); |
1342 | if (off + class->size <= PAGE_SIZE) | 1339 | if (off + class->size <= PAGE_SIZE) |
@@ -1358,21 +1355,28 @@ EXPORT_SYMBOL_GPL(zs_unmap_object); | |||
1358 | static unsigned long obj_malloc(struct size_class *class, | 1355 | static unsigned long obj_malloc(struct size_class *class, |
1359 | struct zspage *zspage, unsigned long handle) | 1356 | struct zspage *zspage, unsigned long handle) |
1360 | { | 1357 | { |
1358 | int i, nr_page, offset; | ||
1361 | unsigned long obj; | 1359 | unsigned long obj; |
1362 | struct link_free *link; | 1360 | struct link_free *link; |
1363 | 1361 | ||
1364 | struct page *m_page; | 1362 | struct page *m_page; |
1365 | unsigned long m_objidx, m_offset; | 1363 | unsigned long m_offset; |
1366 | void *vaddr; | 1364 | void *vaddr; |
1367 | 1365 | ||
1368 | handle |= OBJ_ALLOCATED_TAG; | 1366 | handle |= OBJ_ALLOCATED_TAG; |
1369 | obj = get_freeobj(zspage); | 1367 | obj = get_freeobj(zspage); |
1370 | obj_to_location(obj, &m_page, &m_objidx); | 1368 | |
1371 | m_offset = obj_idx_to_offset(m_page, m_objidx, class->size); | 1369 | offset = obj * class->size; |
1370 | nr_page = offset >> PAGE_SHIFT; | ||
1371 | m_offset = offset & ~PAGE_MASK; | ||
1372 | m_page = get_first_page(zspage); | ||
1373 | |||
1374 | for (i = 0; i < nr_page; i++) | ||
1375 | m_page = get_next_page(m_page); | ||
1372 | 1376 | ||
1373 | vaddr = kmap_atomic(m_page); | 1377 | vaddr = kmap_atomic(m_page); |
1374 | link = (struct link_free *)vaddr + m_offset / sizeof(*link); | 1378 | link = (struct link_free *)vaddr + m_offset / sizeof(*link); |
1375 | set_freeobj(zspage, (unsigned long)link->next); | 1379 | set_freeobj(zspage, link->next >> OBJ_ALLOCATED_TAG); |
1376 | if (!class->huge) | 1380 | if (!class->huge) |
1377 | /* record handle in the header of allocated chunk */ | 1381 | /* record handle in the header of allocated chunk */ |
1378 | link->handle = handle; | 1382 | link->handle = handle; |
@@ -1384,6 +1388,8 @@ static unsigned long obj_malloc(struct size_class *class, | |||
1384 | mod_zspage_inuse(zspage, 1); | 1388 | mod_zspage_inuse(zspage, 1); |
1385 | zs_stat_inc(class, OBJ_USED, 1); | 1389 | zs_stat_inc(class, OBJ_USED, 1); |
1386 | 1390 | ||
1391 | obj = location_to_obj(m_page, obj); | ||
1392 | |||
1387 | return obj; | 1393 | return obj; |
1388 | } | 1394 | } |
1389 | 1395 | ||
@@ -1449,22 +1455,22 @@ static void obj_free(struct size_class *class, unsigned long obj) | |||
1449 | struct link_free *link; | 1455 | struct link_free *link; |
1450 | struct zspage *zspage; | 1456 | struct zspage *zspage; |
1451 | struct page *f_page; | 1457 | struct page *f_page; |
1452 | unsigned long f_objidx, f_offset; | 1458 | unsigned long f_offset; |
1459 | unsigned int f_objidx; | ||
1453 | void *vaddr; | 1460 | void *vaddr; |
1454 | 1461 | ||
1455 | obj &= ~OBJ_ALLOCATED_TAG; | 1462 | obj &= ~OBJ_ALLOCATED_TAG; |
1456 | obj_to_location(obj, &f_page, &f_objidx); | 1463 | obj_to_location(obj, &f_page, &f_objidx); |
1464 | f_offset = (class->size * f_objidx) & ~PAGE_MASK; | ||
1457 | zspage = get_zspage(f_page); | 1465 | zspage = get_zspage(f_page); |
1458 | 1466 | ||
1459 | f_offset = obj_idx_to_offset(f_page, f_objidx, class->size); | ||
1460 | |||
1461 | vaddr = kmap_atomic(f_page); | 1467 | vaddr = kmap_atomic(f_page); |
1462 | 1468 | ||
1463 | /* Insert this object in containing zspage's freelist */ | 1469 | /* Insert this object in containing zspage's freelist */ |
1464 | link = (struct link_free *)(vaddr + f_offset); | 1470 | link = (struct link_free *)(vaddr + f_offset); |
1465 | link->next = (void *)get_freeobj(zspage); | 1471 | link->next = get_freeobj(zspage) << OBJ_ALLOCATED_TAG; |
1466 | kunmap_atomic(vaddr); | 1472 | kunmap_atomic(vaddr); |
1467 | set_freeobj(zspage, obj); | 1473 | set_freeobj(zspage, f_objidx); |
1468 | mod_zspage_inuse(zspage, -1); | 1474 | mod_zspage_inuse(zspage, -1); |
1469 | zs_stat_dec(class, OBJ_USED, 1); | 1475 | zs_stat_dec(class, OBJ_USED, 1); |
1470 | } | 1476 | } |
@@ -1473,7 +1479,8 @@ void zs_free(struct zs_pool *pool, unsigned long handle) | |||
1473 | { | 1479 | { |
1474 | struct zspage *zspage; | 1480 | struct zspage *zspage; |
1475 | struct page *f_page; | 1481 | struct page *f_page; |
1476 | unsigned long obj, f_objidx; | 1482 | unsigned long obj; |
1483 | unsigned int f_objidx; | ||
1477 | int class_idx; | 1484 | int class_idx; |
1478 | struct size_class *class; | 1485 | struct size_class *class; |
1479 | enum fullness_group fullness; | 1486 | enum fullness_group fullness; |
@@ -1510,7 +1517,7 @@ static void zs_object_copy(struct size_class *class, unsigned long dst, | |||
1510 | unsigned long src) | 1517 | unsigned long src) |
1511 | { | 1518 | { |
1512 | struct page *s_page, *d_page; | 1519 | struct page *s_page, *d_page; |
1513 | unsigned long s_objidx, d_objidx; | 1520 | unsigned int s_objidx, d_objidx; |
1514 | unsigned long s_off, d_off; | 1521 | unsigned long s_off, d_off; |
1515 | void *s_addr, *d_addr; | 1522 | void *s_addr, *d_addr; |
1516 | int s_size, d_size, size; | 1523 | int s_size, d_size, size; |
@@ -1521,8 +1528,8 @@ static void zs_object_copy(struct size_class *class, unsigned long dst, | |||
1521 | obj_to_location(src, &s_page, &s_objidx); | 1528 | obj_to_location(src, &s_page, &s_objidx); |
1522 | obj_to_location(dst, &d_page, &d_objidx); | 1529 | obj_to_location(dst, &d_page, &d_objidx); |
1523 | 1530 | ||
1524 | s_off = obj_idx_to_offset(s_page, s_objidx, class->size); | 1531 | s_off = (class->size * s_objidx) & ~PAGE_MASK; |
1525 | d_off = obj_idx_to_offset(d_page, d_objidx, class->size); | 1532 | d_off = (class->size * d_objidx) & ~PAGE_MASK; |
1526 | 1533 | ||
1527 | if (s_off + class->size > PAGE_SIZE) | 1534 | if (s_off + class->size > PAGE_SIZE) |
1528 | s_size = PAGE_SIZE - s_off; | 1535 | s_size = PAGE_SIZE - s_off; |