aboutsummaryrefslogtreecommitdiffstats
path: root/mm/zsmalloc.c
diff options
context:
space:
mode:
authorMinchan Kim <minchan@kernel.org>2016-07-26 18:23:28 -0400
committerLinus Torvalds <torvalds@linux-foundation.org>2016-07-26 19:19:19 -0400
commitbfd093f5e7f09c1e41c43e7605893069975cd734 (patch)
tree0a0fa79efe9fcda15eaae41f4735e7afee01c55c /mm/zsmalloc.c
parent4aa409cab7c39c90f4b725ff22f52bbf5d2fc4e0 (diff)
zsmalloc: use freeobj for index
Zsmalloc stores first free object's <PFN, obj_idx> position into freeobj in each zspage. If we change it with index from first_page instead of position, it makes page migration simple because we don't need to correct other entries for linked list if a page is migrated out. Link: http://lkml.kernel.org/r/1464736881-24886-11-git-send-email-minchan@kernel.org Signed-off-by: Minchan Kim <minchan@kernel.org> Cc: Sergey Senozhatsky <sergey.senozhatsky@gmail.com> Cc: Vlastimil Babka <vbabka@suse.cz> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
Diffstat (limited to 'mm/zsmalloc.c')
-rw-r--r--mm/zsmalloc.c139
1 files changed, 73 insertions, 66 deletions
diff --git a/mm/zsmalloc.c b/mm/zsmalloc.c
index dd3708611f65..c6fb543cfb98 100644
--- a/mm/zsmalloc.c
+++ b/mm/zsmalloc.c
@@ -71,9 +71,7 @@
71 * Object location (<PFN>, <obj_idx>) is encoded as 71 * Object location (<PFN>, <obj_idx>) is encoded as
72 * as single (unsigned long) handle value. 72 * as single (unsigned long) handle value.
73 * 73 *
74 * Note that object index <obj_idx> is relative to system 74 * Note that object index <obj_idx> starts from 0.
75 * page <PFN> it is stored in, so for each sub-page belonging
76 * to a zspage, obj_idx starts with 0.
77 * 75 *
78 * This is made more complicated by various memory models and PAE. 76 * This is made more complicated by various memory models and PAE.
79 */ 77 */
@@ -214,10 +212,10 @@ struct size_class {
214struct link_free { 212struct link_free {
215 union { 213 union {
216 /* 214 /*
217 * Position of next free chunk (encodes <PFN, obj_idx>) 215 * Free object index;
218 * It's valid for non-allocated object 216 * It's valid for non-allocated object
219 */ 217 */
220 void *next; 218 unsigned long next;
221 /* 219 /*
222 * Handle of allocated object. 220 * Handle of allocated object.
223 */ 221 */
@@ -261,7 +259,7 @@ struct zspage {
261 unsigned int class:CLASS_BITS; 259 unsigned int class:CLASS_BITS;
262 }; 260 };
263 unsigned int inuse; 261 unsigned int inuse;
264 void *freeobj; 262 unsigned int freeobj;
265 struct page *first_page; 263 struct page *first_page;
266 struct list_head list; /* fullness list */ 264 struct list_head list; /* fullness list */
267}; 265};
@@ -459,14 +457,14 @@ static inline void set_first_obj_offset(struct page *page, int offset)
459 page->index = offset; 457 page->index = offset;
460} 458}
461 459
462static inline unsigned long get_freeobj(struct zspage *zspage) 460static inline unsigned int get_freeobj(struct zspage *zspage)
463{ 461{
464 return (unsigned long)zspage->freeobj; 462 return zspage->freeobj;
465} 463}
466 464
467static inline void set_freeobj(struct zspage *zspage, unsigned long obj) 465static inline void set_freeobj(struct zspage *zspage, unsigned int obj)
468{ 466{
469 zspage->freeobj = (void *)obj; 467 zspage->freeobj = obj;
470} 468}
471 469
472static void get_zspage_mapping(struct zspage *zspage, 470static void get_zspage_mapping(struct zspage *zspage,
@@ -810,6 +808,10 @@ static int get_pages_per_zspage(int class_size)
810 return max_usedpc_order; 808 return max_usedpc_order;
811} 809}
812 810
811static struct page *get_first_page(struct zspage *zspage)
812{
813 return zspage->first_page;
814}
813 815
814static struct zspage *get_zspage(struct page *page) 816static struct zspage *get_zspage(struct page *page)
815{ 817{
@@ -821,37 +823,33 @@ static struct page *get_next_page(struct page *page)
821 return page->next; 823 return page->next;
822} 824}
823 825
824/* 826/**
825 * Encode <page, obj_idx> as a single handle value. 827 * obj_to_location - get (<page>, <obj_idx>) from encoded object value
826 * We use the least bit of handle for tagging. 828 * @page: page object resides in zspage
829 * @obj_idx: object index
827 */ 830 */
828static void *location_to_obj(struct page *page, unsigned long obj_idx) 831static void obj_to_location(unsigned long obj, struct page **page,
832 unsigned int *obj_idx)
829{ 833{
830 unsigned long obj; 834 obj >>= OBJ_TAG_BITS;
835 *page = pfn_to_page(obj >> OBJ_INDEX_BITS);
836 *obj_idx = (obj & OBJ_INDEX_MASK);
837}
831 838
832 if (!page) { 839/**
833 VM_BUG_ON(obj_idx); 840 * location_to_obj - get obj value encoded from (<page>, <obj_idx>)
834 return NULL; 841 * @page: page object resides in zspage
835 } 842 * @obj_idx: object index
843 */
844static unsigned long location_to_obj(struct page *page, unsigned int obj_idx)
845{
846 unsigned long obj;
836 847
837 obj = page_to_pfn(page) << OBJ_INDEX_BITS; 848 obj = page_to_pfn(page) << OBJ_INDEX_BITS;
838 obj |= ((obj_idx) & OBJ_INDEX_MASK); 849 obj |= obj_idx & OBJ_INDEX_MASK;
839 obj <<= OBJ_TAG_BITS; 850 obj <<= OBJ_TAG_BITS;
840 851
841 return (void *)obj; 852 return obj;
842}
843
844/*
845 * Decode <page, obj_idx> pair from the given object handle. We adjust the
846 * decoded obj_idx back to its original value since it was adjusted in
847 * location_to_obj().
848 */
849static void obj_to_location(unsigned long obj, struct page **page,
850 unsigned long *obj_idx)
851{
852 obj >>= OBJ_TAG_BITS;
853 *page = pfn_to_page(obj >> OBJ_INDEX_BITS);
854 *obj_idx = (obj & OBJ_INDEX_MASK);
855} 853}
856 854
857static unsigned long handle_to_obj(unsigned long handle) 855static unsigned long handle_to_obj(unsigned long handle)
@@ -869,16 +867,6 @@ static unsigned long obj_to_head(struct size_class *class, struct page *page,
869 return *(unsigned long *)obj; 867 return *(unsigned long *)obj;
870} 868}
871 869
872static unsigned long obj_idx_to_offset(struct page *page,
873 unsigned long obj_idx, int class_size)
874{
875 unsigned long off;
876
877 off = get_first_obj_offset(page);
878
879 return off + obj_idx * class_size;
880}
881
882static inline int trypin_tag(unsigned long handle) 870static inline int trypin_tag(unsigned long handle)
883{ 871{
884 return bit_spin_trylock(HANDLE_PIN_BIT, (unsigned long *)handle); 872 return bit_spin_trylock(HANDLE_PIN_BIT, (unsigned long *)handle);
@@ -922,13 +910,13 @@ static void free_zspage(struct zs_pool *pool, struct zspage *zspage)
922/* Initialize a newly allocated zspage */ 910/* Initialize a newly allocated zspage */
923static void init_zspage(struct size_class *class, struct zspage *zspage) 911static void init_zspage(struct size_class *class, struct zspage *zspage)
924{ 912{
913 unsigned int freeobj = 1;
925 unsigned long off = 0; 914 unsigned long off = 0;
926 struct page *page = zspage->first_page; 915 struct page *page = zspage->first_page;
927 916
928 while (page) { 917 while (page) {
929 struct page *next_page; 918 struct page *next_page;
930 struct link_free *link; 919 struct link_free *link;
931 unsigned int i = 1;
932 void *vaddr; 920 void *vaddr;
933 921
934 set_first_obj_offset(page, off); 922 set_first_obj_offset(page, off);
@@ -937,7 +925,7 @@ static void init_zspage(struct size_class *class, struct zspage *zspage)
937 link = (struct link_free *)vaddr + off / sizeof(*link); 925 link = (struct link_free *)vaddr + off / sizeof(*link);
938 926
939 while ((off += class->size) < PAGE_SIZE) { 927 while ((off += class->size) < PAGE_SIZE) {
940 link->next = location_to_obj(page, i++); 928 link->next = freeobj++ << OBJ_ALLOCATED_TAG;
941 link += class->size / sizeof(*link); 929 link += class->size / sizeof(*link);
942 } 930 }
943 931
@@ -947,14 +935,21 @@ static void init_zspage(struct size_class *class, struct zspage *zspage)
947 * page (if present) 935 * page (if present)
948 */ 936 */
949 next_page = get_next_page(page); 937 next_page = get_next_page(page);
950 link->next = location_to_obj(next_page, 0); 938 if (next_page) {
939 link->next = freeobj++ << OBJ_ALLOCATED_TAG;
940 } else {
941 /*
942 * Reset OBJ_ALLOCATED_TAG bit to last link to tell
943 * whether it's allocated object or not.
944 */
945 link->next = -1 << OBJ_ALLOCATED_TAG;
946 }
951 kunmap_atomic(vaddr); 947 kunmap_atomic(vaddr);
952 page = next_page; 948 page = next_page;
953 off %= PAGE_SIZE; 949 off %= PAGE_SIZE;
954 } 950 }
955 951
956 set_freeobj(zspage, 952 set_freeobj(zspage, 0);
957 (unsigned long)location_to_obj(zspage->first_page, 0));
958} 953}
959 954
960static void create_page_chain(struct zspage *zspage, struct page *pages[], 955static void create_page_chain(struct zspage *zspage, struct page *pages[],
@@ -1271,7 +1266,8 @@ void *zs_map_object(struct zs_pool *pool, unsigned long handle,
1271{ 1266{
1272 struct zspage *zspage; 1267 struct zspage *zspage;
1273 struct page *page; 1268 struct page *page;
1274 unsigned long obj, obj_idx, off; 1269 unsigned long obj, off;
1270 unsigned int obj_idx;
1275 1271
1276 unsigned int class_idx; 1272 unsigned int class_idx;
1277 enum fullness_group fg; 1273 enum fullness_group fg;
@@ -1295,7 +1291,7 @@ void *zs_map_object(struct zs_pool *pool, unsigned long handle,
1295 zspage = get_zspage(page); 1291 zspage = get_zspage(page);
1296 get_zspage_mapping(zspage, &class_idx, &fg); 1292 get_zspage_mapping(zspage, &class_idx, &fg);
1297 class = pool->size_class[class_idx]; 1293 class = pool->size_class[class_idx];
1298 off = obj_idx_to_offset(page, obj_idx, class->size); 1294 off = (class->size * obj_idx) & ~PAGE_MASK;
1299 1295
1300 area = &get_cpu_var(zs_map_area); 1296 area = &get_cpu_var(zs_map_area);
1301 area->vm_mm = mm; 1297 area->vm_mm = mm;
@@ -1324,7 +1320,8 @@ void zs_unmap_object(struct zs_pool *pool, unsigned long handle)
1324{ 1320{
1325 struct zspage *zspage; 1321 struct zspage *zspage;
1326 struct page *page; 1322 struct page *page;
1327 unsigned long obj, obj_idx, off; 1323 unsigned long obj, off;
1324 unsigned int obj_idx;
1328 1325
1329 unsigned int class_idx; 1326 unsigned int class_idx;
1330 enum fullness_group fg; 1327 enum fullness_group fg;
@@ -1336,7 +1333,7 @@ void zs_unmap_object(struct zs_pool *pool, unsigned long handle)
1336 zspage = get_zspage(page); 1333 zspage = get_zspage(page);
1337 get_zspage_mapping(zspage, &class_idx, &fg); 1334 get_zspage_mapping(zspage, &class_idx, &fg);
1338 class = pool->size_class[class_idx]; 1335 class = pool->size_class[class_idx];
1339 off = obj_idx_to_offset(page, obj_idx, class->size); 1336 off = (class->size * obj_idx) & ~PAGE_MASK;
1340 1337
1341 area = this_cpu_ptr(&zs_map_area); 1338 area = this_cpu_ptr(&zs_map_area);
1342 if (off + class->size <= PAGE_SIZE) 1339 if (off + class->size <= PAGE_SIZE)
@@ -1358,21 +1355,28 @@ EXPORT_SYMBOL_GPL(zs_unmap_object);
1358static unsigned long obj_malloc(struct size_class *class, 1355static unsigned long obj_malloc(struct size_class *class,
1359 struct zspage *zspage, unsigned long handle) 1356 struct zspage *zspage, unsigned long handle)
1360{ 1357{
1358 int i, nr_page, offset;
1361 unsigned long obj; 1359 unsigned long obj;
1362 struct link_free *link; 1360 struct link_free *link;
1363 1361
1364 struct page *m_page; 1362 struct page *m_page;
1365 unsigned long m_objidx, m_offset; 1363 unsigned long m_offset;
1366 void *vaddr; 1364 void *vaddr;
1367 1365
1368 handle |= OBJ_ALLOCATED_TAG; 1366 handle |= OBJ_ALLOCATED_TAG;
1369 obj = get_freeobj(zspage); 1367 obj = get_freeobj(zspage);
1370 obj_to_location(obj, &m_page, &m_objidx); 1368
1371 m_offset = obj_idx_to_offset(m_page, m_objidx, class->size); 1369 offset = obj * class->size;
1370 nr_page = offset >> PAGE_SHIFT;
1371 m_offset = offset & ~PAGE_MASK;
1372 m_page = get_first_page(zspage);
1373
1374 for (i = 0; i < nr_page; i++)
1375 m_page = get_next_page(m_page);
1372 1376
1373 vaddr = kmap_atomic(m_page); 1377 vaddr = kmap_atomic(m_page);
1374 link = (struct link_free *)vaddr + m_offset / sizeof(*link); 1378 link = (struct link_free *)vaddr + m_offset / sizeof(*link);
1375 set_freeobj(zspage, (unsigned long)link->next); 1379 set_freeobj(zspage, link->next >> OBJ_ALLOCATED_TAG);
1376 if (!class->huge) 1380 if (!class->huge)
1377 /* record handle in the header of allocated chunk */ 1381 /* record handle in the header of allocated chunk */
1378 link->handle = handle; 1382 link->handle = handle;
@@ -1384,6 +1388,8 @@ static unsigned long obj_malloc(struct size_class *class,
1384 mod_zspage_inuse(zspage, 1); 1388 mod_zspage_inuse(zspage, 1);
1385 zs_stat_inc(class, OBJ_USED, 1); 1389 zs_stat_inc(class, OBJ_USED, 1);
1386 1390
1391 obj = location_to_obj(m_page, obj);
1392
1387 return obj; 1393 return obj;
1388} 1394}
1389 1395
@@ -1449,22 +1455,22 @@ static void obj_free(struct size_class *class, unsigned long obj)
1449 struct link_free *link; 1455 struct link_free *link;
1450 struct zspage *zspage; 1456 struct zspage *zspage;
1451 struct page *f_page; 1457 struct page *f_page;
1452 unsigned long f_objidx, f_offset; 1458 unsigned long f_offset;
1459 unsigned int f_objidx;
1453 void *vaddr; 1460 void *vaddr;
1454 1461
1455 obj &= ~OBJ_ALLOCATED_TAG; 1462 obj &= ~OBJ_ALLOCATED_TAG;
1456 obj_to_location(obj, &f_page, &f_objidx); 1463 obj_to_location(obj, &f_page, &f_objidx);
1464 f_offset = (class->size * f_objidx) & ~PAGE_MASK;
1457 zspage = get_zspage(f_page); 1465 zspage = get_zspage(f_page);
1458 1466
1459 f_offset = obj_idx_to_offset(f_page, f_objidx, class->size);
1460
1461 vaddr = kmap_atomic(f_page); 1467 vaddr = kmap_atomic(f_page);
1462 1468
1463 /* Insert this object in containing zspage's freelist */ 1469 /* Insert this object in containing zspage's freelist */
1464 link = (struct link_free *)(vaddr + f_offset); 1470 link = (struct link_free *)(vaddr + f_offset);
1465 link->next = (void *)get_freeobj(zspage); 1471 link->next = get_freeobj(zspage) << OBJ_ALLOCATED_TAG;
1466 kunmap_atomic(vaddr); 1472 kunmap_atomic(vaddr);
1467 set_freeobj(zspage, obj); 1473 set_freeobj(zspage, f_objidx);
1468 mod_zspage_inuse(zspage, -1); 1474 mod_zspage_inuse(zspage, -1);
1469 zs_stat_dec(class, OBJ_USED, 1); 1475 zs_stat_dec(class, OBJ_USED, 1);
1470} 1476}
@@ -1473,7 +1479,8 @@ void zs_free(struct zs_pool *pool, unsigned long handle)
1473{ 1479{
1474 struct zspage *zspage; 1480 struct zspage *zspage;
1475 struct page *f_page; 1481 struct page *f_page;
1476 unsigned long obj, f_objidx; 1482 unsigned long obj;
1483 unsigned int f_objidx;
1477 int class_idx; 1484 int class_idx;
1478 struct size_class *class; 1485 struct size_class *class;
1479 enum fullness_group fullness; 1486 enum fullness_group fullness;
@@ -1510,7 +1517,7 @@ static void zs_object_copy(struct size_class *class, unsigned long dst,
1510 unsigned long src) 1517 unsigned long src)
1511{ 1518{
1512 struct page *s_page, *d_page; 1519 struct page *s_page, *d_page;
1513 unsigned long s_objidx, d_objidx; 1520 unsigned int s_objidx, d_objidx;
1514 unsigned long s_off, d_off; 1521 unsigned long s_off, d_off;
1515 void *s_addr, *d_addr; 1522 void *s_addr, *d_addr;
1516 int s_size, d_size, size; 1523 int s_size, d_size, size;
@@ -1521,8 +1528,8 @@ static void zs_object_copy(struct size_class *class, unsigned long dst,
1521 obj_to_location(src, &s_page, &s_objidx); 1528 obj_to_location(src, &s_page, &s_objidx);
1522 obj_to_location(dst, &d_page, &d_objidx); 1529 obj_to_location(dst, &d_page, &d_objidx);
1523 1530
1524 s_off = obj_idx_to_offset(s_page, s_objidx, class->size); 1531 s_off = (class->size * s_objidx) & ~PAGE_MASK;
1525 d_off = obj_idx_to_offset(d_page, d_objidx, class->size); 1532 d_off = (class->size * d_objidx) & ~PAGE_MASK;
1526 1533
1527 if (s_off + class->size > PAGE_SIZE) 1534 if (s_off + class->size > PAGE_SIZE)
1528 s_size = PAGE_SIZE - s_off; 1535 s_size = PAGE_SIZE - s_off;