aboutsummaryrefslogtreecommitdiffstats
path: root/mm/percpu.c
diff options
context:
space:
mode:
authorTejun Heo <tj@kernel.org>2009-08-14 02:00:51 -0400
committerTejun Heo <tj@kernel.org>2009-08-14 02:00:51 -0400
commitfd1e8a1fe2b54df6c185b4fa65f181f50b9c4d4e (patch)
treed4411507baacaa33c68be7813ad6f9e0d93a74e2 /mm/percpu.c
parent033e48fb82958053113178264ddb9d5038d5e38b (diff)
percpu: introduce pcpu_alloc_info and pcpu_group_info
Till now, non-linear cpu->unit map was expressed using an integer array which maps each cpu to a unit and used only by lpage allocator. Although how many units have been placed in a single contiguos area (group) is known while building unit_map, the information is lost when the result is recorded into the unit_map array. For lpage allocator, as all allocations are done by lpages and whether two adjacent lpages are in the same group or not is irrelevant, this didn't cause any problem. Non-linear cpu->unit mapping will be used for sparse embedding and this grouping information is necessary for that. This patch introduces pcpu_alloc_info which contains all the information necessary for initializing percpu allocator. pcpu_alloc_info contains array of pcpu_group_info which describes how units are grouped and mapped to cpus. pcpu_group_info also has base_offset field to specify its offset from the chunk's base address. pcpu_build_alloc_info() initializes this field as if all groups are allocated back-to-back as is currently done but this will be used to sparsely place groups. pcpu_alloc_info is a rather complex data structure which contains a flexible array which in turn points to nested cpu_map arrays. * pcpu_alloc_alloc_info() and pcpu_free_alloc_info() are provided to help dealing with pcpu_alloc_info. * pcpu_lpage_build_unit_map() is updated to build pcpu_alloc_info, generalized and renamed to pcpu_build_alloc_info(). @cpu_distance_fn may be NULL indicating that all cpus are of LOCAL_DISTANCE. * pcpul_lpage_dump_cfg() is updated to process pcpu_alloc_info, generalized and renamed to pcpu_dump_alloc_info(). It now also prints which group each alloc unit belongs to. * pcpu_setup_first_chunk() now takes pcpu_alloc_info instead of the separate parameters. All first chunk allocators are updated to use pcpu_build_alloc_info() to build alloc_info and call pcpu_setup_first_chunk() with it. This has the side effect of packing units for sparse possible cpus. ie. if cpus 0, 2 and 4 are possible, they'll be assigned unit 0, 1 and 2 instead of 0, 2 and 4. * x86 setup_pcpu_lpage() is updated to deal with alloc_info. * sparc64 setup_per_cpu_areas() is updated to build alloc_info. Although the changes made by this patch are pretty pervasive, it doesn't cause any behavior difference other than packing of sparse cpus. It mostly changes how information is passed among initialization functions and makes room for more flexibility. Signed-off-by: Tejun Heo <tj@kernel.org> Cc: Ingo Molnar <mingo@elte.hu> Cc: David Miller <davem@davemloft.net>
Diffstat (limited to 'mm/percpu.c')
-rw-r--r--mm/percpu.c529
1 files changed, 328 insertions, 201 deletions
diff --git a/mm/percpu.c b/mm/percpu.c
index 2b9c4b2a2fc0..99f7fa682722 100644
--- a/mm/percpu.c
+++ b/mm/percpu.c
@@ -58,6 +58,7 @@
58 58
59#include <linux/bitmap.h> 59#include <linux/bitmap.h>
60#include <linux/bootmem.h> 60#include <linux/bootmem.h>
61#include <linux/err.h>
61#include <linux/list.h> 62#include <linux/list.h>
62#include <linux/log2.h> 63#include <linux/log2.h>
63#include <linux/mm.h> 64#include <linux/mm.h>
@@ -1245,53 +1246,108 @@ static inline size_t pcpu_calc_fc_sizes(size_t static_size,
1245 return size_sum; 1246 return size_sum;
1246} 1247}
1247 1248
1248#ifdef CONFIG_NEED_PER_CPU_LPAGE_FIRST_CHUNK
1249/** 1249/**
1250 * pcpu_lpage_build_unit_map - build unit_map for large page remapping 1250 * pcpu_alloc_alloc_info - allocate percpu allocation info
1251 * @nr_groups: the number of groups
1252 * @nr_units: the number of units
1253 *
1254 * Allocate ai which is large enough for @nr_groups groups containing
1255 * @nr_units units. The returned ai's groups[0].cpu_map points to the
1256 * cpu_map array which is long enough for @nr_units and filled with
1257 * NR_CPUS. It's the caller's responsibility to initialize cpu_map
1258 * pointer of other groups.
1259 *
1260 * RETURNS:
1261 * Pointer to the allocated pcpu_alloc_info on success, NULL on
1262 * failure.
1263 */
1264struct pcpu_alloc_info * __init pcpu_alloc_alloc_info(int nr_groups,
1265 int nr_units)
1266{
1267 struct pcpu_alloc_info *ai;
1268 size_t base_size, ai_size;
1269 void *ptr;
1270 int unit;
1271
1272 base_size = ALIGN(sizeof(*ai) + nr_groups * sizeof(ai->groups[0]),
1273 __alignof__(ai->groups[0].cpu_map[0]));
1274 ai_size = base_size + nr_units * sizeof(ai->groups[0].cpu_map[0]);
1275
1276 ptr = alloc_bootmem_nopanic(PFN_ALIGN(ai_size));
1277 if (!ptr)
1278 return NULL;
1279 ai = ptr;
1280 ptr += base_size;
1281
1282 ai->groups[0].cpu_map = ptr;
1283
1284 for (unit = 0; unit < nr_units; unit++)
1285 ai->groups[0].cpu_map[unit] = NR_CPUS;
1286
1287 ai->nr_groups = nr_groups;
1288 ai->__ai_size = PFN_ALIGN(ai_size);
1289
1290 return ai;
1291}
1292
1293/**
1294 * pcpu_free_alloc_info - free percpu allocation info
1295 * @ai: pcpu_alloc_info to free
1296 *
1297 * Free @ai which was allocated by pcpu_alloc_alloc_info().
1298 */
1299void __init pcpu_free_alloc_info(struct pcpu_alloc_info *ai)
1300{
1301 free_bootmem(__pa(ai), ai->__ai_size);
1302}
1303
1304/**
1305 * pcpu_build_alloc_info - build alloc_info considering distances between CPUs
1251 * @reserved_size: the size of reserved percpu area in bytes 1306 * @reserved_size: the size of reserved percpu area in bytes
1252 * @dyn_sizep: in/out parameter for dynamic size, -1 for auto 1307 * @dyn_size: free size for dynamic allocation in bytes, -1 for auto
1253 * @unit_sizep: out parameter for unit size 1308 * @atom_size: allocation atom size
1254 * @unit_map: unit_map to be filled 1309 * @cpu_distance_fn: callback to determine distance between cpus, optional
1255 * @cpu_distance_fn: callback to determine distance between cpus
1256 * 1310 *
1257 * This function builds cpu -> unit map and determine other parameters 1311 * This function determines grouping of units, their mappings to cpus
1258 * considering needed percpu size, large page size and distances 1312 * and other parameters considering needed percpu size, allocation
1259 * between CPUs in NUMA. 1313 * atom size and distances between CPUs.
1260 * 1314 *
1261 * CPUs which are of LOCAL_DISTANCE both ways are grouped together and 1315 * Groups are always mutliples of atom size and CPUs which are of
1262 * may share units in the same large page. The returned configuration 1316 * LOCAL_DISTANCE both ways are grouped together and share space for
1263 * is guaranteed to have CPUs on different nodes on different large 1317 * units in the same group. The returned configuration is guaranteed
1264 * pages and >=75% usage of allocated virtual address space. 1318 * to have CPUs on different nodes on different groups and >=75% usage
1319 * of allocated virtual address space.
1265 * 1320 *
1266 * RETURNS: 1321 * RETURNS:
1267 * On success, fills in @unit_map, sets *@dyn_sizep, *@unit_sizep and 1322 * On success, pointer to the new allocation_info is returned. On
1268 * returns the number of units to be allocated. -errno on failure. 1323 * failure, ERR_PTR value is returned.
1269 */ 1324 */
1270int __init pcpu_lpage_build_unit_map(size_t reserved_size, ssize_t *dyn_sizep, 1325struct pcpu_alloc_info * __init pcpu_build_alloc_info(
1271 size_t *unit_sizep, size_t lpage_size, 1326 size_t reserved_size, ssize_t dyn_size,
1272 int *unit_map, 1327 size_t atom_size,
1273 pcpu_fc_cpu_distance_fn_t cpu_distance_fn) 1328 pcpu_fc_cpu_distance_fn_t cpu_distance_fn)
1274{ 1329{
1275 static int group_map[NR_CPUS] __initdata; 1330 static int group_map[NR_CPUS] __initdata;
1276 static int group_cnt[NR_CPUS] __initdata; 1331 static int group_cnt[NR_CPUS] __initdata;
1277 const size_t static_size = __per_cpu_end - __per_cpu_start; 1332 const size_t static_size = __per_cpu_end - __per_cpu_start;
1278 int group_cnt_max = 0; 1333 int group_cnt_max = 0, nr_groups = 1, nr_units = 0;
1279 size_t size_sum, min_unit_size, alloc_size; 1334 size_t size_sum, min_unit_size, alloc_size;
1280 int upa, max_upa, uninitialized_var(best_upa); /* units_per_alloc */ 1335 int upa, max_upa, uninitialized_var(best_upa); /* units_per_alloc */
1281 int last_allocs; 1336 int last_allocs, group, unit;
1282 unsigned int cpu, tcpu; 1337 unsigned int cpu, tcpu;
1283 int group, unit; 1338 struct pcpu_alloc_info *ai;
1339 unsigned int *cpu_map;
1284 1340
1285 /* 1341 /*
1286 * Determine min_unit_size, alloc_size and max_upa such that 1342 * Determine min_unit_size, alloc_size and max_upa such that
1287 * alloc_size is multiple of lpage_size and is the smallest 1343 * alloc_size is multiple of atom_size and is the smallest
1288 * which can accomodate 4k aligned segments which are equal to 1344 * which can accomodate 4k aligned segments which are equal to
1289 * or larger than min_unit_size. 1345 * or larger than min_unit_size.
1290 */ 1346 */
1291 size_sum = pcpu_calc_fc_sizes(static_size, reserved_size, dyn_sizep); 1347 size_sum = pcpu_calc_fc_sizes(static_size, reserved_size, &dyn_size);
1292 min_unit_size = max_t(size_t, size_sum, PCPU_MIN_UNIT_SIZE); 1348 min_unit_size = max_t(size_t, size_sum, PCPU_MIN_UNIT_SIZE);
1293 1349
1294 alloc_size = roundup(min_unit_size, lpage_size); 1350 alloc_size = roundup(min_unit_size, atom_size);
1295 upa = alloc_size / min_unit_size; 1351 upa = alloc_size / min_unit_size;
1296 while (alloc_size % upa || ((alloc_size / upa) & ~PAGE_MASK)) 1352 while (alloc_size % upa || ((alloc_size / upa) & ~PAGE_MASK))
1297 upa--; 1353 upa--;
@@ -1304,10 +1360,11 @@ int __init pcpu_lpage_build_unit_map(size_t reserved_size, ssize_t *dyn_sizep,
1304 for_each_possible_cpu(tcpu) { 1360 for_each_possible_cpu(tcpu) {
1305 if (cpu == tcpu) 1361 if (cpu == tcpu)
1306 break; 1362 break;
1307 if (group_map[tcpu] == group && 1363 if (group_map[tcpu] == group && cpu_distance_fn &&
1308 (cpu_distance_fn(cpu, tcpu) > LOCAL_DISTANCE || 1364 (cpu_distance_fn(cpu, tcpu) > LOCAL_DISTANCE ||
1309 cpu_distance_fn(tcpu, cpu) > LOCAL_DISTANCE)) { 1365 cpu_distance_fn(tcpu, cpu) > LOCAL_DISTANCE)) {
1310 group++; 1366 group++;
1367 nr_groups = max(nr_groups, group + 1);
1311 goto next_group; 1368 goto next_group;
1312 } 1369 }
1313 } 1370 }
@@ -1328,7 +1385,7 @@ int __init pcpu_lpage_build_unit_map(size_t reserved_size, ssize_t *dyn_sizep,
1328 if (alloc_size % upa || ((alloc_size / upa) & ~PAGE_MASK)) 1385 if (alloc_size % upa || ((alloc_size / upa) & ~PAGE_MASK))
1329 continue; 1386 continue;
1330 1387
1331 for (group = 0; group_cnt[group]; group++) { 1388 for (group = 0; group < nr_groups; group++) {
1332 int this_allocs = DIV_ROUND_UP(group_cnt[group], upa); 1389 int this_allocs = DIV_ROUND_UP(group_cnt[group], upa);
1333 allocs += this_allocs; 1390 allocs += this_allocs;
1334 wasted += this_allocs * upa - group_cnt[group]; 1391 wasted += this_allocs * upa - group_cnt[group];
@@ -1348,75 +1405,122 @@ int __init pcpu_lpage_build_unit_map(size_t reserved_size, ssize_t *dyn_sizep,
1348 last_allocs = allocs; 1405 last_allocs = allocs;
1349 best_upa = upa; 1406 best_upa = upa;
1350 } 1407 }
1351 *unit_sizep = alloc_size / best_upa; 1408 upa = best_upa;
1409
1410 /* allocate and fill alloc_info */
1411 for (group = 0; group < nr_groups; group++)
1412 nr_units += roundup(group_cnt[group], upa);
1413
1414 ai = pcpu_alloc_alloc_info(nr_groups, nr_units);
1415 if (!ai)
1416 return ERR_PTR(-ENOMEM);
1417 cpu_map = ai->groups[0].cpu_map;
1418
1419 for (group = 0; group < nr_groups; group++) {
1420 ai->groups[group].cpu_map = cpu_map;
1421 cpu_map += roundup(group_cnt[group], upa);
1422 }
1423
1424 ai->static_size = static_size;
1425 ai->reserved_size = reserved_size;
1426 ai->dyn_size = dyn_size;
1427 ai->unit_size = alloc_size / upa;
1428 ai->atom_size = atom_size;
1429 ai->alloc_size = alloc_size;
1430
1431 for (group = 0, unit = 0; group_cnt[group]; group++) {
1432 struct pcpu_group_info *gi = &ai->groups[group];
1433
1434 /*
1435 * Initialize base_offset as if all groups are located
1436 * back-to-back. The caller should update this to
1437 * reflect actual allocation.
1438 */
1439 gi->base_offset = unit * ai->unit_size;
1352 1440
1353 /* assign units to cpus accordingly */
1354 unit = 0;
1355 for (group = 0; group_cnt[group]; group++) {
1356 for_each_possible_cpu(cpu) 1441 for_each_possible_cpu(cpu)
1357 if (group_map[cpu] == group) 1442 if (group_map[cpu] == group)
1358 unit_map[cpu] = unit++; 1443 gi->cpu_map[gi->nr_units++] = cpu;
1359 unit = roundup(unit, best_upa); 1444 gi->nr_units = roundup(gi->nr_units, upa);
1445 unit += gi->nr_units;
1360 } 1446 }
1447 BUG_ON(unit != nr_units);
1361 1448
1362 return unit; /* unit contains aligned number of units */ 1449 return ai;
1363} 1450}
1364 1451
1365static bool __init pcpul_unit_to_cpu(int unit, const int *unit_map, 1452/**
1366 unsigned int *cpup); 1453 * pcpu_dump_alloc_info - print out information about pcpu_alloc_info
1367 1454 * @lvl: loglevel
1368static void __init pcpul_lpage_dump_cfg(const char *lvl, size_t static_size, 1455 * @ai: allocation info to dump
1369 size_t reserved_size, size_t dyn_size, 1456 *
1370 size_t unit_size, size_t lpage_size, 1457 * Print out information about @ai using loglevel @lvl.
1371 const int *unit_map, int nr_units) 1458 */
1459static void pcpu_dump_alloc_info(const char *lvl,
1460 const struct pcpu_alloc_info *ai)
1372{ 1461{
1373 int width = 1, v = nr_units; 1462 int group_width = 1, cpu_width = 1, width;
1374 char empty_str[] = "--------"; 1463 char empty_str[] = "--------";
1375 int upl, lpl; /* units per lpage, lpage per line */ 1464 int alloc = 0, alloc_end = 0;
1376 unsigned int cpu; 1465 int group, v;
1377 int lpage, unit; 1466 int upa, apl; /* units per alloc, allocs per line */
1467
1468 v = ai->nr_groups;
1469 while (v /= 10)
1470 group_width++;
1378 1471
1472 v = num_possible_cpus();
1379 while (v /= 10) 1473 while (v /= 10)
1380 width++; 1474 cpu_width++;
1381 empty_str[min_t(int, width, sizeof(empty_str) - 1)] = '\0'; 1475 empty_str[min_t(int, cpu_width, sizeof(empty_str) - 1)] = '\0';
1382 1476
1383 upl = max_t(int, lpage_size / unit_size, 1); 1477 upa = ai->alloc_size / ai->unit_size;
1384 lpl = rounddown_pow_of_two(max_t(int, 60 / (upl * (width + 1) + 2), 1)); 1478 width = upa * (cpu_width + 1) + group_width + 3;
1479 apl = rounddown_pow_of_two(max(60 / width, 1));
1385 1480
1386 printk("%spcpu-lpage: sta/res/dyn=%zu/%zu/%zu unit=%zu lpage=%zu", lvl, 1481 printk("%spcpu-alloc: s%zu r%zu d%zu u%zu alloc=%zu*%zu",
1387 static_size, reserved_size, dyn_size, unit_size, lpage_size); 1482 lvl, ai->static_size, ai->reserved_size, ai->dyn_size,
1483 ai->unit_size, ai->alloc_size / ai->atom_size, ai->atom_size);
1388 1484
1389 for (lpage = 0, unit = 0; unit < nr_units; unit++) { 1485 for (group = 0; group < ai->nr_groups; group++) {
1390 if (!(unit % upl)) { 1486 const struct pcpu_group_info *gi = &ai->groups[group];
1391 if (!(lpage++ % lpl)) { 1487 int unit = 0, unit_end = 0;
1488
1489 BUG_ON(gi->nr_units % upa);
1490 for (alloc_end += gi->nr_units / upa;
1491 alloc < alloc_end; alloc++) {
1492 if (!(alloc % apl)) {
1392 printk("\n"); 1493 printk("\n");
1393 printk("%spcpu-lpage: ", lvl); 1494 printk("%spcpu-alloc: ", lvl);
1394 } else 1495 }
1395 printk("| "); 1496 printk("[%0*d] ", group_width, group);
1497
1498 for (unit_end += upa; unit < unit_end; unit++)
1499 if (gi->cpu_map[unit] != NR_CPUS)
1500 printk("%0*d ", cpu_width,
1501 gi->cpu_map[unit]);
1502 else
1503 printk("%s ", empty_str);
1396 } 1504 }
1397 if (pcpul_unit_to_cpu(unit, unit_map, &cpu))
1398 printk("%0*d ", width, cpu);
1399 else
1400 printk("%s ", empty_str);
1401 } 1505 }
1402 printk("\n"); 1506 printk("\n");
1403} 1507}
1404#endif
1405 1508
1406/** 1509/**
1407 * pcpu_setup_first_chunk - initialize the first percpu chunk 1510 * pcpu_setup_first_chunk - initialize the first percpu chunk
1408 * @static_size: the size of static percpu area in bytes 1511 * @ai: pcpu_alloc_info describing how to percpu area is shaped
1409 * @reserved_size: the size of reserved percpu area in bytes, 0 for none
1410 * @dyn_size: free size for dynamic allocation in bytes
1411 * @unit_size: unit size in bytes, must be multiple of PAGE_SIZE
1412 * @base_addr: mapped address 1512 * @base_addr: mapped address
1413 * @unit_map: cpu -> unit map, NULL for sequential mapping
1414 * 1513 *
1415 * Initialize the first percpu chunk which contains the kernel static 1514 * Initialize the first percpu chunk which contains the kernel static
1416 * perpcu area. This function is to be called from arch percpu area 1515 * perpcu area. This function is to be called from arch percpu area
1417 * setup path. 1516 * setup path.
1418 * 1517 *
1419 * @reserved_size, if non-zero, specifies the amount of bytes to 1518 * @ai contains all information necessary to initialize the first
1519 * chunk and prime the dynamic percpu allocator.
1520 *
1521 * @ai->static_size is the size of static percpu area.
1522 *
1523 * @ai->reserved_size, if non-zero, specifies the amount of bytes to
1420 * reserve after the static area in the first chunk. This reserves 1524 * reserve after the static area in the first chunk. This reserves
1421 * the first chunk such that it's available only through reserved 1525 * the first chunk such that it's available only through reserved
1422 * percpu allocation. This is primarily used to serve module percpu 1526 * percpu allocation. This is primarily used to serve module percpu
@@ -1424,13 +1528,26 @@ static void __init pcpul_lpage_dump_cfg(const char *lvl, size_t static_size,
1424 * limited offset range for symbol relocations to guarantee module 1528 * limited offset range for symbol relocations to guarantee module
1425 * percpu symbols fall inside the relocatable range. 1529 * percpu symbols fall inside the relocatable range.
1426 * 1530 *
1427 * @dyn_size determines the number of bytes available for dynamic 1531 * @ai->dyn_size determines the number of bytes available for dynamic
1428 * allocation in the first chunk. The area between @static_size + 1532 * allocation in the first chunk. The area between @ai->static_size +
1429 * @reserved_size + @dyn_size and @unit_size is unused. 1533 * @ai->reserved_size + @ai->dyn_size and @ai->unit_size is unused.
1430 * 1534 *
1431 * @unit_size specifies unit size and must be aligned to PAGE_SIZE and 1535 * @ai->unit_size specifies unit size and must be aligned to PAGE_SIZE
1432 * equal to or larger than @static_size + @reserved_size + if 1536 * and equal to or larger than @ai->static_size + @ai->reserved_size +
1433 * non-negative, @dyn_size. 1537 * @ai->dyn_size.
1538 *
1539 * @ai->atom_size is the allocation atom size and used as alignment
1540 * for vm areas.
1541 *
1542 * @ai->alloc_size is the allocation size and always multiple of
1543 * @ai->atom_size. This is larger than @ai->atom_size if
1544 * @ai->unit_size is larger than @ai->atom_size.
1545 *
1546 * @ai->nr_groups and @ai->groups describe virtual memory layout of
1547 * percpu areas. Units which should be colocated are put into the
1548 * same group. Dynamic VM areas will be allocated according to these
1549 * groupings. If @ai->nr_groups is zero, a single group containing
1550 * all units is assumed.
1434 * 1551 *
1435 * The caller should have mapped the first chunk at @base_addr and 1552 * The caller should have mapped the first chunk at @base_addr and
1436 * copied static data to each unit. 1553 * copied static data to each unit.
@@ -1446,70 +1563,63 @@ static void __init pcpul_lpage_dump_cfg(const char *lvl, size_t static_size,
1446 * The determined pcpu_unit_size which can be used to initialize 1563 * The determined pcpu_unit_size which can be used to initialize
1447 * percpu access. 1564 * percpu access.
1448 */ 1565 */
1449size_t __init pcpu_setup_first_chunk(size_t static_size, size_t reserved_size, 1566size_t __init pcpu_setup_first_chunk(const struct pcpu_alloc_info *ai,
1450 size_t dyn_size, size_t unit_size, 1567 void *base_addr)
1451 void *base_addr, const int *unit_map)
1452{ 1568{
1453 static struct vm_struct first_vm; 1569 static struct vm_struct first_vm;
1454 static int smap[2], dmap[2]; 1570 static int smap[2], dmap[2];
1455 size_t size_sum = static_size + reserved_size + dyn_size; 1571 size_t dyn_size = ai->dyn_size;
1572 size_t size_sum = ai->static_size + ai->reserved_size + dyn_size;
1456 struct pcpu_chunk *schunk, *dchunk = NULL; 1573 struct pcpu_chunk *schunk, *dchunk = NULL;
1457 unsigned int cpu, tcpu; 1574 unsigned int cpu;
1458 int i; 1575 int *unit_map;
1576 int group, unit, i;
1459 1577
1460 /* sanity checks */ 1578 /* sanity checks */
1461 BUILD_BUG_ON(ARRAY_SIZE(smap) >= PCPU_DFL_MAP_ALLOC || 1579 BUILD_BUG_ON(ARRAY_SIZE(smap) >= PCPU_DFL_MAP_ALLOC ||
1462 ARRAY_SIZE(dmap) >= PCPU_DFL_MAP_ALLOC); 1580 ARRAY_SIZE(dmap) >= PCPU_DFL_MAP_ALLOC);
1463 BUG_ON(!static_size); 1581 BUG_ON(ai->nr_groups <= 0);
1582 BUG_ON(!ai->static_size);
1464 BUG_ON(!base_addr); 1583 BUG_ON(!base_addr);
1465 BUG_ON(unit_size < size_sum); 1584 BUG_ON(ai->unit_size < size_sum);
1466 BUG_ON(unit_size & ~PAGE_MASK); 1585 BUG_ON(ai->unit_size & ~PAGE_MASK);
1467 BUG_ON(unit_size < PCPU_MIN_UNIT_SIZE); 1586 BUG_ON(ai->unit_size < PCPU_MIN_UNIT_SIZE);
1587
1588 pcpu_dump_alloc_info(KERN_DEBUG, ai);
1468 1589
1469 /* determine number of units and verify and initialize pcpu_unit_map */ 1590 /* determine number of units and verify and initialize pcpu_unit_map */
1470 if (unit_map) { 1591 unit_map = alloc_bootmem(nr_cpu_ids * sizeof(unit_map[0]));
1471 int first_unit = INT_MAX, last_unit = INT_MIN;
1472
1473 for_each_possible_cpu(cpu) {
1474 int unit = unit_map[cpu];
1475
1476 BUG_ON(unit < 0);
1477 for_each_possible_cpu(tcpu) {
1478 if (tcpu == cpu)
1479 break;
1480 /* the mapping should be one-to-one */
1481 BUG_ON(unit_map[tcpu] == unit);
1482 }
1483 1592
1484 if (unit < first_unit) { 1593 for (cpu = 0; cpu < nr_cpu_ids; cpu++)
1485 pcpu_first_unit_cpu = cpu; 1594 unit_map[cpu] = NR_CPUS;
1486 first_unit = unit; 1595 pcpu_first_unit_cpu = NR_CPUS;
1487 }
1488 if (unit > last_unit) {
1489 pcpu_last_unit_cpu = cpu;
1490 last_unit = unit;
1491 }
1492 }
1493 pcpu_nr_units = last_unit + 1;
1494 pcpu_unit_map = unit_map;
1495 } else {
1496 int *identity_map;
1497 1596
1498 /* #units == #cpus, identity mapped */ 1597 for (group = 0, unit = 0; group < ai->nr_groups; group++, unit += i) {
1499 identity_map = alloc_bootmem(nr_cpu_ids * 1598 const struct pcpu_group_info *gi = &ai->groups[group];
1500 sizeof(identity_map[0]));
1501 1599
1502 for_each_possible_cpu(cpu) 1600 for (i = 0; i < gi->nr_units; i++) {
1503 identity_map[cpu] = cpu; 1601 cpu = gi->cpu_map[i];
1602 if (cpu == NR_CPUS)
1603 continue;
1504 1604
1505 pcpu_first_unit_cpu = 0; 1605 BUG_ON(cpu > nr_cpu_ids || !cpu_possible(cpu));
1506 pcpu_last_unit_cpu = pcpu_nr_units - 1; 1606 BUG_ON(unit_map[cpu] != NR_CPUS);
1507 pcpu_nr_units = nr_cpu_ids; 1607
1508 pcpu_unit_map = identity_map; 1608 unit_map[cpu] = unit + i;
1609 if (pcpu_first_unit_cpu == NR_CPUS)
1610 pcpu_first_unit_cpu = cpu;
1611 }
1509 } 1612 }
1613 pcpu_last_unit_cpu = cpu;
1614 pcpu_nr_units = unit;
1615
1616 for_each_possible_cpu(cpu)
1617 BUG_ON(unit_map[cpu] == NR_CPUS);
1618
1619 pcpu_unit_map = unit_map;
1510 1620
1511 /* determine basic parameters */ 1621 /* determine basic parameters */
1512 pcpu_unit_pages = unit_size >> PAGE_SHIFT; 1622 pcpu_unit_pages = ai->unit_size >> PAGE_SHIFT;
1513 pcpu_unit_size = pcpu_unit_pages << PAGE_SHIFT; 1623 pcpu_unit_size = pcpu_unit_pages << PAGE_SHIFT;
1514 pcpu_chunk_size = pcpu_nr_units * pcpu_unit_size; 1624 pcpu_chunk_size = pcpu_nr_units * pcpu_unit_size;
1515 pcpu_chunk_struct_size = sizeof(struct pcpu_chunk) + 1625 pcpu_chunk_struct_size = sizeof(struct pcpu_chunk) +
@@ -1543,17 +1653,17 @@ size_t __init pcpu_setup_first_chunk(size_t static_size, size_t reserved_size,
1543 schunk->immutable = true; 1653 schunk->immutable = true;
1544 bitmap_fill(schunk->populated, pcpu_unit_pages); 1654 bitmap_fill(schunk->populated, pcpu_unit_pages);
1545 1655
1546 if (reserved_size) { 1656 if (ai->reserved_size) {
1547 schunk->free_size = reserved_size; 1657 schunk->free_size = ai->reserved_size;
1548 pcpu_reserved_chunk = schunk; 1658 pcpu_reserved_chunk = schunk;
1549 pcpu_reserved_chunk_limit = static_size + reserved_size; 1659 pcpu_reserved_chunk_limit = ai->static_size + ai->reserved_size;
1550 } else { 1660 } else {
1551 schunk->free_size = dyn_size; 1661 schunk->free_size = dyn_size;
1552 dyn_size = 0; /* dynamic area covered */ 1662 dyn_size = 0; /* dynamic area covered */
1553 } 1663 }
1554 schunk->contig_hint = schunk->free_size; 1664 schunk->contig_hint = schunk->free_size;
1555 1665
1556 schunk->map[schunk->map_used++] = -static_size; 1666 schunk->map[schunk->map_used++] = -ai->static_size;
1557 if (schunk->free_size) 1667 if (schunk->free_size)
1558 schunk->map[schunk->map_used++] = schunk->free_size; 1668 schunk->map[schunk->map_used++] = schunk->free_size;
1559 1669
@@ -1643,44 +1753,47 @@ early_param("percpu_alloc", percpu_alloc_setup);
1643 */ 1753 */
1644ssize_t __init pcpu_embed_first_chunk(size_t reserved_size, ssize_t dyn_size) 1754ssize_t __init pcpu_embed_first_chunk(size_t reserved_size, ssize_t dyn_size)
1645{ 1755{
1646 const size_t static_size = __per_cpu_end - __per_cpu_start; 1756 struct pcpu_alloc_info *ai;
1647 size_t size_sum, unit_size, chunk_size; 1757 size_t size_sum, chunk_size;
1648 void *base; 1758 void *base;
1649 unsigned int cpu; 1759 int unit;
1760 ssize_t ret;
1650 1761
1651 /* determine parameters and allocate */ 1762 ai = pcpu_build_alloc_info(reserved_size, dyn_size, PAGE_SIZE, NULL);
1652 size_sum = pcpu_calc_fc_sizes(static_size, reserved_size, &dyn_size); 1763 if (IS_ERR(ai))
1764 return PTR_ERR(ai);
1765 BUG_ON(ai->nr_groups != 1);
1766 BUG_ON(ai->groups[0].nr_units != num_possible_cpus());
1653 1767
1654 unit_size = max_t(size_t, size_sum, PCPU_MIN_UNIT_SIZE); 1768 size_sum = ai->static_size + ai->reserved_size + ai->dyn_size;
1655 chunk_size = unit_size * nr_cpu_ids; 1769 chunk_size = ai->unit_size * num_possible_cpus();
1656 1770
1657 base = __alloc_bootmem_nopanic(chunk_size, PAGE_SIZE, 1771 base = __alloc_bootmem_nopanic(chunk_size, PAGE_SIZE,
1658 __pa(MAX_DMA_ADDRESS)); 1772 __pa(MAX_DMA_ADDRESS));
1659 if (!base) { 1773 if (!base) {
1660 pr_warning("PERCPU: failed to allocate %zu bytes for " 1774 pr_warning("PERCPU: failed to allocate %zu bytes for "
1661 "embedding\n", chunk_size); 1775 "embedding\n", chunk_size);
1662 return -ENOMEM; 1776 ret = -ENOMEM;
1777 goto out_free_ai;
1663 } 1778 }
1664 1779
1665 /* return the leftover and copy */ 1780 /* return the leftover and copy */
1666 for (cpu = 0; cpu < nr_cpu_ids; cpu++) { 1781 for (unit = 0; unit < num_possible_cpus(); unit++) {
1667 void *ptr = base + cpu * unit_size; 1782 void *ptr = base + unit * ai->unit_size;
1668 1783
1669 if (cpu_possible(cpu)) { 1784 free_bootmem(__pa(ptr + size_sum), ai->unit_size - size_sum);
1670 free_bootmem(__pa(ptr + size_sum), 1785 memcpy(ptr, __per_cpu_load, ai->static_size);
1671 unit_size - size_sum);
1672 memcpy(ptr, __per_cpu_load, static_size);
1673 } else
1674 free_bootmem(__pa(ptr), unit_size);
1675 } 1786 }
1676 1787
1677 /* we're ready, commit */ 1788 /* we're ready, commit */
1678 pr_info("PERCPU: Embedded %zu pages/cpu @%p s%zu r%zu d%zu u%zu\n", 1789 pr_info("PERCPU: Embedded %zu pages/cpu @%p s%zu r%zu d%zu u%zu\n",
1679 PFN_DOWN(size_sum), base, static_size, reserved_size, dyn_size, 1790 PFN_DOWN(size_sum), base, ai->static_size, ai->reserved_size,
1680 unit_size); 1791 ai->dyn_size, ai->unit_size);
1681 1792
1682 return pcpu_setup_first_chunk(static_size, reserved_size, dyn_size, 1793 ret = pcpu_setup_first_chunk(ai, base);
1683 unit_size, base, NULL); 1794out_free_ai:
1795 pcpu_free_alloc_info(ai);
1796 return ret;
1684} 1797}
1685#endif /* CONFIG_NEED_PER_CPU_EMBED_FIRST_CHUNK || 1798#endif /* CONFIG_NEED_PER_CPU_EMBED_FIRST_CHUNK ||
1686 !CONFIG_HAVE_SETUP_PER_CPU_AREA */ 1799 !CONFIG_HAVE_SETUP_PER_CPU_AREA */
@@ -1709,31 +1822,34 @@ ssize_t __init pcpu_page_first_chunk(size_t reserved_size,
1709 pcpu_fc_populate_pte_fn_t populate_pte_fn) 1822 pcpu_fc_populate_pte_fn_t populate_pte_fn)
1710{ 1823{
1711 static struct vm_struct vm; 1824 static struct vm_struct vm;
1712 const size_t static_size = __per_cpu_end - __per_cpu_start; 1825 struct pcpu_alloc_info *ai;
1713 ssize_t dyn_size = -1;
1714 size_t size_sum, unit_size;
1715 char psize_str[16]; 1826 char psize_str[16];
1716 int unit_pages; 1827 int unit_pages;
1717 size_t pages_size; 1828 size_t pages_size;
1718 struct page **pages; 1829 struct page **pages;
1719 unsigned int cpu; 1830 int unit, i, j;
1720 int i, j;
1721 ssize_t ret; 1831 ssize_t ret;
1722 1832
1723 snprintf(psize_str, sizeof(psize_str), "%luK", PAGE_SIZE >> 10); 1833 snprintf(psize_str, sizeof(psize_str), "%luK", PAGE_SIZE >> 10);
1724 1834
1725 size_sum = pcpu_calc_fc_sizes(static_size, reserved_size, &dyn_size); 1835 ai = pcpu_build_alloc_info(reserved_size, -1, PAGE_SIZE, NULL);
1726 unit_size = max_t(size_t, size_sum, PCPU_MIN_UNIT_SIZE); 1836 if (IS_ERR(ai))
1727 unit_pages = unit_size >> PAGE_SHIFT; 1837 return PTR_ERR(ai);
1838 BUG_ON(ai->nr_groups != 1);
1839 BUG_ON(ai->groups[0].nr_units != num_possible_cpus());
1840
1841 unit_pages = ai->unit_size >> PAGE_SHIFT;
1728 1842
1729 /* unaligned allocations can't be freed, round up to page size */ 1843 /* unaligned allocations can't be freed, round up to page size */
1730 pages_size = PFN_ALIGN(unit_pages * nr_cpu_ids * sizeof(pages[0])); 1844 pages_size = PFN_ALIGN(unit_pages * num_possible_cpus() *
1845 sizeof(pages[0]));
1731 pages = alloc_bootmem(pages_size); 1846 pages = alloc_bootmem(pages_size);
1732 1847
1733 /* allocate pages */ 1848 /* allocate pages */
1734 j = 0; 1849 j = 0;
1735 for_each_possible_cpu(cpu) 1850 for (unit = 0; unit < num_possible_cpus(); unit++)
1736 for (i = 0; i < unit_pages; i++) { 1851 for (i = 0; i < unit_pages; i++) {
1852 unsigned int cpu = ai->groups[0].cpu_map[unit];
1737 void *ptr; 1853 void *ptr;
1738 1854
1739 ptr = alloc_fn(cpu, PAGE_SIZE, PAGE_SIZE); 1855 ptr = alloc_fn(cpu, PAGE_SIZE, PAGE_SIZE);
@@ -1747,18 +1863,18 @@ ssize_t __init pcpu_page_first_chunk(size_t reserved_size,
1747 1863
1748 /* allocate vm area, map the pages and copy static data */ 1864 /* allocate vm area, map the pages and copy static data */
1749 vm.flags = VM_ALLOC; 1865 vm.flags = VM_ALLOC;
1750 vm.size = nr_cpu_ids * unit_size; 1866 vm.size = num_possible_cpus() * ai->unit_size;
1751 vm_area_register_early(&vm, PAGE_SIZE); 1867 vm_area_register_early(&vm, PAGE_SIZE);
1752 1868
1753 for_each_possible_cpu(cpu) { 1869 for (unit = 0; unit < num_possible_cpus(); unit++) {
1754 unsigned long unit_addr = 1870 unsigned long unit_addr =
1755 (unsigned long)vm.addr + cpu * unit_size; 1871 (unsigned long)vm.addr + unit * ai->unit_size;
1756 1872
1757 for (i = 0; i < unit_pages; i++) 1873 for (i = 0; i < unit_pages; i++)
1758 populate_pte_fn(unit_addr + (i << PAGE_SHIFT)); 1874 populate_pte_fn(unit_addr + (i << PAGE_SHIFT));
1759 1875
1760 /* pte already populated, the following shouldn't fail */ 1876 /* pte already populated, the following shouldn't fail */
1761 ret = __pcpu_map_pages(unit_addr, &pages[cpu * unit_pages], 1877 ret = __pcpu_map_pages(unit_addr, &pages[unit * unit_pages],
1762 unit_pages); 1878 unit_pages);
1763 if (ret < 0) 1879 if (ret < 0)
1764 panic("failed to map percpu area, err=%zd\n", ret); 1880 panic("failed to map percpu area, err=%zd\n", ret);
@@ -1772,16 +1888,15 @@ ssize_t __init pcpu_page_first_chunk(size_t reserved_size,
1772 */ 1888 */
1773 1889
1774 /* copy static data */ 1890 /* copy static data */
1775 memcpy((void *)unit_addr, __per_cpu_load, static_size); 1891 memcpy((void *)unit_addr, __per_cpu_load, ai->static_size);
1776 } 1892 }
1777 1893
1778 /* we're ready, commit */ 1894 /* we're ready, commit */
1779 pr_info("PERCPU: %d %s pages/cpu @%p s%zu r%zu d%zu\n", 1895 pr_info("PERCPU: %d %s pages/cpu @%p s%zu r%zu d%zu\n",
1780 unit_pages, psize_str, vm.addr, static_size, reserved_size, 1896 unit_pages, psize_str, vm.addr, ai->static_size,
1781 dyn_size); 1897 ai->reserved_size, ai->dyn_size);
1782 1898
1783 ret = pcpu_setup_first_chunk(static_size, reserved_size, dyn_size, 1899 ret = pcpu_setup_first_chunk(ai, vm.addr);
1784 unit_size, vm.addr, NULL);
1785 goto out_free_ar; 1900 goto out_free_ar;
1786 1901
1787enomem: 1902enomem:
@@ -1790,6 +1905,7 @@ enomem:
1790 ret = -ENOMEM; 1905 ret = -ENOMEM;
1791out_free_ar: 1906out_free_ar:
1792 free_bootmem(__pa(pages), pages_size); 1907 free_bootmem(__pa(pages), pages_size);
1908 pcpu_free_alloc_info(ai);
1793 return ret; 1909 return ret;
1794} 1910}
1795#endif /* CONFIG_NEED_PER_CPU_PAGE_FIRST_CHUNK */ 1911#endif /* CONFIG_NEED_PER_CPU_PAGE_FIRST_CHUNK */
@@ -1805,38 +1921,50 @@ static size_t pcpul_lpage_size;
1805static int pcpul_nr_lpages; 1921static int pcpul_nr_lpages;
1806static struct pcpul_ent *pcpul_map; 1922static struct pcpul_ent *pcpul_map;
1807 1923
1808static bool __init pcpul_unit_to_cpu(int unit, const int *unit_map, 1924static bool __init pcpul_unit_to_cpu(int unit, const struct pcpu_alloc_info *ai,
1809 unsigned int *cpup) 1925 unsigned int *cpup)
1810{ 1926{
1811 unsigned int cpu; 1927 int group, cunit;
1812 1928
1813 for_each_possible_cpu(cpu) 1929 for (group = 0, cunit = 0; group < ai->nr_groups; group++) {
1814 if (unit_map[cpu] == unit) { 1930 const struct pcpu_group_info *gi = &ai->groups[group];
1931
1932 if (unit < cunit + gi->nr_units) {
1815 if (cpup) 1933 if (cpup)
1816 *cpup = cpu; 1934 *cpup = gi->cpu_map[unit - cunit];
1817 return true; 1935 return true;
1818 } 1936 }
1937 cunit += gi->nr_units;
1938 }
1819 1939
1820 return false; 1940 return false;
1821} 1941}
1822 1942
1943static int __init pcpul_cpu_to_unit(int cpu, const struct pcpu_alloc_info *ai)
1944{
1945 int group, unit, i;
1946
1947 for (group = 0, unit = 0; group < ai->nr_groups; group++, unit += i) {
1948 const struct pcpu_group_info *gi = &ai->groups[group];
1949
1950 for (i = 0; i < gi->nr_units; i++)
1951 if (gi->cpu_map[i] == cpu)
1952 return unit + i;
1953 }
1954 BUG();
1955}
1956
1823/** 1957/**
1824 * pcpu_lpage_first_chunk - remap the first percpu chunk using large page 1958 * pcpu_lpage_first_chunk - remap the first percpu chunk using large page
1825 * @reserved_size: the size of reserved percpu area in bytes 1959 * @ai: pcpu_alloc_info
1826 * @dyn_size: free size for dynamic allocation in bytes
1827 * @unit_size: unit size in bytes
1828 * @lpage_size: the size of a large page
1829 * @unit_map: cpu -> unit mapping
1830 * @nr_units: the number of units
1831 * @alloc_fn: function to allocate percpu lpage, always called with lpage_size 1960 * @alloc_fn: function to allocate percpu lpage, always called with lpage_size
1832 * @free_fn: function to free percpu memory, @size <= lpage_size 1961 * @free_fn: function to free percpu memory, @size <= lpage_size
1833 * @map_fn: function to map percpu lpage, always called with lpage_size 1962 * @map_fn: function to map percpu lpage, always called with lpage_size
1834 * 1963 *
1835 * This allocator uses large page to build and map the first chunk. 1964 * This allocator uses large page to build and map the first chunk.
1836 * Unlike other helpers, the caller should always specify @dyn_size 1965 * Unlike other helpers, the caller should provide fully initialized
1837 * and @unit_size. These parameters along with @unit_map and 1966 * @ai. This can be done using pcpu_build_alloc_info(). This two
1838 * @nr_units can be determined using pcpu_lpage_build_unit_map(). 1967 * stage initialization is to allow arch code to evaluate the
1839 * This two stage initialization is to allow arch code to evaluate the
1840 * parameters before committing to it. 1968 * parameters before committing to it.
1841 * 1969 *
1842 * Large pages are allocated as directed by @unit_map and other 1970 * Large pages are allocated as directed by @unit_map and other
@@ -1852,27 +1980,26 @@ static bool __init pcpul_unit_to_cpu(int unit, const int *unit_map,
1852 * The determined pcpu_unit_size which can be used to initialize 1980 * The determined pcpu_unit_size which can be used to initialize
1853 * percpu access on success, -errno on failure. 1981 * percpu access on success, -errno on failure.
1854 */ 1982 */
1855ssize_t __init pcpu_lpage_first_chunk(size_t reserved_size, size_t dyn_size, 1983ssize_t __init pcpu_lpage_first_chunk(const struct pcpu_alloc_info *ai,
1856 size_t unit_size, size_t lpage_size,
1857 const int *unit_map, int nr_units,
1858 pcpu_fc_alloc_fn_t alloc_fn, 1984 pcpu_fc_alloc_fn_t alloc_fn,
1859 pcpu_fc_free_fn_t free_fn, 1985 pcpu_fc_free_fn_t free_fn,
1860 pcpu_fc_map_fn_t map_fn) 1986 pcpu_fc_map_fn_t map_fn)
1861{ 1987{
1862 static struct vm_struct vm; 1988 static struct vm_struct vm;
1863 const size_t static_size = __per_cpu_end - __per_cpu_start; 1989 const size_t lpage_size = ai->atom_size;
1864 size_t chunk_size = unit_size * nr_units; 1990 size_t chunk_size, map_size;
1865 size_t map_size;
1866 unsigned int cpu; 1991 unsigned int cpu;
1867 ssize_t ret; 1992 ssize_t ret;
1868 int i, j, unit; 1993 int i, j, unit, nr_units;
1869 1994
1870 pcpul_lpage_dump_cfg(KERN_DEBUG, static_size, reserved_size, dyn_size, 1995 nr_units = 0;
1871 unit_size, lpage_size, unit_map, nr_units); 1996 for (i = 0; i < ai->nr_groups; i++)
1997 nr_units += ai->groups[i].nr_units;
1872 1998
1999 chunk_size = ai->unit_size * nr_units;
1873 BUG_ON(chunk_size % lpage_size); 2000 BUG_ON(chunk_size % lpage_size);
1874 2001
1875 pcpul_size = static_size + reserved_size + dyn_size; 2002 pcpul_size = ai->static_size + ai->reserved_size + ai->dyn_size;
1876 pcpul_lpage_size = lpage_size; 2003 pcpul_lpage_size = lpage_size;
1877 pcpul_nr_lpages = chunk_size / lpage_size; 2004 pcpul_nr_lpages = chunk_size / lpage_size;
1878 2005
@@ -1883,13 +2010,13 @@ ssize_t __init pcpu_lpage_first_chunk(size_t reserved_size, size_t dyn_size,
1883 /* allocate all pages */ 2010 /* allocate all pages */
1884 for (i = 0; i < pcpul_nr_lpages; i++) { 2011 for (i = 0; i < pcpul_nr_lpages; i++) {
1885 size_t offset = i * lpage_size; 2012 size_t offset = i * lpage_size;
1886 int first_unit = offset / unit_size; 2013 int first_unit = offset / ai->unit_size;
1887 int last_unit = (offset + lpage_size - 1) / unit_size; 2014 int last_unit = (offset + lpage_size - 1) / ai->unit_size;
1888 void *ptr; 2015 void *ptr;
1889 2016
1890 /* find out which cpu is mapped to this unit */ 2017 /* find out which cpu is mapped to this unit */
1891 for (unit = first_unit; unit <= last_unit; unit++) 2018 for (unit = first_unit; unit <= last_unit; unit++)
1892 if (pcpul_unit_to_cpu(unit, unit_map, &cpu)) 2019 if (pcpul_unit_to_cpu(unit, ai, &cpu))
1893 goto found; 2020 goto found;
1894 continue; 2021 continue;
1895 found: 2022 found:
@@ -1905,12 +2032,12 @@ ssize_t __init pcpu_lpage_first_chunk(size_t reserved_size, size_t dyn_size,
1905 2032
1906 /* return unused holes */ 2033 /* return unused holes */
1907 for (unit = 0; unit < nr_units; unit++) { 2034 for (unit = 0; unit < nr_units; unit++) {
1908 size_t start = unit * unit_size; 2035 size_t start = unit * ai->unit_size;
1909 size_t end = start + unit_size; 2036 size_t end = start + ai->unit_size;
1910 size_t off, next; 2037 size_t off, next;
1911 2038
1912 /* don't free used part of occupied unit */ 2039 /* don't free used part of occupied unit */
1913 if (pcpul_unit_to_cpu(unit, unit_map, NULL)) 2040 if (pcpul_unit_to_cpu(unit, ai, NULL))
1914 start += pcpul_size; 2041 start += pcpul_size;
1915 2042
1916 /* unit can span more than one page, punch the holes */ 2043 /* unit can span more than one page, punch the holes */
@@ -1925,7 +2052,7 @@ ssize_t __init pcpu_lpage_first_chunk(size_t reserved_size, size_t dyn_size,
1925 /* allocate address, map and copy */ 2052 /* allocate address, map and copy */
1926 vm.flags = VM_ALLOC; 2053 vm.flags = VM_ALLOC;
1927 vm.size = chunk_size; 2054 vm.size = chunk_size;
1928 vm_area_register_early(&vm, unit_size); 2055 vm_area_register_early(&vm, ai->unit_size);
1929 2056
1930 for (i = 0; i < pcpul_nr_lpages; i++) { 2057 for (i = 0; i < pcpul_nr_lpages; i++) {
1931 if (!pcpul_map[i].ptr) 2058 if (!pcpul_map[i].ptr)
@@ -1935,15 +2062,15 @@ ssize_t __init pcpu_lpage_first_chunk(size_t reserved_size, size_t dyn_size,
1935 } 2062 }
1936 2063
1937 for_each_possible_cpu(cpu) 2064 for_each_possible_cpu(cpu)
1938 memcpy(vm.addr + unit_map[cpu] * unit_size, __per_cpu_load, 2065 memcpy(vm.addr + pcpul_cpu_to_unit(cpu, ai) * ai->unit_size,
1939 static_size); 2066 __per_cpu_load, ai->static_size);
1940 2067
1941 /* we're ready, commit */ 2068 /* we're ready, commit */
1942 pr_info("PERCPU: large pages @%p s%zu r%zu d%zu u%zu\n", 2069 pr_info("PERCPU: large pages @%p s%zu r%zu d%zu u%zu\n",
1943 vm.addr, static_size, reserved_size, dyn_size, unit_size); 2070 vm.addr, ai->static_size, ai->reserved_size, ai->dyn_size,
2071 ai->unit_size);
1944 2072
1945 ret = pcpu_setup_first_chunk(static_size, reserved_size, dyn_size, 2073 ret = pcpu_setup_first_chunk(ai, vm.addr);
1946 unit_size, vm.addr, unit_map);
1947 2074
1948 /* 2075 /*
1949 * Sort pcpul_map array for pcpu_lpage_remapped(). Unmapped 2076 * Sort pcpul_map array for pcpu_lpage_remapped(). Unmapped