diff options
Diffstat (limited to 'arch/ia64/sn/kernel/xpc_partition.c')
-rw-r--r-- | arch/ia64/sn/kernel/xpc_partition.c | 475 |
1 files changed, 356 insertions, 119 deletions
diff --git a/arch/ia64/sn/kernel/xpc_partition.c b/arch/ia64/sn/kernel/xpc_partition.c index 578265ea9e67..581e113d2d37 100644 --- a/arch/ia64/sn/kernel/xpc_partition.c +++ b/arch/ia64/sn/kernel/xpc_partition.c | |||
@@ -44,16 +44,19 @@ static u64 xpc_sh2_IPI_access3; | |||
44 | 44 | ||
45 | 45 | ||
46 | /* original protection values for each node */ | 46 | /* original protection values for each node */ |
47 | u64 xpc_prot_vec[MAX_COMPACT_NODES]; | 47 | u64 xpc_prot_vec[MAX_NUMNODES]; |
48 | 48 | ||
49 | 49 | ||
50 | /* this partition's reserved page */ | 50 | /* this partition's reserved page pointers */ |
51 | struct xpc_rsvd_page *xpc_rsvd_page; | 51 | struct xpc_rsvd_page *xpc_rsvd_page; |
52 | 52 | static u64 *xpc_part_nasids; | |
53 | /* this partition's XPC variables (within the reserved page) */ | 53 | static u64 *xpc_mach_nasids; |
54 | struct xpc_vars *xpc_vars; | 54 | struct xpc_vars *xpc_vars; |
55 | struct xpc_vars_part *xpc_vars_part; | 55 | struct xpc_vars_part *xpc_vars_part; |
56 | 56 | ||
57 | static int xp_nasid_mask_bytes; /* actual size in bytes of nasid mask */ | ||
58 | static int xp_nasid_mask_words; /* actual size in words of nasid mask */ | ||
59 | |||
57 | 60 | ||
58 | /* | 61 | /* |
59 | * For performance reasons, each entry of xpc_partitions[] is cacheline | 62 | * For performance reasons, each entry of xpc_partitions[] is cacheline |
@@ -65,20 +68,16 @@ struct xpc_partition xpc_partitions[XP_MAX_PARTITIONS + 1]; | |||
65 | 68 | ||
66 | 69 | ||
67 | /* | 70 | /* |
68 | * Generic buffer used to store a local copy of the remote partitions | 71 | * Generic buffer used to store a local copy of portions of a remote |
69 | * reserved page or XPC variables. | 72 | * partition's reserved page (either its header and part_nasids mask, |
73 | * or its vars). | ||
70 | * | 74 | * |
71 | * xpc_discovery runs only once and is a seperate thread that is | 75 | * xpc_discovery runs only once and is a seperate thread that is |
72 | * very likely going to be processing in parallel with receiving | 76 | * very likely going to be processing in parallel with receiving |
73 | * interrupts. | 77 | * interrupts. |
74 | */ | 78 | */ |
75 | char ____cacheline_aligned | 79 | char ____cacheline_aligned xpc_remote_copy_buffer[XPC_RP_HEADER_SIZE + |
76 | xpc_remote_copy_buffer[XPC_RSVD_PAGE_ALIGNED_SIZE]; | 80 | XP_NASID_MASK_BYTES]; |
77 | |||
78 | |||
79 | /* systune related variables */ | ||
80 | int xpc_hb_interval = XPC_HB_DEFAULT_INTERVAL; | ||
81 | int xpc_hb_check_interval = XPC_HB_CHECK_DEFAULT_TIMEOUT; | ||
82 | 81 | ||
83 | 82 | ||
84 | /* | 83 | /* |
@@ -86,13 +85,16 @@ int xpc_hb_check_interval = XPC_HB_CHECK_DEFAULT_TIMEOUT; | |||
86 | * for that nasid. This function returns 0 on any error. | 85 | * for that nasid. This function returns 0 on any error. |
87 | */ | 86 | */ |
88 | static u64 | 87 | static u64 |
89 | xpc_get_rsvd_page_pa(int nasid, u64 buf, u64 buf_size) | 88 | xpc_get_rsvd_page_pa(int nasid) |
90 | { | 89 | { |
91 | bte_result_t bte_res; | 90 | bte_result_t bte_res; |
92 | s64 status; | 91 | s64 status; |
93 | u64 cookie = 0; | 92 | u64 cookie = 0; |
94 | u64 rp_pa = nasid; /* seed with nasid */ | 93 | u64 rp_pa = nasid; /* seed with nasid */ |
95 | u64 len = 0; | 94 | u64 len = 0; |
95 | u64 buf = buf; | ||
96 | u64 buf_len = 0; | ||
97 | void *buf_base = NULL; | ||
96 | 98 | ||
97 | 99 | ||
98 | while (1) { | 100 | while (1) { |
@@ -108,13 +110,22 @@ xpc_get_rsvd_page_pa(int nasid, u64 buf, u64 buf_size) | |||
108 | break; | 110 | break; |
109 | } | 111 | } |
110 | 112 | ||
111 | if (len > buf_size) { | 113 | if (L1_CACHE_ALIGN(len) > buf_len) { |
112 | dev_err(xpc_part, "len (=0x%016lx) > buf_size\n", len); | 114 | if (buf_base != NULL) { |
113 | status = SALRET_ERROR; | 115 | kfree(buf_base); |
114 | break; | 116 | } |
117 | buf_len = L1_CACHE_ALIGN(len); | ||
118 | buf = (u64) xpc_kmalloc_cacheline_aligned(buf_len, | ||
119 | GFP_KERNEL, &buf_base); | ||
120 | if (buf_base == NULL) { | ||
121 | dev_err(xpc_part, "unable to kmalloc " | ||
122 | "len=0x%016lx\n", buf_len); | ||
123 | status = SALRET_ERROR; | ||
124 | break; | ||
125 | } | ||
115 | } | 126 | } |
116 | 127 | ||
117 | bte_res = xp_bte_copy(rp_pa, ia64_tpa(buf), buf_size, | 128 | bte_res = xp_bte_copy(rp_pa, ia64_tpa(buf), buf_len, |
118 | (BTE_NOTIFY | BTE_WACQUIRE), NULL); | 129 | (BTE_NOTIFY | BTE_WACQUIRE), NULL); |
119 | if (bte_res != BTE_SUCCESS) { | 130 | if (bte_res != BTE_SUCCESS) { |
120 | dev_dbg(xpc_part, "xp_bte_copy failed %i\n", bte_res); | 131 | dev_dbg(xpc_part, "xp_bte_copy failed %i\n", bte_res); |
@@ -123,6 +134,10 @@ xpc_get_rsvd_page_pa(int nasid, u64 buf, u64 buf_size) | |||
123 | } | 134 | } |
124 | } | 135 | } |
125 | 136 | ||
137 | if (buf_base != NULL) { | ||
138 | kfree(buf_base); | ||
139 | } | ||
140 | |||
126 | if (status != SALRET_OK) { | 141 | if (status != SALRET_OK) { |
127 | rp_pa = 0; | 142 | rp_pa = 0; |
128 | } | 143 | } |
@@ -141,15 +156,15 @@ xpc_rsvd_page_init(void) | |||
141 | { | 156 | { |
142 | struct xpc_rsvd_page *rp; | 157 | struct xpc_rsvd_page *rp; |
143 | AMO_t *amos_page; | 158 | AMO_t *amos_page; |
144 | u64 rp_pa, next_cl, nasid_array = 0; | 159 | u64 rp_pa, nasid_array = 0; |
145 | int i, ret; | 160 | int i, ret; |
146 | 161 | ||
147 | 162 | ||
148 | /* get the local reserved page's address */ | 163 | /* get the local reserved page's address */ |
149 | 164 | ||
150 | rp_pa = xpc_get_rsvd_page_pa(cnodeid_to_nasid(0), | 165 | preempt_disable(); |
151 | (u64) xpc_remote_copy_buffer, | 166 | rp_pa = xpc_get_rsvd_page_pa(cpuid_to_nasid(smp_processor_id())); |
152 | XPC_RSVD_PAGE_ALIGNED_SIZE); | 167 | preempt_enable(); |
153 | if (rp_pa == 0) { | 168 | if (rp_pa == 0) { |
154 | dev_err(xpc_part, "SAL failed to locate the reserved page\n"); | 169 | dev_err(xpc_part, "SAL failed to locate the reserved page\n"); |
155 | return NULL; | 170 | return NULL; |
@@ -164,12 +179,19 @@ xpc_rsvd_page_init(void) | |||
164 | 179 | ||
165 | rp->version = XPC_RP_VERSION; | 180 | rp->version = XPC_RP_VERSION; |
166 | 181 | ||
167 | /* | 182 | /* establish the actual sizes of the nasid masks */ |
168 | * Place the XPC variables on the cache line following the | 183 | if (rp->SAL_version == 1) { |
169 | * reserved page structure. | 184 | /* SAL_version 1 didn't set the nasids_size field */ |
170 | */ | 185 | rp->nasids_size = 128; |
171 | next_cl = (u64) rp + XPC_RSVD_PAGE_ALIGNED_SIZE; | 186 | } |
172 | xpc_vars = (struct xpc_vars *) next_cl; | 187 | xp_nasid_mask_bytes = rp->nasids_size; |
188 | xp_nasid_mask_words = xp_nasid_mask_bytes / 8; | ||
189 | |||
190 | /* setup the pointers to the various items in the reserved page */ | ||
191 | xpc_part_nasids = XPC_RP_PART_NASIDS(rp); | ||
192 | xpc_mach_nasids = XPC_RP_MACH_NASIDS(rp); | ||
193 | xpc_vars = XPC_RP_VARS(rp); | ||
194 | xpc_vars_part = XPC_RP_VARS_PART(rp); | ||
173 | 195 | ||
174 | /* | 196 | /* |
175 | * Before clearing xpc_vars, see if a page of AMOs had been previously | 197 | * Before clearing xpc_vars, see if a page of AMOs had been previously |
@@ -221,33 +243,32 @@ xpc_rsvd_page_init(void) | |||
221 | amos_page = (AMO_t *) TO_AMO((u64) amos_page); | 243 | amos_page = (AMO_t *) TO_AMO((u64) amos_page); |
222 | } | 244 | } |
223 | 245 | ||
246 | /* clear xpc_vars */ | ||
224 | memset(xpc_vars, 0, sizeof(struct xpc_vars)); | 247 | memset(xpc_vars, 0, sizeof(struct xpc_vars)); |
225 | 248 | ||
226 | /* | ||
227 | * Place the XPC per partition specific variables on the cache line | ||
228 | * following the XPC variables structure. | ||
229 | */ | ||
230 | next_cl += XPC_VARS_ALIGNED_SIZE; | ||
231 | memset((u64 *) next_cl, 0, sizeof(struct xpc_vars_part) * | ||
232 | XP_MAX_PARTITIONS); | ||
233 | xpc_vars_part = (struct xpc_vars_part *) next_cl; | ||
234 | xpc_vars->vars_part_pa = __pa(next_cl); | ||
235 | |||
236 | xpc_vars->version = XPC_V_VERSION; | 249 | xpc_vars->version = XPC_V_VERSION; |
237 | xpc_vars->act_nasid = cpuid_to_nasid(0); | 250 | xpc_vars->act_nasid = cpuid_to_nasid(0); |
238 | xpc_vars->act_phys_cpuid = cpu_physical_id(0); | 251 | xpc_vars->act_phys_cpuid = cpu_physical_id(0); |
252 | xpc_vars->vars_part_pa = __pa(xpc_vars_part); | ||
253 | xpc_vars->amos_page_pa = ia64_tpa((u64) amos_page); | ||
239 | xpc_vars->amos_page = amos_page; /* save for next load of XPC */ | 254 | xpc_vars->amos_page = amos_page; /* save for next load of XPC */ |
240 | 255 | ||
241 | 256 | ||
242 | /* | 257 | /* clear xpc_vars_part */ |
243 | * Initialize the activation related AMO variables. | 258 | memset((u64 *) xpc_vars_part, 0, sizeof(struct xpc_vars_part) * |
244 | */ | 259 | XP_MAX_PARTITIONS); |
245 | xpc_vars->act_amos = xpc_IPI_init(XP_MAX_PARTITIONS); | 260 | |
246 | for (i = 1; i < XP_NASID_MASK_WORDS; i++) { | 261 | /* initialize the activate IRQ related AMO variables */ |
247 | xpc_IPI_init(i + XP_MAX_PARTITIONS); | 262 | for (i = 0; i < xp_nasid_mask_words; i++) { |
263 | (void) xpc_IPI_init(XPC_ACTIVATE_IRQ_AMOS + i); | ||
248 | } | 264 | } |
249 | /* export AMO page's physical address to other partitions */ | 265 | |
250 | xpc_vars->amos_page_pa = ia64_tpa((u64) xpc_vars->amos_page); | 266 | /* initialize the engaged remote partitions related AMO variables */ |
267 | (void) xpc_IPI_init(XPC_ENGAGED_PARTITIONS_AMO); | ||
268 | (void) xpc_IPI_init(XPC_DISENGAGE_REQUEST_AMO); | ||
269 | |||
270 | /* timestamp of when reserved page was setup by XPC */ | ||
271 | rp->stamp = CURRENT_TIME; | ||
251 | 272 | ||
252 | /* | 273 | /* |
253 | * This signifies to the remote partition that our reserved | 274 | * This signifies to the remote partition that our reserved |
@@ -387,6 +408,11 @@ xpc_check_remote_hb(void) | |||
387 | remote_vars = (struct xpc_vars *) xpc_remote_copy_buffer; | 408 | remote_vars = (struct xpc_vars *) xpc_remote_copy_buffer; |
388 | 409 | ||
389 | for (partid = 1; partid < XP_MAX_PARTITIONS; partid++) { | 410 | for (partid = 1; partid < XP_MAX_PARTITIONS; partid++) { |
411 | |||
412 | if (xpc_exiting) { | ||
413 | break; | ||
414 | } | ||
415 | |||
390 | if (partid == sn_partition_id) { | 416 | if (partid == sn_partition_id) { |
391 | continue; | 417 | continue; |
392 | } | 418 | } |
@@ -401,7 +427,7 @@ xpc_check_remote_hb(void) | |||
401 | /* pull the remote_hb cache line */ | 427 | /* pull the remote_hb cache line */ |
402 | bres = xp_bte_copy(part->remote_vars_pa, | 428 | bres = xp_bte_copy(part->remote_vars_pa, |
403 | ia64_tpa((u64) remote_vars), | 429 | ia64_tpa((u64) remote_vars), |
404 | XPC_VARS_ALIGNED_SIZE, | 430 | XPC_RP_VARS_SIZE, |
405 | (BTE_NOTIFY | BTE_WACQUIRE), NULL); | 431 | (BTE_NOTIFY | BTE_WACQUIRE), NULL); |
406 | if (bres != BTE_SUCCESS) { | 432 | if (bres != BTE_SUCCESS) { |
407 | XPC_DEACTIVATE_PARTITION(part, | 433 | XPC_DEACTIVATE_PARTITION(part, |
@@ -417,7 +443,7 @@ xpc_check_remote_hb(void) | |||
417 | 443 | ||
418 | if (((remote_vars->heartbeat == part->last_heartbeat) && | 444 | if (((remote_vars->heartbeat == part->last_heartbeat) && |
419 | (remote_vars->kdb_status == 0)) || | 445 | (remote_vars->kdb_status == 0)) || |
420 | !XPC_HB_ALLOWED(sn_partition_id, remote_vars)) { | 446 | !xpc_hb_allowed(sn_partition_id, remote_vars)) { |
421 | 447 | ||
422 | XPC_DEACTIVATE_PARTITION(part, xpcNoHeartbeat); | 448 | XPC_DEACTIVATE_PARTITION(part, xpcNoHeartbeat); |
423 | continue; | 449 | continue; |
@@ -429,31 +455,31 @@ xpc_check_remote_hb(void) | |||
429 | 455 | ||
430 | 456 | ||
431 | /* | 457 | /* |
432 | * Get a copy of the remote partition's rsvd page. | 458 | * Get a copy of a portion of the remote partition's rsvd page. |
433 | * | 459 | * |
434 | * remote_rp points to a buffer that is cacheline aligned for BTE copies and | 460 | * remote_rp points to a buffer that is cacheline aligned for BTE copies and |
435 | * assumed to be of size XPC_RSVD_PAGE_ALIGNED_SIZE. | 461 | * is large enough to contain a copy of their reserved page header and |
462 | * part_nasids mask. | ||
436 | */ | 463 | */ |
437 | static enum xpc_retval | 464 | static enum xpc_retval |
438 | xpc_get_remote_rp(int nasid, u64 *discovered_nasids, | 465 | xpc_get_remote_rp(int nasid, u64 *discovered_nasids, |
439 | struct xpc_rsvd_page *remote_rp, u64 *remote_rsvd_page_pa) | 466 | struct xpc_rsvd_page *remote_rp, u64 *remote_rp_pa) |
440 | { | 467 | { |
441 | int bres, i; | 468 | int bres, i; |
442 | 469 | ||
443 | 470 | ||
444 | /* get the reserved page's physical address */ | 471 | /* get the reserved page's physical address */ |
445 | 472 | ||
446 | *remote_rsvd_page_pa = xpc_get_rsvd_page_pa(nasid, (u64) remote_rp, | 473 | *remote_rp_pa = xpc_get_rsvd_page_pa(nasid); |
447 | XPC_RSVD_PAGE_ALIGNED_SIZE); | 474 | if (*remote_rp_pa == 0) { |
448 | if (*remote_rsvd_page_pa == 0) { | ||
449 | return xpcNoRsvdPageAddr; | 475 | return xpcNoRsvdPageAddr; |
450 | } | 476 | } |
451 | 477 | ||
452 | 478 | ||
453 | /* pull over the reserved page structure */ | 479 | /* pull over the reserved page header and part_nasids mask */ |
454 | 480 | ||
455 | bres = xp_bte_copy(*remote_rsvd_page_pa, ia64_tpa((u64) remote_rp), | 481 | bres = xp_bte_copy(*remote_rp_pa, ia64_tpa((u64) remote_rp), |
456 | XPC_RSVD_PAGE_ALIGNED_SIZE, | 482 | XPC_RP_HEADER_SIZE + xp_nasid_mask_bytes, |
457 | (BTE_NOTIFY | BTE_WACQUIRE), NULL); | 483 | (BTE_NOTIFY | BTE_WACQUIRE), NULL); |
458 | if (bres != BTE_SUCCESS) { | 484 | if (bres != BTE_SUCCESS) { |
459 | return xpc_map_bte_errors(bres); | 485 | return xpc_map_bte_errors(bres); |
@@ -461,8 +487,11 @@ xpc_get_remote_rp(int nasid, u64 *discovered_nasids, | |||
461 | 487 | ||
462 | 488 | ||
463 | if (discovered_nasids != NULL) { | 489 | if (discovered_nasids != NULL) { |
464 | for (i = 0; i < XP_NASID_MASK_WORDS; i++) { | 490 | u64 *remote_part_nasids = XPC_RP_PART_NASIDS(remote_rp); |
465 | discovered_nasids[i] |= remote_rp->part_nasids[i]; | 491 | |
492 | |||
493 | for (i = 0; i < xp_nasid_mask_words; i++) { | ||
494 | discovered_nasids[i] |= remote_part_nasids[i]; | ||
466 | } | 495 | } |
467 | } | 496 | } |
468 | 497 | ||
@@ -489,10 +518,10 @@ xpc_get_remote_rp(int nasid, u64 *discovered_nasids, | |||
489 | 518 | ||
490 | 519 | ||
491 | /* | 520 | /* |
492 | * Get a copy of the remote partition's XPC variables. | 521 | * Get a copy of the remote partition's XPC variables from the reserved page. |
493 | * | 522 | * |
494 | * remote_vars points to a buffer that is cacheline aligned for BTE copies and | 523 | * remote_vars points to a buffer that is cacheline aligned for BTE copies and |
495 | * assumed to be of size XPC_VARS_ALIGNED_SIZE. | 524 | * assumed to be of size XPC_RP_VARS_SIZE. |
496 | */ | 525 | */ |
497 | static enum xpc_retval | 526 | static enum xpc_retval |
498 | xpc_get_remote_vars(u64 remote_vars_pa, struct xpc_vars *remote_vars) | 527 | xpc_get_remote_vars(u64 remote_vars_pa, struct xpc_vars *remote_vars) |
@@ -508,7 +537,7 @@ xpc_get_remote_vars(u64 remote_vars_pa, struct xpc_vars *remote_vars) | |||
508 | /* pull over the cross partition variables */ | 537 | /* pull over the cross partition variables */ |
509 | 538 | ||
510 | bres = xp_bte_copy(remote_vars_pa, ia64_tpa((u64) remote_vars), | 539 | bres = xp_bte_copy(remote_vars_pa, ia64_tpa((u64) remote_vars), |
511 | XPC_VARS_ALIGNED_SIZE, | 540 | XPC_RP_VARS_SIZE, |
512 | (BTE_NOTIFY | BTE_WACQUIRE), NULL); | 541 | (BTE_NOTIFY | BTE_WACQUIRE), NULL); |
513 | if (bres != BTE_SUCCESS) { | 542 | if (bres != BTE_SUCCESS) { |
514 | return xpc_map_bte_errors(bres); | 543 | return xpc_map_bte_errors(bres); |
@@ -524,7 +553,56 @@ xpc_get_remote_vars(u64 remote_vars_pa, struct xpc_vars *remote_vars) | |||
524 | 553 | ||
525 | 554 | ||
526 | /* | 555 | /* |
527 | * Prior code has determine the nasid which generated an IPI. Inspect | 556 | * Update the remote partition's info. |
557 | */ | ||
558 | static void | ||
559 | xpc_update_partition_info(struct xpc_partition *part, u8 remote_rp_version, | ||
560 | struct timespec *remote_rp_stamp, u64 remote_rp_pa, | ||
561 | u64 remote_vars_pa, struct xpc_vars *remote_vars) | ||
562 | { | ||
563 | part->remote_rp_version = remote_rp_version; | ||
564 | dev_dbg(xpc_part, " remote_rp_version = 0x%016lx\n", | ||
565 | part->remote_rp_version); | ||
566 | |||
567 | part->remote_rp_stamp = *remote_rp_stamp; | ||
568 | dev_dbg(xpc_part, " remote_rp_stamp (tv_sec = 0x%lx tv_nsec = 0x%lx\n", | ||
569 | part->remote_rp_stamp.tv_sec, part->remote_rp_stamp.tv_nsec); | ||
570 | |||
571 | part->remote_rp_pa = remote_rp_pa; | ||
572 | dev_dbg(xpc_part, " remote_rp_pa = 0x%016lx\n", part->remote_rp_pa); | ||
573 | |||
574 | part->remote_vars_pa = remote_vars_pa; | ||
575 | dev_dbg(xpc_part, " remote_vars_pa = 0x%016lx\n", | ||
576 | part->remote_vars_pa); | ||
577 | |||
578 | part->last_heartbeat = remote_vars->heartbeat; | ||
579 | dev_dbg(xpc_part, " last_heartbeat = 0x%016lx\n", | ||
580 | part->last_heartbeat); | ||
581 | |||
582 | part->remote_vars_part_pa = remote_vars->vars_part_pa; | ||
583 | dev_dbg(xpc_part, " remote_vars_part_pa = 0x%016lx\n", | ||
584 | part->remote_vars_part_pa); | ||
585 | |||
586 | part->remote_act_nasid = remote_vars->act_nasid; | ||
587 | dev_dbg(xpc_part, " remote_act_nasid = 0x%x\n", | ||
588 | part->remote_act_nasid); | ||
589 | |||
590 | part->remote_act_phys_cpuid = remote_vars->act_phys_cpuid; | ||
591 | dev_dbg(xpc_part, " remote_act_phys_cpuid = 0x%x\n", | ||
592 | part->remote_act_phys_cpuid); | ||
593 | |||
594 | part->remote_amos_page_pa = remote_vars->amos_page_pa; | ||
595 | dev_dbg(xpc_part, " remote_amos_page_pa = 0x%lx\n", | ||
596 | part->remote_amos_page_pa); | ||
597 | |||
598 | part->remote_vars_version = remote_vars->version; | ||
599 | dev_dbg(xpc_part, " remote_vars_version = 0x%x\n", | ||
600 | part->remote_vars_version); | ||
601 | } | ||
602 | |||
603 | |||
604 | /* | ||
605 | * Prior code has determined the nasid which generated an IPI. Inspect | ||
528 | * that nasid to determine if its partition needs to be activated or | 606 | * that nasid to determine if its partition needs to be activated or |
529 | * deactivated. | 607 | * deactivated. |
530 | * | 608 | * |
@@ -542,8 +620,12 @@ xpc_identify_act_IRQ_req(int nasid) | |||
542 | { | 620 | { |
543 | struct xpc_rsvd_page *remote_rp; | 621 | struct xpc_rsvd_page *remote_rp; |
544 | struct xpc_vars *remote_vars; | 622 | struct xpc_vars *remote_vars; |
545 | u64 remote_rsvd_page_pa; | 623 | u64 remote_rp_pa; |
546 | u64 remote_vars_pa; | 624 | u64 remote_vars_pa; |
625 | int remote_rp_version; | ||
626 | int reactivate = 0; | ||
627 | int stamp_diff; | ||
628 | struct timespec remote_rp_stamp = { 0, 0 }; | ||
547 | partid_t partid; | 629 | partid_t partid; |
548 | struct xpc_partition *part; | 630 | struct xpc_partition *part; |
549 | enum xpc_retval ret; | 631 | enum xpc_retval ret; |
@@ -553,7 +635,7 @@ xpc_identify_act_IRQ_req(int nasid) | |||
553 | 635 | ||
554 | remote_rp = (struct xpc_rsvd_page *) xpc_remote_copy_buffer; | 636 | remote_rp = (struct xpc_rsvd_page *) xpc_remote_copy_buffer; |
555 | 637 | ||
556 | ret = xpc_get_remote_rp(nasid, NULL, remote_rp, &remote_rsvd_page_pa); | 638 | ret = xpc_get_remote_rp(nasid, NULL, remote_rp, &remote_rp_pa); |
557 | if (ret != xpcSuccess) { | 639 | if (ret != xpcSuccess) { |
558 | dev_warn(xpc_part, "unable to get reserved page from nasid %d, " | 640 | dev_warn(xpc_part, "unable to get reserved page from nasid %d, " |
559 | "which sent interrupt, reason=%d\n", nasid, ret); | 641 | "which sent interrupt, reason=%d\n", nasid, ret); |
@@ -561,6 +643,10 @@ xpc_identify_act_IRQ_req(int nasid) | |||
561 | } | 643 | } |
562 | 644 | ||
563 | remote_vars_pa = remote_rp->vars_pa; | 645 | remote_vars_pa = remote_rp->vars_pa; |
646 | remote_rp_version = remote_rp->version; | ||
647 | if (XPC_SUPPORTS_RP_STAMP(remote_rp_version)) { | ||
648 | remote_rp_stamp = remote_rp->stamp; | ||
649 | } | ||
564 | partid = remote_rp->partid; | 650 | partid = remote_rp->partid; |
565 | part = &xpc_partitions[partid]; | 651 | part = &xpc_partitions[partid]; |
566 | 652 | ||
@@ -586,44 +672,117 @@ xpc_identify_act_IRQ_req(int nasid) | |||
586 | "%ld:0x%lx\n", (int) nasid, (int) partid, part->act_IRQ_rcvd, | 672 | "%ld:0x%lx\n", (int) nasid, (int) partid, part->act_IRQ_rcvd, |
587 | remote_vars->heartbeat, remote_vars->heartbeating_to_mask); | 673 | remote_vars->heartbeat, remote_vars->heartbeating_to_mask); |
588 | 674 | ||
675 | if (xpc_partition_disengaged(part) && | ||
676 | part->act_state == XPC_P_INACTIVE) { | ||
589 | 677 | ||
590 | if (part->act_state == XPC_P_INACTIVE) { | 678 | xpc_update_partition_info(part, remote_rp_version, |
679 | &remote_rp_stamp, remote_rp_pa, | ||
680 | remote_vars_pa, remote_vars); | ||
591 | 681 | ||
592 | part->remote_rp_pa = remote_rsvd_page_pa; | 682 | if (XPC_SUPPORTS_DISENGAGE_REQUEST(part->remote_vars_version)) { |
593 | dev_dbg(xpc_part, " remote_rp_pa = 0x%016lx\n", | 683 | if (xpc_partition_disengage_requested(1UL << partid)) { |
594 | part->remote_rp_pa); | 684 | /* |
685 | * Other side is waiting on us to disengage, | ||
686 | * even though we already have. | ||
687 | */ | ||
688 | return; | ||
689 | } | ||
690 | } else { | ||
691 | /* other side doesn't support disengage requests */ | ||
692 | xpc_clear_partition_disengage_request(1UL << partid); | ||
693 | } | ||
595 | 694 | ||
596 | part->remote_vars_pa = remote_vars_pa; | 695 | xpc_activate_partition(part); |
597 | dev_dbg(xpc_part, " remote_vars_pa = 0x%016lx\n", | 696 | return; |
598 | part->remote_vars_pa); | 697 | } |
599 | 698 | ||
600 | part->last_heartbeat = remote_vars->heartbeat; | 699 | DBUG_ON(part->remote_rp_version == 0); |
601 | dev_dbg(xpc_part, " last_heartbeat = 0x%016lx\n", | 700 | DBUG_ON(part->remote_vars_version == 0); |
602 | part->last_heartbeat); | 701 | |
702 | if (!XPC_SUPPORTS_RP_STAMP(part->remote_rp_version)) { | ||
703 | DBUG_ON(XPC_SUPPORTS_DISENGAGE_REQUEST(part-> | ||
704 | remote_vars_version)); | ||
705 | |||
706 | if (!XPC_SUPPORTS_RP_STAMP(remote_rp_version)) { | ||
707 | DBUG_ON(XPC_SUPPORTS_DISENGAGE_REQUEST(remote_vars-> | ||
708 | version)); | ||
709 | /* see if the other side rebooted */ | ||
710 | if (part->remote_amos_page_pa == | ||
711 | remote_vars->amos_page_pa && | ||
712 | xpc_hb_allowed(sn_partition_id, | ||
713 | remote_vars)) { | ||
714 | /* doesn't look that way, so ignore the IPI */ | ||
715 | return; | ||
716 | } | ||
717 | } | ||
603 | 718 | ||
604 | part->remote_vars_part_pa = remote_vars->vars_part_pa; | 719 | /* |
605 | dev_dbg(xpc_part, " remote_vars_part_pa = 0x%016lx\n", | 720 | * Other side rebooted and previous XPC didn't support the |
606 | part->remote_vars_part_pa); | 721 | * disengage request, so we don't need to do anything special. |
722 | */ | ||
607 | 723 | ||
608 | part->remote_act_nasid = remote_vars->act_nasid; | 724 | xpc_update_partition_info(part, remote_rp_version, |
609 | dev_dbg(xpc_part, " remote_act_nasid = 0x%x\n", | 725 | &remote_rp_stamp, remote_rp_pa, |
610 | part->remote_act_nasid); | 726 | remote_vars_pa, remote_vars); |
727 | part->reactivate_nasid = nasid; | ||
728 | XPC_DEACTIVATE_PARTITION(part, xpcReactivating); | ||
729 | return; | ||
730 | } | ||
611 | 731 | ||
612 | part->remote_act_phys_cpuid = remote_vars->act_phys_cpuid; | 732 | DBUG_ON(!XPC_SUPPORTS_DISENGAGE_REQUEST(part->remote_vars_version)); |
613 | dev_dbg(xpc_part, " remote_act_phys_cpuid = 0x%x\n", | ||
614 | part->remote_act_phys_cpuid); | ||
615 | 733 | ||
616 | part->remote_amos_page_pa = remote_vars->amos_page_pa; | 734 | if (!XPC_SUPPORTS_RP_STAMP(remote_rp_version)) { |
617 | dev_dbg(xpc_part, " remote_amos_page_pa = 0x%lx\n", | 735 | DBUG_ON(!XPC_SUPPORTS_DISENGAGE_REQUEST(remote_vars->version)); |
618 | part->remote_amos_page_pa); | ||
619 | 736 | ||
620 | xpc_activate_partition(part); | 737 | /* |
738 | * Other side rebooted and previous XPC did support the | ||
739 | * disengage request, but the new one doesn't. | ||
740 | */ | ||
741 | |||
742 | xpc_clear_partition_engaged(1UL << partid); | ||
743 | xpc_clear_partition_disengage_request(1UL << partid); | ||
621 | 744 | ||
622 | } else if (part->remote_amos_page_pa != remote_vars->amos_page_pa || | 745 | xpc_update_partition_info(part, remote_rp_version, |
623 | !XPC_HB_ALLOWED(sn_partition_id, remote_vars)) { | 746 | &remote_rp_stamp, remote_rp_pa, |
747 | remote_vars_pa, remote_vars); | ||
748 | reactivate = 1; | ||
749 | |||
750 | } else { | ||
751 | DBUG_ON(!XPC_SUPPORTS_DISENGAGE_REQUEST(remote_vars->version)); | ||
624 | 752 | ||
753 | stamp_diff = xpc_compare_stamps(&part->remote_rp_stamp, | ||
754 | &remote_rp_stamp); | ||
755 | if (stamp_diff != 0) { | ||
756 | DBUG_ON(stamp_diff >= 0); | ||
757 | |||
758 | /* | ||
759 | * Other side rebooted and the previous XPC did support | ||
760 | * the disengage request, as does the new one. | ||
761 | */ | ||
762 | |||
763 | DBUG_ON(xpc_partition_engaged(1UL << partid)); | ||
764 | DBUG_ON(xpc_partition_disengage_requested(1UL << | ||
765 | partid)); | ||
766 | |||
767 | xpc_update_partition_info(part, remote_rp_version, | ||
768 | &remote_rp_stamp, remote_rp_pa, | ||
769 | remote_vars_pa, remote_vars); | ||
770 | reactivate = 1; | ||
771 | } | ||
772 | } | ||
773 | |||
774 | if (!xpc_partition_disengaged(part)) { | ||
775 | /* still waiting on other side to disengage from us */ | ||
776 | return; | ||
777 | } | ||
778 | |||
779 | if (reactivate) { | ||
625 | part->reactivate_nasid = nasid; | 780 | part->reactivate_nasid = nasid; |
626 | XPC_DEACTIVATE_PARTITION(part, xpcReactivating); | 781 | XPC_DEACTIVATE_PARTITION(part, xpcReactivating); |
782 | |||
783 | } else if (XPC_SUPPORTS_DISENGAGE_REQUEST(part->remote_vars_version) && | ||
784 | xpc_partition_disengage_requested(1UL << partid)) { | ||
785 | XPC_DEACTIVATE_PARTITION(part, xpcOtherGoingDown); | ||
627 | } | 786 | } |
628 | } | 787 | } |
629 | 788 | ||
@@ -643,14 +802,17 @@ xpc_identify_act_IRQ_sender(void) | |||
643 | u64 nasid; /* remote nasid */ | 802 | u64 nasid; /* remote nasid */ |
644 | int n_IRQs_detected = 0; | 803 | int n_IRQs_detected = 0; |
645 | AMO_t *act_amos; | 804 | AMO_t *act_amos; |
646 | struct xpc_rsvd_page *rp = (struct xpc_rsvd_page *) xpc_rsvd_page; | ||
647 | 805 | ||
648 | 806 | ||
649 | act_amos = xpc_vars->act_amos; | 807 | act_amos = xpc_vars->amos_page + XPC_ACTIVATE_IRQ_AMOS; |
650 | 808 | ||
651 | 809 | ||
652 | /* scan through act AMO variable looking for non-zero entries */ | 810 | /* scan through act AMO variable looking for non-zero entries */ |
653 | for (word = 0; word < XP_NASID_MASK_WORDS; word++) { | 811 | for (word = 0; word < xp_nasid_mask_words; word++) { |
812 | |||
813 | if (xpc_exiting) { | ||
814 | break; | ||
815 | } | ||
654 | 816 | ||
655 | nasid_mask = xpc_IPI_receive(&act_amos[word]); | 817 | nasid_mask = xpc_IPI_receive(&act_amos[word]); |
656 | if (nasid_mask == 0) { | 818 | if (nasid_mask == 0) { |
@@ -668,7 +830,7 @@ xpc_identify_act_IRQ_sender(void) | |||
668 | * remote nasid in our reserved pages machine mask. | 830 | * remote nasid in our reserved pages machine mask. |
669 | * This is used in the event of module reload. | 831 | * This is used in the event of module reload. |
670 | */ | 832 | */ |
671 | rp->mach_nasids[word] |= nasid_mask; | 833 | xpc_mach_nasids[word] |= nasid_mask; |
672 | 834 | ||
673 | 835 | ||
674 | /* locate the nasid(s) which sent interrupts */ | 836 | /* locate the nasid(s) which sent interrupts */ |
@@ -688,6 +850,55 @@ xpc_identify_act_IRQ_sender(void) | |||
688 | 850 | ||
689 | 851 | ||
690 | /* | 852 | /* |
853 | * See if the other side has responded to a partition disengage request | ||
854 | * from us. | ||
855 | */ | ||
856 | int | ||
857 | xpc_partition_disengaged(struct xpc_partition *part) | ||
858 | { | ||
859 | partid_t partid = XPC_PARTID(part); | ||
860 | int disengaged; | ||
861 | |||
862 | |||
863 | disengaged = (xpc_partition_engaged(1UL << partid) == 0); | ||
864 | if (part->disengage_request_timeout) { | ||
865 | if (!disengaged) { | ||
866 | if (jiffies < part->disengage_request_timeout) { | ||
867 | /* timelimit hasn't been reached yet */ | ||
868 | return 0; | ||
869 | } | ||
870 | |||
871 | /* | ||
872 | * Other side hasn't responded to our disengage | ||
873 | * request in a timely fashion, so assume it's dead. | ||
874 | */ | ||
875 | |||
876 | xpc_clear_partition_engaged(1UL << partid); | ||
877 | disengaged = 1; | ||
878 | } | ||
879 | part->disengage_request_timeout = 0; | ||
880 | |||
881 | /* cancel the timer function, provided it's not us */ | ||
882 | if (!in_interrupt()) { | ||
883 | del_singleshot_timer_sync(&part-> | ||
884 | disengage_request_timer); | ||
885 | } | ||
886 | |||
887 | DBUG_ON(part->act_state != XPC_P_DEACTIVATING && | ||
888 | part->act_state != XPC_P_INACTIVE); | ||
889 | if (part->act_state != XPC_P_INACTIVE) { | ||
890 | xpc_wakeup_channel_mgr(part); | ||
891 | } | ||
892 | |||
893 | if (XPC_SUPPORTS_DISENGAGE_REQUEST(part->remote_vars_version)) { | ||
894 | xpc_cancel_partition_disengage_request(part); | ||
895 | } | ||
896 | } | ||
897 | return disengaged; | ||
898 | } | ||
899 | |||
900 | |||
901 | /* | ||
691 | * Mark specified partition as active. | 902 | * Mark specified partition as active. |
692 | */ | 903 | */ |
693 | enum xpc_retval | 904 | enum xpc_retval |
@@ -721,7 +932,6 @@ xpc_deactivate_partition(const int line, struct xpc_partition *part, | |||
721 | enum xpc_retval reason) | 932 | enum xpc_retval reason) |
722 | { | 933 | { |
723 | unsigned long irq_flags; | 934 | unsigned long irq_flags; |
724 | partid_t partid = XPC_PARTID(part); | ||
725 | 935 | ||
726 | 936 | ||
727 | spin_lock_irqsave(&part->act_lock, irq_flags); | 937 | spin_lock_irqsave(&part->act_lock, irq_flags); |
@@ -749,17 +959,27 @@ xpc_deactivate_partition(const int line, struct xpc_partition *part, | |||
749 | 959 | ||
750 | spin_unlock_irqrestore(&part->act_lock, irq_flags); | 960 | spin_unlock_irqrestore(&part->act_lock, irq_flags); |
751 | 961 | ||
752 | XPC_DISALLOW_HB(partid, xpc_vars); | 962 | if (XPC_SUPPORTS_DISENGAGE_REQUEST(part->remote_vars_version)) { |
963 | xpc_request_partition_disengage(part); | ||
964 | xpc_IPI_send_disengage(part); | ||
753 | 965 | ||
754 | dev_dbg(xpc_part, "bringing partition %d down, reason = %d\n", partid, | 966 | /* set a timelimit on the disengage request */ |
755 | reason); | 967 | part->disengage_request_timeout = jiffies + |
968 | (xpc_disengage_request_timelimit * HZ); | ||
969 | part->disengage_request_timer.expires = | ||
970 | part->disengage_request_timeout; | ||
971 | add_timer(&part->disengage_request_timer); | ||
972 | } | ||
973 | |||
974 | dev_dbg(xpc_part, "bringing partition %d down, reason = %d\n", | ||
975 | XPC_PARTID(part), reason); | ||
756 | 976 | ||
757 | xpc_partition_down(part, reason); | 977 | xpc_partition_going_down(part, reason); |
758 | } | 978 | } |
759 | 979 | ||
760 | 980 | ||
761 | /* | 981 | /* |
762 | * Mark specified partition as active. | 982 | * Mark specified partition as inactive. |
763 | */ | 983 | */ |
764 | void | 984 | void |
765 | xpc_mark_partition_inactive(struct xpc_partition *part) | 985 | xpc_mark_partition_inactive(struct xpc_partition *part) |
@@ -792,9 +1012,10 @@ xpc_discovery(void) | |||
792 | void *remote_rp_base; | 1012 | void *remote_rp_base; |
793 | struct xpc_rsvd_page *remote_rp; | 1013 | struct xpc_rsvd_page *remote_rp; |
794 | struct xpc_vars *remote_vars; | 1014 | struct xpc_vars *remote_vars; |
795 | u64 remote_rsvd_page_pa; | 1015 | u64 remote_rp_pa; |
796 | u64 remote_vars_pa; | 1016 | u64 remote_vars_pa; |
797 | int region; | 1017 | int region; |
1018 | int region_size; | ||
798 | int max_regions; | 1019 | int max_regions; |
799 | int nasid; | 1020 | int nasid; |
800 | struct xpc_rsvd_page *rp; | 1021 | struct xpc_rsvd_page *rp; |
@@ -804,7 +1025,8 @@ xpc_discovery(void) | |||
804 | enum xpc_retval ret; | 1025 | enum xpc_retval ret; |
805 | 1026 | ||
806 | 1027 | ||
807 | remote_rp = xpc_kmalloc_cacheline_aligned(XPC_RSVD_PAGE_ALIGNED_SIZE, | 1028 | remote_rp = xpc_kmalloc_cacheline_aligned(XPC_RP_HEADER_SIZE + |
1029 | xp_nasid_mask_bytes, | ||
808 | GFP_KERNEL, &remote_rp_base); | 1030 | GFP_KERNEL, &remote_rp_base); |
809 | if (remote_rp == NULL) { | 1031 | if (remote_rp == NULL) { |
810 | return; | 1032 | return; |
@@ -812,13 +1034,13 @@ xpc_discovery(void) | |||
812 | remote_vars = (struct xpc_vars *) remote_rp; | 1034 | remote_vars = (struct xpc_vars *) remote_rp; |
813 | 1035 | ||
814 | 1036 | ||
815 | discovered_nasids = kmalloc(sizeof(u64) * XP_NASID_MASK_WORDS, | 1037 | discovered_nasids = kmalloc(sizeof(u64) * xp_nasid_mask_words, |
816 | GFP_KERNEL); | 1038 | GFP_KERNEL); |
817 | if (discovered_nasids == NULL) { | 1039 | if (discovered_nasids == NULL) { |
818 | kfree(remote_rp_base); | 1040 | kfree(remote_rp_base); |
819 | return; | 1041 | return; |
820 | } | 1042 | } |
821 | memset(discovered_nasids, 0, sizeof(u64) * XP_NASID_MASK_WORDS); | 1043 | memset(discovered_nasids, 0, sizeof(u64) * xp_nasid_mask_words); |
822 | 1044 | ||
823 | rp = (struct xpc_rsvd_page *) xpc_rsvd_page; | 1045 | rp = (struct xpc_rsvd_page *) xpc_rsvd_page; |
824 | 1046 | ||
@@ -827,11 +1049,19 @@ xpc_discovery(void) | |||
827 | * nodes that can comprise an access protection grouping. The access | 1049 | * nodes that can comprise an access protection grouping. The access |
828 | * protection is in regards to memory, IOI and IPI. | 1050 | * protection is in regards to memory, IOI and IPI. |
829 | */ | 1051 | */ |
830 | //>>> move the next two #defines into either include/asm-ia64/sn/arch.h or | 1052 | max_regions = 64; |
831 | //>>> include/asm-ia64/sn/addrs.h | 1053 | region_size = sn_region_size; |
832 | #define SH1_MAX_REGIONS 64 | 1054 | |
833 | #define SH2_MAX_REGIONS 256 | 1055 | switch (region_size) { |
834 | max_regions = is_shub2() ? SH2_MAX_REGIONS : SH1_MAX_REGIONS; | 1056 | case 128: |
1057 | max_regions *= 2; | ||
1058 | case 64: | ||
1059 | max_regions *= 2; | ||
1060 | case 32: | ||
1061 | max_regions *= 2; | ||
1062 | region_size = 16; | ||
1063 | DBUG_ON(!is_shub2()); | ||
1064 | } | ||
835 | 1065 | ||
836 | for (region = 0; region < max_regions; region++) { | 1066 | for (region = 0; region < max_regions; region++) { |
837 | 1067 | ||
@@ -841,8 +1071,8 @@ xpc_discovery(void) | |||
841 | 1071 | ||
842 | dev_dbg(xpc_part, "searching region %d\n", region); | 1072 | dev_dbg(xpc_part, "searching region %d\n", region); |
843 | 1073 | ||
844 | for (nasid = (region * sn_region_size * 2); | 1074 | for (nasid = (region * region_size * 2); |
845 | nasid < ((region + 1) * sn_region_size * 2); | 1075 | nasid < ((region + 1) * region_size * 2); |
846 | nasid += 2) { | 1076 | nasid += 2) { |
847 | 1077 | ||
848 | if ((volatile int) xpc_exiting) { | 1078 | if ((volatile int) xpc_exiting) { |
@@ -852,14 +1082,14 @@ xpc_discovery(void) | |||
852 | dev_dbg(xpc_part, "checking nasid %d\n", nasid); | 1082 | dev_dbg(xpc_part, "checking nasid %d\n", nasid); |
853 | 1083 | ||
854 | 1084 | ||
855 | if (XPC_NASID_IN_ARRAY(nasid, rp->part_nasids)) { | 1085 | if (XPC_NASID_IN_ARRAY(nasid, xpc_part_nasids)) { |
856 | dev_dbg(xpc_part, "PROM indicates Nasid %d is " | 1086 | dev_dbg(xpc_part, "PROM indicates Nasid %d is " |
857 | "part of the local partition; skipping " | 1087 | "part of the local partition; skipping " |
858 | "region\n", nasid); | 1088 | "region\n", nasid); |
859 | break; | 1089 | break; |
860 | } | 1090 | } |
861 | 1091 | ||
862 | if (!(XPC_NASID_IN_ARRAY(nasid, rp->mach_nasids))) { | 1092 | if (!(XPC_NASID_IN_ARRAY(nasid, xpc_mach_nasids))) { |
863 | dev_dbg(xpc_part, "PROM indicates Nasid %d was " | 1093 | dev_dbg(xpc_part, "PROM indicates Nasid %d was " |
864 | "not on Numa-Link network at reset\n", | 1094 | "not on Numa-Link network at reset\n", |
865 | nasid); | 1095 | nasid); |
@@ -877,7 +1107,7 @@ xpc_discovery(void) | |||
877 | /* pull over the reserved page structure */ | 1107 | /* pull over the reserved page structure */ |
878 | 1108 | ||
879 | ret = xpc_get_remote_rp(nasid, discovered_nasids, | 1109 | ret = xpc_get_remote_rp(nasid, discovered_nasids, |
880 | remote_rp, &remote_rsvd_page_pa); | 1110 | remote_rp, &remote_rp_pa); |
881 | if (ret != xpcSuccess) { | 1111 | if (ret != xpcSuccess) { |
882 | dev_dbg(xpc_part, "unable to get reserved page " | 1112 | dev_dbg(xpc_part, "unable to get reserved page " |
883 | "from nasid %d, reason=%d\n", nasid, | 1113 | "from nasid %d, reason=%d\n", nasid, |
@@ -948,6 +1178,13 @@ xpc_discovery(void) | |||
948 | remote_vars->act_nasid, | 1178 | remote_vars->act_nasid, |
949 | remote_vars->act_phys_cpuid); | 1179 | remote_vars->act_phys_cpuid); |
950 | 1180 | ||
1181 | if (XPC_SUPPORTS_DISENGAGE_REQUEST(remote_vars-> | ||
1182 | version)) { | ||
1183 | part->remote_amos_page_pa = | ||
1184 | remote_vars->amos_page_pa; | ||
1185 | xpc_mark_partition_disengaged(part); | ||
1186 | xpc_cancel_partition_disengage_request(part); | ||
1187 | } | ||
951 | xpc_IPI_send_activate(remote_vars); | 1188 | xpc_IPI_send_activate(remote_vars); |
952 | } | 1189 | } |
953 | } | 1190 | } |
@@ -974,12 +1211,12 @@ xpc_initiate_partid_to_nasids(partid_t partid, void *nasid_mask) | |||
974 | return xpcPartitionDown; | 1211 | return xpcPartitionDown; |
975 | } | 1212 | } |
976 | 1213 | ||
977 | part_nasid_pa = part->remote_rp_pa + | 1214 | memset(nasid_mask, 0, XP_NASID_MASK_BYTES); |
978 | (u64) &((struct xpc_rsvd_page *) 0)->part_nasids; | 1215 | |
1216 | part_nasid_pa = (u64) XPC_RP_PART_NASIDS(part->remote_rp_pa); | ||
979 | 1217 | ||
980 | bte_res = xp_bte_copy(part_nasid_pa, ia64_tpa((u64) nasid_mask), | 1218 | bte_res = xp_bte_copy(part_nasid_pa, ia64_tpa((u64) nasid_mask), |
981 | L1_CACHE_ALIGN(XP_NASID_MASK_BYTES), | 1219 | xp_nasid_mask_bytes, (BTE_NOTIFY | BTE_WACQUIRE), NULL); |
982 | (BTE_NOTIFY | BTE_WACQUIRE), NULL); | ||
983 | 1220 | ||
984 | return xpc_map_bte_errors(bte_res); | 1221 | return xpc_map_bte_errors(bte_res); |
985 | } | 1222 | } |