diff options
author | Dean Nelson <dcn@sgi.com> | 2008-07-30 01:34:18 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@linux-foundation.org> | 2008-07-30 12:41:50 -0400 |
commit | 5b8669dfd110a62a74eea525a009342f73987ea0 (patch) | |
tree | 04572d8508f450131298b6ec072e97aa9fdba539 /drivers/misc | |
parent | 83469b5525b4a35be40b17cb41d64118d84d9f80 (diff) |
sgi-xp: setup the activate GRU message queue
Setup the activate GRU message queue that is used for partition activation
and channel connection on UV systems.
Signed-off-by: Dean Nelson <dcn@sgi.com>
Cc: Jack Steiner <steiner@sgi.com>
Cc: "Luck, Tony" <tony.luck@intel.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
Diffstat (limited to 'drivers/misc')
-rw-r--r-- | drivers/misc/sgi-xp/xp.h | 3 | ||||
-rw-r--r-- | drivers/misc/sgi-xp/xp_uv.c | 10 | ||||
-rw-r--r-- | drivers/misc/sgi-xp/xpc.h | 158 | ||||
-rw-r--r-- | drivers/misc/sgi-xp/xpc_channel.c | 22 | ||||
-rw-r--r-- | drivers/misc/sgi-xp/xpc_main.c | 329 | ||||
-rw-r--r-- | drivers/misc/sgi-xp/xpc_partition.c | 28 | ||||
-rw-r--r-- | drivers/misc/sgi-xp/xpc_sn2.c | 387 | ||||
-rw-r--r-- | drivers/misc/sgi-xp/xpc_uv.c | 781 |
8 files changed, 1328 insertions, 390 deletions
diff --git a/drivers/misc/sgi-xp/xp.h b/drivers/misc/sgi-xp/xp.h index 45d0a08c2ddd..9ac5758f4d08 100644 --- a/drivers/misc/sgi-xp/xp.h +++ b/drivers/misc/sgi-xp/xp.h | |||
@@ -208,8 +208,9 @@ enum xp_retval { | |||
208 | xpNeedMoreInfo, /* 57: more info is needed by SAL */ | 208 | xpNeedMoreInfo, /* 57: more info is needed by SAL */ |
209 | 209 | ||
210 | xpGruCopyError, /* 58: gru_copy_gru() returned error */ | 210 | xpGruCopyError, /* 58: gru_copy_gru() returned error */ |
211 | xpGruSendMqError, /* 59: gru send message queue related error */ | ||
211 | 212 | ||
212 | xpUnknownReason /* 59: unknown reason - must be last in enum */ | 213 | xpUnknownReason /* 60: unknown reason - must be last in enum */ |
213 | }; | 214 | }; |
214 | 215 | ||
215 | /* | 216 | /* |
diff --git a/drivers/misc/sgi-xp/xp_uv.c b/drivers/misc/sgi-xp/xp_uv.c index 44f2c2b58c2f..d9f7ce2510bc 100644 --- a/drivers/misc/sgi-xp/xp_uv.c +++ b/drivers/misc/sgi-xp/xp_uv.c | |||
@@ -42,15 +42,25 @@ xp_remote_memcpy_uv(unsigned long dst_gpa, const unsigned long src_gpa, | |||
42 | return xpGruCopyError; | 42 | return xpGruCopyError; |
43 | } | 43 | } |
44 | 44 | ||
45 | static int | ||
46 | xp_cpu_to_nasid_uv(int cpuid) | ||
47 | { | ||
48 | /* ??? Is this same as sn2 nasid in mach/part bitmaps set up by SAL? */ | ||
49 | return UV_PNODE_TO_NASID(uv_cpu_to_pnode(cpuid)); | ||
50 | } | ||
51 | |||
45 | enum xp_retval | 52 | enum xp_retval |
46 | xp_init_uv(void) | 53 | xp_init_uv(void) |
47 | { | 54 | { |
48 | BUG_ON(!is_uv()); | 55 | BUG_ON(!is_uv()); |
49 | 56 | ||
50 | xp_max_npartitions = XP_MAX_NPARTITIONS_UV; | 57 | xp_max_npartitions = XP_MAX_NPARTITIONS_UV; |
58 | xp_partition_id = 0; /* !!! not correct value */ | ||
59 | xp_region_size = 0; /* !!! not correct value */ | ||
51 | 60 | ||
52 | xp_pa = xp_pa_uv; | 61 | xp_pa = xp_pa_uv; |
53 | xp_remote_memcpy = xp_remote_memcpy_uv; | 62 | xp_remote_memcpy = xp_remote_memcpy_uv; |
63 | xp_cpu_to_nasid = xp_cpu_to_nasid_uv; | ||
54 | 64 | ||
55 | return xpSuccess; | 65 | return xpSuccess; |
56 | } | 66 | } |
diff --git a/drivers/misc/sgi-xp/xpc.h b/drivers/misc/sgi-xp/xpc.h index 1e48f7765050..4c26181defff 100644 --- a/drivers/misc/sgi-xp/xpc.h +++ b/drivers/misc/sgi-xp/xpc.h | |||
@@ -164,8 +164,8 @@ struct xpc_vars_part_sn2 { | |||
164 | * MAGIC2 indicates that this partition has pulled the remote partititions | 164 | * MAGIC2 indicates that this partition has pulled the remote partititions |
165 | * per partition variables that pertain to this partition. | 165 | * per partition variables that pertain to this partition. |
166 | */ | 166 | */ |
167 | #define XPC_VP_MAGIC1 0x0053524156435058L /* 'XPCVARS\0'L (little endian) */ | 167 | #define XPC_VP_MAGIC1_SN2 0x0053524156435058L /* 'XPCVARS\0'L (little endian) */ |
168 | #define XPC_VP_MAGIC2 0x0073726176435058L /* 'XPCvars\0'L (little endian) */ | 168 | #define XPC_VP_MAGIC2_SN2 0x0073726176435058L /* 'XPCvars\0'L (little endian) */ |
169 | 169 | ||
170 | /* the reserved page sizes and offsets */ | 170 | /* the reserved page sizes and offsets */ |
171 | 171 | ||
@@ -181,6 +181,80 @@ struct xpc_vars_part_sn2 { | |||
181 | xpc_nasid_mask_nlongs)) | 181 | xpc_nasid_mask_nlongs)) |
182 | 182 | ||
183 | /* | 183 | /* |
184 | * The activate_mq is used to send/receive messages that affect XPC's heartbeat, | ||
185 | * partition active state, and channel state. This is UV only. | ||
186 | */ | ||
187 | struct xpc_activate_mq_msghdr_uv { | ||
188 | short partid; /* sender's partid */ | ||
189 | u8 act_state; /* sender's act_state at time msg sent */ | ||
190 | u8 type; /* message's type */ | ||
191 | unsigned long rp_ts_jiffies; /* timestamp of sender's rp setup by XPC */ | ||
192 | }; | ||
193 | |||
194 | /* activate_mq defined message types */ | ||
195 | #define XPC_ACTIVATE_MQ_MSG_SYNC_ACT_STATE_UV 0 | ||
196 | #define XPC_ACTIVATE_MQ_MSG_INC_HEARTBEAT_UV 1 | ||
197 | #define XPC_ACTIVATE_MQ_MSG_OFFLINE_HEARTBEAT_UV 2 | ||
198 | #define XPC_ACTIVATE_MQ_MSG_ONLINE_HEARTBEAT_UV 3 | ||
199 | |||
200 | #define XPC_ACTIVATE_MQ_MSG_ACTIVATE_REQ_UV 4 | ||
201 | #define XPC_ACTIVATE_MQ_MSG_DEACTIVATE_REQ_UV 5 | ||
202 | |||
203 | #define XPC_ACTIVATE_MQ_MSG_CHCTL_CLOSEREQUEST_UV 6 | ||
204 | #define XPC_ACTIVATE_MQ_MSG_CHCTL_CLOSEREPLY_UV 7 | ||
205 | #define XPC_ACTIVATE_MQ_MSG_CHCTL_OPENREQUEST_UV 8 | ||
206 | #define XPC_ACTIVATE_MQ_MSG_CHCTL_OPENREPLY_UV 9 | ||
207 | |||
208 | #define XPC_ACTIVATE_MQ_MSG_MARK_ENGAGED_UV 10 | ||
209 | #define XPC_ACTIVATE_MQ_MSG_MARK_DISENGAGED_UV 11 | ||
210 | |||
211 | struct xpc_activate_mq_msg_uv { | ||
212 | struct xpc_activate_mq_msghdr_uv header; | ||
213 | }; | ||
214 | |||
215 | struct xpc_activate_mq_msg_heartbeat_req_uv { | ||
216 | struct xpc_activate_mq_msghdr_uv header; | ||
217 | u64 heartbeat; | ||
218 | }; | ||
219 | |||
220 | struct xpc_activate_mq_msg_activate_req_uv { | ||
221 | struct xpc_activate_mq_msghdr_uv header; | ||
222 | unsigned long rp_gpa; | ||
223 | unsigned long activate_mq_gpa; | ||
224 | }; | ||
225 | |||
226 | struct xpc_activate_mq_msg_deactivate_req_uv { | ||
227 | struct xpc_activate_mq_msghdr_uv header; | ||
228 | enum xp_retval reason; | ||
229 | }; | ||
230 | |||
231 | struct xpc_activate_mq_msg_chctl_closerequest_uv { | ||
232 | struct xpc_activate_mq_msghdr_uv header; | ||
233 | short ch_number; | ||
234 | enum xp_retval reason; | ||
235 | }; | ||
236 | |||
237 | struct xpc_activate_mq_msg_chctl_closereply_uv { | ||
238 | struct xpc_activate_mq_msghdr_uv header; | ||
239 | short ch_number; | ||
240 | }; | ||
241 | |||
242 | struct xpc_activate_mq_msg_chctl_openrequest_uv { | ||
243 | struct xpc_activate_mq_msghdr_uv header; | ||
244 | short ch_number; | ||
245 | short msg_size; /* size of notify_mq's messages */ | ||
246 | short local_nentries; /* ??? Is this needed? What is? */ | ||
247 | }; | ||
248 | |||
249 | struct xpc_activate_mq_msg_chctl_openreply_uv { | ||
250 | struct xpc_activate_mq_msghdr_uv header; | ||
251 | short ch_number; | ||
252 | short remote_nentries; /* ??? Is this needed? What is? */ | ||
253 | short local_nentries; /* ??? Is this needed? What is? */ | ||
254 | unsigned long local_notify_mq_gpa; | ||
255 | }; | ||
256 | |||
257 | /* | ||
184 | * Functions registered by add_timer() or called by kernel_thread() only | 258 | * Functions registered by add_timer() or called by kernel_thread() only |
185 | * allow for a single 64-bit argument. The following macros can be used to | 259 | * allow for a single 64-bit argument. The following macros can be used to |
186 | * pack and unpack two (32-bit, 16-bit or 8-bit) arguments into or out from | 260 | * pack and unpack two (32-bit, 16-bit or 8-bit) arguments into or out from |
@@ -331,6 +405,18 @@ struct xpc_notify { | |||
331 | */ | 405 | */ |
332 | 406 | ||
333 | struct xpc_channel_sn2 { | 407 | struct xpc_channel_sn2 { |
408 | struct xpc_openclose_args *local_openclose_args; /* args passed on */ | ||
409 | /* opening or closing of channel */ | ||
410 | |||
411 | void *local_msgqueue_base; /* base address of kmalloc'd space */ | ||
412 | struct xpc_msg *local_msgqueue; /* local message queue */ | ||
413 | void *remote_msgqueue_base; /* base address of kmalloc'd space */ | ||
414 | struct xpc_msg *remote_msgqueue; /* cached copy of remote partition's */ | ||
415 | /* local message queue */ | ||
416 | unsigned long remote_msgqueue_pa; /* phys addr of remote partition's */ | ||
417 | /* local message queue */ | ||
418 | |||
419 | struct xpc_notify *notify_queue; /* notify queue for messages sent */ | ||
334 | 420 | ||
335 | /* various flavors of local and remote Get/Put values */ | 421 | /* various flavors of local and remote Get/Put values */ |
336 | 422 | ||
@@ -344,13 +430,14 @@ struct xpc_channel_sn2 { | |||
344 | }; | 430 | }; |
345 | 431 | ||
346 | struct xpc_channel_uv { | 432 | struct xpc_channel_uv { |
347 | /* !!! code is coming */ | 433 | unsigned long remote_notify_mq_gpa; /* gru phys address of remote */ |
434 | /* partition's notify mq */ | ||
348 | }; | 435 | }; |
349 | 436 | ||
350 | struct xpc_channel { | 437 | struct xpc_channel { |
351 | short partid; /* ID of remote partition connected */ | 438 | short partid; /* ID of remote partition connected */ |
352 | spinlock_t lock; /* lock for updating this structure */ | 439 | spinlock_t lock; /* lock for updating this structure */ |
353 | u32 flags; /* general flags */ | 440 | unsigned int flags; /* general flags */ |
354 | 441 | ||
355 | enum xp_retval reason; /* reason why channel is disconnect'g */ | 442 | enum xp_retval reason; /* reason why channel is disconnect'g */ |
356 | int reason_line; /* line# disconnect initiated from */ | 443 | int reason_line; /* line# disconnect initiated from */ |
@@ -361,14 +448,6 @@ struct xpc_channel { | |||
361 | u16 local_nentries; /* #of msg entries in local msg queue */ | 448 | u16 local_nentries; /* #of msg entries in local msg queue */ |
362 | u16 remote_nentries; /* #of msg entries in remote msg queue */ | 449 | u16 remote_nentries; /* #of msg entries in remote msg queue */ |
363 | 450 | ||
364 | void *local_msgqueue_base; /* base address of kmalloc'd space */ | ||
365 | struct xpc_msg *local_msgqueue; /* local message queue */ | ||
366 | void *remote_msgqueue_base; /* base address of kmalloc'd space */ | ||
367 | struct xpc_msg *remote_msgqueue; /* cached copy of remote partition's */ | ||
368 | /* local message queue */ | ||
369 | unsigned long remote_msgqueue_pa; /* phys addr of remote partition's */ | ||
370 | /* local message queue */ | ||
371 | |||
372 | atomic_t references; /* #of external references to queues */ | 451 | atomic_t references; /* #of external references to queues */ |
373 | 452 | ||
374 | atomic_t n_on_msg_allocate_wq; /* #on msg allocation wait queue */ | 453 | atomic_t n_on_msg_allocate_wq; /* #on msg allocation wait queue */ |
@@ -377,19 +456,13 @@ struct xpc_channel { | |||
377 | u8 delayed_chctl_flags; /* chctl flags received, but delayed */ | 456 | u8 delayed_chctl_flags; /* chctl flags received, but delayed */ |
378 | /* action until channel disconnected */ | 457 | /* action until channel disconnected */ |
379 | 458 | ||
380 | /* queue of msg senders who want to be notified when msg received */ | ||
381 | |||
382 | atomic_t n_to_notify; /* #of msg senders to notify */ | 459 | atomic_t n_to_notify; /* #of msg senders to notify */ |
383 | struct xpc_notify *notify_queue; /* notify queue for messages sent */ | ||
384 | 460 | ||
385 | xpc_channel_func func; /* user's channel function */ | 461 | xpc_channel_func func; /* user's channel function */ |
386 | void *key; /* pointer to user's key */ | 462 | void *key; /* pointer to user's key */ |
387 | 463 | ||
388 | struct completion wdisconnect_wait; /* wait for channel disconnect */ | 464 | struct completion wdisconnect_wait; /* wait for channel disconnect */ |
389 | 465 | ||
390 | struct xpc_openclose_args *local_openclose_args; /* args passed on */ | ||
391 | /* opening or closing of channel */ | ||
392 | |||
393 | /* kthread management related fields */ | 466 | /* kthread management related fields */ |
394 | 467 | ||
395 | atomic_t kthreads_assigned; /* #of kthreads assigned to channel */ | 468 | atomic_t kthreads_assigned; /* #of kthreads assigned to channel */ |
@@ -507,6 +580,8 @@ struct xpc_partition_sn2 { | |||
507 | unsigned long remote_GPs_pa; /* phys addr of remote partition's local */ | 580 | unsigned long remote_GPs_pa; /* phys addr of remote partition's local */ |
508 | /* Get/Put values */ | 581 | /* Get/Put values */ |
509 | 582 | ||
583 | void *local_openclose_args_base; /* base address of kmalloc'd space */ | ||
584 | struct xpc_openclose_args *local_openclose_args; /* local's args */ | ||
510 | unsigned long remote_openclose_args_pa; /* phys addr of remote's args */ | 585 | unsigned long remote_openclose_args_pa; /* phys addr of remote's args */ |
511 | 586 | ||
512 | int notify_IRQ_nasid; /* nasid of where to send notify IRQs */ | 587 | int notify_IRQ_nasid; /* nasid of where to send notify IRQs */ |
@@ -520,9 +595,27 @@ struct xpc_partition_sn2 { | |||
520 | }; | 595 | }; |
521 | 596 | ||
522 | struct xpc_partition_uv { | 597 | struct xpc_partition_uv { |
523 | /* !!! code is coming */ | 598 | unsigned long remote_activate_mq_gpa; /* gru phys address of remote */ |
599 | /* partition's activate mq */ | ||
600 | spinlock_t flags_lock; /* protect updating of flags */ | ||
601 | unsigned int flags; /* general flags */ | ||
602 | u8 remote_act_state; /* remote partition's act_state */ | ||
603 | u8 act_state_req; /* act_state request from remote partition */ | ||
604 | enum xp_retval reason; /* reason for deactivate act_state request */ | ||
605 | u64 heartbeat; /* incremented by remote partition */ | ||
524 | }; | 606 | }; |
525 | 607 | ||
608 | /* struct xpc_partition_uv flags */ | ||
609 | |||
610 | #define XPC_P_HEARTBEAT_OFFLINE_UV 0x00000001 | ||
611 | #define XPC_P_ENGAGED_UV 0x00000002 | ||
612 | |||
613 | /* struct xpc_partition_uv act_state change requests */ | ||
614 | |||
615 | #define XPC_P_ASR_ACTIVATE_UV 0x01 | ||
616 | #define XPC_P_ASR_REACTIVATE_UV 0x02 | ||
617 | #define XPC_P_ASR_DEACTIVATE_UV 0x03 | ||
618 | |||
526 | struct xpc_partition { | 619 | struct xpc_partition { |
527 | 620 | ||
528 | /* XPC HB infrastructure */ | 621 | /* XPC HB infrastructure */ |
@@ -556,8 +649,6 @@ struct xpc_partition { | |||
556 | union xpc_channel_ctl_flags chctl; /* chctl flags yet to be processed */ | 649 | union xpc_channel_ctl_flags chctl; /* chctl flags yet to be processed */ |
557 | spinlock_t chctl_lock; /* chctl flags lock */ | 650 | spinlock_t chctl_lock; /* chctl flags lock */ |
558 | 651 | ||
559 | void *local_openclose_args_base; /* base address of kmalloc'd space */ | ||
560 | struct xpc_openclose_args *local_openclose_args; /* local's args */ | ||
561 | void *remote_openclose_args_base; /* base address of kmalloc'd space */ | 652 | void *remote_openclose_args_base; /* base address of kmalloc'd space */ |
562 | struct xpc_openclose_args *remote_openclose_args; /* copy of remote's */ | 653 | struct xpc_openclose_args *remote_openclose_args; /* copy of remote's */ |
563 | /* args */ | 654 | /* args */ |
@@ -616,17 +707,20 @@ extern struct device *xpc_part; | |||
616 | extern struct device *xpc_chan; | 707 | extern struct device *xpc_chan; |
617 | extern int xpc_disengage_timelimit; | 708 | extern int xpc_disengage_timelimit; |
618 | extern int xpc_disengage_timedout; | 709 | extern int xpc_disengage_timedout; |
619 | extern atomic_t xpc_activate_IRQ_rcvd; | 710 | extern int xpc_activate_IRQ_rcvd; |
711 | extern spinlock_t xpc_activate_IRQ_rcvd_lock; | ||
620 | extern wait_queue_head_t xpc_activate_IRQ_wq; | 712 | extern wait_queue_head_t xpc_activate_IRQ_wq; |
621 | extern void *xpc_heartbeating_to_mask; | 713 | extern void *xpc_heartbeating_to_mask; |
714 | extern void *xpc_kzalloc_cacheline_aligned(size_t, gfp_t, void **); | ||
622 | extern void xpc_activate_partition(struct xpc_partition *); | 715 | extern void xpc_activate_partition(struct xpc_partition *); |
623 | extern void xpc_activate_kthreads(struct xpc_channel *, int); | 716 | extern void xpc_activate_kthreads(struct xpc_channel *, int); |
624 | extern void xpc_create_kthreads(struct xpc_channel *, int, int); | 717 | extern void xpc_create_kthreads(struct xpc_channel *, int, int); |
625 | extern void xpc_disconnect_wait(int); | 718 | extern void xpc_disconnect_wait(int); |
719 | extern int (*xpc_setup_partitions_sn) (void); | ||
626 | extern enum xp_retval (*xpc_get_partition_rsvd_page_pa) (void *, u64 *, | 720 | extern enum xp_retval (*xpc_get_partition_rsvd_page_pa) (void *, u64 *, |
627 | unsigned long *, | 721 | unsigned long *, |
628 | size_t *); | 722 | size_t *); |
629 | extern enum xp_retval (*xpc_rsvd_page_init) (struct xpc_rsvd_page *); | 723 | extern int (*xpc_setup_rsvd_page_sn) (struct xpc_rsvd_page *); |
630 | extern void (*xpc_heartbeat_init) (void); | 724 | extern void (*xpc_heartbeat_init) (void); |
631 | extern void (*xpc_heartbeat_exit) (void); | 725 | extern void (*xpc_heartbeat_exit) (void); |
632 | extern void (*xpc_increment_heartbeat) (void); | 726 | extern void (*xpc_increment_heartbeat) (void); |
@@ -635,8 +729,8 @@ extern void (*xpc_online_heartbeat) (void); | |||
635 | extern enum xp_retval (*xpc_get_remote_heartbeat) (struct xpc_partition *); | 729 | extern enum xp_retval (*xpc_get_remote_heartbeat) (struct xpc_partition *); |
636 | extern enum xp_retval (*xpc_make_first_contact) (struct xpc_partition *); | 730 | extern enum xp_retval (*xpc_make_first_contact) (struct xpc_partition *); |
637 | extern u64 (*xpc_get_chctl_all_flags) (struct xpc_partition *); | 731 | extern u64 (*xpc_get_chctl_all_flags) (struct xpc_partition *); |
638 | extern enum xp_retval (*xpc_allocate_msgqueues) (struct xpc_channel *); | 732 | extern enum xp_retval (*xpc_setup_msg_structures) (struct xpc_channel *); |
639 | extern void (*xpc_free_msgqueues) (struct xpc_channel *); | 733 | extern void (*xpc_teardown_msg_structures) (struct xpc_channel *); |
640 | extern void (*xpc_notify_senders_of_disconnect) (struct xpc_channel *); | 734 | extern void (*xpc_notify_senders_of_disconnect) (struct xpc_channel *); |
641 | extern void (*xpc_process_msg_chctl_flags) (struct xpc_partition *, int); | 735 | extern void (*xpc_process_msg_chctl_flags) (struct xpc_partition *, int); |
642 | extern int (*xpc_n_of_deliverable_msgs) (struct xpc_channel *); | 736 | extern int (*xpc_n_of_deliverable_msgs) (struct xpc_channel *); |
@@ -647,9 +741,9 @@ extern void (*xpc_request_partition_reactivation) (struct xpc_partition *); | |||
647 | extern void (*xpc_request_partition_deactivation) (struct xpc_partition *); | 741 | extern void (*xpc_request_partition_deactivation) (struct xpc_partition *); |
648 | extern void (*xpc_cancel_partition_deactivation_request) ( | 742 | extern void (*xpc_cancel_partition_deactivation_request) ( |
649 | struct xpc_partition *); | 743 | struct xpc_partition *); |
650 | extern void (*xpc_process_activate_IRQ_rcvd) (int); | 744 | extern void (*xpc_process_activate_IRQ_rcvd) (void); |
651 | extern enum xp_retval (*xpc_setup_infrastructure) (struct xpc_partition *); | 745 | extern enum xp_retval (*xpc_setup_ch_structures_sn) (struct xpc_partition *); |
652 | extern void (*xpc_teardown_infrastructure) (struct xpc_partition *); | 746 | extern void (*xpc_teardown_ch_structures_sn) (struct xpc_partition *); |
653 | 747 | ||
654 | extern void (*xpc_indicate_partition_engaged) (struct xpc_partition *); | 748 | extern void (*xpc_indicate_partition_engaged) (struct xpc_partition *); |
655 | extern int (*xpc_partition_engaged) (short); | 749 | extern int (*xpc_partition_engaged) (short); |
@@ -665,6 +759,9 @@ extern void (*xpc_send_chctl_openrequest) (struct xpc_channel *, | |||
665 | unsigned long *); | 759 | unsigned long *); |
666 | extern void (*xpc_send_chctl_openreply) (struct xpc_channel *, unsigned long *); | 760 | extern void (*xpc_send_chctl_openreply) (struct xpc_channel *, unsigned long *); |
667 | 761 | ||
762 | extern void (*xpc_save_remote_msgqueue_pa) (struct xpc_channel *, | ||
763 | unsigned long); | ||
764 | |||
668 | extern enum xp_retval (*xpc_send_msg) (struct xpc_channel *, u32, void *, u16, | 765 | extern enum xp_retval (*xpc_send_msg) (struct xpc_channel *, u32, void *, u16, |
669 | u8, xpc_notify_func, void *); | 766 | u8, xpc_notify_func, void *); |
670 | extern void (*xpc_received_msg) (struct xpc_channel *, struct xpc_msg *); | 767 | extern void (*xpc_received_msg) (struct xpc_channel *, struct xpc_msg *); |
@@ -674,7 +771,7 @@ extern int xpc_init_sn2(void); | |||
674 | extern void xpc_exit_sn2(void); | 771 | extern void xpc_exit_sn2(void); |
675 | 772 | ||
676 | /* found in xpc_uv.c */ | 773 | /* found in xpc_uv.c */ |
677 | extern void xpc_init_uv(void); | 774 | extern int xpc_init_uv(void); |
678 | extern void xpc_exit_uv(void); | 775 | extern void xpc_exit_uv(void); |
679 | 776 | ||
680 | /* found in xpc_partition.c */ | 777 | /* found in xpc_partition.c */ |
@@ -684,7 +781,8 @@ extern struct xpc_rsvd_page *xpc_rsvd_page; | |||
684 | extern unsigned long *xpc_mach_nasids; | 781 | extern unsigned long *xpc_mach_nasids; |
685 | extern struct xpc_partition *xpc_partitions; | 782 | extern struct xpc_partition *xpc_partitions; |
686 | extern void *xpc_kmalloc_cacheline_aligned(size_t, gfp_t, void **); | 783 | extern void *xpc_kmalloc_cacheline_aligned(size_t, gfp_t, void **); |
687 | extern struct xpc_rsvd_page *xpc_setup_rsvd_page(void); | 784 | extern int xpc_setup_rsvd_page(void); |
785 | extern void xpc_teardown_rsvd_page(void); | ||
688 | extern int xpc_identify_activate_IRQ_sender(void); | 786 | extern int xpc_identify_activate_IRQ_sender(void); |
689 | extern int xpc_partition_disengaged(struct xpc_partition *); | 787 | extern int xpc_partition_disengaged(struct xpc_partition *); |
690 | extern enum xp_retval xpc_mark_partition_active(struct xpc_partition *); | 788 | extern enum xp_retval xpc_mark_partition_active(struct xpc_partition *); |
diff --git a/drivers/misc/sgi-xp/xpc_channel.c b/drivers/misc/sgi-xp/xpc_channel.c index 17ab75d69e83..73df9fb5ee66 100644 --- a/drivers/misc/sgi-xp/xpc_channel.c +++ b/drivers/misc/sgi-xp/xpc_channel.c | |||
@@ -39,7 +39,7 @@ xpc_process_connect(struct xpc_channel *ch, unsigned long *irq_flags) | |||
39 | 39 | ||
40 | if (!(ch->flags & XPC_C_SETUP)) { | 40 | if (!(ch->flags & XPC_C_SETUP)) { |
41 | spin_unlock_irqrestore(&ch->lock, *irq_flags); | 41 | spin_unlock_irqrestore(&ch->lock, *irq_flags); |
42 | ret = xpc_allocate_msgqueues(ch); | 42 | ret = xpc_setup_msg_structures(ch); |
43 | spin_lock_irqsave(&ch->lock, *irq_flags); | 43 | spin_lock_irqsave(&ch->lock, *irq_flags); |
44 | 44 | ||
45 | if (ret != xpSuccess) | 45 | if (ret != xpSuccess) |
@@ -62,8 +62,6 @@ xpc_process_connect(struct xpc_channel *ch, unsigned long *irq_flags) | |||
62 | if (!(ch->flags & XPC_C_ROPENREPLY)) | 62 | if (!(ch->flags & XPC_C_ROPENREPLY)) |
63 | return; | 63 | return; |
64 | 64 | ||
65 | DBUG_ON(ch->remote_msgqueue_pa == 0); | ||
66 | |||
67 | ch->flags = (XPC_C_CONNECTED | XPC_C_SETUP); /* clear all else */ | 65 | ch->flags = (XPC_C_CONNECTED | XPC_C_SETUP); /* clear all else */ |
68 | 66 | ||
69 | dev_info(xpc_chan, "channel %d to partition %d connected\n", | 67 | dev_info(xpc_chan, "channel %d to partition %d connected\n", |
@@ -134,13 +132,23 @@ xpc_process_disconnect(struct xpc_channel *ch, unsigned long *irq_flags) | |||
134 | spin_lock_irqsave(&ch->lock, *irq_flags); | 132 | spin_lock_irqsave(&ch->lock, *irq_flags); |
135 | } | 133 | } |
136 | 134 | ||
135 | DBUG_ON(atomic_read(&ch->n_to_notify) != 0); | ||
136 | |||
137 | /* it's now safe to free the channel's message queues */ | 137 | /* it's now safe to free the channel's message queues */ |
138 | xpc_free_msgqueues(ch); | 138 | xpc_teardown_msg_structures(ch); |
139 | |||
140 | ch->func = NULL; | ||
141 | ch->key = NULL; | ||
142 | ch->msg_size = 0; | ||
143 | ch->local_nentries = 0; | ||
144 | ch->remote_nentries = 0; | ||
145 | ch->kthreads_assigned_limit = 0; | ||
146 | ch->kthreads_idle_limit = 0; | ||
139 | 147 | ||
140 | /* | 148 | /* |
141 | * Mark the channel disconnected and clear all other flags, including | 149 | * Mark the channel disconnected and clear all other flags, including |
142 | * XPC_C_SETUP (because of call to xpc_free_msgqueues()) but not | 150 | * XPC_C_SETUP (because of call to xpc_teardown_msg_structures()) but |
143 | * including XPC_C_WDISCONNECT (if it was set). | 151 | * not including XPC_C_WDISCONNECT (if it was set). |
144 | */ | 152 | */ |
145 | ch->flags = (XPC_C_DISCONNECTED | (ch->flags & XPC_C_WDISCONNECT)); | 153 | ch->flags = (XPC_C_DISCONNECTED | (ch->flags & XPC_C_WDISCONNECT)); |
146 | 154 | ||
@@ -395,7 +403,7 @@ again: | |||
395 | DBUG_ON(args->remote_nentries == 0); | 403 | DBUG_ON(args->remote_nentries == 0); |
396 | 404 | ||
397 | ch->flags |= XPC_C_ROPENREPLY; | 405 | ch->flags |= XPC_C_ROPENREPLY; |
398 | ch->remote_msgqueue_pa = args->local_msgqueue_pa; | 406 | xpc_save_remote_msgqueue_pa(ch, args->local_msgqueue_pa); |
399 | 407 | ||
400 | if (args->local_nentries < ch->remote_nentries) { | 408 | if (args->local_nentries < ch->remote_nentries) { |
401 | dev_dbg(xpc_chan, "XPC_CHCTL_OPENREPLY: new " | 409 | dev_dbg(xpc_chan, "XPC_CHCTL_OPENREPLY: new " |
diff --git a/drivers/misc/sgi-xp/xpc_main.c b/drivers/misc/sgi-xp/xpc_main.c index b303c130bba8..13ec47928994 100644 --- a/drivers/misc/sgi-xp/xpc_main.c +++ b/drivers/misc/sgi-xp/xpc_main.c | |||
@@ -141,8 +141,9 @@ static struct ctl_table_header *xpc_sysctl; | |||
141 | /* non-zero if any remote partition disengage was timed out */ | 141 | /* non-zero if any remote partition disengage was timed out */ |
142 | int xpc_disengage_timedout; | 142 | int xpc_disengage_timedout; |
143 | 143 | ||
144 | /* #of activate IRQs received */ | 144 | /* #of activate IRQs received and not yet processed */ |
145 | atomic_t xpc_activate_IRQ_rcvd = ATOMIC_INIT(0); | 145 | int xpc_activate_IRQ_rcvd; |
146 | DEFINE_SPINLOCK(xpc_activate_IRQ_rcvd_lock); | ||
146 | 147 | ||
147 | /* IRQ handler notifies this wait queue on receipt of an IRQ */ | 148 | /* IRQ handler notifies this wait queue on receipt of an IRQ */ |
148 | DECLARE_WAIT_QUEUE_HEAD(xpc_activate_IRQ_wq); | 149 | DECLARE_WAIT_QUEUE_HEAD(xpc_activate_IRQ_wq); |
@@ -169,10 +170,11 @@ static struct notifier_block xpc_die_notifier = { | |||
169 | .notifier_call = xpc_system_die, | 170 | .notifier_call = xpc_system_die, |
170 | }; | 171 | }; |
171 | 172 | ||
173 | int (*xpc_setup_partitions_sn) (void); | ||
172 | enum xp_retval (*xpc_get_partition_rsvd_page_pa) (void *buf, u64 *cookie, | 174 | enum xp_retval (*xpc_get_partition_rsvd_page_pa) (void *buf, u64 *cookie, |
173 | unsigned long *rp_pa, | 175 | unsigned long *rp_pa, |
174 | size_t *len); | 176 | size_t *len); |
175 | enum xp_retval (*xpc_rsvd_page_init) (struct xpc_rsvd_page *rp); | 177 | int (*xpc_setup_rsvd_page_sn) (struct xpc_rsvd_page *rp); |
176 | void (*xpc_heartbeat_init) (void); | 178 | void (*xpc_heartbeat_init) (void); |
177 | void (*xpc_heartbeat_exit) (void); | 179 | void (*xpc_heartbeat_exit) (void); |
178 | void (*xpc_increment_heartbeat) (void); | 180 | void (*xpc_increment_heartbeat) (void); |
@@ -183,8 +185,8 @@ enum xp_retval (*xpc_get_remote_heartbeat) (struct xpc_partition *part); | |||
183 | enum xp_retval (*xpc_make_first_contact) (struct xpc_partition *part); | 185 | enum xp_retval (*xpc_make_first_contact) (struct xpc_partition *part); |
184 | void (*xpc_notify_senders_of_disconnect) (struct xpc_channel *ch); | 186 | void (*xpc_notify_senders_of_disconnect) (struct xpc_channel *ch); |
185 | u64 (*xpc_get_chctl_all_flags) (struct xpc_partition *part); | 187 | u64 (*xpc_get_chctl_all_flags) (struct xpc_partition *part); |
186 | enum xp_retval (*xpc_allocate_msgqueues) (struct xpc_channel *ch); | 188 | enum xp_retval (*xpc_setup_msg_structures) (struct xpc_channel *ch); |
187 | void (*xpc_free_msgqueues) (struct xpc_channel *ch); | 189 | void (*xpc_teardown_msg_structures) (struct xpc_channel *ch); |
188 | void (*xpc_process_msg_chctl_flags) (struct xpc_partition *part, int ch_number); | 190 | void (*xpc_process_msg_chctl_flags) (struct xpc_partition *part, int ch_number); |
189 | int (*xpc_n_of_deliverable_msgs) (struct xpc_channel *ch); | 191 | int (*xpc_n_of_deliverable_msgs) (struct xpc_channel *ch); |
190 | struct xpc_msg *(*xpc_get_deliverable_msg) (struct xpc_channel *ch); | 192 | struct xpc_msg *(*xpc_get_deliverable_msg) (struct xpc_channel *ch); |
@@ -196,9 +198,9 @@ void (*xpc_request_partition_reactivation) (struct xpc_partition *part); | |||
196 | void (*xpc_request_partition_deactivation) (struct xpc_partition *part); | 198 | void (*xpc_request_partition_deactivation) (struct xpc_partition *part); |
197 | void (*xpc_cancel_partition_deactivation_request) (struct xpc_partition *part); | 199 | void (*xpc_cancel_partition_deactivation_request) (struct xpc_partition *part); |
198 | 200 | ||
199 | void (*xpc_process_activate_IRQ_rcvd) (int n_IRQs_expected); | 201 | void (*xpc_process_activate_IRQ_rcvd) (void); |
200 | enum xp_retval (*xpc_setup_infrastructure) (struct xpc_partition *part); | 202 | enum xp_retval (*xpc_setup_ch_structures_sn) (struct xpc_partition *part); |
201 | void (*xpc_teardown_infrastructure) (struct xpc_partition *part); | 203 | void (*xpc_teardown_ch_structures_sn) (struct xpc_partition *part); |
202 | 204 | ||
203 | void (*xpc_indicate_partition_engaged) (struct xpc_partition *part); | 205 | void (*xpc_indicate_partition_engaged) (struct xpc_partition *part); |
204 | int (*xpc_partition_engaged) (short partid); | 206 | int (*xpc_partition_engaged) (short partid); |
@@ -215,6 +217,9 @@ void (*xpc_send_chctl_openrequest) (struct xpc_channel *ch, | |||
215 | void (*xpc_send_chctl_openreply) (struct xpc_channel *ch, | 217 | void (*xpc_send_chctl_openreply) (struct xpc_channel *ch, |
216 | unsigned long *irq_flags); | 218 | unsigned long *irq_flags); |
217 | 219 | ||
220 | void (*xpc_save_remote_msgqueue_pa) (struct xpc_channel *ch, | ||
221 | unsigned long msgqueue_pa); | ||
222 | |||
218 | enum xp_retval (*xpc_send_msg) (struct xpc_channel *ch, u32 flags, | 223 | enum xp_retval (*xpc_send_msg) (struct xpc_channel *ch, u32 flags, |
219 | void *payload, u16 payload_size, u8 notify_type, | 224 | void *payload, u16 payload_size, u8 notify_type, |
220 | xpc_notify_func func, void *key); | 225 | xpc_notify_func func, void *key); |
@@ -308,8 +313,6 @@ xpc_check_remote_hb(void) | |||
308 | static int | 313 | static int |
309 | xpc_hb_checker(void *ignore) | 314 | xpc_hb_checker(void *ignore) |
310 | { | 315 | { |
311 | int last_IRQ_count = 0; | ||
312 | int new_IRQ_count; | ||
313 | int force_IRQ = 0; | 316 | int force_IRQ = 0; |
314 | 317 | ||
315 | /* this thread was marked active by xpc_hb_init() */ | 318 | /* this thread was marked active by xpc_hb_init() */ |
@@ -325,43 +328,37 @@ xpc_hb_checker(void *ignore) | |||
325 | dev_dbg(xpc_part, "woke up with %d ticks rem; %d IRQs have " | 328 | dev_dbg(xpc_part, "woke up with %d ticks rem; %d IRQs have " |
326 | "been received\n", | 329 | "been received\n", |
327 | (int)(xpc_hb_check_timeout - jiffies), | 330 | (int)(xpc_hb_check_timeout - jiffies), |
328 | atomic_read(&xpc_activate_IRQ_rcvd) - last_IRQ_count); | 331 | xpc_activate_IRQ_rcvd); |
329 | 332 | ||
330 | /* checking of remote heartbeats is skewed by IRQ handling */ | 333 | /* checking of remote heartbeats is skewed by IRQ handling */ |
331 | if (time_is_before_eq_jiffies(xpc_hb_check_timeout)) { | 334 | if (time_is_before_eq_jiffies(xpc_hb_check_timeout)) { |
335 | xpc_hb_check_timeout = jiffies + | ||
336 | (xpc_hb_check_interval * HZ); | ||
337 | |||
332 | dev_dbg(xpc_part, "checking remote heartbeats\n"); | 338 | dev_dbg(xpc_part, "checking remote heartbeats\n"); |
333 | xpc_check_remote_hb(); | 339 | xpc_check_remote_hb(); |
334 | 340 | ||
335 | /* | 341 | /* |
336 | * We need to periodically recheck to ensure no | 342 | * On sn2 we need to periodically recheck to ensure no |
337 | * IRQ/amo pairs have been missed. That check | 343 | * IRQ/amo pairs have been missed. |
338 | * must always reset xpc_hb_check_timeout. | ||
339 | */ | 344 | */ |
340 | force_IRQ = 1; | 345 | if (is_shub()) |
346 | force_IRQ = 1; | ||
341 | } | 347 | } |
342 | 348 | ||
343 | /* check for outstanding IRQs */ | 349 | /* check for outstanding IRQs */ |
344 | new_IRQ_count = atomic_read(&xpc_activate_IRQ_rcvd); | 350 | if (xpc_activate_IRQ_rcvd > 0 || force_IRQ != 0) { |
345 | if (last_IRQ_count < new_IRQ_count || force_IRQ != 0) { | ||
346 | force_IRQ = 0; | 351 | force_IRQ = 0; |
347 | 352 | dev_dbg(xpc_part, "processing activate IRQs " | |
348 | dev_dbg(xpc_part, "found an IRQ to process; will be " | 353 | "received\n"); |
349 | "resetting xpc_hb_check_timeout\n"); | 354 | xpc_process_activate_IRQ_rcvd(); |
350 | |||
351 | xpc_process_activate_IRQ_rcvd(new_IRQ_count - | ||
352 | last_IRQ_count); | ||
353 | last_IRQ_count = new_IRQ_count; | ||
354 | |||
355 | xpc_hb_check_timeout = jiffies + | ||
356 | (xpc_hb_check_interval * HZ); | ||
357 | } | 355 | } |
358 | 356 | ||
359 | /* wait for IRQ or timeout */ | 357 | /* wait for IRQ or timeout */ |
360 | (void)wait_event_interruptible(xpc_activate_IRQ_wq, | 358 | (void)wait_event_interruptible(xpc_activate_IRQ_wq, |
361 | (last_IRQ_count < atomic_read( | 359 | (time_is_before_eq_jiffies( |
362 | &xpc_activate_IRQ_rcvd) | ||
363 | || time_is_before_eq_jiffies( | ||
364 | xpc_hb_check_timeout) || | 360 | xpc_hb_check_timeout) || |
361 | xpc_activate_IRQ_rcvd > 0 || | ||
365 | xpc_exiting)); | 362 | xpc_exiting)); |
366 | } | 363 | } |
367 | 364 | ||
@@ -437,6 +434,153 @@ xpc_channel_mgr(struct xpc_partition *part) | |||
437 | } | 434 | } |
438 | 435 | ||
439 | /* | 436 | /* |
437 | * Guarantee that the kzalloc'd memory is cacheline aligned. | ||
438 | */ | ||
439 | void * | ||
440 | xpc_kzalloc_cacheline_aligned(size_t size, gfp_t flags, void **base) | ||
441 | { | ||
442 | /* see if kzalloc will give us cachline aligned memory by default */ | ||
443 | *base = kzalloc(size, flags); | ||
444 | if (*base == NULL) | ||
445 | return NULL; | ||
446 | |||
447 | if ((u64)*base == L1_CACHE_ALIGN((u64)*base)) | ||
448 | return *base; | ||
449 | |||
450 | kfree(*base); | ||
451 | |||
452 | /* nope, we'll have to do it ourselves */ | ||
453 | *base = kzalloc(size + L1_CACHE_BYTES, flags); | ||
454 | if (*base == NULL) | ||
455 | return NULL; | ||
456 | |||
457 | return (void *)L1_CACHE_ALIGN((u64)*base); | ||
458 | } | ||
459 | |||
460 | /* | ||
461 | * Setup the channel structures necessary to support XPartition Communication | ||
462 | * between the specified remote partition and the local one. | ||
463 | */ | ||
464 | static enum xp_retval | ||
465 | xpc_setup_ch_structures(struct xpc_partition *part) | ||
466 | { | ||
467 | enum xp_retval ret; | ||
468 | int ch_number; | ||
469 | struct xpc_channel *ch; | ||
470 | short partid = XPC_PARTID(part); | ||
471 | |||
472 | /* | ||
473 | * Allocate all of the channel structures as a contiguous chunk of | ||
474 | * memory. | ||
475 | */ | ||
476 | DBUG_ON(part->channels != NULL); | ||
477 | part->channels = kzalloc(sizeof(struct xpc_channel) * XPC_MAX_NCHANNELS, | ||
478 | GFP_KERNEL); | ||
479 | if (part->channels == NULL) { | ||
480 | dev_err(xpc_chan, "can't get memory for channels\n"); | ||
481 | return xpNoMemory; | ||
482 | } | ||
483 | |||
484 | /* allocate the remote open and close args */ | ||
485 | |||
486 | part->remote_openclose_args = | ||
487 | xpc_kzalloc_cacheline_aligned(XPC_OPENCLOSE_ARGS_SIZE, | ||
488 | GFP_KERNEL, &part-> | ||
489 | remote_openclose_args_base); | ||
490 | if (part->remote_openclose_args == NULL) { | ||
491 | dev_err(xpc_chan, "can't get memory for remote connect args\n"); | ||
492 | ret = xpNoMemory; | ||
493 | goto out_1; | ||
494 | } | ||
495 | |||
496 | part->chctl.all_flags = 0; | ||
497 | spin_lock_init(&part->chctl_lock); | ||
498 | |||
499 | atomic_set(&part->channel_mgr_requests, 1); | ||
500 | init_waitqueue_head(&part->channel_mgr_wq); | ||
501 | |||
502 | part->nchannels = XPC_MAX_NCHANNELS; | ||
503 | |||
504 | atomic_set(&part->nchannels_active, 0); | ||
505 | atomic_set(&part->nchannels_engaged, 0); | ||
506 | |||
507 | for (ch_number = 0; ch_number < part->nchannels; ch_number++) { | ||
508 | ch = &part->channels[ch_number]; | ||
509 | |||
510 | ch->partid = partid; | ||
511 | ch->number = ch_number; | ||
512 | ch->flags = XPC_C_DISCONNECTED; | ||
513 | |||
514 | atomic_set(&ch->kthreads_assigned, 0); | ||
515 | atomic_set(&ch->kthreads_idle, 0); | ||
516 | atomic_set(&ch->kthreads_active, 0); | ||
517 | |||
518 | atomic_set(&ch->references, 0); | ||
519 | atomic_set(&ch->n_to_notify, 0); | ||
520 | |||
521 | spin_lock_init(&ch->lock); | ||
522 | init_completion(&ch->wdisconnect_wait); | ||
523 | |||
524 | atomic_set(&ch->n_on_msg_allocate_wq, 0); | ||
525 | init_waitqueue_head(&ch->msg_allocate_wq); | ||
526 | init_waitqueue_head(&ch->idle_wq); | ||
527 | } | ||
528 | |||
529 | ret = xpc_setup_ch_structures_sn(part); | ||
530 | if (ret != xpSuccess) | ||
531 | goto out_2; | ||
532 | |||
533 | /* | ||
534 | * With the setting of the partition setup_state to XPC_P_SS_SETUP, | ||
535 | * we're declaring that this partition is ready to go. | ||
536 | */ | ||
537 | part->setup_state = XPC_P_SS_SETUP; | ||
538 | |||
539 | return xpSuccess; | ||
540 | |||
541 | /* setup of ch structures failed */ | ||
542 | out_2: | ||
543 | kfree(part->remote_openclose_args_base); | ||
544 | part->remote_openclose_args = NULL; | ||
545 | out_1: | ||
546 | kfree(part->channels); | ||
547 | part->channels = NULL; | ||
548 | return ret; | ||
549 | } | ||
550 | |||
551 | /* | ||
552 | * Teardown the channel structures necessary to support XPartition Communication | ||
553 | * between the specified remote partition and the local one. | ||
554 | */ | ||
555 | static void | ||
556 | xpc_teardown_ch_structures(struct xpc_partition *part) | ||
557 | { | ||
558 | DBUG_ON(atomic_read(&part->nchannels_engaged) != 0); | ||
559 | DBUG_ON(atomic_read(&part->nchannels_active) != 0); | ||
560 | |||
561 | /* | ||
562 | * Make this partition inaccessible to local processes by marking it | ||
563 | * as no longer setup. Then wait before proceeding with the teardown | ||
564 | * until all existing references cease. | ||
565 | */ | ||
566 | DBUG_ON(part->setup_state != XPC_P_SS_SETUP); | ||
567 | part->setup_state = XPC_P_SS_WTEARDOWN; | ||
568 | |||
569 | wait_event(part->teardown_wq, (atomic_read(&part->references) == 0)); | ||
570 | |||
571 | /* now we can begin tearing down the infrastructure */ | ||
572 | |||
573 | xpc_teardown_ch_structures_sn(part); | ||
574 | |||
575 | kfree(part->remote_openclose_args_base); | ||
576 | part->remote_openclose_args = NULL; | ||
577 | kfree(part->channels); | ||
578 | part->channels = NULL; | ||
579 | |||
580 | part->setup_state = XPC_P_SS_TORNDOWN; | ||
581 | } | ||
582 | |||
583 | /* | ||
440 | * When XPC HB determines that a partition has come up, it will create a new | 584 | * When XPC HB determines that a partition has come up, it will create a new |
441 | * kthread and that kthread will call this function to attempt to set up the | 585 | * kthread and that kthread will call this function to attempt to set up the |
442 | * basic infrastructure used for Cross Partition Communication with the newly | 586 | * basic infrastructure used for Cross Partition Communication with the newly |
@@ -476,7 +620,7 @@ xpc_activating(void *__partid) | |||
476 | 620 | ||
477 | xpc_allow_hb(partid); | 621 | xpc_allow_hb(partid); |
478 | 622 | ||
479 | if (xpc_setup_infrastructure(part) == xpSuccess) { | 623 | if (xpc_setup_ch_structures(part) == xpSuccess) { |
480 | (void)xpc_part_ref(part); /* this will always succeed */ | 624 | (void)xpc_part_ref(part); /* this will always succeed */ |
481 | 625 | ||
482 | if (xpc_make_first_contact(part) == xpSuccess) { | 626 | if (xpc_make_first_contact(part) == xpSuccess) { |
@@ -486,7 +630,7 @@ xpc_activating(void *__partid) | |||
486 | } | 630 | } |
487 | 631 | ||
488 | xpc_part_deref(part); | 632 | xpc_part_deref(part); |
489 | xpc_teardown_infrastructure(part); | 633 | xpc_teardown_ch_structures(part); |
490 | } | 634 | } |
491 | 635 | ||
492 | xpc_disallow_hb(partid); | 636 | xpc_disallow_hb(partid); |
@@ -806,6 +950,56 @@ xpc_disconnect_wait(int ch_number) | |||
806 | } | 950 | } |
807 | } | 951 | } |
808 | 952 | ||
953 | static int | ||
954 | xpc_setup_partitions(void) | ||
955 | { | ||
956 | short partid; | ||
957 | struct xpc_partition *part; | ||
958 | |||
959 | xpc_partitions = kzalloc(sizeof(struct xpc_partition) * | ||
960 | xp_max_npartitions, GFP_KERNEL); | ||
961 | if (xpc_partitions == NULL) { | ||
962 | dev_err(xpc_part, "can't get memory for partition structure\n"); | ||
963 | return -ENOMEM; | ||
964 | } | ||
965 | |||
966 | /* | ||
967 | * The first few fields of each entry of xpc_partitions[] need to | ||
968 | * be initialized now so that calls to xpc_connect() and | ||
969 | * xpc_disconnect() can be made prior to the activation of any remote | ||
970 | * partition. NOTE THAT NONE OF THE OTHER FIELDS BELONGING TO THESE | ||
971 | * ENTRIES ARE MEANINGFUL UNTIL AFTER AN ENTRY'S CORRESPONDING | ||
972 | * PARTITION HAS BEEN ACTIVATED. | ||
973 | */ | ||
974 | for (partid = 0; partid < xp_max_npartitions; partid++) { | ||
975 | part = &xpc_partitions[partid]; | ||
976 | |||
977 | DBUG_ON((u64)part != L1_CACHE_ALIGN((u64)part)); | ||
978 | |||
979 | part->activate_IRQ_rcvd = 0; | ||
980 | spin_lock_init(&part->act_lock); | ||
981 | part->act_state = XPC_P_AS_INACTIVE; | ||
982 | XPC_SET_REASON(part, 0, 0); | ||
983 | |||
984 | init_timer(&part->disengage_timer); | ||
985 | part->disengage_timer.function = | ||
986 | xpc_timeout_partition_disengage; | ||
987 | part->disengage_timer.data = (unsigned long)part; | ||
988 | |||
989 | part->setup_state = XPC_P_SS_UNSET; | ||
990 | init_waitqueue_head(&part->teardown_wq); | ||
991 | atomic_set(&part->references, 0); | ||
992 | } | ||
993 | |||
994 | return xpc_setup_partitions_sn(); | ||
995 | } | ||
996 | |||
997 | static void | ||
998 | xpc_teardown_partitions(void) | ||
999 | { | ||
1000 | kfree(xpc_partitions); | ||
1001 | } | ||
1002 | |||
809 | static void | 1003 | static void |
810 | xpc_do_exit(enum xp_retval reason) | 1004 | xpc_do_exit(enum xp_retval reason) |
811 | { | 1005 | { |
@@ -892,8 +1086,7 @@ xpc_do_exit(enum xp_retval reason) | |||
892 | DBUG_ON(xpc_any_partition_engaged()); | 1086 | DBUG_ON(xpc_any_partition_engaged()); |
893 | DBUG_ON(xpc_any_hbs_allowed() != 0); | 1087 | DBUG_ON(xpc_any_hbs_allowed() != 0); |
894 | 1088 | ||
895 | /* a zero timestamp indicates our rsvd page is not initialized */ | 1089 | xpc_teardown_rsvd_page(); |
896 | xpc_rsvd_page->ts_jiffies = 0; | ||
897 | 1090 | ||
898 | if (reason == xpUnloading) { | 1091 | if (reason == xpUnloading) { |
899 | (void)unregister_die_notifier(&xpc_die_notifier); | 1092 | (void)unregister_die_notifier(&xpc_die_notifier); |
@@ -906,7 +1099,7 @@ xpc_do_exit(enum xp_retval reason) | |||
906 | if (xpc_sysctl) | 1099 | if (xpc_sysctl) |
907 | unregister_sysctl_table(xpc_sysctl); | 1100 | unregister_sysctl_table(xpc_sysctl); |
908 | 1101 | ||
909 | kfree(xpc_partitions); | 1102 | xpc_teardown_partitions(); |
910 | 1103 | ||
911 | if (is_shub()) | 1104 | if (is_shub()) |
912 | xpc_exit_sn2(); | 1105 | xpc_exit_sn2(); |
@@ -1062,8 +1255,6 @@ int __init | |||
1062 | xpc_init(void) | 1255 | xpc_init(void) |
1063 | { | 1256 | { |
1064 | int ret; | 1257 | int ret; |
1065 | short partid; | ||
1066 | struct xpc_partition *part; | ||
1067 | struct task_struct *kthread; | 1258 | struct task_struct *kthread; |
1068 | 1259 | ||
1069 | snprintf(xpc_part->bus_id, BUS_ID_SIZE, "part"); | 1260 | snprintf(xpc_part->bus_id, BUS_ID_SIZE, "part"); |
@@ -1076,56 +1267,29 @@ xpc_init(void) | |||
1076 | * further to only support exactly 64 partitions on this | 1267 | * further to only support exactly 64 partitions on this |
1077 | * architecture, no less. | 1268 | * architecture, no less. |
1078 | */ | 1269 | */ |
1079 | if (xp_max_npartitions != 64) | 1270 | if (xp_max_npartitions != 64) { |
1080 | return -EINVAL; | 1271 | dev_err(xpc_part, "max #of partitions not set to 64\n"); |
1081 | 1272 | ret = -EINVAL; | |
1082 | ret = xpc_init_sn2(); | 1273 | } else { |
1083 | if (ret != 0) | 1274 | ret = xpc_init_sn2(); |
1084 | return ret; | 1275 | } |
1085 | 1276 | ||
1086 | } else if (is_uv()) { | 1277 | } else if (is_uv()) { |
1087 | xpc_init_uv(); | 1278 | ret = xpc_init_uv(); |
1088 | 1279 | ||
1089 | } else { | 1280 | } else { |
1090 | return -ENODEV; | 1281 | ret = -ENODEV; |
1091 | } | 1282 | } |
1092 | 1283 | ||
1093 | xpc_partitions = kzalloc(sizeof(struct xpc_partition) * | 1284 | if (ret != 0) |
1094 | xp_max_npartitions, GFP_KERNEL); | 1285 | return ret; |
1095 | if (xpc_partitions == NULL) { | 1286 | |
1287 | ret = xpc_setup_partitions(); | ||
1288 | if (ret != 0) { | ||
1096 | dev_err(xpc_part, "can't get memory for partition structure\n"); | 1289 | dev_err(xpc_part, "can't get memory for partition structure\n"); |
1097 | ret = -ENOMEM; | ||
1098 | goto out_1; | 1290 | goto out_1; |
1099 | } | 1291 | } |
1100 | 1292 | ||
1101 | /* | ||
1102 | * The first few fields of each entry of xpc_partitions[] need to | ||
1103 | * be initialized now so that calls to xpc_connect() and | ||
1104 | * xpc_disconnect() can be made prior to the activation of any remote | ||
1105 | * partition. NOTE THAT NONE OF THE OTHER FIELDS BELONGING TO THESE | ||
1106 | * ENTRIES ARE MEANINGFUL UNTIL AFTER AN ENTRY'S CORRESPONDING | ||
1107 | * PARTITION HAS BEEN ACTIVATED. | ||
1108 | */ | ||
1109 | for (partid = 0; partid < xp_max_npartitions; partid++) { | ||
1110 | part = &xpc_partitions[partid]; | ||
1111 | |||
1112 | DBUG_ON((u64)part != L1_CACHE_ALIGN((u64)part)); | ||
1113 | |||
1114 | part->activate_IRQ_rcvd = 0; | ||
1115 | spin_lock_init(&part->act_lock); | ||
1116 | part->act_state = XPC_P_AS_INACTIVE; | ||
1117 | XPC_SET_REASON(part, 0, 0); | ||
1118 | |||
1119 | init_timer(&part->disengage_timer); | ||
1120 | part->disengage_timer.function = | ||
1121 | xpc_timeout_partition_disengage; | ||
1122 | part->disengage_timer.data = (unsigned long)part; | ||
1123 | |||
1124 | part->setup_state = XPC_P_SS_UNSET; | ||
1125 | init_waitqueue_head(&part->teardown_wq); | ||
1126 | atomic_set(&part->references, 0); | ||
1127 | } | ||
1128 | |||
1129 | xpc_sysctl = register_sysctl_table(xpc_sys_dir); | 1293 | xpc_sysctl = register_sysctl_table(xpc_sys_dir); |
1130 | 1294 | ||
1131 | /* | 1295 | /* |
@@ -1133,10 +1297,9 @@ xpc_init(void) | |||
1133 | * other partitions to discover we are alive and establish initial | 1297 | * other partitions to discover we are alive and establish initial |
1134 | * communications. | 1298 | * communications. |
1135 | */ | 1299 | */ |
1136 | xpc_rsvd_page = xpc_setup_rsvd_page(); | 1300 | ret = xpc_setup_rsvd_page(); |
1137 | if (xpc_rsvd_page == NULL) { | 1301 | if (ret != 0) { |
1138 | dev_err(xpc_part, "can't setup our reserved page\n"); | 1302 | dev_err(xpc_part, "can't setup our reserved page\n"); |
1139 | ret = -EBUSY; | ||
1140 | goto out_2; | 1303 | goto out_2; |
1141 | } | 1304 | } |
1142 | 1305 | ||
@@ -1187,15 +1350,15 @@ xpc_init(void) | |||
1187 | 1350 | ||
1188 | /* initialization was not successful */ | 1351 | /* initialization was not successful */ |
1189 | out_3: | 1352 | out_3: |
1190 | /* a zero timestamp indicates our rsvd page is not initialized */ | 1353 | xpc_teardown_rsvd_page(); |
1191 | xpc_rsvd_page->ts_jiffies = 0; | ||
1192 | 1354 | ||
1193 | (void)unregister_die_notifier(&xpc_die_notifier); | 1355 | (void)unregister_die_notifier(&xpc_die_notifier); |
1194 | (void)unregister_reboot_notifier(&xpc_reboot_notifier); | 1356 | (void)unregister_reboot_notifier(&xpc_reboot_notifier); |
1195 | out_2: | 1357 | out_2: |
1196 | if (xpc_sysctl) | 1358 | if (xpc_sysctl) |
1197 | unregister_sysctl_table(xpc_sysctl); | 1359 | unregister_sysctl_table(xpc_sysctl); |
1198 | kfree(xpc_partitions); | 1360 | |
1361 | xpc_teardown_partitions(); | ||
1199 | out_1: | 1362 | out_1: |
1200 | if (is_shub()) | 1363 | if (is_shub()) |
1201 | xpc_exit_sn2(); | 1364 | xpc_exit_sn2(); |
diff --git a/drivers/misc/sgi-xp/xpc_partition.c b/drivers/misc/sgi-xp/xpc_partition.c index b5fb21641130..6722f6fe4dc7 100644 --- a/drivers/misc/sgi-xp/xpc_partition.c +++ b/drivers/misc/sgi-xp/xpc_partition.c | |||
@@ -73,6 +73,12 @@ xpc_get_rsvd_page_pa(int nasid) | |||
73 | 73 | ||
74 | while (1) { | 74 | while (1) { |
75 | 75 | ||
76 | /* !!! rp_pa will need to be _gpa on UV. | ||
77 | * ??? So do we save it into the architecture specific parts | ||
78 | * ??? of the xpc_partition structure? Do we rename this | ||
79 | * ??? function or have two versions? Rename rp_pa for UV to | ||
80 | * ??? rp_gpa? | ||
81 | */ | ||
76 | ret = xpc_get_partition_rsvd_page_pa(buf, &cookie, &rp_pa, | 82 | ret = xpc_get_partition_rsvd_page_pa(buf, &cookie, &rp_pa, |
77 | &len); | 83 | &len); |
78 | 84 | ||
@@ -118,9 +124,10 @@ xpc_get_rsvd_page_pa(int nasid) | |||
118 | * other partitions to discover we are alive and establish initial | 124 | * other partitions to discover we are alive and establish initial |
119 | * communications. | 125 | * communications. |
120 | */ | 126 | */ |
121 | struct xpc_rsvd_page * | 127 | int |
122 | xpc_setup_rsvd_page(void) | 128 | xpc_setup_rsvd_page(void) |
123 | { | 129 | { |
130 | int ret; | ||
124 | struct xpc_rsvd_page *rp; | 131 | struct xpc_rsvd_page *rp; |
125 | unsigned long rp_pa; | 132 | unsigned long rp_pa; |
126 | unsigned long new_ts_jiffies; | 133 | unsigned long new_ts_jiffies; |
@@ -132,7 +139,7 @@ xpc_setup_rsvd_page(void) | |||
132 | preempt_enable(); | 139 | preempt_enable(); |
133 | if (rp_pa == 0) { | 140 | if (rp_pa == 0) { |
134 | dev_err(xpc_part, "SAL failed to locate the reserved page\n"); | 141 | dev_err(xpc_part, "SAL failed to locate the reserved page\n"); |
135 | return NULL; | 142 | return -ESRCH; |
136 | } | 143 | } |
137 | rp = (struct xpc_rsvd_page *)__va(rp_pa); | 144 | rp = (struct xpc_rsvd_page *)__va(rp_pa); |
138 | 145 | ||
@@ -146,7 +153,7 @@ xpc_setup_rsvd_page(void) | |||
146 | dev_err(xpc_part, "the reserved page's partid of %d is outside " | 153 | dev_err(xpc_part, "the reserved page's partid of %d is outside " |
147 | "supported range (< 0 || >= %d)\n", rp->SAL_partid, | 154 | "supported range (< 0 || >= %d)\n", rp->SAL_partid, |
148 | xp_max_npartitions); | 155 | xp_max_npartitions); |
149 | return NULL; | 156 | return -EINVAL; |
150 | } | 157 | } |
151 | 158 | ||
152 | rp->version = XPC_RP_VERSION; | 159 | rp->version = XPC_RP_VERSION; |
@@ -165,8 +172,9 @@ xpc_setup_rsvd_page(void) | |||
165 | xpc_part_nasids = XPC_RP_PART_NASIDS(rp); | 172 | xpc_part_nasids = XPC_RP_PART_NASIDS(rp); |
166 | xpc_mach_nasids = XPC_RP_MACH_NASIDS(rp); | 173 | xpc_mach_nasids = XPC_RP_MACH_NASIDS(rp); |
167 | 174 | ||
168 | if (xpc_rsvd_page_init(rp) != xpSuccess) | 175 | ret = xpc_setup_rsvd_page_sn(rp); |
169 | return NULL; | 176 | if (ret != 0) |
177 | return ret; | ||
170 | 178 | ||
171 | /* | 179 | /* |
172 | * Set timestamp of when reserved page was setup by XPC. | 180 | * Set timestamp of when reserved page was setup by XPC. |
@@ -178,7 +186,15 @@ xpc_setup_rsvd_page(void) | |||
178 | new_ts_jiffies++; | 186 | new_ts_jiffies++; |
179 | rp->ts_jiffies = new_ts_jiffies; | 187 | rp->ts_jiffies = new_ts_jiffies; |
180 | 188 | ||
181 | return rp; | 189 | xpc_rsvd_page = rp; |
190 | return 0; | ||
191 | } | ||
192 | |||
193 | void | ||
194 | xpc_teardown_rsvd_page(void) | ||
195 | { | ||
196 | /* a zero timestamp indicates our rsvd page is not initialized */ | ||
197 | xpc_rsvd_page->ts_jiffies = 0; | ||
182 | } | 198 | } |
183 | 199 | ||
184 | /* | 200 | /* |
diff --git a/drivers/misc/sgi-xp/xpc_sn2.c b/drivers/misc/sgi-xp/xpc_sn2.c index d1ccadc0857d..8b4b0653d9e9 100644 --- a/drivers/misc/sgi-xp/xpc_sn2.c +++ b/drivers/misc/sgi-xp/xpc_sn2.c | |||
@@ -53,12 +53,19 @@ | |||
53 | * Buffer used to store a local copy of portions of a remote partition's | 53 | * Buffer used to store a local copy of portions of a remote partition's |
54 | * reserved page (either its header and part_nasids mask, or its vars). | 54 | * reserved page (either its header and part_nasids mask, or its vars). |
55 | */ | 55 | */ |
56 | static char *xpc_remote_copy_buffer_sn2; | ||
57 | static void *xpc_remote_copy_buffer_base_sn2; | 56 | static void *xpc_remote_copy_buffer_base_sn2; |
57 | static char *xpc_remote_copy_buffer_sn2; | ||
58 | 58 | ||
59 | static struct xpc_vars_sn2 *xpc_vars_sn2; | 59 | static struct xpc_vars_sn2 *xpc_vars_sn2; |
60 | static struct xpc_vars_part_sn2 *xpc_vars_part_sn2; | 60 | static struct xpc_vars_part_sn2 *xpc_vars_part_sn2; |
61 | 61 | ||
62 | static int | ||
63 | xpc_setup_partitions_sn_sn2(void) | ||
64 | { | ||
65 | /* nothing needs to be done */ | ||
66 | return 0; | ||
67 | } | ||
68 | |||
62 | /* SH_IPI_ACCESS shub register value on startup */ | 69 | /* SH_IPI_ACCESS shub register value on startup */ |
63 | static u64 xpc_sh1_IPI_access_sn2; | 70 | static u64 xpc_sh1_IPI_access_sn2; |
64 | static u64 xpc_sh2_IPI_access0_sn2; | 71 | static u64 xpc_sh2_IPI_access0_sn2; |
@@ -198,7 +205,12 @@ xpc_init_IRQ_amo_sn2(int index) | |||
198 | static irqreturn_t | 205 | static irqreturn_t |
199 | xpc_handle_activate_IRQ_sn2(int irq, void *dev_id) | 206 | xpc_handle_activate_IRQ_sn2(int irq, void *dev_id) |
200 | { | 207 | { |
201 | atomic_inc(&xpc_activate_IRQ_rcvd); | 208 | unsigned long irq_flags; |
209 | |||
210 | spin_lock_irqsave(&xpc_activate_IRQ_rcvd_lock, irq_flags); | ||
211 | xpc_activate_IRQ_rcvd++; | ||
212 | spin_unlock_irqrestore(&xpc_activate_IRQ_rcvd_lock, irq_flags); | ||
213 | |||
202 | wake_up_interruptible(&xpc_activate_IRQ_wq); | 214 | wake_up_interruptible(&xpc_activate_IRQ_wq); |
203 | return IRQ_HANDLED; | 215 | return IRQ_HANDLED; |
204 | } | 216 | } |
@@ -222,6 +234,7 @@ xpc_send_activate_IRQ_sn2(unsigned long amos_page_pa, int from_nasid, | |||
222 | static void | 234 | static void |
223 | xpc_send_local_activate_IRQ_sn2(int from_nasid) | 235 | xpc_send_local_activate_IRQ_sn2(int from_nasid) |
224 | { | 236 | { |
237 | unsigned long irq_flags; | ||
225 | struct amo *amos = (struct amo *)__va(xpc_vars_sn2->amos_page_pa + | 238 | struct amo *amos = (struct amo *)__va(xpc_vars_sn2->amos_page_pa + |
226 | (XPC_ACTIVATE_IRQ_AMOS_SN2 * | 239 | (XPC_ACTIVATE_IRQ_AMOS_SN2 * |
227 | sizeof(struct amo))); | 240 | sizeof(struct amo))); |
@@ -230,7 +243,10 @@ xpc_send_local_activate_IRQ_sn2(int from_nasid) | |||
230 | FETCHOP_STORE_OP(TO_AMO((u64)&amos[BIT_WORD(from_nasid / 2)].variable), | 243 | FETCHOP_STORE_OP(TO_AMO((u64)&amos[BIT_WORD(from_nasid / 2)].variable), |
231 | FETCHOP_OR, BIT_MASK(from_nasid / 2)); | 244 | FETCHOP_OR, BIT_MASK(from_nasid / 2)); |
232 | 245 | ||
233 | atomic_inc(&xpc_activate_IRQ_rcvd); | 246 | spin_lock_irqsave(&xpc_activate_IRQ_rcvd_lock, irq_flags); |
247 | xpc_activate_IRQ_rcvd++; | ||
248 | spin_unlock_irqrestore(&xpc_activate_IRQ_rcvd_lock, irq_flags); | ||
249 | |||
234 | wake_up_interruptible(&xpc_activate_IRQ_wq); | 250 | wake_up_interruptible(&xpc_activate_IRQ_wq); |
235 | } | 251 | } |
236 | 252 | ||
@@ -375,7 +391,7 @@ static void | |||
375 | xpc_send_chctl_closerequest_sn2(struct xpc_channel *ch, | 391 | xpc_send_chctl_closerequest_sn2(struct xpc_channel *ch, |
376 | unsigned long *irq_flags) | 392 | unsigned long *irq_flags) |
377 | { | 393 | { |
378 | struct xpc_openclose_args *args = ch->local_openclose_args; | 394 | struct xpc_openclose_args *args = ch->sn.sn2.local_openclose_args; |
379 | 395 | ||
380 | args->reason = ch->reason; | 396 | args->reason = ch->reason; |
381 | XPC_SEND_NOTIFY_IRQ_SN2(ch, XPC_CHCTL_CLOSEREQUEST, irq_flags); | 397 | XPC_SEND_NOTIFY_IRQ_SN2(ch, XPC_CHCTL_CLOSEREQUEST, irq_flags); |
@@ -390,7 +406,7 @@ xpc_send_chctl_closereply_sn2(struct xpc_channel *ch, unsigned long *irq_flags) | |||
390 | static void | 406 | static void |
391 | xpc_send_chctl_openrequest_sn2(struct xpc_channel *ch, unsigned long *irq_flags) | 407 | xpc_send_chctl_openrequest_sn2(struct xpc_channel *ch, unsigned long *irq_flags) |
392 | { | 408 | { |
393 | struct xpc_openclose_args *args = ch->local_openclose_args; | 409 | struct xpc_openclose_args *args = ch->sn.sn2.local_openclose_args; |
394 | 410 | ||
395 | args->msg_size = ch->msg_size; | 411 | args->msg_size = ch->msg_size; |
396 | args->local_nentries = ch->local_nentries; | 412 | args->local_nentries = ch->local_nentries; |
@@ -400,11 +416,11 @@ xpc_send_chctl_openrequest_sn2(struct xpc_channel *ch, unsigned long *irq_flags) | |||
400 | static void | 416 | static void |
401 | xpc_send_chctl_openreply_sn2(struct xpc_channel *ch, unsigned long *irq_flags) | 417 | xpc_send_chctl_openreply_sn2(struct xpc_channel *ch, unsigned long *irq_flags) |
402 | { | 418 | { |
403 | struct xpc_openclose_args *args = ch->local_openclose_args; | 419 | struct xpc_openclose_args *args = ch->sn.sn2.local_openclose_args; |
404 | 420 | ||
405 | args->remote_nentries = ch->remote_nentries; | 421 | args->remote_nentries = ch->remote_nentries; |
406 | args->local_nentries = ch->local_nentries; | 422 | args->local_nentries = ch->local_nentries; |
407 | args->local_msgqueue_pa = xp_pa(ch->local_msgqueue); | 423 | args->local_msgqueue_pa = xp_pa(ch->sn.sn2.local_msgqueue); |
408 | XPC_SEND_NOTIFY_IRQ_SN2(ch, XPC_CHCTL_OPENREPLY, irq_flags); | 424 | XPC_SEND_NOTIFY_IRQ_SN2(ch, XPC_CHCTL_OPENREPLY, irq_flags); |
409 | } | 425 | } |
410 | 426 | ||
@@ -420,6 +436,13 @@ xpc_send_chctl_local_msgrequest_sn2(struct xpc_channel *ch) | |||
420 | XPC_SEND_LOCAL_NOTIFY_IRQ_SN2(ch, XPC_CHCTL_MSGREQUEST); | 436 | XPC_SEND_LOCAL_NOTIFY_IRQ_SN2(ch, XPC_CHCTL_MSGREQUEST); |
421 | } | 437 | } |
422 | 438 | ||
439 | static void | ||
440 | xpc_save_remote_msgqueue_pa_sn2(struct xpc_channel *ch, | ||
441 | unsigned long msgqueue_pa) | ||
442 | { | ||
443 | ch->sn.sn2.remote_msgqueue_pa = msgqueue_pa; | ||
444 | } | ||
445 | |||
423 | /* | 446 | /* |
424 | * This next set of functions are used to keep track of when a partition is | 447 | * This next set of functions are used to keep track of when a partition is |
425 | * potentially engaged in accessing memory belonging to another partition. | 448 | * potentially engaged in accessing memory belonging to another partition. |
@@ -489,6 +512,17 @@ xpc_indicate_partition_disengaged_sn2(struct xpc_partition *part) | |||
489 | part_sn2->activate_IRQ_phys_cpuid); | 512 | part_sn2->activate_IRQ_phys_cpuid); |
490 | } | 513 | } |
491 | 514 | ||
515 | static void | ||
516 | xpc_assume_partition_disengaged_sn2(short partid) | ||
517 | { | ||
518 | struct amo *amo = xpc_vars_sn2->amos_page + | ||
519 | XPC_ENGAGED_PARTITIONS_AMO_SN2; | ||
520 | |||
521 | /* clear bit(s) based on partid mask in our partition's amo */ | ||
522 | FETCHOP_STORE_OP(TO_AMO((u64)&amo->variable), FETCHOP_AND, | ||
523 | ~BIT(partid)); | ||
524 | } | ||
525 | |||
492 | static int | 526 | static int |
493 | xpc_partition_engaged_sn2(short partid) | 527 | xpc_partition_engaged_sn2(short partid) |
494 | { | 528 | { |
@@ -510,17 +544,6 @@ xpc_any_partition_engaged_sn2(void) | |||
510 | return FETCHOP_LOAD_OP(TO_AMO((u64)&amo->variable), FETCHOP_LOAD) != 0; | 544 | return FETCHOP_LOAD_OP(TO_AMO((u64)&amo->variable), FETCHOP_LOAD) != 0; |
511 | } | 545 | } |
512 | 546 | ||
513 | static void | ||
514 | xpc_assume_partition_disengaged_sn2(short partid) | ||
515 | { | ||
516 | struct amo *amo = xpc_vars_sn2->amos_page + | ||
517 | XPC_ENGAGED_PARTITIONS_AMO_SN2; | ||
518 | |||
519 | /* clear bit(s) based on partid mask in our partition's amo */ | ||
520 | FETCHOP_STORE_OP(TO_AMO((u64)&amo->variable), FETCHOP_AND, | ||
521 | ~BIT(partid)); | ||
522 | } | ||
523 | |||
524 | /* original protection values for each node */ | 547 | /* original protection values for each node */ |
525 | static u64 xpc_prot_vec_sn2[MAX_NUMNODES]; | 548 | static u64 xpc_prot_vec_sn2[MAX_NUMNODES]; |
526 | 549 | ||
@@ -595,8 +618,8 @@ xpc_get_partition_rsvd_page_pa_sn2(void *buf, u64 *cookie, unsigned long *rp_pa, | |||
595 | } | 618 | } |
596 | 619 | ||
597 | 620 | ||
598 | static enum xp_retval | 621 | static int |
599 | xpc_rsvd_page_init_sn2(struct xpc_rsvd_page *rp) | 622 | xpc_setup_rsvd_page_sn_sn2(struct xpc_rsvd_page *rp) |
600 | { | 623 | { |
601 | struct amo *amos_page; | 624 | struct amo *amos_page; |
602 | int i; | 625 | int i; |
@@ -627,7 +650,7 @@ xpc_rsvd_page_init_sn2(struct xpc_rsvd_page *rp) | |||
627 | amos_page = (struct amo *)TO_AMO(uncached_alloc_page(0, 1)); | 650 | amos_page = (struct amo *)TO_AMO(uncached_alloc_page(0, 1)); |
628 | if (amos_page == NULL) { | 651 | if (amos_page == NULL) { |
629 | dev_err(xpc_part, "can't allocate page of amos\n"); | 652 | dev_err(xpc_part, "can't allocate page of amos\n"); |
630 | return xpNoMemory; | 653 | return -ENOMEM; |
631 | } | 654 | } |
632 | 655 | ||
633 | /* | 656 | /* |
@@ -639,7 +662,7 @@ xpc_rsvd_page_init_sn2(struct xpc_rsvd_page *rp) | |||
639 | dev_err(xpc_part, "can't allow amo operations\n"); | 662 | dev_err(xpc_part, "can't allow amo operations\n"); |
640 | uncached_free_page(__IA64_UNCACHED_OFFSET | | 663 | uncached_free_page(__IA64_UNCACHED_OFFSET | |
641 | TO_PHYS((u64)amos_page), 1); | 664 | TO_PHYS((u64)amos_page), 1); |
642 | return ret; | 665 | return -EPERM; |
643 | } | 666 | } |
644 | } | 667 | } |
645 | 668 | ||
@@ -665,7 +688,7 @@ xpc_rsvd_page_init_sn2(struct xpc_rsvd_page *rp) | |||
665 | (void)xpc_init_IRQ_amo_sn2(XPC_ENGAGED_PARTITIONS_AMO_SN2); | 688 | (void)xpc_init_IRQ_amo_sn2(XPC_ENGAGED_PARTITIONS_AMO_SN2); |
666 | (void)xpc_init_IRQ_amo_sn2(XPC_DEACTIVATE_REQUEST_AMO_SN2); | 689 | (void)xpc_init_IRQ_amo_sn2(XPC_DEACTIVATE_REQUEST_AMO_SN2); |
667 | 690 | ||
668 | return xpSuccess; | 691 | return 0; |
669 | } | 692 | } |
670 | 693 | ||
671 | static void | 694 | static void |
@@ -1082,10 +1105,19 @@ xpc_identify_activate_IRQ_sender_sn2(void) | |||
1082 | } | 1105 | } |
1083 | 1106 | ||
1084 | static void | 1107 | static void |
1085 | xpc_process_activate_IRQ_rcvd_sn2(int n_IRQs_expected) | 1108 | xpc_process_activate_IRQ_rcvd_sn2(void) |
1086 | { | 1109 | { |
1110 | unsigned long irq_flags; | ||
1111 | int n_IRQs_expected; | ||
1087 | int n_IRQs_detected; | 1112 | int n_IRQs_detected; |
1088 | 1113 | ||
1114 | DBUG_ON(xpc_activate_IRQ_rcvd == 0); | ||
1115 | |||
1116 | spin_lock_irqsave(&xpc_activate_IRQ_rcvd_lock, irq_flags); | ||
1117 | n_IRQs_expected = xpc_activate_IRQ_rcvd; | ||
1118 | xpc_activate_IRQ_rcvd = 0; | ||
1119 | spin_unlock_irqrestore(&xpc_activate_IRQ_rcvd_lock, irq_flags); | ||
1120 | |||
1089 | n_IRQs_detected = xpc_identify_activate_IRQ_sender_sn2(); | 1121 | n_IRQs_detected = xpc_identify_activate_IRQ_sender_sn2(); |
1090 | if (n_IRQs_detected < n_IRQs_expected) { | 1122 | if (n_IRQs_detected < n_IRQs_expected) { |
1091 | /* retry once to help avoid missing amo */ | 1123 | /* retry once to help avoid missing amo */ |
@@ -1094,116 +1126,63 @@ xpc_process_activate_IRQ_rcvd_sn2(int n_IRQs_expected) | |||
1094 | } | 1126 | } |
1095 | 1127 | ||
1096 | /* | 1128 | /* |
1097 | * Guarantee that the kzalloc'd memory is cacheline aligned. | 1129 | * Setup the channel structures that are sn2 specific. |
1098 | */ | ||
1099 | static void * | ||
1100 | xpc_kzalloc_cacheline_aligned_sn2(size_t size, gfp_t flags, void **base) | ||
1101 | { | ||
1102 | /* see if kzalloc will give us cachline aligned memory by default */ | ||
1103 | *base = kzalloc(size, flags); | ||
1104 | if (*base == NULL) | ||
1105 | return NULL; | ||
1106 | |||
1107 | if ((u64)*base == L1_CACHE_ALIGN((u64)*base)) | ||
1108 | return *base; | ||
1109 | |||
1110 | kfree(*base); | ||
1111 | |||
1112 | /* nope, we'll have to do it ourselves */ | ||
1113 | *base = kzalloc(size + L1_CACHE_BYTES, flags); | ||
1114 | if (*base == NULL) | ||
1115 | return NULL; | ||
1116 | |||
1117 | return (void *)L1_CACHE_ALIGN((u64)*base); | ||
1118 | } | ||
1119 | |||
1120 | /* | ||
1121 | * Setup the infrastructure necessary to support XPartition Communication | ||
1122 | * between the specified remote partition and the local one. | ||
1123 | */ | 1130 | */ |
1124 | static enum xp_retval | 1131 | static enum xp_retval |
1125 | xpc_setup_infrastructure_sn2(struct xpc_partition *part) | 1132 | xpc_setup_ch_structures_sn_sn2(struct xpc_partition *part) |
1126 | { | 1133 | { |
1127 | struct xpc_partition_sn2 *part_sn2 = &part->sn.sn2; | 1134 | struct xpc_partition_sn2 *part_sn2 = &part->sn.sn2; |
1135 | struct xpc_channel_sn2 *ch_sn2; | ||
1128 | enum xp_retval retval; | 1136 | enum xp_retval retval; |
1129 | int ret; | 1137 | int ret; |
1130 | int cpuid; | 1138 | int cpuid; |
1131 | int ch_number; | 1139 | int ch_number; |
1132 | struct xpc_channel *ch; | ||
1133 | struct timer_list *timer; | 1140 | struct timer_list *timer; |
1134 | short partid = XPC_PARTID(part); | 1141 | short partid = XPC_PARTID(part); |
1135 | 1142 | ||
1136 | /* | ||
1137 | * Allocate all of the channel structures as a contiguous chunk of | ||
1138 | * memory. | ||
1139 | */ | ||
1140 | DBUG_ON(part->channels != NULL); | ||
1141 | part->channels = kzalloc(sizeof(struct xpc_channel) * XPC_MAX_NCHANNELS, | ||
1142 | GFP_KERNEL); | ||
1143 | if (part->channels == NULL) { | ||
1144 | dev_err(xpc_chan, "can't get memory for channels\n"); | ||
1145 | return xpNoMemory; | ||
1146 | } | ||
1147 | |||
1148 | /* allocate all the required GET/PUT values */ | 1143 | /* allocate all the required GET/PUT values */ |
1149 | 1144 | ||
1150 | part_sn2->local_GPs = | 1145 | part_sn2->local_GPs = |
1151 | xpc_kzalloc_cacheline_aligned_sn2(XPC_GP_SIZE, GFP_KERNEL, | 1146 | xpc_kzalloc_cacheline_aligned(XPC_GP_SIZE, GFP_KERNEL, |
1152 | &part_sn2->local_GPs_base); | 1147 | &part_sn2->local_GPs_base); |
1153 | if (part_sn2->local_GPs == NULL) { | 1148 | if (part_sn2->local_GPs == NULL) { |
1154 | dev_err(xpc_chan, "can't get memory for local get/put " | 1149 | dev_err(xpc_chan, "can't get memory for local get/put " |
1155 | "values\n"); | 1150 | "values\n"); |
1156 | retval = xpNoMemory; | 1151 | return xpNoMemory; |
1157 | goto out_1; | ||
1158 | } | 1152 | } |
1159 | 1153 | ||
1160 | part_sn2->remote_GPs = | 1154 | part_sn2->remote_GPs = |
1161 | xpc_kzalloc_cacheline_aligned_sn2(XPC_GP_SIZE, GFP_KERNEL, | 1155 | xpc_kzalloc_cacheline_aligned(XPC_GP_SIZE, GFP_KERNEL, |
1162 | &part_sn2->remote_GPs_base); | 1156 | &part_sn2->remote_GPs_base); |
1163 | if (part_sn2->remote_GPs == NULL) { | 1157 | if (part_sn2->remote_GPs == NULL) { |
1164 | dev_err(xpc_chan, "can't get memory for remote get/put " | 1158 | dev_err(xpc_chan, "can't get memory for remote get/put " |
1165 | "values\n"); | 1159 | "values\n"); |
1166 | retval = xpNoMemory; | 1160 | retval = xpNoMemory; |
1167 | goto out_2; | 1161 | goto out_1; |
1168 | } | 1162 | } |
1169 | 1163 | ||
1170 | part_sn2->remote_GPs_pa = 0; | 1164 | part_sn2->remote_GPs_pa = 0; |
1171 | 1165 | ||
1172 | /* allocate all the required open and close args */ | 1166 | /* allocate all the required open and close args */ |
1173 | 1167 | ||
1174 | part->local_openclose_args = | 1168 | part_sn2->local_openclose_args = |
1175 | xpc_kzalloc_cacheline_aligned_sn2(XPC_OPENCLOSE_ARGS_SIZE, | 1169 | xpc_kzalloc_cacheline_aligned(XPC_OPENCLOSE_ARGS_SIZE, |
1176 | GFP_KERNEL, | 1170 | GFP_KERNEL, &part_sn2-> |
1177 | &part->local_openclose_args_base); | 1171 | local_openclose_args_base); |
1178 | if (part->local_openclose_args == NULL) { | 1172 | if (part_sn2->local_openclose_args == NULL) { |
1179 | dev_err(xpc_chan, "can't get memory for local connect args\n"); | 1173 | dev_err(xpc_chan, "can't get memory for local connect args\n"); |
1180 | retval = xpNoMemory; | 1174 | retval = xpNoMemory; |
1181 | goto out_3; | 1175 | goto out_2; |
1182 | } | ||
1183 | |||
1184 | part->remote_openclose_args = | ||
1185 | xpc_kzalloc_cacheline_aligned_sn2(XPC_OPENCLOSE_ARGS_SIZE, | ||
1186 | GFP_KERNEL, | ||
1187 | &part->remote_openclose_args_base); | ||
1188 | if (part->remote_openclose_args == NULL) { | ||
1189 | dev_err(xpc_chan, "can't get memory for remote connect args\n"); | ||
1190 | retval = xpNoMemory; | ||
1191 | goto out_4; | ||
1192 | } | 1176 | } |
1193 | 1177 | ||
1194 | part_sn2->remote_openclose_args_pa = 0; | 1178 | part_sn2->remote_openclose_args_pa = 0; |
1195 | 1179 | ||
1196 | part_sn2->local_chctl_amo_va = xpc_init_IRQ_amo_sn2(partid); | 1180 | part_sn2->local_chctl_amo_va = xpc_init_IRQ_amo_sn2(partid); |
1197 | part->chctl.all_flags = 0; | ||
1198 | spin_lock_init(&part->chctl_lock); | ||
1199 | 1181 | ||
1200 | part_sn2->notify_IRQ_nasid = 0; | 1182 | part_sn2->notify_IRQ_nasid = 0; |
1201 | part_sn2->notify_IRQ_phys_cpuid = 0; | 1183 | part_sn2->notify_IRQ_phys_cpuid = 0; |
1202 | part_sn2->remote_chctl_amo_va = NULL; | 1184 | part_sn2->remote_chctl_amo_va = NULL; |
1203 | 1185 | ||
1204 | atomic_set(&part->channel_mgr_requests, 1); | ||
1205 | init_waitqueue_head(&part->channel_mgr_wq); | ||
1206 | |||
1207 | sprintf(part_sn2->notify_IRQ_owner, "xpc%02d", partid); | 1186 | sprintf(part_sn2->notify_IRQ_owner, "xpc%02d", partid); |
1208 | ret = request_irq(SGI_XPC_NOTIFY, xpc_handle_notify_IRQ_sn2, | 1187 | ret = request_irq(SGI_XPC_NOTIFY, xpc_handle_notify_IRQ_sn2, |
1209 | IRQF_SHARED, part_sn2->notify_IRQ_owner, | 1188 | IRQF_SHARED, part_sn2->notify_IRQ_owner, |
@@ -1212,7 +1191,7 @@ xpc_setup_infrastructure_sn2(struct xpc_partition *part) | |||
1212 | dev_err(xpc_chan, "can't register NOTIFY IRQ handler, " | 1191 | dev_err(xpc_chan, "can't register NOTIFY IRQ handler, " |
1213 | "errno=%d\n", -ret); | 1192 | "errno=%d\n", -ret); |
1214 | retval = xpLackOfResources; | 1193 | retval = xpLackOfResources; |
1215 | goto out_5; | 1194 | goto out_3; |
1216 | } | 1195 | } |
1217 | 1196 | ||
1218 | /* Setup a timer to check for dropped notify IRQs */ | 1197 | /* Setup a timer to check for dropped notify IRQs */ |
@@ -1224,45 +1203,17 @@ xpc_setup_infrastructure_sn2(struct xpc_partition *part) | |||
1224 | timer->expires = jiffies + XPC_DROPPED_NOTIFY_IRQ_WAIT_INTERVAL; | 1203 | timer->expires = jiffies + XPC_DROPPED_NOTIFY_IRQ_WAIT_INTERVAL; |
1225 | add_timer(timer); | 1204 | add_timer(timer); |
1226 | 1205 | ||
1227 | part->nchannels = XPC_MAX_NCHANNELS; | ||
1228 | |||
1229 | atomic_set(&part->nchannels_active, 0); | ||
1230 | atomic_set(&part->nchannels_engaged, 0); | ||
1231 | |||
1232 | for (ch_number = 0; ch_number < part->nchannels; ch_number++) { | 1206 | for (ch_number = 0; ch_number < part->nchannels; ch_number++) { |
1233 | ch = &part->channels[ch_number]; | 1207 | ch_sn2 = &part->channels[ch_number].sn.sn2; |
1234 | |||
1235 | ch->partid = partid; | ||
1236 | ch->number = ch_number; | ||
1237 | ch->flags = XPC_C_DISCONNECTED; | ||
1238 | |||
1239 | ch->sn.sn2.local_GP = &part_sn2->local_GPs[ch_number]; | ||
1240 | ch->local_openclose_args = | ||
1241 | &part->local_openclose_args[ch_number]; | ||
1242 | |||
1243 | atomic_set(&ch->kthreads_assigned, 0); | ||
1244 | atomic_set(&ch->kthreads_idle, 0); | ||
1245 | atomic_set(&ch->kthreads_active, 0); | ||
1246 | 1208 | ||
1247 | atomic_set(&ch->references, 0); | 1209 | ch_sn2->local_GP = &part_sn2->local_GPs[ch_number]; |
1248 | atomic_set(&ch->n_to_notify, 0); | 1210 | ch_sn2->local_openclose_args = |
1211 | &part_sn2->local_openclose_args[ch_number]; | ||
1249 | 1212 | ||
1250 | spin_lock_init(&ch->lock); | 1213 | mutex_init(&ch_sn2->msg_to_pull_mutex); |
1251 | mutex_init(&ch->sn.sn2.msg_to_pull_mutex); | ||
1252 | init_completion(&ch->wdisconnect_wait); | ||
1253 | |||
1254 | atomic_set(&ch->n_on_msg_allocate_wq, 0); | ||
1255 | init_waitqueue_head(&ch->msg_allocate_wq); | ||
1256 | init_waitqueue_head(&ch->idle_wq); | ||
1257 | } | 1214 | } |
1258 | 1215 | ||
1259 | /* | 1216 | /* |
1260 | * With the setting of the partition setup_state to XPC_P_SS_SETUP, | ||
1261 | * we're declaring that this partition is ready to go. | ||
1262 | */ | ||
1263 | part->setup_state = XPC_P_SS_SETUP; | ||
1264 | |||
1265 | /* | ||
1266 | * Setup the per partition specific variables required by the | 1217 | * Setup the per partition specific variables required by the |
1267 | * remote partition to establish channel connections with us. | 1218 | * remote partition to establish channel connections with us. |
1268 | * | 1219 | * |
@@ -1271,7 +1222,7 @@ xpc_setup_infrastructure_sn2(struct xpc_partition *part) | |||
1271 | */ | 1222 | */ |
1272 | xpc_vars_part_sn2[partid].GPs_pa = xp_pa(part_sn2->local_GPs); | 1223 | xpc_vars_part_sn2[partid].GPs_pa = xp_pa(part_sn2->local_GPs); |
1273 | xpc_vars_part_sn2[partid].openclose_args_pa = | 1224 | xpc_vars_part_sn2[partid].openclose_args_pa = |
1274 | xp_pa(part->local_openclose_args); | 1225 | xp_pa(part_sn2->local_openclose_args); |
1275 | xpc_vars_part_sn2[partid].chctl_amo_pa = | 1226 | xpc_vars_part_sn2[partid].chctl_amo_pa = |
1276 | xp_pa(part_sn2->local_chctl_amo_va); | 1227 | xp_pa(part_sn2->local_chctl_amo_va); |
1277 | cpuid = raw_smp_processor_id(); /* any CPU in this partition will do */ | 1228 | cpuid = raw_smp_processor_id(); /* any CPU in this partition will do */ |
@@ -1279,80 +1230,48 @@ xpc_setup_infrastructure_sn2(struct xpc_partition *part) | |||
1279 | xpc_vars_part_sn2[partid].notify_IRQ_phys_cpuid = | 1230 | xpc_vars_part_sn2[partid].notify_IRQ_phys_cpuid = |
1280 | cpu_physical_id(cpuid); | 1231 | cpu_physical_id(cpuid); |
1281 | xpc_vars_part_sn2[partid].nchannels = part->nchannels; | 1232 | xpc_vars_part_sn2[partid].nchannels = part->nchannels; |
1282 | xpc_vars_part_sn2[partid].magic = XPC_VP_MAGIC1; | 1233 | xpc_vars_part_sn2[partid].magic = XPC_VP_MAGIC1_SN2; |
1283 | 1234 | ||
1284 | return xpSuccess; | 1235 | return xpSuccess; |
1285 | 1236 | ||
1286 | /* setup of infrastructure failed */ | 1237 | /* setup of ch structures failed */ |
1287 | out_5: | ||
1288 | kfree(part->remote_openclose_args_base); | ||
1289 | part->remote_openclose_args = NULL; | ||
1290 | out_4: | ||
1291 | kfree(part->local_openclose_args_base); | ||
1292 | part->local_openclose_args = NULL; | ||
1293 | out_3: | 1238 | out_3: |
1239 | kfree(part_sn2->local_openclose_args_base); | ||
1240 | part_sn2->local_openclose_args = NULL; | ||
1241 | out_2: | ||
1294 | kfree(part_sn2->remote_GPs_base); | 1242 | kfree(part_sn2->remote_GPs_base); |
1295 | part_sn2->remote_GPs = NULL; | 1243 | part_sn2->remote_GPs = NULL; |
1296 | out_2: | 1244 | out_1: |
1297 | kfree(part_sn2->local_GPs_base); | 1245 | kfree(part_sn2->local_GPs_base); |
1298 | part_sn2->local_GPs = NULL; | 1246 | part_sn2->local_GPs = NULL; |
1299 | out_1: | ||
1300 | kfree(part->channels); | ||
1301 | part->channels = NULL; | ||
1302 | return retval; | 1247 | return retval; |
1303 | } | 1248 | } |
1304 | 1249 | ||
1305 | /* | 1250 | /* |
1306 | * Teardown the infrastructure necessary to support XPartition Communication | 1251 | * Teardown the channel structures that are sn2 specific. |
1307 | * between the specified remote partition and the local one. | ||
1308 | */ | 1252 | */ |
1309 | static void | 1253 | static void |
1310 | xpc_teardown_infrastructure_sn2(struct xpc_partition *part) | 1254 | xpc_teardown_ch_structures_sn_sn2(struct xpc_partition *part) |
1311 | { | 1255 | { |
1312 | struct xpc_partition_sn2 *part_sn2 = &part->sn.sn2; | 1256 | struct xpc_partition_sn2 *part_sn2 = &part->sn.sn2; |
1313 | short partid = XPC_PARTID(part); | 1257 | short partid = XPC_PARTID(part); |
1314 | 1258 | ||
1315 | /* | 1259 | /* |
1316 | * We start off by making this partition inaccessible to local | 1260 | * Indicate that the variables specific to the remote partition are no |
1317 | * processes by marking it as no longer setup. Then we make it | 1261 | * longer available for its use. |
1318 | * inaccessible to remote processes by clearing the XPC per partition | ||
1319 | * specific variable's magic # (which indicates that these variables | ||
1320 | * are no longer valid) and by ignoring all XPC notify IRQs sent to | ||
1321 | * this partition. | ||
1322 | */ | 1262 | */ |
1323 | |||
1324 | DBUG_ON(atomic_read(&part->nchannels_engaged) != 0); | ||
1325 | DBUG_ON(atomic_read(&part->nchannels_active) != 0); | ||
1326 | DBUG_ON(part->setup_state != XPC_P_SS_SETUP); | ||
1327 | part->setup_state = XPC_P_SS_WTEARDOWN; | ||
1328 | |||
1329 | xpc_vars_part_sn2[partid].magic = 0; | 1263 | xpc_vars_part_sn2[partid].magic = 0; |
1330 | 1264 | ||
1331 | free_irq(SGI_XPC_NOTIFY, (void *)(u64)partid); | ||
1332 | |||
1333 | /* | ||
1334 | * Before proceeding with the teardown we have to wait until all | ||
1335 | * existing references cease. | ||
1336 | */ | ||
1337 | wait_event(part->teardown_wq, (atomic_read(&part->references) == 0)); | ||
1338 | |||
1339 | /* now we can begin tearing down the infrastructure */ | ||
1340 | |||
1341 | part->setup_state = XPC_P_SS_TORNDOWN; | ||
1342 | |||
1343 | /* in case we've still got outstanding timers registered... */ | 1265 | /* in case we've still got outstanding timers registered... */ |
1344 | del_timer_sync(&part_sn2->dropped_notify_IRQ_timer); | 1266 | del_timer_sync(&part_sn2->dropped_notify_IRQ_timer); |
1267 | free_irq(SGI_XPC_NOTIFY, (void *)(u64)partid); | ||
1345 | 1268 | ||
1346 | kfree(part->remote_openclose_args_base); | 1269 | kfree(part_sn2->local_openclose_args_base); |
1347 | part->remote_openclose_args = NULL; | 1270 | part_sn2->local_openclose_args = NULL; |
1348 | kfree(part->local_openclose_args_base); | ||
1349 | part->local_openclose_args = NULL; | ||
1350 | kfree(part_sn2->remote_GPs_base); | 1271 | kfree(part_sn2->remote_GPs_base); |
1351 | part_sn2->remote_GPs = NULL; | 1272 | part_sn2->remote_GPs = NULL; |
1352 | kfree(part_sn2->local_GPs_base); | 1273 | kfree(part_sn2->local_GPs_base); |
1353 | part_sn2->local_GPs = NULL; | 1274 | part_sn2->local_GPs = NULL; |
1354 | kfree(part->channels); | ||
1355 | part->channels = NULL; | ||
1356 | part_sn2->local_chctl_amo_va = NULL; | 1275 | part_sn2->local_chctl_amo_va = NULL; |
1357 | } | 1276 | } |
1358 | 1277 | ||
@@ -1429,8 +1348,8 @@ xpc_pull_remote_vars_part_sn2(struct xpc_partition *part) | |||
1429 | 1348 | ||
1430 | /* see if they've been set up yet */ | 1349 | /* see if they've been set up yet */ |
1431 | 1350 | ||
1432 | if (pulled_entry->magic != XPC_VP_MAGIC1 && | 1351 | if (pulled_entry->magic != XPC_VP_MAGIC1_SN2 && |
1433 | pulled_entry->magic != XPC_VP_MAGIC2) { | 1352 | pulled_entry->magic != XPC_VP_MAGIC2_SN2) { |
1434 | 1353 | ||
1435 | if (pulled_entry->magic != 0) { | 1354 | if (pulled_entry->magic != 0) { |
1436 | dev_dbg(xpc_chan, "partition %d's XPC vars_part for " | 1355 | dev_dbg(xpc_chan, "partition %d's XPC vars_part for " |
@@ -1443,7 +1362,7 @@ xpc_pull_remote_vars_part_sn2(struct xpc_partition *part) | |||
1443 | return xpRetry; | 1362 | return xpRetry; |
1444 | } | 1363 | } |
1445 | 1364 | ||
1446 | if (xpc_vars_part_sn2[partid].magic == XPC_VP_MAGIC1) { | 1365 | if (xpc_vars_part_sn2[partid].magic == XPC_VP_MAGIC1_SN2) { |
1447 | 1366 | ||
1448 | /* validate the variables */ | 1367 | /* validate the variables */ |
1449 | 1368 | ||
@@ -1473,10 +1392,10 @@ xpc_pull_remote_vars_part_sn2(struct xpc_partition *part) | |||
1473 | 1392 | ||
1474 | /* let the other side know that we've pulled their variables */ | 1393 | /* let the other side know that we've pulled their variables */ |
1475 | 1394 | ||
1476 | xpc_vars_part_sn2[partid].magic = XPC_VP_MAGIC2; | 1395 | xpc_vars_part_sn2[partid].magic = XPC_VP_MAGIC2_SN2; |
1477 | } | 1396 | } |
1478 | 1397 | ||
1479 | if (pulled_entry->magic == XPC_VP_MAGIC1) | 1398 | if (pulled_entry->magic == XPC_VP_MAGIC1_SN2) |
1480 | return xpRetry; | 1399 | return xpRetry; |
1481 | 1400 | ||
1482 | return xpSuccess; | 1401 | return xpSuccess; |
@@ -1605,6 +1524,7 @@ xpc_get_chctl_all_flags_sn2(struct xpc_partition *part) | |||
1605 | static enum xp_retval | 1524 | static enum xp_retval |
1606 | xpc_allocate_local_msgqueue_sn2(struct xpc_channel *ch) | 1525 | xpc_allocate_local_msgqueue_sn2(struct xpc_channel *ch) |
1607 | { | 1526 | { |
1527 | struct xpc_channel_sn2 *ch_sn2 = &ch->sn.sn2; | ||
1608 | unsigned long irq_flags; | 1528 | unsigned long irq_flags; |
1609 | int nentries; | 1529 | int nentries; |
1610 | size_t nbytes; | 1530 | size_t nbytes; |
@@ -1612,17 +1532,17 @@ xpc_allocate_local_msgqueue_sn2(struct xpc_channel *ch) | |||
1612 | for (nentries = ch->local_nentries; nentries > 0; nentries--) { | 1532 | for (nentries = ch->local_nentries; nentries > 0; nentries--) { |
1613 | 1533 | ||
1614 | nbytes = nentries * ch->msg_size; | 1534 | nbytes = nentries * ch->msg_size; |
1615 | ch->local_msgqueue = | 1535 | ch_sn2->local_msgqueue = |
1616 | xpc_kzalloc_cacheline_aligned_sn2(nbytes, GFP_KERNEL, | 1536 | xpc_kzalloc_cacheline_aligned(nbytes, GFP_KERNEL, |
1617 | &ch->local_msgqueue_base); | 1537 | &ch_sn2->local_msgqueue_base); |
1618 | if (ch->local_msgqueue == NULL) | 1538 | if (ch_sn2->local_msgqueue == NULL) |
1619 | continue; | 1539 | continue; |
1620 | 1540 | ||
1621 | nbytes = nentries * sizeof(struct xpc_notify); | 1541 | nbytes = nentries * sizeof(struct xpc_notify); |
1622 | ch->notify_queue = kzalloc(nbytes, GFP_KERNEL); | 1542 | ch_sn2->notify_queue = kzalloc(nbytes, GFP_KERNEL); |
1623 | if (ch->notify_queue == NULL) { | 1543 | if (ch_sn2->notify_queue == NULL) { |
1624 | kfree(ch->local_msgqueue_base); | 1544 | kfree(ch_sn2->local_msgqueue_base); |
1625 | ch->local_msgqueue = NULL; | 1545 | ch_sn2->local_msgqueue = NULL; |
1626 | continue; | 1546 | continue; |
1627 | } | 1547 | } |
1628 | 1548 | ||
@@ -1649,6 +1569,7 @@ xpc_allocate_local_msgqueue_sn2(struct xpc_channel *ch) | |||
1649 | static enum xp_retval | 1569 | static enum xp_retval |
1650 | xpc_allocate_remote_msgqueue_sn2(struct xpc_channel *ch) | 1570 | xpc_allocate_remote_msgqueue_sn2(struct xpc_channel *ch) |
1651 | { | 1571 | { |
1572 | struct xpc_channel_sn2 *ch_sn2 = &ch->sn.sn2; | ||
1652 | unsigned long irq_flags; | 1573 | unsigned long irq_flags; |
1653 | int nentries; | 1574 | int nentries; |
1654 | size_t nbytes; | 1575 | size_t nbytes; |
@@ -1658,10 +1579,10 @@ xpc_allocate_remote_msgqueue_sn2(struct xpc_channel *ch) | |||
1658 | for (nentries = ch->remote_nentries; nentries > 0; nentries--) { | 1579 | for (nentries = ch->remote_nentries; nentries > 0; nentries--) { |
1659 | 1580 | ||
1660 | nbytes = nentries * ch->msg_size; | 1581 | nbytes = nentries * ch->msg_size; |
1661 | ch->remote_msgqueue = | 1582 | ch_sn2->remote_msgqueue = |
1662 | xpc_kzalloc_cacheline_aligned_sn2(nbytes, GFP_KERNEL, | 1583 | xpc_kzalloc_cacheline_aligned(nbytes, GFP_KERNEL, &ch_sn2-> |
1663 | &ch->remote_msgqueue_base); | 1584 | remote_msgqueue_base); |
1664 | if (ch->remote_msgqueue == NULL) | 1585 | if (ch_sn2->remote_msgqueue == NULL) |
1665 | continue; | 1586 | continue; |
1666 | 1587 | ||
1667 | spin_lock_irqsave(&ch->lock, irq_flags); | 1588 | spin_lock_irqsave(&ch->lock, irq_flags); |
@@ -1687,8 +1608,9 @@ xpc_allocate_remote_msgqueue_sn2(struct xpc_channel *ch) | |||
1687 | * Note: Assumes all of the channel sizes are filled in. | 1608 | * Note: Assumes all of the channel sizes are filled in. |
1688 | */ | 1609 | */ |
1689 | static enum xp_retval | 1610 | static enum xp_retval |
1690 | xpc_allocate_msgqueues_sn2(struct xpc_channel *ch) | 1611 | xpc_setup_msg_structures_sn2(struct xpc_channel *ch) |
1691 | { | 1612 | { |
1613 | struct xpc_channel_sn2 *ch_sn2 = &ch->sn.sn2; | ||
1692 | enum xp_retval ret; | 1614 | enum xp_retval ret; |
1693 | 1615 | ||
1694 | DBUG_ON(ch->flags & XPC_C_SETUP); | 1616 | DBUG_ON(ch->flags & XPC_C_SETUP); |
@@ -1698,10 +1620,10 @@ xpc_allocate_msgqueues_sn2(struct xpc_channel *ch) | |||
1698 | 1620 | ||
1699 | ret = xpc_allocate_remote_msgqueue_sn2(ch); | 1621 | ret = xpc_allocate_remote_msgqueue_sn2(ch); |
1700 | if (ret != xpSuccess) { | 1622 | if (ret != xpSuccess) { |
1701 | kfree(ch->local_msgqueue_base); | 1623 | kfree(ch_sn2->local_msgqueue_base); |
1702 | ch->local_msgqueue = NULL; | 1624 | ch_sn2->local_msgqueue = NULL; |
1703 | kfree(ch->notify_queue); | 1625 | kfree(ch_sn2->notify_queue); |
1704 | ch->notify_queue = NULL; | 1626 | ch_sn2->notify_queue = NULL; |
1705 | } | 1627 | } |
1706 | } | 1628 | } |
1707 | return ret; | 1629 | return ret; |
@@ -1715,21 +1637,13 @@ xpc_allocate_msgqueues_sn2(struct xpc_channel *ch) | |||
1715 | * they're cleared when XPC_C_DISCONNECTED is cleared. | 1637 | * they're cleared when XPC_C_DISCONNECTED is cleared. |
1716 | */ | 1638 | */ |
1717 | static void | 1639 | static void |
1718 | xpc_free_msgqueues_sn2(struct xpc_channel *ch) | 1640 | xpc_teardown_msg_structures_sn2(struct xpc_channel *ch) |
1719 | { | 1641 | { |
1720 | struct xpc_channel_sn2 *ch_sn2 = &ch->sn.sn2; | 1642 | struct xpc_channel_sn2 *ch_sn2 = &ch->sn.sn2; |
1721 | 1643 | ||
1722 | DBUG_ON(!spin_is_locked(&ch->lock)); | 1644 | DBUG_ON(!spin_is_locked(&ch->lock)); |
1723 | DBUG_ON(atomic_read(&ch->n_to_notify) != 0); | ||
1724 | 1645 | ||
1725 | ch->remote_msgqueue_pa = 0; | 1646 | ch_sn2->remote_msgqueue_pa = 0; |
1726 | ch->func = NULL; | ||
1727 | ch->key = NULL; | ||
1728 | ch->msg_size = 0; | ||
1729 | ch->local_nentries = 0; | ||
1730 | ch->remote_nentries = 0; | ||
1731 | ch->kthreads_assigned_limit = 0; | ||
1732 | ch->kthreads_idle_limit = 0; | ||
1733 | 1647 | ||
1734 | ch_sn2->local_GP->get = 0; | 1648 | ch_sn2->local_GP->get = 0; |
1735 | ch_sn2->local_GP->put = 0; | 1649 | ch_sn2->local_GP->put = 0; |
@@ -1745,12 +1659,12 @@ xpc_free_msgqueues_sn2(struct xpc_channel *ch) | |||
1745 | dev_dbg(xpc_chan, "ch->flags=0x%x, partid=%d, channel=%d\n", | 1659 | dev_dbg(xpc_chan, "ch->flags=0x%x, partid=%d, channel=%d\n", |
1746 | ch->flags, ch->partid, ch->number); | 1660 | ch->flags, ch->partid, ch->number); |
1747 | 1661 | ||
1748 | kfree(ch->local_msgqueue_base); | 1662 | kfree(ch_sn2->local_msgqueue_base); |
1749 | ch->local_msgqueue = NULL; | 1663 | ch_sn2->local_msgqueue = NULL; |
1750 | kfree(ch->remote_msgqueue_base); | 1664 | kfree(ch_sn2->remote_msgqueue_base); |
1751 | ch->remote_msgqueue = NULL; | 1665 | ch_sn2->remote_msgqueue = NULL; |
1752 | kfree(ch->notify_queue); | 1666 | kfree(ch_sn2->notify_queue); |
1753 | ch->notify_queue = NULL; | 1667 | ch_sn2->notify_queue = NULL; |
1754 | } | 1668 | } |
1755 | } | 1669 | } |
1756 | 1670 | ||
@@ -1766,7 +1680,7 @@ xpc_notify_senders_sn2(struct xpc_channel *ch, enum xp_retval reason, s64 put) | |||
1766 | 1680 | ||
1767 | while (++get < put && atomic_read(&ch->n_to_notify) > 0) { | 1681 | while (++get < put && atomic_read(&ch->n_to_notify) > 0) { |
1768 | 1682 | ||
1769 | notify = &ch->notify_queue[get % ch->local_nentries]; | 1683 | notify = &ch->sn.sn2.notify_queue[get % ch->local_nentries]; |
1770 | 1684 | ||
1771 | /* | 1685 | /* |
1772 | * See if the notify entry indicates it was associated with | 1686 | * See if the notify entry indicates it was associated with |
@@ -1818,7 +1732,7 @@ xpc_clear_local_msgqueue_flags_sn2(struct xpc_channel *ch) | |||
1818 | 1732 | ||
1819 | get = ch_sn2->w_remote_GP.get; | 1733 | get = ch_sn2->w_remote_GP.get; |
1820 | do { | 1734 | do { |
1821 | msg = (struct xpc_msg *)((u64)ch->local_msgqueue + | 1735 | msg = (struct xpc_msg *)((u64)ch_sn2->local_msgqueue + |
1822 | (get % ch->local_nentries) * | 1736 | (get % ch->local_nentries) * |
1823 | ch->msg_size); | 1737 | ch->msg_size); |
1824 | msg->flags = 0; | 1738 | msg->flags = 0; |
@@ -1837,7 +1751,7 @@ xpc_clear_remote_msgqueue_flags_sn2(struct xpc_channel *ch) | |||
1837 | 1751 | ||
1838 | put = ch_sn2->w_remote_GP.put; | 1752 | put = ch_sn2->w_remote_GP.put; |
1839 | do { | 1753 | do { |
1840 | msg = (struct xpc_msg *)((u64)ch->remote_msgqueue + | 1754 | msg = (struct xpc_msg *)((u64)ch_sn2->remote_msgqueue + |
1841 | (put % ch->remote_nentries) * | 1755 | (put % ch->remote_nentries) * |
1842 | ch->msg_size); | 1756 | ch->msg_size); |
1843 | msg->flags = 0; | 1757 | msg->flags = 0; |
@@ -1976,8 +1890,9 @@ xpc_pull_remote_msg_sn2(struct xpc_channel *ch, s64 get) | |||
1976 | } | 1890 | } |
1977 | 1891 | ||
1978 | msg_offset = msg_index * ch->msg_size; | 1892 | msg_offset = msg_index * ch->msg_size; |
1979 | msg = (struct xpc_msg *)((u64)ch->remote_msgqueue + msg_offset); | 1893 | msg = (struct xpc_msg *)((u64)ch_sn2->remote_msgqueue + |
1980 | remote_msg_pa = ch->remote_msgqueue_pa + msg_offset; | 1894 | msg_offset); |
1895 | remote_msg_pa = ch_sn2->remote_msgqueue_pa + msg_offset; | ||
1981 | 1896 | ||
1982 | ret = xpc_pull_remote_cachelines_sn2(part, msg, remote_msg_pa, | 1897 | ret = xpc_pull_remote_cachelines_sn2(part, msg, remote_msg_pa, |
1983 | nmsgs * ch->msg_size); | 1898 | nmsgs * ch->msg_size); |
@@ -2001,7 +1916,7 @@ xpc_pull_remote_msg_sn2(struct xpc_channel *ch, s64 get) | |||
2001 | 1916 | ||
2002 | /* return the message we were looking for */ | 1917 | /* return the message we were looking for */ |
2003 | msg_offset = (get % ch->remote_nentries) * ch->msg_size; | 1918 | msg_offset = (get % ch->remote_nentries) * ch->msg_size; |
2004 | msg = (struct xpc_msg *)((u64)ch->remote_msgqueue + msg_offset); | 1919 | msg = (struct xpc_msg *)((u64)ch_sn2->remote_msgqueue + msg_offset); |
2005 | 1920 | ||
2006 | return msg; | 1921 | return msg; |
2007 | } | 1922 | } |
@@ -2080,7 +1995,7 @@ xpc_send_msgs_sn2(struct xpc_channel *ch, s64 initial_put) | |||
2080 | if (put == ch_sn2->w_local_GP.put) | 1995 | if (put == ch_sn2->w_local_GP.put) |
2081 | break; | 1996 | break; |
2082 | 1997 | ||
2083 | msg = (struct xpc_msg *)((u64)ch->local_msgqueue + | 1998 | msg = (struct xpc_msg *)((u64)ch_sn2->local_msgqueue + |
2084 | (put % ch->local_nentries) * | 1999 | (put % ch->local_nentries) * |
2085 | ch->msg_size); | 2000 | ch->msg_size); |
2086 | 2001 | ||
@@ -2182,7 +2097,7 @@ xpc_allocate_msg_sn2(struct xpc_channel *ch, u32 flags, | |||
2182 | } | 2097 | } |
2183 | 2098 | ||
2184 | /* get the message's address and initialize it */ | 2099 | /* get the message's address and initialize it */ |
2185 | msg = (struct xpc_msg *)((u64)ch->local_msgqueue + | 2100 | msg = (struct xpc_msg *)((u64)ch_sn2->local_msgqueue + |
2186 | (put % ch->local_nentries) * ch->msg_size); | 2101 | (put % ch->local_nentries) * ch->msg_size); |
2187 | 2102 | ||
2188 | DBUG_ON(msg->flags != 0); | 2103 | DBUG_ON(msg->flags != 0); |
@@ -2207,6 +2122,7 @@ xpc_send_msg_sn2(struct xpc_channel *ch, u32 flags, void *payload, | |||
2207 | void *key) | 2122 | void *key) |
2208 | { | 2123 | { |
2209 | enum xp_retval ret = xpSuccess; | 2124 | enum xp_retval ret = xpSuccess; |
2125 | struct xpc_channel_sn2 *ch_sn2 = &ch->sn.sn2; | ||
2210 | struct xpc_msg *msg = msg; | 2126 | struct xpc_msg *msg = msg; |
2211 | struct xpc_notify *notify = notify; | 2127 | struct xpc_notify *notify = notify; |
2212 | s64 msg_number; | 2128 | s64 msg_number; |
@@ -2243,7 +2159,7 @@ xpc_send_msg_sn2(struct xpc_channel *ch, u32 flags, void *payload, | |||
2243 | 2159 | ||
2244 | atomic_inc(&ch->n_to_notify); | 2160 | atomic_inc(&ch->n_to_notify); |
2245 | 2161 | ||
2246 | notify = &ch->notify_queue[msg_number % ch->local_nentries]; | 2162 | notify = &ch_sn2->notify_queue[msg_number % ch->local_nentries]; |
2247 | notify->func = func; | 2163 | notify->func = func; |
2248 | notify->key = key; | 2164 | notify->key = key; |
2249 | notify->type = notify_type; | 2165 | notify->type = notify_type; |
@@ -2279,7 +2195,7 @@ xpc_send_msg_sn2(struct xpc_channel *ch, u32 flags, void *payload, | |||
2279 | 2195 | ||
2280 | /* see if the message is next in line to be sent, if so send it */ | 2196 | /* see if the message is next in line to be sent, if so send it */ |
2281 | 2197 | ||
2282 | put = ch->sn.sn2.local_GP->put; | 2198 | put = ch_sn2->local_GP->put; |
2283 | if (put == msg_number) | 2199 | if (put == msg_number) |
2284 | xpc_send_msgs_sn2(ch, put); | 2200 | xpc_send_msgs_sn2(ch, put); |
2285 | 2201 | ||
@@ -2307,7 +2223,7 @@ xpc_acknowledge_msgs_sn2(struct xpc_channel *ch, s64 initial_get, u8 msg_flags) | |||
2307 | if (get == ch_sn2->w_local_GP.get) | 2223 | if (get == ch_sn2->w_local_GP.get) |
2308 | break; | 2224 | break; |
2309 | 2225 | ||
2310 | msg = (struct xpc_msg *)((u64)ch->remote_msgqueue + | 2226 | msg = (struct xpc_msg *)((u64)ch_sn2->remote_msgqueue + |
2311 | (get % ch->remote_nentries) * | 2227 | (get % ch->remote_nentries) * |
2312 | ch->msg_size); | 2228 | ch->msg_size); |
2313 | 2229 | ||
@@ -2385,8 +2301,9 @@ xpc_init_sn2(void) | |||
2385 | int ret; | 2301 | int ret; |
2386 | size_t buf_size; | 2302 | size_t buf_size; |
2387 | 2303 | ||
2304 | xpc_setup_partitions_sn = xpc_setup_partitions_sn_sn2; | ||
2388 | xpc_get_partition_rsvd_page_pa = xpc_get_partition_rsvd_page_pa_sn2; | 2305 | xpc_get_partition_rsvd_page_pa = xpc_get_partition_rsvd_page_pa_sn2; |
2389 | xpc_rsvd_page_init = xpc_rsvd_page_init_sn2; | 2306 | xpc_setup_rsvd_page_sn = xpc_setup_rsvd_page_sn_sn2; |
2390 | xpc_increment_heartbeat = xpc_increment_heartbeat_sn2; | 2307 | xpc_increment_heartbeat = xpc_increment_heartbeat_sn2; |
2391 | xpc_offline_heartbeat = xpc_offline_heartbeat_sn2; | 2308 | xpc_offline_heartbeat = xpc_offline_heartbeat_sn2; |
2392 | xpc_online_heartbeat = xpc_online_heartbeat_sn2; | 2309 | xpc_online_heartbeat = xpc_online_heartbeat_sn2; |
@@ -2403,29 +2320,33 @@ xpc_init_sn2(void) | |||
2403 | xpc_cancel_partition_deactivation_request_sn2; | 2320 | xpc_cancel_partition_deactivation_request_sn2; |
2404 | 2321 | ||
2405 | xpc_process_activate_IRQ_rcvd = xpc_process_activate_IRQ_rcvd_sn2; | 2322 | xpc_process_activate_IRQ_rcvd = xpc_process_activate_IRQ_rcvd_sn2; |
2406 | xpc_setup_infrastructure = xpc_setup_infrastructure_sn2; | 2323 | xpc_setup_ch_structures_sn = xpc_setup_ch_structures_sn_sn2; |
2407 | xpc_teardown_infrastructure = xpc_teardown_infrastructure_sn2; | 2324 | xpc_teardown_ch_structures_sn = xpc_teardown_ch_structures_sn_sn2; |
2408 | xpc_make_first_contact = xpc_make_first_contact_sn2; | 2325 | xpc_make_first_contact = xpc_make_first_contact_sn2; |
2326 | |||
2409 | xpc_get_chctl_all_flags = xpc_get_chctl_all_flags_sn2; | 2327 | xpc_get_chctl_all_flags = xpc_get_chctl_all_flags_sn2; |
2410 | xpc_allocate_msgqueues = xpc_allocate_msgqueues_sn2; | 2328 | xpc_send_chctl_closerequest = xpc_send_chctl_closerequest_sn2; |
2411 | xpc_free_msgqueues = xpc_free_msgqueues_sn2; | 2329 | xpc_send_chctl_closereply = xpc_send_chctl_closereply_sn2; |
2330 | xpc_send_chctl_openrequest = xpc_send_chctl_openrequest_sn2; | ||
2331 | xpc_send_chctl_openreply = xpc_send_chctl_openreply_sn2; | ||
2332 | |||
2333 | xpc_save_remote_msgqueue_pa = xpc_save_remote_msgqueue_pa_sn2; | ||
2334 | |||
2335 | xpc_setup_msg_structures = xpc_setup_msg_structures_sn2; | ||
2336 | xpc_teardown_msg_structures = xpc_teardown_msg_structures_sn2; | ||
2337 | |||
2412 | xpc_notify_senders_of_disconnect = xpc_notify_senders_of_disconnect_sn2; | 2338 | xpc_notify_senders_of_disconnect = xpc_notify_senders_of_disconnect_sn2; |
2413 | xpc_process_msg_chctl_flags = xpc_process_msg_chctl_flags_sn2; | 2339 | xpc_process_msg_chctl_flags = xpc_process_msg_chctl_flags_sn2; |
2414 | xpc_n_of_deliverable_msgs = xpc_n_of_deliverable_msgs_sn2; | 2340 | xpc_n_of_deliverable_msgs = xpc_n_of_deliverable_msgs_sn2; |
2415 | xpc_get_deliverable_msg = xpc_get_deliverable_msg_sn2; | 2341 | xpc_get_deliverable_msg = xpc_get_deliverable_msg_sn2; |
2416 | 2342 | ||
2417 | xpc_indicate_partition_engaged = xpc_indicate_partition_engaged_sn2; | 2343 | xpc_indicate_partition_engaged = xpc_indicate_partition_engaged_sn2; |
2418 | xpc_partition_engaged = xpc_partition_engaged_sn2; | ||
2419 | xpc_any_partition_engaged = xpc_any_partition_engaged_sn2; | ||
2420 | xpc_indicate_partition_disengaged = | 2344 | xpc_indicate_partition_disengaged = |
2421 | xpc_indicate_partition_disengaged_sn2; | 2345 | xpc_indicate_partition_disengaged_sn2; |
2346 | xpc_partition_engaged = xpc_partition_engaged_sn2; | ||
2347 | xpc_any_partition_engaged = xpc_any_partition_engaged_sn2; | ||
2422 | xpc_assume_partition_disengaged = xpc_assume_partition_disengaged_sn2; | 2348 | xpc_assume_partition_disengaged = xpc_assume_partition_disengaged_sn2; |
2423 | 2349 | ||
2424 | xpc_send_chctl_closerequest = xpc_send_chctl_closerequest_sn2; | ||
2425 | xpc_send_chctl_closereply = xpc_send_chctl_closereply_sn2; | ||
2426 | xpc_send_chctl_openrequest = xpc_send_chctl_openrequest_sn2; | ||
2427 | xpc_send_chctl_openreply = xpc_send_chctl_openreply_sn2; | ||
2428 | |||
2429 | xpc_send_msg = xpc_send_msg_sn2; | 2350 | xpc_send_msg = xpc_send_msg_sn2; |
2430 | xpc_received_msg = xpc_received_msg_sn2; | 2351 | xpc_received_msg = xpc_received_msg_sn2; |
2431 | 2352 | ||
diff --git a/drivers/misc/sgi-xp/xpc_uv.c b/drivers/misc/sgi-xp/xpc_uv.c index c2d4ddd6e955..689cb5c68ccf 100644 --- a/drivers/misc/sgi-xp/xpc_uv.c +++ b/drivers/misc/sgi-xp/xpc_uv.c | |||
@@ -14,41 +14,528 @@ | |||
14 | */ | 14 | */ |
15 | 15 | ||
16 | #include <linux/kernel.h> | 16 | #include <linux/kernel.h> |
17 | #include <linux/mm.h> | ||
18 | #include <linux/interrupt.h> | ||
19 | #include <linux/delay.h> | ||
20 | #include <linux/device.h> | ||
17 | #include <asm/uv/uv_hub.h> | 21 | #include <asm/uv/uv_hub.h> |
22 | #include "../sgi-gru/gru.h" | ||
18 | #include "../sgi-gru/grukservices.h" | 23 | #include "../sgi-gru/grukservices.h" |
19 | #include "xpc.h" | 24 | #include "xpc.h" |
20 | 25 | ||
26 | static atomic64_t xpc_heartbeat_uv; | ||
21 | static DECLARE_BITMAP(xpc_heartbeating_to_mask_uv, XP_MAX_NPARTITIONS_UV); | 27 | static DECLARE_BITMAP(xpc_heartbeating_to_mask_uv, XP_MAX_NPARTITIONS_UV); |
22 | 28 | ||
23 | static void *xpc_activate_mq; | 29 | #define XPC_ACTIVATE_MSG_SIZE_UV (1 * GRU_CACHE_LINE_BYTES) |
30 | #define XPC_NOTIFY_MSG_SIZE_UV (2 * GRU_CACHE_LINE_BYTES) | ||
31 | |||
32 | #define XPC_ACTIVATE_MQ_SIZE_UV (4 * XP_MAX_NPARTITIONS_UV * \ | ||
33 | XPC_ACTIVATE_MSG_SIZE_UV) | ||
34 | #define XPC_NOTIFY_MQ_SIZE_UV (4 * XP_MAX_NPARTITIONS_UV * \ | ||
35 | XPC_NOTIFY_MSG_SIZE_UV) | ||
36 | |||
37 | static void *xpc_activate_mq_uv; | ||
38 | static void *xpc_notify_mq_uv; | ||
39 | |||
40 | static int | ||
41 | xpc_setup_partitions_sn_uv(void) | ||
42 | { | ||
43 | short partid; | ||
44 | struct xpc_partition_uv *part_uv; | ||
45 | |||
46 | for (partid = 0; partid < XP_MAX_NPARTITIONS_UV; partid++) { | ||
47 | part_uv = &xpc_partitions[partid].sn.uv; | ||
48 | |||
49 | spin_lock_init(&part_uv->flags_lock); | ||
50 | part_uv->remote_act_state = XPC_P_AS_INACTIVE; | ||
51 | } | ||
52 | return 0; | ||
53 | } | ||
54 | |||
55 | static void * | ||
56 | xpc_create_gru_mq_uv(unsigned int mq_size, int cpuid, unsigned int irq, | ||
57 | irq_handler_t irq_handler) | ||
58 | { | ||
59 | int ret; | ||
60 | int nid; | ||
61 | int mq_order; | ||
62 | struct page *page; | ||
63 | void *mq; | ||
64 | |||
65 | nid = cpu_to_node(cpuid); | ||
66 | mq_order = get_order(mq_size); | ||
67 | page = alloc_pages_node(nid, GFP_KERNEL | __GFP_ZERO | GFP_THISNODE, | ||
68 | mq_order); | ||
69 | if (page == NULL) | ||
70 | return NULL; | ||
71 | |||
72 | mq = page_address(page); | ||
73 | ret = gru_create_message_queue(mq, mq_size); | ||
74 | if (ret != 0) { | ||
75 | dev_err(xpc_part, "gru_create_message_queue() returned " | ||
76 | "error=%d\n", ret); | ||
77 | free_pages((unsigned long)mq, mq_order); | ||
78 | return NULL; | ||
79 | } | ||
80 | |||
81 | /* !!! Need to do some other things to set up IRQ */ | ||
82 | |||
83 | ret = request_irq(irq, irq_handler, 0, "xpc", NULL); | ||
84 | if (ret != 0) { | ||
85 | dev_err(xpc_part, "request_irq(irq=%d) returned error=%d\n", | ||
86 | irq, ret); | ||
87 | free_pages((unsigned long)mq, mq_order); | ||
88 | return NULL; | ||
89 | } | ||
90 | |||
91 | /* !!! enable generation of irq when GRU mq op occurs to this mq */ | ||
92 | |||
93 | /* ??? allow other partitions to access GRU mq? */ | ||
94 | |||
95 | return mq; | ||
96 | } | ||
24 | 97 | ||
25 | static void | 98 | static void |
26 | xpc_send_local_activate_IRQ_uv(struct xpc_partition *part) | 99 | xpc_destroy_gru_mq_uv(void *mq, unsigned int mq_size, unsigned int irq) |
100 | { | ||
101 | /* ??? disallow other partitions to access GRU mq? */ | ||
102 | |||
103 | /* !!! disable generation of irq when GRU mq op occurs to this mq */ | ||
104 | |||
105 | free_irq(irq, NULL); | ||
106 | |||
107 | free_pages((unsigned long)mq, get_order(mq_size)); | ||
108 | } | ||
109 | |||
110 | static enum xp_retval | ||
111 | xpc_send_gru_msg(unsigned long mq_gpa, void *msg, size_t msg_size) | ||
27 | { | 112 | { |
113 | enum xp_retval xp_ret; | ||
114 | int ret; | ||
115 | |||
116 | while (1) { | ||
117 | ret = gru_send_message_gpa(mq_gpa, msg, msg_size); | ||
118 | if (ret == MQE_OK) { | ||
119 | xp_ret = xpSuccess; | ||
120 | break; | ||
121 | } | ||
122 | |||
123 | if (ret == MQE_QUEUE_FULL) { | ||
124 | dev_dbg(xpc_chan, "gru_send_message_gpa() returned " | ||
125 | "error=MQE_QUEUE_FULL\n"); | ||
126 | /* !!! handle QLimit reached; delay & try again */ | ||
127 | /* ??? Do we add a limit to the number of retries? */ | ||
128 | (void)msleep_interruptible(10); | ||
129 | } else if (ret == MQE_CONGESTION) { | ||
130 | dev_dbg(xpc_chan, "gru_send_message_gpa() returned " | ||
131 | "error=MQE_CONGESTION\n"); | ||
132 | /* !!! handle LB Overflow; simply try again */ | ||
133 | /* ??? Do we add a limit to the number of retries? */ | ||
134 | } else { | ||
135 | /* !!! Currently this is MQE_UNEXPECTED_CB_ERR */ | ||
136 | dev_err(xpc_chan, "gru_send_message_gpa() returned " | ||
137 | "error=%d\n", ret); | ||
138 | xp_ret = xpGruSendMqError; | ||
139 | break; | ||
140 | } | ||
141 | } | ||
142 | return xp_ret; | ||
143 | } | ||
144 | |||
145 | static void | ||
146 | xpc_process_activate_IRQ_rcvd_uv(void) | ||
147 | { | ||
148 | unsigned long irq_flags; | ||
149 | short partid; | ||
150 | struct xpc_partition *part; | ||
151 | u8 act_state_req; | ||
152 | |||
153 | DBUG_ON(xpc_activate_IRQ_rcvd == 0); | ||
154 | |||
155 | spin_lock_irqsave(&xpc_activate_IRQ_rcvd_lock, irq_flags); | ||
156 | for (partid = 0; partid < XP_MAX_NPARTITIONS_UV; partid++) { | ||
157 | part = &xpc_partitions[partid]; | ||
158 | |||
159 | if (part->sn.uv.act_state_req == 0) | ||
160 | continue; | ||
161 | |||
162 | xpc_activate_IRQ_rcvd--; | ||
163 | BUG_ON(xpc_activate_IRQ_rcvd < 0); | ||
164 | |||
165 | act_state_req = part->sn.uv.act_state_req; | ||
166 | part->sn.uv.act_state_req = 0; | ||
167 | spin_unlock_irqrestore(&xpc_activate_IRQ_rcvd_lock, irq_flags); | ||
168 | |||
169 | if (act_state_req == XPC_P_ASR_ACTIVATE_UV) { | ||
170 | if (part->act_state == XPC_P_AS_INACTIVE) | ||
171 | xpc_activate_partition(part); | ||
172 | else if (part->act_state == XPC_P_AS_DEACTIVATING) | ||
173 | XPC_DEACTIVATE_PARTITION(part, xpReactivating); | ||
174 | |||
175 | } else if (act_state_req == XPC_P_ASR_REACTIVATE_UV) { | ||
176 | if (part->act_state == XPC_P_AS_INACTIVE) | ||
177 | xpc_activate_partition(part); | ||
178 | else | ||
179 | XPC_DEACTIVATE_PARTITION(part, xpReactivating); | ||
180 | |||
181 | } else if (act_state_req == XPC_P_ASR_DEACTIVATE_UV) { | ||
182 | XPC_DEACTIVATE_PARTITION(part, part->sn.uv.reason); | ||
183 | |||
184 | } else { | ||
185 | BUG(); | ||
186 | } | ||
187 | |||
188 | spin_lock_irqsave(&xpc_activate_IRQ_rcvd_lock, irq_flags); | ||
189 | if (xpc_activate_IRQ_rcvd == 0) | ||
190 | break; | ||
191 | } | ||
192 | spin_unlock_irqrestore(&xpc_activate_IRQ_rcvd_lock, irq_flags); | ||
193 | |||
194 | } | ||
195 | |||
196 | static irqreturn_t | ||
197 | xpc_handle_activate_IRQ_uv(int irq, void *dev_id) | ||
198 | { | ||
199 | unsigned long irq_flags; | ||
200 | struct xpc_activate_mq_msghdr_uv *msg_hdr; | ||
201 | short partid; | ||
202 | struct xpc_partition *part; | ||
203 | struct xpc_partition_uv *part_uv; | ||
204 | struct xpc_openclose_args *args; | ||
205 | int wakeup_hb_checker = 0; | ||
206 | |||
207 | while ((msg_hdr = gru_get_next_message(xpc_activate_mq_uv)) != NULL) { | ||
208 | |||
209 | partid = msg_hdr->partid; | ||
210 | if (partid < 0 || partid >= XP_MAX_NPARTITIONS_UV) { | ||
211 | dev_err(xpc_part, "xpc_handle_activate_IRQ_uv() invalid" | ||
212 | "partid=0x%x passed in message\n", partid); | ||
213 | gru_free_message(xpc_activate_mq_uv, msg_hdr); | ||
214 | continue; | ||
215 | } | ||
216 | part = &xpc_partitions[partid]; | ||
217 | part_uv = &part->sn.uv; | ||
218 | |||
219 | part_uv->remote_act_state = msg_hdr->act_state; | ||
220 | |||
221 | switch (msg_hdr->type) { | ||
222 | case XPC_ACTIVATE_MQ_MSG_SYNC_ACT_STATE_UV: | ||
223 | /* syncing of remote_act_state was just done above */ | ||
224 | break; | ||
225 | |||
226 | case XPC_ACTIVATE_MQ_MSG_INC_HEARTBEAT_UV: { | ||
227 | struct xpc_activate_mq_msg_heartbeat_req_uv *msg; | ||
228 | |||
229 | msg = (struct xpc_activate_mq_msg_heartbeat_req_uv *) | ||
230 | msg_hdr; | ||
231 | part_uv->heartbeat = msg->heartbeat; | ||
232 | break; | ||
233 | } | ||
234 | case XPC_ACTIVATE_MQ_MSG_OFFLINE_HEARTBEAT_UV: { | ||
235 | struct xpc_activate_mq_msg_heartbeat_req_uv *msg; | ||
236 | |||
237 | msg = (struct xpc_activate_mq_msg_heartbeat_req_uv *) | ||
238 | msg_hdr; | ||
239 | part_uv->heartbeat = msg->heartbeat; | ||
240 | spin_lock_irqsave(&part_uv->flags_lock, irq_flags); | ||
241 | part_uv->flags |= XPC_P_HEARTBEAT_OFFLINE_UV; | ||
242 | spin_unlock_irqrestore(&part_uv->flags_lock, irq_flags); | ||
243 | break; | ||
244 | } | ||
245 | case XPC_ACTIVATE_MQ_MSG_ONLINE_HEARTBEAT_UV: { | ||
246 | struct xpc_activate_mq_msg_heartbeat_req_uv *msg; | ||
247 | |||
248 | msg = (struct xpc_activate_mq_msg_heartbeat_req_uv *) | ||
249 | msg_hdr; | ||
250 | part_uv->heartbeat = msg->heartbeat; | ||
251 | spin_lock_irqsave(&part_uv->flags_lock, irq_flags); | ||
252 | part_uv->flags &= ~XPC_P_HEARTBEAT_OFFLINE_UV; | ||
253 | spin_unlock_irqrestore(&part_uv->flags_lock, irq_flags); | ||
254 | break; | ||
255 | } | ||
256 | case XPC_ACTIVATE_MQ_MSG_ACTIVATE_REQ_UV: { | ||
257 | struct xpc_activate_mq_msg_activate_req_uv *msg; | ||
258 | |||
259 | /* | ||
260 | * ??? Do we deal here with ts_jiffies being different | ||
261 | * ??? if act_state != XPC_P_AS_INACTIVE instead of | ||
262 | * ??? below? | ||
263 | */ | ||
264 | msg = (struct xpc_activate_mq_msg_activate_req_uv *) | ||
265 | msg_hdr; | ||
266 | spin_lock_irqsave(&xpc_activate_IRQ_rcvd_lock, | ||
267 | irq_flags); | ||
268 | if (part_uv->act_state_req == 0) | ||
269 | xpc_activate_IRQ_rcvd++; | ||
270 | part_uv->act_state_req = XPC_P_ASR_ACTIVATE_UV; | ||
271 | part->remote_rp_pa = msg->rp_gpa; /* !!! _pa is _gpa */ | ||
272 | part->remote_rp_ts_jiffies = msg_hdr->rp_ts_jiffies; | ||
273 | part_uv->remote_activate_mq_gpa = msg->activate_mq_gpa; | ||
274 | spin_unlock_irqrestore(&xpc_activate_IRQ_rcvd_lock, | ||
275 | irq_flags); | ||
276 | wakeup_hb_checker++; | ||
277 | break; | ||
278 | } | ||
279 | case XPC_ACTIVATE_MQ_MSG_DEACTIVATE_REQ_UV: { | ||
280 | struct xpc_activate_mq_msg_deactivate_req_uv *msg; | ||
281 | |||
282 | msg = (struct xpc_activate_mq_msg_deactivate_req_uv *) | ||
283 | msg_hdr; | ||
284 | spin_lock_irqsave(&xpc_activate_IRQ_rcvd_lock, | ||
285 | irq_flags); | ||
286 | if (part_uv->act_state_req == 0) | ||
287 | xpc_activate_IRQ_rcvd++; | ||
288 | part_uv->act_state_req = XPC_P_ASR_DEACTIVATE_UV; | ||
289 | part_uv->reason = msg->reason; | ||
290 | spin_unlock_irqrestore(&xpc_activate_IRQ_rcvd_lock, | ||
291 | irq_flags); | ||
292 | wakeup_hb_checker++; | ||
293 | break; | ||
294 | } | ||
295 | case XPC_ACTIVATE_MQ_MSG_CHCTL_CLOSEREQUEST_UV: { | ||
296 | struct xpc_activate_mq_msg_chctl_closerequest_uv *msg; | ||
297 | |||
298 | msg = (struct xpc_activate_mq_msg_chctl_closerequest_uv | ||
299 | *)msg_hdr; | ||
300 | args = &part->remote_openclose_args[msg->ch_number]; | ||
301 | args->reason = msg->reason; | ||
302 | |||
303 | spin_lock_irqsave(&part->chctl_lock, irq_flags); | ||
304 | part->chctl.flags[msg->ch_number] |= | ||
305 | XPC_CHCTL_CLOSEREQUEST; | ||
306 | spin_unlock_irqrestore(&part->chctl_lock, irq_flags); | ||
307 | |||
308 | xpc_wakeup_channel_mgr(part); | ||
309 | break; | ||
310 | } | ||
311 | case XPC_ACTIVATE_MQ_MSG_CHCTL_CLOSEREPLY_UV: { | ||
312 | struct xpc_activate_mq_msg_chctl_closereply_uv *msg; | ||
313 | |||
314 | msg = (struct xpc_activate_mq_msg_chctl_closereply_uv *) | ||
315 | msg_hdr; | ||
316 | |||
317 | spin_lock_irqsave(&part->chctl_lock, irq_flags); | ||
318 | part->chctl.flags[msg->ch_number] |= | ||
319 | XPC_CHCTL_CLOSEREPLY; | ||
320 | spin_unlock_irqrestore(&part->chctl_lock, irq_flags); | ||
321 | |||
322 | xpc_wakeup_channel_mgr(part); | ||
323 | break; | ||
324 | } | ||
325 | case XPC_ACTIVATE_MQ_MSG_CHCTL_OPENREQUEST_UV: { | ||
326 | struct xpc_activate_mq_msg_chctl_openrequest_uv *msg; | ||
327 | |||
328 | msg = (struct xpc_activate_mq_msg_chctl_openrequest_uv | ||
329 | *)msg_hdr; | ||
330 | args = &part->remote_openclose_args[msg->ch_number]; | ||
331 | args->msg_size = msg->msg_size; | ||
332 | args->local_nentries = msg->local_nentries; | ||
333 | |||
334 | spin_lock_irqsave(&part->chctl_lock, irq_flags); | ||
335 | part->chctl.flags[msg->ch_number] |= | ||
336 | XPC_CHCTL_OPENREQUEST; | ||
337 | spin_unlock_irqrestore(&part->chctl_lock, irq_flags); | ||
338 | |||
339 | xpc_wakeup_channel_mgr(part); | ||
340 | break; | ||
341 | } | ||
342 | case XPC_ACTIVATE_MQ_MSG_CHCTL_OPENREPLY_UV: { | ||
343 | struct xpc_activate_mq_msg_chctl_openreply_uv *msg; | ||
344 | |||
345 | msg = (struct xpc_activate_mq_msg_chctl_openreply_uv *) | ||
346 | msg_hdr; | ||
347 | args = &part->remote_openclose_args[msg->ch_number]; | ||
348 | args->remote_nentries = msg->remote_nentries; | ||
349 | args->local_nentries = msg->local_nentries; | ||
350 | args->local_msgqueue_pa = msg->local_notify_mq_gpa; | ||
351 | |||
352 | spin_lock_irqsave(&part->chctl_lock, irq_flags); | ||
353 | part->chctl.flags[msg->ch_number] |= | ||
354 | XPC_CHCTL_OPENREPLY; | ||
355 | spin_unlock_irqrestore(&part->chctl_lock, irq_flags); | ||
356 | |||
357 | xpc_wakeup_channel_mgr(part); | ||
358 | break; | ||
359 | } | ||
360 | case XPC_ACTIVATE_MQ_MSG_MARK_ENGAGED_UV: | ||
361 | spin_lock_irqsave(&part_uv->flags_lock, irq_flags); | ||
362 | part_uv->flags |= XPC_P_ENGAGED_UV; | ||
363 | spin_unlock_irqrestore(&part_uv->flags_lock, irq_flags); | ||
364 | break; | ||
365 | |||
366 | case XPC_ACTIVATE_MQ_MSG_MARK_DISENGAGED_UV: | ||
367 | spin_lock_irqsave(&part_uv->flags_lock, irq_flags); | ||
368 | part_uv->flags &= ~XPC_P_ENGAGED_UV; | ||
369 | spin_unlock_irqrestore(&part_uv->flags_lock, irq_flags); | ||
370 | break; | ||
371 | |||
372 | default: | ||
373 | dev_err(xpc_part, "received unknown activate_mq msg " | ||
374 | "type=%d from partition=%d\n", msg_hdr->type, | ||
375 | partid); | ||
376 | } | ||
377 | |||
378 | if (msg_hdr->rp_ts_jiffies != part->remote_rp_ts_jiffies && | ||
379 | part->remote_rp_ts_jiffies != 0) { | ||
380 | /* | ||
381 | * ??? Does what we do here need to be sensitive to | ||
382 | * ??? act_state or remote_act_state? | ||
383 | */ | ||
384 | spin_lock_irqsave(&xpc_activate_IRQ_rcvd_lock, | ||
385 | irq_flags); | ||
386 | if (part_uv->act_state_req == 0) | ||
387 | xpc_activate_IRQ_rcvd++; | ||
388 | part_uv->act_state_req = XPC_P_ASR_REACTIVATE_UV; | ||
389 | spin_unlock_irqrestore(&xpc_activate_IRQ_rcvd_lock, | ||
390 | irq_flags); | ||
391 | wakeup_hb_checker++; | ||
392 | } | ||
393 | |||
394 | gru_free_message(xpc_activate_mq_uv, msg_hdr); | ||
395 | } | ||
396 | |||
397 | if (wakeup_hb_checker) | ||
398 | wake_up_interruptible(&xpc_activate_IRQ_wq); | ||
399 | |||
400 | return IRQ_HANDLED; | ||
401 | } | ||
402 | |||
403 | static enum xp_retval | ||
404 | xpc_send_activate_IRQ_uv(struct xpc_partition *part, void *msg, size_t msg_size, | ||
405 | int msg_type) | ||
406 | { | ||
407 | struct xpc_activate_mq_msghdr_uv *msg_hdr = msg; | ||
408 | |||
409 | DBUG_ON(msg_size > XPC_ACTIVATE_MSG_SIZE_UV); | ||
410 | |||
411 | msg_hdr->type = msg_type; | ||
412 | msg_hdr->partid = XPC_PARTID(part); | ||
413 | msg_hdr->act_state = part->act_state; | ||
414 | msg_hdr->rp_ts_jiffies = xpc_rsvd_page->ts_jiffies; | ||
415 | |||
416 | /* ??? Is holding a spin_lock (ch->lock) during this call a bad idea? */ | ||
417 | return xpc_send_gru_msg(part->sn.uv.remote_activate_mq_gpa, msg, | ||
418 | msg_size); | ||
419 | } | ||
420 | |||
421 | static void | ||
422 | xpc_send_activate_IRQ_part_uv(struct xpc_partition *part, void *msg, | ||
423 | size_t msg_size, int msg_type) | ||
424 | { | ||
425 | enum xp_retval ret; | ||
426 | |||
427 | ret = xpc_send_activate_IRQ_uv(part, msg, msg_size, msg_type); | ||
428 | if (unlikely(ret != xpSuccess)) | ||
429 | XPC_DEACTIVATE_PARTITION(part, ret); | ||
430 | } | ||
431 | |||
432 | static void | ||
433 | xpc_send_activate_IRQ_ch_uv(struct xpc_channel *ch, unsigned long *irq_flags, | ||
434 | void *msg, size_t msg_size, int msg_type) | ||
435 | { | ||
436 | struct xpc_partition *part = &xpc_partitions[ch->number]; | ||
437 | enum xp_retval ret; | ||
438 | |||
439 | ret = xpc_send_activate_IRQ_uv(part, msg, msg_size, msg_type); | ||
440 | if (unlikely(ret != xpSuccess)) { | ||
441 | if (irq_flags != NULL) | ||
442 | spin_unlock_irqrestore(&ch->lock, *irq_flags); | ||
443 | |||
444 | XPC_DEACTIVATE_PARTITION(part, ret); | ||
445 | |||
446 | if (irq_flags != NULL) | ||
447 | spin_lock_irqsave(&ch->lock, *irq_flags); | ||
448 | } | ||
449 | } | ||
450 | |||
451 | static void | ||
452 | xpc_send_local_activate_IRQ_uv(struct xpc_partition *part, int act_state_req) | ||
453 | { | ||
454 | unsigned long irq_flags; | ||
455 | struct xpc_partition_uv *part_uv = &part->sn.uv; | ||
456 | |||
28 | /* | 457 | /* |
29 | * !!! Make our side think that the remote parition sent an activate | 458 | * !!! Make our side think that the remote parition sent an activate |
30 | * !!! message our way. Also do what the activate IRQ handler would | 459 | * !!! message our way by doing what the activate IRQ handler would |
31 | * !!! do had one really been sent. | 460 | * !!! do had one really been sent. |
32 | */ | 461 | */ |
462 | |||
463 | spin_lock_irqsave(&xpc_activate_IRQ_rcvd_lock, irq_flags); | ||
464 | if (part_uv->act_state_req == 0) | ||
465 | xpc_activate_IRQ_rcvd++; | ||
466 | part_uv->act_state_req = act_state_req; | ||
467 | spin_unlock_irqrestore(&xpc_activate_IRQ_rcvd_lock, irq_flags); | ||
468 | |||
469 | wake_up_interruptible(&xpc_activate_IRQ_wq); | ||
33 | } | 470 | } |
34 | 471 | ||
35 | static enum xp_retval | 472 | static enum xp_retval |
36 | xpc_rsvd_page_init_uv(struct xpc_rsvd_page *rp) | 473 | xpc_get_partition_rsvd_page_pa_uv(void *buf, u64 *cookie, unsigned long *rp_pa, |
474 | size_t *len) | ||
37 | { | 475 | { |
38 | /* !!! need to have established xpc_activate_mq earlier */ | 476 | /* !!! call the UV version of sn_partition_reserved_page_pa() */ |
39 | rp->sn.activate_mq_gpa = uv_gpa(xpc_activate_mq); | 477 | return xpUnsupported; |
40 | return xpSuccess; | 478 | } |
479 | |||
480 | static int | ||
481 | xpc_setup_rsvd_page_sn_uv(struct xpc_rsvd_page *rp) | ||
482 | { | ||
483 | rp->sn.activate_mq_gpa = uv_gpa(xpc_activate_mq_uv); | ||
484 | return 0; | ||
485 | } | ||
486 | |||
487 | static void | ||
488 | xpc_send_heartbeat_uv(int msg_type) | ||
489 | { | ||
490 | short partid; | ||
491 | struct xpc_partition *part; | ||
492 | struct xpc_activate_mq_msg_heartbeat_req_uv msg; | ||
493 | |||
494 | /* | ||
495 | * !!! On uv we're broadcasting a heartbeat message every 5 seconds. | ||
496 | * !!! Whereas on sn2 we're bte_copy'ng the heartbeat info every 20 | ||
497 | * !!! seconds. This is an increase in numalink traffic. | ||
498 | * ??? Is this good? | ||
499 | */ | ||
500 | |||
501 | msg.heartbeat = atomic64_inc_return(&xpc_heartbeat_uv); | ||
502 | |||
503 | partid = find_first_bit(xpc_heartbeating_to_mask_uv, | ||
504 | XP_MAX_NPARTITIONS_UV); | ||
505 | |||
506 | while (partid < XP_MAX_NPARTITIONS_UV) { | ||
507 | part = &xpc_partitions[partid]; | ||
508 | |||
509 | xpc_send_activate_IRQ_part_uv(part, &msg, sizeof(msg), | ||
510 | msg_type); | ||
511 | |||
512 | partid = find_next_bit(xpc_heartbeating_to_mask_uv, | ||
513 | XP_MAX_NPARTITIONS_UV, partid + 1); | ||
514 | } | ||
41 | } | 515 | } |
42 | 516 | ||
43 | static void | 517 | static void |
44 | xpc_increment_heartbeat_uv(void) | 518 | xpc_increment_heartbeat_uv(void) |
45 | { | 519 | { |
46 | /* !!! send heartbeat msg to xpc_heartbeating_to_mask partids */ | 520 | xpc_send_heartbeat_uv(XPC_ACTIVATE_MQ_MSG_INC_HEARTBEAT_UV); |
521 | } | ||
522 | |||
523 | static void | ||
524 | xpc_offline_heartbeat_uv(void) | ||
525 | { | ||
526 | xpc_send_heartbeat_uv(XPC_ACTIVATE_MQ_MSG_OFFLINE_HEARTBEAT_UV); | ||
527 | } | ||
528 | |||
529 | static void | ||
530 | xpc_online_heartbeat_uv(void) | ||
531 | { | ||
532 | xpc_send_heartbeat_uv(XPC_ACTIVATE_MQ_MSG_ONLINE_HEARTBEAT_UV); | ||
47 | } | 533 | } |
48 | 534 | ||
49 | static void | 535 | static void |
50 | xpc_heartbeat_init_uv(void) | 536 | xpc_heartbeat_init_uv(void) |
51 | { | 537 | { |
538 | atomic64_set(&xpc_heartbeat_uv, 0); | ||
52 | bitmap_zero(xpc_heartbeating_to_mask_uv, XP_MAX_NPARTITIONS_UV); | 539 | bitmap_zero(xpc_heartbeating_to_mask_uv, XP_MAX_NPARTITIONS_UV); |
53 | xpc_heartbeating_to_mask = &xpc_heartbeating_to_mask_uv[0]; | 540 | xpc_heartbeating_to_mask = &xpc_heartbeating_to_mask_uv[0]; |
54 | } | 541 | } |
@@ -56,48 +543,94 @@ xpc_heartbeat_init_uv(void) | |||
56 | static void | 543 | static void |
57 | xpc_heartbeat_exit_uv(void) | 544 | xpc_heartbeat_exit_uv(void) |
58 | { | 545 | { |
59 | /* !!! send heartbeat_offline msg to xpc_heartbeating_to_mask partids */ | 546 | xpc_send_heartbeat_uv(XPC_ACTIVATE_MQ_MSG_OFFLINE_HEARTBEAT_UV); |
547 | } | ||
548 | |||
549 | static enum xp_retval | ||
550 | xpc_get_remote_heartbeat_uv(struct xpc_partition *part) | ||
551 | { | ||
552 | struct xpc_partition_uv *part_uv = &part->sn.uv; | ||
553 | enum xp_retval ret = xpNoHeartbeat; | ||
554 | |||
555 | if (part_uv->remote_act_state != XPC_P_AS_INACTIVE && | ||
556 | part_uv->remote_act_state != XPC_P_AS_DEACTIVATING) { | ||
557 | |||
558 | if (part_uv->heartbeat != part->last_heartbeat || | ||
559 | (part_uv->flags & XPC_P_HEARTBEAT_OFFLINE_UV)) { | ||
560 | |||
561 | part->last_heartbeat = part_uv->heartbeat; | ||
562 | ret = xpSuccess; | ||
563 | } | ||
564 | } | ||
565 | return ret; | ||
60 | } | 566 | } |
61 | 567 | ||
62 | static void | 568 | static void |
63 | xpc_request_partition_activation_uv(struct xpc_rsvd_page *remote_rp, | 569 | xpc_request_partition_activation_uv(struct xpc_rsvd_page *remote_rp, |
64 | unsigned long remote_rp_pa, int nasid) | 570 | unsigned long remote_rp_gpa, int nasid) |
65 | { | 571 | { |
66 | short partid = remote_rp->SAL_partid; | 572 | short partid = remote_rp->SAL_partid; |
67 | struct xpc_partition *part = &xpc_partitions[partid]; | 573 | struct xpc_partition *part = &xpc_partitions[partid]; |
574 | struct xpc_activate_mq_msg_activate_req_uv msg; | ||
68 | 575 | ||
69 | /* | 576 | part->remote_rp_pa = remote_rp_gpa; /* !!! _pa here is really _gpa */ |
70 | * !!! Setup part structure with the bits of info we can glean from the rp: | 577 | part->remote_rp_ts_jiffies = remote_rp->ts_jiffies; |
71 | * !!! part->remote_rp_pa = remote_rp_pa; | 578 | part->sn.uv.remote_activate_mq_gpa = remote_rp->sn.activate_mq_gpa; |
72 | * !!! part->sn.uv.activate_mq_gpa = remote_rp->sn.activate_mq_gpa; | 579 | |
73 | */ | 580 | /* |
581 | * ??? Is it a good idea to make this conditional on what is | ||
582 | * ??? potentially stale state information? | ||
583 | */ | ||
584 | if (part->sn.uv.remote_act_state == XPC_P_AS_INACTIVE) { | ||
585 | msg.rp_gpa = uv_gpa(xpc_rsvd_page); | ||
586 | msg.activate_mq_gpa = xpc_rsvd_page->sn.activate_mq_gpa; | ||
587 | xpc_send_activate_IRQ_part_uv(part, &msg, sizeof(msg), | ||
588 | XPC_ACTIVATE_MQ_MSG_ACTIVATE_REQ_UV); | ||
589 | } | ||
74 | 590 | ||
75 | xpc_send_local_activate_IRQ_uv(part); | 591 | if (part->act_state == XPC_P_AS_INACTIVE) |
592 | xpc_send_local_activate_IRQ_uv(part, XPC_P_ASR_ACTIVATE_UV); | ||
76 | } | 593 | } |
77 | 594 | ||
78 | static void | 595 | static void |
79 | xpc_request_partition_reactivation_uv(struct xpc_partition *part) | 596 | xpc_request_partition_reactivation_uv(struct xpc_partition *part) |
80 | { | 597 | { |
81 | xpc_send_local_activate_IRQ_uv(part); | 598 | xpc_send_local_activate_IRQ_uv(part, XPC_P_ASR_ACTIVATE_UV); |
599 | } | ||
600 | |||
601 | static void | ||
602 | xpc_request_partition_deactivation_uv(struct xpc_partition *part) | ||
603 | { | ||
604 | struct xpc_activate_mq_msg_deactivate_req_uv msg; | ||
605 | |||
606 | /* | ||
607 | * ??? Is it a good idea to make this conditional on what is | ||
608 | * ??? potentially stale state information? | ||
609 | */ | ||
610 | if (part->sn.uv.remote_act_state != XPC_P_AS_DEACTIVATING && | ||
611 | part->sn.uv.remote_act_state != XPC_P_AS_INACTIVE) { | ||
612 | |||
613 | msg.reason = part->reason; | ||
614 | xpc_send_activate_IRQ_part_uv(part, &msg, sizeof(msg), | ||
615 | XPC_ACTIVATE_MQ_MSG_DEACTIVATE_REQ_UV); | ||
616 | } | ||
82 | } | 617 | } |
83 | 618 | ||
84 | /* | 619 | /* |
85 | * Setup the infrastructure necessary to support XPartition Communication | 620 | * Setup the channel structures that are uv specific. |
86 | * between the specified remote partition and the local one. | ||
87 | */ | 621 | */ |
88 | static enum xp_retval | 622 | static enum xp_retval |
89 | xpc_setup_infrastructure_uv(struct xpc_partition *part) | 623 | xpc_setup_ch_structures_sn_uv(struct xpc_partition *part) |
90 | { | 624 | { |
91 | /* !!! this function needs fleshing out */ | 625 | /* !!! this function needs fleshing out */ |
92 | return xpUnsupported; | 626 | return xpUnsupported; |
93 | } | 627 | } |
94 | 628 | ||
95 | /* | 629 | /* |
96 | * Teardown the infrastructure necessary to support XPartition Communication | 630 | * Teardown the channel structures that are uv specific. |
97 | * between the specified remote partition and the local one. | ||
98 | */ | 631 | */ |
99 | static void | 632 | static void |
100 | xpc_teardown_infrastructure_uv(struct xpc_partition *part) | 633 | xpc_teardown_ch_structures_sn_uv(struct xpc_partition *part) |
101 | { | 634 | { |
102 | /* !!! this function needs fleshing out */ | 635 | /* !!! this function needs fleshing out */ |
103 | return; | 636 | return; |
@@ -106,15 +639,163 @@ xpc_teardown_infrastructure_uv(struct xpc_partition *part) | |||
106 | static enum xp_retval | 639 | static enum xp_retval |
107 | xpc_make_first_contact_uv(struct xpc_partition *part) | 640 | xpc_make_first_contact_uv(struct xpc_partition *part) |
108 | { | 641 | { |
109 | /* !!! this function needs fleshing out */ | 642 | struct xpc_activate_mq_msg_uv msg; |
110 | return xpUnsupported; | 643 | |
644 | /* | ||
645 | * We send a sync msg to get the remote partition's remote_act_state | ||
646 | * updated to our current act_state which at this point should | ||
647 | * be XPC_P_AS_ACTIVATING. | ||
648 | */ | ||
649 | xpc_send_activate_IRQ_part_uv(part, &msg, sizeof(msg), | ||
650 | XPC_ACTIVATE_MQ_MSG_SYNC_ACT_STATE_UV); | ||
651 | |||
652 | while (part->sn.uv.remote_act_state != XPC_P_AS_ACTIVATING) { | ||
653 | |||
654 | dev_dbg(xpc_part, "waiting to make first contact with " | ||
655 | "partition %d\n", XPC_PARTID(part)); | ||
656 | |||
657 | /* wait a 1/4 of a second or so */ | ||
658 | (void)msleep_interruptible(250); | ||
659 | |||
660 | if (part->act_state == XPC_P_AS_DEACTIVATING) | ||
661 | return part->reason; | ||
662 | } | ||
663 | |||
664 | return xpSuccess; | ||
111 | } | 665 | } |
112 | 666 | ||
113 | static u64 | 667 | static u64 |
114 | xpc_get_chctl_all_flags_uv(struct xpc_partition *part) | 668 | xpc_get_chctl_all_flags_uv(struct xpc_partition *part) |
115 | { | 669 | { |
670 | unsigned long irq_flags; | ||
671 | union xpc_channel_ctl_flags chctl; | ||
672 | |||
673 | spin_lock_irqsave(&part->chctl_lock, irq_flags); | ||
674 | chctl = part->chctl; | ||
675 | if (chctl.all_flags != 0) | ||
676 | part->chctl.all_flags = 0; | ||
677 | |||
678 | spin_unlock_irqrestore(&part->chctl_lock, irq_flags); | ||
679 | return chctl.all_flags; | ||
680 | } | ||
681 | |||
682 | static enum xp_retval | ||
683 | xpc_setup_msg_structures_uv(struct xpc_channel *ch) | ||
684 | { | ||
685 | /* !!! this function needs fleshing out */ | ||
686 | return xpUnsupported; | ||
687 | } | ||
688 | |||
689 | static void | ||
690 | xpc_teardown_msg_structures_uv(struct xpc_channel *ch) | ||
691 | { | ||
692 | struct xpc_channel_uv *ch_uv = &ch->sn.uv; | ||
693 | |||
694 | ch_uv->remote_notify_mq_gpa = 0; | ||
695 | |||
116 | /* !!! this function needs fleshing out */ | 696 | /* !!! this function needs fleshing out */ |
117 | return 0UL; | 697 | } |
698 | |||
699 | static void | ||
700 | xpc_send_chctl_closerequest_uv(struct xpc_channel *ch, unsigned long *irq_flags) | ||
701 | { | ||
702 | struct xpc_activate_mq_msg_chctl_closerequest_uv msg; | ||
703 | |||
704 | msg.ch_number = ch->number; | ||
705 | msg.reason = ch->reason; | ||
706 | xpc_send_activate_IRQ_ch_uv(ch, irq_flags, &msg, sizeof(msg), | ||
707 | XPC_ACTIVATE_MQ_MSG_CHCTL_CLOSEREQUEST_UV); | ||
708 | } | ||
709 | |||
710 | static void | ||
711 | xpc_send_chctl_closereply_uv(struct xpc_channel *ch, unsigned long *irq_flags) | ||
712 | { | ||
713 | struct xpc_activate_mq_msg_chctl_closereply_uv msg; | ||
714 | |||
715 | msg.ch_number = ch->number; | ||
716 | xpc_send_activate_IRQ_ch_uv(ch, irq_flags, &msg, sizeof(msg), | ||
717 | XPC_ACTIVATE_MQ_MSG_CHCTL_CLOSEREPLY_UV); | ||
718 | } | ||
719 | |||
720 | static void | ||
721 | xpc_send_chctl_openrequest_uv(struct xpc_channel *ch, unsigned long *irq_flags) | ||
722 | { | ||
723 | struct xpc_activate_mq_msg_chctl_openrequest_uv msg; | ||
724 | |||
725 | msg.ch_number = ch->number; | ||
726 | msg.msg_size = ch->msg_size; | ||
727 | msg.local_nentries = ch->local_nentries; | ||
728 | xpc_send_activate_IRQ_ch_uv(ch, irq_flags, &msg, sizeof(msg), | ||
729 | XPC_ACTIVATE_MQ_MSG_CHCTL_OPENREQUEST_UV); | ||
730 | } | ||
731 | |||
732 | static void | ||
733 | xpc_send_chctl_openreply_uv(struct xpc_channel *ch, unsigned long *irq_flags) | ||
734 | { | ||
735 | struct xpc_activate_mq_msg_chctl_openreply_uv msg; | ||
736 | |||
737 | msg.ch_number = ch->number; | ||
738 | msg.local_nentries = ch->local_nentries; | ||
739 | msg.remote_nentries = ch->remote_nentries; | ||
740 | msg.local_notify_mq_gpa = uv_gpa(xpc_notify_mq_uv); | ||
741 | xpc_send_activate_IRQ_ch_uv(ch, irq_flags, &msg, sizeof(msg), | ||
742 | XPC_ACTIVATE_MQ_MSG_CHCTL_OPENREPLY_UV); | ||
743 | } | ||
744 | |||
745 | static void | ||
746 | xpc_save_remote_msgqueue_pa_uv(struct xpc_channel *ch, | ||
747 | unsigned long msgqueue_pa) | ||
748 | { | ||
749 | ch->sn.uv.remote_notify_mq_gpa = msgqueue_pa; | ||
750 | } | ||
751 | |||
752 | static void | ||
753 | xpc_indicate_partition_engaged_uv(struct xpc_partition *part) | ||
754 | { | ||
755 | struct xpc_activate_mq_msg_uv msg; | ||
756 | |||
757 | xpc_send_activate_IRQ_part_uv(part, &msg, sizeof(msg), | ||
758 | XPC_ACTIVATE_MQ_MSG_MARK_ENGAGED_UV); | ||
759 | } | ||
760 | |||
761 | static void | ||
762 | xpc_indicate_partition_disengaged_uv(struct xpc_partition *part) | ||
763 | { | ||
764 | struct xpc_activate_mq_msg_uv msg; | ||
765 | |||
766 | xpc_send_activate_IRQ_part_uv(part, &msg, sizeof(msg), | ||
767 | XPC_ACTIVATE_MQ_MSG_MARK_DISENGAGED_UV); | ||
768 | } | ||
769 | |||
770 | static void | ||
771 | xpc_assume_partition_disengaged_uv(short partid) | ||
772 | { | ||
773 | struct xpc_partition_uv *part_uv = &xpc_partitions[partid].sn.uv; | ||
774 | unsigned long irq_flags; | ||
775 | |||
776 | spin_lock_irqsave(&part_uv->flags_lock, irq_flags); | ||
777 | part_uv->flags &= ~XPC_P_ENGAGED_UV; | ||
778 | spin_unlock_irqrestore(&part_uv->flags_lock, irq_flags); | ||
779 | } | ||
780 | |||
781 | static int | ||
782 | xpc_partition_engaged_uv(short partid) | ||
783 | { | ||
784 | return (xpc_partitions[partid].sn.uv.flags & XPC_P_ENGAGED_UV) != 0; | ||
785 | } | ||
786 | |||
787 | static int | ||
788 | xpc_any_partition_engaged_uv(void) | ||
789 | { | ||
790 | struct xpc_partition_uv *part_uv; | ||
791 | short partid; | ||
792 | |||
793 | for (partid = 0; partid < XP_MAX_NPARTITIONS_UV; partid++) { | ||
794 | part_uv = &xpc_partitions[partid].sn.uv; | ||
795 | if ((part_uv->flags & XPC_P_ENGAGED_UV) != 0) | ||
796 | return 1; | ||
797 | } | ||
798 | return 0; | ||
118 | } | 799 | } |
119 | 800 | ||
120 | static struct xpc_msg * | 801 | static struct xpc_msg * |
@@ -124,24 +805,64 @@ xpc_get_deliverable_msg_uv(struct xpc_channel *ch) | |||
124 | return NULL; | 805 | return NULL; |
125 | } | 806 | } |
126 | 807 | ||
127 | void | 808 | int |
128 | xpc_init_uv(void) | 809 | xpc_init_uv(void) |
129 | { | 810 | { |
130 | xpc_rsvd_page_init = xpc_rsvd_page_init_uv; | 811 | xpc_setup_partitions_sn = xpc_setup_partitions_sn_uv; |
812 | xpc_process_activate_IRQ_rcvd = xpc_process_activate_IRQ_rcvd_uv; | ||
813 | xpc_get_partition_rsvd_page_pa = xpc_get_partition_rsvd_page_pa_uv; | ||
814 | xpc_setup_rsvd_page_sn = xpc_setup_rsvd_page_sn_uv; | ||
131 | xpc_increment_heartbeat = xpc_increment_heartbeat_uv; | 815 | xpc_increment_heartbeat = xpc_increment_heartbeat_uv; |
816 | xpc_offline_heartbeat = xpc_offline_heartbeat_uv; | ||
817 | xpc_online_heartbeat = xpc_online_heartbeat_uv; | ||
132 | xpc_heartbeat_init = xpc_heartbeat_init_uv; | 818 | xpc_heartbeat_init = xpc_heartbeat_init_uv; |
133 | xpc_heartbeat_exit = xpc_heartbeat_exit_uv; | 819 | xpc_heartbeat_exit = xpc_heartbeat_exit_uv; |
820 | xpc_get_remote_heartbeat = xpc_get_remote_heartbeat_uv; | ||
821 | |||
134 | xpc_request_partition_activation = xpc_request_partition_activation_uv; | 822 | xpc_request_partition_activation = xpc_request_partition_activation_uv; |
135 | xpc_request_partition_reactivation = | 823 | xpc_request_partition_reactivation = |
136 | xpc_request_partition_reactivation_uv; | 824 | xpc_request_partition_reactivation_uv; |
137 | xpc_setup_infrastructure = xpc_setup_infrastructure_uv; | 825 | xpc_request_partition_deactivation = |
138 | xpc_teardown_infrastructure = xpc_teardown_infrastructure_uv; | 826 | xpc_request_partition_deactivation_uv; |
827 | |||
828 | xpc_setup_ch_structures_sn = xpc_setup_ch_structures_sn_uv; | ||
829 | xpc_teardown_ch_structures_sn = xpc_teardown_ch_structures_sn_uv; | ||
830 | |||
139 | xpc_make_first_contact = xpc_make_first_contact_uv; | 831 | xpc_make_first_contact = xpc_make_first_contact_uv; |
832 | |||
140 | xpc_get_chctl_all_flags = xpc_get_chctl_all_flags_uv; | 833 | xpc_get_chctl_all_flags = xpc_get_chctl_all_flags_uv; |
834 | xpc_send_chctl_closerequest = xpc_send_chctl_closerequest_uv; | ||
835 | xpc_send_chctl_closereply = xpc_send_chctl_closereply_uv; | ||
836 | xpc_send_chctl_openrequest = xpc_send_chctl_openrequest_uv; | ||
837 | xpc_send_chctl_openreply = xpc_send_chctl_openreply_uv; | ||
838 | |||
839 | xpc_save_remote_msgqueue_pa = xpc_save_remote_msgqueue_pa_uv; | ||
840 | |||
841 | xpc_setup_msg_structures = xpc_setup_msg_structures_uv; | ||
842 | xpc_teardown_msg_structures = xpc_teardown_msg_structures_uv; | ||
843 | |||
844 | xpc_indicate_partition_engaged = xpc_indicate_partition_engaged_uv; | ||
845 | xpc_indicate_partition_disengaged = | ||
846 | xpc_indicate_partition_disengaged_uv; | ||
847 | xpc_assume_partition_disengaged = xpc_assume_partition_disengaged_uv; | ||
848 | xpc_partition_engaged = xpc_partition_engaged_uv; | ||
849 | xpc_any_partition_engaged = xpc_any_partition_engaged_uv; | ||
850 | |||
141 | xpc_get_deliverable_msg = xpc_get_deliverable_msg_uv; | 851 | xpc_get_deliverable_msg = xpc_get_deliverable_msg_uv; |
852 | |||
853 | /* ??? The cpuid argument's value is 0, is that what we want? */ | ||
854 | /* !!! The irq argument's value isn't correct. */ | ||
855 | xpc_activate_mq_uv = xpc_create_gru_mq_uv(XPC_ACTIVATE_MQ_SIZE_UV, 0, 0, | ||
856 | xpc_handle_activate_IRQ_uv); | ||
857 | if (xpc_activate_mq_uv == NULL) | ||
858 | return -ENOMEM; | ||
859 | |||
860 | return 0; | ||
142 | } | 861 | } |
143 | 862 | ||
144 | void | 863 | void |
145 | xpc_exit_uv(void) | 864 | xpc_exit_uv(void) |
146 | { | 865 | { |
866 | /* !!! The irq argument's value isn't correct. */ | ||
867 | xpc_destroy_gru_mq_uv(xpc_activate_mq_uv, XPC_ACTIVATE_MQ_SIZE_UV, 0); | ||
147 | } | 868 | } |