diff options
| -rw-r--r-- | drivers/misc/sgi-xp/xp.h | 3 | ||||
| -rw-r--r-- | drivers/misc/sgi-xp/xp_uv.c | 10 | ||||
| -rw-r--r-- | drivers/misc/sgi-xp/xpc.h | 158 | ||||
| -rw-r--r-- | drivers/misc/sgi-xp/xpc_channel.c | 22 | ||||
| -rw-r--r-- | drivers/misc/sgi-xp/xpc_main.c | 329 | ||||
| -rw-r--r-- | drivers/misc/sgi-xp/xpc_partition.c | 28 | ||||
| -rw-r--r-- | drivers/misc/sgi-xp/xpc_sn2.c | 387 | ||||
| -rw-r--r-- | drivers/misc/sgi-xp/xpc_uv.c | 781 |
8 files changed, 1328 insertions, 390 deletions
diff --git a/drivers/misc/sgi-xp/xp.h b/drivers/misc/sgi-xp/xp.h index 45d0a08c2ddd..9ac5758f4d08 100644 --- a/drivers/misc/sgi-xp/xp.h +++ b/drivers/misc/sgi-xp/xp.h | |||
| @@ -208,8 +208,9 @@ enum xp_retval { | |||
| 208 | xpNeedMoreInfo, /* 57: more info is needed by SAL */ | 208 | xpNeedMoreInfo, /* 57: more info is needed by SAL */ |
| 209 | 209 | ||
| 210 | xpGruCopyError, /* 58: gru_copy_gru() returned error */ | 210 | xpGruCopyError, /* 58: gru_copy_gru() returned error */ |
| 211 | xpGruSendMqError, /* 59: gru send message queue related error */ | ||
| 211 | 212 | ||
| 212 | xpUnknownReason /* 59: unknown reason - must be last in enum */ | 213 | xpUnknownReason /* 60: unknown reason - must be last in enum */ |
| 213 | }; | 214 | }; |
| 214 | 215 | ||
| 215 | /* | 216 | /* |
diff --git a/drivers/misc/sgi-xp/xp_uv.c b/drivers/misc/sgi-xp/xp_uv.c index 44f2c2b58c2f..d9f7ce2510bc 100644 --- a/drivers/misc/sgi-xp/xp_uv.c +++ b/drivers/misc/sgi-xp/xp_uv.c | |||
| @@ -42,15 +42,25 @@ xp_remote_memcpy_uv(unsigned long dst_gpa, const unsigned long src_gpa, | |||
| 42 | return xpGruCopyError; | 42 | return xpGruCopyError; |
| 43 | } | 43 | } |
| 44 | 44 | ||
| 45 | static int | ||
| 46 | xp_cpu_to_nasid_uv(int cpuid) | ||
| 47 | { | ||
| 48 | /* ??? Is this same as sn2 nasid in mach/part bitmaps set up by SAL? */ | ||
| 49 | return UV_PNODE_TO_NASID(uv_cpu_to_pnode(cpuid)); | ||
| 50 | } | ||
| 51 | |||
| 45 | enum xp_retval | 52 | enum xp_retval |
| 46 | xp_init_uv(void) | 53 | xp_init_uv(void) |
| 47 | { | 54 | { |
| 48 | BUG_ON(!is_uv()); | 55 | BUG_ON(!is_uv()); |
| 49 | 56 | ||
| 50 | xp_max_npartitions = XP_MAX_NPARTITIONS_UV; | 57 | xp_max_npartitions = XP_MAX_NPARTITIONS_UV; |
| 58 | xp_partition_id = 0; /* !!! not correct value */ | ||
| 59 | xp_region_size = 0; /* !!! not correct value */ | ||
| 51 | 60 | ||
| 52 | xp_pa = xp_pa_uv; | 61 | xp_pa = xp_pa_uv; |
| 53 | xp_remote_memcpy = xp_remote_memcpy_uv; | 62 | xp_remote_memcpy = xp_remote_memcpy_uv; |
| 63 | xp_cpu_to_nasid = xp_cpu_to_nasid_uv; | ||
| 54 | 64 | ||
| 55 | return xpSuccess; | 65 | return xpSuccess; |
| 56 | } | 66 | } |
diff --git a/drivers/misc/sgi-xp/xpc.h b/drivers/misc/sgi-xp/xpc.h index 1e48f7765050..4c26181defff 100644 --- a/drivers/misc/sgi-xp/xpc.h +++ b/drivers/misc/sgi-xp/xpc.h | |||
| @@ -164,8 +164,8 @@ struct xpc_vars_part_sn2 { | |||
| 164 | * MAGIC2 indicates that this partition has pulled the remote partititions | 164 | * MAGIC2 indicates that this partition has pulled the remote partititions |
| 165 | * per partition variables that pertain to this partition. | 165 | * per partition variables that pertain to this partition. |
| 166 | */ | 166 | */ |
| 167 | #define XPC_VP_MAGIC1 0x0053524156435058L /* 'XPCVARS\0'L (little endian) */ | 167 | #define XPC_VP_MAGIC1_SN2 0x0053524156435058L /* 'XPCVARS\0'L (little endian) */ |
| 168 | #define XPC_VP_MAGIC2 0x0073726176435058L /* 'XPCvars\0'L (little endian) */ | 168 | #define XPC_VP_MAGIC2_SN2 0x0073726176435058L /* 'XPCvars\0'L (little endian) */ |
| 169 | 169 | ||
| 170 | /* the reserved page sizes and offsets */ | 170 | /* the reserved page sizes and offsets */ |
| 171 | 171 | ||
| @@ -181,6 +181,80 @@ struct xpc_vars_part_sn2 { | |||
| 181 | xpc_nasid_mask_nlongs)) | 181 | xpc_nasid_mask_nlongs)) |
| 182 | 182 | ||
| 183 | /* | 183 | /* |
| 184 | * The activate_mq is used to send/receive messages that affect XPC's heartbeat, | ||
| 185 | * partition active state, and channel state. This is UV only. | ||
| 186 | */ | ||
| 187 | struct xpc_activate_mq_msghdr_uv { | ||
| 188 | short partid; /* sender's partid */ | ||
| 189 | u8 act_state; /* sender's act_state at time msg sent */ | ||
| 190 | u8 type; /* message's type */ | ||
| 191 | unsigned long rp_ts_jiffies; /* timestamp of sender's rp setup by XPC */ | ||
| 192 | }; | ||
| 193 | |||
| 194 | /* activate_mq defined message types */ | ||
| 195 | #define XPC_ACTIVATE_MQ_MSG_SYNC_ACT_STATE_UV 0 | ||
| 196 | #define XPC_ACTIVATE_MQ_MSG_INC_HEARTBEAT_UV 1 | ||
| 197 | #define XPC_ACTIVATE_MQ_MSG_OFFLINE_HEARTBEAT_UV 2 | ||
| 198 | #define XPC_ACTIVATE_MQ_MSG_ONLINE_HEARTBEAT_UV 3 | ||
| 199 | |||
| 200 | #define XPC_ACTIVATE_MQ_MSG_ACTIVATE_REQ_UV 4 | ||
| 201 | #define XPC_ACTIVATE_MQ_MSG_DEACTIVATE_REQ_UV 5 | ||
| 202 | |||
| 203 | #define XPC_ACTIVATE_MQ_MSG_CHCTL_CLOSEREQUEST_UV 6 | ||
| 204 | #define XPC_ACTIVATE_MQ_MSG_CHCTL_CLOSEREPLY_UV 7 | ||
| 205 | #define XPC_ACTIVATE_MQ_MSG_CHCTL_OPENREQUEST_UV 8 | ||
| 206 | #define XPC_ACTIVATE_MQ_MSG_CHCTL_OPENREPLY_UV 9 | ||
| 207 | |||
| 208 | #define XPC_ACTIVATE_MQ_MSG_MARK_ENGAGED_UV 10 | ||
| 209 | #define XPC_ACTIVATE_MQ_MSG_MARK_DISENGAGED_UV 11 | ||
| 210 | |||
| 211 | struct xpc_activate_mq_msg_uv { | ||
| 212 | struct xpc_activate_mq_msghdr_uv header; | ||
| 213 | }; | ||
| 214 | |||
| 215 | struct xpc_activate_mq_msg_heartbeat_req_uv { | ||
| 216 | struct xpc_activate_mq_msghdr_uv header; | ||
| 217 | u64 heartbeat; | ||
| 218 | }; | ||
| 219 | |||
| 220 | struct xpc_activate_mq_msg_activate_req_uv { | ||
| 221 | struct xpc_activate_mq_msghdr_uv header; | ||
| 222 | unsigned long rp_gpa; | ||
| 223 | unsigned long activate_mq_gpa; | ||
| 224 | }; | ||
| 225 | |||
| 226 | struct xpc_activate_mq_msg_deactivate_req_uv { | ||
| 227 | struct xpc_activate_mq_msghdr_uv header; | ||
| 228 | enum xp_retval reason; | ||
| 229 | }; | ||
| 230 | |||
| 231 | struct xpc_activate_mq_msg_chctl_closerequest_uv { | ||
| 232 | struct xpc_activate_mq_msghdr_uv header; | ||
| 233 | short ch_number; | ||
| 234 | enum xp_retval reason; | ||
| 235 | }; | ||
| 236 | |||
| 237 | struct xpc_activate_mq_msg_chctl_closereply_uv { | ||
| 238 | struct xpc_activate_mq_msghdr_uv header; | ||
| 239 | short ch_number; | ||
| 240 | }; | ||
| 241 | |||
| 242 | struct xpc_activate_mq_msg_chctl_openrequest_uv { | ||
| 243 | struct xpc_activate_mq_msghdr_uv header; | ||
| 244 | short ch_number; | ||
| 245 | short msg_size; /* size of notify_mq's messages */ | ||
| 246 | short local_nentries; /* ??? Is this needed? What is? */ | ||
| 247 | }; | ||
| 248 | |||
| 249 | struct xpc_activate_mq_msg_chctl_openreply_uv { | ||
| 250 | struct xpc_activate_mq_msghdr_uv header; | ||
| 251 | short ch_number; | ||
| 252 | short remote_nentries; /* ??? Is this needed? What is? */ | ||
| 253 | short local_nentries; /* ??? Is this needed? What is? */ | ||
| 254 | unsigned long local_notify_mq_gpa; | ||
| 255 | }; | ||
| 256 | |||
| 257 | /* | ||
| 184 | * Functions registered by add_timer() or called by kernel_thread() only | 258 | * Functions registered by add_timer() or called by kernel_thread() only |
| 185 | * allow for a single 64-bit argument. The following macros can be used to | 259 | * allow for a single 64-bit argument. The following macros can be used to |
| 186 | * pack and unpack two (32-bit, 16-bit or 8-bit) arguments into or out from | 260 | * pack and unpack two (32-bit, 16-bit or 8-bit) arguments into or out from |
| @@ -331,6 +405,18 @@ struct xpc_notify { | |||
| 331 | */ | 405 | */ |
| 332 | 406 | ||
| 333 | struct xpc_channel_sn2 { | 407 | struct xpc_channel_sn2 { |
| 408 | struct xpc_openclose_args *local_openclose_args; /* args passed on */ | ||
| 409 | /* opening or closing of channel */ | ||
| 410 | |||
| 411 | void *local_msgqueue_base; /* base address of kmalloc'd space */ | ||
| 412 | struct xpc_msg *local_msgqueue; /* local message queue */ | ||
| 413 | void *remote_msgqueue_base; /* base address of kmalloc'd space */ | ||
| 414 | struct xpc_msg *remote_msgqueue; /* cached copy of remote partition's */ | ||
| 415 | /* local message queue */ | ||
| 416 | unsigned long remote_msgqueue_pa; /* phys addr of remote partition's */ | ||
| 417 | /* local message queue */ | ||
| 418 | |||
| 419 | struct xpc_notify *notify_queue; /* notify queue for messages sent */ | ||
| 334 | 420 | ||
| 335 | /* various flavors of local and remote Get/Put values */ | 421 | /* various flavors of local and remote Get/Put values */ |
| 336 | 422 | ||
| @@ -344,13 +430,14 @@ struct xpc_channel_sn2 { | |||
| 344 | }; | 430 | }; |
| 345 | 431 | ||
| 346 | struct xpc_channel_uv { | 432 | struct xpc_channel_uv { |
| 347 | /* !!! code is coming */ | 433 | unsigned long remote_notify_mq_gpa; /* gru phys address of remote */ |
| 434 | /* partition's notify mq */ | ||
| 348 | }; | 435 | }; |
| 349 | 436 | ||
| 350 | struct xpc_channel { | 437 | struct xpc_channel { |
| 351 | short partid; /* ID of remote partition connected */ | 438 | short partid; /* ID of remote partition connected */ |
| 352 | spinlock_t lock; /* lock for updating this structure */ | 439 | spinlock_t lock; /* lock for updating this structure */ |
| 353 | u32 flags; /* general flags */ | 440 | unsigned int flags; /* general flags */ |
| 354 | 441 | ||
| 355 | enum xp_retval reason; /* reason why channel is disconnect'g */ | 442 | enum xp_retval reason; /* reason why channel is disconnect'g */ |
| 356 | int reason_line; /* line# disconnect initiated from */ | 443 | int reason_line; /* line# disconnect initiated from */ |
| @@ -361,14 +448,6 @@ struct xpc_channel { | |||
| 361 | u16 local_nentries; /* #of msg entries in local msg queue */ | 448 | u16 local_nentries; /* #of msg entries in local msg queue */ |
| 362 | u16 remote_nentries; /* #of msg entries in remote msg queue */ | 449 | u16 remote_nentries; /* #of msg entries in remote msg queue */ |
| 363 | 450 | ||
| 364 | void *local_msgqueue_base; /* base address of kmalloc'd space */ | ||
| 365 | struct xpc_msg *local_msgqueue; /* local message queue */ | ||
| 366 | void *remote_msgqueue_base; /* base address of kmalloc'd space */ | ||
| 367 | struct xpc_msg *remote_msgqueue; /* cached copy of remote partition's */ | ||
| 368 | /* local message queue */ | ||
| 369 | unsigned long remote_msgqueue_pa; /* phys addr of remote partition's */ | ||
| 370 | /* local message queue */ | ||
| 371 | |||
| 372 | atomic_t references; /* #of external references to queues */ | 451 | atomic_t references; /* #of external references to queues */ |
| 373 | 452 | ||
| 374 | atomic_t n_on_msg_allocate_wq; /* #on msg allocation wait queue */ | 453 | atomic_t n_on_msg_allocate_wq; /* #on msg allocation wait queue */ |
| @@ -377,19 +456,13 @@ struct xpc_channel { | |||
| 377 | u8 delayed_chctl_flags; /* chctl flags received, but delayed */ | 456 | u8 delayed_chctl_flags; /* chctl flags received, but delayed */ |
| 378 | /* action until channel disconnected */ | 457 | /* action until channel disconnected */ |
| 379 | 458 | ||
| 380 | /* queue of msg senders who want to be notified when msg received */ | ||
| 381 | |||
| 382 | atomic_t n_to_notify; /* #of msg senders to notify */ | 459 | atomic_t n_to_notify; /* #of msg senders to notify */ |
| 383 | struct xpc_notify *notify_queue; /* notify queue for messages sent */ | ||
| 384 | 460 | ||
| 385 | xpc_channel_func func; /* user's channel function */ | 461 | xpc_channel_func func; /* user's channel function */ |
| 386 | void *key; /* pointer to user's key */ | 462 | void *key; /* pointer to user's key */ |
| 387 | 463 | ||
| 388 | struct completion wdisconnect_wait; /* wait for channel disconnect */ | 464 | struct completion wdisconnect_wait; /* wait for channel disconnect */ |
| 389 | 465 | ||
| 390 | struct xpc_openclose_args *local_openclose_args; /* args passed on */ | ||
| 391 | /* opening or closing of channel */ | ||
| 392 | |||
| 393 | /* kthread management related fields */ | 466 | /* kthread management related fields */ |
| 394 | 467 | ||
| 395 | atomic_t kthreads_assigned; /* #of kthreads assigned to channel */ | 468 | atomic_t kthreads_assigned; /* #of kthreads assigned to channel */ |
| @@ -507,6 +580,8 @@ struct xpc_partition_sn2 { | |||
| 507 | unsigned long remote_GPs_pa; /* phys addr of remote partition's local */ | 580 | unsigned long remote_GPs_pa; /* phys addr of remote partition's local */ |
| 508 | /* Get/Put values */ | 581 | /* Get/Put values */ |
| 509 | 582 | ||
| 583 | void *local_openclose_args_base; /* base address of kmalloc'd space */ | ||
| 584 | struct xpc_openclose_args *local_openclose_args; /* local's args */ | ||
| 510 | unsigned long remote_openclose_args_pa; /* phys addr of remote's args */ | 585 | unsigned long remote_openclose_args_pa; /* phys addr of remote's args */ |
| 511 | 586 | ||
| 512 | int notify_IRQ_nasid; /* nasid of where to send notify IRQs */ | 587 | int notify_IRQ_nasid; /* nasid of where to send notify IRQs */ |
| @@ -520,9 +595,27 @@ struct xpc_partition_sn2 { | |||
| 520 | }; | 595 | }; |
| 521 | 596 | ||
| 522 | struct xpc_partition_uv { | 597 | struct xpc_partition_uv { |
| 523 | /* !!! code is coming */ | 598 | unsigned long remote_activate_mq_gpa; /* gru phys address of remote */ |
| 599 | /* partition's activate mq */ | ||
| 600 | spinlock_t flags_lock; /* protect updating of flags */ | ||
| 601 | unsigned int flags; /* general flags */ | ||
| 602 | u8 remote_act_state; /* remote partition's act_state */ | ||
| 603 | u8 act_state_req; /* act_state request from remote partition */ | ||
| 604 | enum xp_retval reason; /* reason for deactivate act_state request */ | ||
| 605 | u64 heartbeat; /* incremented by remote partition */ | ||
| 524 | }; | 606 | }; |
| 525 | 607 | ||
| 608 | /* struct xpc_partition_uv flags */ | ||
| 609 | |||
| 610 | #define XPC_P_HEARTBEAT_OFFLINE_UV 0x00000001 | ||
| 611 | #define XPC_P_ENGAGED_UV 0x00000002 | ||
| 612 | |||
| 613 | /* struct xpc_partition_uv act_state change requests */ | ||
| 614 | |||
| 615 | #define XPC_P_ASR_ACTIVATE_UV 0x01 | ||
| 616 | #define XPC_P_ASR_REACTIVATE_UV 0x02 | ||
| 617 | #define XPC_P_ASR_DEACTIVATE_UV 0x03 | ||
| 618 | |||
| 526 | struct xpc_partition { | 619 | struct xpc_partition { |
| 527 | 620 | ||
| 528 | /* XPC HB infrastructure */ | 621 | /* XPC HB infrastructure */ |
| @@ -556,8 +649,6 @@ struct xpc_partition { | |||
| 556 | union xpc_channel_ctl_flags chctl; /* chctl flags yet to be processed */ | 649 | union xpc_channel_ctl_flags chctl; /* chctl flags yet to be processed */ |
| 557 | spinlock_t chctl_lock; /* chctl flags lock */ | 650 | spinlock_t chctl_lock; /* chctl flags lock */ |
| 558 | 651 | ||
| 559 | void *local_openclose_args_base; /* base address of kmalloc'd space */ | ||
| 560 | struct xpc_openclose_args *local_openclose_args; /* local's args */ | ||
| 561 | void *remote_openclose_args_base; /* base address of kmalloc'd space */ | 652 | void *remote_openclose_args_base; /* base address of kmalloc'd space */ |
| 562 | struct xpc_openclose_args *remote_openclose_args; /* copy of remote's */ | 653 | struct xpc_openclose_args *remote_openclose_args; /* copy of remote's */ |
| 563 | /* args */ | 654 | /* args */ |
| @@ -616,17 +707,20 @@ extern struct device *xpc_part; | |||
| 616 | extern struct device *xpc_chan; | 707 | extern struct device *xpc_chan; |
| 617 | extern int xpc_disengage_timelimit; | 708 | extern int xpc_disengage_timelimit; |
| 618 | extern int xpc_disengage_timedout; | 709 | extern int xpc_disengage_timedout; |
| 619 | extern atomic_t xpc_activate_IRQ_rcvd; | 710 | extern int xpc_activate_IRQ_rcvd; |
| 711 | extern spinlock_t xpc_activate_IRQ_rcvd_lock; | ||
| 620 | extern wait_queue_head_t xpc_activate_IRQ_wq; | 712 | extern wait_queue_head_t xpc_activate_IRQ_wq; |
| 621 | extern void *xpc_heartbeating_to_mask; | 713 | extern void *xpc_heartbeating_to_mask; |
| 714 | extern void *xpc_kzalloc_cacheline_aligned(size_t, gfp_t, void **); | ||
| 622 | extern void xpc_activate_partition(struct xpc_partition *); | 715 | extern void xpc_activate_partition(struct xpc_partition *); |
| 623 | extern void xpc_activate_kthreads(struct xpc_channel *, int); | 716 | extern void xpc_activate_kthreads(struct xpc_channel *, int); |
| 624 | extern void xpc_create_kthreads(struct xpc_channel *, int, int); | 717 | extern void xpc_create_kthreads(struct xpc_channel *, int, int); |
| 625 | extern void xpc_disconnect_wait(int); | 718 | extern void xpc_disconnect_wait(int); |
| 719 | extern int (*xpc_setup_partitions_sn) (void); | ||
| 626 | extern enum xp_retval (*xpc_get_partition_rsvd_page_pa) (void *, u64 *, | 720 | extern enum xp_retval (*xpc_get_partition_rsvd_page_pa) (void *, u64 *, |
| 627 | unsigned long *, | 721 | unsigned long *, |
| 628 | size_t *); | 722 | size_t *); |
| 629 | extern enum xp_retval (*xpc_rsvd_page_init) (struct xpc_rsvd_page *); | 723 | extern int (*xpc_setup_rsvd_page_sn) (struct xpc_rsvd_page *); |
| 630 | extern void (*xpc_heartbeat_init) (void); | 724 | extern void (*xpc_heartbeat_init) (void); |
| 631 | extern void (*xpc_heartbeat_exit) (void); | 725 | extern void (*xpc_heartbeat_exit) (void); |
| 632 | extern void (*xpc_increment_heartbeat) (void); | 726 | extern void (*xpc_increment_heartbeat) (void); |
| @@ -635,8 +729,8 @@ extern void (*xpc_online_heartbeat) (void); | |||
| 635 | extern enum xp_retval (*xpc_get_remote_heartbeat) (struct xpc_partition *); | 729 | extern enum xp_retval (*xpc_get_remote_heartbeat) (struct xpc_partition *); |
| 636 | extern enum xp_retval (*xpc_make_first_contact) (struct xpc_partition *); | 730 | extern enum xp_retval (*xpc_make_first_contact) (struct xpc_partition *); |
| 637 | extern u64 (*xpc_get_chctl_all_flags) (struct xpc_partition *); | 731 | extern u64 (*xpc_get_chctl_all_flags) (struct xpc_partition *); |
| 638 | extern enum xp_retval (*xpc_allocate_msgqueues) (struct xpc_channel *); | 732 | extern enum xp_retval (*xpc_setup_msg_structures) (struct xpc_channel *); |
| 639 | extern void (*xpc_free_msgqueues) (struct xpc_channel *); | 733 | extern void (*xpc_teardown_msg_structures) (struct xpc_channel *); |
| 640 | extern void (*xpc_notify_senders_of_disconnect) (struct xpc_channel *); | 734 | extern void (*xpc_notify_senders_of_disconnect) (struct xpc_channel *); |
| 641 | extern void (*xpc_process_msg_chctl_flags) (struct xpc_partition *, int); | 735 | extern void (*xpc_process_msg_chctl_flags) (struct xpc_partition *, int); |
| 642 | extern int (*xpc_n_of_deliverable_msgs) (struct xpc_channel *); | 736 | extern int (*xpc_n_of_deliverable_msgs) (struct xpc_channel *); |
| @@ -647,9 +741,9 @@ extern void (*xpc_request_partition_reactivation) (struct xpc_partition *); | |||
| 647 | extern void (*xpc_request_partition_deactivation) (struct xpc_partition *); | 741 | extern void (*xpc_request_partition_deactivation) (struct xpc_partition *); |
| 648 | extern void (*xpc_cancel_partition_deactivation_request) ( | 742 | extern void (*xpc_cancel_partition_deactivation_request) ( |
| 649 | struct xpc_partition *); | 743 | struct xpc_partition *); |
| 650 | extern void (*xpc_process_activate_IRQ_rcvd) (int); | 744 | extern void (*xpc_process_activate_IRQ_rcvd) (void); |
| 651 | extern enum xp_retval (*xpc_setup_infrastructure) (struct xpc_partition *); | 745 | extern enum xp_retval (*xpc_setup_ch_structures_sn) (struct xpc_partition *); |
| 652 | extern void (*xpc_teardown_infrastructure) (struct xpc_partition *); | 746 | extern void (*xpc_teardown_ch_structures_sn) (struct xpc_partition *); |
| 653 | 747 | ||
| 654 | extern void (*xpc_indicate_partition_engaged) (struct xpc_partition *); | 748 | extern void (*xpc_indicate_partition_engaged) (struct xpc_partition *); |
| 655 | extern int (*xpc_partition_engaged) (short); | 749 | extern int (*xpc_partition_engaged) (short); |
| @@ -665,6 +759,9 @@ extern void (*xpc_send_chctl_openrequest) (struct xpc_channel *, | |||
| 665 | unsigned long *); | 759 | unsigned long *); |
| 666 | extern void (*xpc_send_chctl_openreply) (struct xpc_channel *, unsigned long *); | 760 | extern void (*xpc_send_chctl_openreply) (struct xpc_channel *, unsigned long *); |
| 667 | 761 | ||
| 762 | extern void (*xpc_save_remote_msgqueue_pa) (struct xpc_channel *, | ||
| 763 | unsigned long); | ||
| 764 | |||
| 668 | extern enum xp_retval (*xpc_send_msg) (struct xpc_channel *, u32, void *, u16, | 765 | extern enum xp_retval (*xpc_send_msg) (struct xpc_channel *, u32, void *, u16, |
| 669 | u8, xpc_notify_func, void *); | 766 | u8, xpc_notify_func, void *); |
| 670 | extern void (*xpc_received_msg) (struct xpc_channel *, struct xpc_msg *); | 767 | extern void (*xpc_received_msg) (struct xpc_channel *, struct xpc_msg *); |
| @@ -674,7 +771,7 @@ extern int xpc_init_sn2(void); | |||
| 674 | extern void xpc_exit_sn2(void); | 771 | extern void xpc_exit_sn2(void); |
| 675 | 772 | ||
| 676 | /* found in xpc_uv.c */ | 773 | /* found in xpc_uv.c */ |
| 677 | extern void xpc_init_uv(void); | 774 | extern int xpc_init_uv(void); |
| 678 | extern void xpc_exit_uv(void); | 775 | extern void xpc_exit_uv(void); |
| 679 | 776 | ||
| 680 | /* found in xpc_partition.c */ | 777 | /* found in xpc_partition.c */ |
| @@ -684,7 +781,8 @@ extern struct xpc_rsvd_page *xpc_rsvd_page; | |||
| 684 | extern unsigned long *xpc_mach_nasids; | 781 | extern unsigned long *xpc_mach_nasids; |
| 685 | extern struct xpc_partition *xpc_partitions; | 782 | extern struct xpc_partition *xpc_partitions; |
| 686 | extern void *xpc_kmalloc_cacheline_aligned(size_t, gfp_t, void **); | 783 | extern void *xpc_kmalloc_cacheline_aligned(size_t, gfp_t, void **); |
| 687 | extern struct xpc_rsvd_page *xpc_setup_rsvd_page(void); | 784 | extern int xpc_setup_rsvd_page(void); |
| 785 | extern void xpc_teardown_rsvd_page(void); | ||
| 688 | extern int xpc_identify_activate_IRQ_sender(void); | 786 | extern int xpc_identify_activate_IRQ_sender(void); |
| 689 | extern int xpc_partition_disengaged(struct xpc_partition *); | 787 | extern int xpc_partition_disengaged(struct xpc_partition *); |
| 690 | extern enum xp_retval xpc_mark_partition_active(struct xpc_partition *); | 788 | extern enum xp_retval xpc_mark_partition_active(struct xpc_partition *); |
diff --git a/drivers/misc/sgi-xp/xpc_channel.c b/drivers/misc/sgi-xp/xpc_channel.c index 17ab75d69e83..73df9fb5ee66 100644 --- a/drivers/misc/sgi-xp/xpc_channel.c +++ b/drivers/misc/sgi-xp/xpc_channel.c | |||
| @@ -39,7 +39,7 @@ xpc_process_connect(struct xpc_channel *ch, unsigned long *irq_flags) | |||
| 39 | 39 | ||
| 40 | if (!(ch->flags & XPC_C_SETUP)) { | 40 | if (!(ch->flags & XPC_C_SETUP)) { |
| 41 | spin_unlock_irqrestore(&ch->lock, *irq_flags); | 41 | spin_unlock_irqrestore(&ch->lock, *irq_flags); |
| 42 | ret = xpc_allocate_msgqueues(ch); | 42 | ret = xpc_setup_msg_structures(ch); |
| 43 | spin_lock_irqsave(&ch->lock, *irq_flags); | 43 | spin_lock_irqsave(&ch->lock, *irq_flags); |
| 44 | 44 | ||
| 45 | if (ret != xpSuccess) | 45 | if (ret != xpSuccess) |
| @@ -62,8 +62,6 @@ xpc_process_connect(struct xpc_channel *ch, unsigned long *irq_flags) | |||
| 62 | if (!(ch->flags & XPC_C_ROPENREPLY)) | 62 | if (!(ch->flags & XPC_C_ROPENREPLY)) |
| 63 | return; | 63 | return; |
| 64 | 64 | ||
| 65 | DBUG_ON(ch->remote_msgqueue_pa == 0); | ||
| 66 | |||
| 67 | ch->flags = (XPC_C_CONNECTED | XPC_C_SETUP); /* clear all else */ | 65 | ch->flags = (XPC_C_CONNECTED | XPC_C_SETUP); /* clear all else */ |
| 68 | 66 | ||
| 69 | dev_info(xpc_chan, "channel %d to partition %d connected\n", | 67 | dev_info(xpc_chan, "channel %d to partition %d connected\n", |
| @@ -134,13 +132,23 @@ xpc_process_disconnect(struct xpc_channel *ch, unsigned long *irq_flags) | |||
| 134 | spin_lock_irqsave(&ch->lock, *irq_flags); | 132 | spin_lock_irqsave(&ch->lock, *irq_flags); |
| 135 | } | 133 | } |
| 136 | 134 | ||
| 135 | DBUG_ON(atomic_read(&ch->n_to_notify) != 0); | ||
| 136 | |||
| 137 | /* it's now safe to free the channel's message queues */ | 137 | /* it's now safe to free the channel's message queues */ |
| 138 | xpc_free_msgqueues(ch); | 138 | xpc_teardown_msg_structures(ch); |
| 139 | |||
| 140 | ch->func = NULL; | ||
| 141 | ch->key = NULL; | ||
| 142 | ch->msg_size = 0; | ||
| 143 | ch->local_nentries = 0; | ||
| 144 | ch->remote_nentries = 0; | ||
| 145 | ch->kthreads_assigned_limit = 0; | ||
| 146 | ch->kthreads_idle_limit = 0; | ||
| 139 | 147 | ||
| 140 | /* | 148 | /* |
| 141 | * Mark the channel disconnected and clear all other flags, including | 149 | * Mark the channel disconnected and clear all other flags, including |
| 142 | * XPC_C_SETUP (because of call to xpc_free_msgqueues()) but not | 150 | * XPC_C_SETUP (because of call to xpc_teardown_msg_structures()) but |
| 143 | * including XPC_C_WDISCONNECT (if it was set). | 151 | * not including XPC_C_WDISCONNECT (if it was set). |
| 144 | */ | 152 | */ |
| 145 | ch->flags = (XPC_C_DISCONNECTED | (ch->flags & XPC_C_WDISCONNECT)); | 153 | ch->flags = (XPC_C_DISCONNECTED | (ch->flags & XPC_C_WDISCONNECT)); |
| 146 | 154 | ||
| @@ -395,7 +403,7 @@ again: | |||
| 395 | DBUG_ON(args->remote_nentries == 0); | 403 | DBUG_ON(args->remote_nentries == 0); |
| 396 | 404 | ||
| 397 | ch->flags |= XPC_C_ROPENREPLY; | 405 | ch->flags |= XPC_C_ROPENREPLY; |
| 398 | ch->remote_msgqueue_pa = args->local_msgqueue_pa; | 406 | xpc_save_remote_msgqueue_pa(ch, args->local_msgqueue_pa); |
| 399 | 407 | ||
| 400 | if (args->local_nentries < ch->remote_nentries) { | 408 | if (args->local_nentries < ch->remote_nentries) { |
| 401 | dev_dbg(xpc_chan, "XPC_CHCTL_OPENREPLY: new " | 409 | dev_dbg(xpc_chan, "XPC_CHCTL_OPENREPLY: new " |
diff --git a/drivers/misc/sgi-xp/xpc_main.c b/drivers/misc/sgi-xp/xpc_main.c index b303c130bba8..13ec47928994 100644 --- a/drivers/misc/sgi-xp/xpc_main.c +++ b/drivers/misc/sgi-xp/xpc_main.c | |||
| @@ -141,8 +141,9 @@ static struct ctl_table_header *xpc_sysctl; | |||
| 141 | /* non-zero if any remote partition disengage was timed out */ | 141 | /* non-zero if any remote partition disengage was timed out */ |
| 142 | int xpc_disengage_timedout; | 142 | int xpc_disengage_timedout; |
| 143 | 143 | ||
| 144 | /* #of activate IRQs received */ | 144 | /* #of activate IRQs received and not yet processed */ |
| 145 | atomic_t xpc_activate_IRQ_rcvd = ATOMIC_INIT(0); | 145 | int xpc_activate_IRQ_rcvd; |
| 146 | DEFINE_SPINLOCK(xpc_activate_IRQ_rcvd_lock); | ||
| 146 | 147 | ||
| 147 | /* IRQ handler notifies this wait queue on receipt of an IRQ */ | 148 | /* IRQ handler notifies this wait queue on receipt of an IRQ */ |
| 148 | DECLARE_WAIT_QUEUE_HEAD(xpc_activate_IRQ_wq); | 149 | DECLARE_WAIT_QUEUE_HEAD(xpc_activate_IRQ_wq); |
| @@ -169,10 +170,11 @@ static struct notifier_block xpc_die_notifier = { | |||
| 169 | .notifier_call = xpc_system_die, | 170 | .notifier_call = xpc_system_die, |
| 170 | }; | 171 | }; |
| 171 | 172 | ||
| 173 | int (*xpc_setup_partitions_sn) (void); | ||
| 172 | enum xp_retval (*xpc_get_partition_rsvd_page_pa) (void *buf, u64 *cookie, | 174 | enum xp_retval (*xpc_get_partition_rsvd_page_pa) (void *buf, u64 *cookie, |
| 173 | unsigned long *rp_pa, | 175 | unsigned long *rp_pa, |
| 174 | size_t *len); | 176 | size_t *len); |
| 175 | enum xp_retval (*xpc_rsvd_page_init) (struct xpc_rsvd_page *rp); | 177 | int (*xpc_setup_rsvd_page_sn) (struct xpc_rsvd_page *rp); |
| 176 | void (*xpc_heartbeat_init) (void); | 178 | void (*xpc_heartbeat_init) (void); |
| 177 | void (*xpc_heartbeat_exit) (void); | 179 | void (*xpc_heartbeat_exit) (void); |
| 178 | void (*xpc_increment_heartbeat) (void); | 180 | void (*xpc_increment_heartbeat) (void); |
| @@ -183,8 +185,8 @@ enum xp_retval (*xpc_get_remote_heartbeat) (struct xpc_partition *part); | |||
| 183 | enum xp_retval (*xpc_make_first_contact) (struct xpc_partition *part); | 185 | enum xp_retval (*xpc_make_first_contact) (struct xpc_partition *part); |
| 184 | void (*xpc_notify_senders_of_disconnect) (struct xpc_channel *ch); | 186 | void (*xpc_notify_senders_of_disconnect) (struct xpc_channel *ch); |
| 185 | u64 (*xpc_get_chctl_all_flags) (struct xpc_partition *part); | 187 | u64 (*xpc_get_chctl_all_flags) (struct xpc_partition *part); |
| 186 | enum xp_retval (*xpc_allocate_msgqueues) (struct xpc_channel *ch); | 188 | enum xp_retval (*xpc_setup_msg_structures) (struct xpc_channel *ch); |
| 187 | void (*xpc_free_msgqueues) (struct xpc_channel *ch); | 189 | void (*xpc_teardown_msg_structures) (struct xpc_channel *ch); |
| 188 | void (*xpc_process_msg_chctl_flags) (struct xpc_partition *part, int ch_number); | 190 | void (*xpc_process_msg_chctl_flags) (struct xpc_partition *part, int ch_number); |
| 189 | int (*xpc_n_of_deliverable_msgs) (struct xpc_channel *ch); | 191 | int (*xpc_n_of_deliverable_msgs) (struct xpc_channel *ch); |
| 190 | struct xpc_msg *(*xpc_get_deliverable_msg) (struct xpc_channel *ch); | 192 | struct xpc_msg *(*xpc_get_deliverable_msg) (struct xpc_channel *ch); |
| @@ -196,9 +198,9 @@ void (*xpc_request_partition_reactivation) (struct xpc_partition *part); | |||
| 196 | void (*xpc_request_partition_deactivation) (struct xpc_partition *part); | 198 | void (*xpc_request_partition_deactivation) (struct xpc_partition *part); |
| 197 | void (*xpc_cancel_partition_deactivation_request) (struct xpc_partition *part); | 199 | void (*xpc_cancel_partition_deactivation_request) (struct xpc_partition *part); |
| 198 | 200 | ||
| 199 | void (*xpc_process_activate_IRQ_rcvd) (int n_IRQs_expected); | 201 | void (*xpc_process_activate_IRQ_rcvd) (void); |
| 200 | enum xp_retval (*xpc_setup_infrastructure) (struct xpc_partition *part); | 202 | enum xp_retval (*xpc_setup_ch_structures_sn) (struct xpc_partition *part); |
| 201 | void (*xpc_teardown_infrastructure) (struct xpc_partition *part); | 203 | void (*xpc_teardown_ch_structures_sn) (struct xpc_partition *part); |
| 202 | 204 | ||
| 203 | void (*xpc_indicate_partition_engaged) (struct xpc_partition *part); | 205 | void (*xpc_indicate_partition_engaged) (struct xpc_partition *part); |
| 204 | int (*xpc_partition_engaged) (short partid); | 206 | int (*xpc_partition_engaged) (short partid); |
| @@ -215,6 +217,9 @@ void (*xpc_send_chctl_openrequest) (struct xpc_channel *ch, | |||
| 215 | void (*xpc_send_chctl_openreply) (struct xpc_channel *ch, | 217 | void (*xpc_send_chctl_openreply) (struct xpc_channel *ch, |
| 216 | unsigned long *irq_flags); | 218 | unsigned long *irq_flags); |
| 217 | 219 | ||
| 220 | void (*xpc_save_remote_msgqueue_pa) (struct xpc_channel *ch, | ||
| 221 | unsigned long msgqueue_pa); | ||
| 222 | |||
| 218 | enum xp_retval (*xpc_send_msg) (struct xpc_channel *ch, u32 flags, | 223 | enum xp_retval (*xpc_send_msg) (struct xpc_channel *ch, u32 flags, |
| 219 | void *payload, u16 payload_size, u8 notify_type, | 224 | void *payload, u16 payload_size, u8 notify_type, |
| 220 | xpc_notify_func func, void *key); | 225 | xpc_notify_func func, void *key); |
| @@ -308,8 +313,6 @@ xpc_check_remote_hb(void) | |||
| 308 | static int | 313 | static int |
| 309 | xpc_hb_checker(void *ignore) | 314 | xpc_hb_checker(void *ignore) |
| 310 | { | 315 | { |
| 311 | int last_IRQ_count = 0; | ||
| 312 | int new_IRQ_count; | ||
| 313 | int force_IRQ = 0; | 316 | int force_IRQ = 0; |
| 314 | 317 | ||
| 315 | /* this thread was marked active by xpc_hb_init() */ | 318 | /* this thread was marked active by xpc_hb_init() */ |
| @@ -325,43 +328,37 @@ xpc_hb_checker(void *ignore) | |||
| 325 | dev_dbg(xpc_part, "woke up with %d ticks rem; %d IRQs have " | 328 | dev_dbg(xpc_part, "woke up with %d ticks rem; %d IRQs have " |
| 326 | "been received\n", | 329 | "been received\n", |
| 327 | (int)(xpc_hb_check_timeout - jiffies), | 330 | (int)(xpc_hb_check_timeout - jiffies), |
| 328 | atomic_read(&xpc_activate_IRQ_rcvd) - last_IRQ_count); | 331 | xpc_activate_IRQ_rcvd); |
| 329 | 332 | ||
| 330 | /* checking of remote heartbeats is skewed by IRQ handling */ | 333 | /* checking of remote heartbeats is skewed by IRQ handling */ |
| 331 | if (time_is_before_eq_jiffies(xpc_hb_check_timeout)) { | 334 | if (time_is_before_eq_jiffies(xpc_hb_check_timeout)) { |
| 335 | xpc_hb_check_timeout = jiffies + | ||
| 336 | (xpc_hb_check_interval * HZ); | ||
| 337 | |||
| 332 | dev_dbg(xpc_part, "checking remote heartbeats\n"); | 338 | dev_dbg(xpc_part, "checking remote heartbeats\n"); |
| 333 | xpc_check_remote_hb(); | 339 | xpc_check_remote_hb(); |
| 334 | 340 | ||
| 335 | /* | 341 | /* |
| 336 | * We need to periodically recheck to ensure no | 342 | * On sn2 we need to periodically recheck to ensure no |
| 337 | * IRQ/amo pairs have been missed. That check | 343 | * IRQ/amo pairs have been missed. |
| 338 | * must always reset xpc_hb_check_timeout. | ||
| 339 | */ | 344 | */ |
| 340 | force_IRQ = 1; | 345 | if (is_shub()) |
| 346 | force_IRQ = 1; | ||
| 341 | } | 347 | } |
| 342 | 348 | ||
| 343 | /* check for outstanding IRQs */ | 349 | /* check for outstanding IRQs */ |
| 344 | new_IRQ_count = atomic_read(&xpc_activate_IRQ_rcvd); | 350 | if (xpc_activate_IRQ_rcvd > 0 || force_IRQ != 0) { |
| 345 | if (last_IRQ_count < new_IRQ_count || force_IRQ != 0) { | ||
| 346 | force_IRQ = 0; | 351 | force_IRQ = 0; |
| 347 | 352 | dev_dbg(xpc_part, "processing activate IRQs " | |
| 348 | dev_dbg(xpc_part, "found an IRQ to process; will be " | 353 | "received\n"); |
| 349 | "resetting xpc_hb_check_timeout\n"); | 354 | xpc_process_activate_IRQ_rcvd(); |
| 350 | |||
| 351 | xpc_process_activate_IRQ_rcvd(new_IRQ_count - | ||
| 352 | last_IRQ_count); | ||
| 353 | last_IRQ_count = new_IRQ_count; | ||
| 354 | |||
| 355 | xpc_hb_check_timeout = jiffies + | ||
| 356 | (xpc_hb_check_interval * HZ); | ||
| 357 | } | 355 | } |
| 358 | 356 | ||
| 359 | /* wait for IRQ or timeout */ | 357 | /* wait for IRQ or timeout */ |
| 360 | (void)wait_event_interruptible(xpc_activate_IRQ_wq, | 358 | (void)wait_event_interruptible(xpc_activate_IRQ_wq, |
| 361 | (last_IRQ_count < atomic_read( | 359 | (time_is_before_eq_jiffies( |
| 362 | &xpc_activate_IRQ_rcvd) | ||
| 363 | || time_is_before_eq_jiffies( | ||
| 364 | xpc_hb_check_timeout) || | 360 | xpc_hb_check_timeout) || |
| 361 | xpc_activate_IRQ_rcvd > 0 || | ||
| 365 | xpc_exiting)); | 362 | xpc_exiting)); |
| 366 | } | 363 | } |
| 367 | 364 | ||
| @@ -437,6 +434,153 @@ xpc_channel_mgr(struct xpc_partition *part) | |||
| 437 | } | 434 | } |
| 438 | 435 | ||
| 439 | /* | 436 | /* |
| 437 | * Guarantee that the kzalloc'd memory is cacheline aligned. | ||
| 438 | */ | ||
| 439 | void * | ||
| 440 | xpc_kzalloc_cacheline_aligned(size_t size, gfp_t flags, void **base) | ||
| 441 | { | ||
| 442 | /* see if kzalloc will give us cachline aligned memory by default */ | ||
| 443 | *base = kzalloc(size, flags); | ||
| 444 | if (*base == NULL) | ||
| 445 | return NULL; | ||
| 446 | |||
| 447 | if ((u64)*base == L1_CACHE_ALIGN((u64)*base)) | ||
| 448 | return *base; | ||
| 449 | |||
| 450 | kfree(*base); | ||
| 451 | |||
| 452 | /* nope, we'll have to do it ourselves */ | ||
| 453 | *base = kzalloc(size + L1_CACHE_BYTES, flags); | ||
| 454 | if (*base == NULL) | ||
| 455 | return NULL; | ||
| 456 | |||
| 457 | return (void *)L1_CACHE_ALIGN((u64)*base); | ||
| 458 | } | ||
| 459 | |||
| 460 | /* | ||
| 461 | * Setup the channel structures necessary to support XPartition Communication | ||
| 462 | * between the specified remote partition and the local one. | ||
| 463 | */ | ||
| 464 | static enum xp_retval | ||
| 465 | xpc_setup_ch_structures(struct xpc_partition *part) | ||
| 466 | { | ||
| 467 | enum xp_retval ret; | ||
| 468 | int ch_number; | ||
| 469 | struct xpc_channel *ch; | ||
| 470 | short partid = XPC_PARTID(part); | ||
| 471 | |||
| 472 | /* | ||
| 473 | * Allocate all of the channel structures as a contiguous chunk of | ||
| 474 | * memory. | ||
| 475 | */ | ||
| 476 | DBUG_ON(part->channels != NULL); | ||
| 477 | part->channels = kzalloc(sizeof(struct xpc_channel) * XPC_MAX_NCHANNELS, | ||
| 478 | GFP_KERNEL); | ||
| 479 | if (part->channels == NULL) { | ||
| 480 | dev_err(xpc_chan, "can't get memory for channels\n"); | ||
| 481 | return xpNoMemory; | ||
| 482 | } | ||
| 483 | |||
| 484 | /* allocate the remote open and close args */ | ||
| 485 | |||
| 486 | part->remote_openclose_args = | ||
| 487 | xpc_kzalloc_cacheline_aligned(XPC_OPENCLOSE_ARGS_SIZE, | ||
| 488 | GFP_KERNEL, &part-> | ||
| 489 | remote_openclose_args_base); | ||
| 490 | if (part->remote_openclose_args == NULL) { | ||
| 491 | dev_err(xpc_chan, "can't get memory for remote connect args\n"); | ||
| 492 | ret = xpNoMemory; | ||
| 493 | goto out_1; | ||
| 494 | } | ||
| 495 | |||
| 496 | part->chctl.all_flags = 0; | ||
| 497 | spin_lock_init(&part->chctl_lock); | ||
| 498 | |||
| 499 | atomic_set(&part->channel_mgr_requests, 1); | ||
| 500 | init_waitqueue_head(&part->channel_mgr_wq); | ||
| 501 | |||
| 502 | part->nchannels = XPC_MAX_NCHANNELS; | ||
| 503 | |||
| 504 | atomic_set(&part->nchannels_active, 0); | ||
| 505 | atomic_set(&part->nchannels_engaged, 0); | ||
| 506 | |||
| 507 | for (ch_number = 0; ch_number < part->nchannels; ch_number++) { | ||
| 508 | ch = &part->channels[ch_number]; | ||
| 509 | |||
| 510 | ch->partid = partid; | ||
| 511 | ch->number = ch_number; | ||
| 512 | ch->flags = XPC_C_DISCONNECTED; | ||
| 513 | |||
| 514 | atomic_set(&ch->kthreads_assigned, 0); | ||
| 515 | atomic_set(&ch->kthreads_idle, 0); | ||
| 516 | atomic_set(&ch->kthreads_active, 0); | ||
| 517 | |||
| 518 | atomic_set(&ch->references, 0); | ||
| 519 | atomic_set(&ch->n_to_notify, 0); | ||
| 520 | |||
| 521 | spin_lock_init(&ch->lock); | ||
| 522 | init_completion(&ch->wdisconnect_wait); | ||
| 523 | |||
| 524 | atomic_set(&ch->n_on_msg_allocate_wq, 0); | ||
| 525 | init_waitqueue_head(&ch->msg_allocate_wq); | ||
| 526 | init_waitqueue_head(&ch->idle_wq); | ||
| 527 | } | ||
| 528 | |||
| 529 | ret = xpc_setup_ch_structures_sn(part); | ||
| 530 | if (ret != xpSuccess) | ||
| 531 | goto out_2; | ||
| 532 | |||
| 533 | /* | ||
| 534 | * With the setting of the partition setup_state to XPC_P_SS_SETUP, | ||
| 535 | * we're declaring that this partition is ready to go. | ||
| 536 | */ | ||
| 537 | part->setup_state = XPC_P_SS_SETUP; | ||
| 538 | |||
| 539 | return xpSuccess; | ||
| 540 | |||
| 541 | /* setup of ch structures failed */ | ||
| 542 | out_2: | ||
| 543 | kfree(part->remote_openclose_args_base); | ||
| 544 | part->remote_openclose_args = NULL; | ||
| 545 | out_1: | ||
| 546 | kfree(part->channels); | ||
| 547 | part->channels = NULL; | ||
| 548 | return ret; | ||
| 549 | } | ||
| 550 | |||
| 551 | /* | ||
| 552 | * Teardown the channel structures necessary to support XPartition Communication | ||
| 553 | * between the specified remote partition and the local one. | ||
| 554 | */ | ||
| 555 | static void | ||
| 556 | xpc_teardown_ch_structures(struct xpc_partition *part) | ||
| 557 | { | ||
| 558 | DBUG_ON(atomic_read(&part->nchannels_engaged) != 0); | ||
| 559 | DBUG_ON(atomic_read(&part->nchannels_active) != 0); | ||
| 560 | |||
| 561 | /* | ||
| 562 | * Make this partition inaccessible to local processes by marking it | ||
| 563 | * as no longer setup. Then wait before proceeding with the teardown | ||
| 564 | * until all existing references cease. | ||
| 565 | */ | ||
| 566 | DBUG_ON(part->setup_state != XPC_P_SS_SETUP); | ||
| 567 | part->setup_state = XPC_P_SS_WTEARDOWN; | ||
| 568 | |||
| 569 | wait_event(part->teardown_wq, (atomic_read(&part->references) == 0)); | ||
| 570 | |||
| 571 | /* now we can begin tearing down the infrastructure */ | ||
| 572 | |||
| 573 | xpc_teardown_ch_structures_sn(part); | ||
| 574 | |||
| 575 | kfree(part->remote_openclose_args_base); | ||
| 576 | part->remote_openclose_args = NULL; | ||
| 577 | kfree(part->channels); | ||
| 578 | part->channels = NULL; | ||
| 579 | |||
| 580 | part->setup_state = XPC_P_SS_TORNDOWN; | ||
| 581 | } | ||
| 582 | |||
| 583 | /* | ||
| 440 | * When XPC HB determines that a partition has come up, it will create a new | 584 | * When XPC HB determines that a partition has come up, it will create a new |
| 441 | * kthread and that kthread will call this function to attempt to set up the | 585 | * kthread and that kthread will call this function to attempt to set up the |
| 442 | * basic infrastructure used for Cross Partition Communication with the newly | 586 | * basic infrastructure used for Cross Partition Communication with the newly |
| @@ -476,7 +620,7 @@ xpc_activating(void *__partid) | |||
| 476 | 620 | ||
| 477 | xpc_allow_hb(partid); | 621 | xpc_allow_hb(partid); |
| 478 | 622 | ||
| 479 | if (xpc_setup_infrastructure(part) == xpSuccess) { | 623 | if (xpc_setup_ch_structures(part) == xpSuccess) { |
| 480 | (void)xpc_part_ref(part); /* this will always succeed */ | 624 | (void)xpc_part_ref(part); /* this will always succeed */ |
| 481 | 625 | ||
| 482 | if (xpc_make_first_contact(part) == xpSuccess) { | 626 | if (xpc_make_first_contact(part) == xpSuccess) { |
| @@ -486,7 +630,7 @@ xpc_activating(void *__partid) | |||
| 486 | } | 630 | } |
| 487 | 631 | ||
| 488 | xpc_part_deref(part); | 632 | xpc_part_deref(part); |
| 489 | xpc_teardown_infrastructure(part); | 633 | xpc_teardown_ch_structures(part); |
| 490 | } | 634 | } |
| 491 | 635 | ||
| 492 | xpc_disallow_hb(partid); | 636 | xpc_disallow_hb(partid); |
| @@ -806,6 +950,56 @@ xpc_disconnect_wait(int ch_number) | |||
| 806 | } | 950 | } |
| 807 | } | 951 | } |
| 808 | 952 | ||
| 953 | static int | ||
| 954 | xpc_setup_partitions(void) | ||
| 955 | { | ||
| 956 | short partid; | ||
| 957 | struct xpc_partition *part; | ||
| 958 | |||
| 959 | xpc_partitions = kzalloc(sizeof(struct xpc_partition) * | ||
| 960 | xp_max_npartitions, GFP_KERNEL); | ||
| 961 | if (xpc_partitions == NULL) { | ||
| 962 | dev_err(xpc_part, "can't get memory for partition structure\n"); | ||
| 963 | return -ENOMEM; | ||
| 964 | } | ||
| 965 | |||
| 966 | /* | ||
| 967 | * The first few fields of each entry of xpc_partitions[] need to | ||
| 968 | * be initialized now so that calls to xpc_connect() and | ||
| 969 | * xpc_disconnect() can be made prior to the activation of any remote | ||
| 970 | * partition. NOTE THAT NONE OF THE OTHER FIELDS BELONGING TO THESE | ||
| 971 | * ENTRIES ARE MEANINGFUL UNTIL AFTER AN ENTRY'S CORRESPONDING | ||
| 972 | * PARTITION HAS BEEN ACTIVATED. | ||
| 973 | */ | ||
| 974 | for (partid = 0; partid < xp_max_npartitions; partid++) { | ||
| 975 | part = &xpc_partitions[partid]; | ||
| 976 | |||
| 977 | DBUG_ON((u64)part != L1_CACHE_ALIGN((u64)part)); | ||
| 978 | |||
| 979 | part->activate_IRQ_rcvd = 0; | ||
| 980 | spin_lock_init(&part->act_lock); | ||
| 981 | part->act_state = XPC_P_AS_INACTIVE; | ||
| 982 | XPC_SET_REASON(part, 0, 0); | ||
| 983 | |||
| 984 | init_timer(&part->disengage_timer); | ||
| 985 | part->disengage_timer.function = | ||
| 986 | xpc_timeout_partition_disengage; | ||
| 987 | part->disengage_timer.data = (unsigned long)part; | ||
| 988 | |||
| 989 | part->setup_state = XPC_P_SS_UNSET; | ||
| 990 | init_waitqueue_head(&part->teardown_wq); | ||
| 991 | atomic_set(&part->references, 0); | ||
| 992 | } | ||
| 993 | |||
| 994 | return xpc_setup_partitions_sn(); | ||
| 995 | } | ||
| 996 | |||
| 997 | static void | ||
| 998 | xpc_teardown_partitions(void) | ||
| 999 | { | ||
| 1000 | kfree(xpc_partitions); | ||
| 1001 | } | ||
| 1002 | |||
| 809 | static void | 1003 | static void |
| 810 | xpc_do_exit(enum xp_retval reason) | 1004 | xpc_do_exit(enum xp_retval reason) |
| 811 | { | 1005 | { |
| @@ -892,8 +1086,7 @@ xpc_do_exit(enum xp_retval reason) | |||
| 892 | DBUG_ON(xpc_any_partition_engaged()); | 1086 | DBUG_ON(xpc_any_partition_engaged()); |
| 893 | DBUG_ON(xpc_any_hbs_allowed() != 0); | 1087 | DBUG_ON(xpc_any_hbs_allowed() != 0); |
| 894 | 1088 | ||
| 895 | /* a zero timestamp indicates our rsvd page is not initialized */ | 1089 | xpc_teardown_rsvd_page(); |
| 896 | xpc_rsvd_page->ts_jiffies = 0; | ||
| 897 | 1090 | ||
| 898 | if (reason == xpUnloading) { | 1091 | if (reason == xpUnloading) { |
| 899 | (void)unregister_die_notifier(&xpc_die_notifier); | 1092 | (void)unregister_die_notifier(&xpc_die_notifier); |
| @@ -906,7 +1099,7 @@ xpc_do_exit(enum xp_retval reason) | |||
| 906 | if (xpc_sysctl) | 1099 | if (xpc_sysctl) |
| 907 | unregister_sysctl_table(xpc_sysctl); | 1100 | unregister_sysctl_table(xpc_sysctl); |
| 908 | 1101 | ||
| 909 | kfree(xpc_partitions); | 1102 | xpc_teardown_partitions(); |
| 910 | 1103 | ||
| 911 | if (is_shub()) | 1104 | if (is_shub()) |
| 912 | xpc_exit_sn2(); | 1105 | xpc_exit_sn2(); |
| @@ -1062,8 +1255,6 @@ int __init | |||
| 1062 | xpc_init(void) | 1255 | xpc_init(void) |
| 1063 | { | 1256 | { |
| 1064 | int ret; | 1257 | int ret; |
| 1065 | short partid; | ||
| 1066 | struct xpc_partition *part; | ||
| 1067 | struct task_struct *kthread; | 1258 | struct task_struct *kthread; |
| 1068 | 1259 | ||
| 1069 | snprintf(xpc_part->bus_id, BUS_ID_SIZE, "part"); | 1260 | snprintf(xpc_part->bus_id, BUS_ID_SIZE, "part"); |
| @@ -1076,56 +1267,29 @@ xpc_init(void) | |||
| 1076 | * further to only support exactly 64 partitions on this | 1267 | * further to only support exactly 64 partitions on this |
| 1077 | * architecture, no less. | 1268 | * architecture, no less. |
| 1078 | */ | 1269 | */ |
| 1079 | if (xp_max_npartitions != 64) | 1270 | if (xp_max_npartitions != 64) { |
| 1080 | return -EINVAL; | 1271 | dev_err(xpc_part, "max #of partitions not set to 64\n"); |
| 1081 | 1272 | ret = -EINVAL; | |
| 1082 | ret = xpc_init_sn2(); | 1273 | } else { |
| 1083 | if (ret != 0) | 1274 | ret = xpc_init_sn2(); |
| 1084 | return ret; | 1275 | } |
| 1085 | 1276 | ||
| 1086 | } else if (is_uv()) { | 1277 | } else if (is_uv()) { |
| 1087 | xpc_init_uv(); | 1278 | ret = xpc_init_uv(); |
| 1088 | 1279 | ||
| 1089 | } else { | 1280 | } else { |
| 1090 | return -ENODEV; | 1281 | ret = -ENODEV; |
| 1091 | } | 1282 | } |
| 1092 | 1283 | ||
| 1093 | xpc_partitions = kzalloc(sizeof(struct xpc_partition) * | 1284 | if (ret != 0) |
| 1094 | xp_max_npartitions, GFP_KERNEL); | 1285 | return ret; |
| 1095 | if (xpc_partitions == NULL) { | 1286 | |
| 1287 | ret = xpc_setup_partitions(); | ||
| 1288 | if (ret != 0) { | ||
| 1096 | dev_err(xpc_part, "can't get memory for partition structure\n"); | 1289 | dev_err(xpc_part, "can't get memory for partition structure\n"); |
| 1097 | ret = -ENOMEM; | ||
| 1098 | goto out_1; | 1290 | goto out_1; |
| 1099 | } | 1291 | } |
| 1100 | 1292 | ||
| 1101 | /* | ||
| 1102 | * The first few fields of each entry of xpc_partitions[] need to | ||
| 1103 | * be initialized now so that calls to xpc_connect() and | ||
| 1104 | * xpc_disconnect() can be made prior to the activation of any remote | ||
| 1105 | * partition. NOTE THAT NONE OF THE OTHER FIELDS BELONGING TO THESE | ||
| 1106 | * ENTRIES ARE MEANINGFUL UNTIL AFTER AN ENTRY'S CORRESPONDING | ||
| 1107 | * PARTITION HAS BEEN ACTIVATED. | ||
| 1108 | */ | ||
| 1109 | for (partid = 0; partid < xp_max_npartitions; partid++) { | ||
| 1110 | part = &xpc_partitions[partid]; | ||
| 1111 | |||
| 1112 | DBUG_ON((u64)part != L1_CACHE_ALIGN((u64)part)); | ||
| 1113 | |||
| 1114 | part->activate_IRQ_rcvd = 0; | ||
| 1115 | spin_lock_init(&part->act_lock); | ||
| 1116 | part->act_state = XPC_P_AS_INACTIVE; | ||
| 1117 | XPC_SET_REASON(part, 0, 0); | ||
| 1118 | |||
| 1119 | init_timer(&part->disengage_timer); | ||
| 1120 | part->disengage_timer.function = | ||
| 1121 | xpc_timeout_partition_disengage; | ||
| 1122 | part->disengage_timer.data = (unsigned long)part; | ||
| 1123 | |||
| 1124 | part->setup_state = XPC_P_SS_UNSET; | ||
| 1125 | init_waitqueue_head(&part->teardown_wq); | ||
| 1126 | atomic_set(&part->references, 0); | ||
| 1127 | } | ||
| 1128 | |||
| 1129 | xpc_sysctl = register_sysctl_table(xpc_sys_dir); | 1293 | xpc_sysctl = register_sysctl_table(xpc_sys_dir); |
| 1130 | 1294 | ||
| 1131 | /* | 1295 | /* |
| @@ -1133,10 +1297,9 @@ xpc_init(void) | |||
| 1133 | * other partitions to discover we are alive and establish initial | 1297 | * other partitions to discover we are alive and establish initial |
| 1134 | * communications. | 1298 | * communications. |
| 1135 | */ | 1299 | */ |
| 1136 | xpc_rsvd_page = xpc_setup_rsvd_page(); | 1300 | ret = xpc_setup_rsvd_page(); |
| 1137 | if (xpc_rsvd_page == NULL) { | 1301 | if (ret != 0) { |
| 1138 | dev_err(xpc_part, "can't setup our reserved page\n"); | 1302 | dev_err(xpc_part, "can't setup our reserved page\n"); |
| 1139 | ret = -EBUSY; | ||
| 1140 | goto out_2; | 1303 | goto out_2; |
| 1141 | } | 1304 | } |
| 1142 | 1305 | ||
| @@ -1187,15 +1350,15 @@ xpc_init(void) | |||
| 1187 | 1350 | ||
| 1188 | /* initialization was not successful */ | 1351 | /* initialization was not successful */ |
| 1189 | out_3: | 1352 | out_3: |
| 1190 | /* a zero timestamp indicates our rsvd page is not initialized */ | 1353 | xpc_teardown_rsvd_page(); |
| 1191 | xpc_rsvd_page->ts_jiffies = 0; | ||
| 1192 | 1354 | ||
| 1193 | (void)unregister_die_notifier(&xpc_die_notifier); | 1355 | (void)unregister_die_notifier(&xpc_die_notifier); |
| 1194 | (void)unregister_reboot_notifier(&xpc_reboot_notifier); | 1356 | (void)unregister_reboot_notifier(&xpc_reboot_notifier); |
| 1195 | out_2: | 1357 | out_2: |
| 1196 | if (xpc_sysctl) | 1358 | if (xpc_sysctl) |
| 1197 | unregister_sysctl_table(xpc_sysctl); | 1359 | unregister_sysctl_table(xpc_sysctl); |
| 1198 | kfree(xpc_partitions); | 1360 | |
| 1361 | xpc_teardown_partitions(); | ||
| 1199 | out_1: | 1362 | out_1: |
| 1200 | if (is_shub()) | 1363 | if (is_shub()) |
| 1201 | xpc_exit_sn2(); | 1364 | xpc_exit_sn2(); |
diff --git a/drivers/misc/sgi-xp/xpc_partition.c b/drivers/misc/sgi-xp/xpc_partition.c index b5fb21641130..6722f6fe4dc7 100644 --- a/drivers/misc/sgi-xp/xpc_partition.c +++ b/drivers/misc/sgi-xp/xpc_partition.c | |||
| @@ -73,6 +73,12 @@ xpc_get_rsvd_page_pa(int nasid) | |||
| 73 | 73 | ||
| 74 | while (1) { | 74 | while (1) { |
| 75 | 75 | ||
| 76 | /* !!! rp_pa will need to be _gpa on UV. | ||
| 77 | * ??? So do we save it into the architecture specific parts | ||
| 78 | * ??? of the xpc_partition structure? Do we rename this | ||
| 79 | * ??? function or have two versions? Rename rp_pa for UV to | ||
| 80 | * ??? rp_gpa? | ||
| 81 | */ | ||
| 76 | ret = xpc_get_partition_rsvd_page_pa(buf, &cookie, &rp_pa, | 82 | ret = xpc_get_partition_rsvd_page_pa(buf, &cookie, &rp_pa, |
| 77 | &len); | 83 | &len); |
| 78 | 84 | ||
| @@ -118,9 +124,10 @@ xpc_get_rsvd_page_pa(int nasid) | |||
| 118 | * other partitions to discover we are alive and establish initial | 124 | * other partitions to discover we are alive and establish initial |
| 119 | * communications. | 125 | * communications. |
| 120 | */ | 126 | */ |
| 121 | struct xpc_rsvd_page * | 127 | int |
| 122 | xpc_setup_rsvd_page(void) | 128 | xpc_setup_rsvd_page(void) |
| 123 | { | 129 | { |
| 130 | int ret; | ||
| 124 | struct xpc_rsvd_page *rp; | 131 | struct xpc_rsvd_page *rp; |
| 125 | unsigned long rp_pa; | 132 | unsigned long rp_pa; |
| 126 | unsigned long new_ts_jiffies; | 133 | unsigned long new_ts_jiffies; |
| @@ -132,7 +139,7 @@ xpc_setup_rsvd_page(void) | |||
| 132 | preempt_enable(); | 139 | preempt_enable(); |
| 133 | if (rp_pa == 0) { | 140 | if (rp_pa == 0) { |
| 134 | dev_err(xpc_part, "SAL failed to locate the reserved page\n"); | 141 | dev_err(xpc_part, "SAL failed to locate the reserved page\n"); |
| 135 | return NULL; | 142 | return -ESRCH; |
| 136 | } | 143 | } |
| 137 | rp = (struct xpc_rsvd_page *)__va(rp_pa); | 144 | rp = (struct xpc_rsvd_page *)__va(rp_pa); |
| 138 | 145 | ||
| @@ -146,7 +153,7 @@ xpc_setup_rsvd_page(void) | |||
| 146 | dev_err(xpc_part, "the reserved page's partid of %d is outside " | 153 | dev_err(xpc_part, "the reserved page's partid of %d is outside " |
| 147 | "supported range (< 0 || >= %d)\n", rp->SAL_partid, | 154 | "supported range (< 0 || >= %d)\n", rp->SAL_partid, |
| 148 | xp_max_npartitions); | 155 | xp_max_npartitions); |
| 149 | return NULL; | 156 | return -EINVAL; |
| 150 | } | 157 | } |
| 151 | 158 | ||
| 152 | rp->version = XPC_RP_VERSION; | 159 | rp->version = XPC_RP_VERSION; |
| @@ -165,8 +172,9 @@ xpc_setup_rsvd_page(void) | |||
| 165 | xpc_part_nasids = XPC_RP_PART_NASIDS(rp); | 172 | xpc_part_nasids = XPC_RP_PART_NASIDS(rp); |
| 166 | xpc_mach_nasids = XPC_RP_MACH_NASIDS(rp); | 173 | xpc_mach_nasids = XPC_RP_MACH_NASIDS(rp); |
| 167 | 174 | ||
| 168 | if (xpc_rsvd_page_init(rp) != xpSuccess) | 175 | ret = xpc_setup_rsvd_page_sn(rp); |
| 169 | return NULL; | 176 | if (ret != 0) |
| 177 | return ret; | ||
| 170 | 178 | ||
| 171 | /* | 179 | /* |
| 172 | * Set timestamp of when reserved page was setup by XPC. | 180 | * Set timestamp of when reserved page was setup by XPC. |
| @@ -178,7 +186,15 @@ xpc_setup_rsvd_page(void) | |||
| 178 | new_ts_jiffies++; | 186 | new_ts_jiffies++; |
| 179 | rp->ts_jiffies = new_ts_jiffies; | 187 | rp->ts_jiffies = new_ts_jiffies; |
| 180 | 188 | ||
| 181 | return rp; | 189 | xpc_rsvd_page = rp; |
| 190 | return 0; | ||
| 191 | } | ||
| 192 | |||
| 193 | void | ||
| 194 | xpc_teardown_rsvd_page(void) | ||
| 195 | { | ||
| 196 | /* a zero timestamp indicates our rsvd page is not initialized */ | ||
| 197 | xpc_rsvd_page->ts_jiffies = 0; | ||
| 182 | } | 198 | } |
| 183 | 199 | ||
| 184 | /* | 200 | /* |
diff --git a/drivers/misc/sgi-xp/xpc_sn2.c b/drivers/misc/sgi-xp/xpc_sn2.c index d1ccadc0857d..8b4b0653d9e9 100644 --- a/drivers/misc/sgi-xp/xpc_sn2.c +++ b/drivers/misc/sgi-xp/xpc_sn2.c | |||
| @@ -53,12 +53,19 @@ | |||
| 53 | * Buffer used to store a local copy of portions of a remote partition's | 53 | * Buffer used to store a local copy of portions of a remote partition's |
| 54 | * reserved page (either its header and part_nasids mask, or its vars). | 54 | * reserved page (either its header and part_nasids mask, or its vars). |
| 55 | */ | 55 | */ |
| 56 | static char *xpc_remote_copy_buffer_sn2; | ||
| 57 | static void *xpc_remote_copy_buffer_base_sn2; | 56 | static void *xpc_remote_copy_buffer_base_sn2; |
| 57 | static char *xpc_remote_copy_buffer_sn2; | ||
| 58 | 58 | ||
| 59 | static struct xpc_vars_sn2 *xpc_vars_sn2; | 59 | static struct xpc_vars_sn2 *xpc_vars_sn2; |
| 60 | static struct xpc_vars_part_sn2 *xpc_vars_part_sn2; | 60 | static struct xpc_vars_part_sn2 *xpc_vars_part_sn2; |
| 61 | 61 | ||
| 62 | static int | ||
| 63 | xpc_setup_partitions_sn_sn2(void) | ||
| 64 | { | ||
| 65 | /* nothing needs to be done */ | ||
| 66 | return 0; | ||
| 67 | } | ||
| 68 | |||
| 62 | /* SH_IPI_ACCESS shub register value on startup */ | 69 | /* SH_IPI_ACCESS shub register value on startup */ |
| 63 | static u64 xpc_sh1_IPI_access_sn2; | 70 | static u64 xpc_sh1_IPI_access_sn2; |
| 64 | static u64 xpc_sh2_IPI_access0_sn2; | 71 | static u64 xpc_sh2_IPI_access0_sn2; |
| @@ -198,7 +205,12 @@ xpc_init_IRQ_amo_sn2(int index) | |||
| 198 | static irqreturn_t | 205 | static irqreturn_t |
| 199 | xpc_handle_activate_IRQ_sn2(int irq, void *dev_id) | 206 | xpc_handle_activate_IRQ_sn2(int irq, void *dev_id) |
| 200 | { | 207 | { |
| 201 | atomic_inc(&xpc_activate_IRQ_rcvd); | 208 | unsigned long irq_flags; |
| 209 | |||
| 210 | spin_lock_irqsave(&xpc_activate_IRQ_rcvd_lock, irq_flags); | ||
| 211 | xpc_activate_IRQ_rcvd++; | ||
| 212 | spin_unlock_irqrestore(&xpc_activate_IRQ_rcvd_lock, irq_flags); | ||
| 213 | |||
| 202 | wake_up_interruptible(&xpc_activate_IRQ_wq); | 214 | wake_up_interruptible(&xpc_activate_IRQ_wq); |
| 203 | return IRQ_HANDLED; | 215 | return IRQ_HANDLED; |
| 204 | } | 216 | } |
| @@ -222,6 +234,7 @@ xpc_send_activate_IRQ_sn2(unsigned long amos_page_pa, int from_nasid, | |||
| 222 | static void | 234 | static void |
| 223 | xpc_send_local_activate_IRQ_sn2(int from_nasid) | 235 | xpc_send_local_activate_IRQ_sn2(int from_nasid) |
| 224 | { | 236 | { |
| 237 | unsigned long irq_flags; | ||
| 225 | struct amo *amos = (struct amo *)__va(xpc_vars_sn2->amos_page_pa + | 238 | struct amo *amos = (struct amo *)__va(xpc_vars_sn2->amos_page_pa + |
| 226 | (XPC_ACTIVATE_IRQ_AMOS_SN2 * | 239 | (XPC_ACTIVATE_IRQ_AMOS_SN2 * |
| 227 | sizeof(struct amo))); | 240 | sizeof(struct amo))); |
| @@ -230,7 +243,10 @@ xpc_send_local_activate_IRQ_sn2(int from_nasid) | |||
| 230 | FETCHOP_STORE_OP(TO_AMO((u64)&amos[BIT_WORD(from_nasid / 2)].variable), | 243 | FETCHOP_STORE_OP(TO_AMO((u64)&amos[BIT_WORD(from_nasid / 2)].variable), |
| 231 | FETCHOP_OR, BIT_MASK(from_nasid / 2)); | 244 | FETCHOP_OR, BIT_MASK(from_nasid / 2)); |
| 232 | 245 | ||
| 233 | atomic_inc(&xpc_activate_IRQ_rcvd); | 246 | spin_lock_irqsave(&xpc_activate_IRQ_rcvd_lock, irq_flags); |
| 247 | xpc_activate_IRQ_rcvd++; | ||
| 248 | spin_unlock_irqrestore(&xpc_activate_IRQ_rcvd_lock, irq_flags); | ||
| 249 | |||
| 234 | wake_up_interruptible(&xpc_activate_IRQ_wq); | 250 | wake_up_interruptible(&xpc_activate_IRQ_wq); |
| 235 | } | 251 | } |
| 236 | 252 | ||
| @@ -375,7 +391,7 @@ static void | |||
| 375 | xpc_send_chctl_closerequest_sn2(struct xpc_channel *ch, | 391 | xpc_send_chctl_closerequest_sn2(struct xpc_channel *ch, |
| 376 | unsigned long *irq_flags) | 392 | unsigned long *irq_flags) |
| 377 | { | 393 | { |
| 378 | struct xpc_openclose_args *args = ch->local_openclose_args; | 394 | struct xpc_openclose_args *args = ch->sn.sn2.local_openclose_args; |
| 379 | 395 | ||
| 380 | args->reason = ch->reason; | 396 | args->reason = ch->reason; |
| 381 | XPC_SEND_NOTIFY_IRQ_SN2(ch, XPC_CHCTL_CLOSEREQUEST, irq_flags); | 397 | XPC_SEND_NOTIFY_IRQ_SN2(ch, XPC_CHCTL_CLOSEREQUEST, irq_flags); |
| @@ -390,7 +406,7 @@ xpc_send_chctl_closereply_sn2(struct xpc_channel *ch, unsigned long *irq_flags) | |||
| 390 | static void | 406 | static void |
| 391 | xpc_send_chctl_openrequest_sn2(struct xpc_channel *ch, unsigned long *irq_flags) | 407 | xpc_send_chctl_openrequest_sn2(struct xpc_channel *ch, unsigned long *irq_flags) |
| 392 | { | 408 | { |
| 393 | struct xpc_openclose_args *args = ch->local_openclose_args; | 409 | struct xpc_openclose_args *args = ch->sn.sn2.local_openclose_args; |
| 394 | 410 | ||
| 395 | args->msg_size = ch->msg_size; | 411 | args->msg_size = ch->msg_size; |
| 396 | args->local_nentries = ch->local_nentries; | 412 | args->local_nentries = ch->local_nentries; |
| @@ -400,11 +416,11 @@ xpc_send_chctl_openrequest_sn2(struct xpc_channel *ch, unsigned long *irq_flags) | |||
| 400 | static void | 416 | static void |
| 401 | xpc_send_chctl_openreply_sn2(struct xpc_channel *ch, unsigned long *irq_flags) | 417 | xpc_send_chctl_openreply_sn2(struct xpc_channel *ch, unsigned long *irq_flags) |
| 402 | { | 418 | { |
| 403 | struct xpc_openclose_args *args = ch->local_openclose_args; | 419 | struct xpc_openclose_args *args = ch->sn.sn2.local_openclose_args; |
| 404 | 420 | ||
| 405 | args->remote_nentries = ch->remote_nentries; | 421 | args->remote_nentries = ch->remote_nentries; |
| 406 | args->local_nentries = ch->local_nentries; | 422 | args->local_nentries = ch->local_nentries; |
| 407 | args->local_msgqueue_pa = xp_pa(ch->local_msgqueue); | 423 | args->local_msgqueue_pa = xp_pa(ch->sn.sn2.local_msgqueue); |
| 408 | XPC_SEND_NOTIFY_IRQ_SN2(ch, XPC_CHCTL_OPENREPLY, irq_flags); | 424 | XPC_SEND_NOTIFY_IRQ_SN2(ch, XPC_CHCTL_OPENREPLY, irq_flags); |
| 409 | } | 425 | } |
| 410 | 426 | ||
| @@ -420,6 +436,13 @@ xpc_send_chctl_local_msgrequest_sn2(struct xpc_channel *ch) | |||
| 420 | XPC_SEND_LOCAL_NOTIFY_IRQ_SN2(ch, XPC_CHCTL_MSGREQUEST); | 436 | XPC_SEND_LOCAL_NOTIFY_IRQ_SN2(ch, XPC_CHCTL_MSGREQUEST); |
| 421 | } | 437 | } |
| 422 | 438 | ||
| 439 | static void | ||
| 440 | xpc_save_remote_msgqueue_pa_sn2(struct xpc_channel *ch, | ||
| 441 | unsigned long msgqueue_pa) | ||
| 442 | { | ||
| 443 | ch->sn.sn2.remote_msgqueue_pa = msgqueue_pa; | ||
| 444 | } | ||
| 445 | |||
| 423 | /* | 446 | /* |
| 424 | * This next set of functions are used to keep track of when a partition is | 447 | * This next set of functions are used to keep track of when a partition is |
| 425 | * potentially engaged in accessing memory belonging to another partition. | 448 | * potentially engaged in accessing memory belonging to another partition. |
| @@ -489,6 +512,17 @@ xpc_indicate_partition_disengaged_sn2(struct xpc_partition *part) | |||
| 489 | part_sn2->activate_IRQ_phys_cpuid); | 512 | part_sn2->activate_IRQ_phys_cpuid); |
| 490 | } | 513 | } |
| 491 | 514 | ||
| 515 | static void | ||
| 516 | xpc_assume_partition_disengaged_sn2(short partid) | ||
| 517 | { | ||
| 518 | struct amo *amo = xpc_vars_sn2->amos_page + | ||
| 519 | XPC_ENGAGED_PARTITIONS_AMO_SN2; | ||
| 520 | |||
| 521 | /* clear bit(s) based on partid mask in our partition's amo */ | ||
| 522 | FETCHOP_STORE_OP(TO_AMO((u64)&amo->variable), FETCHOP_AND, | ||
| 523 | ~BIT(partid)); | ||
| 524 | } | ||
| 525 | |||
| 492 | static int | 526 | static int |
| 493 | xpc_partition_engaged_sn2(short partid) | 527 | xpc_partition_engaged_sn2(short partid) |
| 494 | { | 528 | { |
| @@ -510,17 +544,6 @@ xpc_any_partition_engaged_sn2(void) | |||
| 510 | return FETCHOP_LOAD_OP(TO_AMO((u64)&amo->variable), FETCHOP_LOAD) != 0; | 544 | return FETCHOP_LOAD_OP(TO_AMO((u64)&amo->variable), FETCHOP_LOAD) != 0; |
| 511 | } | 545 | } |
| 512 | 546 | ||
| 513 | static void | ||
| 514 | xpc_assume_partition_disengaged_sn2(short partid) | ||
| 515 | { | ||
| 516 | struct amo *amo = xpc_vars_sn2->amos_page + | ||
| 517 | XPC_ENGAGED_PARTITIONS_AMO_SN2; | ||
| 518 | |||
| 519 | /* clear bit(s) based on partid mask in our partition's amo */ | ||
| 520 | FETCHOP_STORE_OP(TO_AMO((u64)&amo->variable), FETCHOP_AND, | ||
| 521 | ~BIT(partid)); | ||
| 522 | } | ||
| 523 | |||
| 524 | /* original protection values for each node */ | 547 | /* original protection values for each node */ |
| 525 | static u64 xpc_prot_vec_sn2[MAX_NUMNODES]; | 548 | static u64 xpc_prot_vec_sn2[MAX_NUMNODES]; |
| 526 | 549 | ||
| @@ -595,8 +618,8 @@ xpc_get_partition_rsvd_page_pa_sn2(void *buf, u64 *cookie, unsigned long *rp_pa, | |||
| 595 | } | 618 | } |
| 596 | 619 | ||
| 597 | 620 | ||
| 598 | static enum xp_retval | 621 | static int |
| 599 | xpc_rsvd_page_init_sn2(struct xpc_rsvd_page *rp) | 622 | xpc_setup_rsvd_page_sn_sn2(struct xpc_rsvd_page *rp) |
| 600 | { | 623 | { |
| 601 | struct amo *amos_page; | 624 | struct amo *amos_page; |
| 602 | int i; | 625 | int i; |
| @@ -627,7 +650,7 @@ xpc_rsvd_page_init_sn2(struct xpc_rsvd_page *rp) | |||
| 627 | amos_page = (struct amo *)TO_AMO(uncached_alloc_page(0, 1)); | 650 | amos_page = (struct amo *)TO_AMO(uncached_alloc_page(0, 1)); |
| 628 | if (amos_page == NULL) { | 651 | if (amos_page == NULL) { |
| 629 | dev_err(xpc_part, "can't allocate page of amos\n"); | 652 | dev_err(xpc_part, "can't allocate page of amos\n"); |
| 630 | return xpNoMemory; | 653 | return -ENOMEM; |
| 631 | } | 654 | } |
| 632 | 655 | ||
| 633 | /* | 656 | /* |
| @@ -639,7 +662,7 @@ xpc_rsvd_page_init_sn2(struct xpc_rsvd_page *rp) | |||
| 639 | dev_err(xpc_part, "can't allow amo operations\n"); | 662 | dev_err(xpc_part, "can't allow amo operations\n"); |
| 640 | uncached_free_page(__IA64_UNCACHED_OFFSET | | 663 | uncached_free_page(__IA64_UNCACHED_OFFSET | |
| 641 | TO_PHYS((u64)amos_page), 1); | 664 | TO_PHYS((u64)amos_page), 1); |
| 642 | return ret; | 665 | return -EPERM; |
| 643 | } | 666 | } |
| 644 | } | 667 | } |
| 645 | 668 | ||
| @@ -665,7 +688,7 @@ xpc_rsvd_page_init_sn2(struct xpc_rsvd_page *rp) | |||
| 665 | (void)xpc_init_IRQ_amo_sn2(XPC_ENGAGED_PARTITIONS_AMO_SN2); | 688 | (void)xpc_init_IRQ_amo_sn2(XPC_ENGAGED_PARTITIONS_AMO_SN2); |
| 666 | (void)xpc_init_IRQ_amo_sn2(XPC_DEACTIVATE_REQUEST_AMO_SN2); | 689 | (void)xpc_init_IRQ_amo_sn2(XPC_DEACTIVATE_REQUEST_AMO_SN2); |
| 667 | 690 | ||
| 668 | return xpSuccess; | 691 | return 0; |
| 669 | } | 692 | } |
| 670 | 693 | ||
| 671 | static void | 694 | static void |
| @@ -1082,10 +1105,19 @@ xpc_identify_activate_IRQ_sender_sn2(void) | |||
| 1082 | } | 1105 | } |
| 1083 | 1106 | ||
| 1084 | static void | 1107 | static void |
| 1085 | xpc_process_activate_IRQ_rcvd_sn2(int n_IRQs_expected) | 1108 | xpc_process_activate_IRQ_rcvd_sn2(void) |
| 1086 | { | 1109 | { |
| 1110 | unsigned long irq_flags; | ||
| 1111 | int n_IRQs_expected; | ||
| 1087 | int n_IRQs_detected; | 1112 | int n_IRQs_detected; |
| 1088 | 1113 | ||
| 1114 | DBUG_ON(xpc_activate_IRQ_rcvd == 0); | ||
| 1115 | |||
| 1116 | spin_lock_irqsave(&xpc_activate_IRQ_rcvd_lock, irq_flags); | ||
| 1117 | n_IRQs_expected = xpc_activate_IRQ_rcvd; | ||
| 1118 | xpc_activate_IRQ_rcvd = 0; | ||
| 1119 | spin_unlock_irqrestore(&xpc_activate_IRQ_rcvd_lock, irq_flags); | ||
| 1120 | |||
| 1089 | n_IRQs_detected = xpc_identify_activate_IRQ_sender_sn2(); | 1121 | n_IRQs_detected = xpc_identify_activate_IRQ_sender_sn2(); |
| 1090 | if (n_IRQs_detected < n_IRQs_expected) { | 1122 | if (n_IRQs_detected < n_IRQs_expected) { |
| 1091 | /* retry once to help avoid missing amo */ | 1123 | /* retry once to help avoid missing amo */ |
| @@ -1094,116 +1126,63 @@ xpc_process_activate_IRQ_rcvd_sn2(int n_IRQs_expected) | |||
| 1094 | } | 1126 | } |
| 1095 | 1127 | ||
| 1096 | /* | 1128 | /* |
| 1097 | * Guarantee that the kzalloc'd memory is cacheline aligned. | 1129 | * Setup the channel structures that are sn2 specific. |
| 1098 | */ | ||
| 1099 | static void * | ||
| 1100 | xpc_kzalloc_cacheline_aligned_sn2(size_t size, gfp_t flags, void **base) | ||
| 1101 | { | ||
| 1102 | /* see if kzalloc will give us cachline aligned memory by default */ | ||
| 1103 | *base = kzalloc(size, flags); | ||
| 1104 | if (*base == NULL) | ||
| 1105 | return NULL; | ||
| 1106 | |||
| 1107 | if ((u64)*base == L1_CACHE_ALIGN((u64)*base)) | ||
| 1108 | return *base; | ||
| 1109 | |||
| 1110 | kfree(*base); | ||
| 1111 | |||
| 1112 | /* nope, we'll have to do it ourselves */ | ||
| 1113 | *base = kzalloc(size + L1_CACHE_BYTES, flags); | ||
| 1114 | if (*base == NULL) | ||
| 1115 | return NULL; | ||
| 1116 | |||
| 1117 | return (void *)L1_CACHE_ALIGN((u64)*base); | ||
| 1118 | } | ||
| 1119 | |||
| 1120 | /* | ||
| 1121 | * Setup the infrastructure necessary to support XPartition Communication | ||
| 1122 | * between the specified remote partition and the local one. | ||
| 1123 | */ | 1130 | */ |
| 1124 | static enum xp_retval | 1131 | static enum xp_retval |
| 1125 | xpc_setup_infrastructure_sn2(struct xpc_partition *part) | 1132 | xpc_setup_ch_structures_sn_sn2(struct xpc_partition *part) |
| 1126 | { | 1133 | { |
| 1127 | struct xpc_partition_sn2 *part_sn2 = &part->sn.sn2; | 1134 | struct xpc_partition_sn2 *part_sn2 = &part->sn.sn2; |
| 1135 | struct xpc_channel_sn2 *ch_sn2; | ||
| 1128 | enum xp_retval retval; | 1136 | enum xp_retval retval; |
| 1129 | int ret; | 1137 | int ret; |
| 1130 | int cpuid; | 1138 | int cpuid; |
| 1131 | int ch_number; | 1139 | int ch_number; |
| 1132 | struct xpc_channel *ch; | ||
| 1133 | struct timer_list *timer; | 1140 | struct timer_list *timer; |
| 1134 | short partid = XPC_PARTID(part); | 1141 | short partid = XPC_PARTID(part); |
| 1135 | 1142 | ||
| 1136 | /* | ||
| 1137 | * Allocate all of the channel structures as a contiguous chunk of | ||
| 1138 | * memory. | ||
| 1139 | */ | ||
| 1140 | DBUG_ON(part->channels != NULL); | ||
| 1141 | part->channels = kzalloc(sizeof(struct xpc_channel) * XPC_MAX_NCHANNELS, | ||
| 1142 | GFP_KERNEL); | ||
| 1143 | if (part->channels == NULL) { | ||
| 1144 | dev_err(xpc_chan, "can't get memory for channels\n"); | ||
| 1145 | return xpNoMemory; | ||
| 1146 | } | ||
| 1147 | |||
| 1148 | /* allocate all the required GET/PUT values */ | 1143 | /* allocate all the required GET/PUT values */ |
| 1149 | 1144 | ||
| 1150 | part_sn2->local_GPs = | 1145 | part_sn2->local_GPs = |
| 1151 | xpc_kzalloc_cacheline_aligned_sn2(XPC_GP_SIZE, GFP_KERNEL, | 1146 | xpc_kzalloc_cacheline_aligned(XPC_GP_SIZE, GFP_KERNEL, |
| 1152 | &part_sn2->local_GPs_base); | 1147 | &part_sn2->local_GPs_base); |
| 1153 | if (part_sn2->local_GPs == NULL) { | 1148 | if (part_sn2->local_GPs == NULL) { |
| 1154 | dev_err(xpc_chan, "can't get memory for local get/put " | 1149 | dev_err(xpc_chan, "can't get memory for local get/put " |
| 1155 | "values\n"); | 1150 | "values\n"); |
| 1156 | retval = xpNoMemory; | 1151 | return xpNoMemory; |
| 1157 | goto out_1; | ||
| 1158 | } | 1152 | } |
| 1159 | 1153 | ||
| 1160 | part_sn2->remote_GPs = | 1154 | part_sn2->remote_GPs = |
| 1161 | xpc_kzalloc_cacheline_aligned_sn2(XPC_GP_SIZE, GFP_KERNEL, | 1155 | xpc_kzalloc_cacheline_aligned(XPC_GP_SIZE, GFP_KERNEL, |
| 1162 | &part_sn2->remote_GPs_base); | 1156 | &part_sn2->remote_GPs_base); |
| 1163 | if (part_sn2->remote_GPs == NULL) { | 1157 | if (part_sn2->remote_GPs == NULL) { |
| 1164 | dev_err(xpc_chan, "can't get memory for remote get/put " | 1158 | dev_err(xpc_chan, "can't get memory for remote get/put " |
| 1165 | "values\n"); | 1159 | "values\n"); |
| 1166 | retval = xpNoMemory; | 1160 | retval = xpNoMemory; |
| 1167 | goto out_2; | 1161 | goto out_1; |
| 1168 | } | 1162 | } |
| 1169 | 1163 | ||
| 1170 | part_sn2->remote_GPs_pa = 0; | 1164 | part_sn2->remote_GPs_pa = 0; |
| 1171 | 1165 | ||
| 1172 | /* allocate all the required open and close args */ | 1166 | /* allocate all the required open and close args */ |
| 1173 | 1167 | ||
| 1174 | part->local_openclose_args = | 1168 | part_sn2->local_openclose_args = |
| 1175 | xpc_kzalloc_cacheline_aligned_sn2(XPC_OPENCLOSE_ARGS_SIZE, | 1169 | xpc_kzalloc_cacheline_aligned(XPC_OPENCLOSE_ARGS_SIZE, |
| 1176 | GFP_KERNEL, | 1170 | GFP_KERNEL, &part_sn2-> |
| 1177 | &part->local_openclose_args_base); | 1171 | local_openclose_args_base); |
| 1178 | if (part->local_openclose_args == NULL) { | 1172 | if (part_sn2->local_openclose_args == NULL) { |
| 1179 | dev_err(xpc_chan, "can't get memory for local connect args\n"); | 1173 | dev_err(xpc_chan, "can't get memory for local connect args\n"); |
| 1180 | retval = xpNoMemory; | 1174 | retval = xpNoMemory; |
| 1181 | goto out_3; | 1175 | goto out_2; |
| 1182 | } | ||
| 1183 | |||
| 1184 | part->remote_openclose_args = | ||
| 1185 | xpc_kzalloc_cacheline_aligned_sn2(XPC_OPENCLOSE_ARGS_SIZE, | ||
| 1186 | GFP_KERNEL, | ||
| 1187 | &part->remote_openclose_args_base); | ||
| 1188 | if (part->remote_openclose_args == NULL) { | ||
| 1189 | dev_err(xpc_chan, "can't get memory for remote connect args\n"); | ||
| 1190 | retval = xpNoMemory; | ||
| 1191 | goto out_4; | ||
| 1192 | } | 1176 | } |
| 1193 | 1177 | ||
| 1194 | part_sn2->remote_openclose_args_pa = 0; | 1178 | part_sn2->remote_openclose_args_pa = 0; |
| 1195 | 1179 | ||
| 1196 | part_sn2->local_chctl_amo_va = xpc_init_IRQ_amo_sn2(partid); | 1180 | part_sn2->local_chctl_amo_va = xpc_init_IRQ_amo_sn2(partid); |
| 1197 | part->chctl.all_flags = 0; | ||
| 1198 | spin_lock_init(&part->chctl_lock); | ||
| 1199 | 1181 | ||
| 1200 | part_sn2->notify_IRQ_nasid = 0; | 1182 | part_sn2->notify_IRQ_nasid = 0; |
| 1201 | part_sn2->notify_IRQ_phys_cpuid = 0; | 1183 | part_sn2->notify_IRQ_phys_cpuid = 0; |
| 1202 | part_sn2->remote_chctl_amo_va = NULL; | 1184 | part_sn2->remote_chctl_amo_va = NULL; |
| 1203 | 1185 | ||
| 1204 | atomic_set(&part->channel_mgr_requests, 1); | ||
| 1205 | init_waitqueue_head(&part->channel_mgr_wq); | ||
| 1206 | |||
| 1207 | sprintf(part_sn2->notify_IRQ_owner, "xpc%02d", partid); | 1186 | sprintf(part_sn2->notify_IRQ_owner, "xpc%02d", partid); |
| 1208 | ret = request_irq(SGI_XPC_NOTIFY, xpc_handle_notify_IRQ_sn2, | 1187 | ret = request_irq(SGI_XPC_NOTIFY, xpc_handle_notify_IRQ_sn2, |
| 1209 | IRQF_SHARED, part_sn2->notify_IRQ_owner, | 1188 | IRQF_SHARED, part_sn2->notify_IRQ_owner, |
| @@ -1212,7 +1191,7 @@ xpc_setup_infrastructure_sn2(struct xpc_partition *part) | |||
| 1212 | dev_err(xpc_chan, "can't register NOTIFY IRQ handler, " | 1191 | dev_err(xpc_chan, "can't register NOTIFY IRQ handler, " |
| 1213 | "errno=%d\n", -ret); | 1192 | "errno=%d\n", -ret); |
| 1214 | retval = xpLackOfResources; | 1193 | retval = xpLackOfResources; |
| 1215 | goto out_5; | 1194 | goto out_3; |
| 1216 | } | 1195 | } |
| 1217 | 1196 | ||
| 1218 | /* Setup a timer to check for dropped notify IRQs */ | 1197 | /* Setup a timer to check for dropped notify IRQs */ |
| @@ -1224,45 +1203,17 @@ xpc_setup_infrastructure_sn2(struct xpc_partition *part) | |||
| 1224 | timer->expires = jiffies + XPC_DROPPED_NOTIFY_IRQ_WAIT_INTERVAL; | 1203 | timer->expires = jiffies + XPC_DROPPED_NOTIFY_IRQ_WAIT_INTERVAL; |
| 1225 | add_timer(timer); | 1204 | add_timer(timer); |
| 1226 | 1205 | ||
| 1227 | part->nchannels = XPC_MAX_NCHANNELS; | ||
| 1228 | |||
| 1229 | atomic_set(&part->nchannels_active, 0); | ||
| 1230 | atomic_set(&part->nchannels_engaged, 0); | ||
| 1231 | |||
| 1232 | for (ch_number = 0; ch_number < part->nchannels; ch_number++) { | 1206 | for (ch_number = 0; ch_number < part->nchannels; ch_number++) { |
| 1233 | ch = &part->channels[ch_number]; | 1207 | ch_sn2 = &part->channels[ch_number].sn.sn2; |
| 1234 | |||
| 1235 | ch->partid = partid; | ||
| 1236 | ch->number = ch_number; | ||
| 1237 | ch->flags = XPC_C_DISCONNECTED; | ||
| 1238 | |||
| 1239 | ch->sn.sn2.local_GP = &part_sn2->local_GPs[ch_number]; | ||
| 1240 | ch->local_openclose_args = | ||
| 1241 | &part->local_openclose_args[ch_number]; | ||
| 1242 | |||
| 1243 | atomic_set(&ch->kthreads_assigned, 0); | ||
| 1244 | atomic_set(&ch->kthreads_idle, 0); | ||
| 1245 | atomic_set(&ch->kthreads_active, 0); | ||
| 1246 | 1208 | ||
| 1247 | atomic_set(&ch->references, 0); | 1209 | ch_sn2->local_GP = &part_sn2->local_GPs[ch_number]; |
| 1248 | atomic_set(&ch->n_to_notify, 0); | 1210 | ch_sn2->local_openclose_args = |
| 1211 | &part_sn2->local_openclose_args[ch_number]; | ||
| 1249 | 1212 | ||
| 1250 | spin_lock_init(&ch->lock); | 1213 | mutex_init(&ch_sn2->msg_to_pull_mutex); |
| 1251 | mutex_init(&ch->sn.sn2.msg_to_pull_mutex); | ||
| 1252 | init_completion(&ch->wdisconnect_wait); | ||
| 1253 | |||
| 1254 | atomic_set(&ch->n_on_msg_allocate_wq, 0); | ||
| 1255 | init_waitqueue_head(&ch->msg_allocate_wq); | ||
| 1256 | init_waitqueue_head(&ch->idle_wq); | ||
| 1257 | } | 1214 | } |
| 1258 | 1215 | ||
| 1259 | /* | 1216 | /* |
| 1260 | * With the setting of the partition setup_state to XPC_P_SS_SETUP, | ||
| 1261 | * we're declaring that this partition is ready to go. | ||
| 1262 | */ | ||
| 1263 | part->setup_state = XPC_P_SS_SETUP; | ||
| 1264 | |||
| 1265 | /* | ||
| 1266 | * Setup the per partition specific variables required by the | 1217 | * Setup the per partition specific variables required by the |
| 1267 | * remote partition to establish channel connections with us. | 1218 | * remote partition to establish channel connections with us. |
| 1268 | * | 1219 | * |
| @@ -1271,7 +1222,7 @@ xpc_setup_infrastructure_sn2(struct xpc_partition *part) | |||
| 1271 | */ | 1222 | */ |
| 1272 | xpc_vars_part_sn2[partid].GPs_pa = xp_pa(part_sn2->local_GPs); | 1223 | xpc_vars_part_sn2[partid].GPs_pa = xp_pa(part_sn2->local_GPs); |
| 1273 | xpc_vars_part_sn2[partid].openclose_args_pa = | 1224 | xpc_vars_part_sn2[partid].openclose_args_pa = |
| 1274 | xp_pa(part->local_openclose_args); | 1225 | xp_pa(part_sn2->local_openclose_args); |
| 1275 | xpc_vars_part_sn2[partid].chctl_amo_pa = | 1226 | xpc_vars_part_sn2[partid].chctl_amo_pa = |
| 1276 | xp_pa(part_sn2->local_chctl_amo_va); | 1227 | xp_pa(part_sn2->local_chctl_amo_va); |
| 1277 | cpuid = raw_smp_processor_id(); /* any CPU in this partition will do */ | 1228 | cpuid = raw_smp_processor_id(); /* any CPU in this partition will do */ |
| @@ -1279,80 +1230,48 @@ xpc_setup_infrastructure_sn2(struct xpc_partition *part) | |||
| 1279 | xpc_vars_part_sn2[partid].notify_IRQ_phys_cpuid = | 1230 | xpc_vars_part_sn2[partid].notify_IRQ_phys_cpuid = |
| 1280 | cpu_physical_id(cpuid); | 1231 | cpu_physical_id(cpuid); |
| 1281 | xpc_vars_part_sn2[partid].nchannels = part->nchannels; | 1232 | xpc_vars_part_sn2[partid].nchannels = part->nchannels; |
| 1282 | xpc_vars_part_sn2[partid].magic = XPC_VP_MAGIC1; | 1233 | xpc_vars_part_sn2[partid].magic = XPC_VP_MAGIC1_SN2; |
| 1283 | 1234 | ||
| 1284 | return xpSuccess; | 1235 | return xpSuccess; |
| 1285 | 1236 | ||
| 1286 | /* setup of infrastructure failed */ | 1237 | /* setup of ch structures failed */ |
| 1287 | out_5: | ||
| 1288 | kfree(part->remote_openclose_args_base); | ||
| 1289 | part->remote_openclose_args = NULL; | ||
| 1290 | out_4: | ||
| 1291 | kfree(part->local_openclose_args_base); | ||
| 1292 | part->local_openclose_args = NULL; | ||
| 1293 | out_3: | 1238 | out_3: |
| 1239 | kfree(part_sn2->local_openclose_args_base); | ||
| 1240 | part_sn2->local_openclose_args = NULL; | ||
| 1241 | out_2: | ||
| 1294 | kfree(part_sn2->remote_GPs_base); | 1242 | kfree(part_sn2->remote_GPs_base); |
| 1295 | part_sn2->remote_GPs = NULL; | 1243 | part_sn2->remote_GPs = NULL; |
| 1296 | out_2: | 1244 | out_1: |
| 1297 | kfree(part_sn2->local_GPs_base); | 1245 | kfree(part_sn2->local_GPs_base); |
| 1298 | part_sn2->local_GPs = NULL; | 1246 | part_sn2->local_GPs = NULL; |
| 1299 | out_1: | ||
| 1300 | kfree(part->channels); | ||
| 1301 | part->channels = NULL; | ||
| 1302 | return retval; | 1247 | return retval; |
| 1303 | } | 1248 | } |
| 1304 | 1249 | ||
| 1305 | /* | 1250 | /* |
| 1306 | * Teardown the infrastructure necessary to support XPartition Communication | 1251 | * Teardown the channel structures that are sn2 specific. |
| 1307 | * between the specified remote partition and the local one. | ||
| 1308 | */ | 1252 | */ |
| 1309 | static void | 1253 | static void |
| 1310 | xpc_teardown_infrastructure_sn2(struct xpc_partition *part) | 1254 | xpc_teardown_ch_structures_sn_sn2(struct xpc_partition *part) |
| 1311 | { | 1255 | { |
| 1312 | struct xpc_partition_sn2 *part_sn2 = &part->sn.sn2; | 1256 | struct xpc_partition_sn2 *part_sn2 = &part->sn.sn2; |
| 1313 | short partid = XPC_PARTID(part); | 1257 | short partid = XPC_PARTID(part); |
| 1314 | 1258 | ||
| 1315 | /* | 1259 | /* |
| 1316 | * We start off by making this partition inaccessible to local | 1260 | * Indicate that the variables specific to the remote partition are no |
| 1317 | * processes by marking it as no longer setup. Then we make it | 1261 | * longer available for its use. |
| 1318 | * inaccessible to remote processes by clearing the XPC per partition | ||
| 1319 | * specific variable's magic # (which indicates that these variables | ||
| 1320 | * are no longer valid) and by ignoring all XPC notify IRQs sent to | ||
| 1321 | * this partition. | ||
| 1322 | */ | 1262 | */ |
| 1323 | |||
| 1324 | DBUG_ON(atomic_read(&part->nchannels_engaged) != 0); | ||
| 1325 | DBUG_ON(atomic_read(&part->nchannels_active) != 0); | ||
| 1326 | DBUG_ON(part->setup_state != XPC_P_SS_SETUP); | ||
| 1327 | part->setup_state = XPC_P_SS_WTEARDOWN; | ||
| 1328 | |||
| 1329 | xpc_vars_part_sn2[partid].magic = 0; | 1263 | xpc_vars_part_sn2[partid].magic = 0; |
| 1330 | 1264 | ||
| 1331 | free_irq(SGI_XPC_NOTIFY, (void *)(u64)partid); | ||
| 1332 | |||
| 1333 | /* | ||
| 1334 | * Before proceeding with the teardown we have to wait until all | ||
| 1335 | * existing references cease. | ||
| 1336 | */ | ||
| 1337 | wait_event(part->teardown_wq, (atomic_read(&part->references) == 0)); | ||
| 1338 | |||
| 1339 | /* now we can begin tearing down the infrastructure */ | ||
| 1340 | |||
| 1341 | part->setup_state = XPC_P_SS_TORNDOWN; | ||
| 1342 | |||
| 1343 | /* in case we've still got outstanding timers registered... */ | 1265 | /* in case we've still got outstanding timers registered... */ |
| 1344 | del_timer_sync(&part_sn2->dropped_notify_IRQ_timer); | 1266 | del_timer_sync(&part_sn2->dropped_notify_IRQ_timer); |
| 1267 | free_irq(SGI_XPC_NOTIFY, (void *)(u64)partid); | ||
| 1345 | 1268 | ||
| 1346 | kfree(part->remote_openclose_args_base); | 1269 | kfree(part_sn2->local_openclose_args_base); |
| 1347 | part->remote_openclose_args = NULL; | 1270 | part_sn2->local_openclose_args = NULL; |
| 1348 | kfree(part->local_openclose_args_base); | ||
| 1349 | part->local_openclose_args = NULL; | ||
| 1350 | kfree(part_sn2->remote_GPs_base); | 1271 | kfree(part_sn2->remote_GPs_base); |
| 1351 | part_sn2->remote_GPs = NULL; | 1272 | part_sn2->remote_GPs = NULL; |
| 1352 | kfree(part_sn2->local_GPs_base); | 1273 | kfree(part_sn2->local_GPs_base); |
| 1353 | part_sn2->local_GPs = NULL; | 1274 | part_sn2->local_GPs = NULL; |
| 1354 | kfree(part->channels); | ||
| 1355 | part->channels = NULL; | ||
| 1356 | part_sn2->local_chctl_amo_va = NULL; | 1275 | part_sn2->local_chctl_amo_va = NULL; |
| 1357 | } | 1276 | } |
| 1358 | 1277 | ||
| @@ -1429,8 +1348,8 @@ xpc_pull_remote_vars_part_sn2(struct xpc_partition *part) | |||
| 1429 | 1348 | ||
| 1430 | /* see if they've been set up yet */ | 1349 | /* see if they've been set up yet */ |
| 1431 | 1350 | ||
| 1432 | if (pulled_entry->magic != XPC_VP_MAGIC1 && | 1351 | if (pulled_entry->magic != XPC_VP_MAGIC1_SN2 && |
| 1433 | pulled_entry->magic != XPC_VP_MAGIC2) { | 1352 | pulled_entry->magic != XPC_VP_MAGIC2_SN2) { |
| 1434 | 1353 | ||
| 1435 | if (pulled_entry->magic != 0) { | 1354 | if (pulled_entry->magic != 0) { |
| 1436 | dev_dbg(xpc_chan, "partition %d's XPC vars_part for " | 1355 | dev_dbg(xpc_chan, "partition %d's XPC vars_part for " |
| @@ -1443,7 +1362,7 @@ xpc_pull_remote_vars_part_sn2(struct xpc_partition *part) | |||
| 1443 | return xpRetry; | 1362 | return xpRetry; |
| 1444 | } | 1363 | } |
| 1445 | 1364 | ||
| 1446 | if (xpc_vars_part_sn2[partid].magic == XPC_VP_MAGIC1) { | 1365 | if (xpc_vars_part_sn2[partid].magic == XPC_VP_MAGIC1_SN2) { |
| 1447 | 1366 | ||
| 1448 | /* validate the variables */ | 1367 | /* validate the variables */ |
| 1449 | 1368 | ||
| @@ -1473,10 +1392,10 @@ xpc_pull_remote_vars_part_sn2(struct xpc_partition *part) | |||
| 1473 | 1392 | ||
| 1474 | /* let the other side know that we've pulled their variables */ | 1393 | /* let the other side know that we've pulled their variables */ |
| 1475 | 1394 | ||
| 1476 | xpc_vars_part_sn2[partid].magic = XPC_VP_MAGIC2; | 1395 | xpc_vars_part_sn2[partid].magic = XPC_VP_MAGIC2_SN2; |
| 1477 | } | 1396 | } |
| 1478 | 1397 | ||
| 1479 | if (pulled_entry->magic == XPC_VP_MAGIC1) | 1398 | if (pulled_entry->magic == XPC_VP_MAGIC1_SN2) |
| 1480 | return xpRetry; | 1399 | return xpRetry; |
| 1481 | 1400 | ||
| 1482 | return xpSuccess; | 1401 | return xpSuccess; |
| @@ -1605,6 +1524,7 @@ xpc_get_chctl_all_flags_sn2(struct xpc_partition *part) | |||
| 1605 | static enum xp_retval | 1524 | static enum xp_retval |
| 1606 | xpc_allocate_local_msgqueue_sn2(struct xpc_channel *ch) | 1525 | xpc_allocate_local_msgqueue_sn2(struct xpc_channel *ch) |
| 1607 | { | 1526 | { |
| 1527 | struct xpc_channel_sn2 *ch_sn2 = &ch->sn.sn2; | ||
| 1608 | unsigned long irq_flags; | 1528 | unsigned long irq_flags; |
| 1609 | int nentries; | 1529 | int nentries; |
| 1610 | size_t nbytes; | 1530 | size_t nbytes; |
| @@ -1612,17 +1532,17 @@ xpc_allocate_local_msgqueue_sn2(struct xpc_channel *ch) | |||
| 1612 | for (nentries = ch->local_nentries; nentries > 0; nentries--) { | 1532 | for (nentries = ch->local_nentries; nentries > 0; nentries--) { |
| 1613 | 1533 | ||
| 1614 | nbytes = nentries * ch->msg_size; | 1534 | nbytes = nentries * ch->msg_size; |
| 1615 | ch->local_msgqueue = | 1535 | ch_sn2->local_msgqueue = |
| 1616 | xpc_kzalloc_cacheline_aligned_sn2(nbytes, GFP_KERNEL, | 1536 | xpc_kzalloc_cacheline_aligned(nbytes, GFP_KERNEL, |
| 1617 | &ch->local_msgqueue_base); | 1537 | &ch_sn2->local_msgqueue_base); |
| 1618 | if (ch->local_msgqueue == NULL) | 1538 | if (ch_sn2->local_msgqueue == NULL) |
| 1619 | continue; | 1539 | continue; |
| 1620 | 1540 | ||
| 1621 | nbytes = nentries * sizeof(struct xpc_notify); | 1541 | nbytes = nentries * sizeof(struct xpc_notify); |
| 1622 | ch->notify_queue = kzalloc(nbytes, GFP_KERNEL); | 1542 | ch_sn2->notify_queue = kzalloc(nbytes, GFP_KERNEL); |
| 1623 | if (ch->notify_queue == NULL) { | 1543 | if (ch_sn2->notify_queue == NULL) { |
| 1624 | kfree(ch->local_msgqueue_base); | 1544 | kfree(ch_sn2->local_msgqueue_base); |
| 1625 | ch->local_msgqueue = NULL; | 1545 | ch_sn2->local_msgqueue = NULL; |
| 1626 | continue; | 1546 | continue; |
| 1627 | } | 1547 | } |
| 1628 | 1548 | ||
| @@ -1649,6 +1569,7 @@ xpc_allocate_local_msgqueue_sn2(struct xpc_channel *ch) | |||
| 1649 | static enum xp_retval | 1569 | static enum xp_retval |
| 1650 | xpc_allocate_remote_msgqueue_sn2(struct xpc_channel *ch) | 1570 | xpc_allocate_remote_msgqueue_sn2(struct xpc_channel *ch) |
| 1651 | { | 1571 | { |
| 1572 | struct xpc_channel_sn2 *ch_sn2 = &ch->sn.sn2; | ||
| 1652 | unsigned long irq_flags; | 1573 | unsigned long irq_flags; |
| 1653 | int nentries; | 1574 | int nentries; |
| 1654 | size_t nbytes; | 1575 | size_t nbytes; |
| @@ -1658,10 +1579,10 @@ xpc_allocate_remote_msgqueue_sn2(struct xpc_channel *ch) | |||
| 1658 | for (nentries = ch->remote_nentries; nentries > 0; nentries--) { | 1579 | for (nentries = ch->remote_nentries; nentries > 0; nentries--) { |
| 1659 | 1580 | ||
| 1660 | nbytes = nentries * ch->msg_size; | 1581 | nbytes = nentries * ch->msg_size; |
| 1661 | ch->remote_msgqueue = | 1582 | ch_sn2->remote_msgqueue = |
| 1662 | xpc_kzalloc_cacheline_aligned_sn2(nbytes, GFP_KERNEL, | 1583 | xpc_kzalloc_cacheline_aligned(nbytes, GFP_KERNEL, &ch_sn2-> |
| 1663 | &ch->remote_msgqueue_base); | 1584 | remote_msgqueue_base); |
| 1664 | if (ch->remote_msgqueue == NULL) | 1585 | if (ch_sn2->remote_msgqueue == NULL) |
| 1665 | continue; | 1586 | continue; |
| 1666 | 1587 | ||
| 1667 | spin_lock_irqsave(&ch->lock, irq_flags); | 1588 | spin_lock_irqsave(&ch->lock, irq_flags); |
| @@ -1687,8 +1608,9 @@ xpc_allocate_remote_msgqueue_sn2(struct xpc_channel *ch) | |||
| 1687 | * Note: Assumes all of the channel sizes are filled in. | 1608 | * Note: Assumes all of the channel sizes are filled in. |
| 1688 | */ | 1609 | */ |
| 1689 | static enum xp_retval | 1610 | static enum xp_retval |
| 1690 | xpc_allocate_msgqueues_sn2(struct xpc_channel *ch) | 1611 | xpc_setup_msg_structures_sn2(struct xpc_channel *ch) |
| 1691 | { | 1612 | { |
| 1613 | struct xpc_channel_sn2 *ch_sn2 = &ch->sn.sn2; | ||
| 1692 | enum xp_retval ret; | 1614 | enum xp_retval ret; |
| 1693 | 1615 | ||
| 1694 | DBUG_ON(ch->flags & XPC_C_SETUP); | 1616 | DBUG_ON(ch->flags & XPC_C_SETUP); |
| @@ -1698,10 +1620,10 @@ xpc_allocate_msgqueues_sn2(struct xpc_channel *ch) | |||
| 1698 | 1620 | ||
| 1699 | ret = xpc_allocate_remote_msgqueue_sn2(ch); | 1621 | ret = xpc_allocate_remote_msgqueue_sn2(ch); |
| 1700 | if (ret != xpSuccess) { | 1622 | if (ret != xpSuccess) { |
| 1701 | kfree(ch->local_msgqueue_base); | 1623 | kfree(ch_sn2->local_msgqueue_base); |
| 1702 | ch->local_msgqueue = NULL; | 1624 | ch_sn2->local_msgqueue = NULL; |
| 1703 | kfree(ch->notify_queue); | 1625 | kfree(ch_sn2->notify_queue); |
| 1704 | ch->notify_queue = NULL; | 1626 | ch_sn2->notify_queue = NULL; |
| 1705 | } | 1627 | } |
| 1706 | } | 1628 | } |
| 1707 | return ret; | 1629 | return ret; |
| @@ -1715,21 +1637,13 @@ xpc_allocate_msgqueues_sn2(struct xpc_channel *ch) | |||
| 1715 | * they're cleared when XPC_C_DISCONNECTED is cleared. | 1637 | * they're cleared when XPC_C_DISCONNECTED is cleared. |
| 1716 | */ | 1638 | */ |
| 1717 | static void | 1639 | static void |
| 1718 | xpc_free_msgqueues_sn2(struct xpc_channel *ch) | 1640 | xpc_teardown_msg_structures_sn2(struct xpc_channel *ch) |
| 1719 | { | 1641 | { |
| 1720 | struct xpc_channel_sn2 *ch_sn2 = &ch->sn.sn2; | 1642 | struct xpc_channel_sn2 *ch_sn2 = &ch->sn.sn2; |
| 1721 | 1643 | ||
| 1722 | DBUG_ON(!spin_is_locked(&ch->lock)); | 1644 | DBUG_ON(!spin_is_locked(&ch->lock)); |
| 1723 | DBUG_ON(atomic_read(&ch->n_to_notify) != 0); | ||
| 1724 | 1645 | ||
| 1725 | ch->remote_msgqueue_pa = 0; | 1646 | ch_sn2->remote_msgqueue_pa = 0; |
| 1726 | ch->func = NULL; | ||
| 1727 | ch->key = NULL; | ||
| 1728 | ch->msg_size = 0; | ||
| 1729 | ch->local_nentries = 0; | ||
| 1730 | ch->remote_nentries = 0; | ||
| 1731 | ch->kthreads_assigned_limit = 0; | ||
| 1732 | ch->kthreads_idle_limit = 0; | ||
| 1733 | 1647 | ||
| 1734 | ch_sn2->local_GP->get = 0; | 1648 | ch_sn2->local_GP->get = 0; |
| 1735 | ch_sn2->local_GP->put = 0; | 1649 | ch_sn2->local_GP->put = 0; |
| @@ -1745,12 +1659,12 @@ xpc_free_msgqueues_sn2(struct xpc_channel *ch) | |||
| 1745 | dev_dbg(xpc_chan, "ch->flags=0x%x, partid=%d, channel=%d\n", | 1659 | dev_dbg(xpc_chan, "ch->flags=0x%x, partid=%d, channel=%d\n", |
| 1746 | ch->flags, ch->partid, ch->number); | 1660 | ch->flags, ch->partid, ch->number); |
| 1747 | 1661 | ||
| 1748 | kfree(ch->local_msgqueue_base); | 1662 | kfree(ch_sn2->local_msgqueue_base); |
| 1749 | ch->local_msgqueue = NULL; | 1663 | ch_sn2->local_msgqueue = NULL; |
| 1750 | kfree(ch->remote_msgqueue_base); | 1664 | kfree(ch_sn2->remote_msgqueue_base); |
| 1751 | ch->remote_msgqueue = NULL; | 1665 | ch_sn2->remote_msgqueue = NULL; |
| 1752 | kfree(ch->notify_queue); | 1666 | kfree(ch_sn2->notify_queue); |
| 1753 | ch->notify_queue = NULL; | 1667 | ch_sn2->notify_queue = NULL; |
| 1754 | } | 1668 | } |
| 1755 | } | 1669 | } |
| 1756 | 1670 | ||
| @@ -1766,7 +1680,7 @@ xpc_notify_senders_sn2(struct xpc_channel *ch, enum xp_retval reason, s64 put) | |||
| 1766 | 1680 | ||
| 1767 | while (++get < put && atomic_read(&ch->n_to_notify) > 0) { | 1681 | while (++get < put && atomic_read(&ch->n_to_notify) > 0) { |
| 1768 | 1682 | ||
| 1769 | notify = &ch->notify_queue[get % ch->local_nentries]; | 1683 | notify = &ch->sn.sn2.notify_queue[get % ch->local_nentries]; |
| 1770 | 1684 | ||
| 1771 | /* | 1685 | /* |
| 1772 | * See if the notify entry indicates it was associated with | 1686 | * See if the notify entry indicates it was associated with |
| @@ -1818,7 +1732,7 @@ xpc_clear_local_msgqueue_flags_sn2(struct xpc_channel *ch) | |||
| 1818 | 1732 | ||
| 1819 | get = ch_sn2->w_remote_GP.get; | 1733 | get = ch_sn2->w_remote_GP.get; |
| 1820 | do { | 1734 | do { |
| 1821 | msg = (struct xpc_msg *)((u64)ch->local_msgqueue + | 1735 | msg = (struct xpc_msg *)((u64)ch_sn2->local_msgqueue + |
| 1822 | (get % ch->local_nentries) * | 1736 | (get % ch->local_nentries) * |
| 1823 | ch->msg_size); | 1737 | ch->msg_size); |
| 1824 | msg->flags = 0; | 1738 | msg->flags = 0; |
| @@ -1837,7 +1751,7 @@ xpc_clear_remote_msgqueue_flags_sn2(struct xpc_channel *ch) | |||
| 1837 | 1751 | ||
| 1838 | put = ch_sn2->w_remote_GP.put; | 1752 | put = ch_sn2->w_remote_GP.put; |
| 1839 | do { | 1753 | do { |
| 1840 | msg = (struct xpc_msg *)((u64)ch->remote_msgqueue + | 1754 | msg = (struct xpc_msg *)((u64)ch_sn2->remote_msgqueue + |
| 1841 | (put % ch->remote_nentries) * | 1755 | (put % ch->remote_nentries) * |
| 1842 | ch->msg_size); | 1756 | ch->msg_size); |
| 1843 | msg->flags = 0; | 1757 | msg->flags = 0; |
| @@ -1976,8 +1890,9 @@ xpc_pull_remote_msg_sn2(struct xpc_channel *ch, s64 get) | |||
| 1976 | } | 1890 | } |
| 1977 | 1891 | ||
| 1978 | msg_offset = msg_index * ch->msg_size; | 1892 | msg_offset = msg_index * ch->msg_size; |
| 1979 | msg = (struct xpc_msg *)((u64)ch->remote_msgqueue + msg_offset); | 1893 | msg = (struct xpc_msg *)((u64)ch_sn2->remote_msgqueue + |
| 1980 | remote_msg_pa = ch->remote_msgqueue_pa + msg_offset; | 1894 | msg_offset); |
| 1895 | remote_msg_pa = ch_sn2->remote_msgqueue_pa + msg_offset; | ||
| 1981 | 1896 | ||
| 1982 | ret = xpc_pull_remote_cachelines_sn2(part, msg, remote_msg_pa, | 1897 | ret = xpc_pull_remote_cachelines_sn2(part, msg, remote_msg_pa, |
| 1983 | nmsgs * ch->msg_size); | 1898 | nmsgs * ch->msg_size); |
| @@ -2001,7 +1916,7 @@ xpc_pull_remote_msg_sn2(struct xpc_channel *ch, s64 get) | |||
| 2001 | 1916 | ||
| 2002 | /* return the message we were looking for */ | 1917 | /* return the message we were looking for */ |
| 2003 | msg_offset = (get % ch->remote_nentries) * ch->msg_size; | 1918 | msg_offset = (get % ch->remote_nentries) * ch->msg_size; |
| 2004 | msg = (struct xpc_msg *)((u64)ch->remote_msgqueue + msg_offset); | 1919 | msg = (struct xpc_msg *)((u64)ch_sn2->remote_msgqueue + msg_offset); |
| 2005 | 1920 | ||
| 2006 | return msg; | 1921 | return msg; |
| 2007 | } | 1922 | } |
| @@ -2080,7 +1995,7 @@ xpc_send_msgs_sn2(struct xpc_channel *ch, s64 initial_put) | |||
| 2080 | if (put == ch_sn2->w_local_GP.put) | 1995 | if (put == ch_sn2->w_local_GP.put) |
| 2081 | break; | 1996 | break; |
| 2082 | 1997 | ||
| 2083 | msg = (struct xpc_msg *)((u64)ch->local_msgqueue + | 1998 | msg = (struct xpc_msg *)((u64)ch_sn2->local_msgqueue + |
| 2084 | (put % ch->local_nentries) * | 1999 | (put % ch->local_nentries) * |
| 2085 | ch->msg_size); | 2000 | ch->msg_size); |
| 2086 | 2001 | ||
| @@ -2182,7 +2097,7 @@ xpc_allocate_msg_sn2(struct xpc_channel *ch, u32 flags, | |||
| 2182 | } | 2097 | } |
| 2183 | 2098 | ||
| 2184 | /* get the message's address and initialize it */ | 2099 | /* get the message's address and initialize it */ |
| 2185 | msg = (struct xpc_msg *)((u64)ch->local_msgqueue + | 2100 | msg = (struct xpc_msg *)((u64)ch_sn2->local_msgqueue + |
| 2186 | (put % ch->local_nentries) * ch->msg_size); | 2101 | (put % ch->local_nentries) * ch->msg_size); |
| 2187 | 2102 | ||
| 2188 | DBUG_ON(msg->flags != 0); | 2103 | DBUG_ON(msg->flags != 0); |
| @@ -2207,6 +2122,7 @@ xpc_send_msg_sn2(struct xpc_channel *ch, u32 flags, void *payload, | |||
| 2207 | void *key) | 2122 | void *key) |
| 2208 | { | 2123 | { |
| 2209 | enum xp_retval ret = xpSuccess; | 2124 | enum xp_retval ret = xpSuccess; |
| 2125 | struct xpc_channel_sn2 *ch_sn2 = &ch->sn.sn2; | ||
| 2210 | struct xpc_msg *msg = msg; | 2126 | struct xpc_msg *msg = msg; |
| 2211 | struct xpc_notify *notify = notify; | 2127 | struct xpc_notify *notify = notify; |
| 2212 | s64 msg_number; | 2128 | s64 msg_number; |
| @@ -2243,7 +2159,7 @@ xpc_send_msg_sn2(struct xpc_channel *ch, u32 flags, void *payload, | |||
| 2243 | 2159 | ||
| 2244 | atomic_inc(&ch->n_to_notify); | 2160 | atomic_inc(&ch->n_to_notify); |
| 2245 | 2161 | ||
| 2246 | notify = &ch->notify_queue[msg_number % ch->local_nentries]; | 2162 | notify = &ch_sn2->notify_queue[msg_number % ch->local_nentries]; |
| 2247 | notify->func = func; | 2163 | notify->func = func; |
| 2248 | notify->key = key; | 2164 | notify->key = key; |
| 2249 | notify->type = notify_type; | 2165 | notify->type = notify_type; |
| @@ -2279,7 +2195,7 @@ xpc_send_msg_sn2(struct xpc_channel *ch, u32 flags, void *payload, | |||
| 2279 | 2195 | ||
| 2280 | /* see if the message is next in line to be sent, if so send it */ | 2196 | /* see if the message is next in line to be sent, if so send it */ |
| 2281 | 2197 | ||
| 2282 | put = ch->sn.sn2.local_GP->put; | 2198 | put = ch_sn2->local_GP->put; |
| 2283 | if (put == msg_number) | 2199 | if (put == msg_number) |
| 2284 | xpc_send_msgs_sn2(ch, put); | 2200 | xpc_send_msgs_sn2(ch, put); |
| 2285 | 2201 | ||
| @@ -2307,7 +2223,7 @@ xpc_acknowledge_msgs_sn2(struct xpc_channel *ch, s64 initial_get, u8 msg_flags) | |||
| 2307 | if (get == ch_sn2->w_local_GP.get) | 2223 | if (get == ch_sn2->w_local_GP.get) |
| 2308 | break; | 2224 | break; |
| 2309 | 2225 | ||
| 2310 | msg = (struct xpc_msg *)((u64)ch->remote_msgqueue + | 2226 | msg = (struct xpc_msg *)((u64)ch_sn2->remote_msgqueue + |
| 2311 | (get % ch->remote_nentries) * | 2227 | (get % ch->remote_nentries) * |
| 2312 | ch->msg_size); | 2228 | ch->msg_size); |
| 2313 | 2229 | ||
| @@ -2385,8 +2301,9 @@ xpc_init_sn2(void) | |||
| 2385 | int ret; | 2301 | int ret; |
| 2386 | size_t buf_size; | 2302 | size_t buf_size; |
| 2387 | 2303 | ||
| 2304 | xpc_setup_partitions_sn = xpc_setup_partitions_sn_sn2; | ||
| 2388 | xpc_get_partition_rsvd_page_pa = xpc_get_partition_rsvd_page_pa_sn2; | 2305 | xpc_get_partition_rsvd_page_pa = xpc_get_partition_rsvd_page_pa_sn2; |
| 2389 | xpc_rsvd_page_init = xpc_rsvd_page_init_sn2; | 2306 | xpc_setup_rsvd_page_sn = xpc_setup_rsvd_page_sn_sn2; |
| 2390 | xpc_increment_heartbeat = xpc_increment_heartbeat_sn2; | 2307 | xpc_increment_heartbeat = xpc_increment_heartbeat_sn2; |
| 2391 | xpc_offline_heartbeat = xpc_offline_heartbeat_sn2; | 2308 | xpc_offline_heartbeat = xpc_offline_heartbeat_sn2; |
| 2392 | xpc_online_heartbeat = xpc_online_heartbeat_sn2; | 2309 | xpc_online_heartbeat = xpc_online_heartbeat_sn2; |
| @@ -2403,29 +2320,33 @@ xpc_init_sn2(void) | |||
| 2403 | xpc_cancel_partition_deactivation_request_sn2; | 2320 | xpc_cancel_partition_deactivation_request_sn2; |
| 2404 | 2321 | ||
| 2405 | xpc_process_activate_IRQ_rcvd = xpc_process_activate_IRQ_rcvd_sn2; | 2322 | xpc_process_activate_IRQ_rcvd = xpc_process_activate_IRQ_rcvd_sn2; |
| 2406 | xpc_setup_infrastructure = xpc_setup_infrastructure_sn2; | 2323 | xpc_setup_ch_structures_sn = xpc_setup_ch_structures_sn_sn2; |
| 2407 | xpc_teardown_infrastructure = xpc_teardown_infrastructure_sn2; | 2324 | xpc_teardown_ch_structures_sn = xpc_teardown_ch_structures_sn_sn2; |
| 2408 | xpc_make_first_contact = xpc_make_first_contact_sn2; | 2325 | xpc_make_first_contact = xpc_make_first_contact_sn2; |
| 2326 | |||
| 2409 | xpc_get_chctl_all_flags = xpc_get_chctl_all_flags_sn2; | 2327 | xpc_get_chctl_all_flags = xpc_get_chctl_all_flags_sn2; |
| 2410 | xpc_allocate_msgqueues = xpc_allocate_msgqueues_sn2; | 2328 | xpc_send_chctl_closerequest = xpc_send_chctl_closerequest_sn2; |
| 2411 | xpc_free_msgqueues = xpc_free_msgqueues_sn2; | 2329 | xpc_send_chctl_closereply = xpc_send_chctl_closereply_sn2; |
| 2330 | xpc_send_chctl_openrequest = xpc_send_chctl_openrequest_sn2; | ||
| 2331 | xpc_send_chctl_openreply = xpc_send_chctl_openreply_sn2; | ||
| 2332 | |||
| 2333 | xpc_save_remote_msgqueue_pa = xpc_save_remote_msgqueue_pa_sn2; | ||
| 2334 | |||
| 2335 | xpc_setup_msg_structures = xpc_setup_msg_structures_sn2; | ||
| 2336 | xpc_teardown_msg_structures = xpc_teardown_msg_structures_sn2; | ||
| 2337 | |||
| 2412 | xpc_notify_senders_of_disconnect = xpc_notify_senders_of_disconnect_sn2; | 2338 | xpc_notify_senders_of_disconnect = xpc_notify_senders_of_disconnect_sn2; |
| 2413 | xpc_process_msg_chctl_flags = xpc_process_msg_chctl_flags_sn2; | 2339 | xpc_process_msg_chctl_flags = xpc_process_msg_chctl_flags_sn2; |
| 2414 | xpc_n_of_deliverable_msgs = xpc_n_of_deliverable_msgs_sn2; | 2340 | xpc_n_of_deliverable_msgs = xpc_n_of_deliverable_msgs_sn2; |
| 2415 | xpc_get_deliverable_msg = xpc_get_deliverable_msg_sn2; | 2341 | xpc_get_deliverable_msg = xpc_get_deliverable_msg_sn2; |
| 2416 | 2342 | ||
| 2417 | xpc_indicate_partition_engaged = xpc_indicate_partition_engaged_sn2; | 2343 | xpc_indicate_partition_engaged = xpc_indicate_partition_engaged_sn2; |
| 2418 | xpc_partition_engaged = xpc_partition_engaged_sn2; | ||
| 2419 | xpc_any_partition_engaged = xpc_any_partition_engaged_sn2; | ||
| 2420 | xpc_indicate_partition_disengaged = | 2344 | xpc_indicate_partition_disengaged = |
| 2421 | xpc_indicate_partition_disengaged_sn2; | 2345 | xpc_indicate_partition_disengaged_sn2; |
| 2346 | xpc_partition_engaged = xpc_partition_engaged_sn2; | ||
| 2347 | xpc_any_partition_engaged = xpc_any_partition_engaged_sn2; | ||
| 2422 | xpc_assume_partition_disengaged = xpc_assume_partition_disengaged_sn2; | 2348 | xpc_assume_partition_disengaged = xpc_assume_partition_disengaged_sn2; |
| 2423 | 2349 | ||
| 2424 | xpc_send_chctl_closerequest = xpc_send_chctl_closerequest_sn2; | ||
| 2425 | xpc_send_chctl_closereply = xpc_send_chctl_closereply_sn2; | ||
| 2426 | xpc_send_chctl_openrequest = xpc_send_chctl_openrequest_sn2; | ||
| 2427 | xpc_send_chctl_openreply = xpc_send_chctl_openreply_sn2; | ||
| 2428 | |||
| 2429 | xpc_send_msg = xpc_send_msg_sn2; | 2350 | xpc_send_msg = xpc_send_msg_sn2; |
| 2430 | xpc_received_msg = xpc_received_msg_sn2; | 2351 | xpc_received_msg = xpc_received_msg_sn2; |
| 2431 | 2352 | ||
diff --git a/drivers/misc/sgi-xp/xpc_uv.c b/drivers/misc/sgi-xp/xpc_uv.c index c2d4ddd6e955..689cb5c68ccf 100644 --- a/drivers/misc/sgi-xp/xpc_uv.c +++ b/drivers/misc/sgi-xp/xpc_uv.c | |||
| @@ -14,41 +14,528 @@ | |||
| 14 | */ | 14 | */ |
| 15 | 15 | ||
| 16 | #include <linux/kernel.h> | 16 | #include <linux/kernel.h> |
| 17 | #include <linux/mm.h> | ||
| 18 | #include <linux/interrupt.h> | ||
| 19 | #include <linux/delay.h> | ||
| 20 | #include <linux/device.h> | ||
| 17 | #include <asm/uv/uv_hub.h> | 21 | #include <asm/uv/uv_hub.h> |
| 22 | #include "../sgi-gru/gru.h" | ||
| 18 | #include "../sgi-gru/grukservices.h" | 23 | #include "../sgi-gru/grukservices.h" |
| 19 | #include "xpc.h" | 24 | #include "xpc.h" |
| 20 | 25 | ||
| 26 | static atomic64_t xpc_heartbeat_uv; | ||
| 21 | static DECLARE_BITMAP(xpc_heartbeating_to_mask_uv, XP_MAX_NPARTITIONS_UV); | 27 | static DECLARE_BITMAP(xpc_heartbeating_to_mask_uv, XP_MAX_NPARTITIONS_UV); |
| 22 | 28 | ||
| 23 | static void *xpc_activate_mq; | 29 | #define XPC_ACTIVATE_MSG_SIZE_UV (1 * GRU_CACHE_LINE_BYTES) |
| 30 | #define XPC_NOTIFY_MSG_SIZE_UV (2 * GRU_CACHE_LINE_BYTES) | ||
| 31 | |||
| 32 | #define XPC_ACTIVATE_MQ_SIZE_UV (4 * XP_MAX_NPARTITIONS_UV * \ | ||
| 33 | XPC_ACTIVATE_MSG_SIZE_UV) | ||
| 34 | #define XPC_NOTIFY_MQ_SIZE_UV (4 * XP_MAX_NPARTITIONS_UV * \ | ||
| 35 | XPC_NOTIFY_MSG_SIZE_UV) | ||
| 36 | |||
| 37 | static void *xpc_activate_mq_uv; | ||
| 38 | static void *xpc_notify_mq_uv; | ||
| 39 | |||
| 40 | static int | ||
| 41 | xpc_setup_partitions_sn_uv(void) | ||
| 42 | { | ||
| 43 | short partid; | ||
| 44 | struct xpc_partition_uv *part_uv; | ||
| 45 | |||
| 46 | for (partid = 0; partid < XP_MAX_NPARTITIONS_UV; partid++) { | ||
| 47 | part_uv = &xpc_partitions[partid].sn.uv; | ||
| 48 | |||
| 49 | spin_lock_init(&part_uv->flags_lock); | ||
| 50 | part_uv->remote_act_state = XPC_P_AS_INACTIVE; | ||
| 51 | } | ||
| 52 | return 0; | ||
| 53 | } | ||
| 54 | |||
| 55 | static void * | ||
| 56 | xpc_create_gru_mq_uv(unsigned int mq_size, int cpuid, unsigned int irq, | ||
| 57 | irq_handler_t irq_handler) | ||
| 58 | { | ||
| 59 | int ret; | ||
| 60 | int nid; | ||
| 61 | int mq_order; | ||
| 62 | struct page *page; | ||
| 63 | void *mq; | ||
| 64 | |||
| 65 | nid = cpu_to_node(cpuid); | ||
| 66 | mq_order = get_order(mq_size); | ||
| 67 | page = alloc_pages_node(nid, GFP_KERNEL | __GFP_ZERO | GFP_THISNODE, | ||
| 68 | mq_order); | ||
| 69 | if (page == NULL) | ||
| 70 | return NULL; | ||
| 71 | |||
| 72 | mq = page_address(page); | ||
| 73 | ret = gru_create_message_queue(mq, mq_size); | ||
| 74 | if (ret != 0) { | ||
| 75 | dev_err(xpc_part, "gru_create_message_queue() returned " | ||
| 76 | "error=%d\n", ret); | ||
| 77 | free_pages((unsigned long)mq, mq_order); | ||
| 78 | return NULL; | ||
| 79 | } | ||
| 80 | |||
| 81 | /* !!! Need to do some other things to set up IRQ */ | ||
| 82 | |||
| 83 | ret = request_irq(irq, irq_handler, 0, "xpc", NULL); | ||
| 84 | if (ret != 0) { | ||
| 85 | dev_err(xpc_part, "request_irq(irq=%d) returned error=%d\n", | ||
| 86 | irq, ret); | ||
| 87 | free_pages((unsigned long)mq, mq_order); | ||
| 88 | return NULL; | ||
| 89 | } | ||
| 90 | |||
| 91 | /* !!! enable generation of irq when GRU mq op occurs to this mq */ | ||
| 92 | |||
| 93 | /* ??? allow other partitions to access GRU mq? */ | ||
| 94 | |||
| 95 | return mq; | ||
| 96 | } | ||
| 24 | 97 | ||
| 25 | static void | 98 | static void |
| 26 | xpc_send_local_activate_IRQ_uv(struct xpc_partition *part) | 99 | xpc_destroy_gru_mq_uv(void *mq, unsigned int mq_size, unsigned int irq) |
| 100 | { | ||
| 101 | /* ??? disallow other partitions to access GRU mq? */ | ||
| 102 | |||
| 103 | /* !!! disable generation of irq when GRU mq op occurs to this mq */ | ||
| 104 | |||
| 105 | free_irq(irq, NULL); | ||
| 106 | |||
| 107 | free_pages((unsigned long)mq, get_order(mq_size)); | ||
| 108 | } | ||
| 109 | |||
| 110 | static enum xp_retval | ||
| 111 | xpc_send_gru_msg(unsigned long mq_gpa, void *msg, size_t msg_size) | ||
| 27 | { | 112 | { |
| 113 | enum xp_retval xp_ret; | ||
| 114 | int ret; | ||
| 115 | |||
| 116 | while (1) { | ||
| 117 | ret = gru_send_message_gpa(mq_gpa, msg, msg_size); | ||
| 118 | if (ret == MQE_OK) { | ||
| 119 | xp_ret = xpSuccess; | ||
| 120 | break; | ||
| 121 | } | ||
| 122 | |||
| 123 | if (ret == MQE_QUEUE_FULL) { | ||
| 124 | dev_dbg(xpc_chan, "gru_send_message_gpa() returned " | ||
| 125 | "error=MQE_QUEUE_FULL\n"); | ||
| 126 | /* !!! handle QLimit reached; delay & try again */ | ||
| 127 | /* ??? Do we add a limit to the number of retries? */ | ||
| 128 | (void)msleep_interruptible(10); | ||
| 129 | } else if (ret == MQE_CONGESTION) { | ||
| 130 | dev_dbg(xpc_chan, "gru_send_message_gpa() returned " | ||
| 131 | "error=MQE_CONGESTION\n"); | ||
| 132 | /* !!! handle LB Overflow; simply try again */ | ||
| 133 | /* ??? Do we add a limit to the number of retries? */ | ||
| 134 | } else { | ||
| 135 | /* !!! Currently this is MQE_UNEXPECTED_CB_ERR */ | ||
| 136 | dev_err(xpc_chan, "gru_send_message_gpa() returned " | ||
| 137 | "error=%d\n", ret); | ||
| 138 | xp_ret = xpGruSendMqError; | ||
| 139 | break; | ||
| 140 | } | ||
| 141 | } | ||
| 142 | return xp_ret; | ||
| 143 | } | ||
| 144 | |||
| 145 | static void | ||
| 146 | xpc_process_activate_IRQ_rcvd_uv(void) | ||
| 147 | { | ||
| 148 | unsigned long irq_flags; | ||
| 149 | short partid; | ||
| 150 | struct xpc_partition *part; | ||
| 151 | u8 act_state_req; | ||
| 152 | |||
| 153 | DBUG_ON(xpc_activate_IRQ_rcvd == 0); | ||
| 154 | |||
| 155 | spin_lock_irqsave(&xpc_activate_IRQ_rcvd_lock, irq_flags); | ||
| 156 | for (partid = 0; partid < XP_MAX_NPARTITIONS_UV; partid++) { | ||
| 157 | part = &xpc_partitions[partid]; | ||
| 158 | |||
| 159 | if (part->sn.uv.act_state_req == 0) | ||
| 160 | continue; | ||
| 161 | |||
| 162 | xpc_activate_IRQ_rcvd--; | ||
| 163 | BUG_ON(xpc_activate_IRQ_rcvd < 0); | ||
| 164 | |||
| 165 | act_state_req = part->sn.uv.act_state_req; | ||
| 166 | part->sn.uv.act_state_req = 0; | ||
| 167 | spin_unlock_irqrestore(&xpc_activate_IRQ_rcvd_lock, irq_flags); | ||
| 168 | |||
| 169 | if (act_state_req == XPC_P_ASR_ACTIVATE_UV) { | ||
| 170 | if (part->act_state == XPC_P_AS_INACTIVE) | ||
| 171 | xpc_activate_partition(part); | ||
| 172 | else if (part->act_state == XPC_P_AS_DEACTIVATING) | ||
| 173 | XPC_DEACTIVATE_PARTITION(part, xpReactivating); | ||
| 174 | |||
| 175 | } else if (act_state_req == XPC_P_ASR_REACTIVATE_UV) { | ||
| 176 | if (part->act_state == XPC_P_AS_INACTIVE) | ||
| 177 | xpc_activate_partition(part); | ||
| 178 | else | ||
| 179 | XPC_DEACTIVATE_PARTITION(part, xpReactivating); | ||
| 180 | |||
| 181 | } else if (act_state_req == XPC_P_ASR_DEACTIVATE_UV) { | ||
| 182 | XPC_DEACTIVATE_PARTITION(part, part->sn.uv.reason); | ||
| 183 | |||
| 184 | } else { | ||
| 185 | BUG(); | ||
| 186 | } | ||
| 187 | |||
| 188 | spin_lock_irqsave(&xpc_activate_IRQ_rcvd_lock, irq_flags); | ||
| 189 | if (xpc_activate_IRQ_rcvd == 0) | ||
| 190 | break; | ||
| 191 | } | ||
| 192 | spin_unlock_irqrestore(&xpc_activate_IRQ_rcvd_lock, irq_flags); | ||
| 193 | |||
| 194 | } | ||
| 195 | |||
| 196 | static irqreturn_t | ||
| 197 | xpc_handle_activate_IRQ_uv(int irq, void *dev_id) | ||
| 198 | { | ||
| 199 | unsigned long irq_flags; | ||
| 200 | struct xpc_activate_mq_msghdr_uv *msg_hdr; | ||
| 201 | short partid; | ||
| 202 | struct xpc_partition *part; | ||
| 203 | struct xpc_partition_uv *part_uv; | ||
| 204 | struct xpc_openclose_args *args; | ||
| 205 | int wakeup_hb_checker = 0; | ||
| 206 | |||
| 207 | while ((msg_hdr = gru_get_next_message(xpc_activate_mq_uv)) != NULL) { | ||
| 208 | |||
| 209 | partid = msg_hdr->partid; | ||
| 210 | if (partid < 0 || partid >= XP_MAX_NPARTITIONS_UV) { | ||
| 211 | dev_err(xpc_part, "xpc_handle_activate_IRQ_uv() invalid" | ||
| 212 | "partid=0x%x passed in message\n", partid); | ||
| 213 | gru_free_message(xpc_activate_mq_uv, msg_hdr); | ||
| 214 | continue; | ||
| 215 | } | ||
| 216 | part = &xpc_partitions[partid]; | ||
| 217 | part_uv = &part->sn.uv; | ||
| 218 | |||
| 219 | part_uv->remote_act_state = msg_hdr->act_state; | ||
| 220 | |||
| 221 | switch (msg_hdr->type) { | ||
| 222 | case XPC_ACTIVATE_MQ_MSG_SYNC_ACT_STATE_UV: | ||
| 223 | /* syncing of remote_act_state was just done above */ | ||
| 224 | break; | ||
| 225 | |||
| 226 | case XPC_ACTIVATE_MQ_MSG_INC_HEARTBEAT_UV: { | ||
| 227 | struct xpc_activate_mq_msg_heartbeat_req_uv *msg; | ||
| 228 | |||
| 229 | msg = (struct xpc_activate_mq_msg_heartbeat_req_uv *) | ||
| 230 | msg_hdr; | ||
| 231 | part_uv->heartbeat = msg->heartbeat; | ||
| 232 | break; | ||
| 233 | } | ||
| 234 | case XPC_ACTIVATE_MQ_MSG_OFFLINE_HEARTBEAT_UV: { | ||
| 235 | struct xpc_activate_mq_msg_heartbeat_req_uv *msg; | ||
| 236 | |||
| 237 | msg = (struct xpc_activate_mq_msg_heartbeat_req_uv *) | ||
| 238 | msg_hdr; | ||
| 239 | part_uv->heartbeat = msg->heartbeat; | ||
| 240 | spin_lock_irqsave(&part_uv->flags_lock, irq_flags); | ||
| 241 | part_uv->flags |= XPC_P_HEARTBEAT_OFFLINE_UV; | ||
| 242 | spin_unlock_irqrestore(&part_uv->flags_lock, irq_flags); | ||
| 243 | break; | ||
| 244 | } | ||
| 245 | case XPC_ACTIVATE_MQ_MSG_ONLINE_HEARTBEAT_UV: { | ||
| 246 | struct xpc_activate_mq_msg_heartbeat_req_uv *msg; | ||
| 247 | |||
| 248 | msg = (struct xpc_activate_mq_msg_heartbeat_req_uv *) | ||
| 249 | msg_hdr; | ||
| 250 | part_uv->heartbeat = msg->heartbeat; | ||
| 251 | spin_lock_irqsave(&part_uv->flags_lock, irq_flags); | ||
| 252 | part_uv->flags &= ~XPC_P_HEARTBEAT_OFFLINE_UV; | ||
| 253 | spin_unlock_irqrestore(&part_uv->flags_lock, irq_flags); | ||
| 254 | break; | ||
| 255 | } | ||
| 256 | case XPC_ACTIVATE_MQ_MSG_ACTIVATE_REQ_UV: { | ||
| 257 | struct xpc_activate_mq_msg_activate_req_uv *msg; | ||
| 258 | |||
| 259 | /* | ||
| 260 | * ??? Do we deal here with ts_jiffies being different | ||
| 261 | * ??? if act_state != XPC_P_AS_INACTIVE instead of | ||
| 262 | * ??? below? | ||
| 263 | */ | ||
| 264 | msg = (struct xpc_activate_mq_msg_activate_req_uv *) | ||
| 265 | msg_hdr; | ||
| 266 | spin_lock_irqsave(&xpc_activate_IRQ_rcvd_lock, | ||
| 267 | irq_flags); | ||
| 268 | if (part_uv->act_state_req == 0) | ||
| 269 | xpc_activate_IRQ_rcvd++; | ||
| 270 | part_uv->act_state_req = XPC_P_ASR_ACTIVATE_UV; | ||
| 271 | part->remote_rp_pa = msg->rp_gpa; /* !!! _pa is _gpa */ | ||
| 272 | part->remote_rp_ts_jiffies = msg_hdr->rp_ts_jiffies; | ||
| 273 | part_uv->remote_activate_mq_gpa = msg->activate_mq_gpa; | ||
| 274 | spin_unlock_irqrestore(&xpc_activate_IRQ_rcvd_lock, | ||
| 275 | irq_flags); | ||
| 276 | wakeup_hb_checker++; | ||
| 277 | break; | ||
| 278 | } | ||
| 279 | case XPC_ACTIVATE_MQ_MSG_DEACTIVATE_REQ_UV: { | ||
| 280 | struct xpc_activate_mq_msg_deactivate_req_uv *msg; | ||
| 281 | |||
| 282 | msg = (struct xpc_activate_mq_msg_deactivate_req_uv *) | ||
| 283 | msg_hdr; | ||
| 284 | spin_lock_irqsave(&xpc_activate_IRQ_rcvd_lock, | ||
| 285 | irq_flags); | ||
| 286 | if (part_uv->act_state_req == 0) | ||
| 287 | xpc_activate_IRQ_rcvd++; | ||
| 288 | part_uv->act_state_req = XPC_P_ASR_DEACTIVATE_UV; | ||
| 289 | part_uv->reason = msg->reason; | ||
| 290 | spin_unlock_irqrestore(&xpc_activate_IRQ_rcvd_lock, | ||
| 291 | irq_flags); | ||
| 292 | wakeup_hb_checker++; | ||
| 293 | break; | ||
| 294 | } | ||
| 295 | case XPC_ACTIVATE_MQ_MSG_CHCTL_CLOSEREQUEST_UV: { | ||
| 296 | struct xpc_activate_mq_msg_chctl_closerequest_uv *msg; | ||
| 297 | |||
| 298 | msg = (struct xpc_activate_mq_msg_chctl_closerequest_uv | ||
| 299 | *)msg_hdr; | ||
| 300 | args = &part->remote_openclose_args[msg->ch_number]; | ||
| 301 | args->reason = msg->reason; | ||
| 302 | |||
| 303 | spin_lock_irqsave(&part->chctl_lock, irq_flags); | ||
| 304 | part->chctl.flags[msg->ch_number] |= | ||
| 305 | XPC_CHCTL_CLOSEREQUEST; | ||
| 306 | spin_unlock_irqrestore(&part->chctl_lock, irq_flags); | ||
| 307 | |||
| 308 | xpc_wakeup_channel_mgr(part); | ||
| 309 | break; | ||
| 310 | } | ||
| 311 | case XPC_ACTIVATE_MQ_MSG_CHCTL_CLOSEREPLY_UV: { | ||
| 312 | struct xpc_activate_mq_msg_chctl_closereply_uv *msg; | ||
| 313 | |||
| 314 | msg = (struct xpc_activate_mq_msg_chctl_closereply_uv *) | ||
| 315 | msg_hdr; | ||
| 316 | |||
| 317 | spin_lock_irqsave(&part->chctl_lock, irq_flags); | ||
| 318 | part->chctl.flags[msg->ch_number] |= | ||
| 319 | XPC_CHCTL_CLOSEREPLY; | ||
| 320 | spin_unlock_irqrestore(&part->chctl_lock, irq_flags); | ||
| 321 | |||
| 322 | xpc_wakeup_channel_mgr(part); | ||
| 323 | break; | ||
| 324 | } | ||
| 325 | case XPC_ACTIVATE_MQ_MSG_CHCTL_OPENREQUEST_UV: { | ||
| 326 | struct xpc_activate_mq_msg_chctl_openrequest_uv *msg; | ||
| 327 | |||
| 328 | msg = (struct xpc_activate_mq_msg_chctl_openrequest_uv | ||
| 329 | *)msg_hdr; | ||
| 330 | args = &part->remote_openclose_args[msg->ch_number]; | ||
| 331 | args->msg_size = msg->msg_size; | ||
| 332 | args->local_nentries = msg->local_nentries; | ||
| 333 | |||
| 334 | spin_lock_irqsave(&part->chctl_lock, irq_flags); | ||
| 335 | part->chctl.flags[msg->ch_number] |= | ||
| 336 | XPC_CHCTL_OPENREQUEST; | ||
| 337 | spin_unlock_irqrestore(&part->chctl_lock, irq_flags); | ||
| 338 | |||
| 339 | xpc_wakeup_channel_mgr(part); | ||
| 340 | break; | ||
| 341 | } | ||
| 342 | case XPC_ACTIVATE_MQ_MSG_CHCTL_OPENREPLY_UV: { | ||
| 343 | struct xpc_activate_mq_msg_chctl_openreply_uv *msg; | ||
| 344 | |||
| 345 | msg = (struct xpc_activate_mq_msg_chctl_openreply_uv *) | ||
| 346 | msg_hdr; | ||
| 347 | args = &part->remote_openclose_args[msg->ch_number]; | ||
| 348 | args->remote_nentries = msg->remote_nentries; | ||
| 349 | args->local_nentries = msg->local_nentries; | ||
| 350 | args->local_msgqueue_pa = msg->local_notify_mq_gpa; | ||
| 351 | |||
| 352 | spin_lock_irqsave(&part->chctl_lock, irq_flags); | ||
| 353 | part->chctl.flags[msg->ch_number] |= | ||
| 354 | XPC_CHCTL_OPENREPLY; | ||
| 355 | spin_unlock_irqrestore(&part->chctl_lock, irq_flags); | ||
| 356 | |||
| 357 | xpc_wakeup_channel_mgr(part); | ||
| 358 | break; | ||
| 359 | } | ||
| 360 | case XPC_ACTIVATE_MQ_MSG_MARK_ENGAGED_UV: | ||
| 361 | spin_lock_irqsave(&part_uv->flags_lock, irq_flags); | ||
| 362 | part_uv->flags |= XPC_P_ENGAGED_UV; | ||
| 363 | spin_unlock_irqrestore(&part_uv->flags_lock, irq_flags); | ||
| 364 | break; | ||
| 365 | |||
| 366 | case XPC_ACTIVATE_MQ_MSG_MARK_DISENGAGED_UV: | ||
| 367 | spin_lock_irqsave(&part_uv->flags_lock, irq_flags); | ||
| 368 | part_uv->flags &= ~XPC_P_ENGAGED_UV; | ||
| 369 | spin_unlock_irqrestore(&part_uv->flags_lock, irq_flags); | ||
| 370 | break; | ||
| 371 | |||
| 372 | default: | ||
| 373 | dev_err(xpc_part, "received unknown activate_mq msg " | ||
| 374 | "type=%d from partition=%d\n", msg_hdr->type, | ||
| 375 | partid); | ||
| 376 | } | ||
| 377 | |||
| 378 | if (msg_hdr->rp_ts_jiffies != part->remote_rp_ts_jiffies && | ||
| 379 | part->remote_rp_ts_jiffies != 0) { | ||
| 380 | /* | ||
| 381 | * ??? Does what we do here need to be sensitive to | ||
| 382 | * ??? act_state or remote_act_state? | ||
| 383 | */ | ||
| 384 | spin_lock_irqsave(&xpc_activate_IRQ_rcvd_lock, | ||
| 385 | irq_flags); | ||
| 386 | if (part_uv->act_state_req == 0) | ||
| 387 | xpc_activate_IRQ_rcvd++; | ||
| 388 | part_uv->act_state_req = XPC_P_ASR_REACTIVATE_UV; | ||
| 389 | spin_unlock_irqrestore(&xpc_activate_IRQ_rcvd_lock, | ||
| 390 | irq_flags); | ||
| 391 | wakeup_hb_checker++; | ||
| 392 | } | ||
| 393 | |||
| 394 | gru_free_message(xpc_activate_mq_uv, msg_hdr); | ||
| 395 | } | ||
| 396 | |||
| 397 | if (wakeup_hb_checker) | ||
| 398 | wake_up_interruptible(&xpc_activate_IRQ_wq); | ||
| 399 | |||
| 400 | return IRQ_HANDLED; | ||
| 401 | } | ||
| 402 | |||
| 403 | static enum xp_retval | ||
| 404 | xpc_send_activate_IRQ_uv(struct xpc_partition *part, void *msg, size_t msg_size, | ||
| 405 | int msg_type) | ||
| 406 | { | ||
| 407 | struct xpc_activate_mq_msghdr_uv *msg_hdr = msg; | ||
| 408 | |||
| 409 | DBUG_ON(msg_size > XPC_ACTIVATE_MSG_SIZE_UV); | ||
| 410 | |||
| 411 | msg_hdr->type = msg_type; | ||
| 412 | msg_hdr->partid = XPC_PARTID(part); | ||
| 413 | msg_hdr->act_state = part->act_state; | ||
| 414 | msg_hdr->rp_ts_jiffies = xpc_rsvd_page->ts_jiffies; | ||
| 415 | |||
| 416 | /* ??? Is holding a spin_lock (ch->lock) during this call a bad idea? */ | ||
| 417 | return xpc_send_gru_msg(part->sn.uv.remote_activate_mq_gpa, msg, | ||
| 418 | msg_size); | ||
| 419 | } | ||
| 420 | |||
| 421 | static void | ||
| 422 | xpc_send_activate_IRQ_part_uv(struct xpc_partition *part, void *msg, | ||
| 423 | size_t msg_size, int msg_type) | ||
| 424 | { | ||
| 425 | enum xp_retval ret; | ||
| 426 | |||
| 427 | ret = xpc_send_activate_IRQ_uv(part, msg, msg_size, msg_type); | ||
| 428 | if (unlikely(ret != xpSuccess)) | ||
| 429 | XPC_DEACTIVATE_PARTITION(part, ret); | ||
| 430 | } | ||
| 431 | |||
| 432 | static void | ||
| 433 | xpc_send_activate_IRQ_ch_uv(struct xpc_channel *ch, unsigned long *irq_flags, | ||
| 434 | void *msg, size_t msg_size, int msg_type) | ||
| 435 | { | ||
| 436 | struct xpc_partition *part = &xpc_partitions[ch->number]; | ||
| 437 | enum xp_retval ret; | ||
| 438 | |||
| 439 | ret = xpc_send_activate_IRQ_uv(part, msg, msg_size, msg_type); | ||
| 440 | if (unlikely(ret != xpSuccess)) { | ||
| 441 | if (irq_flags != NULL) | ||
| 442 | spin_unlock_irqrestore(&ch->lock, *irq_flags); | ||
| 443 | |||
| 444 | XPC_DEACTIVATE_PARTITION(part, ret); | ||
| 445 | |||
| 446 | if (irq_flags != NULL) | ||
| 447 | spin_lock_irqsave(&ch->lock, *irq_flags); | ||
| 448 | } | ||
| 449 | } | ||
| 450 | |||
| 451 | static void | ||
| 452 | xpc_send_local_activate_IRQ_uv(struct xpc_partition *part, int act_state_req) | ||
| 453 | { | ||
| 454 | unsigned long irq_flags; | ||
| 455 | struct xpc_partition_uv *part_uv = &part->sn.uv; | ||
| 456 | |||
| 28 | /* | 457 | /* |
| 29 | * !!! Make our side think that the remote parition sent an activate | 458 | * !!! Make our side think that the remote parition sent an activate |
| 30 | * !!! message our way. Also do what the activate IRQ handler would | 459 | * !!! message our way by doing what the activate IRQ handler would |
| 31 | * !!! do had one really been sent. | 460 | * !!! do had one really been sent. |
| 32 | */ | 461 | */ |
| 462 | |||
| 463 | spin_lock_irqsave(&xpc_activate_IRQ_rcvd_lock, irq_flags); | ||
| 464 | if (part_uv->act_state_req == 0) | ||
| 465 | xpc_activate_IRQ_rcvd++; | ||
| 466 | part_uv->act_state_req = act_state_req; | ||
| 467 | spin_unlock_irqrestore(&xpc_activate_IRQ_rcvd_lock, irq_flags); | ||
| 468 | |||
| 469 | wake_up_interruptible(&xpc_activate_IRQ_wq); | ||
| 33 | } | 470 | } |
| 34 | 471 | ||
| 35 | static enum xp_retval | 472 | static enum xp_retval |
| 36 | xpc_rsvd_page_init_uv(struct xpc_rsvd_page *rp) | 473 | xpc_get_partition_rsvd_page_pa_uv(void *buf, u64 *cookie, unsigned long *rp_pa, |
| 474 | size_t *len) | ||
| 37 | { | 475 | { |
| 38 | /* !!! need to have established xpc_activate_mq earlier */ | 476 | /* !!! call the UV version of sn_partition_reserved_page_pa() */ |
| 39 | rp->sn.activate_mq_gpa = uv_gpa(xpc_activate_mq); | 477 | return xpUnsupported; |
| 40 | return xpSuccess; | 478 | } |
| 479 | |||
| 480 | static int | ||
| 481 | xpc_setup_rsvd_page_sn_uv(struct xpc_rsvd_page *rp) | ||
| 482 | { | ||
| 483 | rp->sn.activate_mq_gpa = uv_gpa(xpc_activate_mq_uv); | ||
| 484 | return 0; | ||
| 485 | } | ||
| 486 | |||
| 487 | static void | ||
| 488 | xpc_send_heartbeat_uv(int msg_type) | ||
| 489 | { | ||
| 490 | short partid; | ||
| 491 | struct xpc_partition *part; | ||
| 492 | struct xpc_activate_mq_msg_heartbeat_req_uv msg; | ||
| 493 | |||
| 494 | /* | ||
| 495 | * !!! On uv we're broadcasting a heartbeat message every 5 seconds. | ||
| 496 | * !!! Whereas on sn2 we're bte_copy'ng the heartbeat info every 20 | ||
| 497 | * !!! seconds. This is an increase in numalink traffic. | ||
| 498 | * ??? Is this good? | ||
| 499 | */ | ||
| 500 | |||
| 501 | msg.heartbeat = atomic64_inc_return(&xpc_heartbeat_uv); | ||
| 502 | |||
| 503 | partid = find_first_bit(xpc_heartbeating_to_mask_uv, | ||
| 504 | XP_MAX_NPARTITIONS_UV); | ||
| 505 | |||
| 506 | while (partid < XP_MAX_NPARTITIONS_UV) { | ||
| 507 | part = &xpc_partitions[partid]; | ||
| 508 | |||
| 509 | xpc_send_activate_IRQ_part_uv(part, &msg, sizeof(msg), | ||
| 510 | msg_type); | ||
| 511 | |||
| 512 | partid = find_next_bit(xpc_heartbeating_to_mask_uv, | ||
| 513 | XP_MAX_NPARTITIONS_UV, partid + 1); | ||
| 514 | } | ||
| 41 | } | 515 | } |
| 42 | 516 | ||
| 43 | static void | 517 | static void |
| 44 | xpc_increment_heartbeat_uv(void) | 518 | xpc_increment_heartbeat_uv(void) |
| 45 | { | 519 | { |
| 46 | /* !!! send heartbeat msg to xpc_heartbeating_to_mask partids */ | 520 | xpc_send_heartbeat_uv(XPC_ACTIVATE_MQ_MSG_INC_HEARTBEAT_UV); |
| 521 | } | ||
| 522 | |||
| 523 | static void | ||
| 524 | xpc_offline_heartbeat_uv(void) | ||
| 525 | { | ||
| 526 | xpc_send_heartbeat_uv(XPC_ACTIVATE_MQ_MSG_OFFLINE_HEARTBEAT_UV); | ||
| 527 | } | ||
| 528 | |||
| 529 | static void | ||
| 530 | xpc_online_heartbeat_uv(void) | ||
| 531 | { | ||
| 532 | xpc_send_heartbeat_uv(XPC_ACTIVATE_MQ_MSG_ONLINE_HEARTBEAT_UV); | ||
| 47 | } | 533 | } |
| 48 | 534 | ||
| 49 | static void | 535 | static void |
| 50 | xpc_heartbeat_init_uv(void) | 536 | xpc_heartbeat_init_uv(void) |
| 51 | { | 537 | { |
| 538 | atomic64_set(&xpc_heartbeat_uv, 0); | ||
| 52 | bitmap_zero(xpc_heartbeating_to_mask_uv, XP_MAX_NPARTITIONS_UV); | 539 | bitmap_zero(xpc_heartbeating_to_mask_uv, XP_MAX_NPARTITIONS_UV); |
| 53 | xpc_heartbeating_to_mask = &xpc_heartbeating_to_mask_uv[0]; | 540 | xpc_heartbeating_to_mask = &xpc_heartbeating_to_mask_uv[0]; |
| 54 | } | 541 | } |
| @@ -56,48 +543,94 @@ xpc_heartbeat_init_uv(void) | |||
| 56 | static void | 543 | static void |
| 57 | xpc_heartbeat_exit_uv(void) | 544 | xpc_heartbeat_exit_uv(void) |
| 58 | { | 545 | { |
| 59 | /* !!! send heartbeat_offline msg to xpc_heartbeating_to_mask partids */ | 546 | xpc_send_heartbeat_uv(XPC_ACTIVATE_MQ_MSG_OFFLINE_HEARTBEAT_UV); |
| 547 | } | ||
| 548 | |||
| 549 | static enum xp_retval | ||
| 550 | xpc_get_remote_heartbeat_uv(struct xpc_partition *part) | ||
| 551 | { | ||
| 552 | struct xpc_partition_uv *part_uv = &part->sn.uv; | ||
| 553 | enum xp_retval ret = xpNoHeartbeat; | ||
| 554 | |||
| 555 | if (part_uv->remote_act_state != XPC_P_AS_INACTIVE && | ||
| 556 | part_uv->remote_act_state != XPC_P_AS_DEACTIVATING) { | ||
| 557 | |||
| 558 | if (part_uv->heartbeat != part->last_heartbeat || | ||
| 559 | (part_uv->flags & XPC_P_HEARTBEAT_OFFLINE_UV)) { | ||
| 560 | |||
| 561 | part->last_heartbeat = part_uv->heartbeat; | ||
| 562 | ret = xpSuccess; | ||
| 563 | } | ||
| 564 | } | ||
| 565 | return ret; | ||
| 60 | } | 566 | } |
| 61 | 567 | ||
| 62 | static void | 568 | static void |
| 63 | xpc_request_partition_activation_uv(struct xpc_rsvd_page *remote_rp, | 569 | xpc_request_partition_activation_uv(struct xpc_rsvd_page *remote_rp, |
| 64 | unsigned long remote_rp_pa, int nasid) | 570 | unsigned long remote_rp_gpa, int nasid) |
| 65 | { | 571 | { |
| 66 | short partid = remote_rp->SAL_partid; | 572 | short partid = remote_rp->SAL_partid; |
| 67 | struct xpc_partition *part = &xpc_partitions[partid]; | 573 | struct xpc_partition *part = &xpc_partitions[partid]; |
| 574 | struct xpc_activate_mq_msg_activate_req_uv msg; | ||
| 68 | 575 | ||
| 69 | /* | 576 | part->remote_rp_pa = remote_rp_gpa; /* !!! _pa here is really _gpa */ |
| 70 | * !!! Setup part structure with the bits of info we can glean from the rp: | 577 | part->remote_rp_ts_jiffies = remote_rp->ts_jiffies; |
| 71 | * !!! part->remote_rp_pa = remote_rp_pa; | 578 | part->sn.uv.remote_activate_mq_gpa = remote_rp->sn.activate_mq_gpa; |
| 72 | * !!! part->sn.uv.activate_mq_gpa = remote_rp->sn.activate_mq_gpa; | 579 | |
| 73 | */ | 580 | /* |
| 581 | * ??? Is it a good idea to make this conditional on what is | ||
| 582 | * ??? potentially stale state information? | ||
| 583 | */ | ||
| 584 | if (part->sn.uv.remote_act_state == XPC_P_AS_INACTIVE) { | ||
| 585 | msg.rp_gpa = uv_gpa(xpc_rsvd_page); | ||
| 586 | msg.activate_mq_gpa = xpc_rsvd_page->sn.activate_mq_gpa; | ||
| 587 | xpc_send_activate_IRQ_part_uv(part, &msg, sizeof(msg), | ||
| 588 | XPC_ACTIVATE_MQ_MSG_ACTIVATE_REQ_UV); | ||
| 589 | } | ||
| 74 | 590 | ||
| 75 | xpc_send_local_activate_IRQ_uv(part); | 591 | if (part->act_state == XPC_P_AS_INACTIVE) |
| 592 | xpc_send_local_activate_IRQ_uv(part, XPC_P_ASR_ACTIVATE_UV); | ||
| 76 | } | 593 | } |
| 77 | 594 | ||
| 78 | static void | 595 | static void |
| 79 | xpc_request_partition_reactivation_uv(struct xpc_partition *part) | 596 | xpc_request_partition_reactivation_uv(struct xpc_partition *part) |
| 80 | { | 597 | { |
| 81 | xpc_send_local_activate_IRQ_uv(part); | 598 | xpc_send_local_activate_IRQ_uv(part, XPC_P_ASR_ACTIVATE_UV); |
| 599 | } | ||
| 600 | |||
| 601 | static void | ||
| 602 | xpc_request_partition_deactivation_uv(struct xpc_partition *part) | ||
| 603 | { | ||
| 604 | struct xpc_activate_mq_msg_deactivate_req_uv msg; | ||
| 605 | |||
| 606 | /* | ||
| 607 | * ??? Is it a good idea to make this conditional on what is | ||
| 608 | * ??? potentially stale state information? | ||
| 609 | */ | ||
| 610 | if (part->sn.uv.remote_act_state != XPC_P_AS_DEACTIVATING && | ||
| 611 | part->sn.uv.remote_act_state != XPC_P_AS_INACTIVE) { | ||
| 612 | |||
| 613 | msg.reason = part->reason; | ||
| 614 | xpc_send_activate_IRQ_part_uv(part, &msg, sizeof(msg), | ||
| 615 | XPC_ACTIVATE_MQ_MSG_DEACTIVATE_REQ_UV); | ||
| 616 | } | ||
| 82 | } | 617 | } |
| 83 | 618 | ||
| 84 | /* | 619 | /* |
| 85 | * Setup the infrastructure necessary to support XPartition Communication | 620 | * Setup the channel structures that are uv specific. |
| 86 | * between the specified remote partition and the local one. | ||
| 87 | */ | 621 | */ |
| 88 | static enum xp_retval | 622 | static enum xp_retval |
| 89 | xpc_setup_infrastructure_uv(struct xpc_partition *part) | 623 | xpc_setup_ch_structures_sn_uv(struct xpc_partition *part) |
| 90 | { | 624 | { |
| 91 | /* !!! this function needs fleshing out */ | 625 | /* !!! this function needs fleshing out */ |
| 92 | return xpUnsupported; | 626 | return xpUnsupported; |
| 93 | } | 627 | } |
| 94 | 628 | ||
| 95 | /* | 629 | /* |
| 96 | * Teardown the infrastructure necessary to support XPartition Communication | 630 | * Teardown the channel structures that are uv specific. |
| 97 | * between the specified remote partition and the local one. | ||
| 98 | */ | 631 | */ |
| 99 | static void | 632 | static void |
| 100 | xpc_teardown_infrastructure_uv(struct xpc_partition *part) | 633 | xpc_teardown_ch_structures_sn_uv(struct xpc_partition *part) |
| 101 | { | 634 | { |
| 102 | /* !!! this function needs fleshing out */ | 635 | /* !!! this function needs fleshing out */ |
| 103 | return; | 636 | return; |
| @@ -106,15 +639,163 @@ xpc_teardown_infrastructure_uv(struct xpc_partition *part) | |||
| 106 | static enum xp_retval | 639 | static enum xp_retval |
| 107 | xpc_make_first_contact_uv(struct xpc_partition *part) | 640 | xpc_make_first_contact_uv(struct xpc_partition *part) |
| 108 | { | 641 | { |
| 109 | /* !!! this function needs fleshing out */ | 642 | struct xpc_activate_mq_msg_uv msg; |
| 110 | return xpUnsupported; | 643 | |
| 644 | /* | ||
| 645 | * We send a sync msg to get the remote partition's remote_act_state | ||
| 646 | * updated to our current act_state which at this point should | ||
| 647 | * be XPC_P_AS_ACTIVATING. | ||
| 648 | */ | ||
| 649 | xpc_send_activate_IRQ_part_uv(part, &msg, sizeof(msg), | ||
| 650 | XPC_ACTIVATE_MQ_MSG_SYNC_ACT_STATE_UV); | ||
| 651 | |||
| 652 | while (part->sn.uv.remote_act_state != XPC_P_AS_ACTIVATING) { | ||
| 653 | |||
| 654 | dev_dbg(xpc_part, "waiting to make first contact with " | ||
| 655 | "partition %d\n", XPC_PARTID(part)); | ||
| 656 | |||
| 657 | /* wait a 1/4 of a second or so */ | ||
| 658 | (void)msleep_interruptible(250); | ||
| 659 | |||
| 660 | if (part->act_state == XPC_P_AS_DEACTIVATING) | ||
| 661 | return part->reason; | ||
| 662 | } | ||
| 663 | |||
| 664 | return xpSuccess; | ||
| 111 | } | 665 | } |
| 112 | 666 | ||
| 113 | static u64 | 667 | static u64 |
| 114 | xpc_get_chctl_all_flags_uv(struct xpc_partition *part) | 668 | xpc_get_chctl_all_flags_uv(struct xpc_partition *part) |
| 115 | { | 669 | { |
| 670 | unsigned long irq_flags; | ||
| 671 | union xpc_channel_ctl_flags chctl; | ||
| 672 | |||
| 673 | spin_lock_irqsave(&part->chctl_lock, irq_flags); | ||
| 674 | chctl = part->chctl; | ||
| 675 | if (chctl.all_flags != 0) | ||
| 676 | part->chctl.all_flags = 0; | ||
| 677 | |||
| 678 | spin_unlock_irqrestore(&part->chctl_lock, irq_flags); | ||
| 679 | return chctl.all_flags; | ||
| 680 | } | ||
| 681 | |||
| 682 | static enum xp_retval | ||
| 683 | xpc_setup_msg_structures_uv(struct xpc_channel *ch) | ||
| 684 | { | ||
| 685 | /* !!! this function needs fleshing out */ | ||
| 686 | return xpUnsupported; | ||
| 687 | } | ||
| 688 | |||
| 689 | static void | ||
| 690 | xpc_teardown_msg_structures_uv(struct xpc_channel *ch) | ||
| 691 | { | ||
| 692 | struct xpc_channel_uv *ch_uv = &ch->sn.uv; | ||
| 693 | |||
| 694 | ch_uv->remote_notify_mq_gpa = 0; | ||
| 695 | |||
| 116 | /* !!! this function needs fleshing out */ | 696 | /* !!! this function needs fleshing out */ |
| 117 | return 0UL; | 697 | } |
| 698 | |||
| 699 | static void | ||
| 700 | xpc_send_chctl_closerequest_uv(struct xpc_channel *ch, unsigned long *irq_flags) | ||
| 701 | { | ||
| 702 | struct xpc_activate_mq_msg_chctl_closerequest_uv msg; | ||
| 703 | |||
| 704 | msg.ch_number = ch->number; | ||
| 705 | msg.reason = ch->reason; | ||
| 706 | xpc_send_activate_IRQ_ch_uv(ch, irq_flags, &msg, sizeof(msg), | ||
| 707 | XPC_ACTIVATE_MQ_MSG_CHCTL_CLOSEREQUEST_UV); | ||
| 708 | } | ||
| 709 | |||
| 710 | static void | ||
| 711 | xpc_send_chctl_closereply_uv(struct xpc_channel *ch, unsigned long *irq_flags) | ||
| 712 | { | ||
| 713 | struct xpc_activate_mq_msg_chctl_closereply_uv msg; | ||
| 714 | |||
| 715 | msg.ch_number = ch->number; | ||
| 716 | xpc_send_activate_IRQ_ch_uv(ch, irq_flags, &msg, sizeof(msg), | ||
| 717 | XPC_ACTIVATE_MQ_MSG_CHCTL_CLOSEREPLY_UV); | ||
| 718 | } | ||
| 719 | |||
| 720 | static void | ||
| 721 | xpc_send_chctl_openrequest_uv(struct xpc_channel *ch, unsigned long *irq_flags) | ||
| 722 | { | ||
| 723 | struct xpc_activate_mq_msg_chctl_openrequest_uv msg; | ||
| 724 | |||
| 725 | msg.ch_number = ch->number; | ||
| 726 | msg.msg_size = ch->msg_size; | ||
| 727 | msg.local_nentries = ch->local_nentries; | ||
| 728 | xpc_send_activate_IRQ_ch_uv(ch, irq_flags, &msg, sizeof(msg), | ||
| 729 | XPC_ACTIVATE_MQ_MSG_CHCTL_OPENREQUEST_UV); | ||
| 730 | } | ||
| 731 | |||
| 732 | static void | ||
| 733 | xpc_send_chctl_openreply_uv(struct xpc_channel *ch, unsigned long *irq_flags) | ||
| 734 | { | ||
| 735 | struct xpc_activate_mq_msg_chctl_openreply_uv msg; | ||
| 736 | |||
| 737 | msg.ch_number = ch->number; | ||
| 738 | msg.local_nentries = ch->local_nentries; | ||
| 739 | msg.remote_nentries = ch->remote_nentries; | ||
| 740 | msg.local_notify_mq_gpa = uv_gpa(xpc_notify_mq_uv); | ||
| 741 | xpc_send_activate_IRQ_ch_uv(ch, irq_flags, &msg, sizeof(msg), | ||
| 742 | XPC_ACTIVATE_MQ_MSG_CHCTL_OPENREPLY_UV); | ||
| 743 | } | ||
| 744 | |||
| 745 | static void | ||
| 746 | xpc_save_remote_msgqueue_pa_uv(struct xpc_channel *ch, | ||
| 747 | unsigned long msgqueue_pa) | ||
| 748 | { | ||
| 749 | ch->sn.uv.remote_notify_mq_gpa = msgqueue_pa; | ||
| 750 | } | ||
| 751 | |||
| 752 | static void | ||
| 753 | xpc_indicate_partition_engaged_uv(struct xpc_partition *part) | ||
| 754 | { | ||
| 755 | struct xpc_activate_mq_msg_uv msg; | ||
| 756 | |||
| 757 | xpc_send_activate_IRQ_part_uv(part, &msg, sizeof(msg), | ||
| 758 | XPC_ACTIVATE_MQ_MSG_MARK_ENGAGED_UV); | ||
| 759 | } | ||
| 760 | |||
| 761 | static void | ||
| 762 | xpc_indicate_partition_disengaged_uv(struct xpc_partition *part) | ||
| 763 | { | ||
| 764 | struct xpc_activate_mq_msg_uv msg; | ||
| 765 | |||
| 766 | xpc_send_activate_IRQ_part_uv(part, &msg, sizeof(msg), | ||
| 767 | XPC_ACTIVATE_MQ_MSG_MARK_DISENGAGED_UV); | ||
| 768 | } | ||
| 769 | |||
| 770 | static void | ||
| 771 | xpc_assume_partition_disengaged_uv(short partid) | ||
| 772 | { | ||
| 773 | struct xpc_partition_uv *part_uv = &xpc_partitions[partid].sn.uv; | ||
| 774 | unsigned long irq_flags; | ||
| 775 | |||
| 776 | spin_lock_irqsave(&part_uv->flags_lock, irq_flags); | ||
| 777 | part_uv->flags &= ~XPC_P_ENGAGED_UV; | ||
| 778 | spin_unlock_irqrestore(&part_uv->flags_lock, irq_flags); | ||
| 779 | } | ||
| 780 | |||
| 781 | static int | ||
| 782 | xpc_partition_engaged_uv(short partid) | ||
| 783 | { | ||
| 784 | return (xpc_partitions[partid].sn.uv.flags & XPC_P_ENGAGED_UV) != 0; | ||
| 785 | } | ||
| 786 | |||
| 787 | static int | ||
| 788 | xpc_any_partition_engaged_uv(void) | ||
| 789 | { | ||
| 790 | struct xpc_partition_uv *part_uv; | ||
| 791 | short partid; | ||
| 792 | |||
| 793 | for (partid = 0; partid < XP_MAX_NPARTITIONS_UV; partid++) { | ||
| 794 | part_uv = &xpc_partitions[partid].sn.uv; | ||
| 795 | if ((part_uv->flags & XPC_P_ENGAGED_UV) != 0) | ||
| 796 | return 1; | ||
| 797 | } | ||
| 798 | return 0; | ||
| 118 | } | 799 | } |
| 119 | 800 | ||
| 120 | static struct xpc_msg * | 801 | static struct xpc_msg * |
| @@ -124,24 +805,64 @@ xpc_get_deliverable_msg_uv(struct xpc_channel *ch) | |||
| 124 | return NULL; | 805 | return NULL; |
| 125 | } | 806 | } |
| 126 | 807 | ||
| 127 | void | 808 | int |
| 128 | xpc_init_uv(void) | 809 | xpc_init_uv(void) |
| 129 | { | 810 | { |
| 130 | xpc_rsvd_page_init = xpc_rsvd_page_init_uv; | 811 | xpc_setup_partitions_sn = xpc_setup_partitions_sn_uv; |
| 812 | xpc_process_activate_IRQ_rcvd = xpc_process_activate_IRQ_rcvd_uv; | ||
| 813 | xpc_get_partition_rsvd_page_pa = xpc_get_partition_rsvd_page_pa_uv; | ||
| 814 | xpc_setup_rsvd_page_sn = xpc_setup_rsvd_page_sn_uv; | ||
| 131 | xpc_increment_heartbeat = xpc_increment_heartbeat_uv; | 815 | xpc_increment_heartbeat = xpc_increment_heartbeat_uv; |
| 816 | xpc_offline_heartbeat = xpc_offline_heartbeat_uv; | ||
| 817 | xpc_online_heartbeat = xpc_online_heartbeat_uv; | ||
| 132 | xpc_heartbeat_init = xpc_heartbeat_init_uv; | 818 | xpc_heartbeat_init = xpc_heartbeat_init_uv; |
| 133 | xpc_heartbeat_exit = xpc_heartbeat_exit_uv; | 819 | xpc_heartbeat_exit = xpc_heartbeat_exit_uv; |
| 820 | xpc_get_remote_heartbeat = xpc_get_remote_heartbeat_uv; | ||
| 821 | |||
| 134 | xpc_request_partition_activation = xpc_request_partition_activation_uv; | 822 | xpc_request_partition_activation = xpc_request_partition_activation_uv; |
| 135 | xpc_request_partition_reactivation = | 823 | xpc_request_partition_reactivation = |
| 136 | xpc_request_partition_reactivation_uv; | 824 | xpc_request_partition_reactivation_uv; |
| 137 | xpc_setup_infrastructure = xpc_setup_infrastructure_uv; | 825 | xpc_request_partition_deactivation = |
| 138 | xpc_teardown_infrastructure = xpc_teardown_infrastructure_uv; | 826 | xpc_request_partition_deactivation_uv; |
| 827 | |||
| 828 | xpc_setup_ch_structures_sn = xpc_setup_ch_structures_sn_uv; | ||
| 829 | xpc_teardown_ch_structures_sn = xpc_teardown_ch_structures_sn_uv; | ||
| 830 | |||
| 139 | xpc_make_first_contact = xpc_make_first_contact_uv; | 831 | xpc_make_first_contact = xpc_make_first_contact_uv; |
| 832 | |||
| 140 | xpc_get_chctl_all_flags = xpc_get_chctl_all_flags_uv; | 833 | xpc_get_chctl_all_flags = xpc_get_chctl_all_flags_uv; |
| 834 | xpc_send_chctl_closerequest = xpc_send_chctl_closerequest_uv; | ||
| 835 | xpc_send_chctl_closereply = xpc_send_chctl_closereply_uv; | ||
| 836 | xpc_send_chctl_openrequest = xpc_send_chctl_openrequest_uv; | ||
| 837 | xpc_send_chctl_openreply = xpc_send_chctl_openreply_uv; | ||
| 838 | |||
| 839 | xpc_save_remote_msgqueue_pa = xpc_save_remote_msgqueue_pa_uv; | ||
| 840 | |||
| 841 | xpc_setup_msg_structures = xpc_setup_msg_structures_uv; | ||
| 842 | xpc_teardown_msg_structures = xpc_teardown_msg_structures_uv; | ||
| 843 | |||
| 844 | xpc_indicate_partition_engaged = xpc_indicate_partition_engaged_uv; | ||
| 845 | xpc_indicate_partition_disengaged = | ||
| 846 | xpc_indicate_partition_disengaged_uv; | ||
| 847 | xpc_assume_partition_disengaged = xpc_assume_partition_disengaged_uv; | ||
| 848 | xpc_partition_engaged = xpc_partition_engaged_uv; | ||
| 849 | xpc_any_partition_engaged = xpc_any_partition_engaged_uv; | ||
| 850 | |||
| 141 | xpc_get_deliverable_msg = xpc_get_deliverable_msg_uv; | 851 | xpc_get_deliverable_msg = xpc_get_deliverable_msg_uv; |
| 852 | |||
| 853 | /* ??? The cpuid argument's value is 0, is that what we want? */ | ||
| 854 | /* !!! The irq argument's value isn't correct. */ | ||
| 855 | xpc_activate_mq_uv = xpc_create_gru_mq_uv(XPC_ACTIVATE_MQ_SIZE_UV, 0, 0, | ||
| 856 | xpc_handle_activate_IRQ_uv); | ||
| 857 | if (xpc_activate_mq_uv == NULL) | ||
| 858 | return -ENOMEM; | ||
| 859 | |||
| 860 | return 0; | ||
| 142 | } | 861 | } |
| 143 | 862 | ||
| 144 | void | 863 | void |
| 145 | xpc_exit_uv(void) | 864 | xpc_exit_uv(void) |
| 146 | { | 865 | { |
| 866 | /* !!! The irq argument's value isn't correct. */ | ||
| 867 | xpc_destroy_gru_mq_uv(xpc_activate_mq_uv, XPC_ACTIVATE_MQ_SIZE_UV, 0); | ||
| 147 | } | 868 | } |
