diff options
| -rw-r--r-- | arch/x86/include/asm/uv/uv_bau.h | 9 | ||||
| -rw-r--r-- | arch/x86/kernel/apic/x2apic_uv_x.c | 61 | ||||
| -rw-r--r-- | arch/x86/platform/uv/tlb_uv.c | 22 |
3 files changed, 57 insertions, 35 deletions
diff --git a/arch/x86/include/asm/uv/uv_bau.h b/arch/x86/include/asm/uv/uv_bau.h index 42d412fd8b02..ce1d54c8a433 100644 --- a/arch/x86/include/asm/uv/uv_bau.h +++ b/arch/x86/include/asm/uv/uv_bau.h | |||
| @@ -26,20 +26,22 @@ | |||
| 26 | * BAU_SB_DESCRIPTOR_BASE register, set 1 is located at BASE + 512, | 26 | * BAU_SB_DESCRIPTOR_BASE register, set 1 is located at BASE + 512, |
| 27 | * set 2 is at BASE + 2*512, set 3 at BASE + 3*512, and so on. | 27 | * set 2 is at BASE + 2*512, set 3 at BASE + 3*512, and so on. |
| 28 | * | 28 | * |
| 29 | * We will use 31 sets, one for sending BAU messages from each of the 32 | 29 | * We will use one set for sending BAU messages from each of the |
| 30 | * cpu's on the uvhub. | 30 | * cpu's on the uvhub. |
| 31 | * | 31 | * |
| 32 | * TLB shootdown will use the first of the 8 descriptors of each set. | 32 | * TLB shootdown will use the first of the 8 descriptors of each set. |
| 33 | * Each of the descriptors is 64 bytes in size (8*64 = 512 bytes in a set). | 33 | * Each of the descriptors is 64 bytes in size (8*64 = 512 bytes in a set). |
| 34 | */ | 34 | */ |
| 35 | 35 | ||
| 36 | #define MAX_CPUS_PER_UVHUB 64 | ||
| 37 | #define MAX_CPUS_PER_SOCKET 32 | ||
| 38 | #define UV_ADP_SIZE 64 /* hardware-provided max. */ | ||
| 39 | #define UV_CPUS_PER_ACT_STATUS 32 /* hardware-provided max. */ | ||
| 36 | #define UV_ITEMS_PER_DESCRIPTOR 8 | 40 | #define UV_ITEMS_PER_DESCRIPTOR 8 |
| 37 | /* the 'throttle' to prevent the hardware stay-busy bug */ | 41 | /* the 'throttle' to prevent the hardware stay-busy bug */ |
| 38 | #define MAX_BAU_CONCURRENT 3 | 42 | #define MAX_BAU_CONCURRENT 3 |
| 39 | #define UV_CPUS_PER_ACT_STATUS 32 | ||
| 40 | #define UV_ACT_STATUS_MASK 0x3 | 43 | #define UV_ACT_STATUS_MASK 0x3 |
| 41 | #define UV_ACT_STATUS_SIZE 2 | 44 | #define UV_ACT_STATUS_SIZE 2 |
| 42 | #define UV_ADP_SIZE 32 | ||
| 43 | #define UV_DISTRIBUTION_SIZE 256 | 45 | #define UV_DISTRIBUTION_SIZE 256 |
| 44 | #define UV_SW_ACK_NPENDING 8 | 46 | #define UV_SW_ACK_NPENDING 8 |
| 45 | #define UV_NET_ENDPOINT_INTD 0x38 | 47 | #define UV_NET_ENDPOINT_INTD 0x38 |
| @@ -100,7 +102,6 @@ | |||
| 100 | * number of destination side software ack resources | 102 | * number of destination side software ack resources |
| 101 | */ | 103 | */ |
| 102 | #define DEST_NUM_RESOURCES 8 | 104 | #define DEST_NUM_RESOURCES 8 |
| 103 | #define MAX_CPUS_PER_NODE 32 | ||
| 104 | /* | 105 | /* |
| 105 | * completion statuses for sending a TLB flush message | 106 | * completion statuses for sending a TLB flush message |
| 106 | */ | 107 | */ |
diff --git a/arch/x86/kernel/apic/x2apic_uv_x.c b/arch/x86/kernel/apic/x2apic_uv_x.c index c1c52c341f40..2a3f2a7db243 100644 --- a/arch/x86/kernel/apic/x2apic_uv_x.c +++ b/arch/x86/kernel/apic/x2apic_uv_x.c | |||
| @@ -48,6 +48,16 @@ unsigned int uv_apicid_hibits; | |||
| 48 | EXPORT_SYMBOL_GPL(uv_apicid_hibits); | 48 | EXPORT_SYMBOL_GPL(uv_apicid_hibits); |
| 49 | static DEFINE_SPINLOCK(uv_nmi_lock); | 49 | static DEFINE_SPINLOCK(uv_nmi_lock); |
| 50 | 50 | ||
| 51 | static unsigned long __init uv_early_read_mmr(unsigned long addr) | ||
| 52 | { | ||
| 53 | unsigned long val, *mmr; | ||
| 54 | |||
| 55 | mmr = early_ioremap(UV_LOCAL_MMR_BASE | addr, sizeof(*mmr)); | ||
| 56 | val = *mmr; | ||
| 57 | early_iounmap(mmr, sizeof(*mmr)); | ||
| 58 | return val; | ||
| 59 | } | ||
| 60 | |||
| 51 | static inline bool is_GRU_range(u64 start, u64 end) | 61 | static inline bool is_GRU_range(u64 start, u64 end) |
| 52 | { | 62 | { |
| 53 | return start >= gru_start_paddr && end <= gru_end_paddr; | 63 | return start >= gru_start_paddr && end <= gru_end_paddr; |
| @@ -58,28 +68,24 @@ static bool uv_is_untracked_pat_range(u64 start, u64 end) | |||
| 58 | return is_ISA_range(start, end) || is_GRU_range(start, end); | 68 | return is_ISA_range(start, end) || is_GRU_range(start, end); |
| 59 | } | 69 | } |
| 60 | 70 | ||
| 61 | static int early_get_nodeid(void) | 71 | static int __init early_get_pnodeid(void) |
| 62 | { | 72 | { |
| 63 | union uvh_node_id_u node_id; | 73 | union uvh_node_id_u node_id; |
| 64 | unsigned long *mmr; | 74 | union uvh_rh_gam_config_mmr_u m_n_config; |
| 65 | 75 | int pnode; | |
| 66 | mmr = early_ioremap(UV_LOCAL_MMR_BASE | UVH_NODE_ID, sizeof(*mmr)); | ||
| 67 | node_id.v = *mmr; | ||
| 68 | early_iounmap(mmr, sizeof(*mmr)); | ||
| 69 | 76 | ||
| 70 | /* Currently, all blades have same revision number */ | 77 | /* Currently, all blades have same revision number */ |
| 78 | node_id.v = uv_early_read_mmr(UVH_NODE_ID); | ||
| 79 | m_n_config.v = uv_early_read_mmr(UVH_RH_GAM_CONFIG_MMR); | ||
| 71 | uv_min_hub_revision_id = node_id.s.revision; | 80 | uv_min_hub_revision_id = node_id.s.revision; |
| 72 | 81 | ||
| 73 | return node_id.s.node_id; | 82 | pnode = (node_id.s.node_id >> 1) & ((1 << m_n_config.s.n_skt) - 1); |
| 83 | return pnode; | ||
| 74 | } | 84 | } |
| 75 | 85 | ||
| 76 | static void __init early_get_apic_pnode_shift(void) | 86 | static void __init early_get_apic_pnode_shift(void) |
| 77 | { | 87 | { |
| 78 | unsigned long *mmr; | 88 | uvh_apicid.v = uv_early_read_mmr(UVH_APICID); |
| 79 | |||
| 80 | mmr = early_ioremap(UV_LOCAL_MMR_BASE | UVH_APICID, sizeof(*mmr)); | ||
| 81 | uvh_apicid.v = *mmr; | ||
| 82 | early_iounmap(mmr, sizeof(*mmr)); | ||
| 83 | if (!uvh_apicid.v) | 89 | if (!uvh_apicid.v) |
| 84 | /* | 90 | /* |
| 85 | * Old bios, use default value | 91 | * Old bios, use default value |
| @@ -95,21 +101,17 @@ static void __init early_get_apic_pnode_shift(void) | |||
| 95 | static void __init uv_set_apicid_hibit(void) | 101 | static void __init uv_set_apicid_hibit(void) |
| 96 | { | 102 | { |
| 97 | union uvh_lb_target_physical_apic_id_mask_u apicid_mask; | 103 | union uvh_lb_target_physical_apic_id_mask_u apicid_mask; |
| 98 | unsigned long *mmr; | ||
| 99 | 104 | ||
| 100 | mmr = early_ioremap(UV_LOCAL_MMR_BASE | | 105 | apicid_mask.v = uv_early_read_mmr(UVH_LB_TARGET_PHYSICAL_APIC_ID_MASK); |
| 101 | UVH_LB_TARGET_PHYSICAL_APIC_ID_MASK, sizeof(*mmr)); | ||
| 102 | apicid_mask.v = *mmr; | ||
| 103 | early_iounmap(mmr, sizeof(*mmr)); | ||
| 104 | uv_apicid_hibits = apicid_mask.s.bit_enables & UV_APICID_HIBIT_MASK; | 106 | uv_apicid_hibits = apicid_mask.s.bit_enables & UV_APICID_HIBIT_MASK; |
| 105 | } | 107 | } |
| 106 | 108 | ||
| 107 | static int __init uv_acpi_madt_oem_check(char *oem_id, char *oem_table_id) | 109 | static int __init uv_acpi_madt_oem_check(char *oem_id, char *oem_table_id) |
| 108 | { | 110 | { |
| 109 | int nodeid; | 111 | int pnodeid; |
| 110 | 112 | ||
| 111 | if (!strcmp(oem_id, "SGI")) { | 113 | if (!strcmp(oem_id, "SGI")) { |
| 112 | nodeid = early_get_nodeid(); | 114 | pnodeid = early_get_pnodeid(); |
| 113 | early_get_apic_pnode_shift(); | 115 | early_get_apic_pnode_shift(); |
| 114 | x86_platform.is_untracked_pat_range = uv_is_untracked_pat_range; | 116 | x86_platform.is_untracked_pat_range = uv_is_untracked_pat_range; |
| 115 | x86_platform.nmi_init = uv_nmi_init; | 117 | x86_platform.nmi_init = uv_nmi_init; |
| @@ -119,7 +121,7 @@ static int __init uv_acpi_madt_oem_check(char *oem_id, char *oem_table_id) | |||
| 119 | uv_system_type = UV_X2APIC; | 121 | uv_system_type = UV_X2APIC; |
| 120 | else if (!strcmp(oem_table_id, "UVH")) { | 122 | else if (!strcmp(oem_table_id, "UVH")) { |
| 121 | __get_cpu_var(x2apic_extra_bits) = | 123 | __get_cpu_var(x2apic_extra_bits) = |
| 122 | nodeid << (uvh_apicid.s.pnode_shift - 1); | 124 | pnodeid << uvh_apicid.s.pnode_shift; |
| 123 | uv_system_type = UV_NON_UNIQUE_APIC; | 125 | uv_system_type = UV_NON_UNIQUE_APIC; |
| 124 | uv_set_apicid_hibit(); | 126 | uv_set_apicid_hibit(); |
| 125 | return 1; | 127 | return 1; |
| @@ -682,27 +684,32 @@ void uv_nmi_init(void) | |||
| 682 | void __init uv_system_init(void) | 684 | void __init uv_system_init(void) |
| 683 | { | 685 | { |
| 684 | union uvh_rh_gam_config_mmr_u m_n_config; | 686 | union uvh_rh_gam_config_mmr_u m_n_config; |
| 687 | union uvh_rh_gam_mmioh_overlay_config_mmr_u mmioh; | ||
| 685 | union uvh_node_id_u node_id; | 688 | union uvh_node_id_u node_id; |
| 686 | unsigned long gnode_upper, lowmem_redir_base, lowmem_redir_size; | 689 | unsigned long gnode_upper, lowmem_redir_base, lowmem_redir_size; |
| 687 | int bytes, nid, cpu, lcpu, pnode, blade, i, j, m_val, n_val; | 690 | int bytes, nid, cpu, lcpu, pnode, blade, i, j, m_val, n_val, n_io; |
| 688 | int gnode_extra, max_pnode = 0; | 691 | int gnode_extra, max_pnode = 0; |
| 689 | unsigned long mmr_base, present, paddr; | 692 | unsigned long mmr_base, present, paddr; |
| 690 | unsigned short pnode_mask; | 693 | unsigned short pnode_mask, pnode_io_mask; |
| 691 | 694 | ||
| 692 | map_low_mmrs(); | 695 | map_low_mmrs(); |
| 693 | 696 | ||
| 694 | m_n_config.v = uv_read_local_mmr(UVH_RH_GAM_CONFIG_MMR ); | 697 | m_n_config.v = uv_read_local_mmr(UVH_RH_GAM_CONFIG_MMR ); |
| 695 | m_val = m_n_config.s.m_skt; | 698 | m_val = m_n_config.s.m_skt; |
| 696 | n_val = m_n_config.s.n_skt; | 699 | n_val = m_n_config.s.n_skt; |
| 700 | mmioh.v = uv_read_local_mmr(UVH_RH_GAM_MMIOH_OVERLAY_CONFIG_MMR); | ||
| 701 | n_io = mmioh.s.n_io; | ||
| 697 | mmr_base = | 702 | mmr_base = |
| 698 | uv_read_local_mmr(UVH_RH_GAM_MMR_OVERLAY_CONFIG_MMR) & | 703 | uv_read_local_mmr(UVH_RH_GAM_MMR_OVERLAY_CONFIG_MMR) & |
| 699 | ~UV_MMR_ENABLE; | 704 | ~UV_MMR_ENABLE; |
| 700 | pnode_mask = (1 << n_val) - 1; | 705 | pnode_mask = (1 << n_val) - 1; |
| 706 | pnode_io_mask = (1 << n_io) - 1; | ||
| 707 | |||
| 701 | node_id.v = uv_read_local_mmr(UVH_NODE_ID); | 708 | node_id.v = uv_read_local_mmr(UVH_NODE_ID); |
| 702 | gnode_extra = (node_id.s.node_id & ~((1 << n_val) - 1)) >> 1; | 709 | gnode_extra = (node_id.s.node_id & ~((1 << n_val) - 1)) >> 1; |
| 703 | gnode_upper = ((unsigned long)gnode_extra << m_val); | 710 | gnode_upper = ((unsigned long)gnode_extra << m_val); |
| 704 | printk(KERN_DEBUG "UV: N %d, M %d, gnode_upper 0x%lx, gnode_extra 0x%x\n", | 711 | printk(KERN_INFO "UV: N %d, M %d, N_IO: %d, gnode_upper 0x%lx, gnode_extra 0x%x, pnode_mask 0x%x, pnode_io_mask 0x%x\n", |
| 705 | n_val, m_val, gnode_upper, gnode_extra); | 712 | n_val, m_val, n_io, gnode_upper, gnode_extra, pnode_mask, pnode_io_mask); |
| 706 | 713 | ||
| 707 | printk(KERN_DEBUG "UV: global MMR base 0x%lx\n", mmr_base); | 714 | printk(KERN_DEBUG "UV: global MMR base 0x%lx\n", mmr_base); |
| 708 | 715 | ||
| @@ -735,7 +742,7 @@ void __init uv_system_init(void) | |||
| 735 | for (j = 0; j < 64; j++) { | 742 | for (j = 0; j < 64; j++) { |
| 736 | if (!test_bit(j, &present)) | 743 | if (!test_bit(j, &present)) |
| 737 | continue; | 744 | continue; |
| 738 | pnode = (i * 64 + j); | 745 | pnode = (i * 64 + j) & pnode_mask; |
| 739 | uv_blade_info[blade].pnode = pnode; | 746 | uv_blade_info[blade].pnode = pnode; |
| 740 | uv_blade_info[blade].nr_possible_cpus = 0; | 747 | uv_blade_info[blade].nr_possible_cpus = 0; |
| 741 | uv_blade_info[blade].nr_online_cpus = 0; | 748 | uv_blade_info[blade].nr_online_cpus = 0; |
| @@ -756,6 +763,7 @@ void __init uv_system_init(void) | |||
| 756 | /* | 763 | /* |
| 757 | * apic_pnode_shift must be set before calling uv_apicid_to_pnode(); | 764 | * apic_pnode_shift must be set before calling uv_apicid_to_pnode(); |
| 758 | */ | 765 | */ |
| 766 | uv_cpu_hub_info(cpu)->pnode_mask = pnode_mask; | ||
| 759 | uv_cpu_hub_info(cpu)->apic_pnode_shift = uvh_apicid.s.pnode_shift; | 767 | uv_cpu_hub_info(cpu)->apic_pnode_shift = uvh_apicid.s.pnode_shift; |
| 760 | pnode = uv_apicid_to_pnode(apicid); | 768 | pnode = uv_apicid_to_pnode(apicid); |
| 761 | blade = boot_pnode_to_blade(pnode); | 769 | blade = boot_pnode_to_blade(pnode); |
| @@ -772,7 +780,6 @@ void __init uv_system_init(void) | |||
| 772 | uv_cpu_hub_info(cpu)->numa_blade_id = blade; | 780 | uv_cpu_hub_info(cpu)->numa_blade_id = blade; |
| 773 | uv_cpu_hub_info(cpu)->blade_processor_id = lcpu; | 781 | uv_cpu_hub_info(cpu)->blade_processor_id = lcpu; |
| 774 | uv_cpu_hub_info(cpu)->pnode = pnode; | 782 | uv_cpu_hub_info(cpu)->pnode = pnode; |
| 775 | uv_cpu_hub_info(cpu)->pnode_mask = pnode_mask; | ||
| 776 | uv_cpu_hub_info(cpu)->gpa_mask = (1UL << (m_val + n_val)) - 1; | 783 | uv_cpu_hub_info(cpu)->gpa_mask = (1UL << (m_val + n_val)) - 1; |
| 777 | uv_cpu_hub_info(cpu)->gnode_upper = gnode_upper; | 784 | uv_cpu_hub_info(cpu)->gnode_upper = gnode_upper; |
| 778 | uv_cpu_hub_info(cpu)->gnode_extra = gnode_extra; | 785 | uv_cpu_hub_info(cpu)->gnode_extra = gnode_extra; |
| @@ -796,7 +803,7 @@ void __init uv_system_init(void) | |||
| 796 | 803 | ||
| 797 | map_gru_high(max_pnode); | 804 | map_gru_high(max_pnode); |
| 798 | map_mmr_high(max_pnode); | 805 | map_mmr_high(max_pnode); |
| 799 | map_mmioh_high(max_pnode); | 806 | map_mmioh_high(max_pnode & pnode_io_mask); |
| 800 | 807 | ||
| 801 | uv_cpu_init(); | 808 | uv_cpu_init(); |
| 802 | uv_scir_register_cpu_notifier(); | 809 | uv_scir_register_cpu_notifier(); |
diff --git a/arch/x86/platform/uv/tlb_uv.c b/arch/x86/platform/uv/tlb_uv.c index ba9caa808a9c..df58e9cad96a 100644 --- a/arch/x86/platform/uv/tlb_uv.c +++ b/arch/x86/platform/uv/tlb_uv.c | |||
| @@ -1341,7 +1341,7 @@ uv_activation_descriptor_init(int node, int pnode) | |||
| 1341 | 1341 | ||
| 1342 | /* | 1342 | /* |
| 1343 | * each bau_desc is 64 bytes; there are 8 (UV_ITEMS_PER_DESCRIPTOR) | 1343 | * each bau_desc is 64 bytes; there are 8 (UV_ITEMS_PER_DESCRIPTOR) |
| 1344 | * per cpu; and up to 32 (UV_ADP_SIZE) cpu's per uvhub | 1344 | * per cpu; and one per cpu on the uvhub (UV_ADP_SIZE) |
| 1345 | */ | 1345 | */ |
| 1346 | bau_desc = kmalloc_node(sizeof(struct bau_desc) * UV_ADP_SIZE | 1346 | bau_desc = kmalloc_node(sizeof(struct bau_desc) * UV_ADP_SIZE |
| 1347 | * UV_ITEMS_PER_DESCRIPTOR, GFP_KERNEL, node); | 1347 | * UV_ITEMS_PER_DESCRIPTOR, GFP_KERNEL, node); |
| @@ -1490,7 +1490,7 @@ calculate_destination_timeout(void) | |||
| 1490 | /* | 1490 | /* |
| 1491 | * initialize the bau_control structure for each cpu | 1491 | * initialize the bau_control structure for each cpu |
| 1492 | */ | 1492 | */ |
| 1493 | static void __init uv_init_per_cpu(int nuvhubs) | 1493 | static int __init uv_init_per_cpu(int nuvhubs) |
| 1494 | { | 1494 | { |
| 1495 | int i; | 1495 | int i; |
| 1496 | int cpu; | 1496 | int cpu; |
| @@ -1507,7 +1507,7 @@ static void __init uv_init_per_cpu(int nuvhubs) | |||
| 1507 | struct bau_control *smaster = NULL; | 1507 | struct bau_control *smaster = NULL; |
| 1508 | struct socket_desc { | 1508 | struct socket_desc { |
| 1509 | short num_cpus; | 1509 | short num_cpus; |
| 1510 | short cpu_number[16]; | 1510 | short cpu_number[MAX_CPUS_PER_SOCKET]; |
| 1511 | }; | 1511 | }; |
| 1512 | struct uvhub_desc { | 1512 | struct uvhub_desc { |
| 1513 | unsigned short socket_mask; | 1513 | unsigned short socket_mask; |
| @@ -1540,6 +1540,10 @@ static void __init uv_init_per_cpu(int nuvhubs) | |||
| 1540 | sdp = &bdp->socket[socket]; | 1540 | sdp = &bdp->socket[socket]; |
| 1541 | sdp->cpu_number[sdp->num_cpus] = cpu; | 1541 | sdp->cpu_number[sdp->num_cpus] = cpu; |
| 1542 | sdp->num_cpus++; | 1542 | sdp->num_cpus++; |
| 1543 | if (sdp->num_cpus > MAX_CPUS_PER_SOCKET) { | ||
| 1544 | printk(KERN_EMERG "%d cpus per socket invalid\n", sdp->num_cpus); | ||
| 1545 | return 1; | ||
| 1546 | } | ||
| 1543 | } | 1547 | } |
| 1544 | for (uvhub = 0; uvhub < nuvhubs; uvhub++) { | 1548 | for (uvhub = 0; uvhub < nuvhubs; uvhub++) { |
| 1545 | if (!(*(uvhub_mask + (uvhub/8)) & (1 << (uvhub%8)))) | 1549 | if (!(*(uvhub_mask + (uvhub/8)) & (1 << (uvhub%8)))) |
| @@ -1570,6 +1574,12 @@ static void __init uv_init_per_cpu(int nuvhubs) | |||
| 1570 | bcp->uvhub_master = hmaster; | 1574 | bcp->uvhub_master = hmaster; |
| 1571 | bcp->uvhub_cpu = uv_cpu_hub_info(cpu)-> | 1575 | bcp->uvhub_cpu = uv_cpu_hub_info(cpu)-> |
| 1572 | blade_processor_id; | 1576 | blade_processor_id; |
| 1577 | if (bcp->uvhub_cpu >= MAX_CPUS_PER_UVHUB) { | ||
| 1578 | printk(KERN_EMERG | ||
| 1579 | "%d cpus per uvhub invalid\n", | ||
| 1580 | bcp->uvhub_cpu); | ||
| 1581 | return 1; | ||
| 1582 | } | ||
| 1573 | } | 1583 | } |
| 1574 | nextsocket: | 1584 | nextsocket: |
| 1575 | socket++; | 1585 | socket++; |
| @@ -1595,6 +1605,7 @@ nextsocket: | |||
| 1595 | bcp->congested_reps = congested_reps; | 1605 | bcp->congested_reps = congested_reps; |
| 1596 | bcp->congested_period = congested_period; | 1606 | bcp->congested_period = congested_period; |
| 1597 | } | 1607 | } |
| 1608 | return 0; | ||
| 1598 | } | 1609 | } |
| 1599 | 1610 | ||
| 1600 | /* | 1611 | /* |
| @@ -1625,7 +1636,10 @@ static int __init uv_bau_init(void) | |||
| 1625 | spin_lock_init(&disable_lock); | 1636 | spin_lock_init(&disable_lock); |
| 1626 | congested_cycles = microsec_2_cycles(congested_response_us); | 1637 | congested_cycles = microsec_2_cycles(congested_response_us); |
| 1627 | 1638 | ||
| 1628 | uv_init_per_cpu(nuvhubs); | 1639 | if (uv_init_per_cpu(nuvhubs)) { |
| 1640 | nobau = 1; | ||
| 1641 | return 0; | ||
| 1642 | } | ||
| 1629 | 1643 | ||
| 1630 | uv_partition_base_pnode = 0x7fffffff; | 1644 | uv_partition_base_pnode = 0x7fffffff; |
| 1631 | for (uvhub = 0; uvhub < nuvhubs; uvhub++) | 1645 | for (uvhub = 0; uvhub < nuvhubs; uvhub++) |
