diff options
171 files changed, 1886 insertions, 671 deletions
| @@ -1172,7 +1172,7 @@ headers_check_all: headers_install_all | |||
| 1172 | PHONY += headers_check | 1172 | PHONY += headers_check |
| 1173 | headers_check: headers_install | 1173 | headers_check: headers_install |
| 1174 | $(Q)$(MAKE) $(hdr-inst)=include/uapi HDRCHECK=1 | 1174 | $(Q)$(MAKE) $(hdr-inst)=include/uapi HDRCHECK=1 |
| 1175 | $(Q)$(MAKE) $(hdr-inst)=arch/$(hdr-arch)/include/uapi/ $(hdr-dst) HDRCHECK=1 | 1175 | $(Q)$(MAKE) $(hdr-inst)=arch/$(hdr-arch)/include/uapi $(hdr-dst) HDRCHECK=1 |
| 1176 | 1176 | ||
| 1177 | # --------------------------------------------------------------------------- | 1177 | # --------------------------------------------------------------------------- |
| 1178 | # Kernel selftest | 1178 | # Kernel selftest |
diff --git a/arch/arm64/net/bpf_jit_comp.c b/arch/arm64/net/bpf_jit_comp.c index c6e53580aefe..71f930501ade 100644 --- a/arch/arm64/net/bpf_jit_comp.c +++ b/arch/arm64/net/bpf_jit_comp.c | |||
| @@ -253,8 +253,9 @@ static int emit_bpf_tail_call(struct jit_ctx *ctx) | |||
| 253 | */ | 253 | */ |
| 254 | off = offsetof(struct bpf_array, ptrs); | 254 | off = offsetof(struct bpf_array, ptrs); |
| 255 | emit_a64_mov_i64(tmp, off, ctx); | 255 | emit_a64_mov_i64(tmp, off, ctx); |
| 256 | emit(A64_LDR64(tmp, r2, tmp), ctx); | 256 | emit(A64_ADD(1, tmp, r2, tmp), ctx); |
| 257 | emit(A64_LDR64(prg, tmp, r3), ctx); | 257 | emit(A64_LSL(1, prg, r3, 3), ctx); |
| 258 | emit(A64_LDR64(prg, tmp, prg), ctx); | ||
| 258 | emit(A64_CBZ(1, prg, jmp_offset), ctx); | 259 | emit(A64_CBZ(1, prg, jmp_offset), ctx); |
| 259 | 260 | ||
| 260 | /* goto *(prog->bpf_func + prologue_size); */ | 261 | /* goto *(prog->bpf_func + prologue_size); */ |
diff --git a/arch/powerpc/include/asm/module.h b/arch/powerpc/include/asm/module.h index 53885512b8d3..6c0132c7212f 100644 --- a/arch/powerpc/include/asm/module.h +++ b/arch/powerpc/include/asm/module.h | |||
| @@ -14,6 +14,10 @@ | |||
| 14 | #include <asm-generic/module.h> | 14 | #include <asm-generic/module.h> |
| 15 | 15 | ||
| 16 | 16 | ||
| 17 | #ifdef CC_USING_MPROFILE_KERNEL | ||
| 18 | #define MODULE_ARCH_VERMAGIC "mprofile-kernel" | ||
| 19 | #endif | ||
| 20 | |||
| 17 | #ifndef __powerpc64__ | 21 | #ifndef __powerpc64__ |
| 18 | /* | 22 | /* |
| 19 | * Thanks to Paul M for explaining this. | 23 | * Thanks to Paul M for explaining this. |
diff --git a/arch/powerpc/include/asm/page.h b/arch/powerpc/include/asm/page.h index 2a32483c7b6c..8da5d4c1cab2 100644 --- a/arch/powerpc/include/asm/page.h +++ b/arch/powerpc/include/asm/page.h | |||
| @@ -132,7 +132,19 @@ extern long long virt_phys_offset; | |||
| 132 | #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) | 132 | #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) |
| 133 | #define virt_to_page(kaddr) pfn_to_page(virt_to_pfn(kaddr)) | 133 | #define virt_to_page(kaddr) pfn_to_page(virt_to_pfn(kaddr)) |
| 134 | #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) | 134 | #define pfn_to_kaddr(pfn) __va((pfn) << PAGE_SHIFT) |
| 135 | |||
| 136 | #ifdef CONFIG_PPC_BOOK3S_64 | ||
| 137 | /* | ||
| 138 | * On hash the vmalloc and other regions alias to the kernel region when passed | ||
| 139 | * through __pa(), which virt_to_pfn() uses. That means virt_addr_valid() can | ||
| 140 | * return true for some vmalloc addresses, which is incorrect. So explicitly | ||
| 141 | * check that the address is in the kernel region. | ||
| 142 | */ | ||
| 143 | #define virt_addr_valid(kaddr) (REGION_ID(kaddr) == KERNEL_REGION_ID && \ | ||
| 144 | pfn_valid(virt_to_pfn(kaddr))) | ||
| 145 | #else | ||
| 135 | #define virt_addr_valid(kaddr) pfn_valid(virt_to_pfn(kaddr)) | 146 | #define virt_addr_valid(kaddr) pfn_valid(virt_to_pfn(kaddr)) |
| 147 | #endif | ||
| 136 | 148 | ||
| 137 | /* | 149 | /* |
| 138 | * On Book-E parts we need __va to parse the device tree and we can't | 150 | * On Book-E parts we need __va to parse the device tree and we can't |
diff --git a/arch/powerpc/kernel/idle_book3s.S b/arch/powerpc/kernel/idle_book3s.S index 07d4e0ad60db..4898d676dcae 100644 --- a/arch/powerpc/kernel/idle_book3s.S +++ b/arch/powerpc/kernel/idle_book3s.S | |||
| @@ -416,7 +416,7 @@ power9_dd1_recover_paca: | |||
| 416 | * which needs to be restored from the stack. | 416 | * which needs to be restored from the stack. |
| 417 | */ | 417 | */ |
| 418 | li r3, 1 | 418 | li r3, 1 |
| 419 | stb r0,PACA_NAPSTATELOST(r13) | 419 | stb r3,PACA_NAPSTATELOST(r13) |
| 420 | blr | 420 | blr |
| 421 | 421 | ||
| 422 | /* | 422 | /* |
diff --git a/arch/powerpc/kernel/kprobes.c b/arch/powerpc/kernel/kprobes.c index 160ae0fa7d0d..fc4343514bed 100644 --- a/arch/powerpc/kernel/kprobes.c +++ b/arch/powerpc/kernel/kprobes.c | |||
| @@ -305,16 +305,17 @@ int kprobe_handler(struct pt_regs *regs) | |||
| 305 | save_previous_kprobe(kcb); | 305 | save_previous_kprobe(kcb); |
| 306 | set_current_kprobe(p, regs, kcb); | 306 | set_current_kprobe(p, regs, kcb); |
| 307 | kprobes_inc_nmissed_count(p); | 307 | kprobes_inc_nmissed_count(p); |
| 308 | prepare_singlestep(p, regs); | ||
| 309 | kcb->kprobe_status = KPROBE_REENTER; | 308 | kcb->kprobe_status = KPROBE_REENTER; |
| 310 | if (p->ainsn.boostable >= 0) { | 309 | if (p->ainsn.boostable >= 0) { |
| 311 | ret = try_to_emulate(p, regs); | 310 | ret = try_to_emulate(p, regs); |
| 312 | 311 | ||
| 313 | if (ret > 0) { | 312 | if (ret > 0) { |
| 314 | restore_previous_kprobe(kcb); | 313 | restore_previous_kprobe(kcb); |
| 314 | preempt_enable_no_resched(); | ||
| 315 | return 1; | 315 | return 1; |
| 316 | } | 316 | } |
| 317 | } | 317 | } |
| 318 | prepare_singlestep(p, regs); | ||
| 318 | return 1; | 319 | return 1; |
| 319 | } else { | 320 | } else { |
| 320 | if (*addr != BREAKPOINT_INSTRUCTION) { | 321 | if (*addr != BREAKPOINT_INSTRUCTION) { |
diff --git a/arch/powerpc/kernel/process.c b/arch/powerpc/kernel/process.c index d645da302bf2..baae104b16c7 100644 --- a/arch/powerpc/kernel/process.c +++ b/arch/powerpc/kernel/process.c | |||
| @@ -864,6 +864,25 @@ static void tm_reclaim_thread(struct thread_struct *thr, | |||
| 864 | if (!MSR_TM_SUSPENDED(mfmsr())) | 864 | if (!MSR_TM_SUSPENDED(mfmsr())) |
| 865 | return; | 865 | return; |
| 866 | 866 | ||
| 867 | /* | ||
| 868 | * If we are in a transaction and FP is off then we can't have | ||
| 869 | * used FP inside that transaction. Hence the checkpointed | ||
| 870 | * state is the same as the live state. We need to copy the | ||
| 871 | * live state to the checkpointed state so that when the | ||
| 872 | * transaction is restored, the checkpointed state is correct | ||
| 873 | * and the aborted transaction sees the correct state. We use | ||
| 874 | * ckpt_regs.msr here as that's what tm_reclaim will use to | ||
| 875 | * determine if it's going to write the checkpointed state or | ||
| 876 | * not. So either this will write the checkpointed registers, | ||
| 877 | * or reclaim will. Similarly for VMX. | ||
| 878 | */ | ||
| 879 | if ((thr->ckpt_regs.msr & MSR_FP) == 0) | ||
| 880 | memcpy(&thr->ckfp_state, &thr->fp_state, | ||
| 881 | sizeof(struct thread_fp_state)); | ||
| 882 | if ((thr->ckpt_regs.msr & MSR_VEC) == 0) | ||
| 883 | memcpy(&thr->ckvr_state, &thr->vr_state, | ||
| 884 | sizeof(struct thread_vr_state)); | ||
| 885 | |||
| 867 | giveup_all(container_of(thr, struct task_struct, thread)); | 886 | giveup_all(container_of(thr, struct task_struct, thread)); |
| 868 | 887 | ||
| 869 | tm_reclaim(thr, thr->ckpt_regs.msr, cause); | 888 | tm_reclaim(thr, thr->ckpt_regs.msr, cause); |
diff --git a/arch/powerpc/mm/dump_linuxpagetables.c b/arch/powerpc/mm/dump_linuxpagetables.c index d659345a98d6..44fe4833910f 100644 --- a/arch/powerpc/mm/dump_linuxpagetables.c +++ b/arch/powerpc/mm/dump_linuxpagetables.c | |||
| @@ -16,6 +16,7 @@ | |||
| 16 | */ | 16 | */ |
| 17 | #include <linux/debugfs.h> | 17 | #include <linux/debugfs.h> |
| 18 | #include <linux/fs.h> | 18 | #include <linux/fs.h> |
| 19 | #include <linux/hugetlb.h> | ||
| 19 | #include <linux/io.h> | 20 | #include <linux/io.h> |
| 20 | #include <linux/mm.h> | 21 | #include <linux/mm.h> |
| 21 | #include <linux/sched.h> | 22 | #include <linux/sched.h> |
| @@ -391,7 +392,7 @@ static void walk_pmd(struct pg_state *st, pud_t *pud, unsigned long start) | |||
| 391 | 392 | ||
| 392 | for (i = 0; i < PTRS_PER_PMD; i++, pmd++) { | 393 | for (i = 0; i < PTRS_PER_PMD; i++, pmd++) { |
| 393 | addr = start + i * PMD_SIZE; | 394 | addr = start + i * PMD_SIZE; |
| 394 | if (!pmd_none(*pmd)) | 395 | if (!pmd_none(*pmd) && !pmd_huge(*pmd)) |
| 395 | /* pmd exists */ | 396 | /* pmd exists */ |
| 396 | walk_pte(st, pmd, addr); | 397 | walk_pte(st, pmd, addr); |
| 397 | else | 398 | else |
| @@ -407,7 +408,7 @@ static void walk_pud(struct pg_state *st, pgd_t *pgd, unsigned long start) | |||
| 407 | 408 | ||
| 408 | for (i = 0; i < PTRS_PER_PUD; i++, pud++) { | 409 | for (i = 0; i < PTRS_PER_PUD; i++, pud++) { |
| 409 | addr = start + i * PUD_SIZE; | 410 | addr = start + i * PUD_SIZE; |
| 410 | if (!pud_none(*pud)) | 411 | if (!pud_none(*pud) && !pud_huge(*pud)) |
| 411 | /* pud exists */ | 412 | /* pud exists */ |
| 412 | walk_pmd(st, pud, addr); | 413 | walk_pmd(st, pud, addr); |
| 413 | else | 414 | else |
| @@ -427,7 +428,7 @@ static void walk_pagetables(struct pg_state *st) | |||
| 427 | */ | 428 | */ |
| 428 | for (i = 0; i < PTRS_PER_PGD; i++, pgd++) { | 429 | for (i = 0; i < PTRS_PER_PGD; i++, pgd++) { |
| 429 | addr = KERN_VIRT_START + i * PGDIR_SIZE; | 430 | addr = KERN_VIRT_START + i * PGDIR_SIZE; |
| 430 | if (!pgd_none(*pgd)) | 431 | if (!pgd_none(*pgd) && !pgd_huge(*pgd)) |
| 431 | /* pgd exists */ | 432 | /* pgd exists */ |
| 432 | walk_pud(st, pgd, addr); | 433 | walk_pud(st, pgd, addr); |
| 433 | else | 434 | else |
diff --git a/arch/s390/include/asm/debug.h b/arch/s390/include/asm/debug.h index 0206c8052328..df7b54ea956d 100644 --- a/arch/s390/include/asm/debug.h +++ b/arch/s390/include/asm/debug.h | |||
| @@ -10,6 +10,7 @@ | |||
| 10 | #include <linux/spinlock.h> | 10 | #include <linux/spinlock.h> |
| 11 | #include <linux/kernel.h> | 11 | #include <linux/kernel.h> |
| 12 | #include <linux/time.h> | 12 | #include <linux/time.h> |
| 13 | #include <linux/refcount.h> | ||
| 13 | #include <uapi/asm/debug.h> | 14 | #include <uapi/asm/debug.h> |
| 14 | 15 | ||
| 15 | #define DEBUG_MAX_LEVEL 6 /* debug levels range from 0 to 6 */ | 16 | #define DEBUG_MAX_LEVEL 6 /* debug levels range from 0 to 6 */ |
| @@ -31,7 +32,7 @@ struct debug_view; | |||
| 31 | typedef struct debug_info { | 32 | typedef struct debug_info { |
| 32 | struct debug_info* next; | 33 | struct debug_info* next; |
| 33 | struct debug_info* prev; | 34 | struct debug_info* prev; |
| 34 | atomic_t ref_count; | 35 | refcount_t ref_count; |
| 35 | spinlock_t lock; | 36 | spinlock_t lock; |
| 36 | int level; | 37 | int level; |
| 37 | int nr_areas; | 38 | int nr_areas; |
diff --git a/arch/s390/include/asm/dis.h b/arch/s390/include/asm/dis.h index 60323c21938b..37f617dfbede 100644 --- a/arch/s390/include/asm/dis.h +++ b/arch/s390/include/asm/dis.h | |||
| @@ -40,6 +40,8 @@ static inline int insn_length(unsigned char code) | |||
| 40 | return ((((int) code + 64) >> 7) + 1) << 1; | 40 | return ((((int) code + 64) >> 7) + 1) << 1; |
| 41 | } | 41 | } |
| 42 | 42 | ||
| 43 | struct pt_regs; | ||
| 44 | |||
| 43 | void show_code(struct pt_regs *regs); | 45 | void show_code(struct pt_regs *regs); |
| 44 | void print_fn_code(unsigned char *code, unsigned long len); | 46 | void print_fn_code(unsigned char *code, unsigned long len); |
| 45 | int insn_to_mnemonic(unsigned char *instruction, char *buf, unsigned int len); | 47 | int insn_to_mnemonic(unsigned char *instruction, char *buf, unsigned int len); |
diff --git a/arch/s390/include/asm/kprobes.h b/arch/s390/include/asm/kprobes.h index 1293c4066cfc..28792ef82c83 100644 --- a/arch/s390/include/asm/kprobes.h +++ b/arch/s390/include/asm/kprobes.h | |||
| @@ -27,12 +27,21 @@ | |||
| 27 | * 2005-Dec Used as a template for s390 by Mike Grundy | 27 | * 2005-Dec Used as a template for s390 by Mike Grundy |
| 28 | * <grundym@us.ibm.com> | 28 | * <grundym@us.ibm.com> |
| 29 | */ | 29 | */ |
| 30 | #include <linux/types.h> | ||
| 30 | #include <asm-generic/kprobes.h> | 31 | #include <asm-generic/kprobes.h> |
| 31 | 32 | ||
| 32 | #define BREAKPOINT_INSTRUCTION 0x0002 | 33 | #define BREAKPOINT_INSTRUCTION 0x0002 |
| 33 | 34 | ||
| 35 | #define FIXUP_PSW_NORMAL 0x08 | ||
| 36 | #define FIXUP_BRANCH_NOT_TAKEN 0x04 | ||
| 37 | #define FIXUP_RETURN_REGISTER 0x02 | ||
| 38 | #define FIXUP_NOT_REQUIRED 0x01 | ||
| 39 | |||
| 40 | int probe_is_prohibited_opcode(u16 *insn); | ||
| 41 | int probe_get_fixup_type(u16 *insn); | ||
| 42 | int probe_is_insn_relative_long(u16 *insn); | ||
| 43 | |||
| 34 | #ifdef CONFIG_KPROBES | 44 | #ifdef CONFIG_KPROBES |
| 35 | #include <linux/types.h> | ||
| 36 | #include <linux/ptrace.h> | 45 | #include <linux/ptrace.h> |
| 37 | #include <linux/percpu.h> | 46 | #include <linux/percpu.h> |
| 38 | #include <linux/sched/task_stack.h> | 47 | #include <linux/sched/task_stack.h> |
| @@ -56,11 +65,6 @@ typedef u16 kprobe_opcode_t; | |||
| 56 | 65 | ||
| 57 | #define KPROBE_SWAP_INST 0x10 | 66 | #define KPROBE_SWAP_INST 0x10 |
| 58 | 67 | ||
| 59 | #define FIXUP_PSW_NORMAL 0x08 | ||
| 60 | #define FIXUP_BRANCH_NOT_TAKEN 0x04 | ||
| 61 | #define FIXUP_RETURN_REGISTER 0x02 | ||
| 62 | #define FIXUP_NOT_REQUIRED 0x01 | ||
| 63 | |||
| 64 | /* Architecture specific copy of original instruction */ | 68 | /* Architecture specific copy of original instruction */ |
| 65 | struct arch_specific_insn { | 69 | struct arch_specific_insn { |
| 66 | /* copy of original instruction */ | 70 | /* copy of original instruction */ |
| @@ -90,10 +94,6 @@ int kprobe_fault_handler(struct pt_regs *regs, int trapnr); | |||
| 90 | int kprobe_exceptions_notify(struct notifier_block *self, | 94 | int kprobe_exceptions_notify(struct notifier_block *self, |
| 91 | unsigned long val, void *data); | 95 | unsigned long val, void *data); |
| 92 | 96 | ||
| 93 | int probe_is_prohibited_opcode(u16 *insn); | ||
| 94 | int probe_get_fixup_type(u16 *insn); | ||
| 95 | int probe_is_insn_relative_long(u16 *insn); | ||
| 96 | |||
| 97 | #define flush_insn_slot(p) do { } while (0) | 97 | #define flush_insn_slot(p) do { } while (0) |
| 98 | 98 | ||
| 99 | #endif /* CONFIG_KPROBES */ | 99 | #endif /* CONFIG_KPROBES */ |
diff --git a/arch/s390/include/asm/sysinfo.h b/arch/s390/include/asm/sysinfo.h index 73bff45ced55..e784bed6ed7f 100644 --- a/arch/s390/include/asm/sysinfo.h +++ b/arch/s390/include/asm/sysinfo.h | |||
| @@ -146,7 +146,7 @@ extern int topology_max_mnest; | |||
| 146 | * Returns the maximum nesting level supported by the cpu topology code. | 146 | * Returns the maximum nesting level supported by the cpu topology code. |
| 147 | * The current maximum level is 4 which is the drawer level. | 147 | * The current maximum level is 4 which is the drawer level. |
| 148 | */ | 148 | */ |
| 149 | static inline int topology_mnest_limit(void) | 149 | static inline unsigned char topology_mnest_limit(void) |
| 150 | { | 150 | { |
| 151 | return min(topology_max_mnest, 4); | 151 | return min(topology_max_mnest, 4); |
| 152 | } | 152 | } |
diff --git a/arch/s390/kernel/debug.c b/arch/s390/kernel/debug.c index 530226b6cb19..86b3e74f569e 100644 --- a/arch/s390/kernel/debug.c +++ b/arch/s390/kernel/debug.c | |||
| @@ -277,7 +277,7 @@ debug_info_alloc(const char *name, int pages_per_area, int nr_areas, | |||
| 277 | memset(rc->views, 0, DEBUG_MAX_VIEWS * sizeof(struct debug_view *)); | 277 | memset(rc->views, 0, DEBUG_MAX_VIEWS * sizeof(struct debug_view *)); |
| 278 | memset(rc->debugfs_entries, 0 ,DEBUG_MAX_VIEWS * | 278 | memset(rc->debugfs_entries, 0 ,DEBUG_MAX_VIEWS * |
| 279 | sizeof(struct dentry*)); | 279 | sizeof(struct dentry*)); |
| 280 | atomic_set(&(rc->ref_count), 0); | 280 | refcount_set(&(rc->ref_count), 0); |
| 281 | 281 | ||
| 282 | return rc; | 282 | return rc; |
| 283 | 283 | ||
| @@ -361,7 +361,7 @@ debug_info_create(const char *name, int pages_per_area, int nr_areas, | |||
| 361 | debug_area_last = rc; | 361 | debug_area_last = rc; |
| 362 | rc->next = NULL; | 362 | rc->next = NULL; |
| 363 | 363 | ||
| 364 | debug_info_get(rc); | 364 | refcount_set(&rc->ref_count, 1); |
| 365 | out: | 365 | out: |
| 366 | return rc; | 366 | return rc; |
| 367 | } | 367 | } |
| @@ -416,7 +416,7 @@ static void | |||
| 416 | debug_info_get(debug_info_t * db_info) | 416 | debug_info_get(debug_info_t * db_info) |
| 417 | { | 417 | { |
| 418 | if (db_info) | 418 | if (db_info) |
| 419 | atomic_inc(&db_info->ref_count); | 419 | refcount_inc(&db_info->ref_count); |
| 420 | } | 420 | } |
| 421 | 421 | ||
| 422 | /* | 422 | /* |
| @@ -431,7 +431,7 @@ debug_info_put(debug_info_t *db_info) | |||
| 431 | 431 | ||
| 432 | if (!db_info) | 432 | if (!db_info) |
| 433 | return; | 433 | return; |
| 434 | if (atomic_dec_and_test(&db_info->ref_count)) { | 434 | if (refcount_dec_and_test(&db_info->ref_count)) { |
| 435 | for (i = 0; i < DEBUG_MAX_VIEWS; i++) { | 435 | for (i = 0; i < DEBUG_MAX_VIEWS; i++) { |
| 436 | if (!db_info->views[i]) | 436 | if (!db_info->views[i]) |
| 437 | continue; | 437 | continue; |
diff --git a/arch/s390/kernel/entry.S b/arch/s390/kernel/entry.S index a5f5d3bb3dbc..e408d9cc5b96 100644 --- a/arch/s390/kernel/entry.S +++ b/arch/s390/kernel/entry.S | |||
| @@ -312,6 +312,7 @@ ENTRY(system_call) | |||
| 312 | lg %r14,__LC_VDSO_PER_CPU | 312 | lg %r14,__LC_VDSO_PER_CPU |
| 313 | lmg %r0,%r10,__PT_R0(%r11) | 313 | lmg %r0,%r10,__PT_R0(%r11) |
| 314 | mvc __LC_RETURN_PSW(16),__PT_PSW(%r11) | 314 | mvc __LC_RETURN_PSW(16),__PT_PSW(%r11) |
| 315 | .Lsysc_exit_timer: | ||
| 315 | stpt __LC_EXIT_TIMER | 316 | stpt __LC_EXIT_TIMER |
| 316 | mvc __VDSO_ECTG_BASE(16,%r14),__LC_EXIT_TIMER | 317 | mvc __VDSO_ECTG_BASE(16,%r14),__LC_EXIT_TIMER |
| 317 | lmg %r11,%r15,__PT_R11(%r11) | 318 | lmg %r11,%r15,__PT_R11(%r11) |
| @@ -623,6 +624,7 @@ ENTRY(io_int_handler) | |||
| 623 | lg %r14,__LC_VDSO_PER_CPU | 624 | lg %r14,__LC_VDSO_PER_CPU |
| 624 | lmg %r0,%r10,__PT_R0(%r11) | 625 | lmg %r0,%r10,__PT_R0(%r11) |
| 625 | mvc __LC_RETURN_PSW(16),__PT_PSW(%r11) | 626 | mvc __LC_RETURN_PSW(16),__PT_PSW(%r11) |
| 627 | .Lio_exit_timer: | ||
| 626 | stpt __LC_EXIT_TIMER | 628 | stpt __LC_EXIT_TIMER |
| 627 | mvc __VDSO_ECTG_BASE(16,%r14),__LC_EXIT_TIMER | 629 | mvc __VDSO_ECTG_BASE(16,%r14),__LC_EXIT_TIMER |
| 628 | lmg %r11,%r15,__PT_R11(%r11) | 630 | lmg %r11,%r15,__PT_R11(%r11) |
| @@ -1174,15 +1176,23 @@ cleanup_critical: | |||
| 1174 | br %r14 | 1176 | br %r14 |
| 1175 | 1177 | ||
| 1176 | .Lcleanup_sysc_restore: | 1178 | .Lcleanup_sysc_restore: |
| 1179 | # check if stpt has been executed | ||
| 1177 | clg %r9,BASED(.Lcleanup_sysc_restore_insn) | 1180 | clg %r9,BASED(.Lcleanup_sysc_restore_insn) |
| 1181 | jh 0f | ||
| 1182 | mvc __LC_EXIT_TIMER(8),__LC_ASYNC_ENTER_TIMER | ||
| 1183 | cghi %r11,__LC_SAVE_AREA_ASYNC | ||
| 1178 | je 0f | 1184 | je 0f |
| 1185 | mvc __LC_EXIT_TIMER(8),__LC_MCCK_ENTER_TIMER | ||
| 1186 | 0: clg %r9,BASED(.Lcleanup_sysc_restore_insn+8) | ||
| 1187 | je 1f | ||
| 1179 | lg %r9,24(%r11) # get saved pointer to pt_regs | 1188 | lg %r9,24(%r11) # get saved pointer to pt_regs |
| 1180 | mvc __LC_RETURN_PSW(16),__PT_PSW(%r9) | 1189 | mvc __LC_RETURN_PSW(16),__PT_PSW(%r9) |
| 1181 | mvc 0(64,%r11),__PT_R8(%r9) | 1190 | mvc 0(64,%r11),__PT_R8(%r9) |
| 1182 | lmg %r0,%r7,__PT_R0(%r9) | 1191 | lmg %r0,%r7,__PT_R0(%r9) |
| 1183 | 0: lmg %r8,%r9,__LC_RETURN_PSW | 1192 | 1: lmg %r8,%r9,__LC_RETURN_PSW |
| 1184 | br %r14 | 1193 | br %r14 |
| 1185 | .Lcleanup_sysc_restore_insn: | 1194 | .Lcleanup_sysc_restore_insn: |
| 1195 | .quad .Lsysc_exit_timer | ||
| 1186 | .quad .Lsysc_done - 4 | 1196 | .quad .Lsysc_done - 4 |
| 1187 | 1197 | ||
| 1188 | .Lcleanup_io_tif: | 1198 | .Lcleanup_io_tif: |
| @@ -1190,15 +1200,20 @@ cleanup_critical: | |||
| 1190 | br %r14 | 1200 | br %r14 |
| 1191 | 1201 | ||
| 1192 | .Lcleanup_io_restore: | 1202 | .Lcleanup_io_restore: |
| 1203 | # check if stpt has been executed | ||
| 1193 | clg %r9,BASED(.Lcleanup_io_restore_insn) | 1204 | clg %r9,BASED(.Lcleanup_io_restore_insn) |
| 1194 | je 0f | 1205 | jh 0f |
| 1206 | mvc __LC_EXIT_TIMER(8),__LC_MCCK_ENTER_TIMER | ||
| 1207 | 0: clg %r9,BASED(.Lcleanup_io_restore_insn+8) | ||
| 1208 | je 1f | ||
| 1195 | lg %r9,24(%r11) # get saved r11 pointer to pt_regs | 1209 | lg %r9,24(%r11) # get saved r11 pointer to pt_regs |
| 1196 | mvc __LC_RETURN_PSW(16),__PT_PSW(%r9) | 1210 | mvc __LC_RETURN_PSW(16),__PT_PSW(%r9) |
| 1197 | mvc 0(64,%r11),__PT_R8(%r9) | 1211 | mvc 0(64,%r11),__PT_R8(%r9) |
| 1198 | lmg %r0,%r7,__PT_R0(%r9) | 1212 | lmg %r0,%r7,__PT_R0(%r9) |
| 1199 | 0: lmg %r8,%r9,__LC_RETURN_PSW | 1213 | 1: lmg %r8,%r9,__LC_RETURN_PSW |
| 1200 | br %r14 | 1214 | br %r14 |
| 1201 | .Lcleanup_io_restore_insn: | 1215 | .Lcleanup_io_restore_insn: |
| 1216 | .quad .Lio_exit_timer | ||
| 1202 | .quad .Lio_done - 4 | 1217 | .quad .Lio_done - 4 |
| 1203 | 1218 | ||
| 1204 | .Lcleanup_idle: | 1219 | .Lcleanup_idle: |
diff --git a/arch/s390/kernel/ftrace.c b/arch/s390/kernel/ftrace.c index 27477f34cc0a..d03a6d12c4bd 100644 --- a/arch/s390/kernel/ftrace.c +++ b/arch/s390/kernel/ftrace.c | |||
| @@ -173,6 +173,8 @@ int __init ftrace_dyn_arch_init(void) | |||
| 173 | return 0; | 173 | return 0; |
| 174 | } | 174 | } |
| 175 | 175 | ||
| 176 | #ifdef CONFIG_MODULES | ||
| 177 | |||
| 176 | static int __init ftrace_plt_init(void) | 178 | static int __init ftrace_plt_init(void) |
| 177 | { | 179 | { |
| 178 | unsigned int *ip; | 180 | unsigned int *ip; |
| @@ -191,6 +193,8 @@ static int __init ftrace_plt_init(void) | |||
| 191 | } | 193 | } |
| 192 | device_initcall(ftrace_plt_init); | 194 | device_initcall(ftrace_plt_init); |
| 193 | 195 | ||
| 196 | #endif /* CONFIG_MODULES */ | ||
| 197 | |||
| 194 | #ifdef CONFIG_FUNCTION_GRAPH_TRACER | 198 | #ifdef CONFIG_FUNCTION_GRAPH_TRACER |
| 195 | /* | 199 | /* |
| 196 | * Hook the return address and push it in the stack of return addresses | 200 | * Hook the return address and push it in the stack of return addresses |
diff --git a/arch/s390/kernel/vmlinux.lds.S b/arch/s390/kernel/vmlinux.lds.S index 72307f108c40..6e2c42bd1c3b 100644 --- a/arch/s390/kernel/vmlinux.lds.S +++ b/arch/s390/kernel/vmlinux.lds.S | |||
| @@ -31,8 +31,14 @@ SECTIONS | |||
| 31 | { | 31 | { |
| 32 | . = 0x00000000; | 32 | . = 0x00000000; |
| 33 | .text : { | 33 | .text : { |
| 34 | _text = .; /* Text and read-only data */ | 34 | /* Text and read-only data */ |
| 35 | HEAD_TEXT | 35 | HEAD_TEXT |
| 36 | /* | ||
| 37 | * E.g. perf doesn't like symbols starting at address zero, | ||
| 38 | * therefore skip the initial PSW and channel program located | ||
| 39 | * at address zero and let _text start at 0x200. | ||
| 40 | */ | ||
| 41 | _text = 0x200; | ||
| 36 | TEXT_TEXT | 42 | TEXT_TEXT |
| 37 | SCHED_TEXT | 43 | SCHED_TEXT |
| 38 | CPUIDLE_TEXT | 44 | CPUIDLE_TEXT |
diff --git a/arch/s390/lib/probes.c b/arch/s390/lib/probes.c index ae90e1ae3607..1963ddbf4ab3 100644 --- a/arch/s390/lib/probes.c +++ b/arch/s390/lib/probes.c | |||
| @@ -4,6 +4,7 @@ | |||
| 4 | * Copyright IBM Corp. 2014 | 4 | * Copyright IBM Corp. 2014 |
| 5 | */ | 5 | */ |
| 6 | 6 | ||
| 7 | #include <linux/errno.h> | ||
| 7 | #include <asm/kprobes.h> | 8 | #include <asm/kprobes.h> |
| 8 | #include <asm/dis.h> | 9 | #include <asm/dis.h> |
| 9 | 10 | ||
diff --git a/arch/s390/lib/uaccess.c b/arch/s390/lib/uaccess.c index 1e5bb2b86c42..b3bd3f23b8e8 100644 --- a/arch/s390/lib/uaccess.c +++ b/arch/s390/lib/uaccess.c | |||
| @@ -337,8 +337,8 @@ long __strncpy_from_user(char *dst, const char __user *src, long size) | |||
| 337 | return 0; | 337 | return 0; |
| 338 | done = 0; | 338 | done = 0; |
| 339 | do { | 339 | do { |
| 340 | offset = (size_t)src & ~PAGE_MASK; | 340 | offset = (size_t)src & (L1_CACHE_BYTES - 1); |
| 341 | len = min(size - done, PAGE_SIZE - offset); | 341 | len = min(size - done, L1_CACHE_BYTES - offset); |
| 342 | if (copy_from_user(dst, src, len)) | 342 | if (copy_from_user(dst, src, len)) |
| 343 | return -EFAULT; | 343 | return -EFAULT; |
| 344 | len_str = strnlen(dst, len); | 344 | len_str = strnlen(dst, len); |
diff --git a/arch/sparc/include/asm/hugetlb.h b/arch/sparc/include/asm/hugetlb.h index dcbf985ab243..d1f837dc77a4 100644 --- a/arch/sparc/include/asm/hugetlb.h +++ b/arch/sparc/include/asm/hugetlb.h | |||
| @@ -24,9 +24,11 @@ static inline int is_hugepage_only_range(struct mm_struct *mm, | |||
| 24 | static inline int prepare_hugepage_range(struct file *file, | 24 | static inline int prepare_hugepage_range(struct file *file, |
| 25 | unsigned long addr, unsigned long len) | 25 | unsigned long addr, unsigned long len) |
| 26 | { | 26 | { |
| 27 | if (len & ~HPAGE_MASK) | 27 | struct hstate *h = hstate_file(file); |
| 28 | |||
| 29 | if (len & ~huge_page_mask(h)) | ||
| 28 | return -EINVAL; | 30 | return -EINVAL; |
| 29 | if (addr & ~HPAGE_MASK) | 31 | if (addr & ~huge_page_mask(h)) |
| 30 | return -EINVAL; | 32 | return -EINVAL; |
| 31 | return 0; | 33 | return 0; |
| 32 | } | 34 | } |
diff --git a/arch/sparc/include/asm/pgtable_32.h b/arch/sparc/include/asm/pgtable_32.h index ce6f56980aef..cf190728360b 100644 --- a/arch/sparc/include/asm/pgtable_32.h +++ b/arch/sparc/include/asm/pgtable_32.h | |||
| @@ -91,9 +91,9 @@ extern unsigned long pfn_base; | |||
| 91 | * ZERO_PAGE is a global shared page that is always zero: used | 91 | * ZERO_PAGE is a global shared page that is always zero: used |
| 92 | * for zero-mapped memory areas etc.. | 92 | * for zero-mapped memory areas etc.. |
| 93 | */ | 93 | */ |
| 94 | extern unsigned long empty_zero_page; | 94 | extern unsigned long empty_zero_page[PAGE_SIZE / sizeof(unsigned long)]; |
| 95 | 95 | ||
| 96 | #define ZERO_PAGE(vaddr) (virt_to_page(&empty_zero_page)) | 96 | #define ZERO_PAGE(vaddr) (virt_to_page(empty_zero_page)) |
| 97 | 97 | ||
| 98 | /* | 98 | /* |
| 99 | * In general all page table modifications should use the V8 atomic | 99 | * In general all page table modifications should use the V8 atomic |
diff --git a/arch/sparc/include/asm/setup.h b/arch/sparc/include/asm/setup.h index 478bf6bb4598..3fae200dd251 100644 --- a/arch/sparc/include/asm/setup.h +++ b/arch/sparc/include/asm/setup.h | |||
| @@ -16,7 +16,7 @@ extern char reboot_command[]; | |||
| 16 | */ | 16 | */ |
| 17 | extern unsigned char boot_cpu_id; | 17 | extern unsigned char boot_cpu_id; |
| 18 | 18 | ||
| 19 | extern unsigned long empty_zero_page; | 19 | extern unsigned long empty_zero_page[PAGE_SIZE / sizeof(unsigned long)]; |
| 20 | 20 | ||
| 21 | extern int serial_console; | 21 | extern int serial_console; |
| 22 | static inline int con_is_present(void) | 22 | static inline int con_is_present(void) |
diff --git a/arch/sparc/kernel/ftrace.c b/arch/sparc/kernel/ftrace.c index 6bcff698069b..cec54dc4ab81 100644 --- a/arch/sparc/kernel/ftrace.c +++ b/arch/sparc/kernel/ftrace.c | |||
| @@ -130,17 +130,16 @@ unsigned long prepare_ftrace_return(unsigned long parent, | |||
| 130 | if (unlikely(atomic_read(¤t->tracing_graph_pause))) | 130 | if (unlikely(atomic_read(¤t->tracing_graph_pause))) |
| 131 | return parent + 8UL; | 131 | return parent + 8UL; |
| 132 | 132 | ||
| 133 | if (ftrace_push_return_trace(parent, self_addr, &trace.depth, | ||
| 134 | frame_pointer, NULL) == -EBUSY) | ||
| 135 | return parent + 8UL; | ||
| 136 | |||
| 137 | trace.func = self_addr; | 133 | trace.func = self_addr; |
| 134 | trace.depth = current->curr_ret_stack + 1; | ||
| 138 | 135 | ||
| 139 | /* Only trace if the calling function expects to */ | 136 | /* Only trace if the calling function expects to */ |
| 140 | if (!ftrace_graph_entry(&trace)) { | 137 | if (!ftrace_graph_entry(&trace)) |
| 141 | current->curr_ret_stack--; | 138 | return parent + 8UL; |
| 139 | |||
| 140 | if (ftrace_push_return_trace(parent, self_addr, &trace.depth, | ||
| 141 | frame_pointer, NULL) == -EBUSY) | ||
| 142 | return parent + 8UL; | 142 | return parent + 8UL; |
| 143 | } | ||
| 144 | 143 | ||
| 145 | return return_hooker; | 144 | return return_hooker; |
| 146 | } | 145 | } |
diff --git a/arch/sparc/mm/init_32.c b/arch/sparc/mm/init_32.c index c6afe98de4d9..3bd0d513bddb 100644 --- a/arch/sparc/mm/init_32.c +++ b/arch/sparc/mm/init_32.c | |||
| @@ -290,7 +290,7 @@ void __init mem_init(void) | |||
| 290 | 290 | ||
| 291 | 291 | ||
| 292 | /* Saves us work later. */ | 292 | /* Saves us work later. */ |
| 293 | memset((void *)&empty_zero_page, 0, PAGE_SIZE); | 293 | memset((void *)empty_zero_page, 0, PAGE_SIZE); |
| 294 | 294 | ||
| 295 | i = last_valid_pfn >> ((20 - PAGE_SHIFT) + 5); | 295 | i = last_valid_pfn >> ((20 - PAGE_SHIFT) + 5); |
| 296 | i += 1; | 296 | i += 1; |
diff --git a/drivers/edac/amd64_edac.c b/drivers/edac/amd64_edac.c index 82dab1692264..3aea55698165 100644 --- a/drivers/edac/amd64_edac.c +++ b/drivers/edac/amd64_edac.c | |||
| @@ -782,24 +782,26 @@ static void debug_dump_dramcfg_low(struct amd64_pvt *pvt, u32 dclr, int chan) | |||
| 782 | 782 | ||
| 783 | static void debug_display_dimm_sizes_df(struct amd64_pvt *pvt, u8 ctrl) | 783 | static void debug_display_dimm_sizes_df(struct amd64_pvt *pvt, u8 ctrl) |
| 784 | { | 784 | { |
| 785 | u32 *dcsb = ctrl ? pvt->csels[1].csbases : pvt->csels[0].csbases; | 785 | int dimm, size0, size1, cs0, cs1; |
| 786 | int dimm, size0, size1; | ||
| 787 | 786 | ||
| 788 | edac_printk(KERN_DEBUG, EDAC_MC, "UMC%d chip selects:\n", ctrl); | 787 | edac_printk(KERN_DEBUG, EDAC_MC, "UMC%d chip selects:\n", ctrl); |
| 789 | 788 | ||
| 790 | for (dimm = 0; dimm < 4; dimm++) { | 789 | for (dimm = 0; dimm < 4; dimm++) { |
| 791 | size0 = 0; | 790 | size0 = 0; |
| 791 | cs0 = dimm * 2; | ||
| 792 | 792 | ||
| 793 | if (dcsb[dimm*2] & DCSB_CS_ENABLE) | 793 | if (csrow_enabled(cs0, ctrl, pvt)) |
| 794 | size0 = pvt->ops->dbam_to_cs(pvt, ctrl, 0, dimm); | 794 | size0 = pvt->ops->dbam_to_cs(pvt, ctrl, 0, cs0); |
| 795 | 795 | ||
| 796 | size1 = 0; | 796 | size1 = 0; |
| 797 | if (dcsb[dimm*2 + 1] & DCSB_CS_ENABLE) | 797 | cs1 = dimm * 2 + 1; |
| 798 | size1 = pvt->ops->dbam_to_cs(pvt, ctrl, 0, dimm); | 798 | |
| 799 | if (csrow_enabled(cs1, ctrl, pvt)) | ||
| 800 | size1 = pvt->ops->dbam_to_cs(pvt, ctrl, 0, cs1); | ||
| 799 | 801 | ||
| 800 | amd64_info(EDAC_MC ": %d: %5dMB %d: %5dMB\n", | 802 | amd64_info(EDAC_MC ": %d: %5dMB %d: %5dMB\n", |
| 801 | dimm * 2, size0, | 803 | cs0, size0, |
| 802 | dimm * 2 + 1, size1); | 804 | cs1, size1); |
| 803 | } | 805 | } |
| 804 | } | 806 | } |
| 805 | 807 | ||
| @@ -2756,26 +2758,22 @@ skip: | |||
| 2756 | * encompasses | 2758 | * encompasses |
| 2757 | * | 2759 | * |
| 2758 | */ | 2760 | */ |
| 2759 | static u32 get_csrow_nr_pages(struct amd64_pvt *pvt, u8 dct, int csrow_nr) | 2761 | static u32 get_csrow_nr_pages(struct amd64_pvt *pvt, u8 dct, int csrow_nr_orig) |
| 2760 | { | 2762 | { |
| 2761 | u32 cs_mode, nr_pages; | ||
| 2762 | u32 dbam = dct ? pvt->dbam1 : pvt->dbam0; | 2763 | u32 dbam = dct ? pvt->dbam1 : pvt->dbam0; |
| 2764 | int csrow_nr = csrow_nr_orig; | ||
| 2765 | u32 cs_mode, nr_pages; | ||
| 2763 | 2766 | ||
| 2767 | if (!pvt->umc) | ||
| 2768 | csrow_nr >>= 1; | ||
| 2764 | 2769 | ||
| 2765 | /* | 2770 | cs_mode = DBAM_DIMM(csrow_nr, dbam); |
| 2766 | * The math on this doesn't look right on the surface because x/2*4 can | ||
| 2767 | * be simplified to x*2 but this expression makes use of the fact that | ||
| 2768 | * it is integral math where 1/2=0. This intermediate value becomes the | ||
| 2769 | * number of bits to shift the DBAM register to extract the proper CSROW | ||
| 2770 | * field. | ||
| 2771 | */ | ||
| 2772 | cs_mode = DBAM_DIMM(csrow_nr / 2, dbam); | ||
| 2773 | 2771 | ||
| 2774 | nr_pages = pvt->ops->dbam_to_cs(pvt, dct, cs_mode, (csrow_nr / 2)) | 2772 | nr_pages = pvt->ops->dbam_to_cs(pvt, dct, cs_mode, csrow_nr); |
| 2775 | << (20 - PAGE_SHIFT); | 2773 | nr_pages <<= 20 - PAGE_SHIFT; |
| 2776 | 2774 | ||
| 2777 | edac_dbg(0, "csrow: %d, channel: %d, DBAM idx: %d\n", | 2775 | edac_dbg(0, "csrow: %d, channel: %d, DBAM idx: %d\n", |
| 2778 | csrow_nr, dct, cs_mode); | 2776 | csrow_nr_orig, dct, cs_mode); |
| 2779 | edac_dbg(0, "nr_pages/channel: %u\n", nr_pages); | 2777 | edac_dbg(0, "nr_pages/channel: %u\n", nr_pages); |
| 2780 | 2778 | ||
| 2781 | return nr_pages; | 2779 | return nr_pages; |
diff --git a/drivers/firmware/efi/efi-pstore.c b/drivers/firmware/efi/efi-pstore.c index ed3137c1ceb0..ab3a951a17e6 100644 --- a/drivers/firmware/efi/efi-pstore.c +++ b/drivers/firmware/efi/efi-pstore.c | |||
| @@ -155,19 +155,14 @@ static int efi_pstore_scan_sysfs_exit(struct efivar_entry *pos, | |||
| 155 | * efi_pstore_sysfs_entry_iter | 155 | * efi_pstore_sysfs_entry_iter |
| 156 | * | 156 | * |
| 157 | * @record: pstore record to pass to callback | 157 | * @record: pstore record to pass to callback |
| 158 | * @pos: entry to begin iterating from | ||
| 159 | * | 158 | * |
| 160 | * You MUST call efivar_enter_iter_begin() before this function, and | 159 | * You MUST call efivar_enter_iter_begin() before this function, and |
| 161 | * efivar_entry_iter_end() afterwards. | 160 | * efivar_entry_iter_end() afterwards. |
| 162 | * | 161 | * |
| 163 | * It is possible to begin iteration from an arbitrary entry within | ||
| 164 | * the list by passing @pos. @pos is updated on return to point to | ||
| 165 | * the next entry of the last one passed to efi_pstore_read_func(). | ||
| 166 | * To begin iterating from the beginning of the list @pos must be %NULL. | ||
| 167 | */ | 162 | */ |
| 168 | static int efi_pstore_sysfs_entry_iter(struct pstore_record *record, | 163 | static int efi_pstore_sysfs_entry_iter(struct pstore_record *record) |
| 169 | struct efivar_entry **pos) | ||
| 170 | { | 164 | { |
| 165 | struct efivar_entry **pos = (struct efivar_entry **)&record->psi->data; | ||
| 171 | struct efivar_entry *entry, *n; | 166 | struct efivar_entry *entry, *n; |
| 172 | struct list_head *head = &efivar_sysfs_list; | 167 | struct list_head *head = &efivar_sysfs_list; |
| 173 | int size = 0; | 168 | int size = 0; |
| @@ -218,7 +213,6 @@ static int efi_pstore_sysfs_entry_iter(struct pstore_record *record, | |||
| 218 | */ | 213 | */ |
| 219 | static ssize_t efi_pstore_read(struct pstore_record *record) | 214 | static ssize_t efi_pstore_read(struct pstore_record *record) |
| 220 | { | 215 | { |
| 221 | struct efivar_entry *entry = (struct efivar_entry *)record->psi->data; | ||
| 222 | ssize_t size; | 216 | ssize_t size; |
| 223 | 217 | ||
| 224 | record->buf = kzalloc(EFIVARS_DATA_SIZE_MAX, GFP_KERNEL); | 218 | record->buf = kzalloc(EFIVARS_DATA_SIZE_MAX, GFP_KERNEL); |
| @@ -229,7 +223,7 @@ static ssize_t efi_pstore_read(struct pstore_record *record) | |||
| 229 | size = -EINTR; | 223 | size = -EINTR; |
| 230 | goto out; | 224 | goto out; |
| 231 | } | 225 | } |
| 232 | size = efi_pstore_sysfs_entry_iter(record, &entry); | 226 | size = efi_pstore_sysfs_entry_iter(record); |
| 233 | efivar_entry_iter_end(); | 227 | efivar_entry_iter_end(); |
| 234 | 228 | ||
| 235 | out: | 229 | out: |
diff --git a/drivers/hwmon/coretemp.c b/drivers/hwmon/coretemp.c index 3ac4c03ba77b..c13a4fd86b3c 100644 --- a/drivers/hwmon/coretemp.c +++ b/drivers/hwmon/coretemp.c | |||
| @@ -605,6 +605,13 @@ static int coretemp_cpu_online(unsigned int cpu) | |||
| 605 | struct platform_data *pdata; | 605 | struct platform_data *pdata; |
| 606 | 606 | ||
| 607 | /* | 607 | /* |
| 608 | * Don't execute this on resume as the offline callback did | ||
| 609 | * not get executed on suspend. | ||
| 610 | */ | ||
| 611 | if (cpuhp_tasks_frozen) | ||
| 612 | return 0; | ||
| 613 | |||
| 614 | /* | ||
| 608 | * CPUID.06H.EAX[0] indicates whether the CPU has thermal | 615 | * CPUID.06H.EAX[0] indicates whether the CPU has thermal |
| 609 | * sensors. We check this bit only, all the early CPUs | 616 | * sensors. We check this bit only, all the early CPUs |
| 610 | * without thermal sensors will be filtered out. | 617 | * without thermal sensors will be filtered out. |
| @@ -654,6 +661,13 @@ static int coretemp_cpu_offline(unsigned int cpu) | |||
| 654 | struct temp_data *tdata; | 661 | struct temp_data *tdata; |
| 655 | int indx, target; | 662 | int indx, target; |
| 656 | 663 | ||
| 664 | /* | ||
| 665 | * Don't execute this on suspend as the device remove locks | ||
| 666 | * up the machine. | ||
| 667 | */ | ||
| 668 | if (cpuhp_tasks_frozen) | ||
| 669 | return 0; | ||
| 670 | |||
| 657 | /* If the physical CPU device does not exist, just return */ | 671 | /* If the physical CPU device does not exist, just return */ |
| 658 | if (!pdev) | 672 | if (!pdev) |
| 659 | return 0; | 673 | return 0; |
diff --git a/drivers/i2c/busses/i2c-mv64xxx.c b/drivers/i2c/busses/i2c-mv64xxx.c index cf737ec8563b..5c4db65c5019 100644 --- a/drivers/i2c/busses/i2c-mv64xxx.c +++ b/drivers/i2c/busses/i2c-mv64xxx.c | |||
| @@ -819,7 +819,6 @@ mv64xxx_of_config(struct mv64xxx_i2c_data *drv_data, | |||
| 819 | rc = -EINVAL; | 819 | rc = -EINVAL; |
| 820 | goto out; | 820 | goto out; |
| 821 | } | 821 | } |
| 822 | drv_data->irq = irq_of_parse_and_map(np, 0); | ||
| 823 | 822 | ||
| 824 | drv_data->rstc = devm_reset_control_get_optional(dev, NULL); | 823 | drv_data->rstc = devm_reset_control_get_optional(dev, NULL); |
| 825 | if (IS_ERR(drv_data->rstc)) { | 824 | if (IS_ERR(drv_data->rstc)) { |
| @@ -902,10 +901,11 @@ mv64xxx_i2c_probe(struct platform_device *pd) | |||
| 902 | if (!IS_ERR(drv_data->clk)) | 901 | if (!IS_ERR(drv_data->clk)) |
| 903 | clk_prepare_enable(drv_data->clk); | 902 | clk_prepare_enable(drv_data->clk); |
| 904 | 903 | ||
| 904 | drv_data->irq = platform_get_irq(pd, 0); | ||
| 905 | |||
| 905 | if (pdata) { | 906 | if (pdata) { |
| 906 | drv_data->freq_m = pdata->freq_m; | 907 | drv_data->freq_m = pdata->freq_m; |
| 907 | drv_data->freq_n = pdata->freq_n; | 908 | drv_data->freq_n = pdata->freq_n; |
| 908 | drv_data->irq = platform_get_irq(pd, 0); | ||
| 909 | drv_data->adapter.timeout = msecs_to_jiffies(pdata->timeout); | 909 | drv_data->adapter.timeout = msecs_to_jiffies(pdata->timeout); |
| 910 | drv_data->offload_enabled = false; | 910 | drv_data->offload_enabled = false; |
| 911 | memcpy(&drv_data->reg_offsets, &mv64xxx_i2c_regs_mv64xxx, sizeof(drv_data->reg_offsets)); | 911 | memcpy(&drv_data->reg_offsets, &mv64xxx_i2c_regs_mv64xxx, sizeof(drv_data->reg_offsets)); |
| @@ -915,7 +915,7 @@ mv64xxx_i2c_probe(struct platform_device *pd) | |||
| 915 | goto exit_clk; | 915 | goto exit_clk; |
| 916 | } | 916 | } |
| 917 | if (drv_data->irq < 0) { | 917 | if (drv_data->irq < 0) { |
| 918 | rc = -ENXIO; | 918 | rc = drv_data->irq; |
| 919 | goto exit_reset; | 919 | goto exit_reset; |
| 920 | } | 920 | } |
| 921 | 921 | ||
diff --git a/drivers/i2c/busses/i2c-xgene-slimpro.c b/drivers/i2c/busses/i2c-xgene-slimpro.c index dbe7e44c9321..6ba6c83ca8f1 100644 --- a/drivers/i2c/busses/i2c-xgene-slimpro.c +++ b/drivers/i2c/busses/i2c-xgene-slimpro.c | |||
| @@ -416,6 +416,7 @@ static int xgene_slimpro_i2c_probe(struct platform_device *pdev) | |||
| 416 | adapter->class = I2C_CLASS_HWMON; | 416 | adapter->class = I2C_CLASS_HWMON; |
| 417 | adapter->dev.parent = &pdev->dev; | 417 | adapter->dev.parent = &pdev->dev; |
| 418 | adapter->dev.of_node = pdev->dev.of_node; | 418 | adapter->dev.of_node = pdev->dev.of_node; |
| 419 | ACPI_COMPANION_SET(&adapter->dev, ACPI_COMPANION(&pdev->dev)); | ||
| 419 | i2c_set_adapdata(adapter, ctx); | 420 | i2c_set_adapdata(adapter, ctx); |
| 420 | rc = i2c_add_adapter(adapter); | 421 | rc = i2c_add_adapter(adapter); |
| 421 | if (rc) { | 422 | if (rc) { |
diff --git a/drivers/i2c/i2c-mux.c b/drivers/i2c/i2c-mux.c index 26f7237558ba..9669ca4937b8 100644 --- a/drivers/i2c/i2c-mux.c +++ b/drivers/i2c/i2c-mux.c | |||
| @@ -395,18 +395,20 @@ int i2c_mux_add_adapter(struct i2c_mux_core *muxc, | |||
| 395 | if (force_nr) { | 395 | if (force_nr) { |
| 396 | priv->adap.nr = force_nr; | 396 | priv->adap.nr = force_nr; |
| 397 | ret = i2c_add_numbered_adapter(&priv->adap); | 397 | ret = i2c_add_numbered_adapter(&priv->adap); |
| 398 | dev_err(&parent->dev, | 398 | if (ret < 0) { |
| 399 | "failed to add mux-adapter %u as bus %u (error=%d)\n", | 399 | dev_err(&parent->dev, |
| 400 | chan_id, force_nr, ret); | 400 | "failed to add mux-adapter %u as bus %u (error=%d)\n", |
| 401 | chan_id, force_nr, ret); | ||
| 402 | goto err_free_priv; | ||
| 403 | } | ||
| 401 | } else { | 404 | } else { |
| 402 | ret = i2c_add_adapter(&priv->adap); | 405 | ret = i2c_add_adapter(&priv->adap); |
| 403 | dev_err(&parent->dev, | 406 | if (ret < 0) { |
| 404 | "failed to add mux-adapter %u (error=%d)\n", | 407 | dev_err(&parent->dev, |
| 405 | chan_id, ret); | 408 | "failed to add mux-adapter %u (error=%d)\n", |
| 406 | } | 409 | chan_id, ret); |
| 407 | if (ret < 0) { | 410 | goto err_free_priv; |
| 408 | kfree(priv); | 411 | } |
| 409 | return ret; | ||
| 410 | } | 412 | } |
| 411 | 413 | ||
| 412 | WARN(sysfs_create_link(&priv->adap.dev.kobj, &muxc->dev->kobj, | 414 | WARN(sysfs_create_link(&priv->adap.dev.kobj, &muxc->dev->kobj, |
| @@ -422,6 +424,10 @@ int i2c_mux_add_adapter(struct i2c_mux_core *muxc, | |||
| 422 | 424 | ||
| 423 | muxc->adapter[muxc->num_adapters++] = &priv->adap; | 425 | muxc->adapter[muxc->num_adapters++] = &priv->adap; |
| 424 | return 0; | 426 | return 0; |
| 427 | |||
| 428 | err_free_priv: | ||
| 429 | kfree(priv); | ||
| 430 | return ret; | ||
| 425 | } | 431 | } |
| 426 | EXPORT_SYMBOL_GPL(i2c_mux_add_adapter); | 432 | EXPORT_SYMBOL_GPL(i2c_mux_add_adapter); |
| 427 | 433 | ||
diff --git a/drivers/i2c/muxes/i2c-mux-reg.c b/drivers/i2c/muxes/i2c-mux-reg.c index 406d5059072c..d97031804de8 100644 --- a/drivers/i2c/muxes/i2c-mux-reg.c +++ b/drivers/i2c/muxes/i2c-mux-reg.c | |||
| @@ -196,20 +196,25 @@ static int i2c_mux_reg_probe(struct platform_device *pdev) | |||
| 196 | res = platform_get_resource(pdev, IORESOURCE_MEM, 0); | 196 | res = platform_get_resource(pdev, IORESOURCE_MEM, 0); |
| 197 | mux->data.reg_size = resource_size(res); | 197 | mux->data.reg_size = resource_size(res); |
| 198 | mux->data.reg = devm_ioremap_resource(&pdev->dev, res); | 198 | mux->data.reg = devm_ioremap_resource(&pdev->dev, res); |
| 199 | if (IS_ERR(mux->data.reg)) | 199 | if (IS_ERR(mux->data.reg)) { |
| 200 | return PTR_ERR(mux->data.reg); | 200 | ret = PTR_ERR(mux->data.reg); |
| 201 | goto err_put_parent; | ||
| 202 | } | ||
| 201 | } | 203 | } |
| 202 | 204 | ||
| 203 | if (mux->data.reg_size != 4 && mux->data.reg_size != 2 && | 205 | if (mux->data.reg_size != 4 && mux->data.reg_size != 2 && |
| 204 | mux->data.reg_size != 1) { | 206 | mux->data.reg_size != 1) { |
| 205 | dev_err(&pdev->dev, "Invalid register size\n"); | 207 | dev_err(&pdev->dev, "Invalid register size\n"); |
| 206 | return -EINVAL; | 208 | ret = -EINVAL; |
| 209 | goto err_put_parent; | ||
| 207 | } | 210 | } |
| 208 | 211 | ||
| 209 | muxc = i2c_mux_alloc(parent, &pdev->dev, mux->data.n_values, 0, 0, | 212 | muxc = i2c_mux_alloc(parent, &pdev->dev, mux->data.n_values, 0, 0, |
| 210 | i2c_mux_reg_select, NULL); | 213 | i2c_mux_reg_select, NULL); |
| 211 | if (!muxc) | 214 | if (!muxc) { |
| 212 | return -ENOMEM; | 215 | ret = -ENOMEM; |
| 216 | goto err_put_parent; | ||
| 217 | } | ||
| 213 | muxc->priv = mux; | 218 | muxc->priv = mux; |
| 214 | 219 | ||
| 215 | platform_set_drvdata(pdev, muxc); | 220 | platform_set_drvdata(pdev, muxc); |
| @@ -223,7 +228,7 @@ static int i2c_mux_reg_probe(struct platform_device *pdev) | |||
| 223 | 228 | ||
| 224 | ret = i2c_mux_add_adapter(muxc, nr, mux->data.values[i], class); | 229 | ret = i2c_mux_add_adapter(muxc, nr, mux->data.values[i], class); |
| 225 | if (ret) | 230 | if (ret) |
| 226 | goto add_adapter_failed; | 231 | goto err_del_mux_adapters; |
| 227 | } | 232 | } |
| 228 | 233 | ||
| 229 | dev_dbg(&pdev->dev, "%d port mux on %s adapter\n", | 234 | dev_dbg(&pdev->dev, "%d port mux on %s adapter\n", |
| @@ -231,8 +236,10 @@ static int i2c_mux_reg_probe(struct platform_device *pdev) | |||
| 231 | 236 | ||
| 232 | return 0; | 237 | return 0; |
| 233 | 238 | ||
| 234 | add_adapter_failed: | 239 | err_del_mux_adapters: |
| 235 | i2c_mux_del_adapters(muxc); | 240 | i2c_mux_del_adapters(muxc); |
| 241 | err_put_parent: | ||
| 242 | i2c_put_adapter(parent); | ||
| 236 | 243 | ||
| 237 | return ret; | 244 | return ret; |
| 238 | } | 245 | } |
diff --git a/drivers/md/dm-bufio.c b/drivers/md/dm-bufio.c index 5db11a405129..cd8139593ccd 100644 --- a/drivers/md/dm-bufio.c +++ b/drivers/md/dm-bufio.c | |||
| @@ -218,7 +218,7 @@ static DEFINE_SPINLOCK(param_spinlock); | |||
| 218 | * Buffers are freed after this timeout | 218 | * Buffers are freed after this timeout |
| 219 | */ | 219 | */ |
| 220 | static unsigned dm_bufio_max_age = DM_BUFIO_DEFAULT_AGE_SECS; | 220 | static unsigned dm_bufio_max_age = DM_BUFIO_DEFAULT_AGE_SECS; |
| 221 | static unsigned dm_bufio_retain_bytes = DM_BUFIO_DEFAULT_RETAIN_BYTES; | 221 | static unsigned long dm_bufio_retain_bytes = DM_BUFIO_DEFAULT_RETAIN_BYTES; |
| 222 | 222 | ||
| 223 | static unsigned long dm_bufio_peak_allocated; | 223 | static unsigned long dm_bufio_peak_allocated; |
| 224 | static unsigned long dm_bufio_allocated_kmem_cache; | 224 | static unsigned long dm_bufio_allocated_kmem_cache; |
| @@ -1558,10 +1558,10 @@ static bool __try_evict_buffer(struct dm_buffer *b, gfp_t gfp) | |||
| 1558 | return true; | 1558 | return true; |
| 1559 | } | 1559 | } |
| 1560 | 1560 | ||
| 1561 | static unsigned get_retain_buffers(struct dm_bufio_client *c) | 1561 | static unsigned long get_retain_buffers(struct dm_bufio_client *c) |
| 1562 | { | 1562 | { |
| 1563 | unsigned retain_bytes = ACCESS_ONCE(dm_bufio_retain_bytes); | 1563 | unsigned long retain_bytes = ACCESS_ONCE(dm_bufio_retain_bytes); |
| 1564 | return retain_bytes / c->block_size; | 1564 | return retain_bytes >> (c->sectors_per_block_bits + SECTOR_SHIFT); |
| 1565 | } | 1565 | } |
| 1566 | 1566 | ||
| 1567 | static unsigned long __scan(struct dm_bufio_client *c, unsigned long nr_to_scan, | 1567 | static unsigned long __scan(struct dm_bufio_client *c, unsigned long nr_to_scan, |
| @@ -1571,7 +1571,7 @@ static unsigned long __scan(struct dm_bufio_client *c, unsigned long nr_to_scan, | |||
| 1571 | struct dm_buffer *b, *tmp; | 1571 | struct dm_buffer *b, *tmp; |
| 1572 | unsigned long freed = 0; | 1572 | unsigned long freed = 0; |
| 1573 | unsigned long count = nr_to_scan; | 1573 | unsigned long count = nr_to_scan; |
| 1574 | unsigned retain_target = get_retain_buffers(c); | 1574 | unsigned long retain_target = get_retain_buffers(c); |
| 1575 | 1575 | ||
| 1576 | for (l = 0; l < LIST_SIZE; l++) { | 1576 | for (l = 0; l < LIST_SIZE; l++) { |
| 1577 | list_for_each_entry_safe_reverse(b, tmp, &c->lru[l], lru_list) { | 1577 | list_for_each_entry_safe_reverse(b, tmp, &c->lru[l], lru_list) { |
| @@ -1794,8 +1794,8 @@ static bool older_than(struct dm_buffer *b, unsigned long age_hz) | |||
| 1794 | static void __evict_old_buffers(struct dm_bufio_client *c, unsigned long age_hz) | 1794 | static void __evict_old_buffers(struct dm_bufio_client *c, unsigned long age_hz) |
| 1795 | { | 1795 | { |
| 1796 | struct dm_buffer *b, *tmp; | 1796 | struct dm_buffer *b, *tmp; |
| 1797 | unsigned retain_target = get_retain_buffers(c); | 1797 | unsigned long retain_target = get_retain_buffers(c); |
| 1798 | unsigned count; | 1798 | unsigned long count; |
| 1799 | LIST_HEAD(write_list); | 1799 | LIST_HEAD(write_list); |
| 1800 | 1800 | ||
| 1801 | dm_bufio_lock(c); | 1801 | dm_bufio_lock(c); |
| @@ -1955,7 +1955,7 @@ MODULE_PARM_DESC(max_cache_size_bytes, "Size of metadata cache"); | |||
| 1955 | module_param_named(max_age_seconds, dm_bufio_max_age, uint, S_IRUGO | S_IWUSR); | 1955 | module_param_named(max_age_seconds, dm_bufio_max_age, uint, S_IRUGO | S_IWUSR); |
| 1956 | MODULE_PARM_DESC(max_age_seconds, "Max age of a buffer in seconds"); | 1956 | MODULE_PARM_DESC(max_age_seconds, "Max age of a buffer in seconds"); |
| 1957 | 1957 | ||
| 1958 | module_param_named(retain_bytes, dm_bufio_retain_bytes, uint, S_IRUGO | S_IWUSR); | 1958 | module_param_named(retain_bytes, dm_bufio_retain_bytes, ulong, S_IRUGO | S_IWUSR); |
| 1959 | MODULE_PARM_DESC(retain_bytes, "Try to keep at least this many bytes cached in memory"); | 1959 | MODULE_PARM_DESC(retain_bytes, "Try to keep at least this many bytes cached in memory"); |
| 1960 | 1960 | ||
| 1961 | module_param_named(peak_allocated_bytes, dm_bufio_peak_allocated, ulong, S_IRUGO | S_IWUSR); | 1961 | module_param_named(peak_allocated_bytes, dm_bufio_peak_allocated, ulong, S_IRUGO | S_IWUSR); |
diff --git a/drivers/md/dm-cache-background-tracker.c b/drivers/md/dm-cache-background-tracker.c index 9b1afdfb13f0..707233891291 100644 --- a/drivers/md/dm-cache-background-tracker.c +++ b/drivers/md/dm-cache-background-tracker.c | |||
| @@ -33,6 +33,11 @@ struct background_tracker *btracker_create(unsigned max_work) | |||
| 33 | { | 33 | { |
| 34 | struct background_tracker *b = kmalloc(sizeof(*b), GFP_KERNEL); | 34 | struct background_tracker *b = kmalloc(sizeof(*b), GFP_KERNEL); |
| 35 | 35 | ||
| 36 | if (!b) { | ||
| 37 | DMERR("couldn't create background_tracker"); | ||
| 38 | return NULL; | ||
| 39 | } | ||
| 40 | |||
| 36 | b->max_work = max_work; | 41 | b->max_work = max_work; |
| 37 | atomic_set(&b->pending_promotes, 0); | 42 | atomic_set(&b->pending_promotes, 0); |
| 38 | atomic_set(&b->pending_writebacks, 0); | 43 | atomic_set(&b->pending_writebacks, 0); |
diff --git a/drivers/md/dm-cache-policy-smq.c b/drivers/md/dm-cache-policy-smq.c index 72479bd61e11..e5eb9c9b4bc8 100644 --- a/drivers/md/dm-cache-policy-smq.c +++ b/drivers/md/dm-cache-policy-smq.c | |||
| @@ -1120,8 +1120,6 @@ static bool clean_target_met(struct smq_policy *mq, bool idle) | |||
| 1120 | * Cache entries may not be populated. So we cannot rely on the | 1120 | * Cache entries may not be populated. So we cannot rely on the |
| 1121 | * size of the clean queue. | 1121 | * size of the clean queue. |
| 1122 | */ | 1122 | */ |
| 1123 | unsigned nr_clean; | ||
| 1124 | |||
| 1125 | if (idle) { | 1123 | if (idle) { |
| 1126 | /* | 1124 | /* |
| 1127 | * We'd like to clean everything. | 1125 | * We'd like to clean everything. |
| @@ -1129,18 +1127,16 @@ static bool clean_target_met(struct smq_policy *mq, bool idle) | |||
| 1129 | return q_size(&mq->dirty) == 0u; | 1127 | return q_size(&mq->dirty) == 0u; |
| 1130 | } | 1128 | } |
| 1131 | 1129 | ||
| 1132 | nr_clean = from_cblock(mq->cache_size) - q_size(&mq->dirty); | 1130 | /* |
| 1133 | return (nr_clean + btracker_nr_writebacks_queued(mq->bg_work)) >= | 1131 | * If we're busy we don't worry about cleaning at all. |
| 1134 | percent_to_target(mq, CLEAN_TARGET); | 1132 | */ |
| 1133 | return true; | ||
| 1135 | } | 1134 | } |
| 1136 | 1135 | ||
| 1137 | static bool free_target_met(struct smq_policy *mq, bool idle) | 1136 | static bool free_target_met(struct smq_policy *mq) |
| 1138 | { | 1137 | { |
| 1139 | unsigned nr_free; | 1138 | unsigned nr_free; |
| 1140 | 1139 | ||
| 1141 | if (!idle) | ||
| 1142 | return true; | ||
| 1143 | |||
| 1144 | nr_free = from_cblock(mq->cache_size) - mq->cache_alloc.nr_allocated; | 1140 | nr_free = from_cblock(mq->cache_size) - mq->cache_alloc.nr_allocated; |
| 1145 | return (nr_free + btracker_nr_demotions_queued(mq->bg_work)) >= | 1141 | return (nr_free + btracker_nr_demotions_queued(mq->bg_work)) >= |
| 1146 | percent_to_target(mq, FREE_TARGET); | 1142 | percent_to_target(mq, FREE_TARGET); |
| @@ -1190,9 +1186,9 @@ static void queue_demotion(struct smq_policy *mq) | |||
| 1190 | if (unlikely(WARN_ON_ONCE(!mq->migrations_allowed))) | 1186 | if (unlikely(WARN_ON_ONCE(!mq->migrations_allowed))) |
| 1191 | return; | 1187 | return; |
| 1192 | 1188 | ||
| 1193 | e = q_peek(&mq->clean, mq->clean.nr_levels, true); | 1189 | e = q_peek(&mq->clean, mq->clean.nr_levels / 2, true); |
| 1194 | if (!e) { | 1190 | if (!e) { |
| 1195 | if (!clean_target_met(mq, false)) | 1191 | if (!clean_target_met(mq, true)) |
| 1196 | queue_writeback(mq); | 1192 | queue_writeback(mq); |
| 1197 | return; | 1193 | return; |
| 1198 | } | 1194 | } |
| @@ -1220,7 +1216,7 @@ static void queue_promotion(struct smq_policy *mq, dm_oblock_t oblock, | |||
| 1220 | * We always claim to be 'idle' to ensure some demotions happen | 1216 | * We always claim to be 'idle' to ensure some demotions happen |
| 1221 | * with continuous loads. | 1217 | * with continuous loads. |
| 1222 | */ | 1218 | */ |
| 1223 | if (!free_target_met(mq, true)) | 1219 | if (!free_target_met(mq)) |
| 1224 | queue_demotion(mq); | 1220 | queue_demotion(mq); |
| 1225 | return; | 1221 | return; |
| 1226 | } | 1222 | } |
| @@ -1421,14 +1417,10 @@ static int smq_get_background_work(struct dm_cache_policy *p, bool idle, | |||
| 1421 | spin_lock_irqsave(&mq->lock, flags); | 1417 | spin_lock_irqsave(&mq->lock, flags); |
| 1422 | r = btracker_issue(mq->bg_work, result); | 1418 | r = btracker_issue(mq->bg_work, result); |
| 1423 | if (r == -ENODATA) { | 1419 | if (r == -ENODATA) { |
| 1424 | /* find some writeback work to do */ | 1420 | if (!clean_target_met(mq, idle)) { |
| 1425 | if (mq->migrations_allowed && !free_target_met(mq, idle)) | ||
| 1426 | queue_demotion(mq); | ||
| 1427 | |||
| 1428 | else if (!clean_target_met(mq, idle)) | ||
| 1429 | queue_writeback(mq); | 1421 | queue_writeback(mq); |
| 1430 | 1422 | r = btracker_issue(mq->bg_work, result); | |
| 1431 | r = btracker_issue(mq->bg_work, result); | 1423 | } |
| 1432 | } | 1424 | } |
| 1433 | spin_unlock_irqrestore(&mq->lock, flags); | 1425 | spin_unlock_irqrestore(&mq->lock, flags); |
| 1434 | 1426 | ||
| @@ -1452,6 +1444,7 @@ static void __complete_background_work(struct smq_policy *mq, | |||
| 1452 | clear_pending(mq, e); | 1444 | clear_pending(mq, e); |
| 1453 | if (success) { | 1445 | if (success) { |
| 1454 | e->oblock = work->oblock; | 1446 | e->oblock = work->oblock; |
| 1447 | e->level = NR_CACHE_LEVELS - 1; | ||
| 1455 | push(mq, e); | 1448 | push(mq, e); |
| 1456 | // h, q, a | 1449 | // h, q, a |
| 1457 | } else { | 1450 | } else { |
diff --git a/drivers/md/dm-cache-target.c b/drivers/md/dm-cache-target.c index 1db375f50a13..d682a0511381 100644 --- a/drivers/md/dm-cache-target.c +++ b/drivers/md/dm-cache-target.c | |||
| @@ -94,6 +94,9 @@ static void iot_io_begin(struct io_tracker *iot, sector_t len) | |||
| 94 | 94 | ||
| 95 | static void __iot_io_end(struct io_tracker *iot, sector_t len) | 95 | static void __iot_io_end(struct io_tracker *iot, sector_t len) |
| 96 | { | 96 | { |
| 97 | if (!len) | ||
| 98 | return; | ||
| 99 | |||
| 97 | iot->in_flight -= len; | 100 | iot->in_flight -= len; |
| 98 | if (!iot->in_flight) | 101 | if (!iot->in_flight) |
| 99 | iot->idle_time = jiffies; | 102 | iot->idle_time = jiffies; |
| @@ -474,7 +477,7 @@ struct cache { | |||
| 474 | spinlock_t invalidation_lock; | 477 | spinlock_t invalidation_lock; |
| 475 | struct list_head invalidation_requests; | 478 | struct list_head invalidation_requests; |
| 476 | 479 | ||
| 477 | struct io_tracker origin_tracker; | 480 | struct io_tracker tracker; |
| 478 | 481 | ||
| 479 | struct work_struct commit_ws; | 482 | struct work_struct commit_ws; |
| 480 | struct batcher committer; | 483 | struct batcher committer; |
| @@ -901,8 +904,7 @@ static dm_oblock_t get_bio_block(struct cache *cache, struct bio *bio) | |||
| 901 | 904 | ||
| 902 | static bool accountable_bio(struct cache *cache, struct bio *bio) | 905 | static bool accountable_bio(struct cache *cache, struct bio *bio) |
| 903 | { | 906 | { |
| 904 | return ((bio->bi_bdev == cache->origin_dev->bdev) && | 907 | return bio_op(bio) != REQ_OP_DISCARD; |
| 905 | bio_op(bio) != REQ_OP_DISCARD); | ||
| 906 | } | 908 | } |
| 907 | 909 | ||
| 908 | static void accounted_begin(struct cache *cache, struct bio *bio) | 910 | static void accounted_begin(struct cache *cache, struct bio *bio) |
| @@ -912,7 +914,7 @@ static void accounted_begin(struct cache *cache, struct bio *bio) | |||
| 912 | 914 | ||
| 913 | if (accountable_bio(cache, bio)) { | 915 | if (accountable_bio(cache, bio)) { |
| 914 | pb->len = bio_sectors(bio); | 916 | pb->len = bio_sectors(bio); |
| 915 | iot_io_begin(&cache->origin_tracker, pb->len); | 917 | iot_io_begin(&cache->tracker, pb->len); |
| 916 | } | 918 | } |
| 917 | } | 919 | } |
| 918 | 920 | ||
| @@ -921,7 +923,7 @@ static void accounted_complete(struct cache *cache, struct bio *bio) | |||
| 921 | size_t pb_data_size = get_per_bio_data_size(cache); | 923 | size_t pb_data_size = get_per_bio_data_size(cache); |
| 922 | struct per_bio_data *pb = get_per_bio_data(bio, pb_data_size); | 924 | struct per_bio_data *pb = get_per_bio_data(bio, pb_data_size); |
| 923 | 925 | ||
| 924 | iot_io_end(&cache->origin_tracker, pb->len); | 926 | iot_io_end(&cache->tracker, pb->len); |
| 925 | } | 927 | } |
| 926 | 928 | ||
| 927 | static void accounted_request(struct cache *cache, struct bio *bio) | 929 | static void accounted_request(struct cache *cache, struct bio *bio) |
| @@ -1716,20 +1718,19 @@ static int invalidate_start(struct cache *cache, dm_cblock_t cblock, | |||
| 1716 | 1718 | ||
| 1717 | enum busy { | 1719 | enum busy { |
| 1718 | IDLE, | 1720 | IDLE, |
| 1719 | MODERATE, | ||
| 1720 | BUSY | 1721 | BUSY |
| 1721 | }; | 1722 | }; |
| 1722 | 1723 | ||
| 1723 | static enum busy spare_migration_bandwidth(struct cache *cache) | 1724 | static enum busy spare_migration_bandwidth(struct cache *cache) |
| 1724 | { | 1725 | { |
| 1725 | bool idle = iot_idle_for(&cache->origin_tracker, HZ); | 1726 | bool idle = iot_idle_for(&cache->tracker, HZ); |
| 1726 | sector_t current_volume = (atomic_read(&cache->nr_io_migrations) + 1) * | 1727 | sector_t current_volume = (atomic_read(&cache->nr_io_migrations) + 1) * |
| 1727 | cache->sectors_per_block; | 1728 | cache->sectors_per_block; |
| 1728 | 1729 | ||
| 1729 | if (current_volume <= cache->migration_threshold) | 1730 | if (idle && current_volume <= cache->migration_threshold) |
| 1730 | return idle ? IDLE : MODERATE; | 1731 | return IDLE; |
| 1731 | else | 1732 | else |
| 1732 | return idle ? MODERATE : BUSY; | 1733 | return BUSY; |
| 1733 | } | 1734 | } |
| 1734 | 1735 | ||
| 1735 | static void inc_hit_counter(struct cache *cache, struct bio *bio) | 1736 | static void inc_hit_counter(struct cache *cache, struct bio *bio) |
| @@ -2045,8 +2046,6 @@ static void check_migrations(struct work_struct *ws) | |||
| 2045 | 2046 | ||
| 2046 | for (;;) { | 2047 | for (;;) { |
| 2047 | b = spare_migration_bandwidth(cache); | 2048 | b = spare_migration_bandwidth(cache); |
| 2048 | if (b == BUSY) | ||
| 2049 | break; | ||
| 2050 | 2049 | ||
| 2051 | r = policy_get_background_work(cache->policy, b == IDLE, &op); | 2050 | r = policy_get_background_work(cache->policy, b == IDLE, &op); |
| 2052 | if (r == -ENODATA) | 2051 | if (r == -ENODATA) |
| @@ -2717,7 +2716,7 @@ static int cache_create(struct cache_args *ca, struct cache **result) | |||
| 2717 | 2716 | ||
| 2718 | batcher_init(&cache->committer, commit_op, cache, | 2717 | batcher_init(&cache->committer, commit_op, cache, |
| 2719 | issue_op, cache, cache->wq); | 2718 | issue_op, cache, cache->wq); |
| 2720 | iot_init(&cache->origin_tracker); | 2719 | iot_init(&cache->tracker); |
| 2721 | 2720 | ||
| 2722 | init_rwsem(&cache->background_work_lock); | 2721 | init_rwsem(&cache->background_work_lock); |
| 2723 | prevent_background_work(cache); | 2722 | prevent_background_work(cache); |
| @@ -2941,7 +2940,7 @@ static void cache_postsuspend(struct dm_target *ti) | |||
| 2941 | 2940 | ||
| 2942 | cancel_delayed_work(&cache->waker); | 2941 | cancel_delayed_work(&cache->waker); |
| 2943 | flush_workqueue(cache->wq); | 2942 | flush_workqueue(cache->wq); |
| 2944 | WARN_ON(cache->origin_tracker.in_flight); | 2943 | WARN_ON(cache->tracker.in_flight); |
| 2945 | 2944 | ||
| 2946 | /* | 2945 | /* |
| 2947 | * If it's a flush suspend there won't be any deferred bios, so this | 2946 | * If it's a flush suspend there won't be any deferred bios, so this |
diff --git a/drivers/md/dm-mpath.c b/drivers/md/dm-mpath.c index 926a6bcb32c8..3df056b73b66 100644 --- a/drivers/md/dm-mpath.c +++ b/drivers/md/dm-mpath.c | |||
| @@ -447,7 +447,7 @@ failed: | |||
| 447 | * it has been invoked. | 447 | * it has been invoked. |
| 448 | */ | 448 | */ |
| 449 | #define dm_report_EIO(m) \ | 449 | #define dm_report_EIO(m) \ |
| 450 | ({ \ | 450 | do { \ |
| 451 | struct mapped_device *md = dm_table_get_md((m)->ti->table); \ | 451 | struct mapped_device *md = dm_table_get_md((m)->ti->table); \ |
| 452 | \ | 452 | \ |
| 453 | pr_debug("%s: returning EIO; QIFNP = %d; SQIFNP = %d; DNFS = %d\n", \ | 453 | pr_debug("%s: returning EIO; QIFNP = %d; SQIFNP = %d; DNFS = %d\n", \ |
| @@ -455,8 +455,7 @@ failed: | |||
| 455 | test_bit(MPATHF_QUEUE_IF_NO_PATH, &(m)->flags), \ | 455 | test_bit(MPATHF_QUEUE_IF_NO_PATH, &(m)->flags), \ |
| 456 | test_bit(MPATHF_SAVED_QUEUE_IF_NO_PATH, &(m)->flags), \ | 456 | test_bit(MPATHF_SAVED_QUEUE_IF_NO_PATH, &(m)->flags), \ |
| 457 | dm_noflush_suspending((m)->ti)); \ | 457 | dm_noflush_suspending((m)->ti)); \ |
| 458 | -EIO; \ | 458 | } while (0) |
| 459 | }) | ||
| 460 | 459 | ||
| 461 | /* | 460 | /* |
| 462 | * Map cloned requests (request-based multipath) | 461 | * Map cloned requests (request-based multipath) |
| @@ -481,7 +480,8 @@ static int multipath_clone_and_map(struct dm_target *ti, struct request *rq, | |||
| 481 | if (!pgpath) { | 480 | if (!pgpath) { |
| 482 | if (test_bit(MPATHF_QUEUE_IF_NO_PATH, &m->flags)) | 481 | if (test_bit(MPATHF_QUEUE_IF_NO_PATH, &m->flags)) |
| 483 | return DM_MAPIO_DELAY_REQUEUE; | 482 | return DM_MAPIO_DELAY_REQUEUE; |
| 484 | return dm_report_EIO(m); /* Failed */ | 483 | dm_report_EIO(m); /* Failed */ |
| 484 | return DM_MAPIO_KILL; | ||
| 485 | } else if (test_bit(MPATHF_QUEUE_IO, &m->flags) || | 485 | } else if (test_bit(MPATHF_QUEUE_IO, &m->flags) || |
| 486 | test_bit(MPATHF_PG_INIT_REQUIRED, &m->flags)) { | 486 | test_bit(MPATHF_PG_INIT_REQUIRED, &m->flags)) { |
| 487 | if (pg_init_all_paths(m)) | 487 | if (pg_init_all_paths(m)) |
| @@ -558,7 +558,8 @@ static int __multipath_map_bio(struct multipath *m, struct bio *bio, struct dm_m | |||
| 558 | if (!pgpath) { | 558 | if (!pgpath) { |
| 559 | if (test_bit(MPATHF_QUEUE_IF_NO_PATH, &m->flags)) | 559 | if (test_bit(MPATHF_QUEUE_IF_NO_PATH, &m->flags)) |
| 560 | return DM_MAPIO_REQUEUE; | 560 | return DM_MAPIO_REQUEUE; |
| 561 | return dm_report_EIO(m); | 561 | dm_report_EIO(m); |
| 562 | return -EIO; | ||
| 562 | } | 563 | } |
| 563 | 564 | ||
| 564 | mpio->pgpath = pgpath; | 565 | mpio->pgpath = pgpath; |
| @@ -1493,7 +1494,7 @@ static int multipath_end_io(struct dm_target *ti, struct request *clone, | |||
| 1493 | if (atomic_read(&m->nr_valid_paths) == 0 && | 1494 | if (atomic_read(&m->nr_valid_paths) == 0 && |
| 1494 | !test_bit(MPATHF_QUEUE_IF_NO_PATH, &m->flags)) { | 1495 | !test_bit(MPATHF_QUEUE_IF_NO_PATH, &m->flags)) { |
| 1495 | if (error == -EIO) | 1496 | if (error == -EIO) |
| 1496 | error = dm_report_EIO(m); | 1497 | dm_report_EIO(m); |
| 1497 | /* complete with the original error */ | 1498 | /* complete with the original error */ |
| 1498 | r = DM_ENDIO_DONE; | 1499 | r = DM_ENDIO_DONE; |
| 1499 | } | 1500 | } |
| @@ -1524,8 +1525,10 @@ static int do_end_io_bio(struct multipath *m, struct bio *clone, | |||
| 1524 | fail_path(mpio->pgpath); | 1525 | fail_path(mpio->pgpath); |
| 1525 | 1526 | ||
| 1526 | if (atomic_read(&m->nr_valid_paths) == 0 && | 1527 | if (atomic_read(&m->nr_valid_paths) == 0 && |
| 1527 | !test_bit(MPATHF_QUEUE_IF_NO_PATH, &m->flags)) | 1528 | !test_bit(MPATHF_QUEUE_IF_NO_PATH, &m->flags)) { |
| 1528 | return dm_report_EIO(m); | 1529 | dm_report_EIO(m); |
| 1530 | return -EIO; | ||
| 1531 | } | ||
| 1529 | 1532 | ||
| 1530 | /* Queue for the daemon to resubmit */ | 1533 | /* Queue for the daemon to resubmit */ |
| 1531 | dm_bio_restore(get_bio_details_from_bio(clone), clone); | 1534 | dm_bio_restore(get_bio_details_from_bio(clone), clone); |
diff --git a/drivers/md/dm-rq.c b/drivers/md/dm-rq.c index 2af27026aa2e..b639fa7246ee 100644 --- a/drivers/md/dm-rq.c +++ b/drivers/md/dm-rq.c | |||
| @@ -507,6 +507,7 @@ static int map_request(struct dm_rq_target_io *tio) | |||
| 507 | case DM_MAPIO_KILL: | 507 | case DM_MAPIO_KILL: |
| 508 | /* The target wants to complete the I/O */ | 508 | /* The target wants to complete the I/O */ |
| 509 | dm_kill_unmapped_request(rq, -EIO); | 509 | dm_kill_unmapped_request(rq, -EIO); |
| 510 | break; | ||
| 510 | default: | 511 | default: |
| 511 | DMWARN("unimplemented target map return value: %d", r); | 512 | DMWARN("unimplemented target map return value: %d", r); |
| 512 | BUG(); | 513 | BUG(); |
diff --git a/drivers/md/dm-thin-metadata.c b/drivers/md/dm-thin-metadata.c index 0f0251d0d337..d31d18d9727c 100644 --- a/drivers/md/dm-thin-metadata.c +++ b/drivers/md/dm-thin-metadata.c | |||
| @@ -484,11 +484,11 @@ static int __write_initial_superblock(struct dm_pool_metadata *pmd) | |||
| 484 | if (r < 0) | 484 | if (r < 0) |
| 485 | return r; | 485 | return r; |
| 486 | 486 | ||
| 487 | r = save_sm_roots(pmd); | 487 | r = dm_tm_pre_commit(pmd->tm); |
| 488 | if (r < 0) | 488 | if (r < 0) |
| 489 | return r; | 489 | return r; |
| 490 | 490 | ||
| 491 | r = dm_tm_pre_commit(pmd->tm); | 491 | r = save_sm_roots(pmd); |
| 492 | if (r < 0) | 492 | if (r < 0) |
| 493 | return r; | 493 | return r; |
| 494 | 494 | ||
diff --git a/drivers/md/md.c b/drivers/md/md.c index 82f798be964f..10367ffe92e3 100644 --- a/drivers/md/md.c +++ b/drivers/md/md.c | |||
| @@ -8022,18 +8022,15 @@ EXPORT_SYMBOL(md_write_end); | |||
| 8022 | * may proceed without blocking. It is important to call this before | 8022 | * may proceed without blocking. It is important to call this before |
| 8023 | * attempting a GFP_KERNEL allocation while holding the mddev lock. | 8023 | * attempting a GFP_KERNEL allocation while holding the mddev lock. |
| 8024 | * Must be called with mddev_lock held. | 8024 | * Must be called with mddev_lock held. |
| 8025 | * | ||
| 8026 | * In the ->external case MD_SB_CHANGE_PENDING can not be cleared until mddev->lock | ||
| 8027 | * is dropped, so return -EAGAIN after notifying userspace. | ||
| 8028 | */ | 8025 | */ |
| 8029 | int md_allow_write(struct mddev *mddev) | 8026 | void md_allow_write(struct mddev *mddev) |
| 8030 | { | 8027 | { |
| 8031 | if (!mddev->pers) | 8028 | if (!mddev->pers) |
| 8032 | return 0; | 8029 | return; |
| 8033 | if (mddev->ro) | 8030 | if (mddev->ro) |
| 8034 | return 0; | 8031 | return; |
| 8035 | if (!mddev->pers->sync_request) | 8032 | if (!mddev->pers->sync_request) |
| 8036 | return 0; | 8033 | return; |
| 8037 | 8034 | ||
| 8038 | spin_lock(&mddev->lock); | 8035 | spin_lock(&mddev->lock); |
| 8039 | if (mddev->in_sync) { | 8036 | if (mddev->in_sync) { |
| @@ -8046,13 +8043,12 @@ int md_allow_write(struct mddev *mddev) | |||
| 8046 | spin_unlock(&mddev->lock); | 8043 | spin_unlock(&mddev->lock); |
| 8047 | md_update_sb(mddev, 0); | 8044 | md_update_sb(mddev, 0); |
| 8048 | sysfs_notify_dirent_safe(mddev->sysfs_state); | 8045 | sysfs_notify_dirent_safe(mddev->sysfs_state); |
| 8046 | /* wait for the dirty state to be recorded in the metadata */ | ||
| 8047 | wait_event(mddev->sb_wait, | ||
| 8048 | !test_bit(MD_SB_CHANGE_CLEAN, &mddev->sb_flags) && | ||
| 8049 | !test_bit(MD_SB_CHANGE_PENDING, &mddev->sb_flags)); | ||
| 8049 | } else | 8050 | } else |
| 8050 | spin_unlock(&mddev->lock); | 8051 | spin_unlock(&mddev->lock); |
| 8051 | |||
| 8052 | if (test_bit(MD_SB_CHANGE_PENDING, &mddev->sb_flags)) | ||
| 8053 | return -EAGAIN; | ||
| 8054 | else | ||
| 8055 | return 0; | ||
| 8056 | } | 8052 | } |
| 8057 | EXPORT_SYMBOL_GPL(md_allow_write); | 8053 | EXPORT_SYMBOL_GPL(md_allow_write); |
| 8058 | 8054 | ||
diff --git a/drivers/md/md.h b/drivers/md/md.h index 4e75d121bfcc..11f15146ce51 100644 --- a/drivers/md/md.h +++ b/drivers/md/md.h | |||
| @@ -665,7 +665,7 @@ extern int sync_page_io(struct md_rdev *rdev, sector_t sector, int size, | |||
| 665 | bool metadata_op); | 665 | bool metadata_op); |
| 666 | extern void md_do_sync(struct md_thread *thread); | 666 | extern void md_do_sync(struct md_thread *thread); |
| 667 | extern void md_new_event(struct mddev *mddev); | 667 | extern void md_new_event(struct mddev *mddev); |
| 668 | extern int md_allow_write(struct mddev *mddev); | 668 | extern void md_allow_write(struct mddev *mddev); |
| 669 | extern void md_wait_for_blocked_rdev(struct md_rdev *rdev, struct mddev *mddev); | 669 | extern void md_wait_for_blocked_rdev(struct md_rdev *rdev, struct mddev *mddev); |
| 670 | extern void md_set_array_sectors(struct mddev *mddev, sector_t array_sectors); | 670 | extern void md_set_array_sectors(struct mddev *mddev, sector_t array_sectors); |
| 671 | extern int md_check_no_bitmap(struct mddev *mddev); | 671 | extern int md_check_no_bitmap(struct mddev *mddev); |
diff --git a/drivers/md/persistent-data/dm-space-map-disk.c b/drivers/md/persistent-data/dm-space-map-disk.c index ebb280a14325..32adf6b4a9c7 100644 --- a/drivers/md/persistent-data/dm-space-map-disk.c +++ b/drivers/md/persistent-data/dm-space-map-disk.c | |||
| @@ -142,10 +142,23 @@ static int sm_disk_inc_block(struct dm_space_map *sm, dm_block_t b) | |||
| 142 | 142 | ||
| 143 | static int sm_disk_dec_block(struct dm_space_map *sm, dm_block_t b) | 143 | static int sm_disk_dec_block(struct dm_space_map *sm, dm_block_t b) |
| 144 | { | 144 | { |
| 145 | int r; | ||
| 146 | uint32_t old_count; | ||
| 145 | enum allocation_event ev; | 147 | enum allocation_event ev; |
| 146 | struct sm_disk *smd = container_of(sm, struct sm_disk, sm); | 148 | struct sm_disk *smd = container_of(sm, struct sm_disk, sm); |
| 147 | 149 | ||
| 148 | return sm_ll_dec(&smd->ll, b, &ev); | 150 | r = sm_ll_dec(&smd->ll, b, &ev); |
| 151 | if (!r && (ev == SM_FREE)) { | ||
| 152 | /* | ||
| 153 | * It's only free if it's also free in the last | ||
| 154 | * transaction. | ||
| 155 | */ | ||
| 156 | r = sm_ll_lookup(&smd->old_ll, b, &old_count); | ||
| 157 | if (!r && !old_count) | ||
| 158 | smd->nr_allocated_this_transaction--; | ||
| 159 | } | ||
| 160 | |||
| 161 | return r; | ||
| 149 | } | 162 | } |
| 150 | 163 | ||
| 151 | static int sm_disk_new_block(struct dm_space_map *sm, dm_block_t *b) | 164 | static int sm_disk_new_block(struct dm_space_map *sm, dm_block_t *b) |
diff --git a/drivers/md/raid0.c b/drivers/md/raid0.c index 84e58596594d..d6c0bc76e837 100644 --- a/drivers/md/raid0.c +++ b/drivers/md/raid0.c | |||
| @@ -385,7 +385,7 @@ static int raid0_run(struct mddev *mddev) | |||
| 385 | blk_queue_max_hw_sectors(mddev->queue, mddev->chunk_sectors); | 385 | blk_queue_max_hw_sectors(mddev->queue, mddev->chunk_sectors); |
| 386 | blk_queue_max_write_same_sectors(mddev->queue, mddev->chunk_sectors); | 386 | blk_queue_max_write_same_sectors(mddev->queue, mddev->chunk_sectors); |
| 387 | blk_queue_max_write_zeroes_sectors(mddev->queue, mddev->chunk_sectors); | 387 | blk_queue_max_write_zeroes_sectors(mddev->queue, mddev->chunk_sectors); |
| 388 | blk_queue_max_discard_sectors(mddev->queue, mddev->chunk_sectors); | 388 | blk_queue_max_discard_sectors(mddev->queue, UINT_MAX); |
| 389 | 389 | ||
| 390 | blk_queue_io_min(mddev->queue, mddev->chunk_sectors << 9); | 390 | blk_queue_io_min(mddev->queue, mddev->chunk_sectors << 9); |
| 391 | blk_queue_io_opt(mddev->queue, | 391 | blk_queue_io_opt(mddev->queue, |
| @@ -459,6 +459,95 @@ static inline int is_io_in_chunk_boundary(struct mddev *mddev, | |||
| 459 | } | 459 | } |
| 460 | } | 460 | } |
| 461 | 461 | ||
| 462 | static void raid0_handle_discard(struct mddev *mddev, struct bio *bio) | ||
| 463 | { | ||
| 464 | struct r0conf *conf = mddev->private; | ||
| 465 | struct strip_zone *zone; | ||
| 466 | sector_t start = bio->bi_iter.bi_sector; | ||
| 467 | sector_t end; | ||
| 468 | unsigned int stripe_size; | ||
| 469 | sector_t first_stripe_index, last_stripe_index; | ||
| 470 | sector_t start_disk_offset; | ||
| 471 | unsigned int start_disk_index; | ||
| 472 | sector_t end_disk_offset; | ||
| 473 | unsigned int end_disk_index; | ||
| 474 | unsigned int disk; | ||
| 475 | |||
| 476 | zone = find_zone(conf, &start); | ||
| 477 | |||
| 478 | if (bio_end_sector(bio) > zone->zone_end) { | ||
| 479 | struct bio *split = bio_split(bio, | ||
| 480 | zone->zone_end - bio->bi_iter.bi_sector, GFP_NOIO, | ||
| 481 | mddev->bio_set); | ||
| 482 | bio_chain(split, bio); | ||
| 483 | generic_make_request(bio); | ||
| 484 | bio = split; | ||
| 485 | end = zone->zone_end; | ||
| 486 | } else | ||
| 487 | end = bio_end_sector(bio); | ||
| 488 | |||
| 489 | if (zone != conf->strip_zone) | ||
| 490 | end = end - zone[-1].zone_end; | ||
| 491 | |||
| 492 | /* Now start and end is the offset in zone */ | ||
| 493 | stripe_size = zone->nb_dev * mddev->chunk_sectors; | ||
| 494 | |||
| 495 | first_stripe_index = start; | ||
| 496 | sector_div(first_stripe_index, stripe_size); | ||
| 497 | last_stripe_index = end; | ||
| 498 | sector_div(last_stripe_index, stripe_size); | ||
| 499 | |||
| 500 | start_disk_index = (int)(start - first_stripe_index * stripe_size) / | ||
| 501 | mddev->chunk_sectors; | ||
| 502 | start_disk_offset = ((int)(start - first_stripe_index * stripe_size) % | ||
| 503 | mddev->chunk_sectors) + | ||
| 504 | first_stripe_index * mddev->chunk_sectors; | ||
| 505 | end_disk_index = (int)(end - last_stripe_index * stripe_size) / | ||
| 506 | mddev->chunk_sectors; | ||
| 507 | end_disk_offset = ((int)(end - last_stripe_index * stripe_size) % | ||
| 508 | mddev->chunk_sectors) + | ||
| 509 | last_stripe_index * mddev->chunk_sectors; | ||
| 510 | |||
| 511 | for (disk = 0; disk < zone->nb_dev; disk++) { | ||
| 512 | sector_t dev_start, dev_end; | ||
| 513 | struct bio *discard_bio = NULL; | ||
| 514 | struct md_rdev *rdev; | ||
| 515 | |||
| 516 | if (disk < start_disk_index) | ||
| 517 | dev_start = (first_stripe_index + 1) * | ||
| 518 | mddev->chunk_sectors; | ||
| 519 | else if (disk > start_disk_index) | ||
| 520 | dev_start = first_stripe_index * mddev->chunk_sectors; | ||
| 521 | else | ||
| 522 | dev_start = start_disk_offset; | ||
| 523 | |||
| 524 | if (disk < end_disk_index) | ||
| 525 | dev_end = (last_stripe_index + 1) * mddev->chunk_sectors; | ||
| 526 | else if (disk > end_disk_index) | ||
| 527 | dev_end = last_stripe_index * mddev->chunk_sectors; | ||
| 528 | else | ||
| 529 | dev_end = end_disk_offset; | ||
| 530 | |||
| 531 | if (dev_end <= dev_start) | ||
| 532 | continue; | ||
| 533 | |||
| 534 | rdev = conf->devlist[(zone - conf->strip_zone) * | ||
| 535 | conf->strip_zone[0].nb_dev + disk]; | ||
| 536 | if (__blkdev_issue_discard(rdev->bdev, | ||
| 537 | dev_start + zone->dev_start + rdev->data_offset, | ||
| 538 | dev_end - dev_start, GFP_NOIO, 0, &discard_bio) || | ||
| 539 | !discard_bio) | ||
| 540 | continue; | ||
| 541 | bio_chain(discard_bio, bio); | ||
| 542 | if (mddev->gendisk) | ||
| 543 | trace_block_bio_remap(bdev_get_queue(rdev->bdev), | ||
| 544 | discard_bio, disk_devt(mddev->gendisk), | ||
| 545 | bio->bi_iter.bi_sector); | ||
| 546 | generic_make_request(discard_bio); | ||
| 547 | } | ||
| 548 | bio_endio(bio); | ||
| 549 | } | ||
| 550 | |||
| 462 | static void raid0_make_request(struct mddev *mddev, struct bio *bio) | 551 | static void raid0_make_request(struct mddev *mddev, struct bio *bio) |
| 463 | { | 552 | { |
| 464 | struct strip_zone *zone; | 553 | struct strip_zone *zone; |
| @@ -473,6 +562,11 @@ static void raid0_make_request(struct mddev *mddev, struct bio *bio) | |||
| 473 | return; | 562 | return; |
| 474 | } | 563 | } |
| 475 | 564 | ||
| 565 | if (unlikely((bio_op(bio) == REQ_OP_DISCARD))) { | ||
| 566 | raid0_handle_discard(mddev, bio); | ||
| 567 | return; | ||
| 568 | } | ||
| 569 | |||
| 476 | bio_sector = bio->bi_iter.bi_sector; | 570 | bio_sector = bio->bi_iter.bi_sector; |
| 477 | sector = bio_sector; | 571 | sector = bio_sector; |
| 478 | chunk_sects = mddev->chunk_sectors; | 572 | chunk_sects = mddev->chunk_sectors; |
| @@ -498,19 +592,13 @@ static void raid0_make_request(struct mddev *mddev, struct bio *bio) | |||
| 498 | bio->bi_iter.bi_sector = sector + zone->dev_start + | 592 | bio->bi_iter.bi_sector = sector + zone->dev_start + |
| 499 | tmp_dev->data_offset; | 593 | tmp_dev->data_offset; |
| 500 | 594 | ||
| 501 | if (unlikely((bio_op(bio) == REQ_OP_DISCARD) && | 595 | if (mddev->gendisk) |
| 502 | !blk_queue_discard(bdev_get_queue(bio->bi_bdev)))) { | 596 | trace_block_bio_remap(bdev_get_queue(bio->bi_bdev), |
| 503 | /* Just ignore it */ | 597 | bio, disk_devt(mddev->gendisk), |
| 504 | bio_endio(bio); | 598 | bio_sector); |
| 505 | } else { | 599 | mddev_check_writesame(mddev, bio); |
| 506 | if (mddev->gendisk) | 600 | mddev_check_write_zeroes(mddev, bio); |
| 507 | trace_block_bio_remap(bdev_get_queue(bio->bi_bdev), | 601 | generic_make_request(bio); |
| 508 | bio, disk_devt(mddev->gendisk), | ||
| 509 | bio_sector); | ||
| 510 | mddev_check_writesame(mddev, bio); | ||
| 511 | mddev_check_write_zeroes(mddev, bio); | ||
| 512 | generic_make_request(bio); | ||
| 513 | } | ||
| 514 | } | 602 | } |
| 515 | 603 | ||
| 516 | static void raid0_status(struct seq_file *seq, struct mddev *mddev) | 604 | static void raid0_status(struct seq_file *seq, struct mddev *mddev) |
diff --git a/drivers/md/raid1.c b/drivers/md/raid1.c index 7ed59351fe97..af5056d56878 100644 --- a/drivers/md/raid1.c +++ b/drivers/md/raid1.c | |||
| @@ -666,8 +666,11 @@ static int read_balance(struct r1conf *conf, struct r1bio *r1_bio, int *max_sect | |||
| 666 | break; | 666 | break; |
| 667 | } | 667 | } |
| 668 | continue; | 668 | continue; |
| 669 | } else | 669 | } else { |
| 670 | if ((sectors > best_good_sectors) && (best_disk >= 0)) | ||
| 671 | best_disk = -1; | ||
| 670 | best_good_sectors = sectors; | 672 | best_good_sectors = sectors; |
| 673 | } | ||
| 671 | 674 | ||
| 672 | if (best_disk >= 0) | 675 | if (best_disk >= 0) |
| 673 | /* At least two disks to choose from so failfast is OK */ | 676 | /* At least two disks to choose from so failfast is OK */ |
| @@ -1529,17 +1532,16 @@ static void raid1_write_request(struct mddev *mddev, struct bio *bio, | |||
| 1529 | plug = container_of(cb, struct raid1_plug_cb, cb); | 1532 | plug = container_of(cb, struct raid1_plug_cb, cb); |
| 1530 | else | 1533 | else |
| 1531 | plug = NULL; | 1534 | plug = NULL; |
| 1532 | spin_lock_irqsave(&conf->device_lock, flags); | ||
| 1533 | if (plug) { | 1535 | if (plug) { |
| 1534 | bio_list_add(&plug->pending, mbio); | 1536 | bio_list_add(&plug->pending, mbio); |
| 1535 | plug->pending_cnt++; | 1537 | plug->pending_cnt++; |
| 1536 | } else { | 1538 | } else { |
| 1539 | spin_lock_irqsave(&conf->device_lock, flags); | ||
| 1537 | bio_list_add(&conf->pending_bio_list, mbio); | 1540 | bio_list_add(&conf->pending_bio_list, mbio); |
| 1538 | conf->pending_count++; | 1541 | conf->pending_count++; |
| 1539 | } | 1542 | spin_unlock_irqrestore(&conf->device_lock, flags); |
| 1540 | spin_unlock_irqrestore(&conf->device_lock, flags); | ||
| 1541 | if (!plug) | ||
| 1542 | md_wakeup_thread(mddev->thread); | 1543 | md_wakeup_thread(mddev->thread); |
| 1544 | } | ||
| 1543 | } | 1545 | } |
| 1544 | 1546 | ||
| 1545 | r1_bio_write_done(r1_bio); | 1547 | r1_bio_write_done(r1_bio); |
| @@ -3197,7 +3199,7 @@ static int raid1_reshape(struct mddev *mddev) | |||
| 3197 | struct r1conf *conf = mddev->private; | 3199 | struct r1conf *conf = mddev->private; |
| 3198 | int cnt, raid_disks; | 3200 | int cnt, raid_disks; |
| 3199 | unsigned long flags; | 3201 | unsigned long flags; |
| 3200 | int d, d2, err; | 3202 | int d, d2; |
| 3201 | 3203 | ||
| 3202 | /* Cannot change chunk_size, layout, or level */ | 3204 | /* Cannot change chunk_size, layout, or level */ |
| 3203 | if (mddev->chunk_sectors != mddev->new_chunk_sectors || | 3205 | if (mddev->chunk_sectors != mddev->new_chunk_sectors || |
| @@ -3209,11 +3211,8 @@ static int raid1_reshape(struct mddev *mddev) | |||
| 3209 | return -EINVAL; | 3211 | return -EINVAL; |
| 3210 | } | 3212 | } |
| 3211 | 3213 | ||
| 3212 | if (!mddev_is_clustered(mddev)) { | 3214 | if (!mddev_is_clustered(mddev)) |
| 3213 | err = md_allow_write(mddev); | 3215 | md_allow_write(mddev); |
| 3214 | if (err) | ||
| 3215 | return err; | ||
| 3216 | } | ||
| 3217 | 3216 | ||
| 3218 | raid_disks = mddev->raid_disks + mddev->delta_disks; | 3217 | raid_disks = mddev->raid_disks + mddev->delta_disks; |
| 3219 | 3218 | ||
diff --git a/drivers/md/raid10.c b/drivers/md/raid10.c index 6b86a0032cf8..4343d7ff9916 100644 --- a/drivers/md/raid10.c +++ b/drivers/md/raid10.c | |||
| @@ -1282,17 +1282,16 @@ static void raid10_write_one_disk(struct mddev *mddev, struct r10bio *r10_bio, | |||
| 1282 | plug = container_of(cb, struct raid10_plug_cb, cb); | 1282 | plug = container_of(cb, struct raid10_plug_cb, cb); |
| 1283 | else | 1283 | else |
| 1284 | plug = NULL; | 1284 | plug = NULL; |
| 1285 | spin_lock_irqsave(&conf->device_lock, flags); | ||
| 1286 | if (plug) { | 1285 | if (plug) { |
| 1287 | bio_list_add(&plug->pending, mbio); | 1286 | bio_list_add(&plug->pending, mbio); |
| 1288 | plug->pending_cnt++; | 1287 | plug->pending_cnt++; |
| 1289 | } else { | 1288 | } else { |
| 1289 | spin_lock_irqsave(&conf->device_lock, flags); | ||
| 1290 | bio_list_add(&conf->pending_bio_list, mbio); | 1290 | bio_list_add(&conf->pending_bio_list, mbio); |
| 1291 | conf->pending_count++; | 1291 | conf->pending_count++; |
| 1292 | } | 1292 | spin_unlock_irqrestore(&conf->device_lock, flags); |
| 1293 | spin_unlock_irqrestore(&conf->device_lock, flags); | ||
| 1294 | if (!plug) | ||
| 1295 | md_wakeup_thread(mddev->thread); | 1293 | md_wakeup_thread(mddev->thread); |
| 1294 | } | ||
| 1296 | } | 1295 | } |
| 1297 | 1296 | ||
| 1298 | static void raid10_write_request(struct mddev *mddev, struct bio *bio, | 1297 | static void raid10_write_request(struct mddev *mddev, struct bio *bio, |
diff --git a/drivers/md/raid5-cache.c b/drivers/md/raid5-cache.c index 26ba09282e7c..4c00bc248287 100644 --- a/drivers/md/raid5-cache.c +++ b/drivers/md/raid5-cache.c | |||
| @@ -24,6 +24,7 @@ | |||
| 24 | #include "md.h" | 24 | #include "md.h" |
| 25 | #include "raid5.h" | 25 | #include "raid5.h" |
| 26 | #include "bitmap.h" | 26 | #include "bitmap.h" |
| 27 | #include "raid5-log.h" | ||
| 27 | 28 | ||
| 28 | /* | 29 | /* |
| 29 | * metadata/data stored in disk with 4k size unit (a block) regardless | 30 | * metadata/data stored in disk with 4k size unit (a block) regardless |
| @@ -622,20 +623,30 @@ static void r5l_do_submit_io(struct r5l_log *log, struct r5l_io_unit *io) | |||
| 622 | __r5l_set_io_unit_state(io, IO_UNIT_IO_START); | 623 | __r5l_set_io_unit_state(io, IO_UNIT_IO_START); |
| 623 | spin_unlock_irqrestore(&log->io_list_lock, flags); | 624 | spin_unlock_irqrestore(&log->io_list_lock, flags); |
| 624 | 625 | ||
| 626 | /* | ||
| 627 | * In case of journal device failures, submit_bio will get error | ||
| 628 | * and calls endio, then active stripes will continue write | ||
| 629 | * process. Therefore, it is not necessary to check Faulty bit | ||
| 630 | * of journal device here. | ||
| 631 | * | ||
| 632 | * We can't check split_bio after current_bio is submitted. If | ||
| 633 | * io->split_bio is null, after current_bio is submitted, current_bio | ||
| 634 | * might already be completed and the io_unit is freed. We submit | ||
| 635 | * split_bio first to avoid the issue. | ||
| 636 | */ | ||
| 637 | if (io->split_bio) { | ||
| 638 | if (io->has_flush) | ||
| 639 | io->split_bio->bi_opf |= REQ_PREFLUSH; | ||
| 640 | if (io->has_fua) | ||
| 641 | io->split_bio->bi_opf |= REQ_FUA; | ||
| 642 | submit_bio(io->split_bio); | ||
| 643 | } | ||
| 644 | |||
| 625 | if (io->has_flush) | 645 | if (io->has_flush) |
| 626 | io->current_bio->bi_opf |= REQ_PREFLUSH; | 646 | io->current_bio->bi_opf |= REQ_PREFLUSH; |
| 627 | if (io->has_fua) | 647 | if (io->has_fua) |
| 628 | io->current_bio->bi_opf |= REQ_FUA; | 648 | io->current_bio->bi_opf |= REQ_FUA; |
| 629 | submit_bio(io->current_bio); | 649 | submit_bio(io->current_bio); |
| 630 | |||
| 631 | if (!io->split_bio) | ||
| 632 | return; | ||
| 633 | |||
| 634 | if (io->has_flush) | ||
| 635 | io->split_bio->bi_opf |= REQ_PREFLUSH; | ||
| 636 | if (io->has_fua) | ||
| 637 | io->split_bio->bi_opf |= REQ_FUA; | ||
| 638 | submit_bio(io->split_bio); | ||
| 639 | } | 650 | } |
| 640 | 651 | ||
| 641 | /* deferred io_unit will be dispatched here */ | 652 | /* deferred io_unit will be dispatched here */ |
| @@ -670,6 +681,11 @@ static void r5c_disable_writeback_async(struct work_struct *work) | |||
| 670 | return; | 681 | return; |
| 671 | pr_info("md/raid:%s: Disabling writeback cache for degraded array.\n", | 682 | pr_info("md/raid:%s: Disabling writeback cache for degraded array.\n", |
| 672 | mdname(mddev)); | 683 | mdname(mddev)); |
| 684 | |||
| 685 | /* wait superblock change before suspend */ | ||
| 686 | wait_event(mddev->sb_wait, | ||
| 687 | !test_bit(MD_SB_CHANGE_PENDING, &mddev->sb_flags)); | ||
| 688 | |||
| 673 | mddev_suspend(mddev); | 689 | mddev_suspend(mddev); |
| 674 | log->r5c_journal_mode = R5C_JOURNAL_MODE_WRITE_THROUGH; | 690 | log->r5c_journal_mode = R5C_JOURNAL_MODE_WRITE_THROUGH; |
| 675 | mddev_resume(mddev); | 691 | mddev_resume(mddev); |
| @@ -2621,8 +2637,11 @@ int r5c_try_caching_write(struct r5conf *conf, | |||
| 2621 | * When run in degraded mode, array is set to write-through mode. | 2637 | * When run in degraded mode, array is set to write-through mode. |
| 2622 | * This check helps drain pending write safely in the transition to | 2638 | * This check helps drain pending write safely in the transition to |
| 2623 | * write-through mode. | 2639 | * write-through mode. |
| 2640 | * | ||
| 2641 | * When a stripe is syncing, the write is also handled in write | ||
| 2642 | * through mode. | ||
| 2624 | */ | 2643 | */ |
| 2625 | if (s->failed) { | 2644 | if (s->failed || test_bit(STRIPE_SYNCING, &sh->state)) { |
| 2626 | r5c_make_stripe_write_out(sh); | 2645 | r5c_make_stripe_write_out(sh); |
| 2627 | return -EAGAIN; | 2646 | return -EAGAIN; |
| 2628 | } | 2647 | } |
| @@ -2825,6 +2844,9 @@ void r5c_finish_stripe_write_out(struct r5conf *conf, | |||
| 2825 | } | 2844 | } |
| 2826 | 2845 | ||
| 2827 | r5l_append_flush_payload(log, sh->sector); | 2846 | r5l_append_flush_payload(log, sh->sector); |
| 2847 | /* stripe is flused to raid disks, we can do resync now */ | ||
| 2848 | if (test_bit(STRIPE_SYNC_REQUESTED, &sh->state)) | ||
| 2849 | set_bit(STRIPE_HANDLE, &sh->state); | ||
| 2828 | } | 2850 | } |
| 2829 | 2851 | ||
| 2830 | int r5c_cache_data(struct r5l_log *log, struct stripe_head *sh) | 2852 | int r5c_cache_data(struct r5l_log *log, struct stripe_head *sh) |
| @@ -2973,7 +2995,7 @@ ioerr: | |||
| 2973 | return ret; | 2995 | return ret; |
| 2974 | } | 2996 | } |
| 2975 | 2997 | ||
| 2976 | void r5c_update_on_rdev_error(struct mddev *mddev) | 2998 | void r5c_update_on_rdev_error(struct mddev *mddev, struct md_rdev *rdev) |
| 2977 | { | 2999 | { |
| 2978 | struct r5conf *conf = mddev->private; | 3000 | struct r5conf *conf = mddev->private; |
| 2979 | struct r5l_log *log = conf->log; | 3001 | struct r5l_log *log = conf->log; |
| @@ -2981,7 +3003,8 @@ void r5c_update_on_rdev_error(struct mddev *mddev) | |||
| 2981 | if (!log) | 3003 | if (!log) |
| 2982 | return; | 3004 | return; |
| 2983 | 3005 | ||
| 2984 | if (raid5_calc_degraded(conf) > 0 && | 3006 | if ((raid5_calc_degraded(conf) > 0 || |
| 3007 | test_bit(Journal, &rdev->flags)) && | ||
| 2985 | conf->log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_BACK) | 3008 | conf->log->r5c_journal_mode == R5C_JOURNAL_MODE_WRITE_BACK) |
| 2986 | schedule_work(&log->disable_writeback_work); | 3009 | schedule_work(&log->disable_writeback_work); |
| 2987 | } | 3010 | } |
diff --git a/drivers/md/raid5-log.h b/drivers/md/raid5-log.h index 27097101ccca..328d67aedda4 100644 --- a/drivers/md/raid5-log.h +++ b/drivers/md/raid5-log.h | |||
| @@ -28,7 +28,8 @@ extern void r5c_flush_cache(struct r5conf *conf, int num); | |||
| 28 | extern void r5c_check_stripe_cache_usage(struct r5conf *conf); | 28 | extern void r5c_check_stripe_cache_usage(struct r5conf *conf); |
| 29 | extern void r5c_check_cached_full_stripe(struct r5conf *conf); | 29 | extern void r5c_check_cached_full_stripe(struct r5conf *conf); |
| 30 | extern struct md_sysfs_entry r5c_journal_mode; | 30 | extern struct md_sysfs_entry r5c_journal_mode; |
| 31 | extern void r5c_update_on_rdev_error(struct mddev *mddev); | 31 | extern void r5c_update_on_rdev_error(struct mddev *mddev, |
| 32 | struct md_rdev *rdev); | ||
| 32 | extern bool r5c_big_stripe_cached(struct r5conf *conf, sector_t sect); | 33 | extern bool r5c_big_stripe_cached(struct r5conf *conf, sector_t sect); |
| 33 | 34 | ||
| 34 | extern struct dma_async_tx_descriptor * | 35 | extern struct dma_async_tx_descriptor * |
diff --git a/drivers/md/raid5.c b/drivers/md/raid5.c index 2e38cfac5b1d..9c4f7659f8b1 100644 --- a/drivers/md/raid5.c +++ b/drivers/md/raid5.c | |||
| @@ -103,8 +103,7 @@ static inline void unlock_device_hash_lock(struct r5conf *conf, int hash) | |||
| 103 | static inline void lock_all_device_hash_locks_irq(struct r5conf *conf) | 103 | static inline void lock_all_device_hash_locks_irq(struct r5conf *conf) |
| 104 | { | 104 | { |
| 105 | int i; | 105 | int i; |
| 106 | local_irq_disable(); | 106 | spin_lock_irq(conf->hash_locks); |
| 107 | spin_lock(conf->hash_locks); | ||
| 108 | for (i = 1; i < NR_STRIPE_HASH_LOCKS; i++) | 107 | for (i = 1; i < NR_STRIPE_HASH_LOCKS; i++) |
| 109 | spin_lock_nest_lock(conf->hash_locks + i, conf->hash_locks); | 108 | spin_lock_nest_lock(conf->hash_locks + i, conf->hash_locks); |
| 110 | spin_lock(&conf->device_lock); | 109 | spin_lock(&conf->device_lock); |
| @@ -114,9 +113,9 @@ static inline void unlock_all_device_hash_locks_irq(struct r5conf *conf) | |||
| 114 | { | 113 | { |
| 115 | int i; | 114 | int i; |
| 116 | spin_unlock(&conf->device_lock); | 115 | spin_unlock(&conf->device_lock); |
| 117 | for (i = NR_STRIPE_HASH_LOCKS; i; i--) | 116 | for (i = NR_STRIPE_HASH_LOCKS - 1; i; i--) |
| 118 | spin_unlock(conf->hash_locks + i - 1); | 117 | spin_unlock(conf->hash_locks + i); |
| 119 | local_irq_enable(); | 118 | spin_unlock_irq(conf->hash_locks); |
| 120 | } | 119 | } |
| 121 | 120 | ||
| 122 | /* Find first data disk in a raid6 stripe */ | 121 | /* Find first data disk in a raid6 stripe */ |
| @@ -234,11 +233,15 @@ static void do_release_stripe(struct r5conf *conf, struct stripe_head *sh, | |||
| 234 | if (test_bit(R5_InJournal, &sh->dev[i].flags)) | 233 | if (test_bit(R5_InJournal, &sh->dev[i].flags)) |
| 235 | injournal++; | 234 | injournal++; |
| 236 | /* | 235 | /* |
| 237 | * When quiesce in r5c write back, set STRIPE_HANDLE for stripes with | 236 | * In the following cases, the stripe cannot be released to cached |
| 238 | * data in journal, so they are not released to cached lists | 237 | * lists. Therefore, we make the stripe write out and set |
| 238 | * STRIPE_HANDLE: | ||
| 239 | * 1. when quiesce in r5c write back; | ||
| 240 | * 2. when resync is requested fot the stripe. | ||
| 239 | */ | 241 | */ |
| 240 | if (conf->quiesce && r5c_is_writeback(conf->log) && | 242 | if (test_bit(STRIPE_SYNC_REQUESTED, &sh->state) || |
| 241 | !test_bit(STRIPE_HANDLE, &sh->state) && injournal != 0) { | 243 | (conf->quiesce && r5c_is_writeback(conf->log) && |
| 244 | !test_bit(STRIPE_HANDLE, &sh->state) && injournal != 0)) { | ||
| 242 | if (test_bit(STRIPE_R5C_CACHING, &sh->state)) | 245 | if (test_bit(STRIPE_R5C_CACHING, &sh->state)) |
| 243 | r5c_make_stripe_write_out(sh); | 246 | r5c_make_stripe_write_out(sh); |
| 244 | set_bit(STRIPE_HANDLE, &sh->state); | 247 | set_bit(STRIPE_HANDLE, &sh->state); |
| @@ -714,12 +717,11 @@ static bool is_full_stripe_write(struct stripe_head *sh) | |||
| 714 | 717 | ||
| 715 | static void lock_two_stripes(struct stripe_head *sh1, struct stripe_head *sh2) | 718 | static void lock_two_stripes(struct stripe_head *sh1, struct stripe_head *sh2) |
| 716 | { | 719 | { |
| 717 | local_irq_disable(); | ||
| 718 | if (sh1 > sh2) { | 720 | if (sh1 > sh2) { |
| 719 | spin_lock(&sh2->stripe_lock); | 721 | spin_lock_irq(&sh2->stripe_lock); |
| 720 | spin_lock_nested(&sh1->stripe_lock, 1); | 722 | spin_lock_nested(&sh1->stripe_lock, 1); |
| 721 | } else { | 723 | } else { |
| 722 | spin_lock(&sh1->stripe_lock); | 724 | spin_lock_irq(&sh1->stripe_lock); |
| 723 | spin_lock_nested(&sh2->stripe_lock, 1); | 725 | spin_lock_nested(&sh2->stripe_lock, 1); |
| 724 | } | 726 | } |
| 725 | } | 727 | } |
| @@ -727,8 +729,7 @@ static void lock_two_stripes(struct stripe_head *sh1, struct stripe_head *sh2) | |||
| 727 | static void unlock_two_stripes(struct stripe_head *sh1, struct stripe_head *sh2) | 729 | static void unlock_two_stripes(struct stripe_head *sh1, struct stripe_head *sh2) |
| 728 | { | 730 | { |
| 729 | spin_unlock(&sh1->stripe_lock); | 731 | spin_unlock(&sh1->stripe_lock); |
| 730 | spin_unlock(&sh2->stripe_lock); | 732 | spin_unlock_irq(&sh2->stripe_lock); |
| 731 | local_irq_enable(); | ||
| 732 | } | 733 | } |
| 733 | 734 | ||
| 734 | /* Only freshly new full stripe normal write stripe can be added to a batch list */ | 735 | /* Only freshly new full stripe normal write stripe can be added to a batch list */ |
| @@ -2312,14 +2313,12 @@ static int resize_stripes(struct r5conf *conf, int newsize) | |||
| 2312 | struct stripe_head *osh, *nsh; | 2313 | struct stripe_head *osh, *nsh; |
| 2313 | LIST_HEAD(newstripes); | 2314 | LIST_HEAD(newstripes); |
| 2314 | struct disk_info *ndisks; | 2315 | struct disk_info *ndisks; |
| 2315 | int err; | 2316 | int err = 0; |
| 2316 | struct kmem_cache *sc; | 2317 | struct kmem_cache *sc; |
| 2317 | int i; | 2318 | int i; |
| 2318 | int hash, cnt; | 2319 | int hash, cnt; |
| 2319 | 2320 | ||
| 2320 | err = md_allow_write(conf->mddev); | 2321 | md_allow_write(conf->mddev); |
| 2321 | if (err) | ||
| 2322 | return err; | ||
| 2323 | 2322 | ||
| 2324 | /* Step 1 */ | 2323 | /* Step 1 */ |
| 2325 | sc = kmem_cache_create(conf->cache_name[1-conf->active_name], | 2324 | sc = kmem_cache_create(conf->cache_name[1-conf->active_name], |
| @@ -2694,7 +2693,7 @@ static void raid5_error(struct mddev *mddev, struct md_rdev *rdev) | |||
| 2694 | bdevname(rdev->bdev, b), | 2693 | bdevname(rdev->bdev, b), |
| 2695 | mdname(mddev), | 2694 | mdname(mddev), |
| 2696 | conf->raid_disks - mddev->degraded); | 2695 | conf->raid_disks - mddev->degraded); |
| 2697 | r5c_update_on_rdev_error(mddev); | 2696 | r5c_update_on_rdev_error(mddev, rdev); |
| 2698 | } | 2697 | } |
| 2699 | 2698 | ||
| 2700 | /* | 2699 | /* |
| @@ -3055,6 +3054,11 @@ sector_t raid5_compute_blocknr(struct stripe_head *sh, int i, int previous) | |||
| 3055 | * When LOG_CRITICAL, stripes with injournal == 0 will be sent to | 3054 | * When LOG_CRITICAL, stripes with injournal == 0 will be sent to |
| 3056 | * no_space_stripes list. | 3055 | * no_space_stripes list. |
| 3057 | * | 3056 | * |
| 3057 | * 3. during journal failure | ||
| 3058 | * In journal failure, we try to flush all cached data to raid disks | ||
| 3059 | * based on data in stripe cache. The array is read-only to upper | ||
| 3060 | * layers, so we would skip all pending writes. | ||
| 3061 | * | ||
| 3058 | */ | 3062 | */ |
| 3059 | static inline bool delay_towrite(struct r5conf *conf, | 3063 | static inline bool delay_towrite(struct r5conf *conf, |
| 3060 | struct r5dev *dev, | 3064 | struct r5dev *dev, |
| @@ -3068,6 +3072,9 @@ static inline bool delay_towrite(struct r5conf *conf, | |||
| 3068 | if (test_bit(R5C_LOG_CRITICAL, &conf->cache_state) && | 3072 | if (test_bit(R5C_LOG_CRITICAL, &conf->cache_state) && |
| 3069 | s->injournal > 0) | 3073 | s->injournal > 0) |
| 3070 | return true; | 3074 | return true; |
| 3075 | /* case 3 above */ | ||
| 3076 | if (s->log_failed && s->injournal) | ||
| 3077 | return true; | ||
| 3071 | return false; | 3078 | return false; |
| 3072 | } | 3079 | } |
| 3073 | 3080 | ||
| @@ -4653,8 +4660,13 @@ static void handle_stripe(struct stripe_head *sh) | |||
| 4653 | 4660 | ||
| 4654 | if (test_bit(STRIPE_SYNC_REQUESTED, &sh->state) && !sh->batch_head) { | 4661 | if (test_bit(STRIPE_SYNC_REQUESTED, &sh->state) && !sh->batch_head) { |
| 4655 | spin_lock(&sh->stripe_lock); | 4662 | spin_lock(&sh->stripe_lock); |
| 4656 | /* Cannot process 'sync' concurrently with 'discard' */ | 4663 | /* |
| 4657 | if (!test_bit(STRIPE_DISCARD, &sh->state) && | 4664 | * Cannot process 'sync' concurrently with 'discard'. |
| 4665 | * Flush data in r5cache before 'sync'. | ||
| 4666 | */ | ||
| 4667 | if (!test_bit(STRIPE_R5C_PARTIAL_STRIPE, &sh->state) && | ||
| 4668 | !test_bit(STRIPE_R5C_FULL_STRIPE, &sh->state) && | ||
| 4669 | !test_bit(STRIPE_DISCARD, &sh->state) && | ||
| 4658 | test_and_clear_bit(STRIPE_SYNC_REQUESTED, &sh->state)) { | 4670 | test_and_clear_bit(STRIPE_SYNC_REQUESTED, &sh->state)) { |
| 4659 | set_bit(STRIPE_SYNCING, &sh->state); | 4671 | set_bit(STRIPE_SYNCING, &sh->state); |
| 4660 | clear_bit(STRIPE_INSYNC, &sh->state); | 4672 | clear_bit(STRIPE_INSYNC, &sh->state); |
| @@ -4701,10 +4713,15 @@ static void handle_stripe(struct stripe_head *sh) | |||
| 4701 | " to_write=%d failed=%d failed_num=%d,%d\n", | 4713 | " to_write=%d failed=%d failed_num=%d,%d\n", |
| 4702 | s.locked, s.uptodate, s.to_read, s.to_write, s.failed, | 4714 | s.locked, s.uptodate, s.to_read, s.to_write, s.failed, |
| 4703 | s.failed_num[0], s.failed_num[1]); | 4715 | s.failed_num[0], s.failed_num[1]); |
| 4704 | /* check if the array has lost more than max_degraded devices and, | 4716 | /* |
| 4717 | * check if the array has lost more than max_degraded devices and, | ||
| 4705 | * if so, some requests might need to be failed. | 4718 | * if so, some requests might need to be failed. |
| 4719 | * | ||
| 4720 | * When journal device failed (log_failed), we will only process | ||
| 4721 | * the stripe if there is data need write to raid disks | ||
| 4706 | */ | 4722 | */ |
| 4707 | if (s.failed > conf->max_degraded || s.log_failed) { | 4723 | if (s.failed > conf->max_degraded || |
| 4724 | (s.log_failed && s.injournal == 0)) { | ||
| 4708 | sh->check_state = 0; | 4725 | sh->check_state = 0; |
| 4709 | sh->reconstruct_state = 0; | 4726 | sh->reconstruct_state = 0; |
| 4710 | break_stripe_batch_list(sh, 0); | 4727 | break_stripe_batch_list(sh, 0); |
| @@ -5277,8 +5294,10 @@ static struct stripe_head *__get_priority_stripe(struct r5conf *conf, int group) | |||
| 5277 | struct stripe_head *sh, *tmp; | 5294 | struct stripe_head *sh, *tmp; |
| 5278 | struct list_head *handle_list = NULL; | 5295 | struct list_head *handle_list = NULL; |
| 5279 | struct r5worker_group *wg; | 5296 | struct r5worker_group *wg; |
| 5280 | bool second_try = !r5c_is_writeback(conf->log); | 5297 | bool second_try = !r5c_is_writeback(conf->log) && |
| 5281 | bool try_loprio = test_bit(R5C_LOG_TIGHT, &conf->cache_state); | 5298 | !r5l_log_disk_error(conf); |
| 5299 | bool try_loprio = test_bit(R5C_LOG_TIGHT, &conf->cache_state) || | ||
| 5300 | r5l_log_disk_error(conf); | ||
| 5282 | 5301 | ||
| 5283 | again: | 5302 | again: |
| 5284 | wg = NULL; | 5303 | wg = NULL; |
| @@ -6313,7 +6332,6 @@ int | |||
| 6313 | raid5_set_cache_size(struct mddev *mddev, int size) | 6332 | raid5_set_cache_size(struct mddev *mddev, int size) |
| 6314 | { | 6333 | { |
| 6315 | struct r5conf *conf = mddev->private; | 6334 | struct r5conf *conf = mddev->private; |
| 6316 | int err; | ||
| 6317 | 6335 | ||
| 6318 | if (size <= 16 || size > 32768) | 6336 | if (size <= 16 || size > 32768) |
| 6319 | return -EINVAL; | 6337 | return -EINVAL; |
| @@ -6325,10 +6343,7 @@ raid5_set_cache_size(struct mddev *mddev, int size) | |||
| 6325 | ; | 6343 | ; |
| 6326 | mutex_unlock(&conf->cache_size_mutex); | 6344 | mutex_unlock(&conf->cache_size_mutex); |
| 6327 | 6345 | ||
| 6328 | 6346 | md_allow_write(mddev); | |
| 6329 | err = md_allow_write(mddev); | ||
| 6330 | if (err) | ||
| 6331 | return err; | ||
| 6332 | 6347 | ||
| 6333 | mutex_lock(&conf->cache_size_mutex); | 6348 | mutex_lock(&conf->cache_size_mutex); |
| 6334 | while (size > conf->max_nr_stripes) | 6349 | while (size > conf->max_nr_stripes) |
| @@ -7530,7 +7545,9 @@ static int raid5_remove_disk(struct mddev *mddev, struct md_rdev *rdev) | |||
| 7530 | * neilb: there is no locking about new writes here, | 7545 | * neilb: there is no locking about new writes here, |
| 7531 | * so this cannot be safe. | 7546 | * so this cannot be safe. |
| 7532 | */ | 7547 | */ |
| 7533 | if (atomic_read(&conf->active_stripes)) { | 7548 | if (atomic_read(&conf->active_stripes) || |
| 7549 | atomic_read(&conf->r5c_cached_full_stripes) || | ||
| 7550 | atomic_read(&conf->r5c_cached_partial_stripes)) { | ||
| 7534 | return -EBUSY; | 7551 | return -EBUSY; |
| 7535 | } | 7552 | } |
| 7536 | log_exit(conf); | 7553 | log_exit(conf); |
diff --git a/drivers/net/dsa/mv88e6xxx/chip.c b/drivers/net/dsa/mv88e6xxx/chip.c index 19581d783d8e..d034d8cd7d22 100644 --- a/drivers/net/dsa/mv88e6xxx/chip.c +++ b/drivers/net/dsa/mv88e6xxx/chip.c | |||
| @@ -849,6 +849,9 @@ static uint64_t _mv88e6xxx_get_ethtool_stat(struct mv88e6xxx_chip *chip, | |||
| 849 | mv88e6xxx_g1_stats_read(chip, reg, &low); | 849 | mv88e6xxx_g1_stats_read(chip, reg, &low); |
| 850 | if (s->sizeof_stat == 8) | 850 | if (s->sizeof_stat == 8) |
| 851 | mv88e6xxx_g1_stats_read(chip, reg + 1, &high); | 851 | mv88e6xxx_g1_stats_read(chip, reg + 1, &high); |
| 852 | break; | ||
| 853 | default: | ||
| 854 | return UINT64_MAX; | ||
| 852 | } | 855 | } |
| 853 | value = (((u64)high) << 16) | low; | 856 | value = (((u64)high) << 16) | low; |
| 854 | return value; | 857 | return value; |
diff --git a/drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c b/drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c index 4ee15ff06a44..faeb4935ef3e 100644 --- a/drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c +++ b/drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c | |||
| @@ -200,29 +200,18 @@ err_exit: | |||
| 200 | static int hw_atl_a0_hw_offload_set(struct aq_hw_s *self, | 200 | static int hw_atl_a0_hw_offload_set(struct aq_hw_s *self, |
| 201 | struct aq_nic_cfg_s *aq_nic_cfg) | 201 | struct aq_nic_cfg_s *aq_nic_cfg) |
| 202 | { | 202 | { |
| 203 | int err = 0; | ||
| 204 | |||
| 205 | /* TX checksums offloads*/ | 203 | /* TX checksums offloads*/ |
| 206 | tpo_ipv4header_crc_offload_en_set(self, 1); | 204 | tpo_ipv4header_crc_offload_en_set(self, 1); |
| 207 | tpo_tcp_udp_crc_offload_en_set(self, 1); | 205 | tpo_tcp_udp_crc_offload_en_set(self, 1); |
| 208 | if (err < 0) | ||
| 209 | goto err_exit; | ||
| 210 | 206 | ||
| 211 | /* RX checksums offloads*/ | 207 | /* RX checksums offloads*/ |
| 212 | rpo_ipv4header_crc_offload_en_set(self, 1); | 208 | rpo_ipv4header_crc_offload_en_set(self, 1); |
| 213 | rpo_tcp_udp_crc_offload_en_set(self, 1); | 209 | rpo_tcp_udp_crc_offload_en_set(self, 1); |
| 214 | if (err < 0) | ||
| 215 | goto err_exit; | ||
| 216 | 210 | ||
| 217 | /* LSO offloads*/ | 211 | /* LSO offloads*/ |
| 218 | tdm_large_send_offload_en_set(self, 0xFFFFFFFFU); | 212 | tdm_large_send_offload_en_set(self, 0xFFFFFFFFU); |
| 219 | if (err < 0) | ||
| 220 | goto err_exit; | ||
| 221 | |||
| 222 | err = aq_hw_err_from_flags(self); | ||
| 223 | 213 | ||
| 224 | err_exit: | 214 | return aq_hw_err_from_flags(self); |
| 225 | return err; | ||
| 226 | } | 215 | } |
| 227 | 216 | ||
| 228 | static int hw_atl_a0_hw_init_tx_path(struct aq_hw_s *self) | 217 | static int hw_atl_a0_hw_init_tx_path(struct aq_hw_s *self) |
diff --git a/drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c b/drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c index 42150708191d..1bceb7358e5c 100644 --- a/drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c +++ b/drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c | |||
| @@ -200,25 +200,18 @@ err_exit: | |||
| 200 | static int hw_atl_b0_hw_offload_set(struct aq_hw_s *self, | 200 | static int hw_atl_b0_hw_offload_set(struct aq_hw_s *self, |
| 201 | struct aq_nic_cfg_s *aq_nic_cfg) | 201 | struct aq_nic_cfg_s *aq_nic_cfg) |
| 202 | { | 202 | { |
| 203 | int err = 0; | ||
| 204 | unsigned int i; | 203 | unsigned int i; |
| 205 | 204 | ||
| 206 | /* TX checksums offloads*/ | 205 | /* TX checksums offloads*/ |
| 207 | tpo_ipv4header_crc_offload_en_set(self, 1); | 206 | tpo_ipv4header_crc_offload_en_set(self, 1); |
| 208 | tpo_tcp_udp_crc_offload_en_set(self, 1); | 207 | tpo_tcp_udp_crc_offload_en_set(self, 1); |
| 209 | if (err < 0) | ||
| 210 | goto err_exit; | ||
| 211 | 208 | ||
| 212 | /* RX checksums offloads*/ | 209 | /* RX checksums offloads*/ |
| 213 | rpo_ipv4header_crc_offload_en_set(self, 1); | 210 | rpo_ipv4header_crc_offload_en_set(self, 1); |
| 214 | rpo_tcp_udp_crc_offload_en_set(self, 1); | 211 | rpo_tcp_udp_crc_offload_en_set(self, 1); |
| 215 | if (err < 0) | ||
| 216 | goto err_exit; | ||
| 217 | 212 | ||
| 218 | /* LSO offloads*/ | 213 | /* LSO offloads*/ |
| 219 | tdm_large_send_offload_en_set(self, 0xFFFFFFFFU); | 214 | tdm_large_send_offload_en_set(self, 0xFFFFFFFFU); |
| 220 | if (err < 0) | ||
| 221 | goto err_exit; | ||
| 222 | 215 | ||
| 223 | /* LRO offloads */ | 216 | /* LRO offloads */ |
| 224 | { | 217 | { |
| @@ -245,10 +238,7 @@ static int hw_atl_b0_hw_offload_set(struct aq_hw_s *self, | |||
| 245 | 238 | ||
| 246 | rpo_lro_en_set(self, aq_nic_cfg->is_lro ? 0xFFFFFFFFU : 0U); | 239 | rpo_lro_en_set(self, aq_nic_cfg->is_lro ? 0xFFFFFFFFU : 0U); |
| 247 | } | 240 | } |
| 248 | err = aq_hw_err_from_flags(self); | 241 | return aq_hw_err_from_flags(self); |
| 249 | |||
| 250 | err_exit: | ||
| 251 | return err; | ||
| 252 | } | 242 | } |
| 253 | 243 | ||
| 254 | static int hw_atl_b0_hw_init_tx_path(struct aq_hw_s *self) | 244 | static int hw_atl_b0_hw_init_tx_path(struct aq_hw_s *self) |
diff --git a/drivers/net/ethernet/broadcom/bnxt/bnxt.c b/drivers/net/ethernet/broadcom/bnxt/bnxt.c index b56c54d68d5e..03f55daecb20 100644 --- a/drivers/net/ethernet/broadcom/bnxt/bnxt.c +++ b/drivers/net/ethernet/broadcom/bnxt/bnxt.c | |||
| @@ -7630,8 +7630,6 @@ static int bnxt_init_one(struct pci_dev *pdev, const struct pci_device_id *ent) | |||
| 7630 | dev->min_mtu = ETH_ZLEN; | 7630 | dev->min_mtu = ETH_ZLEN; |
| 7631 | dev->max_mtu = BNXT_MAX_MTU; | 7631 | dev->max_mtu = BNXT_MAX_MTU; |
| 7632 | 7632 | ||
| 7633 | bnxt_dcb_init(bp); | ||
| 7634 | |||
| 7635 | #ifdef CONFIG_BNXT_SRIOV | 7633 | #ifdef CONFIG_BNXT_SRIOV |
| 7636 | init_waitqueue_head(&bp->sriov_cfg_wait); | 7634 | init_waitqueue_head(&bp->sriov_cfg_wait); |
| 7637 | #endif | 7635 | #endif |
| @@ -7669,6 +7667,7 @@ static int bnxt_init_one(struct pci_dev *pdev, const struct pci_device_id *ent) | |||
| 7669 | bnxt_hwrm_func_qcfg(bp); | 7667 | bnxt_hwrm_func_qcfg(bp); |
| 7670 | bnxt_hwrm_port_led_qcaps(bp); | 7668 | bnxt_hwrm_port_led_qcaps(bp); |
| 7671 | bnxt_ethtool_init(bp); | 7669 | bnxt_ethtool_init(bp); |
| 7670 | bnxt_dcb_init(bp); | ||
| 7672 | 7671 | ||
| 7673 | bnxt_set_rx_skb_mode(bp, false); | 7672 | bnxt_set_rx_skb_mode(bp, false); |
| 7674 | bnxt_set_tpa_flags(bp); | 7673 | bnxt_set_tpa_flags(bp); |
diff --git a/drivers/net/ethernet/broadcom/bnxt/bnxt_dcb.c b/drivers/net/ethernet/broadcom/bnxt/bnxt_dcb.c index 46de2f8ff024..5c6dd0ce209f 100644 --- a/drivers/net/ethernet/broadcom/bnxt/bnxt_dcb.c +++ b/drivers/net/ethernet/broadcom/bnxt/bnxt_dcb.c | |||
| @@ -553,8 +553,10 @@ static u8 bnxt_dcbnl_setdcbx(struct net_device *dev, u8 mode) | |||
| 553 | if ((mode & DCB_CAP_DCBX_VER_CEE) || !(mode & DCB_CAP_DCBX_VER_IEEE)) | 553 | if ((mode & DCB_CAP_DCBX_VER_CEE) || !(mode & DCB_CAP_DCBX_VER_IEEE)) |
| 554 | return 1; | 554 | return 1; |
| 555 | 555 | ||
| 556 | if ((mode & DCB_CAP_DCBX_HOST) && BNXT_VF(bp)) | 556 | if (mode & DCB_CAP_DCBX_HOST) { |
| 557 | return 1; | 557 | if (BNXT_VF(bp) || (bp->flags & BNXT_FLAG_FW_LLDP_AGENT)) |
| 558 | return 1; | ||
| 559 | } | ||
| 558 | 560 | ||
| 559 | if (mode == bp->dcbx_cap) | 561 | if (mode == bp->dcbx_cap) |
| 560 | return 0; | 562 | return 0; |
diff --git a/drivers/net/ethernet/chelsio/cxgb4/t4fw_version.h b/drivers/net/ethernet/chelsio/cxgb4/t4fw_version.h index fa376444e57c..3549d3876278 100644 --- a/drivers/net/ethernet/chelsio/cxgb4/t4fw_version.h +++ b/drivers/net/ethernet/chelsio/cxgb4/t4fw_version.h | |||
| @@ -37,7 +37,7 @@ | |||
| 37 | 37 | ||
| 38 | #define T4FW_VERSION_MAJOR 0x01 | 38 | #define T4FW_VERSION_MAJOR 0x01 |
| 39 | #define T4FW_VERSION_MINOR 0x10 | 39 | #define T4FW_VERSION_MINOR 0x10 |
| 40 | #define T4FW_VERSION_MICRO 0x21 | 40 | #define T4FW_VERSION_MICRO 0x2B |
| 41 | #define T4FW_VERSION_BUILD 0x00 | 41 | #define T4FW_VERSION_BUILD 0x00 |
| 42 | 42 | ||
| 43 | #define T4FW_MIN_VERSION_MAJOR 0x01 | 43 | #define T4FW_MIN_VERSION_MAJOR 0x01 |
| @@ -46,7 +46,7 @@ | |||
| 46 | 46 | ||
| 47 | #define T5FW_VERSION_MAJOR 0x01 | 47 | #define T5FW_VERSION_MAJOR 0x01 |
| 48 | #define T5FW_VERSION_MINOR 0x10 | 48 | #define T5FW_VERSION_MINOR 0x10 |
| 49 | #define T5FW_VERSION_MICRO 0x21 | 49 | #define T5FW_VERSION_MICRO 0x2B |
| 50 | #define T5FW_VERSION_BUILD 0x00 | 50 | #define T5FW_VERSION_BUILD 0x00 |
| 51 | 51 | ||
| 52 | #define T5FW_MIN_VERSION_MAJOR 0x00 | 52 | #define T5FW_MIN_VERSION_MAJOR 0x00 |
| @@ -55,7 +55,7 @@ | |||
| 55 | 55 | ||
| 56 | #define T6FW_VERSION_MAJOR 0x01 | 56 | #define T6FW_VERSION_MAJOR 0x01 |
| 57 | #define T6FW_VERSION_MINOR 0x10 | 57 | #define T6FW_VERSION_MINOR 0x10 |
| 58 | #define T6FW_VERSION_MICRO 0x21 | 58 | #define T6FW_VERSION_MICRO 0x2B |
| 59 | #define T6FW_VERSION_BUILD 0x00 | 59 | #define T6FW_VERSION_BUILD 0x00 |
| 60 | 60 | ||
| 61 | #define T6FW_MIN_VERSION_MAJOR 0x00 | 61 | #define T6FW_MIN_VERSION_MAJOR 0x00 |
diff --git a/drivers/net/ethernet/faraday/ftmac100.c b/drivers/net/ethernet/faraday/ftmac100.c index 6ac336b546e6..1536356e2ea8 100644 --- a/drivers/net/ethernet/faraday/ftmac100.c +++ b/drivers/net/ethernet/faraday/ftmac100.c | |||
| @@ -1174,11 +1174,17 @@ static int ftmac100_remove(struct platform_device *pdev) | |||
| 1174 | return 0; | 1174 | return 0; |
| 1175 | } | 1175 | } |
| 1176 | 1176 | ||
| 1177 | static const struct of_device_id ftmac100_of_ids[] = { | ||
| 1178 | { .compatible = "andestech,atmac100" }, | ||
| 1179 | { } | ||
| 1180 | }; | ||
| 1181 | |||
| 1177 | static struct platform_driver ftmac100_driver = { | 1182 | static struct platform_driver ftmac100_driver = { |
| 1178 | .probe = ftmac100_probe, | 1183 | .probe = ftmac100_probe, |
| 1179 | .remove = ftmac100_remove, | 1184 | .remove = ftmac100_remove, |
| 1180 | .driver = { | 1185 | .driver = { |
| 1181 | .name = DRV_NAME, | 1186 | .name = DRV_NAME, |
| 1187 | .of_match_table = ftmac100_of_ids | ||
| 1182 | }, | 1188 | }, |
| 1183 | }; | 1189 | }; |
| 1184 | 1190 | ||
| @@ -1202,3 +1208,4 @@ module_exit(ftmac100_exit); | |||
| 1202 | MODULE_AUTHOR("Po-Yu Chuang <ratbert@faraday-tech.com>"); | 1208 | MODULE_AUTHOR("Po-Yu Chuang <ratbert@faraday-tech.com>"); |
| 1203 | MODULE_DESCRIPTION("FTMAC100 driver"); | 1209 | MODULE_DESCRIPTION("FTMAC100 driver"); |
| 1204 | MODULE_LICENSE("GPL"); | 1210 | MODULE_LICENSE("GPL"); |
| 1211 | MODULE_DEVICE_TABLE(of, ftmac100_of_ids); | ||
diff --git a/drivers/net/ethernet/mellanox/mlx4/main.c b/drivers/net/ethernet/mellanox/mlx4/main.c index 703205475524..83aab1e4c8c8 100644 --- a/drivers/net/ethernet/mellanox/mlx4/main.c +++ b/drivers/net/ethernet/mellanox/mlx4/main.c | |||
| @@ -2862,12 +2862,10 @@ static void mlx4_enable_msi_x(struct mlx4_dev *dev) | |||
| 2862 | int port = 0; | 2862 | int port = 0; |
| 2863 | 2863 | ||
| 2864 | if (msi_x) { | 2864 | if (msi_x) { |
| 2865 | int nreq = dev->caps.num_ports * num_online_cpus() + 1; | 2865 | int nreq = min3(dev->caps.num_ports * |
| 2866 | 2866 | (int)num_online_cpus() + 1, | |
| 2867 | nreq = min_t(int, dev->caps.num_eqs - dev->caps.reserved_eqs, | 2867 | dev->caps.num_eqs - dev->caps.reserved_eqs, |
| 2868 | nreq); | 2868 | MAX_MSIX); |
| 2869 | if (nreq > MAX_MSIX) | ||
| 2870 | nreq = MAX_MSIX; | ||
| 2871 | 2869 | ||
| 2872 | entries = kcalloc(nreq, sizeof *entries, GFP_KERNEL); | 2870 | entries = kcalloc(nreq, sizeof *entries, GFP_KERNEL); |
| 2873 | if (!entries) | 2871 | if (!entries) |
diff --git a/drivers/net/ethernet/mellanox/mlx5/core/Kconfig b/drivers/net/ethernet/mellanox/mlx5/core/Kconfig index fc52d742b7f7..27251a78075c 100644 --- a/drivers/net/ethernet/mellanox/mlx5/core/Kconfig +++ b/drivers/net/ethernet/mellanox/mlx5/core/Kconfig | |||
| @@ -13,7 +13,7 @@ config MLX5_CORE | |||
| 13 | 13 | ||
| 14 | config MLX5_CORE_EN | 14 | config MLX5_CORE_EN |
| 15 | bool "Mellanox Technologies ConnectX-4 Ethernet support" | 15 | bool "Mellanox Technologies ConnectX-4 Ethernet support" |
| 16 | depends on NETDEVICES && ETHERNET && PCI && MLX5_CORE | 16 | depends on NETDEVICES && ETHERNET && INET && PCI && MLX5_CORE |
| 17 | depends on IPV6=y || IPV6=n || MLX5_CORE=m | 17 | depends on IPV6=y || IPV6=n || MLX5_CORE=m |
| 18 | imply PTP_1588_CLOCK | 18 | imply PTP_1588_CLOCK |
| 19 | default n | 19 | default n |
diff --git a/drivers/net/ethernet/mellanox/mlx5/core/en.h b/drivers/net/ethernet/mellanox/mlx5/core/en.h index 0099a3e397bc..2fd044b23875 100644 --- a/drivers/net/ethernet/mellanox/mlx5/core/en.h +++ b/drivers/net/ethernet/mellanox/mlx5/core/en.h | |||
| @@ -1003,7 +1003,7 @@ int mlx5e_create_direct_tirs(struct mlx5e_priv *priv); | |||
| 1003 | void mlx5e_destroy_direct_tirs(struct mlx5e_priv *priv); | 1003 | void mlx5e_destroy_direct_tirs(struct mlx5e_priv *priv); |
| 1004 | void mlx5e_destroy_rqt(struct mlx5e_priv *priv, struct mlx5e_rqt *rqt); | 1004 | void mlx5e_destroy_rqt(struct mlx5e_priv *priv, struct mlx5e_rqt *rqt); |
| 1005 | 1005 | ||
| 1006 | int mlx5e_create_ttc_table(struct mlx5e_priv *priv, u32 underlay_qpn); | 1006 | int mlx5e_create_ttc_table(struct mlx5e_priv *priv); |
| 1007 | void mlx5e_destroy_ttc_table(struct mlx5e_priv *priv); | 1007 | void mlx5e_destroy_ttc_table(struct mlx5e_priv *priv); |
| 1008 | 1008 | ||
| 1009 | int mlx5e_create_tis(struct mlx5_core_dev *mdev, int tc, | 1009 | int mlx5e_create_tis(struct mlx5_core_dev *mdev, int tc, |
diff --git a/drivers/net/ethernet/mellanox/mlx5/core/en_ethtool.c b/drivers/net/ethernet/mellanox/mlx5/core/en_ethtool.c index ce7b09d72ff6..8209affa75c3 100644 --- a/drivers/net/ethernet/mellanox/mlx5/core/en_ethtool.c +++ b/drivers/net/ethernet/mellanox/mlx5/core/en_ethtool.c | |||
| @@ -794,7 +794,6 @@ static void get_supported(u32 eth_proto_cap, | |||
| 794 | ptys2ethtool_supported_port(link_ksettings, eth_proto_cap); | 794 | ptys2ethtool_supported_port(link_ksettings, eth_proto_cap); |
| 795 | ptys2ethtool_supported_link(supported, eth_proto_cap); | 795 | ptys2ethtool_supported_link(supported, eth_proto_cap); |
| 796 | ethtool_link_ksettings_add_link_mode(link_ksettings, supported, Pause); | 796 | ethtool_link_ksettings_add_link_mode(link_ksettings, supported, Pause); |
| 797 | ethtool_link_ksettings_add_link_mode(link_ksettings, supported, Asym_Pause); | ||
| 798 | } | 797 | } |
| 799 | 798 | ||
| 800 | static void get_advertising(u32 eth_proto_cap, u8 tx_pause, | 799 | static void get_advertising(u32 eth_proto_cap, u8 tx_pause, |
| @@ -804,7 +803,7 @@ static void get_advertising(u32 eth_proto_cap, u8 tx_pause, | |||
| 804 | unsigned long *advertising = link_ksettings->link_modes.advertising; | 803 | unsigned long *advertising = link_ksettings->link_modes.advertising; |
| 805 | 804 | ||
| 806 | ptys2ethtool_adver_link(advertising, eth_proto_cap); | 805 | ptys2ethtool_adver_link(advertising, eth_proto_cap); |
| 807 | if (tx_pause) | 806 | if (rx_pause) |
| 808 | ethtool_link_ksettings_add_link_mode(link_ksettings, advertising, Pause); | 807 | ethtool_link_ksettings_add_link_mode(link_ksettings, advertising, Pause); |
| 809 | if (tx_pause ^ rx_pause) | 808 | if (tx_pause ^ rx_pause) |
| 810 | ethtool_link_ksettings_add_link_mode(link_ksettings, advertising, Asym_Pause); | 809 | ethtool_link_ksettings_add_link_mode(link_ksettings, advertising, Asym_Pause); |
| @@ -849,6 +848,8 @@ static int mlx5e_get_link_ksettings(struct net_device *netdev, | |||
| 849 | struct mlx5e_priv *priv = netdev_priv(netdev); | 848 | struct mlx5e_priv *priv = netdev_priv(netdev); |
| 850 | struct mlx5_core_dev *mdev = priv->mdev; | 849 | struct mlx5_core_dev *mdev = priv->mdev; |
| 851 | u32 out[MLX5_ST_SZ_DW(ptys_reg)] = {0}; | 850 | u32 out[MLX5_ST_SZ_DW(ptys_reg)] = {0}; |
| 851 | u32 rx_pause = 0; | ||
| 852 | u32 tx_pause = 0; | ||
| 852 | u32 eth_proto_cap; | 853 | u32 eth_proto_cap; |
| 853 | u32 eth_proto_admin; | 854 | u32 eth_proto_admin; |
| 854 | u32 eth_proto_lp; | 855 | u32 eth_proto_lp; |
| @@ -871,11 +872,13 @@ static int mlx5e_get_link_ksettings(struct net_device *netdev, | |||
| 871 | an_disable_admin = MLX5_GET(ptys_reg, out, an_disable_admin); | 872 | an_disable_admin = MLX5_GET(ptys_reg, out, an_disable_admin); |
| 872 | an_status = MLX5_GET(ptys_reg, out, an_status); | 873 | an_status = MLX5_GET(ptys_reg, out, an_status); |
| 873 | 874 | ||
| 875 | mlx5_query_port_pause(mdev, &rx_pause, &tx_pause); | ||
| 876 | |||
| 874 | ethtool_link_ksettings_zero_link_mode(link_ksettings, supported); | 877 | ethtool_link_ksettings_zero_link_mode(link_ksettings, supported); |
| 875 | ethtool_link_ksettings_zero_link_mode(link_ksettings, advertising); | 878 | ethtool_link_ksettings_zero_link_mode(link_ksettings, advertising); |
| 876 | 879 | ||
| 877 | get_supported(eth_proto_cap, link_ksettings); | 880 | get_supported(eth_proto_cap, link_ksettings); |
| 878 | get_advertising(eth_proto_admin, 0, 0, link_ksettings); | 881 | get_advertising(eth_proto_admin, tx_pause, rx_pause, link_ksettings); |
| 879 | get_speed_duplex(netdev, eth_proto_oper, link_ksettings); | 882 | get_speed_duplex(netdev, eth_proto_oper, link_ksettings); |
| 880 | 883 | ||
| 881 | eth_proto_oper = eth_proto_oper ? eth_proto_oper : eth_proto_cap; | 884 | eth_proto_oper = eth_proto_oper ? eth_proto_oper : eth_proto_cap; |
diff --git a/drivers/net/ethernet/mellanox/mlx5/core/en_fs.c b/drivers/net/ethernet/mellanox/mlx5/core/en_fs.c index 576d6787b484..53ed58320a24 100644 --- a/drivers/net/ethernet/mellanox/mlx5/core/en_fs.c +++ b/drivers/net/ethernet/mellanox/mlx5/core/en_fs.c | |||
| @@ -800,7 +800,7 @@ void mlx5e_destroy_ttc_table(struct mlx5e_priv *priv) | |||
| 800 | mlx5e_destroy_flow_table(&ttc->ft); | 800 | mlx5e_destroy_flow_table(&ttc->ft); |
| 801 | } | 801 | } |
| 802 | 802 | ||
| 803 | int mlx5e_create_ttc_table(struct mlx5e_priv *priv, u32 underlay_qpn) | 803 | int mlx5e_create_ttc_table(struct mlx5e_priv *priv) |
| 804 | { | 804 | { |
| 805 | struct mlx5e_ttc_table *ttc = &priv->fs.ttc; | 805 | struct mlx5e_ttc_table *ttc = &priv->fs.ttc; |
| 806 | struct mlx5_flow_table_attr ft_attr = {}; | 806 | struct mlx5_flow_table_attr ft_attr = {}; |
| @@ -810,7 +810,6 @@ int mlx5e_create_ttc_table(struct mlx5e_priv *priv, u32 underlay_qpn) | |||
| 810 | ft_attr.max_fte = MLX5E_TTC_TABLE_SIZE; | 810 | ft_attr.max_fte = MLX5E_TTC_TABLE_SIZE; |
| 811 | ft_attr.level = MLX5E_TTC_FT_LEVEL; | 811 | ft_attr.level = MLX5E_TTC_FT_LEVEL; |
| 812 | ft_attr.prio = MLX5E_NIC_PRIO; | 812 | ft_attr.prio = MLX5E_NIC_PRIO; |
| 813 | ft_attr.underlay_qpn = underlay_qpn; | ||
| 814 | 813 | ||
| 815 | ft->t = mlx5_create_flow_table(priv->fs.ns, &ft_attr); | 814 | ft->t = mlx5_create_flow_table(priv->fs.ns, &ft_attr); |
| 816 | if (IS_ERR(ft->t)) { | 815 | if (IS_ERR(ft->t)) { |
| @@ -1147,7 +1146,7 @@ int mlx5e_create_flow_steering(struct mlx5e_priv *priv) | |||
| 1147 | priv->netdev->hw_features &= ~NETIF_F_NTUPLE; | 1146 | priv->netdev->hw_features &= ~NETIF_F_NTUPLE; |
| 1148 | } | 1147 | } |
| 1149 | 1148 | ||
| 1150 | err = mlx5e_create_ttc_table(priv, 0); | 1149 | err = mlx5e_create_ttc_table(priv); |
| 1151 | if (err) { | 1150 | if (err) { |
| 1152 | netdev_err(priv->netdev, "Failed to create ttc table, err=%d\n", | 1151 | netdev_err(priv->netdev, "Failed to create ttc table, err=%d\n", |
| 1153 | err); | 1152 | err); |
diff --git a/drivers/net/ethernet/mellanox/mlx5/core/en_main.c b/drivers/net/ethernet/mellanox/mlx5/core/en_main.c index a61b71b6fff3..41cd22a223dc 100644 --- a/drivers/net/ethernet/mellanox/mlx5/core/en_main.c +++ b/drivers/net/ethernet/mellanox/mlx5/core/en_main.c | |||
| @@ -2976,7 +2976,7 @@ static int mlx5e_setup_tc(struct net_device *netdev, u8 tc) | |||
| 2976 | new_channels.params = priv->channels.params; | 2976 | new_channels.params = priv->channels.params; |
| 2977 | new_channels.params.num_tc = tc ? tc : 1; | 2977 | new_channels.params.num_tc = tc ? tc : 1; |
| 2978 | 2978 | ||
| 2979 | if (test_bit(MLX5E_STATE_OPENED, &priv->state)) { | 2979 | if (!test_bit(MLX5E_STATE_OPENED, &priv->state)) { |
| 2980 | priv->channels.params = new_channels.params; | 2980 | priv->channels.params = new_channels.params; |
| 2981 | goto out; | 2981 | goto out; |
| 2982 | } | 2982 | } |
diff --git a/drivers/net/ethernet/mellanox/mlx5/core/fs_cmd.c b/drivers/net/ethernet/mellanox/mlx5/core/fs_cmd.c index 19e3d2fc2099..fcec7bedd3cd 100644 --- a/drivers/net/ethernet/mellanox/mlx5/core/fs_cmd.c +++ b/drivers/net/ethernet/mellanox/mlx5/core/fs_cmd.c | |||
| @@ -40,28 +40,25 @@ | |||
| 40 | #include "eswitch.h" | 40 | #include "eswitch.h" |
| 41 | 41 | ||
| 42 | int mlx5_cmd_update_root_ft(struct mlx5_core_dev *dev, | 42 | int mlx5_cmd_update_root_ft(struct mlx5_core_dev *dev, |
| 43 | struct mlx5_flow_table *ft) | 43 | struct mlx5_flow_table *ft, u32 underlay_qpn) |
| 44 | { | 44 | { |
| 45 | u32 in[MLX5_ST_SZ_DW(set_flow_table_root_in)] = {0}; | 45 | u32 in[MLX5_ST_SZ_DW(set_flow_table_root_in)] = {0}; |
| 46 | u32 out[MLX5_ST_SZ_DW(set_flow_table_root_out)] = {0}; | 46 | u32 out[MLX5_ST_SZ_DW(set_flow_table_root_out)] = {0}; |
| 47 | 47 | ||
| 48 | if ((MLX5_CAP_GEN(dev, port_type) == MLX5_CAP_PORT_TYPE_IB) && | 48 | if ((MLX5_CAP_GEN(dev, port_type) == MLX5_CAP_PORT_TYPE_IB) && |
| 49 | ft->underlay_qpn == 0) | 49 | underlay_qpn == 0) |
| 50 | return 0; | 50 | return 0; |
| 51 | 51 | ||
| 52 | MLX5_SET(set_flow_table_root_in, in, opcode, | 52 | MLX5_SET(set_flow_table_root_in, in, opcode, |
| 53 | MLX5_CMD_OP_SET_FLOW_TABLE_ROOT); | 53 | MLX5_CMD_OP_SET_FLOW_TABLE_ROOT); |
| 54 | MLX5_SET(set_flow_table_root_in, in, table_type, ft->type); | 54 | MLX5_SET(set_flow_table_root_in, in, table_type, ft->type); |
| 55 | MLX5_SET(set_flow_table_root_in, in, table_id, ft->id); | 55 | MLX5_SET(set_flow_table_root_in, in, table_id, ft->id); |
| 56 | MLX5_SET(set_flow_table_root_in, in, underlay_qpn, underlay_qpn); | ||
| 56 | if (ft->vport) { | 57 | if (ft->vport) { |
| 57 | MLX5_SET(set_flow_table_root_in, in, vport_number, ft->vport); | 58 | MLX5_SET(set_flow_table_root_in, in, vport_number, ft->vport); |
| 58 | MLX5_SET(set_flow_table_root_in, in, other_vport, 1); | 59 | MLX5_SET(set_flow_table_root_in, in, other_vport, 1); |
| 59 | } | 60 | } |
| 60 | 61 | ||
| 61 | if ((MLX5_CAP_GEN(dev, port_type) == MLX5_CAP_PORT_TYPE_IB) && | ||
| 62 | ft->underlay_qpn != 0) | ||
| 63 | MLX5_SET(set_flow_table_root_in, in, underlay_qpn, ft->underlay_qpn); | ||
| 64 | |||
| 65 | return mlx5_cmd_exec(dev, in, sizeof(in), out, sizeof(out)); | 62 | return mlx5_cmd_exec(dev, in, sizeof(in), out, sizeof(out)); |
| 66 | } | 63 | } |
| 67 | 64 | ||
diff --git a/drivers/net/ethernet/mellanox/mlx5/core/fs_cmd.h b/drivers/net/ethernet/mellanox/mlx5/core/fs_cmd.h index 8fad80688536..0f98a7cf4877 100644 --- a/drivers/net/ethernet/mellanox/mlx5/core/fs_cmd.h +++ b/drivers/net/ethernet/mellanox/mlx5/core/fs_cmd.h | |||
| @@ -71,7 +71,8 @@ int mlx5_cmd_delete_fte(struct mlx5_core_dev *dev, | |||
| 71 | unsigned int index); | 71 | unsigned int index); |
| 72 | 72 | ||
| 73 | int mlx5_cmd_update_root_ft(struct mlx5_core_dev *dev, | 73 | int mlx5_cmd_update_root_ft(struct mlx5_core_dev *dev, |
| 74 | struct mlx5_flow_table *ft); | 74 | struct mlx5_flow_table *ft, |
| 75 | u32 underlay_qpn); | ||
| 75 | 76 | ||
| 76 | int mlx5_cmd_fc_alloc(struct mlx5_core_dev *dev, u16 *id); | 77 | int mlx5_cmd_fc_alloc(struct mlx5_core_dev *dev, u16 *id); |
| 77 | int mlx5_cmd_fc_free(struct mlx5_core_dev *dev, u16 id); | 78 | int mlx5_cmd_fc_free(struct mlx5_core_dev *dev, u16 id); |
diff --git a/drivers/net/ethernet/mellanox/mlx5/core/fs_core.c b/drivers/net/ethernet/mellanox/mlx5/core/fs_core.c index b8a176503d38..0e487e8ca634 100644 --- a/drivers/net/ethernet/mellanox/mlx5/core/fs_core.c +++ b/drivers/net/ethernet/mellanox/mlx5/core/fs_core.c | |||
| @@ -650,7 +650,7 @@ static int update_root_ft_create(struct mlx5_flow_table *ft, struct fs_prio | |||
| 650 | if (ft->level >= min_level) | 650 | if (ft->level >= min_level) |
| 651 | return 0; | 651 | return 0; |
| 652 | 652 | ||
| 653 | err = mlx5_cmd_update_root_ft(root->dev, ft); | 653 | err = mlx5_cmd_update_root_ft(root->dev, ft, root->underlay_qpn); |
| 654 | if (err) | 654 | if (err) |
| 655 | mlx5_core_warn(root->dev, "Update root flow table of id=%u failed\n", | 655 | mlx5_core_warn(root->dev, "Update root flow table of id=%u failed\n", |
| 656 | ft->id); | 656 | ft->id); |
| @@ -818,8 +818,6 @@ static struct mlx5_flow_table *__mlx5_create_flow_table(struct mlx5_flow_namespa | |||
| 818 | goto unlock_root; | 818 | goto unlock_root; |
| 819 | } | 819 | } |
| 820 | 820 | ||
| 821 | ft->underlay_qpn = ft_attr->underlay_qpn; | ||
| 822 | |||
| 823 | tree_init_node(&ft->node, 1, del_flow_table); | 821 | tree_init_node(&ft->node, 1, del_flow_table); |
| 824 | log_table_sz = ft->max_fte ? ilog2(ft->max_fte) : 0; | 822 | log_table_sz = ft->max_fte ? ilog2(ft->max_fte) : 0; |
| 825 | next_ft = find_next_chained_ft(fs_prio); | 823 | next_ft = find_next_chained_ft(fs_prio); |
| @@ -1489,7 +1487,8 @@ static int update_root_ft_destroy(struct mlx5_flow_table *ft) | |||
| 1489 | 1487 | ||
| 1490 | new_root_ft = find_next_ft(ft); | 1488 | new_root_ft = find_next_ft(ft); |
| 1491 | if (new_root_ft) { | 1489 | if (new_root_ft) { |
| 1492 | int err = mlx5_cmd_update_root_ft(root->dev, new_root_ft); | 1490 | int err = mlx5_cmd_update_root_ft(root->dev, new_root_ft, |
| 1491 | root->underlay_qpn); | ||
| 1493 | 1492 | ||
| 1494 | if (err) { | 1493 | if (err) { |
| 1495 | mlx5_core_warn(root->dev, "Update root flow table of id=%u failed\n", | 1494 | mlx5_core_warn(root->dev, "Update root flow table of id=%u failed\n", |
| @@ -2062,3 +2061,21 @@ err: | |||
| 2062 | mlx5_cleanup_fs(dev); | 2061 | mlx5_cleanup_fs(dev); |
| 2063 | return err; | 2062 | return err; |
| 2064 | } | 2063 | } |
| 2064 | |||
| 2065 | int mlx5_fs_add_rx_underlay_qpn(struct mlx5_core_dev *dev, u32 underlay_qpn) | ||
| 2066 | { | ||
| 2067 | struct mlx5_flow_root_namespace *root = dev->priv.steering->root_ns; | ||
| 2068 | |||
| 2069 | root->underlay_qpn = underlay_qpn; | ||
| 2070 | return 0; | ||
| 2071 | } | ||
| 2072 | EXPORT_SYMBOL(mlx5_fs_add_rx_underlay_qpn); | ||
| 2073 | |||
| 2074 | int mlx5_fs_remove_rx_underlay_qpn(struct mlx5_core_dev *dev, u32 underlay_qpn) | ||
| 2075 | { | ||
| 2076 | struct mlx5_flow_root_namespace *root = dev->priv.steering->root_ns; | ||
| 2077 | |||
| 2078 | root->underlay_qpn = 0; | ||
| 2079 | return 0; | ||
| 2080 | } | ||
| 2081 | EXPORT_SYMBOL(mlx5_fs_remove_rx_underlay_qpn); | ||
diff --git a/drivers/net/ethernet/mellanox/mlx5/core/fs_core.h b/drivers/net/ethernet/mellanox/mlx5/core/fs_core.h index 81eafc7b9dd9..990acee6fb09 100644 --- a/drivers/net/ethernet/mellanox/mlx5/core/fs_core.h +++ b/drivers/net/ethernet/mellanox/mlx5/core/fs_core.h | |||
| @@ -118,7 +118,6 @@ struct mlx5_flow_table { | |||
| 118 | /* FWD rules that point on this flow table */ | 118 | /* FWD rules that point on this flow table */ |
| 119 | struct list_head fwd_rules; | 119 | struct list_head fwd_rules; |
| 120 | u32 flags; | 120 | u32 flags; |
| 121 | u32 underlay_qpn; | ||
| 122 | }; | 121 | }; |
| 123 | 122 | ||
| 124 | struct mlx5_fc_cache { | 123 | struct mlx5_fc_cache { |
| @@ -195,6 +194,7 @@ struct mlx5_flow_root_namespace { | |||
| 195 | struct mlx5_flow_table *root_ft; | 194 | struct mlx5_flow_table *root_ft; |
| 196 | /* Should be held when chaining flow tables */ | 195 | /* Should be held when chaining flow tables */ |
| 197 | struct mutex chain_lock; | 196 | struct mutex chain_lock; |
| 197 | u32 underlay_qpn; | ||
| 198 | }; | 198 | }; |
| 199 | 199 | ||
| 200 | int mlx5_init_fc_stats(struct mlx5_core_dev *dev); | 200 | int mlx5_init_fc_stats(struct mlx5_core_dev *dev); |
diff --git a/drivers/net/ethernet/mellanox/mlx5/core/ipoib.c b/drivers/net/ethernet/mellanox/mlx5/core/ipoib.c index 019c230da498..cc1858752e70 100644 --- a/drivers/net/ethernet/mellanox/mlx5/core/ipoib.c +++ b/drivers/net/ethernet/mellanox/mlx5/core/ipoib.c | |||
| @@ -66,6 +66,10 @@ static void mlx5i_init(struct mlx5_core_dev *mdev, | |||
| 66 | 66 | ||
| 67 | mlx5e_build_nic_params(mdev, &priv->channels.params, profile->max_nch(mdev)); | 67 | mlx5e_build_nic_params(mdev, &priv->channels.params, profile->max_nch(mdev)); |
| 68 | 68 | ||
| 69 | /* Override RQ params as IPoIB supports only LINKED LIST RQ for now */ | ||
| 70 | mlx5e_set_rq_type_params(mdev, &priv->channels.params, MLX5_WQ_TYPE_LINKED_LIST); | ||
| 71 | priv->channels.params.lro_en = false; | ||
| 72 | |||
| 69 | mutex_init(&priv->state_lock); | 73 | mutex_init(&priv->state_lock); |
| 70 | 74 | ||
| 71 | netdev->hw_features |= NETIF_F_SG; | 75 | netdev->hw_features |= NETIF_F_SG; |
| @@ -156,6 +160,8 @@ out: | |||
| 156 | 160 | ||
| 157 | static void mlx5i_destroy_underlay_qp(struct mlx5_core_dev *mdev, struct mlx5_core_qp *qp) | 161 | static void mlx5i_destroy_underlay_qp(struct mlx5_core_dev *mdev, struct mlx5_core_qp *qp) |
| 158 | { | 162 | { |
| 163 | mlx5_fs_remove_rx_underlay_qpn(mdev, qp->qpn); | ||
| 164 | |||
| 159 | mlx5_core_destroy_qp(mdev, qp); | 165 | mlx5_core_destroy_qp(mdev, qp); |
| 160 | } | 166 | } |
| 161 | 167 | ||
| @@ -170,6 +176,8 @@ static int mlx5i_init_tx(struct mlx5e_priv *priv) | |||
| 170 | return err; | 176 | return err; |
| 171 | } | 177 | } |
| 172 | 178 | ||
| 179 | mlx5_fs_add_rx_underlay_qpn(priv->mdev, ipriv->qp.qpn); | ||
| 180 | |||
| 173 | err = mlx5e_create_tis(priv->mdev, 0 /* tc */, ipriv->qp.qpn, &priv->tisn[0]); | 181 | err = mlx5e_create_tis(priv->mdev, 0 /* tc */, ipriv->qp.qpn, &priv->tisn[0]); |
| 174 | if (err) { | 182 | if (err) { |
| 175 | mlx5_core_warn(priv->mdev, "create tis failed, %d\n", err); | 183 | mlx5_core_warn(priv->mdev, "create tis failed, %d\n", err); |
| @@ -189,7 +197,6 @@ static void mlx5i_cleanup_tx(struct mlx5e_priv *priv) | |||
| 189 | 197 | ||
| 190 | static int mlx5i_create_flow_steering(struct mlx5e_priv *priv) | 198 | static int mlx5i_create_flow_steering(struct mlx5e_priv *priv) |
| 191 | { | 199 | { |
| 192 | struct mlx5i_priv *ipriv = priv->ppriv; | ||
| 193 | int err; | 200 | int err; |
| 194 | 201 | ||
| 195 | priv->fs.ns = mlx5_get_flow_namespace(priv->mdev, | 202 | priv->fs.ns = mlx5_get_flow_namespace(priv->mdev, |
| @@ -205,7 +212,7 @@ static int mlx5i_create_flow_steering(struct mlx5e_priv *priv) | |||
| 205 | priv->netdev->hw_features &= ~NETIF_F_NTUPLE; | 212 | priv->netdev->hw_features &= ~NETIF_F_NTUPLE; |
| 206 | } | 213 | } |
| 207 | 214 | ||
| 208 | err = mlx5e_create_ttc_table(priv, ipriv->qp.qpn); | 215 | err = mlx5e_create_ttc_table(priv); |
| 209 | if (err) { | 216 | if (err) { |
| 210 | netdev_err(priv->netdev, "Failed to create ttc table, err=%d\n", | 217 | netdev_err(priv->netdev, "Failed to create ttc table, err=%d\n", |
| 211 | err); | 218 | err); |
diff --git a/drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c b/drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c index ea56f6ade6b4..5f0a7bc692a4 100644 --- a/drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c +++ b/drivers/net/ethernet/mellanox/mlxsw/spectrum_dpipe.c | |||
| @@ -199,10 +199,11 @@ static int mlxsw_sp_erif_entry_get(struct mlxsw_sp *mlxsw_sp, | |||
| 199 | 199 | ||
| 200 | entry->counter_valid = false; | 200 | entry->counter_valid = false; |
| 201 | entry->counter = 0; | 201 | entry->counter = 0; |
| 202 | entry->index = mlxsw_sp_rif_index(rif); | ||
| 203 | |||
| 202 | if (!counters_enabled) | 204 | if (!counters_enabled) |
| 203 | return 0; | 205 | return 0; |
| 204 | 206 | ||
| 205 | entry->index = mlxsw_sp_rif_index(rif); | ||
| 206 | err = mlxsw_sp_rif_counter_value_get(mlxsw_sp, rif, | 207 | err = mlxsw_sp_rif_counter_value_get(mlxsw_sp, rif, |
| 207 | MLXSW_SP_RIF_COUNTER_EGRESS, | 208 | MLXSW_SP_RIF_COUNTER_EGRESS, |
| 208 | &cnt); | 209 | &cnt); |
diff --git a/drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c b/drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c index 33cec1cc1642..9f89c4137d21 100644 --- a/drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c +++ b/drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c | |||
| @@ -206,6 +206,9 @@ void mlxsw_sp_rif_counter_free(struct mlxsw_sp *mlxsw_sp, | |||
| 206 | { | 206 | { |
| 207 | unsigned int *p_counter_index; | 207 | unsigned int *p_counter_index; |
| 208 | 208 | ||
| 209 | if (!mlxsw_sp_rif_counter_valid_get(rif, dir)) | ||
| 210 | return; | ||
| 211 | |||
| 209 | p_counter_index = mlxsw_sp_rif_p_counter_get(rif, dir); | 212 | p_counter_index = mlxsw_sp_rif_p_counter_get(rif, dir); |
| 210 | if (WARN_ON(!p_counter_index)) | 213 | if (WARN_ON(!p_counter_index)) |
| 211 | return; | 214 | return; |
diff --git a/drivers/net/ethernet/mellanox/mlxsw/spectrum_switchdev.c b/drivers/net/ethernet/mellanox/mlxsw/spectrum_switchdev.c index 0d8411f1f954..f4bb0c0b7c1d 100644 --- a/drivers/net/ethernet/mellanox/mlxsw/spectrum_switchdev.c +++ b/drivers/net/ethernet/mellanox/mlxsw/spectrum_switchdev.c | |||
| @@ -1497,8 +1497,7 @@ do_fdb_op: | |||
| 1497 | err = mlxsw_sp_port_fdb_uc_op(mlxsw_sp, local_port, mac, fid, | 1497 | err = mlxsw_sp_port_fdb_uc_op(mlxsw_sp, local_port, mac, fid, |
| 1498 | adding, true); | 1498 | adding, true); |
| 1499 | if (err) { | 1499 | if (err) { |
| 1500 | if (net_ratelimit()) | 1500 | dev_err_ratelimited(mlxsw_sp->bus_info->dev, "Failed to set FDB entry\n"); |
| 1501 | netdev_err(mlxsw_sp_port->dev, "Failed to set FDB entry\n"); | ||
| 1502 | return; | 1501 | return; |
| 1503 | } | 1502 | } |
| 1504 | 1503 | ||
| @@ -1558,8 +1557,7 @@ do_fdb_op: | |||
| 1558 | err = mlxsw_sp_port_fdb_uc_lag_op(mlxsw_sp, lag_id, mac, fid, lag_vid, | 1557 | err = mlxsw_sp_port_fdb_uc_lag_op(mlxsw_sp, lag_id, mac, fid, lag_vid, |
| 1559 | adding, true); | 1558 | adding, true); |
| 1560 | if (err) { | 1559 | if (err) { |
| 1561 | if (net_ratelimit()) | 1560 | dev_err_ratelimited(mlxsw_sp->bus_info->dev, "Failed to set FDB entry\n"); |
| 1562 | netdev_err(mlxsw_sp_port->dev, "Failed to set FDB entry\n"); | ||
| 1563 | return; | 1561 | return; |
| 1564 | } | 1562 | } |
| 1565 | 1563 | ||
diff --git a/drivers/net/ethernet/qlogic/netxen/netxen_nic_ctx.c b/drivers/net/ethernet/qlogic/netxen/netxen_nic_ctx.c index b8d5270359cd..e30676515529 100644 --- a/drivers/net/ethernet/qlogic/netxen/netxen_nic_ctx.c +++ b/drivers/net/ethernet/qlogic/netxen/netxen_nic_ctx.c | |||
| @@ -247,7 +247,7 @@ nx_fw_cmd_set_mtu(struct netxen_adapter *adapter, int mtu) | |||
| 247 | cmd.req.arg3 = 0; | 247 | cmd.req.arg3 = 0; |
| 248 | 248 | ||
| 249 | if (recv_ctx->state == NX_HOST_CTX_STATE_ACTIVE) | 249 | if (recv_ctx->state == NX_HOST_CTX_STATE_ACTIVE) |
| 250 | netxen_issue_cmd(adapter, &cmd); | 250 | rcode = netxen_issue_cmd(adapter, &cmd); |
| 251 | 251 | ||
| 252 | if (rcode != NX_RCODE_SUCCESS) | 252 | if (rcode != NX_RCODE_SUCCESS) |
| 253 | return -EIO; | 253 | return -EIO; |
diff --git a/drivers/net/ethernet/qlogic/qed/qed_init_fw_funcs.c b/drivers/net/ethernet/qlogic/qed/qed_init_fw_funcs.c index 67200c5498ab..0a8fde629991 100644 --- a/drivers/net/ethernet/qlogic/qed/qed_init_fw_funcs.c +++ b/drivers/net/ethernet/qlogic/qed/qed_init_fw_funcs.c | |||
| @@ -983,7 +983,7 @@ void qed_set_rfs_mode_disable(struct qed_hwfn *p_hwfn, | |||
| 983 | memset(&camline, 0, sizeof(union gft_cam_line_union)); | 983 | memset(&camline, 0, sizeof(union gft_cam_line_union)); |
| 984 | qed_wr(p_hwfn, p_ptt, PRS_REG_GFT_CAM + CAM_LINE_SIZE * pf_id, | 984 | qed_wr(p_hwfn, p_ptt, PRS_REG_GFT_CAM + CAM_LINE_SIZE * pf_id, |
| 985 | camline.cam_line_mapped.camline); | 985 | camline.cam_line_mapped.camline); |
| 986 | memset(&ramline, 0, sizeof(union gft_cam_line_union)); | 986 | memset(&ramline, 0, sizeof(ramline)); |
| 987 | 987 | ||
| 988 | for (i = 0; i < RAM_LINE_SIZE / REG_SIZE; i++) { | 988 | for (i = 0; i < RAM_LINE_SIZE / REG_SIZE; i++) { |
| 989 | u32 hw_addr = PRS_REG_GFT_PROFILE_MASK_RAM; | 989 | u32 hw_addr = PRS_REG_GFT_PROFILE_MASK_RAM; |
diff --git a/drivers/net/ethernet/qlogic/qlcnic/qlcnic.h b/drivers/net/ethernet/qlogic/qlcnic/qlcnic.h index 49bad00a0f8f..7245b1072518 100644 --- a/drivers/net/ethernet/qlogic/qlcnic/qlcnic.h +++ b/drivers/net/ethernet/qlogic/qlcnic/qlcnic.h | |||
| @@ -37,8 +37,8 @@ | |||
| 37 | 37 | ||
| 38 | #define _QLCNIC_LINUX_MAJOR 5 | 38 | #define _QLCNIC_LINUX_MAJOR 5 |
| 39 | #define _QLCNIC_LINUX_MINOR 3 | 39 | #define _QLCNIC_LINUX_MINOR 3 |
| 40 | #define _QLCNIC_LINUX_SUBVERSION 65 | 40 | #define _QLCNIC_LINUX_SUBVERSION 66 |
| 41 | #define QLCNIC_LINUX_VERSIONID "5.3.65" | 41 | #define QLCNIC_LINUX_VERSIONID "5.3.66" |
| 42 | #define QLCNIC_DRV_IDC_VER 0x01 | 42 | #define QLCNIC_DRV_IDC_VER 0x01 |
| 43 | #define QLCNIC_DRIVER_VERSION ((_QLCNIC_LINUX_MAJOR << 16) |\ | 43 | #define QLCNIC_DRIVER_VERSION ((_QLCNIC_LINUX_MAJOR << 16) |\ |
| 44 | (_QLCNIC_LINUX_MINOR << 8) | (_QLCNIC_LINUX_SUBVERSION)) | 44 | (_QLCNIC_LINUX_MINOR << 8) | (_QLCNIC_LINUX_SUBVERSION)) |
diff --git a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c index 718bf58a7da6..4fb68797630e 100644 --- a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c +++ b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c | |||
| @@ -3168,6 +3168,40 @@ int qlcnic_83xx_flash_read32(struct qlcnic_adapter *adapter, u32 flash_addr, | |||
| 3168 | return 0; | 3168 | return 0; |
| 3169 | } | 3169 | } |
| 3170 | 3170 | ||
| 3171 | void qlcnic_83xx_get_port_type(struct qlcnic_adapter *adapter) | ||
| 3172 | { | ||
| 3173 | struct qlcnic_hardware_context *ahw = adapter->ahw; | ||
| 3174 | struct qlcnic_cmd_args cmd; | ||
| 3175 | u32 config; | ||
| 3176 | int err; | ||
| 3177 | |||
| 3178 | err = qlcnic_alloc_mbx_args(&cmd, adapter, QLCNIC_CMD_GET_LINK_STATUS); | ||
| 3179 | if (err) | ||
| 3180 | return; | ||
| 3181 | |||
| 3182 | err = qlcnic_issue_cmd(adapter, &cmd); | ||
| 3183 | if (err) { | ||
| 3184 | dev_info(&adapter->pdev->dev, | ||
| 3185 | "Get Link Status Command failed: 0x%x\n", err); | ||
| 3186 | goto out; | ||
| 3187 | } else { | ||
| 3188 | config = cmd.rsp.arg[3]; | ||
| 3189 | |||
| 3190 | switch (QLC_83XX_SFP_MODULE_TYPE(config)) { | ||
| 3191 | case QLC_83XX_MODULE_FIBRE_1000BASE_SX: | ||
| 3192 | case QLC_83XX_MODULE_FIBRE_1000BASE_LX: | ||
| 3193 | case QLC_83XX_MODULE_FIBRE_1000BASE_CX: | ||
| 3194 | case QLC_83XX_MODULE_TP_1000BASE_T: | ||
| 3195 | ahw->port_type = QLCNIC_GBE; | ||
| 3196 | break; | ||
| 3197 | default: | ||
| 3198 | ahw->port_type = QLCNIC_XGBE; | ||
| 3199 | } | ||
| 3200 | } | ||
| 3201 | out: | ||
| 3202 | qlcnic_free_mbx_args(&cmd); | ||
| 3203 | } | ||
| 3204 | |||
| 3171 | int qlcnic_83xx_test_link(struct qlcnic_adapter *adapter) | 3205 | int qlcnic_83xx_test_link(struct qlcnic_adapter *adapter) |
| 3172 | { | 3206 | { |
| 3173 | u8 pci_func; | 3207 | u8 pci_func; |
diff --git a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.h b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.h index 3dfe8e27b51c..b75a81246856 100644 --- a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.h +++ b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.h | |||
| @@ -637,6 +637,7 @@ void qlcnic_83xx_get_pauseparam(struct qlcnic_adapter *, | |||
| 637 | int qlcnic_83xx_set_pauseparam(struct qlcnic_adapter *, | 637 | int qlcnic_83xx_set_pauseparam(struct qlcnic_adapter *, |
| 638 | struct ethtool_pauseparam *); | 638 | struct ethtool_pauseparam *); |
| 639 | int qlcnic_83xx_test_link(struct qlcnic_adapter *); | 639 | int qlcnic_83xx_test_link(struct qlcnic_adapter *); |
| 640 | void qlcnic_83xx_get_port_type(struct qlcnic_adapter *adapter); | ||
| 640 | int qlcnic_83xx_reg_test(struct qlcnic_adapter *); | 641 | int qlcnic_83xx_reg_test(struct qlcnic_adapter *); |
| 641 | int qlcnic_83xx_get_regs_len(struct qlcnic_adapter *); | 642 | int qlcnic_83xx_get_regs_len(struct qlcnic_adapter *); |
| 642 | int qlcnic_83xx_get_registers(struct qlcnic_adapter *, u32 *); | 643 | int qlcnic_83xx_get_registers(struct qlcnic_adapter *, u32 *); |
diff --git a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_ethtool.c b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_ethtool.c index 9a869c15d8bf..7f7deeaf1cf0 100644 --- a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_ethtool.c +++ b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_ethtool.c | |||
| @@ -486,6 +486,9 @@ static int qlcnic_set_link_ksettings(struct net_device *dev, | |||
| 486 | u32 ret = 0; | 486 | u32 ret = 0; |
| 487 | struct qlcnic_adapter *adapter = netdev_priv(dev); | 487 | struct qlcnic_adapter *adapter = netdev_priv(dev); |
| 488 | 488 | ||
| 489 | if (qlcnic_83xx_check(adapter)) | ||
| 490 | qlcnic_83xx_get_port_type(adapter); | ||
| 491 | |||
| 489 | if (adapter->ahw->port_type != QLCNIC_GBE) | 492 | if (adapter->ahw->port_type != QLCNIC_GBE) |
| 490 | return -EOPNOTSUPP; | 493 | return -EOPNOTSUPP; |
| 491 | 494 | ||
diff --git a/drivers/net/ethernet/qualcomm/qca_spi.c b/drivers/net/ethernet/qualcomm/qca_spi.c index 513e6c74e199..24ca7df15d07 100644 --- a/drivers/net/ethernet/qualcomm/qca_spi.c +++ b/drivers/net/ethernet/qualcomm/qca_spi.c | |||
| @@ -296,8 +296,9 @@ qcaspi_receive(struct qcaspi *qca) | |||
| 296 | 296 | ||
| 297 | /* Allocate rx SKB if we don't have one available. */ | 297 | /* Allocate rx SKB if we don't have one available. */ |
| 298 | if (!qca->rx_skb) { | 298 | if (!qca->rx_skb) { |
| 299 | qca->rx_skb = netdev_alloc_skb(net_dev, | 299 | qca->rx_skb = netdev_alloc_skb_ip_align(net_dev, |
| 300 | net_dev->mtu + VLAN_ETH_HLEN); | 300 | net_dev->mtu + |
| 301 | VLAN_ETH_HLEN); | ||
| 301 | if (!qca->rx_skb) { | 302 | if (!qca->rx_skb) { |
| 302 | netdev_dbg(net_dev, "out of RX resources\n"); | 303 | netdev_dbg(net_dev, "out of RX resources\n"); |
| 303 | qca->stats.out_of_mem++; | 304 | qca->stats.out_of_mem++; |
| @@ -377,7 +378,7 @@ qcaspi_receive(struct qcaspi *qca) | |||
| 377 | qca->rx_skb, qca->rx_skb->dev); | 378 | qca->rx_skb, qca->rx_skb->dev); |
| 378 | qca->rx_skb->ip_summed = CHECKSUM_UNNECESSARY; | 379 | qca->rx_skb->ip_summed = CHECKSUM_UNNECESSARY; |
| 379 | netif_rx_ni(qca->rx_skb); | 380 | netif_rx_ni(qca->rx_skb); |
| 380 | qca->rx_skb = netdev_alloc_skb(net_dev, | 381 | qca->rx_skb = netdev_alloc_skb_ip_align(net_dev, |
| 381 | net_dev->mtu + VLAN_ETH_HLEN); | 382 | net_dev->mtu + VLAN_ETH_HLEN); |
| 382 | if (!qca->rx_skb) { | 383 | if (!qca->rx_skb) { |
| 383 | netdev_dbg(net_dev, "out of RX resources\n"); | 384 | netdev_dbg(net_dev, "out of RX resources\n"); |
| @@ -759,7 +760,8 @@ qcaspi_netdev_init(struct net_device *dev) | |||
| 759 | if (!qca->rx_buffer) | 760 | if (!qca->rx_buffer) |
| 760 | return -ENOBUFS; | 761 | return -ENOBUFS; |
| 761 | 762 | ||
| 762 | qca->rx_skb = netdev_alloc_skb(dev, qca->net_dev->mtu + VLAN_ETH_HLEN); | 763 | qca->rx_skb = netdev_alloc_skb_ip_align(dev, qca->net_dev->mtu + |
| 764 | VLAN_ETH_HLEN); | ||
| 763 | if (!qca->rx_skb) { | 765 | if (!qca->rx_skb) { |
| 764 | kfree(qca->rx_buffer); | 766 | kfree(qca->rx_buffer); |
| 765 | netdev_info(qca->net_dev, "Failed to allocate RX sk_buff.\n"); | 767 | netdev_info(qca->net_dev, "Failed to allocate RX sk_buff.\n"); |
diff --git a/drivers/net/ethernet/renesas/sh_eth.c b/drivers/net/ethernet/renesas/sh_eth.c index f68c4db656ed..2d686ccf971b 100644 --- a/drivers/net/ethernet/renesas/sh_eth.c +++ b/drivers/net/ethernet/renesas/sh_eth.c | |||
| @@ -3220,7 +3220,8 @@ static int sh_eth_drv_probe(struct platform_device *pdev) | |||
| 3220 | /* MDIO bus init */ | 3220 | /* MDIO bus init */ |
| 3221 | ret = sh_mdio_init(mdp, pd); | 3221 | ret = sh_mdio_init(mdp, pd); |
| 3222 | if (ret) { | 3222 | if (ret) { |
| 3223 | dev_err(&ndev->dev, "failed to initialise MDIO\n"); | 3223 | if (ret != -EPROBE_DEFER) |
| 3224 | dev_err(&pdev->dev, "MDIO init failed: %d\n", ret); | ||
| 3224 | goto out_release; | 3225 | goto out_release; |
| 3225 | } | 3226 | } |
| 3226 | 3227 | ||
diff --git a/drivers/net/ethernet/sfc/nic.h b/drivers/net/ethernet/sfc/nic.h index 7b916aa21bde..4d7fb8af880d 100644 --- a/drivers/net/ethernet/sfc/nic.h +++ b/drivers/net/ethernet/sfc/nic.h | |||
| @@ -18,8 +18,12 @@ | |||
| 18 | #include "mcdi.h" | 18 | #include "mcdi.h" |
| 19 | 19 | ||
| 20 | enum { | 20 | enum { |
| 21 | EFX_REV_SIENA_A0 = 0, | 21 | /* Revisions 0-2 were Falcon A0, A1 and B0 respectively. |
| 22 | EFX_REV_HUNT_A0 = 1, | 22 | * They are not supported by this driver but these revision numbers |
| 23 | * form part of the ethtool API for register dumping. | ||
| 24 | */ | ||
| 25 | EFX_REV_SIENA_A0 = 3, | ||
| 26 | EFX_REV_HUNT_A0 = 4, | ||
| 23 | }; | 27 | }; |
| 24 | 28 | ||
| 25 | static inline int efx_nic_rev(struct efx_nic *efx) | 29 | static inline int efx_nic_rev(struct efx_nic *efx) |
diff --git a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c b/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c index cd8c60132390..a74c481401c4 100644 --- a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c +++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c | |||
| @@ -3725,7 +3725,7 @@ static void sysfs_display_ring(void *head, int size, int extend_desc, | |||
| 3725 | ep++; | 3725 | ep++; |
| 3726 | } else { | 3726 | } else { |
| 3727 | seq_printf(seq, "%d [0x%x]: 0x%x 0x%x 0x%x 0x%x\n", | 3727 | seq_printf(seq, "%d [0x%x]: 0x%x 0x%x 0x%x 0x%x\n", |
| 3728 | i, (unsigned int)virt_to_phys(ep), | 3728 | i, (unsigned int)virt_to_phys(p), |
| 3729 | le32_to_cpu(p->des0), le32_to_cpu(p->des1), | 3729 | le32_to_cpu(p->des0), le32_to_cpu(p->des1), |
| 3730 | le32_to_cpu(p->des2), le32_to_cpu(p->des3)); | 3730 | le32_to_cpu(p->des2), le32_to_cpu(p->des3)); |
| 3731 | p++; | 3731 | p++; |
diff --git a/drivers/net/ethernet/sun/ldmvsw.c b/drivers/net/ethernet/sun/ldmvsw.c index 5a90fed06260..5b56c24b6ed2 100644 --- a/drivers/net/ethernet/sun/ldmvsw.c +++ b/drivers/net/ethernet/sun/ldmvsw.c | |||
| @@ -411,13 +411,14 @@ static int vsw_port_remove(struct vio_dev *vdev) | |||
| 411 | 411 | ||
| 412 | if (port) { | 412 | if (port) { |
| 413 | del_timer_sync(&port->vio.timer); | 413 | del_timer_sync(&port->vio.timer); |
| 414 | del_timer_sync(&port->clean_timer); | ||
| 414 | 415 | ||
| 415 | napi_disable(&port->napi); | 416 | napi_disable(&port->napi); |
| 417 | unregister_netdev(port->dev); | ||
| 416 | 418 | ||
| 417 | list_del_rcu(&port->list); | 419 | list_del_rcu(&port->list); |
| 418 | 420 | ||
| 419 | synchronize_rcu(); | 421 | synchronize_rcu(); |
| 420 | del_timer_sync(&port->clean_timer); | ||
| 421 | spin_lock_irqsave(&port->vp->lock, flags); | 422 | spin_lock_irqsave(&port->vp->lock, flags); |
| 422 | sunvnet_port_rm_txq_common(port); | 423 | sunvnet_port_rm_txq_common(port); |
| 423 | spin_unlock_irqrestore(&port->vp->lock, flags); | 424 | spin_unlock_irqrestore(&port->vp->lock, flags); |
| @@ -427,7 +428,6 @@ static int vsw_port_remove(struct vio_dev *vdev) | |||
| 427 | 428 | ||
| 428 | dev_set_drvdata(&vdev->dev, NULL); | 429 | dev_set_drvdata(&vdev->dev, NULL); |
| 429 | 430 | ||
| 430 | unregister_netdev(port->dev); | ||
| 431 | free_netdev(port->dev); | 431 | free_netdev(port->dev); |
| 432 | } | 432 | } |
| 433 | 433 | ||
diff --git a/drivers/net/ethernet/ti/netcp_core.c b/drivers/net/ethernet/ti/netcp_core.c index 729a7da90b5b..e6222e535019 100644 --- a/drivers/net/ethernet/ti/netcp_core.c +++ b/drivers/net/ethernet/ti/netcp_core.c | |||
| @@ -1353,9 +1353,10 @@ int netcp_txpipe_open(struct netcp_tx_pipe *tx_pipe) | |||
| 1353 | 1353 | ||
| 1354 | tx_pipe->dma_channel = knav_dma_open_channel(dev, | 1354 | tx_pipe->dma_channel = knav_dma_open_channel(dev, |
| 1355 | tx_pipe->dma_chan_name, &config); | 1355 | tx_pipe->dma_chan_name, &config); |
| 1356 | if (IS_ERR_OR_NULL(tx_pipe->dma_channel)) { | 1356 | if (IS_ERR(tx_pipe->dma_channel)) { |
| 1357 | dev_err(dev, "failed opening tx chan(%s)\n", | 1357 | dev_err(dev, "failed opening tx chan(%s)\n", |
| 1358 | tx_pipe->dma_chan_name); | 1358 | tx_pipe->dma_chan_name); |
| 1359 | ret = PTR_ERR(tx_pipe->dma_channel); | ||
| 1359 | goto err; | 1360 | goto err; |
| 1360 | } | 1361 | } |
| 1361 | 1362 | ||
| @@ -1673,9 +1674,10 @@ static int netcp_setup_navigator_resources(struct net_device *ndev) | |||
| 1673 | 1674 | ||
| 1674 | netcp->rx_channel = knav_dma_open_channel(netcp->netcp_device->device, | 1675 | netcp->rx_channel = knav_dma_open_channel(netcp->netcp_device->device, |
| 1675 | netcp->dma_chan_name, &config); | 1676 | netcp->dma_chan_name, &config); |
| 1676 | if (IS_ERR_OR_NULL(netcp->rx_channel)) { | 1677 | if (IS_ERR(netcp->rx_channel)) { |
| 1677 | dev_err(netcp->ndev_dev, "failed opening rx chan(%s\n", | 1678 | dev_err(netcp->ndev_dev, "failed opening rx chan(%s\n", |
| 1678 | netcp->dma_chan_name); | 1679 | netcp->dma_chan_name); |
| 1680 | ret = PTR_ERR(netcp->rx_channel); | ||
| 1679 | goto fail; | 1681 | goto fail; |
| 1680 | } | 1682 | } |
| 1681 | 1683 | ||
diff --git a/drivers/net/ethernet/ti/netcp_ethss.c b/drivers/net/ethernet/ti/netcp_ethss.c index 897176fc5043..dd92950a4615 100644 --- a/drivers/net/ethernet/ti/netcp_ethss.c +++ b/drivers/net/ethernet/ti/netcp_ethss.c | |||
| @@ -2651,7 +2651,6 @@ static int gbe_hwtstamp_set(struct gbe_intf *gbe_intf, struct ifreq *ifr) | |||
| 2651 | case HWTSTAMP_FILTER_NONE: | 2651 | case HWTSTAMP_FILTER_NONE: |
| 2652 | cpts_rx_enable(cpts, 0); | 2652 | cpts_rx_enable(cpts, 0); |
| 2653 | break; | 2653 | break; |
| 2654 | case HWTSTAMP_FILTER_ALL: | ||
| 2655 | case HWTSTAMP_FILTER_PTP_V1_L4_EVENT: | 2654 | case HWTSTAMP_FILTER_PTP_V1_L4_EVENT: |
| 2656 | case HWTSTAMP_FILTER_PTP_V1_L4_SYNC: | 2655 | case HWTSTAMP_FILTER_PTP_V1_L4_SYNC: |
| 2657 | case HWTSTAMP_FILTER_PTP_V1_L4_DELAY_REQ: | 2656 | case HWTSTAMP_FILTER_PTP_V1_L4_DELAY_REQ: |
diff --git a/drivers/net/irda/irda-usb.c b/drivers/net/irda/irda-usb.c index 8716b8c07feb..6f3c805f7211 100644 --- a/drivers/net/irda/irda-usb.c +++ b/drivers/net/irda/irda-usb.c | |||
| @@ -1077,7 +1077,7 @@ static int stir421x_patch_device(struct irda_usb_cb *self) | |||
| 1077 | * are "42101001.sb" or "42101002.sb" | 1077 | * are "42101001.sb" or "42101002.sb" |
| 1078 | */ | 1078 | */ |
| 1079 | sprintf(stir421x_fw_name, "4210%4X.sb", | 1079 | sprintf(stir421x_fw_name, "4210%4X.sb", |
| 1080 | self->usbdev->descriptor.bcdDevice); | 1080 | le16_to_cpu(self->usbdev->descriptor.bcdDevice)); |
| 1081 | ret = request_firmware(&fw, stir421x_fw_name, &self->usbdev->dev); | 1081 | ret = request_firmware(&fw, stir421x_fw_name, &self->usbdev->dev); |
| 1082 | if (ret < 0) | 1082 | if (ret < 0) |
| 1083 | return ret; | 1083 | return ret; |
diff --git a/drivers/net/macvlan.c b/drivers/net/macvlan.c index b34eaaae03fd..346ad2ff3998 100644 --- a/drivers/net/macvlan.c +++ b/drivers/net/macvlan.c | |||
| @@ -789,10 +789,12 @@ static int macvlan_change_mtu(struct net_device *dev, int new_mtu) | |||
| 789 | */ | 789 | */ |
| 790 | static struct lock_class_key macvlan_netdev_addr_lock_key; | 790 | static struct lock_class_key macvlan_netdev_addr_lock_key; |
| 791 | 791 | ||
| 792 | #define ALWAYS_ON_FEATURES \ | 792 | #define ALWAYS_ON_OFFLOADS \ |
| 793 | (NETIF_F_SG | NETIF_F_HW_CSUM | NETIF_F_GSO_SOFTWARE | NETIF_F_LLTX | \ | 793 | (NETIF_F_SG | NETIF_F_HW_CSUM | NETIF_F_GSO_SOFTWARE | \ |
| 794 | NETIF_F_GSO_ROBUST) | 794 | NETIF_F_GSO_ROBUST) |
| 795 | 795 | ||
| 796 | #define ALWAYS_ON_FEATURES (ALWAYS_ON_OFFLOADS | NETIF_F_LLTX) | ||
| 797 | |||
| 796 | #define MACVLAN_FEATURES \ | 798 | #define MACVLAN_FEATURES \ |
| 797 | (NETIF_F_SG | NETIF_F_HW_CSUM | NETIF_F_HIGHDMA | NETIF_F_FRAGLIST | \ | 799 | (NETIF_F_SG | NETIF_F_HW_CSUM | NETIF_F_HIGHDMA | NETIF_F_FRAGLIST | \ |
| 798 | NETIF_F_GSO | NETIF_F_TSO | NETIF_F_UFO | NETIF_F_LRO | \ | 800 | NETIF_F_GSO | NETIF_F_TSO | NETIF_F_UFO | NETIF_F_LRO | \ |
| @@ -827,6 +829,7 @@ static int macvlan_init(struct net_device *dev) | |||
| 827 | dev->features |= ALWAYS_ON_FEATURES; | 829 | dev->features |= ALWAYS_ON_FEATURES; |
| 828 | dev->hw_features |= NETIF_F_LRO; | 830 | dev->hw_features |= NETIF_F_LRO; |
| 829 | dev->vlan_features = lowerdev->vlan_features & MACVLAN_FEATURES; | 831 | dev->vlan_features = lowerdev->vlan_features & MACVLAN_FEATURES; |
| 832 | dev->vlan_features |= ALWAYS_ON_OFFLOADS; | ||
| 830 | dev->gso_max_size = lowerdev->gso_max_size; | 833 | dev->gso_max_size = lowerdev->gso_max_size; |
| 831 | dev->gso_max_segs = lowerdev->gso_max_segs; | 834 | dev->gso_max_segs = lowerdev->gso_max_segs; |
| 832 | dev->hard_header_len = lowerdev->hard_header_len; | 835 | dev->hard_header_len = lowerdev->hard_header_len; |
diff --git a/drivers/net/phy/mdio-mux.c b/drivers/net/phy/mdio-mux.c index 963838d4fac1..599ce24c514f 100644 --- a/drivers/net/phy/mdio-mux.c +++ b/drivers/net/phy/mdio-mux.c | |||
| @@ -122,10 +122,9 @@ int mdio_mux_init(struct device *dev, | |||
| 122 | pb = devm_kzalloc(dev, sizeof(*pb), GFP_KERNEL); | 122 | pb = devm_kzalloc(dev, sizeof(*pb), GFP_KERNEL); |
| 123 | if (pb == NULL) { | 123 | if (pb == NULL) { |
| 124 | ret_val = -ENOMEM; | 124 | ret_val = -ENOMEM; |
| 125 | goto err_parent_bus; | 125 | goto err_pb_kz; |
| 126 | } | 126 | } |
| 127 | 127 | ||
| 128 | |||
| 129 | pb->switch_data = data; | 128 | pb->switch_data = data; |
| 130 | pb->switch_fn = switch_fn; | 129 | pb->switch_fn = switch_fn; |
| 131 | pb->current_child = -1; | 130 | pb->current_child = -1; |
| @@ -154,6 +153,7 @@ int mdio_mux_init(struct device *dev, | |||
| 154 | cb->mii_bus = mdiobus_alloc(); | 153 | cb->mii_bus = mdiobus_alloc(); |
| 155 | if (!cb->mii_bus) { | 154 | if (!cb->mii_bus) { |
| 156 | ret_val = -ENOMEM; | 155 | ret_val = -ENOMEM; |
| 156 | devm_kfree(dev, cb); | ||
| 157 | of_node_put(child_bus_node); | 157 | of_node_put(child_bus_node); |
| 158 | break; | 158 | break; |
| 159 | } | 159 | } |
| @@ -170,7 +170,6 @@ int mdio_mux_init(struct device *dev, | |||
| 170 | mdiobus_free(cb->mii_bus); | 170 | mdiobus_free(cb->mii_bus); |
| 171 | devm_kfree(dev, cb); | 171 | devm_kfree(dev, cb); |
| 172 | } else { | 172 | } else { |
| 173 | of_node_get(child_bus_node); | ||
| 174 | cb->next = pb->children; | 173 | cb->next = pb->children; |
| 175 | pb->children = cb; | 174 | pb->children = cb; |
| 176 | } | 175 | } |
| @@ -181,9 +180,11 @@ int mdio_mux_init(struct device *dev, | |||
| 181 | return 0; | 180 | return 0; |
| 182 | } | 181 | } |
| 183 | 182 | ||
| 183 | devm_kfree(dev, pb); | ||
| 184 | err_pb_kz: | ||
| 184 | /* balance the reference of_mdio_find_bus() took */ | 185 | /* balance the reference of_mdio_find_bus() took */ |
| 185 | put_device(&pb->mii_bus->dev); | 186 | if (!mux_bus) |
| 186 | 187 | put_device(&parent_bus->dev); | |
| 187 | err_parent_bus: | 188 | err_parent_bus: |
| 188 | of_node_put(parent_bus_node); | 189 | of_node_put(parent_bus_node); |
| 189 | return ret_val; | 190 | return ret_val; |
diff --git a/drivers/net/phy/mdio_bus.c b/drivers/net/phy/mdio_bus.c index a898e5c4ef1b..8e73f5f36e71 100644 --- a/drivers/net/phy/mdio_bus.c +++ b/drivers/net/phy/mdio_bus.c | |||
| @@ -364,9 +364,6 @@ int __mdiobus_register(struct mii_bus *bus, struct module *owner) | |||
| 364 | 364 | ||
| 365 | mutex_init(&bus->mdio_lock); | 365 | mutex_init(&bus->mdio_lock); |
| 366 | 366 | ||
| 367 | if (bus->reset) | ||
| 368 | bus->reset(bus); | ||
| 369 | |||
| 370 | /* de-assert bus level PHY GPIO resets */ | 367 | /* de-assert bus level PHY GPIO resets */ |
| 371 | if (bus->num_reset_gpios > 0) { | 368 | if (bus->num_reset_gpios > 0) { |
| 372 | bus->reset_gpiod = devm_kcalloc(&bus->dev, | 369 | bus->reset_gpiod = devm_kcalloc(&bus->dev, |
| @@ -396,6 +393,9 @@ int __mdiobus_register(struct mii_bus *bus, struct module *owner) | |||
| 396 | } | 393 | } |
| 397 | } | 394 | } |
| 398 | 395 | ||
| 396 | if (bus->reset) | ||
| 397 | bus->reset(bus); | ||
| 398 | |||
| 399 | for (i = 0; i < PHY_MAX_ADDR; i++) { | 399 | for (i = 0; i < PHY_MAX_ADDR; i++) { |
| 400 | if ((bus->phy_mask & (1 << i)) == 0) { | 400 | if ((bus->phy_mask & (1 << i)) == 0) { |
| 401 | struct phy_device *phydev; | 401 | struct phy_device *phydev; |
diff --git a/drivers/net/usb/ch9200.c b/drivers/net/usb/ch9200.c index c4f1c363e24b..9df3c1ffff35 100644 --- a/drivers/net/usb/ch9200.c +++ b/drivers/net/usb/ch9200.c | |||
| @@ -310,8 +310,8 @@ static int get_mac_address(struct usbnet *dev, unsigned char *data) | |||
| 310 | int rd_mac_len = 0; | 310 | int rd_mac_len = 0; |
| 311 | 311 | ||
| 312 | netdev_dbg(dev->net, "get_mac_address:\n\tusbnet VID:%0x PID:%0x\n", | 312 | netdev_dbg(dev->net, "get_mac_address:\n\tusbnet VID:%0x PID:%0x\n", |
| 313 | dev->udev->descriptor.idVendor, | 313 | le16_to_cpu(dev->udev->descriptor.idVendor), |
| 314 | dev->udev->descriptor.idProduct); | 314 | le16_to_cpu(dev->udev->descriptor.idProduct)); |
| 315 | 315 | ||
| 316 | memset(mac_addr, 0, sizeof(mac_addr)); | 316 | memset(mac_addr, 0, sizeof(mac_addr)); |
| 317 | rd_mac_len = control_read(dev, REQUEST_READ, 0, | 317 | rd_mac_len = control_read(dev, REQUEST_READ, 0, |
diff --git a/drivers/net/usb/qmi_wwan.c b/drivers/net/usb/qmi_wwan.c index d7165767ca9d..8f923a147fa9 100644 --- a/drivers/net/usb/qmi_wwan.c +++ b/drivers/net/usb/qmi_wwan.c | |||
| @@ -1196,6 +1196,8 @@ static const struct usb_device_id products[] = { | |||
| 1196 | {QMI_FIXED_INTF(0x1199, 0x9071, 10)}, /* Sierra Wireless MC74xx */ | 1196 | {QMI_FIXED_INTF(0x1199, 0x9071, 10)}, /* Sierra Wireless MC74xx */ |
| 1197 | {QMI_FIXED_INTF(0x1199, 0x9079, 8)}, /* Sierra Wireless EM74xx */ | 1197 | {QMI_FIXED_INTF(0x1199, 0x9079, 8)}, /* Sierra Wireless EM74xx */ |
| 1198 | {QMI_FIXED_INTF(0x1199, 0x9079, 10)}, /* Sierra Wireless EM74xx */ | 1198 | {QMI_FIXED_INTF(0x1199, 0x9079, 10)}, /* Sierra Wireless EM74xx */ |
| 1199 | {QMI_FIXED_INTF(0x1199, 0x907b, 8)}, /* Sierra Wireless EM74xx */ | ||
| 1200 | {QMI_FIXED_INTF(0x1199, 0x907b, 10)}, /* Sierra Wireless EM74xx */ | ||
| 1199 | {QMI_FIXED_INTF(0x1bbb, 0x011e, 4)}, /* Telekom Speedstick LTE II (Alcatel One Touch L100V LTE) */ | 1201 | {QMI_FIXED_INTF(0x1bbb, 0x011e, 4)}, /* Telekom Speedstick LTE II (Alcatel One Touch L100V LTE) */ |
| 1200 | {QMI_FIXED_INTF(0x1bbb, 0x0203, 2)}, /* Alcatel L800MA */ | 1202 | {QMI_FIXED_INTF(0x1bbb, 0x0203, 2)}, /* Alcatel L800MA */ |
| 1201 | {QMI_FIXED_INTF(0x2357, 0x0201, 4)}, /* TP-LINK HSUPA Modem MA180 */ | 1203 | {QMI_FIXED_INTF(0x2357, 0x0201, 4)}, /* TP-LINK HSUPA Modem MA180 */ |
diff --git a/drivers/net/vmxnet3/vmxnet3_drv.c b/drivers/net/vmxnet3/vmxnet3_drv.c index 25bc764ae7dc..d1c7029ded7c 100644 --- a/drivers/net/vmxnet3/vmxnet3_drv.c +++ b/drivers/net/vmxnet3/vmxnet3_drv.c | |||
| @@ -2962,6 +2962,11 @@ vmxnet3_force_close(struct vmxnet3_adapter *adapter) | |||
| 2962 | /* we need to enable NAPI, otherwise dev_close will deadlock */ | 2962 | /* we need to enable NAPI, otherwise dev_close will deadlock */ |
| 2963 | for (i = 0; i < adapter->num_rx_queues; i++) | 2963 | for (i = 0; i < adapter->num_rx_queues; i++) |
| 2964 | napi_enable(&adapter->rx_queue[i].napi); | 2964 | napi_enable(&adapter->rx_queue[i].napi); |
| 2965 | /* | ||
| 2966 | * Need to clear the quiesce bit to ensure that vmxnet3_close | ||
| 2967 | * can quiesce the device properly | ||
| 2968 | */ | ||
| 2969 | clear_bit(VMXNET3_STATE_BIT_QUIESCED, &adapter->state); | ||
| 2965 | dev_close(adapter->netdev); | 2970 | dev_close(adapter->netdev); |
| 2966 | } | 2971 | } |
| 2967 | 2972 | ||
diff --git a/drivers/net/vrf.c b/drivers/net/vrf.c index ceda5861da78..db882493875c 100644 --- a/drivers/net/vrf.c +++ b/drivers/net/vrf.c | |||
| @@ -989,6 +989,7 @@ static u32 vrf_fib_table(const struct net_device *dev) | |||
| 989 | 989 | ||
| 990 | static int vrf_rcv_finish(struct net *net, struct sock *sk, struct sk_buff *skb) | 990 | static int vrf_rcv_finish(struct net *net, struct sock *sk, struct sk_buff *skb) |
| 991 | { | 991 | { |
| 992 | kfree_skb(skb); | ||
| 992 | return 0; | 993 | return 0; |
| 993 | } | 994 | } |
| 994 | 995 | ||
| @@ -998,7 +999,7 @@ static struct sk_buff *vrf_rcv_nfhook(u8 pf, unsigned int hook, | |||
| 998 | { | 999 | { |
| 999 | struct net *net = dev_net(dev); | 1000 | struct net *net = dev_net(dev); |
| 1000 | 1001 | ||
| 1001 | if (NF_HOOK(pf, hook, net, NULL, skb, dev, NULL, vrf_rcv_finish) < 0) | 1002 | if (nf_hook(pf, hook, net, NULL, skb, dev, NULL, vrf_rcv_finish) != 1) |
| 1002 | skb = NULL; /* kfree_skb(skb) handled by nf code */ | 1003 | skb = NULL; /* kfree_skb(skb) handled by nf code */ |
| 1003 | 1004 | ||
| 1004 | return skb; | 1005 | return skb; |
diff --git a/drivers/net/xen-netfront.c b/drivers/net/xen-netfront.c index 6ffc482550c1..7b61adb6270c 100644 --- a/drivers/net/xen-netfront.c +++ b/drivers/net/xen-netfront.c | |||
| @@ -1934,8 +1934,7 @@ abort_transaction_no_dev_fatal: | |||
| 1934 | xennet_disconnect_backend(info); | 1934 | xennet_disconnect_backend(info); |
| 1935 | xennet_destroy_queues(info); | 1935 | xennet_destroy_queues(info); |
| 1936 | out: | 1936 | out: |
| 1937 | unregister_netdev(info->netdev); | 1937 | device_unregister(&dev->dev); |
| 1938 | xennet_free_netdev(info->netdev); | ||
| 1939 | return err; | 1938 | return err; |
| 1940 | } | 1939 | } |
| 1941 | 1940 | ||
diff --git a/drivers/s390/cio/ccwgroup.c b/drivers/s390/cio/ccwgroup.c index e443b0d0b236..34b9ad6b3143 100644 --- a/drivers/s390/cio/ccwgroup.c +++ b/drivers/s390/cio/ccwgroup.c | |||
| @@ -35,7 +35,7 @@ static struct bus_type ccwgroup_bus_type; | |||
| 35 | static void __ccwgroup_remove_symlinks(struct ccwgroup_device *gdev) | 35 | static void __ccwgroup_remove_symlinks(struct ccwgroup_device *gdev) |
| 36 | { | 36 | { |
| 37 | int i; | 37 | int i; |
| 38 | char str[8]; | 38 | char str[16]; |
| 39 | 39 | ||
| 40 | for (i = 0; i < gdev->count; i++) { | 40 | for (i = 0; i < gdev->count; i++) { |
| 41 | sprintf(str, "cdev%d", i); | 41 | sprintf(str, "cdev%d", i); |
| @@ -238,7 +238,7 @@ static void ccwgroup_release(struct device *dev) | |||
| 238 | 238 | ||
| 239 | static int __ccwgroup_create_symlinks(struct ccwgroup_device *gdev) | 239 | static int __ccwgroup_create_symlinks(struct ccwgroup_device *gdev) |
| 240 | { | 240 | { |
| 241 | char str[8]; | 241 | char str[16]; |
| 242 | int i, rc; | 242 | int i, rc; |
| 243 | 243 | ||
| 244 | for (i = 0; i < gdev->count; i++) { | 244 | for (i = 0; i < gdev->count; i++) { |
diff --git a/drivers/s390/cio/qdio_debug.h b/drivers/s390/cio/qdio_debug.h index f33ce8577619..1d595d17bf11 100644 --- a/drivers/s390/cio/qdio_debug.h +++ b/drivers/s390/cio/qdio_debug.h | |||
| @@ -11,7 +11,7 @@ | |||
| 11 | #include "qdio.h" | 11 | #include "qdio.h" |
| 12 | 12 | ||
| 13 | /* that gives us 15 characters in the text event views */ | 13 | /* that gives us 15 characters in the text event views */ |
| 14 | #define QDIO_DBF_LEN 16 | 14 | #define QDIO_DBF_LEN 32 |
| 15 | 15 | ||
| 16 | extern debug_info_t *qdio_dbf_setup; | 16 | extern debug_info_t *qdio_dbf_setup; |
| 17 | extern debug_info_t *qdio_dbf_error; | 17 | extern debug_info_t *qdio_dbf_error; |
diff --git a/drivers/s390/net/qeth_core.h b/drivers/s390/net/qeth_core.h index f6aa21176d89..30bc6105aac3 100644 --- a/drivers/s390/net/qeth_core.h +++ b/drivers/s390/net/qeth_core.h | |||
| @@ -701,6 +701,7 @@ enum qeth_discipline_id { | |||
| 701 | }; | 701 | }; |
| 702 | 702 | ||
| 703 | struct qeth_discipline { | 703 | struct qeth_discipline { |
| 704 | const struct device_type *devtype; | ||
| 704 | void (*start_poll)(struct ccw_device *, int, unsigned long); | 705 | void (*start_poll)(struct ccw_device *, int, unsigned long); |
| 705 | qdio_handler_t *input_handler; | 706 | qdio_handler_t *input_handler; |
| 706 | qdio_handler_t *output_handler; | 707 | qdio_handler_t *output_handler; |
| @@ -875,6 +876,9 @@ extern struct qeth_discipline qeth_l2_discipline; | |||
| 875 | extern struct qeth_discipline qeth_l3_discipline; | 876 | extern struct qeth_discipline qeth_l3_discipline; |
| 876 | extern const struct attribute_group *qeth_generic_attr_groups[]; | 877 | extern const struct attribute_group *qeth_generic_attr_groups[]; |
| 877 | extern const struct attribute_group *qeth_osn_attr_groups[]; | 878 | extern const struct attribute_group *qeth_osn_attr_groups[]; |
| 879 | extern const struct attribute_group qeth_device_attr_group; | ||
| 880 | extern const struct attribute_group qeth_device_blkt_group; | ||
| 881 | extern const struct device_type qeth_generic_devtype; | ||
| 878 | extern struct workqueue_struct *qeth_wq; | 882 | extern struct workqueue_struct *qeth_wq; |
| 879 | 883 | ||
| 880 | int qeth_card_hw_is_reachable(struct qeth_card *); | 884 | int qeth_card_hw_is_reachable(struct qeth_card *); |
diff --git a/drivers/s390/net/qeth_core_main.c b/drivers/s390/net/qeth_core_main.c index 38114a8d56e0..fc6d85f2b38d 100644 --- a/drivers/s390/net/qeth_core_main.c +++ b/drivers/s390/net/qeth_core_main.c | |||
| @@ -5530,10 +5530,12 @@ void qeth_core_free_discipline(struct qeth_card *card) | |||
| 5530 | card->discipline = NULL; | 5530 | card->discipline = NULL; |
| 5531 | } | 5531 | } |
| 5532 | 5532 | ||
| 5533 | static const struct device_type qeth_generic_devtype = { | 5533 | const struct device_type qeth_generic_devtype = { |
| 5534 | .name = "qeth_generic", | 5534 | .name = "qeth_generic", |
| 5535 | .groups = qeth_generic_attr_groups, | 5535 | .groups = qeth_generic_attr_groups, |
| 5536 | }; | 5536 | }; |
| 5537 | EXPORT_SYMBOL_GPL(qeth_generic_devtype); | ||
| 5538 | |||
| 5537 | static const struct device_type qeth_osn_devtype = { | 5539 | static const struct device_type qeth_osn_devtype = { |
| 5538 | .name = "qeth_osn", | 5540 | .name = "qeth_osn", |
| 5539 | .groups = qeth_osn_attr_groups, | 5541 | .groups = qeth_osn_attr_groups, |
| @@ -5659,23 +5661,22 @@ static int qeth_core_probe_device(struct ccwgroup_device *gdev) | |||
| 5659 | goto err_card; | 5661 | goto err_card; |
| 5660 | } | 5662 | } |
| 5661 | 5663 | ||
| 5662 | if (card->info.type == QETH_CARD_TYPE_OSN) | ||
| 5663 | gdev->dev.type = &qeth_osn_devtype; | ||
| 5664 | else | ||
| 5665 | gdev->dev.type = &qeth_generic_devtype; | ||
| 5666 | |||
| 5667 | switch (card->info.type) { | 5664 | switch (card->info.type) { |
| 5668 | case QETH_CARD_TYPE_OSN: | 5665 | case QETH_CARD_TYPE_OSN: |
| 5669 | case QETH_CARD_TYPE_OSM: | 5666 | case QETH_CARD_TYPE_OSM: |
| 5670 | rc = qeth_core_load_discipline(card, QETH_DISCIPLINE_LAYER2); | 5667 | rc = qeth_core_load_discipline(card, QETH_DISCIPLINE_LAYER2); |
| 5671 | if (rc) | 5668 | if (rc) |
| 5672 | goto err_card; | 5669 | goto err_card; |
| 5670 | |||
| 5671 | gdev->dev.type = (card->info.type != QETH_CARD_TYPE_OSN) | ||
| 5672 | ? card->discipline->devtype | ||
| 5673 | : &qeth_osn_devtype; | ||
| 5673 | rc = card->discipline->setup(card->gdev); | 5674 | rc = card->discipline->setup(card->gdev); |
| 5674 | if (rc) | 5675 | if (rc) |
| 5675 | goto err_disc; | 5676 | goto err_disc; |
| 5676 | case QETH_CARD_TYPE_OSD: | 5677 | break; |
| 5677 | case QETH_CARD_TYPE_OSX: | ||
| 5678 | default: | 5678 | default: |
| 5679 | gdev->dev.type = &qeth_generic_devtype; | ||
| 5679 | break; | 5680 | break; |
| 5680 | } | 5681 | } |
| 5681 | 5682 | ||
| @@ -5731,8 +5732,10 @@ static int qeth_core_set_online(struct ccwgroup_device *gdev) | |||
| 5731 | if (rc) | 5732 | if (rc) |
| 5732 | goto err; | 5733 | goto err; |
| 5733 | rc = card->discipline->setup(card->gdev); | 5734 | rc = card->discipline->setup(card->gdev); |
| 5734 | if (rc) | 5735 | if (rc) { |
| 5736 | qeth_core_free_discipline(card); | ||
| 5735 | goto err; | 5737 | goto err; |
| 5738 | } | ||
| 5736 | } | 5739 | } |
| 5737 | rc = card->discipline->set_online(gdev); | 5740 | rc = card->discipline->set_online(gdev); |
| 5738 | err: | 5741 | err: |
diff --git a/drivers/s390/net/qeth_core_sys.c b/drivers/s390/net/qeth_core_sys.c index 75b29fd2fcf4..db6a285d41e0 100644 --- a/drivers/s390/net/qeth_core_sys.c +++ b/drivers/s390/net/qeth_core_sys.c | |||
| @@ -413,12 +413,16 @@ static ssize_t qeth_dev_layer2_store(struct device *dev, | |||
| 413 | 413 | ||
| 414 | if (card->options.layer2 == newdis) | 414 | if (card->options.layer2 == newdis) |
| 415 | goto out; | 415 | goto out; |
| 416 | else { | 416 | if (card->info.type == QETH_CARD_TYPE_OSM) { |
| 417 | card->info.mac_bits = 0; | 417 | /* fixed layer, can't switch */ |
| 418 | if (card->discipline) { | 418 | rc = -EOPNOTSUPP; |
| 419 | card->discipline->remove(card->gdev); | 419 | goto out; |
| 420 | qeth_core_free_discipline(card); | 420 | } |
| 421 | } | 421 | |
| 422 | card->info.mac_bits = 0; | ||
| 423 | if (card->discipline) { | ||
| 424 | card->discipline->remove(card->gdev); | ||
| 425 | qeth_core_free_discipline(card); | ||
| 422 | } | 426 | } |
| 423 | 427 | ||
| 424 | rc = qeth_core_load_discipline(card, newdis); | 428 | rc = qeth_core_load_discipline(card, newdis); |
| @@ -426,6 +430,8 @@ static ssize_t qeth_dev_layer2_store(struct device *dev, | |||
| 426 | goto out; | 430 | goto out; |
| 427 | 431 | ||
| 428 | rc = card->discipline->setup(card->gdev); | 432 | rc = card->discipline->setup(card->gdev); |
| 433 | if (rc) | ||
| 434 | qeth_core_free_discipline(card); | ||
| 429 | out: | 435 | out: |
| 430 | mutex_unlock(&card->discipline_mutex); | 436 | mutex_unlock(&card->discipline_mutex); |
| 431 | return rc ? rc : count; | 437 | return rc ? rc : count; |
| @@ -703,10 +709,11 @@ static struct attribute *qeth_blkt_device_attrs[] = { | |||
| 703 | &dev_attr_inter_jumbo.attr, | 709 | &dev_attr_inter_jumbo.attr, |
| 704 | NULL, | 710 | NULL, |
| 705 | }; | 711 | }; |
| 706 | static struct attribute_group qeth_device_blkt_group = { | 712 | const struct attribute_group qeth_device_blkt_group = { |
| 707 | .name = "blkt", | 713 | .name = "blkt", |
| 708 | .attrs = qeth_blkt_device_attrs, | 714 | .attrs = qeth_blkt_device_attrs, |
| 709 | }; | 715 | }; |
| 716 | EXPORT_SYMBOL_GPL(qeth_device_blkt_group); | ||
| 710 | 717 | ||
| 711 | static struct attribute *qeth_device_attrs[] = { | 718 | static struct attribute *qeth_device_attrs[] = { |
| 712 | &dev_attr_state.attr, | 719 | &dev_attr_state.attr, |
| @@ -726,9 +733,10 @@ static struct attribute *qeth_device_attrs[] = { | |||
| 726 | &dev_attr_switch_attrs.attr, | 733 | &dev_attr_switch_attrs.attr, |
| 727 | NULL, | 734 | NULL, |
| 728 | }; | 735 | }; |
| 729 | static struct attribute_group qeth_device_attr_group = { | 736 | const struct attribute_group qeth_device_attr_group = { |
| 730 | .attrs = qeth_device_attrs, | 737 | .attrs = qeth_device_attrs, |
| 731 | }; | 738 | }; |
| 739 | EXPORT_SYMBOL_GPL(qeth_device_attr_group); | ||
| 732 | 740 | ||
| 733 | const struct attribute_group *qeth_generic_attr_groups[] = { | 741 | const struct attribute_group *qeth_generic_attr_groups[] = { |
| 734 | &qeth_device_attr_group, | 742 | &qeth_device_attr_group, |
diff --git a/drivers/s390/net/qeth_l2.h b/drivers/s390/net/qeth_l2.h index 29d9fb3890ad..0d59f9a45ea9 100644 --- a/drivers/s390/net/qeth_l2.h +++ b/drivers/s390/net/qeth_l2.h | |||
| @@ -8,6 +8,8 @@ | |||
| 8 | 8 | ||
| 9 | #include "qeth_core.h" | 9 | #include "qeth_core.h" |
| 10 | 10 | ||
| 11 | extern const struct attribute_group *qeth_l2_attr_groups[]; | ||
| 12 | |||
| 11 | int qeth_l2_create_device_attributes(struct device *); | 13 | int qeth_l2_create_device_attributes(struct device *); |
| 12 | void qeth_l2_remove_device_attributes(struct device *); | 14 | void qeth_l2_remove_device_attributes(struct device *); |
| 13 | void qeth_l2_setup_bridgeport_attrs(struct qeth_card *card); | 15 | void qeth_l2_setup_bridgeport_attrs(struct qeth_card *card); |
diff --git a/drivers/s390/net/qeth_l2_main.c b/drivers/s390/net/qeth_l2_main.c index 1b07f382d74c..bd2df62a5cdf 100644 --- a/drivers/s390/net/qeth_l2_main.c +++ b/drivers/s390/net/qeth_l2_main.c | |||
| @@ -880,11 +880,21 @@ static int qeth_l2_stop(struct net_device *dev) | |||
| 880 | return 0; | 880 | return 0; |
| 881 | } | 881 | } |
| 882 | 882 | ||
| 883 | static const struct device_type qeth_l2_devtype = { | ||
| 884 | .name = "qeth_layer2", | ||
| 885 | .groups = qeth_l2_attr_groups, | ||
| 886 | }; | ||
| 887 | |||
| 883 | static int qeth_l2_probe_device(struct ccwgroup_device *gdev) | 888 | static int qeth_l2_probe_device(struct ccwgroup_device *gdev) |
| 884 | { | 889 | { |
| 885 | struct qeth_card *card = dev_get_drvdata(&gdev->dev); | 890 | struct qeth_card *card = dev_get_drvdata(&gdev->dev); |
| 891 | int rc; | ||
| 886 | 892 | ||
| 887 | qeth_l2_create_device_attributes(&gdev->dev); | 893 | if (gdev->dev.type == &qeth_generic_devtype) { |
| 894 | rc = qeth_l2_create_device_attributes(&gdev->dev); | ||
| 895 | if (rc) | ||
| 896 | return rc; | ||
| 897 | } | ||
| 888 | INIT_LIST_HEAD(&card->vid_list); | 898 | INIT_LIST_HEAD(&card->vid_list); |
| 889 | hash_init(card->mac_htable); | 899 | hash_init(card->mac_htable); |
| 890 | card->options.layer2 = 1; | 900 | card->options.layer2 = 1; |
| @@ -896,7 +906,8 @@ static void qeth_l2_remove_device(struct ccwgroup_device *cgdev) | |||
| 896 | { | 906 | { |
| 897 | struct qeth_card *card = dev_get_drvdata(&cgdev->dev); | 907 | struct qeth_card *card = dev_get_drvdata(&cgdev->dev); |
| 898 | 908 | ||
| 899 | qeth_l2_remove_device_attributes(&cgdev->dev); | 909 | if (cgdev->dev.type == &qeth_generic_devtype) |
| 910 | qeth_l2_remove_device_attributes(&cgdev->dev); | ||
| 900 | qeth_set_allowed_threads(card, 0, 1); | 911 | qeth_set_allowed_threads(card, 0, 1); |
| 901 | wait_event(card->wait_q, qeth_threads_running(card, 0xffffffff) == 0); | 912 | wait_event(card->wait_q, qeth_threads_running(card, 0xffffffff) == 0); |
| 902 | 913 | ||
| @@ -954,7 +965,6 @@ static int qeth_l2_setup_netdev(struct qeth_card *card) | |||
| 954 | case QETH_CARD_TYPE_OSN: | 965 | case QETH_CARD_TYPE_OSN: |
| 955 | card->dev = alloc_netdev(0, "osn%d", NET_NAME_UNKNOWN, | 966 | card->dev = alloc_netdev(0, "osn%d", NET_NAME_UNKNOWN, |
| 956 | ether_setup); | 967 | ether_setup); |
| 957 | card->dev->flags |= IFF_NOARP; | ||
| 958 | break; | 968 | break; |
| 959 | default: | 969 | default: |
| 960 | card->dev = alloc_etherdev(0); | 970 | card->dev = alloc_etherdev(0); |
| @@ -969,9 +979,12 @@ static int qeth_l2_setup_netdev(struct qeth_card *card) | |||
| 969 | card->dev->min_mtu = 64; | 979 | card->dev->min_mtu = 64; |
| 970 | card->dev->max_mtu = ETH_MAX_MTU; | 980 | card->dev->max_mtu = ETH_MAX_MTU; |
| 971 | card->dev->netdev_ops = &qeth_l2_netdev_ops; | 981 | card->dev->netdev_ops = &qeth_l2_netdev_ops; |
| 972 | card->dev->ethtool_ops = | 982 | if (card->info.type == QETH_CARD_TYPE_OSN) { |
| 973 | (card->info.type != QETH_CARD_TYPE_OSN) ? | 983 | card->dev->ethtool_ops = &qeth_l2_osn_ops; |
| 974 | &qeth_l2_ethtool_ops : &qeth_l2_osn_ops; | 984 | card->dev->flags |= IFF_NOARP; |
| 985 | } else { | ||
| 986 | card->dev->ethtool_ops = &qeth_l2_ethtool_ops; | ||
| 987 | } | ||
| 975 | card->dev->features |= NETIF_F_HW_VLAN_CTAG_FILTER; | 988 | card->dev->features |= NETIF_F_HW_VLAN_CTAG_FILTER; |
| 976 | if (card->info.type == QETH_CARD_TYPE_OSD && !card->info.guestlan) { | 989 | if (card->info.type == QETH_CARD_TYPE_OSD && !card->info.guestlan) { |
| 977 | card->dev->hw_features = NETIF_F_SG; | 990 | card->dev->hw_features = NETIF_F_SG; |
| @@ -1269,6 +1282,7 @@ static int qeth_l2_control_event(struct qeth_card *card, | |||
| 1269 | } | 1282 | } |
| 1270 | 1283 | ||
| 1271 | struct qeth_discipline qeth_l2_discipline = { | 1284 | struct qeth_discipline qeth_l2_discipline = { |
| 1285 | .devtype = &qeth_l2_devtype, | ||
| 1272 | .start_poll = qeth_qdio_start_poll, | 1286 | .start_poll = qeth_qdio_start_poll, |
| 1273 | .input_handler = (qdio_handler_t *) qeth_qdio_input_handler, | 1287 | .input_handler = (qdio_handler_t *) qeth_qdio_input_handler, |
| 1274 | .output_handler = (qdio_handler_t *) qeth_qdio_output_handler, | 1288 | .output_handler = (qdio_handler_t *) qeth_qdio_output_handler, |
diff --git a/drivers/s390/net/qeth_l2_sys.c b/drivers/s390/net/qeth_l2_sys.c index 687972356d6b..9696baa49e2d 100644 --- a/drivers/s390/net/qeth_l2_sys.c +++ b/drivers/s390/net/qeth_l2_sys.c | |||
| @@ -269,3 +269,11 @@ void qeth_l2_setup_bridgeport_attrs(struct qeth_card *card) | |||
| 269 | } else | 269 | } else |
| 270 | qeth_bridgeport_an_set(card, 0); | 270 | qeth_bridgeport_an_set(card, 0); |
| 271 | } | 271 | } |
| 272 | |||
| 273 | const struct attribute_group *qeth_l2_attr_groups[] = { | ||
| 274 | &qeth_device_attr_group, | ||
| 275 | &qeth_device_blkt_group, | ||
| 276 | /* l2 specific, see l2_{create,remove}_device_attributes(): */ | ||
| 277 | &qeth_l2_bridgeport_attr_group, | ||
| 278 | NULL, | ||
| 279 | }; | ||
diff --git a/drivers/s390/net/qeth_l3_main.c b/drivers/s390/net/qeth_l3_main.c index 6e0354ef4b86..d8df1e635163 100644 --- a/drivers/s390/net/qeth_l3_main.c +++ b/drivers/s390/net/qeth_l3_main.c | |||
| @@ -3039,8 +3039,13 @@ static int qeth_l3_setup_netdev(struct qeth_card *card) | |||
| 3039 | static int qeth_l3_probe_device(struct ccwgroup_device *gdev) | 3039 | static int qeth_l3_probe_device(struct ccwgroup_device *gdev) |
| 3040 | { | 3040 | { |
| 3041 | struct qeth_card *card = dev_get_drvdata(&gdev->dev); | 3041 | struct qeth_card *card = dev_get_drvdata(&gdev->dev); |
| 3042 | int rc; | ||
| 3042 | 3043 | ||
| 3043 | qeth_l3_create_device_attributes(&gdev->dev); | 3044 | rc = qeth_l3_create_device_attributes(&gdev->dev); |
| 3045 | if (rc) | ||
| 3046 | return rc; | ||
| 3047 | hash_init(card->ip_htable); | ||
| 3048 | hash_init(card->ip_mc_htable); | ||
| 3044 | card->options.layer2 = 0; | 3049 | card->options.layer2 = 0; |
| 3045 | card->info.hwtrap = 0; | 3050 | card->info.hwtrap = 0; |
| 3046 | return 0; | 3051 | return 0; |
| @@ -3306,6 +3311,7 @@ static int qeth_l3_control_event(struct qeth_card *card, | |||
| 3306 | } | 3311 | } |
| 3307 | 3312 | ||
| 3308 | struct qeth_discipline qeth_l3_discipline = { | 3313 | struct qeth_discipline qeth_l3_discipline = { |
| 3314 | .devtype = &qeth_generic_devtype, | ||
| 3309 | .start_poll = qeth_qdio_start_poll, | 3315 | .start_poll = qeth_qdio_start_poll, |
| 3310 | .input_handler = (qdio_handler_t *) qeth_qdio_input_handler, | 3316 | .input_handler = (qdio_handler_t *) qeth_qdio_input_handler, |
| 3311 | .output_handler = (qdio_handler_t *) qeth_qdio_output_handler, | 3317 | .output_handler = (qdio_handler_t *) qeth_qdio_output_handler, |
diff --git a/drivers/s390/virtio/virtio_ccw.c b/drivers/s390/virtio/virtio_ccw.c index 2a76ea78a0bf..b18fe2014cf2 100644 --- a/drivers/s390/virtio/virtio_ccw.c +++ b/drivers/s390/virtio/virtio_ccw.c | |||
| @@ -87,7 +87,7 @@ struct vq_info_block { | |||
| 87 | } __packed; | 87 | } __packed; |
| 88 | 88 | ||
| 89 | struct virtio_feature_desc { | 89 | struct virtio_feature_desc { |
| 90 | __u32 features; | 90 | __le32 features; |
| 91 | __u8 index; | 91 | __u8 index; |
| 92 | } __packed; | 92 | } __packed; |
| 93 | 93 | ||
diff --git a/drivers/soc/ti/knav_dma.c b/drivers/soc/ti/knav_dma.c index ecebe2eecc3a..026182d3b27c 100644 --- a/drivers/soc/ti/knav_dma.c +++ b/drivers/soc/ti/knav_dma.c | |||
| @@ -413,7 +413,7 @@ static int of_channel_match_helper(struct device_node *np, const char *name, | |||
| 413 | * @name: slave channel name | 413 | * @name: slave channel name |
| 414 | * @config: dma configuration parameters | 414 | * @config: dma configuration parameters |
| 415 | * | 415 | * |
| 416 | * Returns pointer to appropriate DMA channel on success or NULL. | 416 | * Returns pointer to appropriate DMA channel on success or error. |
| 417 | */ | 417 | */ |
| 418 | void *knav_dma_open_channel(struct device *dev, const char *name, | 418 | void *knav_dma_open_channel(struct device *dev, const char *name, |
| 419 | struct knav_dma_cfg *config) | 419 | struct knav_dma_cfg *config) |
diff --git a/fs/cifs/cifsacl.c b/fs/cifs/cifsacl.c index 15bac390dff9..b98436f5c7c7 100644 --- a/fs/cifs/cifsacl.c +++ b/fs/cifs/cifsacl.c | |||
| @@ -1135,20 +1135,19 @@ cifs_acl_to_fattr(struct cifs_sb_info *cifs_sb, struct cifs_fattr *fattr, | |||
| 1135 | u32 acllen = 0; | 1135 | u32 acllen = 0; |
| 1136 | int rc = 0; | 1136 | int rc = 0; |
| 1137 | struct tcon_link *tlink = cifs_sb_tlink(cifs_sb); | 1137 | struct tcon_link *tlink = cifs_sb_tlink(cifs_sb); |
| 1138 | struct cifs_tcon *tcon; | 1138 | struct smb_version_operations *ops; |
| 1139 | 1139 | ||
| 1140 | cifs_dbg(NOISY, "converting ACL to mode for %s\n", path); | 1140 | cifs_dbg(NOISY, "converting ACL to mode for %s\n", path); |
| 1141 | 1141 | ||
| 1142 | if (IS_ERR(tlink)) | 1142 | if (IS_ERR(tlink)) |
| 1143 | return PTR_ERR(tlink); | 1143 | return PTR_ERR(tlink); |
| 1144 | tcon = tlink_tcon(tlink); | ||
| 1145 | 1144 | ||
| 1146 | if (pfid && (tcon->ses->server->ops->get_acl_by_fid)) | 1145 | ops = tlink_tcon(tlink)->ses->server->ops; |
| 1147 | pntsd = tcon->ses->server->ops->get_acl_by_fid(cifs_sb, pfid, | 1146 | |
| 1148 | &acllen); | 1147 | if (pfid && (ops->get_acl_by_fid)) |
| 1149 | else if (tcon->ses->server->ops->get_acl) | 1148 | pntsd = ops->get_acl_by_fid(cifs_sb, pfid, &acllen); |
| 1150 | pntsd = tcon->ses->server->ops->get_acl(cifs_sb, inode, path, | 1149 | else if (ops->get_acl) |
| 1151 | &acllen); | 1150 | pntsd = ops->get_acl(cifs_sb, inode, path, &acllen); |
| 1152 | else { | 1151 | else { |
| 1153 | cifs_put_tlink(tlink); | 1152 | cifs_put_tlink(tlink); |
| 1154 | return -EOPNOTSUPP; | 1153 | return -EOPNOTSUPP; |
| @@ -1181,23 +1180,23 @@ id_mode_to_cifs_acl(struct inode *inode, const char *path, __u64 nmode, | |||
| 1181 | struct cifs_ntsd *pnntsd = NULL; /* modified acl to be sent to server */ | 1180 | struct cifs_ntsd *pnntsd = NULL; /* modified acl to be sent to server */ |
| 1182 | struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb); | 1181 | struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb); |
| 1183 | struct tcon_link *tlink = cifs_sb_tlink(cifs_sb); | 1182 | struct tcon_link *tlink = cifs_sb_tlink(cifs_sb); |
| 1184 | struct cifs_tcon *tcon; | 1183 | struct smb_version_operations *ops; |
| 1185 | 1184 | ||
| 1186 | if (IS_ERR(tlink)) | 1185 | if (IS_ERR(tlink)) |
| 1187 | return PTR_ERR(tlink); | 1186 | return PTR_ERR(tlink); |
| 1188 | tcon = tlink_tcon(tlink); | 1187 | |
| 1188 | ops = tlink_tcon(tlink)->ses->server->ops; | ||
| 1189 | 1189 | ||
| 1190 | cifs_dbg(NOISY, "set ACL from mode for %s\n", path); | 1190 | cifs_dbg(NOISY, "set ACL from mode for %s\n", path); |
| 1191 | 1191 | ||
| 1192 | /* Get the security descriptor */ | 1192 | /* Get the security descriptor */ |
| 1193 | 1193 | ||
| 1194 | if (tcon->ses->server->ops->get_acl == NULL) { | 1194 | if (ops->get_acl == NULL) { |
| 1195 | cifs_put_tlink(tlink); | 1195 | cifs_put_tlink(tlink); |
| 1196 | return -EOPNOTSUPP; | 1196 | return -EOPNOTSUPP; |
| 1197 | } | 1197 | } |
| 1198 | 1198 | ||
| 1199 | pntsd = tcon->ses->server->ops->get_acl(cifs_sb, inode, path, | 1199 | pntsd = ops->get_acl(cifs_sb, inode, path, &secdesclen); |
| 1200 | &secdesclen); | ||
| 1201 | if (IS_ERR(pntsd)) { | 1200 | if (IS_ERR(pntsd)) { |
| 1202 | rc = PTR_ERR(pntsd); | 1201 | rc = PTR_ERR(pntsd); |
| 1203 | cifs_dbg(VFS, "%s: error %d getting sec desc\n", __func__, rc); | 1202 | cifs_dbg(VFS, "%s: error %d getting sec desc\n", __func__, rc); |
| @@ -1224,13 +1223,12 @@ id_mode_to_cifs_acl(struct inode *inode, const char *path, __u64 nmode, | |||
| 1224 | 1223 | ||
| 1225 | cifs_dbg(NOISY, "build_sec_desc rc: %d\n", rc); | 1224 | cifs_dbg(NOISY, "build_sec_desc rc: %d\n", rc); |
| 1226 | 1225 | ||
| 1227 | if (tcon->ses->server->ops->set_acl == NULL) | 1226 | if (ops->set_acl == NULL) |
| 1228 | rc = -EOPNOTSUPP; | 1227 | rc = -EOPNOTSUPP; |
| 1229 | 1228 | ||
| 1230 | if (!rc) { | 1229 | if (!rc) { |
| 1231 | /* Set the security descriptor */ | 1230 | /* Set the security descriptor */ |
| 1232 | rc = tcon->ses->server->ops->set_acl(pnntsd, secdesclen, inode, | 1231 | rc = ops->set_acl(pnntsd, secdesclen, inode, path, aclflag); |
| 1233 | path, aclflag); | ||
| 1234 | cifs_dbg(NOISY, "set_cifs_acl rc: %d\n", rc); | 1232 | cifs_dbg(NOISY, "set_cifs_acl rc: %d\n", rc); |
| 1235 | } | 1233 | } |
| 1236 | cifs_put_tlink(tlink); | 1234 | cifs_put_tlink(tlink); |
diff --git a/fs/cifs/cifsglob.h b/fs/cifs/cifsglob.h index 8be55be70faf..bcc7d9acad64 100644 --- a/fs/cifs/cifsglob.h +++ b/fs/cifs/cifsglob.h | |||
| @@ -418,7 +418,7 @@ struct smb_version_operations { | |||
| 418 | int (*validate_negotiate)(const unsigned int, struct cifs_tcon *); | 418 | int (*validate_negotiate)(const unsigned int, struct cifs_tcon *); |
| 419 | ssize_t (*query_all_EAs)(const unsigned int, struct cifs_tcon *, | 419 | ssize_t (*query_all_EAs)(const unsigned int, struct cifs_tcon *, |
| 420 | const unsigned char *, const unsigned char *, char *, | 420 | const unsigned char *, const unsigned char *, char *, |
| 421 | size_t, const struct nls_table *, int); | 421 | size_t, struct cifs_sb_info *); |
| 422 | int (*set_EA)(const unsigned int, struct cifs_tcon *, const char *, | 422 | int (*set_EA)(const unsigned int, struct cifs_tcon *, const char *, |
| 423 | const char *, const void *, const __u16, | 423 | const char *, const void *, const __u16, |
| 424 | const struct nls_table *, int); | 424 | const struct nls_table *, int); |
diff --git a/fs/cifs/cifsproto.h b/fs/cifs/cifsproto.h index e49958c3f8bb..6eb3147132e3 100644 --- a/fs/cifs/cifsproto.h +++ b/fs/cifs/cifsproto.h | |||
| @@ -480,8 +480,7 @@ extern int CIFSSMBCopy(unsigned int xid, | |||
| 480 | extern ssize_t CIFSSMBQAllEAs(const unsigned int xid, struct cifs_tcon *tcon, | 480 | extern ssize_t CIFSSMBQAllEAs(const unsigned int xid, struct cifs_tcon *tcon, |
| 481 | const unsigned char *searchName, | 481 | const unsigned char *searchName, |
| 482 | const unsigned char *ea_name, char *EAData, | 482 | const unsigned char *ea_name, char *EAData, |
| 483 | size_t bufsize, const struct nls_table *nls_codepage, | 483 | size_t bufsize, struct cifs_sb_info *cifs_sb); |
| 484 | int remap_special_chars); | ||
| 485 | extern int CIFSSMBSetEA(const unsigned int xid, struct cifs_tcon *tcon, | 484 | extern int CIFSSMBSetEA(const unsigned int xid, struct cifs_tcon *tcon, |
| 486 | const char *fileName, const char *ea_name, | 485 | const char *fileName, const char *ea_name, |
| 487 | const void *ea_value, const __u16 ea_value_len, | 486 | const void *ea_value, const __u16 ea_value_len, |
diff --git a/fs/cifs/cifssmb.c b/fs/cifs/cifssmb.c index 4c01b3f9abf0..fbb0d4cbda41 100644 --- a/fs/cifs/cifssmb.c +++ b/fs/cifs/cifssmb.c | |||
| @@ -697,9 +697,7 @@ cifs_echo_callback(struct mid_q_entry *mid) | |||
| 697 | { | 697 | { |
| 698 | struct TCP_Server_Info *server = mid->callback_data; | 698 | struct TCP_Server_Info *server = mid->callback_data; |
| 699 | 699 | ||
| 700 | mutex_lock(&server->srv_mutex); | ||
| 701 | DeleteMidQEntry(mid); | 700 | DeleteMidQEntry(mid); |
| 702 | mutex_unlock(&server->srv_mutex); | ||
| 703 | add_credits(server, 1, CIFS_ECHO_OP); | 701 | add_credits(server, 1, CIFS_ECHO_OP); |
| 704 | } | 702 | } |
| 705 | 703 | ||
| @@ -1599,9 +1597,7 @@ cifs_readv_callback(struct mid_q_entry *mid) | |||
| 1599 | } | 1597 | } |
| 1600 | 1598 | ||
| 1601 | queue_work(cifsiod_wq, &rdata->work); | 1599 | queue_work(cifsiod_wq, &rdata->work); |
| 1602 | mutex_lock(&server->srv_mutex); | ||
| 1603 | DeleteMidQEntry(mid); | 1600 | DeleteMidQEntry(mid); |
| 1604 | mutex_unlock(&server->srv_mutex); | ||
| 1605 | add_credits(server, 1, 0); | 1601 | add_credits(server, 1, 0); |
| 1606 | } | 1602 | } |
| 1607 | 1603 | ||
| @@ -2058,7 +2054,6 @@ cifs_writev_callback(struct mid_q_entry *mid) | |||
| 2058 | { | 2054 | { |
| 2059 | struct cifs_writedata *wdata = mid->callback_data; | 2055 | struct cifs_writedata *wdata = mid->callback_data; |
| 2060 | struct cifs_tcon *tcon = tlink_tcon(wdata->cfile->tlink); | 2056 | struct cifs_tcon *tcon = tlink_tcon(wdata->cfile->tlink); |
| 2061 | struct TCP_Server_Info *server = tcon->ses->server; | ||
| 2062 | unsigned int written; | 2057 | unsigned int written; |
| 2063 | WRITE_RSP *smb = (WRITE_RSP *)mid->resp_buf; | 2058 | WRITE_RSP *smb = (WRITE_RSP *)mid->resp_buf; |
| 2064 | 2059 | ||
| @@ -2095,9 +2090,7 @@ cifs_writev_callback(struct mid_q_entry *mid) | |||
| 2095 | } | 2090 | } |
| 2096 | 2091 | ||
| 2097 | queue_work(cifsiod_wq, &wdata->work); | 2092 | queue_work(cifsiod_wq, &wdata->work); |
| 2098 | mutex_lock(&server->srv_mutex); | ||
| 2099 | DeleteMidQEntry(mid); | 2093 | DeleteMidQEntry(mid); |
| 2100 | mutex_unlock(&server->srv_mutex); | ||
| 2101 | add_credits(tcon->ses->server, 1, 0); | 2094 | add_credits(tcon->ses->server, 1, 0); |
| 2102 | } | 2095 | } |
| 2103 | 2096 | ||
| @@ -6076,11 +6069,13 @@ ssize_t | |||
| 6076 | CIFSSMBQAllEAs(const unsigned int xid, struct cifs_tcon *tcon, | 6069 | CIFSSMBQAllEAs(const unsigned int xid, struct cifs_tcon *tcon, |
| 6077 | const unsigned char *searchName, const unsigned char *ea_name, | 6070 | const unsigned char *searchName, const unsigned char *ea_name, |
| 6078 | char *EAData, size_t buf_size, | 6071 | char *EAData, size_t buf_size, |
| 6079 | const struct nls_table *nls_codepage, int remap) | 6072 | struct cifs_sb_info *cifs_sb) |
| 6080 | { | 6073 | { |
| 6081 | /* BB assumes one setup word */ | 6074 | /* BB assumes one setup word */ |
| 6082 | TRANSACTION2_QPI_REQ *pSMB = NULL; | 6075 | TRANSACTION2_QPI_REQ *pSMB = NULL; |
| 6083 | TRANSACTION2_QPI_RSP *pSMBr = NULL; | 6076 | TRANSACTION2_QPI_RSP *pSMBr = NULL; |
| 6077 | int remap = cifs_remap(cifs_sb); | ||
| 6078 | struct nls_table *nls_codepage = cifs_sb->local_nls; | ||
| 6084 | int rc = 0; | 6079 | int rc = 0; |
| 6085 | int bytes_returned; | 6080 | int bytes_returned; |
| 6086 | int list_len; | 6081 | int list_len; |
diff --git a/fs/cifs/file.c b/fs/cifs/file.c index 6ef78ad838e6..0fd081bd2a2f 100644 --- a/fs/cifs/file.c +++ b/fs/cifs/file.c | |||
| @@ -582,7 +582,7 @@ cifs_relock_file(struct cifsFileInfo *cfile) | |||
| 582 | struct cifs_tcon *tcon = tlink_tcon(cfile->tlink); | 582 | struct cifs_tcon *tcon = tlink_tcon(cfile->tlink); |
| 583 | int rc = 0; | 583 | int rc = 0; |
| 584 | 584 | ||
| 585 | down_read(&cinode->lock_sem); | 585 | down_read_nested(&cinode->lock_sem, SINGLE_DEPTH_NESTING); |
| 586 | if (cinode->can_cache_brlcks) { | 586 | if (cinode->can_cache_brlcks) { |
| 587 | /* can cache locks - no need to relock */ | 587 | /* can cache locks - no need to relock */ |
| 588 | up_read(&cinode->lock_sem); | 588 | up_read(&cinode->lock_sem); |
diff --git a/fs/cifs/inode.c b/fs/cifs/inode.c index c3b2fa0b2ec8..4d1fcd76d022 100644 --- a/fs/cifs/inode.c +++ b/fs/cifs/inode.c | |||
| @@ -563,8 +563,7 @@ static int cifs_sfu_mode(struct cifs_fattr *fattr, const unsigned char *path, | |||
| 563 | 563 | ||
| 564 | rc = tcon->ses->server->ops->query_all_EAs(xid, tcon, path, | 564 | rc = tcon->ses->server->ops->query_all_EAs(xid, tcon, path, |
| 565 | "SETFILEBITS", ea_value, 4 /* size of buf */, | 565 | "SETFILEBITS", ea_value, 4 /* size of buf */, |
| 566 | cifs_sb->local_nls, | 566 | cifs_sb); |
| 567 | cifs_remap(cifs_sb)); | ||
| 568 | cifs_put_tlink(tlink); | 567 | cifs_put_tlink(tlink); |
| 569 | if (rc < 0) | 568 | if (rc < 0) |
| 570 | return (int)rc; | 569 | return (int)rc; |
diff --git a/fs/cifs/smb2pdu.c b/fs/cifs/smb2pdu.c index 48ff7703b919..e4afdaae743f 100644 --- a/fs/cifs/smb2pdu.c +++ b/fs/cifs/smb2pdu.c | |||
| @@ -1240,15 +1240,19 @@ SMB2_tcon(const unsigned int xid, struct cifs_ses *ses, const char *tree, | |||
| 1240 | goto tcon_exit; | 1240 | goto tcon_exit; |
| 1241 | } | 1241 | } |
| 1242 | 1242 | ||
| 1243 | if (rsp->ShareType & SMB2_SHARE_TYPE_DISK) | 1243 | switch (rsp->ShareType) { |
| 1244 | case SMB2_SHARE_TYPE_DISK: | ||
| 1244 | cifs_dbg(FYI, "connection to disk share\n"); | 1245 | cifs_dbg(FYI, "connection to disk share\n"); |
| 1245 | else if (rsp->ShareType & SMB2_SHARE_TYPE_PIPE) { | 1246 | break; |
| 1247 | case SMB2_SHARE_TYPE_PIPE: | ||
| 1246 | tcon->ipc = true; | 1248 | tcon->ipc = true; |
| 1247 | cifs_dbg(FYI, "connection to pipe share\n"); | 1249 | cifs_dbg(FYI, "connection to pipe share\n"); |
| 1248 | } else if (rsp->ShareType & SMB2_SHARE_TYPE_PRINT) { | 1250 | break; |
| 1249 | tcon->print = true; | 1251 | case SMB2_SHARE_TYPE_PRINT: |
| 1252 | tcon->ipc = true; | ||
| 1250 | cifs_dbg(FYI, "connection to printer\n"); | 1253 | cifs_dbg(FYI, "connection to printer\n"); |
| 1251 | } else { | 1254 | break; |
| 1255 | default: | ||
| 1252 | cifs_dbg(VFS, "unknown share type %d\n", rsp->ShareType); | 1256 | cifs_dbg(VFS, "unknown share type %d\n", rsp->ShareType); |
| 1253 | rc = -EOPNOTSUPP; | 1257 | rc = -EOPNOTSUPP; |
| 1254 | goto tcon_error_exit; | 1258 | goto tcon_error_exit; |
| @@ -2173,9 +2177,7 @@ smb2_echo_callback(struct mid_q_entry *mid) | |||
| 2173 | if (mid->mid_state == MID_RESPONSE_RECEIVED) | 2177 | if (mid->mid_state == MID_RESPONSE_RECEIVED) |
| 2174 | credits_received = le16_to_cpu(rsp->hdr.sync_hdr.CreditRequest); | 2178 | credits_received = le16_to_cpu(rsp->hdr.sync_hdr.CreditRequest); |
| 2175 | 2179 | ||
| 2176 | mutex_lock(&server->srv_mutex); | ||
| 2177 | DeleteMidQEntry(mid); | 2180 | DeleteMidQEntry(mid); |
| 2178 | mutex_unlock(&server->srv_mutex); | ||
| 2179 | add_credits(server, credits_received, CIFS_ECHO_OP); | 2181 | add_credits(server, credits_received, CIFS_ECHO_OP); |
| 2180 | } | 2182 | } |
| 2181 | 2183 | ||
| @@ -2433,9 +2435,7 @@ smb2_readv_callback(struct mid_q_entry *mid) | |||
| 2433 | cifs_stats_fail_inc(tcon, SMB2_READ_HE); | 2435 | cifs_stats_fail_inc(tcon, SMB2_READ_HE); |
| 2434 | 2436 | ||
| 2435 | queue_work(cifsiod_wq, &rdata->work); | 2437 | queue_work(cifsiod_wq, &rdata->work); |
| 2436 | mutex_lock(&server->srv_mutex); | ||
| 2437 | DeleteMidQEntry(mid); | 2438 | DeleteMidQEntry(mid); |
| 2438 | mutex_unlock(&server->srv_mutex); | ||
| 2439 | add_credits(server, credits_received, 0); | 2439 | add_credits(server, credits_received, 0); |
| 2440 | } | 2440 | } |
| 2441 | 2441 | ||
| @@ -2594,7 +2594,6 @@ smb2_writev_callback(struct mid_q_entry *mid) | |||
| 2594 | { | 2594 | { |
| 2595 | struct cifs_writedata *wdata = mid->callback_data; | 2595 | struct cifs_writedata *wdata = mid->callback_data; |
| 2596 | struct cifs_tcon *tcon = tlink_tcon(wdata->cfile->tlink); | 2596 | struct cifs_tcon *tcon = tlink_tcon(wdata->cfile->tlink); |
| 2597 | struct TCP_Server_Info *server = tcon->ses->server; | ||
| 2598 | unsigned int written; | 2597 | unsigned int written; |
| 2599 | struct smb2_write_rsp *rsp = (struct smb2_write_rsp *)mid->resp_buf; | 2598 | struct smb2_write_rsp *rsp = (struct smb2_write_rsp *)mid->resp_buf; |
| 2600 | unsigned int credits_received = 1; | 2599 | unsigned int credits_received = 1; |
| @@ -2634,9 +2633,7 @@ smb2_writev_callback(struct mid_q_entry *mid) | |||
| 2634 | cifs_stats_fail_inc(tcon, SMB2_WRITE_HE); | 2633 | cifs_stats_fail_inc(tcon, SMB2_WRITE_HE); |
| 2635 | 2634 | ||
| 2636 | queue_work(cifsiod_wq, &wdata->work); | 2635 | queue_work(cifsiod_wq, &wdata->work); |
| 2637 | mutex_lock(&server->srv_mutex); | ||
| 2638 | DeleteMidQEntry(mid); | 2636 | DeleteMidQEntry(mid); |
| 2639 | mutex_unlock(&server->srv_mutex); | ||
| 2640 | add_credits(tcon->ses->server, credits_received, 0); | 2637 | add_credits(tcon->ses->server, credits_received, 0); |
| 2641 | } | 2638 | } |
| 2642 | 2639 | ||
diff --git a/fs/cifs/transport.c b/fs/cifs/transport.c index 4d64b5b8fc9c..47a125ece11e 100644 --- a/fs/cifs/transport.c +++ b/fs/cifs/transport.c | |||
| @@ -94,7 +94,7 @@ DeleteMidQEntry(struct mid_q_entry *midEntry) | |||
| 94 | now = jiffies; | 94 | now = jiffies; |
| 95 | /* commands taking longer than one second are indications that | 95 | /* commands taking longer than one second are indications that |
| 96 | something is wrong, unless it is quite a slow link or server */ | 96 | something is wrong, unless it is quite a slow link or server */ |
| 97 | if ((now - midEntry->when_alloc) > HZ) { | 97 | if (time_after(now, midEntry->when_alloc + HZ)) { |
| 98 | if ((cifsFYI & CIFS_TIMER) && (midEntry->command != command)) { | 98 | if ((cifsFYI & CIFS_TIMER) && (midEntry->command != command)) { |
| 99 | pr_debug(" CIFS slow rsp: cmd %d mid %llu", | 99 | pr_debug(" CIFS slow rsp: cmd %d mid %llu", |
| 100 | midEntry->command, midEntry->mid); | 100 | midEntry->command, midEntry->mid); |
| @@ -613,9 +613,7 @@ cifs_sync_mid_result(struct mid_q_entry *mid, struct TCP_Server_Info *server) | |||
| 613 | } | 613 | } |
| 614 | spin_unlock(&GlobalMid_Lock); | 614 | spin_unlock(&GlobalMid_Lock); |
| 615 | 615 | ||
| 616 | mutex_lock(&server->srv_mutex); | ||
| 617 | DeleteMidQEntry(mid); | 616 | DeleteMidQEntry(mid); |
| 618 | mutex_unlock(&server->srv_mutex); | ||
| 619 | return rc; | 617 | return rc; |
| 620 | } | 618 | } |
| 621 | 619 | ||
diff --git a/fs/cifs/xattr.c b/fs/cifs/xattr.c index 20af5187ba63..3cb5c9e2d4e7 100644 --- a/fs/cifs/xattr.c +++ b/fs/cifs/xattr.c | |||
| @@ -235,8 +235,7 @@ static int cifs_xattr_get(const struct xattr_handler *handler, | |||
| 235 | 235 | ||
| 236 | if (pTcon->ses->server->ops->query_all_EAs) | 236 | if (pTcon->ses->server->ops->query_all_EAs) |
| 237 | rc = pTcon->ses->server->ops->query_all_EAs(xid, pTcon, | 237 | rc = pTcon->ses->server->ops->query_all_EAs(xid, pTcon, |
| 238 | full_path, name, value, size, | 238 | full_path, name, value, size, cifs_sb); |
| 239 | cifs_sb->local_nls, cifs_remap(cifs_sb)); | ||
| 240 | break; | 239 | break; |
| 241 | 240 | ||
| 242 | case XATTR_CIFS_ACL: { | 241 | case XATTR_CIFS_ACL: { |
| @@ -336,8 +335,7 @@ ssize_t cifs_listxattr(struct dentry *direntry, char *data, size_t buf_size) | |||
| 336 | 335 | ||
| 337 | if (pTcon->ses->server->ops->query_all_EAs) | 336 | if (pTcon->ses->server->ops->query_all_EAs) |
| 338 | rc = pTcon->ses->server->ops->query_all_EAs(xid, pTcon, | 337 | rc = pTcon->ses->server->ops->query_all_EAs(xid, pTcon, |
| 339 | full_path, NULL, data, buf_size, | 338 | full_path, NULL, data, buf_size, cifs_sb); |
| 340 | cifs_sb->local_nls, cifs_remap(cifs_sb)); | ||
| 341 | list_ea_exit: | 339 | list_ea_exit: |
| 342 | kfree(full_path); | 340 | kfree(full_path); |
| 343 | free_xid(xid); | 341 | free_xid(xid); |
diff --git a/include/linux/bpf_verifier.h b/include/linux/bpf_verifier.h index 5efb4db44e1e..d5093b52b485 100644 --- a/include/linux/bpf_verifier.h +++ b/include/linux/bpf_verifier.h | |||
| @@ -40,6 +40,9 @@ struct bpf_reg_state { | |||
| 40 | */ | 40 | */ |
| 41 | s64 min_value; | 41 | s64 min_value; |
| 42 | u64 max_value; | 42 | u64 max_value; |
| 43 | u32 min_align; | ||
| 44 | u32 aux_off; | ||
| 45 | u32 aux_off_align; | ||
| 43 | }; | 46 | }; |
| 44 | 47 | ||
| 45 | enum bpf_stack_slot_type { | 48 | enum bpf_stack_slot_type { |
| @@ -87,6 +90,7 @@ struct bpf_verifier_env { | |||
| 87 | struct bpf_prog *prog; /* eBPF program being verified */ | 90 | struct bpf_prog *prog; /* eBPF program being verified */ |
| 88 | struct bpf_verifier_stack_elem *head; /* stack of verifier states to be processed */ | 91 | struct bpf_verifier_stack_elem *head; /* stack of verifier states to be processed */ |
| 89 | int stack_size; /* number of states to be processed */ | 92 | int stack_size; /* number of states to be processed */ |
| 93 | bool strict_alignment; /* perform strict pointer alignment checks */ | ||
| 90 | struct bpf_verifier_state cur_state; /* current verifier state */ | 94 | struct bpf_verifier_state cur_state; /* current verifier state */ |
| 91 | struct bpf_verifier_state_list **explored_states; /* search pruning optimization */ | 95 | struct bpf_verifier_state_list **explored_states; /* search pruning optimization */ |
| 92 | const struct bpf_ext_analyzer_ops *analyzer_ops; /* external analyzer ops */ | 96 | const struct bpf_ext_analyzer_ops *analyzer_ops; /* external analyzer ops */ |
diff --git a/include/linux/mlx5/fs.h b/include/linux/mlx5/fs.h index 1b166d2e19c5..b25e7baa273e 100644 --- a/include/linux/mlx5/fs.h +++ b/include/linux/mlx5/fs.h | |||
| @@ -109,7 +109,6 @@ struct mlx5_flow_table_attr { | |||
| 109 | int max_fte; | 109 | int max_fte; |
| 110 | u32 level; | 110 | u32 level; |
| 111 | u32 flags; | 111 | u32 flags; |
| 112 | u32 underlay_qpn; | ||
| 113 | }; | 112 | }; |
| 114 | 113 | ||
| 115 | struct mlx5_flow_table * | 114 | struct mlx5_flow_table * |
| @@ -167,4 +166,7 @@ struct mlx5_fc *mlx5_fc_create(struct mlx5_core_dev *dev, bool aging); | |||
| 167 | void mlx5_fc_destroy(struct mlx5_core_dev *dev, struct mlx5_fc *counter); | 166 | void mlx5_fc_destroy(struct mlx5_core_dev *dev, struct mlx5_fc *counter); |
| 168 | void mlx5_fc_query_cached(struct mlx5_fc *counter, | 167 | void mlx5_fc_query_cached(struct mlx5_fc *counter, |
| 169 | u64 *bytes, u64 *packets, u64 *lastuse); | 168 | u64 *bytes, u64 *packets, u64 *lastuse); |
| 169 | int mlx5_fs_add_rx_underlay_qpn(struct mlx5_core_dev *dev, u32 underlay_qpn); | ||
| 170 | int mlx5_fs_remove_rx_underlay_qpn(struct mlx5_core_dev *dev, u32 underlay_qpn); | ||
| 171 | |||
| 170 | #endif | 172 | #endif |
diff --git a/include/linux/netdevice.h b/include/linux/netdevice.h index 9c23bd2efb56..3f39d27decf4 100644 --- a/include/linux/netdevice.h +++ b/include/linux/netdevice.h | |||
| @@ -3296,11 +3296,15 @@ int dev_get_phys_port_id(struct net_device *dev, | |||
| 3296 | int dev_get_phys_port_name(struct net_device *dev, | 3296 | int dev_get_phys_port_name(struct net_device *dev, |
| 3297 | char *name, size_t len); | 3297 | char *name, size_t len); |
| 3298 | int dev_change_proto_down(struct net_device *dev, bool proto_down); | 3298 | int dev_change_proto_down(struct net_device *dev, bool proto_down); |
| 3299 | int dev_change_xdp_fd(struct net_device *dev, struct netlink_ext_ack *extack, | ||
| 3300 | int fd, u32 flags); | ||
| 3301 | struct sk_buff *validate_xmit_skb_list(struct sk_buff *skb, struct net_device *dev); | 3299 | struct sk_buff *validate_xmit_skb_list(struct sk_buff *skb, struct net_device *dev); |
| 3302 | struct sk_buff *dev_hard_start_xmit(struct sk_buff *skb, struct net_device *dev, | 3300 | struct sk_buff *dev_hard_start_xmit(struct sk_buff *skb, struct net_device *dev, |
| 3303 | struct netdev_queue *txq, int *ret); | 3301 | struct netdev_queue *txq, int *ret); |
| 3302 | |||
| 3303 | typedef int (*xdp_op_t)(struct net_device *dev, struct netdev_xdp *xdp); | ||
| 3304 | int dev_change_xdp_fd(struct net_device *dev, struct netlink_ext_ack *extack, | ||
| 3305 | int fd, u32 flags); | ||
| 3306 | bool __dev_xdp_attached(struct net_device *dev, xdp_op_t xdp_op); | ||
| 3307 | |||
| 3304 | int __dev_forward_skb(struct net_device *dev, struct sk_buff *skb); | 3308 | int __dev_forward_skb(struct net_device *dev, struct sk_buff *skb); |
| 3305 | int dev_forward_skb(struct net_device *dev, struct sk_buff *skb); | 3309 | int dev_forward_skb(struct net_device *dev, struct sk_buff *skb); |
| 3306 | bool is_skb_forwardable(const struct net_device *dev, | 3310 | bool is_skb_forwardable(const struct net_device *dev, |
diff --git a/include/net/x25.h b/include/net/x25.h index c383aa4edbf0..6d30a01d281d 100644 --- a/include/net/x25.h +++ b/include/net/x25.h | |||
| @@ -298,10 +298,10 @@ void x25_check_rbuf(struct sock *); | |||
| 298 | 298 | ||
| 299 | /* sysctl_net_x25.c */ | 299 | /* sysctl_net_x25.c */ |
| 300 | #ifdef CONFIG_SYSCTL | 300 | #ifdef CONFIG_SYSCTL |
| 301 | void x25_register_sysctl(void); | 301 | int x25_register_sysctl(void); |
| 302 | void x25_unregister_sysctl(void); | 302 | void x25_unregister_sysctl(void); |
| 303 | #else | 303 | #else |
| 304 | static inline void x25_register_sysctl(void) {}; | 304 | static inline int x25_register_sysctl(void) { return 0; }; |
| 305 | static inline void x25_unregister_sysctl(void) {}; | 305 | static inline void x25_unregister_sysctl(void) {}; |
| 306 | #endif /* CONFIG_SYSCTL */ | 306 | #endif /* CONFIG_SYSCTL */ |
| 307 | 307 | ||
diff --git a/include/uapi/linux/bpf.h b/include/uapi/linux/bpf.h index 945a1f5f63c5..94dfa9def355 100644 --- a/include/uapi/linux/bpf.h +++ b/include/uapi/linux/bpf.h | |||
| @@ -132,6 +132,13 @@ enum bpf_attach_type { | |||
| 132 | */ | 132 | */ |
| 133 | #define BPF_F_ALLOW_OVERRIDE (1U << 0) | 133 | #define BPF_F_ALLOW_OVERRIDE (1U << 0) |
| 134 | 134 | ||
| 135 | /* If BPF_F_STRICT_ALIGNMENT is used in BPF_PROG_LOAD command, the | ||
| 136 | * verifier will perform strict alignment checking as if the kernel | ||
| 137 | * has been built with CONFIG_EFFICIENT_UNALIGNED_ACCESS not set, | ||
| 138 | * and NET_IP_ALIGN defined to 2. | ||
| 139 | */ | ||
| 140 | #define BPF_F_STRICT_ALIGNMENT (1U << 0) | ||
| 141 | |||
| 135 | #define BPF_PSEUDO_MAP_FD 1 | 142 | #define BPF_PSEUDO_MAP_FD 1 |
| 136 | 143 | ||
| 137 | /* flags for BPF_MAP_UPDATE_ELEM command */ | 144 | /* flags for BPF_MAP_UPDATE_ELEM command */ |
| @@ -177,6 +184,7 @@ union bpf_attr { | |||
| 177 | __u32 log_size; /* size of user buffer */ | 184 | __u32 log_size; /* size of user buffer */ |
| 178 | __aligned_u64 log_buf; /* user supplied buffer */ | 185 | __aligned_u64 log_buf; /* user supplied buffer */ |
| 179 | __u32 kern_version; /* checked when prog_type=kprobe */ | 186 | __u32 kern_version; /* checked when prog_type=kprobe */ |
| 187 | __u32 prog_flags; | ||
| 180 | }; | 188 | }; |
| 181 | 189 | ||
| 182 | struct { /* anonymous struct used by BPF_OBJ_* commands */ | 190 | struct { /* anonymous struct used by BPF_OBJ_* commands */ |
diff --git a/include/uapi/linux/if_link.h b/include/uapi/linux/if_link.h index 8e56ac70e0d1..15ac20382aba 100644 --- a/include/uapi/linux/if_link.h +++ b/include/uapi/linux/if_link.h | |||
| @@ -888,9 +888,18 @@ enum { | |||
| 888 | /* XDP section */ | 888 | /* XDP section */ |
| 889 | 889 | ||
| 890 | #define XDP_FLAGS_UPDATE_IF_NOEXIST (1U << 0) | 890 | #define XDP_FLAGS_UPDATE_IF_NOEXIST (1U << 0) |
| 891 | #define XDP_FLAGS_SKB_MODE (2U << 0) | 891 | #define XDP_FLAGS_SKB_MODE (1U << 1) |
| 892 | #define XDP_FLAGS_DRV_MODE (1U << 2) | ||
| 892 | #define XDP_FLAGS_MASK (XDP_FLAGS_UPDATE_IF_NOEXIST | \ | 893 | #define XDP_FLAGS_MASK (XDP_FLAGS_UPDATE_IF_NOEXIST | \ |
| 893 | XDP_FLAGS_SKB_MODE) | 894 | XDP_FLAGS_SKB_MODE | \ |
| 895 | XDP_FLAGS_DRV_MODE) | ||
| 896 | |||
| 897 | /* These are stored into IFLA_XDP_ATTACHED on dump. */ | ||
| 898 | enum { | ||
| 899 | XDP_ATTACHED_NONE = 0, | ||
| 900 | XDP_ATTACHED_DRV, | ||
| 901 | XDP_ATTACHED_SKB, | ||
| 902 | }; | ||
| 894 | 903 | ||
| 895 | enum { | 904 | enum { |
| 896 | IFLA_XDP_UNSPEC, | 905 | IFLA_XDP_UNSPEC, |
diff --git a/kernel/bpf/syscall.c b/kernel/bpf/syscall.c index fd2411fd6914..265a0d854e33 100644 --- a/kernel/bpf/syscall.c +++ b/kernel/bpf/syscall.c | |||
| @@ -783,7 +783,7 @@ struct bpf_prog *bpf_prog_get_type(u32 ufd, enum bpf_prog_type type) | |||
| 783 | EXPORT_SYMBOL_GPL(bpf_prog_get_type); | 783 | EXPORT_SYMBOL_GPL(bpf_prog_get_type); |
| 784 | 784 | ||
| 785 | /* last field in 'union bpf_attr' used by this command */ | 785 | /* last field in 'union bpf_attr' used by this command */ |
| 786 | #define BPF_PROG_LOAD_LAST_FIELD kern_version | 786 | #define BPF_PROG_LOAD_LAST_FIELD prog_flags |
| 787 | 787 | ||
| 788 | static int bpf_prog_load(union bpf_attr *attr) | 788 | static int bpf_prog_load(union bpf_attr *attr) |
| 789 | { | 789 | { |
| @@ -796,6 +796,9 @@ static int bpf_prog_load(union bpf_attr *attr) | |||
| 796 | if (CHECK_ATTR(BPF_PROG_LOAD)) | 796 | if (CHECK_ATTR(BPF_PROG_LOAD)) |
| 797 | return -EINVAL; | 797 | return -EINVAL; |
| 798 | 798 | ||
| 799 | if (attr->prog_flags & ~BPF_F_STRICT_ALIGNMENT) | ||
| 800 | return -EINVAL; | ||
| 801 | |||
| 799 | /* copy eBPF program license from user space */ | 802 | /* copy eBPF program license from user space */ |
| 800 | if (strncpy_from_user(license, u64_to_user_ptr(attr->license), | 803 | if (strncpy_from_user(license, u64_to_user_ptr(attr->license), |
| 801 | sizeof(license) - 1) < 0) | 804 | sizeof(license) - 1) < 0) |
diff --git a/kernel/bpf/verifier.c b/kernel/bpf/verifier.c index c5b56c92f8e2..1eddb713b815 100644 --- a/kernel/bpf/verifier.c +++ b/kernel/bpf/verifier.c | |||
| @@ -140,7 +140,7 @@ struct bpf_verifier_stack_elem { | |||
| 140 | struct bpf_verifier_stack_elem *next; | 140 | struct bpf_verifier_stack_elem *next; |
| 141 | }; | 141 | }; |
| 142 | 142 | ||
| 143 | #define BPF_COMPLEXITY_LIMIT_INSNS 65536 | 143 | #define BPF_COMPLEXITY_LIMIT_INSNS 98304 |
| 144 | #define BPF_COMPLEXITY_LIMIT_STACK 1024 | 144 | #define BPF_COMPLEXITY_LIMIT_STACK 1024 |
| 145 | 145 | ||
| 146 | #define BPF_MAP_PTR_POISON ((void *)0xeB9F + POISON_POINTER_DELTA) | 146 | #define BPF_MAP_PTR_POISON ((void *)0xeB9F + POISON_POINTER_DELTA) |
| @@ -241,6 +241,12 @@ static void print_verifier_state(struct bpf_verifier_state *state) | |||
| 241 | if (reg->max_value != BPF_REGISTER_MAX_RANGE) | 241 | if (reg->max_value != BPF_REGISTER_MAX_RANGE) |
| 242 | verbose(",max_value=%llu", | 242 | verbose(",max_value=%llu", |
| 243 | (unsigned long long)reg->max_value); | 243 | (unsigned long long)reg->max_value); |
| 244 | if (reg->min_align) | ||
| 245 | verbose(",min_align=%u", reg->min_align); | ||
| 246 | if (reg->aux_off) | ||
| 247 | verbose(",aux_off=%u", reg->aux_off); | ||
| 248 | if (reg->aux_off_align) | ||
| 249 | verbose(",aux_off_align=%u", reg->aux_off_align); | ||
| 244 | } | 250 | } |
| 245 | for (i = 0; i < MAX_BPF_STACK; i += BPF_REG_SIZE) { | 251 | for (i = 0; i < MAX_BPF_STACK; i += BPF_REG_SIZE) { |
| 246 | if (state->stack_slot_type[i] == STACK_SPILL) | 252 | if (state->stack_slot_type[i] == STACK_SPILL) |
| @@ -466,6 +472,9 @@ static void init_reg_state(struct bpf_reg_state *regs) | |||
| 466 | regs[i].imm = 0; | 472 | regs[i].imm = 0; |
| 467 | regs[i].min_value = BPF_REGISTER_MIN_RANGE; | 473 | regs[i].min_value = BPF_REGISTER_MIN_RANGE; |
| 468 | regs[i].max_value = BPF_REGISTER_MAX_RANGE; | 474 | regs[i].max_value = BPF_REGISTER_MAX_RANGE; |
| 475 | regs[i].min_align = 0; | ||
| 476 | regs[i].aux_off = 0; | ||
| 477 | regs[i].aux_off_align = 0; | ||
| 469 | } | 478 | } |
| 470 | 479 | ||
| 471 | /* frame pointer */ | 480 | /* frame pointer */ |
| @@ -492,6 +501,7 @@ static void reset_reg_range_values(struct bpf_reg_state *regs, u32 regno) | |||
| 492 | { | 501 | { |
| 493 | regs[regno].min_value = BPF_REGISTER_MIN_RANGE; | 502 | regs[regno].min_value = BPF_REGISTER_MIN_RANGE; |
| 494 | regs[regno].max_value = BPF_REGISTER_MAX_RANGE; | 503 | regs[regno].max_value = BPF_REGISTER_MAX_RANGE; |
| 504 | regs[regno].min_align = 0; | ||
| 495 | } | 505 | } |
| 496 | 506 | ||
| 497 | static void mark_reg_unknown_value_and_range(struct bpf_reg_state *regs, | 507 | static void mark_reg_unknown_value_and_range(struct bpf_reg_state *regs, |
| @@ -779,17 +789,33 @@ static bool is_pointer_value(struct bpf_verifier_env *env, int regno) | |||
| 779 | } | 789 | } |
| 780 | 790 | ||
| 781 | static int check_pkt_ptr_alignment(const struct bpf_reg_state *reg, | 791 | static int check_pkt_ptr_alignment(const struct bpf_reg_state *reg, |
| 782 | int off, int size) | 792 | int off, int size, bool strict) |
| 783 | { | 793 | { |
| 784 | if (reg->id && size != 1) { | 794 | int ip_align; |
| 785 | verbose("Unknown alignment. Only byte-sized access allowed in packet access.\n"); | 795 | int reg_off; |
| 786 | return -EACCES; | 796 | |
| 797 | /* Byte size accesses are always allowed. */ | ||
| 798 | if (!strict || size == 1) | ||
| 799 | return 0; | ||
| 800 | |||
| 801 | reg_off = reg->off; | ||
| 802 | if (reg->id) { | ||
| 803 | if (reg->aux_off_align % size) { | ||
| 804 | verbose("Packet access is only %u byte aligned, %d byte access not allowed\n", | ||
| 805 | reg->aux_off_align, size); | ||
| 806 | return -EACCES; | ||
| 807 | } | ||
| 808 | reg_off += reg->aux_off; | ||
| 787 | } | 809 | } |
| 788 | 810 | ||
| 789 | /* skb->data is NET_IP_ALIGN-ed */ | 811 | /* skb->data is NET_IP_ALIGN-ed, but for strict alignment checking |
| 790 | if ((NET_IP_ALIGN + reg->off + off) % size != 0) { | 812 | * we force this to 2 which is universally what architectures use |
| 813 | * when they don't set CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS. | ||
| 814 | */ | ||
| 815 | ip_align = strict ? 2 : NET_IP_ALIGN; | ||
| 816 | if ((ip_align + reg_off + off) % size != 0) { | ||
| 791 | verbose("misaligned packet access off %d+%d+%d size %d\n", | 817 | verbose("misaligned packet access off %d+%d+%d size %d\n", |
| 792 | NET_IP_ALIGN, reg->off, off, size); | 818 | ip_align, reg_off, off, size); |
| 793 | return -EACCES; | 819 | return -EACCES; |
| 794 | } | 820 | } |
| 795 | 821 | ||
| @@ -797,9 +823,9 @@ static int check_pkt_ptr_alignment(const struct bpf_reg_state *reg, | |||
| 797 | } | 823 | } |
| 798 | 824 | ||
| 799 | static int check_val_ptr_alignment(const struct bpf_reg_state *reg, | 825 | static int check_val_ptr_alignment(const struct bpf_reg_state *reg, |
| 800 | int size) | 826 | int size, bool strict) |
| 801 | { | 827 | { |
| 802 | if (size != 1) { | 828 | if (strict && size != 1) { |
| 803 | verbose("Unknown alignment. Only byte-sized access allowed in value access.\n"); | 829 | verbose("Unknown alignment. Only byte-sized access allowed in value access.\n"); |
| 804 | return -EACCES; | 830 | return -EACCES; |
| 805 | } | 831 | } |
| @@ -807,16 +833,20 @@ static int check_val_ptr_alignment(const struct bpf_reg_state *reg, | |||
| 807 | return 0; | 833 | return 0; |
| 808 | } | 834 | } |
| 809 | 835 | ||
| 810 | static int check_ptr_alignment(const struct bpf_reg_state *reg, | 836 | static int check_ptr_alignment(struct bpf_verifier_env *env, |
| 837 | const struct bpf_reg_state *reg, | ||
| 811 | int off, int size) | 838 | int off, int size) |
| 812 | { | 839 | { |
| 840 | bool strict = env->strict_alignment; | ||
| 841 | |||
| 842 | if (!IS_ENABLED(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS)) | ||
| 843 | strict = true; | ||
| 844 | |||
| 813 | switch (reg->type) { | 845 | switch (reg->type) { |
| 814 | case PTR_TO_PACKET: | 846 | case PTR_TO_PACKET: |
| 815 | return IS_ENABLED(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) ? 0 : | 847 | return check_pkt_ptr_alignment(reg, off, size, strict); |
| 816 | check_pkt_ptr_alignment(reg, off, size); | ||
| 817 | case PTR_TO_MAP_VALUE_ADJ: | 848 | case PTR_TO_MAP_VALUE_ADJ: |
| 818 | return IS_ENABLED(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) ? 0 : | 849 | return check_val_ptr_alignment(reg, size, strict); |
| 819 | check_val_ptr_alignment(reg, size); | ||
| 820 | default: | 850 | default: |
| 821 | if (off % size != 0) { | 851 | if (off % size != 0) { |
| 822 | verbose("misaligned access off %d size %d\n", | 852 | verbose("misaligned access off %d size %d\n", |
| @@ -849,7 +879,7 @@ static int check_mem_access(struct bpf_verifier_env *env, u32 regno, int off, | |||
| 849 | if (size < 0) | 879 | if (size < 0) |
| 850 | return size; | 880 | return size; |
| 851 | 881 | ||
| 852 | err = check_ptr_alignment(reg, off, size); | 882 | err = check_ptr_alignment(env, reg, off, size); |
| 853 | if (err) | 883 | if (err) |
| 854 | return err; | 884 | return err; |
| 855 | 885 | ||
| @@ -883,6 +913,8 @@ static int check_mem_access(struct bpf_verifier_env *env, u32 regno, int off, | |||
| 883 | value_regno); | 913 | value_regno); |
| 884 | /* note that reg.[id|off|range] == 0 */ | 914 | /* note that reg.[id|off|range] == 0 */ |
| 885 | state->regs[value_regno].type = reg_type; | 915 | state->regs[value_regno].type = reg_type; |
| 916 | state->regs[value_regno].aux_off = 0; | ||
| 917 | state->regs[value_regno].aux_off_align = 0; | ||
| 886 | } | 918 | } |
| 887 | 919 | ||
| 888 | } else if (reg->type == FRAME_PTR || reg->type == PTR_TO_STACK) { | 920 | } else if (reg->type == FRAME_PTR || reg->type == PTR_TO_STACK) { |
| @@ -1455,6 +1487,8 @@ add_imm: | |||
| 1455 | */ | 1487 | */ |
| 1456 | dst_reg->off += imm; | 1488 | dst_reg->off += imm; |
| 1457 | } else { | 1489 | } else { |
| 1490 | bool had_id; | ||
| 1491 | |||
| 1458 | if (src_reg->type == PTR_TO_PACKET) { | 1492 | if (src_reg->type == PTR_TO_PACKET) { |
| 1459 | /* R6=pkt(id=0,off=0,r=62) R7=imm22; r7 += r6 */ | 1493 | /* R6=pkt(id=0,off=0,r=62) R7=imm22; r7 += r6 */ |
| 1460 | tmp_reg = *dst_reg; /* save r7 state */ | 1494 | tmp_reg = *dst_reg; /* save r7 state */ |
| @@ -1488,14 +1522,23 @@ add_imm: | |||
| 1488 | src_reg->imm); | 1522 | src_reg->imm); |
| 1489 | return -EACCES; | 1523 | return -EACCES; |
| 1490 | } | 1524 | } |
| 1525 | |||
| 1526 | had_id = (dst_reg->id != 0); | ||
| 1527 | |||
| 1491 | /* dst_reg stays as pkt_ptr type and since some positive | 1528 | /* dst_reg stays as pkt_ptr type and since some positive |
| 1492 | * integer value was added to the pointer, increment its 'id' | 1529 | * integer value was added to the pointer, increment its 'id' |
| 1493 | */ | 1530 | */ |
| 1494 | dst_reg->id = ++env->id_gen; | 1531 | dst_reg->id = ++env->id_gen; |
| 1495 | 1532 | ||
| 1496 | /* something was added to pkt_ptr, set range and off to zero */ | 1533 | /* something was added to pkt_ptr, set range to zero */ |
| 1534 | dst_reg->aux_off += dst_reg->off; | ||
| 1497 | dst_reg->off = 0; | 1535 | dst_reg->off = 0; |
| 1498 | dst_reg->range = 0; | 1536 | dst_reg->range = 0; |
| 1537 | if (had_id) | ||
| 1538 | dst_reg->aux_off_align = min(dst_reg->aux_off_align, | ||
| 1539 | src_reg->min_align); | ||
| 1540 | else | ||
| 1541 | dst_reg->aux_off_align = src_reg->min_align; | ||
| 1499 | } | 1542 | } |
| 1500 | return 0; | 1543 | return 0; |
| 1501 | } | 1544 | } |
| @@ -1669,6 +1712,13 @@ static void check_reg_overflow(struct bpf_reg_state *reg) | |||
| 1669 | reg->min_value = BPF_REGISTER_MIN_RANGE; | 1712 | reg->min_value = BPF_REGISTER_MIN_RANGE; |
| 1670 | } | 1713 | } |
| 1671 | 1714 | ||
| 1715 | static u32 calc_align(u32 imm) | ||
| 1716 | { | ||
| 1717 | if (!imm) | ||
| 1718 | return 1U << 31; | ||
| 1719 | return imm - ((imm - 1) & imm); | ||
| 1720 | } | ||
| 1721 | |||
| 1672 | static void adjust_reg_min_max_vals(struct bpf_verifier_env *env, | 1722 | static void adjust_reg_min_max_vals(struct bpf_verifier_env *env, |
| 1673 | struct bpf_insn *insn) | 1723 | struct bpf_insn *insn) |
| 1674 | { | 1724 | { |
| @@ -1676,8 +1726,10 @@ static void adjust_reg_min_max_vals(struct bpf_verifier_env *env, | |||
| 1676 | s64 min_val = BPF_REGISTER_MIN_RANGE; | 1726 | s64 min_val = BPF_REGISTER_MIN_RANGE; |
| 1677 | u64 max_val = BPF_REGISTER_MAX_RANGE; | 1727 | u64 max_val = BPF_REGISTER_MAX_RANGE; |
| 1678 | u8 opcode = BPF_OP(insn->code); | 1728 | u8 opcode = BPF_OP(insn->code); |
| 1729 | u32 dst_align, src_align; | ||
| 1679 | 1730 | ||
| 1680 | dst_reg = ®s[insn->dst_reg]; | 1731 | dst_reg = ®s[insn->dst_reg]; |
| 1732 | src_align = 0; | ||
| 1681 | if (BPF_SRC(insn->code) == BPF_X) { | 1733 | if (BPF_SRC(insn->code) == BPF_X) { |
| 1682 | check_reg_overflow(®s[insn->src_reg]); | 1734 | check_reg_overflow(®s[insn->src_reg]); |
| 1683 | min_val = regs[insn->src_reg].min_value; | 1735 | min_val = regs[insn->src_reg].min_value; |
| @@ -1693,12 +1745,18 @@ static void adjust_reg_min_max_vals(struct bpf_verifier_env *env, | |||
| 1693 | regs[insn->src_reg].type != UNKNOWN_VALUE) { | 1745 | regs[insn->src_reg].type != UNKNOWN_VALUE) { |
| 1694 | min_val = BPF_REGISTER_MIN_RANGE; | 1746 | min_val = BPF_REGISTER_MIN_RANGE; |
| 1695 | max_val = BPF_REGISTER_MAX_RANGE; | 1747 | max_val = BPF_REGISTER_MAX_RANGE; |
| 1748 | src_align = 0; | ||
| 1749 | } else { | ||
| 1750 | src_align = regs[insn->src_reg].min_align; | ||
| 1696 | } | 1751 | } |
| 1697 | } else if (insn->imm < BPF_REGISTER_MAX_RANGE && | 1752 | } else if (insn->imm < BPF_REGISTER_MAX_RANGE && |
| 1698 | (s64)insn->imm > BPF_REGISTER_MIN_RANGE) { | 1753 | (s64)insn->imm > BPF_REGISTER_MIN_RANGE) { |
| 1699 | min_val = max_val = insn->imm; | 1754 | min_val = max_val = insn->imm; |
| 1755 | src_align = calc_align(insn->imm); | ||
| 1700 | } | 1756 | } |
| 1701 | 1757 | ||
| 1758 | dst_align = dst_reg->min_align; | ||
| 1759 | |||
| 1702 | /* We don't know anything about what was done to this register, mark it | 1760 | /* We don't know anything about what was done to this register, mark it |
| 1703 | * as unknown. | 1761 | * as unknown. |
| 1704 | */ | 1762 | */ |
| @@ -1723,18 +1781,21 @@ static void adjust_reg_min_max_vals(struct bpf_verifier_env *env, | |||
| 1723 | dst_reg->min_value += min_val; | 1781 | dst_reg->min_value += min_val; |
| 1724 | if (dst_reg->max_value != BPF_REGISTER_MAX_RANGE) | 1782 | if (dst_reg->max_value != BPF_REGISTER_MAX_RANGE) |
| 1725 | dst_reg->max_value += max_val; | 1783 | dst_reg->max_value += max_val; |
| 1784 | dst_reg->min_align = min(src_align, dst_align); | ||
| 1726 | break; | 1785 | break; |
| 1727 | case BPF_SUB: | 1786 | case BPF_SUB: |
| 1728 | if (dst_reg->min_value != BPF_REGISTER_MIN_RANGE) | 1787 | if (dst_reg->min_value != BPF_REGISTER_MIN_RANGE) |
| 1729 | dst_reg->min_value -= min_val; | 1788 | dst_reg->min_value -= min_val; |
| 1730 | if (dst_reg->max_value != BPF_REGISTER_MAX_RANGE) | 1789 | if (dst_reg->max_value != BPF_REGISTER_MAX_RANGE) |
| 1731 | dst_reg->max_value -= max_val; | 1790 | dst_reg->max_value -= max_val; |
| 1791 | dst_reg->min_align = min(src_align, dst_align); | ||
| 1732 | break; | 1792 | break; |
| 1733 | case BPF_MUL: | 1793 | case BPF_MUL: |
| 1734 | if (dst_reg->min_value != BPF_REGISTER_MIN_RANGE) | 1794 | if (dst_reg->min_value != BPF_REGISTER_MIN_RANGE) |
| 1735 | dst_reg->min_value *= min_val; | 1795 | dst_reg->min_value *= min_val; |
| 1736 | if (dst_reg->max_value != BPF_REGISTER_MAX_RANGE) | 1796 | if (dst_reg->max_value != BPF_REGISTER_MAX_RANGE) |
| 1737 | dst_reg->max_value *= max_val; | 1797 | dst_reg->max_value *= max_val; |
| 1798 | dst_reg->min_align = max(src_align, dst_align); | ||
| 1738 | break; | 1799 | break; |
| 1739 | case BPF_AND: | 1800 | case BPF_AND: |
| 1740 | /* Disallow AND'ing of negative numbers, ain't nobody got time | 1801 | /* Disallow AND'ing of negative numbers, ain't nobody got time |
| @@ -1746,17 +1807,23 @@ static void adjust_reg_min_max_vals(struct bpf_verifier_env *env, | |||
| 1746 | else | 1807 | else |
| 1747 | dst_reg->min_value = 0; | 1808 | dst_reg->min_value = 0; |
| 1748 | dst_reg->max_value = max_val; | 1809 | dst_reg->max_value = max_val; |
| 1810 | dst_reg->min_align = max(src_align, dst_align); | ||
| 1749 | break; | 1811 | break; |
| 1750 | case BPF_LSH: | 1812 | case BPF_LSH: |
| 1751 | /* Gotta have special overflow logic here, if we're shifting | 1813 | /* Gotta have special overflow logic here, if we're shifting |
| 1752 | * more than MAX_RANGE then just assume we have an invalid | 1814 | * more than MAX_RANGE then just assume we have an invalid |
| 1753 | * range. | 1815 | * range. |
| 1754 | */ | 1816 | */ |
| 1755 | if (min_val > ilog2(BPF_REGISTER_MAX_RANGE)) | 1817 | if (min_val > ilog2(BPF_REGISTER_MAX_RANGE)) { |
| 1756 | dst_reg->min_value = BPF_REGISTER_MIN_RANGE; | 1818 | dst_reg->min_value = BPF_REGISTER_MIN_RANGE; |
| 1757 | else if (dst_reg->min_value != BPF_REGISTER_MIN_RANGE) | 1819 | dst_reg->min_align = 1; |
| 1758 | dst_reg->min_value <<= min_val; | 1820 | } else { |
| 1759 | 1821 | if (dst_reg->min_value != BPF_REGISTER_MIN_RANGE) | |
| 1822 | dst_reg->min_value <<= min_val; | ||
| 1823 | if (!dst_reg->min_align) | ||
| 1824 | dst_reg->min_align = 1; | ||
| 1825 | dst_reg->min_align <<= min_val; | ||
| 1826 | } | ||
| 1760 | if (max_val > ilog2(BPF_REGISTER_MAX_RANGE)) | 1827 | if (max_val > ilog2(BPF_REGISTER_MAX_RANGE)) |
| 1761 | dst_reg->max_value = BPF_REGISTER_MAX_RANGE; | 1828 | dst_reg->max_value = BPF_REGISTER_MAX_RANGE; |
| 1762 | else if (dst_reg->max_value != BPF_REGISTER_MAX_RANGE) | 1829 | else if (dst_reg->max_value != BPF_REGISTER_MAX_RANGE) |
| @@ -1766,11 +1833,19 @@ static void adjust_reg_min_max_vals(struct bpf_verifier_env *env, | |||
| 1766 | /* RSH by a negative number is undefined, and the BPF_RSH is an | 1833 | /* RSH by a negative number is undefined, and the BPF_RSH is an |
| 1767 | * unsigned shift, so make the appropriate casts. | 1834 | * unsigned shift, so make the appropriate casts. |
| 1768 | */ | 1835 | */ |
| 1769 | if (min_val < 0 || dst_reg->min_value < 0) | 1836 | if (min_val < 0 || dst_reg->min_value < 0) { |
| 1770 | dst_reg->min_value = BPF_REGISTER_MIN_RANGE; | 1837 | dst_reg->min_value = BPF_REGISTER_MIN_RANGE; |
| 1771 | else | 1838 | } else { |
| 1772 | dst_reg->min_value = | 1839 | dst_reg->min_value = |
| 1773 | (u64)(dst_reg->min_value) >> min_val; | 1840 | (u64)(dst_reg->min_value) >> min_val; |
| 1841 | } | ||
| 1842 | if (min_val < 0) { | ||
| 1843 | dst_reg->min_align = 1; | ||
| 1844 | } else { | ||
| 1845 | dst_reg->min_align >>= (u64) min_val; | ||
| 1846 | if (!dst_reg->min_align) | ||
| 1847 | dst_reg->min_align = 1; | ||
| 1848 | } | ||
| 1774 | if (dst_reg->max_value != BPF_REGISTER_MAX_RANGE) | 1849 | if (dst_reg->max_value != BPF_REGISTER_MAX_RANGE) |
| 1775 | dst_reg->max_value >>= max_val; | 1850 | dst_reg->max_value >>= max_val; |
| 1776 | break; | 1851 | break; |
| @@ -1872,6 +1947,7 @@ static int check_alu_op(struct bpf_verifier_env *env, struct bpf_insn *insn) | |||
| 1872 | regs[insn->dst_reg].imm = insn->imm; | 1947 | regs[insn->dst_reg].imm = insn->imm; |
| 1873 | regs[insn->dst_reg].max_value = insn->imm; | 1948 | regs[insn->dst_reg].max_value = insn->imm; |
| 1874 | regs[insn->dst_reg].min_value = insn->imm; | 1949 | regs[insn->dst_reg].min_value = insn->imm; |
| 1950 | regs[insn->dst_reg].min_align = calc_align(insn->imm); | ||
| 1875 | } | 1951 | } |
| 1876 | 1952 | ||
| 1877 | } else if (opcode > BPF_END) { | 1953 | } else if (opcode > BPF_END) { |
| @@ -2564,6 +2640,7 @@ peek_stack: | |||
| 2564 | env->explored_states[t + 1] = STATE_LIST_MARK; | 2640 | env->explored_states[t + 1] = STATE_LIST_MARK; |
| 2565 | } else { | 2641 | } else { |
| 2566 | /* conditional jump with two edges */ | 2642 | /* conditional jump with two edges */ |
| 2643 | env->explored_states[t] = STATE_LIST_MARK; | ||
| 2567 | ret = push_insn(t, t + 1, FALLTHROUGH, env); | 2644 | ret = push_insn(t, t + 1, FALLTHROUGH, env); |
| 2568 | if (ret == 1) | 2645 | if (ret == 1) |
| 2569 | goto peek_stack; | 2646 | goto peek_stack; |
| @@ -2722,6 +2799,12 @@ static bool states_equal(struct bpf_verifier_env *env, | |||
| 2722 | rcur->type != NOT_INIT)) | 2799 | rcur->type != NOT_INIT)) |
| 2723 | continue; | 2800 | continue; |
| 2724 | 2801 | ||
| 2802 | /* Don't care about the reg->id in this case. */ | ||
| 2803 | if (rold->type == PTR_TO_MAP_VALUE_OR_NULL && | ||
| 2804 | rcur->type == PTR_TO_MAP_VALUE_OR_NULL && | ||
| 2805 | rold->map_ptr == rcur->map_ptr) | ||
| 2806 | continue; | ||
| 2807 | |||
| 2725 | if (rold->type == PTR_TO_PACKET && rcur->type == PTR_TO_PACKET && | 2808 | if (rold->type == PTR_TO_PACKET && rcur->type == PTR_TO_PACKET && |
| 2726 | compare_ptrs_to_packet(rold, rcur)) | 2809 | compare_ptrs_to_packet(rold, rcur)) |
| 2727 | continue; | 2810 | continue; |
| @@ -2856,8 +2939,15 @@ static int do_check(struct bpf_verifier_env *env) | |||
| 2856 | goto process_bpf_exit; | 2939 | goto process_bpf_exit; |
| 2857 | } | 2940 | } |
| 2858 | 2941 | ||
| 2859 | if (log_level && do_print_state) { | 2942 | if (need_resched()) |
| 2860 | verbose("\nfrom %d to %d:", prev_insn_idx, insn_idx); | 2943 | cond_resched(); |
| 2944 | |||
| 2945 | if (log_level > 1 || (log_level && do_print_state)) { | ||
| 2946 | if (log_level > 1) | ||
| 2947 | verbose("%d:", insn_idx); | ||
| 2948 | else | ||
| 2949 | verbose("\nfrom %d to %d:", | ||
| 2950 | prev_insn_idx, insn_idx); | ||
| 2861 | print_verifier_state(&env->cur_state); | 2951 | print_verifier_state(&env->cur_state); |
| 2862 | do_print_state = false; | 2952 | do_print_state = false; |
| 2863 | } | 2953 | } |
| @@ -3494,6 +3584,10 @@ int bpf_check(struct bpf_prog **prog, union bpf_attr *attr) | |||
| 3494 | } else { | 3584 | } else { |
| 3495 | log_level = 0; | 3585 | log_level = 0; |
| 3496 | } | 3586 | } |
| 3587 | if (attr->prog_flags & BPF_F_STRICT_ALIGNMENT) | ||
| 3588 | env->strict_alignment = true; | ||
| 3589 | else | ||
| 3590 | env->strict_alignment = false; | ||
| 3497 | 3591 | ||
| 3498 | ret = replace_map_fd_with_map_ptr(env); | 3592 | ret = replace_map_fd_with_map_ptr(env); |
| 3499 | if (ret < 0) | 3593 | if (ret < 0) |
| @@ -3599,6 +3693,7 @@ int bpf_analyzer(struct bpf_prog *prog, const struct bpf_ext_analyzer_ops *ops, | |||
| 3599 | mutex_lock(&bpf_verifier_lock); | 3693 | mutex_lock(&bpf_verifier_lock); |
| 3600 | 3694 | ||
| 3601 | log_level = 0; | 3695 | log_level = 0; |
| 3696 | env->strict_alignment = false; | ||
| 3602 | 3697 | ||
| 3603 | env->explored_states = kcalloc(env->prog->len, | 3698 | env->explored_states = kcalloc(env->prog->len, |
| 3604 | sizeof(struct bpf_verifier_state_list *), | 3699 | sizeof(struct bpf_verifier_state_list *), |
diff --git a/kernel/fork.c b/kernel/fork.c index 06d759ab4c62..aa1076c5e4a9 100644 --- a/kernel/fork.c +++ b/kernel/fork.c | |||
| @@ -1845,11 +1845,13 @@ static __latent_entropy struct task_struct *copy_process( | |||
| 1845 | */ | 1845 | */ |
| 1846 | recalc_sigpending(); | 1846 | recalc_sigpending(); |
| 1847 | if (signal_pending(current)) { | 1847 | if (signal_pending(current)) { |
| 1848 | spin_unlock(¤t->sighand->siglock); | ||
| 1849 | write_unlock_irq(&tasklist_lock); | ||
| 1850 | retval = -ERESTARTNOINTR; | 1848 | retval = -ERESTARTNOINTR; |
| 1851 | goto bad_fork_cancel_cgroup; | 1849 | goto bad_fork_cancel_cgroup; |
| 1852 | } | 1850 | } |
| 1851 | if (unlikely(!(ns_of_pid(pid)->nr_hashed & PIDNS_HASH_ADDING))) { | ||
| 1852 | retval = -ENOMEM; | ||
| 1853 | goto bad_fork_cancel_cgroup; | ||
| 1854 | } | ||
| 1853 | 1855 | ||
| 1854 | if (likely(p->pid)) { | 1856 | if (likely(p->pid)) { |
| 1855 | ptrace_init_task(p, (clone_flags & CLONE_PTRACE) || trace); | 1857 | ptrace_init_task(p, (clone_flags & CLONE_PTRACE) || trace); |
| @@ -1907,6 +1909,8 @@ static __latent_entropy struct task_struct *copy_process( | |||
| 1907 | return p; | 1909 | return p; |
| 1908 | 1910 | ||
| 1909 | bad_fork_cancel_cgroup: | 1911 | bad_fork_cancel_cgroup: |
| 1912 | spin_unlock(¤t->sighand->siglock); | ||
| 1913 | write_unlock_irq(&tasklist_lock); | ||
| 1910 | cgroup_cancel_fork(p); | 1914 | cgroup_cancel_fork(p); |
| 1911 | bad_fork_free_pid: | 1915 | bad_fork_free_pid: |
| 1912 | cgroup_threadgroup_change_end(current); | 1916 | cgroup_threadgroup_change_end(current); |
diff --git a/kernel/pid_namespace.c b/kernel/pid_namespace.c index d1f3e9f558b8..74a5a7255b4d 100644 --- a/kernel/pid_namespace.c +++ b/kernel/pid_namespace.c | |||
| @@ -277,7 +277,7 @@ void zap_pid_ns_processes(struct pid_namespace *pid_ns) | |||
| 277 | * if reparented. | 277 | * if reparented. |
| 278 | */ | 278 | */ |
| 279 | for (;;) { | 279 | for (;;) { |
| 280 | set_current_state(TASK_UNINTERRUPTIBLE); | 280 | set_current_state(TASK_INTERRUPTIBLE); |
| 281 | if (pid_ns->nr_hashed == init_pids) | 281 | if (pid_ns->nr_hashed == init_pids) |
| 282 | break; | 282 | break; |
| 283 | schedule(); | 283 | schedule(); |
diff --git a/net/bridge/br_netlink.c b/net/bridge/br_netlink.c index c5ce7745b230..574f78824d8a 100644 --- a/net/bridge/br_netlink.c +++ b/net/bridge/br_netlink.c | |||
| @@ -835,6 +835,13 @@ static int br_validate(struct nlattr *tb[], struct nlattr *data[]) | |||
| 835 | return -EPROTONOSUPPORT; | 835 | return -EPROTONOSUPPORT; |
| 836 | } | 836 | } |
| 837 | } | 837 | } |
| 838 | |||
| 839 | if (data[IFLA_BR_VLAN_DEFAULT_PVID]) { | ||
| 840 | __u16 defpvid = nla_get_u16(data[IFLA_BR_VLAN_DEFAULT_PVID]); | ||
| 841 | |||
| 842 | if (defpvid >= VLAN_VID_MASK) | ||
| 843 | return -EINVAL; | ||
| 844 | } | ||
| 838 | #endif | 845 | #endif |
| 839 | 846 | ||
| 840 | return 0; | 847 | return 0; |
diff --git a/net/core/dev.c b/net/core/dev.c index 96cf83da0d66..fca407b4a6ea 100644 --- a/net/core/dev.c +++ b/net/core/dev.c | |||
| @@ -6852,6 +6852,32 @@ int dev_change_proto_down(struct net_device *dev, bool proto_down) | |||
| 6852 | } | 6852 | } |
| 6853 | EXPORT_SYMBOL(dev_change_proto_down); | 6853 | EXPORT_SYMBOL(dev_change_proto_down); |
| 6854 | 6854 | ||
| 6855 | bool __dev_xdp_attached(struct net_device *dev, xdp_op_t xdp_op) | ||
| 6856 | { | ||
| 6857 | struct netdev_xdp xdp; | ||
| 6858 | |||
| 6859 | memset(&xdp, 0, sizeof(xdp)); | ||
| 6860 | xdp.command = XDP_QUERY_PROG; | ||
| 6861 | |||
| 6862 | /* Query must always succeed. */ | ||
| 6863 | WARN_ON(xdp_op(dev, &xdp) < 0); | ||
| 6864 | return xdp.prog_attached; | ||
| 6865 | } | ||
| 6866 | |||
| 6867 | static int dev_xdp_install(struct net_device *dev, xdp_op_t xdp_op, | ||
| 6868 | struct netlink_ext_ack *extack, | ||
| 6869 | struct bpf_prog *prog) | ||
| 6870 | { | ||
| 6871 | struct netdev_xdp xdp; | ||
| 6872 | |||
| 6873 | memset(&xdp, 0, sizeof(xdp)); | ||
| 6874 | xdp.command = XDP_SETUP_PROG; | ||
| 6875 | xdp.extack = extack; | ||
| 6876 | xdp.prog = prog; | ||
| 6877 | |||
| 6878 | return xdp_op(dev, &xdp); | ||
| 6879 | } | ||
| 6880 | |||
| 6855 | /** | 6881 | /** |
| 6856 | * dev_change_xdp_fd - set or clear a bpf program for a device rx path | 6882 | * dev_change_xdp_fd - set or clear a bpf program for a device rx path |
| 6857 | * @dev: device | 6883 | * @dev: device |
| @@ -6864,41 +6890,34 @@ EXPORT_SYMBOL(dev_change_proto_down); | |||
| 6864 | int dev_change_xdp_fd(struct net_device *dev, struct netlink_ext_ack *extack, | 6890 | int dev_change_xdp_fd(struct net_device *dev, struct netlink_ext_ack *extack, |
| 6865 | int fd, u32 flags) | 6891 | int fd, u32 flags) |
| 6866 | { | 6892 | { |
| 6867 | int (*xdp_op)(struct net_device *dev, struct netdev_xdp *xdp); | ||
| 6868 | const struct net_device_ops *ops = dev->netdev_ops; | 6893 | const struct net_device_ops *ops = dev->netdev_ops; |
| 6869 | struct bpf_prog *prog = NULL; | 6894 | struct bpf_prog *prog = NULL; |
| 6870 | struct netdev_xdp xdp; | 6895 | xdp_op_t xdp_op, xdp_chk; |
| 6871 | int err; | 6896 | int err; |
| 6872 | 6897 | ||
| 6873 | ASSERT_RTNL(); | 6898 | ASSERT_RTNL(); |
| 6874 | 6899 | ||
| 6875 | xdp_op = ops->ndo_xdp; | 6900 | xdp_op = xdp_chk = ops->ndo_xdp; |
| 6901 | if (!xdp_op && (flags & XDP_FLAGS_DRV_MODE)) | ||
| 6902 | return -EOPNOTSUPP; | ||
| 6876 | if (!xdp_op || (flags & XDP_FLAGS_SKB_MODE)) | 6903 | if (!xdp_op || (flags & XDP_FLAGS_SKB_MODE)) |
| 6877 | xdp_op = generic_xdp_install; | 6904 | xdp_op = generic_xdp_install; |
| 6905 | if (xdp_op == xdp_chk) | ||
| 6906 | xdp_chk = generic_xdp_install; | ||
| 6878 | 6907 | ||
| 6879 | if (fd >= 0) { | 6908 | if (fd >= 0) { |
| 6880 | if (flags & XDP_FLAGS_UPDATE_IF_NOEXIST) { | 6909 | if (xdp_chk && __dev_xdp_attached(dev, xdp_chk)) |
| 6881 | memset(&xdp, 0, sizeof(xdp)); | 6910 | return -EEXIST; |
| 6882 | xdp.command = XDP_QUERY_PROG; | 6911 | if ((flags & XDP_FLAGS_UPDATE_IF_NOEXIST) && |
| 6883 | 6912 | __dev_xdp_attached(dev, xdp_op)) | |
| 6884 | err = xdp_op(dev, &xdp); | 6913 | return -EBUSY; |
| 6885 | if (err < 0) | ||
| 6886 | return err; | ||
| 6887 | if (xdp.prog_attached) | ||
| 6888 | return -EBUSY; | ||
| 6889 | } | ||
| 6890 | 6914 | ||
| 6891 | prog = bpf_prog_get_type(fd, BPF_PROG_TYPE_XDP); | 6915 | prog = bpf_prog_get_type(fd, BPF_PROG_TYPE_XDP); |
| 6892 | if (IS_ERR(prog)) | 6916 | if (IS_ERR(prog)) |
| 6893 | return PTR_ERR(prog); | 6917 | return PTR_ERR(prog); |
| 6894 | } | 6918 | } |
| 6895 | 6919 | ||
| 6896 | memset(&xdp, 0, sizeof(xdp)); | 6920 | err = dev_xdp_install(dev, xdp_op, extack, prog); |
| 6897 | xdp.command = XDP_SETUP_PROG; | ||
| 6898 | xdp.extack = extack; | ||
| 6899 | xdp.prog = prog; | ||
| 6900 | |||
| 6901 | err = xdp_op(dev, &xdp); | ||
| 6902 | if (err < 0 && prog) | 6921 | if (err < 0 && prog) |
| 6903 | bpf_prog_put(prog); | 6922 | bpf_prog_put(prog); |
| 6904 | 6923 | ||
diff --git a/net/core/neighbour.c b/net/core/neighbour.c index 58b0bcc125b5..d274f81fcc2c 100644 --- a/net/core/neighbour.c +++ b/net/core/neighbour.c | |||
| @@ -1132,10 +1132,6 @@ int neigh_update(struct neighbour *neigh, const u8 *lladdr, u8 new, | |||
| 1132 | lladdr = neigh->ha; | 1132 | lladdr = neigh->ha; |
| 1133 | } | 1133 | } |
| 1134 | 1134 | ||
| 1135 | if (new & NUD_CONNECTED) | ||
| 1136 | neigh->confirmed = jiffies; | ||
| 1137 | neigh->updated = jiffies; | ||
| 1138 | |||
| 1139 | /* If entry was valid and address is not changed, | 1135 | /* If entry was valid and address is not changed, |
| 1140 | do not change entry state, if new one is STALE. | 1136 | do not change entry state, if new one is STALE. |
| 1141 | */ | 1137 | */ |
| @@ -1157,6 +1153,16 @@ int neigh_update(struct neighbour *neigh, const u8 *lladdr, u8 new, | |||
| 1157 | } | 1153 | } |
| 1158 | } | 1154 | } |
| 1159 | 1155 | ||
| 1156 | /* Update timestamps only once we know we will make a change to the | ||
| 1157 | * neighbour entry. Otherwise we risk to move the locktime window with | ||
| 1158 | * noop updates and ignore relevant ARP updates. | ||
| 1159 | */ | ||
| 1160 | if (new != old || lladdr != neigh->ha) { | ||
| 1161 | if (new & NUD_CONNECTED) | ||
| 1162 | neigh->confirmed = jiffies; | ||
| 1163 | neigh->updated = jiffies; | ||
| 1164 | } | ||
| 1165 | |||
| 1160 | if (new != old) { | 1166 | if (new != old) { |
| 1161 | neigh_del_timer(neigh); | 1167 | neigh_del_timer(neigh); |
| 1162 | if (new & NUD_PROBE) | 1168 | if (new & NUD_PROBE) |
diff --git a/net/core/rtnetlink.c b/net/core/rtnetlink.c index bcb0f610ee42..49a279a7cc15 100644 --- a/net/core/rtnetlink.c +++ b/net/core/rtnetlink.c | |||
| @@ -899,8 +899,7 @@ static size_t rtnl_port_size(const struct net_device *dev, | |||
| 899 | static size_t rtnl_xdp_size(void) | 899 | static size_t rtnl_xdp_size(void) |
| 900 | { | 900 | { |
| 901 | size_t xdp_size = nla_total_size(0) + /* nest IFLA_XDP */ | 901 | size_t xdp_size = nla_total_size(0) + /* nest IFLA_XDP */ |
| 902 | nla_total_size(1) + /* XDP_ATTACHED */ | 902 | nla_total_size(1); /* XDP_ATTACHED */ |
| 903 | nla_total_size(4); /* XDP_FLAGS */ | ||
| 904 | 903 | ||
| 905 | return xdp_size; | 904 | return xdp_size; |
| 906 | } | 905 | } |
| @@ -1247,37 +1246,34 @@ static int rtnl_fill_link_ifmap(struct sk_buff *skb, struct net_device *dev) | |||
| 1247 | return 0; | 1246 | return 0; |
| 1248 | } | 1247 | } |
| 1249 | 1248 | ||
| 1249 | static u8 rtnl_xdp_attached_mode(struct net_device *dev) | ||
| 1250 | { | ||
| 1251 | const struct net_device_ops *ops = dev->netdev_ops; | ||
| 1252 | |||
| 1253 | ASSERT_RTNL(); | ||
| 1254 | |||
| 1255 | if (rcu_access_pointer(dev->xdp_prog)) | ||
| 1256 | return XDP_ATTACHED_SKB; | ||
| 1257 | if (ops->ndo_xdp && __dev_xdp_attached(dev, ops->ndo_xdp)) | ||
| 1258 | return XDP_ATTACHED_DRV; | ||
| 1259 | |||
| 1260 | return XDP_ATTACHED_NONE; | ||
| 1261 | } | ||
| 1262 | |||
| 1250 | static int rtnl_xdp_fill(struct sk_buff *skb, struct net_device *dev) | 1263 | static int rtnl_xdp_fill(struct sk_buff *skb, struct net_device *dev) |
| 1251 | { | 1264 | { |
| 1252 | struct nlattr *xdp; | 1265 | struct nlattr *xdp; |
| 1253 | u32 xdp_flags = 0; | ||
| 1254 | u8 val = 0; | ||
| 1255 | int err; | 1266 | int err; |
| 1256 | 1267 | ||
| 1257 | xdp = nla_nest_start(skb, IFLA_XDP); | 1268 | xdp = nla_nest_start(skb, IFLA_XDP); |
| 1258 | if (!xdp) | 1269 | if (!xdp) |
| 1259 | return -EMSGSIZE; | 1270 | return -EMSGSIZE; |
| 1260 | if (rcu_access_pointer(dev->xdp_prog)) { | 1271 | |
| 1261 | xdp_flags = XDP_FLAGS_SKB_MODE; | 1272 | err = nla_put_u8(skb, IFLA_XDP_ATTACHED, |
| 1262 | val = 1; | 1273 | rtnl_xdp_attached_mode(dev)); |
| 1263 | } else if (dev->netdev_ops->ndo_xdp) { | ||
| 1264 | struct netdev_xdp xdp_op = {}; | ||
| 1265 | |||
| 1266 | xdp_op.command = XDP_QUERY_PROG; | ||
| 1267 | err = dev->netdev_ops->ndo_xdp(dev, &xdp_op); | ||
| 1268 | if (err) | ||
| 1269 | goto err_cancel; | ||
| 1270 | val = xdp_op.prog_attached; | ||
| 1271 | } | ||
| 1272 | err = nla_put_u8(skb, IFLA_XDP_ATTACHED, val); | ||
| 1273 | if (err) | 1274 | if (err) |
| 1274 | goto err_cancel; | 1275 | goto err_cancel; |
| 1275 | 1276 | ||
| 1276 | if (xdp_flags) { | ||
| 1277 | err = nla_put_u32(skb, IFLA_XDP_FLAGS, xdp_flags); | ||
| 1278 | if (err) | ||
| 1279 | goto err_cancel; | ||
| 1280 | } | ||
| 1281 | nla_nest_end(skb, xdp); | 1277 | nla_nest_end(skb, xdp); |
| 1282 | return 0; | 1278 | return 0; |
| 1283 | 1279 | ||
| @@ -1631,13 +1627,13 @@ static int rtnl_dump_ifinfo(struct sk_buff *skb, struct netlink_callback *cb) | |||
| 1631 | cb->nlh->nlmsg_seq, 0, | 1627 | cb->nlh->nlmsg_seq, 0, |
| 1632 | flags, | 1628 | flags, |
| 1633 | ext_filter_mask); | 1629 | ext_filter_mask); |
| 1634 | /* If we ran out of room on the first message, | ||
| 1635 | * we're in trouble | ||
| 1636 | */ | ||
| 1637 | WARN_ON((err == -EMSGSIZE) && (skb->len == 0)); | ||
| 1638 | 1630 | ||
| 1639 | if (err < 0) | 1631 | if (err < 0) { |
| 1640 | goto out; | 1632 | if (likely(skb->len)) |
| 1633 | goto out; | ||
| 1634 | |||
| 1635 | goto out_err; | ||
| 1636 | } | ||
| 1641 | 1637 | ||
| 1642 | nl_dump_check_consistent(cb, nlmsg_hdr(skb)); | 1638 | nl_dump_check_consistent(cb, nlmsg_hdr(skb)); |
| 1643 | cont: | 1639 | cont: |
| @@ -1645,10 +1641,12 @@ cont: | |||
| 1645 | } | 1641 | } |
| 1646 | } | 1642 | } |
| 1647 | out: | 1643 | out: |
| 1644 | err = skb->len; | ||
| 1645 | out_err: | ||
| 1648 | cb->args[1] = idx; | 1646 | cb->args[1] = idx; |
| 1649 | cb->args[0] = h; | 1647 | cb->args[0] = h; |
| 1650 | 1648 | ||
| 1651 | return skb->len; | 1649 | return err; |
| 1652 | } | 1650 | } |
| 1653 | 1651 | ||
| 1654 | int rtnl_nla_parse_ifla(struct nlattr **tb, const struct nlattr *head, int len, | 1652 | int rtnl_nla_parse_ifla(struct nlattr **tb, const struct nlattr *head, int len, |
| @@ -2199,6 +2197,11 @@ static int do_setlink(const struct sk_buff *skb, | |||
| 2199 | err = -EINVAL; | 2197 | err = -EINVAL; |
| 2200 | goto errout; | 2198 | goto errout; |
| 2201 | } | 2199 | } |
| 2200 | if ((xdp_flags & XDP_FLAGS_SKB_MODE) && | ||
| 2201 | (xdp_flags & XDP_FLAGS_DRV_MODE)) { | ||
| 2202 | err = -EINVAL; | ||
| 2203 | goto errout; | ||
| 2204 | } | ||
| 2202 | } | 2205 | } |
| 2203 | 2206 | ||
| 2204 | if (xdp[IFLA_XDP_FD]) { | 2207 | if (xdp[IFLA_XDP_FD]) { |
| @@ -3452,8 +3455,12 @@ static int rtnl_bridge_getlink(struct sk_buff *skb, struct netlink_callback *cb) | |||
| 3452 | err = br_dev->netdev_ops->ndo_bridge_getlink( | 3455 | err = br_dev->netdev_ops->ndo_bridge_getlink( |
| 3453 | skb, portid, seq, dev, | 3456 | skb, portid, seq, dev, |
| 3454 | filter_mask, NLM_F_MULTI); | 3457 | filter_mask, NLM_F_MULTI); |
| 3455 | if (err < 0 && err != -EOPNOTSUPP) | 3458 | if (err < 0 && err != -EOPNOTSUPP) { |
| 3456 | break; | 3459 | if (likely(skb->len)) |
| 3460 | break; | ||
| 3461 | |||
| 3462 | goto out_err; | ||
| 3463 | } | ||
| 3457 | } | 3464 | } |
| 3458 | idx++; | 3465 | idx++; |
| 3459 | } | 3466 | } |
| @@ -3464,16 +3471,22 @@ static int rtnl_bridge_getlink(struct sk_buff *skb, struct netlink_callback *cb) | |||
| 3464 | seq, dev, | 3471 | seq, dev, |
| 3465 | filter_mask, | 3472 | filter_mask, |
| 3466 | NLM_F_MULTI); | 3473 | NLM_F_MULTI); |
| 3467 | if (err < 0 && err != -EOPNOTSUPP) | 3474 | if (err < 0 && err != -EOPNOTSUPP) { |
| 3468 | break; | 3475 | if (likely(skb->len)) |
| 3476 | break; | ||
| 3477 | |||
| 3478 | goto out_err; | ||
| 3479 | } | ||
| 3469 | } | 3480 | } |
| 3470 | idx++; | 3481 | idx++; |
| 3471 | } | 3482 | } |
| 3472 | } | 3483 | } |
| 3484 | err = skb->len; | ||
| 3485 | out_err: | ||
| 3473 | rcu_read_unlock(); | 3486 | rcu_read_unlock(); |
| 3474 | cb->args[0] = idx; | 3487 | cb->args[0] = idx; |
| 3475 | 3488 | ||
| 3476 | return skb->len; | 3489 | return err; |
| 3477 | } | 3490 | } |
| 3478 | 3491 | ||
| 3479 | static inline size_t bridge_nlmsg_size(void) | 3492 | static inline size_t bridge_nlmsg_size(void) |
diff --git a/net/core/sock.c b/net/core/sock.c index 79c6aee6af9b..727f924b7f91 100644 --- a/net/core/sock.c +++ b/net/core/sock.c | |||
| @@ -139,10 +139,7 @@ | |||
| 139 | 139 | ||
| 140 | #include <trace/events/sock.h> | 140 | #include <trace/events/sock.h> |
| 141 | 141 | ||
| 142 | #ifdef CONFIG_INET | ||
| 143 | #include <net/tcp.h> | 142 | #include <net/tcp.h> |
| 144 | #endif | ||
| 145 | |||
| 146 | #include <net/busy_poll.h> | 143 | #include <net/busy_poll.h> |
| 147 | 144 | ||
| 148 | static DEFINE_MUTEX(proto_list_mutex); | 145 | static DEFINE_MUTEX(proto_list_mutex); |
| @@ -1803,28 +1800,24 @@ EXPORT_SYMBOL(skb_set_owner_w); | |||
| 1803 | * delay queue. We want to allow the owner socket to send more | 1800 | * delay queue. We want to allow the owner socket to send more |
| 1804 | * packets, as if they were already TX completed by a typical driver. | 1801 | * packets, as if they were already TX completed by a typical driver. |
| 1805 | * But we also want to keep skb->sk set because some packet schedulers | 1802 | * But we also want to keep skb->sk set because some packet schedulers |
| 1806 | * rely on it (sch_fq for example). So we set skb->truesize to a small | 1803 | * rely on it (sch_fq for example). |
| 1807 | * amount (1) and decrease sk_wmem_alloc accordingly. | ||
| 1808 | */ | 1804 | */ |
| 1809 | void skb_orphan_partial(struct sk_buff *skb) | 1805 | void skb_orphan_partial(struct sk_buff *skb) |
| 1810 | { | 1806 | { |
| 1811 | /* If this skb is a TCP pure ACK or already went here, | 1807 | if (skb_is_tcp_pure_ack(skb)) |
| 1812 | * we have nothing to do. 2 is already a very small truesize. | ||
| 1813 | */ | ||
| 1814 | if (skb->truesize <= 2) | ||
| 1815 | return; | 1808 | return; |
| 1816 | 1809 | ||
| 1817 | /* TCP stack sets skb->ooo_okay based on sk_wmem_alloc, | ||
| 1818 | * so we do not completely orphan skb, but transfert all | ||
| 1819 | * accounted bytes but one, to avoid unexpected reorders. | ||
| 1820 | */ | ||
| 1821 | if (skb->destructor == sock_wfree | 1810 | if (skb->destructor == sock_wfree |
| 1822 | #ifdef CONFIG_INET | 1811 | #ifdef CONFIG_INET |
| 1823 | || skb->destructor == tcp_wfree | 1812 | || skb->destructor == tcp_wfree |
| 1824 | #endif | 1813 | #endif |
| 1825 | ) { | 1814 | ) { |
| 1826 | atomic_sub(skb->truesize - 1, &skb->sk->sk_wmem_alloc); | 1815 | struct sock *sk = skb->sk; |
| 1827 | skb->truesize = 1; | 1816 | |
| 1817 | if (atomic_inc_not_zero(&sk->sk_refcnt)) { | ||
| 1818 | atomic_sub(skb->truesize, &sk->sk_wmem_alloc); | ||
| 1819 | skb->destructor = sock_efree; | ||
| 1820 | } | ||
| 1828 | } else { | 1821 | } else { |
| 1829 | skb_orphan(skb); | 1822 | skb_orphan(skb); |
| 1830 | } | 1823 | } |
diff --git a/net/dccp/ipv6.c b/net/dccp/ipv6.c index 840f14aaa016..992621172220 100644 --- a/net/dccp/ipv6.c +++ b/net/dccp/ipv6.c | |||
| @@ -426,6 +426,9 @@ static struct sock *dccp_v6_request_recv_sock(const struct sock *sk, | |||
| 426 | newsk->sk_backlog_rcv = dccp_v4_do_rcv; | 426 | newsk->sk_backlog_rcv = dccp_v4_do_rcv; |
| 427 | newnp->pktoptions = NULL; | 427 | newnp->pktoptions = NULL; |
| 428 | newnp->opt = NULL; | 428 | newnp->opt = NULL; |
| 429 | newnp->ipv6_mc_list = NULL; | ||
| 430 | newnp->ipv6_ac_list = NULL; | ||
| 431 | newnp->ipv6_fl_list = NULL; | ||
| 429 | newnp->mcast_oif = inet6_iif(skb); | 432 | newnp->mcast_oif = inet6_iif(skb); |
| 430 | newnp->mcast_hops = ipv6_hdr(skb)->hop_limit; | 433 | newnp->mcast_hops = ipv6_hdr(skb)->hop_limit; |
| 431 | 434 | ||
| @@ -490,6 +493,9 @@ static struct sock *dccp_v6_request_recv_sock(const struct sock *sk, | |||
| 490 | /* Clone RX bits */ | 493 | /* Clone RX bits */ |
| 491 | newnp->rxopt.all = np->rxopt.all; | 494 | newnp->rxopt.all = np->rxopt.all; |
| 492 | 495 | ||
| 496 | newnp->ipv6_mc_list = NULL; | ||
| 497 | newnp->ipv6_ac_list = NULL; | ||
| 498 | newnp->ipv6_fl_list = NULL; | ||
| 493 | newnp->pktoptions = NULL; | 499 | newnp->pktoptions = NULL; |
| 494 | newnp->opt = NULL; | 500 | newnp->opt = NULL; |
| 495 | newnp->mcast_oif = inet6_iif(skb); | 501 | newnp->mcast_oif = inet6_iif(skb); |
diff --git a/net/ipv4/arp.c b/net/ipv4/arp.c index 0937b34c27ca..d54345a06f72 100644 --- a/net/ipv4/arp.c +++ b/net/ipv4/arp.c | |||
| @@ -653,6 +653,7 @@ static int arp_process(struct net *net, struct sock *sk, struct sk_buff *skb) | |||
| 653 | unsigned char *arp_ptr; | 653 | unsigned char *arp_ptr; |
| 654 | struct rtable *rt; | 654 | struct rtable *rt; |
| 655 | unsigned char *sha; | 655 | unsigned char *sha; |
| 656 | unsigned char *tha = NULL; | ||
| 656 | __be32 sip, tip; | 657 | __be32 sip, tip; |
| 657 | u16 dev_type = dev->type; | 658 | u16 dev_type = dev->type; |
| 658 | int addr_type; | 659 | int addr_type; |
| @@ -724,6 +725,7 @@ static int arp_process(struct net *net, struct sock *sk, struct sk_buff *skb) | |||
| 724 | break; | 725 | break; |
| 725 | #endif | 726 | #endif |
| 726 | default: | 727 | default: |
| 728 | tha = arp_ptr; | ||
| 727 | arp_ptr += dev->addr_len; | 729 | arp_ptr += dev->addr_len; |
| 728 | } | 730 | } |
| 729 | memcpy(&tip, arp_ptr, 4); | 731 | memcpy(&tip, arp_ptr, 4); |
| @@ -842,8 +844,18 @@ static int arp_process(struct net *net, struct sock *sk, struct sk_buff *skb) | |||
| 842 | It is possible, that this option should be enabled for some | 844 | It is possible, that this option should be enabled for some |
| 843 | devices (strip is candidate) | 845 | devices (strip is candidate) |
| 844 | */ | 846 | */ |
| 845 | is_garp = arp->ar_op == htons(ARPOP_REQUEST) && tip == sip && | 847 | is_garp = tip == sip && addr_type == RTN_UNICAST; |
| 846 | addr_type == RTN_UNICAST; | 848 | |
| 849 | /* Unsolicited ARP _replies_ also require target hwaddr to be | ||
| 850 | * the same as source. | ||
| 851 | */ | ||
| 852 | if (is_garp && arp->ar_op == htons(ARPOP_REPLY)) | ||
| 853 | is_garp = | ||
| 854 | /* IPv4 over IEEE 1394 doesn't provide target | ||
| 855 | * hardware address field in its ARP payload. | ||
| 856 | */ | ||
| 857 | tha && | ||
| 858 | !memcmp(tha, sha, dev->addr_len); | ||
| 847 | 859 | ||
| 848 | if (!n && | 860 | if (!n && |
| 849 | ((arp->ar_op == htons(ARPOP_REPLY) && | 861 | ((arp->ar_op == htons(ARPOP_REPLY) && |
diff --git a/net/ipv4/fib_frontend.c b/net/ipv4/fib_frontend.c index 39bd1edee676..83e3ed258467 100644 --- a/net/ipv4/fib_frontend.c +++ b/net/ipv4/fib_frontend.c | |||
| @@ -763,7 +763,7 @@ static int inet_dump_fib(struct sk_buff *skb, struct netlink_callback *cb) | |||
| 763 | unsigned int e = 0, s_e; | 763 | unsigned int e = 0, s_e; |
| 764 | struct fib_table *tb; | 764 | struct fib_table *tb; |
| 765 | struct hlist_head *head; | 765 | struct hlist_head *head; |
| 766 | int dumped = 0; | 766 | int dumped = 0, err; |
| 767 | 767 | ||
| 768 | if (nlmsg_len(cb->nlh) >= sizeof(struct rtmsg) && | 768 | if (nlmsg_len(cb->nlh) >= sizeof(struct rtmsg) && |
| 769 | ((struct rtmsg *) nlmsg_data(cb->nlh))->rtm_flags & RTM_F_CLONED) | 769 | ((struct rtmsg *) nlmsg_data(cb->nlh))->rtm_flags & RTM_F_CLONED) |
| @@ -783,20 +783,27 @@ static int inet_dump_fib(struct sk_buff *skb, struct netlink_callback *cb) | |||
| 783 | if (dumped) | 783 | if (dumped) |
| 784 | memset(&cb->args[2], 0, sizeof(cb->args) - | 784 | memset(&cb->args[2], 0, sizeof(cb->args) - |
| 785 | 2 * sizeof(cb->args[0])); | 785 | 2 * sizeof(cb->args[0])); |
| 786 | if (fib_table_dump(tb, skb, cb) < 0) | 786 | err = fib_table_dump(tb, skb, cb); |
| 787 | goto out; | 787 | if (err < 0) { |
| 788 | if (likely(skb->len)) | ||
| 789 | goto out; | ||
| 790 | |||
| 791 | goto out_err; | ||
| 792 | } | ||
| 788 | dumped = 1; | 793 | dumped = 1; |
| 789 | next: | 794 | next: |
| 790 | e++; | 795 | e++; |
| 791 | } | 796 | } |
| 792 | } | 797 | } |
| 793 | out: | 798 | out: |
| 799 | err = skb->len; | ||
| 800 | out_err: | ||
| 794 | rcu_read_unlock(); | 801 | rcu_read_unlock(); |
| 795 | 802 | ||
| 796 | cb->args[1] = e; | 803 | cb->args[1] = e; |
| 797 | cb->args[0] = h; | 804 | cb->args[0] = h; |
| 798 | 805 | ||
| 799 | return skb->len; | 806 | return err; |
| 800 | } | 807 | } |
| 801 | 808 | ||
| 802 | /* Prepare and feed intra-kernel routing request. | 809 | /* Prepare and feed intra-kernel routing request. |
diff --git a/net/ipv4/fib_trie.c b/net/ipv4/fib_trie.c index 1201409ba1dc..51182ff2b441 100644 --- a/net/ipv4/fib_trie.c +++ b/net/ipv4/fib_trie.c | |||
| @@ -1983,6 +1983,8 @@ static int fn_trie_dump_leaf(struct key_vector *l, struct fib_table *tb, | |||
| 1983 | 1983 | ||
| 1984 | /* rcu_read_lock is hold by caller */ | 1984 | /* rcu_read_lock is hold by caller */ |
| 1985 | hlist_for_each_entry_rcu(fa, &l->leaf, fa_list) { | 1985 | hlist_for_each_entry_rcu(fa, &l->leaf, fa_list) { |
| 1986 | int err; | ||
| 1987 | |||
| 1986 | if (i < s_i) { | 1988 | if (i < s_i) { |
| 1987 | i++; | 1989 | i++; |
| 1988 | continue; | 1990 | continue; |
| @@ -1993,17 +1995,14 @@ static int fn_trie_dump_leaf(struct key_vector *l, struct fib_table *tb, | |||
| 1993 | continue; | 1995 | continue; |
| 1994 | } | 1996 | } |
| 1995 | 1997 | ||
| 1996 | if (fib_dump_info(skb, NETLINK_CB(cb->skb).portid, | 1998 | err = fib_dump_info(skb, NETLINK_CB(cb->skb).portid, |
| 1997 | cb->nlh->nlmsg_seq, | 1999 | cb->nlh->nlmsg_seq, RTM_NEWROUTE, |
| 1998 | RTM_NEWROUTE, | 2000 | tb->tb_id, fa->fa_type, |
| 1999 | tb->tb_id, | 2001 | xkey, KEYLENGTH - fa->fa_slen, |
| 2000 | fa->fa_type, | 2002 | fa->fa_tos, fa->fa_info, NLM_F_MULTI); |
| 2001 | xkey, | 2003 | if (err < 0) { |
| 2002 | KEYLENGTH - fa->fa_slen, | ||
| 2003 | fa->fa_tos, | ||
| 2004 | fa->fa_info, NLM_F_MULTI) < 0) { | ||
| 2005 | cb->args[4] = i; | 2004 | cb->args[4] = i; |
| 2006 | return -1; | 2005 | return err; |
| 2007 | } | 2006 | } |
| 2008 | i++; | 2007 | i++; |
| 2009 | } | 2008 | } |
| @@ -2025,10 +2024,13 @@ int fib_table_dump(struct fib_table *tb, struct sk_buff *skb, | |||
| 2025 | t_key key = cb->args[3]; | 2024 | t_key key = cb->args[3]; |
| 2026 | 2025 | ||
| 2027 | while ((l = leaf_walk_rcu(&tp, key)) != NULL) { | 2026 | while ((l = leaf_walk_rcu(&tp, key)) != NULL) { |
| 2028 | if (fn_trie_dump_leaf(l, tb, skb, cb) < 0) { | 2027 | int err; |
| 2028 | |||
| 2029 | err = fn_trie_dump_leaf(l, tb, skb, cb); | ||
| 2030 | if (err < 0) { | ||
| 2029 | cb->args[3] = key; | 2031 | cb->args[3] = key; |
| 2030 | cb->args[2] = count; | 2032 | cb->args[2] = count; |
| 2031 | return -1; | 2033 | return err; |
| 2032 | } | 2034 | } |
| 2033 | 2035 | ||
| 2034 | ++count; | 2036 | ++count; |
diff --git a/net/ipv4/ipmr.c b/net/ipv4/ipmr.c index 3a02d52ed50e..551de4d023a8 100644 --- a/net/ipv4/ipmr.c +++ b/net/ipv4/ipmr.c | |||
| @@ -1980,6 +1980,20 @@ int ip_mr_input(struct sk_buff *skb) | |||
| 1980 | struct net *net = dev_net(skb->dev); | 1980 | struct net *net = dev_net(skb->dev); |
| 1981 | int local = skb_rtable(skb)->rt_flags & RTCF_LOCAL; | 1981 | int local = skb_rtable(skb)->rt_flags & RTCF_LOCAL; |
| 1982 | struct mr_table *mrt; | 1982 | struct mr_table *mrt; |
| 1983 | struct net_device *dev; | ||
| 1984 | |||
| 1985 | /* skb->dev passed in is the loX master dev for vrfs. | ||
| 1986 | * As there are no vifs associated with loopback devices, | ||
| 1987 | * get the proper interface that does have a vif associated with it. | ||
| 1988 | */ | ||
| 1989 | dev = skb->dev; | ||
| 1990 | if (netif_is_l3_master(skb->dev)) { | ||
| 1991 | dev = dev_get_by_index_rcu(net, IPCB(skb)->iif); | ||
| 1992 | if (!dev) { | ||
| 1993 | kfree_skb(skb); | ||
| 1994 | return -ENODEV; | ||
| 1995 | } | ||
| 1996 | } | ||
| 1983 | 1997 | ||
| 1984 | /* Packet is looped back after forward, it should not be | 1998 | /* Packet is looped back after forward, it should not be |
| 1985 | * forwarded second time, but still can be delivered locally. | 1999 | * forwarded second time, but still can be delivered locally. |
| @@ -2017,7 +2031,7 @@ int ip_mr_input(struct sk_buff *skb) | |||
| 2017 | /* already under rcu_read_lock() */ | 2031 | /* already under rcu_read_lock() */ |
| 2018 | cache = ipmr_cache_find(mrt, ip_hdr(skb)->saddr, ip_hdr(skb)->daddr); | 2032 | cache = ipmr_cache_find(mrt, ip_hdr(skb)->saddr, ip_hdr(skb)->daddr); |
| 2019 | if (!cache) { | 2033 | if (!cache) { |
| 2020 | int vif = ipmr_find_vif(mrt, skb->dev); | 2034 | int vif = ipmr_find_vif(mrt, dev); |
| 2021 | 2035 | ||
| 2022 | if (vif >= 0) | 2036 | if (vif >= 0) |
| 2023 | cache = ipmr_cache_find_any(mrt, ip_hdr(skb)->daddr, | 2037 | cache = ipmr_cache_find_any(mrt, ip_hdr(skb)->daddr, |
| @@ -2037,7 +2051,7 @@ int ip_mr_input(struct sk_buff *skb) | |||
| 2037 | } | 2051 | } |
| 2038 | 2052 | ||
| 2039 | read_lock(&mrt_lock); | 2053 | read_lock(&mrt_lock); |
| 2040 | vif = ipmr_find_vif(mrt, skb->dev); | 2054 | vif = ipmr_find_vif(mrt, dev); |
| 2041 | if (vif >= 0) { | 2055 | if (vif >= 0) { |
| 2042 | int err2 = ipmr_cache_unresolved(mrt, vif, skb); | 2056 | int err2 = ipmr_cache_unresolved(mrt, vif, skb); |
| 2043 | read_unlock(&mrt_lock); | 2057 | read_unlock(&mrt_lock); |
diff --git a/net/ipv4/tcp_input.c b/net/ipv4/tcp_input.c index 5a3ad09e2786..174d4376baa5 100644 --- a/net/ipv4/tcp_input.c +++ b/net/ipv4/tcp_input.c | |||
| @@ -1179,13 +1179,14 @@ static int tcp_match_skb_to_sack(struct sock *sk, struct sk_buff *skb, | |||
| 1179 | */ | 1179 | */ |
| 1180 | if (pkt_len > mss) { | 1180 | if (pkt_len > mss) { |
| 1181 | unsigned int new_len = (pkt_len / mss) * mss; | 1181 | unsigned int new_len = (pkt_len / mss) * mss; |
| 1182 | if (!in_sack && new_len < pkt_len) { | 1182 | if (!in_sack && new_len < pkt_len) |
| 1183 | new_len += mss; | 1183 | new_len += mss; |
| 1184 | if (new_len >= skb->len) | ||
| 1185 | return 0; | ||
| 1186 | } | ||
| 1187 | pkt_len = new_len; | 1184 | pkt_len = new_len; |
| 1188 | } | 1185 | } |
| 1186 | |||
| 1187 | if (pkt_len >= skb->len && !in_sack) | ||
| 1188 | return 0; | ||
| 1189 | |||
| 1189 | err = tcp_fragment(sk, skb, pkt_len, mss, GFP_ATOMIC); | 1190 | err = tcp_fragment(sk, skb, pkt_len, mss, GFP_ATOMIC); |
| 1190 | if (err < 0) | 1191 | if (err < 0) |
| 1191 | return err; | 1192 | return err; |
| @@ -3189,7 +3190,7 @@ static int tcp_clean_rtx_queue(struct sock *sk, int prior_fackets, | |||
| 3189 | int delta; | 3190 | int delta; |
| 3190 | 3191 | ||
| 3191 | /* Non-retransmitted hole got filled? That's reordering */ | 3192 | /* Non-retransmitted hole got filled? That's reordering */ |
| 3192 | if (reord < prior_fackets) | 3193 | if (reord < prior_fackets && reord <= tp->fackets_out) |
| 3193 | tcp_update_reordering(sk, tp->fackets_out - reord, 0); | 3194 | tcp_update_reordering(sk, tp->fackets_out - reord, 0); |
| 3194 | 3195 | ||
| 3195 | delta = tcp_is_fack(tp) ? pkts_acked : | 3196 | delta = tcp_is_fack(tp) ? pkts_acked : |
diff --git a/net/ipv4/udp.c b/net/ipv4/udp.c index ea6e4cff9faf..1d6219bf2d6b 100644 --- a/net/ipv4/udp.c +++ b/net/ipv4/udp.c | |||
| @@ -1612,7 +1612,7 @@ static void udp_v4_rehash(struct sock *sk) | |||
| 1612 | udp_lib_rehash(sk, new_hash); | 1612 | udp_lib_rehash(sk, new_hash); |
| 1613 | } | 1613 | } |
| 1614 | 1614 | ||
| 1615 | int __udp_queue_rcv_skb(struct sock *sk, struct sk_buff *skb) | 1615 | static int __udp_queue_rcv_skb(struct sock *sk, struct sk_buff *skb) |
| 1616 | { | 1616 | { |
| 1617 | int rc; | 1617 | int rc; |
| 1618 | 1618 | ||
| @@ -1657,7 +1657,7 @@ EXPORT_SYMBOL(udp_encap_enable); | |||
| 1657 | * Note that in the success and error cases, the skb is assumed to | 1657 | * Note that in the success and error cases, the skb is assumed to |
| 1658 | * have either been requeued or freed. | 1658 | * have either been requeued or freed. |
| 1659 | */ | 1659 | */ |
| 1660 | int udp_queue_rcv_skb(struct sock *sk, struct sk_buff *skb) | 1660 | static int udp_queue_rcv_skb(struct sock *sk, struct sk_buff *skb) |
| 1661 | { | 1661 | { |
| 1662 | struct udp_sock *up = udp_sk(sk); | 1662 | struct udp_sock *up = udp_sk(sk); |
| 1663 | int is_udplite = IS_UDPLITE(sk); | 1663 | int is_udplite = IS_UDPLITE(sk); |
diff --git a/net/ipv4/udp_impl.h b/net/ipv4/udp_impl.h index feb50a16398d..a8cf8c6fb60c 100644 --- a/net/ipv4/udp_impl.h +++ b/net/ipv4/udp_impl.h | |||
| @@ -25,7 +25,6 @@ int udp_recvmsg(struct sock *sk, struct msghdr *msg, size_t len, int noblock, | |||
| 25 | int flags, int *addr_len); | 25 | int flags, int *addr_len); |
| 26 | int udp_sendpage(struct sock *sk, struct page *page, int offset, size_t size, | 26 | int udp_sendpage(struct sock *sk, struct page *page, int offset, size_t size, |
| 27 | int flags); | 27 | int flags); |
| 28 | int __udp_queue_rcv_skb(struct sock *sk, struct sk_buff *skb); | ||
| 29 | void udp_destroy_sock(struct sock *sk); | 28 | void udp_destroy_sock(struct sock *sk); |
| 30 | 29 | ||
| 31 | #ifdef CONFIG_PROC_FS | 30 | #ifdef CONFIG_PROC_FS |
diff --git a/net/ipv6/addrconf.c b/net/ipv6/addrconf.c index 8d297a79b568..6a4fb1e629fb 100644 --- a/net/ipv6/addrconf.c +++ b/net/ipv6/addrconf.c | |||
| @@ -1022,7 +1022,10 @@ ipv6_add_addr(struct inet6_dev *idev, const struct in6_addr *addr, | |||
| 1022 | INIT_HLIST_NODE(&ifa->addr_lst); | 1022 | INIT_HLIST_NODE(&ifa->addr_lst); |
| 1023 | ifa->scope = scope; | 1023 | ifa->scope = scope; |
| 1024 | ifa->prefix_len = pfxlen; | 1024 | ifa->prefix_len = pfxlen; |
| 1025 | ifa->flags = flags | IFA_F_TENTATIVE; | 1025 | ifa->flags = flags; |
| 1026 | /* No need to add the TENTATIVE flag for addresses with NODAD */ | ||
| 1027 | if (!(flags & IFA_F_NODAD)) | ||
| 1028 | ifa->flags |= IFA_F_TENTATIVE; | ||
| 1026 | ifa->valid_lft = valid_lft; | 1029 | ifa->valid_lft = valid_lft; |
| 1027 | ifa->prefered_lft = prefered_lft; | 1030 | ifa->prefered_lft = prefered_lft; |
| 1028 | ifa->cstamp = ifa->tstamp = jiffies; | 1031 | ifa->cstamp = ifa->tstamp = jiffies; |
diff --git a/net/ipv6/ip6_offload.c b/net/ipv6/ip6_offload.c index 93e58a5e1837..280268f1dd7b 100644 --- a/net/ipv6/ip6_offload.c +++ b/net/ipv6/ip6_offload.c | |||
| @@ -63,7 +63,6 @@ static struct sk_buff *ipv6_gso_segment(struct sk_buff *skb, | |||
| 63 | const struct net_offload *ops; | 63 | const struct net_offload *ops; |
| 64 | int proto; | 64 | int proto; |
| 65 | struct frag_hdr *fptr; | 65 | struct frag_hdr *fptr; |
| 66 | unsigned int unfrag_ip6hlen; | ||
| 67 | unsigned int payload_len; | 66 | unsigned int payload_len; |
| 68 | u8 *prevhdr; | 67 | u8 *prevhdr; |
| 69 | int offset = 0; | 68 | int offset = 0; |
| @@ -116,8 +115,10 @@ static struct sk_buff *ipv6_gso_segment(struct sk_buff *skb, | |||
| 116 | skb->network_header = (u8 *)ipv6h - skb->head; | 115 | skb->network_header = (u8 *)ipv6h - skb->head; |
| 117 | 116 | ||
| 118 | if (udpfrag) { | 117 | if (udpfrag) { |
| 119 | unfrag_ip6hlen = ip6_find_1stfragopt(skb, &prevhdr); | 118 | int err = ip6_find_1stfragopt(skb, &prevhdr); |
| 120 | fptr = (struct frag_hdr *)((u8 *)ipv6h + unfrag_ip6hlen); | 119 | if (err < 0) |
| 120 | return ERR_PTR(err); | ||
| 121 | fptr = (struct frag_hdr *)((u8 *)ipv6h + err); | ||
| 121 | fptr->frag_off = htons(offset); | 122 | fptr->frag_off = htons(offset); |
| 122 | if (skb->next) | 123 | if (skb->next) |
| 123 | fptr->frag_off |= htons(IP6_MF); | 124 | fptr->frag_off |= htons(IP6_MF); |
diff --git a/net/ipv6/ip6_output.c b/net/ipv6/ip6_output.c index 58f6288e9ba5..d4a31becbd25 100644 --- a/net/ipv6/ip6_output.c +++ b/net/ipv6/ip6_output.c | |||
| @@ -597,7 +597,10 @@ int ip6_fragment(struct net *net, struct sock *sk, struct sk_buff *skb, | |||
| 597 | int ptr, offset = 0, err = 0; | 597 | int ptr, offset = 0, err = 0; |
| 598 | u8 *prevhdr, nexthdr = 0; | 598 | u8 *prevhdr, nexthdr = 0; |
| 599 | 599 | ||
| 600 | hlen = ip6_find_1stfragopt(skb, &prevhdr); | 600 | err = ip6_find_1stfragopt(skb, &prevhdr); |
| 601 | if (err < 0) | ||
| 602 | goto fail; | ||
| 603 | hlen = err; | ||
| 601 | nexthdr = *prevhdr; | 604 | nexthdr = *prevhdr; |
| 602 | 605 | ||
| 603 | mtu = ip6_skb_dst_mtu(skb); | 606 | mtu = ip6_skb_dst_mtu(skb); |
diff --git a/net/ipv6/output_core.c b/net/ipv6/output_core.c index cd4252346a32..e9065b8d3af8 100644 --- a/net/ipv6/output_core.c +++ b/net/ipv6/output_core.c | |||
| @@ -79,14 +79,13 @@ EXPORT_SYMBOL(ipv6_select_ident); | |||
| 79 | int ip6_find_1stfragopt(struct sk_buff *skb, u8 **nexthdr) | 79 | int ip6_find_1stfragopt(struct sk_buff *skb, u8 **nexthdr) |
| 80 | { | 80 | { |
| 81 | u16 offset = sizeof(struct ipv6hdr); | 81 | u16 offset = sizeof(struct ipv6hdr); |
| 82 | struct ipv6_opt_hdr *exthdr = | ||
| 83 | (struct ipv6_opt_hdr *)(ipv6_hdr(skb) + 1); | ||
| 84 | unsigned int packet_len = skb_tail_pointer(skb) - | 82 | unsigned int packet_len = skb_tail_pointer(skb) - |
| 85 | skb_network_header(skb); | 83 | skb_network_header(skb); |
| 86 | int found_rhdr = 0; | 84 | int found_rhdr = 0; |
| 87 | *nexthdr = &ipv6_hdr(skb)->nexthdr; | 85 | *nexthdr = &ipv6_hdr(skb)->nexthdr; |
| 88 | 86 | ||
| 89 | while (offset + 1 <= packet_len) { | 87 | while (offset <= packet_len) { |
| 88 | struct ipv6_opt_hdr *exthdr; | ||
| 90 | 89 | ||
| 91 | switch (**nexthdr) { | 90 | switch (**nexthdr) { |
| 92 | 91 | ||
| @@ -107,13 +106,16 @@ int ip6_find_1stfragopt(struct sk_buff *skb, u8 **nexthdr) | |||
| 107 | return offset; | 106 | return offset; |
| 108 | } | 107 | } |
| 109 | 108 | ||
| 110 | offset += ipv6_optlen(exthdr); | 109 | if (offset + sizeof(struct ipv6_opt_hdr) > packet_len) |
| 111 | *nexthdr = &exthdr->nexthdr; | 110 | return -EINVAL; |
| 111 | |||
| 112 | exthdr = (struct ipv6_opt_hdr *)(skb_network_header(skb) + | 112 | exthdr = (struct ipv6_opt_hdr *)(skb_network_header(skb) + |
| 113 | offset); | 113 | offset); |
| 114 | offset += ipv6_optlen(exthdr); | ||
| 115 | *nexthdr = &exthdr->nexthdr; | ||
| 114 | } | 116 | } |
| 115 | 117 | ||
| 116 | return offset; | 118 | return -EINVAL; |
| 117 | } | 119 | } |
| 118 | EXPORT_SYMBOL(ip6_find_1stfragopt); | 120 | EXPORT_SYMBOL(ip6_find_1stfragopt); |
| 119 | 121 | ||
diff --git a/net/ipv6/tcp_ipv6.c b/net/ipv6/tcp_ipv6.c index 7a8237acd210..4f4310a36a04 100644 --- a/net/ipv6/tcp_ipv6.c +++ b/net/ipv6/tcp_ipv6.c | |||
| @@ -1062,6 +1062,7 @@ static struct sock *tcp_v6_syn_recv_sock(const struct sock *sk, struct sk_buff * | |||
| 1062 | newtp->af_specific = &tcp_sock_ipv6_mapped_specific; | 1062 | newtp->af_specific = &tcp_sock_ipv6_mapped_specific; |
| 1063 | #endif | 1063 | #endif |
| 1064 | 1064 | ||
| 1065 | newnp->ipv6_mc_list = NULL; | ||
| 1065 | newnp->ipv6_ac_list = NULL; | 1066 | newnp->ipv6_ac_list = NULL; |
| 1066 | newnp->ipv6_fl_list = NULL; | 1067 | newnp->ipv6_fl_list = NULL; |
| 1067 | newnp->pktoptions = NULL; | 1068 | newnp->pktoptions = NULL; |
| @@ -1131,6 +1132,7 @@ static struct sock *tcp_v6_syn_recv_sock(const struct sock *sk, struct sk_buff * | |||
| 1131 | First: no IPv4 options. | 1132 | First: no IPv4 options. |
| 1132 | */ | 1133 | */ |
| 1133 | newinet->inet_opt = NULL; | 1134 | newinet->inet_opt = NULL; |
| 1135 | newnp->ipv6_mc_list = NULL; | ||
| 1134 | newnp->ipv6_ac_list = NULL; | 1136 | newnp->ipv6_ac_list = NULL; |
| 1135 | newnp->ipv6_fl_list = NULL; | 1137 | newnp->ipv6_fl_list = NULL; |
| 1136 | 1138 | ||
diff --git a/net/ipv6/udp.c b/net/ipv6/udp.c index 04862abfe4ec..06ec39b79609 100644 --- a/net/ipv6/udp.c +++ b/net/ipv6/udp.c | |||
| @@ -526,7 +526,7 @@ out: | |||
| 526 | return; | 526 | return; |
| 527 | } | 527 | } |
| 528 | 528 | ||
| 529 | int __udpv6_queue_rcv_skb(struct sock *sk, struct sk_buff *skb) | 529 | static int __udpv6_queue_rcv_skb(struct sock *sk, struct sk_buff *skb) |
| 530 | { | 530 | { |
| 531 | int rc; | 531 | int rc; |
| 532 | 532 | ||
| @@ -569,7 +569,7 @@ void udpv6_encap_enable(void) | |||
| 569 | } | 569 | } |
| 570 | EXPORT_SYMBOL(udpv6_encap_enable); | 570 | EXPORT_SYMBOL(udpv6_encap_enable); |
| 571 | 571 | ||
| 572 | int udpv6_queue_rcv_skb(struct sock *sk, struct sk_buff *skb) | 572 | static int udpv6_queue_rcv_skb(struct sock *sk, struct sk_buff *skb) |
| 573 | { | 573 | { |
| 574 | struct udp_sock *up = udp_sk(sk); | 574 | struct udp_sock *up = udp_sk(sk); |
| 575 | int is_udplite = IS_UDPLITE(sk); | 575 | int is_udplite = IS_UDPLITE(sk); |
diff --git a/net/ipv6/udp_impl.h b/net/ipv6/udp_impl.h index e78bdc76dcc3..f180b3d85e31 100644 --- a/net/ipv6/udp_impl.h +++ b/net/ipv6/udp_impl.h | |||
| @@ -26,7 +26,6 @@ int compat_udpv6_getsockopt(struct sock *sk, int level, int optname, | |||
| 26 | int udpv6_sendmsg(struct sock *sk, struct msghdr *msg, size_t len); | 26 | int udpv6_sendmsg(struct sock *sk, struct msghdr *msg, size_t len); |
| 27 | int udpv6_recvmsg(struct sock *sk, struct msghdr *msg, size_t len, int noblock, | 27 | int udpv6_recvmsg(struct sock *sk, struct msghdr *msg, size_t len, int noblock, |
| 28 | int flags, int *addr_len); | 28 | int flags, int *addr_len); |
| 29 | int __udpv6_queue_rcv_skb(struct sock *sk, struct sk_buff *skb); | ||
| 30 | void udpv6_destroy_sock(struct sock *sk); | 29 | void udpv6_destroy_sock(struct sock *sk); |
| 31 | 30 | ||
| 32 | #ifdef CONFIG_PROC_FS | 31 | #ifdef CONFIG_PROC_FS |
diff --git a/net/ipv6/udp_offload.c b/net/ipv6/udp_offload.c index ac858c480f2f..a2267f80febb 100644 --- a/net/ipv6/udp_offload.c +++ b/net/ipv6/udp_offload.c | |||
| @@ -29,6 +29,7 @@ static struct sk_buff *udp6_ufo_fragment(struct sk_buff *skb, | |||
| 29 | u8 frag_hdr_sz = sizeof(struct frag_hdr); | 29 | u8 frag_hdr_sz = sizeof(struct frag_hdr); |
| 30 | __wsum csum; | 30 | __wsum csum; |
| 31 | int tnl_hlen; | 31 | int tnl_hlen; |
| 32 | int err; | ||
| 32 | 33 | ||
| 33 | mss = skb_shinfo(skb)->gso_size; | 34 | mss = skb_shinfo(skb)->gso_size; |
| 34 | if (unlikely(skb->len <= mss)) | 35 | if (unlikely(skb->len <= mss)) |
| @@ -90,7 +91,10 @@ static struct sk_buff *udp6_ufo_fragment(struct sk_buff *skb, | |||
| 90 | /* Find the unfragmentable header and shift it left by frag_hdr_sz | 91 | /* Find the unfragmentable header and shift it left by frag_hdr_sz |
| 91 | * bytes to insert fragment header. | 92 | * bytes to insert fragment header. |
| 92 | */ | 93 | */ |
| 93 | unfrag_ip6hlen = ip6_find_1stfragopt(skb, &prevhdr); | 94 | err = ip6_find_1stfragopt(skb, &prevhdr); |
| 95 | if (err < 0) | ||
| 96 | return ERR_PTR(err); | ||
| 97 | unfrag_ip6hlen = err; | ||
| 94 | nexthdr = *prevhdr; | 98 | nexthdr = *prevhdr; |
| 95 | *prevhdr = NEXTHDR_FRAGMENT; | 99 | *prevhdr = NEXTHDR_FRAGMENT; |
| 96 | unfrag_len = (skb_network_header(skb) - skb_mac_header(skb)) + | 100 | unfrag_len = (skb_network_header(skb) - skb_mac_header(skb)) + |
diff --git a/net/packet/af_packet.c b/net/packet/af_packet.c index f4001763134d..e3eeed19cc7a 100644 --- a/net/packet/af_packet.c +++ b/net/packet/af_packet.c | |||
| @@ -2658,13 +2658,6 @@ static int tpacket_snd(struct packet_sock *po, struct msghdr *msg) | |||
| 2658 | dev = dev_get_by_index(sock_net(&po->sk), saddr->sll_ifindex); | 2658 | dev = dev_get_by_index(sock_net(&po->sk), saddr->sll_ifindex); |
| 2659 | } | 2659 | } |
| 2660 | 2660 | ||
| 2661 | sockc.tsflags = po->sk.sk_tsflags; | ||
| 2662 | if (msg->msg_controllen) { | ||
| 2663 | err = sock_cmsg_send(&po->sk, msg, &sockc); | ||
| 2664 | if (unlikely(err)) | ||
| 2665 | goto out; | ||
| 2666 | } | ||
| 2667 | |||
| 2668 | err = -ENXIO; | 2661 | err = -ENXIO; |
| 2669 | if (unlikely(dev == NULL)) | 2662 | if (unlikely(dev == NULL)) |
| 2670 | goto out; | 2663 | goto out; |
| @@ -2672,6 +2665,13 @@ static int tpacket_snd(struct packet_sock *po, struct msghdr *msg) | |||
| 2672 | if (unlikely(!(dev->flags & IFF_UP))) | 2665 | if (unlikely(!(dev->flags & IFF_UP))) |
| 2673 | goto out_put; | 2666 | goto out_put; |
| 2674 | 2667 | ||
| 2668 | sockc.tsflags = po->sk.sk_tsflags; | ||
| 2669 | if (msg->msg_controllen) { | ||
| 2670 | err = sock_cmsg_send(&po->sk, msg, &sockc); | ||
| 2671 | if (unlikely(err)) | ||
| 2672 | goto out_put; | ||
| 2673 | } | ||
| 2674 | |||
| 2675 | if (po->sk.sk_socket->type == SOCK_RAW) | 2675 | if (po->sk.sk_socket->type == SOCK_RAW) |
| 2676 | reserve = dev->hard_header_len; | 2676 | reserve = dev->hard_header_len; |
| 2677 | size_max = po->tx_ring.frame_size | 2677 | size_max = po->tx_ring.frame_size |
diff --git a/net/sched/sch_api.c b/net/sched/sch_api.c index bbe57d57b67f..e88342fde1bc 100644 --- a/net/sched/sch_api.c +++ b/net/sched/sch_api.c | |||
| @@ -1831,6 +1831,12 @@ static int tc_dump_tclass_root(struct Qdisc *root, struct sk_buff *skb, | |||
| 1831 | if (!qdisc_dev(root)) | 1831 | if (!qdisc_dev(root)) |
| 1832 | return 0; | 1832 | return 0; |
| 1833 | 1833 | ||
| 1834 | if (tcm->tcm_parent) { | ||
| 1835 | q = qdisc_match_from_root(root, TC_H_MAJ(tcm->tcm_parent)); | ||
| 1836 | if (q && tc_dump_tclass_qdisc(q, skb, tcm, cb, t_p, s_t) < 0) | ||
| 1837 | return -1; | ||
| 1838 | return 0; | ||
| 1839 | } | ||
| 1834 | hash_for_each(qdisc_dev(root)->qdisc_hash, b, q, hash) { | 1840 | hash_for_each(qdisc_dev(root)->qdisc_hash, b, q, hash) { |
| 1835 | if (tc_dump_tclass_qdisc(q, skb, tcm, cb, t_p, s_t) < 0) | 1841 | if (tc_dump_tclass_qdisc(q, skb, tcm, cb, t_p, s_t) < 0) |
| 1836 | return -1; | 1842 | return -1; |
diff --git a/net/sctp/ipv6.c b/net/sctp/ipv6.c index 961ee59f696a..f5b45b8b8b16 100644 --- a/net/sctp/ipv6.c +++ b/net/sctp/ipv6.c | |||
| @@ -240,12 +240,10 @@ static void sctp_v6_get_dst(struct sctp_transport *t, union sctp_addr *saddr, | |||
| 240 | struct sctp_bind_addr *bp; | 240 | struct sctp_bind_addr *bp; |
| 241 | struct ipv6_pinfo *np = inet6_sk(sk); | 241 | struct ipv6_pinfo *np = inet6_sk(sk); |
| 242 | struct sctp_sockaddr_entry *laddr; | 242 | struct sctp_sockaddr_entry *laddr; |
| 243 | union sctp_addr *baddr = NULL; | ||
| 244 | union sctp_addr *daddr = &t->ipaddr; | 243 | union sctp_addr *daddr = &t->ipaddr; |
| 245 | union sctp_addr dst_saddr; | 244 | union sctp_addr dst_saddr; |
| 246 | struct in6_addr *final_p, final; | 245 | struct in6_addr *final_p, final; |
| 247 | __u8 matchlen = 0; | 246 | __u8 matchlen = 0; |
| 248 | __u8 bmatchlen; | ||
| 249 | sctp_scope_t scope; | 247 | sctp_scope_t scope; |
| 250 | 248 | ||
| 251 | memset(fl6, 0, sizeof(struct flowi6)); | 249 | memset(fl6, 0, sizeof(struct flowi6)); |
| @@ -312,23 +310,37 @@ static void sctp_v6_get_dst(struct sctp_transport *t, union sctp_addr *saddr, | |||
| 312 | */ | 310 | */ |
| 313 | rcu_read_lock(); | 311 | rcu_read_lock(); |
| 314 | list_for_each_entry_rcu(laddr, &bp->address_list, list) { | 312 | list_for_each_entry_rcu(laddr, &bp->address_list, list) { |
| 315 | if (!laddr->valid) | 313 | struct dst_entry *bdst; |
| 314 | __u8 bmatchlen; | ||
| 315 | |||
| 316 | if (!laddr->valid || | ||
| 317 | laddr->state != SCTP_ADDR_SRC || | ||
| 318 | laddr->a.sa.sa_family != AF_INET6 || | ||
| 319 | scope > sctp_scope(&laddr->a)) | ||
| 316 | continue; | 320 | continue; |
| 317 | if ((laddr->state == SCTP_ADDR_SRC) && | 321 | |
| 318 | (laddr->a.sa.sa_family == AF_INET6) && | 322 | fl6->saddr = laddr->a.v6.sin6_addr; |
| 319 | (scope <= sctp_scope(&laddr->a))) { | 323 | fl6->fl6_sport = laddr->a.v6.sin6_port; |
| 320 | bmatchlen = sctp_v6_addr_match_len(daddr, &laddr->a); | ||
| 321 | if (!baddr || (matchlen < bmatchlen)) { | ||
| 322 | baddr = &laddr->a; | ||
| 323 | matchlen = bmatchlen; | ||
| 324 | } | ||
| 325 | } | ||
| 326 | } | ||
| 327 | if (baddr) { | ||
| 328 | fl6->saddr = baddr->v6.sin6_addr; | ||
| 329 | fl6->fl6_sport = baddr->v6.sin6_port; | ||
| 330 | final_p = fl6_update_dst(fl6, rcu_dereference(np->opt), &final); | 324 | final_p = fl6_update_dst(fl6, rcu_dereference(np->opt), &final); |
| 331 | dst = ip6_dst_lookup_flow(sk, fl6, final_p); | 325 | bdst = ip6_dst_lookup_flow(sk, fl6, final_p); |
| 326 | |||
| 327 | if (!IS_ERR(bdst) && | ||
| 328 | ipv6_chk_addr(dev_net(bdst->dev), | ||
| 329 | &laddr->a.v6.sin6_addr, bdst->dev, 1)) { | ||
| 330 | if (!IS_ERR_OR_NULL(dst)) | ||
| 331 | dst_release(dst); | ||
| 332 | dst = bdst; | ||
| 333 | break; | ||
| 334 | } | ||
| 335 | |||
| 336 | bmatchlen = sctp_v6_addr_match_len(daddr, &laddr->a); | ||
| 337 | if (matchlen > bmatchlen) | ||
| 338 | continue; | ||
| 339 | |||
| 340 | if (!IS_ERR_OR_NULL(dst)) | ||
| 341 | dst_release(dst); | ||
| 342 | dst = bdst; | ||
| 343 | matchlen = bmatchlen; | ||
| 332 | } | 344 | } |
| 333 | rcu_read_unlock(); | 345 | rcu_read_unlock(); |
| 334 | 346 | ||
| @@ -665,6 +677,9 @@ static struct sock *sctp_v6_create_accept_sk(struct sock *sk, | |||
| 665 | newnp = inet6_sk(newsk); | 677 | newnp = inet6_sk(newsk); |
| 666 | 678 | ||
| 667 | memcpy(newnp, np, sizeof(struct ipv6_pinfo)); | 679 | memcpy(newnp, np, sizeof(struct ipv6_pinfo)); |
| 680 | newnp->ipv6_mc_list = NULL; | ||
| 681 | newnp->ipv6_ac_list = NULL; | ||
| 682 | newnp->ipv6_fl_list = NULL; | ||
| 668 | 683 | ||
| 669 | rcu_read_lock(); | 684 | rcu_read_lock(); |
| 670 | opt = rcu_dereference(np->opt); | 685 | opt = rcu_dereference(np->opt); |
diff --git a/net/smc/Kconfig b/net/smc/Kconfig index c717ef0896aa..33954852f3f8 100644 --- a/net/smc/Kconfig +++ b/net/smc/Kconfig | |||
| @@ -8,6 +8,10 @@ config SMC | |||
| 8 | The Linux implementation of the SMC-R solution is designed as | 8 | The Linux implementation of the SMC-R solution is designed as |
| 9 | a separate socket family SMC. | 9 | a separate socket family SMC. |
| 10 | 10 | ||
| 11 | Warning: SMC will expose all memory for remote reads and writes | ||
| 12 | once a connection is established. Don't enable this option except | ||
| 13 | for tightly controlled lab environment. | ||
| 14 | |||
| 11 | Select this option if you want to run SMC socket applications | 15 | Select this option if you want to run SMC socket applications |
| 12 | 16 | ||
| 13 | config SMC_DIAG | 17 | config SMC_DIAG |
diff --git a/net/smc/smc_clc.c b/net/smc/smc_clc.c index e41f594a1e1d..03ec058d18df 100644 --- a/net/smc/smc_clc.c +++ b/net/smc/smc_clc.c | |||
| @@ -204,7 +204,7 @@ int smc_clc_send_confirm(struct smc_sock *smc) | |||
| 204 | memcpy(&cclc.lcl.mac, &link->smcibdev->mac[link->ibport - 1], ETH_ALEN); | 204 | memcpy(&cclc.lcl.mac, &link->smcibdev->mac[link->ibport - 1], ETH_ALEN); |
| 205 | hton24(cclc.qpn, link->roce_qp->qp_num); | 205 | hton24(cclc.qpn, link->roce_qp->qp_num); |
| 206 | cclc.rmb_rkey = | 206 | cclc.rmb_rkey = |
| 207 | htonl(conn->rmb_desc->mr_rx[SMC_SINGLE_LINK]->rkey); | 207 | htonl(conn->rmb_desc->rkey[SMC_SINGLE_LINK]); |
| 208 | cclc.conn_idx = 1; /* for now: 1 RMB = 1 RMBE */ | 208 | cclc.conn_idx = 1; /* for now: 1 RMB = 1 RMBE */ |
| 209 | cclc.rmbe_alert_token = htonl(conn->alert_token_local); | 209 | cclc.rmbe_alert_token = htonl(conn->alert_token_local); |
| 210 | cclc.qp_mtu = min(link->path_mtu, link->peer_mtu); | 210 | cclc.qp_mtu = min(link->path_mtu, link->peer_mtu); |
| @@ -256,7 +256,7 @@ int smc_clc_send_accept(struct smc_sock *new_smc, int srv_first_contact) | |||
| 256 | memcpy(&aclc.lcl.mac, link->smcibdev->mac[link->ibport - 1], ETH_ALEN); | 256 | memcpy(&aclc.lcl.mac, link->smcibdev->mac[link->ibport - 1], ETH_ALEN); |
| 257 | hton24(aclc.qpn, link->roce_qp->qp_num); | 257 | hton24(aclc.qpn, link->roce_qp->qp_num); |
| 258 | aclc.rmb_rkey = | 258 | aclc.rmb_rkey = |
| 259 | htonl(conn->rmb_desc->mr_rx[SMC_SINGLE_LINK]->rkey); | 259 | htonl(conn->rmb_desc->rkey[SMC_SINGLE_LINK]); |
| 260 | aclc.conn_idx = 1; /* as long as 1 RMB = 1 RMBE */ | 260 | aclc.conn_idx = 1; /* as long as 1 RMB = 1 RMBE */ |
| 261 | aclc.rmbe_alert_token = htonl(conn->alert_token_local); | 261 | aclc.rmbe_alert_token = htonl(conn->alert_token_local); |
| 262 | aclc.qp_mtu = link->path_mtu; | 262 | aclc.qp_mtu = link->path_mtu; |
diff --git a/net/smc/smc_core.c b/net/smc/smc_core.c index 65020e93ff21..3ac09a629ea1 100644 --- a/net/smc/smc_core.c +++ b/net/smc/smc_core.c | |||
| @@ -613,19 +613,8 @@ int smc_rmb_create(struct smc_sock *smc) | |||
| 613 | rmb_desc = NULL; | 613 | rmb_desc = NULL; |
| 614 | continue; /* if mapping failed, try smaller one */ | 614 | continue; /* if mapping failed, try smaller one */ |
| 615 | } | 615 | } |
| 616 | rc = smc_ib_get_memory_region(lgr->lnk[SMC_SINGLE_LINK].roce_pd, | 616 | rmb_desc->rkey[SMC_SINGLE_LINK] = |
| 617 | IB_ACCESS_REMOTE_WRITE | | 617 | lgr->lnk[SMC_SINGLE_LINK].roce_pd->unsafe_global_rkey; |
| 618 | IB_ACCESS_LOCAL_WRITE, | ||
| 619 | &rmb_desc->mr_rx[SMC_SINGLE_LINK]); | ||
| 620 | if (rc) { | ||
| 621 | smc_ib_buf_unmap(lgr->lnk[SMC_SINGLE_LINK].smcibdev, | ||
| 622 | tmp_bufsize, rmb_desc, | ||
| 623 | DMA_FROM_DEVICE); | ||
| 624 | kfree(rmb_desc->cpu_addr); | ||
| 625 | kfree(rmb_desc); | ||
| 626 | rmb_desc = NULL; | ||
| 627 | continue; | ||
| 628 | } | ||
| 629 | rmb_desc->used = 1; | 618 | rmb_desc->used = 1; |
| 630 | write_lock_bh(&lgr->rmbs_lock); | 619 | write_lock_bh(&lgr->rmbs_lock); |
| 631 | list_add(&rmb_desc->list, | 620 | list_add(&rmb_desc->list, |
| @@ -668,6 +657,7 @@ int smc_rmb_rtoken_handling(struct smc_connection *conn, | |||
| 668 | 657 | ||
| 669 | for (i = 0; i < SMC_RMBS_PER_LGR_MAX; i++) { | 658 | for (i = 0; i < SMC_RMBS_PER_LGR_MAX; i++) { |
| 670 | if ((lgr->rtokens[i][SMC_SINGLE_LINK].rkey == rkey) && | 659 | if ((lgr->rtokens[i][SMC_SINGLE_LINK].rkey == rkey) && |
| 660 | (lgr->rtokens[i][SMC_SINGLE_LINK].dma_addr == dma_addr) && | ||
| 671 | test_bit(i, lgr->rtokens_used_mask)) { | 661 | test_bit(i, lgr->rtokens_used_mask)) { |
| 672 | conn->rtoken_idx = i; | 662 | conn->rtoken_idx = i; |
| 673 | return 0; | 663 | return 0; |
diff --git a/net/smc/smc_core.h b/net/smc/smc_core.h index 27eb38056a27..b013cb43a327 100644 --- a/net/smc/smc_core.h +++ b/net/smc/smc_core.h | |||
| @@ -93,7 +93,7 @@ struct smc_buf_desc { | |||
| 93 | u64 dma_addr[SMC_LINKS_PER_LGR_MAX]; | 93 | u64 dma_addr[SMC_LINKS_PER_LGR_MAX]; |
| 94 | /* mapped address of buffer */ | 94 | /* mapped address of buffer */ |
| 95 | void *cpu_addr; /* virtual address of buffer */ | 95 | void *cpu_addr; /* virtual address of buffer */ |
| 96 | struct ib_mr *mr_rx[SMC_LINKS_PER_LGR_MAX]; | 96 | u32 rkey[SMC_LINKS_PER_LGR_MAX]; |
| 97 | /* for rmb only: | 97 | /* for rmb only: |
| 98 | * rkey provided to peer | 98 | * rkey provided to peer |
| 99 | */ | 99 | */ |
diff --git a/net/smc/smc_ib.c b/net/smc/smc_ib.c index cb69ab977cd7..b31715505a35 100644 --- a/net/smc/smc_ib.c +++ b/net/smc/smc_ib.c | |||
| @@ -37,24 +37,6 @@ u8 local_systemid[SMC_SYSTEMID_LEN] = SMC_LOCAL_SYSTEMID_RESET; /* unique system | |||
| 37 | * identifier | 37 | * identifier |
| 38 | */ | 38 | */ |
| 39 | 39 | ||
| 40 | int smc_ib_get_memory_region(struct ib_pd *pd, int access_flags, | ||
| 41 | struct ib_mr **mr) | ||
| 42 | { | ||
| 43 | int rc; | ||
| 44 | |||
| 45 | if (*mr) | ||
| 46 | return 0; /* already done */ | ||
| 47 | |||
| 48 | /* obtain unique key - | ||
| 49 | * next invocation of get_dma_mr returns a different key! | ||
| 50 | */ | ||
| 51 | *mr = pd->device->get_dma_mr(pd, access_flags); | ||
| 52 | rc = PTR_ERR_OR_ZERO(*mr); | ||
| 53 | if (IS_ERR(*mr)) | ||
| 54 | *mr = NULL; | ||
| 55 | return rc; | ||
| 56 | } | ||
| 57 | |||
| 58 | static int smc_ib_modify_qp_init(struct smc_link *lnk) | 40 | static int smc_ib_modify_qp_init(struct smc_link *lnk) |
| 59 | { | 41 | { |
| 60 | struct ib_qp_attr qp_attr; | 42 | struct ib_qp_attr qp_attr; |
| @@ -210,7 +192,8 @@ int smc_ib_create_protection_domain(struct smc_link *lnk) | |||
| 210 | { | 192 | { |
| 211 | int rc; | 193 | int rc; |
| 212 | 194 | ||
| 213 | lnk->roce_pd = ib_alloc_pd(lnk->smcibdev->ibdev, 0); | 195 | lnk->roce_pd = ib_alloc_pd(lnk->smcibdev->ibdev, |
| 196 | IB_PD_UNSAFE_GLOBAL_RKEY); | ||
| 214 | rc = PTR_ERR_OR_ZERO(lnk->roce_pd); | 197 | rc = PTR_ERR_OR_ZERO(lnk->roce_pd); |
| 215 | if (IS_ERR(lnk->roce_pd)) | 198 | if (IS_ERR(lnk->roce_pd)) |
| 216 | lnk->roce_pd = NULL; | 199 | lnk->roce_pd = NULL; |
diff --git a/net/smc/smc_ib.h b/net/smc/smc_ib.h index 7e1f0e24d177..b567152a526d 100644 --- a/net/smc/smc_ib.h +++ b/net/smc/smc_ib.h | |||
| @@ -61,8 +61,6 @@ void smc_ib_dealloc_protection_domain(struct smc_link *lnk); | |||
| 61 | int smc_ib_create_protection_domain(struct smc_link *lnk); | 61 | int smc_ib_create_protection_domain(struct smc_link *lnk); |
| 62 | void smc_ib_destroy_queue_pair(struct smc_link *lnk); | 62 | void smc_ib_destroy_queue_pair(struct smc_link *lnk); |
| 63 | int smc_ib_create_queue_pair(struct smc_link *lnk); | 63 | int smc_ib_create_queue_pair(struct smc_link *lnk); |
| 64 | int smc_ib_get_memory_region(struct ib_pd *pd, int access_flags, | ||
| 65 | struct ib_mr **mr); | ||
| 66 | int smc_ib_ready_link(struct smc_link *lnk); | 64 | int smc_ib_ready_link(struct smc_link *lnk); |
| 67 | int smc_ib_modify_qp_rts(struct smc_link *lnk); | 65 | int smc_ib_modify_qp_rts(struct smc_link *lnk); |
| 68 | int smc_ib_modify_qp_reset(struct smc_link *lnk); | 66 | int smc_ib_modify_qp_reset(struct smc_link *lnk); |
diff --git a/net/tipc/socket.c b/net/tipc/socket.c index 0d4f2f455a7c..1b92b72e812f 100644 --- a/net/tipc/socket.c +++ b/net/tipc/socket.c | |||
| @@ -362,25 +362,25 @@ static int tipc_sk_sock_err(struct socket *sock, long *timeout) | |||
| 362 | return 0; | 362 | return 0; |
| 363 | } | 363 | } |
| 364 | 364 | ||
| 365 | #define tipc_wait_for_cond(sock_, timeout_, condition_) \ | 365 | #define tipc_wait_for_cond(sock_, timeo_, condition_) \ |
| 366 | ({ \ | 366 | ({ \ |
| 367 | int rc_ = 0; \ | 367 | struct sock *sk_; \ |
| 368 | int done_ = 0; \ | 368 | int rc_; \ |
| 369 | \ | 369 | \ |
| 370 | while (!(condition_) && !done_) { \ | 370 | while ((rc_ = !(condition_))) { \ |
| 371 | struct sock *sk_ = sock->sk; \ | 371 | DEFINE_WAIT_FUNC(wait_, woken_wake_function); \ |
| 372 | DEFINE_WAIT_FUNC(wait_, woken_wake_function); \ | 372 | sk_ = (sock_)->sk; \ |
| 373 | \ | 373 | rc_ = tipc_sk_sock_err((sock_), timeo_); \ |
| 374 | rc_ = tipc_sk_sock_err(sock_, timeout_); \ | 374 | if (rc_) \ |
| 375 | if (rc_) \ | 375 | break; \ |
| 376 | break; \ | 376 | prepare_to_wait(sk_sleep(sk_), &wait_, TASK_INTERRUPTIBLE); \ |
| 377 | prepare_to_wait(sk_sleep(sk_), &wait_, \ | 377 | release_sock(sk_); \ |
| 378 | TASK_INTERRUPTIBLE); \ | 378 | *(timeo_) = wait_woken(&wait_, TASK_INTERRUPTIBLE, *(timeo_)); \ |
| 379 | done_ = sk_wait_event(sk_, timeout_, \ | 379 | sched_annotate_sleep(); \ |
| 380 | (condition_), &wait_); \ | 380 | lock_sock(sk_); \ |
| 381 | remove_wait_queue(sk_sleep(sk_), &wait_); \ | 381 | remove_wait_queue(sk_sleep(sk_), &wait_); \ |
| 382 | } \ | 382 | } \ |
| 383 | rc_; \ | 383 | rc_; \ |
| 384 | }) | 384 | }) |
| 385 | 385 | ||
| 386 | /** | 386 | /** |
diff --git a/net/x25/af_x25.c b/net/x25/af_x25.c index 8b911c29860e..5a1a98df3499 100644 --- a/net/x25/af_x25.c +++ b/net/x25/af_x25.c | |||
| @@ -1791,32 +1791,40 @@ void x25_kill_by_neigh(struct x25_neigh *nb) | |||
| 1791 | 1791 | ||
| 1792 | static int __init x25_init(void) | 1792 | static int __init x25_init(void) |
| 1793 | { | 1793 | { |
| 1794 | int rc = proto_register(&x25_proto, 0); | 1794 | int rc; |
| 1795 | 1795 | ||
| 1796 | if (rc != 0) | 1796 | rc = proto_register(&x25_proto, 0); |
| 1797 | if (rc) | ||
| 1797 | goto out; | 1798 | goto out; |
| 1798 | 1799 | ||
| 1799 | rc = sock_register(&x25_family_ops); | 1800 | rc = sock_register(&x25_family_ops); |
| 1800 | if (rc != 0) | 1801 | if (rc) |
| 1801 | goto out_proto; | 1802 | goto out_proto; |
| 1802 | 1803 | ||
| 1803 | dev_add_pack(&x25_packet_type); | 1804 | dev_add_pack(&x25_packet_type); |
| 1804 | 1805 | ||
| 1805 | rc = register_netdevice_notifier(&x25_dev_notifier); | 1806 | rc = register_netdevice_notifier(&x25_dev_notifier); |
| 1806 | if (rc != 0) | 1807 | if (rc) |
| 1807 | goto out_sock; | 1808 | goto out_sock; |
| 1808 | 1809 | ||
| 1809 | pr_info("Linux Version 0.2\n"); | 1810 | rc = x25_register_sysctl(); |
| 1811 | if (rc) | ||
| 1812 | goto out_dev; | ||
| 1810 | 1813 | ||
| 1811 | x25_register_sysctl(); | ||
| 1812 | rc = x25_proc_init(); | 1814 | rc = x25_proc_init(); |
| 1813 | if (rc != 0) | 1815 | if (rc) |
| 1814 | goto out_dev; | 1816 | goto out_sysctl; |
| 1817 | |||
| 1818 | pr_info("Linux Version 0.2\n"); | ||
| 1819 | |||
| 1815 | out: | 1820 | out: |
| 1816 | return rc; | 1821 | return rc; |
| 1822 | out_sysctl: | ||
| 1823 | x25_unregister_sysctl(); | ||
| 1817 | out_dev: | 1824 | out_dev: |
| 1818 | unregister_netdevice_notifier(&x25_dev_notifier); | 1825 | unregister_netdevice_notifier(&x25_dev_notifier); |
| 1819 | out_sock: | 1826 | out_sock: |
| 1827 | dev_remove_pack(&x25_packet_type); | ||
| 1820 | sock_unregister(AF_X25); | 1828 | sock_unregister(AF_X25); |
| 1821 | out_proto: | 1829 | out_proto: |
| 1822 | proto_unregister(&x25_proto); | 1830 | proto_unregister(&x25_proto); |
diff --git a/net/x25/sysctl_net_x25.c b/net/x25/sysctl_net_x25.c index a06dfe143c67..ba078c85f0a1 100644 --- a/net/x25/sysctl_net_x25.c +++ b/net/x25/sysctl_net_x25.c | |||
| @@ -73,9 +73,12 @@ static struct ctl_table x25_table[] = { | |||
| 73 | { }, | 73 | { }, |
| 74 | }; | 74 | }; |
| 75 | 75 | ||
| 76 | void __init x25_register_sysctl(void) | 76 | int __init x25_register_sysctl(void) |
| 77 | { | 77 | { |
| 78 | x25_table_header = register_net_sysctl(&init_net, "net/x25", x25_table); | 78 | x25_table_header = register_net_sysctl(&init_net, "net/x25", x25_table); |
| 79 | if (!x25_table_header) | ||
| 80 | return -ENOMEM; | ||
| 81 | return 0; | ||
| 79 | } | 82 | } |
| 80 | 83 | ||
| 81 | void x25_unregister_sysctl(void) | 84 | void x25_unregister_sysctl(void) |
diff --git a/samples/bpf/cookie_uid_helper_example.c b/samples/bpf/cookie_uid_helper_example.c index b08ab4e88929..9d751e209f31 100644 --- a/samples/bpf/cookie_uid_helper_example.c +++ b/samples/bpf/cookie_uid_helper_example.c | |||
| @@ -306,7 +306,9 @@ int main(int argc, char *argv[]) | |||
| 306 | prog_attach_iptables(argv[2]); | 306 | prog_attach_iptables(argv[2]); |
| 307 | if (cfg_test_traffic) { | 307 | if (cfg_test_traffic) { |
| 308 | if (signal(SIGINT, finish) == SIG_ERR) | 308 | if (signal(SIGINT, finish) == SIG_ERR) |
| 309 | error(1, errno, "register handler failed"); | 309 | error(1, errno, "register SIGINT handler failed"); |
| 310 | if (signal(SIGTERM, finish) == SIG_ERR) | ||
| 311 | error(1, errno, "register SIGTERM handler failed"); | ||
| 310 | while (!test_finish) { | 312 | while (!test_finish) { |
| 311 | print_table(); | 313 | print_table(); |
| 312 | printf("\n"); | 314 | printf("\n"); |
diff --git a/samples/bpf/offwaketime_user.c b/samples/bpf/offwaketime_user.c index 9cce2a66bd66..512f87a5fd20 100644 --- a/samples/bpf/offwaketime_user.c +++ b/samples/bpf/offwaketime_user.c | |||
| @@ -100,6 +100,7 @@ int main(int argc, char **argv) | |||
| 100 | setrlimit(RLIMIT_MEMLOCK, &r); | 100 | setrlimit(RLIMIT_MEMLOCK, &r); |
| 101 | 101 | ||
| 102 | signal(SIGINT, int_exit); | 102 | signal(SIGINT, int_exit); |
| 103 | signal(SIGTERM, int_exit); | ||
| 103 | 104 | ||
| 104 | if (load_kallsyms()) { | 105 | if (load_kallsyms()) { |
| 105 | printf("failed to process /proc/kallsyms\n"); | 106 | printf("failed to process /proc/kallsyms\n"); |
diff --git a/samples/bpf/sampleip_user.c b/samples/bpf/sampleip_user.c index be59d7dcbdde..4ed690b907ff 100644 --- a/samples/bpf/sampleip_user.c +++ b/samples/bpf/sampleip_user.c | |||
| @@ -180,6 +180,7 @@ int main(int argc, char **argv) | |||
| 180 | return 1; | 180 | return 1; |
| 181 | } | 181 | } |
| 182 | signal(SIGINT, int_exit); | 182 | signal(SIGINT, int_exit); |
| 183 | signal(SIGTERM, int_exit); | ||
| 183 | 184 | ||
| 184 | /* do sampling */ | 185 | /* do sampling */ |
| 185 | printf("Sampling at %d Hertz for %d seconds. Ctrl-C also ends.\n", | 186 | printf("Sampling at %d Hertz for %d seconds. Ctrl-C also ends.\n", |
diff --git a/samples/bpf/trace_event_user.c b/samples/bpf/trace_event_user.c index 0c5561d193a4..fa4336423da5 100644 --- a/samples/bpf/trace_event_user.c +++ b/samples/bpf/trace_event_user.c | |||
| @@ -192,6 +192,7 @@ int main(int argc, char **argv) | |||
| 192 | setrlimit(RLIMIT_MEMLOCK, &r); | 192 | setrlimit(RLIMIT_MEMLOCK, &r); |
| 193 | 193 | ||
| 194 | signal(SIGINT, int_exit); | 194 | signal(SIGINT, int_exit); |
| 195 | signal(SIGTERM, int_exit); | ||
| 195 | 196 | ||
| 196 | if (load_kallsyms()) { | 197 | if (load_kallsyms()) { |
| 197 | printf("failed to process /proc/kallsyms\n"); | 198 | printf("failed to process /proc/kallsyms\n"); |
diff --git a/samples/bpf/tracex2_user.c b/samples/bpf/tracex2_user.c index 7fee0f1ba9a3..7321a3f253c9 100644 --- a/samples/bpf/tracex2_user.c +++ b/samples/bpf/tracex2_user.c | |||
| @@ -127,6 +127,7 @@ int main(int ac, char **argv) | |||
| 127 | } | 127 | } |
| 128 | 128 | ||
| 129 | signal(SIGINT, int_exit); | 129 | signal(SIGINT, int_exit); |
| 130 | signal(SIGTERM, int_exit); | ||
| 130 | 131 | ||
| 131 | /* start 'ping' in the background to have some kfree_skb events */ | 132 | /* start 'ping' in the background to have some kfree_skb events */ |
| 132 | f = popen("ping -c5 localhost", "r"); | 133 | f = popen("ping -c5 localhost", "r"); |
diff --git a/samples/bpf/xdp1_user.c b/samples/bpf/xdp1_user.c index 378850c70eb8..2431c0321b71 100644 --- a/samples/bpf/xdp1_user.c +++ b/samples/bpf/xdp1_user.c | |||
| @@ -62,13 +62,14 @@ static void usage(const char *prog) | |||
| 62 | fprintf(stderr, | 62 | fprintf(stderr, |
| 63 | "usage: %s [OPTS] IFINDEX\n\n" | 63 | "usage: %s [OPTS] IFINDEX\n\n" |
| 64 | "OPTS:\n" | 64 | "OPTS:\n" |
| 65 | " -S use skb-mode\n", | 65 | " -S use skb-mode\n" |
| 66 | " -N enforce native mode\n", | ||
| 66 | prog); | 67 | prog); |
| 67 | } | 68 | } |
| 68 | 69 | ||
| 69 | int main(int argc, char **argv) | 70 | int main(int argc, char **argv) |
| 70 | { | 71 | { |
| 71 | const char *optstr = "S"; | 72 | const char *optstr = "SN"; |
| 72 | char filename[256]; | 73 | char filename[256]; |
| 73 | int opt; | 74 | int opt; |
| 74 | 75 | ||
| @@ -77,6 +78,9 @@ int main(int argc, char **argv) | |||
| 77 | case 'S': | 78 | case 'S': |
| 78 | xdp_flags |= XDP_FLAGS_SKB_MODE; | 79 | xdp_flags |= XDP_FLAGS_SKB_MODE; |
| 79 | break; | 80 | break; |
| 81 | case 'N': | ||
| 82 | xdp_flags |= XDP_FLAGS_DRV_MODE; | ||
| 83 | break; | ||
| 80 | default: | 84 | default: |
| 81 | usage(basename(argv[0])); | 85 | usage(basename(argv[0])); |
| 82 | return 1; | 86 | return 1; |
| @@ -102,6 +106,7 @@ int main(int argc, char **argv) | |||
| 102 | } | 106 | } |
| 103 | 107 | ||
| 104 | signal(SIGINT, int_exit); | 108 | signal(SIGINT, int_exit); |
| 109 | signal(SIGTERM, int_exit); | ||
| 105 | 110 | ||
| 106 | if (set_link_xdp_fd(ifindex, prog_fd[0], xdp_flags) < 0) { | 111 | if (set_link_xdp_fd(ifindex, prog_fd[0], xdp_flags) < 0) { |
| 107 | printf("link set xdp fd failed\n"); | 112 | printf("link set xdp fd failed\n"); |
diff --git a/samples/bpf/xdp_tx_iptunnel_user.c b/samples/bpf/xdp_tx_iptunnel_user.c index 92b8bde9337c..715cd12eaca5 100644 --- a/samples/bpf/xdp_tx_iptunnel_user.c +++ b/samples/bpf/xdp_tx_iptunnel_user.c | |||
| @@ -79,6 +79,8 @@ static void usage(const char *cmd) | |||
| 79 | printf(" -m <dest-MAC> Used in sending the IP Tunneled pkt\n"); | 79 | printf(" -m <dest-MAC> Used in sending the IP Tunneled pkt\n"); |
| 80 | printf(" -T <stop-after-X-seconds> Default: 0 (forever)\n"); | 80 | printf(" -T <stop-after-X-seconds> Default: 0 (forever)\n"); |
| 81 | printf(" -P <IP-Protocol> Default is TCP\n"); | 81 | printf(" -P <IP-Protocol> Default is TCP\n"); |
| 82 | printf(" -S use skb-mode\n"); | ||
| 83 | printf(" -N enforce native mode\n"); | ||
| 82 | printf(" -h Display this help\n"); | 84 | printf(" -h Display this help\n"); |
| 83 | } | 85 | } |
| 84 | 86 | ||
| @@ -138,7 +140,7 @@ int main(int argc, char **argv) | |||
| 138 | { | 140 | { |
| 139 | unsigned char opt_flags[256] = {}; | 141 | unsigned char opt_flags[256] = {}; |
| 140 | unsigned int kill_after_s = 0; | 142 | unsigned int kill_after_s = 0; |
| 141 | const char *optstr = "i:a:p:s:d:m:T:P:Sh"; | 143 | const char *optstr = "i:a:p:s:d:m:T:P:SNh"; |
| 142 | int min_port = 0, max_port = 0; | 144 | int min_port = 0, max_port = 0; |
| 143 | struct iptnl_info tnl = {}; | 145 | struct iptnl_info tnl = {}; |
| 144 | struct rlimit r = {RLIM_INFINITY, RLIM_INFINITY}; | 146 | struct rlimit r = {RLIM_INFINITY, RLIM_INFINITY}; |
| @@ -206,6 +208,9 @@ int main(int argc, char **argv) | |||
| 206 | case 'S': | 208 | case 'S': |
| 207 | xdp_flags |= XDP_FLAGS_SKB_MODE; | 209 | xdp_flags |= XDP_FLAGS_SKB_MODE; |
| 208 | break; | 210 | break; |
| 211 | case 'N': | ||
| 212 | xdp_flags |= XDP_FLAGS_DRV_MODE; | ||
| 213 | break; | ||
| 209 | default: | 214 | default: |
| 210 | usage(argv[0]); | 215 | usage(argv[0]); |
| 211 | return 1; | 216 | return 1; |
| @@ -239,6 +244,7 @@ int main(int argc, char **argv) | |||
| 239 | } | 244 | } |
| 240 | 245 | ||
| 241 | signal(SIGINT, int_exit); | 246 | signal(SIGINT, int_exit); |
| 247 | signal(SIGTERM, int_exit); | ||
| 242 | 248 | ||
| 243 | while (min_port <= max_port) { | 249 | while (min_port <= max_port) { |
| 244 | vip.dport = htons(min_port++); | 250 | vip.dport = htons(min_port++); |
diff --git a/scripts/Makefile.headersinst b/scripts/Makefile.headersinst index 6ba97a1f9c5a..ce753a408c56 100644 --- a/scripts/Makefile.headersinst +++ b/scripts/Makefile.headersinst | |||
| @@ -8,6 +8,29 @@ | |||
| 8 | # | 8 | # |
| 9 | # ========================================================================== | 9 | # ========================================================================== |
| 10 | 10 | ||
| 11 | PHONY := __headers | ||
| 12 | __headers: | ||
| 13 | |||
| 14 | include scripts/Kbuild.include | ||
| 15 | |||
| 16 | srcdir := $(srctree)/$(obj) | ||
| 17 | subdirs := $(patsubst $(srcdir)/%/.,%,$(wildcard $(srcdir)/*/.)) | ||
| 18 | # caller may set destination dir (when installing to asm/) | ||
| 19 | _dst := $(if $(dst),$(dst),$(obj)) | ||
| 20 | |||
| 21 | # Recursion | ||
| 22 | __headers: $(subdirs) | ||
| 23 | |||
| 24 | .PHONY: $(subdirs) | ||
| 25 | $(subdirs): | ||
| 26 | $(Q)$(MAKE) $(hdr-inst)=$(obj)/$@ dst=$(_dst)/$@ | ||
| 27 | |||
| 28 | # Skip header install/check for include/uapi and arch/$(hdr-arch)/include/uapi. | ||
| 29 | # We have only sub-directories there. | ||
| 30 | skip-inst := $(if $(filter %/uapi,$(obj)),1) | ||
| 31 | |||
| 32 | ifeq ($(skip-inst),) | ||
| 33 | |||
| 11 | # generated header directory | 34 | # generated header directory |
| 12 | gen := $(if $(gen),$(gen),$(subst include/,include/generated/,$(obj))) | 35 | gen := $(if $(gen),$(gen),$(subst include/,include/generated/,$(obj))) |
| 13 | 36 | ||
| @@ -15,21 +38,14 @@ gen := $(if $(gen),$(gen),$(subst include/,include/generated/,$(obj))) | |||
| 15 | kbuild-file := $(srctree)/$(obj)/Kbuild | 38 | kbuild-file := $(srctree)/$(obj)/Kbuild |
| 16 | -include $(kbuild-file) | 39 | -include $(kbuild-file) |
| 17 | 40 | ||
| 18 | # called may set destination dir (when installing to asm/) | ||
| 19 | _dst := $(if $(dst),$(dst),$(obj)) | ||
| 20 | |||
| 21 | old-kbuild-file := $(srctree)/$(subst uapi/,,$(obj))/Kbuild | 41 | old-kbuild-file := $(srctree)/$(subst uapi/,,$(obj))/Kbuild |
| 22 | ifneq ($(wildcard $(old-kbuild-file)),) | 42 | ifneq ($(wildcard $(old-kbuild-file)),) |
| 23 | include $(old-kbuild-file) | 43 | include $(old-kbuild-file) |
| 24 | endif | 44 | endif |
| 25 | 45 | ||
| 26 | include scripts/Kbuild.include | ||
| 27 | |||
| 28 | installdir := $(INSTALL_HDR_PATH)/$(subst uapi/,,$(_dst)) | 46 | installdir := $(INSTALL_HDR_PATH)/$(subst uapi/,,$(_dst)) |
| 29 | 47 | ||
| 30 | srcdir := $(srctree)/$(obj) | ||
| 31 | gendir := $(objtree)/$(gen) | 48 | gendir := $(objtree)/$(gen) |
| 32 | subdirs := $(patsubst $(srcdir)/%/.,%,$(wildcard $(srcdir)/*/.)) | ||
| 33 | header-files := $(notdir $(wildcard $(srcdir)/*.h)) | 49 | header-files := $(notdir $(wildcard $(srcdir)/*.h)) |
| 34 | header-files += $(notdir $(wildcard $(srcdir)/*.agh)) | 50 | header-files += $(notdir $(wildcard $(srcdir)/*.agh)) |
| 35 | header-files := $(filter-out $(no-export-headers), $(header-files)) | 51 | header-files := $(filter-out $(no-export-headers), $(header-files)) |
| @@ -88,11 +104,9 @@ quiet_cmd_check = CHECK $(printdir) ($(words $(all-files)) files) | |||
| 88 | $(PERL) $< $(INSTALL_HDR_PATH)/include $(SRCARCH); \ | 104 | $(PERL) $< $(INSTALL_HDR_PATH)/include $(SRCARCH); \ |
| 89 | touch $@ | 105 | touch $@ |
| 90 | 106 | ||
| 91 | PHONY += __headersinst __headerscheck | ||
| 92 | |||
| 93 | ifndef HDRCHECK | 107 | ifndef HDRCHECK |
| 94 | # Rules for installing headers | 108 | # Rules for installing headers |
| 95 | __headersinst: $(subdirs) $(install-file) | 109 | __headers: $(install-file) |
| 96 | @: | 110 | @: |
| 97 | 111 | ||
| 98 | targets += $(install-file) | 112 | targets += $(install-file) |
| @@ -104,7 +118,7 @@ $(install-file): scripts/headers_install.sh \ | |||
| 104 | $(call if_changed,install) | 118 | $(call if_changed,install) |
| 105 | 119 | ||
| 106 | else | 120 | else |
| 107 | __headerscheck: $(subdirs) $(check-file) | 121 | __headers: $(check-file) |
| 108 | @: | 122 | @: |
| 109 | 123 | ||
| 110 | targets += $(check-file) | 124 | targets += $(check-file) |
| @@ -113,11 +127,6 @@ $(check-file): scripts/headers_check.pl $(output-files) FORCE | |||
| 113 | 127 | ||
| 114 | endif | 128 | endif |
| 115 | 129 | ||
| 116 | # Recursion | ||
| 117 | .PHONY: $(subdirs) | ||
| 118 | $(subdirs): | ||
| 119 | $(Q)$(MAKE) $(hdr-inst)=$(obj)/$@ dst=$(_dst)/$@ | ||
| 120 | |||
| 121 | targets := $(wildcard $(sort $(targets))) | 130 | targets := $(wildcard $(sort $(targets))) |
| 122 | cmd_files := $(wildcard \ | 131 | cmd_files := $(wildcard \ |
| 123 | $(foreach f,$(targets),$(dir $(f)).$(notdir $(f)).cmd)) | 132 | $(foreach f,$(targets),$(dir $(f)).$(notdir $(f)).cmd)) |
| @@ -126,6 +135,8 @@ ifneq ($(cmd_files),) | |||
| 126 | include $(cmd_files) | 135 | include $(cmd_files) |
| 127 | endif | 136 | endif |
| 128 | 137 | ||
| 138 | endif # skip-inst | ||
| 139 | |||
| 129 | .PHONY: $(PHONY) | 140 | .PHONY: $(PHONY) |
| 130 | PHONY += FORCE | 141 | PHONY += FORCE |
| 131 | FORCE: ; | 142 | FORCE: ; |
diff --git a/tools/build/feature/test-bpf.c b/tools/build/feature/test-bpf.c index ebc6dceddb58..7598361ef1f1 100644 --- a/tools/build/feature/test-bpf.c +++ b/tools/build/feature/test-bpf.c | |||
| @@ -29,6 +29,7 @@ int main(void) | |||
| 29 | attr.log_size = 0; | 29 | attr.log_size = 0; |
| 30 | attr.log_level = 0; | 30 | attr.log_level = 0; |
| 31 | attr.kern_version = 0; | 31 | attr.kern_version = 0; |
| 32 | attr.prog_flags = 0; | ||
| 32 | 33 | ||
| 33 | /* | 34 | /* |
| 34 | * Test existence of __NR_bpf and BPF_PROG_LOAD. | 35 | * Test existence of __NR_bpf and BPF_PROG_LOAD. |
diff --git a/tools/include/uapi/linux/bpf.h b/tools/include/uapi/linux/bpf.h index e553529929f6..94dfa9def355 100644 --- a/tools/include/uapi/linux/bpf.h +++ b/tools/include/uapi/linux/bpf.h | |||
| @@ -132,6 +132,13 @@ enum bpf_attach_type { | |||
| 132 | */ | 132 | */ |
| 133 | #define BPF_F_ALLOW_OVERRIDE (1U << 0) | 133 | #define BPF_F_ALLOW_OVERRIDE (1U << 0) |
| 134 | 134 | ||
| 135 | /* If BPF_F_STRICT_ALIGNMENT is used in BPF_PROG_LOAD command, the | ||
| 136 | * verifier will perform strict alignment checking as if the kernel | ||
| 137 | * has been built with CONFIG_EFFICIENT_UNALIGNED_ACCESS not set, | ||
| 138 | * and NET_IP_ALIGN defined to 2. | ||
| 139 | */ | ||
| 140 | #define BPF_F_STRICT_ALIGNMENT (1U << 0) | ||
| 141 | |||
| 135 | #define BPF_PSEUDO_MAP_FD 1 | 142 | #define BPF_PSEUDO_MAP_FD 1 |
| 136 | 143 | ||
| 137 | /* flags for BPF_MAP_UPDATE_ELEM command */ | 144 | /* flags for BPF_MAP_UPDATE_ELEM command */ |
| @@ -177,6 +184,7 @@ union bpf_attr { | |||
| 177 | __u32 log_size; /* size of user buffer */ | 184 | __u32 log_size; /* size of user buffer */ |
| 178 | __aligned_u64 log_buf; /* user supplied buffer */ | 185 | __aligned_u64 log_buf; /* user supplied buffer */ |
| 179 | __u32 kern_version; /* checked when prog_type=kprobe */ | 186 | __u32 kern_version; /* checked when prog_type=kprobe */ |
| 187 | __u32 prog_flags; | ||
| 180 | }; | 188 | }; |
| 181 | 189 | ||
| 182 | struct { /* anonymous struct used by BPF_OBJ_* commands */ | 190 | struct { /* anonymous struct used by BPF_OBJ_* commands */ |
| @@ -481,8 +489,7 @@ union bpf_attr { | |||
| 481 | * u32 bpf_get_socket_uid(skb) | 489 | * u32 bpf_get_socket_uid(skb) |
| 482 | * Get the owner uid of the socket stored inside sk_buff. | 490 | * Get the owner uid of the socket stored inside sk_buff. |
| 483 | * @skb: pointer to skb | 491 | * @skb: pointer to skb |
| 484 | * Return: uid of the socket owner on success or 0 if the socket pointer | 492 | * Return: uid of the socket owner on success or overflowuid if failed. |
| 485 | * inside sk_buff is NULL | ||
| 486 | */ | 493 | */ |
| 487 | #define __BPF_FUNC_MAPPER(FN) \ | 494 | #define __BPF_FUNC_MAPPER(FN) \ |
| 488 | FN(unspec), \ | 495 | FN(unspec), \ |
diff --git a/tools/lib/bpf/bpf.c b/tools/lib/bpf/bpf.c index 4fe444b8092e..6e178987af8e 100644 --- a/tools/lib/bpf/bpf.c +++ b/tools/lib/bpf/bpf.c | |||
| @@ -117,6 +117,28 @@ int bpf_load_program(enum bpf_prog_type type, const struct bpf_insn *insns, | |||
| 117 | return sys_bpf(BPF_PROG_LOAD, &attr, sizeof(attr)); | 117 | return sys_bpf(BPF_PROG_LOAD, &attr, sizeof(attr)); |
| 118 | } | 118 | } |
| 119 | 119 | ||
| 120 | int bpf_verify_program(enum bpf_prog_type type, const struct bpf_insn *insns, | ||
| 121 | size_t insns_cnt, int strict_alignment, | ||
| 122 | const char *license, __u32 kern_version, | ||
| 123 | char *log_buf, size_t log_buf_sz) | ||
| 124 | { | ||
| 125 | union bpf_attr attr; | ||
| 126 | |||
| 127 | bzero(&attr, sizeof(attr)); | ||
| 128 | attr.prog_type = type; | ||
| 129 | attr.insn_cnt = (__u32)insns_cnt; | ||
| 130 | attr.insns = ptr_to_u64(insns); | ||
| 131 | attr.license = ptr_to_u64(license); | ||
| 132 | attr.log_buf = ptr_to_u64(log_buf); | ||
| 133 | attr.log_size = log_buf_sz; | ||
| 134 | attr.log_level = 2; | ||
| 135 | log_buf[0] = 0; | ||
| 136 | attr.kern_version = kern_version; | ||
| 137 | attr.prog_flags = strict_alignment ? BPF_F_STRICT_ALIGNMENT : 0; | ||
| 138 | |||
| 139 | return sys_bpf(BPF_PROG_LOAD, &attr, sizeof(attr)); | ||
| 140 | } | ||
| 141 | |||
| 120 | int bpf_map_update_elem(int fd, const void *key, const void *value, | 142 | int bpf_map_update_elem(int fd, const void *key, const void *value, |
| 121 | __u64 flags) | 143 | __u64 flags) |
| 122 | { | 144 | { |
diff --git a/tools/lib/bpf/bpf.h b/tools/lib/bpf/bpf.h index edb4daeff7a5..972bd8333eb7 100644 --- a/tools/lib/bpf/bpf.h +++ b/tools/lib/bpf/bpf.h | |||
| @@ -35,6 +35,10 @@ int bpf_load_program(enum bpf_prog_type type, const struct bpf_insn *insns, | |||
| 35 | size_t insns_cnt, const char *license, | 35 | size_t insns_cnt, const char *license, |
| 36 | __u32 kern_version, char *log_buf, | 36 | __u32 kern_version, char *log_buf, |
| 37 | size_t log_buf_sz); | 37 | size_t log_buf_sz); |
| 38 | int bpf_verify_program(enum bpf_prog_type type, const struct bpf_insn *insns, | ||
| 39 | size_t insns_cnt, int strict_alignment, | ||
| 40 | const char *license, __u32 kern_version, | ||
| 41 | char *log_buf, size_t log_buf_sz); | ||
| 38 | 42 | ||
| 39 | int bpf_map_update_elem(int fd, const void *key, const void *value, | 43 | int bpf_map_update_elem(int fd, const void *key, const void *value, |
| 40 | __u64 flags); | 44 | __u64 flags); |
diff --git a/tools/testing/selftests/bpf/Makefile b/tools/testing/selftests/bpf/Makefile index 91edd0566237..f389b02d43a0 100644 --- a/tools/testing/selftests/bpf/Makefile +++ b/tools/testing/selftests/bpf/Makefile | |||
| @@ -11,7 +11,8 @@ endif | |||
| 11 | CFLAGS += -Wall -O2 -I$(APIDIR) -I$(LIBDIR) -I$(GENDIR) $(GENFLAGS) -I../../../include | 11 | CFLAGS += -Wall -O2 -I$(APIDIR) -I$(LIBDIR) -I$(GENDIR) $(GENFLAGS) -I../../../include |
| 12 | LDLIBS += -lcap -lelf | 12 | LDLIBS += -lcap -lelf |
| 13 | 13 | ||
| 14 | TEST_GEN_PROGS = test_verifier test_tag test_maps test_lru_map test_lpm_map test_progs | 14 | TEST_GEN_PROGS = test_verifier test_tag test_maps test_lru_map test_lpm_map test_progs \ |
| 15 | test_align | ||
| 15 | 16 | ||
| 16 | TEST_GEN_FILES = test_pkt_access.o test_xdp.o test_l4lb.o test_tcp_estats.o | 17 | TEST_GEN_FILES = test_pkt_access.o test_xdp.o test_l4lb.o test_tcp_estats.o |
| 17 | 18 | ||
| @@ -34,6 +35,7 @@ $(BPFOBJ): force | |||
| 34 | CLANG ?= clang | 35 | CLANG ?= clang |
| 35 | 36 | ||
| 36 | %.o: %.c | 37 | %.o: %.c |
| 37 | $(CLANG) -I. -I../../../include/uapi -I../../../../samples/bpf/ \ | 38 | $(CLANG) -I. -I./include/uapi -I../../../include/uapi \ |
| 39 | -I../../../../samples/bpf/ \ | ||
| 38 | -Wno-compare-distinct-pointer-types \ | 40 | -Wno-compare-distinct-pointer-types \ |
| 39 | -O2 -target bpf -c $< -o $@ | 41 | -O2 -target bpf -c $< -o $@ |
diff --git a/tools/testing/selftests/bpf/include/uapi/linux/types.h b/tools/testing/selftests/bpf/include/uapi/linux/types.h new file mode 100644 index 000000000000..51841848fbfe --- /dev/null +++ b/tools/testing/selftests/bpf/include/uapi/linux/types.h | |||
| @@ -0,0 +1,22 @@ | |||
| 1 | #ifndef _UAPI_LINUX_TYPES_H | ||
| 2 | #define _UAPI_LINUX_TYPES_H | ||
| 3 | |||
| 4 | #include <asm-generic/int-ll64.h> | ||
| 5 | |||
| 6 | /* copied from linux:include/uapi/linux/types.h */ | ||
| 7 | #define __bitwise | ||
| 8 | typedef __u16 __bitwise __le16; | ||
| 9 | typedef __u16 __bitwise __be16; | ||
| 10 | typedef __u32 __bitwise __le32; | ||
| 11 | typedef __u32 __bitwise __be32; | ||
| 12 | typedef __u64 __bitwise __le64; | ||
| 13 | typedef __u64 __bitwise __be64; | ||
| 14 | |||
| 15 | typedef __u16 __bitwise __sum16; | ||
| 16 | typedef __u32 __bitwise __wsum; | ||
| 17 | |||
| 18 | #define __aligned_u64 __u64 __attribute__((aligned(8))) | ||
| 19 | #define __aligned_be64 __be64 __attribute__((aligned(8))) | ||
| 20 | #define __aligned_le64 __le64 __attribute__((aligned(8))) | ||
| 21 | |||
| 22 | #endif /* _UAPI_LINUX_TYPES_H */ | ||
diff --git a/tools/testing/selftests/bpf/test_align.c b/tools/testing/selftests/bpf/test_align.c new file mode 100644 index 000000000000..9644d4e069de --- /dev/null +++ b/tools/testing/selftests/bpf/test_align.c | |||
| @@ -0,0 +1,453 @@ | |||
| 1 | #include <asm/types.h> | ||
| 2 | #include <linux/types.h> | ||
| 3 | #include <stdint.h> | ||
| 4 | #include <stdio.h> | ||
| 5 | #include <stdlib.h> | ||
| 6 | #include <unistd.h> | ||
| 7 | #include <errno.h> | ||
| 8 | #include <string.h> | ||
| 9 | #include <stddef.h> | ||
| 10 | #include <stdbool.h> | ||
| 11 | |||
| 12 | #include <linux/unistd.h> | ||
| 13 | #include <linux/filter.h> | ||
| 14 | #include <linux/bpf_perf_event.h> | ||
| 15 | #include <linux/bpf.h> | ||
| 16 | |||
| 17 | #include <bpf/bpf.h> | ||
| 18 | |||
| 19 | #include "../../../include/linux/filter.h" | ||
| 20 | |||
| 21 | #ifndef ARRAY_SIZE | ||
| 22 | # define ARRAY_SIZE(x) (sizeof(x) / sizeof((x)[0])) | ||
| 23 | #endif | ||
| 24 | |||
| 25 | #define MAX_INSNS 512 | ||
| 26 | #define MAX_MATCHES 16 | ||
| 27 | |||
| 28 | struct bpf_align_test { | ||
| 29 | const char *descr; | ||
| 30 | struct bpf_insn insns[MAX_INSNS]; | ||
| 31 | enum { | ||
| 32 | UNDEF, | ||
| 33 | ACCEPT, | ||
| 34 | REJECT | ||
| 35 | } result; | ||
| 36 | enum bpf_prog_type prog_type; | ||
| 37 | const char *matches[MAX_MATCHES]; | ||
| 38 | }; | ||
| 39 | |||
| 40 | static struct bpf_align_test tests[] = { | ||
| 41 | { | ||
| 42 | .descr = "mov", | ||
| 43 | .insns = { | ||
| 44 | BPF_MOV64_IMM(BPF_REG_3, 2), | ||
| 45 | BPF_MOV64_IMM(BPF_REG_3, 4), | ||
| 46 | BPF_MOV64_IMM(BPF_REG_3, 8), | ||
| 47 | BPF_MOV64_IMM(BPF_REG_3, 16), | ||
| 48 | BPF_MOV64_IMM(BPF_REG_3, 32), | ||
| 49 | BPF_MOV64_IMM(BPF_REG_0, 0), | ||
| 50 | BPF_EXIT_INSN(), | ||
| 51 | }, | ||
| 52 | .prog_type = BPF_PROG_TYPE_SCHED_CLS, | ||
| 53 | .matches = { | ||
| 54 | "1: R1=ctx R3=imm2,min_value=2,max_value=2,min_align=2 R10=fp", | ||
| 55 | "2: R1=ctx R3=imm4,min_value=4,max_value=4,min_align=4 R10=fp", | ||
| 56 | "3: R1=ctx R3=imm8,min_value=8,max_value=8,min_align=8 R10=fp", | ||
| 57 | "4: R1=ctx R3=imm16,min_value=16,max_value=16,min_align=16 R10=fp", | ||
| 58 | "5: R1=ctx R3=imm32,min_value=32,max_value=32,min_align=32 R10=fp", | ||
| 59 | }, | ||
| 60 | }, | ||
| 61 | { | ||
| 62 | .descr = "shift", | ||
| 63 | .insns = { | ||
| 64 | BPF_MOV64_IMM(BPF_REG_3, 1), | ||
| 65 | BPF_ALU64_IMM(BPF_LSH, BPF_REG_3, 1), | ||
| 66 | BPF_ALU64_IMM(BPF_LSH, BPF_REG_3, 1), | ||
| 67 | BPF_ALU64_IMM(BPF_LSH, BPF_REG_3, 1), | ||
| 68 | BPF_ALU64_IMM(BPF_LSH, BPF_REG_3, 1), | ||
| 69 | BPF_ALU64_IMM(BPF_RSH, BPF_REG_3, 4), | ||
| 70 | BPF_MOV64_IMM(BPF_REG_4, 32), | ||
| 71 | BPF_ALU64_IMM(BPF_RSH, BPF_REG_4, 1), | ||
| 72 | BPF_ALU64_IMM(BPF_RSH, BPF_REG_4, 1), | ||
| 73 | BPF_ALU64_IMM(BPF_RSH, BPF_REG_4, 1), | ||
| 74 | BPF_ALU64_IMM(BPF_RSH, BPF_REG_4, 1), | ||
| 75 | BPF_MOV64_IMM(BPF_REG_0, 0), | ||
| 76 | BPF_EXIT_INSN(), | ||
| 77 | }, | ||
| 78 | .prog_type = BPF_PROG_TYPE_SCHED_CLS, | ||
| 79 | .matches = { | ||
| 80 | "1: R1=ctx R3=imm1,min_value=1,max_value=1,min_align=1 R10=fp", | ||
| 81 | "2: R1=ctx R3=imm2,min_value=2,max_value=2,min_align=2 R10=fp", | ||
| 82 | "3: R1=ctx R3=imm4,min_value=4,max_value=4,min_align=4 R10=fp", | ||
| 83 | "4: R1=ctx R3=imm8,min_value=8,max_value=8,min_align=8 R10=fp", | ||
| 84 | "5: R1=ctx R3=imm16,min_value=16,max_value=16,min_align=16 R10=fp", | ||
| 85 | "6: R1=ctx R3=imm1,min_value=1,max_value=1,min_align=1 R10=fp", | ||
| 86 | "7: R1=ctx R3=imm1,min_value=1,max_value=1,min_align=1 R4=imm32,min_value=32,max_value=32,min_align=32 R10=fp", | ||
| 87 | "8: R1=ctx R3=imm1,min_value=1,max_value=1,min_align=1 R4=imm16,min_value=16,max_value=16,min_align=16 R10=fp", | ||
| 88 | "9: R1=ctx R3=imm1,min_value=1,max_value=1,min_align=1 R4=imm8,min_value=8,max_value=8,min_align=8 R10=fp", | ||
| 89 | "10: R1=ctx R3=imm1,min_value=1,max_value=1,min_align=1 R4=imm4,min_value=4,max_value=4,min_align=4 R10=fp", | ||
| 90 | "11: R1=ctx R3=imm1,min_value=1,max_value=1,min_align=1 R4=imm2,min_value=2,max_value=2,min_align=2 R10=fp", | ||
| 91 | }, | ||
| 92 | }, | ||
| 93 | { | ||
| 94 | .descr = "addsub", | ||
| 95 | .insns = { | ||
| 96 | BPF_MOV64_IMM(BPF_REG_3, 4), | ||
| 97 | BPF_ALU64_IMM(BPF_ADD, BPF_REG_3, 4), | ||
| 98 | BPF_ALU64_IMM(BPF_ADD, BPF_REG_3, 2), | ||
| 99 | BPF_MOV64_IMM(BPF_REG_4, 8), | ||
| 100 | BPF_ALU64_IMM(BPF_ADD, BPF_REG_4, 4), | ||
| 101 | BPF_ALU64_IMM(BPF_ADD, BPF_REG_4, 2), | ||
| 102 | BPF_MOV64_IMM(BPF_REG_0, 0), | ||
| 103 | BPF_EXIT_INSN(), | ||
| 104 | }, | ||
| 105 | .prog_type = BPF_PROG_TYPE_SCHED_CLS, | ||
| 106 | .matches = { | ||
| 107 | "1: R1=ctx R3=imm4,min_value=4,max_value=4,min_align=4 R10=fp", | ||
| 108 | "2: R1=ctx R3=imm8,min_value=8,max_value=8,min_align=4 R10=fp", | ||
| 109 | "3: R1=ctx R3=imm10,min_value=10,max_value=10,min_align=2 R10=fp", | ||
| 110 | "4: R1=ctx R3=imm10,min_value=10,max_value=10,min_align=2 R4=imm8,min_value=8,max_value=8,min_align=8 R10=fp", | ||
| 111 | "5: R1=ctx R3=imm10,min_value=10,max_value=10,min_align=2 R4=imm12,min_value=12,max_value=12,min_align=4 R10=fp", | ||
| 112 | "6: R1=ctx R3=imm10,min_value=10,max_value=10,min_align=2 R4=imm14,min_value=14,max_value=14,min_align=2 R10=fp", | ||
| 113 | }, | ||
| 114 | }, | ||
| 115 | { | ||
| 116 | .descr = "mul", | ||
| 117 | .insns = { | ||
| 118 | BPF_MOV64_IMM(BPF_REG_3, 7), | ||
| 119 | BPF_ALU64_IMM(BPF_MUL, BPF_REG_3, 1), | ||
| 120 | BPF_ALU64_IMM(BPF_MUL, BPF_REG_3, 2), | ||
| 121 | BPF_ALU64_IMM(BPF_MUL, BPF_REG_3, 4), | ||
| 122 | BPF_MOV64_IMM(BPF_REG_0, 0), | ||
| 123 | BPF_EXIT_INSN(), | ||
| 124 | }, | ||
| 125 | .prog_type = BPF_PROG_TYPE_SCHED_CLS, | ||
| 126 | .matches = { | ||
| 127 | "1: R1=ctx R3=imm7,min_value=7,max_value=7,min_align=1 R10=fp", | ||
| 128 | "2: R1=ctx R3=imm7,min_value=7,max_value=7,min_align=1 R10=fp", | ||
| 129 | "3: R1=ctx R3=imm14,min_value=14,max_value=14,min_align=2 R10=fp", | ||
| 130 | "4: R1=ctx R3=imm56,min_value=56,max_value=56,min_align=4 R10=fp", | ||
| 131 | }, | ||
| 132 | }, | ||
| 133 | |||
| 134 | #define PREP_PKT_POINTERS \ | ||
| 135 | BPF_LDX_MEM(BPF_W, BPF_REG_2, BPF_REG_1, \ | ||
| 136 | offsetof(struct __sk_buff, data)), \ | ||
| 137 | BPF_LDX_MEM(BPF_W, BPF_REG_3, BPF_REG_1, \ | ||
| 138 | offsetof(struct __sk_buff, data_end)) | ||
| 139 | |||
| 140 | #define LOAD_UNKNOWN(DST_REG) \ | ||
| 141 | PREP_PKT_POINTERS, \ | ||
| 142 | BPF_MOV64_REG(BPF_REG_0, BPF_REG_2), \ | ||
| 143 | BPF_ALU64_IMM(BPF_ADD, BPF_REG_0, 8), \ | ||
| 144 | BPF_JMP_REG(BPF_JGE, BPF_REG_3, BPF_REG_0, 1), \ | ||
| 145 | BPF_EXIT_INSN(), \ | ||
| 146 | BPF_LDX_MEM(BPF_B, DST_REG, BPF_REG_2, 0) | ||
| 147 | |||
| 148 | { | ||
| 149 | .descr = "unknown shift", | ||
| 150 | .insns = { | ||
| 151 | LOAD_UNKNOWN(BPF_REG_3), | ||
| 152 | BPF_ALU64_IMM(BPF_LSH, BPF_REG_3, 1), | ||
| 153 | BPF_ALU64_IMM(BPF_LSH, BPF_REG_3, 1), | ||
| 154 | BPF_ALU64_IMM(BPF_LSH, BPF_REG_3, 1), | ||
| 155 | BPF_ALU64_IMM(BPF_LSH, BPF_REG_3, 1), | ||
| 156 | LOAD_UNKNOWN(BPF_REG_4), | ||
| 157 | BPF_ALU64_IMM(BPF_LSH, BPF_REG_4, 5), | ||
| 158 | BPF_ALU64_IMM(BPF_RSH, BPF_REG_4, 1), | ||
| 159 | BPF_ALU64_IMM(BPF_RSH, BPF_REG_4, 1), | ||
| 160 | BPF_ALU64_IMM(BPF_RSH, BPF_REG_4, 1), | ||
| 161 | BPF_ALU64_IMM(BPF_RSH, BPF_REG_4, 1), | ||
| 162 | BPF_MOV64_IMM(BPF_REG_0, 0), | ||
| 163 | BPF_EXIT_INSN(), | ||
| 164 | }, | ||
| 165 | .prog_type = BPF_PROG_TYPE_SCHED_CLS, | ||
| 166 | .matches = { | ||
| 167 | "7: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=inv56 R10=fp", | ||
| 168 | "8: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=inv55,min_align=2 R10=fp", | ||
| 169 | "9: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=inv54,min_align=4 R10=fp", | ||
| 170 | "10: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=inv53,min_align=8 R10=fp", | ||
| 171 | "11: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=inv52,min_align=16 R10=fp", | ||
| 172 | "18: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=pkt_end R4=inv56 R10=fp", | ||
| 173 | "19: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=pkt_end R4=inv51,min_align=32 R10=fp", | ||
| 174 | "20: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=pkt_end R4=inv52,min_align=16 R10=fp", | ||
| 175 | "21: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=pkt_end R4=inv53,min_align=8 R10=fp", | ||
| 176 | "22: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=pkt_end R4=inv54,min_align=4 R10=fp", | ||
| 177 | "23: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=pkt_end R4=inv55,min_align=2 R10=fp", | ||
| 178 | }, | ||
| 179 | }, | ||
| 180 | { | ||
| 181 | .descr = "unknown mul", | ||
| 182 | .insns = { | ||
| 183 | LOAD_UNKNOWN(BPF_REG_3), | ||
| 184 | BPF_MOV64_REG(BPF_REG_4, BPF_REG_3), | ||
| 185 | BPF_ALU64_IMM(BPF_MUL, BPF_REG_4, 1), | ||
| 186 | BPF_MOV64_REG(BPF_REG_4, BPF_REG_3), | ||
| 187 | BPF_ALU64_IMM(BPF_MUL, BPF_REG_4, 2), | ||
| 188 | BPF_MOV64_REG(BPF_REG_4, BPF_REG_3), | ||
| 189 | BPF_ALU64_IMM(BPF_MUL, BPF_REG_4, 4), | ||
| 190 | BPF_MOV64_REG(BPF_REG_4, BPF_REG_3), | ||
| 191 | BPF_ALU64_IMM(BPF_MUL, BPF_REG_4, 8), | ||
| 192 | BPF_ALU64_IMM(BPF_MUL, BPF_REG_4, 2), | ||
| 193 | BPF_MOV64_IMM(BPF_REG_0, 0), | ||
| 194 | BPF_EXIT_INSN(), | ||
| 195 | }, | ||
| 196 | .prog_type = BPF_PROG_TYPE_SCHED_CLS, | ||
| 197 | .matches = { | ||
| 198 | "7: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=inv56 R10=fp", | ||
| 199 | "8: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=inv56 R4=inv56 R10=fp", | ||
| 200 | "9: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=inv56 R4=inv55,min_align=1 R10=fp", | ||
| 201 | "10: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=inv56 R4=inv56 R10=fp", | ||
| 202 | "11: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=inv56 R4=inv54,min_align=2 R10=fp", | ||
| 203 | "12: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=inv56 R4=inv56 R10=fp", | ||
| 204 | "13: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=inv56 R4=inv53,min_align=4 R10=fp", | ||
| 205 | "14: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=inv56 R4=inv56 R10=fp", | ||
| 206 | "15: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=inv56 R4=inv52,min_align=8 R10=fp", | ||
| 207 | "16: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=inv56 R4=inv50,min_align=8 R10=fp" | ||
| 208 | }, | ||
| 209 | }, | ||
| 210 | { | ||
| 211 | .descr = "packet const offset", | ||
| 212 | .insns = { | ||
| 213 | PREP_PKT_POINTERS, | ||
| 214 | BPF_MOV64_REG(BPF_REG_5, BPF_REG_2), | ||
| 215 | |||
| 216 | BPF_MOV64_IMM(BPF_REG_0, 0), | ||
| 217 | |||
| 218 | /* Skip over ethernet header. */ | ||
| 219 | BPF_ALU64_IMM(BPF_ADD, BPF_REG_5, 14), | ||
| 220 | BPF_MOV64_REG(BPF_REG_4, BPF_REG_5), | ||
| 221 | BPF_ALU64_IMM(BPF_ADD, BPF_REG_4, 4), | ||
| 222 | BPF_JMP_REG(BPF_JGE, BPF_REG_3, BPF_REG_4, 1), | ||
| 223 | BPF_EXIT_INSN(), | ||
| 224 | |||
| 225 | BPF_LDX_MEM(BPF_B, BPF_REG_4, BPF_REG_5, 0), | ||
| 226 | BPF_LDX_MEM(BPF_B, BPF_REG_4, BPF_REG_5, 1), | ||
| 227 | BPF_LDX_MEM(BPF_B, BPF_REG_4, BPF_REG_5, 2), | ||
| 228 | BPF_LDX_MEM(BPF_B, BPF_REG_4, BPF_REG_5, 3), | ||
| 229 | BPF_LDX_MEM(BPF_H, BPF_REG_4, BPF_REG_5, 0), | ||
| 230 | BPF_LDX_MEM(BPF_H, BPF_REG_4, BPF_REG_5, 2), | ||
| 231 | BPF_LDX_MEM(BPF_W, BPF_REG_4, BPF_REG_5, 0), | ||
| 232 | |||
| 233 | BPF_MOV64_IMM(BPF_REG_0, 0), | ||
| 234 | BPF_EXIT_INSN(), | ||
| 235 | }, | ||
| 236 | .prog_type = BPF_PROG_TYPE_SCHED_CLS, | ||
| 237 | .matches = { | ||
| 238 | "4: R0=imm0,min_value=0,max_value=0,min_align=2147483648 R1=ctx R2=pkt(id=0,off=0,r=0) R3=pkt_end R5=pkt(id=0,off=0,r=0) R10=fp", | ||
| 239 | "5: R0=imm0,min_value=0,max_value=0,min_align=2147483648 R1=ctx R2=pkt(id=0,off=0,r=0) R3=pkt_end R5=pkt(id=0,off=14,r=0) R10=fp", | ||
| 240 | "6: R0=imm0,min_value=0,max_value=0,min_align=2147483648 R1=ctx R2=pkt(id=0,off=0,r=0) R3=pkt_end R4=pkt(id=0,off=14,r=0) R5=pkt(id=0,off=14,r=0) R10=fp", | ||
| 241 | "10: R0=imm0,min_value=0,max_value=0,min_align=2147483648 R1=ctx R2=pkt(id=0,off=0,r=18) R3=pkt_end R4=inv56 R5=pkt(id=0,off=14,r=18) R10=fp", | ||
| 242 | "14: R0=imm0,min_value=0,max_value=0,min_align=2147483648 R1=ctx R2=pkt(id=0,off=0,r=18) R3=pkt_end R4=inv48 R5=pkt(id=0,off=14,r=18) R10=fp", | ||
| 243 | "15: R0=imm0,min_value=0,max_value=0,min_align=2147483648 R1=ctx R2=pkt(id=0,off=0,r=18) R3=pkt_end R4=inv48 R5=pkt(id=0,off=14,r=18) R10=fp", | ||
| 244 | }, | ||
| 245 | }, | ||
| 246 | { | ||
| 247 | .descr = "packet variable offset", | ||
| 248 | .insns = { | ||
| 249 | LOAD_UNKNOWN(BPF_REG_6), | ||
| 250 | BPF_ALU64_IMM(BPF_LSH, BPF_REG_6, 2), | ||
| 251 | |||
| 252 | /* First, add a constant to the R5 packet pointer, | ||
| 253 | * then a variable with a known alignment. | ||
| 254 | */ | ||
| 255 | BPF_MOV64_REG(BPF_REG_5, BPF_REG_2), | ||
| 256 | BPF_ALU64_IMM(BPF_ADD, BPF_REG_5, 14), | ||
| 257 | BPF_ALU64_REG(BPF_ADD, BPF_REG_5, BPF_REG_6), | ||
| 258 | BPF_MOV64_REG(BPF_REG_4, BPF_REG_5), | ||
| 259 | BPF_ALU64_IMM(BPF_ADD, BPF_REG_4, 4), | ||
| 260 | BPF_JMP_REG(BPF_JGE, BPF_REG_3, BPF_REG_4, 1), | ||
| 261 | BPF_EXIT_INSN(), | ||
| 262 | BPF_LDX_MEM(BPF_W, BPF_REG_4, BPF_REG_5, 0), | ||
| 263 | |||
| 264 | /* Now, test in the other direction. Adding first | ||
| 265 | * the variable offset to R5, then the constant. | ||
| 266 | */ | ||
| 267 | BPF_MOV64_REG(BPF_REG_5, BPF_REG_2), | ||
| 268 | BPF_ALU64_REG(BPF_ADD, BPF_REG_5, BPF_REG_6), | ||
| 269 | BPF_ALU64_IMM(BPF_ADD, BPF_REG_5, 14), | ||
| 270 | BPF_MOV64_REG(BPF_REG_4, BPF_REG_5), | ||
| 271 | BPF_ALU64_IMM(BPF_ADD, BPF_REG_4, 4), | ||
| 272 | BPF_JMP_REG(BPF_JGE, BPF_REG_3, BPF_REG_4, 1), | ||
| 273 | BPF_EXIT_INSN(), | ||
| 274 | BPF_LDX_MEM(BPF_W, BPF_REG_4, BPF_REG_5, 0), | ||
| 275 | |||
| 276 | /* Test multiple accumulations of unknown values | ||
| 277 | * into a packet pointer. | ||
| 278 | */ | ||
| 279 | BPF_MOV64_REG(BPF_REG_5, BPF_REG_2), | ||
| 280 | BPF_ALU64_IMM(BPF_ADD, BPF_REG_5, 14), | ||
| 281 | BPF_ALU64_REG(BPF_ADD, BPF_REG_5, BPF_REG_6), | ||
| 282 | BPF_ALU64_IMM(BPF_ADD, BPF_REG_5, 4), | ||
| 283 | BPF_ALU64_REG(BPF_ADD, BPF_REG_5, BPF_REG_6), | ||
| 284 | BPF_MOV64_REG(BPF_REG_4, BPF_REG_5), | ||
| 285 | BPF_ALU64_IMM(BPF_ADD, BPF_REG_4, 4), | ||
| 286 | BPF_JMP_REG(BPF_JGE, BPF_REG_3, BPF_REG_4, 1), | ||
| 287 | BPF_EXIT_INSN(), | ||
| 288 | BPF_LDX_MEM(BPF_W, BPF_REG_4, BPF_REG_5, 0), | ||
| 289 | |||
| 290 | BPF_MOV64_IMM(BPF_REG_0, 0), | ||
| 291 | BPF_EXIT_INSN(), | ||
| 292 | }, | ||
| 293 | .prog_type = BPF_PROG_TYPE_SCHED_CLS, | ||
| 294 | .matches = { | ||
| 295 | /* Calculated offset in R6 has unknown value, but known | ||
| 296 | * alignment of 4. | ||
| 297 | */ | ||
| 298 | "8: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=pkt_end R6=inv54,min_align=4 R10=fp", | ||
| 299 | |||
| 300 | /* Offset is added to packet pointer R5, resulting in known | ||
| 301 | * auxiliary alignment and offset. | ||
| 302 | */ | ||
| 303 | "11: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=pkt_end R5=pkt(id=1,off=0,r=0),aux_off=14,aux_off_align=4 R6=inv54,min_align=4 R10=fp", | ||
| 304 | |||
| 305 | /* At the time the word size load is performed from R5, | ||
| 306 | * it's total offset is NET_IP_ALIGN + reg->off (0) + | ||
| 307 | * reg->aux_off (14) which is 16. Then the variable | ||
| 308 | * offset is considered using reg->aux_off_align which | ||
| 309 | * is 4 and meets the load's requirements. | ||
| 310 | */ | ||
| 311 | "15: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=pkt_end R4=pkt(id=1,off=4,r=4),aux_off=14,aux_off_align=4 R5=pkt(id=1,off=0,r=4),aux_off=14,aux_off_align=4 R6=inv54,min_align=4 R10=fp", | ||
| 312 | |||
| 313 | |||
| 314 | /* Variable offset is added to R5 packet pointer, | ||
| 315 | * resulting in auxiliary alignment of 4. | ||
| 316 | */ | ||
| 317 | "18: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=pkt_end R4=inv,aux_off=14,aux_off_align=4 R5=pkt(id=2,off=0,r=0),aux_off_align=4 R6=inv54,min_align=4 R10=fp", | ||
| 318 | |||
| 319 | /* Constant offset is added to R5, resulting in | ||
| 320 | * reg->off of 14. | ||
| 321 | */ | ||
| 322 | "19: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=pkt_end R4=inv,aux_off=14,aux_off_align=4 R5=pkt(id=2,off=14,r=0),aux_off_align=4 R6=inv54,min_align=4 R10=fp", | ||
| 323 | |||
| 324 | /* At the time the word size load is performed from R5, | ||
| 325 | * it's total offset is NET_IP_ALIGN + reg->off (14) which | ||
| 326 | * is 16. Then the variable offset is considered using | ||
| 327 | * reg->aux_off_align which is 4 and meets the load's | ||
| 328 | * requirements. | ||
| 329 | */ | ||
| 330 | "23: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=pkt_end R4=pkt(id=2,off=18,r=18),aux_off_align=4 R5=pkt(id=2,off=14,r=18),aux_off_align=4 R6=inv54,min_align=4 R10=fp", | ||
| 331 | |||
| 332 | /* Constant offset is added to R5 packet pointer, | ||
| 333 | * resulting in reg->off value of 14. | ||
| 334 | */ | ||
| 335 | "26: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=pkt_end R4=inv,aux_off_align=4 R5=pkt(id=0,off=14,r=8) R6=inv54,min_align=4 R10=fp", | ||
| 336 | /* Variable offset is added to R5, resulting in an | ||
| 337 | * auxiliary offset of 14, and an auxiliary alignment of 4. | ||
| 338 | */ | ||
| 339 | "27: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=pkt_end R4=inv,aux_off_align=4 R5=pkt(id=3,off=0,r=0),aux_off=14,aux_off_align=4 R6=inv54,min_align=4 R10=fp", | ||
| 340 | /* Constant is added to R5 again, setting reg->off to 4. */ | ||
| 341 | "28: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=pkt_end R4=inv,aux_off_align=4 R5=pkt(id=3,off=4,r=0),aux_off=14,aux_off_align=4 R6=inv54,min_align=4 R10=fp", | ||
| 342 | /* And once more we add a variable, which causes an accumulation | ||
| 343 | * of reg->off into reg->aux_off_align, with resulting value of | ||
| 344 | * 18. The auxiliary alignment stays at 4. | ||
| 345 | */ | ||
| 346 | "29: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=pkt_end R4=inv,aux_off_align=4 R5=pkt(id=4,off=0,r=0),aux_off=18,aux_off_align=4 R6=inv54,min_align=4 R10=fp", | ||
| 347 | /* At the time the word size load is performed from R5, | ||
| 348 | * it's total offset is NET_IP_ALIGN + reg->off (0) + | ||
| 349 | * reg->aux_off (18) which is 20. Then the variable offset | ||
| 350 | * is considered using reg->aux_off_align which is 4 and meets | ||
| 351 | * the load's requirements. | ||
| 352 | */ | ||
| 353 | "33: R0=pkt(id=0,off=8,r=8) R1=ctx R2=pkt(id=0,off=0,r=8) R3=pkt_end R4=pkt(id=4,off=4,r=4),aux_off=18,aux_off_align=4 R5=pkt(id=4,off=0,r=4),aux_off=18,aux_off_align=4 R6=inv54,min_align=4 R10=fp", | ||
| 354 | }, | ||
| 355 | }, | ||
| 356 | }; | ||
| 357 | |||
| 358 | static int probe_filter_length(const struct bpf_insn *fp) | ||
| 359 | { | ||
| 360 | int len; | ||
| 361 | |||
| 362 | for (len = MAX_INSNS - 1; len > 0; --len) | ||
| 363 | if (fp[len].code != 0 || fp[len].imm != 0) | ||
| 364 | break; | ||
| 365 | return len + 1; | ||
| 366 | } | ||
| 367 | |||
| 368 | static char bpf_vlog[32768]; | ||
| 369 | |||
| 370 | static int do_test_single(struct bpf_align_test *test) | ||
| 371 | { | ||
| 372 | struct bpf_insn *prog = test->insns; | ||
| 373 | int prog_type = test->prog_type; | ||
| 374 | int prog_len, i; | ||
| 375 | int fd_prog; | ||
| 376 | int ret; | ||
| 377 | |||
| 378 | prog_len = probe_filter_length(prog); | ||
| 379 | fd_prog = bpf_verify_program(prog_type ? : BPF_PROG_TYPE_SOCKET_FILTER, | ||
| 380 | prog, prog_len, 1, "GPL", 0, | ||
| 381 | bpf_vlog, sizeof(bpf_vlog)); | ||
| 382 | if (fd_prog < 0) { | ||
| 383 | printf("Failed to load program.\n"); | ||
| 384 | printf("%s", bpf_vlog); | ||
| 385 | ret = 1; | ||
| 386 | } else { | ||
| 387 | ret = 0; | ||
| 388 | for (i = 0; i < MAX_MATCHES; i++) { | ||
| 389 | const char *t, *m = test->matches[i]; | ||
| 390 | |||
| 391 | if (!m) | ||
| 392 | break; | ||
| 393 | t = strstr(bpf_vlog, m); | ||
| 394 | if (!t) { | ||
| 395 | printf("Failed to find match: %s\n", m); | ||
| 396 | ret = 1; | ||
| 397 | printf("%s", bpf_vlog); | ||
| 398 | break; | ||
| 399 | } | ||
| 400 | } | ||
| 401 | close(fd_prog); | ||
| 402 | } | ||
| 403 | return ret; | ||
| 404 | } | ||
| 405 | |||
| 406 | static int do_test(unsigned int from, unsigned int to) | ||
| 407 | { | ||
| 408 | int all_pass = 0; | ||
| 409 | int all_fail = 0; | ||
| 410 | unsigned int i; | ||
| 411 | |||
| 412 | for (i = from; i < to; i++) { | ||
| 413 | struct bpf_align_test *test = &tests[i]; | ||
| 414 | int fail; | ||
| 415 | |||
| 416 | printf("Test %3d: %s ... ", | ||
| 417 | i, test->descr); | ||
| 418 | fail = do_test_single(test); | ||
| 419 | if (fail) { | ||
| 420 | all_fail++; | ||
| 421 | printf("FAIL\n"); | ||
| 422 | } else { | ||
| 423 | all_pass++; | ||
| 424 | printf("PASS\n"); | ||
| 425 | } | ||
| 426 | } | ||
| 427 | printf("Results: %d pass %d fail\n", | ||
| 428 | all_pass, all_fail); | ||
| 429 | return 0; | ||
| 430 | } | ||
| 431 | |||
| 432 | int main(int argc, char **argv) | ||
| 433 | { | ||
| 434 | unsigned int from = 0, to = ARRAY_SIZE(tests); | ||
| 435 | |||
| 436 | if (argc == 3) { | ||
| 437 | unsigned int l = atoi(argv[argc - 2]); | ||
| 438 | unsigned int u = atoi(argv[argc - 1]); | ||
| 439 | |||
| 440 | if (l < to && u < to) { | ||
| 441 | from = l; | ||
| 442 | to = u + 1; | ||
| 443 | } | ||
| 444 | } else if (argc == 2) { | ||
| 445 | unsigned int t = atoi(argv[argc - 1]); | ||
| 446 | |||
| 447 | if (t < to) { | ||
| 448 | from = t; | ||
| 449 | to = t + 1; | ||
| 450 | } | ||
| 451 | } | ||
| 452 | return do_test(from, to); | ||
| 453 | } | ||
diff --git a/tools/testing/selftests/bpf/test_pkt_access.c b/tools/testing/selftests/bpf/test_pkt_access.c index 39387bb7e08c..6e11ba11709e 100644 --- a/tools/testing/selftests/bpf/test_pkt_access.c +++ b/tools/testing/selftests/bpf/test_pkt_access.c | |||
| @@ -5,6 +5,7 @@ | |||
| 5 | * License as published by the Free Software Foundation. | 5 | * License as published by the Free Software Foundation. |
| 6 | */ | 6 | */ |
| 7 | #include <stddef.h> | 7 | #include <stddef.h> |
| 8 | #include <string.h> | ||
| 8 | #include <linux/bpf.h> | 9 | #include <linux/bpf.h> |
| 9 | #include <linux/if_ether.h> | 10 | #include <linux/if_ether.h> |
| 10 | #include <linux/if_packet.h> | 11 | #include <linux/if_packet.h> |
diff --git a/tools/testing/selftests/powerpc/tm/.gitignore b/tools/testing/selftests/powerpc/tm/.gitignore index 427621792229..2f1f7b013293 100644 --- a/tools/testing/selftests/powerpc/tm/.gitignore +++ b/tools/testing/selftests/powerpc/tm/.gitignore | |||
| @@ -11,3 +11,4 @@ tm-signal-context-chk-fpu | |||
| 11 | tm-signal-context-chk-gpr | 11 | tm-signal-context-chk-gpr |
| 12 | tm-signal-context-chk-vmx | 12 | tm-signal-context-chk-vmx |
| 13 | tm-signal-context-chk-vsx | 13 | tm-signal-context-chk-vsx |
| 14 | tm-vmx-unavail | ||
diff --git a/tools/testing/selftests/powerpc/tm/Makefile b/tools/testing/selftests/powerpc/tm/Makefile index 5576ee6a51f2..958c11c14acd 100644 --- a/tools/testing/selftests/powerpc/tm/Makefile +++ b/tools/testing/selftests/powerpc/tm/Makefile | |||
| @@ -2,7 +2,8 @@ SIGNAL_CONTEXT_CHK_TESTS := tm-signal-context-chk-gpr tm-signal-context-chk-fpu | |||
| 2 | tm-signal-context-chk-vmx tm-signal-context-chk-vsx | 2 | tm-signal-context-chk-vmx tm-signal-context-chk-vsx |
| 3 | 3 | ||
| 4 | TEST_GEN_PROGS := tm-resched-dscr tm-syscall tm-signal-msr-resv tm-signal-stack \ | 4 | TEST_GEN_PROGS := tm-resched-dscr tm-syscall tm-signal-msr-resv tm-signal-stack \ |
| 5 | tm-vmxcopy tm-fork tm-tar tm-tmspr $(SIGNAL_CONTEXT_CHK_TESTS) | 5 | tm-vmxcopy tm-fork tm-tar tm-tmspr tm-vmx-unavail \ |
| 6 | $(SIGNAL_CONTEXT_CHK_TESTS) | ||
| 6 | 7 | ||
| 7 | include ../../lib.mk | 8 | include ../../lib.mk |
| 8 | 9 | ||
| @@ -13,6 +14,7 @@ CFLAGS += -mhtm | |||
| 13 | $(OUTPUT)/tm-syscall: tm-syscall-asm.S | 14 | $(OUTPUT)/tm-syscall: tm-syscall-asm.S |
| 14 | $(OUTPUT)/tm-syscall: CFLAGS += -I../../../../../usr/include | 15 | $(OUTPUT)/tm-syscall: CFLAGS += -I../../../../../usr/include |
| 15 | $(OUTPUT)/tm-tmspr: CFLAGS += -pthread | 16 | $(OUTPUT)/tm-tmspr: CFLAGS += -pthread |
| 17 | $(OUTPUT)/tm-vmx-unavail: CFLAGS += -pthread -m64 | ||
| 16 | 18 | ||
| 17 | SIGNAL_CONTEXT_CHK_TESTS := $(patsubst %,$(OUTPUT)/%,$(SIGNAL_CONTEXT_CHK_TESTS)) | 19 | SIGNAL_CONTEXT_CHK_TESTS := $(patsubst %,$(OUTPUT)/%,$(SIGNAL_CONTEXT_CHK_TESTS)) |
| 18 | $(SIGNAL_CONTEXT_CHK_TESTS): tm-signal.S | 20 | $(SIGNAL_CONTEXT_CHK_TESTS): tm-signal.S |
diff --git a/tools/testing/selftests/powerpc/tm/tm-vmx-unavail.c b/tools/testing/selftests/powerpc/tm/tm-vmx-unavail.c new file mode 100644 index 000000000000..137185ba4937 --- /dev/null +++ b/tools/testing/selftests/powerpc/tm/tm-vmx-unavail.c | |||
| @@ -0,0 +1,118 @@ | |||
| 1 | /* | ||
| 2 | * Copyright 2017, Michael Neuling, IBM Corp. | ||
| 3 | * Licensed under GPLv2. | ||
| 4 | * Original: Breno Leitao <brenohl@br.ibm.com> & | ||
| 5 | * Gustavo Bueno Romero <gromero@br.ibm.com> | ||
| 6 | * Edited: Michael Neuling | ||
| 7 | * | ||
| 8 | * Force VMX unavailable during a transaction and see if it corrupts | ||
| 9 | * the checkpointed VMX register state after the abort. | ||
| 10 | */ | ||
| 11 | |||
| 12 | #include <inttypes.h> | ||
| 13 | #include <htmintrin.h> | ||
| 14 | #include <string.h> | ||
| 15 | #include <stdlib.h> | ||
| 16 | #include <stdio.h> | ||
| 17 | #include <pthread.h> | ||
| 18 | #include <sys/mman.h> | ||
| 19 | #include <unistd.h> | ||
| 20 | #include <pthread.h> | ||
| 21 | |||
| 22 | #include "tm.h" | ||
| 23 | #include "utils.h" | ||
| 24 | |||
| 25 | int passed; | ||
| 26 | |||
| 27 | void *worker(void *unused) | ||
| 28 | { | ||
| 29 | __int128 vmx0; | ||
| 30 | uint64_t texasr; | ||
| 31 | |||
| 32 | asm goto ( | ||
| 33 | "li 3, 1;" /* Stick non-zero value in VMX0 */ | ||
| 34 | "std 3, 0(%[vmx0_ptr]);" | ||
| 35 | "lvx 0, 0, %[vmx0_ptr];" | ||
| 36 | |||
| 37 | /* Wait here a bit so we get scheduled out 255 times */ | ||
| 38 | "lis 3, 0x3fff;" | ||
| 39 | "1: ;" | ||
| 40 | "addi 3, 3, -1;" | ||
| 41 | "cmpdi 3, 0;" | ||
| 42 | "bne 1b;" | ||
| 43 | |||
| 44 | /* Kernel will hopefully turn VMX off now */ | ||
| 45 | |||
| 46 | "tbegin. ;" | ||
| 47 | "beq failure;" | ||
| 48 | |||
| 49 | /* Cause VMX unavail. Any VMX instruction */ | ||
| 50 | "vaddcuw 0,0,0;" | ||
| 51 | |||
| 52 | "tend. ;" | ||
| 53 | "b %l[success];" | ||
| 54 | |||
| 55 | /* Check VMX0 sanity after abort */ | ||
| 56 | "failure: ;" | ||
| 57 | "lvx 1, 0, %[vmx0_ptr];" | ||
| 58 | "vcmpequb. 2, 0, 1;" | ||
| 59 | "bc 4, 24, %l[value_mismatch];" | ||
| 60 | "b %l[value_match];" | ||
| 61 | : | ||
| 62 | : [vmx0_ptr] "r"(&vmx0) | ||
| 63 | : "r3" | ||
| 64 | : success, value_match, value_mismatch | ||
| 65 | ); | ||
| 66 | |||
| 67 | /* HTM aborted and VMX0 is corrupted */ | ||
| 68 | value_mismatch: | ||
| 69 | texasr = __builtin_get_texasr(); | ||
| 70 | |||
| 71 | printf("\n\n==============\n\n"); | ||
| 72 | printf("Failure with error: %lx\n", _TEXASR_FAILURE_CODE(texasr)); | ||
| 73 | printf("Summary error : %lx\n", _TEXASR_FAILURE_SUMMARY(texasr)); | ||
| 74 | printf("TFIAR exact : %lx\n\n", _TEXASR_TFIAR_EXACT(texasr)); | ||
| 75 | |||
| 76 | passed = 0; | ||
| 77 | return NULL; | ||
| 78 | |||
| 79 | /* HTM aborted but VMX0 is correct */ | ||
| 80 | value_match: | ||
| 81 | // printf("!"); | ||
| 82 | return NULL; | ||
| 83 | |||
| 84 | success: | ||
| 85 | // printf("."); | ||
| 86 | return NULL; | ||
| 87 | } | ||
| 88 | |||
| 89 | int tm_vmx_unavail_test() | ||
| 90 | { | ||
| 91 | int threads; | ||
| 92 | pthread_t *thread; | ||
| 93 | |||
| 94 | SKIP_IF(!have_htm()); | ||
| 95 | |||
| 96 | passed = 1; | ||
| 97 | |||
| 98 | threads = sysconf(_SC_NPROCESSORS_ONLN) * 4; | ||
| 99 | thread = malloc(sizeof(pthread_t)*threads); | ||
| 100 | if (!thread) | ||
| 101 | return EXIT_FAILURE; | ||
| 102 | |||
| 103 | for (uint64_t i = 0; i < threads; i++) | ||
| 104 | pthread_create(&thread[i], NULL, &worker, NULL); | ||
| 105 | |||
| 106 | for (uint64_t i = 0; i < threads; i++) | ||
| 107 | pthread_join(thread[i], NULL); | ||
| 108 | |||
| 109 | free(thread); | ||
| 110 | |||
| 111 | return passed ? EXIT_SUCCESS : EXIT_FAILURE; | ||
| 112 | } | ||
| 113 | |||
| 114 | |||
| 115 | int main(int argc, char **argv) | ||
| 116 | { | ||
| 117 | return test_harness(tm_vmx_unavail_test, "tm_vmx_unavail_test"); | ||
| 118 | } | ||
