diff options
author | Anthony G. Basile <blueness@gentoo.org> | 2017-04-14 10:48:43 -0400 |
---|---|---|
committer | Anthony G. Basile <blueness@gentoo.org> | 2017-04-14 10:48:43 -0400 |
commit | c0ac9f36b469dcdb20ee37316e77a21994bafc50 (patch) | |
tree | 40b0abb35bca57d337945ff9b90646ba39f53fd4 | |
parent | grsecurity-3.1-4.9.21-201704091948 (diff) | |
download | hardened-patchset-c0ac9f36b469dcdb20ee37316e77a21994bafc50.tar.gz hardened-patchset-c0ac9f36b469dcdb20ee37316e77a21994bafc50.tar.bz2 hardened-patchset-c0ac9f36b469dcdb20ee37316e77a21994bafc50.zip |
grsecurity-3.1-4.9.22-20170412083620170412
-rw-r--r-- | 4.9.21/1020_linux-4.9.21.patch | 4658 | ||||
-rw-r--r-- | 4.9.22/0000_README (renamed from 4.9.21/0000_README) | 6 | ||||
-rw-r--r-- | 4.9.22/4420_grsecurity-3.1-4.9.22-201704120836.patch (renamed from 4.9.21/4420_grsecurity-3.1-4.9.21-201704091948.patch) | 146 | ||||
-rw-r--r-- | 4.9.22/4425_grsec_remove_EI_PAX.patch (renamed from 4.9.21/4425_grsec_remove_EI_PAX.patch) | 0 | ||||
-rw-r--r-- | 4.9.22/4426_default_XATTR_PAX_FLAGS.patch (renamed from 4.9.21/4426_default_XATTR_PAX_FLAGS.patch) | 0 | ||||
-rw-r--r-- | 4.9.22/4427_force_XATTR_PAX_tmpfs.patch (renamed from 4.9.21/4427_force_XATTR_PAX_tmpfs.patch) | 0 | ||||
-rw-r--r-- | 4.9.22/4430_grsec-remove-localversion-grsec.patch (renamed from 4.9.21/4430_grsec-remove-localversion-grsec.patch) | 0 | ||||
-rw-r--r-- | 4.9.22/4435_grsec-mute-warnings.patch (renamed from 4.9.21/4435_grsec-mute-warnings.patch) | 0 | ||||
-rw-r--r-- | 4.9.22/4440_grsec-remove-protected-paths.patch (renamed from 4.9.21/4440_grsec-remove-protected-paths.patch) | 0 | ||||
-rw-r--r-- | 4.9.22/4450_grsec-kconfig-default-gids.patch (renamed from 4.9.21/4450_grsec-kconfig-default-gids.patch) | 0 | ||||
-rw-r--r-- | 4.9.22/4465_selinux-avc_audit-log-curr_ip.patch (renamed from 4.9.21/4465_selinux-avc_audit-log-curr_ip.patch) | 0 | ||||
-rw-r--r-- | 4.9.22/4470_disable-compat_vdso.patch (renamed from 4.9.21/4470_disable-compat_vdso.patch) | 0 | ||||
-rw-r--r-- | 4.9.22/4475_emutramp_default_on.patch (renamed from 4.9.21/4475_emutramp_default_on.patch) | 0 |
13 files changed, 74 insertions, 4736 deletions
diff --git a/4.9.21/1020_linux-4.9.21.patch b/4.9.21/1020_linux-4.9.21.patch deleted file mode 100644 index 42defb2..0000000 --- a/4.9.21/1020_linux-4.9.21.patch +++ /dev/null @@ -1,4658 +0,0 @@ -diff --git a/Makefile b/Makefile -index 4496018..1523557 100644 ---- a/Makefile -+++ b/Makefile -@@ -1,6 +1,6 @@ - VERSION = 4 - PATCHLEVEL = 9 --SUBLEVEL = 20 -+SUBLEVEL = 21 - EXTRAVERSION = - NAME = Roaring Lionus - -diff --git a/arch/arm/boot/dts/bcm5301x.dtsi b/arch/arm/boot/dts/bcm5301x.dtsi -index ae4b388..4616452 100644 ---- a/arch/arm/boot/dts/bcm5301x.dtsi -+++ b/arch/arm/boot/dts/bcm5301x.dtsi -@@ -66,14 +66,14 @@ - timer@20200 { - compatible = "arm,cortex-a9-global-timer"; - reg = <0x20200 0x100>; -- interrupts = <GIC_PPI 11 IRQ_TYPE_LEVEL_HIGH>; -+ interrupts = <GIC_PPI 11 IRQ_TYPE_EDGE_RISING>; - clocks = <&periph_clk>; - }; - - local-timer@20600 { - compatible = "arm,cortex-a9-twd-timer"; - reg = <0x20600 0x100>; -- interrupts = <GIC_PPI 13 IRQ_TYPE_LEVEL_HIGH>; -+ interrupts = <GIC_PPI 13 IRQ_TYPE_EDGE_RISING>; - clocks = <&periph_clk>; - }; - -diff --git a/arch/arm/mach-bcm/bcm_5301x.c b/arch/arm/mach-bcm/bcm_5301x.c -index c8830a2..fe067f6 100644 ---- a/arch/arm/mach-bcm/bcm_5301x.c -+++ b/arch/arm/mach-bcm/bcm_5301x.c -@@ -9,14 +9,42 @@ - #include <asm/hardware/cache-l2x0.h> - - #include <asm/mach/arch.h> -+#include <asm/siginfo.h> -+#include <asm/signal.h> -+ -+#define FSR_EXTERNAL (1 << 12) -+#define FSR_READ (0 << 10) -+#define FSR_IMPRECISE 0x0406 - - static const char *const bcm5301x_dt_compat[] __initconst = { - "brcm,bcm4708", - NULL, - }; - -+static int bcm5301x_abort_handler(unsigned long addr, unsigned int fsr, -+ struct pt_regs *regs) -+{ -+ /* -+ * We want to ignore aborts forwarded from the PCIe bus that are -+ * expected and shouldn't really be passed by the PCIe controller. -+ * The biggest disadvantage is the same FSR code may be reported when -+ * reading non-existing APB register and we shouldn't ignore that. -+ */ -+ if (fsr == (FSR_EXTERNAL | FSR_READ | FSR_IMPRECISE)) -+ return 0; -+ -+ return 1; -+} -+ -+static void __init bcm5301x_init_early(void) -+{ -+ hook_fault_code(16 + 6, bcm5301x_abort_handler, SIGBUS, BUS_OBJERR, -+ "imprecise external abort"); -+} -+ - DT_MACHINE_START(BCM5301X, "BCM5301X") - .l2c_aux_val = 0, - .l2c_aux_mask = ~0, - .dt_compat = bcm5301x_dt_compat, -+ .init_early = bcm5301x_init_early, - MACHINE_END -diff --git a/arch/mips/lantiq/irq.c b/arch/mips/lantiq/irq.c -index 8ac0e59..0ddf369 100644 ---- a/arch/mips/lantiq/irq.c -+++ b/arch/mips/lantiq/irq.c -@@ -269,6 +269,11 @@ static void ltq_hw5_irqdispatch(void) - DEFINE_HWx_IRQDISPATCH(5) - #endif - -+static void ltq_hw_irq_handler(struct irq_desc *desc) -+{ -+ ltq_hw_irqdispatch(irq_desc_get_irq(desc) - 2); -+} -+ - #ifdef CONFIG_MIPS_MT_SMP - void __init arch_init_ipiirq(int irq, struct irqaction *action) - { -@@ -313,23 +318,19 @@ static struct irqaction irq_call = { - asmlinkage void plat_irq_dispatch(void) - { - unsigned int pending = read_c0_status() & read_c0_cause() & ST0_IM; -- unsigned int i; -- -- if ((MIPS_CPU_TIMER_IRQ == 7) && (pending & CAUSEF_IP7)) { -- do_IRQ(MIPS_CPU_TIMER_IRQ); -- goto out; -- } else { -- for (i = 0; i < MAX_IM; i++) { -- if (pending & (CAUSEF_IP2 << i)) { -- ltq_hw_irqdispatch(i); -- goto out; -- } -- } -+ int irq; -+ -+ if (!pending) { -+ spurious_interrupt(); -+ return; - } -- pr_alert("Spurious IRQ: CAUSE=0x%08x\n", read_c0_status()); - --out: -- return; -+ pending >>= CAUSEB_IP; -+ while (pending) { -+ irq = fls(pending) - 1; -+ do_IRQ(MIPS_CPU_IRQ_BASE + irq); -+ pending &= ~BIT(irq); -+ } - } - - static int icu_map(struct irq_domain *d, unsigned int irq, irq_hw_number_t hw) -@@ -354,11 +355,6 @@ static const struct irq_domain_ops irq_domain_ops = { - .map = icu_map, - }; - --static struct irqaction cascade = { -- .handler = no_action, -- .name = "cascade", --}; -- - int __init icu_of_init(struct device_node *node, struct device_node *parent) - { - struct device_node *eiu_node; -@@ -390,7 +386,7 @@ int __init icu_of_init(struct device_node *node, struct device_node *parent) - mips_cpu_irq_init(); - - for (i = 0; i < MAX_IM; i++) -- setup_irq(i + 2, &cascade); -+ irq_set_chained_handler(i + 2, ltq_hw_irq_handler); - - if (cpu_has_vint) { - pr_info("Setting up vectored interrupts\n"); -diff --git a/arch/parisc/include/asm/uaccess.h b/arch/parisc/include/asm/uaccess.h -index 9a2aee1..7fcf512 100644 ---- a/arch/parisc/include/asm/uaccess.h -+++ b/arch/parisc/include/asm/uaccess.h -@@ -68,6 +68,15 @@ struct exception_table_entry { - ".previous\n" - - /* -+ * ASM_EXCEPTIONTABLE_ENTRY_EFAULT() creates a special exception table entry -+ * (with lowest bit set) for which the fault handler in fixup_exception() will -+ * load -EFAULT into %r8 for a read or write fault, and zeroes the target -+ * register in case of a read fault in get_user(). -+ */ -+#define ASM_EXCEPTIONTABLE_ENTRY_EFAULT( fault_addr, except_addr )\ -+ ASM_EXCEPTIONTABLE_ENTRY( fault_addr, except_addr + 1) -+ -+/* - * The page fault handler stores, in a per-cpu area, the following information - * if a fixup routine is available. - */ -@@ -94,7 +103,7 @@ struct exception_data { - #define __get_user(x, ptr) \ - ({ \ - register long __gu_err __asm__ ("r8") = 0; \ -- register long __gu_val __asm__ ("r9") = 0; \ -+ register long __gu_val; \ - \ - load_sr2(); \ - switch (sizeof(*(ptr))) { \ -@@ -110,22 +119,23 @@ struct exception_data { - }) - - #define __get_user_asm(ldx, ptr) \ -- __asm__("\n1:\t" ldx "\t0(%%sr2,%2),%0\n\t" \ -- ASM_EXCEPTIONTABLE_ENTRY(1b, fixup_get_user_skip_1)\ -+ __asm__("1: " ldx " 0(%%sr2,%2),%0\n" \ -+ "9:\n" \ -+ ASM_EXCEPTIONTABLE_ENTRY_EFAULT(1b, 9b) \ - : "=r"(__gu_val), "=r"(__gu_err) \ -- : "r"(ptr), "1"(__gu_err) \ -- : "r1"); -+ : "r"(ptr), "1"(__gu_err)); - - #if !defined(CONFIG_64BIT) - - #define __get_user_asm64(ptr) \ -- __asm__("\n1:\tldw 0(%%sr2,%2),%0" \ -- "\n2:\tldw 4(%%sr2,%2),%R0\n\t" \ -- ASM_EXCEPTIONTABLE_ENTRY(1b, fixup_get_user_skip_2)\ -- ASM_EXCEPTIONTABLE_ENTRY(2b, fixup_get_user_skip_1)\ -+ __asm__(" copy %%r0,%R0\n" \ -+ "1: ldw 0(%%sr2,%2),%0\n" \ -+ "2: ldw 4(%%sr2,%2),%R0\n" \ -+ "9:\n" \ -+ ASM_EXCEPTIONTABLE_ENTRY_EFAULT(1b, 9b) \ -+ ASM_EXCEPTIONTABLE_ENTRY_EFAULT(2b, 9b) \ - : "=r"(__gu_val), "=r"(__gu_err) \ -- : "r"(ptr), "1"(__gu_err) \ -- : "r1"); -+ : "r"(ptr), "1"(__gu_err)); - - #endif /* !defined(CONFIG_64BIT) */ - -@@ -151,32 +161,31 @@ struct exception_data { - * The "__put_user/kernel_asm()" macros tell gcc they read from memory - * instead of writing. This is because they do not write to any memory - * gcc knows about, so there are no aliasing issues. These macros must -- * also be aware that "fixup_put_user_skip_[12]" are executed in the -- * context of the fault, and any registers used there must be listed -- * as clobbers. In this case only "r1" is used by the current routines. -- * r8/r9 are already listed as err/val. -+ * also be aware that fixups are executed in the context of the fault, -+ * and any registers used there must be listed as clobbers. -+ * r8 is already listed as err. - */ - - #define __put_user_asm(stx, x, ptr) \ - __asm__ __volatile__ ( \ -- "\n1:\t" stx "\t%2,0(%%sr2,%1)\n\t" \ -- ASM_EXCEPTIONTABLE_ENTRY(1b, fixup_put_user_skip_1)\ -+ "1: " stx " %2,0(%%sr2,%1)\n" \ -+ "9:\n" \ -+ ASM_EXCEPTIONTABLE_ENTRY_EFAULT(1b, 9b) \ - : "=r"(__pu_err) \ -- : "r"(ptr), "r"(x), "0"(__pu_err) \ -- : "r1") -+ : "r"(ptr), "r"(x), "0"(__pu_err)) - - - #if !defined(CONFIG_64BIT) - - #define __put_user_asm64(__val, ptr) do { \ - __asm__ __volatile__ ( \ -- "\n1:\tstw %2,0(%%sr2,%1)" \ -- "\n2:\tstw %R2,4(%%sr2,%1)\n\t" \ -- ASM_EXCEPTIONTABLE_ENTRY(1b, fixup_put_user_skip_2)\ -- ASM_EXCEPTIONTABLE_ENTRY(2b, fixup_put_user_skip_1)\ -+ "1: stw %2,0(%%sr2,%1)\n" \ -+ "2: stw %R2,4(%%sr2,%1)\n" \ -+ "9:\n" \ -+ ASM_EXCEPTIONTABLE_ENTRY_EFAULT(1b, 9b) \ -+ ASM_EXCEPTIONTABLE_ENTRY_EFAULT(2b, 9b) \ - : "=r"(__pu_err) \ -- : "r"(ptr), "r"(__val), "0"(__pu_err) \ -- : "r1"); \ -+ : "r"(ptr), "r"(__val), "0"(__pu_err)); \ - } while (0) - - #endif /* !defined(CONFIG_64BIT) */ -diff --git a/arch/parisc/kernel/parisc_ksyms.c b/arch/parisc/kernel/parisc_ksyms.c -index 3cad8aa..4e6f0d9 100644 ---- a/arch/parisc/kernel/parisc_ksyms.c -+++ b/arch/parisc/kernel/parisc_ksyms.c -@@ -47,16 +47,6 @@ EXPORT_SYMBOL(__cmpxchg_u64); - EXPORT_SYMBOL(lclear_user); - EXPORT_SYMBOL(lstrnlen_user); - --/* Global fixups - defined as int to avoid creation of function pointers */ --extern int fixup_get_user_skip_1; --extern int fixup_get_user_skip_2; --extern int fixup_put_user_skip_1; --extern int fixup_put_user_skip_2; --EXPORT_SYMBOL(fixup_get_user_skip_1); --EXPORT_SYMBOL(fixup_get_user_skip_2); --EXPORT_SYMBOL(fixup_put_user_skip_1); --EXPORT_SYMBOL(fixup_put_user_skip_2); -- - #ifndef CONFIG_64BIT - /* Needed so insmod can set dp value */ - extern int $global$; -diff --git a/arch/parisc/kernel/process.c b/arch/parisc/kernel/process.c -index e81afc37..e7ffde2 100644 ---- a/arch/parisc/kernel/process.c -+++ b/arch/parisc/kernel/process.c -@@ -140,6 +140,8 @@ void machine_power_off(void) - printk(KERN_EMERG "System shut down completed.\n" - "Please power this system off now."); - -+ /* prevent soft lockup/stalled CPU messages for endless loop. */ -+ rcu_sysrq_start(); - for (;;); - } - -diff --git a/arch/parisc/lib/Makefile b/arch/parisc/lib/Makefile -index 8fa92b8..f2dac4d 100644 ---- a/arch/parisc/lib/Makefile -+++ b/arch/parisc/lib/Makefile -@@ -2,7 +2,7 @@ - # Makefile for parisc-specific library files - # - --lib-y := lusercopy.o bitops.o checksum.o io.o memset.o fixup.o memcpy.o \ -+lib-y := lusercopy.o bitops.o checksum.o io.o memset.o memcpy.o \ - ucmpdi2.o delay.o - - obj-y := iomap.o -diff --git a/arch/parisc/lib/fixup.S b/arch/parisc/lib/fixup.S -deleted file mode 100644 -index a5b72f2..0000000 ---- a/arch/parisc/lib/fixup.S -+++ /dev/null -@@ -1,98 +0,0 @@ --/* -- * Linux/PA-RISC Project (http://www.parisc-linux.org/) -- * -- * Copyright (C) 2004 Randolph Chung <tausq@debian.org> -- * -- * This program is free software; you can redistribute it and/or modify -- * it under the terms of the GNU General Public License as published by -- * the Free Software Foundation; either version 2, or (at your option) -- * any later version. -- * -- * This program is distributed in the hope that it will be useful, -- * but WITHOUT ANY WARRANTY; without even the implied warranty of -- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the -- * GNU General Public License for more details. -- * -- * You should have received a copy of the GNU General Public License -- * along with this program; if not, write to the Free Software -- * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA. -- * -- * Fixup routines for kernel exception handling. -- */ --#include <asm/asm-offsets.h> --#include <asm/assembly.h> --#include <asm/errno.h> --#include <linux/linkage.h> -- --#ifdef CONFIG_SMP -- .macro get_fault_ip t1 t2 -- loadgp -- addil LT%__per_cpu_offset,%r27 -- LDREG RT%__per_cpu_offset(%r1),\t1 -- /* t2 = smp_processor_id() */ -- mfctl 30,\t2 -- ldw TI_CPU(\t2),\t2 --#ifdef CONFIG_64BIT -- extrd,u \t2,63,32,\t2 --#endif -- /* t2 = &__per_cpu_offset[smp_processor_id()]; */ -- LDREGX \t2(\t1),\t2 -- addil LT%exception_data,%r27 -- LDREG RT%exception_data(%r1),\t1 -- /* t1 = this_cpu_ptr(&exception_data) */ -- add,l \t1,\t2,\t1 -- /* %r27 = t1->fault_gp - restore gp */ -- LDREG EXCDATA_GP(\t1), %r27 -- /* t1 = t1->fault_ip */ -- LDREG EXCDATA_IP(\t1), \t1 -- .endm --#else -- .macro get_fault_ip t1 t2 -- loadgp -- /* t1 = this_cpu_ptr(&exception_data) */ -- addil LT%exception_data,%r27 -- LDREG RT%exception_data(%r1),\t2 -- /* %r27 = t2->fault_gp - restore gp */ -- LDREG EXCDATA_GP(\t2), %r27 -- /* t1 = t2->fault_ip */ -- LDREG EXCDATA_IP(\t2), \t1 -- .endm --#endif -- -- .level LEVEL -- -- .text -- .section .fixup, "ax" -- -- /* get_user() fixups, store -EFAULT in r8, and 0 in r9 */ --ENTRY_CFI(fixup_get_user_skip_1) -- get_fault_ip %r1,%r8 -- ldo 4(%r1), %r1 -- ldi -EFAULT, %r8 -- bv %r0(%r1) -- copy %r0, %r9 --ENDPROC_CFI(fixup_get_user_skip_1) -- --ENTRY_CFI(fixup_get_user_skip_2) -- get_fault_ip %r1,%r8 -- ldo 8(%r1), %r1 -- ldi -EFAULT, %r8 -- bv %r0(%r1) -- copy %r0, %r9 --ENDPROC_CFI(fixup_get_user_skip_2) -- -- /* put_user() fixups, store -EFAULT in r8 */ --ENTRY_CFI(fixup_put_user_skip_1) -- get_fault_ip %r1,%r8 -- ldo 4(%r1), %r1 -- bv %r0(%r1) -- ldi -EFAULT, %r8 --ENDPROC_CFI(fixup_put_user_skip_1) -- --ENTRY_CFI(fixup_put_user_skip_2) -- get_fault_ip %r1,%r8 -- ldo 8(%r1), %r1 -- bv %r0(%r1) -- ldi -EFAULT, %r8 --ENDPROC_CFI(fixup_put_user_skip_2) -- -diff --git a/arch/parisc/lib/lusercopy.S b/arch/parisc/lib/lusercopy.S -index 56845de..f01188c 100644 ---- a/arch/parisc/lib/lusercopy.S -+++ b/arch/parisc/lib/lusercopy.S -@@ -5,6 +5,8 @@ - * Copyright (C) 2000 Richard Hirst <rhirst with parisc-linux.org> - * Copyright (C) 2001 Matthieu Delahaye <delahaym at esiee.fr> - * Copyright (C) 2003 Randolph Chung <tausq with parisc-linux.org> -+ * Copyright (C) 2017 Helge Deller <deller@gmx.de> -+ * Copyright (C) 2017 John David Anglin <dave.anglin@bell.net> - * - * - * This program is free software; you can redistribute it and/or modify -@@ -132,4 +134,320 @@ ENDPROC_CFI(lstrnlen_user) - - .procend - -+ -+ -+/* -+ * unsigned long pa_memcpy(void *dstp, const void *srcp, unsigned long len) -+ * -+ * Inputs: -+ * - sr1 already contains space of source region -+ * - sr2 already contains space of destination region -+ * -+ * Returns: -+ * - number of bytes that could not be copied. -+ * On success, this will be zero. -+ * -+ * This code is based on a C-implementation of a copy routine written by -+ * Randolph Chung, which in turn was derived from the glibc. -+ * -+ * Several strategies are tried to try to get the best performance for various -+ * conditions. In the optimal case, we copy by loops that copy 32- or 16-bytes -+ * at a time using general registers. Unaligned copies are handled either by -+ * aligning the destination and then using shift-and-write method, or in a few -+ * cases by falling back to a byte-at-a-time copy. -+ * -+ * Testing with various alignments and buffer sizes shows that this code is -+ * often >10x faster than a simple byte-at-a-time copy, even for strangely -+ * aligned operands. It is interesting to note that the glibc version of memcpy -+ * (written in C) is actually quite fast already. This routine is able to beat -+ * it by 30-40% for aligned copies because of the loop unrolling, but in some -+ * cases the glibc version is still slightly faster. This lends more -+ * credibility that gcc can generate very good code as long as we are careful. -+ * -+ * Possible optimizations: -+ * - add cache prefetching -+ * - try not to use the post-increment address modifiers; they may create -+ * additional interlocks. Assumption is that those were only efficient on old -+ * machines (pre PA8000 processors) -+ */ -+ -+ dst = arg0 -+ src = arg1 -+ len = arg2 -+ end = arg3 -+ t1 = r19 -+ t2 = r20 -+ t3 = r21 -+ t4 = r22 -+ srcspc = sr1 -+ dstspc = sr2 -+ -+ t0 = r1 -+ a1 = t1 -+ a2 = t2 -+ a3 = t3 -+ a0 = t4 -+ -+ save_src = ret0 -+ save_dst = ret1 -+ save_len = r31 -+ -+ENTRY_CFI(pa_memcpy) -+ .proc -+ .callinfo NO_CALLS -+ .entry -+ -+ /* Last destination address */ -+ add dst,len,end -+ -+ /* short copy with less than 16 bytes? */ -+ cmpib,>>=,n 15,len,.Lbyte_loop -+ -+ /* same alignment? */ -+ xor src,dst,t0 -+ extru t0,31,2,t1 -+ cmpib,<>,n 0,t1,.Lunaligned_copy -+ -+#ifdef CONFIG_64BIT -+ /* only do 64-bit copies if we can get aligned. */ -+ extru t0,31,3,t1 -+ cmpib,<>,n 0,t1,.Lalign_loop32 -+ -+ /* loop until we are 64-bit aligned */ -+.Lalign_loop64: -+ extru dst,31,3,t1 -+ cmpib,=,n 0,t1,.Lcopy_loop_16 -+20: ldb,ma 1(srcspc,src),t1 -+21: stb,ma t1,1(dstspc,dst) -+ b .Lalign_loop64 -+ ldo -1(len),len -+ -+ ASM_EXCEPTIONTABLE_ENTRY(20b,.Lcopy_done) -+ ASM_EXCEPTIONTABLE_ENTRY(21b,.Lcopy_done) -+ -+ ldi 31,t0 -+.Lcopy_loop_16: -+ cmpb,COND(>>=),n t0,len,.Lword_loop -+ -+10: ldd 0(srcspc,src),t1 -+11: ldd 8(srcspc,src),t2 -+ ldo 16(src),src -+12: std,ma t1,8(dstspc,dst) -+13: std,ma t2,8(dstspc,dst) -+14: ldd 0(srcspc,src),t1 -+15: ldd 8(srcspc,src),t2 -+ ldo 16(src),src -+16: std,ma t1,8(dstspc,dst) -+17: std,ma t2,8(dstspc,dst) -+ -+ ASM_EXCEPTIONTABLE_ENTRY(10b,.Lcopy_done) -+ ASM_EXCEPTIONTABLE_ENTRY(11b,.Lcopy16_fault) -+ ASM_EXCEPTIONTABLE_ENTRY(12b,.Lcopy_done) -+ ASM_EXCEPTIONTABLE_ENTRY(13b,.Lcopy_done) -+ ASM_EXCEPTIONTABLE_ENTRY(14b,.Lcopy_done) -+ ASM_EXCEPTIONTABLE_ENTRY(15b,.Lcopy16_fault) -+ ASM_EXCEPTIONTABLE_ENTRY(16b,.Lcopy_done) -+ ASM_EXCEPTIONTABLE_ENTRY(17b,.Lcopy_done) -+ -+ b .Lcopy_loop_16 -+ ldo -32(len),len -+ -+.Lword_loop: -+ cmpib,COND(>>=),n 3,len,.Lbyte_loop -+20: ldw,ma 4(srcspc,src),t1 -+21: stw,ma t1,4(dstspc,dst) -+ b .Lword_loop -+ ldo -4(len),len -+ -+ ASM_EXCEPTIONTABLE_ENTRY(20b,.Lcopy_done) -+ ASM_EXCEPTIONTABLE_ENTRY(21b,.Lcopy_done) -+ -+#endif /* CONFIG_64BIT */ -+ -+ /* loop until we are 32-bit aligned */ -+.Lalign_loop32: -+ extru dst,31,2,t1 -+ cmpib,=,n 0,t1,.Lcopy_loop_4 -+20: ldb,ma 1(srcspc,src),t1 -+21: stb,ma t1,1(dstspc,dst) -+ b .Lalign_loop32 -+ ldo -1(len),len -+ -+ ASM_EXCEPTIONTABLE_ENTRY(20b,.Lcopy_done) -+ ASM_EXCEPTIONTABLE_ENTRY(21b,.Lcopy_done) -+ -+ -+.Lcopy_loop_4: -+ cmpib,COND(>>=),n 15,len,.Lbyte_loop -+ -+10: ldw 0(srcspc,src),t1 -+11: ldw 4(srcspc,src),t2 -+12: stw,ma t1,4(dstspc,dst) -+13: stw,ma t2,4(dstspc,dst) -+14: ldw 8(srcspc,src),t1 -+15: ldw 12(srcspc,src),t2 -+ ldo 16(src),src -+16: stw,ma t1,4(dstspc,dst) -+17: stw,ma t2,4(dstspc,dst) -+ -+ ASM_EXCEPTIONTABLE_ENTRY(10b,.Lcopy_done) -+ ASM_EXCEPTIONTABLE_ENTRY(11b,.Lcopy8_fault) -+ ASM_EXCEPTIONTABLE_ENTRY(12b,.Lcopy_done) -+ ASM_EXCEPTIONTABLE_ENTRY(13b,.Lcopy_done) -+ ASM_EXCEPTIONTABLE_ENTRY(14b,.Lcopy_done) -+ ASM_EXCEPTIONTABLE_ENTRY(15b,.Lcopy8_fault) -+ ASM_EXCEPTIONTABLE_ENTRY(16b,.Lcopy_done) -+ ASM_EXCEPTIONTABLE_ENTRY(17b,.Lcopy_done) -+ -+ b .Lcopy_loop_4 -+ ldo -16(len),len -+ -+.Lbyte_loop: -+ cmpclr,COND(<>) len,%r0,%r0 -+ b,n .Lcopy_done -+20: ldb 0(srcspc,src),t1 -+ ldo 1(src),src -+21: stb,ma t1,1(dstspc,dst) -+ b .Lbyte_loop -+ ldo -1(len),len -+ -+ ASM_EXCEPTIONTABLE_ENTRY(20b,.Lcopy_done) -+ ASM_EXCEPTIONTABLE_ENTRY(21b,.Lcopy_done) -+ -+.Lcopy_done: -+ bv %r0(%r2) -+ sub end,dst,ret0 -+ -+ -+ /* src and dst are not aligned the same way. */ -+ /* need to go the hard way */ -+.Lunaligned_copy: -+ /* align until dst is 32bit-word-aligned */ -+ extru dst,31,2,t1 -+ cmpib,COND(=),n 0,t1,.Lcopy_dstaligned -+20: ldb 0(srcspc,src),t1 -+ ldo 1(src),src -+21: stb,ma t1,1(dstspc,dst) -+ b .Lunaligned_copy -+ ldo -1(len),len -+ -+ ASM_EXCEPTIONTABLE_ENTRY(20b,.Lcopy_done) -+ ASM_EXCEPTIONTABLE_ENTRY(21b,.Lcopy_done) -+ -+.Lcopy_dstaligned: -+ -+ /* store src, dst and len in safe place */ -+ copy src,save_src -+ copy dst,save_dst -+ copy len,save_len -+ -+ /* len now needs give number of words to copy */ -+ SHRREG len,2,len -+ -+ /* -+ * Copy from a not-aligned src to an aligned dst using shifts. -+ * Handles 4 words per loop. -+ */ -+ -+ depw,z src,28,2,t0 -+ subi 32,t0,t0 -+ mtsar t0 -+ extru len,31,2,t0 -+ cmpib,= 2,t0,.Lcase2 -+ /* Make src aligned by rounding it down. */ -+ depi 0,31,2,src -+ -+ cmpiclr,<> 3,t0,%r0 -+ b,n .Lcase3 -+ cmpiclr,<> 1,t0,%r0 -+ b,n .Lcase1 -+.Lcase0: -+ cmpb,= %r0,len,.Lcda_finish -+ nop -+ -+1: ldw,ma 4(srcspc,src), a3 -+ ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfault) -+1: ldw,ma 4(srcspc,src), a0 -+ ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfault) -+ b,n .Ldo3 -+.Lcase1: -+1: ldw,ma 4(srcspc,src), a2 -+ ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfault) -+1: ldw,ma 4(srcspc,src), a3 -+ ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfault) -+ ldo -1(len),len -+ cmpb,=,n %r0,len,.Ldo0 -+.Ldo4: -+1: ldw,ma 4(srcspc,src), a0 -+ ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfault) -+ shrpw a2, a3, %sar, t0 -+1: stw,ma t0, 4(dstspc,dst) -+ ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcopy_done) -+.Ldo3: -+1: ldw,ma 4(srcspc,src), a1 -+ ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfault) -+ shrpw a3, a0, %sar, t0 -+1: stw,ma t0, 4(dstspc,dst) -+ ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcopy_done) -+.Ldo2: -+1: ldw,ma 4(srcspc,src), a2 -+ ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfault) -+ shrpw a0, a1, %sar, t0 -+1: stw,ma t0, 4(dstspc,dst) -+ ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcopy_done) -+.Ldo1: -+1: ldw,ma 4(srcspc,src), a3 -+ ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfault) -+ shrpw a1, a2, %sar, t0 -+1: stw,ma t0, 4(dstspc,dst) -+ ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcopy_done) -+ ldo -4(len),len -+ cmpb,<> %r0,len,.Ldo4 -+ nop -+.Ldo0: -+ shrpw a2, a3, %sar, t0 -+1: stw,ma t0, 4(dstspc,dst) -+ ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcopy_done) -+ -+.Lcda_rdfault: -+.Lcda_finish: -+ /* calculate new src, dst and len and jump to byte-copy loop */ -+ sub dst,save_dst,t0 -+ add save_src,t0,src -+ b .Lbyte_loop -+ sub save_len,t0,len -+ -+.Lcase3: -+1: ldw,ma 4(srcspc,src), a0 -+ ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfault) -+1: ldw,ma 4(srcspc,src), a1 -+ ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfault) -+ b .Ldo2 -+ ldo 1(len),len -+.Lcase2: -+1: ldw,ma 4(srcspc,src), a1 -+ ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfault) -+1: ldw,ma 4(srcspc,src), a2 -+ ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfault) -+ b .Ldo1 -+ ldo 2(len),len -+ -+ -+ /* fault exception fixup handlers: */ -+#ifdef CONFIG_64BIT -+.Lcopy16_fault: -+10: b .Lcopy_done -+ std,ma t1,8(dstspc,dst) -+ ASM_EXCEPTIONTABLE_ENTRY(10b,.Lcopy_done) -+#endif -+ -+.Lcopy8_fault: -+10: b .Lcopy_done -+ stw,ma t1,4(dstspc,dst) -+ ASM_EXCEPTIONTABLE_ENTRY(10b,.Lcopy_done) -+ -+ .exit -+ENDPROC_CFI(pa_memcpy) -+ .procend -+ - .end -diff --git a/arch/parisc/lib/memcpy.c b/arch/parisc/lib/memcpy.c -index f82ff10..b3d47ec 100644 ---- a/arch/parisc/lib/memcpy.c -+++ b/arch/parisc/lib/memcpy.c -@@ -2,7 +2,7 @@ - * Optimized memory copy routines. - * - * Copyright (C) 2004 Randolph Chung <tausq@debian.org> -- * Copyright (C) 2013 Helge Deller <deller@gmx.de> -+ * Copyright (C) 2013-2017 Helge Deller <deller@gmx.de> - * - * This program is free software; you can redistribute it and/or modify - * it under the terms of the GNU General Public License as published by -@@ -21,474 +21,21 @@ - * Portions derived from the GNU C Library - * Copyright (C) 1991, 1997, 2003 Free Software Foundation, Inc. - * -- * Several strategies are tried to try to get the best performance for various -- * conditions. In the optimal case, we copy 64-bytes in an unrolled loop using -- * fp regs. This is followed by loops that copy 32- or 16-bytes at a time using -- * general registers. Unaligned copies are handled either by aligning the -- * destination and then using shift-and-write method, or in a few cases by -- * falling back to a byte-at-a-time copy. -- * -- * I chose to implement this in C because it is easier to maintain and debug, -- * and in my experiments it appears that the C code generated by gcc (3.3/3.4 -- * at the time of writing) is fairly optimal. Unfortunately some of the -- * semantics of the copy routine (exception handling) is difficult to express -- * in C, so we have to play some tricks to get it to work. -- * -- * All the loads and stores are done via explicit asm() code in order to use -- * the right space registers. -- * -- * Testing with various alignments and buffer sizes shows that this code is -- * often >10x faster than a simple byte-at-a-time copy, even for strangely -- * aligned operands. It is interesting to note that the glibc version -- * of memcpy (written in C) is actually quite fast already. This routine is -- * able to beat it by 30-40% for aligned copies because of the loop unrolling, -- * but in some cases the glibc version is still slightly faster. This lends -- * more credibility that gcc can generate very good code as long as we are -- * careful. -- * -- * TODO: -- * - cache prefetching needs more experimentation to get optimal settings -- * - try not to use the post-increment address modifiers; they create additional -- * interlocks -- * - replace byte-copy loops with stybs sequences - */ - --#ifdef __KERNEL__ - #include <linux/module.h> - #include <linux/compiler.h> - #include <linux/uaccess.h> --#define s_space "%%sr1" --#define d_space "%%sr2" --#else --#include "memcpy.h" --#define s_space "%%sr0" --#define d_space "%%sr0" --#define pa_memcpy new2_copy --#endif - - DECLARE_PER_CPU(struct exception_data, exception_data); - --#define preserve_branch(label) do { \ -- volatile int dummy = 0; \ -- /* The following branch is never taken, it's just here to */ \ -- /* prevent gcc from optimizing away our exception code. */ \ -- if (unlikely(dummy != dummy)) \ -- goto label; \ --} while (0) -- - #define get_user_space() (segment_eq(get_fs(), KERNEL_DS) ? 0 : mfsp(3)) - #define get_kernel_space() (0) - --#define MERGE(w0, sh_1, w1, sh_2) ({ \ -- unsigned int _r; \ -- asm volatile ( \ -- "mtsar %3\n" \ -- "shrpw %1, %2, %%sar, %0\n" \ -- : "=r"(_r) \ -- : "r"(w0), "r"(w1), "r"(sh_2) \ -- ); \ -- _r; \ --}) --#define THRESHOLD 16 -- --#ifdef DEBUG_MEMCPY --#define DPRINTF(fmt, args...) do { printk(KERN_DEBUG "%s:%d:%s ", __FILE__, __LINE__, __func__ ); printk(KERN_DEBUG fmt, ##args ); } while (0) --#else --#define DPRINTF(fmt, args...) --#endif -- --#define def_load_ai_insn(_insn,_sz,_tt,_s,_a,_t,_e) \ -- __asm__ __volatile__ ( \ -- "1:\t" #_insn ",ma " #_sz "(" _s ",%1), %0\n\t" \ -- ASM_EXCEPTIONTABLE_ENTRY(1b,_e) \ -- : _tt(_t), "+r"(_a) \ -- : \ -- : "r8") -- --#define def_store_ai_insn(_insn,_sz,_tt,_s,_a,_t,_e) \ -- __asm__ __volatile__ ( \ -- "1:\t" #_insn ",ma %1, " #_sz "(" _s ",%0)\n\t" \ -- ASM_EXCEPTIONTABLE_ENTRY(1b,_e) \ -- : "+r"(_a) \ -- : _tt(_t) \ -- : "r8") -- --#define ldbma(_s, _a, _t, _e) def_load_ai_insn(ldbs,1,"=r",_s,_a,_t,_e) --#define stbma(_s, _t, _a, _e) def_store_ai_insn(stbs,1,"r",_s,_a,_t,_e) --#define ldwma(_s, _a, _t, _e) def_load_ai_insn(ldw,4,"=r",_s,_a,_t,_e) --#define stwma(_s, _t, _a, _e) def_store_ai_insn(stw,4,"r",_s,_a,_t,_e) --#define flddma(_s, _a, _t, _e) def_load_ai_insn(fldd,8,"=f",_s,_a,_t,_e) --#define fstdma(_s, _t, _a, _e) def_store_ai_insn(fstd,8,"f",_s,_a,_t,_e) -- --#define def_load_insn(_insn,_tt,_s,_o,_a,_t,_e) \ -- __asm__ __volatile__ ( \ -- "1:\t" #_insn " " #_o "(" _s ",%1), %0\n\t" \ -- ASM_EXCEPTIONTABLE_ENTRY(1b,_e) \ -- : _tt(_t) \ -- : "r"(_a) \ -- : "r8") -- --#define def_store_insn(_insn,_tt,_s,_t,_o,_a,_e) \ -- __asm__ __volatile__ ( \ -- "1:\t" #_insn " %0, " #_o "(" _s ",%1)\n\t" \ -- ASM_EXCEPTIONTABLE_ENTRY(1b,_e) \ -- : \ -- : _tt(_t), "r"(_a) \ -- : "r8") -- --#define ldw(_s,_o,_a,_t,_e) def_load_insn(ldw,"=r",_s,_o,_a,_t,_e) --#define stw(_s,_t,_o,_a,_e) def_store_insn(stw,"r",_s,_t,_o,_a,_e) -- --#ifdef CONFIG_PREFETCH --static inline void prefetch_src(const void *addr) --{ -- __asm__("ldw 0(" s_space ",%0), %%r0" : : "r" (addr)); --} -- --static inline void prefetch_dst(const void *addr) --{ -- __asm__("ldd 0(" d_space ",%0), %%r0" : : "r" (addr)); --} --#else --#define prefetch_src(addr) do { } while(0) --#define prefetch_dst(addr) do { } while(0) --#endif -- --#define PA_MEMCPY_OK 0 --#define PA_MEMCPY_LOAD_ERROR 1 --#define PA_MEMCPY_STORE_ERROR 2 -- --/* Copy from a not-aligned src to an aligned dst, using shifts. Handles 4 words -- * per loop. This code is derived from glibc. -- */ --static noinline unsigned long copy_dstaligned(unsigned long dst, -- unsigned long src, unsigned long len) --{ -- /* gcc complains that a2 and a3 may be uninitialized, but actually -- * they cannot be. Initialize a2/a3 to shut gcc up. -- */ -- register unsigned int a0, a1, a2 = 0, a3 = 0; -- int sh_1, sh_2; -- -- /* prefetch_src((const void *)src); */ -- -- /* Calculate how to shift a word read at the memory operation -- aligned srcp to make it aligned for copy. */ -- sh_1 = 8 * (src % sizeof(unsigned int)); -- sh_2 = 8 * sizeof(unsigned int) - sh_1; -- -- /* Make src aligned by rounding it down. */ -- src &= -sizeof(unsigned int); -- -- switch (len % 4) -- { -- case 2: -- /* a1 = ((unsigned int *) src)[0]; -- a2 = ((unsigned int *) src)[1]; */ -- ldw(s_space, 0, src, a1, cda_ldw_exc); -- ldw(s_space, 4, src, a2, cda_ldw_exc); -- src -= 1 * sizeof(unsigned int); -- dst -= 3 * sizeof(unsigned int); -- len += 2; -- goto do1; -- case 3: -- /* a0 = ((unsigned int *) src)[0]; -- a1 = ((unsigned int *) src)[1]; */ -- ldw(s_space, 0, src, a0, cda_ldw_exc); -- ldw(s_space, 4, src, a1, cda_ldw_exc); -- src -= 0 * sizeof(unsigned int); -- dst -= 2 * sizeof(unsigned int); -- len += 1; -- goto do2; -- case 0: -- if (len == 0) -- return PA_MEMCPY_OK; -- /* a3 = ((unsigned int *) src)[0]; -- a0 = ((unsigned int *) src)[1]; */ -- ldw(s_space, 0, src, a3, cda_ldw_exc); -- ldw(s_space, 4, src, a0, cda_ldw_exc); -- src -=-1 * sizeof(unsigned int); -- dst -= 1 * sizeof(unsigned int); -- len += 0; -- goto do3; -- case 1: -- /* a2 = ((unsigned int *) src)[0]; -- a3 = ((unsigned int *) src)[1]; */ -- ldw(s_space, 0, src, a2, cda_ldw_exc); -- ldw(s_space, 4, src, a3, cda_ldw_exc); -- src -=-2 * sizeof(unsigned int); -- dst -= 0 * sizeof(unsigned int); -- len -= 1; -- if (len == 0) -- goto do0; -- goto do4; /* No-op. */ -- } -- -- do -- { -- /* prefetch_src((const void *)(src + 4 * sizeof(unsigned int))); */ --do4: -- /* a0 = ((unsigned int *) src)[0]; */ -- ldw(s_space, 0, src, a0, cda_ldw_exc); -- /* ((unsigned int *) dst)[0] = MERGE (a2, sh_1, a3, sh_2); */ -- stw(d_space, MERGE (a2, sh_1, a3, sh_2), 0, dst, cda_stw_exc); --do3: -- /* a1 = ((unsigned int *) src)[1]; */ -- ldw(s_space, 4, src, a1, cda_ldw_exc); -- /* ((unsigned int *) dst)[1] = MERGE (a3, sh_1, a0, sh_2); */ -- stw(d_space, MERGE (a3, sh_1, a0, sh_2), 4, dst, cda_stw_exc); --do2: -- /* a2 = ((unsigned int *) src)[2]; */ -- ldw(s_space, 8, src, a2, cda_ldw_exc); -- /* ((unsigned int *) dst)[2] = MERGE (a0, sh_1, a1, sh_2); */ -- stw(d_space, MERGE (a0, sh_1, a1, sh_2), 8, dst, cda_stw_exc); --do1: -- /* a3 = ((unsigned int *) src)[3]; */ -- ldw(s_space, 12, src, a3, cda_ldw_exc); -- /* ((unsigned int *) dst)[3] = MERGE (a1, sh_1, a2, sh_2); */ -- stw(d_space, MERGE (a1, sh_1, a2, sh_2), 12, dst, cda_stw_exc); -- -- src += 4 * sizeof(unsigned int); -- dst += 4 * sizeof(unsigned int); -- len -= 4; -- } -- while (len != 0); -- --do0: -- /* ((unsigned int *) dst)[0] = MERGE (a2, sh_1, a3, sh_2); */ -- stw(d_space, MERGE (a2, sh_1, a3, sh_2), 0, dst, cda_stw_exc); -- -- preserve_branch(handle_load_error); -- preserve_branch(handle_store_error); -- -- return PA_MEMCPY_OK; -- --handle_load_error: -- __asm__ __volatile__ ("cda_ldw_exc:\n"); -- return PA_MEMCPY_LOAD_ERROR; -- --handle_store_error: -- __asm__ __volatile__ ("cda_stw_exc:\n"); -- return PA_MEMCPY_STORE_ERROR; --} -- -- --/* Returns PA_MEMCPY_OK, PA_MEMCPY_LOAD_ERROR or PA_MEMCPY_STORE_ERROR. -- * In case of an access fault the faulty address can be read from the per_cpu -- * exception data struct. */ --static noinline unsigned long pa_memcpy_internal(void *dstp, const void *srcp, -- unsigned long len) --{ -- register unsigned long src, dst, t1, t2, t3; -- register unsigned char *pcs, *pcd; -- register unsigned int *pws, *pwd; -- register double *pds, *pdd; -- unsigned long ret; -- -- src = (unsigned long)srcp; -- dst = (unsigned long)dstp; -- pcs = (unsigned char *)srcp; -- pcd = (unsigned char *)dstp; -- -- /* prefetch_src((const void *)srcp); */ -- -- if (len < THRESHOLD) -- goto byte_copy; -- -- /* Check alignment */ -- t1 = (src ^ dst); -- if (unlikely(t1 & (sizeof(double)-1))) -- goto unaligned_copy; -- -- /* src and dst have same alignment. */ -- -- /* Copy bytes till we are double-aligned. */ -- t2 = src & (sizeof(double) - 1); -- if (unlikely(t2 != 0)) { -- t2 = sizeof(double) - t2; -- while (t2 && len) { -- /* *pcd++ = *pcs++; */ -- ldbma(s_space, pcs, t3, pmc_load_exc); -- len--; -- stbma(d_space, t3, pcd, pmc_store_exc); -- t2--; -- } -- } -- -- pds = (double *)pcs; -- pdd = (double *)pcd; -- --#if 0 -- /* Copy 8 doubles at a time */ -- while (len >= 8*sizeof(double)) { -- register double r1, r2, r3, r4, r5, r6, r7, r8; -- /* prefetch_src((char *)pds + L1_CACHE_BYTES); */ -- flddma(s_space, pds, r1, pmc_load_exc); -- flddma(s_space, pds, r2, pmc_load_exc); -- flddma(s_space, pds, r3, pmc_load_exc); -- flddma(s_space, pds, r4, pmc_load_exc); -- fstdma(d_space, r1, pdd, pmc_store_exc); -- fstdma(d_space, r2, pdd, pmc_store_exc); -- fstdma(d_space, r3, pdd, pmc_store_exc); -- fstdma(d_space, r4, pdd, pmc_store_exc); -- --#if 0 -- if (L1_CACHE_BYTES <= 32) -- prefetch_src((char *)pds + L1_CACHE_BYTES); --#endif -- flddma(s_space, pds, r5, pmc_load_exc); -- flddma(s_space, pds, r6, pmc_load_exc); -- flddma(s_space, pds, r7, pmc_load_exc); -- flddma(s_space, pds, r8, pmc_load_exc); -- fstdma(d_space, r5, pdd, pmc_store_exc); -- fstdma(d_space, r6, pdd, pmc_store_exc); -- fstdma(d_space, r7, pdd, pmc_store_exc); -- fstdma(d_space, r8, pdd, pmc_store_exc); -- len -= 8*sizeof(double); -- } --#endif -- -- pws = (unsigned int *)pds; -- pwd = (unsigned int *)pdd; -- --word_copy: -- while (len >= 8*sizeof(unsigned int)) { -- register unsigned int r1,r2,r3,r4,r5,r6,r7,r8; -- /* prefetch_src((char *)pws + L1_CACHE_BYTES); */ -- ldwma(s_space, pws, r1, pmc_load_exc); -- ldwma(s_space, pws, r2, pmc_load_exc); -- ldwma(s_space, pws, r3, pmc_load_exc); -- ldwma(s_space, pws, r4, pmc_load_exc); -- stwma(d_space, r1, pwd, pmc_store_exc); -- stwma(d_space, r2, pwd, pmc_store_exc); -- stwma(d_space, r3, pwd, pmc_store_exc); -- stwma(d_space, r4, pwd, pmc_store_exc); -- -- ldwma(s_space, pws, r5, pmc_load_exc); -- ldwma(s_space, pws, r6, pmc_load_exc); -- ldwma(s_space, pws, r7, pmc_load_exc); -- ldwma(s_space, pws, r8, pmc_load_exc); -- stwma(d_space, r5, pwd, pmc_store_exc); -- stwma(d_space, r6, pwd, pmc_store_exc); -- stwma(d_space, r7, pwd, pmc_store_exc); -- stwma(d_space, r8, pwd, pmc_store_exc); -- len -= 8*sizeof(unsigned int); -- } -- -- while (len >= 4*sizeof(unsigned int)) { -- register unsigned int r1,r2,r3,r4; -- ldwma(s_space, pws, r1, pmc_load_exc); -- ldwma(s_space, pws, r2, pmc_load_exc); -- ldwma(s_space, pws, r3, pmc_load_exc); -- ldwma(s_space, pws, r4, pmc_load_exc); -- stwma(d_space, r1, pwd, pmc_store_exc); -- stwma(d_space, r2, pwd, pmc_store_exc); -- stwma(d_space, r3, pwd, pmc_store_exc); -- stwma(d_space, r4, pwd, pmc_store_exc); -- len -= 4*sizeof(unsigned int); -- } -- -- pcs = (unsigned char *)pws; -- pcd = (unsigned char *)pwd; -- --byte_copy: -- while (len) { -- /* *pcd++ = *pcs++; */ -- ldbma(s_space, pcs, t3, pmc_load_exc); -- stbma(d_space, t3, pcd, pmc_store_exc); -- len--; -- } -- -- return PA_MEMCPY_OK; -- --unaligned_copy: -- /* possibly we are aligned on a word, but not on a double... */ -- if (likely((t1 & (sizeof(unsigned int)-1)) == 0)) { -- t2 = src & (sizeof(unsigned int) - 1); -- -- if (unlikely(t2 != 0)) { -- t2 = sizeof(unsigned int) - t2; -- while (t2) { -- /* *pcd++ = *pcs++; */ -- ldbma(s_space, pcs, t3, pmc_load_exc); -- stbma(d_space, t3, pcd, pmc_store_exc); -- len--; -- t2--; -- } -- } -- -- pws = (unsigned int *)pcs; -- pwd = (unsigned int *)pcd; -- goto word_copy; -- } -- -- /* Align the destination. */ -- if (unlikely((dst & (sizeof(unsigned int) - 1)) != 0)) { -- t2 = sizeof(unsigned int) - (dst & (sizeof(unsigned int) - 1)); -- while (t2) { -- /* *pcd++ = *pcs++; */ -- ldbma(s_space, pcs, t3, pmc_load_exc); -- stbma(d_space, t3, pcd, pmc_store_exc); -- len--; -- t2--; -- } -- dst = (unsigned long)pcd; -- src = (unsigned long)pcs; -- } -- -- ret = copy_dstaligned(dst, src, len / sizeof(unsigned int)); -- if (ret) -- return ret; -- -- pcs += (len & -sizeof(unsigned int)); -- pcd += (len & -sizeof(unsigned int)); -- len %= sizeof(unsigned int); -- -- preserve_branch(handle_load_error); -- preserve_branch(handle_store_error); -- -- goto byte_copy; -- --handle_load_error: -- __asm__ __volatile__ ("pmc_load_exc:\n"); -- return PA_MEMCPY_LOAD_ERROR; -- --handle_store_error: -- __asm__ __volatile__ ("pmc_store_exc:\n"); -- return PA_MEMCPY_STORE_ERROR; --} -- -- - /* Returns 0 for success, otherwise, returns number of bytes not transferred. */ --static unsigned long pa_memcpy(void *dstp, const void *srcp, unsigned long len) --{ -- unsigned long ret, fault_addr, reference; -- struct exception_data *d; -- -- ret = pa_memcpy_internal(dstp, srcp, len); -- if (likely(ret == PA_MEMCPY_OK)) -- return 0; -- -- /* if a load or store fault occured we can get the faulty addr */ -- d = this_cpu_ptr(&exception_data); -- fault_addr = d->fault_addr; -- -- /* error in load or store? */ -- if (ret == PA_MEMCPY_LOAD_ERROR) -- reference = (unsigned long) srcp; -- else -- reference = (unsigned long) dstp; -+extern unsigned long pa_memcpy(void *dst, const void *src, -+ unsigned long len); - -- DPRINTF("pa_memcpy: fault type = %lu, len=%lu fault_addr=%lu ref=%lu\n", -- ret, len, fault_addr, reference); -- -- if (fault_addr >= reference) -- return len - (fault_addr - reference); -- else -- return len; --} -- --#ifdef __KERNEL__ - unsigned long __copy_to_user(void __user *dst, const void *src, - unsigned long len) - { -@@ -537,5 +84,3 @@ long probe_kernel_read(void *dst, const void *src, size_t size) - - return __probe_kernel_read(dst, src, size); - } -- --#endif -diff --git a/arch/parisc/mm/fault.c b/arch/parisc/mm/fault.c -index 1a0b4f6..040c48f 100644 ---- a/arch/parisc/mm/fault.c -+++ b/arch/parisc/mm/fault.c -@@ -149,6 +149,23 @@ int fixup_exception(struct pt_regs *regs) - d->fault_space = regs->isr; - d->fault_addr = regs->ior; - -+ /* -+ * Fix up get_user() and put_user(). -+ * ASM_EXCEPTIONTABLE_ENTRY_EFAULT() sets the least-significant -+ * bit in the relative address of the fixup routine to indicate -+ * that %r8 should be loaded with -EFAULT to report a userspace -+ * access error. -+ */ -+ if (fix->fixup & 1) { -+ regs->gr[8] = -EFAULT; -+ -+ /* zero target register for get_user() */ -+ if (parisc_acctyp(0, regs->iir) == VM_READ) { -+ int treg = regs->iir & 0x1f; -+ regs->gr[treg] = 0; -+ } -+ } -+ - regs->iaoq[0] = (unsigned long)&fix->fixup + fix->fixup; - regs->iaoq[0] &= ~3; - /* -diff --git a/arch/x86/lib/memcpy_64.S b/arch/x86/lib/memcpy_64.S -index 779782f..9a53a06 100644 ---- a/arch/x86/lib/memcpy_64.S -+++ b/arch/x86/lib/memcpy_64.S -@@ -290,7 +290,7 @@ EXPORT_SYMBOL_GPL(memcpy_mcsafe_unrolled) - _ASM_EXTABLE_FAULT(.L_copy_leading_bytes, .L_memcpy_mcsafe_fail) - _ASM_EXTABLE_FAULT(.L_cache_w0, .L_memcpy_mcsafe_fail) - _ASM_EXTABLE_FAULT(.L_cache_w1, .L_memcpy_mcsafe_fail) -- _ASM_EXTABLE_FAULT(.L_cache_w3, .L_memcpy_mcsafe_fail) -+ _ASM_EXTABLE_FAULT(.L_cache_w2, .L_memcpy_mcsafe_fail) - _ASM_EXTABLE_FAULT(.L_cache_w3, .L_memcpy_mcsafe_fail) - _ASM_EXTABLE_FAULT(.L_cache_w4, .L_memcpy_mcsafe_fail) - _ASM_EXTABLE_FAULT(.L_cache_w5, .L_memcpy_mcsafe_fail) -diff --git a/arch/x86/mm/kaslr.c b/arch/x86/mm/kaslr.c -index 887e571..aed2064 100644 ---- a/arch/x86/mm/kaslr.c -+++ b/arch/x86/mm/kaslr.c -@@ -48,7 +48,7 @@ static const unsigned long vaddr_start = __PAGE_OFFSET_BASE; - #if defined(CONFIG_X86_ESPFIX64) - static const unsigned long vaddr_end = ESPFIX_BASE_ADDR; - #elif defined(CONFIG_EFI) --static const unsigned long vaddr_end = EFI_VA_START; -+static const unsigned long vaddr_end = EFI_VA_END; - #else - static const unsigned long vaddr_end = __START_KERNEL_map; - #endif -@@ -105,7 +105,7 @@ void __init kernel_randomize_memory(void) - */ - BUILD_BUG_ON(vaddr_start >= vaddr_end); - BUILD_BUG_ON(IS_ENABLED(CONFIG_X86_ESPFIX64) && -- vaddr_end >= EFI_VA_START); -+ vaddr_end >= EFI_VA_END); - BUILD_BUG_ON((IS_ENABLED(CONFIG_X86_ESPFIX64) || - IS_ENABLED(CONFIG_EFI)) && - vaddr_end >= __START_KERNEL_map); -diff --git a/arch/x86/xen/setup.c b/arch/x86/xen/setup.c -index f8960fc..9f21b0c 100644 ---- a/arch/x86/xen/setup.c -+++ b/arch/x86/xen/setup.c -@@ -713,10 +713,9 @@ static void __init xen_reserve_xen_mfnlist(void) - size = PFN_PHYS(xen_start_info->nr_p2m_frames); - } - -- if (!xen_is_e820_reserved(start, size)) { -- memblock_reserve(start, size); -+ memblock_reserve(start, size); -+ if (!xen_is_e820_reserved(start, size)) - return; -- } - - #ifdef CONFIG_X86_32 - /* -@@ -727,6 +726,7 @@ static void __init xen_reserve_xen_mfnlist(void) - BUG(); - #else - xen_relocate_p2m(); -+ memblock_free(start, size); - #endif - } - -diff --git a/block/bio.c b/block/bio.c -index db85c57..655c901 100644 ---- a/block/bio.c -+++ b/block/bio.c -@@ -372,10 +372,14 @@ static void punt_bios_to_rescuer(struct bio_set *bs) - bio_list_init(&punt); - bio_list_init(&nopunt); - -- while ((bio = bio_list_pop(current->bio_list))) -+ while ((bio = bio_list_pop(¤t->bio_list[0]))) - bio_list_add(bio->bi_pool == bs ? &punt : &nopunt, bio); -+ current->bio_list[0] = nopunt; - -- *current->bio_list = nopunt; -+ bio_list_init(&nopunt); -+ while ((bio = bio_list_pop(¤t->bio_list[1]))) -+ bio_list_add(bio->bi_pool == bs ? &punt : &nopunt, bio); -+ current->bio_list[1] = nopunt; - - spin_lock(&bs->rescue_lock); - bio_list_merge(&bs->rescue_list, &punt); -@@ -462,7 +466,9 @@ struct bio *bio_alloc_bioset(gfp_t gfp_mask, int nr_iovecs, struct bio_set *bs) - * we retry with the original gfp_flags. - */ - -- if (current->bio_list && !bio_list_empty(current->bio_list)) -+ if (current->bio_list && -+ (!bio_list_empty(¤t->bio_list[0]) || -+ !bio_list_empty(¤t->bio_list[1]))) - gfp_mask &= ~__GFP_DIRECT_RECLAIM; - - p = mempool_alloc(bs->bio_pool, gfp_mask); -diff --git a/block/blk-core.c b/block/blk-core.c -index 14d7c07..d1f2801 100644 ---- a/block/blk-core.c -+++ b/block/blk-core.c -@@ -1994,7 +1994,14 @@ generic_make_request_checks(struct bio *bio) - */ - blk_qc_t generic_make_request(struct bio *bio) - { -- struct bio_list bio_list_on_stack; -+ /* -+ * bio_list_on_stack[0] contains bios submitted by the current -+ * make_request_fn. -+ * bio_list_on_stack[1] contains bios that were submitted before -+ * the current make_request_fn, but that haven't been processed -+ * yet. -+ */ -+ struct bio_list bio_list_on_stack[2]; - blk_qc_t ret = BLK_QC_T_NONE; - - if (!generic_make_request_checks(bio)) -@@ -2011,7 +2018,7 @@ blk_qc_t generic_make_request(struct bio *bio) - * should be added at the tail - */ - if (current->bio_list) { -- bio_list_add(current->bio_list, bio); -+ bio_list_add(¤t->bio_list[0], bio); - goto out; - } - -@@ -2030,23 +2037,39 @@ blk_qc_t generic_make_request(struct bio *bio) - * bio_list, and call into ->make_request() again. - */ - BUG_ON(bio->bi_next); -- bio_list_init(&bio_list_on_stack); -- current->bio_list = &bio_list_on_stack; -+ bio_list_init(&bio_list_on_stack[0]); -+ current->bio_list = bio_list_on_stack; - do { - struct request_queue *q = bdev_get_queue(bio->bi_bdev); - - if (likely(blk_queue_enter(q, false) == 0)) { -+ struct bio_list lower, same; -+ -+ /* Create a fresh bio_list for all subordinate requests */ -+ bio_list_on_stack[1] = bio_list_on_stack[0]; -+ bio_list_init(&bio_list_on_stack[0]); - ret = q->make_request_fn(q, bio); - - blk_queue_exit(q); - -- bio = bio_list_pop(current->bio_list); -+ /* sort new bios into those for a lower level -+ * and those for the same level -+ */ -+ bio_list_init(&lower); -+ bio_list_init(&same); -+ while ((bio = bio_list_pop(&bio_list_on_stack[0])) != NULL) -+ if (q == bdev_get_queue(bio->bi_bdev)) -+ bio_list_add(&same, bio); -+ else -+ bio_list_add(&lower, bio); -+ /* now assemble so we handle the lowest level first */ -+ bio_list_merge(&bio_list_on_stack[0], &lower); -+ bio_list_merge(&bio_list_on_stack[0], &same); -+ bio_list_merge(&bio_list_on_stack[0], &bio_list_on_stack[1]); - } else { -- struct bio *bio_next = bio_list_pop(current->bio_list); -- - bio_io_error(bio); -- bio = bio_next; - } -+ bio = bio_list_pop(&bio_list_on_stack[0]); - } while (bio); - current->bio_list = NULL; /* deactivate */ - -diff --git a/drivers/acpi/Makefile b/drivers/acpi/Makefile -index 9ed0878..4c5678c 100644 ---- a/drivers/acpi/Makefile -+++ b/drivers/acpi/Makefile -@@ -2,7 +2,6 @@ - # Makefile for the Linux ACPI interpreter - # - --ccflags-y := -Os - ccflags-$(CONFIG_ACPI_DEBUG) += -DACPI_DEBUG_OUTPUT - - # -diff --git a/drivers/acpi/acpi_platform.c b/drivers/acpi/acpi_platform.c -index b4c1a6a..03250e1 100644 ---- a/drivers/acpi/acpi_platform.c -+++ b/drivers/acpi/acpi_platform.c -@@ -25,9 +25,11 @@ - ACPI_MODULE_NAME("platform"); - - static const struct acpi_device_id forbidden_id_list[] = { -- {"PNP0000", 0}, /* PIC */ -- {"PNP0100", 0}, /* Timer */ -- {"PNP0200", 0}, /* AT DMA Controller */ -+ {"PNP0000", 0}, /* PIC */ -+ {"PNP0100", 0}, /* Timer */ -+ {"PNP0200", 0}, /* AT DMA Controller */ -+ {"ACPI0009", 0}, /* IOxAPIC */ -+ {"ACPI000A", 0}, /* IOAPIC */ - {"", 0}, - }; - -diff --git a/drivers/gpu/drm/etnaviv/etnaviv_gpu.c b/drivers/gpu/drm/etnaviv/etnaviv_gpu.c -index b1254f8..b87d278 100644 ---- a/drivers/gpu/drm/etnaviv/etnaviv_gpu.c -+++ b/drivers/gpu/drm/etnaviv/etnaviv_gpu.c -@@ -1299,6 +1299,8 @@ int etnaviv_gpu_submit(struct etnaviv_gpu *gpu, - goto out_pm_put; - } - -+ mutex_lock(&gpu->lock); -+ - fence = etnaviv_gpu_fence_alloc(gpu); - if (!fence) { - event_free(gpu, event); -@@ -1306,8 +1308,6 @@ int etnaviv_gpu_submit(struct etnaviv_gpu *gpu, - goto out_pm_put; - } - -- mutex_lock(&gpu->lock); -- - gpu->event[event].fence = fence; - submit->fence = fence->seqno; - gpu->active_fence = submit->fence; -diff --git a/drivers/gpu/drm/radeon/radeon_ttm.c b/drivers/gpu/drm/radeon/radeon_ttm.c -index 3de5e6e..4ce04e0 100644 ---- a/drivers/gpu/drm/radeon/radeon_ttm.c -+++ b/drivers/gpu/drm/radeon/radeon_ttm.c -@@ -213,8 +213,8 @@ static void radeon_evict_flags(struct ttm_buffer_object *bo, - rbo->placement.num_busy_placement = 0; - for (i = 0; i < rbo->placement.num_placement; i++) { - if (rbo->placements[i].flags & TTM_PL_FLAG_VRAM) { -- if (rbo->placements[0].fpfn < fpfn) -- rbo->placements[0].fpfn = fpfn; -+ if (rbo->placements[i].fpfn < fpfn) -+ rbo->placements[i].fpfn = fpfn; - } else { - rbo->placement.busy_placement = - &rbo->placements[i]; -diff --git a/drivers/gpu/drm/vc4/vc4_crtc.c b/drivers/gpu/drm/vc4/vc4_crtc.c -index 7aadce1..c7e6c98 100644 ---- a/drivers/gpu/drm/vc4/vc4_crtc.c -+++ b/drivers/gpu/drm/vc4/vc4_crtc.c -@@ -842,6 +842,17 @@ static void vc4_crtc_destroy_state(struct drm_crtc *crtc, - drm_atomic_helper_crtc_destroy_state(crtc, state); - } - -+static void -+vc4_crtc_reset(struct drm_crtc *crtc) -+{ -+ if (crtc->state) -+ __drm_atomic_helper_crtc_destroy_state(crtc->state); -+ -+ crtc->state = kzalloc(sizeof(struct vc4_crtc_state), GFP_KERNEL); -+ if (crtc->state) -+ crtc->state->crtc = crtc; -+} -+ - static const struct drm_crtc_funcs vc4_crtc_funcs = { - .set_config = drm_atomic_helper_set_config, - .destroy = vc4_crtc_destroy, -@@ -849,7 +860,7 @@ static const struct drm_crtc_funcs vc4_crtc_funcs = { - .set_property = NULL, - .cursor_set = NULL, /* handled by drm_mode_cursor_universal */ - .cursor_move = NULL, /* handled by drm_mode_cursor_universal */ -- .reset = drm_atomic_helper_crtc_reset, -+ .reset = vc4_crtc_reset, - .atomic_duplicate_state = vc4_crtc_duplicate_state, - .atomic_destroy_state = vc4_crtc_destroy_state, - .gamma_set = vc4_crtc_gamma_set, -diff --git a/drivers/hid/wacom_sys.c b/drivers/hid/wacom_sys.c -index 5e7a564..0c535d0 100644 ---- a/drivers/hid/wacom_sys.c -+++ b/drivers/hid/wacom_sys.c -@@ -2017,6 +2017,14 @@ static int wacom_parse_and_register(struct wacom *wacom, bool wireless) - - wacom_update_name(wacom, wireless ? " (WL)" : ""); - -+ /* pen only Bamboo neither support touch nor pad */ -+ if ((features->type == BAMBOO_PEN) && -+ ((features->device_type & WACOM_DEVICETYPE_TOUCH) || -+ (features->device_type & WACOM_DEVICETYPE_PAD))) { -+ error = -ENODEV; -+ goto fail; -+ } -+ - error = wacom_add_shared_data(hdev); - if (error) - goto fail; -@@ -2064,14 +2072,6 @@ static int wacom_parse_and_register(struct wacom *wacom, bool wireless) - goto fail_quirks; - } - -- /* pen only Bamboo neither support touch nor pad */ -- if ((features->type == BAMBOO_PEN) && -- ((features->device_type & WACOM_DEVICETYPE_TOUCH) || -- (features->device_type & WACOM_DEVICETYPE_PAD))) { -- error = -ENODEV; -- goto fail_quirks; -- } -- - if (features->device_type & WACOM_DEVICETYPE_WL_MONITOR) - error = hid_hw_open(hdev); - -diff --git a/drivers/md/dm.c b/drivers/md/dm.c -index 628ba00..e66f404 100644 ---- a/drivers/md/dm.c -+++ b/drivers/md/dm.c -@@ -986,26 +986,29 @@ static void flush_current_bio_list(struct blk_plug_cb *cb, bool from_schedule) - struct dm_offload *o = container_of(cb, struct dm_offload, cb); - struct bio_list list; - struct bio *bio; -+ int i; - - INIT_LIST_HEAD(&o->cb.list); - - if (unlikely(!current->bio_list)) - return; - -- list = *current->bio_list; -- bio_list_init(current->bio_list); -- -- while ((bio = bio_list_pop(&list))) { -- struct bio_set *bs = bio->bi_pool; -- if (unlikely(!bs) || bs == fs_bio_set) { -- bio_list_add(current->bio_list, bio); -- continue; -+ for (i = 0; i < 2; i++) { -+ list = current->bio_list[i]; -+ bio_list_init(¤t->bio_list[i]); -+ -+ while ((bio = bio_list_pop(&list))) { -+ struct bio_set *bs = bio->bi_pool; -+ if (unlikely(!bs) || bs == fs_bio_set) { -+ bio_list_add(¤t->bio_list[i], bio); -+ continue; -+ } -+ -+ spin_lock(&bs->rescue_lock); -+ bio_list_add(&bs->rescue_list, bio); -+ queue_work(bs->rescue_workqueue, &bs->rescue_work); -+ spin_unlock(&bs->rescue_lock); - } -- -- spin_lock(&bs->rescue_lock); -- bio_list_add(&bs->rescue_list, bio); -- queue_work(bs->rescue_workqueue, &bs->rescue_work); -- spin_unlock(&bs->rescue_lock); - } - } - -diff --git a/drivers/md/raid10.c b/drivers/md/raid10.c -index 55b5e0e..4c4aab0 100644 ---- a/drivers/md/raid10.c -+++ b/drivers/md/raid10.c -@@ -941,7 +941,8 @@ static void wait_barrier(struct r10conf *conf) - !conf->barrier || - (atomic_read(&conf->nr_pending) && - current->bio_list && -- !bio_list_empty(current->bio_list)), -+ (!bio_list_empty(¤t->bio_list[0]) || -+ !bio_list_empty(¤t->bio_list[1]))), - conf->resync_lock); - conf->nr_waiting--; - if (!conf->nr_waiting) -diff --git a/drivers/mmc/host/sdhci-of-at91.c b/drivers/mmc/host/sdhci-of-at91.c -index 387ae1c..a8b430f 100644 ---- a/drivers/mmc/host/sdhci-of-at91.c -+++ b/drivers/mmc/host/sdhci-of-at91.c -@@ -29,6 +29,8 @@ - - #include "sdhci-pltfm.h" - -+#define SDMMC_MC1R 0x204 -+#define SDMMC_MC1R_DDR BIT(3) - #define SDMMC_CACR 0x230 - #define SDMMC_CACR_CAPWREN BIT(0) - #define SDMMC_CACR_KEY (0x46 << 8) -@@ -103,11 +105,18 @@ static void sdhci_at91_set_power(struct sdhci_host *host, unsigned char mode, - sdhci_set_power_noreg(host, mode, vdd); - } - -+void sdhci_at91_set_uhs_signaling(struct sdhci_host *host, unsigned int timing) -+{ -+ if (timing == MMC_TIMING_MMC_DDR52) -+ sdhci_writeb(host, SDMMC_MC1R_DDR, SDMMC_MC1R); -+ sdhci_set_uhs_signaling(host, timing); -+} -+ - static const struct sdhci_ops sdhci_at91_sama5d2_ops = { - .set_clock = sdhci_at91_set_clock, - .set_bus_width = sdhci_set_bus_width, - .reset = sdhci_reset, -- .set_uhs_signaling = sdhci_set_uhs_signaling, -+ .set_uhs_signaling = sdhci_at91_set_uhs_signaling, - .set_power = sdhci_at91_set_power, - }; - -diff --git a/drivers/mmc/host/sdhci.c b/drivers/mmc/host/sdhci.c -index a983ba0..7d275e7 100644 ---- a/drivers/mmc/host/sdhci.c -+++ b/drivers/mmc/host/sdhci.c -@@ -1823,6 +1823,9 @@ static void sdhci_enable_sdio_irq(struct mmc_host *mmc, int enable) - struct sdhci_host *host = mmc_priv(mmc); - unsigned long flags; - -+ if (enable) -+ pm_runtime_get_noresume(host->mmc->parent); -+ - spin_lock_irqsave(&host->lock, flags); - if (enable) - host->flags |= SDHCI_SDIO_IRQ_ENABLED; -@@ -1831,6 +1834,9 @@ static void sdhci_enable_sdio_irq(struct mmc_host *mmc, int enable) - - sdhci_enable_sdio_irq_nolock(host, enable); - spin_unlock_irqrestore(&host->lock, flags); -+ -+ if (!enable) -+ pm_runtime_put_noidle(host->mmc->parent); - } - - static int sdhci_start_signal_voltage_switch(struct mmc_host *mmc, -diff --git a/drivers/nvme/host/core.c b/drivers/nvme/host/core.c -index da10b48..bde769b 100644 ---- a/drivers/nvme/host/core.c -+++ b/drivers/nvme/host/core.c -@@ -2057,9 +2057,9 @@ void nvme_kill_queues(struct nvme_ctrl *ctrl) - * Revalidating a dead namespace sets capacity to 0. This will - * end buffered writers dirtying pages that can't be synced. - */ -- if (ns->disk && !test_and_set_bit(NVME_NS_DEAD, &ns->flags)) -- revalidate_disk(ns->disk); -- -+ if (!ns->disk || test_and_set_bit(NVME_NS_DEAD, &ns->flags)) -+ continue; -+ revalidate_disk(ns->disk); - blk_set_queue_dying(ns->queue); - blk_mq_abort_requeue_list(ns->queue); - blk_mq_start_stopped_hw_queues(ns->queue, true); -diff --git a/drivers/nvme/host/pci.c b/drivers/nvme/host/pci.c -index 5e52034..8a9c186 100644 ---- a/drivers/nvme/host/pci.c -+++ b/drivers/nvme/host/pci.c -@@ -1983,8 +1983,10 @@ static void nvme_remove(struct pci_dev *pdev) - - pci_set_drvdata(pdev, NULL); - -- if (!pci_device_is_present(pdev)) -+ if (!pci_device_is_present(pdev)) { - nvme_change_ctrl_state(&dev->ctrl, NVME_CTRL_DEAD); -+ nvme_dev_disable(dev, false); -+ } - - flush_work(&dev->reset_work); - nvme_uninit_ctrl(&dev->ctrl); -diff --git a/drivers/pci/host/pcie-iproc-bcma.c b/drivers/pci/host/pcie-iproc-bcma.c -index 8ce0890..46ca8ed 100644 ---- a/drivers/pci/host/pcie-iproc-bcma.c -+++ b/drivers/pci/host/pcie-iproc-bcma.c -@@ -44,8 +44,7 @@ static int iproc_pcie_bcma_probe(struct bcma_device *bdev) - { - struct device *dev = &bdev->dev; - struct iproc_pcie *pcie; -- LIST_HEAD(res); -- struct resource res_mem; -+ LIST_HEAD(resources); - int ret; - - pcie = devm_kzalloc(dev, sizeof(*pcie), GFP_KERNEL); -@@ -62,22 +61,23 @@ static int iproc_pcie_bcma_probe(struct bcma_device *bdev) - - pcie->base_addr = bdev->addr; - -- res_mem.start = bdev->addr_s[0]; -- res_mem.end = bdev->addr_s[0] + SZ_128M - 1; -- res_mem.name = "PCIe MEM space"; -- res_mem.flags = IORESOURCE_MEM; -- pci_add_resource(&res, &res_mem); -+ pcie->mem.start = bdev->addr_s[0]; -+ pcie->mem.end = bdev->addr_s[0] + SZ_128M - 1; -+ pcie->mem.name = "PCIe MEM space"; -+ pcie->mem.flags = IORESOURCE_MEM; -+ pci_add_resource(&resources, &pcie->mem); - - pcie->map_irq = iproc_pcie_bcma_map_irq; - -- ret = iproc_pcie_setup(pcie, &res); -- if (ret) -+ ret = iproc_pcie_setup(pcie, &resources); -+ if (ret) { - dev_err(dev, "PCIe controller setup failed\n"); -- -- pci_free_resource_list(&res); -+ pci_free_resource_list(&resources); -+ return ret; -+ } - - bcma_set_drvdata(bdev, pcie); -- return ret; -+ return 0; - } - - static void iproc_pcie_bcma_remove(struct bcma_device *bdev) -diff --git a/drivers/pci/host/pcie-iproc-platform.c b/drivers/pci/host/pcie-iproc-platform.c -index a3de087..7dcaddc 100644 ---- a/drivers/pci/host/pcie-iproc-platform.c -+++ b/drivers/pci/host/pcie-iproc-platform.c -@@ -46,7 +46,7 @@ static int iproc_pcie_pltfm_probe(struct platform_device *pdev) - struct device_node *np = dev->of_node; - struct resource reg; - resource_size_t iobase = 0; -- LIST_HEAD(res); -+ LIST_HEAD(resources); - int ret; - - of_id = of_match_device(iproc_pcie_of_match_table, dev); -@@ -108,23 +108,24 @@ static int iproc_pcie_pltfm_probe(struct platform_device *pdev) - pcie->phy = NULL; - } - -- ret = of_pci_get_host_bridge_resources(np, 0, 0xff, &res, &iobase); -+ ret = of_pci_get_host_bridge_resources(np, 0, 0xff, &resources, -+ &iobase); - if (ret) { -- dev_err(dev, -- "unable to get PCI host bridge resources\n"); -+ dev_err(dev, "unable to get PCI host bridge resources\n"); - return ret; - } - - pcie->map_irq = of_irq_parse_and_map_pci; - -- ret = iproc_pcie_setup(pcie, &res); -- if (ret) -+ ret = iproc_pcie_setup(pcie, &resources); -+ if (ret) { - dev_err(dev, "PCIe controller setup failed\n"); -- -- pci_free_resource_list(&res); -+ pci_free_resource_list(&resources); -+ return ret; -+ } - - platform_set_drvdata(pdev, pcie); -- return ret; -+ return 0; - } - - static int iproc_pcie_pltfm_remove(struct platform_device *pdev) -diff --git a/drivers/pci/host/pcie-iproc.h b/drivers/pci/host/pcie-iproc.h -index e84d93c..fa42267 100644 ---- a/drivers/pci/host/pcie-iproc.h -+++ b/drivers/pci/host/pcie-iproc.h -@@ -68,6 +68,7 @@ struct iproc_pcie { - #ifdef CONFIG_ARM - struct pci_sys_data sysdata; - #endif -+ struct resource mem; - struct pci_bus *root_bus; - struct phy *phy; - int (*map_irq)(const struct pci_dev *, u8, u8); -diff --git a/drivers/scsi/device_handler/scsi_dh_alua.c b/drivers/scsi/device_handler/scsi_dh_alua.c -index 7bb2068..d314579 100644 ---- a/drivers/scsi/device_handler/scsi_dh_alua.c -+++ b/drivers/scsi/device_handler/scsi_dh_alua.c -@@ -113,7 +113,7 @@ struct alua_queue_data { - #define ALUA_POLICY_SWITCH_ALL 1 - - static void alua_rtpg_work(struct work_struct *work); --static void alua_rtpg_queue(struct alua_port_group *pg, -+static bool alua_rtpg_queue(struct alua_port_group *pg, - struct scsi_device *sdev, - struct alua_queue_data *qdata, bool force); - static void alua_check(struct scsi_device *sdev, bool force); -@@ -862,7 +862,13 @@ static void alua_rtpg_work(struct work_struct *work) - kref_put(&pg->kref, release_port_group); - } - --static void alua_rtpg_queue(struct alua_port_group *pg, -+/** -+ * alua_rtpg_queue() - cause RTPG to be submitted asynchronously -+ * -+ * Returns true if and only if alua_rtpg_work() will be called asynchronously. -+ * That function is responsible for calling @qdata->fn(). -+ */ -+static bool alua_rtpg_queue(struct alua_port_group *pg, - struct scsi_device *sdev, - struct alua_queue_data *qdata, bool force) - { -@@ -870,8 +876,8 @@ static void alua_rtpg_queue(struct alua_port_group *pg, - unsigned long flags; - struct workqueue_struct *alua_wq = kaluad_wq; - -- if (!pg) -- return; -+ if (!pg || scsi_device_get(sdev)) -+ return false; - - spin_lock_irqsave(&pg->lock, flags); - if (qdata) { -@@ -884,14 +890,12 @@ static void alua_rtpg_queue(struct alua_port_group *pg, - pg->flags |= ALUA_PG_RUN_RTPG; - kref_get(&pg->kref); - pg->rtpg_sdev = sdev; -- scsi_device_get(sdev); - start_queue = 1; - } else if (!(pg->flags & ALUA_PG_RUN_RTPG) && force) { - pg->flags |= ALUA_PG_RUN_RTPG; - /* Do not queue if the worker is already running */ - if (!(pg->flags & ALUA_PG_RUNNING)) { - kref_get(&pg->kref); -- sdev = NULL; - start_queue = 1; - } - } -@@ -900,13 +904,17 @@ static void alua_rtpg_queue(struct alua_port_group *pg, - alua_wq = kaluad_sync_wq; - spin_unlock_irqrestore(&pg->lock, flags); - -- if (start_queue && -- !queue_delayed_work(alua_wq, &pg->rtpg_work, -- msecs_to_jiffies(ALUA_RTPG_DELAY_MSECS))) { -- if (sdev) -- scsi_device_put(sdev); -- kref_put(&pg->kref, release_port_group); -+ if (start_queue) { -+ if (queue_delayed_work(alua_wq, &pg->rtpg_work, -+ msecs_to_jiffies(ALUA_RTPG_DELAY_MSECS))) -+ sdev = NULL; -+ else -+ kref_put(&pg->kref, release_port_group); - } -+ if (sdev) -+ scsi_device_put(sdev); -+ -+ return true; - } - - /* -@@ -1007,11 +1015,13 @@ static int alua_activate(struct scsi_device *sdev, - mutex_unlock(&h->init_mutex); - goto out; - } -- fn = NULL; - rcu_read_unlock(); - mutex_unlock(&h->init_mutex); - -- alua_rtpg_queue(pg, sdev, qdata, true); -+ if (alua_rtpg_queue(pg, sdev, qdata, true)) -+ fn = NULL; -+ else -+ err = SCSI_DH_DEV_OFFLINED; - kref_put(&pg->kref, release_port_group); - out: - if (fn) -diff --git a/drivers/scsi/libsas/sas_ata.c b/drivers/scsi/libsas/sas_ata.c -index 763f012..87f5e694 100644 ---- a/drivers/scsi/libsas/sas_ata.c -+++ b/drivers/scsi/libsas/sas_ata.c -@@ -221,7 +221,7 @@ static unsigned int sas_ata_qc_issue(struct ata_queued_cmd *qc) - task->num_scatter = qc->n_elem; - } else { - for_each_sg(qc->sg, sg, qc->n_elem, si) -- xfer += sg->length; -+ xfer += sg_dma_len(sg); - - task->total_xfer_len = xfer; - task->num_scatter = si; -diff --git a/drivers/scsi/qla2xxx/qla_attr.c b/drivers/scsi/qla2xxx/qla_attr.c -index fe7469c..ad33238 100644 ---- a/drivers/scsi/qla2xxx/qla_attr.c -+++ b/drivers/scsi/qla2xxx/qla_attr.c -@@ -2153,8 +2153,6 @@ qla24xx_vport_delete(struct fc_vport *fc_vport) - "Timer for the VP[%d] has stopped\n", vha->vp_idx); - } - -- BUG_ON(atomic_read(&vha->vref_count)); -- - qla2x00_free_fcports(vha); - - mutex_lock(&ha->vport_lock); -diff --git a/drivers/scsi/qla2xxx/qla_def.h b/drivers/scsi/qla2xxx/qla_def.h -index 73b12e4..8e63a7b 100644 ---- a/drivers/scsi/qla2xxx/qla_def.h -+++ b/drivers/scsi/qla2xxx/qla_def.h -@@ -3742,6 +3742,7 @@ typedef struct scsi_qla_host { - struct qla8044_reset_template reset_tmplt; - struct qla_tgt_counters tgt_counters; - uint16_t bbcr; -+ wait_queue_head_t vref_waitq; - } scsi_qla_host_t; - - struct qla27xx_image_status { -@@ -3780,6 +3781,7 @@ struct qla_tgt_vp_map { - mb(); \ - if (__vha->flags.delete_progress) { \ - atomic_dec(&__vha->vref_count); \ -+ wake_up(&__vha->vref_waitq); \ - __bail = 1; \ - } else { \ - __bail = 0; \ -@@ -3788,6 +3790,7 @@ struct qla_tgt_vp_map { - - #define QLA_VHA_MARK_NOT_BUSY(__vha) do { \ - atomic_dec(&__vha->vref_count); \ -+ wake_up(&__vha->vref_waitq); \ - } while (0) - - /* -diff --git a/drivers/scsi/qla2xxx/qla_init.c b/drivers/scsi/qla2xxx/qla_init.c -index 5b09296..8f12f6b 100644 ---- a/drivers/scsi/qla2xxx/qla_init.c -+++ b/drivers/scsi/qla2xxx/qla_init.c -@@ -4356,6 +4356,7 @@ qla2x00_update_fcports(scsi_qla_host_t *base_vha) - } - } - atomic_dec(&vha->vref_count); -+ wake_up(&vha->vref_waitq); - } - spin_unlock_irqrestore(&ha->vport_slock, flags); - } -diff --git a/drivers/scsi/qla2xxx/qla_mid.c b/drivers/scsi/qla2xxx/qla_mid.c -index cf7ba52..3dfb54a 100644 ---- a/drivers/scsi/qla2xxx/qla_mid.c -+++ b/drivers/scsi/qla2xxx/qla_mid.c -@@ -74,13 +74,14 @@ qla24xx_deallocate_vp_id(scsi_qla_host_t *vha) - * ensures no active vp_list traversal while the vport is removed - * from the queue) - */ -- spin_lock_irqsave(&ha->vport_slock, flags); -- while (atomic_read(&vha->vref_count)) { -- spin_unlock_irqrestore(&ha->vport_slock, flags); -- -- msleep(500); -+ wait_event_timeout(vha->vref_waitq, atomic_read(&vha->vref_count), -+ 10*HZ); - -- spin_lock_irqsave(&ha->vport_slock, flags); -+ spin_lock_irqsave(&ha->vport_slock, flags); -+ if (atomic_read(&vha->vref_count)) { -+ ql_dbg(ql_dbg_vport, vha, 0xfffa, -+ "vha->vref_count=%u timeout\n", vha->vref_count.counter); -+ vha->vref_count = (atomic_t)ATOMIC_INIT(0); - } - list_del(&vha->list); - qlt_update_vp_map(vha, RESET_VP_IDX); -@@ -269,6 +270,7 @@ qla2x00_alert_all_vps(struct rsp_que *rsp, uint16_t *mb) - - spin_lock_irqsave(&ha->vport_slock, flags); - atomic_dec(&vha->vref_count); -+ wake_up(&vha->vref_waitq); - } - i++; - } -diff --git a/drivers/scsi/qla2xxx/qla_os.c b/drivers/scsi/qla2xxx/qla_os.c -index bea819e..4f361d8 100644 ---- a/drivers/scsi/qla2xxx/qla_os.c -+++ b/drivers/scsi/qla2xxx/qla_os.c -@@ -4045,6 +4045,7 @@ struct scsi_qla_host *qla2x00_create_host(struct scsi_host_template *sht, - - spin_lock_init(&vha->work_lock); - spin_lock_init(&vha->cmd_list_lock); -+ init_waitqueue_head(&vha->vref_waitq); - - sprintf(vha->host_str, "%s_%ld", QLA2XXX_DRIVER_NAME, vha->host_no); - ql_dbg(ql_dbg_init, vha, 0x0041, -diff --git a/drivers/scsi/sg.c b/drivers/scsi/sg.c -index 121de0a..f753df2 100644 ---- a/drivers/scsi/sg.c -+++ b/drivers/scsi/sg.c -@@ -998,6 +998,8 @@ sg_ioctl(struct file *filp, unsigned int cmd_in, unsigned long arg) - result = get_user(val, ip); - if (result) - return result; -+ if (val > SG_MAX_CDB_SIZE) -+ return -ENOMEM; - sfp->next_cmd_len = (val > 0) ? val : 0; - return 0; - case SG_GET_VERSION_NUM: -diff --git a/drivers/tty/serial/atmel_serial.c b/drivers/tty/serial/atmel_serial.c -index fabbe76..4d079cd 100644 ---- a/drivers/tty/serial/atmel_serial.c -+++ b/drivers/tty/serial/atmel_serial.c -@@ -1938,6 +1938,11 @@ static void atmel_flush_buffer(struct uart_port *port) - atmel_uart_writel(port, ATMEL_PDC_TCR, 0); - atmel_port->pdc_tx.ofs = 0; - } -+ /* -+ * in uart_flush_buffer(), the xmit circular buffer has just -+ * been cleared, so we have to reset tx_len accordingly. -+ */ -+ atmel_port->tx_len = 0; - } - - /* -@@ -2471,6 +2476,9 @@ static void atmel_console_write(struct console *co, const char *s, u_int count) - pdc_tx = atmel_uart_readl(port, ATMEL_PDC_PTSR) & ATMEL_PDC_TXTEN; - atmel_uart_writel(port, ATMEL_PDC_PTCR, ATMEL_PDC_TXTDIS); - -+ /* Make sure that tx path is actually able to send characters */ -+ atmel_uart_writel(port, ATMEL_US_CR, ATMEL_US_TXEN); -+ - uart_console_write(port, s, count, atmel_console_putchar); - - /* -diff --git a/drivers/tty/serial/mxs-auart.c b/drivers/tty/serial/mxs-auart.c -index 770454e..07390f8 100644 ---- a/drivers/tty/serial/mxs-auart.c -+++ b/drivers/tty/serial/mxs-auart.c -@@ -1085,7 +1085,7 @@ static void mxs_auart_settermios(struct uart_port *u, - AUART_LINECTRL_BAUD_DIV_MAX); - baud_max = u->uartclk * 32 / AUART_LINECTRL_BAUD_DIV_MIN; - baud = uart_get_baud_rate(u, termios, old, baud_min, baud_max); -- div = u->uartclk * 32 / baud; -+ div = DIV_ROUND_CLOSEST(u->uartclk * 32, baud); - } - - ctrl |= AUART_LINECTRL_BAUD_DIVFRAC(div & 0x3F); -diff --git a/drivers/usb/core/hcd.c b/drivers/usb/core/hcd.c -index 479e223..f029aad 100644 ---- a/drivers/usb/core/hcd.c -+++ b/drivers/usb/core/hcd.c -@@ -520,8 +520,10 @@ static int rh_call_control (struct usb_hcd *hcd, struct urb *urb) - */ - tbuf_size = max_t(u16, sizeof(struct usb_hub_descriptor), wLength); - tbuf = kzalloc(tbuf_size, GFP_KERNEL); -- if (!tbuf) -- return -ENOMEM; -+ if (!tbuf) { -+ status = -ENOMEM; -+ goto err_alloc; -+ } - - bufp = tbuf; - -@@ -734,6 +736,7 @@ static int rh_call_control (struct usb_hcd *hcd, struct urb *urb) - } - - kfree(tbuf); -+ err_alloc: - - /* any errors get returned through the urb completion */ - spin_lock_irq(&hcd_root_hub_lock); -diff --git a/fs/nfs/nfs4proc.c b/fs/nfs/nfs4proc.c -index 1536aeb..4e894d3 100644 ---- a/fs/nfs/nfs4proc.c -+++ b/fs/nfs/nfs4proc.c -@@ -2532,17 +2532,14 @@ static void nfs41_check_delegation_stateid(struct nfs4_state *state) - } - - nfs4_stateid_copy(&stateid, &delegation->stateid); -- if (test_bit(NFS_DELEGATION_REVOKED, &delegation->flags)) { -+ if (test_bit(NFS_DELEGATION_REVOKED, &delegation->flags) || -+ !test_and_clear_bit(NFS_DELEGATION_TEST_EXPIRED, -+ &delegation->flags)) { - rcu_read_unlock(); - nfs_finish_clear_delegation_stateid(state, &stateid); - return; - } - -- if (!test_and_clear_bit(NFS_DELEGATION_TEST_EXPIRED, &delegation->flags)) { -- rcu_read_unlock(); -- return; -- } -- - cred = get_rpccred(delegation->cred); - rcu_read_unlock(); - status = nfs41_test_and_free_expired_stateid(server, &stateid, cred); -diff --git a/fs/nfsd/nfsproc.c b/fs/nfsd/nfsproc.c -index 010aff5..536009e 100644 ---- a/fs/nfsd/nfsproc.c -+++ b/fs/nfsd/nfsproc.c -@@ -790,6 +790,7 @@ nfserrno (int errno) - { nfserr_serverfault, -ESERVERFAULT }, - { nfserr_serverfault, -ENFILE }, - { nfserr_io, -EUCLEAN }, -+ { nfserr_perm, -ENOKEY }, - }; - int i; - -diff --git a/fs/xfs/libxfs/xfs_ag_resv.c b/fs/xfs/libxfs/xfs_ag_resv.c -index d346d42..33db69b 100644 ---- a/fs/xfs/libxfs/xfs_ag_resv.c -+++ b/fs/xfs/libxfs/xfs_ag_resv.c -@@ -39,6 +39,7 @@ - #include "xfs_rmap_btree.h" - #include "xfs_btree.h" - #include "xfs_refcount_btree.h" -+#include "xfs_ialloc_btree.h" - - /* - * Per-AG Block Reservations -@@ -200,22 +201,30 @@ __xfs_ag_resv_init( - struct xfs_mount *mp = pag->pag_mount; - struct xfs_ag_resv *resv; - int error; -+ xfs_extlen_t reserved; - -- resv = xfs_perag_resv(pag, type); - if (used > ask) - ask = used; -- resv->ar_asked = ask; -- resv->ar_reserved = resv->ar_orig_reserved = ask - used; -- mp->m_ag_max_usable -= ask; -+ reserved = ask - used; - -- trace_xfs_ag_resv_init(pag, type, ask); -- -- error = xfs_mod_fdblocks(mp, -(int64_t)resv->ar_reserved, true); -- if (error) -+ error = xfs_mod_fdblocks(mp, -(int64_t)reserved, true); -+ if (error) { - trace_xfs_ag_resv_init_error(pag->pag_mount, pag->pag_agno, - error, _RET_IP_); -+ xfs_warn(mp, -+"Per-AG reservation for AG %u failed. Filesystem may run out of space.", -+ pag->pag_agno); -+ return error; -+ } - -- return error; -+ mp->m_ag_max_usable -= ask; -+ -+ resv = xfs_perag_resv(pag, type); -+ resv->ar_asked = ask; -+ resv->ar_reserved = resv->ar_orig_reserved = reserved; -+ -+ trace_xfs_ag_resv_init(pag, type, ask); -+ return 0; - } - - /* Create a per-AG block reservation. */ -@@ -223,6 +232,8 @@ int - xfs_ag_resv_init( - struct xfs_perag *pag) - { -+ struct xfs_mount *mp = pag->pag_mount; -+ xfs_agnumber_t agno = pag->pag_agno; - xfs_extlen_t ask; - xfs_extlen_t used; - int error = 0; -@@ -231,23 +242,45 @@ xfs_ag_resv_init( - if (pag->pag_meta_resv.ar_asked == 0) { - ask = used = 0; - -- error = xfs_refcountbt_calc_reserves(pag->pag_mount, -- pag->pag_agno, &ask, &used); -+ error = xfs_refcountbt_calc_reserves(mp, agno, &ask, &used); - if (error) - goto out; - -- error = __xfs_ag_resv_init(pag, XFS_AG_RESV_METADATA, -- ask, used); -+ error = xfs_finobt_calc_reserves(mp, agno, &ask, &used); - if (error) - goto out; -+ -+ error = __xfs_ag_resv_init(pag, XFS_AG_RESV_METADATA, -+ ask, used); -+ if (error) { -+ /* -+ * Because we didn't have per-AG reservations when the -+ * finobt feature was added we might not be able to -+ * reserve all needed blocks. Warn and fall back to the -+ * old and potentially buggy code in that case, but -+ * ensure we do have the reservation for the refcountbt. -+ */ -+ ask = used = 0; -+ -+ mp->m_inotbt_nores = true; -+ -+ error = xfs_refcountbt_calc_reserves(mp, agno, &ask, -+ &used); -+ if (error) -+ goto out; -+ -+ error = __xfs_ag_resv_init(pag, XFS_AG_RESV_METADATA, -+ ask, used); -+ if (error) -+ goto out; -+ } - } - - /* Create the AGFL metadata reservation */ - if (pag->pag_agfl_resv.ar_asked == 0) { - ask = used = 0; - -- error = xfs_rmapbt_calc_reserves(pag->pag_mount, pag->pag_agno, -- &ask, &used); -+ error = xfs_rmapbt_calc_reserves(mp, agno, &ask, &used); - if (error) - goto out; - -@@ -256,9 +289,16 @@ xfs_ag_resv_init( - goto out; - } - -+#ifdef DEBUG -+ /* need to read in the AGF for the ASSERT below to work */ -+ error = xfs_alloc_pagf_init(pag->pag_mount, NULL, pag->pag_agno, 0); -+ if (error) -+ return error; -+ - ASSERT(xfs_perag_resv(pag, XFS_AG_RESV_METADATA)->ar_reserved + - xfs_perag_resv(pag, XFS_AG_RESV_AGFL)->ar_reserved <= - pag->pagf_freeblks + pag->pagf_flcount); -+#endif - out: - return error; - } -diff --git a/fs/xfs/libxfs/xfs_bmap.c b/fs/xfs/libxfs/xfs_bmap.c -index f52fd63..5a508b0 100644 ---- a/fs/xfs/libxfs/xfs_bmap.c -+++ b/fs/xfs/libxfs/xfs_bmap.c -@@ -769,8 +769,8 @@ xfs_bmap_extents_to_btree( - args.type = XFS_ALLOCTYPE_START_BNO; - args.fsbno = XFS_INO_TO_FSB(mp, ip->i_ino); - } else if (dfops->dop_low) { --try_another_ag: - args.type = XFS_ALLOCTYPE_START_BNO; -+try_another_ag: - args.fsbno = *firstblock; - } else { - args.type = XFS_ALLOCTYPE_NEAR_BNO; -@@ -796,17 +796,19 @@ xfs_bmap_extents_to_btree( - if (xfs_sb_version_hasreflink(&cur->bc_mp->m_sb) && - args.fsbno == NULLFSBLOCK && - args.type == XFS_ALLOCTYPE_NEAR_BNO) { -- dfops->dop_low = true; -+ args.type = XFS_ALLOCTYPE_FIRST_AG; - goto try_another_ag; - } -+ if (WARN_ON_ONCE(args.fsbno == NULLFSBLOCK)) { -+ xfs_iroot_realloc(ip, -1, whichfork); -+ xfs_btree_del_cursor(cur, XFS_BTREE_ERROR); -+ return -ENOSPC; -+ } - /* - * Allocation can't fail, the space was reserved. - */ -- ASSERT(args.fsbno != NULLFSBLOCK); - ASSERT(*firstblock == NULLFSBLOCK || -- args.agno == XFS_FSB_TO_AGNO(mp, *firstblock) || -- (dfops->dop_low && -- args.agno > XFS_FSB_TO_AGNO(mp, *firstblock))); -+ args.agno >= XFS_FSB_TO_AGNO(mp, *firstblock)); - *firstblock = cur->bc_private.b.firstblock = args.fsbno; - cur->bc_private.b.allocated++; - ip->i_d.di_nblocks++; -@@ -1278,7 +1280,6 @@ xfs_bmap_read_extents( - /* REFERENCED */ - xfs_extnum_t room; /* number of entries there's room for */ - -- bno = NULLFSBLOCK; - mp = ip->i_mount; - ifp = XFS_IFORK_PTR(ip, whichfork); - exntf = (whichfork != XFS_DATA_FORK) ? XFS_EXTFMT_NOSTATE : -@@ -1291,9 +1292,7 @@ xfs_bmap_read_extents( - ASSERT(level > 0); - pp = XFS_BMAP_BROOT_PTR_ADDR(mp, block, 1, ifp->if_broot_bytes); - bno = be64_to_cpu(*pp); -- ASSERT(bno != NULLFSBLOCK); -- ASSERT(XFS_FSB_TO_AGNO(mp, bno) < mp->m_sb.sb_agcount); -- ASSERT(XFS_FSB_TO_AGBNO(mp, bno) < mp->m_sb.sb_agblocks); -+ - /* - * Go down the tree until leaf level is reached, following the first - * pointer (leftmost) at each level. -@@ -1955,6 +1954,7 @@ xfs_bmap_add_extent_delay_real( - */ - trace_xfs_bmap_pre_update(bma->ip, bma->idx, state, _THIS_IP_); - xfs_bmbt_set_startblock(ep, new->br_startblock); -+ xfs_bmbt_set_state(ep, new->br_state); - trace_xfs_bmap_post_update(bma->ip, bma->idx, state, _THIS_IP_); - - (*nextents)++; -@@ -2293,6 +2293,7 @@ STATIC int /* error */ - xfs_bmap_add_extent_unwritten_real( - struct xfs_trans *tp, - xfs_inode_t *ip, /* incore inode pointer */ -+ int whichfork, - xfs_extnum_t *idx, /* extent number to update/insert */ - xfs_btree_cur_t **curp, /* if *curp is null, not a btree */ - xfs_bmbt_irec_t *new, /* new data to add to file extents */ -@@ -2312,12 +2313,14 @@ xfs_bmap_add_extent_unwritten_real( - /* left is 0, right is 1, prev is 2 */ - int rval=0; /* return value (logging flags) */ - int state = 0;/* state bits, accessed thru macros */ -- struct xfs_mount *mp = tp->t_mountp; -+ struct xfs_mount *mp = ip->i_mount; - - *logflagsp = 0; - - cur = *curp; -- ifp = XFS_IFORK_PTR(ip, XFS_DATA_FORK); -+ ifp = XFS_IFORK_PTR(ip, whichfork); -+ if (whichfork == XFS_COW_FORK) -+ state |= BMAP_COWFORK; - - ASSERT(*idx >= 0); - ASSERT(*idx <= xfs_iext_count(ifp)); -@@ -2376,7 +2379,7 @@ xfs_bmap_add_extent_unwritten_real( - * Don't set contiguous if the combined extent would be too large. - * Also check for all-three-contiguous being too large. - */ -- if (*idx < xfs_iext_count(&ip->i_df) - 1) { -+ if (*idx < xfs_iext_count(ifp) - 1) { - state |= BMAP_RIGHT_VALID; - xfs_bmbt_get_all(xfs_iext_get_ext(ifp, *idx + 1), &RIGHT); - if (isnullstartblock(RIGHT.br_startblock)) -@@ -2416,7 +2419,8 @@ xfs_bmap_add_extent_unwritten_real( - trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_); - - xfs_iext_remove(ip, *idx + 1, 2, state); -- ip->i_d.di_nextents -= 2; -+ XFS_IFORK_NEXT_SET(ip, whichfork, -+ XFS_IFORK_NEXTENTS(ip, whichfork) - 2); - if (cur == NULL) - rval = XFS_ILOG_CORE | XFS_ILOG_DEXT; - else { -@@ -2459,7 +2463,8 @@ xfs_bmap_add_extent_unwritten_real( - trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_); - - xfs_iext_remove(ip, *idx + 1, 1, state); -- ip->i_d.di_nextents--; -+ XFS_IFORK_NEXT_SET(ip, whichfork, -+ XFS_IFORK_NEXTENTS(ip, whichfork) - 1); - if (cur == NULL) - rval = XFS_ILOG_CORE | XFS_ILOG_DEXT; - else { -@@ -2494,7 +2499,8 @@ xfs_bmap_add_extent_unwritten_real( - xfs_bmbt_set_state(ep, newext); - trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_); - xfs_iext_remove(ip, *idx + 1, 1, state); -- ip->i_d.di_nextents--; -+ XFS_IFORK_NEXT_SET(ip, whichfork, -+ XFS_IFORK_NEXTENTS(ip, whichfork) - 1); - if (cur == NULL) - rval = XFS_ILOG_CORE | XFS_ILOG_DEXT; - else { -@@ -2606,7 +2612,8 @@ xfs_bmap_add_extent_unwritten_real( - trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_); - - xfs_iext_insert(ip, *idx, 1, new, state); -- ip->i_d.di_nextents++; -+ XFS_IFORK_NEXT_SET(ip, whichfork, -+ XFS_IFORK_NEXTENTS(ip, whichfork) + 1); - if (cur == NULL) - rval = XFS_ILOG_CORE | XFS_ILOG_DEXT; - else { -@@ -2684,7 +2691,8 @@ xfs_bmap_add_extent_unwritten_real( - ++*idx; - xfs_iext_insert(ip, *idx, 1, new, state); - -- ip->i_d.di_nextents++; -+ XFS_IFORK_NEXT_SET(ip, whichfork, -+ XFS_IFORK_NEXTENTS(ip, whichfork) + 1); - if (cur == NULL) - rval = XFS_ILOG_CORE | XFS_ILOG_DEXT; - else { -@@ -2732,7 +2740,8 @@ xfs_bmap_add_extent_unwritten_real( - ++*idx; - xfs_iext_insert(ip, *idx, 2, &r[0], state); - -- ip->i_d.di_nextents += 2; -+ XFS_IFORK_NEXT_SET(ip, whichfork, -+ XFS_IFORK_NEXTENTS(ip, whichfork) + 2); - if (cur == NULL) - rval = XFS_ILOG_CORE | XFS_ILOG_DEXT; - else { -@@ -2786,17 +2795,17 @@ xfs_bmap_add_extent_unwritten_real( - } - - /* update reverse mappings */ -- error = xfs_rmap_convert_extent(mp, dfops, ip, XFS_DATA_FORK, new); -+ error = xfs_rmap_convert_extent(mp, dfops, ip, whichfork, new); - if (error) - goto done; - - /* convert to a btree if necessary */ -- if (xfs_bmap_needs_btree(ip, XFS_DATA_FORK)) { -+ if (xfs_bmap_needs_btree(ip, whichfork)) { - int tmp_logflags; /* partial log flag return val */ - - ASSERT(cur == NULL); - error = xfs_bmap_extents_to_btree(tp, ip, first, dfops, &cur, -- 0, &tmp_logflags, XFS_DATA_FORK); -+ 0, &tmp_logflags, whichfork); - *logflagsp |= tmp_logflags; - if (error) - goto done; -@@ -2808,7 +2817,7 @@ xfs_bmap_add_extent_unwritten_real( - *curp = cur; - } - -- xfs_bmap_check_leaf_extents(*curp, ip, XFS_DATA_FORK); -+ xfs_bmap_check_leaf_extents(*curp, ip, whichfork); - done: - *logflagsp |= rval; - return error; -@@ -2900,7 +2909,8 @@ xfs_bmap_add_extent_hole_delay( - oldlen = startblockval(left.br_startblock) + - startblockval(new->br_startblock) + - startblockval(right.br_startblock); -- newlen = xfs_bmap_worst_indlen(ip, temp); -+ newlen = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(ip, temp), -+ oldlen); - xfs_bmbt_set_startblock(xfs_iext_get_ext(ifp, *idx), - nullstartblock((int)newlen)); - trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_); -@@ -2921,7 +2931,8 @@ xfs_bmap_add_extent_hole_delay( - xfs_bmbt_set_blockcount(xfs_iext_get_ext(ifp, *idx), temp); - oldlen = startblockval(left.br_startblock) + - startblockval(new->br_startblock); -- newlen = xfs_bmap_worst_indlen(ip, temp); -+ newlen = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(ip, temp), -+ oldlen); - xfs_bmbt_set_startblock(xfs_iext_get_ext(ifp, *idx), - nullstartblock((int)newlen)); - trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_); -@@ -2937,7 +2948,8 @@ xfs_bmap_add_extent_hole_delay( - temp = new->br_blockcount + right.br_blockcount; - oldlen = startblockval(new->br_startblock) + - startblockval(right.br_startblock); -- newlen = xfs_bmap_worst_indlen(ip, temp); -+ newlen = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(ip, temp), -+ oldlen); - xfs_bmbt_set_allf(xfs_iext_get_ext(ifp, *idx), - new->br_startoff, - nullstartblock((int)newlen), temp, right.br_state); -@@ -3913,17 +3925,13 @@ xfs_bmap_btalloc( - * the first block that was allocated. - */ - ASSERT(*ap->firstblock == NULLFSBLOCK || -- XFS_FSB_TO_AGNO(mp, *ap->firstblock) == -- XFS_FSB_TO_AGNO(mp, args.fsbno) || -- (ap->dfops->dop_low && -- XFS_FSB_TO_AGNO(mp, *ap->firstblock) < -- XFS_FSB_TO_AGNO(mp, args.fsbno))); -+ XFS_FSB_TO_AGNO(mp, *ap->firstblock) <= -+ XFS_FSB_TO_AGNO(mp, args.fsbno)); - - ap->blkno = args.fsbno; - if (*ap->firstblock == NULLFSBLOCK) - *ap->firstblock = args.fsbno; -- ASSERT(nullfb || fb_agno == args.agno || -- (ap->dfops->dop_low && fb_agno < args.agno)); -+ ASSERT(nullfb || fb_agno <= args.agno); - ap->length = args.len; - if (!(ap->flags & XFS_BMAPI_COWFORK)) - ap->ip->i_d.di_nblocks += args.len; -@@ -4249,6 +4257,19 @@ xfs_bmapi_read( - return 0; - } - -+/* -+ * Add a delayed allocation extent to an inode. Blocks are reserved from the -+ * global pool and the extent inserted into the inode in-core extent tree. -+ * -+ * On entry, got refers to the first extent beyond the offset of the extent to -+ * allocate or eof is specified if no such extent exists. On return, got refers -+ * to the extent record that was inserted to the inode fork. -+ * -+ * Note that the allocated extent may have been merged with contiguous extents -+ * during insertion into the inode fork. Thus, got does not reflect the current -+ * state of the inode fork on return. If necessary, the caller can use lastx to -+ * look up the updated record in the inode fork. -+ */ - int - xfs_bmapi_reserve_delalloc( - struct xfs_inode *ip, -@@ -4335,13 +4356,8 @@ xfs_bmapi_reserve_delalloc( - got->br_startblock = nullstartblock(indlen); - got->br_blockcount = alen; - got->br_state = XFS_EXT_NORM; -- xfs_bmap_add_extent_hole_delay(ip, whichfork, lastx, got); - -- /* -- * Update our extent pointer, given that xfs_bmap_add_extent_hole_delay -- * might have merged it into one of the neighbouring ones. -- */ -- xfs_bmbt_get_all(xfs_iext_get_ext(ifp, *lastx), got); -+ xfs_bmap_add_extent_hole_delay(ip, whichfork, lastx, got); - - /* - * Tag the inode if blocks were preallocated. Note that COW fork -@@ -4353,10 +4369,6 @@ xfs_bmapi_reserve_delalloc( - if (whichfork == XFS_COW_FORK && (prealloc || aoff < off || alen > len)) - xfs_inode_set_cowblocks_tag(ip); - -- ASSERT(got->br_startoff <= aoff); -- ASSERT(got->br_startoff + got->br_blockcount >= aoff + alen); -- ASSERT(isnullstartblock(got->br_startblock)); -- ASSERT(got->br_state == XFS_EXT_NORM); - return 0; - - out_unreserve_blocks: -@@ -4461,10 +4473,16 @@ xfs_bmapi_allocate( - bma->got.br_state = XFS_EXT_NORM; - - /* -- * A wasdelay extent has been initialized, so shouldn't be flagged -- * as unwritten. -+ * In the data fork, a wasdelay extent has been initialized, so -+ * shouldn't be flagged as unwritten. -+ * -+ * For the cow fork, however, we convert delalloc reservations -+ * (extents allocated for speculative preallocation) to -+ * allocated unwritten extents, and only convert the unwritten -+ * extents to real extents when we're about to write the data. - */ -- if (!bma->wasdel && (bma->flags & XFS_BMAPI_PREALLOC) && -+ if ((!bma->wasdel || (bma->flags & XFS_BMAPI_COWFORK)) && -+ (bma->flags & XFS_BMAPI_PREALLOC) && - xfs_sb_version_hasextflgbit(&mp->m_sb)) - bma->got.br_state = XFS_EXT_UNWRITTEN; - -@@ -4515,8 +4533,6 @@ xfs_bmapi_convert_unwritten( - (XFS_BMAPI_PREALLOC | XFS_BMAPI_CONVERT)) - return 0; - -- ASSERT(whichfork != XFS_COW_FORK); -- - /* - * Modify (by adding) the state flag, if writing. - */ -@@ -4541,8 +4557,8 @@ xfs_bmapi_convert_unwritten( - return error; - } - -- error = xfs_bmap_add_extent_unwritten_real(bma->tp, bma->ip, &bma->idx, -- &bma->cur, mval, bma->firstblock, bma->dfops, -+ error = xfs_bmap_add_extent_unwritten_real(bma->tp, bma->ip, whichfork, -+ &bma->idx, &bma->cur, mval, bma->firstblock, bma->dfops, - &tmp_logflags); - /* - * Log the inode core unconditionally in the unwritten extent conversion -@@ -4551,8 +4567,12 @@ xfs_bmapi_convert_unwritten( - * in the transaction for the sake of fsync(), even if nothing has - * changed, because fsync() will not force the log for this transaction - * unless it sees the inode pinned. -+ * -+ * Note: If we're only converting cow fork extents, there aren't -+ * any on-disk updates to make, so we don't need to log anything. - */ -- bma->logflags |= tmp_logflags | XFS_ILOG_CORE; -+ if (whichfork != XFS_COW_FORK) -+ bma->logflags |= tmp_logflags | XFS_ILOG_CORE; - if (error) - return error; - -@@ -4626,15 +4646,15 @@ xfs_bmapi_write( - ASSERT(*nmap >= 1); - ASSERT(*nmap <= XFS_BMAP_MAX_NMAP); - ASSERT(!(flags & XFS_BMAPI_IGSTATE)); -- ASSERT(tp != NULL); -+ ASSERT(tp != NULL || -+ (flags & (XFS_BMAPI_CONVERT | XFS_BMAPI_COWFORK)) == -+ (XFS_BMAPI_CONVERT | XFS_BMAPI_COWFORK)); - ASSERT(len > 0); - ASSERT(XFS_IFORK_FORMAT(ip, whichfork) != XFS_DINODE_FMT_LOCAL); - ASSERT(xfs_isilocked(ip, XFS_ILOCK_EXCL)); - ASSERT(!(flags & XFS_BMAPI_REMAP) || whichfork == XFS_DATA_FORK); - ASSERT(!(flags & XFS_BMAPI_PREALLOC) || !(flags & XFS_BMAPI_REMAP)); - ASSERT(!(flags & XFS_BMAPI_CONVERT) || !(flags & XFS_BMAPI_REMAP)); -- ASSERT(!(flags & XFS_BMAPI_PREALLOC) || whichfork != XFS_COW_FORK); -- ASSERT(!(flags & XFS_BMAPI_CONVERT) || whichfork != XFS_COW_FORK); - - /* zeroing is for currently only for data extents, not metadata */ - ASSERT((flags & (XFS_BMAPI_METADATA | XFS_BMAPI_ZERO)) != -@@ -4840,13 +4860,9 @@ xfs_bmapi_write( - if (bma.cur) { - if (!error) { - ASSERT(*firstblock == NULLFSBLOCK || -- XFS_FSB_TO_AGNO(mp, *firstblock) == -+ XFS_FSB_TO_AGNO(mp, *firstblock) <= - XFS_FSB_TO_AGNO(mp, -- bma.cur->bc_private.b.firstblock) || -- (dfops->dop_low && -- XFS_FSB_TO_AGNO(mp, *firstblock) < -- XFS_FSB_TO_AGNO(mp, -- bma.cur->bc_private.b.firstblock))); -+ bma.cur->bc_private.b.firstblock)); - *firstblock = bma.cur->bc_private.b.firstblock; - } - xfs_btree_del_cursor(bma.cur, -@@ -4881,34 +4897,59 @@ xfs_bmap_split_indlen( - xfs_filblks_t len2 = *indlen2; - xfs_filblks_t nres = len1 + len2; /* new total res. */ - xfs_filblks_t stolen = 0; -+ xfs_filblks_t resfactor; - - /* - * Steal as many blocks as we can to try and satisfy the worst case - * indlen for both new extents. - */ -- while (nres > ores && avail) { -- nres--; -- avail--; -- stolen++; -- } -+ if (ores < nres && avail) -+ stolen = XFS_FILBLKS_MIN(nres - ores, avail); -+ ores += stolen; -+ -+ /* nothing else to do if we've satisfied the new reservation */ -+ if (ores >= nres) -+ return stolen; -+ -+ /* -+ * We can't meet the total required reservation for the two extents. -+ * Calculate the percent of the overall shortage between both extents -+ * and apply this percentage to each of the requested indlen values. -+ * This distributes the shortage fairly and reduces the chances that one -+ * of the two extents is left with nothing when extents are repeatedly -+ * split. -+ */ -+ resfactor = (ores * 100); -+ do_div(resfactor, nres); -+ len1 *= resfactor; -+ do_div(len1, 100); -+ len2 *= resfactor; -+ do_div(len2, 100); -+ ASSERT(len1 + len2 <= ores); -+ ASSERT(len1 < *indlen1 && len2 < *indlen2); - - /* -- * The only blocks available are those reserved for the original -- * extent and what we can steal from the extent being removed. -- * If this still isn't enough to satisfy the combined -- * requirements for the two new extents, skim blocks off of each -- * of the new reservations until they match what is available. -+ * Hand out the remainder to each extent. If one of the two reservations -+ * is zero, we want to make sure that one gets a block first. The loop -+ * below starts with len1, so hand len2 a block right off the bat if it -+ * is zero. - */ -- while (nres > ores) { -- if (len1) { -- len1--; -- nres--; -+ ores -= (len1 + len2); -+ ASSERT((*indlen1 - len1) + (*indlen2 - len2) >= ores); -+ if (ores && !len2 && *indlen2) { -+ len2++; -+ ores--; -+ } -+ while (ores) { -+ if (len1 < *indlen1) { -+ len1++; -+ ores--; - } -- if (nres == ores) -+ if (!ores) - break; -- if (len2) { -- len2--; -- nres--; -+ if (len2 < *indlen2) { -+ len2++; -+ ores--; - } - } - -@@ -5656,8 +5697,8 @@ __xfs_bunmapi( - } - del.br_state = XFS_EXT_UNWRITTEN; - error = xfs_bmap_add_extent_unwritten_real(tp, ip, -- &lastx, &cur, &del, firstblock, dfops, -- &logflags); -+ whichfork, &lastx, &cur, &del, -+ firstblock, dfops, &logflags); - if (error) - goto error0; - goto nodelete; -@@ -5714,8 +5755,9 @@ __xfs_bunmapi( - prev.br_state = XFS_EXT_UNWRITTEN; - lastx--; - error = xfs_bmap_add_extent_unwritten_real(tp, -- ip, &lastx, &cur, &prev, -- firstblock, dfops, &logflags); -+ ip, whichfork, &lastx, &cur, -+ &prev, firstblock, dfops, -+ &logflags); - if (error) - goto error0; - goto nodelete; -@@ -5723,8 +5765,9 @@ __xfs_bunmapi( - ASSERT(del.br_state == XFS_EXT_NORM); - del.br_state = XFS_EXT_UNWRITTEN; - error = xfs_bmap_add_extent_unwritten_real(tp, -- ip, &lastx, &cur, &del, -- firstblock, dfops, &logflags); -+ ip, whichfork, &lastx, &cur, -+ &del, firstblock, dfops, -+ &logflags); - if (error) - goto error0; - goto nodelete; -diff --git a/fs/xfs/libxfs/xfs_bmap_btree.c b/fs/xfs/libxfs/xfs_bmap_btree.c -index f76c169..5c39186 100644 ---- a/fs/xfs/libxfs/xfs_bmap_btree.c -+++ b/fs/xfs/libxfs/xfs_bmap_btree.c -@@ -453,8 +453,8 @@ xfs_bmbt_alloc_block( - - if (args.fsbno == NULLFSBLOCK) { - args.fsbno = be64_to_cpu(start->l); --try_another_ag: - args.type = XFS_ALLOCTYPE_START_BNO; -+try_another_ag: - /* - * Make sure there is sufficient room left in the AG to - * complete a full tree split for an extent insert. If -@@ -494,8 +494,8 @@ xfs_bmbt_alloc_block( - if (xfs_sb_version_hasreflink(&cur->bc_mp->m_sb) && - args.fsbno == NULLFSBLOCK && - args.type == XFS_ALLOCTYPE_NEAR_BNO) { -- cur->bc_private.b.dfops->dop_low = true; - args.fsbno = cur->bc_private.b.firstblock; -+ args.type = XFS_ALLOCTYPE_FIRST_AG; - goto try_another_ag; - } - -@@ -512,7 +512,7 @@ xfs_bmbt_alloc_block( - goto error0; - cur->bc_private.b.dfops->dop_low = true; - } -- if (args.fsbno == NULLFSBLOCK) { -+ if (WARN_ON_ONCE(args.fsbno == NULLFSBLOCK)) { - XFS_BTREE_TRACE_CURSOR(cur, XBT_EXIT); - *stat = 0; - return 0; -diff --git a/fs/xfs/libxfs/xfs_btree.c b/fs/xfs/libxfs/xfs_btree.c -index 21e6a6a..2849d3f 100644 ---- a/fs/xfs/libxfs/xfs_btree.c -+++ b/fs/xfs/libxfs/xfs_btree.c -@@ -810,7 +810,8 @@ xfs_btree_read_bufl( - xfs_daddr_t d; /* real disk block address */ - int error; - -- ASSERT(fsbno != NULLFSBLOCK); -+ if (!XFS_FSB_SANITY_CHECK(mp, fsbno)) -+ return -EFSCORRUPTED; - d = XFS_FSB_TO_DADDR(mp, fsbno); - error = xfs_trans_read_buf(mp, tp, mp->m_ddev_targp, d, - mp->m_bsize, lock, &bp, ops); -diff --git a/fs/xfs/libxfs/xfs_btree.h b/fs/xfs/libxfs/xfs_btree.h -index c2b01d1..3b0fc1a 100644 ---- a/fs/xfs/libxfs/xfs_btree.h -+++ b/fs/xfs/libxfs/xfs_btree.h -@@ -491,7 +491,7 @@ static inline int xfs_btree_get_level(struct xfs_btree_block *block) - #define XFS_FILBLKS_MAX(a,b) max_t(xfs_filblks_t, (a), (b)) - - #define XFS_FSB_SANITY_CHECK(mp,fsb) \ -- (XFS_FSB_TO_AGNO(mp, fsb) < mp->m_sb.sb_agcount && \ -+ (fsb && XFS_FSB_TO_AGNO(mp, fsb) < mp->m_sb.sb_agcount && \ - XFS_FSB_TO_AGBNO(mp, fsb) < mp->m_sb.sb_agblocks) - - /* -diff --git a/fs/xfs/libxfs/xfs_da_btree.c b/fs/xfs/libxfs/xfs_da_btree.c -index f2dc1a9..1bdf288 100644 ---- a/fs/xfs/libxfs/xfs_da_btree.c -+++ b/fs/xfs/libxfs/xfs_da_btree.c -@@ -2633,7 +2633,7 @@ xfs_da_read_buf( - /* - * Readahead the dir/attr block. - */ --xfs_daddr_t -+int - xfs_da_reada_buf( - struct xfs_inode *dp, - xfs_dablk_t bno, -@@ -2664,7 +2664,5 @@ xfs_da_reada_buf( - if (mapp != &map) - kmem_free(mapp); - -- if (error) -- return -1; -- return mappedbno; -+ return error; - } -diff --git a/fs/xfs/libxfs/xfs_da_btree.h b/fs/xfs/libxfs/xfs_da_btree.h -index 98c75cb..4e29cb6 100644 ---- a/fs/xfs/libxfs/xfs_da_btree.h -+++ b/fs/xfs/libxfs/xfs_da_btree.h -@@ -201,7 +201,7 @@ int xfs_da_read_buf(struct xfs_trans *trans, struct xfs_inode *dp, - xfs_dablk_t bno, xfs_daddr_t mappedbno, - struct xfs_buf **bpp, int whichfork, - const struct xfs_buf_ops *ops); --xfs_daddr_t xfs_da_reada_buf(struct xfs_inode *dp, xfs_dablk_t bno, -+int xfs_da_reada_buf(struct xfs_inode *dp, xfs_dablk_t bno, - xfs_daddr_t mapped_bno, int whichfork, - const struct xfs_buf_ops *ops); - int xfs_da_shrink_inode(xfs_da_args_t *args, xfs_dablk_t dead_blkno, -diff --git a/fs/xfs/libxfs/xfs_dir2_node.c b/fs/xfs/libxfs/xfs_dir2_node.c -index 75a5574..bbd1238 100644 ---- a/fs/xfs/libxfs/xfs_dir2_node.c -+++ b/fs/xfs/libxfs/xfs_dir2_node.c -@@ -155,6 +155,42 @@ const struct xfs_buf_ops xfs_dir3_free_buf_ops = { - .verify_write = xfs_dir3_free_write_verify, - }; - -+/* Everything ok in the free block header? */ -+static bool -+xfs_dir3_free_header_check( -+ struct xfs_inode *dp, -+ xfs_dablk_t fbno, -+ struct xfs_buf *bp) -+{ -+ struct xfs_mount *mp = dp->i_mount; -+ unsigned int firstdb; -+ int maxbests; -+ -+ maxbests = dp->d_ops->free_max_bests(mp->m_dir_geo); -+ firstdb = (xfs_dir2_da_to_db(mp->m_dir_geo, fbno) - -+ xfs_dir2_byte_to_db(mp->m_dir_geo, XFS_DIR2_FREE_OFFSET)) * -+ maxbests; -+ if (xfs_sb_version_hascrc(&mp->m_sb)) { -+ struct xfs_dir3_free_hdr *hdr3 = bp->b_addr; -+ -+ if (be32_to_cpu(hdr3->firstdb) != firstdb) -+ return false; -+ if (be32_to_cpu(hdr3->nvalid) > maxbests) -+ return false; -+ if (be32_to_cpu(hdr3->nvalid) < be32_to_cpu(hdr3->nused)) -+ return false; -+ } else { -+ struct xfs_dir2_free_hdr *hdr = bp->b_addr; -+ -+ if (be32_to_cpu(hdr->firstdb) != firstdb) -+ return false; -+ if (be32_to_cpu(hdr->nvalid) > maxbests) -+ return false; -+ if (be32_to_cpu(hdr->nvalid) < be32_to_cpu(hdr->nused)) -+ return false; -+ } -+ return true; -+} - - static int - __xfs_dir3_free_read( -@@ -168,11 +204,22 @@ __xfs_dir3_free_read( - - err = xfs_da_read_buf(tp, dp, fbno, mappedbno, bpp, - XFS_DATA_FORK, &xfs_dir3_free_buf_ops); -+ if (err || !*bpp) -+ return err; -+ -+ /* Check things that we can't do in the verifier. */ -+ if (!xfs_dir3_free_header_check(dp, fbno, *bpp)) { -+ xfs_buf_ioerror(*bpp, -EFSCORRUPTED); -+ xfs_verifier_error(*bpp); -+ xfs_trans_brelse(tp, *bpp); -+ return -EFSCORRUPTED; -+ } - - /* try read returns without an error or *bpp if it lands in a hole */ -- if (!err && tp && *bpp) -+ if (tp) - xfs_trans_buf_set_type(tp, *bpp, XFS_BLFT_DIR_FREE_BUF); -- return err; -+ -+ return 0; - } - - int -diff --git a/fs/xfs/libxfs/xfs_ialloc.c b/fs/xfs/libxfs/xfs_ialloc.c -index d45c037..a2818f6 100644 ---- a/fs/xfs/libxfs/xfs_ialloc.c -+++ b/fs/xfs/libxfs/xfs_ialloc.c -@@ -51,8 +51,7 @@ xfs_ialloc_cluster_alignment( - struct xfs_mount *mp) - { - if (xfs_sb_version_hasalign(&mp->m_sb) && -- mp->m_sb.sb_inoalignmt >= -- XFS_B_TO_FSBT(mp, mp->m_inode_cluster_size)) -+ mp->m_sb.sb_inoalignmt >= xfs_icluster_size_fsb(mp)) - return mp->m_sb.sb_inoalignmt; - return 1; - } -diff --git a/fs/xfs/libxfs/xfs_ialloc_btree.c b/fs/xfs/libxfs/xfs_ialloc_btree.c -index 6c6b959..b9c351f 100644 ---- a/fs/xfs/libxfs/xfs_ialloc_btree.c -+++ b/fs/xfs/libxfs/xfs_ialloc_btree.c -@@ -82,11 +82,12 @@ xfs_finobt_set_root( - } - - STATIC int --xfs_inobt_alloc_block( -+__xfs_inobt_alloc_block( - struct xfs_btree_cur *cur, - union xfs_btree_ptr *start, - union xfs_btree_ptr *new, -- int *stat) -+ int *stat, -+ enum xfs_ag_resv_type resv) - { - xfs_alloc_arg_t args; /* block allocation args */ - int error; /* error return value */ -@@ -103,6 +104,7 @@ xfs_inobt_alloc_block( - args.maxlen = 1; - args.prod = 1; - args.type = XFS_ALLOCTYPE_NEAR_BNO; -+ args.resv = resv; - - error = xfs_alloc_vextent(&args); - if (error) { -@@ -123,6 +125,27 @@ xfs_inobt_alloc_block( - } - - STATIC int -+xfs_inobt_alloc_block( -+ struct xfs_btree_cur *cur, -+ union xfs_btree_ptr *start, -+ union xfs_btree_ptr *new, -+ int *stat) -+{ -+ return __xfs_inobt_alloc_block(cur, start, new, stat, XFS_AG_RESV_NONE); -+} -+ -+STATIC int -+xfs_finobt_alloc_block( -+ struct xfs_btree_cur *cur, -+ union xfs_btree_ptr *start, -+ union xfs_btree_ptr *new, -+ int *stat) -+{ -+ return __xfs_inobt_alloc_block(cur, start, new, stat, -+ XFS_AG_RESV_METADATA); -+} -+ -+STATIC int - xfs_inobt_free_block( - struct xfs_btree_cur *cur, - struct xfs_buf *bp) -@@ -328,7 +351,7 @@ static const struct xfs_btree_ops xfs_finobt_ops = { - - .dup_cursor = xfs_inobt_dup_cursor, - .set_root = xfs_finobt_set_root, -- .alloc_block = xfs_inobt_alloc_block, -+ .alloc_block = xfs_finobt_alloc_block, - .free_block = xfs_inobt_free_block, - .get_minrecs = xfs_inobt_get_minrecs, - .get_maxrecs = xfs_inobt_get_maxrecs, -@@ -478,3 +501,64 @@ xfs_inobt_rec_check_count( - return 0; - } - #endif /* DEBUG */ -+ -+static xfs_extlen_t -+xfs_inobt_max_size( -+ struct xfs_mount *mp) -+{ -+ /* Bail out if we're uninitialized, which can happen in mkfs. */ -+ if (mp->m_inobt_mxr[0] == 0) -+ return 0; -+ -+ return xfs_btree_calc_size(mp, mp->m_inobt_mnr, -+ (uint64_t)mp->m_sb.sb_agblocks * mp->m_sb.sb_inopblock / -+ XFS_INODES_PER_CHUNK); -+} -+ -+static int -+xfs_inobt_count_blocks( -+ struct xfs_mount *mp, -+ xfs_agnumber_t agno, -+ xfs_btnum_t btnum, -+ xfs_extlen_t *tree_blocks) -+{ -+ struct xfs_buf *agbp; -+ struct xfs_btree_cur *cur; -+ int error; -+ -+ error = xfs_ialloc_read_agi(mp, NULL, agno, &agbp); -+ if (error) -+ return error; -+ -+ cur = xfs_inobt_init_cursor(mp, NULL, agbp, agno, btnum); -+ error = xfs_btree_count_blocks(cur, tree_blocks); -+ xfs_btree_del_cursor(cur, error ? XFS_BTREE_ERROR : XFS_BTREE_NOERROR); -+ xfs_buf_relse(agbp); -+ -+ return error; -+} -+ -+/* -+ * Figure out how many blocks to reserve and how many are used by this btree. -+ */ -+int -+xfs_finobt_calc_reserves( -+ struct xfs_mount *mp, -+ xfs_agnumber_t agno, -+ xfs_extlen_t *ask, -+ xfs_extlen_t *used) -+{ -+ xfs_extlen_t tree_len = 0; -+ int error; -+ -+ if (!xfs_sb_version_hasfinobt(&mp->m_sb)) -+ return 0; -+ -+ error = xfs_inobt_count_blocks(mp, agno, XFS_BTNUM_FINO, &tree_len); -+ if (error) -+ return error; -+ -+ *ask += xfs_inobt_max_size(mp); -+ *used += tree_len; -+ return 0; -+} -diff --git a/fs/xfs/libxfs/xfs_ialloc_btree.h b/fs/xfs/libxfs/xfs_ialloc_btree.h -index bd88453..aa81e2e 100644 ---- a/fs/xfs/libxfs/xfs_ialloc_btree.h -+++ b/fs/xfs/libxfs/xfs_ialloc_btree.h -@@ -72,4 +72,7 @@ int xfs_inobt_rec_check_count(struct xfs_mount *, - #define xfs_inobt_rec_check_count(mp, rec) 0 - #endif /* DEBUG */ - -+int xfs_finobt_calc_reserves(struct xfs_mount *mp, xfs_agnumber_t agno, -+ xfs_extlen_t *ask, xfs_extlen_t *used); -+ - #endif /* __XFS_IALLOC_BTREE_H__ */ -diff --git a/fs/xfs/libxfs/xfs_inode_fork.c b/fs/xfs/libxfs/xfs_inode_fork.c -index 222e103..25c1e07 100644 ---- a/fs/xfs/libxfs/xfs_inode_fork.c -+++ b/fs/xfs/libxfs/xfs_inode_fork.c -@@ -26,6 +26,7 @@ - #include "xfs_inode.h" - #include "xfs_trans.h" - #include "xfs_inode_item.h" -+#include "xfs_btree.h" - #include "xfs_bmap_btree.h" - #include "xfs_bmap.h" - #include "xfs_error.h" -@@ -429,11 +430,13 @@ xfs_iformat_btree( - /* REFERENCED */ - int nrecs; - int size; -+ int level; - - ifp = XFS_IFORK_PTR(ip, whichfork); - dfp = (xfs_bmdr_block_t *)XFS_DFORK_PTR(dip, whichfork); - size = XFS_BMAP_BROOT_SPACE(mp, dfp); - nrecs = be16_to_cpu(dfp->bb_numrecs); -+ level = be16_to_cpu(dfp->bb_level); - - /* - * blow out if -- fork has less extents than can fit in -@@ -446,7 +449,8 @@ xfs_iformat_btree( - XFS_IFORK_MAXEXT(ip, whichfork) || - XFS_BMDR_SPACE_CALC(nrecs) > - XFS_DFORK_SIZE(dip, mp, whichfork) || -- XFS_IFORK_NEXTENTS(ip, whichfork) > ip->i_d.di_nblocks)) { -+ XFS_IFORK_NEXTENTS(ip, whichfork) > ip->i_d.di_nblocks) || -+ level == 0 || level > XFS_BTREE_MAXLEVELS) { - xfs_warn(mp, "corrupt inode %Lu (btree).", - (unsigned long long) ip->i_ino); - XFS_CORRUPTION_ERROR("xfs_iformat_btree", XFS_ERRLEVEL_LOW, -@@ -497,15 +501,14 @@ xfs_iread_extents( - * We know that the size is valid (it's checked in iformat_btree) - */ - ifp->if_bytes = ifp->if_real_bytes = 0; -- ifp->if_flags |= XFS_IFEXTENTS; - xfs_iext_add(ifp, 0, nextents); - error = xfs_bmap_read_extents(tp, ip, whichfork); - if (error) { - xfs_iext_destroy(ifp); -- ifp->if_flags &= ~XFS_IFEXTENTS; - return error; - } - xfs_validate_extents(ifp, nextents, XFS_EXTFMT_INODE(ip)); -+ ifp->if_flags |= XFS_IFEXTENTS; - return 0; - } - /* -diff --git a/fs/xfs/xfs_aops.c b/fs/xfs/xfs_aops.c -index 06763f5..0457abe 100644 ---- a/fs/xfs/xfs_aops.c -+++ b/fs/xfs/xfs_aops.c -@@ -279,54 +279,49 @@ xfs_end_io( - struct xfs_ioend *ioend = - container_of(work, struct xfs_ioend, io_work); - struct xfs_inode *ip = XFS_I(ioend->io_inode); -+ xfs_off_t offset = ioend->io_offset; -+ size_t size = ioend->io_size; - int error = ioend->io_bio->bi_error; - - /* -- * Set an error if the mount has shut down and proceed with end I/O -- * processing so it can perform whatever cleanups are necessary. -+ * Just clean up the in-memory strutures if the fs has been shut down. - */ -- if (XFS_FORCED_SHUTDOWN(ip->i_mount)) -+ if (XFS_FORCED_SHUTDOWN(ip->i_mount)) { - error = -EIO; -+ goto done; -+ } - - /* -- * For a CoW extent, we need to move the mapping from the CoW fork -- * to the data fork. If instead an error happened, just dump the -- * new blocks. -+ * Clean up any COW blocks on an I/O error. - */ -- if (ioend->io_type == XFS_IO_COW) { -- if (error) -- goto done; -- if (ioend->io_bio->bi_error) { -- error = xfs_reflink_cancel_cow_range(ip, -- ioend->io_offset, ioend->io_size); -- goto done; -+ if (unlikely(error)) { -+ switch (ioend->io_type) { -+ case XFS_IO_COW: -+ xfs_reflink_cancel_cow_range(ip, offset, size, true); -+ break; - } -- error = xfs_reflink_end_cow(ip, ioend->io_offset, -- ioend->io_size); -- if (error) -- goto done; -+ -+ goto done; - } - - /* -- * For unwritten extents we need to issue transactions to convert a -- * range to normal written extens after the data I/O has finished. -- * Detecting and handling completion IO errors is done individually -- * for each case as different cleanup operations need to be performed -- * on error. -+ * Success: commit the COW or unwritten blocks if needed. - */ -- if (ioend->io_type == XFS_IO_UNWRITTEN) { -- if (error) -- goto done; -- error = xfs_iomap_write_unwritten(ip, ioend->io_offset, -- ioend->io_size); -- } else if (ioend->io_append_trans) { -- error = xfs_setfilesize_ioend(ioend, error); -- } else { -- ASSERT(!xfs_ioend_is_append(ioend) || -- ioend->io_type == XFS_IO_COW); -+ switch (ioend->io_type) { -+ case XFS_IO_COW: -+ error = xfs_reflink_end_cow(ip, offset, size); -+ break; -+ case XFS_IO_UNWRITTEN: -+ error = xfs_iomap_write_unwritten(ip, offset, size); -+ break; -+ default: -+ ASSERT(!xfs_ioend_is_append(ioend) || ioend->io_append_trans); -+ break; - } - - done: -+ if (ioend->io_append_trans) -+ error = xfs_setfilesize_ioend(ioend, error); - xfs_destroy_ioend(ioend, error); - } - -@@ -486,6 +481,12 @@ xfs_submit_ioend( - struct xfs_ioend *ioend, - int status) - { -+ /* Convert CoW extents to regular */ -+ if (!status && ioend->io_type == XFS_IO_COW) { -+ status = xfs_reflink_convert_cow(XFS_I(ioend->io_inode), -+ ioend->io_offset, ioend->io_size); -+ } -+ - /* Reserve log space if we might write beyond the on-disk inode size. */ - if (!status && - ioend->io_type != XFS_IO_UNWRITTEN && -@@ -1257,44 +1258,6 @@ xfs_map_trim_size( - bh_result->b_size = mapping_size; - } - --/* Bounce unaligned directio writes to the page cache. */ --static int --xfs_bounce_unaligned_dio_write( -- struct xfs_inode *ip, -- xfs_fileoff_t offset_fsb, -- struct xfs_bmbt_irec *imap) --{ -- struct xfs_bmbt_irec irec; -- xfs_fileoff_t delta; -- bool shared; -- bool x; -- int error; -- -- irec = *imap; -- if (offset_fsb > irec.br_startoff) { -- delta = offset_fsb - irec.br_startoff; -- irec.br_blockcount -= delta; -- irec.br_startblock += delta; -- irec.br_startoff = offset_fsb; -- } -- error = xfs_reflink_trim_around_shared(ip, &irec, &shared, &x); -- if (error) -- return error; -- -- /* -- * We're here because we're trying to do a directio write to a -- * region that isn't aligned to a filesystem block. If any part -- * of the extent is shared, fall back to buffered mode to handle -- * the RMW. This is done by returning -EREMCHG ("remote addr -- * changed"), which is caught further up the call stack. -- */ -- if (shared) { -- trace_xfs_reflink_bounce_dio_write(ip, imap); -- return -EREMCHG; -- } -- return 0; --} -- - STATIC int - __xfs_get_blocks( - struct inode *inode, -@@ -1432,13 +1395,6 @@ __xfs_get_blocks( - if (imap.br_startblock != HOLESTARTBLOCK && - imap.br_startblock != DELAYSTARTBLOCK && - (create || !ISUNWRITTEN(&imap))) { -- if (create && direct && !is_cow) { -- error = xfs_bounce_unaligned_dio_write(ip, offset_fsb, -- &imap); -- if (error) -- return error; -- } -- - xfs_map_buffer(inode, bh_result, &imap, offset); - if (ISUNWRITTEN(&imap)) - set_buffer_unwritten(bh_result); -diff --git a/fs/xfs/xfs_bmap_util.c b/fs/xfs/xfs_bmap_util.c -index efb8ccd..5c395e4 100644 ---- a/fs/xfs/xfs_bmap_util.c -+++ b/fs/xfs/xfs_bmap_util.c -@@ -917,17 +917,18 @@ xfs_can_free_eofblocks(struct xfs_inode *ip, bool force) - */ - int - xfs_free_eofblocks( -- xfs_mount_t *mp, -- xfs_inode_t *ip, -- bool need_iolock) -+ struct xfs_inode *ip) - { -- xfs_trans_t *tp; -- int error; -- xfs_fileoff_t end_fsb; -- xfs_fileoff_t last_fsb; -- xfs_filblks_t map_len; -- int nimaps; -- xfs_bmbt_irec_t imap; -+ struct xfs_trans *tp; -+ int error; -+ xfs_fileoff_t end_fsb; -+ xfs_fileoff_t last_fsb; -+ xfs_filblks_t map_len; -+ int nimaps; -+ struct xfs_bmbt_irec imap; -+ struct xfs_mount *mp = ip->i_mount; -+ -+ ASSERT(xfs_isilocked(ip, XFS_IOLOCK_EXCL)); - - /* - * Figure out if there are any blocks beyond the end -@@ -944,6 +945,10 @@ xfs_free_eofblocks( - error = xfs_bmapi_read(ip, end_fsb, map_len, &imap, &nimaps, 0); - xfs_iunlock(ip, XFS_ILOCK_SHARED); - -+ /* -+ * If there are blocks after the end of file, truncate the file to its -+ * current size to free them up. -+ */ - if (!error && (nimaps != 0) && - (imap.br_startblock != HOLESTARTBLOCK || - ip->i_delayed_blks)) { -@@ -954,22 +959,13 @@ xfs_free_eofblocks( - if (error) - return error; - -- /* -- * There are blocks after the end of file. -- * Free them up now by truncating the file to -- * its current size. -- */ -- if (need_iolock) { -- if (!xfs_ilock_nowait(ip, XFS_IOLOCK_EXCL)) -- return -EAGAIN; -- } -+ /* wait on dio to ensure i_size has settled */ -+ inode_dio_wait(VFS_I(ip)); - - error = xfs_trans_alloc(mp, &M_RES(mp)->tr_itruncate, 0, 0, 0, - &tp); - if (error) { - ASSERT(XFS_FORCED_SHUTDOWN(mp)); -- if (need_iolock) -- xfs_iunlock(ip, XFS_IOLOCK_EXCL); - return error; - } - -@@ -997,8 +993,6 @@ xfs_free_eofblocks( - } - - xfs_iunlock(ip, XFS_ILOCK_EXCL); -- if (need_iolock) -- xfs_iunlock(ip, XFS_IOLOCK_EXCL); - } - return error; - } -@@ -1393,10 +1387,16 @@ xfs_shift_file_space( - xfs_fileoff_t stop_fsb; - xfs_fileoff_t next_fsb; - xfs_fileoff_t shift_fsb; -+ uint resblks; - - ASSERT(direction == SHIFT_LEFT || direction == SHIFT_RIGHT); - - if (direction == SHIFT_LEFT) { -+ /* -+ * Reserve blocks to cover potential extent merges after left -+ * shift operations. -+ */ -+ resblks = XFS_DIOSTRAT_SPACE_RES(mp, 0); - next_fsb = XFS_B_TO_FSB(mp, offset + len); - stop_fsb = XFS_B_TO_FSB(mp, VFS_I(ip)->i_size); - } else { -@@ -1404,6 +1404,7 @@ xfs_shift_file_space( - * If right shift, delegate the work of initialization of - * next_fsb to xfs_bmap_shift_extent as it has ilock held. - */ -+ resblks = 0; - next_fsb = NULLFSBLOCK; - stop_fsb = XFS_B_TO_FSB(mp, offset); - } -@@ -1415,7 +1416,7 @@ xfs_shift_file_space( - * into the accessible region of the file. - */ - if (xfs_can_free_eofblocks(ip, true)) { -- error = xfs_free_eofblocks(mp, ip, false); -+ error = xfs_free_eofblocks(ip); - if (error) - return error; - } -@@ -1445,21 +1446,14 @@ xfs_shift_file_space( - } - - while (!error && !done) { -- /* -- * We would need to reserve permanent block for transaction. -- * This will come into picture when after shifting extent into -- * hole we found that adjacent extents can be merged which -- * may lead to freeing of a block during record update. -- */ -- error = xfs_trans_alloc(mp, &M_RES(mp)->tr_write, -- XFS_DIOSTRAT_SPACE_RES(mp, 0), 0, 0, &tp); -+ error = xfs_trans_alloc(mp, &M_RES(mp)->tr_write, resblks, 0, 0, -+ &tp); - if (error) - break; - - xfs_ilock(ip, XFS_ILOCK_EXCL); - error = xfs_trans_reserve_quota(tp, mp, ip->i_udquot, -- ip->i_gdquot, ip->i_pdquot, -- XFS_DIOSTRAT_SPACE_RES(mp, 0), 0, -+ ip->i_gdquot, ip->i_pdquot, resblks, 0, - XFS_QMOPT_RES_REGBLKS); - if (error) - goto out_trans_cancel; -diff --git a/fs/xfs/xfs_bmap_util.h b/fs/xfs/xfs_bmap_util.h -index 68a621a..f100539 100644 ---- a/fs/xfs/xfs_bmap_util.h -+++ b/fs/xfs/xfs_bmap_util.h -@@ -63,8 +63,7 @@ int xfs_insert_file_space(struct xfs_inode *, xfs_off_t offset, - - /* EOF block manipulation functions */ - bool xfs_can_free_eofblocks(struct xfs_inode *ip, bool force); --int xfs_free_eofblocks(struct xfs_mount *mp, struct xfs_inode *ip, -- bool need_iolock); -+int xfs_free_eofblocks(struct xfs_inode *ip); - - int xfs_swap_extents(struct xfs_inode *ip, struct xfs_inode *tip, - struct xfs_swapext *sx); -diff --git a/fs/xfs/xfs_buf_item.c b/fs/xfs/xfs_buf_item.c -index 2975cb2..0306168 100644 ---- a/fs/xfs/xfs_buf_item.c -+++ b/fs/xfs/xfs_buf_item.c -@@ -1162,6 +1162,7 @@ xfs_buf_iodone_callbacks( - */ - bp->b_last_error = 0; - bp->b_retries = 0; -+ bp->b_first_retry_time = 0; - - xfs_buf_do_callbacks(bp); - bp->b_fspriv = NULL; -diff --git a/fs/xfs/xfs_extent_busy.c b/fs/xfs/xfs_extent_busy.c -index 162dc18..29c2f99 100644 ---- a/fs/xfs/xfs_extent_busy.c -+++ b/fs/xfs/xfs_extent_busy.c -@@ -45,18 +45,7 @@ xfs_extent_busy_insert( - struct rb_node **rbp; - struct rb_node *parent = NULL; - -- new = kmem_zalloc(sizeof(struct xfs_extent_busy), KM_MAYFAIL); -- if (!new) { -- /* -- * No Memory! Since it is now not possible to track the free -- * block, make this a synchronous transaction to insure that -- * the block is not reused before this transaction commits. -- */ -- trace_xfs_extent_busy_enomem(tp->t_mountp, agno, bno, len); -- xfs_trans_set_sync(tp); -- return; -- } -- -+ new = kmem_zalloc(sizeof(struct xfs_extent_busy), KM_SLEEP); - new->agno = agno; - new->bno = bno; - new->length = len; -diff --git a/fs/xfs/xfs_file.c b/fs/xfs/xfs_file.c -index 9a5d64b..1209ad2 100644 ---- a/fs/xfs/xfs_file.c -+++ b/fs/xfs/xfs_file.c -@@ -554,6 +554,15 @@ xfs_file_dio_aio_write( - if ((iocb->ki_pos & mp->m_blockmask) || - ((iocb->ki_pos + count) & mp->m_blockmask)) { - unaligned_io = 1; -+ -+ /* -+ * We can't properly handle unaligned direct I/O to reflink -+ * files yet, as we can't unshare a partial block. -+ */ -+ if (xfs_is_reflink_inode(ip)) { -+ trace_xfs_reflink_bounce_dio_write(ip, iocb->ki_pos, count); -+ return -EREMCHG; -+ } - iolock = XFS_IOLOCK_EXCL; - } else { - iolock = XFS_IOLOCK_SHARED; -@@ -675,8 +684,10 @@ xfs_file_buffered_aio_write( - struct xfs_inode *ip = XFS_I(inode); - ssize_t ret; - int enospc = 0; -- int iolock = XFS_IOLOCK_EXCL; -+ int iolock; - -+write_retry: -+ iolock = XFS_IOLOCK_EXCL; - xfs_rw_ilock(ip, iolock); - - ret = xfs_file_aio_write_checks(iocb, from, &iolock); -@@ -686,7 +697,6 @@ xfs_file_buffered_aio_write( - /* We can write back this queue in page reclaim */ - current->backing_dev_info = inode_to_bdi(inode); - --write_retry: - trace_xfs_file_buffered_write(ip, iov_iter_count(from), iocb->ki_pos); - ret = iomap_file_buffered_write(iocb, from, &xfs_iomap_ops); - if (likely(ret >= 0)) -@@ -702,18 +712,21 @@ xfs_file_buffered_aio_write( - * running at the same time. - */ - if (ret == -EDQUOT && !enospc) { -+ xfs_rw_iunlock(ip, iolock); - enospc = xfs_inode_free_quota_eofblocks(ip); - if (enospc) - goto write_retry; - enospc = xfs_inode_free_quota_cowblocks(ip); - if (enospc) - goto write_retry; -+ iolock = 0; - } else if (ret == -ENOSPC && !enospc) { - struct xfs_eofblocks eofb = {0}; - - enospc = 1; - xfs_flush_inodes(ip->i_mount); -- eofb.eof_scan_owner = ip->i_ino; /* for locking */ -+ -+ xfs_rw_iunlock(ip, iolock); - eofb.eof_flags = XFS_EOF_FLAGS_SYNC; - xfs_icache_free_eofblocks(ip->i_mount, &eofb); - goto write_retry; -@@ -721,7 +734,8 @@ xfs_file_buffered_aio_write( - - current->backing_dev_info = NULL; - out: -- xfs_rw_iunlock(ip, iolock); -+ if (iolock) -+ xfs_rw_iunlock(ip, iolock); - return ret; - } - -@@ -987,9 +1001,9 @@ xfs_dir_open( - */ - mode = xfs_ilock_data_map_shared(ip); - if (ip->i_d.di_nextents > 0) -- xfs_dir3_data_readahead(ip, 0, -1); -+ error = xfs_dir3_data_readahead(ip, 0, -1); - xfs_iunlock(ip, mode); -- return 0; -+ return error; - } - - STATIC int -diff --git a/fs/xfs/xfs_icache.c b/fs/xfs/xfs_icache.c -index 29cc988..3fb1f3f 100644 ---- a/fs/xfs/xfs_icache.c -+++ b/fs/xfs/xfs_icache.c -@@ -1324,13 +1324,10 @@ xfs_inode_free_eofblocks( - int flags, - void *args) - { -- int ret; -+ int ret = 0; - struct xfs_eofblocks *eofb = args; -- bool need_iolock = true; - int match; - -- ASSERT(!eofb || (eofb && eofb->eof_scan_owner != 0)); -- - if (!xfs_can_free_eofblocks(ip, false)) { - /* inode could be preallocated or append-only */ - trace_xfs_inode_free_eofblocks_invalid(ip); -@@ -1358,21 +1355,19 @@ xfs_inode_free_eofblocks( - if (eofb->eof_flags & XFS_EOF_FLAGS_MINFILESIZE && - XFS_ISIZE(ip) < eofb->eof_min_file_size) - return 0; -- -- /* -- * A scan owner implies we already hold the iolock. Skip it in -- * xfs_free_eofblocks() to avoid deadlock. This also eliminates -- * the possibility of EAGAIN being returned. -- */ -- if (eofb->eof_scan_owner == ip->i_ino) -- need_iolock = false; - } - -- ret = xfs_free_eofblocks(ip->i_mount, ip, need_iolock); -- -- /* don't revisit the inode if we're not waiting */ -- if (ret == -EAGAIN && !(flags & SYNC_WAIT)) -- ret = 0; -+ /* -+ * If the caller is waiting, return -EAGAIN to keep the background -+ * scanner moving and revisit the inode in a subsequent pass. -+ */ -+ if (!xfs_ilock_nowait(ip, XFS_IOLOCK_EXCL)) { -+ if (flags & SYNC_WAIT) -+ ret = -EAGAIN; -+ return ret; -+ } -+ ret = xfs_free_eofblocks(ip); -+ xfs_iunlock(ip, XFS_IOLOCK_EXCL); - - return ret; - } -@@ -1419,15 +1414,10 @@ __xfs_inode_free_quota_eofblocks( - struct xfs_eofblocks eofb = {0}; - struct xfs_dquot *dq; - -- ASSERT(xfs_isilocked(ip, XFS_IOLOCK_EXCL)); -- - /* -- * Set the scan owner to avoid a potential livelock. Otherwise, the scan -- * can repeatedly trylock on the inode we're currently processing. We -- * run a sync scan to increase effectiveness and use the union filter to -+ * Run a sync scan to increase effectiveness and use the union filter to - * cover all applicable quotas in a single scan. - */ -- eofb.eof_scan_owner = ip->i_ino; - eofb.eof_flags = XFS_EOF_FLAGS_UNION|XFS_EOF_FLAGS_SYNC; - - if (XFS_IS_UQUOTA_ENFORCED(ip->i_mount)) { -@@ -1579,12 +1569,9 @@ xfs_inode_free_cowblocks( - { - int ret; - struct xfs_eofblocks *eofb = args; -- bool need_iolock = true; - int match; - struct xfs_ifork *ifp = XFS_IFORK_PTR(ip, XFS_COW_FORK); - -- ASSERT(!eofb || (eofb && eofb->eof_scan_owner != 0)); -- - /* - * Just clear the tag if we have an empty cow fork or none at all. It's - * possible the inode was fully unshared since it was originally tagged. -@@ -1617,28 +1604,16 @@ xfs_inode_free_cowblocks( - if (eofb->eof_flags & XFS_EOF_FLAGS_MINFILESIZE && - XFS_ISIZE(ip) < eofb->eof_min_file_size) - return 0; -- -- /* -- * A scan owner implies we already hold the iolock. Skip it in -- * xfs_free_eofblocks() to avoid deadlock. This also eliminates -- * the possibility of EAGAIN being returned. -- */ -- if (eofb->eof_scan_owner == ip->i_ino) -- need_iolock = false; - } - - /* Free the CoW blocks */ -- if (need_iolock) { -- xfs_ilock(ip, XFS_IOLOCK_EXCL); -- xfs_ilock(ip, XFS_MMAPLOCK_EXCL); -- } -+ xfs_ilock(ip, XFS_IOLOCK_EXCL); -+ xfs_ilock(ip, XFS_MMAPLOCK_EXCL); - -- ret = xfs_reflink_cancel_cow_range(ip, 0, NULLFILEOFF); -+ ret = xfs_reflink_cancel_cow_range(ip, 0, NULLFILEOFF, false); - -- if (need_iolock) { -- xfs_iunlock(ip, XFS_MMAPLOCK_EXCL); -- xfs_iunlock(ip, XFS_IOLOCK_EXCL); -- } -+ xfs_iunlock(ip, XFS_MMAPLOCK_EXCL); -+ xfs_iunlock(ip, XFS_IOLOCK_EXCL); - - return ret; - } -diff --git a/fs/xfs/xfs_icache.h b/fs/xfs/xfs_icache.h -index a1e02f4..8a7c849 100644 ---- a/fs/xfs/xfs_icache.h -+++ b/fs/xfs/xfs_icache.h -@@ -27,7 +27,6 @@ struct xfs_eofblocks { - kgid_t eof_gid; - prid_t eof_prid; - __u64 eof_min_file_size; -- xfs_ino_t eof_scan_owner; - }; - - #define SYNC_WAIT 0x0001 /* wait for i/o to complete */ -@@ -102,7 +101,6 @@ xfs_fs_eofblocks_from_user( - dst->eof_flags = src->eof_flags; - dst->eof_prid = src->eof_prid; - dst->eof_min_file_size = src->eof_min_file_size; -- dst->eof_scan_owner = NULLFSINO; - - dst->eof_uid = INVALID_UID; - if (src->eof_flags & XFS_EOF_FLAGS_UID) { -diff --git a/fs/xfs/xfs_inode.c b/fs/xfs/xfs_inode.c -index 512ff13..e50636c 100644 ---- a/fs/xfs/xfs_inode.c -+++ b/fs/xfs/xfs_inode.c -@@ -1624,7 +1624,7 @@ xfs_itruncate_extents( - - /* Remove all pending CoW reservations. */ - error = xfs_reflink_cancel_cow_blocks(ip, &tp, first_unmap_block, -- last_block); -+ last_block, true); - if (error) - goto out; - -@@ -1701,32 +1701,34 @@ xfs_release( - if (xfs_can_free_eofblocks(ip, false)) { - - /* -+ * Check if the inode is being opened, written and closed -+ * frequently and we have delayed allocation blocks outstanding -+ * (e.g. streaming writes from the NFS server), truncating the -+ * blocks past EOF will cause fragmentation to occur. -+ * -+ * In this case don't do the truncation, but we have to be -+ * careful how we detect this case. Blocks beyond EOF show up as -+ * i_delayed_blks even when the inode is clean, so we need to -+ * truncate them away first before checking for a dirty release. -+ * Hence on the first dirty close we will still remove the -+ * speculative allocation, but after that we will leave it in -+ * place. -+ */ -+ if (xfs_iflags_test(ip, XFS_IDIRTY_RELEASE)) -+ return 0; -+ /* - * If we can't get the iolock just skip truncating the blocks - * past EOF because we could deadlock with the mmap_sem -- * otherwise. We'll get another chance to drop them once the -+ * otherwise. We'll get another chance to drop them once the - * last reference to the inode is dropped, so we'll never leak - * blocks permanently. -- * -- * Further, check if the inode is being opened, written and -- * closed frequently and we have delayed allocation blocks -- * outstanding (e.g. streaming writes from the NFS server), -- * truncating the blocks past EOF will cause fragmentation to -- * occur. -- * -- * In this case don't do the truncation, either, but we have to -- * be careful how we detect this case. Blocks beyond EOF show -- * up as i_delayed_blks even when the inode is clean, so we -- * need to truncate them away first before checking for a dirty -- * release. Hence on the first dirty close we will still remove -- * the speculative allocation, but after that we will leave it -- * in place. - */ -- if (xfs_iflags_test(ip, XFS_IDIRTY_RELEASE)) -- return 0; -- -- error = xfs_free_eofblocks(mp, ip, true); -- if (error && error != -EAGAIN) -- return error; -+ if (xfs_ilock_nowait(ip, XFS_IOLOCK_EXCL)) { -+ error = xfs_free_eofblocks(ip); -+ xfs_iunlock(ip, XFS_IOLOCK_EXCL); -+ if (error) -+ return error; -+ } - - /* delalloc blocks after truncation means it really is dirty */ - if (ip->i_delayed_blks) -@@ -1801,22 +1803,23 @@ xfs_inactive_ifree( - int error; - - /* -- * The ifree transaction might need to allocate blocks for record -- * insertion to the finobt. We don't want to fail here at ENOSPC, so -- * allow ifree to dip into the reserved block pool if necessary. -- * -- * Freeing large sets of inodes generally means freeing inode chunks, -- * directory and file data blocks, so this should be relatively safe. -- * Only under severe circumstances should it be possible to free enough -- * inodes to exhaust the reserve block pool via finobt expansion while -- * at the same time not creating free space in the filesystem. -+ * We try to use a per-AG reservation for any block needed by the finobt -+ * tree, but as the finobt feature predates the per-AG reservation -+ * support a degraded file system might not have enough space for the -+ * reservation at mount time. In that case try to dip into the reserved -+ * pool and pray. - * - * Send a warning if the reservation does happen to fail, as the inode - * now remains allocated and sits on the unlinked list until the fs is - * repaired. - */ -- error = xfs_trans_alloc(mp, &M_RES(mp)->tr_ifree, -- XFS_IFREE_SPACE_RES(mp), 0, XFS_TRANS_RESERVE, &tp); -+ if (unlikely(mp->m_inotbt_nores)) { -+ error = xfs_trans_alloc(mp, &M_RES(mp)->tr_ifree, -+ XFS_IFREE_SPACE_RES(mp), 0, XFS_TRANS_RESERVE, -+ &tp); -+ } else { -+ error = xfs_trans_alloc(mp, &M_RES(mp)->tr_ifree, 0, 0, 0, &tp); -+ } - if (error) { - if (error == -ENOSPC) { - xfs_warn_ratelimited(mp, -@@ -1912,8 +1915,11 @@ xfs_inactive( - * cache. Post-eof blocks must be freed, lest we end up with - * broken free space accounting. - */ -- if (xfs_can_free_eofblocks(ip, true)) -- xfs_free_eofblocks(mp, ip, false); -+ if (xfs_can_free_eofblocks(ip, true)) { -+ xfs_ilock(ip, XFS_IOLOCK_EXCL); -+ xfs_free_eofblocks(ip); -+ xfs_iunlock(ip, XFS_IOLOCK_EXCL); -+ } - - return; - } -diff --git a/fs/xfs/xfs_iomap.c b/fs/xfs/xfs_iomap.c -index e888961..3605624 100644 ---- a/fs/xfs/xfs_iomap.c -+++ b/fs/xfs/xfs_iomap.c -@@ -637,6 +637,11 @@ xfs_file_iomap_begin_delay( - goto out_unlock; - } - -+ /* -+ * Flag newly allocated delalloc blocks with IOMAP_F_NEW so we punch -+ * them out if the write happens to fail. -+ */ -+ iomap->flags = IOMAP_F_NEW; - trace_xfs_iomap_alloc(ip, offset, count, 0, &got); - done: - if (isnullstartblock(got.br_startblock)) -@@ -685,7 +690,7 @@ xfs_iomap_write_allocate( - int nres; - - if (whichfork == XFS_COW_FORK) -- flags |= XFS_BMAPI_COWFORK; -+ flags |= XFS_BMAPI_COWFORK | XFS_BMAPI_PREALLOC; - - /* - * Make sure that the dquots are there. -@@ -1061,7 +1066,8 @@ xfs_file_iomap_end_delalloc( - struct xfs_inode *ip, - loff_t offset, - loff_t length, -- ssize_t written) -+ ssize_t written, -+ struct iomap *iomap) - { - struct xfs_mount *mp = ip->i_mount; - xfs_fileoff_t start_fsb; -@@ -1080,14 +1086,14 @@ xfs_file_iomap_end_delalloc( - end_fsb = XFS_B_TO_FSB(mp, offset + length); - - /* -- * Trim back delalloc blocks if we didn't manage to write the whole -- * range reserved. -+ * Trim delalloc blocks if they were allocated by this write and we -+ * didn't manage to write the whole range. - * - * We don't need to care about racing delalloc as we hold i_mutex - * across the reserve/allocate/unreserve calls. If there are delalloc - * blocks in the range, they are ours. - */ -- if (start_fsb < end_fsb) { -+ if ((iomap->flags & IOMAP_F_NEW) && start_fsb < end_fsb) { - truncate_pagecache_range(VFS_I(ip), XFS_FSB_TO_B(mp, start_fsb), - XFS_FSB_TO_B(mp, end_fsb) - 1); - -@@ -1117,7 +1123,7 @@ xfs_file_iomap_end( - { - if ((flags & IOMAP_WRITE) && iomap->type == IOMAP_DELALLOC) - return xfs_file_iomap_end_delalloc(XFS_I(inode), offset, -- length, written); -+ length, written, iomap); - return 0; - } - -diff --git a/fs/xfs/xfs_mount.c b/fs/xfs/xfs_mount.c -index b341f10..13796f2 100644 ---- a/fs/xfs/xfs_mount.c -+++ b/fs/xfs/xfs_mount.c -@@ -502,8 +502,7 @@ STATIC void - xfs_set_inoalignment(xfs_mount_t *mp) - { - if (xfs_sb_version_hasalign(&mp->m_sb) && -- mp->m_sb.sb_inoalignmt >= -- XFS_B_TO_FSBT(mp, mp->m_inode_cluster_size)) -+ mp->m_sb.sb_inoalignmt >= xfs_icluster_size_fsb(mp)) - mp->m_inoalign_mask = mp->m_sb.sb_inoalignmt - 1; - else - mp->m_inoalign_mask = 0; -diff --git a/fs/xfs/xfs_mount.h b/fs/xfs/xfs_mount.h -index 819b80b..1bf878b 100644 ---- a/fs/xfs/xfs_mount.h -+++ b/fs/xfs/xfs_mount.h -@@ -140,6 +140,7 @@ typedef struct xfs_mount { - int m_fixedfsid[2]; /* unchanged for life of FS */ - uint m_dmevmask; /* DMI events for this FS */ - __uint64_t m_flags; /* global mount flags */ -+ bool m_inotbt_nores; /* no per-AG finobt resv. */ - int m_ialloc_inos; /* inodes in inode allocation */ - int m_ialloc_blks; /* blocks in inode allocation */ - int m_ialloc_min_blks;/* min blocks in sparse inode -diff --git a/fs/xfs/xfs_reflink.c b/fs/xfs/xfs_reflink.c -index 4d3f74e..2252f16 100644 ---- a/fs/xfs/xfs_reflink.c -+++ b/fs/xfs/xfs_reflink.c -@@ -82,11 +82,22 @@ - * mappings are a reservation against the free space in the filesystem; - * adjacent mappings can also be combined into fewer larger mappings. - * -+ * As an optimization, the CoW extent size hint (cowextsz) creates -+ * outsized aligned delalloc reservations in the hope of landing out of -+ * order nearby CoW writes in a single extent on disk, thereby reducing -+ * fragmentation and improving future performance. -+ * -+ * D: --RRRRRRSSSRRRRRRRR--- (data fork) -+ * C: ------DDDDDDD--------- (CoW fork) -+ * - * When dirty pages are being written out (typically in writepage), the -- * delalloc reservations are converted into real mappings by allocating -- * blocks and replacing the delalloc mapping with real ones. A delalloc -- * mapping can be replaced by several real ones if the free space is -- * fragmented. -+ * delalloc reservations are converted into unwritten mappings by -+ * allocating blocks and replacing the delalloc mapping with real ones. -+ * A delalloc mapping can be replaced by several unwritten ones if the -+ * free space is fragmented. -+ * -+ * D: --RRRRRRSSSRRRRRRRR--- -+ * C: ------UUUUUUU--------- - * - * We want to adapt the delalloc mechanism for copy-on-write, since the - * write paths are similar. The first two steps (creating the reservation -@@ -101,13 +112,29 @@ - * Block-aligned directio writes will use the same mechanism as buffered - * writes. - * -+ * Just prior to submitting the actual disk write requests, we convert -+ * the extents representing the range of the file actually being written -+ * (as opposed to extra pieces created for the cowextsize hint) to real -+ * extents. This will become important in the next step: -+ * -+ * D: --RRRRRRSSSRRRRRRRR--- -+ * C: ------UUrrUUU--------- -+ * - * CoW remapping must be done after the data block write completes, - * because we don't want to destroy the old data fork map until we're sure - * the new block has been written. Since the new mappings are kept in a - * separate fork, we can simply iterate these mappings to find the ones - * that cover the file blocks that we just CoW'd. For each extent, simply - * unmap the corresponding range in the data fork, map the new range into -- * the data fork, and remove the extent from the CoW fork. -+ * the data fork, and remove the extent from the CoW fork. Because of -+ * the presence of the cowextsize hint, however, we must be careful -+ * only to remap the blocks that we've actually written out -- we must -+ * never remap delalloc reservations nor CoW staging blocks that have -+ * yet to be written. This corresponds exactly to the real extents in -+ * the CoW fork: -+ * -+ * D: --RRRRRRrrSRRRRRRRR--- -+ * C: ------UU--UUU--------- - * - * Since the remapping operation can be applied to an arbitrary file - * range, we record the need for the remap step as a flag in the ioend -@@ -296,6 +323,65 @@ xfs_reflink_reserve_cow( - return 0; - } - -+/* Convert part of an unwritten CoW extent to a real one. */ -+STATIC int -+xfs_reflink_convert_cow_extent( -+ struct xfs_inode *ip, -+ struct xfs_bmbt_irec *imap, -+ xfs_fileoff_t offset_fsb, -+ xfs_filblks_t count_fsb, -+ struct xfs_defer_ops *dfops) -+{ -+ struct xfs_bmbt_irec irec = *imap; -+ xfs_fsblock_t first_block; -+ int nimaps = 1; -+ -+ if (imap->br_state == XFS_EXT_NORM) -+ return 0; -+ -+ xfs_trim_extent(&irec, offset_fsb, count_fsb); -+ trace_xfs_reflink_convert_cow(ip, &irec); -+ if (irec.br_blockcount == 0) -+ return 0; -+ return xfs_bmapi_write(NULL, ip, irec.br_startoff, irec.br_blockcount, -+ XFS_BMAPI_COWFORK | XFS_BMAPI_CONVERT, &first_block, -+ 0, &irec, &nimaps, dfops); -+} -+ -+/* Convert all of the unwritten CoW extents in a file's range to real ones. */ -+int -+xfs_reflink_convert_cow( -+ struct xfs_inode *ip, -+ xfs_off_t offset, -+ xfs_off_t count) -+{ -+ struct xfs_bmbt_irec got; -+ struct xfs_defer_ops dfops; -+ struct xfs_mount *mp = ip->i_mount; -+ struct xfs_ifork *ifp = XFS_IFORK_PTR(ip, XFS_COW_FORK); -+ xfs_fileoff_t offset_fsb = XFS_B_TO_FSBT(mp, offset); -+ xfs_fileoff_t end_fsb = XFS_B_TO_FSB(mp, offset + count); -+ xfs_extnum_t idx; -+ bool found; -+ int error = 0; -+ -+ xfs_ilock(ip, XFS_ILOCK_EXCL); -+ -+ /* Convert all the extents to real from unwritten. */ -+ for (found = xfs_iext_lookup_extent(ip, ifp, offset_fsb, &idx, &got); -+ found && got.br_startoff < end_fsb; -+ found = xfs_iext_get_extent(ifp, ++idx, &got)) { -+ error = xfs_reflink_convert_cow_extent(ip, &got, offset_fsb, -+ end_fsb - offset_fsb, &dfops); -+ if (error) -+ break; -+ } -+ -+ /* Finish up. */ -+ xfs_iunlock(ip, XFS_ILOCK_EXCL); -+ return error; -+} -+ - /* Allocate all CoW reservations covering a range of blocks in a file. */ - static int - __xfs_reflink_allocate_cow( -@@ -328,6 +414,7 @@ __xfs_reflink_allocate_cow( - goto out_unlock; - ASSERT(nimaps == 1); - -+ /* Make sure there's a CoW reservation for it. */ - error = xfs_reflink_reserve_cow(ip, &imap, &shared); - if (error) - goto out_trans_cancel; -@@ -337,14 +424,16 @@ __xfs_reflink_allocate_cow( - goto out_trans_cancel; - } - -+ /* Allocate the entire reservation as unwritten blocks. */ - xfs_trans_ijoin(tp, ip, 0); - error = xfs_bmapi_write(tp, ip, imap.br_startoff, imap.br_blockcount, -- XFS_BMAPI_COWFORK, &first_block, -+ XFS_BMAPI_COWFORK | XFS_BMAPI_PREALLOC, &first_block, - XFS_EXTENTADD_SPACE_RES(mp, XFS_DATA_FORK), - &imap, &nimaps, &dfops); - if (error) - goto out_trans_cancel; - -+ /* Finish up. */ - error = xfs_defer_finish(&tp, &dfops, NULL); - if (error) - goto out_trans_cancel; -@@ -389,11 +478,12 @@ xfs_reflink_allocate_cow_range( - if (error) { - trace_xfs_reflink_allocate_cow_range_error(ip, error, - _RET_IP_); -- break; -+ return error; - } - } - -- return error; -+ /* Convert the CoW extents to regular. */ -+ return xfs_reflink_convert_cow(ip, offset, count); - } - - /* -@@ -481,14 +571,18 @@ xfs_reflink_trim_irec_to_next_cow( - } - - /* -- * Cancel all pending CoW reservations for some block range of an inode. -+ * Cancel CoW reservations for some block range of an inode. -+ * -+ * If cancel_real is true this function cancels all COW fork extents for the -+ * inode; if cancel_real is false, real extents are not cleared. - */ - int - xfs_reflink_cancel_cow_blocks( - struct xfs_inode *ip, - struct xfs_trans **tpp, - xfs_fileoff_t offset_fsb, -- xfs_fileoff_t end_fsb) -+ xfs_fileoff_t end_fsb, -+ bool cancel_real) - { - struct xfs_ifork *ifp = XFS_IFORK_PTR(ip, XFS_COW_FORK); - struct xfs_bmbt_irec got, prev, del; -@@ -515,7 +609,7 @@ xfs_reflink_cancel_cow_blocks( - &idx, &got, &del); - if (error) - break; -- } else { -+ } else if (del.br_state == XFS_EXT_UNWRITTEN || cancel_real) { - xfs_trans_ijoin(*tpp, ip, 0); - xfs_defer_init(&dfops, &firstfsb); - -@@ -558,13 +652,17 @@ xfs_reflink_cancel_cow_blocks( - } - - /* -- * Cancel all pending CoW reservations for some byte range of an inode. -+ * Cancel CoW reservations for some byte range of an inode. -+ * -+ * If cancel_real is true this function cancels all COW fork extents for the -+ * inode; if cancel_real is false, real extents are not cleared. - */ - int - xfs_reflink_cancel_cow_range( - struct xfs_inode *ip, - xfs_off_t offset, -- xfs_off_t count) -+ xfs_off_t count, -+ bool cancel_real) - { - struct xfs_trans *tp; - xfs_fileoff_t offset_fsb; -@@ -590,7 +688,8 @@ xfs_reflink_cancel_cow_range( - xfs_trans_ijoin(tp, ip, 0); - - /* Scrape out the old CoW reservations */ -- error = xfs_reflink_cancel_cow_blocks(ip, &tp, offset_fsb, end_fsb); -+ error = xfs_reflink_cancel_cow_blocks(ip, &tp, offset_fsb, end_fsb, -+ cancel_real); - if (error) - goto out_cancel; - -@@ -669,6 +768,16 @@ xfs_reflink_end_cow( - - ASSERT(!isnullstartblock(got.br_startblock)); - -+ /* -+ * Don't remap unwritten extents; these are -+ * speculatively preallocated CoW extents that have been -+ * allocated but have not yet been involved in a write. -+ */ -+ if (got.br_state == XFS_EXT_UNWRITTEN) { -+ idx--; -+ goto next_extent; -+ } -+ - /* Unmap the old blocks in the data fork. */ - xfs_defer_init(&dfops, &firstfsb); - rlen = del.br_blockcount; -@@ -885,13 +994,14 @@ STATIC int - xfs_reflink_update_dest( - struct xfs_inode *dest, - xfs_off_t newlen, -- xfs_extlen_t cowextsize) -+ xfs_extlen_t cowextsize, -+ bool is_dedupe) - { - struct xfs_mount *mp = dest->i_mount; - struct xfs_trans *tp; - int error; - -- if (newlen <= i_size_read(VFS_I(dest)) && cowextsize == 0) -+ if (is_dedupe && newlen <= i_size_read(VFS_I(dest)) && cowextsize == 0) - return 0; - - error = xfs_trans_alloc(mp, &M_RES(mp)->tr_ichange, 0, 0, 0, &tp); -@@ -912,6 +1022,10 @@ xfs_reflink_update_dest( - dest->i_d.di_flags2 |= XFS_DIFLAG2_COWEXTSIZE; - } - -+ if (!is_dedupe) { -+ xfs_trans_ichgtime(tp, dest, -+ XFS_ICHGTIME_MOD | XFS_ICHGTIME_CHG); -+ } - xfs_trans_log_inode(tp, dest, XFS_ILOG_CORE); - - error = xfs_trans_commit(tp); -@@ -1428,7 +1542,8 @@ xfs_reflink_remap_range( - !(dest->i_d.di_flags2 & XFS_DIFLAG2_COWEXTSIZE)) - cowextsize = src->i_d.di_cowextsize; - -- ret = xfs_reflink_update_dest(dest, pos_out + len, cowextsize); -+ ret = xfs_reflink_update_dest(dest, pos_out + len, cowextsize, -+ is_dedupe); - - out_unlock: - xfs_iunlock(src, XFS_MMAPLOCK_EXCL); -@@ -1580,7 +1695,7 @@ xfs_reflink_clear_inode_flag( - * We didn't find any shared blocks so turn off the reflink flag. - * First, get rid of any leftover CoW mappings. - */ -- error = xfs_reflink_cancel_cow_blocks(ip, tpp, 0, NULLFILEOFF); -+ error = xfs_reflink_cancel_cow_blocks(ip, tpp, 0, NULLFILEOFF, true); - if (error) - return error; - -diff --git a/fs/xfs/xfs_reflink.h b/fs/xfs/xfs_reflink.h -index 97ea9b4..a57966f 100644 ---- a/fs/xfs/xfs_reflink.h -+++ b/fs/xfs/xfs_reflink.h -@@ -30,6 +30,8 @@ extern int xfs_reflink_reserve_cow(struct xfs_inode *ip, - struct xfs_bmbt_irec *imap, bool *shared); - extern int xfs_reflink_allocate_cow_range(struct xfs_inode *ip, - xfs_off_t offset, xfs_off_t count); -+extern int xfs_reflink_convert_cow(struct xfs_inode *ip, xfs_off_t offset, -+ xfs_off_t count); - extern bool xfs_reflink_find_cow_mapping(struct xfs_inode *ip, xfs_off_t offset, - struct xfs_bmbt_irec *imap, bool *need_alloc); - extern int xfs_reflink_trim_irec_to_next_cow(struct xfs_inode *ip, -@@ -37,9 +39,9 @@ extern int xfs_reflink_trim_irec_to_next_cow(struct xfs_inode *ip, - - extern int xfs_reflink_cancel_cow_blocks(struct xfs_inode *ip, - struct xfs_trans **tpp, xfs_fileoff_t offset_fsb, -- xfs_fileoff_t end_fsb); -+ xfs_fileoff_t end_fsb, bool cancel_real); - extern int xfs_reflink_cancel_cow_range(struct xfs_inode *ip, xfs_off_t offset, -- xfs_off_t count); -+ xfs_off_t count, bool cancel_real); - extern int xfs_reflink_end_cow(struct xfs_inode *ip, xfs_off_t offset, - xfs_off_t count); - extern int xfs_reflink_recover_cow(struct xfs_mount *mp); -diff --git a/fs/xfs/xfs_super.c b/fs/xfs/xfs_super.c -index ade4691..dbbd3f1 100644 ---- a/fs/xfs/xfs_super.c -+++ b/fs/xfs/xfs_super.c -@@ -948,7 +948,7 @@ xfs_fs_destroy_inode( - XFS_STATS_INC(ip->i_mount, vn_remove); - - if (xfs_is_reflink_inode(ip)) { -- error = xfs_reflink_cancel_cow_range(ip, 0, NULLFILEOFF); -+ error = xfs_reflink_cancel_cow_range(ip, 0, NULLFILEOFF, true); - if (error && !XFS_FORCED_SHUTDOWN(ip->i_mount)) - xfs_warn(ip->i_mount, - "Error %d while evicting CoW blocks for inode %llu.", -diff --git a/fs/xfs/xfs_trace.h b/fs/xfs/xfs_trace.h -index 0907752..828f383 100644 ---- a/fs/xfs/xfs_trace.h -+++ b/fs/xfs/xfs_trace.h -@@ -3183,6 +3183,7 @@ DECLARE_EVENT_CLASS(xfs_inode_irec_class, - __field(xfs_fileoff_t, lblk) - __field(xfs_extlen_t, len) - __field(xfs_fsblock_t, pblk) -+ __field(int, state) - ), - TP_fast_assign( - __entry->dev = VFS_I(ip)->i_sb->s_dev; -@@ -3190,13 +3191,15 @@ DECLARE_EVENT_CLASS(xfs_inode_irec_class, - __entry->lblk = irec->br_startoff; - __entry->len = irec->br_blockcount; - __entry->pblk = irec->br_startblock; -+ __entry->state = irec->br_state; - ), -- TP_printk("dev %d:%d ino 0x%llx lblk 0x%llx len 0x%x pblk %llu", -+ TP_printk("dev %d:%d ino 0x%llx lblk 0x%llx len 0x%x pblk %llu st %d", - MAJOR(__entry->dev), MINOR(__entry->dev), - __entry->ino, - __entry->lblk, - __entry->len, -- __entry->pblk) -+ __entry->pblk, -+ __entry->state) - ); - #define DEFINE_INODE_IREC_EVENT(name) \ - DEFINE_EVENT(xfs_inode_irec_class, name, \ -@@ -3345,11 +3348,12 @@ DEFINE_INODE_IREC_EVENT(xfs_reflink_trim_around_shared); - DEFINE_INODE_IREC_EVENT(xfs_reflink_cow_alloc); - DEFINE_INODE_IREC_EVENT(xfs_reflink_cow_found); - DEFINE_INODE_IREC_EVENT(xfs_reflink_cow_enospc); -+DEFINE_INODE_IREC_EVENT(xfs_reflink_convert_cow); - - DEFINE_RW_EVENT(xfs_reflink_reserve_cow); - DEFINE_RW_EVENT(xfs_reflink_allocate_cow_range); - --DEFINE_INODE_IREC_EVENT(xfs_reflink_bounce_dio_write); -+DEFINE_SIMPLE_IO_EVENT(xfs_reflink_bounce_dio_write); - DEFINE_IOMAP_EVENT(xfs_reflink_find_cow_mapping); - DEFINE_INODE_IREC_EVENT(xfs_reflink_trim_irec); - -diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h -index 01c0b9c..8c58db2 100644 ---- a/include/linux/kvm_host.h -+++ b/include/linux/kvm_host.h -@@ -162,8 +162,8 @@ int kvm_io_bus_read(struct kvm_vcpu *vcpu, enum kvm_bus bus_idx, gpa_t addr, - int len, void *val); - int kvm_io_bus_register_dev(struct kvm *kvm, enum kvm_bus bus_idx, gpa_t addr, - int len, struct kvm_io_device *dev); --int kvm_io_bus_unregister_dev(struct kvm *kvm, enum kvm_bus bus_idx, -- struct kvm_io_device *dev); -+void kvm_io_bus_unregister_dev(struct kvm *kvm, enum kvm_bus bus_idx, -+ struct kvm_io_device *dev); - struct kvm_io_device *kvm_io_bus_get_dev(struct kvm *kvm, enum kvm_bus bus_idx, - gpa_t addr); - -diff --git a/include/linux/memcontrol.h b/include/linux/memcontrol.h -index 2546988..8b35bdb 100644 ---- a/include/linux/memcontrol.h -+++ b/include/linux/memcontrol.h -@@ -739,6 +739,12 @@ static inline bool mem_cgroup_oom_synchronize(bool wait) - return false; - } - -+static inline void mem_cgroup_update_page_stat(struct page *page, -+ enum mem_cgroup_stat_index idx, -+ int nr) -+{ -+} -+ - static inline void mem_cgroup_inc_page_stat(struct page *page, - enum mem_cgroup_stat_index idx) - { -diff --git a/kernel/padata.c b/kernel/padata.c -index 7848f05..b4a3c0a 100644 ---- a/kernel/padata.c -+++ b/kernel/padata.c -@@ -190,19 +190,20 @@ static struct padata_priv *padata_get_next(struct parallel_data *pd) - - reorder = &next_queue->reorder; - -+ spin_lock(&reorder->lock); - if (!list_empty(&reorder->list)) { - padata = list_entry(reorder->list.next, - struct padata_priv, list); - -- spin_lock(&reorder->lock); - list_del_init(&padata->list); - atomic_dec(&pd->reorder_objects); -- spin_unlock(&reorder->lock); - - pd->processed++; - -+ spin_unlock(&reorder->lock); - goto out; - } -+ spin_unlock(&reorder->lock); - - if (__this_cpu_read(pd->pqueue->cpu_index) == next_queue->cpu_index) { - padata = ERR_PTR(-ENODATA); -diff --git a/lib/syscall.c b/lib/syscall.c -index 63239e0..a72cd09 100644 ---- a/lib/syscall.c -+++ b/lib/syscall.c -@@ -11,6 +11,7 @@ static int collect_syscall(struct task_struct *target, long *callno, - - if (!try_get_task_stack(target)) { - /* Task has no stack, so the task isn't in a syscall. */ -+ *sp = *pc = 0; - *callno = -1; - return 0; - } -diff --git a/mm/hugetlb.c b/mm/hugetlb.c -index b6adedb..65c36ac 100644 ---- a/mm/hugetlb.c -+++ b/mm/hugetlb.c -@@ -4471,6 +4471,7 @@ follow_huge_pmd(struct mm_struct *mm, unsigned long address, - { - struct page *page = NULL; - spinlock_t *ptl; -+ pte_t pte; - retry: - ptl = pmd_lockptr(mm, pmd); - spin_lock(ptl); -@@ -4480,12 +4481,13 @@ follow_huge_pmd(struct mm_struct *mm, unsigned long address, - */ - if (!pmd_huge(*pmd)) - goto out; -- if (pmd_present(*pmd)) { -+ pte = huge_ptep_get((pte_t *)pmd); -+ if (pte_present(pte)) { - page = pmd_page(*pmd) + ((address & ~PMD_MASK) >> PAGE_SHIFT); - if (flags & FOLL_GET) - get_page(page); - } else { -- if (is_hugetlb_entry_migration(huge_ptep_get((pte_t *)pmd))) { -+ if (is_hugetlb_entry_migration(pte)) { - spin_unlock(ptl); - __migration_entry_wait(mm, (pte_t *)pmd, ptl); - goto retry; -diff --git a/mm/rmap.c b/mm/rmap.c -index 1ef3640..cd37c1c 100644 ---- a/mm/rmap.c -+++ b/mm/rmap.c -@@ -1295,7 +1295,7 @@ void page_add_file_rmap(struct page *page, bool compound) - goto out; - } - __mod_node_page_state(page_pgdat(page), NR_FILE_MAPPED, nr); -- mem_cgroup_inc_page_stat(page, MEM_CGROUP_STAT_FILE_MAPPED); -+ mem_cgroup_update_page_stat(page, MEM_CGROUP_STAT_FILE_MAPPED, nr); - out: - unlock_page_memcg(page); - } -@@ -1335,7 +1335,7 @@ static void page_remove_file_rmap(struct page *page, bool compound) - * pte lock(a spinlock) is held, which implies preemption disabled. - */ - __mod_node_page_state(page_pgdat(page), NR_FILE_MAPPED, -nr); -- mem_cgroup_dec_page_stat(page, MEM_CGROUP_STAT_FILE_MAPPED); -+ mem_cgroup_update_page_stat(page, MEM_CGROUP_STAT_FILE_MAPPED, -nr); - - if (unlikely(PageMlocked(page))) - clear_page_mlock(page); -diff --git a/mm/workingset.c b/mm/workingset.c -index 33f6f4d..4c4f056 100644 ---- a/mm/workingset.c -+++ b/mm/workingset.c -@@ -492,7 +492,7 @@ static int __init workingset_init(void) - pr_info("workingset: timestamp_bits=%d max_order=%d bucket_order=%u\n", - timestamp_bits, max_order, bucket_order); - -- ret = list_lru_init_key(&workingset_shadow_nodes, &shadow_nodes_key); -+ ret = __list_lru_init(&workingset_shadow_nodes, true, &shadow_nodes_key); - if (ret) - goto err; - ret = register_shrinker(&workingset_shadow_shrinker); -diff --git a/net/ceph/messenger.c b/net/ceph/messenger.c -index 2efb335..25a30be 100644 ---- a/net/ceph/messenger.c -+++ b/net/ceph/messenger.c -@@ -7,6 +7,7 @@ - #include <linux/kthread.h> - #include <linux/net.h> - #include <linux/nsproxy.h> -+#include <linux/sched.h> - #include <linux/slab.h> - #include <linux/socket.h> - #include <linux/string.h> -@@ -469,11 +470,16 @@ static int ceph_tcp_connect(struct ceph_connection *con) - { - struct sockaddr_storage *paddr = &con->peer_addr.in_addr; - struct socket *sock; -+ unsigned int noio_flag; - int ret; - - BUG_ON(con->sock); -+ -+ /* sock_create_kern() allocates with GFP_KERNEL */ -+ noio_flag = memalloc_noio_save(); - ret = sock_create_kern(read_pnet(&con->msgr->net), paddr->ss_family, - SOCK_STREAM, IPPROTO_TCP, &sock); -+ memalloc_noio_restore(noio_flag); - if (ret) - return ret; - sock->sk->sk_allocation = GFP_NOFS; -diff --git a/sound/core/seq/seq_fifo.c b/sound/core/seq/seq_fifo.c -index 3f4efcb..3490d21 100644 ---- a/sound/core/seq/seq_fifo.c -+++ b/sound/core/seq/seq_fifo.c -@@ -265,6 +265,10 @@ int snd_seq_fifo_resize(struct snd_seq_fifo *f, int poolsize) - /* NOTE: overflow flag is not cleared */ - spin_unlock_irqrestore(&f->lock, flags); - -+ /* close the old pool and wait until all users are gone */ -+ snd_seq_pool_mark_closing(oldpool); -+ snd_use_lock_sync(&f->use_lock); -+ - /* release cells in old pool */ - for (cell = oldhead; cell; cell = next) { - next = cell->next; -diff --git a/sound/pci/hda/patch_realtek.c b/sound/pci/hda/patch_realtek.c -index 112caa2..bb1aad3 100644 ---- a/sound/pci/hda/patch_realtek.c -+++ b/sound/pci/hda/patch_realtek.c -@@ -4846,6 +4846,7 @@ enum { - ALC292_FIXUP_DISABLE_AAMIX, - ALC293_FIXUP_DISABLE_AAMIX_MULTIJACK, - ALC298_FIXUP_DELL1_MIC_NO_PRESENCE, -+ ALC298_FIXUP_DELL_AIO_MIC_NO_PRESENCE, - ALC275_FIXUP_DELL_XPS, - ALC256_FIXUP_DELL_XPS_13_HEADPHONE_NOISE, - ALC293_FIXUP_LENOVO_SPK_NOISE, -@@ -5446,6 +5447,15 @@ static const struct hda_fixup alc269_fixups[] = { - .chained = true, - .chain_id = ALC269_FIXUP_HEADSET_MODE - }, -+ [ALC298_FIXUP_DELL_AIO_MIC_NO_PRESENCE] = { -+ .type = HDA_FIXUP_PINS, -+ .v.pins = (const struct hda_pintbl[]) { -+ { 0x18, 0x01a1913c }, /* use as headset mic, without its own jack detect */ -+ { } -+ }, -+ .chained = true, -+ .chain_id = ALC269_FIXUP_HEADSET_MODE -+ }, - [ALC275_FIXUP_DELL_XPS] = { - .type = HDA_FIXUP_VERBS, - .v.verbs = (const struct hda_verb[]) { -@@ -5518,7 +5528,7 @@ static const struct hda_fixup alc269_fixups[] = { - .type = HDA_FIXUP_FUNC, - .v.func = alc298_fixup_speaker_volume, - .chained = true, -- .chain_id = ALC298_FIXUP_DELL1_MIC_NO_PRESENCE, -+ .chain_id = ALC298_FIXUP_DELL_AIO_MIC_NO_PRESENCE, - }, - [ALC256_FIXUP_DELL_INSPIRON_7559_SUBWOOFER] = { - .type = HDA_FIXUP_PINS, -diff --git a/sound/soc/atmel/atmel-classd.c b/sound/soc/atmel/atmel-classd.c -index 89ac5f5..7ae46c2 100644 ---- a/sound/soc/atmel/atmel-classd.c -+++ b/sound/soc/atmel/atmel-classd.c -@@ -349,7 +349,7 @@ static int atmel_classd_codec_dai_digital_mute(struct snd_soc_dai *codec_dai, - } - - #define CLASSD_ACLK_RATE_11M2896_MPY_8 (112896 * 100 * 8) --#define CLASSD_ACLK_RATE_12M288_MPY_8 (12228 * 1000 * 8) -+#define CLASSD_ACLK_RATE_12M288_MPY_8 (12288 * 1000 * 8) - - static struct { - int rate; -diff --git a/sound/soc/intel/skylake/skl-topology.c b/sound/soc/intel/skylake/skl-topology.c -index b5b1934..bef8a45 100644 ---- a/sound/soc/intel/skylake/skl-topology.c -+++ b/sound/soc/intel/skylake/skl-topology.c -@@ -448,7 +448,7 @@ static int skl_tplg_set_module_init_data(struct snd_soc_dapm_widget *w) - if (bc->set_params != SKL_PARAM_INIT) - continue; - -- mconfig->formats_config.caps = (u32 *)&bc->params; -+ mconfig->formats_config.caps = (u32 *)bc->params; - mconfig->formats_config.caps_size = bc->size; - - break; -diff --git a/virt/kvm/eventfd.c b/virt/kvm/eventfd.c -index a29786d..4d28a9d 100644 ---- a/virt/kvm/eventfd.c -+++ b/virt/kvm/eventfd.c -@@ -870,7 +870,8 @@ kvm_deassign_ioeventfd_idx(struct kvm *kvm, enum kvm_bus bus_idx, - continue; - - kvm_io_bus_unregister_dev(kvm, bus_idx, &p->dev); -- kvm->buses[bus_idx]->ioeventfd_count--; -+ if (kvm->buses[bus_idx]) -+ kvm->buses[bus_idx]->ioeventfd_count--; - ioeventfd_release(p); - ret = 0; - break; -diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c -index 7f9ee29..f4c6d4f 100644 ---- a/virt/kvm/kvm_main.c -+++ b/virt/kvm/kvm_main.c -@@ -720,8 +720,11 @@ static void kvm_destroy_vm(struct kvm *kvm) - list_del(&kvm->vm_list); - spin_unlock(&kvm_lock); - kvm_free_irq_routing(kvm); -- for (i = 0; i < KVM_NR_BUSES; i++) -- kvm_io_bus_destroy(kvm->buses[i]); -+ for (i = 0; i < KVM_NR_BUSES; i++) { -+ if (kvm->buses[i]) -+ kvm_io_bus_destroy(kvm->buses[i]); -+ kvm->buses[i] = NULL; -+ } - kvm_coalesced_mmio_free(kvm); - #if defined(CONFIG_MMU_NOTIFIER) && defined(KVM_ARCH_WANT_MMU_NOTIFIER) - mmu_notifier_unregister(&kvm->mmu_notifier, kvm->mm); -@@ -3463,6 +3466,8 @@ int kvm_io_bus_write(struct kvm_vcpu *vcpu, enum kvm_bus bus_idx, gpa_t addr, - }; - - bus = srcu_dereference(vcpu->kvm->buses[bus_idx], &vcpu->kvm->srcu); -+ if (!bus) -+ return -ENOMEM; - r = __kvm_io_bus_write(vcpu, bus, &range, val); - return r < 0 ? r : 0; - } -@@ -3480,6 +3485,8 @@ int kvm_io_bus_write_cookie(struct kvm_vcpu *vcpu, enum kvm_bus bus_idx, - }; - - bus = srcu_dereference(vcpu->kvm->buses[bus_idx], &vcpu->kvm->srcu); -+ if (!bus) -+ return -ENOMEM; - - /* First try the device referenced by cookie. */ - if ((cookie >= 0) && (cookie < bus->dev_count) && -@@ -3530,6 +3537,8 @@ int kvm_io_bus_read(struct kvm_vcpu *vcpu, enum kvm_bus bus_idx, gpa_t addr, - }; - - bus = srcu_dereference(vcpu->kvm->buses[bus_idx], &vcpu->kvm->srcu); -+ if (!bus) -+ return -ENOMEM; - r = __kvm_io_bus_read(vcpu, bus, &range, val); - return r < 0 ? r : 0; - } -@@ -3542,6 +3551,9 @@ int kvm_io_bus_register_dev(struct kvm *kvm, enum kvm_bus bus_idx, gpa_t addr, - struct kvm_io_bus *new_bus, *bus; - - bus = kvm->buses[bus_idx]; -+ if (!bus) -+ return -ENOMEM; -+ - /* exclude ioeventfd which is limited by maximum fd */ - if (bus->dev_count - bus->ioeventfd_count > NR_IOBUS_DEVS - 1) - return -ENOSPC; -@@ -3561,37 +3573,41 @@ int kvm_io_bus_register_dev(struct kvm *kvm, enum kvm_bus bus_idx, gpa_t addr, - } - - /* Caller must hold slots_lock. */ --int kvm_io_bus_unregister_dev(struct kvm *kvm, enum kvm_bus bus_idx, -- struct kvm_io_device *dev) -+void kvm_io_bus_unregister_dev(struct kvm *kvm, enum kvm_bus bus_idx, -+ struct kvm_io_device *dev) - { -- int i, r; -+ int i; - struct kvm_io_bus *new_bus, *bus; - - bus = kvm->buses[bus_idx]; -- r = -ENOENT; -+ if (!bus) -+ return; -+ - for (i = 0; i < bus->dev_count; i++) - if (bus->range[i].dev == dev) { -- r = 0; - break; - } - -- if (r) -- return r; -+ if (i == bus->dev_count) -+ return; - - new_bus = kmalloc(sizeof(*bus) + ((bus->dev_count - 1) * - sizeof(struct kvm_io_range)), GFP_KERNEL); -- if (!new_bus) -- return -ENOMEM; -+ if (!new_bus) { -+ pr_err("kvm: failed to shrink bus, removing it completely\n"); -+ goto broken; -+ } - - memcpy(new_bus, bus, sizeof(*bus) + i * sizeof(struct kvm_io_range)); - new_bus->dev_count--; - memcpy(new_bus->range + i, bus->range + i + 1, - (new_bus->dev_count - i) * sizeof(struct kvm_io_range)); - -+broken: - rcu_assign_pointer(kvm->buses[bus_idx], new_bus); - synchronize_srcu_expedited(&kvm->srcu); - kfree(bus); -- return r; -+ return; - } - - struct kvm_io_device *kvm_io_bus_get_dev(struct kvm *kvm, enum kvm_bus bus_idx, -@@ -3604,6 +3620,8 @@ struct kvm_io_device *kvm_io_bus_get_dev(struct kvm *kvm, enum kvm_bus bus_idx, - srcu_idx = srcu_read_lock(&kvm->srcu); - - bus = srcu_dereference(kvm->buses[bus_idx], &kvm->srcu); -+ if (!bus) -+ goto out_unlock; - - dev_idx = kvm_io_bus_get_first_dev(bus, addr, 1); - if (dev_idx < 0) diff --git a/4.9.21/0000_README b/4.9.22/0000_README index 40c05c0..0713033 100644 --- a/4.9.21/0000_README +++ b/4.9.22/0000_README @@ -2,14 +2,10 @@ README ----------------------------------------------------------------------------- Individual Patch Descriptions: ----------------------------------------------------------------------------- -Patch: 4420_grsecurity-3.1-4.9.21-201704091948.patch +Patch: 4420_grsecurity-3.1-4.9.22-201704120836.patch From: http://www.grsecurity.net Desc: hardened-sources base patch from upstream grsecurity -Patch: 1020_linux-4.9.21.patch -From: http://www.kernel.org -Desc: Linux 4.9.21 - Patch: 4425_grsec_remove_EI_PAX.patch From: Anthony G. Basile <blueness@gentoo.org> Desc: Remove EI_PAX option and force off diff --git a/4.9.21/4420_grsecurity-3.1-4.9.21-201704091948.patch b/4.9.22/4420_grsecurity-3.1-4.9.22-201704120836.patch index b1e5afd..17d7169 100644 --- a/4.9.21/4420_grsecurity-3.1-4.9.21-201704091948.patch +++ b/4.9.22/4420_grsecurity-3.1-4.9.22-201704120836.patch @@ -310,10 +310,10 @@ index 9b9c479..5a635ff 100644 A typical pattern in a Kbuild file looks like this: diff --git a/Documentation/kernel-parameters.txt b/Documentation/kernel-parameters.txt -index 65b05ba..725a42a 100644 +index a6fadef..fdf6042 100644 --- a/Documentation/kernel-parameters.txt +++ b/Documentation/kernel-parameters.txt -@@ -1426,6 +1426,12 @@ bytes respectively. Such letter suffixes can also be entirely omitted. +@@ -1436,6 +1436,12 @@ bytes respectively. Such letter suffixes can also be entirely omitted. [KNL] Should the hard-lockup detector generate backtraces on all cpus. Format: <integer> @@ -326,7 +326,7 @@ index 65b05ba..725a42a 100644 hashdist= [KNL,NUMA] Large hashes allocated during boot are distributed across NUMA nodes. Defaults on -@@ -2655,6 +2661,10 @@ bytes respectively. Such letter suffixes can also be entirely omitted. +@@ -2665,6 +2671,10 @@ bytes respectively. Such letter suffixes can also be entirely omitted. noexec=on: enable non-executable mappings (default) noexec=off: disable non-executable mappings @@ -337,7 +337,7 @@ index 65b05ba..725a42a 100644 nosmap [X86] Disable SMAP (Supervisor Mode Access Prevention) even if it is supported by processor. -@@ -2963,6 +2973,35 @@ bytes respectively. Such letter suffixes can also be entirely omitted. +@@ -2973,6 +2983,35 @@ bytes respectively. Such letter suffixes can also be entirely omitted. the specified number of seconds. This is to be used if your oopses keep scrolling off the screen. @@ -419,7 +419,7 @@ index 3d0ae15..84e5412 100644 cmd_syscalls = $(CONFIG_SHELL) $< $(CC) $(c_flags) $(missing_syscalls_flags) diff --git a/Makefile b/Makefile -index 1523557..8b62811 100644 +index 4bf4648..77b4b78 100644 --- a/Makefile +++ b/Makefile @@ -302,7 +302,9 @@ CONFIG_SHELL := $(shell if [ -x "$$BASH" ]; then echo $$BASH; \ @@ -2556,7 +2556,7 @@ index 5af0ed1..cea83883 100644 #define PSR_ENDIAN_MASK 0x00000200 /* Endianness state mask */ diff --git a/arch/arm/kernel/armksyms.c b/arch/arm/kernel/armksyms.c -index 7e45f69..2c047db 100644 +index 8e8d20c..fd4b52d 100644 --- a/arch/arm/kernel/armksyms.c +++ b/arch/arm/kernel/armksyms.c @@ -59,7 +59,7 @@ EXPORT_SYMBOL(arm_delay_ops); @@ -6230,7 +6230,7 @@ index 5c3f688..f8cc1b3 100644 # platform specific definitions include arch/mips/Kbuild.platforms diff --git a/arch/mips/Kconfig b/arch/mips/Kconfig -index b3c5bde..d6b5104 100644 +index 9a6e11b6..cda6b13 100644 --- a/arch/mips/Kconfig +++ b/arch/mips/Kconfig @@ -49,6 +49,7 @@ config MIPS @@ -7447,7 +7447,7 @@ index 4472a7f..c5905e6 100644 } /* Arrange for an interrupt in a short while */ diff --git a/arch/mips/kernel/traps.c b/arch/mips/kernel/traps.c -index 3905003..7c0cc88 100644 +index ec87ef9..dac8bcd 100644 --- a/arch/mips/kernel/traps.c +++ b/arch/mips/kernel/traps.c @@ -702,7 +702,18 @@ asmlinkage void do_ov(struct pt_regs *regs) @@ -10213,7 +10213,7 @@ index c4a93d6..4d2a9b4 100644 #endif /* __ASM_EXEC_H */ diff --git a/arch/s390/include/asm/uaccess.h b/arch/s390/include/asm/uaccess.h -index 52d7c87..577d292 100644 +index a7ef702..8c25ce4 100644 --- a/arch/s390/include/asm/uaccess.h +++ b/arch/s390/include/asm/uaccess.h @@ -59,6 +59,7 @@ static inline int __range_ok(unsigned long addr, unsigned long size) @@ -34097,7 +34097,7 @@ index 5b2cc88..bdbf72f 100644 return ret; diff --git a/arch/x86/kernel/reboot.c b/arch/x86/kernel/reboot.c -index e244c19..2cf3a26 100644 +index 067f981..5486a61 100644 --- a/arch/x86/kernel/reboot.c +++ b/arch/x86/kernel/reboot.c @@ -83,6 +83,11 @@ static int __init set_bios_reboot(const struct dmi_system_id *d) @@ -34152,7 +34152,7 @@ index e244c19..2cf3a26 100644 /* Acer */ { /* Handle reboot issue on Acer Aspire one */ -@@ -540,7 +567,7 @@ void __attribute__((weak)) mach_reboot_fixups(void) +@@ -556,7 +583,7 @@ void __attribute__((weak)) mach_reboot_fixups(void) * This means that this function can never return, it can misbehave * by not rebooting properly and hanging. */ @@ -34161,7 +34161,7 @@ index e244c19..2cf3a26 100644 { int i; int attempt = 0; -@@ -669,13 +696,13 @@ void native_machine_shutdown(void) +@@ -685,13 +712,13 @@ void native_machine_shutdown(void) #endif } @@ -34177,7 +34177,7 @@ index e244c19..2cf3a26 100644 { pr_notice("machine restart\n"); -@@ -684,7 +711,7 @@ static void native_machine_restart(char *__unused) +@@ -700,7 +727,7 @@ static void native_machine_restart(char *__unused) __machine_emergency_restart(0); } @@ -34186,7 +34186,7 @@ index e244c19..2cf3a26 100644 { /* Stop other cpus and apics */ machine_shutdown(); -@@ -694,7 +721,7 @@ static void native_machine_halt(void) +@@ -710,7 +737,7 @@ static void native_machine_halt(void) stop_this_cpu(NULL); } @@ -34195,7 +34195,7 @@ index e244c19..2cf3a26 100644 { if (pm_power_off) { if (!reboot_force) -@@ -703,6 +730,7 @@ static void native_machine_power_off(void) +@@ -719,6 +746,7 @@ static void native_machine_power_off(void) } /* A fallback in case there is no PM info available */ tboot_shutdown(TB_SHUTDOWN_HALT); @@ -46176,7 +46176,7 @@ index 56190d0..406f04c 100644 }; #endif diff --git a/drivers/acpi/button.c b/drivers/acpi/button.c -index e19f530..cc2d296 100644 +index 6d5a8c1..c763e4a 100644 --- a/drivers/acpi/button.c +++ b/drivers/acpi/button.c @@ -556,7 +556,7 @@ static int acpi_button_remove(struct acpi_device *device) @@ -46334,10 +46334,10 @@ index 7cfbda4..74f738c 100644 set_no_mwait, "Extensa 5220", { DMI_MATCH(DMI_BIOS_VENDOR, "Phoenix Technologies LTD"), diff --git a/drivers/acpi/sleep.c b/drivers/acpi/sleep.c -index 54abb26..6f7015f 100644 +index a4327af..d0ebae1 100644 --- a/drivers/acpi/sleep.c +++ b/drivers/acpi/sleep.c -@@ -154,7 +154,7 @@ static int __init init_nvs_nosave(const struct dmi_system_id *d) +@@ -160,7 +160,7 @@ static int __init init_nvs_nosave(const struct dmi_system_id *d) return 0; } @@ -46347,7 +46347,7 @@ index 54abb26..6f7015f 100644 .callback = init_old_suspend_ordering, .ident = "Abit KN9 (nForce4 variant)", diff --git a/drivers/acpi/sysfs.c b/drivers/acpi/sysfs.c -index 703c26e..47be990 100644 +index cf05ae9..8a581b4 100644 --- a/drivers/acpi/sysfs.c +++ b/drivers/acpi/sysfs.c @@ -227,7 +227,7 @@ module_param_cb(trace_method_name, ¶m_ops_trace_method, &trace_method_name, @@ -46391,7 +46391,7 @@ index 703c26e..47be990 100644 static void delete_gpe_attr_array(void) { -@@ -797,13 +797,13 @@ static void __exit interrupt_stats_exit(void) +@@ -853,13 +853,13 @@ static void __exit interrupt_stats_exit(void) } static ssize_t @@ -49305,7 +49305,7 @@ index d28922d..3c343d6 100644 if (cmd != SIOCWANDEV) diff --git a/drivers/char/random.c b/drivers/char/random.c -index d6876d5..a326099 100644 +index 08d1dd5..0b23b06 100644 --- a/drivers/char/random.c +++ b/drivers/char/random.c @@ -292,9 +292,6 @@ @@ -53169,10 +53169,10 @@ index 0e8c4d9..f9d98b7 100644 { struct mtk_hdmi *hdmi = hdmi_ctx_from_conn(conn); diff --git a/drivers/gpu/drm/mga/mga_drv.c b/drivers/gpu/drm/mga/mga_drv.c -index 25b2a1a..58ae09b 100644 +index 63ba0699..35e349f 100644 --- a/drivers/gpu/drm/mga/mga_drv.c +++ b/drivers/gpu/drm/mga/mga_drv.c -@@ -92,7 +92,10 @@ static struct pci_driver mga_pci_driver = { +@@ -89,7 +89,10 @@ static struct pci_driver mga_pci_driver = { static int __init mga_init(void) { @@ -55267,10 +55267,10 @@ index 5f962bf..b095fc5 100644 /* copy over all the bus versions */ if (dev->bus && dev->bus->pm) { diff --git a/drivers/hid/hid-core.c b/drivers/hid/hid-core.c -index 2b89c70..cc99025 100644 +index a5dd7e6..de73abd 100644 --- a/drivers/hid/hid-core.c +++ b/drivers/hid/hid-core.c -@@ -2645,7 +2645,7 @@ EXPORT_SYMBOL_GPL(hid_ignore); +@@ -2643,7 +2643,7 @@ EXPORT_SYMBOL_GPL(hid_ignore); int hid_add_device(struct hid_device *hdev) { @@ -55279,7 +55279,7 @@ index 2b89c70..cc99025 100644 int ret; if (WARN_ON(hdev->status & HID_STAT_ADDED)) -@@ -2689,7 +2689,7 @@ int hid_add_device(struct hid_device *hdev) +@@ -2687,7 +2687,7 @@ int hid_add_device(struct hid_device *hdev) /* XXX hack, any other cleaner solution after the driver core * is converted to allow more than 20 bytes as the device name? */ dev_set_name(&hdev->dev, "%04X:%04X:%04X.%04X", hdev->bus, @@ -55929,10 +55929,10 @@ index d127ace..6ee866f 100644 int i, j = 1; diff --git a/drivers/idle/intel_idle.c b/drivers/idle/intel_idle.c -index 4466a2f..6771887 100644 +index 5ded9b2..8410378 100644 --- a/drivers/idle/intel_idle.c +++ b/drivers/idle/intel_idle.c -@@ -1208,36 +1208,46 @@ static void bxt_idle_state_table_update(void) +@@ -1257,36 +1257,46 @@ static void bxt_idle_state_table_update(void) rdmsrl(MSR_PKGC6_IRTL, msr); usec = irtl_2_usec(msr); if (usec) { @@ -55989,7 +55989,7 @@ index 4466a2f..6771887 100644 } } -@@ -1280,8 +1290,10 @@ static void sklh_idle_state_table_update(void) +@@ -1329,8 +1339,10 @@ static void sklh_idle_state_table_update(void) return; } @@ -61799,7 +61799,7 @@ index e477af8..a5b1fce 100644 DMEMIT("%u ", test_bit(MPATHF_QUEUE_IF_NO_PATH, &m->flags) + (m->pg_init_retries > 0) * 2 + diff --git a/drivers/md/dm-raid.c b/drivers/md/dm-raid.c -index 15daa36..d19c604 100644 +index ee75e35..7e87ea2 100644 --- a/drivers/md/dm-raid.c +++ b/drivers/md/dm-raid.c @@ -3196,7 +3196,7 @@ static void raid_status(struct dm_target *ti, status_type_t type, @@ -68503,10 +68503,10 @@ index e2509bb..8357fef 100644 if (ring->free_tx_desc == mlx4_en_recycle_tx_desc) return done < budget; diff --git a/drivers/net/ethernet/mellanox/mlx4/main.c b/drivers/net/ethernet/mellanox/mlx4/main.c -index 75d07fa..d766d8e 100644 +index b2ca8a6..4f1c28e 100644 --- a/drivers/net/ethernet/mellanox/mlx4/main.c +++ b/drivers/net/ethernet/mellanox/mlx4/main.c -@@ -4069,7 +4069,7 @@ static const struct pci_device_id mlx4_pci_table[] = { +@@ -4071,7 +4071,7 @@ static const struct pci_device_id mlx4_pci_table[] = { MODULE_DEVICE_TABLE(pci, mlx4_pci_table); static pci_ers_result_t mlx4_pci_err_detected(struct pci_dev *pdev, @@ -75679,10 +75679,10 @@ index a66be13..124be13 100644 /* * Early Macbook Pros with switchable graphics use nvidia diff --git a/drivers/platform/x86/asus-wmi.c b/drivers/platform/x86/asus-wmi.c -index ce6ca31..3e69fbe 100644 +index 8499d3a..5e2f9e4 100644 --- a/drivers/platform/x86/asus-wmi.c +++ b/drivers/platform/x86/asus-wmi.c -@@ -1872,6 +1872,10 @@ static int show_dsts(struct seq_file *m, void *data) +@@ -1900,6 +1900,10 @@ static int show_dsts(struct seq_file *m, void *data) int err; u32 retval = -1; @@ -75693,7 +75693,7 @@ index ce6ca31..3e69fbe 100644 err = asus_wmi_get_devstate(asus, asus->debug.dev_id, &retval); if (err < 0) -@@ -1888,6 +1892,10 @@ static int show_devs(struct seq_file *m, void *data) +@@ -1916,6 +1920,10 @@ static int show_devs(struct seq_file *m, void *data) int err; u32 retval = -1; @@ -75704,7 +75704,7 @@ index ce6ca31..3e69fbe 100644 err = asus_wmi_set_devstate(asus->debug.dev_id, asus->debug.ctrl_param, &retval); -@@ -1912,6 +1920,10 @@ static int show_call(struct seq_file *m, void *data) +@@ -1940,6 +1948,10 @@ static int show_call(struct seq_file *m, void *data) union acpi_object *obj; acpi_status status; @@ -105880,10 +105880,10 @@ index 5d456eb..6b33055 100644 } diff --git a/fs/cifs/smb2pdu.c b/fs/cifs/smb2pdu.c -index 8745722..407a2a7 100644 +index bdd3292..eed47d8 100644 --- a/fs/cifs/smb2pdu.c +++ b/fs/cifs/smb2pdu.c -@@ -2613,8 +2613,7 @@ SMB2_query_directory(const unsigned int xid, struct cifs_tcon *tcon, +@@ -2617,8 +2617,7 @@ SMB2_query_directory(const unsigned int xid, struct cifs_tcon *tcon, default: cifs_dbg(VFS, "info level %u isn't supported\n", srch_inf->info_level); @@ -120913,10 +120913,10 @@ index d3ed817..d6038021 100644 } putname(tmp); diff --git a/fs/orangefs/super.c b/fs/orangefs/super.c -index 67c2435..a1f43ba 100644 +index cd261c8..7a1c069 100644 --- a/fs/orangefs/super.c +++ b/fs/orangefs/super.c -@@ -573,10 +573,12 @@ void orangefs_kill_sb(struct super_block *sb) +@@ -578,10 +578,12 @@ void orangefs_kill_sb(struct super_block *sb) int orangefs_inode_cache_initialize(void) { @@ -136875,7 +136875,7 @@ index 10e449c..f2a0d40 100644 /** * drm_connector_helper_add - sets the helper vtable for a connector diff --git a/include/drm/i915_pciids.h b/include/drm/i915_pciids.h -index 0d5f426..ac3b055 100644 +index 61766a4..2f88e2b 100644 --- a/include/drm/i915_pciids.h +++ b/include/drm/i915_pciids.h @@ -37,7 +37,7 @@ @@ -142589,10 +142589,10 @@ index 55107a8..eb06178 100644 extern bool qid_valid(struct kqid qid); diff --git a/include/linux/random.h b/include/linux/random.h -index 7bd2403..211378b 100644 +index 16ab429..147be00 100644 --- a/include/linux/random.h +++ b/include/linux/random.h -@@ -63,6 +63,11 @@ void prandom_seed_full_state(struct rnd_state __percpu *pcpu_state); +@@ -62,6 +62,11 @@ void prandom_seed_full_state(struct rnd_state __percpu *pcpu_state); #define prandom_init_once(pcpu_state) \ DO_ONCE(prandom_seed_full_state, (pcpu_state)) @@ -142604,7 +142604,7 @@ index 7bd2403..211378b 100644 /** * prandom_u32_max - returns a pseudo-random number in interval [0, ep_ro) * @ep_ro: right open interval endpoint -@@ -75,7 +80,7 @@ void prandom_seed_full_state(struct rnd_state __percpu *pcpu_state); +@@ -74,7 +79,7 @@ void prandom_seed_full_state(struct rnd_state __percpu *pcpu_state); * * Returns: pseudo-random number in interval [0, ep_ro) */ @@ -147446,7 +147446,7 @@ index b32ad7d..05f6420 100644 next_state = Reset; return 0; diff --git a/init/main.c b/init/main.c -index 2858be7..75fba2e 100644 +index ae3996a..5a222fc 100644 --- a/init/main.c +++ b/init/main.c @@ -11,6 +11,10 @@ @@ -147548,7 +147548,7 @@ index 2858be7..75fba2e 100644 add_latent_entropy(); return ret; -@@ -894,8 +940,8 @@ static int run_init_process(const char *init_filename) +@@ -893,8 +939,8 @@ static int run_init_process(const char *init_filename) { argv_init[0] = init_filename; return do_execve(getname_kernel(init_filename), @@ -147559,7 +147559,7 @@ index 2858be7..75fba2e 100644 } static int try_to_run_init_process(const char *init_filename) -@@ -912,6 +958,10 @@ static int try_to_run_init_process(const char *init_filename) +@@ -911,6 +957,10 @@ static int try_to_run_init_process(const char *init_filename) return ret; } @@ -147570,7 +147570,7 @@ index 2858be7..75fba2e 100644 static noinline void __init kernel_init_freeable(void); #ifdef CONFIG_DEBUG_RODATA -@@ -960,6 +1010,11 @@ static int __ref kernel_init(void *unused) +@@ -959,6 +1009,11 @@ static int __ref kernel_init(void *unused) ramdisk_execute_command, ret); } @@ -147582,7 +147582,7 @@ index 2858be7..75fba2e 100644 /* * We try each of these until one succeeds. * -@@ -1017,7 +1072,7 @@ static noinline void __init kernel_init_freeable(void) +@@ -1016,7 +1071,7 @@ static noinline void __init kernel_init_freeable(void) do_basic_setup(); /* Open the /dev/console on the rootfs, this should never fail */ @@ -147591,7 +147591,7 @@ index 2858be7..75fba2e 100644 pr_err("Warning: unable to open an initial console.\n"); (void) sys_dup(0); -@@ -1030,11 +1085,13 @@ static noinline void __init kernel_init_freeable(void) +@@ -1029,11 +1084,13 @@ static noinline void __init kernel_init_freeable(void) if (!ramdisk_execute_command) ramdisk_execute_command = "/init"; @@ -151902,10 +151902,10 @@ index 2dbccf2..f98676c 100644 } diff --git a/kernel/ptrace.c b/kernel/ptrace.c -index 49ba7c1..4936c74 100644 +index a5caece..a072a59 100644 --- a/kernel/ptrace.c +++ b/kernel/ptrace.c -@@ -244,7 +244,35 @@ static int ptrace_check_attach(struct task_struct *child, bool ignore_state) +@@ -250,7 +250,35 @@ static int ptrace_check_attach(struct task_struct *child, bool ignore_state) return ret; } @@ -151942,7 +151942,7 @@ index 49ba7c1..4936c74 100644 { if (mode & PTRACE_MODE_NOAUDIT) return has_ns_capability_noaudit(current, ns, CAP_SYS_PTRACE); -@@ -259,6 +287,7 @@ static int __ptrace_may_access(struct task_struct *task, unsigned int mode) +@@ -265,6 +293,7 @@ static int __ptrace_may_access(struct task_struct *task, unsigned int mode) struct mm_struct *mm; kuid_t caller_uid; kgid_t caller_gid; @@ -151950,7 +151950,7 @@ index 49ba7c1..4936c74 100644 if (!(mode & PTRACE_MODE_FSCREDS) == !(mode & PTRACE_MODE_REALCREDS)) { WARN(1, "denying ptrace access check without PTRACE_MODE_*CREDS\n"); -@@ -301,16 +330,27 @@ static int __ptrace_may_access(struct task_struct *task, unsigned int mode) +@@ -307,16 +336,27 @@ static int __ptrace_may_access(struct task_struct *task, unsigned int mode) gid_eq(caller_gid, tcred->sgid) && gid_eq(caller_gid, tcred->gid)) goto ok; @@ -151981,7 +151981,7 @@ index 49ba7c1..4936c74 100644 return -EPERM; return security_ptrace_access_check(task, mode); -@@ -571,7 +611,7 @@ int ptrace_readdata(struct task_struct *tsk, unsigned long src, char __user *dst +@@ -577,7 +617,7 @@ int ptrace_readdata(struct task_struct *tsk, unsigned long src, char __user *dst break; return -EIO; } @@ -151990,7 +151990,7 @@ index 49ba7c1..4936c74 100644 return -EFAULT; copied += retval; src += retval; -@@ -873,7 +913,7 @@ int ptrace_request(struct task_struct *child, long request, +@@ -879,7 +919,7 @@ int ptrace_request(struct task_struct *child, long request, bool seized = child->ptrace & PT_SEIZED; int ret = -EIO; siginfo_t siginfo, *si; @@ -151999,7 +151999,7 @@ index 49ba7c1..4936c74 100644 unsigned long __user *datalp = datavp; unsigned long flags; -@@ -1124,14 +1164,21 @@ SYSCALL_DEFINE4(ptrace, long, request, long, pid, unsigned long, addr, +@@ -1130,14 +1170,21 @@ SYSCALL_DEFINE4(ptrace, long, request, long, pid, unsigned long, addr, goto out; } @@ -152022,7 +152022,7 @@ index 49ba7c1..4936c74 100644 goto out_put_task_struct; } -@@ -1159,7 +1206,7 @@ int generic_ptrace_peekdata(struct task_struct *tsk, unsigned long addr, +@@ -1165,7 +1212,7 @@ int generic_ptrace_peekdata(struct task_struct *tsk, unsigned long addr, copied = ptrace_access_vm(tsk, addr, &tmp, sizeof(tmp), FOLL_FORCE); if (copied != sizeof(tmp)) return -EIO; @@ -152031,7 +152031,7 @@ index 49ba7c1..4936c74 100644 } int generic_ptrace_pokedata(struct task_struct *tsk, unsigned long addr, -@@ -1255,7 +1302,7 @@ int compat_ptrace_request(struct task_struct *child, compat_long_t request, +@@ -1261,7 +1308,7 @@ int compat_ptrace_request(struct task_struct *child, compat_long_t request, } COMPAT_SYSCALL_DEFINE4(ptrace, compat_long_t, request, compat_long_t, pid, @@ -152040,7 +152040,7 @@ index 49ba7c1..4936c74 100644 { struct task_struct *child; long ret; -@@ -1271,14 +1318,21 @@ COMPAT_SYSCALL_DEFINE4(ptrace, compat_long_t, request, compat_long_t, pid, +@@ -1277,14 +1324,21 @@ COMPAT_SYSCALL_DEFINE4(ptrace, compat_long_t, request, compat_long_t, pid, goto out; } @@ -154574,7 +154574,7 @@ index da87b3c..2f63cc9 100644 /* make curr_ret_stack visible before we add the ret_stack */ smp_wmb(); diff --git a/kernel/trace/ring_buffer.c b/kernel/trace/ring_buffer.c -index 9c14373..5ddd763 100644 +index f30847a..9e62b08 100644 --- a/kernel/trace/ring_buffer.c +++ b/kernel/trace/ring_buffer.c @@ -296,9 +296,9 @@ struct buffer_data_page { @@ -158157,7 +158157,7 @@ index cbb1e5e..e8b4bc4 100644 +} +EXPORT_SYMBOL(__access_ok); diff --git a/mm/mempolicy.c b/mm/mempolicy.c -index f757047..47a173d 100644 +index 23471526..5cb985f 100644 --- a/mm/mempolicy.c +++ b/mm/mempolicy.c @@ -732,6 +732,10 @@ static int mbind_range(struct mm_struct *mm, unsigned long start, @@ -160004,7 +160004,7 @@ index 439cc63..d342d3b 100644 struct bdi_writeback *wb = dtc->wb; unsigned long write_bw = wb->avg_write_bandwidth; diff --git a/mm/page_alloc.c b/mm/page_alloc.c -index 1460e6a..154adc1f 100644 +index e5b159b..90fc606 100644 --- a/mm/page_alloc.c +++ b/mm/page_alloc.c @@ -64,6 +64,7 @@ @@ -168490,7 +168490,7 @@ index 34c2add..322c6a3 100644 /* number of interfaces with corresponding FIF_ flags */ int fif_fcsfail, fif_plcpfail, fif_control, fif_other_bss, fif_pspoll, diff --git a/net/mac80211/iface.c b/net/mac80211/iface.c -index 638ec07..706e042 100644 +index 8d7747e..d12b9be 100644 --- a/net/mac80211/iface.c +++ b/net/mac80211/iface.c @@ -560,7 +560,7 @@ int ieee80211_do_open(struct wireless_dev *wdev, bool coming_up) @@ -168520,7 +168520,7 @@ index 638ec07..706e042 100644 if (hw_reconf_flags) ieee80211_hw_config(local, hw_reconf_flags); -@@ -756,7 +756,7 @@ int ieee80211_do_open(struct wireless_dev *wdev, bool coming_up) +@@ -757,7 +757,7 @@ int ieee80211_do_open(struct wireless_dev *wdev, bool coming_up) err_del_interface: drv_remove_interface(local, sdata); err_stop: @@ -168529,7 +168529,7 @@ index 638ec07..706e042 100644 drv_stop(local); err_del_bss: sdata->bss = NULL; -@@ -925,7 +925,7 @@ static void ieee80211_do_stop(struct ieee80211_sub_if_data *sdata, +@@ -926,7 +926,7 @@ static void ieee80211_do_stop(struct ieee80211_sub_if_data *sdata, } if (going_down) @@ -168538,7 +168538,7 @@ index 638ec07..706e042 100644 switch (sdata->vif.type) { case NL80211_IFTYPE_AP_VLAN: -@@ -1007,7 +1007,7 @@ static void ieee80211_do_stop(struct ieee80211_sub_if_data *sdata, +@@ -1008,7 +1008,7 @@ static void ieee80211_do_stop(struct ieee80211_sub_if_data *sdata, spin_unlock_bh(&fq->lock); } @@ -168547,7 +168547,7 @@ index 638ec07..706e042 100644 ieee80211_clear_tx_pending(local); /* -@@ -1050,7 +1050,7 @@ static void ieee80211_do_stop(struct ieee80211_sub_if_data *sdata, +@@ -1051,7 +1051,7 @@ static void ieee80211_do_stop(struct ieee80211_sub_if_data *sdata, if (cancel_scan) flush_delayed_work(&local->scan_work); @@ -168556,7 +168556,7 @@ index 638ec07..706e042 100644 ieee80211_stop_device(local); /* no reconfiguring after stop! */ -@@ -1061,7 +1061,7 @@ static void ieee80211_do_stop(struct ieee80211_sub_if_data *sdata, +@@ -1062,7 +1062,7 @@ static void ieee80211_do_stop(struct ieee80211_sub_if_data *sdata, ieee80211_configure_filter(local); ieee80211_hw_config(local, hw_reconf_flags); @@ -168565,7 +168565,7 @@ index 638ec07..706e042 100644 ieee80211_add_virtual_monitor(local); } -@@ -1948,8 +1948,8 @@ void ieee80211_remove_interfaces(struct ieee80211_local *local) +@@ -1949,8 +1949,8 @@ void ieee80211_remove_interfaces(struct ieee80211_local *local) */ cfg80211_shutdown_all_interfaces(local->hw.wiphy); @@ -171088,7 +171088,7 @@ index 9d88c62..53396b6 100644 #define gssx_dec_release_handle NULL #define gssx_enc_get_mic NULL diff --git a/net/sunrpc/auth_gss/svcauth_gss.c b/net/sunrpc/auth_gss/svcauth_gss.c -index 6fdffde..41a4b00 100644 +index 1530825..1e2fe0d50 100644 --- a/net/sunrpc/auth_gss/svcauth_gss.c +++ b/net/sunrpc/auth_gss/svcauth_gss.c @@ -1149,7 +1149,7 @@ static int gss_proxy_save_rsc(struct cache_detail *cd, @@ -171340,7 +171340,7 @@ index 2ecb994..5afb49f 100644 return -ENOMEM; diff --git a/net/sunrpc/svc.c b/net/sunrpc/svc.c -index 7c8070e..a3dd59d 100644 +index 75f290b..6ca88e9 100644 --- a/net/sunrpc/svc.c +++ b/net/sunrpc/svc.c @@ -50,7 +50,7 @@ EXPORT_SYMBOL_GPL(svc_pool_map); @@ -171361,7 +171361,7 @@ index 7c8070e..a3dd59d 100644 { int *ip = (int *)kp->arg; -@@ -1183,7 +1183,9 @@ svc_process_common(struct svc_rqst *rqstp, struct kvec *argv, struct kvec *resv) +@@ -1182,7 +1182,9 @@ svc_process_common(struct svc_rqst *rqstp, struct kvec *argv, struct kvec *resv) svc_putnl(resv, RPC_SUCCESS); /* Bump per-procedure stats counter */ diff --git a/4.9.21/4425_grsec_remove_EI_PAX.patch b/4.9.22/4425_grsec_remove_EI_PAX.patch index 594598a..594598a 100644 --- a/4.9.21/4425_grsec_remove_EI_PAX.patch +++ b/4.9.22/4425_grsec_remove_EI_PAX.patch diff --git a/4.9.21/4426_default_XATTR_PAX_FLAGS.patch b/4.9.22/4426_default_XATTR_PAX_FLAGS.patch index f7e97b5..f7e97b5 100644 --- a/4.9.21/4426_default_XATTR_PAX_FLAGS.patch +++ b/4.9.22/4426_default_XATTR_PAX_FLAGS.patch diff --git a/4.9.21/4427_force_XATTR_PAX_tmpfs.patch b/4.9.22/4427_force_XATTR_PAX_tmpfs.patch index 3871139..3871139 100644 --- a/4.9.21/4427_force_XATTR_PAX_tmpfs.patch +++ b/4.9.22/4427_force_XATTR_PAX_tmpfs.patch diff --git a/4.9.21/4430_grsec-remove-localversion-grsec.patch b/4.9.22/4430_grsec-remove-localversion-grsec.patch index 31cf878..31cf878 100644 --- a/4.9.21/4430_grsec-remove-localversion-grsec.patch +++ b/4.9.22/4430_grsec-remove-localversion-grsec.patch diff --git a/4.9.21/4435_grsec-mute-warnings.patch b/4.9.22/4435_grsec-mute-warnings.patch index 8929222..8929222 100644 --- a/4.9.21/4435_grsec-mute-warnings.patch +++ b/4.9.22/4435_grsec-mute-warnings.patch diff --git a/4.9.21/4440_grsec-remove-protected-paths.patch b/4.9.22/4440_grsec-remove-protected-paths.patch index 741546d..741546d 100644 --- a/4.9.21/4440_grsec-remove-protected-paths.patch +++ b/4.9.22/4440_grsec-remove-protected-paths.patch diff --git a/4.9.21/4450_grsec-kconfig-default-gids.patch b/4.9.22/4450_grsec-kconfig-default-gids.patch index cee6e27..cee6e27 100644 --- a/4.9.21/4450_grsec-kconfig-default-gids.patch +++ b/4.9.22/4450_grsec-kconfig-default-gids.patch diff --git a/4.9.21/4465_selinux-avc_audit-log-curr_ip.patch b/4.9.22/4465_selinux-avc_audit-log-curr_ip.patch index 06a5294..06a5294 100644 --- a/4.9.21/4465_selinux-avc_audit-log-curr_ip.patch +++ b/4.9.22/4465_selinux-avc_audit-log-curr_ip.patch diff --git a/4.9.21/4470_disable-compat_vdso.patch b/4.9.22/4470_disable-compat_vdso.patch index a1401d8..a1401d8 100644 --- a/4.9.21/4470_disable-compat_vdso.patch +++ b/4.9.22/4470_disable-compat_vdso.patch diff --git a/4.9.21/4475_emutramp_default_on.patch b/4.9.22/4475_emutramp_default_on.patch index feb8c7b..feb8c7b 100644 --- a/4.9.21/4475_emutramp_default_on.patch +++ b/4.9.22/4475_emutramp_default_on.patch |