diff options
author | Mike Pagano <mpagano@gentoo.org> | 2022-01-27 06:36:55 -0500 |
---|---|---|
committer | Mike Pagano <mpagano@gentoo.org> | 2022-01-27 06:36:55 -0500 |
commit | 23b5b031e20c79ab7455e7e41e9eaf0641343aef (patch) | |
tree | 9d68afb89013aa02274896b646b1349b66f9d22d | |
parent | Linux patch 5.15.16 (diff) | |
download | linux-patches-23b5b031e20c79ab7455e7e41e9eaf0641343aef.tar.gz linux-patches-23b5b031e20c79ab7455e7e41e9eaf0641343aef.tar.bz2 linux-patches-23b5b031e20c79ab7455e7e41e9eaf0641343aef.zip |
Linux patch 5.15.17
Signed-off-by: Mike Pagano <mpagano@gentoo.org>
-rw-r--r-- | 0000_README | 4 | ||||
-rw-r--r-- | 1016_linux-5.15.17.patch | 34359 |
2 files changed, 34363 insertions, 0 deletions
diff --git a/0000_README b/0000_README index 7901341f..c0a48663 100644 --- a/0000_README +++ b/0000_README @@ -107,6 +107,10 @@ Patch: 1015_linux-5.15.16.patch From: http://www.kernel.org Desc: Linux 5.15.16 +Patch: 1016_linux-5.15.17.patch +From: http://www.kernel.org +Desc: Linux 5.15.17 + Patch: 1500_XATTR_USER_PREFIX.patch From: https://bugs.gentoo.org/show_bug.cgi?id=470644 Desc: Support for namespace user.pax.* on tmpfs. diff --git a/1016_linux-5.15.17.patch b/1016_linux-5.15.17.patch new file mode 100644 index 00000000..43b919c4 --- /dev/null +++ b/1016_linux-5.15.17.patch @@ -0,0 +1,34359 @@ +diff --git a/Documentation/admin-guide/cifs/usage.rst b/Documentation/admin-guide/cifs/usage.rst +index f170d88202588..3766bf8a1c20e 100644 +--- a/Documentation/admin-guide/cifs/usage.rst ++++ b/Documentation/admin-guide/cifs/usage.rst +@@ -734,10 +734,9 @@ SecurityFlags Flags which control security negotiation and + using weaker password hashes is 0x37037 (lanman, + plaintext, ntlm, ntlmv2, signing allowed). Some + SecurityFlags require the corresponding menuconfig +- options to be enabled (lanman and plaintext require +- CONFIG_CIFS_WEAK_PW_HASH for example). Enabling +- plaintext authentication currently requires also +- enabling lanman authentication in the security flags ++ options to be enabled. Enabling plaintext ++ authentication currently requires also enabling ++ lanman authentication in the security flags + because the cifs module only supports sending + laintext passwords using the older lanman dialect + form of the session setup SMB. (e.g. for authentication +diff --git a/Documentation/admin-guide/devices.txt b/Documentation/admin-guide/devices.txt +index 922c23bb4372a..c07dc0ee860e7 100644 +--- a/Documentation/admin-guide/devices.txt ++++ b/Documentation/admin-guide/devices.txt +@@ -2339,13 +2339,7 @@ + disks (see major number 3) except that the limit on + partitions is 31. + +- 162 char Raw block device interface +- 0 = /dev/rawctl Raw I/O control device +- 1 = /dev/raw/raw1 First raw I/O device +- 2 = /dev/raw/raw2 Second raw I/O device +- ... +- max minor number of raw device is set by kernel config +- MAX_RAW_DEVS or raw module parameter 'max_raw_devs' ++ 162 char Used for (now removed) raw block device interface + + 163 char + +diff --git a/Documentation/admin-guide/hw-vuln/spectre.rst b/Documentation/admin-guide/hw-vuln/spectre.rst +index e05e581af5cfe..985181dba0bac 100644 +--- a/Documentation/admin-guide/hw-vuln/spectre.rst ++++ b/Documentation/admin-guide/hw-vuln/spectre.rst +@@ -468,7 +468,7 @@ Spectre variant 2 + before invoking any firmware code to prevent Spectre variant 2 exploits + using the firmware. + +- Using kernel address space randomization (CONFIG_RANDOMIZE_SLAB=y ++ Using kernel address space randomization (CONFIG_RANDOMIZE_BASE=y + and CONFIG_SLAB_FREELIST_RANDOM=y in the kernel configuration) makes + attacks on the kernel generally more difficult. + +diff --git a/Documentation/devicetree/bindings/display/amlogic,meson-dw-hdmi.yaml b/Documentation/devicetree/bindings/display/amlogic,meson-dw-hdmi.yaml +index cf5a208f2f105..343598c9f473b 100644 +--- a/Documentation/devicetree/bindings/display/amlogic,meson-dw-hdmi.yaml ++++ b/Documentation/devicetree/bindings/display/amlogic,meson-dw-hdmi.yaml +@@ -10,6 +10,9 @@ title: Amlogic specific extensions to the Synopsys Designware HDMI Controller + maintainers: + - Neil Armstrong <narmstrong@baylibre.com> + ++allOf: ++ - $ref: /schemas/sound/name-prefix.yaml# ++ + description: | + The Amlogic Meson Synopsys Designware Integration is composed of + - A Synopsys DesignWare HDMI Controller IP +@@ -99,6 +102,8 @@ properties: + "#sound-dai-cells": + const: 0 + ++ sound-name-prefix: true ++ + required: + - compatible + - reg +diff --git a/Documentation/devicetree/bindings/display/amlogic,meson-vpu.yaml b/Documentation/devicetree/bindings/display/amlogic,meson-vpu.yaml +index 851cb07812173..047fd69e03770 100644 +--- a/Documentation/devicetree/bindings/display/amlogic,meson-vpu.yaml ++++ b/Documentation/devicetree/bindings/display/amlogic,meson-vpu.yaml +@@ -78,6 +78,10 @@ properties: + interrupts: + maxItems: 1 + ++ amlogic,canvas: ++ description: should point to a canvas provider node ++ $ref: /schemas/types.yaml#/definitions/phandle ++ + power-domains: + maxItems: 1 + description: phandle to the associated power domain +@@ -106,6 +110,7 @@ required: + - port@1 + - "#address-cells" + - "#size-cells" ++ - amlogic,canvas + + additionalProperties: false + +@@ -118,6 +123,7 @@ examples: + interrupts = <3>; + #address-cells = <1>; + #size-cells = <0>; ++ amlogic,canvas = <&canvas>; + + /* CVBS VDAC output port */ + port@0 { +diff --git a/Documentation/devicetree/bindings/input/hid-over-i2c.txt b/Documentation/devicetree/bindings/input/hid-over-i2c.txt +index c76bafaf98d2f..34c43d3bddfd1 100644 +--- a/Documentation/devicetree/bindings/input/hid-over-i2c.txt ++++ b/Documentation/devicetree/bindings/input/hid-over-i2c.txt +@@ -32,6 +32,8 @@ device-specific compatible properties, which should be used in addition to the + - vdd-supply: phandle of the regulator that provides the supply voltage. + - post-power-on-delay-ms: time required by the device after enabling its regulators + or powering it on, before it is ready for communication. ++- touchscreen-inverted-x: See touchscreen.txt ++- touchscreen-inverted-y: See touchscreen.txt + + Example: + +diff --git a/Documentation/devicetree/bindings/thermal/thermal-zones.yaml b/Documentation/devicetree/bindings/thermal/thermal-zones.yaml +index a07de5ed0ca6a..2d34f3ccb2572 100644 +--- a/Documentation/devicetree/bindings/thermal/thermal-zones.yaml ++++ b/Documentation/devicetree/bindings/thermal/thermal-zones.yaml +@@ -199,12 +199,11 @@ patternProperties: + + contribution: + $ref: /schemas/types.yaml#/definitions/uint32 +- minimum: 0 +- maximum: 100 + description: +- The percentage contribution of the cooling devices at the +- specific trip temperature referenced in this map +- to this thermal zone ++ The cooling contribution to the thermal zone of the referred ++ cooling device at the referred trip point. The contribution is ++ a ratio of the sum of all cooling contributions within a ++ thermal zone. + + required: + - trip +diff --git a/Documentation/devicetree/bindings/watchdog/samsung-wdt.yaml b/Documentation/devicetree/bindings/watchdog/samsung-wdt.yaml +index 76cb9586ee00c..93cd77a6e92c0 100644 +--- a/Documentation/devicetree/bindings/watchdog/samsung-wdt.yaml ++++ b/Documentation/devicetree/bindings/watchdog/samsung-wdt.yaml +@@ -39,8 +39,8 @@ properties: + samsung,syscon-phandle: + $ref: /schemas/types.yaml#/definitions/phandle + description: +- Phandle to the PMU system controller node (in case of Exynos5250 +- and Exynos5420). ++ Phandle to the PMU system controller node (in case of Exynos5250, ++ Exynos5420 and Exynos7). + + required: + - compatible +@@ -58,6 +58,7 @@ allOf: + enum: + - samsung,exynos5250-wdt + - samsung,exynos5420-wdt ++ - samsung,exynos7-wdt + then: + required: + - samsung,syscon-phandle +diff --git a/Documentation/driver-api/dmaengine/dmatest.rst b/Documentation/driver-api/dmaengine/dmatest.rst +index ee268d445d38b..d2e1d8b58e7dc 100644 +--- a/Documentation/driver-api/dmaengine/dmatest.rst ++++ b/Documentation/driver-api/dmaengine/dmatest.rst +@@ -143,13 +143,14 @@ Part 5 - Handling channel allocation + Allocating Channels + ------------------- + +-Channels are required to be configured prior to starting the test run. +-Attempting to run the test without configuring the channels will fail. ++Channels do not need to be configured prior to starting a test run. Attempting ++to run the test without configuring the channels will result in testing any ++channels that are available. + + Example:: + + % echo 1 > /sys/module/dmatest/parameters/run +- dmatest: Could not start test, no channels configured ++ dmatest: No channels configured, continue with any + + Channels are registered using the "channel" parameter. Channels can be requested by their + name, once requested, the channel is registered and a pending thread is added to the test list. +diff --git a/Documentation/driver-api/firewire.rst b/Documentation/driver-api/firewire.rst +index 94a2d7f01d999..d3cfa73cbb2b4 100644 +--- a/Documentation/driver-api/firewire.rst ++++ b/Documentation/driver-api/firewire.rst +@@ -19,7 +19,7 @@ of kernel interfaces is available via exported symbols in `firewire-core` module + Firewire char device data structures + ==================================== + +-.. include:: /ABI/stable/firewire-cdev ++.. include:: ../ABI/stable/firewire-cdev + :literal: + + .. kernel-doc:: include/uapi/linux/firewire-cdev.h +@@ -28,7 +28,7 @@ Firewire char device data structures + Firewire device probing and sysfs interfaces + ============================================ + +-.. include:: /ABI/stable/sysfs-bus-firewire ++.. include:: ../ABI/stable/sysfs-bus-firewire + :literal: + + .. kernel-doc:: drivers/firewire/core-device.c +diff --git a/Documentation/firmware-guide/acpi/dsd/data-node-references.rst b/Documentation/firmware-guide/acpi/dsd/data-node-references.rst +index b7ad47df49de0..8b65b32e6e40e 100644 +--- a/Documentation/firmware-guide/acpi/dsd/data-node-references.rst ++++ b/Documentation/firmware-guide/acpi/dsd/data-node-references.rst +@@ -5,7 +5,7 @@ + Referencing hierarchical data nodes + =================================== + +-:Copyright: |copy| 2018 Intel Corporation ++:Copyright: |copy| 2018, 2021 Intel Corporation + :Author: Sakari Ailus <sakari.ailus@linux.intel.com> + + ACPI in general allows referring to device objects in the tree only. +@@ -52,12 +52,14 @@ the ANOD object which is also the final target node of the reference. + Name (NOD0, Package() { + ToUUID("daffd814-6eba-4d8c-8a91-bc9bbf4aa301"), + Package () { ++ Package () { "reg", 0 }, + Package () { "random-property", 3 }, + } + }) + Name (NOD1, Package() { + ToUUID("dbb8e3e6-5886-4ba6-8795-1319f52a966b"), + Package () { ++ Package () { "reg", 1 }, + Package () { "anothernode", "ANOD" }, + } + }) +@@ -74,7 +76,11 @@ the ANOD object which is also the final target node of the reference. + Name (_DSD, Package () { + ToUUID("daffd814-6eba-4d8c-8a91-bc9bbf4aa301"), + Package () { +- Package () { "reference", ^DEV0, "node@1", "anothernode" }, ++ Package () { ++ "reference", Package () { ++ ^DEV0, "node@1", "anothernode" ++ } ++ }, + } + }) + } +diff --git a/Documentation/trace/coresight/coresight-config.rst b/Documentation/trace/coresight/coresight-config.rst +index a4e3ef2952401..6ed13398ca2ce 100644 +--- a/Documentation/trace/coresight/coresight-config.rst ++++ b/Documentation/trace/coresight/coresight-config.rst +@@ -211,19 +211,13 @@ also declared in the perf 'cs_etm' event infrastructure so that they can + be selected when running trace under perf:: + + $ ls /sys/devices/cs_etm +- configurations format perf_event_mux_interval_ms sinks type +- events nr_addr_filters power ++ cpu0 cpu2 events nr_addr_filters power subsystem uevent ++ cpu1 cpu3 format perf_event_mux_interval_ms sinks type + +-Key directories here are 'configurations' - which lists the loaded +-configurations, and 'events' - a generic perf directory which allows +-selection on the perf command line.:: ++The key directory here is 'events' - a generic perf directory which allows ++selection on the perf command line. As with the sinks entries, this provides ++a hash of the configuration name. + +- $ ls configurations/ +- autofdo +- $ cat configurations/autofdo +- 0xa7c3dddd +- +-As with the sinks entries, this provides a hash of the configuration name. + The entry in the 'events' directory uses perfs built in syntax generator + to substitute the syntax for the name when evaluating the command:: + +diff --git a/Makefile b/Makefile +index af173c9df9422..088197ed3f66c 100644 +--- a/Makefile ++++ b/Makefile +@@ -1,7 +1,7 @@ + # SPDX-License-Identifier: GPL-2.0 + VERSION = 5 + PATCHLEVEL = 15 +-SUBLEVEL = 16 ++SUBLEVEL = 17 + EXTRAVERSION = + NAME = Trick or Treat + +diff --git a/arch/arm/Kconfig.debug b/arch/arm/Kconfig.debug +index 98436702e0c7e..644875d73ba15 100644 +--- a/arch/arm/Kconfig.debug ++++ b/arch/arm/Kconfig.debug +@@ -410,12 +410,12 @@ choice + Say Y here if you want kernel low-level debugging support + on i.MX25. + +- config DEBUG_IMX21_IMX27_UART +- bool "i.MX21 and i.MX27 Debug UART" +- depends on SOC_IMX21 || SOC_IMX27 ++ config DEBUG_IMX27_UART ++ bool "i.MX27 Debug UART" ++ depends on SOC_IMX27 + help + Say Y here if you want kernel low-level debugging support +- on i.MX21 or i.MX27. ++ on i.MX27. + + config DEBUG_IMX28_UART + bool "i.MX28 Debug UART" +@@ -1481,7 +1481,7 @@ config DEBUG_IMX_UART_PORT + int "i.MX Debug UART Port Selection" + depends on DEBUG_IMX1_UART || \ + DEBUG_IMX25_UART || \ +- DEBUG_IMX21_IMX27_UART || \ ++ DEBUG_IMX27_UART || \ + DEBUG_IMX31_UART || \ + DEBUG_IMX35_UART || \ + DEBUG_IMX50_UART || \ +@@ -1540,12 +1540,12 @@ config DEBUG_LL_INCLUDE + default "debug/icedcc.S" if DEBUG_ICEDCC + default "debug/imx.S" if DEBUG_IMX1_UART || \ + DEBUG_IMX25_UART || \ +- DEBUG_IMX21_IMX27_UART || \ ++ DEBUG_IMX27_UART || \ + DEBUG_IMX31_UART || \ + DEBUG_IMX35_UART || \ + DEBUG_IMX50_UART || \ + DEBUG_IMX51_UART || \ +- DEBUG_IMX53_UART ||\ ++ DEBUG_IMX53_UART || \ + DEBUG_IMX6Q_UART || \ + DEBUG_IMX6SL_UART || \ + DEBUG_IMX6SX_UART || \ +diff --git a/arch/arm/boot/compressed/efi-header.S b/arch/arm/boot/compressed/efi-header.S +index c0e7a745103e2..230030c130853 100644 +--- a/arch/arm/boot/compressed/efi-header.S ++++ b/arch/arm/boot/compressed/efi-header.S +@@ -9,16 +9,22 @@ + #include <linux/sizes.h> + + .macro __nop +-#ifdef CONFIG_EFI_STUB +- @ This is almost but not quite a NOP, since it does clobber the +- @ condition flags. But it is the best we can do for EFI, since +- @ PE/COFF expects the magic string "MZ" at offset 0, while the +- @ ARM/Linux boot protocol expects an executable instruction +- @ there. +- .inst MZ_MAGIC | (0x1310 << 16) @ tstne r0, #0x4d000 +-#else + AR_CLASS( mov r0, r0 ) + M_CLASS( nop.w ) ++ .endm ++ ++ .macro __initial_nops ++#ifdef CONFIG_EFI_STUB ++ @ This is a two-instruction NOP, which happens to bear the ++ @ PE/COFF signature "MZ" in the first two bytes, so the kernel ++ @ is accepted as an EFI binary. Booting via the UEFI stub ++ @ will not execute those instructions, but the ARM/Linux ++ @ boot protocol does, so we need some NOPs here. ++ .inst MZ_MAGIC | (0xe225 << 16) @ eor r5, r5, 0x4d000 ++ eor r5, r5, 0x4d000 @ undo previous insn ++#else ++ __nop ++ __nop + #endif + .endm + +diff --git a/arch/arm/boot/compressed/head.S b/arch/arm/boot/compressed/head.S +index b1cb1972361b8..bf79f2f78d232 100644 +--- a/arch/arm/boot/compressed/head.S ++++ b/arch/arm/boot/compressed/head.S +@@ -203,7 +203,8 @@ start: + * were patching the initial instructions of the kernel, i.e + * had started to exploit this "patch area". + */ +- .rept 7 ++ __initial_nops ++ .rept 5 + __nop + .endr + #ifndef CONFIG_THUMB2_KERNEL +diff --git a/arch/arm/boot/dts/armada-38x.dtsi b/arch/arm/boot/dts/armada-38x.dtsi +index 9b1a24cc5e91f..df3c8d1d8f641 100644 +--- a/arch/arm/boot/dts/armada-38x.dtsi ++++ b/arch/arm/boot/dts/armada-38x.dtsi +@@ -168,7 +168,7 @@ + }; + + uart0: serial@12000 { +- compatible = "marvell,armada-38x-uart"; ++ compatible = "marvell,armada-38x-uart", "ns16550a"; + reg = <0x12000 0x100>; + reg-shift = <2>; + interrupts = <GIC_SPI 12 IRQ_TYPE_LEVEL_HIGH>; +@@ -178,7 +178,7 @@ + }; + + uart1: serial@12100 { +- compatible = "marvell,armada-38x-uart"; ++ compatible = "marvell,armada-38x-uart", "ns16550a"; + reg = <0x12100 0x100>; + reg-shift = <2>; + interrupts = <GIC_SPI 13 IRQ_TYPE_LEVEL_HIGH>; +diff --git a/arch/arm/boot/dts/gemini-nas4220b.dts b/arch/arm/boot/dts/gemini-nas4220b.dts +index 13112a8a5dd88..6544c730340fa 100644 +--- a/arch/arm/boot/dts/gemini-nas4220b.dts ++++ b/arch/arm/boot/dts/gemini-nas4220b.dts +@@ -84,7 +84,7 @@ + partitions { + compatible = "redboot-fis"; + /* Eraseblock at 0xfe0000 */ +- fis-index-block = <0x1fc>; ++ fis-index-block = <0x7f>; + }; + }; + +diff --git a/arch/arm/boot/dts/omap3-n900.dts b/arch/arm/boot/dts/omap3-n900.dts +index 32335d4ce478b..d40c3d2c4914e 100644 +--- a/arch/arm/boot/dts/omap3-n900.dts ++++ b/arch/arm/boot/dts/omap3-n900.dts +@@ -8,6 +8,7 @@ + + #include "omap34xx.dtsi" + #include <dt-bindings/input/input.h> ++#include <dt-bindings/leds/common.h> + + /* + * Default secure signed bootloader (Nokia X-Loader) does not enable L3 firewall +@@ -630,63 +631,92 @@ + }; + + lp5523: lp5523@32 { ++ #address-cells = <1>; ++ #size-cells = <0>; + compatible = "national,lp5523"; + reg = <0x32>; + clock-mode = /bits/ 8 <0>; /* LP55XX_CLOCK_AUTO */ +- enable-gpio = <&gpio2 9 GPIO_ACTIVE_HIGH>; /* 41 */ ++ enable-gpios = <&gpio2 9 GPIO_ACTIVE_HIGH>; /* 41 */ + +- chan0 { ++ led@0 { ++ reg = <0>; + chan-name = "lp5523:kb1"; + led-cur = /bits/ 8 <50>; + max-cur = /bits/ 8 <100>; ++ color = <LED_COLOR_ID_WHITE>; ++ function = LED_FUNCTION_KBD_BACKLIGHT; + }; + +- chan1 { ++ led@1 { ++ reg = <1>; + chan-name = "lp5523:kb2"; + led-cur = /bits/ 8 <50>; + max-cur = /bits/ 8 <100>; ++ color = <LED_COLOR_ID_WHITE>; ++ function = LED_FUNCTION_KBD_BACKLIGHT; + }; + +- chan2 { ++ led@2 { ++ reg = <2>; + chan-name = "lp5523:kb3"; + led-cur = /bits/ 8 <50>; + max-cur = /bits/ 8 <100>; ++ color = <LED_COLOR_ID_WHITE>; ++ function = LED_FUNCTION_KBD_BACKLIGHT; + }; + +- chan3 { ++ led@3 { ++ reg = <3>; + chan-name = "lp5523:kb4"; + led-cur = /bits/ 8 <50>; + max-cur = /bits/ 8 <100>; ++ color = <LED_COLOR_ID_WHITE>; ++ function = LED_FUNCTION_KBD_BACKLIGHT; + }; + +- chan4 { ++ led@4 { ++ reg = <4>; + chan-name = "lp5523:b"; + led-cur = /bits/ 8 <50>; + max-cur = /bits/ 8 <100>; ++ color = <LED_COLOR_ID_BLUE>; ++ function = LED_FUNCTION_STATUS; + }; + +- chan5 { ++ led@5 { ++ reg = <5>; + chan-name = "lp5523:g"; + led-cur = /bits/ 8 <50>; + max-cur = /bits/ 8 <100>; ++ color = <LED_COLOR_ID_GREEN>; ++ function = LED_FUNCTION_STATUS; + }; + +- chan6 { ++ led@6 { ++ reg = <6>; + chan-name = "lp5523:r"; + led-cur = /bits/ 8 <50>; + max-cur = /bits/ 8 <100>; ++ color = <LED_COLOR_ID_RED>; ++ function = LED_FUNCTION_STATUS; + }; + +- chan7 { ++ led@7 { ++ reg = <7>; + chan-name = "lp5523:kb5"; + led-cur = /bits/ 8 <50>; + max-cur = /bits/ 8 <100>; ++ color = <LED_COLOR_ID_WHITE>; ++ function = LED_FUNCTION_KBD_BACKLIGHT; + }; + +- chan8 { ++ led@8 { ++ reg = <8>; + chan-name = "lp5523:kb6"; + led-cur = /bits/ 8 <50>; + max-cur = /bits/ 8 <100>; ++ color = <LED_COLOR_ID_WHITE>; ++ function = LED_FUNCTION_KBD_BACKLIGHT; + }; + }; + +diff --git a/arch/arm/boot/dts/qcom-sdx55.dtsi b/arch/arm/boot/dts/qcom-sdx55.dtsi +index 1e6ce035f76a9..b5b784c5c65e4 100644 +--- a/arch/arm/boot/dts/qcom-sdx55.dtsi ++++ b/arch/arm/boot/dts/qcom-sdx55.dtsi +@@ -334,12 +334,10 @@ + clocks = <&rpmhcc RPMH_IPA_CLK>; + clock-names = "core"; + +- interconnects = <&system_noc MASTER_IPA &system_noc SLAVE_SNOC_MEM_NOC_GC>, +- <&mem_noc MASTER_SNOC_GC_MEM_NOC &mc_virt SLAVE_EBI_CH0>, ++ interconnects = <&system_noc MASTER_IPA &mc_virt SLAVE_EBI_CH0>, + <&system_noc MASTER_IPA &system_noc SLAVE_OCIMEM>, + <&mem_noc MASTER_AMPSS_M0 &system_noc SLAVE_IPA_CFG>; +- interconnect-names = "memory-a", +- "memory-b", ++ interconnect-names = "memory", + "imem", + "config"; + +diff --git a/arch/arm/boot/dts/sama7g5-pinfunc.h b/arch/arm/boot/dts/sama7g5-pinfunc.h +index 22fe9e522a97b..4eb30445d2057 100644 +--- a/arch/arm/boot/dts/sama7g5-pinfunc.h ++++ b/arch/arm/boot/dts/sama7g5-pinfunc.h +@@ -765,7 +765,7 @@ + #define PIN_PD20__PCK0 PINMUX_PIN(PIN_PD20, 1, 3) + #define PIN_PD20__FLEXCOM2_IO3 PINMUX_PIN(PIN_PD20, 2, 2) + #define PIN_PD20__PWMH3 PINMUX_PIN(PIN_PD20, 3, 4) +-#define PIN_PD20__CANTX4 PINMUX_PIN(PIN_PD20, 5, 2) ++#define PIN_PD20__CANTX4 PINMUX_PIN(PIN_PD20, 4, 2) + #define PIN_PD20__FLEXCOM5_IO0 PINMUX_PIN(PIN_PD20, 6, 5) + #define PIN_PD21 117 + #define PIN_PD21__GPIO PINMUX_PIN(PIN_PD21, 0, 0) +diff --git a/arch/arm/boot/dts/stm32f429-disco.dts b/arch/arm/boot/dts/stm32f429-disco.dts +index 075ac57d0bf4a..6435e099c6326 100644 +--- a/arch/arm/boot/dts/stm32f429-disco.dts ++++ b/arch/arm/boot/dts/stm32f429-disco.dts +@@ -192,7 +192,7 @@ + + display: display@1{ + /* Connect panel-ilitek-9341 to ltdc */ +- compatible = "st,sf-tc240t-9370-t"; ++ compatible = "st,sf-tc240t-9370-t", "ilitek,ili9341"; + reg = <1>; + spi-3wire; + spi-max-frequency = <10000000>; +diff --git a/arch/arm/configs/cm_x300_defconfig b/arch/arm/configs/cm_x300_defconfig +index 502a9d870ca44..45769d0ddd4ef 100644 +--- a/arch/arm/configs/cm_x300_defconfig ++++ b/arch/arm/configs/cm_x300_defconfig +@@ -146,7 +146,6 @@ CONFIG_NFS_V3_ACL=y + CONFIG_NFS_V4=y + CONFIG_ROOT_NFS=y + CONFIG_CIFS=m +-CONFIG_CIFS_WEAK_PW_HASH=y + CONFIG_PARTITION_ADVANCED=y + CONFIG_NLS_CODEPAGE_437=m + CONFIG_NLS_ISO8859_1=m +diff --git a/arch/arm/configs/ezx_defconfig b/arch/arm/configs/ezx_defconfig +index a49e699e52de3..ec84d80096b1c 100644 +--- a/arch/arm/configs/ezx_defconfig ++++ b/arch/arm/configs/ezx_defconfig +@@ -314,7 +314,6 @@ CONFIG_NFSD_V3_ACL=y + CONFIG_SMB_FS=m + CONFIG_CIFS=m + CONFIG_CIFS_STATS=y +-CONFIG_CIFS_WEAK_PW_HASH=y + CONFIG_CIFS_XATTR=y + CONFIG_CIFS_POSIX=y + CONFIG_NLS_CODEPAGE_437=m +diff --git a/arch/arm/configs/imote2_defconfig b/arch/arm/configs/imote2_defconfig +index 118c4c927f264..6db871d4e0775 100644 +--- a/arch/arm/configs/imote2_defconfig ++++ b/arch/arm/configs/imote2_defconfig +@@ -288,7 +288,6 @@ CONFIG_NFSD_V3_ACL=y + CONFIG_SMB_FS=m + CONFIG_CIFS=m + CONFIG_CIFS_STATS=y +-CONFIG_CIFS_WEAK_PW_HASH=y + CONFIG_CIFS_XATTR=y + CONFIG_CIFS_POSIX=y + CONFIG_NLS_CODEPAGE_437=m +diff --git a/arch/arm/configs/nhk8815_defconfig b/arch/arm/configs/nhk8815_defconfig +index 23595fc5a29a9..907d6512821ad 100644 +--- a/arch/arm/configs/nhk8815_defconfig ++++ b/arch/arm/configs/nhk8815_defconfig +@@ -127,7 +127,6 @@ CONFIG_NFS_FS=y + CONFIG_NFS_V3_ACL=y + CONFIG_ROOT_NFS=y + CONFIG_CIFS=m +-CONFIG_CIFS_WEAK_PW_HASH=y + CONFIG_NLS_CODEPAGE_437=y + CONFIG_NLS_ASCII=y + CONFIG_NLS_ISO8859_1=y +diff --git a/arch/arm/configs/pxa_defconfig b/arch/arm/configs/pxa_defconfig +index 58f4834289e63..dedaaae3d0d8a 100644 +--- a/arch/arm/configs/pxa_defconfig ++++ b/arch/arm/configs/pxa_defconfig +@@ -699,7 +699,6 @@ CONFIG_NFSD_V3_ACL=y + CONFIG_NFSD_V4=y + CONFIG_CIFS=m + CONFIG_CIFS_STATS=y +-CONFIG_CIFS_WEAK_PW_HASH=y + CONFIG_CIFS_XATTR=y + CONFIG_CIFS_POSIX=y + CONFIG_NLS_DEFAULT="utf8" +diff --git a/arch/arm/configs/spear13xx_defconfig b/arch/arm/configs/spear13xx_defconfig +index 3b206a31902ff..065553326b391 100644 +--- a/arch/arm/configs/spear13xx_defconfig ++++ b/arch/arm/configs/spear13xx_defconfig +@@ -61,7 +61,6 @@ CONFIG_SERIAL_AMBA_PL011=y + CONFIG_SERIAL_AMBA_PL011_CONSOLE=y + # CONFIG_HW_RANDOM is not set + CONFIG_RAW_DRIVER=y +-CONFIG_MAX_RAW_DEVS=8192 + CONFIG_I2C=y + CONFIG_I2C_DESIGNWARE_PLATFORM=y + CONFIG_SPI=y +diff --git a/arch/arm/configs/spear3xx_defconfig b/arch/arm/configs/spear3xx_defconfig +index fc5f71c765edc..afca722d6605c 100644 +--- a/arch/arm/configs/spear3xx_defconfig ++++ b/arch/arm/configs/spear3xx_defconfig +@@ -41,7 +41,6 @@ CONFIG_SERIAL_AMBA_PL011=y + CONFIG_SERIAL_AMBA_PL011_CONSOLE=y + # CONFIG_HW_RANDOM is not set + CONFIG_RAW_DRIVER=y +-CONFIG_MAX_RAW_DEVS=8192 + CONFIG_I2C=y + CONFIG_I2C_DESIGNWARE_PLATFORM=y + CONFIG_SPI=y +diff --git a/arch/arm/configs/spear6xx_defconfig b/arch/arm/configs/spear6xx_defconfig +index 52a56b8ce6a71..bc32c02cb86b1 100644 +--- a/arch/arm/configs/spear6xx_defconfig ++++ b/arch/arm/configs/spear6xx_defconfig +@@ -36,7 +36,6 @@ CONFIG_INPUT_FF_MEMLESS=y + CONFIG_SERIAL_AMBA_PL011=y + CONFIG_SERIAL_AMBA_PL011_CONSOLE=y + CONFIG_RAW_DRIVER=y +-CONFIG_MAX_RAW_DEVS=8192 + CONFIG_I2C=y + CONFIG_I2C_DESIGNWARE_PLATFORM=y + CONFIG_SPI=y +diff --git a/arch/arm/include/debug/imx-uart.h b/arch/arm/include/debug/imx-uart.h +index c8eb83d4b8964..3edbb3c5b42bf 100644 +--- a/arch/arm/include/debug/imx-uart.h ++++ b/arch/arm/include/debug/imx-uart.h +@@ -11,13 +11,6 @@ + #define IMX1_UART_BASE_ADDR(n) IMX1_UART##n##_BASE_ADDR + #define IMX1_UART_BASE(n) IMX1_UART_BASE_ADDR(n) + +-#define IMX21_UART1_BASE_ADDR 0x1000a000 +-#define IMX21_UART2_BASE_ADDR 0x1000b000 +-#define IMX21_UART3_BASE_ADDR 0x1000c000 +-#define IMX21_UART4_BASE_ADDR 0x1000d000 +-#define IMX21_UART_BASE_ADDR(n) IMX21_UART##n##_BASE_ADDR +-#define IMX21_UART_BASE(n) IMX21_UART_BASE_ADDR(n) +- + #define IMX25_UART1_BASE_ADDR 0x43f90000 + #define IMX25_UART2_BASE_ADDR 0x43f94000 + #define IMX25_UART3_BASE_ADDR 0x5000c000 +@@ -26,6 +19,13 @@ + #define IMX25_UART_BASE_ADDR(n) IMX25_UART##n##_BASE_ADDR + #define IMX25_UART_BASE(n) IMX25_UART_BASE_ADDR(n) + ++#define IMX27_UART1_BASE_ADDR 0x1000a000 ++#define IMX27_UART2_BASE_ADDR 0x1000b000 ++#define IMX27_UART3_BASE_ADDR 0x1000c000 ++#define IMX27_UART4_BASE_ADDR 0x1000d000 ++#define IMX27_UART_BASE_ADDR(n) IMX27_UART##n##_BASE_ADDR ++#define IMX27_UART_BASE(n) IMX27_UART_BASE_ADDR(n) ++ + #define IMX31_UART1_BASE_ADDR 0x43f90000 + #define IMX31_UART2_BASE_ADDR 0x43f94000 + #define IMX31_UART3_BASE_ADDR 0x5000c000 +@@ -112,10 +112,10 @@ + + #ifdef CONFIG_DEBUG_IMX1_UART + #define UART_PADDR IMX_DEBUG_UART_BASE(IMX1) +-#elif defined(CONFIG_DEBUG_IMX21_IMX27_UART) +-#define UART_PADDR IMX_DEBUG_UART_BASE(IMX21) + #elif defined(CONFIG_DEBUG_IMX25_UART) + #define UART_PADDR IMX_DEBUG_UART_BASE(IMX25) ++#elif defined(CONFIG_DEBUG_IMX27_UART) ++#define UART_PADDR IMX_DEBUG_UART_BASE(IMX27) + #elif defined(CONFIG_DEBUG_IMX31_UART) + #define UART_PADDR IMX_DEBUG_UART_BASE(IMX31) + #elif defined(CONFIG_DEBUG_IMX35_UART) +diff --git a/arch/arm/mach-shmobile/regulator-quirk-rcar-gen2.c b/arch/arm/mach-shmobile/regulator-quirk-rcar-gen2.c +index ee949255ced3f..09ef73b99dd86 100644 +--- a/arch/arm/mach-shmobile/regulator-quirk-rcar-gen2.c ++++ b/arch/arm/mach-shmobile/regulator-quirk-rcar-gen2.c +@@ -154,8 +154,10 @@ static int __init rcar_gen2_regulator_quirk(void) + return -ENODEV; + + for_each_matching_node_and_match(np, rcar_gen2_quirk_match, &id) { +- if (!of_device_is_available(np)) ++ if (!of_device_is_available(np)) { ++ of_node_put(np); + break; ++ } + + ret = of_property_read_u32(np, "reg", &addr); + if (ret) /* Skip invalid entry and continue */ +@@ -164,6 +166,7 @@ static int __init rcar_gen2_regulator_quirk(void) + quirk = kzalloc(sizeof(*quirk), GFP_KERNEL); + if (!quirk) { + ret = -ENOMEM; ++ of_node_put(np); + goto err_mem; + } + +diff --git a/arch/arm64/boot/dts/amlogic/meson-g12-common.dtsi b/arch/arm64/boot/dts/amlogic/meson-g12-common.dtsi +index 00c6f53290d43..428449d98c0ae 100644 +--- a/arch/arm64/boot/dts/amlogic/meson-g12-common.dtsi ++++ b/arch/arm64/boot/dts/amlogic/meson-g12-common.dtsi +@@ -58,7 +58,7 @@ + secure-monitor = <&sm>; + }; + +- gpu_opp_table: gpu-opp-table { ++ gpu_opp_table: opp-table-gpu { + compatible = "operating-points-v2"; + + opp-124999998 { +diff --git a/arch/arm64/boot/dts/amlogic/meson-g12b-odroid-n2.dtsi b/arch/arm64/boot/dts/amlogic/meson-g12b-odroid-n2.dtsi +index 4f33820aba1f1..a84ed3578425e 100644 +--- a/arch/arm64/boot/dts/amlogic/meson-g12b-odroid-n2.dtsi ++++ b/arch/arm64/boot/dts/amlogic/meson-g12b-odroid-n2.dtsi +@@ -607,7 +607,7 @@ + pinctrl-0 = <&nor_pins>; + pinctrl-names = "default"; + +- mx25u64: spi-flash@0 { ++ mx25u64: flash@0 { + #address-cells = <1>; + #size-cells = <1>; + compatible = "mxicy,mx25u6435f", "jedec,spi-nor"; +diff --git a/arch/arm64/boot/dts/amlogic/meson-gxbb-wetek.dtsi b/arch/arm64/boot/dts/amlogic/meson-gxbb-wetek.dtsi +index a350fee1264d7..a4d34398da358 100644 +--- a/arch/arm64/boot/dts/amlogic/meson-gxbb-wetek.dtsi ++++ b/arch/arm64/boot/dts/amlogic/meson-gxbb-wetek.dtsi +@@ -6,6 +6,7 @@ + */ + + #include "meson-gxbb.dtsi" ++#include <dt-bindings/gpio/gpio.h> + + / { + aliases { +@@ -64,6 +65,7 @@ + regulator-name = "VDDIO_AO18"; + regulator-min-microvolt = <1800000>; + regulator-max-microvolt = <1800000>; ++ regulator-always-on; + }; + + vcc_3v3: regulator-vcc_3v3 { +@@ -161,6 +163,7 @@ + status = "okay"; + pinctrl-0 = <&hdmi_hpd_pins>, <&hdmi_i2c_pins>; + pinctrl-names = "default"; ++ hdmi-supply = <&vddio_ao18>; + }; + + &hdmi_tx_tmds_port { +diff --git a/arch/arm64/boot/dts/freescale/fsl-ls1028a-qds.dts b/arch/arm64/boot/dts/freescale/fsl-ls1028a-qds.dts +index bfd14b64567e4..2f92e62ecafe9 100644 +--- a/arch/arm64/boot/dts/freescale/fsl-ls1028a-qds.dts ++++ b/arch/arm64/boot/dts/freescale/fsl-ls1028a-qds.dts +@@ -272,11 +272,6 @@ + vcc-supply = <&sb_3v3>; + }; + +- rtc@51 { +- compatible = "nxp,pcf2129"; +- reg = <0x51>; +- }; +- + eeprom@56 { + compatible = "atmel,24c512"; + reg = <0x56>; +@@ -318,6 +313,15 @@ + + }; + ++&i2c1 { ++ status = "okay"; ++ ++ rtc@51 { ++ compatible = "nxp,pcf2129"; ++ reg = <0x51>; ++ }; ++}; ++ + &enetc_port1 { + phy-handle = <&qds_phy1>; + phy-connection-type = "rgmii-id"; +diff --git a/arch/arm64/boot/dts/marvell/cn9130.dtsi b/arch/arm64/boot/dts/marvell/cn9130.dtsi +index a2b7e5ec979d3..327b04134134f 100644 +--- a/arch/arm64/boot/dts/marvell/cn9130.dtsi ++++ b/arch/arm64/boot/dts/marvell/cn9130.dtsi +@@ -11,6 +11,13 @@ + model = "Marvell Armada CN9130 SoC"; + compatible = "marvell,cn9130", "marvell,armada-ap807-quad", + "marvell,armada-ap807"; ++ ++ aliases { ++ gpio1 = &cp0_gpio1; ++ gpio2 = &cp0_gpio2; ++ spi1 = &cp0_spi0; ++ spi2 = &cp0_spi1; ++ }; + }; + + /* +@@ -35,3 +42,11 @@ + #undef CP11X_PCIE0_BASE + #undef CP11X_PCIE1_BASE + #undef CP11X_PCIE2_BASE ++ ++&cp0_gpio1 { ++ status = "okay"; ++}; ++ ++&cp0_gpio2 { ++ status = "okay"; ++}; +diff --git a/arch/arm64/boot/dts/nvidia/tegra186.dtsi b/arch/arm64/boot/dts/nvidia/tegra186.dtsi +index e94f8add1a400..062e87e893316 100644 +--- a/arch/arm64/boot/dts/nvidia/tegra186.dtsi ++++ b/arch/arm64/boot/dts/nvidia/tegra186.dtsi +@@ -1079,7 +1079,7 @@ + + ccplex@e000000 { + compatible = "nvidia,tegra186-ccplex-cluster"; +- reg = <0x0 0x0e000000 0x0 0x3fffff>; ++ reg = <0x0 0x0e000000 0x0 0x400000>; + + nvidia,bpmp = <&bpmp>; + }; +diff --git a/arch/arm64/boot/dts/nvidia/tegra194.dtsi b/arch/arm64/boot/dts/nvidia/tegra194.dtsi +index c8250a3f7891f..510d2974470cd 100644 +--- a/arch/arm64/boot/dts/nvidia/tegra194.dtsi ++++ b/arch/arm64/boot/dts/nvidia/tegra194.dtsi +@@ -818,9 +818,8 @@ + <&bpmp TEGRA194_CLK_HDA2CODEC_2X>; + clock-names = "hda", "hda2hdmi", "hda2codec_2x"; + resets = <&bpmp TEGRA194_RESET_HDA>, +- <&bpmp TEGRA194_RESET_HDA2HDMICODEC>, +- <&bpmp TEGRA194_RESET_HDA2CODEC_2X>; +- reset-names = "hda", "hda2hdmi", "hda2codec_2x"; ++ <&bpmp TEGRA194_RESET_HDA2HDMICODEC>; ++ reset-names = "hda", "hda2hdmi"; + power-domains = <&bpmp TEGRA194_POWER_DOMAIN_DISP>; + interconnects = <&mc TEGRA194_MEMORY_CLIENT_HDAR &emc>, + <&mc TEGRA194_MEMORY_CLIENT_HDAW &emc>; +diff --git a/arch/arm64/boot/dts/qcom/ipq6018.dtsi b/arch/arm64/boot/dts/qcom/ipq6018.dtsi +index 7b6205c180df1..ce4c2b4a5fc07 100644 +--- a/arch/arm64/boot/dts/qcom/ipq6018.dtsi ++++ b/arch/arm64/boot/dts/qcom/ipq6018.dtsi +@@ -221,7 +221,7 @@ + interrupts = <GIC_SPI 208 IRQ_TYPE_LEVEL_HIGH>; + gpio-controller; + #gpio-cells = <2>; +- gpio-ranges = <&tlmm 0 80>; ++ gpio-ranges = <&tlmm 0 0 80>; + interrupt-controller; + #interrupt-cells = <2>; + +diff --git a/arch/arm64/boot/dts/qcom/msm8916.dtsi b/arch/arm64/boot/dts/qcom/msm8916.dtsi +index 427bb20626549..8b27242724641 100644 +--- a/arch/arm64/boot/dts/qcom/msm8916.dtsi ++++ b/arch/arm64/boot/dts/qcom/msm8916.dtsi +@@ -19,8 +19,8 @@ + #size-cells = <2>; + + aliases { +- sdhc1 = &sdhc_1; /* SDC1 eMMC slot */ +- sdhc2 = &sdhc_2; /* SDC2 SD card slot */ ++ mmc0 = &sdhc_1; /* SDC1 eMMC slot */ ++ mmc1 = &sdhc_2; /* SDC2 SD card slot */ + }; + + chosen { }; +diff --git a/arch/arm64/boot/dts/qcom/msm8996.dtsi b/arch/arm64/boot/dts/qcom/msm8996.dtsi +index f8d28dd76cfa8..6077c36019514 100644 +--- a/arch/arm64/boot/dts/qcom/msm8996.dtsi ++++ b/arch/arm64/boot/dts/qcom/msm8996.dtsi +@@ -965,9 +965,6 @@ + nvmem-cells = <&speedbin_efuse>; + nvmem-cell-names = "speed_bin"; + +- qcom,gpu-quirk-two-pass-use-wfi; +- qcom,gpu-quirk-fault-detect-mask; +- + operating-points-v2 = <&gpu_opp_table>; + + status = "disabled"; +diff --git a/arch/arm64/boot/dts/qcom/sc7280.dtsi b/arch/arm64/boot/dts/qcom/sc7280.dtsi +index f58336536a92a..692973c4f4344 100644 +--- a/arch/arm64/boot/dts/qcom/sc7280.dtsi ++++ b/arch/arm64/boot/dts/qcom/sc7280.dtsi +@@ -429,7 +429,7 @@ + <&rpmhcc RPMH_CXO_CLK_A>, <&sleep_clk>, + <0>, <0>, <0>, <0>, <0>, <0>; + clock-names = "bi_tcxo", "bi_tcxo_ao", "sleep_clk", +- "pcie_0_pipe_clk", "pcie_1_pipe-clk", ++ "pcie_0_pipe_clk", "pcie_1_pipe_clk", + "ufs_phy_rx_symbol_0_clk", "ufs_phy_rx_symbol_1_clk", + "ufs_phy_tx_symbol_0_clk", + "usb3_phy_wrapper_gcc_usb30_pipe_clk"; +diff --git a/arch/arm64/boot/dts/qcom/sdm850-lenovo-yoga-c630.dts b/arch/arm64/boot/dts/qcom/sdm850-lenovo-yoga-c630.dts +index 2ba23aa582a18..617a634ac9051 100644 +--- a/arch/arm64/boot/dts/qcom/sdm850-lenovo-yoga-c630.dts ++++ b/arch/arm64/boot/dts/qcom/sdm850-lenovo-yoga-c630.dts +@@ -518,6 +518,10 @@ + dai@1 { + reg = <1>; + }; ++ ++ dai@2 { ++ reg = <2>; ++ }; + }; + + &sound { +@@ -530,6 +534,7 @@ + "SpkrLeft IN", "SPK1 OUT", + "SpkrRight IN", "SPK2 OUT", + "MM_DL1", "MultiMedia1 Playback", ++ "MM_DL3", "MultiMedia3 Playback", + "MultiMedia2 Capture", "MM_UL2"; + + mm1-dai-link { +@@ -546,6 +551,13 @@ + }; + }; + ++ mm3-dai-link { ++ link-name = "MultiMedia3"; ++ cpu { ++ sound-dai = <&q6asmdai MSM_FRONTEND_DAI_MULTIMEDIA3>; ++ }; ++ }; ++ + slim-dai-link { + link-name = "SLIM Playback"; + cpu { +@@ -575,6 +587,21 @@ + sound-dai = <&wcd9340 1>; + }; + }; ++ ++ slim-wcd-dai-link { ++ link-name = "SLIM WCD Playback"; ++ cpu { ++ sound-dai = <&q6afedai SLIMBUS_1_RX>; ++ }; ++ ++ platform { ++ sound-dai = <&q6routing>; ++ }; ++ ++ codec { ++ sound-dai = <&wcd9340 2>; ++ }; ++ }; + }; + + &tlmm { +diff --git a/arch/arm64/boot/dts/qcom/sm8350.dtsi b/arch/arm64/boot/dts/qcom/sm8350.dtsi +index e91cd8a5e5356..296ffb0e9888c 100644 +--- a/arch/arm64/boot/dts/qcom/sm8350.dtsi ++++ b/arch/arm64/boot/dts/qcom/sm8350.dtsi +@@ -2185,7 +2185,7 @@ + }; + }; + +- camera-thermal-bottom { ++ cam-thermal-bottom { + polling-delay-passive = <250>; + polling-delay = <1000>; + +diff --git a/arch/arm64/boot/dts/renesas/cat875.dtsi b/arch/arm64/boot/dts/renesas/cat875.dtsi +index 801ea54b027c4..20f8adc635e72 100644 +--- a/arch/arm64/boot/dts/renesas/cat875.dtsi ++++ b/arch/arm64/boot/dts/renesas/cat875.dtsi +@@ -18,6 +18,7 @@ + pinctrl-names = "default"; + renesas,no-ether-link; + phy-handle = <&phy0>; ++ phy-mode = "rgmii-id"; + status = "okay"; + + phy0: ethernet-phy@0 { +diff --git a/arch/arm64/boot/dts/renesas/r8a774a1.dtsi b/arch/arm64/boot/dts/renesas/r8a774a1.dtsi +index 6f4fffacfca21..e70aa5a087402 100644 +--- a/arch/arm64/boot/dts/renesas/r8a774a1.dtsi ++++ b/arch/arm64/boot/dts/renesas/r8a774a1.dtsi +@@ -2784,7 +2784,7 @@ + }; + + thermal-zones { +- sensor_thermal1: sensor-thermal1 { ++ sensor1_thermal: sensor1-thermal { + polling-delay-passive = <250>; + polling-delay = <1000>; + thermal-sensors = <&tsc 0>; +@@ -2799,7 +2799,7 @@ + }; + }; + +- sensor_thermal2: sensor-thermal2 { ++ sensor2_thermal: sensor2-thermal { + polling-delay-passive = <250>; + polling-delay = <1000>; + thermal-sensors = <&tsc 1>; +@@ -2814,7 +2814,7 @@ + }; + }; + +- sensor_thermal3: sensor-thermal3 { ++ sensor3_thermal: sensor3-thermal { + polling-delay-passive = <250>; + polling-delay = <1000>; + thermal-sensors = <&tsc 2>; +diff --git a/arch/arm64/boot/dts/renesas/r8a774b1.dtsi b/arch/arm64/boot/dts/renesas/r8a774b1.dtsi +index 0f7bdfc90a0dc..6c5694fa66900 100644 +--- a/arch/arm64/boot/dts/renesas/r8a774b1.dtsi ++++ b/arch/arm64/boot/dts/renesas/r8a774b1.dtsi +@@ -2629,7 +2629,7 @@ + }; + + thermal-zones { +- sensor_thermal1: sensor-thermal1 { ++ sensor1_thermal: sensor1-thermal { + polling-delay-passive = <250>; + polling-delay = <1000>; + thermal-sensors = <&tsc 0>; +@@ -2644,7 +2644,7 @@ + }; + }; + +- sensor_thermal2: sensor-thermal2 { ++ sensor2_thermal: sensor2-thermal { + polling-delay-passive = <250>; + polling-delay = <1000>; + thermal-sensors = <&tsc 1>; +@@ -2659,7 +2659,7 @@ + }; + }; + +- sensor_thermal3: sensor-thermal3 { ++ sensor3_thermal: sensor3-thermal { + polling-delay-passive = <250>; + polling-delay = <1000>; + thermal-sensors = <&tsc 2>; +diff --git a/arch/arm64/boot/dts/renesas/r8a774e1.dtsi b/arch/arm64/boot/dts/renesas/r8a774e1.dtsi +index 379a1300272ba..62209ab6deb9a 100644 +--- a/arch/arm64/boot/dts/renesas/r8a774e1.dtsi ++++ b/arch/arm64/boot/dts/renesas/r8a774e1.dtsi +@@ -2904,7 +2904,7 @@ + }; + + thermal-zones { +- sensor_thermal1: sensor-thermal1 { ++ sensor1_thermal: sensor1-thermal { + polling-delay-passive = <250>; + polling-delay = <1000>; + thermal-sensors = <&tsc 0>; +@@ -2919,7 +2919,7 @@ + }; + }; + +- sensor_thermal2: sensor-thermal2 { ++ sensor2_thermal: sensor2-thermal { + polling-delay-passive = <250>; + polling-delay = <1000>; + thermal-sensors = <&tsc 1>; +@@ -2934,7 +2934,7 @@ + }; + }; + +- sensor_thermal3: sensor-thermal3 { ++ sensor3_thermal: sensor3-thermal { + polling-delay-passive = <250>; + polling-delay = <1000>; + thermal-sensors = <&tsc 2>; +diff --git a/arch/arm64/boot/dts/renesas/r8a77951.dtsi b/arch/arm64/boot/dts/renesas/r8a77951.dtsi +index 1768a3e6bb8da..193d81be40fc4 100644 +--- a/arch/arm64/boot/dts/renesas/r8a77951.dtsi ++++ b/arch/arm64/boot/dts/renesas/r8a77951.dtsi +@@ -3375,7 +3375,7 @@ + }; + + thermal-zones { +- sensor_thermal1: sensor-thermal1 { ++ sensor1_thermal: sensor1-thermal { + polling-delay-passive = <250>; + polling-delay = <1000>; + thermal-sensors = <&tsc 0>; +@@ -3390,7 +3390,7 @@ + }; + }; + +- sensor_thermal2: sensor-thermal2 { ++ sensor2_thermal: sensor2-thermal { + polling-delay-passive = <250>; + polling-delay = <1000>; + thermal-sensors = <&tsc 1>; +@@ -3405,7 +3405,7 @@ + }; + }; + +- sensor_thermal3: sensor-thermal3 { ++ sensor3_thermal: sensor3-thermal { + polling-delay-passive = <250>; + polling-delay = <1000>; + thermal-sensors = <&tsc 2>; +diff --git a/arch/arm64/boot/dts/renesas/r8a77960.dtsi b/arch/arm64/boot/dts/renesas/r8a77960.dtsi +index 2bd8169735d35..b526e4f0ee6a8 100644 +--- a/arch/arm64/boot/dts/renesas/r8a77960.dtsi ++++ b/arch/arm64/boot/dts/renesas/r8a77960.dtsi +@@ -2972,7 +2972,7 @@ + }; + + thermal-zones { +- sensor_thermal1: sensor-thermal1 { ++ sensor1_thermal: sensor1-thermal { + polling-delay-passive = <250>; + polling-delay = <1000>; + thermal-sensors = <&tsc 0>; +@@ -2987,7 +2987,7 @@ + }; + }; + +- sensor_thermal2: sensor-thermal2 { ++ sensor2_thermal: sensor2-thermal { + polling-delay-passive = <250>; + polling-delay = <1000>; + thermal-sensors = <&tsc 1>; +@@ -3002,7 +3002,7 @@ + }; + }; + +- sensor_thermal3: sensor-thermal3 { ++ sensor3_thermal: sensor3-thermal { + polling-delay-passive = <250>; + polling-delay = <1000>; + thermal-sensors = <&tsc 2>; +diff --git a/arch/arm64/boot/dts/renesas/r8a77961.dtsi b/arch/arm64/boot/dts/renesas/r8a77961.dtsi +index 041473aa5cd09..21fc95397c3c2 100644 +--- a/arch/arm64/boot/dts/renesas/r8a77961.dtsi ++++ b/arch/arm64/boot/dts/renesas/r8a77961.dtsi +@@ -2719,7 +2719,7 @@ + }; + + thermal-zones { +- sensor_thermal1: sensor-thermal1 { ++ sensor1_thermal: sensor1-thermal { + polling-delay-passive = <250>; + polling-delay = <1000>; + thermal-sensors = <&tsc 0>; +@@ -2734,7 +2734,7 @@ + }; + }; + +- sensor_thermal2: sensor-thermal2 { ++ sensor2_thermal: sensor2-thermal { + polling-delay-passive = <250>; + polling-delay = <1000>; + thermal-sensors = <&tsc 1>; +@@ -2749,7 +2749,7 @@ + }; + }; + +- sensor_thermal3: sensor-thermal3 { ++ sensor3_thermal: sensor3-thermal { + polling-delay-passive = <250>; + polling-delay = <1000>; + thermal-sensors = <&tsc 2>; +diff --git a/arch/arm64/boot/dts/renesas/r8a77965.dtsi b/arch/arm64/boot/dts/renesas/r8a77965.dtsi +index 08df75606430b..f9679a4dd85fa 100644 +--- a/arch/arm64/boot/dts/renesas/r8a77965.dtsi ++++ b/arch/arm64/boot/dts/renesas/r8a77965.dtsi +@@ -2784,7 +2784,7 @@ + }; + + thermal-zones { +- sensor_thermal1: sensor-thermal1 { ++ sensor1_thermal: sensor1-thermal { + polling-delay-passive = <250>; + polling-delay = <1000>; + thermal-sensors = <&tsc 0>; +@@ -2799,7 +2799,7 @@ + }; + }; + +- sensor_thermal2: sensor-thermal2 { ++ sensor2_thermal: sensor2-thermal { + polling-delay-passive = <250>; + polling-delay = <1000>; + thermal-sensors = <&tsc 1>; +@@ -2814,7 +2814,7 @@ + }; + }; + +- sensor_thermal3: sensor-thermal3 { ++ sensor3_thermal: sensor3-thermal { + polling-delay-passive = <250>; + polling-delay = <1000>; + thermal-sensors = <&tsc 2>; +diff --git a/arch/arm64/boot/dts/renesas/r8a77980.dtsi b/arch/arm64/boot/dts/renesas/r8a77980.dtsi +index 6347d15e66b64..21fe602bd25af 100644 +--- a/arch/arm64/boot/dts/renesas/r8a77980.dtsi ++++ b/arch/arm64/boot/dts/renesas/r8a77980.dtsi +@@ -1580,7 +1580,7 @@ + }; + + thermal-zones { +- thermal-sensor-1 { ++ sensor1_thermal: sensor1-thermal { + polling-delay-passive = <250>; + polling-delay = <1000>; + thermal-sensors = <&tsc 0>; +@@ -1599,7 +1599,7 @@ + }; + }; + +- thermal-sensor-2 { ++ sensor2_thermal: sensor2-thermal { + polling-delay-passive = <250>; + polling-delay = <1000>; + thermal-sensors = <&tsc 1>; +diff --git a/arch/arm64/boot/dts/renesas/r8a779a0.dtsi b/arch/arm64/boot/dts/renesas/r8a779a0.dtsi +index 631d520cebee5..26899fb768a73 100644 +--- a/arch/arm64/boot/dts/renesas/r8a779a0.dtsi ++++ b/arch/arm64/boot/dts/renesas/r8a779a0.dtsi +@@ -1149,7 +1149,7 @@ + }; + + thermal-zones { +- sensor_thermal1: sensor-thermal1 { ++ sensor1_thermal: sensor1-thermal { + polling-delay-passive = <250>; + polling-delay = <1000>; + thermal-sensors = <&tsc 0>; +@@ -1163,7 +1163,7 @@ + }; + }; + +- sensor_thermal2: sensor-thermal2 { ++ sensor2_thermal: sensor2-thermal { + polling-delay-passive = <250>; + polling-delay = <1000>; + thermal-sensors = <&tsc 1>; +@@ -1177,7 +1177,7 @@ + }; + }; + +- sensor_thermal3: sensor-thermal3 { ++ sensor3_thermal: sensor3-thermal { + polling-delay-passive = <250>; + polling-delay = <1000>; + thermal-sensors = <&tsc 2>; +@@ -1191,7 +1191,7 @@ + }; + }; + +- sensor_thermal4: sensor-thermal4 { ++ sensor4_thermal: sensor4-thermal { + polling-delay-passive = <250>; + polling-delay = <1000>; + thermal-sensors = <&tsc 3>; +@@ -1205,7 +1205,7 @@ + }; + }; + +- sensor_thermal5: sensor-thermal5 { ++ sensor5_thermal: sensor5-thermal { + polling-delay-passive = <250>; + polling-delay = <1000>; + thermal-sensors = <&tsc 4>; +diff --git a/arch/arm64/boot/dts/ti/k3-am642.dtsi b/arch/arm64/boot/dts/ti/k3-am642.dtsi +index e2b397c884018..8a76f4821b11b 100644 +--- a/arch/arm64/boot/dts/ti/k3-am642.dtsi ++++ b/arch/arm64/boot/dts/ti/k3-am642.dtsi +@@ -60,6 +60,6 @@ + cache-level = <2>; + cache-size = <0x40000>; + cache-line-size = <64>; +- cache-sets = <512>; ++ cache-sets = <256>; + }; + }; +diff --git a/arch/arm64/boot/dts/ti/k3-j7200-main.dtsi b/arch/arm64/boot/dts/ti/k3-j7200-main.dtsi +index 874cba75e9a5a..7daa280220442 100644 +--- a/arch/arm64/boot/dts/ti/k3-j7200-main.dtsi ++++ b/arch/arm64/boot/dts/ti/k3-j7200-main.dtsi +@@ -32,7 +32,7 @@ + #size-cells = <1>; + ranges = <0x00 0x00 0x00100000 0x1c000>; + +- serdes_ln_ctrl: serdes-ln-ctrl@4080 { ++ serdes_ln_ctrl: mux-controller@4080 { + compatible = "mmio-mux"; + #mux-control-cells = <1>; + mux-reg-masks = <0x4080 0x3>, <0x4084 0x3>, /* SERDES0 lane0/1 select */ +diff --git a/arch/arm64/boot/dts/ti/k3-j7200.dtsi b/arch/arm64/boot/dts/ti/k3-j7200.dtsi +index b7005b8031495..7586b5aea446f 100644 +--- a/arch/arm64/boot/dts/ti/k3-j7200.dtsi ++++ b/arch/arm64/boot/dts/ti/k3-j7200.dtsi +@@ -60,7 +60,7 @@ + i-cache-sets = <256>; + d-cache-size = <0x8000>; + d-cache-line-size = <64>; +- d-cache-sets = <128>; ++ d-cache-sets = <256>; + next-level-cache = <&L2_0>; + }; + +@@ -74,7 +74,7 @@ + i-cache-sets = <256>; + d-cache-size = <0x8000>; + d-cache-line-size = <64>; +- d-cache-sets = <128>; ++ d-cache-sets = <256>; + next-level-cache = <&L2_0>; + }; + }; +@@ -84,7 +84,7 @@ + cache-level = <2>; + cache-size = <0x100000>; + cache-line-size = <64>; +- cache-sets = <2048>; ++ cache-sets = <1024>; + next-level-cache = <&msmc_l3>; + }; + +diff --git a/arch/arm64/boot/dts/ti/k3-j721e-main.dtsi b/arch/arm64/boot/dts/ti/k3-j721e-main.dtsi +index 08c8d1b47dcd9..e85c89eebfa31 100644 +--- a/arch/arm64/boot/dts/ti/k3-j721e-main.dtsi ++++ b/arch/arm64/boot/dts/ti/k3-j721e-main.dtsi +@@ -42,7 +42,7 @@ + #size-cells = <1>; + ranges = <0x0 0x0 0x00100000 0x1c000>; + +- serdes_ln_ctrl: mux@4080 { ++ serdes_ln_ctrl: mux-controller@4080 { + compatible = "mmio-mux"; + reg = <0x00004080 0x50>; + #mux-control-cells = <1>; +diff --git a/arch/arm64/boot/dts/ti/k3-j721e.dtsi b/arch/arm64/boot/dts/ti/k3-j721e.dtsi +index f0587fde147e6..69ce048a2136e 100644 +--- a/arch/arm64/boot/dts/ti/k3-j721e.dtsi ++++ b/arch/arm64/boot/dts/ti/k3-j721e.dtsi +@@ -61,7 +61,7 @@ + i-cache-sets = <256>; + d-cache-size = <0x8000>; + d-cache-line-size = <64>; +- d-cache-sets = <128>; ++ d-cache-sets = <256>; + next-level-cache = <&L2_0>; + }; + +@@ -75,7 +75,7 @@ + i-cache-sets = <256>; + d-cache-size = <0x8000>; + d-cache-line-size = <64>; +- d-cache-sets = <128>; ++ d-cache-sets = <256>; + next-level-cache = <&L2_0>; + }; + }; +@@ -85,7 +85,7 @@ + cache-level = <2>; + cache-size = <0x100000>; + cache-line-size = <64>; +- cache-sets = <2048>; ++ cache-sets = <1024>; + next-level-cache = <&msmc_l3>; + }; + +diff --git a/arch/arm64/include/asm/mte-kasan.h b/arch/arm64/include/asm/mte-kasan.h +index 22420e1f8c037..26e013e540ae2 100644 +--- a/arch/arm64/include/asm/mte-kasan.h ++++ b/arch/arm64/include/asm/mte-kasan.h +@@ -84,10 +84,12 @@ static inline void __dc_gzva(u64 p) + static inline void mte_set_mem_tag_range(void *addr, size_t size, u8 tag, + bool init) + { +- u64 curr, mask, dczid_bs, end1, end2, end3; ++ u64 curr, mask, dczid, dczid_bs, dczid_dzp, end1, end2, end3; + + /* Read DC G(Z)VA block size from the system register. */ +- dczid_bs = 4ul << (read_cpuid(DCZID_EL0) & 0xf); ++ dczid = read_cpuid(DCZID_EL0); ++ dczid_bs = 4ul << (dczid & 0xf); ++ dczid_dzp = (dczid >> 4) & 1; + + curr = (u64)__tag_set(addr, tag); + mask = dczid_bs - 1; +@@ -106,7 +108,7 @@ static inline void mte_set_mem_tag_range(void *addr, size_t size, u8 tag, + */ + #define SET_MEMTAG_RANGE(stg_post, dc_gva) \ + do { \ +- if (size >= 2 * dczid_bs) { \ ++ if (!dczid_dzp && size >= 2 * dczid_bs) {\ + do { \ + curr = stg_post(curr); \ + } while (curr < end1); \ +diff --git a/arch/arm64/kernel/process.c b/arch/arm64/kernel/process.c +index 40adb8cdbf5af..23efabcb00b85 100644 +--- a/arch/arm64/kernel/process.c ++++ b/arch/arm64/kernel/process.c +@@ -439,34 +439,26 @@ static void entry_task_switch(struct task_struct *next) + + /* + * ARM erratum 1418040 handling, affecting the 32bit view of CNTVCT. +- * Assuming the virtual counter is enabled at the beginning of times: +- * +- * - disable access when switching from a 64bit task to a 32bit task +- * - enable access when switching from a 32bit task to a 64bit task ++ * Ensure access is disabled when switching to a 32bit task, ensure ++ * access is enabled when switching to a 64bit task. + */ +-static void erratum_1418040_thread_switch(struct task_struct *prev, +- struct task_struct *next) ++static void erratum_1418040_thread_switch(struct task_struct *next) + { +- bool prev32, next32; +- u64 val; +- +- if (!IS_ENABLED(CONFIG_ARM64_ERRATUM_1418040)) +- return; +- +- prev32 = is_compat_thread(task_thread_info(prev)); +- next32 = is_compat_thread(task_thread_info(next)); +- +- if (prev32 == next32 || !this_cpu_has_cap(ARM64_WORKAROUND_1418040)) ++ if (!IS_ENABLED(CONFIG_ARM64_ERRATUM_1418040) || ++ !this_cpu_has_cap(ARM64_WORKAROUND_1418040)) + return; + +- val = read_sysreg(cntkctl_el1); +- +- if (!next32) +- val |= ARCH_TIMER_USR_VCT_ACCESS_EN; ++ if (is_compat_thread(task_thread_info(next))) ++ sysreg_clear_set(cntkctl_el1, ARCH_TIMER_USR_VCT_ACCESS_EN, 0); + else +- val &= ~ARCH_TIMER_USR_VCT_ACCESS_EN; ++ sysreg_clear_set(cntkctl_el1, 0, ARCH_TIMER_USR_VCT_ACCESS_EN); ++} + +- write_sysreg(val, cntkctl_el1); ++static void erratum_1418040_new_exec(void) ++{ ++ preempt_disable(); ++ erratum_1418040_thread_switch(current); ++ preempt_enable(); + } + + /* +@@ -501,7 +493,7 @@ __notrace_funcgraph struct task_struct *__switch_to(struct task_struct *prev, + contextidr_thread_switch(next); + entry_task_switch(next); + ssbs_thread_switch(next); +- erratum_1418040_thread_switch(prev, next); ++ erratum_1418040_thread_switch(next); + ptrauth_thread_switch_user(next); + + /* +@@ -613,6 +605,7 @@ void arch_setup_new_exec(void) + current->mm->context.flags = mmflags; + ptrauth_thread_init_user(); + mte_thread_init_user(); ++ erratum_1418040_new_exec(); + + if (task_spec_ssb_noexec(current)) { + arch_prctl_spec_ctrl_set(current, PR_SPEC_STORE_BYPASS, +diff --git a/arch/arm64/lib/clear_page.S b/arch/arm64/lib/clear_page.S +index b84b179edba3a..1fd5d790ab800 100644 +--- a/arch/arm64/lib/clear_page.S ++++ b/arch/arm64/lib/clear_page.S +@@ -16,6 +16,7 @@ + */ + SYM_FUNC_START_PI(clear_page) + mrs x1, dczid_el0 ++ tbnz x1, #4, 2f /* Branch if DC ZVA is prohibited */ + and w1, w1, #0xf + mov x2, #4 + lsl x1, x2, x1 +@@ -25,5 +26,14 @@ SYM_FUNC_START_PI(clear_page) + tst x0, #(PAGE_SIZE - 1) + b.ne 1b + ret ++ ++2: stnp xzr, xzr, [x0] ++ stnp xzr, xzr, [x0, #16] ++ stnp xzr, xzr, [x0, #32] ++ stnp xzr, xzr, [x0, #48] ++ add x0, x0, #64 ++ tst x0, #(PAGE_SIZE - 1) ++ b.ne 2b ++ ret + SYM_FUNC_END_PI(clear_page) + EXPORT_SYMBOL(clear_page) +diff --git a/arch/arm64/lib/mte.S b/arch/arm64/lib/mte.S +index e83643b3995f4..f531dcb95174a 100644 +--- a/arch/arm64/lib/mte.S ++++ b/arch/arm64/lib/mte.S +@@ -43,17 +43,23 @@ SYM_FUNC_END(mte_clear_page_tags) + * x0 - address to the beginning of the page + */ + SYM_FUNC_START(mte_zero_clear_page_tags) ++ and x0, x0, #(1 << MTE_TAG_SHIFT) - 1 // clear the tag + mrs x1, dczid_el0 ++ tbnz x1, #4, 2f // Branch if DC GZVA is prohibited + and w1, w1, #0xf + mov x2, #4 + lsl x1, x2, x1 +- and x0, x0, #(1 << MTE_TAG_SHIFT) - 1 // clear the tag + + 1: dc gzva, x0 + add x0, x0, x1 + tst x0, #(PAGE_SIZE - 1) + b.ne 1b + ret ++ ++2: stz2g x0, [x0], #(MTE_GRANULE_SIZE * 2) ++ tst x0, #(PAGE_SIZE - 1) ++ b.ne 2b ++ ret + SYM_FUNC_END(mte_zero_clear_page_tags) + + /* +diff --git a/arch/mips/Kconfig b/arch/mips/Kconfig +index 23654ccdbfb12..393eb2133243f 100644 +--- a/arch/mips/Kconfig ++++ b/arch/mips/Kconfig +@@ -1993,6 +1993,10 @@ config SYS_HAS_CPU_MIPS64_R1 + config SYS_HAS_CPU_MIPS64_R2 + bool + ++config SYS_HAS_CPU_MIPS64_R5 ++ bool ++ select ARCH_HAS_SYNC_DMA_FOR_CPU if DMA_NONCOHERENT ++ + config SYS_HAS_CPU_MIPS64_R6 + bool + select ARCH_HAS_SYNC_DMA_FOR_CPU if DMA_NONCOHERENT +@@ -2157,7 +2161,7 @@ config CPU_SUPPORTS_ADDRWINCFG + bool + config CPU_SUPPORTS_HUGEPAGES + bool +- depends on !(32BIT && (ARCH_PHYS_ADDR_T_64BIT || EVA)) ++ depends on !(32BIT && (PHYS_ADDR_T_64BIT || EVA)) + config MIPS_PGD_C0_CONTEXT + bool + depends on 64BIT +diff --git a/arch/mips/bcm63xx/clk.c b/arch/mips/bcm63xx/clk.c +index 1c91064cb448b..6e6756e8fa0a9 100644 +--- a/arch/mips/bcm63xx/clk.c ++++ b/arch/mips/bcm63xx/clk.c +@@ -387,6 +387,12 @@ struct clk *clk_get_parent(struct clk *clk) + } + EXPORT_SYMBOL(clk_get_parent); + ++int clk_set_parent(struct clk *clk, struct clk *parent) ++{ ++ return 0; ++} ++EXPORT_SYMBOL(clk_set_parent); ++ + unsigned long clk_get_rate(struct clk *clk) + { + if (!clk) +diff --git a/arch/mips/boot/compressed/Makefile b/arch/mips/boot/compressed/Makefile +index 9112bdb86be45..705b9e7f8035a 100644 +--- a/arch/mips/boot/compressed/Makefile ++++ b/arch/mips/boot/compressed/Makefile +@@ -56,7 +56,7 @@ $(obj)/uart-ath79.c: $(srctree)/arch/mips/ath79/early_printk.c + + vmlinuzobjs-$(CONFIG_KERNEL_XZ) += $(obj)/ashldi3.o + +-vmlinuzobjs-$(CONFIG_KERNEL_ZSTD) += $(obj)/bswapdi.o ++vmlinuzobjs-$(CONFIG_KERNEL_ZSTD) += $(obj)/bswapdi.o $(obj)/ashldi3.o $(obj)/clz_ctz.o + + extra-y += ashldi3.c + $(obj)/ashldi3.c: $(obj)/%.c: $(srctree)/lib/%.c FORCE +diff --git a/arch/mips/boot/compressed/clz_ctz.c b/arch/mips/boot/compressed/clz_ctz.c +new file mode 100644 +index 0000000000000..b4a1b6eb2f8ad +--- /dev/null ++++ b/arch/mips/boot/compressed/clz_ctz.c +@@ -0,0 +1,2 @@ ++// SPDX-License-Identifier: GPL-2.0-only ++#include "../../../../lib/clz_ctz.c" +diff --git a/arch/mips/cavium-octeon/octeon-platform.c b/arch/mips/cavium-octeon/octeon-platform.c +index d56e9b9d2e434..a994022e32c9f 100644 +--- a/arch/mips/cavium-octeon/octeon-platform.c ++++ b/arch/mips/cavium-octeon/octeon-platform.c +@@ -328,6 +328,7 @@ static int __init octeon_ehci_device_init(void) + + pd->dev.platform_data = &octeon_ehci_pdata; + octeon_ehci_hw_start(&pd->dev); ++ put_device(&pd->dev); + + return ret; + } +@@ -391,6 +392,7 @@ static int __init octeon_ohci_device_init(void) + + pd->dev.platform_data = &octeon_ohci_pdata; + octeon_ohci_hw_start(&pd->dev); ++ put_device(&pd->dev); + + return ret; + } +diff --git a/arch/mips/cavium-octeon/octeon-usb.c b/arch/mips/cavium-octeon/octeon-usb.c +index 6e4d3619137af..4df919d26b082 100644 +--- a/arch/mips/cavium-octeon/octeon-usb.c ++++ b/arch/mips/cavium-octeon/octeon-usb.c +@@ -537,6 +537,7 @@ static int __init dwc3_octeon_device_init(void) + devm_iounmap(&pdev->dev, base); + devm_release_mem_region(&pdev->dev, res->start, + resource_size(res)); ++ put_device(&pdev->dev); + } + } while (node != NULL); + +diff --git a/arch/mips/configs/fuloong2e_defconfig b/arch/mips/configs/fuloong2e_defconfig +index 5c24ac7fdf56d..ba47c5e929b7f 100644 +--- a/arch/mips/configs/fuloong2e_defconfig ++++ b/arch/mips/configs/fuloong2e_defconfig +@@ -206,7 +206,6 @@ CONFIG_NFSD_V3_ACL=y + CONFIG_NFSD_V4=y + CONFIG_CIFS=m + CONFIG_CIFS_STATS2=y +-CONFIG_CIFS_WEAK_PW_HASH=y + CONFIG_CIFS_XATTR=y + CONFIG_CIFS_POSIX=y + CONFIG_CIFS_DEBUG2=y +diff --git a/arch/mips/configs/malta_qemu_32r6_defconfig b/arch/mips/configs/malta_qemu_32r6_defconfig +index 614af02d83e6e..6fb9bc29f4a03 100644 +--- a/arch/mips/configs/malta_qemu_32r6_defconfig ++++ b/arch/mips/configs/malta_qemu_32r6_defconfig +@@ -165,7 +165,6 @@ CONFIG_TMPFS=y + CONFIG_NFS_FS=y + CONFIG_ROOT_NFS=y + CONFIG_CIFS=m +-CONFIG_CIFS_WEAK_PW_HASH=y + CONFIG_CIFS_XATTR=y + CONFIG_CIFS_POSIX=y + CONFIG_NLS_CODEPAGE_437=m +diff --git a/arch/mips/configs/maltaaprp_defconfig b/arch/mips/configs/maltaaprp_defconfig +index 9c051f8fd3300..eb72df528243a 100644 +--- a/arch/mips/configs/maltaaprp_defconfig ++++ b/arch/mips/configs/maltaaprp_defconfig +@@ -166,7 +166,6 @@ CONFIG_TMPFS=y + CONFIG_NFS_FS=y + CONFIG_ROOT_NFS=y + CONFIG_CIFS=m +-CONFIG_CIFS_WEAK_PW_HASH=y + CONFIG_CIFS_XATTR=y + CONFIG_CIFS_POSIX=y + CONFIG_NLS_CODEPAGE_437=m +diff --git a/arch/mips/configs/maltasmvp_defconfig b/arch/mips/configs/maltasmvp_defconfig +index 2e90d97551d6f..1fb40d310f49c 100644 +--- a/arch/mips/configs/maltasmvp_defconfig ++++ b/arch/mips/configs/maltasmvp_defconfig +@@ -167,7 +167,6 @@ CONFIG_TMPFS=y + CONFIG_NFS_FS=y + CONFIG_ROOT_NFS=y + CONFIG_CIFS=m +-CONFIG_CIFS_WEAK_PW_HASH=y + CONFIG_CIFS_XATTR=y + CONFIG_CIFS_POSIX=y + CONFIG_NLS_CODEPAGE_437=m +diff --git a/arch/mips/configs/maltasmvp_eva_defconfig b/arch/mips/configs/maltasmvp_eva_defconfig +index d1f7fdb27284b..75cb778c61496 100644 +--- a/arch/mips/configs/maltasmvp_eva_defconfig ++++ b/arch/mips/configs/maltasmvp_eva_defconfig +@@ -169,7 +169,6 @@ CONFIG_TMPFS=y + CONFIG_NFS_FS=y + CONFIG_ROOT_NFS=y + CONFIG_CIFS=m +-CONFIG_CIFS_WEAK_PW_HASH=y + CONFIG_CIFS_XATTR=y + CONFIG_CIFS_POSIX=y + CONFIG_NLS_CODEPAGE_437=m +diff --git a/arch/mips/configs/maltaup_defconfig b/arch/mips/configs/maltaup_defconfig +index 48e5bd4924522..7b4f247dc60cc 100644 +--- a/arch/mips/configs/maltaup_defconfig ++++ b/arch/mips/configs/maltaup_defconfig +@@ -165,7 +165,6 @@ CONFIG_TMPFS=y + CONFIG_NFS_FS=y + CONFIG_ROOT_NFS=y + CONFIG_CIFS=m +-CONFIG_CIFS_WEAK_PW_HASH=y + CONFIG_CIFS_XATTR=y + CONFIG_CIFS_POSIX=y + CONFIG_NLS_CODEPAGE_437=m +diff --git a/arch/mips/include/asm/mach-loongson64/kernel-entry-init.h b/arch/mips/include/asm/mach-loongson64/kernel-entry-init.h +index 13373c5144f89..efb41b3519747 100644 +--- a/arch/mips/include/asm/mach-loongson64/kernel-entry-init.h ++++ b/arch/mips/include/asm/mach-loongson64/kernel-entry-init.h +@@ -32,7 +32,7 @@ + nop + /* Loongson-3A R2/R3 */ + andi t0, (PRID_IMP_MASK | PRID_REV_MASK) +- slti t0, (PRID_IMP_LOONGSON_64C | PRID_REV_LOONGSON3A_R2_0) ++ slti t0, t0, (PRID_IMP_LOONGSON_64C | PRID_REV_LOONGSON3A_R2_0) + bnez t0, 2f + nop + 1: +@@ -63,7 +63,7 @@ + nop + /* Loongson-3A R2/R3 */ + andi t0, (PRID_IMP_MASK | PRID_REV_MASK) +- slti t0, (PRID_IMP_LOONGSON_64C | PRID_REV_LOONGSON3A_R2_0) ++ slti t0, t0, (PRID_IMP_LOONGSON_64C | PRID_REV_LOONGSON3A_R2_0) + bnez t0, 2f + nop + 1: +diff --git a/arch/mips/include/asm/octeon/cvmx-bootinfo.h b/arch/mips/include/asm/octeon/cvmx-bootinfo.h +index 0e6bf220db618..6c61e0a639249 100644 +--- a/arch/mips/include/asm/octeon/cvmx-bootinfo.h ++++ b/arch/mips/include/asm/octeon/cvmx-bootinfo.h +@@ -318,7 +318,7 @@ enum cvmx_chip_types_enum { + + /* Functions to return string based on type */ + #define ENUM_BRD_TYPE_CASE(x) \ +- case x: return(#x + 16); /* Skip CVMX_BOARD_TYPE_ */ ++ case x: return (&#x[16]); /* Skip CVMX_BOARD_TYPE_ */ + static inline const char *cvmx_board_type_to_string(enum + cvmx_board_types_enum type) + { +@@ -410,7 +410,7 @@ static inline const char *cvmx_board_type_to_string(enum + } + + #define ENUM_CHIP_TYPE_CASE(x) \ +- case x: return(#x + 15); /* Skip CVMX_CHIP_TYPE */ ++ case x: return (&#x[15]); /* Skip CVMX_CHIP_TYPE */ + static inline const char *cvmx_chip_type_to_string(enum + cvmx_chip_types_enum type) + { +diff --git a/arch/mips/lantiq/clk.c b/arch/mips/lantiq/clk.c +index 4916cccf378fd..7a623684d9b5e 100644 +--- a/arch/mips/lantiq/clk.c ++++ b/arch/mips/lantiq/clk.c +@@ -164,6 +164,12 @@ struct clk *clk_get_parent(struct clk *clk) + } + EXPORT_SYMBOL(clk_get_parent); + ++int clk_set_parent(struct clk *clk, struct clk *parent) ++{ ++ return 0; ++} ++EXPORT_SYMBOL(clk_set_parent); ++ + static inline u32 get_counter_resolution(void) + { + u32 res; +diff --git a/arch/openrisc/include/asm/syscalls.h b/arch/openrisc/include/asm/syscalls.h +index 3a7eeae6f56a8..aa1c7e98722e3 100644 +--- a/arch/openrisc/include/asm/syscalls.h ++++ b/arch/openrisc/include/asm/syscalls.h +@@ -22,9 +22,11 @@ asmlinkage long sys_or1k_atomic(unsigned long type, unsigned long *v1, + + asmlinkage long __sys_clone(unsigned long clone_flags, unsigned long newsp, + void __user *parent_tid, void __user *child_tid, int tls); ++asmlinkage long __sys_clone3(struct clone_args __user *uargs, size_t size); + asmlinkage long __sys_fork(void); + + #define sys_clone __sys_clone ++#define sys_clone3 __sys_clone3 + #define sys_fork __sys_fork + + #endif /* __ASM_OPENRISC_SYSCALLS_H */ +diff --git a/arch/openrisc/kernel/entry.S b/arch/openrisc/kernel/entry.S +index edaa775a648e6..c68f3349c1741 100644 +--- a/arch/openrisc/kernel/entry.S ++++ b/arch/openrisc/kernel/entry.S +@@ -1170,6 +1170,11 @@ ENTRY(__sys_clone) + l.j _fork_save_extra_regs_and_call + l.nop + ++ENTRY(__sys_clone3) ++ l.movhi r29,hi(sys_clone3) ++ l.j _fork_save_extra_regs_and_call ++ l.ori r29,r29,lo(sys_clone3) ++ + ENTRY(__sys_fork) + l.movhi r29,hi(sys_fork) + l.ori r29,r29,lo(sys_fork) +diff --git a/arch/parisc/include/asm/special_insns.h b/arch/parisc/include/asm/special_insns.h +index a303ae9a77f41..16ee41e77174f 100644 +--- a/arch/parisc/include/asm/special_insns.h ++++ b/arch/parisc/include/asm/special_insns.h +@@ -2,28 +2,32 @@ + #ifndef __PARISC_SPECIAL_INSNS_H + #define __PARISC_SPECIAL_INSNS_H + +-#define lpa(va) ({ \ +- unsigned long pa; \ +- __asm__ __volatile__( \ +- "copy %%r0,%0\n\t" \ +- "lpa %%r0(%1),%0" \ +- : "=r" (pa) \ +- : "r" (va) \ +- : "memory" \ +- ); \ +- pa; \ ++#define lpa(va) ({ \ ++ unsigned long pa; \ ++ __asm__ __volatile__( \ ++ "copy %%r0,%0\n" \ ++ "8:\tlpa %%r0(%1),%0\n" \ ++ "9:\n" \ ++ ASM_EXCEPTIONTABLE_ENTRY(8b, 9b) \ ++ : "=&r" (pa) \ ++ : "r" (va) \ ++ : "memory" \ ++ ); \ ++ pa; \ + }) + +-#define lpa_user(va) ({ \ +- unsigned long pa; \ +- __asm__ __volatile__( \ +- "copy %%r0,%0\n\t" \ +- "lpa %%r0(%%sr3,%1),%0" \ +- : "=r" (pa) \ +- : "r" (va) \ +- : "memory" \ +- ); \ +- pa; \ ++#define lpa_user(va) ({ \ ++ unsigned long pa; \ ++ __asm__ __volatile__( \ ++ "copy %%r0,%0\n" \ ++ "8:\tlpa %%r0(%%sr3,%1),%0\n" \ ++ "9:\n" \ ++ ASM_EXCEPTIONTABLE_ENTRY(8b, 9b) \ ++ : "=&r" (pa) \ ++ : "r" (va) \ ++ : "memory" \ ++ ); \ ++ pa; \ + }) + + #define mfctl(reg) ({ \ +diff --git a/arch/parisc/kernel/traps.c b/arch/parisc/kernel/traps.c +index 197cb8480350c..afe8b902a8fc4 100644 +--- a/arch/parisc/kernel/traps.c ++++ b/arch/parisc/kernel/traps.c +@@ -784,7 +784,7 @@ void notrace handle_interruption(int code, struct pt_regs *regs) + * unless pagefault_disable() was called before. + */ + +- if (fault_space == 0 && !faulthandler_disabled()) ++ if (faulthandler_disabled() || fault_space == 0) + { + /* Clean up and return if in exception table. */ + if (fixup_exception(regs)) +diff --git a/arch/powerpc/boot/dts/fsl/qoriq-fman3l-0.dtsi b/arch/powerpc/boot/dts/fsl/qoriq-fman3l-0.dtsi +index c90702b04a530..48e5cd61599c6 100644 +--- a/arch/powerpc/boot/dts/fsl/qoriq-fman3l-0.dtsi ++++ b/arch/powerpc/boot/dts/fsl/qoriq-fman3l-0.dtsi +@@ -79,6 +79,7 @@ fman0: fman@400000 { + #size-cells = <0>; + compatible = "fsl,fman-memac-mdio", "fsl,fman-xmdio"; + reg = <0xfc000 0x1000>; ++ fsl,erratum-a009885; + }; + + xmdio0: mdio@fd000 { +@@ -86,6 +87,7 @@ fman0: fman@400000 { + #size-cells = <0>; + compatible = "fsl,fman-memac-mdio", "fsl,fman-xmdio"; + reg = <0xfd000 0x1000>; ++ fsl,erratum-a009885; + }; + }; + +diff --git a/arch/powerpc/configs/ppc6xx_defconfig b/arch/powerpc/configs/ppc6xx_defconfig +index 6697c5e6682f1..bb549cb1c3e33 100644 +--- a/arch/powerpc/configs/ppc6xx_defconfig ++++ b/arch/powerpc/configs/ppc6xx_defconfig +@@ -1022,7 +1022,6 @@ CONFIG_NFSD=m + CONFIG_NFSD_V3_ACL=y + CONFIG_NFSD_V4=y + CONFIG_CIFS=m +-CONFIG_CIFS_WEAK_PW_HASH=y + CONFIG_CIFS_UPCALL=y + CONFIG_CIFS_XATTR=y + CONFIG_CIFS_POSIX=y +diff --git a/arch/powerpc/configs/pseries_defconfig b/arch/powerpc/configs/pseries_defconfig +index b183629f1bcfb..d0494fbb49617 100644 +--- a/arch/powerpc/configs/pseries_defconfig ++++ b/arch/powerpc/configs/pseries_defconfig +@@ -190,7 +190,6 @@ CONFIG_HVCS=m + CONFIG_VIRTIO_CONSOLE=m + CONFIG_IBM_BSR=m + CONFIG_RAW_DRIVER=y +-CONFIG_MAX_RAW_DEVS=1024 + CONFIG_I2C_CHARDEV=y + CONFIG_FB=y + CONFIG_FIRMWARE_EDID=y +diff --git a/arch/powerpc/include/asm/hw_irq.h b/arch/powerpc/include/asm/hw_irq.h +index 21cc571ea9c2d..5c98a950eca0d 100644 +--- a/arch/powerpc/include/asm/hw_irq.h ++++ b/arch/powerpc/include/asm/hw_irq.h +@@ -224,6 +224,42 @@ static inline bool arch_irqs_disabled(void) + return arch_irqs_disabled_flags(arch_local_save_flags()); + } + ++static inline void set_pmi_irq_pending(void) ++{ ++ /* ++ * Invoked from PMU callback functions to set PMI bit in the paca. ++ * This has to be called with irq's disabled (via hard_irq_disable()). ++ */ ++ if (IS_ENABLED(CONFIG_PPC_IRQ_SOFT_MASK_DEBUG)) ++ WARN_ON_ONCE(mfmsr() & MSR_EE); ++ ++ get_paca()->irq_happened |= PACA_IRQ_PMI; ++} ++ ++static inline void clear_pmi_irq_pending(void) ++{ ++ /* ++ * Invoked from PMU callback functions to clear the pending PMI bit ++ * in the paca. ++ */ ++ if (IS_ENABLED(CONFIG_PPC_IRQ_SOFT_MASK_DEBUG)) ++ WARN_ON_ONCE(mfmsr() & MSR_EE); ++ ++ get_paca()->irq_happened &= ~PACA_IRQ_PMI; ++} ++ ++static inline bool pmi_irq_pending(void) ++{ ++ /* ++ * Invoked from PMU callback functions to check if there is a pending ++ * PMI bit in the paca. ++ */ ++ if (get_paca()->irq_happened & PACA_IRQ_PMI) ++ return true; ++ ++ return false; ++} ++ + #ifdef CONFIG_PPC_BOOK3S + /* + * To support disabling and enabling of irq with PMI, set of +@@ -408,6 +444,10 @@ static inline void do_hard_irq_enable(void) + BUILD_BUG(); + } + ++static inline void clear_pmi_irq_pending(void) { } ++static inline void set_pmi_irq_pending(void) { } ++static inline bool pmi_irq_pending(void) { return false; } ++ + static inline void irq_soft_mask_regs_set_state(struct pt_regs *regs, unsigned long val) + { + } +diff --git a/arch/powerpc/kernel/btext.c b/arch/powerpc/kernel/btext.c +index 803c2a45b22ac..1cffb5e7c38d6 100644 +--- a/arch/powerpc/kernel/btext.c ++++ b/arch/powerpc/kernel/btext.c +@@ -241,8 +241,10 @@ int __init btext_find_display(int allow_nonstdout) + rc = btext_initialize(np); + printk("result: %d\n", rc); + } +- if (rc == 0) ++ if (rc == 0) { ++ of_node_put(np); + break; ++ } + } + return rc; + } +diff --git a/arch/powerpc/kernel/fadump.c b/arch/powerpc/kernel/fadump.c +index b7ceb041743c9..60f5fc14aa235 100644 +--- a/arch/powerpc/kernel/fadump.c ++++ b/arch/powerpc/kernel/fadump.c +@@ -1641,6 +1641,14 @@ int __init setup_fadump(void) + else if (fw_dump.reserve_dump_area_size) + fw_dump.ops->fadump_init_mem_struct(&fw_dump); + ++ /* ++ * In case of panic, fadump is triggered via ppc_panic_event() ++ * panic notifier. Setting crash_kexec_post_notifiers to 'true' ++ * lets panic() function take crash friendly path before panic ++ * notifiers are invoked. ++ */ ++ crash_kexec_post_notifiers = true; ++ + return 1; + } + subsys_initcall(setup_fadump); +diff --git a/arch/powerpc/kernel/head_40x.S b/arch/powerpc/kernel/head_40x.S +index 7d72ee5ab387c..e783860bea838 100644 +--- a/arch/powerpc/kernel/head_40x.S ++++ b/arch/powerpc/kernel/head_40x.S +@@ -27,6 +27,7 @@ + + #include <linux/init.h> + #include <linux/pgtable.h> ++#include <linux/sizes.h> + #include <asm/processor.h> + #include <asm/page.h> + #include <asm/mmu.h> +@@ -650,7 +651,7 @@ start_here: + b . /* prevent prefetch past rfi */ + + /* Set up the initial MMU state so we can do the first level of +- * kernel initialization. This maps the first 16 MBytes of memory 1:1 ++ * kernel initialization. This maps the first 32 MBytes of memory 1:1 + * virtual to physical and more importantly sets the cache mode. + */ + initial_mmu: +@@ -687,6 +688,12 @@ initial_mmu: + tlbwe r4,r0,TLB_DATA /* Load the data portion of the entry */ + tlbwe r3,r0,TLB_TAG /* Load the tag portion of the entry */ + ++ li r0,62 /* TLB slot 62 */ ++ addis r4,r4,SZ_16M@h ++ addis r3,r3,SZ_16M@h ++ tlbwe r4,r0,TLB_DATA /* Load the data portion of the entry */ ++ tlbwe r3,r0,TLB_TAG /* Load the tag portion of the entry */ ++ + isync + + /* Establish the exception vector base +diff --git a/arch/powerpc/kernel/interrupt.c b/arch/powerpc/kernel/interrupt.c +index 835b626cd4760..df048e331cbfe 100644 +--- a/arch/powerpc/kernel/interrupt.c ++++ b/arch/powerpc/kernel/interrupt.c +@@ -148,7 +148,7 @@ notrace long system_call_exception(long r3, long r4, long r5, + */ + if (IS_ENABLED(CONFIG_PPC_TRANSACTIONAL_MEM) && + unlikely(MSR_TM_TRANSACTIONAL(regs->msr))) +- current_thread_info()->flags |= _TIF_RESTOREALL; ++ set_bits(_TIF_RESTOREALL, ¤t_thread_info()->flags); + + /* + * If the system call was made with a transaction active, doom it and +diff --git a/arch/powerpc/kernel/interrupt_64.S b/arch/powerpc/kernel/interrupt_64.S +index ec950b08a8dcc..4b1ff94e67eb4 100644 +--- a/arch/powerpc/kernel/interrupt_64.S ++++ b/arch/powerpc/kernel/interrupt_64.S +@@ -30,21 +30,23 @@ COMPAT_SYS_CALL_TABLE: + .ifc \srr,srr + mfspr r11,SPRN_SRR0 + ld r12,_NIP(r1) ++ clrrdi r12,r12,2 + 100: tdne r11,r12 +- EMIT_BUG_ENTRY 100b,__FILE__,__LINE__,(BUGFLAG_WARNING | BUGFLAG_ONCE) ++ EMIT_WARN_ENTRY 100b,__FILE__,__LINE__,(BUGFLAG_WARNING | BUGFLAG_ONCE) + mfspr r11,SPRN_SRR1 + ld r12,_MSR(r1) + 100: tdne r11,r12 +- EMIT_BUG_ENTRY 100b,__FILE__,__LINE__,(BUGFLAG_WARNING | BUGFLAG_ONCE) ++ EMIT_WARN_ENTRY 100b,__FILE__,__LINE__,(BUGFLAG_WARNING | BUGFLAG_ONCE) + .else + mfspr r11,SPRN_HSRR0 + ld r12,_NIP(r1) ++ clrrdi r12,r12,2 + 100: tdne r11,r12 +- EMIT_BUG_ENTRY 100b,__FILE__,__LINE__,(BUGFLAG_WARNING | BUGFLAG_ONCE) ++ EMIT_WARN_ENTRY 100b,__FILE__,__LINE__,(BUGFLAG_WARNING | BUGFLAG_ONCE) + mfspr r11,SPRN_HSRR1 + ld r12,_MSR(r1) + 100: tdne r11,r12 +- EMIT_BUG_ENTRY 100b,__FILE__,__LINE__,(BUGFLAG_WARNING | BUGFLAG_ONCE) ++ EMIT_WARN_ENTRY 100b,__FILE__,__LINE__,(BUGFLAG_WARNING | BUGFLAG_ONCE) + .endif + #endif + .endm +diff --git a/arch/powerpc/kernel/module.c b/arch/powerpc/kernel/module.c +index ed04a3ba66fe8..40a583e9d3c70 100644 +--- a/arch/powerpc/kernel/module.c ++++ b/arch/powerpc/kernel/module.c +@@ -90,16 +90,17 @@ int module_finalize(const Elf_Ehdr *hdr, + } + + static __always_inline void * +-__module_alloc(unsigned long size, unsigned long start, unsigned long end) ++__module_alloc(unsigned long size, unsigned long start, unsigned long end, bool nowarn) + { + pgprot_t prot = strict_module_rwx_enabled() ? PAGE_KERNEL : PAGE_KERNEL_EXEC; ++ gfp_t gfp = GFP_KERNEL | (nowarn ? __GFP_NOWARN : 0); + + /* + * Don't do huge page allocations for modules yet until more testing + * is done. STRICT_MODULE_RWX may require extra work to support this + * too. + */ +- return __vmalloc_node_range(size, 1, start, end, GFP_KERNEL, prot, ++ return __vmalloc_node_range(size, 1, start, end, gfp, prot, + VM_FLUSH_RESET_PERMS | VM_NO_HUGE_VMAP, + NUMA_NO_NODE, __builtin_return_address(0)); + } +@@ -114,13 +115,13 @@ void *module_alloc(unsigned long size) + + /* First try within 32M limit from _etext to avoid branch trampolines */ + if (MODULES_VADDR < PAGE_OFFSET && MODULES_END > limit) +- ptr = __module_alloc(size, limit, MODULES_END); ++ ptr = __module_alloc(size, limit, MODULES_END, true); + + if (!ptr) +- ptr = __module_alloc(size, MODULES_VADDR, MODULES_END); ++ ptr = __module_alloc(size, MODULES_VADDR, MODULES_END, false); + + return ptr; + #else +- return __module_alloc(size, VMALLOC_START, VMALLOC_END); ++ return __module_alloc(size, VMALLOC_START, VMALLOC_END, false); + #endif + } +diff --git a/arch/powerpc/kernel/prom_init.c b/arch/powerpc/kernel/prom_init.c +index 18b04b08b9833..f845065c860e3 100644 +--- a/arch/powerpc/kernel/prom_init.c ++++ b/arch/powerpc/kernel/prom_init.c +@@ -2991,7 +2991,7 @@ static void __init fixup_device_tree_efika_add_phy(void) + + /* Check if the phy-handle property exists - bail if it does */ + rv = prom_getprop(node, "phy-handle", prop, sizeof(prop)); +- if (!rv) ++ if (rv <= 0) + return; + + /* +diff --git a/arch/powerpc/kernel/smp.c b/arch/powerpc/kernel/smp.c +index 605bab448f847..fb95f92dcfac6 100644 +--- a/arch/powerpc/kernel/smp.c ++++ b/arch/powerpc/kernel/smp.c +@@ -61,6 +61,7 @@ + #include <asm/cpu_has_feature.h> + #include <asm/ftrace.h> + #include <asm/kup.h> ++#include <asm/fadump.h> + + #ifdef DEBUG + #include <asm/udbg.h> +@@ -620,6 +621,45 @@ void crash_send_ipi(void (*crash_ipi_callback)(struct pt_regs *)) + } + #endif + ++#ifdef CONFIG_NMI_IPI ++static void crash_stop_this_cpu(struct pt_regs *regs) ++#else ++static void crash_stop_this_cpu(void *dummy) ++#endif ++{ ++ /* ++ * Just busy wait here and avoid marking CPU as offline to ensure ++ * register data is captured appropriately. ++ */ ++ while (1) ++ cpu_relax(); ++} ++ ++void crash_smp_send_stop(void) ++{ ++ static bool stopped = false; ++ ++ /* ++ * In case of fadump, register data for all CPUs is captured by f/w ++ * on ibm,os-term rtas call. Skip IPI callbacks to other CPUs before ++ * this rtas call to avoid tricky post processing of those CPUs' ++ * backtraces. ++ */ ++ if (should_fadump_crash()) ++ return; ++ ++ if (stopped) ++ return; ++ ++ stopped = true; ++ ++#ifdef CONFIG_NMI_IPI ++ smp_send_nmi_ipi(NMI_IPI_ALL_OTHERS, crash_stop_this_cpu, 1000000); ++#else ++ smp_call_function(crash_stop_this_cpu, NULL, 0); ++#endif /* CONFIG_NMI_IPI */ ++} ++ + #ifdef CONFIG_NMI_IPI + static void nmi_stop_this_cpu(struct pt_regs *regs) + { +@@ -1640,10 +1680,12 @@ void start_secondary(void *unused) + BUG(); + } + ++#ifdef CONFIG_PROFILING + int setup_profiling_timer(unsigned int multiplier) + { + return 0; + } ++#endif + + static void fixup_topology(void) + { +diff --git a/arch/powerpc/kernel/watchdog.c b/arch/powerpc/kernel/watchdog.c +index 3fa6d240bade2..ad94a2c6b7337 100644 +--- a/arch/powerpc/kernel/watchdog.c ++++ b/arch/powerpc/kernel/watchdog.c +@@ -135,6 +135,10 @@ static void set_cpumask_stuck(const struct cpumask *cpumask, u64 tb) + { + cpumask_or(&wd_smp_cpus_stuck, &wd_smp_cpus_stuck, cpumask); + cpumask_andnot(&wd_smp_cpus_pending, &wd_smp_cpus_pending, cpumask); ++ /* ++ * See wd_smp_clear_cpu_pending() ++ */ ++ smp_mb(); + if (cpumask_empty(&wd_smp_cpus_pending)) { + wd_smp_last_reset_tb = tb; + cpumask_andnot(&wd_smp_cpus_pending, +@@ -221,13 +225,44 @@ static void wd_smp_clear_cpu_pending(int cpu, u64 tb) + + cpumask_clear_cpu(cpu, &wd_smp_cpus_stuck); + wd_smp_unlock(&flags); ++ } else { ++ /* ++ * The last CPU to clear pending should have reset the ++ * watchdog so we generally should not find it empty ++ * here if our CPU was clear. However it could happen ++ * due to a rare race with another CPU taking the ++ * last CPU out of the mask concurrently. ++ * ++ * We can't add a warning for it. But just in case ++ * there is a problem with the watchdog that is causing ++ * the mask to not be reset, try to kick it along here. ++ */ ++ if (unlikely(cpumask_empty(&wd_smp_cpus_pending))) ++ goto none_pending; + } + return; + } ++ + cpumask_clear_cpu(cpu, &wd_smp_cpus_pending); ++ ++ /* ++ * Order the store to clear pending with the load(s) to check all ++ * words in the pending mask to check they are all empty. This orders ++ * with the same barrier on another CPU. This prevents two CPUs ++ * clearing the last 2 pending bits, but neither seeing the other's ++ * store when checking if the mask is empty, and missing an empty ++ * mask, which ends with a false positive. ++ */ ++ smp_mb(); + if (cpumask_empty(&wd_smp_cpus_pending)) { + unsigned long flags; + ++none_pending: ++ /* ++ * Double check under lock because more than one CPU could see ++ * a clear mask with the lockless check after clearing their ++ * pending bits. ++ */ + wd_smp_lock(&flags); + if (cpumask_empty(&wd_smp_cpus_pending)) { + wd_smp_last_reset_tb = tb; +@@ -318,8 +353,12 @@ void arch_touch_nmi_watchdog(void) + { + unsigned long ticks = tb_ticks_per_usec * wd_timer_period_ms * 1000; + int cpu = smp_processor_id(); +- u64 tb = get_tb(); ++ u64 tb; + ++ if (!cpumask_test_cpu(cpu, &watchdog_cpumask)) ++ return; ++ ++ tb = get_tb(); + if (tb - per_cpu(wd_timer_tb, cpu) >= ticks) { + per_cpu(wd_timer_tb, cpu) = tb; + wd_smp_clear_cpu_pending(cpu, tb); +diff --git a/arch/powerpc/kvm/book3s_hv.c b/arch/powerpc/kvm/book3s_hv.c +index 7b74fc0a986b8..94da0d25eb125 100644 +--- a/arch/powerpc/kvm/book3s_hv.c ++++ b/arch/powerpc/kvm/book3s_hv.c +@@ -4861,8 +4861,12 @@ static int kvmppc_core_prepare_memory_region_hv(struct kvm *kvm, + unsigned long npages = mem->memory_size >> PAGE_SHIFT; + + if (change == KVM_MR_CREATE) { +- slot->arch.rmap = vzalloc(array_size(npages, +- sizeof(*slot->arch.rmap))); ++ unsigned long size = array_size(npages, sizeof(*slot->arch.rmap)); ++ ++ if ((size >> PAGE_SHIFT) > totalram_pages()) ++ return -ENOMEM; ++ ++ slot->arch.rmap = vzalloc(size); + if (!slot->arch.rmap) + return -ENOMEM; + } +diff --git a/arch/powerpc/kvm/book3s_hv_nested.c b/arch/powerpc/kvm/book3s_hv_nested.c +index ed8a2c9f56299..89295b52a97c3 100644 +--- a/arch/powerpc/kvm/book3s_hv_nested.c ++++ b/arch/powerpc/kvm/book3s_hv_nested.c +@@ -582,7 +582,7 @@ long kvmhv_copy_tofrom_guest_nested(struct kvm_vcpu *vcpu) + if (eaddr & (0xFFFUL << 52)) + return H_PARAMETER; + +- buf = kzalloc(n, GFP_KERNEL); ++ buf = kzalloc(n, GFP_KERNEL | __GFP_NOWARN); + if (!buf) + return H_NO_MEM; + +diff --git a/arch/powerpc/mm/book3s64/radix_pgtable.c b/arch/powerpc/mm/book3s64/radix_pgtable.c +index ae20add7954a0..795d18a84f556 100644 +--- a/arch/powerpc/mm/book3s64/radix_pgtable.c ++++ b/arch/powerpc/mm/book3s64/radix_pgtable.c +@@ -1093,7 +1093,7 @@ int pud_set_huge(pud_t *pud, phys_addr_t addr, pgprot_t prot) + + int pud_clear_huge(pud_t *pud) + { +- if (pud_huge(*pud)) { ++ if (pud_is_leaf(*pud)) { + pud_clear(pud); + return 1; + } +@@ -1140,7 +1140,7 @@ int pmd_set_huge(pmd_t *pmd, phys_addr_t addr, pgprot_t prot) + + int pmd_clear_huge(pmd_t *pmd) + { +- if (pmd_huge(*pmd)) { ++ if (pmd_is_leaf(*pmd)) { + pmd_clear(pmd); + return 1; + } +diff --git a/arch/powerpc/mm/kasan/book3s_32.c b/arch/powerpc/mm/kasan/book3s_32.c +index 202bd260a0095..35b287b0a8da4 100644 +--- a/arch/powerpc/mm/kasan/book3s_32.c ++++ b/arch/powerpc/mm/kasan/book3s_32.c +@@ -19,7 +19,8 @@ int __init kasan_init_region(void *start, size_t size) + block = memblock_alloc(k_size, k_size_base); + + if (block && k_size_base >= SZ_128K && k_start == ALIGN(k_start, k_size_base)) { +- int k_size_more = 1 << (ffs(k_size - k_size_base) - 1); ++ int shift = ffs(k_size - k_size_base); ++ int k_size_more = shift ? 1 << (shift - 1) : 0; + + setbat(-1, k_start, __pa(block), k_size_base, PAGE_KERNEL); + if (k_size_more >= SZ_128K) +diff --git a/arch/powerpc/mm/pgtable_64.c b/arch/powerpc/mm/pgtable_64.c +index 78c8cf01db5f9..175aabf101e87 100644 +--- a/arch/powerpc/mm/pgtable_64.c ++++ b/arch/powerpc/mm/pgtable_64.c +@@ -102,7 +102,8 @@ EXPORT_SYMBOL(__pte_frag_size_shift); + struct page *p4d_page(p4d_t p4d) + { + if (p4d_is_leaf(p4d)) { +- VM_WARN_ON(!p4d_huge(p4d)); ++ if (!IS_ENABLED(CONFIG_HAVE_ARCH_HUGE_VMAP)) ++ VM_WARN_ON(!p4d_huge(p4d)); + return pte_page(p4d_pte(p4d)); + } + return virt_to_page(p4d_pgtable(p4d)); +@@ -112,7 +113,8 @@ struct page *p4d_page(p4d_t p4d) + struct page *pud_page(pud_t pud) + { + if (pud_is_leaf(pud)) { +- VM_WARN_ON(!pud_huge(pud)); ++ if (!IS_ENABLED(CONFIG_HAVE_ARCH_HUGE_VMAP)) ++ VM_WARN_ON(!pud_huge(pud)); + return pte_page(pud_pte(pud)); + } + return virt_to_page(pud_pgtable(pud)); +@@ -125,7 +127,13 @@ struct page *pud_page(pud_t pud) + struct page *pmd_page(pmd_t pmd) + { + if (pmd_is_leaf(pmd)) { +- VM_WARN_ON(!(pmd_large(pmd) || pmd_huge(pmd))); ++ /* ++ * vmalloc_to_page may be called on any vmap address (not only ++ * vmalloc), and it uses pmd_page() etc., when huge vmap is ++ * enabled so these checks can't be used. ++ */ ++ if (!IS_ENABLED(CONFIG_HAVE_ARCH_HUGE_VMAP)) ++ VM_WARN_ON(!(pmd_large(pmd) || pmd_huge(pmd))); + return pte_page(pmd_pte(pmd)); + } + return virt_to_page(pmd_page_vaddr(pmd)); +diff --git a/arch/powerpc/perf/core-book3s.c b/arch/powerpc/perf/core-book3s.c +index 73e62e9b179bc..bef6b1abce702 100644 +--- a/arch/powerpc/perf/core-book3s.c ++++ b/arch/powerpc/perf/core-book3s.c +@@ -857,6 +857,19 @@ static void write_pmc(int idx, unsigned long val) + } + } + ++static int any_pmc_overflown(struct cpu_hw_events *cpuhw) ++{ ++ int i, idx; ++ ++ for (i = 0; i < cpuhw->n_events; i++) { ++ idx = cpuhw->event[i]->hw.idx; ++ if ((idx) && ((int)read_pmc(idx) < 0)) ++ return idx; ++ } ++ ++ return 0; ++} ++ + /* Called from sysrq_handle_showregs() */ + void perf_event_print_debug(void) + { +@@ -1281,11 +1294,13 @@ static void power_pmu_disable(struct pmu *pmu) + + /* + * Set the 'freeze counters' bit, clear EBE/BHRBA/PMCC/PMAO/FC56 ++ * Also clear PMXE to disable PMI's getting triggered in some ++ * corner cases during PMU disable. + */ + val = mmcr0 = mfspr(SPRN_MMCR0); + val |= MMCR0_FC; + val &= ~(MMCR0_EBE | MMCR0_BHRBA | MMCR0_PMCC | MMCR0_PMAO | +- MMCR0_FC56); ++ MMCR0_PMXE | MMCR0_FC56); + /* Set mmcr0 PMCCEXT for p10 */ + if (ppmu->flags & PPMU_ARCH_31) + val |= MMCR0_PMCCEXT; +@@ -1299,6 +1314,23 @@ static void power_pmu_disable(struct pmu *pmu) + mb(); + isync(); + ++ /* ++ * Some corner cases could clear the PMU counter overflow ++ * while a masked PMI is pending. One such case is when ++ * a PMI happens during interrupt replay and perf counter ++ * values are cleared by PMU callbacks before replay. ++ * ++ * If any PMC corresponding to the active PMU events are ++ * overflown, disable the interrupt by clearing the paca ++ * bit for PMI since we are disabling the PMU now. ++ * Otherwise provide a warning if there is PMI pending, but ++ * no counter is found overflown. ++ */ ++ if (any_pmc_overflown(cpuhw)) ++ clear_pmi_irq_pending(); ++ else ++ WARN_ON(pmi_irq_pending()); ++ + val = mmcra = cpuhw->mmcr.mmcra; + + /* +@@ -1390,6 +1422,15 @@ static void power_pmu_enable(struct pmu *pmu) + * (possibly updated for removal of events). + */ + if (!cpuhw->n_added) { ++ /* ++ * If there is any active event with an overflown PMC ++ * value, set back PACA_IRQ_PMI which would have been ++ * cleared in power_pmu_disable(). ++ */ ++ hard_irq_disable(); ++ if (any_pmc_overflown(cpuhw)) ++ set_pmi_irq_pending(); ++ + mtspr(SPRN_MMCRA, cpuhw->mmcr.mmcra & ~MMCRA_SAMPLE_ENABLE); + mtspr(SPRN_MMCR1, cpuhw->mmcr.mmcr1); + if (ppmu->flags & PPMU_ARCH_31) +@@ -2337,6 +2378,14 @@ static void __perf_event_interrupt(struct pt_regs *regs) + break; + } + } ++ ++ /* ++ * Clear PACA_IRQ_PMI in case it was set by ++ * set_pmi_irq_pending() when PMU was enabled ++ * after accounting for interrupts. ++ */ ++ clear_pmi_irq_pending(); ++ + if (!active) + /* reset non active counters that have overflowed */ + write_pmc(i + 1, 0); +@@ -2356,6 +2405,13 @@ static void __perf_event_interrupt(struct pt_regs *regs) + } + } + } ++ ++ /* ++ * During system wide profling or while specific CPU is monitored for an ++ * event, some corner cases could cause PMC to overflow in idle path. This ++ * will trigger a PMI after waking up from idle. Since counter values are _not_ ++ * saved/restored in idle path, can lead to below "Can't find PMC" message. ++ */ + if (unlikely(!found) && !arch_irq_disabled_regs(regs)) + printk_ratelimited(KERN_WARNING "Can't find PMC that caused IRQ\n"); + +diff --git a/arch/powerpc/platforms/cell/iommu.c b/arch/powerpc/platforms/cell/iommu.c +index fa08699aedeb8..d32f24de84798 100644 +--- a/arch/powerpc/platforms/cell/iommu.c ++++ b/arch/powerpc/platforms/cell/iommu.c +@@ -977,6 +977,7 @@ static int __init cell_iommu_fixed_mapping_init(void) + if (hbase < dbase || (hend > (dbase + dsize))) { + pr_debug("iommu: hash window doesn't fit in" + "real DMA window\n"); ++ of_node_put(np); + return -1; + } + } +diff --git a/arch/powerpc/platforms/cell/pervasive.c b/arch/powerpc/platforms/cell/pervasive.c +index 5b9a7e9f144b3..dff8d5e7ab82b 100644 +--- a/arch/powerpc/platforms/cell/pervasive.c ++++ b/arch/powerpc/platforms/cell/pervasive.c +@@ -78,6 +78,7 @@ static int cbe_system_reset_exception(struct pt_regs *regs) + switch (regs->msr & SRR1_WAKEMASK) { + case SRR1_WAKEDEC: + set_dec(1); ++ break; + case SRR1_WAKEEE: + /* + * Handle these when interrupts get re-enabled and we take +diff --git a/arch/powerpc/platforms/embedded6xx/hlwd-pic.c b/arch/powerpc/platforms/embedded6xx/hlwd-pic.c +index 15396333a90bd..a4b020e4b6af0 100644 +--- a/arch/powerpc/platforms/embedded6xx/hlwd-pic.c ++++ b/arch/powerpc/platforms/embedded6xx/hlwd-pic.c +@@ -214,6 +214,7 @@ void hlwd_pic_probe(void) + irq_set_chained_handler(cascade_virq, + hlwd_pic_irq_cascade); + hlwd_irq_host = host; ++ of_node_put(np); + break; + } + } +diff --git a/arch/powerpc/platforms/powermac/low_i2c.c b/arch/powerpc/platforms/powermac/low_i2c.c +index f77a59b5c2e1a..df89d916236d9 100644 +--- a/arch/powerpc/platforms/powermac/low_i2c.c ++++ b/arch/powerpc/platforms/powermac/low_i2c.c +@@ -582,6 +582,7 @@ static void __init kw_i2c_add(struct pmac_i2c_host_kw *host, + bus->close = kw_i2c_close; + bus->xfer = kw_i2c_xfer; + mutex_init(&bus->mutex); ++ lockdep_register_key(&bus->lock_key); + lockdep_set_class(&bus->mutex, &bus->lock_key); + if (controller == busnode) + bus->flags = pmac_i2c_multibus; +@@ -810,6 +811,7 @@ static void __init pmu_i2c_probe(void) + bus->hostdata = bus + 1; + bus->xfer = pmu_i2c_xfer; + mutex_init(&bus->mutex); ++ lockdep_register_key(&bus->lock_key); + lockdep_set_class(&bus->mutex, &bus->lock_key); + bus->flags = pmac_i2c_multibus; + list_add(&bus->link, &pmac_i2c_busses); +@@ -933,6 +935,7 @@ static void __init smu_i2c_probe(void) + bus->hostdata = bus + 1; + bus->xfer = smu_i2c_xfer; + mutex_init(&bus->mutex); ++ lockdep_register_key(&bus->lock_key); + lockdep_set_class(&bus->mutex, &bus->lock_key); + bus->flags = 0; + list_add(&bus->link, &pmac_i2c_busses); +diff --git a/arch/powerpc/platforms/powernv/opal-lpc.c b/arch/powerpc/platforms/powernv/opal-lpc.c +index 1e5d51db40f84..5390c888db162 100644 +--- a/arch/powerpc/platforms/powernv/opal-lpc.c ++++ b/arch/powerpc/platforms/powernv/opal-lpc.c +@@ -396,6 +396,7 @@ void __init opal_lpc_init(void) + if (!of_get_property(np, "primary", NULL)) + continue; + opal_lpc_chip_id = of_get_ibm_chip_id(np); ++ of_node_put(np); + break; + } + if (opal_lpc_chip_id < 0) +diff --git a/arch/powerpc/sysdev/xive/spapr.c b/arch/powerpc/sysdev/xive/spapr.c +index f143b6f111ac0..1179632560b8d 100644 +--- a/arch/powerpc/sysdev/xive/spapr.c ++++ b/arch/powerpc/sysdev/xive/spapr.c +@@ -653,6 +653,9 @@ static int xive_spapr_debug_show(struct seq_file *m, void *private) + struct xive_irq_bitmap *xibm; + char *buf = kmalloc(PAGE_SIZE, GFP_KERNEL); + ++ if (!buf) ++ return -ENOMEM; ++ + list_for_each_entry(xibm, &xive_irq_bitmaps, list) { + memset(buf, 0, PAGE_SIZE); + bitmap_print_to_pagebuf(true, buf, xibm->bitmap, xibm->count); +diff --git a/arch/riscv/Kconfig b/arch/riscv/Kconfig +index f076cee11af69..20bc4b8c13b92 100644 +--- a/arch/riscv/Kconfig ++++ b/arch/riscv/Kconfig +@@ -158,10 +158,9 @@ config PA_BITS + + config PAGE_OFFSET + hex +- default 0xC0000000 if 32BIT && MAXPHYSMEM_1GB ++ default 0xC0000000 if 32BIT + default 0x80000000 if 64BIT && !MMU +- default 0xffffffff80000000 if 64BIT && MAXPHYSMEM_2GB +- default 0xffffffe000000000 if 64BIT && MAXPHYSMEM_128GB ++ default 0xffffffe000000000 if 64BIT + + config KASAN_SHADOW_OFFSET + hex +@@ -270,24 +269,6 @@ config MODULE_SECTIONS + bool + select HAVE_MOD_ARCH_SPECIFIC + +-choice +- prompt "Maximum Physical Memory" +- default MAXPHYSMEM_1GB if 32BIT +- default MAXPHYSMEM_2GB if 64BIT && CMODEL_MEDLOW +- default MAXPHYSMEM_128GB if 64BIT && CMODEL_MEDANY +- +- config MAXPHYSMEM_1GB +- depends on 32BIT +- bool "1GiB" +- config MAXPHYSMEM_2GB +- depends on 64BIT && CMODEL_MEDLOW +- bool "2GiB" +- config MAXPHYSMEM_128GB +- depends on 64BIT && CMODEL_MEDANY +- bool "128GiB" +-endchoice +- +- + config SMP + bool "Symmetric Multi-Processing" + help +diff --git a/arch/riscv/boot/dts/microchip/microchip-mpfs.dtsi b/arch/riscv/boot/dts/microchip/microchip-mpfs.dtsi +index b12fd594e7172..4ef4bcb748729 100644 +--- a/arch/riscv/boot/dts/microchip/microchip-mpfs.dtsi ++++ b/arch/riscv/boot/dts/microchip/microchip-mpfs.dtsi +@@ -9,9 +9,6 @@ + model = "Microchip PolarFire SoC"; + compatible = "microchip,mpfs"; + +- chosen { +- }; +- + cpus { + #address-cells = <1>; + #size-cells = <0>; +diff --git a/arch/riscv/configs/nommu_k210_defconfig b/arch/riscv/configs/nommu_k210_defconfig +index b16a2a12c82a8..3b9f83221f9c2 100644 +--- a/arch/riscv/configs/nommu_k210_defconfig ++++ b/arch/riscv/configs/nommu_k210_defconfig +@@ -29,8 +29,6 @@ CONFIG_EMBEDDED=y + CONFIG_SLOB=y + # CONFIG_MMU is not set + CONFIG_SOC_CANAAN=y +-CONFIG_SOC_CANAAN_K210_DTB_SOURCE="k210_generic" +-CONFIG_MAXPHYSMEM_2GB=y + CONFIG_SMP=y + CONFIG_NR_CPUS=2 + CONFIG_CMDLINE="earlycon console=ttySIF0" +diff --git a/arch/riscv/configs/nommu_k210_sdcard_defconfig b/arch/riscv/configs/nommu_k210_sdcard_defconfig +index 61f887f654199..d68b743d580f8 100644 +--- a/arch/riscv/configs/nommu_k210_sdcard_defconfig ++++ b/arch/riscv/configs/nommu_k210_sdcard_defconfig +@@ -21,8 +21,6 @@ CONFIG_EMBEDDED=y + CONFIG_SLOB=y + # CONFIG_MMU is not set + CONFIG_SOC_CANAAN=y +-CONFIG_SOC_CANAAN_K210_DTB_SOURCE="k210_generic" +-CONFIG_MAXPHYSMEM_2GB=y + CONFIG_SMP=y + CONFIG_NR_CPUS=2 + CONFIG_CMDLINE="earlycon console=ttySIF0 rootdelay=2 root=/dev/mmcblk0p1 ro" +diff --git a/arch/riscv/configs/nommu_virt_defconfig b/arch/riscv/configs/nommu_virt_defconfig +index e046a0babde43..f224be697785f 100644 +--- a/arch/riscv/configs/nommu_virt_defconfig ++++ b/arch/riscv/configs/nommu_virt_defconfig +@@ -27,7 +27,6 @@ CONFIG_SLOB=y + # CONFIG_SLAB_MERGE_DEFAULT is not set + # CONFIG_MMU is not set + CONFIG_SOC_VIRT=y +-CONFIG_MAXPHYSMEM_2GB=y + CONFIG_SMP=y + CONFIG_CMDLINE="root=/dev/vda rw earlycon=uart8250,mmio,0x10000000,115200n8 console=ttyS0" + CONFIG_CMDLINE_FORCE=y +diff --git a/arch/riscv/include/asm/smp.h b/arch/riscv/include/asm/smp.h +index a7d2811f35365..62d0e6e61da83 100644 +--- a/arch/riscv/include/asm/smp.h ++++ b/arch/riscv/include/asm/smp.h +@@ -43,7 +43,6 @@ void arch_send_call_function_ipi_mask(struct cpumask *mask); + void arch_send_call_function_single_ipi(int cpu); + + int riscv_hartid_to_cpuid(int hartid); +-void riscv_cpuid_to_hartid_mask(const struct cpumask *in, struct cpumask *out); + + /* Set custom IPI operations */ + void riscv_set_ipi_ops(const struct riscv_ipi_ops *ops); +@@ -85,13 +84,6 @@ static inline unsigned long cpuid_to_hartid_map(int cpu) + return boot_cpu_hartid; + } + +-static inline void riscv_cpuid_to_hartid_mask(const struct cpumask *in, +- struct cpumask *out) +-{ +- cpumask_clear(out); +- cpumask_set_cpu(boot_cpu_hartid, out); +-} +- + static inline void riscv_set_ipi_ops(const struct riscv_ipi_ops *ops) + { + } +@@ -102,6 +94,8 @@ static inline void riscv_clear_ipi(void) + + #endif /* CONFIG_SMP */ + ++void riscv_cpuid_to_hartid_mask(const struct cpumask *in, struct cpumask *out); ++ + #if defined(CONFIG_HOTPLUG_CPU) && (CONFIG_SMP) + bool cpu_has_hotplug(unsigned int cpu); + #else +diff --git a/arch/riscv/kernel/kexec_relocate.S b/arch/riscv/kernel/kexec_relocate.S +index a80b52a74f58c..059c5e216ae75 100644 +--- a/arch/riscv/kernel/kexec_relocate.S ++++ b/arch/riscv/kernel/kexec_relocate.S +@@ -159,25 +159,15 @@ SYM_CODE_START(riscv_kexec_norelocate) + * s0: (const) Phys address to jump to + * s1: (const) Phys address of the FDT image + * s2: (const) The hartid of the current hart +- * s3: (const) kernel_map.va_pa_offset, used when switching MMU off + */ + mv s0, a1 + mv s1, a2 + mv s2, a3 +- mv s3, a4 + + /* Disable / cleanup interrupts */ + csrw CSR_SIE, zero + csrw CSR_SIP, zero + +- /* Switch to physical addressing */ +- la s4, 1f +- sub s4, s4, s3 +- csrw CSR_STVEC, s4 +- csrw CSR_SATP, zero +- +-.align 2 +-1: + /* Pass the arguments to the next kernel / Cleanup*/ + mv a0, s2 + mv a1, s1 +@@ -214,7 +204,15 @@ SYM_CODE_START(riscv_kexec_norelocate) + csrw CSR_SCAUSE, zero + csrw CSR_SSCRATCH, zero + +- jalr zero, a2, 0 ++ /* ++ * Switch to physical addressing ++ * This will also trigger a jump to CSR_STVEC ++ * which in this case is the address of the new ++ * kernel. ++ */ ++ csrw CSR_STVEC, a2 ++ csrw CSR_SATP, zero ++ + SYM_CODE_END(riscv_kexec_norelocate) + + .section ".rodata" +diff --git a/arch/riscv/kernel/machine_kexec.c b/arch/riscv/kernel/machine_kexec.c +index e6eca271a4d60..cbef0fc73afa8 100644 +--- a/arch/riscv/kernel/machine_kexec.c ++++ b/arch/riscv/kernel/machine_kexec.c +@@ -169,7 +169,8 @@ machine_kexec(struct kimage *image) + struct kimage_arch *internal = &image->arch; + unsigned long jump_addr = (unsigned long) image->start; + unsigned long first_ind_entry = (unsigned long) &image->head; +- unsigned long this_hart_id = raw_smp_processor_id(); ++ unsigned long this_cpu_id = smp_processor_id(); ++ unsigned long this_hart_id = cpuid_to_hartid_map(this_cpu_id); + unsigned long fdt_addr = internal->fdt_addr; + void *control_code_buffer = page_address(image->control_code_page); + riscv_kexec_method kexec_method = NULL; +diff --git a/arch/riscv/kernel/setup.c b/arch/riscv/kernel/setup.c +index b9620e5f00baf..6c5caf5eb9061 100644 +--- a/arch/riscv/kernel/setup.c ++++ b/arch/riscv/kernel/setup.c +@@ -59,6 +59,16 @@ atomic_t hart_lottery __section(".sdata") + unsigned long boot_cpu_hartid; + static DEFINE_PER_CPU(struct cpu, cpu_devices); + ++void riscv_cpuid_to_hartid_mask(const struct cpumask *in, struct cpumask *out) ++{ ++ int cpu; ++ ++ cpumask_clear(out); ++ for_each_cpu(cpu, in) ++ cpumask_set_cpu(cpuid_to_hartid_map(cpu), out); ++} ++EXPORT_SYMBOL_GPL(riscv_cpuid_to_hartid_mask); ++ + /* + * Place kernel memory regions on the resource tree so that + * kexec-tools can retrieve them from /proc/iomem. While there +diff --git a/arch/riscv/kernel/smp.c b/arch/riscv/kernel/smp.c +index 921d9d7df4001..d0147294691d9 100644 +--- a/arch/riscv/kernel/smp.c ++++ b/arch/riscv/kernel/smp.c +@@ -59,16 +59,6 @@ int riscv_hartid_to_cpuid(int hartid) + return -ENOENT; + } + +-void riscv_cpuid_to_hartid_mask(const struct cpumask *in, struct cpumask *out) +-{ +- int cpu; +- +- cpumask_clear(out); +- for_each_cpu(cpu, in) +- cpumask_set_cpu(cpuid_to_hartid_map(cpu), out); +-} +-EXPORT_SYMBOL_GPL(riscv_cpuid_to_hartid_mask); +- + bool arch_match_cpu_phys_id(int cpu, u64 phys_id) + { + return phys_id == cpuid_to_hartid_map(cpu); +diff --git a/arch/riscv/mm/init.c b/arch/riscv/mm/init.c +index c0cddf0fc22db..5e7decd875258 100644 +--- a/arch/riscv/mm/init.c ++++ b/arch/riscv/mm/init.c +@@ -187,10 +187,10 @@ static void __init setup_bootmem(void) + + + phys_ram_end = memblock_end_of_DRAM(); +-#ifndef CONFIG_64BIT + #ifndef CONFIG_XIP_KERNEL + phys_ram_base = memblock_start_of_DRAM(); + #endif ++#ifndef CONFIG_64BIT + /* + * memblock allocator is not aware of the fact that last 4K bytes of + * the addressable memory can not be mapped because of IS_ERR_VALUE +@@ -813,13 +813,22 @@ static void __init reserve_crashkernel(void) + /* + * Current riscv boot protocol requires 2MB alignment for + * RV64 and 4MB alignment for RV32 (hugepage size) ++ * ++ * Try to alloc from 32bit addressible physical memory so that ++ * swiotlb can work on the crash kernel. + */ + crash_base = memblock_phys_alloc_range(crash_size, PMD_SIZE, +- search_start, search_end); ++ search_start, ++ min(search_end, (unsigned long) SZ_4G)); + if (crash_base == 0) { +- pr_warn("crashkernel: couldn't allocate %lldKB\n", +- crash_size >> 10); +- return; ++ /* Try again without restricting region to 32bit addressible memory */ ++ crash_base = memblock_phys_alloc_range(crash_size, PMD_SIZE, ++ search_start, search_end); ++ if (crash_base == 0) { ++ pr_warn("crashkernel: couldn't allocate %lldKB\n", ++ crash_size >> 10); ++ return; ++ } + } + + pr_info("crashkernel: reserved 0x%016llx - 0x%016llx (%lld MB)\n", +diff --git a/arch/s390/mm/pgalloc.c b/arch/s390/mm/pgalloc.c +index 781965f7210eb..91e478e09b54b 100644 +--- a/arch/s390/mm/pgalloc.c ++++ b/arch/s390/mm/pgalloc.c +@@ -244,13 +244,15 @@ void page_table_free(struct mm_struct *mm, unsigned long *table) + /* Free 2K page table fragment of a 4K page */ + bit = ((unsigned long) table & ~PAGE_MASK)/(PTRS_PER_PTE*sizeof(pte_t)); + spin_lock_bh(&mm->context.lock); +- mask = atomic_xor_bits(&page->_refcount, 1U << (bit + 24)); ++ mask = atomic_xor_bits(&page->_refcount, 0x11U << (bit + 24)); + mask >>= 24; + if (mask & 3) + list_add(&page->lru, &mm->context.pgtable_list); + else + list_del(&page->lru); + spin_unlock_bh(&mm->context.lock); ++ mask = atomic_xor_bits(&page->_refcount, 0x10U << (bit + 24)); ++ mask >>= 24; + if (mask != 0) + return; + } else { +diff --git a/arch/sh/configs/titan_defconfig b/arch/sh/configs/titan_defconfig +index ba887f1351be6..cd5c58916c65a 100644 +--- a/arch/sh/configs/titan_defconfig ++++ b/arch/sh/configs/titan_defconfig +@@ -242,7 +242,6 @@ CONFIG_NFSD=y + CONFIG_NFSD_V3=y + CONFIG_SMB_FS=m + CONFIG_CIFS=m +-CONFIG_CIFS_WEAK_PW_HASH=y + CONFIG_PARTITION_ADVANCED=y + CONFIG_NLS_CODEPAGE_437=m + CONFIG_NLS_ASCII=m +diff --git a/arch/um/.gitignore b/arch/um/.gitignore +index 6323e5571887e..d69ea5b562cee 100644 +--- a/arch/um/.gitignore ++++ b/arch/um/.gitignore +@@ -2,3 +2,4 @@ + kernel/config.c + kernel/config.tmp + kernel/vmlinux.lds ++kernel/capflags.c +diff --git a/arch/um/drivers/virt-pci.c b/arch/um/drivers/virt-pci.c +index c080666330234..0ab58016db22f 100644 +--- a/arch/um/drivers/virt-pci.c ++++ b/arch/um/drivers/virt-pci.c +@@ -181,15 +181,15 @@ static unsigned long um_pci_cfgspace_read(void *priv, unsigned int offset, + /* buf->data is maximum size - we may only use parts of it */ + struct um_pci_message_buffer *buf; + u8 *data; +- unsigned long ret = ~0ULL; ++ unsigned long ret = ULONG_MAX; + + if (!dev) +- return ~0ULL; ++ return ULONG_MAX; + + buf = get_cpu_var(um_pci_msg_bufs); + data = buf->data; + +- memset(data, 0xff, sizeof(data)); ++ memset(buf->data, 0xff, sizeof(buf->data)); + + switch (size) { + case 1: +@@ -304,7 +304,7 @@ static unsigned long um_pci_bar_read(void *priv, unsigned int offset, + /* buf->data is maximum size - we may only use parts of it */ + struct um_pci_message_buffer *buf; + u8 *data; +- unsigned long ret = ~0ULL; ++ unsigned long ret = ULONG_MAX; + + buf = get_cpu_var(um_pci_msg_bufs); + data = buf->data; +diff --git a/arch/um/drivers/virtio_uml.c b/arch/um/drivers/virtio_uml.c +index d51e445df7976..7755cb4ff9fc6 100644 +--- a/arch/um/drivers/virtio_uml.c ++++ b/arch/um/drivers/virtio_uml.c +@@ -1090,6 +1090,8 @@ static void virtio_uml_release_dev(struct device *d) + container_of(d, struct virtio_device, dev); + struct virtio_uml_device *vu_dev = to_virtio_uml_device(vdev); + ++ time_travel_propagate_time(); ++ + /* might not have been opened due to not negotiating the feature */ + if (vu_dev->req_fd >= 0) { + um_free_irq(vu_dev->irq, vu_dev); +@@ -1136,6 +1138,8 @@ static int virtio_uml_probe(struct platform_device *pdev) + vu_dev->pdev = pdev; + vu_dev->req_fd = -1; + ++ time_travel_propagate_time(); ++ + do { + rc = os_connect_socket(pdata->socket_path); + } while (rc == -EINTR); +diff --git a/arch/um/include/asm/delay.h b/arch/um/include/asm/delay.h +index 56fc2b8f2dd01..e79b2ab6f40c8 100644 +--- a/arch/um/include/asm/delay.h ++++ b/arch/um/include/asm/delay.h +@@ -14,7 +14,7 @@ static inline void um_ndelay(unsigned long nsecs) + ndelay(nsecs); + } + #undef ndelay +-#define ndelay um_ndelay ++#define ndelay(n) um_ndelay(n) + + static inline void um_udelay(unsigned long usecs) + { +@@ -26,5 +26,5 @@ static inline void um_udelay(unsigned long usecs) + udelay(usecs); + } + #undef udelay +-#define udelay um_udelay ++#define udelay(n) um_udelay(n) + #endif /* __UM_DELAY_H */ +diff --git a/arch/um/include/asm/irqflags.h b/arch/um/include/asm/irqflags.h +index dab5744e9253d..1e69ef5bc35e0 100644 +--- a/arch/um/include/asm/irqflags.h ++++ b/arch/um/include/asm/irqflags.h +@@ -3,7 +3,7 @@ + #define __UM_IRQFLAGS_H + + extern int signals_enabled; +-int set_signals(int enable); ++int um_set_signals(int enable); + void block_signals(void); + void unblock_signals(void); + +@@ -16,7 +16,7 @@ static inline unsigned long arch_local_save_flags(void) + #define arch_local_irq_restore arch_local_irq_restore + static inline void arch_local_irq_restore(unsigned long flags) + { +- set_signals(flags); ++ um_set_signals(flags); + } + + #define arch_local_irq_enable arch_local_irq_enable +diff --git a/arch/um/include/shared/longjmp.h b/arch/um/include/shared/longjmp.h +index bdb2869b72b31..8863319039f3d 100644 +--- a/arch/um/include/shared/longjmp.h ++++ b/arch/um/include/shared/longjmp.h +@@ -18,7 +18,7 @@ extern void longjmp(jmp_buf, int); + enable = *(volatile int *)&signals_enabled; \ + n = setjmp(*buf); \ + if(n != 0) \ +- set_signals_trace(enable); \ ++ um_set_signals_trace(enable); \ + n; }) + + #endif +diff --git a/arch/um/include/shared/os.h b/arch/um/include/shared/os.h +index 96d400387c93e..03ffbdddcc480 100644 +--- a/arch/um/include/shared/os.h ++++ b/arch/um/include/shared/os.h +@@ -238,8 +238,8 @@ extern void send_sigio_to_self(void); + extern int change_sig(int signal, int on); + extern void block_signals(void); + extern void unblock_signals(void); +-extern int set_signals(int enable); +-extern int set_signals_trace(int enable); ++extern int um_set_signals(int enable); ++extern int um_set_signals_trace(int enable); + extern int os_is_signal_stack(void); + extern void deliver_alarm(void); + extern void register_pm_wake_signal(void); +diff --git a/arch/um/include/shared/registers.h b/arch/um/include/shared/registers.h +index 0c50fa6e8a55b..fbb709a222839 100644 +--- a/arch/um/include/shared/registers.h ++++ b/arch/um/include/shared/registers.h +@@ -16,8 +16,8 @@ extern int restore_fp_registers(int pid, unsigned long *fp_regs); + extern int save_fpx_registers(int pid, unsigned long *fp_regs); + extern int restore_fpx_registers(int pid, unsigned long *fp_regs); + extern int save_registers(int pid, struct uml_pt_regs *regs); +-extern int restore_registers(int pid, struct uml_pt_regs *regs); +-extern int init_registers(int pid); ++extern int restore_pid_registers(int pid, struct uml_pt_regs *regs); ++extern int init_pid_registers(int pid); + extern void get_safe_registers(unsigned long *regs, unsigned long *fp_regs); + extern unsigned long get_thread_reg(int reg, jmp_buf *buf); + extern int get_fp_registers(int pid, unsigned long *regs); +diff --git a/arch/um/kernel/ksyms.c b/arch/um/kernel/ksyms.c +index b1e5634398d09..3a85bde3e1734 100644 +--- a/arch/um/kernel/ksyms.c ++++ b/arch/um/kernel/ksyms.c +@@ -6,7 +6,7 @@ + #include <linux/module.h> + #include <os.h> + +-EXPORT_SYMBOL(set_signals); ++EXPORT_SYMBOL(um_set_signals); + EXPORT_SYMBOL(signals_enabled); + + EXPORT_SYMBOL(os_stat_fd); +diff --git a/arch/um/os-Linux/registers.c b/arch/um/os-Linux/registers.c +index 2d9270508e156..b123955be7acc 100644 +--- a/arch/um/os-Linux/registers.c ++++ b/arch/um/os-Linux/registers.c +@@ -21,7 +21,7 @@ int save_registers(int pid, struct uml_pt_regs *regs) + return 0; + } + +-int restore_registers(int pid, struct uml_pt_regs *regs) ++int restore_pid_registers(int pid, struct uml_pt_regs *regs) + { + int err; + +@@ -36,7 +36,7 @@ int restore_registers(int pid, struct uml_pt_regs *regs) + static unsigned long exec_regs[MAX_REG_NR]; + static unsigned long exec_fp_regs[FP_SIZE]; + +-int init_registers(int pid) ++int init_pid_registers(int pid) + { + int err; + +diff --git a/arch/um/os-Linux/sigio.c b/arch/um/os-Linux/sigio.c +index 6597ea1986ffa..9e71794839e87 100644 +--- a/arch/um/os-Linux/sigio.c ++++ b/arch/um/os-Linux/sigio.c +@@ -132,7 +132,7 @@ static void update_thread(void) + int n; + char c; + +- flags = set_signals_trace(0); ++ flags = um_set_signals_trace(0); + CATCH_EINTR(n = write(sigio_private[0], &c, sizeof(c))); + if (n != sizeof(c)) { + printk(UM_KERN_ERR "update_thread : write failed, err = %d\n", +@@ -147,7 +147,7 @@ static void update_thread(void) + goto fail; + } + +- set_signals_trace(flags); ++ um_set_signals_trace(flags); + return; + fail: + /* Critical section start */ +@@ -161,7 +161,7 @@ static void update_thread(void) + close(write_sigio_fds[0]); + close(write_sigio_fds[1]); + /* Critical section end */ +- set_signals_trace(flags); ++ um_set_signals_trace(flags); + } + + int __add_sigio_fd(int fd) +diff --git a/arch/um/os-Linux/signal.c b/arch/um/os-Linux/signal.c +index 6cf098c23a394..24a403a70a020 100644 +--- a/arch/um/os-Linux/signal.c ++++ b/arch/um/os-Linux/signal.c +@@ -94,7 +94,7 @@ void sig_handler(int sig, struct siginfo *si, mcontext_t *mc) + + sig_handler_common(sig, si, mc); + +- set_signals_trace(enabled); ++ um_set_signals_trace(enabled); + } + + static void timer_real_alarm_handler(mcontext_t *mc) +@@ -126,7 +126,7 @@ void timer_alarm_handler(int sig, struct siginfo *unused_si, mcontext_t *mc) + + signals_active &= ~SIGALRM_MASK; + +- set_signals_trace(enabled); ++ um_set_signals_trace(enabled); + } + + void deliver_alarm(void) { +@@ -348,7 +348,7 @@ void unblock_signals(void) + } + } + +-int set_signals(int enable) ++int um_set_signals(int enable) + { + int ret; + if (signals_enabled == enable) +@@ -362,7 +362,7 @@ int set_signals(int enable) + return ret; + } + +-int set_signals_trace(int enable) ++int um_set_signals_trace(int enable) + { + int ret; + if (signals_enabled == enable) +diff --git a/arch/um/os-Linux/start_up.c b/arch/um/os-Linux/start_up.c +index 8a72c99994eb1..e3ee4db58b40d 100644 +--- a/arch/um/os-Linux/start_up.c ++++ b/arch/um/os-Linux/start_up.c +@@ -368,7 +368,7 @@ void __init os_early_checks(void) + check_tmpexec(); + + pid = start_ptraced_child(); +- if (init_registers(pid)) ++ if (init_pid_registers(pid)) + fatal("Failed to initialize default registers"); + stop_ptraced_child(pid, 1, 1); + } +diff --git a/arch/x86/boot/compressed/Makefile b/arch/x86/boot/compressed/Makefile +index 431bf7f846c3c..e118136460518 100644 +--- a/arch/x86/boot/compressed/Makefile ++++ b/arch/x86/boot/compressed/Makefile +@@ -28,7 +28,11 @@ KCOV_INSTRUMENT := n + targets := vmlinux vmlinux.bin vmlinux.bin.gz vmlinux.bin.bz2 vmlinux.bin.lzma \ + vmlinux.bin.xz vmlinux.bin.lzo vmlinux.bin.lz4 vmlinux.bin.zst + +-KBUILD_CFLAGS := -m$(BITS) -O2 ++# CLANG_FLAGS must come before any cc-disable-warning or cc-option calls in ++# case of cross compiling, as it has the '--target=' flag, which is needed to ++# avoid errors with '-march=i386', and future flags may depend on the target to ++# be valid. ++KBUILD_CFLAGS := -m$(BITS) -O2 $(CLANG_FLAGS) + KBUILD_CFLAGS += -fno-strict-aliasing -fPIE + KBUILD_CFLAGS += -Wundef + KBUILD_CFLAGS += -DDISABLE_BRANCH_PROFILING +@@ -47,7 +51,6 @@ KBUILD_CFLAGS += -D__DISABLE_EXPORTS + # Disable relocation relaxation in case the link is not PIE. + KBUILD_CFLAGS += $(call as-option,-Wa$(comma)-mrelax-relocations=no) + KBUILD_CFLAGS += -include $(srctree)/include/linux/hidden.h +-KBUILD_CFLAGS += $(CLANG_FLAGS) + + # sev.c indirectly inludes inat-table.h which is generated during + # compilation and stored in $(objtree). Add the directory to the includes so +diff --git a/arch/x86/configs/i386_defconfig b/arch/x86/configs/i386_defconfig +index e81885384f604..99398cbdae434 100644 +--- a/arch/x86/configs/i386_defconfig ++++ b/arch/x86/configs/i386_defconfig +@@ -262,3 +262,4 @@ CONFIG_BLK_DEV_IO_TRACE=y + CONFIG_PROVIDE_OHCI1394_DMA_INIT=y + CONFIG_EARLY_PRINTK_DBGP=y + CONFIG_DEBUG_BOOT_PARAMS=y ++CONFIG_KALLSYMS_ALL=y +diff --git a/arch/x86/configs/x86_64_defconfig b/arch/x86/configs/x86_64_defconfig +index e8a7a0af2bdaa..d7298b104a456 100644 +--- a/arch/x86/configs/x86_64_defconfig ++++ b/arch/x86/configs/x86_64_defconfig +@@ -258,3 +258,4 @@ CONFIG_BLK_DEV_IO_TRACE=y + CONFIG_PROVIDE_OHCI1394_DMA_INIT=y + CONFIG_EARLY_PRINTK_DBGP=y + CONFIG_DEBUG_BOOT_PARAMS=y ++CONFIG_KALLSYMS_ALL=y +diff --git a/arch/x86/crypto/aesni-intel_glue.c b/arch/x86/crypto/aesni-intel_glue.c +index e09f4672dd382..41901ba9d3a2c 100644 +--- a/arch/x86/crypto/aesni-intel_glue.c ++++ b/arch/x86/crypto/aesni-intel_glue.c +@@ -1107,7 +1107,7 @@ static struct aead_alg aesni_aeads[] = { { + .cra_flags = CRYPTO_ALG_INTERNAL, + .cra_blocksize = 1, + .cra_ctxsize = sizeof(struct aesni_rfc4106_gcm_ctx), +- .cra_alignmask = AESNI_ALIGN - 1, ++ .cra_alignmask = 0, + .cra_module = THIS_MODULE, + }, + }, { +@@ -1124,7 +1124,7 @@ static struct aead_alg aesni_aeads[] = { { + .cra_flags = CRYPTO_ALG_INTERNAL, + .cra_blocksize = 1, + .cra_ctxsize = sizeof(struct generic_gcmaes_ctx), +- .cra_alignmask = AESNI_ALIGN - 1, ++ .cra_alignmask = 0, + .cra_module = THIS_MODULE, + }, + } }; +diff --git a/arch/x86/include/asm/realmode.h b/arch/x86/include/asm/realmode.h +index 5db5d083c8732..331474b150f16 100644 +--- a/arch/x86/include/asm/realmode.h ++++ b/arch/x86/include/asm/realmode.h +@@ -89,6 +89,7 @@ static inline void set_real_mode_mem(phys_addr_t mem) + } + + void reserve_real_mode(void); ++void load_trampoline_pgtable(void); + + #endif /* __ASSEMBLY__ */ + +diff --git a/arch/x86/include/asm/topology.h b/arch/x86/include/asm/topology.h +index 9239399e54914..55160445ea78b 100644 +--- a/arch/x86/include/asm/topology.h ++++ b/arch/x86/include/asm/topology.h +@@ -218,7 +218,7 @@ static inline void arch_set_max_freq_ratio(bool turbo_disabled) + } + #endif + +-#ifdef CONFIG_ACPI_CPPC_LIB ++#if defined(CONFIG_ACPI_CPPC_LIB) && defined(CONFIG_SMP) + void init_freq_invariance_cppc(void); + #define init_freq_invariance_cppc init_freq_invariance_cppc + #endif +diff --git a/arch/x86/include/asm/uaccess.h b/arch/x86/include/asm/uaccess.h +index 5c95d242f38d7..bb1430283c726 100644 +--- a/arch/x86/include/asm/uaccess.h ++++ b/arch/x86/include/asm/uaccess.h +@@ -314,11 +314,12 @@ do { \ + do { \ + __chk_user_ptr(ptr); \ + switch (size) { \ +- unsigned char x_u8__; \ +- case 1: \ ++ case 1: { \ ++ unsigned char x_u8__; \ + __get_user_asm(x_u8__, ptr, "b", "=q", label); \ + (x) = x_u8__; \ + break; \ ++ } \ + case 2: \ + __get_user_asm(x, ptr, "w", "=r", label); \ + break; \ +diff --git a/arch/x86/kernel/cpu/mce/core.c b/arch/x86/kernel/cpu/mce/core.c +index 193204aee8801..e23e74e2f928d 100644 +--- a/arch/x86/kernel/cpu/mce/core.c ++++ b/arch/x86/kernel/cpu/mce/core.c +@@ -295,11 +295,17 @@ static void wait_for_panic(void) + panic("Panicing machine check CPU died"); + } + +-static void mce_panic(const char *msg, struct mce *final, char *exp) ++static noinstr void mce_panic(const char *msg, struct mce *final, char *exp) + { +- int apei_err = 0; + struct llist_node *pending; + struct mce_evt_llist *l; ++ int apei_err = 0; ++ ++ /* ++ * Allow instrumentation around external facilities usage. Not that it ++ * matters a whole lot since the machine is going to panic anyway. ++ */ ++ instrumentation_begin(); + + if (!fake_panic) { + /* +@@ -314,7 +320,7 @@ static void mce_panic(const char *msg, struct mce *final, char *exp) + } else { + /* Don't log too much for fake panic */ + if (atomic_inc_return(&mce_fake_panicked) > 1) +- return; ++ goto out; + } + pending = mce_gen_pool_prepare_records(); + /* First print corrected ones that are still unlogged */ +@@ -352,6 +358,9 @@ static void mce_panic(const char *msg, struct mce *final, char *exp) + panic(msg); + } else + pr_emerg(HW_ERR "Fake kernel panic: %s\n", msg); ++ ++out: ++ instrumentation_end(); + } + + /* Support code for software error injection */ +@@ -682,7 +691,7 @@ static struct notifier_block mce_default_nb = { + /* + * Read ADDR and MISC registers. + */ +-static void mce_read_aux(struct mce *m, int i) ++static noinstr void mce_read_aux(struct mce *m, int i) + { + if (m->status & MCI_STATUS_MISCV) + m->misc = mce_rdmsrl(msr_ops.misc(i)); +@@ -1072,10 +1081,13 @@ static int mce_start(int *no_way_out) + * Synchronize between CPUs after main scanning loop. + * This invokes the bulk of the Monarch processing. + */ +-static int mce_end(int order) ++static noinstr int mce_end(int order) + { +- int ret = -1; + u64 timeout = (u64)mca_cfg.monarch_timeout * NSEC_PER_USEC; ++ int ret = -1; ++ ++ /* Allow instrumentation around external facilities. */ ++ instrumentation_begin(); + + if (!timeout) + goto reset; +@@ -1119,7 +1131,8 @@ static int mce_end(int order) + /* + * Don't reset anything. That's done by the Monarch. + */ +- return 0; ++ ret = 0; ++ goto out; + } + + /* +@@ -1135,6 +1148,10 @@ reset: + * Let others run again. + */ + atomic_set(&mce_executing, 0); ++ ++out: ++ instrumentation_end(); ++ + return ret; + } + +@@ -1454,6 +1471,14 @@ noinstr void do_machine_check(struct pt_regs *regs) + if (worst != MCE_AR_SEVERITY && !kill_current_task) + goto out; + ++ /* ++ * Enable instrumentation around the external facilities like ++ * task_work_add() (via queue_task_work()), fixup_exception() etc. ++ * For now, that is. Fixing this properly would need a lot more involved ++ * reorganization. ++ */ ++ instrumentation_begin(); ++ + /* Fault was in user mode and we need to take some action */ + if ((m.cs & 3) == 3) { + /* If this triggers there is no way to recover. Die hard. */ +@@ -1479,6 +1504,9 @@ noinstr void do_machine_check(struct pt_regs *regs) + if (m.kflags & MCE_IN_KERNEL_COPYIN) + queue_task_work(&m, msg, kill_current_task); + } ++ ++ instrumentation_end(); ++ + out: + mce_wrmsrl(MSR_IA32_MCG_STATUS, 0); + } +diff --git a/arch/x86/kernel/cpu/mce/inject.c b/arch/x86/kernel/cpu/mce/inject.c +index 0bfc14041bbb4..b63b548497c14 100644 +--- a/arch/x86/kernel/cpu/mce/inject.c ++++ b/arch/x86/kernel/cpu/mce/inject.c +@@ -350,7 +350,7 @@ static ssize_t flags_write(struct file *filp, const char __user *ubuf, + char buf[MAX_FLAG_OPT_SIZE], *__buf; + int err; + +- if (cnt > MAX_FLAG_OPT_SIZE) ++ if (!cnt || cnt > MAX_FLAG_OPT_SIZE) + return -EINVAL; + + if (copy_from_user(&buf, ubuf, cnt)) +diff --git a/arch/x86/kernel/early-quirks.c b/arch/x86/kernel/early-quirks.c +index 391a4e2b86049..8690fab95ae4b 100644 +--- a/arch/x86/kernel/early-quirks.c ++++ b/arch/x86/kernel/early-quirks.c +@@ -515,6 +515,7 @@ static const struct intel_early_ops gen11_early_ops __initconst = { + .stolen_size = gen9_stolen_size, + }; + ++/* Intel integrated GPUs for which we need to reserve "stolen memory" */ + static const struct pci_device_id intel_early_ids[] __initconst = { + INTEL_I830_IDS(&i830_early_ops), + INTEL_I845G_IDS(&i845_early_ops), +@@ -591,6 +592,13 @@ static void __init intel_graphics_quirks(int num, int slot, int func) + u16 device; + int i; + ++ /* ++ * Reserve "stolen memory" for an integrated GPU. If we've already ++ * found one, there's nothing to do for other (discrete) GPUs. ++ */ ++ if (resource_size(&intel_graphics_stolen_res)) ++ return; ++ + device = read_pci_config_16(num, slot, func, PCI_DEVICE_ID); + + for (i = 0; i < ARRAY_SIZE(intel_early_ids); i++) { +@@ -703,7 +711,7 @@ static struct chipset early_qrk[] __initdata = { + { PCI_VENDOR_ID_INTEL, 0x3406, PCI_CLASS_BRIDGE_HOST, + PCI_BASE_CLASS_BRIDGE, 0, intel_remapping_check }, + { PCI_VENDOR_ID_INTEL, PCI_ANY_ID, PCI_CLASS_DISPLAY_VGA, PCI_ANY_ID, +- QFLAG_APPLY_ONCE, intel_graphics_quirks }, ++ 0, intel_graphics_quirks }, + /* + * HPET on the current version of the Baytrail platform has accuracy + * problems: it will halt in deep idle state - so we disable it. +diff --git a/arch/x86/kernel/reboot.c b/arch/x86/kernel/reboot.c +index 0a40df66a40de..fa700b46588e0 100644 +--- a/arch/x86/kernel/reboot.c ++++ b/arch/x86/kernel/reboot.c +@@ -113,17 +113,9 @@ void __noreturn machine_real_restart(unsigned int type) + spin_unlock(&rtc_lock); + + /* +- * Switch back to the initial page table. ++ * Switch to the trampoline page table. + */ +-#ifdef CONFIG_X86_32 +- load_cr3(initial_page_table); +-#else +- write_cr3(real_mode_header->trampoline_pgd); +- +- /* Exiting long mode will fail if CR4.PCIDE is set. */ +- if (boot_cpu_has(X86_FEATURE_PCID)) +- cr4_clear_bits(X86_CR4_PCIDE); +-#endif ++ load_trampoline_pgtable(); + + /* Jump to the identity-mapped low memory code */ + #ifdef CONFIG_X86_32 +diff --git a/arch/x86/kvm/mmu/tdp_mmu.c b/arch/x86/kvm/mmu/tdp_mmu.c +index 4954cd01c1f0e..d479b2b121259 100644 +--- a/arch/x86/kvm/mmu/tdp_mmu.c ++++ b/arch/x86/kvm/mmu/tdp_mmu.c +@@ -1493,12 +1493,12 @@ static bool write_protect_gfn(struct kvm *kvm, struct kvm_mmu_page *root, + !is_last_spte(iter.old_spte, iter.level)) + continue; + +- if (!is_writable_pte(iter.old_spte)) +- break; +- + new_spte = iter.old_spte & + ~(PT_WRITABLE_MASK | shadow_mmu_writable_mask); + ++ if (new_spte == iter.old_spte) ++ break; ++ + tdp_mmu_set_spte(kvm, &iter, new_spte); + spte_set = true; + } +diff --git a/arch/x86/kvm/vmx/posted_intr.c b/arch/x86/kvm/vmx/posted_intr.c +index 1c94783b5a54c..21ea58d25771f 100644 +--- a/arch/x86/kvm/vmx/posted_intr.c ++++ b/arch/x86/kvm/vmx/posted_intr.c +@@ -15,7 +15,7 @@ + * can find which vCPU should be waken up. + */ + static DEFINE_PER_CPU(struct list_head, blocked_vcpu_on_cpu); +-static DEFINE_PER_CPU(spinlock_t, blocked_vcpu_on_cpu_lock); ++static DEFINE_PER_CPU(raw_spinlock_t, blocked_vcpu_on_cpu_lock); + + static inline struct pi_desc *vcpu_to_pi_desc(struct kvm_vcpu *vcpu) + { +@@ -121,9 +121,9 @@ static void __pi_post_block(struct kvm_vcpu *vcpu) + new.control) != old.control); + + if (!WARN_ON_ONCE(vcpu->pre_pcpu == -1)) { +- spin_lock(&per_cpu(blocked_vcpu_on_cpu_lock, vcpu->pre_pcpu)); ++ raw_spin_lock(&per_cpu(blocked_vcpu_on_cpu_lock, vcpu->pre_pcpu)); + list_del(&vcpu->blocked_vcpu_list); +- spin_unlock(&per_cpu(blocked_vcpu_on_cpu_lock, vcpu->pre_pcpu)); ++ raw_spin_unlock(&per_cpu(blocked_vcpu_on_cpu_lock, vcpu->pre_pcpu)); + vcpu->pre_pcpu = -1; + } + } +@@ -154,11 +154,11 @@ int pi_pre_block(struct kvm_vcpu *vcpu) + local_irq_disable(); + if (!WARN_ON_ONCE(vcpu->pre_pcpu != -1)) { + vcpu->pre_pcpu = vcpu->cpu; +- spin_lock(&per_cpu(blocked_vcpu_on_cpu_lock, vcpu->pre_pcpu)); ++ raw_spin_lock(&per_cpu(blocked_vcpu_on_cpu_lock, vcpu->pre_pcpu)); + list_add_tail(&vcpu->blocked_vcpu_list, + &per_cpu(blocked_vcpu_on_cpu, + vcpu->pre_pcpu)); +- spin_unlock(&per_cpu(blocked_vcpu_on_cpu_lock, vcpu->pre_pcpu)); ++ raw_spin_unlock(&per_cpu(blocked_vcpu_on_cpu_lock, vcpu->pre_pcpu)); + } + + do { +@@ -215,7 +215,7 @@ void pi_wakeup_handler(void) + struct kvm_vcpu *vcpu; + int cpu = smp_processor_id(); + +- spin_lock(&per_cpu(blocked_vcpu_on_cpu_lock, cpu)); ++ raw_spin_lock(&per_cpu(blocked_vcpu_on_cpu_lock, cpu)); + list_for_each_entry(vcpu, &per_cpu(blocked_vcpu_on_cpu, cpu), + blocked_vcpu_list) { + struct pi_desc *pi_desc = vcpu_to_pi_desc(vcpu); +@@ -223,13 +223,13 @@ void pi_wakeup_handler(void) + if (pi_test_on(pi_desc) == 1) + kvm_vcpu_kick(vcpu); + } +- spin_unlock(&per_cpu(blocked_vcpu_on_cpu_lock, cpu)); ++ raw_spin_unlock(&per_cpu(blocked_vcpu_on_cpu_lock, cpu)); + } + + void __init pi_init_cpu(int cpu) + { + INIT_LIST_HEAD(&per_cpu(blocked_vcpu_on_cpu, cpu)); +- spin_lock_init(&per_cpu(blocked_vcpu_on_cpu_lock, cpu)); ++ raw_spin_lock_init(&per_cpu(blocked_vcpu_on_cpu_lock, cpu)); + } + + bool pi_has_pending_interrupt(struct kvm_vcpu *vcpu) +diff --git a/arch/x86/realmode/init.c b/arch/x86/realmode/init.c +index d3eee1ebcf1d5..1d20ed4b28729 100644 +--- a/arch/x86/realmode/init.c ++++ b/arch/x86/realmode/init.c +@@ -17,6 +17,32 @@ u32 *trampoline_cr4_features; + /* Hold the pgd entry used on booting additional CPUs */ + pgd_t trampoline_pgd_entry; + ++void load_trampoline_pgtable(void) ++{ ++#ifdef CONFIG_X86_32 ++ load_cr3(initial_page_table); ++#else ++ /* ++ * This function is called before exiting to real-mode and that will ++ * fail with CR4.PCIDE still set. ++ */ ++ if (boot_cpu_has(X86_FEATURE_PCID)) ++ cr4_clear_bits(X86_CR4_PCIDE); ++ ++ write_cr3(real_mode_header->trampoline_pgd); ++#endif ++ ++ /* ++ * The CR3 write above will not flush global TLB entries. ++ * Stale, global entries from previous page tables may still be ++ * present. Flush those stale entries. ++ * ++ * This ensures that memory accessed while running with ++ * trampoline_pgd is *actually* mapped into trampoline_pgd. ++ */ ++ __flush_tlb_all(); ++} ++ + void __init reserve_real_mode(void) + { + phys_addr_t mem; +diff --git a/arch/x86/um/syscalls_64.c b/arch/x86/um/syscalls_64.c +index 58f51667e2e4b..8249685b40960 100644 +--- a/arch/x86/um/syscalls_64.c ++++ b/arch/x86/um/syscalls_64.c +@@ -11,6 +11,7 @@ + #include <linux/uaccess.h> + #include <asm/prctl.h> /* XXX This should get the constants from libc */ + #include <os.h> ++#include <registers.h> + + long arch_prctl(struct task_struct *task, int option, + unsigned long __user *arg2) +@@ -35,7 +36,7 @@ long arch_prctl(struct task_struct *task, int option, + switch (option) { + case ARCH_SET_FS: + case ARCH_SET_GS: +- ret = restore_registers(pid, ¤t->thread.regs.regs); ++ ret = restore_pid_registers(pid, ¤t->thread.regs.regs); + if (ret) + return ret; + break; +diff --git a/block/bfq-iosched.c b/block/bfq-iosched.c +index 480e1a1348596..ea9a086d0498f 100644 +--- a/block/bfq-iosched.c ++++ b/block/bfq-iosched.c +@@ -5991,48 +5991,7 @@ static void bfq_insert_request(struct blk_mq_hw_ctx *hctx, struct request *rq, + + spin_lock_irq(&bfqd->lock); + bfqq = bfq_init_rq(rq); +- +- /* +- * Reqs with at_head or passthrough flags set are to be put +- * directly into dispatch list. Additional case for putting rq +- * directly into the dispatch queue: the only active +- * bfq_queues are bfqq and either its waker bfq_queue or one +- * of its woken bfq_queues. The rationale behind this +- * additional condition is as follows: +- * - consider a bfq_queue, say Q1, detected as a waker of +- * another bfq_queue, say Q2 +- * - by definition of a waker, Q1 blocks the I/O of Q2, i.e., +- * some I/O of Q1 needs to be completed for new I/O of Q2 +- * to arrive. A notable example of waker is journald +- * - so, Q1 and Q2 are in any respect the queues of two +- * cooperating processes (or of two cooperating sets of +- * processes): the goal of Q1's I/O is doing what needs to +- * be done so that new Q2's I/O can finally be +- * issued. Therefore, if the service of Q1's I/O is delayed, +- * then Q2's I/O is delayed too. Conversely, if Q2's I/O is +- * delayed, the goal of Q1's I/O is hindered. +- * - as a consequence, if some I/O of Q1/Q2 arrives while +- * Q2/Q1 is the only queue in service, there is absolutely +- * no point in delaying the service of such an I/O. The +- * only possible result is a throughput loss +- * - so, when the above condition holds, the best option is to +- * have the new I/O dispatched as soon as possible +- * - the most effective and efficient way to attain the above +- * goal is to put the new I/O directly in the dispatch +- * list +- * - as an additional restriction, Q1 and Q2 must be the only +- * busy queues for this commit to put the I/O of Q2/Q1 in +- * the dispatch list. This is necessary, because, if also +- * other queues are waiting for service, then putting new +- * I/O directly in the dispatch list may evidently cause a +- * violation of service guarantees for the other queues +- */ +- if (!bfqq || +- (bfqq != bfqd->in_service_queue && +- bfqd->in_service_queue != NULL && +- bfq_tot_busy_queues(bfqd) == 1 + bfq_bfqq_busy(bfqq) && +- (bfqq->waker_bfqq == bfqd->in_service_queue || +- bfqd->in_service_queue->waker_bfqq == bfqq)) || at_head) { ++ if (!bfqq || at_head) { + if (at_head) + list_add(&rq->queuelist, &bfqd->dispatch); + else +@@ -6059,7 +6018,6 @@ static void bfq_insert_request(struct blk_mq_hw_ctx *hctx, struct request *rq, + * merge). + */ + cmd_flags = rq->cmd_flags; +- + spin_unlock_irq(&bfqd->lock); + + bfq_update_insert_stats(q, bfqq, idle_timer_disabled, +diff --git a/block/blk-flush.c b/block/blk-flush.c +index 4201728bf3a5a..94a86acbb7f67 100644 +--- a/block/blk-flush.c ++++ b/block/blk-flush.c +@@ -235,8 +235,10 @@ static void flush_end_io(struct request *flush_rq, blk_status_t error) + * avoiding use-after-free. + */ + WRITE_ONCE(flush_rq->state, MQ_RQ_IDLE); +- if (fq->rq_status != BLK_STS_OK) ++ if (fq->rq_status != BLK_STS_OK) { + error = fq->rq_status; ++ fq->rq_status = BLK_STS_OK; ++ } + + if (!q->elevator) { + flush_rq->tag = BLK_MQ_NO_TAG; +diff --git a/block/blk-pm.c b/block/blk-pm.c +index 17bd020268d42..2dad62cc15727 100644 +--- a/block/blk-pm.c ++++ b/block/blk-pm.c +@@ -163,27 +163,19 @@ EXPORT_SYMBOL(blk_pre_runtime_resume); + /** + * blk_post_runtime_resume - Post runtime resume processing + * @q: the queue of the device +- * @err: return value of the device's runtime_resume function + * + * Description: +- * Update the queue's runtime status according to the return value of the +- * device's runtime_resume function. If the resume was successful, call +- * blk_set_runtime_active() to do the real work of restarting the queue. ++ * For historical reasons, this routine merely calls blk_set_runtime_active() ++ * to do the real work of restarting the queue. It does this regardless of ++ * whether the device's runtime-resume succeeded; even if it failed the ++ * driver or error handler will need to communicate with the device. + * + * This function should be called near the end of the device's + * runtime_resume callback. + */ +-void blk_post_runtime_resume(struct request_queue *q, int err) ++void blk_post_runtime_resume(struct request_queue *q) + { +- if (!q->dev) +- return; +- if (!err) { +- blk_set_runtime_active(q); +- } else { +- spin_lock_irq(&q->queue_lock); +- q->rpm_status = RPM_SUSPENDED; +- spin_unlock_irq(&q->queue_lock); +- } ++ blk_set_runtime_active(q); + } + EXPORT_SYMBOL(blk_post_runtime_resume); + +@@ -201,7 +193,7 @@ EXPORT_SYMBOL(blk_post_runtime_resume); + * runtime PM status and re-enable peeking requests from the queue. It + * should be called before first request is added to the queue. + * +- * This function is also called by blk_post_runtime_resume() for successful ++ * This function is also called by blk_post_runtime_resume() for + * runtime resumes. It does everything necessary to restart the queue. + */ + void blk_set_runtime_active(struct request_queue *q) +diff --git a/block/genhd.c b/block/genhd.c +index f091a60dcf1ea..de789d1a1e3d2 100644 +--- a/block/genhd.c ++++ b/block/genhd.c +@@ -420,6 +420,8 @@ int device_add_disk(struct device *parent, struct gendisk *disk, + DISK_MAX_PARTS); + disk->minors = DISK_MAX_PARTS; + } ++ if (disk->first_minor + disk->minors > MINORMASK + 1) ++ return -EINVAL; + } else { + if (WARN_ON(disk->minors)) + return -EINVAL; +@@ -432,10 +434,6 @@ int device_add_disk(struct device *parent, struct gendisk *disk, + disk->flags |= GENHD_FL_EXT_DEVT; + } + +- ret = disk_alloc_events(disk); +- if (ret) +- goto out_free_ext_minor; +- + /* delay uevents, until we scanned partition table */ + dev_set_uevent_suppress(ddev, 1); + +@@ -446,7 +444,12 @@ int device_add_disk(struct device *parent, struct gendisk *disk, + ddev->devt = MKDEV(disk->major, disk->first_minor); + ret = device_add(ddev); + if (ret) +- goto out_disk_release_events; ++ goto out_free_ext_minor; ++ ++ ret = disk_alloc_events(disk); ++ if (ret) ++ goto out_device_del; ++ + if (!sysfs_deprecated) { + ret = sysfs_create_link(block_depr, &ddev->kobj, + kobject_name(&ddev->kobj)); +@@ -534,8 +537,6 @@ out_del_block_link: + sysfs_remove_link(block_depr, dev_name(ddev)); + out_device_del: + device_del(ddev); +-out_disk_release_events: +- disk_release_events(disk); + out_free_ext_minor: + if (disk->major == BLOCK_EXT_MAJOR) + blk_free_ext_minor(disk->first_minor); +diff --git a/block/mq-deadline.c b/block/mq-deadline.c +index 7f3c3932b723e..cd2342d297048 100644 +--- a/block/mq-deadline.c ++++ b/block/mq-deadline.c +@@ -811,7 +811,7 @@ SHOW_JIFFIES(deadline_read_expire_show, dd->fifo_expire[DD_READ]); + SHOW_JIFFIES(deadline_write_expire_show, dd->fifo_expire[DD_WRITE]); + SHOW_INT(deadline_writes_starved_show, dd->writes_starved); + SHOW_INT(deadline_front_merges_show, dd->front_merges); +-SHOW_INT(deadline_async_depth_show, dd->front_merges); ++SHOW_INT(deadline_async_depth_show, dd->async_depth); + SHOW_INT(deadline_fifo_batch_show, dd->fifo_batch); + #undef SHOW_INT + #undef SHOW_JIFFIES +@@ -840,7 +840,7 @@ STORE_JIFFIES(deadline_read_expire_store, &dd->fifo_expire[DD_READ], 0, INT_MAX) + STORE_JIFFIES(deadline_write_expire_store, &dd->fifo_expire[DD_WRITE], 0, INT_MAX); + STORE_INT(deadline_writes_starved_store, &dd->writes_starved, INT_MIN, INT_MAX); + STORE_INT(deadline_front_merges_store, &dd->front_merges, 0, 1); +-STORE_INT(deadline_async_depth_store, &dd->front_merges, 1, INT_MAX); ++STORE_INT(deadline_async_depth_store, &dd->async_depth, 1, INT_MAX); + STORE_INT(deadline_fifo_batch_store, &dd->fifo_batch, 0, INT_MAX); + #undef STORE_FUNCTION + #undef STORE_INT +diff --git a/crypto/jitterentropy.c b/crypto/jitterentropy.c +index a11b3208760f3..f6d3a84e3c214 100644 +--- a/crypto/jitterentropy.c ++++ b/crypto/jitterentropy.c +@@ -265,7 +265,6 @@ static int jent_stuck(struct rand_data *ec, __u64 current_delta) + { + __u64 delta2 = jent_delta(ec->last_delta, current_delta); + __u64 delta3 = jent_delta(ec->last_delta2, delta2); +- unsigned int delta_masked = current_delta & JENT_APT_WORD_MASK; + + ec->last_delta = current_delta; + ec->last_delta2 = delta2; +@@ -274,7 +273,7 @@ static int jent_stuck(struct rand_data *ec, __u64 current_delta) + * Insert the result of the comparison of two back-to-back time + * deltas. + */ +- jent_apt_insert(ec, delta_masked); ++ jent_apt_insert(ec, current_delta); + + if (!current_delta || !delta2 || !delta3) { + /* RCT with a stuck bit */ +diff --git a/drivers/acpi/acpica/exfield.c b/drivers/acpi/acpica/exfield.c +index 06f3c9df1e22d..8618500f23b39 100644 +--- a/drivers/acpi/acpica/exfield.c ++++ b/drivers/acpi/acpica/exfield.c +@@ -330,12 +330,7 @@ acpi_ex_write_data_to_field(union acpi_operand_object *source_desc, + obj_desc->field.base_byte_offset, + source_desc->buffer.pointer, data_length); + +- if ((obj_desc->field.region_obj->region.address == +- PCC_MASTER_SUBSPACE +- && MASTER_SUBSPACE_COMMAND(obj_desc->field. +- base_byte_offset)) +- || GENERIC_SUBSPACE_COMMAND(obj_desc->field. +- base_byte_offset)) { ++ if (MASTER_SUBSPACE_COMMAND(obj_desc->field.base_byte_offset)) { + + /* Perform the write */ + +diff --git a/drivers/acpi/acpica/exoparg1.c b/drivers/acpi/acpica/exoparg1.c +index b639e930d6429..44b7c350ed5ca 100644 +--- a/drivers/acpi/acpica/exoparg1.c ++++ b/drivers/acpi/acpica/exoparg1.c +@@ -1007,7 +1007,8 @@ acpi_status acpi_ex_opcode_1A_0T_1R(struct acpi_walk_state *walk_state) + (walk_state, return_desc, + &temp_desc); + if (ACPI_FAILURE(status)) { +- goto cleanup; ++ return_ACPI_STATUS ++ (status); + } + + return_desc = temp_desc; +diff --git a/drivers/acpi/acpica/hwesleep.c b/drivers/acpi/acpica/hwesleep.c +index 808fdf54aeebf..7ee2939c08cd4 100644 +--- a/drivers/acpi/acpica/hwesleep.c ++++ b/drivers/acpi/acpica/hwesleep.c +@@ -104,7 +104,9 @@ acpi_status acpi_hw_extended_sleep(u8 sleep_state) + + /* Flush caches, as per ACPI specification */ + +- ACPI_FLUSH_CPU_CACHE(); ++ if (sleep_state < ACPI_STATE_S4) { ++ ACPI_FLUSH_CPU_CACHE(); ++ } + + status = acpi_os_enter_sleep(sleep_state, sleep_control, 0); + if (status == AE_CTRL_TERMINATE) { +diff --git a/drivers/acpi/acpica/hwsleep.c b/drivers/acpi/acpica/hwsleep.c +index 34a3825f25d37..5efa3d8e483e0 100644 +--- a/drivers/acpi/acpica/hwsleep.c ++++ b/drivers/acpi/acpica/hwsleep.c +@@ -110,7 +110,9 @@ acpi_status acpi_hw_legacy_sleep(u8 sleep_state) + + /* Flush caches, as per ACPI specification */ + +- ACPI_FLUSH_CPU_CACHE(); ++ if (sleep_state < ACPI_STATE_S4) { ++ ACPI_FLUSH_CPU_CACHE(); ++ } + + status = acpi_os_enter_sleep(sleep_state, pm1a_control, pm1b_control); + if (status == AE_CTRL_TERMINATE) { +diff --git a/drivers/acpi/acpica/hwxfsleep.c b/drivers/acpi/acpica/hwxfsleep.c +index e4cde23a29061..ba77598ee43e8 100644 +--- a/drivers/acpi/acpica/hwxfsleep.c ++++ b/drivers/acpi/acpica/hwxfsleep.c +@@ -162,8 +162,6 @@ acpi_status acpi_enter_sleep_state_s4bios(void) + return_ACPI_STATUS(status); + } + +- ACPI_FLUSH_CPU_CACHE(); +- + status = acpi_hw_write_port(acpi_gbl_FADT.smi_command, + (u32)acpi_gbl_FADT.s4_bios_request, 8); + if (ACPI_FAILURE(status)) { +diff --git a/drivers/acpi/acpica/utdelete.c b/drivers/acpi/acpica/utdelete.c +index e5ba9795ec696..8d7736d2d2699 100644 +--- a/drivers/acpi/acpica/utdelete.c ++++ b/drivers/acpi/acpica/utdelete.c +@@ -422,6 +422,7 @@ acpi_ut_update_ref_count(union acpi_operand_object *object, u32 action) + ACPI_WARNING((AE_INFO, + "Obj %p, Reference Count is already zero, cannot decrement\n", + object)); ++ return; + } + + ACPI_DEBUG_PRINT_RAW((ACPI_DB_ALLOCATIONS, +diff --git a/drivers/acpi/battery.c b/drivers/acpi/battery.c +index 8afa85d6eb6a7..ead0114f27c9f 100644 +--- a/drivers/acpi/battery.c ++++ b/drivers/acpi/battery.c +@@ -53,6 +53,7 @@ static int battery_bix_broken_package; + static int battery_notification_delay_ms; + static int battery_ac_is_broken; + static int battery_check_pmic = 1; ++static int battery_quirk_notcharging; + static unsigned int cache_time = 1000; + module_param(cache_time, uint, 0644); + MODULE_PARM_DESC(cache_time, "cache time in milliseconds"); +@@ -217,6 +218,8 @@ static int acpi_battery_get_property(struct power_supply *psy, + val->intval = POWER_SUPPLY_STATUS_CHARGING; + else if (acpi_battery_is_charged(battery)) + val->intval = POWER_SUPPLY_STATUS_FULL; ++ else if (battery_quirk_notcharging) ++ val->intval = POWER_SUPPLY_STATUS_NOT_CHARGING; + else + val->intval = POWER_SUPPLY_STATUS_UNKNOWN; + break; +@@ -1111,6 +1114,12 @@ battery_do_not_check_pmic_quirk(const struct dmi_system_id *d) + return 0; + } + ++static int __init battery_quirk_not_charging(const struct dmi_system_id *d) ++{ ++ battery_quirk_notcharging = 1; ++ return 0; ++} ++ + static const struct dmi_system_id bat_dmi_table[] __initconst = { + { + /* NEC LZ750/LS */ +@@ -1155,6 +1164,19 @@ static const struct dmi_system_id bat_dmi_table[] __initconst = { + DMI_MATCH(DMI_PRODUCT_VERSION, "Lenovo MIIX 320-10ICR"), + }, + }, ++ { ++ /* ++ * On Lenovo ThinkPads the BIOS specification defines ++ * a state when the bits for charging and discharging ++ * are both set to 0. That state is "Not Charging". ++ */ ++ .callback = battery_quirk_not_charging, ++ .ident = "Lenovo ThinkPad", ++ .matches = { ++ DMI_MATCH(DMI_SYS_VENDOR, "LENOVO"), ++ DMI_MATCH(DMI_PRODUCT_VERSION, "ThinkPad"), ++ }, ++ }, + {}, + }; + +diff --git a/drivers/acpi/bus.c b/drivers/acpi/bus.c +index fa923a9292244..dd535b4b9a160 100644 +--- a/drivers/acpi/bus.c ++++ b/drivers/acpi/bus.c +@@ -98,8 +98,8 @@ int acpi_bus_get_status(struct acpi_device *device) + acpi_status status; + unsigned long long sta; + +- if (acpi_device_always_present(device)) { +- acpi_set_device_status(device, ACPI_STA_DEFAULT); ++ if (acpi_device_override_status(device, &sta)) { ++ acpi_set_device_status(device, sta); + return 0; + } + +diff --git a/drivers/acpi/cppc_acpi.c b/drivers/acpi/cppc_acpi.c +index 3fbb17ecce2d5..6fe28a2d387bd 100644 +--- a/drivers/acpi/cppc_acpi.c ++++ b/drivers/acpi/cppc_acpi.c +@@ -411,7 +411,7 @@ bool acpi_cpc_valid(void) + struct cpc_desc *cpc_ptr; + int cpu; + +- for_each_possible_cpu(cpu) { ++ for_each_present_cpu(cpu) { + cpc_ptr = per_cpu(cpc_desc_ptr, cpu); + if (!cpc_ptr) + return false; +diff --git a/drivers/acpi/ec.c b/drivers/acpi/ec.c +index e629e891d1bb3..9b859ff976e89 100644 +--- a/drivers/acpi/ec.c ++++ b/drivers/acpi/ec.c +@@ -166,6 +166,7 @@ struct acpi_ec_query { + struct transaction transaction; + struct work_struct work; + struct acpi_ec_query_handler *handler; ++ struct acpi_ec *ec; + }; + + static int acpi_ec_query(struct acpi_ec *ec, u8 *data); +@@ -452,6 +453,7 @@ static void acpi_ec_submit_query(struct acpi_ec *ec) + ec_dbg_evt("Command(%s) submitted/blocked", + acpi_ec_cmd_string(ACPI_EC_COMMAND_QUERY)); + ec->nr_pending_queries++; ++ ec->events_in_progress++; + queue_work(ec_wq, &ec->work); + } + } +@@ -518,7 +520,7 @@ static void acpi_ec_enable_event(struct acpi_ec *ec) + #ifdef CONFIG_PM_SLEEP + static void __acpi_ec_flush_work(void) + { +- drain_workqueue(ec_wq); /* flush ec->work */ ++ flush_workqueue(ec_wq); /* flush ec->work */ + flush_workqueue(ec_query_wq); /* flush queries */ + } + +@@ -1103,7 +1105,7 @@ void acpi_ec_remove_query_handler(struct acpi_ec *ec, u8 query_bit) + } + EXPORT_SYMBOL_GPL(acpi_ec_remove_query_handler); + +-static struct acpi_ec_query *acpi_ec_create_query(u8 *pval) ++static struct acpi_ec_query *acpi_ec_create_query(struct acpi_ec *ec, u8 *pval) + { + struct acpi_ec_query *q; + struct transaction *t; +@@ -1111,11 +1113,13 @@ static struct acpi_ec_query *acpi_ec_create_query(u8 *pval) + q = kzalloc(sizeof (struct acpi_ec_query), GFP_KERNEL); + if (!q) + return NULL; ++ + INIT_WORK(&q->work, acpi_ec_event_processor); + t = &q->transaction; + t->command = ACPI_EC_COMMAND_QUERY; + t->rdata = pval; + t->rlen = 1; ++ q->ec = ec; + return q; + } + +@@ -1132,13 +1136,21 @@ static void acpi_ec_event_processor(struct work_struct *work) + { + struct acpi_ec_query *q = container_of(work, struct acpi_ec_query, work); + struct acpi_ec_query_handler *handler = q->handler; ++ struct acpi_ec *ec = q->ec; + + ec_dbg_evt("Query(0x%02x) started", handler->query_bit); ++ + if (handler->func) + handler->func(handler->data); + else if (handler->handle) + acpi_evaluate_object(handler->handle, NULL, NULL, NULL); ++ + ec_dbg_evt("Query(0x%02x) stopped", handler->query_bit); ++ ++ spin_lock_irq(&ec->lock); ++ ec->queries_in_progress--; ++ spin_unlock_irq(&ec->lock); ++ + acpi_ec_delete_query(q); + } + +@@ -1148,7 +1160,7 @@ static int acpi_ec_query(struct acpi_ec *ec, u8 *data) + int result; + struct acpi_ec_query *q; + +- q = acpi_ec_create_query(&value); ++ q = acpi_ec_create_query(ec, &value); + if (!q) + return -ENOMEM; + +@@ -1170,19 +1182,20 @@ static int acpi_ec_query(struct acpi_ec *ec, u8 *data) + } + + /* +- * It is reported that _Qxx are evaluated in a parallel way on +- * Windows: ++ * It is reported that _Qxx are evaluated in a parallel way on Windows: + * https://bugzilla.kernel.org/show_bug.cgi?id=94411 + * +- * Put this log entry before schedule_work() in order to make +- * it appearing before any other log entries occurred during the +- * work queue execution. ++ * Put this log entry before queue_work() to make it appear in the log ++ * before any other messages emitted during workqueue handling. + */ + ec_dbg_evt("Query(0x%02x) scheduled", value); +- if (!queue_work(ec_query_wq, &q->work)) { +- ec_dbg_evt("Query(0x%02x) overlapped", value); +- result = -EBUSY; +- } ++ ++ spin_lock_irq(&ec->lock); ++ ++ ec->queries_in_progress++; ++ queue_work(ec_query_wq, &q->work); ++ ++ spin_unlock_irq(&ec->lock); + + err_exit: + if (result) +@@ -1240,6 +1253,10 @@ static void acpi_ec_event_handler(struct work_struct *work) + ec_dbg_evt("Event stopped"); + + acpi_ec_check_event(ec); ++ ++ spin_lock_irqsave(&ec->lock, flags); ++ ec->events_in_progress--; ++ spin_unlock_irqrestore(&ec->lock, flags); + } + + static void acpi_ec_handle_interrupt(struct acpi_ec *ec) +@@ -2021,6 +2038,7 @@ void acpi_ec_set_gpe_wake_mask(u8 action) + + bool acpi_ec_dispatch_gpe(void) + { ++ bool work_in_progress; + u32 ret; + + if (!first_ec) +@@ -2041,8 +2059,19 @@ bool acpi_ec_dispatch_gpe(void) + if (ret == ACPI_INTERRUPT_HANDLED) + pm_pr_dbg("ACPI EC GPE dispatched\n"); + +- /* Flush the event and query workqueues. */ +- acpi_ec_flush_work(); ++ /* Drain EC work. */ ++ do { ++ acpi_ec_flush_work(); ++ ++ pm_pr_dbg("ACPI EC work flushed\n"); ++ ++ spin_lock_irq(&first_ec->lock); ++ ++ work_in_progress = first_ec->events_in_progress + ++ first_ec->queries_in_progress > 0; ++ ++ spin_unlock_irq(&first_ec->lock); ++ } while (work_in_progress && !pm_wakeup_pending()); + + return false; + } +diff --git a/drivers/acpi/internal.h b/drivers/acpi/internal.h +index d91b560e88674..54b2be94d23dc 100644 +--- a/drivers/acpi/internal.h ++++ b/drivers/acpi/internal.h +@@ -183,6 +183,8 @@ struct acpi_ec { + struct work_struct work; + unsigned long timestamp; + unsigned long nr_pending_queries; ++ unsigned int events_in_progress; ++ unsigned int queries_in_progress; + bool busy_polling; + unsigned int polling_guard; + }; +diff --git a/drivers/acpi/scan.c b/drivers/acpi/scan.c +index 5b54c80b9d32a..6e9cd41c5f9b1 100644 +--- a/drivers/acpi/scan.c ++++ b/drivers/acpi/scan.c +@@ -1690,6 +1690,7 @@ static bool acpi_device_enumeration_by_parent(struct acpi_device *device) + { + struct list_head resource_list; + bool is_serial_bus_slave = false; ++ static const struct acpi_device_id ignore_serial_bus_ids[] = { + /* + * These devices have multiple I2cSerialBus resources and an i2c-client + * must be instantiated for each, each with its own i2c_device_id. +@@ -1698,11 +1699,18 @@ static bool acpi_device_enumeration_by_parent(struct acpi_device *device) + * drivers/platform/x86/i2c-multi-instantiate.c driver, which knows + * which i2c_device_id to use for each resource. + */ +- static const struct acpi_device_id i2c_multi_instantiate_ids[] = { + {"BSG1160", }, + {"BSG2150", }, + {"INT33FE", }, + {"INT3515", }, ++ /* ++ * HIDs of device with an UartSerialBusV2 resource for which userspace ++ * expects a regular tty cdev to be created (instead of the in kernel ++ * serdev) and which have a kernel driver which expects a platform_dev ++ * such as the rfkill-gpio driver. ++ */ ++ {"BCM4752", }, ++ {"LNV4752", }, + {} + }; + +@@ -1716,8 +1724,7 @@ static bool acpi_device_enumeration_by_parent(struct acpi_device *device) + fwnode_property_present(&device->fwnode, "baud"))) + return true; + +- /* Instantiate a pdev for the i2c-multi-instantiate drv to bind to */ +- if (!acpi_match_device_ids(device, i2c_multi_instantiate_ids)) ++ if (!acpi_match_device_ids(device, ignore_serial_bus_ids)) + return false; + + INIT_LIST_HEAD(&resource_list); +diff --git a/drivers/acpi/x86/utils.c b/drivers/acpi/x86/utils.c +index f22f23933063b..b3fb428461c6f 100644 +--- a/drivers/acpi/x86/utils.c ++++ b/drivers/acpi/x86/utils.c +@@ -22,58 +22,71 @@ + * Some BIOS-es (temporarily) hide specific APCI devices to work around Windows + * driver bugs. We use DMI matching to match known cases of this. + * +- * We work around this by always reporting ACPI_STA_DEFAULT for these +- * devices. Note this MUST only be done for devices where this is safe. ++ * Likewise sometimes some not-actually present devices are sometimes ++ * reported as present, which may cause issues. + * +- * This forcing of devices to be present is limited to specific CPU (SoC) +- * models both to avoid potentially causing trouble on other models and +- * because some HIDs are re-used on different SoCs for completely +- * different devices. ++ * We work around this by using the below quirk list to override the status ++ * reported by the _STA method with a fixed value (ACPI_STA_DEFAULT or 0). ++ * Note this MUST only be done for devices where this is safe. ++ * ++ * This status overriding is limited to specific CPU (SoC) models both to ++ * avoid potentially causing trouble on other models and because some HIDs ++ * are re-used on different SoCs for completely different devices. + */ +-struct always_present_id { ++struct override_status_id { + struct acpi_device_id hid[2]; + struct x86_cpu_id cpu_ids[2]; + struct dmi_system_id dmi_ids[2]; /* Optional */ + const char *uid; ++ const char *path; ++ unsigned long long status; + }; + +-#define X86_MATCH(model) X86_MATCH_INTEL_FAM6_MODEL(model, NULL) +- +-#define ENTRY(hid, uid, cpu_models, dmi...) { \ ++#define ENTRY(status, hid, uid, path, cpu_model, dmi...) { \ + { { hid, }, {} }, \ +- { cpu_models, {} }, \ ++ { X86_MATCH_INTEL_FAM6_MODEL(cpu_model, NULL), {} }, \ + { { .matches = dmi }, {} }, \ + uid, \ ++ path, \ ++ status, \ + } + +-static const struct always_present_id always_present_ids[] = { ++#define PRESENT_ENTRY_HID(hid, uid, cpu_model, dmi...) \ ++ ENTRY(ACPI_STA_DEFAULT, hid, uid, NULL, cpu_model, dmi) ++ ++#define NOT_PRESENT_ENTRY_HID(hid, uid, cpu_model, dmi...) \ ++ ENTRY(0, hid, uid, NULL, cpu_model, dmi) ++ ++#define PRESENT_ENTRY_PATH(path, cpu_model, dmi...) \ ++ ENTRY(ACPI_STA_DEFAULT, "", NULL, path, cpu_model, dmi) ++ ++#define NOT_PRESENT_ENTRY_PATH(path, cpu_model, dmi...) \ ++ ENTRY(0, "", NULL, path, cpu_model, dmi) ++ ++static const struct override_status_id override_status_ids[] = { + /* + * Bay / Cherry Trail PWM directly poked by GPU driver in win10, + * but Linux uses a separate PWM driver, harmless if not used. + */ +- ENTRY("80860F09", "1", X86_MATCH(ATOM_SILVERMONT), {}), +- ENTRY("80862288", "1", X86_MATCH(ATOM_AIRMONT), {}), ++ PRESENT_ENTRY_HID("80860F09", "1", ATOM_SILVERMONT, {}), ++ PRESENT_ENTRY_HID("80862288", "1", ATOM_AIRMONT, {}), + +- /* Lenovo Yoga Book uses PWM2 for keyboard backlight control */ +- ENTRY("80862289", "2", X86_MATCH(ATOM_AIRMONT), { +- DMI_MATCH(DMI_PRODUCT_NAME, "Lenovo YB1-X9"), +- }), + /* + * The INT0002 device is necessary to clear wakeup interrupt sources + * on Cherry Trail devices, without it we get nobody cared IRQ msgs. + */ +- ENTRY("INT0002", "1", X86_MATCH(ATOM_AIRMONT), {}), ++ PRESENT_ENTRY_HID("INT0002", "1", ATOM_AIRMONT, {}), + /* + * On the Dell Venue 11 Pro 7130 and 7139, the DSDT hides + * the touchscreen ACPI device until a certain time + * after _SB.PCI0.GFX0.LCD.LCD1._ON gets called has passed + * *and* _STA has been called at least 3 times since. + */ +- ENTRY("SYNA7500", "1", X86_MATCH(HASWELL_L), { ++ PRESENT_ENTRY_HID("SYNA7500", "1", HASWELL_L, { + DMI_MATCH(DMI_SYS_VENDOR, "Dell Inc."), + DMI_MATCH(DMI_PRODUCT_NAME, "Venue 11 Pro 7130"), + }), +- ENTRY("SYNA7500", "1", X86_MATCH(HASWELL_L), { ++ PRESENT_ENTRY_HID("SYNA7500", "1", HASWELL_L, { + DMI_MATCH(DMI_SYS_VENDOR, "Dell Inc."), + DMI_MATCH(DMI_PRODUCT_NAME, "Venue 11 Pro 7139"), + }), +@@ -81,54 +94,83 @@ static const struct always_present_id always_present_ids[] = { + /* + * The GPD win BIOS dated 20170221 has disabled the accelerometer, the + * drivers sometimes cause crashes under Windows and this is how the +- * manufacturer has solved this :| Note that the the DMI data is less +- * generic then it seems, a board_vendor of "AMI Corporation" is quite +- * rare and a board_name of "Default String" also is rare. ++ * manufacturer has solved this :| The DMI match may not seem unique, ++ * but it is. In the 67000+ DMI decode dumps from linux-hardware.org ++ * only 116 have board_vendor set to "AMI Corporation" and of those 116 ++ * only the GPD win and pocket entries' board_name is "Default string". + * + * Unfortunately the GPD pocket also uses these strings and its BIOS + * was copy-pasted from the GPD win, so it has a disabled KIOX000A + * node which we should not enable, thus we also check the BIOS date. + */ +- ENTRY("KIOX000A", "1", X86_MATCH(ATOM_AIRMONT), { ++ PRESENT_ENTRY_HID("KIOX000A", "1", ATOM_AIRMONT, { + DMI_MATCH(DMI_BOARD_VENDOR, "AMI Corporation"), + DMI_MATCH(DMI_BOARD_NAME, "Default string"), + DMI_MATCH(DMI_PRODUCT_NAME, "Default string"), + DMI_MATCH(DMI_BIOS_DATE, "02/21/2017") + }), +- ENTRY("KIOX000A", "1", X86_MATCH(ATOM_AIRMONT), { ++ PRESENT_ENTRY_HID("KIOX000A", "1", ATOM_AIRMONT, { + DMI_MATCH(DMI_BOARD_VENDOR, "AMI Corporation"), + DMI_MATCH(DMI_BOARD_NAME, "Default string"), + DMI_MATCH(DMI_PRODUCT_NAME, "Default string"), + DMI_MATCH(DMI_BIOS_DATE, "03/20/2017") + }), +- ENTRY("KIOX000A", "1", X86_MATCH(ATOM_AIRMONT), { ++ PRESENT_ENTRY_HID("KIOX000A", "1", ATOM_AIRMONT, { + DMI_MATCH(DMI_BOARD_VENDOR, "AMI Corporation"), + DMI_MATCH(DMI_BOARD_NAME, "Default string"), + DMI_MATCH(DMI_PRODUCT_NAME, "Default string"), + DMI_MATCH(DMI_BIOS_DATE, "05/25/2017") + }), ++ ++ /* ++ * The GPD win/pocket have a PCI wifi card, but its DSDT has the SDIO ++ * mmc controller enabled and that has a child-device which _PS3 ++ * method sets a GPIO causing the PCI wifi card to turn off. ++ * See above remark about uniqueness of the DMI match. ++ */ ++ NOT_PRESENT_ENTRY_PATH("\\_SB_.PCI0.SDHB.BRC1", ATOM_AIRMONT, { ++ DMI_EXACT_MATCH(DMI_BOARD_VENDOR, "AMI Corporation"), ++ DMI_EXACT_MATCH(DMI_BOARD_NAME, "Default string"), ++ DMI_EXACT_MATCH(DMI_BOARD_SERIAL, "Default string"), ++ DMI_EXACT_MATCH(DMI_PRODUCT_NAME, "Default string"), ++ }), + }; + +-bool acpi_device_always_present(struct acpi_device *adev) ++bool acpi_device_override_status(struct acpi_device *adev, unsigned long long *status) + { + bool ret = false; + unsigned int i; + +- for (i = 0; i < ARRAY_SIZE(always_present_ids); i++) { +- if (acpi_match_device_ids(adev, always_present_ids[i].hid)) ++ for (i = 0; i < ARRAY_SIZE(override_status_ids); i++) { ++ if (!x86_match_cpu(override_status_ids[i].cpu_ids)) + continue; + +- if (!adev->pnp.unique_id || +- strcmp(adev->pnp.unique_id, always_present_ids[i].uid)) ++ if (override_status_ids[i].dmi_ids[0].matches[0].slot && ++ !dmi_check_system(override_status_ids[i].dmi_ids)) + continue; + +- if (!x86_match_cpu(always_present_ids[i].cpu_ids)) +- continue; ++ if (override_status_ids[i].path) { ++ struct acpi_buffer path = { ACPI_ALLOCATE_BUFFER, NULL }; ++ bool match; + +- if (always_present_ids[i].dmi_ids[0].matches[0].slot && +- !dmi_check_system(always_present_ids[i].dmi_ids)) +- continue; ++ if (acpi_get_name(adev->handle, ACPI_FULL_PATHNAME, &path)) ++ continue; ++ ++ match = strcmp((char *)path.pointer, override_status_ids[i].path) == 0; ++ kfree(path.pointer); ++ ++ if (!match) ++ continue; ++ } else { ++ if (acpi_match_device_ids(adev, override_status_ids[i].hid)) ++ continue; ++ ++ if (!adev->pnp.unique_id || ++ strcmp(adev->pnp.unique_id, override_status_ids[i].uid)) ++ continue; ++ } + ++ *status = override_status_ids[i].status; + ret = true; + break; + } +diff --git a/drivers/android/binder.c b/drivers/android/binder.c +index c75fb600740cc..99ae919255f4d 100644 +--- a/drivers/android/binder.c ++++ b/drivers/android/binder.c +@@ -1608,15 +1608,21 @@ static void binder_cleanup_transaction(struct binder_transaction *t, + /** + * binder_get_object() - gets object and checks for valid metadata + * @proc: binder_proc owning the buffer ++ * @u: sender's user pointer to base of buffer + * @buffer: binder_buffer that we're parsing. + * @offset: offset in the @buffer at which to validate an object. + * @object: struct binder_object to read into + * +- * Return: If there's a valid metadata object at @offset in @buffer, the ++ * Copy the binder object at the given offset into @object. If @u is ++ * provided then the copy is from the sender's buffer. If not, then ++ * it is copied from the target's @buffer. ++ * ++ * Return: If there's a valid metadata object at @offset, the + * size of that object. Otherwise, it returns zero. The object + * is read into the struct binder_object pointed to by @object. + */ + static size_t binder_get_object(struct binder_proc *proc, ++ const void __user *u, + struct binder_buffer *buffer, + unsigned long offset, + struct binder_object *object) +@@ -1626,10 +1632,16 @@ static size_t binder_get_object(struct binder_proc *proc, + size_t object_size = 0; + + read_size = min_t(size_t, sizeof(*object), buffer->data_size - offset); +- if (offset > buffer->data_size || read_size < sizeof(*hdr) || +- binder_alloc_copy_from_buffer(&proc->alloc, object, buffer, +- offset, read_size)) ++ if (offset > buffer->data_size || read_size < sizeof(*hdr)) + return 0; ++ if (u) { ++ if (copy_from_user(object, u + offset, read_size)) ++ return 0; ++ } else { ++ if (binder_alloc_copy_from_buffer(&proc->alloc, object, buffer, ++ offset, read_size)) ++ return 0; ++ } + + /* Ok, now see if we read a complete object. */ + hdr = &object->hdr; +@@ -1702,7 +1714,7 @@ static struct binder_buffer_object *binder_validate_ptr( + b, buffer_offset, + sizeof(object_offset))) + return NULL; +- object_size = binder_get_object(proc, b, object_offset, object); ++ object_size = binder_get_object(proc, NULL, b, object_offset, object); + if (!object_size || object->hdr.type != BINDER_TYPE_PTR) + return NULL; + if (object_offsetp) +@@ -1767,7 +1779,8 @@ static bool binder_validate_fixup(struct binder_proc *proc, + unsigned long buffer_offset; + struct binder_object last_object; + struct binder_buffer_object *last_bbo; +- size_t object_size = binder_get_object(proc, b, last_obj_offset, ++ size_t object_size = binder_get_object(proc, NULL, b, ++ last_obj_offset, + &last_object); + if (object_size != sizeof(*last_bbo)) + return false; +@@ -1882,7 +1895,7 @@ static void binder_transaction_buffer_release(struct binder_proc *proc, + if (!binder_alloc_copy_from_buffer(&proc->alloc, &object_offset, + buffer, buffer_offset, + sizeof(object_offset))) +- object_size = binder_get_object(proc, buffer, ++ object_size = binder_get_object(proc, NULL, buffer, + object_offset, &object); + if (object_size == 0) { + pr_err("transaction release %d bad object at offset %lld, size %zd\n", +@@ -2269,8 +2282,8 @@ static int binder_translate_fd_array(struct binder_fd_array_object *fda, + if (!ret) + ret = binder_translate_fd(fd, offset, t, thread, + in_reply_to); +- if (ret < 0) +- return ret; ++ if (ret) ++ return ret > 0 ? -EINVAL : ret; + } + return 0; + } +@@ -2455,6 +2468,7 @@ static void binder_transaction(struct binder_proc *proc, + binder_size_t off_start_offset, off_end_offset; + binder_size_t off_min; + binder_size_t sg_buf_offset, sg_buf_end_offset; ++ binder_size_t user_offset = 0; + struct binder_proc *target_proc = NULL; + struct binder_thread *target_thread = NULL; + struct binder_node *target_node = NULL; +@@ -2469,6 +2483,8 @@ static void binder_transaction(struct binder_proc *proc, + int t_debug_id = atomic_inc_return(&binder_last_id); + char *secctx = NULL; + u32 secctx_sz = 0; ++ const void __user *user_buffer = (const void __user *) ++ (uintptr_t)tr->data.ptr.buffer; + + e = binder_transaction_log_add(&binder_transaction_log); + e->debug_id = t_debug_id; +@@ -2780,19 +2796,6 @@ static void binder_transaction(struct binder_proc *proc, + t->buffer->clear_on_free = !!(t->flags & TF_CLEAR_BUF); + trace_binder_transaction_alloc_buf(t->buffer); + +- if (binder_alloc_copy_user_to_buffer( +- &target_proc->alloc, +- t->buffer, 0, +- (const void __user *) +- (uintptr_t)tr->data.ptr.buffer, +- tr->data_size)) { +- binder_user_error("%d:%d got transaction with invalid data ptr\n", +- proc->pid, thread->pid); +- return_error = BR_FAILED_REPLY; +- return_error_param = -EFAULT; +- return_error_line = __LINE__; +- goto err_copy_data_failed; +- } + if (binder_alloc_copy_user_to_buffer( + &target_proc->alloc, + t->buffer, +@@ -2837,6 +2840,7 @@ static void binder_transaction(struct binder_proc *proc, + size_t object_size; + struct binder_object object; + binder_size_t object_offset; ++ binder_size_t copy_size; + + if (binder_alloc_copy_from_buffer(&target_proc->alloc, + &object_offset, +@@ -2848,8 +2852,27 @@ static void binder_transaction(struct binder_proc *proc, + return_error_line = __LINE__; + goto err_bad_offset; + } +- object_size = binder_get_object(target_proc, t->buffer, +- object_offset, &object); ++ ++ /* ++ * Copy the source user buffer up to the next object ++ * that will be processed. ++ */ ++ copy_size = object_offset - user_offset; ++ if (copy_size && (user_offset > object_offset || ++ binder_alloc_copy_user_to_buffer( ++ &target_proc->alloc, ++ t->buffer, user_offset, ++ user_buffer + user_offset, ++ copy_size))) { ++ binder_user_error("%d:%d got transaction with invalid data ptr\n", ++ proc->pid, thread->pid); ++ return_error = BR_FAILED_REPLY; ++ return_error_param = -EFAULT; ++ return_error_line = __LINE__; ++ goto err_copy_data_failed; ++ } ++ object_size = binder_get_object(target_proc, user_buffer, ++ t->buffer, object_offset, &object); + if (object_size == 0 || object_offset < off_min) { + binder_user_error("%d:%d got transaction with invalid offset (%lld, min %lld max %lld) or object.\n", + proc->pid, thread->pid, +@@ -2861,6 +2884,11 @@ static void binder_transaction(struct binder_proc *proc, + return_error_line = __LINE__; + goto err_bad_offset; + } ++ /* ++ * Set offset to the next buffer fragment to be ++ * copied ++ */ ++ user_offset = object_offset + object_size; + + hdr = &object.hdr; + off_min = object_offset + object_size; +@@ -2956,9 +2984,14 @@ static void binder_transaction(struct binder_proc *proc, + } + ret = binder_translate_fd_array(fda, parent, t, thread, + in_reply_to); +- if (ret < 0) { ++ if (!ret) ++ ret = binder_alloc_copy_to_buffer(&target_proc->alloc, ++ t->buffer, ++ object_offset, ++ fda, sizeof(*fda)); ++ if (ret) { + return_error = BR_FAILED_REPLY; +- return_error_param = ret; ++ return_error_param = ret > 0 ? -EINVAL : ret; + return_error_line = __LINE__; + goto err_translate_failed; + } +@@ -3028,6 +3061,19 @@ static void binder_transaction(struct binder_proc *proc, + goto err_bad_object_type; + } + } ++ /* Done processing objects, copy the rest of the buffer */ ++ if (binder_alloc_copy_user_to_buffer( ++ &target_proc->alloc, ++ t->buffer, user_offset, ++ user_buffer + user_offset, ++ tr->data_size - user_offset)) { ++ binder_user_error("%d:%d got transaction with invalid data ptr\n", ++ proc->pid, thread->pid); ++ return_error = BR_FAILED_REPLY; ++ return_error_param = -EFAULT; ++ return_error_line = __LINE__; ++ goto err_copy_data_failed; ++ } + if (t->buffer->oneway_spam_suspect) + tcomplete->type = BINDER_WORK_TRANSACTION_ONEWAY_SPAM_SUSPECT; + else +diff --git a/drivers/base/core.c b/drivers/base/core.c +index 63577de268565..8e73a34e10055 100644 +--- a/drivers/base/core.c ++++ b/drivers/base/core.c +@@ -485,8 +485,7 @@ static void device_link_release_fn(struct work_struct *work) + /* Ensure that all references to the link object have been dropped. */ + device_link_synchronize_removal(); + +- while (refcount_dec_not_one(&link->rpm_active)) +- pm_runtime_put(link->supplier); ++ pm_runtime_release_supplier(link, true); + + put_device(link->consumer); + put_device(link->supplier); +diff --git a/drivers/base/power/runtime.c b/drivers/base/power/runtime.c +index ec94049442b99..44ae3909e64bb 100644 +--- a/drivers/base/power/runtime.c ++++ b/drivers/base/power/runtime.c +@@ -305,19 +305,40 @@ static int rpm_get_suppliers(struct device *dev) + return 0; + } + ++/** ++ * pm_runtime_release_supplier - Drop references to device link's supplier. ++ * @link: Target device link. ++ * @check_idle: Whether or not to check if the supplier device is idle. ++ * ++ * Drop all runtime PM references associated with @link to its supplier device ++ * and if @check_idle is set, check if that device is idle (and so it can be ++ * suspended). ++ */ ++void pm_runtime_release_supplier(struct device_link *link, bool check_idle) ++{ ++ struct device *supplier = link->supplier; ++ ++ /* ++ * The additional power.usage_count check is a safety net in case ++ * the rpm_active refcount becomes saturated, in which case ++ * refcount_dec_not_one() would return true forever, but it is not ++ * strictly necessary. ++ */ ++ while (refcount_dec_not_one(&link->rpm_active) && ++ atomic_read(&supplier->power.usage_count) > 0) ++ pm_runtime_put_noidle(supplier); ++ ++ if (check_idle) ++ pm_request_idle(supplier); ++} ++ + static void __rpm_put_suppliers(struct device *dev, bool try_to_suspend) + { + struct device_link *link; + + list_for_each_entry_rcu(link, &dev->links.suppliers, c_node, +- device_links_read_lock_held()) { +- +- while (refcount_dec_not_one(&link->rpm_active)) +- pm_runtime_put_noidle(link->supplier); +- +- if (try_to_suspend) +- pm_request_idle(link->supplier); +- } ++ device_links_read_lock_held()) ++ pm_runtime_release_supplier(link, try_to_suspend); + } + + static void rpm_put_suppliers(struct device *dev) +@@ -1770,9 +1791,7 @@ void pm_runtime_drop_link(struct device_link *link) + return; + + pm_runtime_drop_link_count(link->consumer); +- +- while (refcount_dec_not_one(&link->rpm_active)) +- pm_runtime_put(link->supplier); ++ pm_runtime_release_supplier(link, true); + } + + static bool pm_runtime_need_not_resume(struct device *dev) +diff --git a/drivers/base/property.c b/drivers/base/property.c +index 453918eb7390c..4c77837769c6e 100644 +--- a/drivers/base/property.c ++++ b/drivers/base/property.c +@@ -1269,8 +1269,10 @@ fwnode_graph_devcon_match(struct fwnode_handle *fwnode, const char *con_id, + + fwnode_graph_for_each_endpoint(fwnode, ep) { + node = fwnode_graph_get_remote_port_parent(ep); +- if (!fwnode_device_is_available(node)) ++ if (!fwnode_device_is_available(node)) { ++ fwnode_handle_put(node); + continue; ++ } + + ret = match(node, con_id, data); + fwnode_handle_put(node); +diff --git a/drivers/base/regmap/regmap.c b/drivers/base/regmap/regmap.c +index 21a0c2562ec06..f7811641ed5ae 100644 +--- a/drivers/base/regmap/regmap.c ++++ b/drivers/base/regmap/regmap.c +@@ -647,6 +647,7 @@ int regmap_attach_dev(struct device *dev, struct regmap *map, + if (ret) + return ret; + ++ regmap_debugfs_exit(map); + regmap_debugfs_init(map); + + /* Add a devres resource for dev_get_regmap() */ +diff --git a/drivers/base/swnode.c b/drivers/base/swnode.c +index c46f6a8e14d23..3ba1232ce8451 100644 +--- a/drivers/base/swnode.c ++++ b/drivers/base/swnode.c +@@ -535,7 +535,7 @@ software_node_get_reference_args(const struct fwnode_handle *fwnode, + return -ENOENT; + + if (nargs_prop) { +- error = property_entry_read_int_array(swnode->node->properties, ++ error = property_entry_read_int_array(ref->node->properties, + nargs_prop, sizeof(u32), + &nargs_prop_val, 1); + if (error) +diff --git a/drivers/block/floppy.c b/drivers/block/floppy.c +index fb2aafabfebc1..4a6a74177b3c9 100644 +--- a/drivers/block/floppy.c ++++ b/drivers/block/floppy.c +@@ -1014,7 +1014,7 @@ static DECLARE_DELAYED_WORK(fd_timer, fd_timer_workfn); + static void cancel_activity(void) + { + do_floppy = NULL; +- cancel_delayed_work_sync(&fd_timer); ++ cancel_delayed_work(&fd_timer); + cancel_work_sync(&floppy_work); + } + +@@ -3080,6 +3080,8 @@ static void raw_cmd_free(struct floppy_raw_cmd **ptr) + } + } + ++#define MAX_LEN (1UL << MAX_ORDER << PAGE_SHIFT) ++ + static int raw_cmd_copyin(int cmd, void __user *param, + struct floppy_raw_cmd **rcmd) + { +@@ -3107,7 +3109,7 @@ loop: + ptr->resultcode = 0; + + if (ptr->flags & (FD_RAW_READ | FD_RAW_WRITE)) { +- if (ptr->length <= 0) ++ if (ptr->length <= 0 || ptr->length >= MAX_LEN) + return -EINVAL; + ptr->kernel_data = (char *)fd_dma_mem_alloc(ptr->length); + fallback_on_nodma_alloc(&ptr->kernel_data, ptr->length); +diff --git a/drivers/bluetooth/btintel.c b/drivers/bluetooth/btintel.c +index 525be2e1fbb25..e73d4c719b0ad 100644 +--- a/drivers/bluetooth/btintel.c ++++ b/drivers/bluetooth/btintel.c +@@ -2330,10 +2330,14 @@ static int btintel_setup_combined(struct hci_dev *hdev) + case 0x12: /* ThP */ + case 0x13: /* HrP */ + case 0x14: /* CcP */ +- /* Some legacy bootloader devices from JfP supports both old +- * and TLV based HCI_Intel_Read_Version command. But we don't +- * want to use the TLV based setup routines for those legacy +- * bootloader device. ++ /* Some legacy bootloader devices starting from JfP, ++ * the operational firmware supports both old and TLV based ++ * HCI_Intel_Read_Version command based on the command ++ * parameter. ++ * ++ * For upgrading firmware case, the TLV based version cannot ++ * be used because the firmware filename for legacy bootloader ++ * is based on the old format. + * + * Also, it is not easy to convert TLV based version from the + * legacy version format. +@@ -2345,6 +2349,20 @@ static int btintel_setup_combined(struct hci_dev *hdev) + err = btintel_read_version(hdev, &ver); + if (err) + return err; ++ ++ /* Apply the device specific HCI quirks ++ * ++ * All Legacy bootloader devices support WBS ++ */ ++ set_bit(HCI_QUIRK_WIDEBAND_SPEECH_SUPPORTED, &hdev->quirks); ++ ++ /* Valid LE States quirk for JfP/ThP familiy */ ++ if (ver.hw_variant == 0x11 || ver.hw_variant == 0x12) ++ set_bit(HCI_QUIRK_VALID_LE_STATES, &hdev->quirks); ++ ++ /* Setup MSFT Extension support */ ++ btintel_set_msft_opcode(hdev, ver.hw_variant); ++ + err = btintel_bootloader_setup(hdev, &ver); + break; + case 0x17: +diff --git a/drivers/bluetooth/btmtksdio.c b/drivers/bluetooth/btmtksdio.c +index 9872ef18f9fea..1cbdeca1fdc4a 100644 +--- a/drivers/bluetooth/btmtksdio.c ++++ b/drivers/bluetooth/btmtksdio.c +@@ -1042,6 +1042,8 @@ static int btmtksdio_runtime_suspend(struct device *dev) + if (!bdev) + return 0; + ++ sdio_set_host_pm_flags(func, MMC_PM_KEEP_POWER); ++ + sdio_claim_host(bdev->func); + + sdio_writel(bdev->func, C_FW_OWN_REQ_SET, MTK_REG_CHLPCR, &err); +diff --git a/drivers/bluetooth/btusb.c b/drivers/bluetooth/btusb.c +index fa44828562d32..ac90392cce339 100644 +--- a/drivers/bluetooth/btusb.c ++++ b/drivers/bluetooth/btusb.c +@@ -2561,6 +2561,7 @@ static int btusb_mtk_setup_firmware_79xx(struct hci_dev *hdev, const char *fwnam + } else { + bt_dev_err(hdev, "Failed wmt patch dwnld status (%d)", + status); ++ err = -EIO; + goto err_release_fw; + } + } +@@ -2856,6 +2857,10 @@ static int btusb_mtk_setup(struct hci_dev *hdev) + "mediatek/BT_RAM_CODE_MT%04x_1_%x_hdr.bin", + dev_id & 0xffff, (fw_version & 0xff) + 1); + err = btusb_mtk_setup_firmware_79xx(hdev, fw_bin_name); ++ if (err < 0) { ++ bt_dev_err(hdev, "Failed to set up firmware (%d)", err); ++ return err; ++ } + + /* It's Device EndPoint Reset Option Register */ + btusb_mtk_uhw_reg_write(data, MTK_EP_RST_OPT, MTK_EP_RST_IN_OUT_OPT); +diff --git a/drivers/bluetooth/hci_bcm.c b/drivers/bluetooth/hci_bcm.c +index ef54afa293574..7abf99f0ee399 100644 +--- a/drivers/bluetooth/hci_bcm.c ++++ b/drivers/bluetooth/hci_bcm.c +@@ -1188,7 +1188,12 @@ static int bcm_probe(struct platform_device *pdev) + return -ENOMEM; + + dev->dev = &pdev->dev; +- dev->irq = platform_get_irq(pdev, 0); ++ ++ ret = platform_get_irq(pdev, 0); ++ if (ret < 0) ++ return ret; ++ ++ dev->irq = ret; + + /* Initialize routing field to an unused value */ + dev->pcm_int_params[0] = 0xff; +diff --git a/drivers/bluetooth/hci_qca.c b/drivers/bluetooth/hci_qca.c +index 53deea2eb7b4d..8eb7fddfb9300 100644 +--- a/drivers/bluetooth/hci_qca.c ++++ b/drivers/bluetooth/hci_qca.c +@@ -1927,6 +1927,9 @@ static int qca_power_off(struct hci_dev *hdev) + hu->hdev->hw_error = NULL; + hu->hdev->cmd_timeout = NULL; + ++ del_timer_sync(&qca->wake_retrans_timer); ++ del_timer_sync(&qca->tx_idle_timer); ++ + /* Stop sending shutdown command if soc crashes. */ + if (soc_type != QCA_ROME + && qca->memdump_state == QCA_MEMDUMP_IDLE) { +@@ -2055,14 +2058,14 @@ static int qca_serdev_probe(struct serdev_device *serdev) + + qcadev->bt_en = devm_gpiod_get_optional(&serdev->dev, "enable", + GPIOD_OUT_LOW); +- if (!qcadev->bt_en && data->soc_type == QCA_WCN6750) { ++ if (IS_ERR_OR_NULL(qcadev->bt_en) && data->soc_type == QCA_WCN6750) { + dev_err(&serdev->dev, "failed to acquire BT_EN gpio\n"); + power_ctrl_enabled = false; + } + + qcadev->sw_ctrl = devm_gpiod_get_optional(&serdev->dev, "swctrl", + GPIOD_IN); +- if (!qcadev->sw_ctrl && data->soc_type == QCA_WCN6750) ++ if (IS_ERR_OR_NULL(qcadev->sw_ctrl) && data->soc_type == QCA_WCN6750) + dev_warn(&serdev->dev, "failed to acquire SW_CTRL gpio\n"); + + qcadev->susclk = devm_clk_get_optional(&serdev->dev, NULL); +@@ -2084,7 +2087,7 @@ static int qca_serdev_probe(struct serdev_device *serdev) + + qcadev->bt_en = devm_gpiod_get_optional(&serdev->dev, "enable", + GPIOD_OUT_LOW); +- if (!qcadev->bt_en) { ++ if (IS_ERR_OR_NULL(qcadev->bt_en)) { + dev_warn(&serdev->dev, "failed to acquire enable gpio\n"); + power_ctrl_enabled = false; + } +diff --git a/drivers/bluetooth/hci_vhci.c b/drivers/bluetooth/hci_vhci.c +index 8ab26dec5f6e8..8469f9876dd26 100644 +--- a/drivers/bluetooth/hci_vhci.c ++++ b/drivers/bluetooth/hci_vhci.c +@@ -121,6 +121,8 @@ static int __vhci_create_device(struct vhci_data *data, __u8 opcode) + if (opcode & 0x80) + set_bit(HCI_QUIRK_RAW_DEVICE, &hdev->quirks); + ++ set_bit(HCI_QUIRK_VALID_LE_STATES, &hdev->quirks); ++ + if (hci_register_dev(hdev) < 0) { + BT_ERR("Can't register HCI device"); + hci_free_dev(hdev); +diff --git a/drivers/bluetooth/virtio_bt.c b/drivers/bluetooth/virtio_bt.c +index 57908ce4fae85..076e4942a3f0e 100644 +--- a/drivers/bluetooth/virtio_bt.c ++++ b/drivers/bluetooth/virtio_bt.c +@@ -202,6 +202,9 @@ static void virtbt_rx_handle(struct virtio_bluetooth *vbt, struct sk_buff *skb) + hci_skb_pkt_type(skb) = pkt_type; + hci_recv_frame(vbt->hdev, skb); + break; ++ default: ++ kfree_skb(skb); ++ break; + } + } + +diff --git a/drivers/bus/mhi/core/init.c b/drivers/bus/mhi/core/init.c +index 5aaca6d0f52b2..f1ec344175928 100644 +--- a/drivers/bus/mhi/core/init.c ++++ b/drivers/bus/mhi/core/init.c +@@ -788,6 +788,7 @@ static int parse_ch_cfg(struct mhi_controller *mhi_cntrl, + mhi_chan->offload_ch = ch_cfg->offload_channel; + mhi_chan->db_cfg.reset_req = ch_cfg->doorbell_mode_switch; + mhi_chan->pre_alloc = ch_cfg->auto_queue; ++ mhi_chan->wake_capable = ch_cfg->wake_capable; + + /* + * If MHI host allocates buffers, then the channel direction +diff --git a/drivers/bus/mhi/core/pm.c b/drivers/bus/mhi/core/pm.c +index 547e6e769546a..bb9a2043f3a20 100644 +--- a/drivers/bus/mhi/core/pm.c ++++ b/drivers/bus/mhi/core/pm.c +@@ -1053,7 +1053,7 @@ int mhi_async_power_up(struct mhi_controller *mhi_cntrl) + enum mhi_ee_type current_ee; + enum dev_st_transition next_state; + struct device *dev = &mhi_cntrl->mhi_dev->dev; +- u32 val; ++ u32 interval_us = 25000; /* poll register field every 25 milliseconds */ + int ret; + + dev_info(dev, "Requested to power ON\n"); +@@ -1070,10 +1070,6 @@ int mhi_async_power_up(struct mhi_controller *mhi_cntrl) + mutex_lock(&mhi_cntrl->pm_mutex); + mhi_cntrl->pm_state = MHI_PM_DISABLE; + +- ret = mhi_init_irq_setup(mhi_cntrl); +- if (ret) +- goto error_setup_irq; +- + /* Setup BHI INTVEC */ + write_lock_irq(&mhi_cntrl->pm_lock); + mhi_write_reg(mhi_cntrl, mhi_cntrl->bhi, BHI_INTVEC, 0); +@@ -1087,7 +1083,7 @@ int mhi_async_power_up(struct mhi_controller *mhi_cntrl) + dev_err(dev, "%s is not a valid EE for power on\n", + TO_MHI_EXEC_STR(current_ee)); + ret = -EIO; +- goto error_async_power_up; ++ goto error_exit; + } + + state = mhi_get_mhi_state(mhi_cntrl); +@@ -1096,20 +1092,12 @@ int mhi_async_power_up(struct mhi_controller *mhi_cntrl) + + if (state == MHI_STATE_SYS_ERR) { + mhi_set_mhi_state(mhi_cntrl, MHI_STATE_RESET); +- ret = wait_event_timeout(mhi_cntrl->state_event, +- MHI_PM_IN_FATAL_STATE(mhi_cntrl->pm_state) || +- mhi_read_reg_field(mhi_cntrl, +- mhi_cntrl->regs, +- MHICTRL, +- MHICTRL_RESET_MASK, +- MHICTRL_RESET_SHIFT, +- &val) || +- !val, +- msecs_to_jiffies(mhi_cntrl->timeout_ms)); +- if (!ret) { +- ret = -EIO; ++ ret = mhi_poll_reg_field(mhi_cntrl, mhi_cntrl->regs, MHICTRL, ++ MHICTRL_RESET_MASK, MHICTRL_RESET_SHIFT, 0, ++ interval_us); ++ if (ret) { + dev_info(dev, "Failed to reset MHI due to syserr state\n"); +- goto error_async_power_up; ++ goto error_exit; + } + + /* +@@ -1119,6 +1107,10 @@ int mhi_async_power_up(struct mhi_controller *mhi_cntrl) + mhi_write_reg(mhi_cntrl, mhi_cntrl->bhi, BHI_INTVEC, 0); + } + ++ ret = mhi_init_irq_setup(mhi_cntrl); ++ if (ret) ++ goto error_exit; ++ + /* Transition to next state */ + next_state = MHI_IN_PBL(current_ee) ? + DEV_ST_TRANSITION_PBL : DEV_ST_TRANSITION_READY; +@@ -1131,10 +1123,7 @@ int mhi_async_power_up(struct mhi_controller *mhi_cntrl) + + return 0; + +-error_async_power_up: +- mhi_deinit_free_irq(mhi_cntrl); +- +-error_setup_irq: ++error_exit: + mhi_cntrl->pm_state = MHI_PM_DISABLE; + mutex_unlock(&mhi_cntrl->pm_mutex); + +diff --git a/drivers/bus/mhi/pci_generic.c b/drivers/bus/mhi/pci_generic.c +index 4c577a7317091..b8b2811c7c0a9 100644 +--- a/drivers/bus/mhi/pci_generic.c ++++ b/drivers/bus/mhi/pci_generic.c +@@ -1018,7 +1018,7 @@ static int __maybe_unused mhi_pci_freeze(struct device *dev) + * context. + */ + if (test_and_clear_bit(MHI_PCI_DEV_STARTED, &mhi_pdev->status)) { +- mhi_power_down(mhi_cntrl, false); ++ mhi_power_down(mhi_cntrl, true); + mhi_unprepare_after_power_down(mhi_cntrl); + } + +diff --git a/drivers/char/mwave/3780i.h b/drivers/char/mwave/3780i.h +index 9ccb6b270b071..95164246afd1a 100644 +--- a/drivers/char/mwave/3780i.h ++++ b/drivers/char/mwave/3780i.h +@@ -68,7 +68,7 @@ typedef struct { + unsigned char ClockControl:1; /* RW: Clock control: 0=normal, 1=stop 3780i clocks */ + unsigned char SoftReset:1; /* RW: Soft reset 0=normal, 1=soft reset active */ + unsigned char ConfigMode:1; /* RW: Configuration mode, 0=normal, 1=config mode */ +- unsigned char Reserved:5; /* 0: Reserved */ ++ unsigned short Reserved:13; /* 0: Reserved */ + } DSP_ISA_SLAVE_CONTROL; + + +diff --git a/drivers/char/random.c b/drivers/char/random.c +index 7470ee24db2f9..a27ae3999ff32 100644 +--- a/drivers/char/random.c ++++ b/drivers/char/random.c +@@ -912,12 +912,14 @@ static struct crng_state *select_crng(void) + + /* + * crng_fast_load() can be called by code in the interrupt service +- * path. So we can't afford to dilly-dally. ++ * path. So we can't afford to dilly-dally. Returns the number of ++ * bytes processed from cp. + */ +-static int crng_fast_load(const char *cp, size_t len) ++static size_t crng_fast_load(const char *cp, size_t len) + { + unsigned long flags; + char *p; ++ size_t ret = 0; + + if (!spin_trylock_irqsave(&primary_crng.lock, flags)) + return 0; +@@ -928,7 +930,7 @@ static int crng_fast_load(const char *cp, size_t len) + p = (unsigned char *) &primary_crng.state[4]; + while (len > 0 && crng_init_cnt < CRNG_INIT_CNT_THRESH) { + p[crng_init_cnt % CHACHA_KEY_SIZE] ^= *cp; +- cp++; crng_init_cnt++; len--; ++ cp++; crng_init_cnt++; len--; ret++; + } + spin_unlock_irqrestore(&primary_crng.lock, flags); + if (crng_init_cnt >= CRNG_INIT_CNT_THRESH) { +@@ -936,7 +938,7 @@ static int crng_fast_load(const char *cp, size_t len) + crng_init = 1; + pr_notice("fast init done\n"); + } +- return 1; ++ return ret; + } + + /* +@@ -1287,7 +1289,7 @@ void add_interrupt_randomness(int irq, int irq_flags) + if (unlikely(crng_init == 0)) { + if ((fast_pool->count >= 64) && + crng_fast_load((char *) fast_pool->pool, +- sizeof(fast_pool->pool))) { ++ sizeof(fast_pool->pool)) > 0) { + fast_pool->count = 0; + fast_pool->last = now; + } +@@ -2295,8 +2297,11 @@ void add_hwgenerator_randomness(const char *buffer, size_t count, + struct entropy_store *poolp = &input_pool; + + if (unlikely(crng_init == 0)) { +- crng_fast_load(buffer, count); +- return; ++ size_t ret = crng_fast_load(buffer, count); ++ count -= ret; ++ buffer += ret; ++ if (!count || crng_init == 0) ++ return; + } + + /* Suspend writing if we're above the trickle threshold. +diff --git a/drivers/char/tpm/tpm-chip.c b/drivers/char/tpm/tpm-chip.c +index ddaeceb7e1091..df37e7b6a10a5 100644 +--- a/drivers/char/tpm/tpm-chip.c ++++ b/drivers/char/tpm/tpm-chip.c +@@ -474,13 +474,21 @@ static void tpm_del_char_device(struct tpm_chip *chip) + + /* Make the driver uncallable. */ + down_write(&chip->ops_sem); +- if (chip->flags & TPM_CHIP_FLAG_TPM2) { +- if (!tpm_chip_start(chip)) { +- tpm2_shutdown(chip, TPM2_SU_CLEAR); +- tpm_chip_stop(chip); ++ ++ /* ++ * Check if chip->ops is still valid: In case that the controller ++ * drivers shutdown handler unregisters the controller in its ++ * shutdown handler we are called twice and chip->ops to NULL. ++ */ ++ if (chip->ops) { ++ if (chip->flags & TPM_CHIP_FLAG_TPM2) { ++ if (!tpm_chip_start(chip)) { ++ tpm2_shutdown(chip, TPM2_SU_CLEAR); ++ tpm_chip_stop(chip); ++ } + } ++ chip->ops = NULL; + } +- chip->ops = NULL; + up_write(&chip->ops_sem); + } + +diff --git a/drivers/char/tpm/tpm_tis_core.c b/drivers/char/tpm/tpm_tis_core.c +index b2659a4c40168..dc56b976d8162 100644 +--- a/drivers/char/tpm/tpm_tis_core.c ++++ b/drivers/char/tpm/tpm_tis_core.c +@@ -950,9 +950,11 @@ int tpm_tis_core_init(struct device *dev, struct tpm_tis_data *priv, int irq, + priv->timeout_max = TPM_TIMEOUT_USECS_MAX; + priv->phy_ops = phy_ops; + ++ dev_set_drvdata(&chip->dev, priv); ++ + rc = tpm_tis_read32(priv, TPM_DID_VID(0), &vendor); + if (rc < 0) +- goto out_err; ++ return rc; + + priv->manufacturer_id = vendor; + +@@ -962,8 +964,6 @@ int tpm_tis_core_init(struct device *dev, struct tpm_tis_data *priv, int irq, + priv->timeout_max = TIS_TIMEOUT_MAX_ATML; + } + +- dev_set_drvdata(&chip->dev, priv); +- + if (is_bsw()) { + priv->ilb_base_addr = ioremap(INTEL_LEGACY_BLK_BASE_ADDR, + ILB_REMAP_SIZE); +@@ -994,7 +994,15 @@ int tpm_tis_core_init(struct device *dev, struct tpm_tis_data *priv, int irq, + intmask |= TPM_INTF_CMD_READY_INT | TPM_INTF_LOCALITY_CHANGE_INT | + TPM_INTF_DATA_AVAIL_INT | TPM_INTF_STS_VALID_INT; + intmask &= ~TPM_GLOBAL_INT_ENABLE; ++ ++ rc = request_locality(chip, 0); ++ if (rc < 0) { ++ rc = -ENODEV; ++ goto out_err; ++ } ++ + tpm_tis_write32(priv, TPM_INT_ENABLE(priv->locality), intmask); ++ release_locality(chip, 0); + + rc = tpm_chip_start(chip); + if (rc) +diff --git a/drivers/clk/bcm/clk-bcm2835.c b/drivers/clk/bcm/clk-bcm2835.c +index a254512965eb8..3667b4d731e71 100644 +--- a/drivers/clk/bcm/clk-bcm2835.c ++++ b/drivers/clk/bcm/clk-bcm2835.c +@@ -932,8 +932,7 @@ static int bcm2835_clock_is_on(struct clk_hw *hw) + + static u32 bcm2835_clock_choose_div(struct clk_hw *hw, + unsigned long rate, +- unsigned long parent_rate, +- bool round_up) ++ unsigned long parent_rate) + { + struct bcm2835_clock *clock = bcm2835_clock_from_hw(hw); + const struct bcm2835_clock_data *data = clock->data; +@@ -945,10 +944,6 @@ static u32 bcm2835_clock_choose_div(struct clk_hw *hw, + + rem = do_div(temp, rate); + div = temp; +- +- /* Round up and mask off the unused bits */ +- if (round_up && ((div & unused_frac_mask) != 0 || rem != 0)) +- div += unused_frac_mask + 1; + div &= ~unused_frac_mask; + + /* different clamping limits apply for a mash clock */ +@@ -1079,7 +1074,7 @@ static int bcm2835_clock_set_rate(struct clk_hw *hw, + struct bcm2835_clock *clock = bcm2835_clock_from_hw(hw); + struct bcm2835_cprman *cprman = clock->cprman; + const struct bcm2835_clock_data *data = clock->data; +- u32 div = bcm2835_clock_choose_div(hw, rate, parent_rate, false); ++ u32 div = bcm2835_clock_choose_div(hw, rate, parent_rate); + u32 ctl; + + spin_lock(&cprman->regs_lock); +@@ -1130,7 +1125,7 @@ static unsigned long bcm2835_clock_choose_div_and_prate(struct clk_hw *hw, + + if (!(BIT(parent_idx) & data->set_rate_parent)) { + *prate = clk_hw_get_rate(parent); +- *div = bcm2835_clock_choose_div(hw, rate, *prate, true); ++ *div = bcm2835_clock_choose_div(hw, rate, *prate); + + *avgrate = bcm2835_clock_rate_from_divisor(clock, *prate, *div); + +@@ -1216,7 +1211,7 @@ static int bcm2835_clock_determine_rate(struct clk_hw *hw, + rate = bcm2835_clock_choose_div_and_prate(hw, i, req->rate, + &div, &prate, + &avgrate); +- if (rate > best_rate && rate <= req->rate) { ++ if (abs(req->rate - rate) < abs(req->rate - best_rate)) { + best_parent = parent; + best_prate = prate; + best_rate = rate; +diff --git a/drivers/clk/clk-bm1880.c b/drivers/clk/clk-bm1880.c +index e6d6599d310a1..fad78a22218e8 100644 +--- a/drivers/clk/clk-bm1880.c ++++ b/drivers/clk/clk-bm1880.c +@@ -522,14 +522,6 @@ static struct clk_hw *bm1880_clk_register_pll(struct bm1880_pll_hw_clock *pll_cl + return hw; + } + +-static void bm1880_clk_unregister_pll(struct clk_hw *hw) +-{ +- struct bm1880_pll_hw_clock *pll_hw = to_bm1880_pll_clk(hw); +- +- clk_hw_unregister(hw); +- kfree(pll_hw); +-} +- + static int bm1880_clk_register_plls(struct bm1880_pll_hw_clock *clks, + int num_clks, + struct bm1880_clock_data *data) +@@ -555,7 +547,7 @@ static int bm1880_clk_register_plls(struct bm1880_pll_hw_clock *clks, + + err_clk: + while (i--) +- bm1880_clk_unregister_pll(data->hw_data.hws[clks[i].pll.id]); ++ clk_hw_unregister(data->hw_data.hws[clks[i].pll.id]); + + return PTR_ERR(hw); + } +@@ -695,14 +687,6 @@ static struct clk_hw *bm1880_clk_register_div(struct bm1880_div_hw_clock *div_cl + return hw; + } + +-static void bm1880_clk_unregister_div(struct clk_hw *hw) +-{ +- struct bm1880_div_hw_clock *div_hw = to_bm1880_div_clk(hw); +- +- clk_hw_unregister(hw); +- kfree(div_hw); +-} +- + static int bm1880_clk_register_divs(struct bm1880_div_hw_clock *clks, + int num_clks, + struct bm1880_clock_data *data) +@@ -729,7 +713,7 @@ static int bm1880_clk_register_divs(struct bm1880_div_hw_clock *clks, + + err_clk: + while (i--) +- bm1880_clk_unregister_div(data->hw_data.hws[clks[i].div.id]); ++ clk_hw_unregister(data->hw_data.hws[clks[i].div.id]); + + return PTR_ERR(hw); + } +diff --git a/drivers/clk/clk-si5341.c b/drivers/clk/clk-si5341.c +index 57ae183982d8c..f7b41366666e5 100644 +--- a/drivers/clk/clk-si5341.c ++++ b/drivers/clk/clk-si5341.c +@@ -1740,7 +1740,7 @@ static int si5341_probe(struct i2c_client *client, + clk_prepare(data->clk[i].hw.clk); + } + +- err = of_clk_add_hw_provider(client->dev.of_node, of_clk_si5341_get, ++ err = devm_of_clk_add_hw_provider(&client->dev, of_clk_si5341_get, + data); + if (err) { + dev_err(&client->dev, "unable to add clk provider\n"); +diff --git a/drivers/clk/clk-stm32f4.c b/drivers/clk/clk-stm32f4.c +index af46176ad0539..473dfe632cc57 100644 +--- a/drivers/clk/clk-stm32f4.c ++++ b/drivers/clk/clk-stm32f4.c +@@ -129,7 +129,6 @@ static const struct stm32f4_gate_data stm32f429_gates[] __initconst = { + { STM32F4_RCC_APB2ENR, 20, "spi5", "apb2_div" }, + { STM32F4_RCC_APB2ENR, 21, "spi6", "apb2_div" }, + { STM32F4_RCC_APB2ENR, 22, "sai1", "apb2_div" }, +- { STM32F4_RCC_APB2ENR, 26, "ltdc", "apb2_div" }, + }; + + static const struct stm32f4_gate_data stm32f469_gates[] __initconst = { +@@ -211,7 +210,6 @@ static const struct stm32f4_gate_data stm32f469_gates[] __initconst = { + { STM32F4_RCC_APB2ENR, 20, "spi5", "apb2_div" }, + { STM32F4_RCC_APB2ENR, 21, "spi6", "apb2_div" }, + { STM32F4_RCC_APB2ENR, 22, "sai1", "apb2_div" }, +- { STM32F4_RCC_APB2ENR, 26, "ltdc", "apb2_div" }, + }; + + static const struct stm32f4_gate_data stm32f746_gates[] __initconst = { +@@ -286,7 +284,6 @@ static const struct stm32f4_gate_data stm32f746_gates[] __initconst = { + { STM32F4_RCC_APB2ENR, 21, "spi6", "apb2_div" }, + { STM32F4_RCC_APB2ENR, 22, "sai1", "apb2_div" }, + { STM32F4_RCC_APB2ENR, 23, "sai2", "apb2_div" }, +- { STM32F4_RCC_APB2ENR, 26, "ltdc", "apb2_div" }, + }; + + static const struct stm32f4_gate_data stm32f769_gates[] __initconst = { +@@ -364,7 +361,6 @@ static const struct stm32f4_gate_data stm32f769_gates[] __initconst = { + { STM32F4_RCC_APB2ENR, 21, "spi6", "apb2_div" }, + { STM32F4_RCC_APB2ENR, 22, "sai1", "apb2_div" }, + { STM32F4_RCC_APB2ENR, 23, "sai2", "apb2_div" }, +- { STM32F4_RCC_APB2ENR, 26, "ltdc", "apb2_div" }, + { STM32F4_RCC_APB2ENR, 30, "mdio", "apb2_div" }, + }; + +diff --git a/drivers/clk/clk.c b/drivers/clk/clk.c +index a277fd4f2f0a6..ac11cefc31911 100644 +--- a/drivers/clk/clk.c ++++ b/drivers/clk/clk.c +@@ -3340,6 +3340,24 @@ static int __init clk_debug_init(void) + { + struct clk_core *core; + ++#ifdef CLOCK_ALLOW_WRITE_DEBUGFS ++ pr_warn("\n"); ++ pr_warn("********************************************************************\n"); ++ pr_warn("** NOTICE NOTICE NOTICE NOTICE NOTICE NOTICE NOTICE **\n"); ++ pr_warn("** **\n"); ++ pr_warn("** WRITEABLE clk DebugFS SUPPORT HAS BEEN ENABLED IN THIS KERNEL **\n"); ++ pr_warn("** **\n"); ++ pr_warn("** This means that this kernel is built to expose clk operations **\n"); ++ pr_warn("** such as parent or rate setting, enabling, disabling, etc. **\n"); ++ pr_warn("** to userspace, which may compromise security on your system. **\n"); ++ pr_warn("** **\n"); ++ pr_warn("** If you see this message and you are not debugging the **\n"); ++ pr_warn("** kernel, report this immediately to your vendor! **\n"); ++ pr_warn("** **\n"); ++ pr_warn("** NOTICE NOTICE NOTICE NOTICE NOTICE NOTICE NOTICE **\n"); ++ pr_warn("********************************************************************\n"); ++#endif ++ + rootdir = debugfs_create_dir("clk", NULL); + + debugfs_create_file("clk_summary", 0444, rootdir, &all_lists, +diff --git a/drivers/clk/imx/clk-imx8mn.c b/drivers/clk/imx/clk-imx8mn.c +index c55577604e16a..021355a247081 100644 +--- a/drivers/clk/imx/clk-imx8mn.c ++++ b/drivers/clk/imx/clk-imx8mn.c +@@ -277,9 +277,9 @@ static const char * const imx8mn_pdm_sels[] = {"osc_24m", "sys_pll2_100m", "audi + + static const char * const imx8mn_dram_core_sels[] = {"dram_pll_out", "dram_alt_root", }; + +-static const char * const imx8mn_clko1_sels[] = {"osc_24m", "sys_pll1_800m", "osc_27m", +- "sys_pll1_200m", "audio_pll2_out", "vpu_pll", +- "sys_pll1_80m", }; ++static const char * const imx8mn_clko1_sels[] = {"osc_24m", "sys_pll1_800m", "dummy", ++ "sys_pll1_200m", "audio_pll2_out", "sys_pll2_500m", ++ "dummy", "sys_pll1_80m", }; + static const char * const imx8mn_clko2_sels[] = {"osc_24m", "sys_pll2_200m", "sys_pll1_400m", + "sys_pll2_166m", "sys_pll3_out", "audio_pll1_out", + "video_pll1_out", "osc_32k", }; +diff --git a/drivers/clk/meson/gxbb.c b/drivers/clk/meson/gxbb.c +index d6eed760327d0..608e0e8ca49a8 100644 +--- a/drivers/clk/meson/gxbb.c ++++ b/drivers/clk/meson/gxbb.c +@@ -713,6 +713,35 @@ static struct clk_regmap gxbb_mpll_prediv = { + }; + + static struct clk_regmap gxbb_mpll0_div = { ++ .data = &(struct meson_clk_mpll_data){ ++ .sdm = { ++ .reg_off = HHI_MPLL_CNTL7, ++ .shift = 0, ++ .width = 14, ++ }, ++ .sdm_en = { ++ .reg_off = HHI_MPLL_CNTL, ++ .shift = 25, ++ .width = 1, ++ }, ++ .n2 = { ++ .reg_off = HHI_MPLL_CNTL7, ++ .shift = 16, ++ .width = 9, ++ }, ++ .lock = &meson_clk_lock, ++ }, ++ .hw.init = &(struct clk_init_data){ ++ .name = "mpll0_div", ++ .ops = &meson_clk_mpll_ops, ++ .parent_hws = (const struct clk_hw *[]) { ++ &gxbb_mpll_prediv.hw ++ }, ++ .num_parents = 1, ++ }, ++}; ++ ++static struct clk_regmap gxl_mpll0_div = { + .data = &(struct meson_clk_mpll_data){ + .sdm = { + .reg_off = HHI_MPLL_CNTL7, +@@ -749,7 +778,16 @@ static struct clk_regmap gxbb_mpll0 = { + .hw.init = &(struct clk_init_data){ + .name = "mpll0", + .ops = &clk_regmap_gate_ops, +- .parent_hws = (const struct clk_hw *[]) { &gxbb_mpll0_div.hw }, ++ .parent_data = &(const struct clk_parent_data) { ++ /* ++ * Note: ++ * GXL and GXBB have different SDM_EN registers. We ++ * fallback to the global naming string mechanism so ++ * mpll0_div picks up the appropriate one. ++ */ ++ .name = "mpll0_div", ++ .index = -1, ++ }, + .num_parents = 1, + .flags = CLK_SET_RATE_PARENT, + }, +@@ -3044,7 +3082,7 @@ static struct clk_hw_onecell_data gxl_hw_onecell_data = { + [CLKID_VAPB_1] = &gxbb_vapb_1.hw, + [CLKID_VAPB_SEL] = &gxbb_vapb_sel.hw, + [CLKID_VAPB] = &gxbb_vapb.hw, +- [CLKID_MPLL0_DIV] = &gxbb_mpll0_div.hw, ++ [CLKID_MPLL0_DIV] = &gxl_mpll0_div.hw, + [CLKID_MPLL1_DIV] = &gxbb_mpll1_div.hw, + [CLKID_MPLL2_DIV] = &gxbb_mpll2_div.hw, + [CLKID_MPLL_PREDIV] = &gxbb_mpll_prediv.hw, +@@ -3439,7 +3477,7 @@ static struct clk_regmap *const gxl_clk_regmaps[] = { + &gxbb_mpll0, + &gxbb_mpll1, + &gxbb_mpll2, +- &gxbb_mpll0_div, ++ &gxl_mpll0_div, + &gxbb_mpll1_div, + &gxbb_mpll2_div, + &gxbb_cts_amclk_div, +diff --git a/drivers/clk/qcom/gcc-sc7280.c b/drivers/clk/qcom/gcc-sc7280.c +index 6cefcdc869905..ce7c5ba2b9b7a 100644 +--- a/drivers/clk/qcom/gcc-sc7280.c ++++ b/drivers/clk/qcom/gcc-sc7280.c +@@ -2998,7 +2998,7 @@ static struct clk_branch gcc_cfg_noc_lpass_clk = { + .enable_mask = BIT(0), + .hw.init = &(struct clk_init_data){ + .name = "gcc_cfg_noc_lpass_clk", +- .ops = &clk_branch2_ops, ++ .ops = &clk_branch2_aon_ops, + }, + }, + }; +diff --git a/drivers/clk/renesas/rzg2l-cpg.c b/drivers/clk/renesas/rzg2l-cpg.c +index 761922ea5db76..1c92e73cd2b8c 100644 +--- a/drivers/clk/renesas/rzg2l-cpg.c ++++ b/drivers/clk/renesas/rzg2l-cpg.c +@@ -638,10 +638,16 @@ static void rzg2l_cpg_detach_dev(struct generic_pm_domain *unused, struct device + pm_clk_destroy(dev); + } + ++static void rzg2l_cpg_genpd_remove(void *data) ++{ ++ pm_genpd_remove(data); ++} ++ + static int __init rzg2l_cpg_add_clk_domain(struct device *dev) + { + struct device_node *np = dev->of_node; + struct generic_pm_domain *genpd; ++ int ret; + + genpd = devm_kzalloc(dev, sizeof(*genpd), GFP_KERNEL); + if (!genpd) +@@ -652,10 +658,15 @@ static int __init rzg2l_cpg_add_clk_domain(struct device *dev) + GENPD_FLAG_ACTIVE_WAKEUP; + genpd->attach_dev = rzg2l_cpg_attach_dev; + genpd->detach_dev = rzg2l_cpg_detach_dev; +- pm_genpd_init(genpd, &pm_domain_always_on_gov, false); ++ ret = pm_genpd_init(genpd, &pm_domain_always_on_gov, false); ++ if (ret) ++ return ret; + +- of_genpd_add_provider_simple(np, genpd); +- return 0; ++ ret = devm_add_action_or_reset(dev, rzg2l_cpg_genpd_remove, genpd); ++ if (ret) ++ return ret; ++ ++ return of_genpd_add_provider_simple(np, genpd); + } + + static int __init rzg2l_cpg_probe(struct platform_device *pdev) +diff --git a/drivers/cpufreq/cpufreq.c b/drivers/cpufreq/cpufreq.c +index fcb44352623ee..eeac6d8092298 100644 +--- a/drivers/cpufreq/cpufreq.c ++++ b/drivers/cpufreq/cpufreq.c +@@ -1402,7 +1402,7 @@ static int cpufreq_online(unsigned int cpu) + + ret = freq_qos_add_request(&policy->constraints, + policy->min_freq_req, FREQ_QOS_MIN, +- policy->min); ++ FREQ_QOS_MIN_DEFAULT_VALUE); + if (ret < 0) { + /* + * So we don't call freq_qos_remove_request() for an +@@ -1422,7 +1422,7 @@ static int cpufreq_online(unsigned int cpu) + + ret = freq_qos_add_request(&policy->constraints, + policy->max_freq_req, FREQ_QOS_MAX, +- policy->max); ++ FREQ_QOS_MAX_DEFAULT_VALUE); + if (ret < 0) { + policy->max_freq_req = NULL; + goto out_destroy_policy; +diff --git a/drivers/cpufreq/qcom-cpufreq-hw.c b/drivers/cpufreq/qcom-cpufreq-hw.c +index a2be0df7e1747..35d93361fda1a 100644 +--- a/drivers/cpufreq/qcom-cpufreq-hw.c ++++ b/drivers/cpufreq/qcom-cpufreq-hw.c +@@ -304,7 +304,8 @@ static void qcom_lmh_dcvs_notify(struct qcom_cpufreq_data *data) + if (capacity > max_capacity) + capacity = max_capacity; + +- arch_set_thermal_pressure(policy->cpus, max_capacity - capacity); ++ arch_set_thermal_pressure(policy->related_cpus, ++ max_capacity - capacity); + + /* + * In the unlikely case policy is unregistered do not enable +@@ -342,9 +343,9 @@ static irqreturn_t qcom_lmh_dcvs_handle_irq(int irq, void *data) + + /* Disable interrupt and enable polling */ + disable_irq_nosync(c_data->throttle_irq); +- qcom_lmh_dcvs_notify(c_data); ++ schedule_delayed_work(&c_data->throttle_work, 0); + +- return 0; ++ return IRQ_HANDLED; + } + + static const struct qcom_cpufreq_soc_data qcom_soc_data = { +diff --git a/drivers/crypto/atmel-aes.c b/drivers/crypto/atmel-aes.c +index 9391ccc03382d..fe05584031914 100644 +--- a/drivers/crypto/atmel-aes.c ++++ b/drivers/crypto/atmel-aes.c +@@ -960,6 +960,7 @@ static int atmel_aes_handle_queue(struct atmel_aes_dev *dd, + ctx = crypto_tfm_ctx(areq->tfm); + + dd->areq = areq; ++ dd->ctx = ctx; + start_async = (areq != new_areq); + dd->is_async = start_async; + +@@ -1274,7 +1275,6 @@ static int atmel_aes_init_tfm(struct crypto_skcipher *tfm) + + crypto_skcipher_set_reqsize(tfm, sizeof(struct atmel_aes_reqctx)); + ctx->base.dd = dd; +- ctx->base.dd->ctx = &ctx->base; + ctx->base.start = atmel_aes_start; + + return 0; +@@ -1291,7 +1291,6 @@ static int atmel_aes_ctr_init_tfm(struct crypto_skcipher *tfm) + + crypto_skcipher_set_reqsize(tfm, sizeof(struct atmel_aes_reqctx)); + ctx->base.dd = dd; +- ctx->base.dd->ctx = &ctx->base; + ctx->base.start = atmel_aes_ctr_start; + + return 0; +@@ -1783,7 +1782,6 @@ static int atmel_aes_gcm_init(struct crypto_aead *tfm) + + crypto_aead_set_reqsize(tfm, sizeof(struct atmel_aes_reqctx)); + ctx->base.dd = dd; +- ctx->base.dd->ctx = &ctx->base; + ctx->base.start = atmel_aes_gcm_start; + + return 0; +@@ -1927,7 +1925,6 @@ static int atmel_aes_xts_init_tfm(struct crypto_skcipher *tfm) + crypto_skcipher_set_reqsize(tfm, sizeof(struct atmel_aes_reqctx) + + crypto_skcipher_reqsize(ctx->fallback_tfm)); + ctx->base.dd = dd; +- ctx->base.dd->ctx = &ctx->base; + ctx->base.start = atmel_aes_xts_start; + + return 0; +@@ -2154,7 +2151,6 @@ static int atmel_aes_authenc_init_tfm(struct crypto_aead *tfm, + crypto_aead_set_reqsize(tfm, (sizeof(struct atmel_aes_authenc_reqctx) + + auth_reqsize)); + ctx->base.dd = dd; +- ctx->base.dd->ctx = &ctx->base; + ctx->base.start = atmel_aes_authenc_start; + + return 0; +diff --git a/drivers/crypto/caam/caamalg.c b/drivers/crypto/caam/caamalg.c +index 8697ae53b0633..d3d8bb0a69900 100644 +--- a/drivers/crypto/caam/caamalg.c ++++ b/drivers/crypto/caam/caamalg.c +@@ -1533,6 +1533,9 @@ static int aead_do_one_req(struct crypto_engine *engine, void *areq) + + ret = caam_jr_enqueue(ctx->jrdev, desc, aead_crypt_done, req); + ++ if (ret == -ENOSPC && engine->retry_support) ++ return ret; ++ + if (ret != -EINPROGRESS) { + aead_unmap(ctx->jrdev, rctx->edesc, req); + kfree(rctx->edesc); +@@ -1762,6 +1765,9 @@ static int skcipher_do_one_req(struct crypto_engine *engine, void *areq) + + ret = caam_jr_enqueue(ctx->jrdev, desc, skcipher_crypt_done, req); + ++ if (ret == -ENOSPC && engine->retry_support) ++ return ret; ++ + if (ret != -EINPROGRESS) { + skcipher_unmap(ctx->jrdev, rctx->edesc, req); + kfree(rctx->edesc); +diff --git a/drivers/crypto/caam/caamalg_qi2.c b/drivers/crypto/caam/caamalg_qi2.c +index 8b8ed77d8715d..6753f0e6e55d1 100644 +--- a/drivers/crypto/caam/caamalg_qi2.c ++++ b/drivers/crypto/caam/caamalg_qi2.c +@@ -5470,7 +5470,7 @@ int dpaa2_caam_enqueue(struct device *dev, struct caam_request *req) + dpaa2_fd_set_len(&fd, dpaa2_fl_get_len(&req->fd_flt[1])); + dpaa2_fd_set_flc(&fd, req->flc_dma); + +- ppriv = this_cpu_ptr(priv->ppriv); ++ ppriv = raw_cpu_ptr(priv->ppriv); + for (i = 0; i < (priv->dpseci_attr.num_tx_queues << 1); i++) { + err = dpaa2_io_service_enqueue_fq(ppriv->dpio, ppriv->req_fqid, + &fd); +diff --git a/drivers/crypto/caam/caamhash.c b/drivers/crypto/caam/caamhash.c +index e8a6d8bc43b5d..36ef738e4a181 100644 +--- a/drivers/crypto/caam/caamhash.c ++++ b/drivers/crypto/caam/caamhash.c +@@ -765,6 +765,9 @@ static int ahash_do_one_req(struct crypto_engine *engine, void *areq) + + ret = caam_jr_enqueue(jrdev, desc, state->ahash_op_done, req); + ++ if (ret == -ENOSPC && engine->retry_support) ++ return ret; ++ + if (ret != -EINPROGRESS) { + ahash_unmap(jrdev, state->edesc, req, 0); + kfree(state->edesc); +diff --git a/drivers/crypto/caam/caampkc.c b/drivers/crypto/caam/caampkc.c +index bf6275ffc4aad..8867275767101 100644 +--- a/drivers/crypto/caam/caampkc.c ++++ b/drivers/crypto/caam/caampkc.c +@@ -380,6 +380,9 @@ static int akcipher_do_one_req(struct crypto_engine *engine, void *areq) + + ret = caam_jr_enqueue(jrdev, desc, req_ctx->akcipher_op_done, req); + ++ if (ret == -ENOSPC && engine->retry_support) ++ return ret; ++ + if (ret != -EINPROGRESS) { + rsa_pub_unmap(jrdev, req_ctx->edesc, req); + rsa_io_unmap(jrdev, req_ctx->edesc, req); +diff --git a/drivers/crypto/ccp/sev-dev.c b/drivers/crypto/ccp/sev-dev.c +index 2ecb0e1f65d8d..e2806ca3300a8 100644 +--- a/drivers/crypto/ccp/sev-dev.c ++++ b/drivers/crypto/ccp/sev-dev.c +@@ -241,7 +241,7 @@ static int __sev_platform_init_locked(int *error) + struct psp_device *psp = psp_master; + struct sev_data_init data; + struct sev_device *sev; +- int rc = 0; ++ int psp_ret, rc = 0; + + if (!psp || !psp->sev_data) + return -ENODEV; +@@ -266,7 +266,21 @@ static int __sev_platform_init_locked(int *error) + data.tmr_len = SEV_ES_TMR_SIZE; + } + +- rc = __sev_do_cmd_locked(SEV_CMD_INIT, &data, error); ++ rc = __sev_do_cmd_locked(SEV_CMD_INIT, &data, &psp_ret); ++ if (rc && psp_ret == SEV_RET_SECURE_DATA_INVALID) { ++ /* ++ * Initialization command returned an integrity check failure ++ * status code, meaning that firmware load and validation of SEV ++ * related persistent data has failed. Retrying the ++ * initialization function should succeed by replacing the state ++ * with a reset state. ++ */ ++ dev_dbg(sev->dev, "SEV: retrying INIT command"); ++ rc = __sev_do_cmd_locked(SEV_CMD_INIT, &data, &psp_ret); ++ } ++ if (error) ++ *error = psp_ret; ++ + if (rc) + return rc; + +@@ -1091,18 +1105,6 @@ void sev_pci_init(void) + + /* Initialize the platform */ + rc = sev_platform_init(&error); +- if (rc && (error == SEV_RET_SECURE_DATA_INVALID)) { +- /* +- * INIT command returned an integrity check failure +- * status code, meaning that firmware load and +- * validation of SEV related persistent data has +- * failed and persistent state has been erased. +- * Retrying INIT command here should succeed. +- */ +- dev_dbg(sev->dev, "SEV: retrying INIT command"); +- rc = sev_platform_init(&error); +- } +- + if (rc) { + dev_err(sev->dev, "SEV: failed to INIT error %#x\n", error); + return; +diff --git a/drivers/crypto/hisilicon/hpre/hpre_crypto.c b/drivers/crypto/hisilicon/hpre/hpre_crypto.c +index a032c192ef1d6..7ba7641723a0b 100644 +--- a/drivers/crypto/hisilicon/hpre/hpre_crypto.c ++++ b/drivers/crypto/hisilicon/hpre/hpre_crypto.c +@@ -1865,7 +1865,7 @@ static int hpre_curve25519_src_init(struct hpre_asym_request *hpre_req, + */ + if (memcmp(ptr, p, ctx->key_sz) == 0) { + dev_err(dev, "gx is p!\n"); +- return -EINVAL; ++ goto err; + } else if (memcmp(ptr, p, ctx->key_sz) > 0) { + hpre_curve25519_src_modulo_p(ptr); + } +diff --git a/drivers/crypto/hisilicon/qm.c b/drivers/crypto/hisilicon/qm.c +index 369562d34d66a..ff1122153fbec 100644 +--- a/drivers/crypto/hisilicon/qm.c ++++ b/drivers/crypto/hisilicon/qm.c +@@ -5986,7 +5986,7 @@ int hisi_qm_resume(struct device *dev) + if (ret) + pci_err(pdev, "failed to start qm(%d)\n", ret); + +- return 0; ++ return ret; + } + EXPORT_SYMBOL_GPL(hisi_qm_resume); + +diff --git a/drivers/crypto/marvell/octeontx2/otx2_cptpf_main.c b/drivers/crypto/marvell/octeontx2/otx2_cptpf_main.c +index 146a55ac4b9b0..be1ad55a208f6 100644 +--- a/drivers/crypto/marvell/octeontx2/otx2_cptpf_main.c ++++ b/drivers/crypto/marvell/octeontx2/otx2_cptpf_main.c +@@ -494,12 +494,11 @@ static ssize_t kvf_limits_store(struct device *dev, + { + struct otx2_cptpf_dev *cptpf = dev_get_drvdata(dev); + int lfs_num; ++ int ret; + +- if (kstrtoint(buf, 0, &lfs_num)) { +- dev_err(dev, "lfs count %d must be in range [1 - %d]\n", +- lfs_num, num_online_cpus()); +- return -EINVAL; +- } ++ ret = kstrtoint(buf, 0, &lfs_num); ++ if (ret) ++ return ret; + if (lfs_num < 1 || lfs_num > num_online_cpus()) { + dev_err(dev, "lfs count %d must be in range [1 - %d]\n", + lfs_num, num_online_cpus()); +diff --git a/drivers/crypto/marvell/octeontx2/otx2_cptpf_ucode.c b/drivers/crypto/marvell/octeontx2/otx2_cptpf_ucode.c +index dff34b3ec09e1..7c1b92aaab398 100644 +--- a/drivers/crypto/marvell/octeontx2/otx2_cptpf_ucode.c ++++ b/drivers/crypto/marvell/octeontx2/otx2_cptpf_ucode.c +@@ -29,7 +29,8 @@ static struct otx2_cpt_bitmap get_cores_bmap(struct device *dev, + bool found = false; + int i; + +- if (eng_grp->g->engs_num > OTX2_CPT_MAX_ENGINES) { ++ if (eng_grp->g->engs_num < 0 || ++ eng_grp->g->engs_num > OTX2_CPT_MAX_ENGINES) { + dev_err(dev, "unsupported number of engines %d on octeontx2\n", + eng_grp->g->engs_num); + return bmap; +diff --git a/drivers/crypto/omap-aes.c b/drivers/crypto/omap-aes.c +index 9b968ac4ee7b6..a196bb8b17010 100644 +--- a/drivers/crypto/omap-aes.c ++++ b/drivers/crypto/omap-aes.c +@@ -1302,7 +1302,7 @@ static int omap_aes_suspend(struct device *dev) + + static int omap_aes_resume(struct device *dev) + { +- pm_runtime_resume_and_get(dev); ++ pm_runtime_get_sync(dev); + return 0; + } + #endif +diff --git a/drivers/crypto/qat/qat_common/adf_pf2vf_msg.c b/drivers/crypto/qat/qat_common/adf_pf2vf_msg.c +index 5a41beb8f20f6..7ec81989beb03 100644 +--- a/drivers/crypto/qat/qat_common/adf_pf2vf_msg.c ++++ b/drivers/crypto/qat/qat_common/adf_pf2vf_msg.c +@@ -117,37 +117,19 @@ static int __adf_iov_putmsg(struct adf_accel_dev *accel_dev, u32 msg, u8 vf_nr) + + mutex_lock(lock); + +- /* Check if PF2VF CSR is in use by remote function */ ++ /* Check if the PFVF CSR is in use by remote function */ + val = ADF_CSR_RD(pmisc_bar_addr, pf2vf_offset); + if ((val & remote_in_use_mask) == remote_in_use_pattern) { + dev_dbg(&GET_DEV(accel_dev), +- "PF2VF CSR in use by remote function\n"); ++ "PFVF CSR in use by remote function\n"); + ret = -EBUSY; + goto out; + } + +- /* Attempt to get ownership of PF2VF CSR */ + msg &= ~local_in_use_mask; + msg |= local_in_use_pattern; +- ADF_CSR_WR(pmisc_bar_addr, pf2vf_offset, msg); + +- /* Wait in case remote func also attempting to get ownership */ +- msleep(ADF_IOV_MSG_COLLISION_DETECT_DELAY); +- +- val = ADF_CSR_RD(pmisc_bar_addr, pf2vf_offset); +- if ((val & local_in_use_mask) != local_in_use_pattern) { +- dev_dbg(&GET_DEV(accel_dev), +- "PF2VF CSR in use by remote - collision detected\n"); +- ret = -EBUSY; +- goto out; +- } +- +- /* +- * This function now owns the PV2VF CSR. The IN_USE_BY pattern must +- * remain in the PF2VF CSR for all writes including ACK from remote +- * until this local function relinquishes the CSR. Send the message +- * by interrupting the remote. +- */ ++ /* Attempt to get ownership of the PFVF CSR */ + ADF_CSR_WR(pmisc_bar_addr, pf2vf_offset, msg | int_bit); + + /* Wait for confirmation from remote func it received the message */ +@@ -156,6 +138,12 @@ static int __adf_iov_putmsg(struct adf_accel_dev *accel_dev, u32 msg, u8 vf_nr) + val = ADF_CSR_RD(pmisc_bar_addr, pf2vf_offset); + } while ((val & int_bit) && (count++ < ADF_IOV_MSG_ACK_MAX_RETRY)); + ++ if (val & int_bit) { ++ dev_dbg(&GET_DEV(accel_dev), "ACK not received from remote\n"); ++ val &= ~int_bit; ++ ret = -EIO; ++ } ++ + if (val != msg) { + dev_dbg(&GET_DEV(accel_dev), + "Collision - PFVF CSR overwritten by remote function\n"); +@@ -163,13 +151,7 @@ static int __adf_iov_putmsg(struct adf_accel_dev *accel_dev, u32 msg, u8 vf_nr) + goto out; + } + +- if (val & int_bit) { +- dev_dbg(&GET_DEV(accel_dev), "ACK not received from remote\n"); +- val &= ~int_bit; +- ret = -EIO; +- } +- +- /* Finished with PF2VF CSR; relinquish it and leave msg in CSR */ ++ /* Finished with the PFVF CSR; relinquish it and leave msg in CSR */ + ADF_CSR_WR(pmisc_bar_addr, pf2vf_offset, val & ~local_in_use_mask); + out: + mutex_unlock(lock); +@@ -177,12 +159,13 @@ out: + } + + /** +- * adf_iov_putmsg() - send PF2VF message ++ * adf_iov_putmsg() - send PFVF message + * @accel_dev: Pointer to acceleration device. + * @msg: Message to send +- * @vf_nr: VF number to which the message will be sent ++ * @vf_nr: VF number to which the message will be sent if on PF, ignored ++ * otherwise + * +- * Function sends a message from the PF to a VF ++ * Function sends a message through the PFVF channel + * + * Return: 0 on success, error code otherwise. + */ +diff --git a/drivers/crypto/qce/aead.c b/drivers/crypto/qce/aead.c +index 290e2446a2f35..97a530171f07a 100644 +--- a/drivers/crypto/qce/aead.c ++++ b/drivers/crypto/qce/aead.c +@@ -802,8 +802,8 @@ static int qce_aead_register_one(const struct qce_aead_def *def, struct qce_devi + + ret = crypto_register_aead(alg); + if (ret) { +- kfree(tmpl); + dev_err(qce->dev, "%s registration failed\n", alg->base.cra_name); ++ kfree(tmpl); + return ret; + } + +diff --git a/drivers/crypto/qce/sha.c b/drivers/crypto/qce/sha.c +index 8e6fcf2c21cc0..59159f5e64e52 100644 +--- a/drivers/crypto/qce/sha.c ++++ b/drivers/crypto/qce/sha.c +@@ -498,8 +498,8 @@ static int qce_ahash_register_one(const struct qce_ahash_def *def, + + ret = crypto_register_ahash(alg); + if (ret) { +- kfree(tmpl); + dev_err(qce->dev, "%s registration failed\n", base->cra_name); ++ kfree(tmpl); + return ret; + } + +diff --git a/drivers/crypto/qce/skcipher.c b/drivers/crypto/qce/skcipher.c +index 8ff10928f581d..3d27cd5210ef5 100644 +--- a/drivers/crypto/qce/skcipher.c ++++ b/drivers/crypto/qce/skcipher.c +@@ -484,8 +484,8 @@ static int qce_skcipher_register_one(const struct qce_skcipher_def *def, + + ret = crypto_register_skcipher(alg); + if (ret) { +- kfree(tmpl); + dev_err(qce->dev, "%s registration failed\n", alg->base.cra_name); ++ kfree(tmpl); + return ret; + } + +diff --git a/drivers/crypto/stm32/stm32-crc32.c b/drivers/crypto/stm32/stm32-crc32.c +index 75867c0b00172..be1bf39a317de 100644 +--- a/drivers/crypto/stm32/stm32-crc32.c ++++ b/drivers/crypto/stm32/stm32-crc32.c +@@ -279,7 +279,7 @@ static struct shash_alg algs[] = { + .digestsize = CHKSUM_DIGEST_SIZE, + .base = { + .cra_name = "crc32", +- .cra_driver_name = DRIVER_NAME, ++ .cra_driver_name = "stm32-crc32-crc32", + .cra_priority = 200, + .cra_flags = CRYPTO_ALG_OPTIONAL_KEY, + .cra_blocksize = CHKSUM_BLOCK_SIZE, +@@ -301,7 +301,7 @@ static struct shash_alg algs[] = { + .digestsize = CHKSUM_DIGEST_SIZE, + .base = { + .cra_name = "crc32c", +- .cra_driver_name = DRIVER_NAME, ++ .cra_driver_name = "stm32-crc32-crc32c", + .cra_priority = 200, + .cra_flags = CRYPTO_ALG_OPTIONAL_KEY, + .cra_blocksize = CHKSUM_BLOCK_SIZE, +diff --git a/drivers/crypto/stm32/stm32-cryp.c b/drivers/crypto/stm32/stm32-cryp.c +index 7389a0536ff02..81eb136b6c11d 100644 +--- a/drivers/crypto/stm32/stm32-cryp.c ++++ b/drivers/crypto/stm32/stm32-cryp.c +@@ -37,7 +37,6 @@ + /* Mode mask = bits [15..0] */ + #define FLG_MODE_MASK GENMASK(15, 0) + /* Bit [31..16] status */ +-#define FLG_CCM_PADDED_WA BIT(16) + + /* Registers */ + #define CRYP_CR 0x00000000 +@@ -105,8 +104,6 @@ + /* Misc */ + #define AES_BLOCK_32 (AES_BLOCK_SIZE / sizeof(u32)) + #define GCM_CTR_INIT 2 +-#define _walked_in (cryp->in_walk.offset - cryp->in_sg->offset) +-#define _walked_out (cryp->out_walk.offset - cryp->out_sg->offset) + #define CRYP_AUTOSUSPEND_DELAY 50 + + struct stm32_cryp_caps { +@@ -144,26 +141,16 @@ struct stm32_cryp { + size_t authsize; + size_t hw_blocksize; + +- size_t total_in; +- size_t total_in_save; +- size_t total_out; +- size_t total_out_save; ++ size_t payload_in; ++ size_t header_in; ++ size_t payload_out; + +- struct scatterlist *in_sg; + struct scatterlist *out_sg; +- struct scatterlist *out_sg_save; +- +- struct scatterlist in_sgl; +- struct scatterlist out_sgl; +- bool sgs_copied; +- +- int in_sg_len; +- int out_sg_len; + + struct scatter_walk in_walk; + struct scatter_walk out_walk; + +- u32 last_ctr[4]; ++ __be32 last_ctr[4]; + u32 gcm_ctr; + }; + +@@ -262,6 +249,7 @@ static inline int stm32_cryp_wait_output(struct stm32_cryp *cryp) + } + + static int stm32_cryp_read_auth_tag(struct stm32_cryp *cryp); ++static void stm32_cryp_finish_req(struct stm32_cryp *cryp, int err); + + static struct stm32_cryp *stm32_cryp_find_dev(struct stm32_cryp_ctx *ctx) + { +@@ -283,103 +271,6 @@ static struct stm32_cryp *stm32_cryp_find_dev(struct stm32_cryp_ctx *ctx) + return cryp; + } + +-static int stm32_cryp_check_aligned(struct scatterlist *sg, size_t total, +- size_t align) +-{ +- int len = 0; +- +- if (!total) +- return 0; +- +- if (!IS_ALIGNED(total, align)) +- return -EINVAL; +- +- while (sg) { +- if (!IS_ALIGNED(sg->offset, sizeof(u32))) +- return -EINVAL; +- +- if (!IS_ALIGNED(sg->length, align)) +- return -EINVAL; +- +- len += sg->length; +- sg = sg_next(sg); +- } +- +- if (len != total) +- return -EINVAL; +- +- return 0; +-} +- +-static int stm32_cryp_check_io_aligned(struct stm32_cryp *cryp) +-{ +- int ret; +- +- ret = stm32_cryp_check_aligned(cryp->in_sg, cryp->total_in, +- cryp->hw_blocksize); +- if (ret) +- return ret; +- +- ret = stm32_cryp_check_aligned(cryp->out_sg, cryp->total_out, +- cryp->hw_blocksize); +- +- return ret; +-} +- +-static void sg_copy_buf(void *buf, struct scatterlist *sg, +- unsigned int start, unsigned int nbytes, int out) +-{ +- struct scatter_walk walk; +- +- if (!nbytes) +- return; +- +- scatterwalk_start(&walk, sg); +- scatterwalk_advance(&walk, start); +- scatterwalk_copychunks(buf, &walk, nbytes, out); +- scatterwalk_done(&walk, out, 0); +-} +- +-static int stm32_cryp_copy_sgs(struct stm32_cryp *cryp) +-{ +- void *buf_in, *buf_out; +- int pages, total_in, total_out; +- +- if (!stm32_cryp_check_io_aligned(cryp)) { +- cryp->sgs_copied = 0; +- return 0; +- } +- +- total_in = ALIGN(cryp->total_in, cryp->hw_blocksize); +- pages = total_in ? get_order(total_in) : 1; +- buf_in = (void *)__get_free_pages(GFP_ATOMIC, pages); +- +- total_out = ALIGN(cryp->total_out, cryp->hw_blocksize); +- pages = total_out ? get_order(total_out) : 1; +- buf_out = (void *)__get_free_pages(GFP_ATOMIC, pages); +- +- if (!buf_in || !buf_out) { +- dev_err(cryp->dev, "Can't allocate pages when unaligned\n"); +- cryp->sgs_copied = 0; +- return -EFAULT; +- } +- +- sg_copy_buf(buf_in, cryp->in_sg, 0, cryp->total_in, 0); +- +- sg_init_one(&cryp->in_sgl, buf_in, total_in); +- cryp->in_sg = &cryp->in_sgl; +- cryp->in_sg_len = 1; +- +- sg_init_one(&cryp->out_sgl, buf_out, total_out); +- cryp->out_sg_save = cryp->out_sg; +- cryp->out_sg = &cryp->out_sgl; +- cryp->out_sg_len = 1; +- +- cryp->sgs_copied = 1; +- +- return 0; +-} +- + static void stm32_cryp_hw_write_iv(struct stm32_cryp *cryp, __be32 *iv) + { + if (!iv) +@@ -481,16 +372,99 @@ static int stm32_cryp_gcm_init(struct stm32_cryp *cryp, u32 cfg) + + /* Wait for end of processing */ + ret = stm32_cryp_wait_enable(cryp); +- if (ret) ++ if (ret) { + dev_err(cryp->dev, "Timeout (gcm init)\n"); ++ return ret; ++ } + +- return ret; ++ /* Prepare next phase */ ++ if (cryp->areq->assoclen) { ++ cfg |= CR_PH_HEADER; ++ stm32_cryp_write(cryp, CRYP_CR, cfg); ++ } else if (stm32_cryp_get_input_text_len(cryp)) { ++ cfg |= CR_PH_PAYLOAD; ++ stm32_cryp_write(cryp, CRYP_CR, cfg); ++ } ++ ++ return 0; ++} ++ ++static void stm32_crypt_gcmccm_end_header(struct stm32_cryp *cryp) ++{ ++ u32 cfg; ++ int err; ++ ++ /* Check if whole header written */ ++ if (!cryp->header_in) { ++ /* Wait for completion */ ++ err = stm32_cryp_wait_busy(cryp); ++ if (err) { ++ dev_err(cryp->dev, "Timeout (gcm/ccm header)\n"); ++ stm32_cryp_write(cryp, CRYP_IMSCR, 0); ++ stm32_cryp_finish_req(cryp, err); ++ return; ++ } ++ ++ if (stm32_cryp_get_input_text_len(cryp)) { ++ /* Phase 3 : payload */ ++ cfg = stm32_cryp_read(cryp, CRYP_CR); ++ cfg &= ~CR_CRYPEN; ++ stm32_cryp_write(cryp, CRYP_CR, cfg); ++ ++ cfg &= ~CR_PH_MASK; ++ cfg |= CR_PH_PAYLOAD | CR_CRYPEN; ++ stm32_cryp_write(cryp, CRYP_CR, cfg); ++ } else { ++ /* ++ * Phase 4 : tag. ++ * Nothing to read, nothing to write, caller have to ++ * end request ++ */ ++ } ++ } ++} ++ ++static void stm32_cryp_write_ccm_first_header(struct stm32_cryp *cryp) ++{ ++ unsigned int i; ++ size_t written; ++ size_t len; ++ u32 alen = cryp->areq->assoclen; ++ u32 block[AES_BLOCK_32] = {0}; ++ u8 *b8 = (u8 *)block; ++ ++ if (alen <= 65280) { ++ /* Write first u32 of B1 */ ++ b8[0] = (alen >> 8) & 0xFF; ++ b8[1] = alen & 0xFF; ++ len = 2; ++ } else { ++ /* Build the two first u32 of B1 */ ++ b8[0] = 0xFF; ++ b8[1] = 0xFE; ++ b8[2] = (alen & 0xFF000000) >> 24; ++ b8[3] = (alen & 0x00FF0000) >> 16; ++ b8[4] = (alen & 0x0000FF00) >> 8; ++ b8[5] = alen & 0x000000FF; ++ len = 6; ++ } ++ ++ written = min_t(size_t, AES_BLOCK_SIZE - len, alen); ++ ++ scatterwalk_copychunks((char *)block + len, &cryp->in_walk, written, 0); ++ for (i = 0; i < AES_BLOCK_32; i++) ++ stm32_cryp_write(cryp, CRYP_DIN, block[i]); ++ ++ cryp->header_in -= written; ++ ++ stm32_crypt_gcmccm_end_header(cryp); + } + + static int stm32_cryp_ccm_init(struct stm32_cryp *cryp, u32 cfg) + { + int ret; +- u8 iv[AES_BLOCK_SIZE], b0[AES_BLOCK_SIZE]; ++ u32 iv_32[AES_BLOCK_32], b0_32[AES_BLOCK_32]; ++ u8 *iv = (u8 *)iv_32, *b0 = (u8 *)b0_32; + __be32 *bd; + u32 *d; + unsigned int i, textlen; +@@ -531,10 +505,24 @@ static int stm32_cryp_ccm_init(struct stm32_cryp *cryp, u32 cfg) + + /* Wait for end of processing */ + ret = stm32_cryp_wait_enable(cryp); +- if (ret) ++ if (ret) { + dev_err(cryp->dev, "Timeout (ccm init)\n"); ++ return ret; ++ } + +- return ret; ++ /* Prepare next phase */ ++ if (cryp->areq->assoclen) { ++ cfg |= CR_PH_HEADER | CR_CRYPEN; ++ stm32_cryp_write(cryp, CRYP_CR, cfg); ++ ++ /* Write first (special) block (may move to next phase [payload]) */ ++ stm32_cryp_write_ccm_first_header(cryp); ++ } else if (stm32_cryp_get_input_text_len(cryp)) { ++ cfg |= CR_PH_PAYLOAD; ++ stm32_cryp_write(cryp, CRYP_CR, cfg); ++ } ++ ++ return 0; + } + + static int stm32_cryp_hw_init(struct stm32_cryp *cryp) +@@ -542,7 +530,7 @@ static int stm32_cryp_hw_init(struct stm32_cryp *cryp) + int ret; + u32 cfg, hw_mode; + +- pm_runtime_resume_and_get(cryp->dev); ++ pm_runtime_get_sync(cryp->dev); + + /* Disable interrupt */ + stm32_cryp_write(cryp, CRYP_IMSCR, 0); +@@ -605,16 +593,6 @@ static int stm32_cryp_hw_init(struct stm32_cryp *cryp) + if (ret) + return ret; + +- /* Phase 2 : header (authenticated data) */ +- if (cryp->areq->assoclen) { +- cfg |= CR_PH_HEADER; +- } else if (stm32_cryp_get_input_text_len(cryp)) { +- cfg |= CR_PH_PAYLOAD; +- stm32_cryp_write(cryp, CRYP_CR, cfg); +- } else { +- cfg |= CR_PH_INIT; +- } +- + break; + + case CR_DES_CBC: +@@ -633,8 +611,6 @@ static int stm32_cryp_hw_init(struct stm32_cryp *cryp) + + stm32_cryp_write(cryp, CRYP_CR, cfg); + +- cryp->flags &= ~FLG_CCM_PADDED_WA; +- + return 0; + } + +@@ -644,28 +620,9 @@ static void stm32_cryp_finish_req(struct stm32_cryp *cryp, int err) + /* Phase 4 : output tag */ + err = stm32_cryp_read_auth_tag(cryp); + +- if (!err && (!(is_gcm(cryp) || is_ccm(cryp)))) ++ if (!err && (!(is_gcm(cryp) || is_ccm(cryp) || is_ecb(cryp)))) + stm32_cryp_get_iv(cryp); + +- if (cryp->sgs_copied) { +- void *buf_in, *buf_out; +- int pages, len; +- +- buf_in = sg_virt(&cryp->in_sgl); +- buf_out = sg_virt(&cryp->out_sgl); +- +- sg_copy_buf(buf_out, cryp->out_sg_save, 0, +- cryp->total_out_save, 1); +- +- len = ALIGN(cryp->total_in_save, cryp->hw_blocksize); +- pages = len ? get_order(len) : 1; +- free_pages((unsigned long)buf_in, pages); +- +- len = ALIGN(cryp->total_out_save, cryp->hw_blocksize); +- pages = len ? get_order(len) : 1; +- free_pages((unsigned long)buf_out, pages); +- } +- + pm_runtime_mark_last_busy(cryp->dev); + pm_runtime_put_autosuspend(cryp->dev); + +@@ -674,8 +631,6 @@ static void stm32_cryp_finish_req(struct stm32_cryp *cryp, int err) + else + crypto_finalize_skcipher_request(cryp->engine, cryp->req, + err); +- +- memset(cryp->ctx->key, 0, cryp->ctx->keylen); + } + + static int stm32_cryp_cpu_start(struct stm32_cryp *cryp) +@@ -801,7 +756,20 @@ static int stm32_cryp_aes_aead_setkey(struct crypto_aead *tfm, const u8 *key, + static int stm32_cryp_aes_gcm_setauthsize(struct crypto_aead *tfm, + unsigned int authsize) + { +- return authsize == AES_BLOCK_SIZE ? 0 : -EINVAL; ++ switch (authsize) { ++ case 4: ++ case 8: ++ case 12: ++ case 13: ++ case 14: ++ case 15: ++ case 16: ++ break; ++ default: ++ return -EINVAL; ++ } ++ ++ return 0; + } + + static int stm32_cryp_aes_ccm_setauthsize(struct crypto_aead *tfm, +@@ -825,31 +793,61 @@ static int stm32_cryp_aes_ccm_setauthsize(struct crypto_aead *tfm, + + static int stm32_cryp_aes_ecb_encrypt(struct skcipher_request *req) + { ++ if (req->cryptlen % AES_BLOCK_SIZE) ++ return -EINVAL; ++ ++ if (req->cryptlen == 0) ++ return 0; ++ + return stm32_cryp_crypt(req, FLG_AES | FLG_ECB | FLG_ENCRYPT); + } + + static int stm32_cryp_aes_ecb_decrypt(struct skcipher_request *req) + { ++ if (req->cryptlen % AES_BLOCK_SIZE) ++ return -EINVAL; ++ ++ if (req->cryptlen == 0) ++ return 0; ++ + return stm32_cryp_crypt(req, FLG_AES | FLG_ECB); + } + + static int stm32_cryp_aes_cbc_encrypt(struct skcipher_request *req) + { ++ if (req->cryptlen % AES_BLOCK_SIZE) ++ return -EINVAL; ++ ++ if (req->cryptlen == 0) ++ return 0; ++ + return stm32_cryp_crypt(req, FLG_AES | FLG_CBC | FLG_ENCRYPT); + } + + static int stm32_cryp_aes_cbc_decrypt(struct skcipher_request *req) + { ++ if (req->cryptlen % AES_BLOCK_SIZE) ++ return -EINVAL; ++ ++ if (req->cryptlen == 0) ++ return 0; ++ + return stm32_cryp_crypt(req, FLG_AES | FLG_CBC); + } + + static int stm32_cryp_aes_ctr_encrypt(struct skcipher_request *req) + { ++ if (req->cryptlen == 0) ++ return 0; ++ + return stm32_cryp_crypt(req, FLG_AES | FLG_CTR | FLG_ENCRYPT); + } + + static int stm32_cryp_aes_ctr_decrypt(struct skcipher_request *req) + { ++ if (req->cryptlen == 0) ++ return 0; ++ + return stm32_cryp_crypt(req, FLG_AES | FLG_CTR); + } + +@@ -863,53 +861,122 @@ static int stm32_cryp_aes_gcm_decrypt(struct aead_request *req) + return stm32_cryp_aead_crypt(req, FLG_AES | FLG_GCM); + } + ++static inline int crypto_ccm_check_iv(const u8 *iv) ++{ ++ /* 2 <= L <= 8, so 1 <= L' <= 7. */ ++ if (iv[0] < 1 || iv[0] > 7) ++ return -EINVAL; ++ ++ return 0; ++} ++ + static int stm32_cryp_aes_ccm_encrypt(struct aead_request *req) + { ++ int err; ++ ++ err = crypto_ccm_check_iv(req->iv); ++ if (err) ++ return err; ++ + return stm32_cryp_aead_crypt(req, FLG_AES | FLG_CCM | FLG_ENCRYPT); + } + + static int stm32_cryp_aes_ccm_decrypt(struct aead_request *req) + { ++ int err; ++ ++ err = crypto_ccm_check_iv(req->iv); ++ if (err) ++ return err; ++ + return stm32_cryp_aead_crypt(req, FLG_AES | FLG_CCM); + } + + static int stm32_cryp_des_ecb_encrypt(struct skcipher_request *req) + { ++ if (req->cryptlen % DES_BLOCK_SIZE) ++ return -EINVAL; ++ ++ if (req->cryptlen == 0) ++ return 0; ++ + return stm32_cryp_crypt(req, FLG_DES | FLG_ECB | FLG_ENCRYPT); + } + + static int stm32_cryp_des_ecb_decrypt(struct skcipher_request *req) + { ++ if (req->cryptlen % DES_BLOCK_SIZE) ++ return -EINVAL; ++ ++ if (req->cryptlen == 0) ++ return 0; ++ + return stm32_cryp_crypt(req, FLG_DES | FLG_ECB); + } + + static int stm32_cryp_des_cbc_encrypt(struct skcipher_request *req) + { ++ if (req->cryptlen % DES_BLOCK_SIZE) ++ return -EINVAL; ++ ++ if (req->cryptlen == 0) ++ return 0; ++ + return stm32_cryp_crypt(req, FLG_DES | FLG_CBC | FLG_ENCRYPT); + } + + static int stm32_cryp_des_cbc_decrypt(struct skcipher_request *req) + { ++ if (req->cryptlen % DES_BLOCK_SIZE) ++ return -EINVAL; ++ ++ if (req->cryptlen == 0) ++ return 0; ++ + return stm32_cryp_crypt(req, FLG_DES | FLG_CBC); + } + + static int stm32_cryp_tdes_ecb_encrypt(struct skcipher_request *req) + { ++ if (req->cryptlen % DES_BLOCK_SIZE) ++ return -EINVAL; ++ ++ if (req->cryptlen == 0) ++ return 0; ++ + return stm32_cryp_crypt(req, FLG_TDES | FLG_ECB | FLG_ENCRYPT); + } + + static int stm32_cryp_tdes_ecb_decrypt(struct skcipher_request *req) + { ++ if (req->cryptlen % DES_BLOCK_SIZE) ++ return -EINVAL; ++ ++ if (req->cryptlen == 0) ++ return 0; ++ + return stm32_cryp_crypt(req, FLG_TDES | FLG_ECB); + } + + static int stm32_cryp_tdes_cbc_encrypt(struct skcipher_request *req) + { ++ if (req->cryptlen % DES_BLOCK_SIZE) ++ return -EINVAL; ++ ++ if (req->cryptlen == 0) ++ return 0; ++ + return stm32_cryp_crypt(req, FLG_TDES | FLG_CBC | FLG_ENCRYPT); + } + + static int stm32_cryp_tdes_cbc_decrypt(struct skcipher_request *req) + { ++ if (req->cryptlen % DES_BLOCK_SIZE) ++ return -EINVAL; ++ ++ if (req->cryptlen == 0) ++ return 0; ++ + return stm32_cryp_crypt(req, FLG_TDES | FLG_CBC); + } + +@@ -919,6 +986,7 @@ static int stm32_cryp_prepare_req(struct skcipher_request *req, + struct stm32_cryp_ctx *ctx; + struct stm32_cryp *cryp; + struct stm32_cryp_reqctx *rctx; ++ struct scatterlist *in_sg; + int ret; + + if (!req && !areq) +@@ -944,76 +1012,55 @@ static int stm32_cryp_prepare_req(struct skcipher_request *req, + if (req) { + cryp->req = req; + cryp->areq = NULL; +- cryp->total_in = req->cryptlen; +- cryp->total_out = cryp->total_in; ++ cryp->header_in = 0; ++ cryp->payload_in = req->cryptlen; ++ cryp->payload_out = req->cryptlen; ++ cryp->authsize = 0; + } else { + /* + * Length of input and output data: + * Encryption case: +- * INPUT = AssocData || PlainText ++ * INPUT = AssocData || PlainText + * <- assoclen -> <- cryptlen -> +- * <------- total_in -----------> + * +- * OUTPUT = AssocData || CipherText || AuthTag +- * <- assoclen -> <- cryptlen -> <- authsize -> +- * <---------------- total_out -----------------> ++ * OUTPUT = AssocData || CipherText || AuthTag ++ * <- assoclen -> <-- cryptlen --> <- authsize -> + * + * Decryption case: +- * INPUT = AssocData || CipherText || AuthTag +- * <- assoclen -> <--------- cryptlen ---------> +- * <- authsize -> +- * <---------------- total_in ------------------> ++ * INPUT = AssocData || CipherTex || AuthTag ++ * <- assoclen ---> <---------- cryptlen ----------> + * +- * OUTPUT = AssocData || PlainText +- * <- assoclen -> <- crypten - authsize -> +- * <---------- total_out -----------------> ++ * OUTPUT = AssocData || PlainText ++ * <- assoclen -> <- cryptlen - authsize -> + */ + cryp->areq = areq; + cryp->req = NULL; + cryp->authsize = crypto_aead_authsize(crypto_aead_reqtfm(areq)); +- cryp->total_in = areq->assoclen + areq->cryptlen; +- if (is_encrypt(cryp)) +- /* Append auth tag to output */ +- cryp->total_out = cryp->total_in + cryp->authsize; +- else +- /* No auth tag in output */ +- cryp->total_out = cryp->total_in - cryp->authsize; ++ if (is_encrypt(cryp)) { ++ cryp->payload_in = areq->cryptlen; ++ cryp->header_in = areq->assoclen; ++ cryp->payload_out = areq->cryptlen; ++ } else { ++ cryp->payload_in = areq->cryptlen - cryp->authsize; ++ cryp->header_in = areq->assoclen; ++ cryp->payload_out = cryp->payload_in; ++ } + } + +- cryp->total_in_save = cryp->total_in; +- cryp->total_out_save = cryp->total_out; ++ in_sg = req ? req->src : areq->src; ++ scatterwalk_start(&cryp->in_walk, in_sg); + +- cryp->in_sg = req ? req->src : areq->src; + cryp->out_sg = req ? req->dst : areq->dst; +- cryp->out_sg_save = cryp->out_sg; +- +- cryp->in_sg_len = sg_nents_for_len(cryp->in_sg, cryp->total_in); +- if (cryp->in_sg_len < 0) { +- dev_err(cryp->dev, "Cannot get in_sg_len\n"); +- ret = cryp->in_sg_len; +- return ret; +- } +- +- cryp->out_sg_len = sg_nents_for_len(cryp->out_sg, cryp->total_out); +- if (cryp->out_sg_len < 0) { +- dev_err(cryp->dev, "Cannot get out_sg_len\n"); +- ret = cryp->out_sg_len; +- return ret; +- } +- +- ret = stm32_cryp_copy_sgs(cryp); +- if (ret) +- return ret; +- +- scatterwalk_start(&cryp->in_walk, cryp->in_sg); + scatterwalk_start(&cryp->out_walk, cryp->out_sg); + + if (is_gcm(cryp) || is_ccm(cryp)) { + /* In output, jump after assoc data */ +- scatterwalk_advance(&cryp->out_walk, cryp->areq->assoclen); +- cryp->total_out -= cryp->areq->assoclen; ++ scatterwalk_copychunks(NULL, &cryp->out_walk, cryp->areq->assoclen, 2); + } + ++ if (is_ctr(cryp)) ++ memset(cryp->last_ctr, 0, sizeof(cryp->last_ctr)); ++ + ret = stm32_cryp_hw_init(cryp); + return ret; + } +@@ -1061,8 +1108,7 @@ static int stm32_cryp_aead_one_req(struct crypto_engine *engine, void *areq) + if (!cryp) + return -ENODEV; + +- if (unlikely(!cryp->areq->assoclen && +- !stm32_cryp_get_input_text_len(cryp))) { ++ if (unlikely(!cryp->payload_in && !cryp->header_in)) { + /* No input data to process: get tag and finish */ + stm32_cryp_finish_req(cryp, 0); + return 0; +@@ -1071,43 +1117,10 @@ static int stm32_cryp_aead_one_req(struct crypto_engine *engine, void *areq) + return stm32_cryp_cpu_start(cryp); + } + +-static u32 *stm32_cryp_next_out(struct stm32_cryp *cryp, u32 *dst, +- unsigned int n) +-{ +- scatterwalk_advance(&cryp->out_walk, n); +- +- if (unlikely(cryp->out_sg->length == _walked_out)) { +- cryp->out_sg = sg_next(cryp->out_sg); +- if (cryp->out_sg) { +- scatterwalk_start(&cryp->out_walk, cryp->out_sg); +- return (sg_virt(cryp->out_sg) + _walked_out); +- } +- } +- +- return (u32 *)((u8 *)dst + n); +-} +- +-static u32 *stm32_cryp_next_in(struct stm32_cryp *cryp, u32 *src, +- unsigned int n) +-{ +- scatterwalk_advance(&cryp->in_walk, n); +- +- if (unlikely(cryp->in_sg->length == _walked_in)) { +- cryp->in_sg = sg_next(cryp->in_sg); +- if (cryp->in_sg) { +- scatterwalk_start(&cryp->in_walk, cryp->in_sg); +- return (sg_virt(cryp->in_sg) + _walked_in); +- } +- } +- +- return (u32 *)((u8 *)src + n); +-} +- + static int stm32_cryp_read_auth_tag(struct stm32_cryp *cryp) + { +- u32 cfg, size_bit, *dst, d32; +- u8 *d8; +- unsigned int i, j; ++ u32 cfg, size_bit; ++ unsigned int i; + int ret = 0; + + /* Update Config */ +@@ -1130,7 +1143,7 @@ static int stm32_cryp_read_auth_tag(struct stm32_cryp *cryp) + stm32_cryp_write(cryp, CRYP_DIN, size_bit); + + size_bit = is_encrypt(cryp) ? cryp->areq->cryptlen : +- cryp->areq->cryptlen - AES_BLOCK_SIZE; ++ cryp->areq->cryptlen - cryp->authsize; + size_bit *= 8; + if (cryp->caps->swap_final) + size_bit = (__force u32)cpu_to_be32(size_bit); +@@ -1139,11 +1152,9 @@ static int stm32_cryp_read_auth_tag(struct stm32_cryp *cryp) + stm32_cryp_write(cryp, CRYP_DIN, size_bit); + } else { + /* CCM: write CTR0 */ +- u8 iv[AES_BLOCK_SIZE]; +- u32 *iv32 = (u32 *)iv; +- __be32 *biv; +- +- biv = (void *)iv; ++ u32 iv32[AES_BLOCK_32]; ++ u8 *iv = (u8 *)iv32; ++ __be32 *biv = (__be32 *)iv32; + + memcpy(iv, cryp->areq->iv, AES_BLOCK_SIZE); + memset(iv + AES_BLOCK_SIZE - 1 - iv[0], 0, iv[0] + 1); +@@ -1165,39 +1176,18 @@ static int stm32_cryp_read_auth_tag(struct stm32_cryp *cryp) + } + + if (is_encrypt(cryp)) { ++ u32 out_tag[AES_BLOCK_32]; ++ + /* Get and write tag */ +- dst = sg_virt(cryp->out_sg) + _walked_out; ++ for (i = 0; i < AES_BLOCK_32; i++) ++ out_tag[i] = stm32_cryp_read(cryp, CRYP_DOUT); + +- for (i = 0; i < AES_BLOCK_32; i++) { +- if (cryp->total_out >= sizeof(u32)) { +- /* Read a full u32 */ +- *dst = stm32_cryp_read(cryp, CRYP_DOUT); +- +- dst = stm32_cryp_next_out(cryp, dst, +- sizeof(u32)); +- cryp->total_out -= sizeof(u32); +- } else if (!cryp->total_out) { +- /* Empty fifo out (data from input padding) */ +- stm32_cryp_read(cryp, CRYP_DOUT); +- } else { +- /* Read less than an u32 */ +- d32 = stm32_cryp_read(cryp, CRYP_DOUT); +- d8 = (u8 *)&d32; +- +- for (j = 0; j < cryp->total_out; j++) { +- *((u8 *)dst) = *(d8++); +- dst = stm32_cryp_next_out(cryp, dst, 1); +- } +- cryp->total_out = 0; +- } +- } ++ scatterwalk_copychunks(out_tag, &cryp->out_walk, cryp->authsize, 1); + } else { + /* Get and check tag */ + u32 in_tag[AES_BLOCK_32], out_tag[AES_BLOCK_32]; + +- scatterwalk_map_and_copy(in_tag, cryp->in_sg, +- cryp->total_in_save - cryp->authsize, +- cryp->authsize, 0); ++ scatterwalk_copychunks(in_tag, &cryp->in_walk, cryp->authsize, 0); + + for (i = 0; i < AES_BLOCK_32; i++) + out_tag[i] = stm32_cryp_read(cryp, CRYP_DOUT); +@@ -1217,115 +1207,59 @@ static void stm32_cryp_check_ctr_counter(struct stm32_cryp *cryp) + { + u32 cr; + +- if (unlikely(cryp->last_ctr[3] == 0xFFFFFFFF)) { +- cryp->last_ctr[3] = 0; +- cryp->last_ctr[2]++; +- if (!cryp->last_ctr[2]) { +- cryp->last_ctr[1]++; +- if (!cryp->last_ctr[1]) +- cryp->last_ctr[0]++; +- } ++ if (unlikely(cryp->last_ctr[3] == cpu_to_be32(0xFFFFFFFF))) { ++ /* ++ * In this case, we need to increment manually the ctr counter, ++ * as HW doesn't handle the U32 carry. ++ */ ++ crypto_inc((u8 *)cryp->last_ctr, sizeof(cryp->last_ctr)); + + cr = stm32_cryp_read(cryp, CRYP_CR); + stm32_cryp_write(cryp, CRYP_CR, cr & ~CR_CRYPEN); + +- stm32_cryp_hw_write_iv(cryp, (__be32 *)cryp->last_ctr); ++ stm32_cryp_hw_write_iv(cryp, cryp->last_ctr); + + stm32_cryp_write(cryp, CRYP_CR, cr); + } + +- cryp->last_ctr[0] = stm32_cryp_read(cryp, CRYP_IV0LR); +- cryp->last_ctr[1] = stm32_cryp_read(cryp, CRYP_IV0RR); +- cryp->last_ctr[2] = stm32_cryp_read(cryp, CRYP_IV1LR); +- cryp->last_ctr[3] = stm32_cryp_read(cryp, CRYP_IV1RR); ++ /* The IV registers are BE */ ++ cryp->last_ctr[0] = cpu_to_be32(stm32_cryp_read(cryp, CRYP_IV0LR)); ++ cryp->last_ctr[1] = cpu_to_be32(stm32_cryp_read(cryp, CRYP_IV0RR)); ++ cryp->last_ctr[2] = cpu_to_be32(stm32_cryp_read(cryp, CRYP_IV1LR)); ++ cryp->last_ctr[3] = cpu_to_be32(stm32_cryp_read(cryp, CRYP_IV1RR)); + } + +-static bool stm32_cryp_irq_read_data(struct stm32_cryp *cryp) ++static void stm32_cryp_irq_read_data(struct stm32_cryp *cryp) + { +- unsigned int i, j; +- u32 d32, *dst; +- u8 *d8; +- size_t tag_size; +- +- /* Do no read tag now (if any) */ +- if (is_encrypt(cryp) && (is_gcm(cryp) || is_ccm(cryp))) +- tag_size = cryp->authsize; +- else +- tag_size = 0; +- +- dst = sg_virt(cryp->out_sg) + _walked_out; ++ unsigned int i; ++ u32 block[AES_BLOCK_32]; + +- for (i = 0; i < cryp->hw_blocksize / sizeof(u32); i++) { +- if (likely(cryp->total_out - tag_size >= sizeof(u32))) { +- /* Read a full u32 */ +- *dst = stm32_cryp_read(cryp, CRYP_DOUT); ++ for (i = 0; i < cryp->hw_blocksize / sizeof(u32); i++) ++ block[i] = stm32_cryp_read(cryp, CRYP_DOUT); + +- dst = stm32_cryp_next_out(cryp, dst, sizeof(u32)); +- cryp->total_out -= sizeof(u32); +- } else if (cryp->total_out == tag_size) { +- /* Empty fifo out (data from input padding) */ +- d32 = stm32_cryp_read(cryp, CRYP_DOUT); +- } else { +- /* Read less than an u32 */ +- d32 = stm32_cryp_read(cryp, CRYP_DOUT); +- d8 = (u8 *)&d32; +- +- for (j = 0; j < cryp->total_out - tag_size; j++) { +- *((u8 *)dst) = *(d8++); +- dst = stm32_cryp_next_out(cryp, dst, 1); +- } +- cryp->total_out = tag_size; +- } +- } +- +- return !(cryp->total_out - tag_size) || !cryp->total_in; ++ scatterwalk_copychunks(block, &cryp->out_walk, min_t(size_t, cryp->hw_blocksize, ++ cryp->payload_out), 1); ++ cryp->payload_out -= min_t(size_t, cryp->hw_blocksize, ++ cryp->payload_out); + } + + static void stm32_cryp_irq_write_block(struct stm32_cryp *cryp) + { +- unsigned int i, j; +- u32 *src; +- u8 d8[4]; +- size_t tag_size; +- +- /* Do no write tag (if any) */ +- if (is_decrypt(cryp) && (is_gcm(cryp) || is_ccm(cryp))) +- tag_size = cryp->authsize; +- else +- tag_size = 0; +- +- src = sg_virt(cryp->in_sg) + _walked_in; ++ unsigned int i; ++ u32 block[AES_BLOCK_32] = {0}; + +- for (i = 0; i < cryp->hw_blocksize / sizeof(u32); i++) { +- if (likely(cryp->total_in - tag_size >= sizeof(u32))) { +- /* Write a full u32 */ +- stm32_cryp_write(cryp, CRYP_DIN, *src); ++ scatterwalk_copychunks(block, &cryp->in_walk, min_t(size_t, cryp->hw_blocksize, ++ cryp->payload_in), 0); ++ for (i = 0; i < cryp->hw_blocksize / sizeof(u32); i++) ++ stm32_cryp_write(cryp, CRYP_DIN, block[i]); + +- src = stm32_cryp_next_in(cryp, src, sizeof(u32)); +- cryp->total_in -= sizeof(u32); +- } else if (cryp->total_in == tag_size) { +- /* Write padding data */ +- stm32_cryp_write(cryp, CRYP_DIN, 0); +- } else { +- /* Write less than an u32 */ +- memset(d8, 0, sizeof(u32)); +- for (j = 0; j < cryp->total_in - tag_size; j++) { +- d8[j] = *((u8 *)src); +- src = stm32_cryp_next_in(cryp, src, 1); +- } +- +- stm32_cryp_write(cryp, CRYP_DIN, *(u32 *)d8); +- cryp->total_in = tag_size; +- } +- } ++ cryp->payload_in -= min_t(size_t, cryp->hw_blocksize, cryp->payload_in); + } + + static void stm32_cryp_irq_write_gcm_padded_data(struct stm32_cryp *cryp) + { + int err; +- u32 cfg, tmp[AES_BLOCK_32]; +- size_t total_in_ori = cryp->total_in; +- struct scatterlist *out_sg_ori = cryp->out_sg; ++ u32 cfg, block[AES_BLOCK_32] = {0}; + unsigned int i; + + /* 'Special workaround' procedure described in the datasheet */ +@@ -1350,18 +1284,25 @@ static void stm32_cryp_irq_write_gcm_padded_data(struct stm32_cryp *cryp) + + /* b) pad and write the last block */ + stm32_cryp_irq_write_block(cryp); +- cryp->total_in = total_in_ori; ++ /* wait end of process */ + err = stm32_cryp_wait_output(cryp); + if (err) { +- dev_err(cryp->dev, "Timeout (write gcm header)\n"); ++ dev_err(cryp->dev, "Timeout (write gcm last data)\n"); + return stm32_cryp_finish_req(cryp, err); + } + + /* c) get and store encrypted data */ +- stm32_cryp_irq_read_data(cryp); +- scatterwalk_map_and_copy(tmp, out_sg_ori, +- cryp->total_in_save - total_in_ori, +- total_in_ori, 0); ++ /* ++ * Same code as stm32_cryp_irq_read_data(), but we want to store ++ * block value ++ */ ++ for (i = 0; i < cryp->hw_blocksize / sizeof(u32); i++) ++ block[i] = stm32_cryp_read(cryp, CRYP_DOUT); ++ ++ scatterwalk_copychunks(block, &cryp->out_walk, min_t(size_t, cryp->hw_blocksize, ++ cryp->payload_out), 1); ++ cryp->payload_out -= min_t(size_t, cryp->hw_blocksize, ++ cryp->payload_out); + + /* d) change mode back to AES GCM */ + cfg &= ~CR_ALGO_MASK; +@@ -1374,19 +1315,13 @@ static void stm32_cryp_irq_write_gcm_padded_data(struct stm32_cryp *cryp) + stm32_cryp_write(cryp, CRYP_CR, cfg); + + /* f) write padded data */ +- for (i = 0; i < AES_BLOCK_32; i++) { +- if (cryp->total_in) +- stm32_cryp_write(cryp, CRYP_DIN, tmp[i]); +- else +- stm32_cryp_write(cryp, CRYP_DIN, 0); +- +- cryp->total_in -= min_t(size_t, sizeof(u32), cryp->total_in); +- } ++ for (i = 0; i < AES_BLOCK_32; i++) ++ stm32_cryp_write(cryp, CRYP_DIN, block[i]); + + /* g) Empty fifo out */ + err = stm32_cryp_wait_output(cryp); + if (err) { +- dev_err(cryp->dev, "Timeout (write gcm header)\n"); ++ dev_err(cryp->dev, "Timeout (write gcm padded data)\n"); + return stm32_cryp_finish_req(cryp, err); + } + +@@ -1399,16 +1334,14 @@ static void stm32_cryp_irq_write_gcm_padded_data(struct stm32_cryp *cryp) + + static void stm32_cryp_irq_set_npblb(struct stm32_cryp *cryp) + { +- u32 cfg, payload_bytes; ++ u32 cfg; + + /* disable ip, set NPBLB and reneable ip */ + cfg = stm32_cryp_read(cryp, CRYP_CR); + cfg &= ~CR_CRYPEN; + stm32_cryp_write(cryp, CRYP_CR, cfg); + +- payload_bytes = is_decrypt(cryp) ? cryp->total_in - cryp->authsize : +- cryp->total_in; +- cfg |= (cryp->hw_blocksize - payload_bytes) << CR_NBPBL_SHIFT; ++ cfg |= (cryp->hw_blocksize - cryp->payload_in) << CR_NBPBL_SHIFT; + cfg |= CR_CRYPEN; + stm32_cryp_write(cryp, CRYP_CR, cfg); + } +@@ -1417,13 +1350,11 @@ static void stm32_cryp_irq_write_ccm_padded_data(struct stm32_cryp *cryp) + { + int err = 0; + u32 cfg, iv1tmp; +- u32 cstmp1[AES_BLOCK_32], cstmp2[AES_BLOCK_32], tmp[AES_BLOCK_32]; +- size_t last_total_out, total_in_ori = cryp->total_in; +- struct scatterlist *out_sg_ori = cryp->out_sg; ++ u32 cstmp1[AES_BLOCK_32], cstmp2[AES_BLOCK_32]; ++ u32 block[AES_BLOCK_32] = {0}; + unsigned int i; + + /* 'Special workaround' procedure described in the datasheet */ +- cryp->flags |= FLG_CCM_PADDED_WA; + + /* a) disable ip */ + stm32_cryp_write(cryp, CRYP_IMSCR, 0); +@@ -1453,7 +1384,7 @@ static void stm32_cryp_irq_write_ccm_padded_data(struct stm32_cryp *cryp) + + /* b) pad and write the last block */ + stm32_cryp_irq_write_block(cryp); +- cryp->total_in = total_in_ori; ++ /* wait end of process */ + err = stm32_cryp_wait_output(cryp); + if (err) { + dev_err(cryp->dev, "Timeout (wite ccm padded data)\n"); +@@ -1461,13 +1392,16 @@ static void stm32_cryp_irq_write_ccm_padded_data(struct stm32_cryp *cryp) + } + + /* c) get and store decrypted data */ +- last_total_out = cryp->total_out; +- stm32_cryp_irq_read_data(cryp); ++ /* ++ * Same code as stm32_cryp_irq_read_data(), but we want to store ++ * block value ++ */ ++ for (i = 0; i < cryp->hw_blocksize / sizeof(u32); i++) ++ block[i] = stm32_cryp_read(cryp, CRYP_DOUT); + +- memset(tmp, 0, sizeof(tmp)); +- scatterwalk_map_and_copy(tmp, out_sg_ori, +- cryp->total_out_save - last_total_out, +- last_total_out, 0); ++ scatterwalk_copychunks(block, &cryp->out_walk, min_t(size_t, cryp->hw_blocksize, ++ cryp->payload_out), 1); ++ cryp->payload_out -= min_t(size_t, cryp->hw_blocksize, cryp->payload_out); + + /* d) Load again CRYP_CSGCMCCMxR */ + for (i = 0; i < ARRAY_SIZE(cstmp2); i++) +@@ -1484,10 +1418,10 @@ static void stm32_cryp_irq_write_ccm_padded_data(struct stm32_cryp *cryp) + stm32_cryp_write(cryp, CRYP_CR, cfg); + + /* g) XOR and write padded data */ +- for (i = 0; i < ARRAY_SIZE(tmp); i++) { +- tmp[i] ^= cstmp1[i]; +- tmp[i] ^= cstmp2[i]; +- stm32_cryp_write(cryp, CRYP_DIN, tmp[i]); ++ for (i = 0; i < ARRAY_SIZE(block); i++) { ++ block[i] ^= cstmp1[i]; ++ block[i] ^= cstmp2[i]; ++ stm32_cryp_write(cryp, CRYP_DIN, block[i]); + } + + /* h) wait for completion */ +@@ -1501,30 +1435,34 @@ static void stm32_cryp_irq_write_ccm_padded_data(struct stm32_cryp *cryp) + + static void stm32_cryp_irq_write_data(struct stm32_cryp *cryp) + { +- if (unlikely(!cryp->total_in)) { ++ if (unlikely(!cryp->payload_in)) { + dev_warn(cryp->dev, "No more data to process\n"); + return; + } + +- if (unlikely(cryp->total_in < AES_BLOCK_SIZE && ++ if (unlikely(cryp->payload_in < AES_BLOCK_SIZE && + (stm32_cryp_get_hw_mode(cryp) == CR_AES_GCM) && + is_encrypt(cryp))) { + /* Padding for AES GCM encryption */ +- if (cryp->caps->padding_wa) ++ if (cryp->caps->padding_wa) { + /* Special case 1 */ +- return stm32_cryp_irq_write_gcm_padded_data(cryp); ++ stm32_cryp_irq_write_gcm_padded_data(cryp); ++ return; ++ } + + /* Setting padding bytes (NBBLB) */ + stm32_cryp_irq_set_npblb(cryp); + } + +- if (unlikely((cryp->total_in - cryp->authsize < AES_BLOCK_SIZE) && ++ if (unlikely((cryp->payload_in < AES_BLOCK_SIZE) && + (stm32_cryp_get_hw_mode(cryp) == CR_AES_CCM) && + is_decrypt(cryp))) { + /* Padding for AES CCM decryption */ +- if (cryp->caps->padding_wa) ++ if (cryp->caps->padding_wa) { + /* Special case 2 */ +- return stm32_cryp_irq_write_ccm_padded_data(cryp); ++ stm32_cryp_irq_write_ccm_padded_data(cryp); ++ return; ++ } + + /* Setting padding bytes (NBBLB) */ + stm32_cryp_irq_set_npblb(cryp); +@@ -1536,192 +1474,60 @@ static void stm32_cryp_irq_write_data(struct stm32_cryp *cryp) + stm32_cryp_irq_write_block(cryp); + } + +-static void stm32_cryp_irq_write_gcm_header(struct stm32_cryp *cryp) ++static void stm32_cryp_irq_write_gcmccm_header(struct stm32_cryp *cryp) + { +- int err; +- unsigned int i, j; +- u32 cfg, *src; +- +- src = sg_virt(cryp->in_sg) + _walked_in; +- +- for (i = 0; i < AES_BLOCK_32; i++) { +- stm32_cryp_write(cryp, CRYP_DIN, *src); +- +- src = stm32_cryp_next_in(cryp, src, sizeof(u32)); +- cryp->total_in -= min_t(size_t, sizeof(u32), cryp->total_in); +- +- /* Check if whole header written */ +- if ((cryp->total_in_save - cryp->total_in) == +- cryp->areq->assoclen) { +- /* Write padding if needed */ +- for (j = i + 1; j < AES_BLOCK_32; j++) +- stm32_cryp_write(cryp, CRYP_DIN, 0); +- +- /* Wait for completion */ +- err = stm32_cryp_wait_busy(cryp); +- if (err) { +- dev_err(cryp->dev, "Timeout (gcm header)\n"); +- return stm32_cryp_finish_req(cryp, err); +- } +- +- if (stm32_cryp_get_input_text_len(cryp)) { +- /* Phase 3 : payload */ +- cfg = stm32_cryp_read(cryp, CRYP_CR); +- cfg &= ~CR_CRYPEN; +- stm32_cryp_write(cryp, CRYP_CR, cfg); +- +- cfg &= ~CR_PH_MASK; +- cfg |= CR_PH_PAYLOAD; +- cfg |= CR_CRYPEN; +- stm32_cryp_write(cryp, CRYP_CR, cfg); +- } else { +- /* Phase 4 : tag */ +- stm32_cryp_write(cryp, CRYP_IMSCR, 0); +- stm32_cryp_finish_req(cryp, 0); +- } +- +- break; +- } +- +- if (!cryp->total_in) +- break; +- } +-} ++ unsigned int i; ++ u32 block[AES_BLOCK_32] = {0}; ++ size_t written; + +-static void stm32_cryp_irq_write_ccm_header(struct stm32_cryp *cryp) +-{ +- int err; +- unsigned int i = 0, j, k; +- u32 alen, cfg, *src; +- u8 d8[4]; +- +- src = sg_virt(cryp->in_sg) + _walked_in; +- alen = cryp->areq->assoclen; +- +- if (!_walked_in) { +- if (cryp->areq->assoclen <= 65280) { +- /* Write first u32 of B1 */ +- d8[0] = (alen >> 8) & 0xFF; +- d8[1] = alen & 0xFF; +- d8[2] = *((u8 *)src); +- src = stm32_cryp_next_in(cryp, src, 1); +- d8[3] = *((u8 *)src); +- src = stm32_cryp_next_in(cryp, src, 1); +- +- stm32_cryp_write(cryp, CRYP_DIN, *(u32 *)d8); +- i++; +- +- cryp->total_in -= min_t(size_t, 2, cryp->total_in); +- } else { +- /* Build the two first u32 of B1 */ +- d8[0] = 0xFF; +- d8[1] = 0xFE; +- d8[2] = alen & 0xFF000000; +- d8[3] = alen & 0x00FF0000; +- +- stm32_cryp_write(cryp, CRYP_DIN, *(u32 *)d8); +- i++; +- +- d8[0] = alen & 0x0000FF00; +- d8[1] = alen & 0x000000FF; +- d8[2] = *((u8 *)src); +- src = stm32_cryp_next_in(cryp, src, 1); +- d8[3] = *((u8 *)src); +- src = stm32_cryp_next_in(cryp, src, 1); +- +- stm32_cryp_write(cryp, CRYP_DIN, *(u32 *)d8); +- i++; +- +- cryp->total_in -= min_t(size_t, 2, cryp->total_in); +- } +- } ++ written = min_t(size_t, AES_BLOCK_SIZE, cryp->header_in); + +- /* Write next u32 */ +- for (; i < AES_BLOCK_32; i++) { +- /* Build an u32 */ +- memset(d8, 0, sizeof(u32)); +- for (k = 0; k < sizeof(u32); k++) { +- d8[k] = *((u8 *)src); +- src = stm32_cryp_next_in(cryp, src, 1); +- +- cryp->total_in -= min_t(size_t, 1, cryp->total_in); +- if ((cryp->total_in_save - cryp->total_in) == alen) +- break; +- } ++ scatterwalk_copychunks(block, &cryp->in_walk, written, 0); ++ for (i = 0; i < AES_BLOCK_32; i++) ++ stm32_cryp_write(cryp, CRYP_DIN, block[i]); + +- stm32_cryp_write(cryp, CRYP_DIN, *(u32 *)d8); +- +- if ((cryp->total_in_save - cryp->total_in) == alen) { +- /* Write padding if needed */ +- for (j = i + 1; j < AES_BLOCK_32; j++) +- stm32_cryp_write(cryp, CRYP_DIN, 0); +- +- /* Wait for completion */ +- err = stm32_cryp_wait_busy(cryp); +- if (err) { +- dev_err(cryp->dev, "Timeout (ccm header)\n"); +- return stm32_cryp_finish_req(cryp, err); +- } +- +- if (stm32_cryp_get_input_text_len(cryp)) { +- /* Phase 3 : payload */ +- cfg = stm32_cryp_read(cryp, CRYP_CR); +- cfg &= ~CR_CRYPEN; +- stm32_cryp_write(cryp, CRYP_CR, cfg); +- +- cfg &= ~CR_PH_MASK; +- cfg |= CR_PH_PAYLOAD; +- cfg |= CR_CRYPEN; +- stm32_cryp_write(cryp, CRYP_CR, cfg); +- } else { +- /* Phase 4 : tag */ +- stm32_cryp_write(cryp, CRYP_IMSCR, 0); +- stm32_cryp_finish_req(cryp, 0); +- } ++ cryp->header_in -= written; + +- break; +- } +- } ++ stm32_crypt_gcmccm_end_header(cryp); + } + + static irqreturn_t stm32_cryp_irq_thread(int irq, void *arg) + { + struct stm32_cryp *cryp = arg; + u32 ph; ++ u32 it_mask = stm32_cryp_read(cryp, CRYP_IMSCR); + + if (cryp->irq_status & MISR_OUT) + /* Output FIFO IRQ: read data */ +- if (unlikely(stm32_cryp_irq_read_data(cryp))) { +- /* All bytes processed, finish */ +- stm32_cryp_write(cryp, CRYP_IMSCR, 0); +- stm32_cryp_finish_req(cryp, 0); +- return IRQ_HANDLED; +- } ++ stm32_cryp_irq_read_data(cryp); + + if (cryp->irq_status & MISR_IN) { +- if (is_gcm(cryp)) { ++ if (is_gcm(cryp) || is_ccm(cryp)) { + ph = stm32_cryp_read(cryp, CRYP_CR) & CR_PH_MASK; + if (unlikely(ph == CR_PH_HEADER)) + /* Write Header */ +- stm32_cryp_irq_write_gcm_header(cryp); +- else +- /* Input FIFO IRQ: write data */ +- stm32_cryp_irq_write_data(cryp); +- cryp->gcm_ctr++; +- } else if (is_ccm(cryp)) { +- ph = stm32_cryp_read(cryp, CRYP_CR) & CR_PH_MASK; +- if (unlikely(ph == CR_PH_HEADER)) +- /* Write Header */ +- stm32_cryp_irq_write_ccm_header(cryp); ++ stm32_cryp_irq_write_gcmccm_header(cryp); + else + /* Input FIFO IRQ: write data */ + stm32_cryp_irq_write_data(cryp); ++ if (is_gcm(cryp)) ++ cryp->gcm_ctr++; + } else { + /* Input FIFO IRQ: write data */ + stm32_cryp_irq_write_data(cryp); + } + } + ++ /* Mask useless interrupts */ ++ if (!cryp->payload_in && !cryp->header_in) ++ it_mask &= ~IMSCR_IN; ++ if (!cryp->payload_out) ++ it_mask &= ~IMSCR_OUT; ++ stm32_cryp_write(cryp, CRYP_IMSCR, it_mask); ++ ++ if (!cryp->payload_in && !cryp->header_in && !cryp->payload_out) ++ stm32_cryp_finish_req(cryp, 0); ++ + return IRQ_HANDLED; + } + +@@ -1742,7 +1548,7 @@ static struct skcipher_alg crypto_algs[] = { + .base.cra_flags = CRYPTO_ALG_ASYNC, + .base.cra_blocksize = AES_BLOCK_SIZE, + .base.cra_ctxsize = sizeof(struct stm32_cryp_ctx), +- .base.cra_alignmask = 0xf, ++ .base.cra_alignmask = 0, + .base.cra_module = THIS_MODULE, + + .init = stm32_cryp_init_tfm, +@@ -1759,7 +1565,7 @@ static struct skcipher_alg crypto_algs[] = { + .base.cra_flags = CRYPTO_ALG_ASYNC, + .base.cra_blocksize = AES_BLOCK_SIZE, + .base.cra_ctxsize = sizeof(struct stm32_cryp_ctx), +- .base.cra_alignmask = 0xf, ++ .base.cra_alignmask = 0, + .base.cra_module = THIS_MODULE, + + .init = stm32_cryp_init_tfm, +@@ -1777,7 +1583,7 @@ static struct skcipher_alg crypto_algs[] = { + .base.cra_flags = CRYPTO_ALG_ASYNC, + .base.cra_blocksize = 1, + .base.cra_ctxsize = sizeof(struct stm32_cryp_ctx), +- .base.cra_alignmask = 0xf, ++ .base.cra_alignmask = 0, + .base.cra_module = THIS_MODULE, + + .init = stm32_cryp_init_tfm, +@@ -1795,7 +1601,7 @@ static struct skcipher_alg crypto_algs[] = { + .base.cra_flags = CRYPTO_ALG_ASYNC, + .base.cra_blocksize = DES_BLOCK_SIZE, + .base.cra_ctxsize = sizeof(struct stm32_cryp_ctx), +- .base.cra_alignmask = 0xf, ++ .base.cra_alignmask = 0, + .base.cra_module = THIS_MODULE, + + .init = stm32_cryp_init_tfm, +@@ -1812,7 +1618,7 @@ static struct skcipher_alg crypto_algs[] = { + .base.cra_flags = CRYPTO_ALG_ASYNC, + .base.cra_blocksize = DES_BLOCK_SIZE, + .base.cra_ctxsize = sizeof(struct stm32_cryp_ctx), +- .base.cra_alignmask = 0xf, ++ .base.cra_alignmask = 0, + .base.cra_module = THIS_MODULE, + + .init = stm32_cryp_init_tfm, +@@ -1830,7 +1636,7 @@ static struct skcipher_alg crypto_algs[] = { + .base.cra_flags = CRYPTO_ALG_ASYNC, + .base.cra_blocksize = DES_BLOCK_SIZE, + .base.cra_ctxsize = sizeof(struct stm32_cryp_ctx), +- .base.cra_alignmask = 0xf, ++ .base.cra_alignmask = 0, + .base.cra_module = THIS_MODULE, + + .init = stm32_cryp_init_tfm, +@@ -1847,7 +1653,7 @@ static struct skcipher_alg crypto_algs[] = { + .base.cra_flags = CRYPTO_ALG_ASYNC, + .base.cra_blocksize = DES_BLOCK_SIZE, + .base.cra_ctxsize = sizeof(struct stm32_cryp_ctx), +- .base.cra_alignmask = 0xf, ++ .base.cra_alignmask = 0, + .base.cra_module = THIS_MODULE, + + .init = stm32_cryp_init_tfm, +@@ -1877,7 +1683,7 @@ static struct aead_alg aead_algs[] = { + .cra_flags = CRYPTO_ALG_ASYNC, + .cra_blocksize = 1, + .cra_ctxsize = sizeof(struct stm32_cryp_ctx), +- .cra_alignmask = 0xf, ++ .cra_alignmask = 0, + .cra_module = THIS_MODULE, + }, + }, +@@ -1897,7 +1703,7 @@ static struct aead_alg aead_algs[] = { + .cra_flags = CRYPTO_ALG_ASYNC, + .cra_blocksize = 1, + .cra_ctxsize = sizeof(struct stm32_cryp_ctx), +- .cra_alignmask = 0xf, ++ .cra_alignmask = 0, + .cra_module = THIS_MODULE, + }, + }, +@@ -2025,8 +1831,6 @@ err_engine1: + list_del(&cryp->list); + spin_unlock(&cryp_list.lock); + +- pm_runtime_disable(dev); +- pm_runtime_put_noidle(dev); + pm_runtime_disable(dev); + pm_runtime_put_noidle(dev); + +diff --git a/drivers/crypto/stm32/stm32-hash.c b/drivers/crypto/stm32/stm32-hash.c +index 389de9e3302d5..d33006d43f761 100644 +--- a/drivers/crypto/stm32/stm32-hash.c ++++ b/drivers/crypto/stm32/stm32-hash.c +@@ -813,7 +813,7 @@ static void stm32_hash_finish_req(struct ahash_request *req, int err) + static int stm32_hash_hw_init(struct stm32_hash_dev *hdev, + struct stm32_hash_request_ctx *rctx) + { +- pm_runtime_resume_and_get(hdev->dev); ++ pm_runtime_get_sync(hdev->dev); + + if (!(HASH_FLAGS_INIT & hdev->flags)) { + stm32_hash_write(hdev, HASH_CR, HASH_CR_INIT); +@@ -962,7 +962,7 @@ static int stm32_hash_export(struct ahash_request *req, void *out) + u32 *preg; + unsigned int i; + +- pm_runtime_resume_and_get(hdev->dev); ++ pm_runtime_get_sync(hdev->dev); + + while ((stm32_hash_read(hdev, HASH_SR) & HASH_SR_BUSY)) + cpu_relax(); +@@ -1000,7 +1000,7 @@ static int stm32_hash_import(struct ahash_request *req, const void *in) + + preg = rctx->hw_context; + +- pm_runtime_resume_and_get(hdev->dev); ++ pm_runtime_get_sync(hdev->dev); + + stm32_hash_write(hdev, HASH_IMR, *preg++); + stm32_hash_write(hdev, HASH_STR, *preg++); +diff --git a/drivers/cxl/pmem.c b/drivers/cxl/pmem.c +index 9652c3ee41e7f..2bb2f9a0499f7 100644 +--- a/drivers/cxl/pmem.c ++++ b/drivers/cxl/pmem.c +@@ -149,14 +149,24 @@ static void cxl_nvb_update_state(struct work_struct *work) + put_device(&cxl_nvb->dev); + } + ++static void cxl_nvdimm_bridge_state_work(struct cxl_nvdimm_bridge *cxl_nvb) ++{ ++ /* ++ * Take a reference that the workqueue will drop if new work ++ * gets queued. ++ */ ++ get_device(&cxl_nvb->dev); ++ if (!queue_work(cxl_pmem_wq, &cxl_nvb->state_work)) ++ put_device(&cxl_nvb->dev); ++} ++ + static void cxl_nvdimm_bridge_remove(struct device *dev) + { + struct cxl_nvdimm_bridge *cxl_nvb = to_cxl_nvdimm_bridge(dev); + + if (cxl_nvb->state == CXL_NVB_ONLINE) + cxl_nvb->state = CXL_NVB_OFFLINE; +- if (queue_work(cxl_pmem_wq, &cxl_nvb->state_work)) +- get_device(&cxl_nvb->dev); ++ cxl_nvdimm_bridge_state_work(cxl_nvb); + } + + static int cxl_nvdimm_bridge_probe(struct device *dev) +@@ -177,8 +187,7 @@ static int cxl_nvdimm_bridge_probe(struct device *dev) + } + + cxl_nvb->state = CXL_NVB_ONLINE; +- if (queue_work(cxl_pmem_wq, &cxl_nvb->state_work)) +- get_device(&cxl_nvb->dev); ++ cxl_nvdimm_bridge_state_work(cxl_nvb); + + return 0; + } +diff --git a/drivers/dma-buf/dma-fence-array.c b/drivers/dma-buf/dma-fence-array.c +index d3fbd950be944..3e07f961e2f3d 100644 +--- a/drivers/dma-buf/dma-fence-array.c ++++ b/drivers/dma-buf/dma-fence-array.c +@@ -104,7 +104,11 @@ static bool dma_fence_array_signaled(struct dma_fence *fence) + { + struct dma_fence_array *array = to_dma_fence_array(fence); + +- return atomic_read(&array->num_pending) <= 0; ++ if (atomic_read(&array->num_pending) > 0) ++ return false; ++ ++ dma_fence_array_clear_pending_error(array); ++ return true; + } + + static void dma_fence_array_release(struct dma_fence *fence) +diff --git a/drivers/dma/at_xdmac.c b/drivers/dma/at_xdmac.c +index e18abbd56fb51..8177aed160060 100644 +--- a/drivers/dma/at_xdmac.c ++++ b/drivers/dma/at_xdmac.c +@@ -99,6 +99,7 @@ + #define AT_XDMAC_CNDC_NDE (0x1 << 0) /* Channel x Next Descriptor Enable */ + #define AT_XDMAC_CNDC_NDSUP (0x1 << 1) /* Channel x Next Descriptor Source Update */ + #define AT_XDMAC_CNDC_NDDUP (0x1 << 2) /* Channel x Next Descriptor Destination Update */ ++#define AT_XDMAC_CNDC_NDVIEW_MASK GENMASK(28, 27) + #define AT_XDMAC_CNDC_NDVIEW_NDV0 (0x0 << 3) /* Channel x Next Descriptor View 0 */ + #define AT_XDMAC_CNDC_NDVIEW_NDV1 (0x1 << 3) /* Channel x Next Descriptor View 1 */ + #define AT_XDMAC_CNDC_NDVIEW_NDV2 (0x2 << 3) /* Channel x Next Descriptor View 2 */ +@@ -252,15 +253,15 @@ struct at_xdmac { + + /* Linked List Descriptor */ + struct at_xdmac_lld { +- dma_addr_t mbr_nda; /* Next Descriptor Member */ +- u32 mbr_ubc; /* Microblock Control Member */ +- dma_addr_t mbr_sa; /* Source Address Member */ +- dma_addr_t mbr_da; /* Destination Address Member */ +- u32 mbr_cfg; /* Configuration Register */ +- u32 mbr_bc; /* Block Control Register */ +- u32 mbr_ds; /* Data Stride Register */ +- u32 mbr_sus; /* Source Microblock Stride Register */ +- u32 mbr_dus; /* Destination Microblock Stride Register */ ++ u32 mbr_nda; /* Next Descriptor Member */ ++ u32 mbr_ubc; /* Microblock Control Member */ ++ u32 mbr_sa; /* Source Address Member */ ++ u32 mbr_da; /* Destination Address Member */ ++ u32 mbr_cfg; /* Configuration Register */ ++ u32 mbr_bc; /* Block Control Register */ ++ u32 mbr_ds; /* Data Stride Register */ ++ u32 mbr_sus; /* Source Microblock Stride Register */ ++ u32 mbr_dus; /* Destination Microblock Stride Register */ + }; + + /* 64-bit alignment needed to update CNDA and CUBC registers in an atomic way. */ +@@ -385,9 +386,6 @@ static void at_xdmac_start_xfer(struct at_xdmac_chan *atchan, + + dev_vdbg(chan2dev(&atchan->chan), "%s: desc 0x%p\n", __func__, first); + +- if (at_xdmac_chan_is_enabled(atchan)) +- return; +- + /* Set transfer as active to not try to start it again. */ + first->active_xfer = true; + +@@ -405,7 +403,8 @@ static void at_xdmac_start_xfer(struct at_xdmac_chan *atchan, + */ + if (at_xdmac_chan_is_cyclic(atchan)) + reg = AT_XDMAC_CNDC_NDVIEW_NDV1; +- else if (first->lld.mbr_ubc & AT_XDMAC_MBR_UBC_NDV3) ++ else if ((first->lld.mbr_ubc & ++ AT_XDMAC_CNDC_NDVIEW_MASK) == AT_XDMAC_MBR_UBC_NDV3) + reg = AT_XDMAC_CNDC_NDVIEW_NDV3; + else + reg = AT_XDMAC_CNDC_NDVIEW_NDV2; +@@ -476,13 +475,12 @@ static dma_cookie_t at_xdmac_tx_submit(struct dma_async_tx_descriptor *tx) + spin_lock_irqsave(&atchan->lock, irqflags); + cookie = dma_cookie_assign(tx); + ++ list_add_tail(&desc->xfer_node, &atchan->xfers_list); ++ spin_unlock_irqrestore(&atchan->lock, irqflags); ++ + dev_vdbg(chan2dev(tx->chan), "%s: atchan 0x%p, add desc 0x%p to xfers_list\n", + __func__, atchan, desc); +- list_add_tail(&desc->xfer_node, &atchan->xfers_list); +- if (list_is_singular(&atchan->xfers_list)) +- at_xdmac_start_xfer(atchan, desc); + +- spin_unlock_irqrestore(&atchan->lock, irqflags); + return cookie; + } + +@@ -1623,14 +1621,17 @@ static void at_xdmac_handle_cyclic(struct at_xdmac_chan *atchan) + struct at_xdmac_desc *desc; + struct dma_async_tx_descriptor *txd; + +- if (!list_empty(&atchan->xfers_list)) { +- desc = list_first_entry(&atchan->xfers_list, +- struct at_xdmac_desc, xfer_node); +- txd = &desc->tx_dma_desc; +- +- if (txd->flags & DMA_PREP_INTERRUPT) +- dmaengine_desc_get_callback_invoke(txd, NULL); ++ spin_lock_irq(&atchan->lock); ++ if (list_empty(&atchan->xfers_list)) { ++ spin_unlock_irq(&atchan->lock); ++ return; + } ++ desc = list_first_entry(&atchan->xfers_list, struct at_xdmac_desc, ++ xfer_node); ++ spin_unlock_irq(&atchan->lock); ++ txd = &desc->tx_dma_desc; ++ if (txd->flags & DMA_PREP_INTERRUPT) ++ dmaengine_desc_get_callback_invoke(txd, NULL); + } + + static void at_xdmac_handle_error(struct at_xdmac_chan *atchan) +@@ -1784,11 +1785,9 @@ static void at_xdmac_issue_pending(struct dma_chan *chan) + + dev_dbg(chan2dev(&atchan->chan), "%s\n", __func__); + +- if (!at_xdmac_chan_is_cyclic(atchan)) { +- spin_lock_irqsave(&atchan->lock, flags); +- at_xdmac_advance_work(atchan); +- spin_unlock_irqrestore(&atchan->lock, flags); +- } ++ spin_lock_irqsave(&atchan->lock, flags); ++ at_xdmac_advance_work(atchan); ++ spin_unlock_irqrestore(&atchan->lock, flags); + + return; + } +diff --git a/drivers/dma/idxd/device.c b/drivers/dma/idxd/device.c +index 419b206f8a42d..b468ca36d3a01 100644 +--- a/drivers/dma/idxd/device.c ++++ b/drivers/dma/idxd/device.c +@@ -394,8 +394,6 @@ static void idxd_wq_disable_cleanup(struct idxd_wq *wq) + lockdep_assert_held(&wq->wq_lock); + memset(wq->wqcfg, 0, idxd->wqcfg_size); + wq->type = IDXD_WQT_NONE; +- wq->size = 0; +- wq->group = NULL; + wq->threshold = 0; + wq->priority = 0; + wq->ats_dis = 0; +@@ -404,6 +402,15 @@ static void idxd_wq_disable_cleanup(struct idxd_wq *wq) + memset(wq->name, 0, WQ_NAME_SIZE); + } + ++static void idxd_wq_device_reset_cleanup(struct idxd_wq *wq) ++{ ++ lockdep_assert_held(&wq->wq_lock); ++ ++ idxd_wq_disable_cleanup(wq); ++ wq->size = 0; ++ wq->group = NULL; ++} ++ + static void idxd_wq_ref_release(struct percpu_ref *ref) + { + struct idxd_wq *wq = container_of(ref, struct idxd_wq, wq_active); +@@ -711,6 +718,7 @@ static void idxd_device_wqs_clear_state(struct idxd_device *idxd) + + if (wq->state == IDXD_WQ_ENABLED) { + idxd_wq_disable_cleanup(wq); ++ idxd_wq_device_reset_cleanup(wq); + wq->state = IDXD_WQ_DISABLED; + } + } +diff --git a/drivers/dma/mmp_pdma.c b/drivers/dma/mmp_pdma.c +index 89f1814ff27a0..26d11885c50ec 100644 +--- a/drivers/dma/mmp_pdma.c ++++ b/drivers/dma/mmp_pdma.c +@@ -727,12 +727,6 @@ static int mmp_pdma_config_write(struct dma_chan *dchan, + + chan->dir = direction; + chan->dev_addr = addr; +- /* FIXME: drivers should be ported over to use the filter +- * function. Once that's done, the following two lines can +- * be removed. +- */ +- if (cfg->slave_id) +- chan->drcmr = cfg->slave_id; + + return 0; + } +diff --git a/drivers/dma/pxa_dma.c b/drivers/dma/pxa_dma.c +index 4a2a796e348c1..aa6e552249ab9 100644 +--- a/drivers/dma/pxa_dma.c ++++ b/drivers/dma/pxa_dma.c +@@ -910,13 +910,6 @@ static void pxad_get_config(struct pxad_chan *chan, + *dcmd |= PXA_DCMD_BURST16; + else if (maxburst == 32) + *dcmd |= PXA_DCMD_BURST32; +- +- /* FIXME: drivers should be ported over to use the filter +- * function. Once that's done, the following two lines can +- * be removed. +- */ +- if (chan->cfg.slave_id) +- chan->drcmr = chan->cfg.slave_id; + } + + static struct dma_async_tx_descriptor * +diff --git a/drivers/dma/stm32-mdma.c b/drivers/dma/stm32-mdma.c +index 18cbd1e43c2e8..f17a9ffcd00da 100644 +--- a/drivers/dma/stm32-mdma.c ++++ b/drivers/dma/stm32-mdma.c +@@ -184,7 +184,7 @@ + #define STM32_MDMA_CTBR(x) (0x68 + 0x40 * (x)) + #define STM32_MDMA_CTBR_DBUS BIT(17) + #define STM32_MDMA_CTBR_SBUS BIT(16) +-#define STM32_MDMA_CTBR_TSEL_MASK GENMASK(7, 0) ++#define STM32_MDMA_CTBR_TSEL_MASK GENMASK(5, 0) + #define STM32_MDMA_CTBR_TSEL(n) STM32_MDMA_SET(n, \ + STM32_MDMA_CTBR_TSEL_MASK) + +diff --git a/drivers/dma/uniphier-xdmac.c b/drivers/dma/uniphier-xdmac.c +index d6b8a202474f4..290836b7e1be2 100644 +--- a/drivers/dma/uniphier-xdmac.c ++++ b/drivers/dma/uniphier-xdmac.c +@@ -131,8 +131,9 @@ uniphier_xdmac_next_desc(struct uniphier_xdmac_chan *xc) + static void uniphier_xdmac_chan_start(struct uniphier_xdmac_chan *xc, + struct uniphier_xdmac_desc *xd) + { +- u32 src_mode, src_addr, src_width; +- u32 dst_mode, dst_addr, dst_width; ++ u32 src_mode, src_width; ++ u32 dst_mode, dst_width; ++ dma_addr_t src_addr, dst_addr; + u32 val, its, tnum; + enum dma_slave_buswidth buswidth; + +diff --git a/drivers/edac/synopsys_edac.c b/drivers/edac/synopsys_edac.c +index 7d08627e738b3..a5486d86fdd2f 100644 +--- a/drivers/edac/synopsys_edac.c ++++ b/drivers/edac/synopsys_edac.c +@@ -1352,8 +1352,7 @@ static int mc_probe(struct platform_device *pdev) + } + } + +- if (of_device_is_compatible(pdev->dev.of_node, +- "xlnx,zynqmp-ddrc-2.40a")) ++ if (priv->p_data->quirks & DDR_ECC_INTR_SUPPORT) + setup_address_map(priv); + #endif + +diff --git a/drivers/firmware/efi/efi-init.c b/drivers/firmware/efi/efi-init.c +index b19ce1a83f91a..b2c829e95bd14 100644 +--- a/drivers/firmware/efi/efi-init.c ++++ b/drivers/firmware/efi/efi-init.c +@@ -235,6 +235,11 @@ void __init efi_init(void) + } + + reserve_regions(); ++ /* ++ * For memblock manipulation, the cap should come after the memblock_add(). ++ * And now, memblock is fully populated, it is time to do capping. ++ */ ++ early_init_dt_check_for_usable_mem_range(); + efi_esrt_init(); + efi_mokvar_table_init(); + +diff --git a/drivers/firmware/google/Kconfig b/drivers/firmware/google/Kconfig +index 97968aece54f8..931544c9f63d4 100644 +--- a/drivers/firmware/google/Kconfig ++++ b/drivers/firmware/google/Kconfig +@@ -3,9 +3,9 @@ menuconfig GOOGLE_FIRMWARE + bool "Google Firmware Drivers" + default n + help +- These firmware drivers are used by Google's servers. They are +- only useful if you are working directly on one of their +- proprietary servers. If in doubt, say "N". ++ These firmware drivers are used by Google servers, ++ Chromebooks and other devices using coreboot firmware. ++ If in doubt, say "N". + + if GOOGLE_FIRMWARE + +diff --git a/drivers/firmware/sysfb_simplefb.c b/drivers/firmware/sysfb_simplefb.c +index b86761904949c..303a491e520d1 100644 +--- a/drivers/firmware/sysfb_simplefb.c ++++ b/drivers/firmware/sysfb_simplefb.c +@@ -113,12 +113,16 @@ __init int sysfb_create_simplefb(const struct screen_info *si, + sysfb_apply_efi_quirks(pd); + + ret = platform_device_add_resources(pd, &res, 1); +- if (ret) ++ if (ret) { ++ platform_device_put(pd); + return ret; ++ } + + ret = platform_device_add_data(pd, mode, sizeof(*mode)); +- if (ret) ++ if (ret) { ++ platform_device_put(pd); + return ret; ++ } + + return platform_device_add(pd); + } +diff --git a/drivers/gpio/gpio-aspeed-sgpio.c b/drivers/gpio/gpio-aspeed-sgpio.c +index b3a9b8488f11d..454cefbeecf0e 100644 +--- a/drivers/gpio/gpio-aspeed-sgpio.c ++++ b/drivers/gpio/gpio-aspeed-sgpio.c +@@ -31,7 +31,7 @@ struct aspeed_sgpio { + struct gpio_chip chip; + struct irq_chip intc; + struct clk *pclk; +- spinlock_t lock; ++ raw_spinlock_t lock; + void __iomem *base; + int irq; + }; +@@ -173,12 +173,12 @@ static int aspeed_sgpio_get(struct gpio_chip *gc, unsigned int offset) + enum aspeed_sgpio_reg reg; + int rc = 0; + +- spin_lock_irqsave(&gpio->lock, flags); ++ raw_spin_lock_irqsave(&gpio->lock, flags); + + reg = aspeed_sgpio_is_input(offset) ? reg_val : reg_rdata; + rc = !!(ioread32(bank_reg(gpio, bank, reg)) & GPIO_BIT(offset)); + +- spin_unlock_irqrestore(&gpio->lock, flags); ++ raw_spin_unlock_irqrestore(&gpio->lock, flags); + + return rc; + } +@@ -215,11 +215,11 @@ static void aspeed_sgpio_set(struct gpio_chip *gc, unsigned int offset, int val) + struct aspeed_sgpio *gpio = gpiochip_get_data(gc); + unsigned long flags; + +- spin_lock_irqsave(&gpio->lock, flags); ++ raw_spin_lock_irqsave(&gpio->lock, flags); + + sgpio_set_value(gc, offset, val); + +- spin_unlock_irqrestore(&gpio->lock, flags); ++ raw_spin_unlock_irqrestore(&gpio->lock, flags); + } + + static int aspeed_sgpio_dir_in(struct gpio_chip *gc, unsigned int offset) +@@ -236,9 +236,9 @@ static int aspeed_sgpio_dir_out(struct gpio_chip *gc, unsigned int offset, int v + /* No special action is required for setting the direction; we'll + * error-out in sgpio_set_value if this isn't an output GPIO */ + +- spin_lock_irqsave(&gpio->lock, flags); ++ raw_spin_lock_irqsave(&gpio->lock, flags); + rc = sgpio_set_value(gc, offset, val); +- spin_unlock_irqrestore(&gpio->lock, flags); ++ raw_spin_unlock_irqrestore(&gpio->lock, flags); + + return rc; + } +@@ -277,11 +277,11 @@ static void aspeed_sgpio_irq_ack(struct irq_data *d) + + status_addr = bank_reg(gpio, bank, reg_irq_status); + +- spin_lock_irqsave(&gpio->lock, flags); ++ raw_spin_lock_irqsave(&gpio->lock, flags); + + iowrite32(bit, status_addr); + +- spin_unlock_irqrestore(&gpio->lock, flags); ++ raw_spin_unlock_irqrestore(&gpio->lock, flags); + } + + static void aspeed_sgpio_irq_set_mask(struct irq_data *d, bool set) +@@ -296,7 +296,7 @@ static void aspeed_sgpio_irq_set_mask(struct irq_data *d, bool set) + irqd_to_aspeed_sgpio_data(d, &gpio, &bank, &bit, &offset); + addr = bank_reg(gpio, bank, reg_irq_enable); + +- spin_lock_irqsave(&gpio->lock, flags); ++ raw_spin_lock_irqsave(&gpio->lock, flags); + + reg = ioread32(addr); + if (set) +@@ -306,7 +306,7 @@ static void aspeed_sgpio_irq_set_mask(struct irq_data *d, bool set) + + iowrite32(reg, addr); + +- spin_unlock_irqrestore(&gpio->lock, flags); ++ raw_spin_unlock_irqrestore(&gpio->lock, flags); + } + + static void aspeed_sgpio_irq_mask(struct irq_data *d) +@@ -355,7 +355,7 @@ static int aspeed_sgpio_set_type(struct irq_data *d, unsigned int type) + return -EINVAL; + } + +- spin_lock_irqsave(&gpio->lock, flags); ++ raw_spin_lock_irqsave(&gpio->lock, flags); + + addr = bank_reg(gpio, bank, reg_irq_type0); + reg = ioread32(addr); +@@ -372,7 +372,7 @@ static int aspeed_sgpio_set_type(struct irq_data *d, unsigned int type) + reg = (reg & ~bit) | type2; + iowrite32(reg, addr); + +- spin_unlock_irqrestore(&gpio->lock, flags); ++ raw_spin_unlock_irqrestore(&gpio->lock, flags); + + irq_set_handler_locked(d, handler); + +@@ -467,7 +467,7 @@ static int aspeed_sgpio_reset_tolerance(struct gpio_chip *chip, + + reg = bank_reg(gpio, to_bank(offset), reg_tolerance); + +- spin_lock_irqsave(&gpio->lock, flags); ++ raw_spin_lock_irqsave(&gpio->lock, flags); + + val = readl(reg); + +@@ -478,7 +478,7 @@ static int aspeed_sgpio_reset_tolerance(struct gpio_chip *chip, + + writel(val, reg); + +- spin_unlock_irqrestore(&gpio->lock, flags); ++ raw_spin_unlock_irqrestore(&gpio->lock, flags); + + return 0; + } +@@ -575,7 +575,7 @@ static int __init aspeed_sgpio_probe(struct platform_device *pdev) + iowrite32(FIELD_PREP(ASPEED_SGPIO_CLK_DIV_MASK, sgpio_clk_div) | gpio_cnt_regval | + ASPEED_SGPIO_ENABLE, gpio->base + ASPEED_SGPIO_CTRL); + +- spin_lock_init(&gpio->lock); ++ raw_spin_lock_init(&gpio->lock); + + gpio->chip.parent = &pdev->dev; + gpio->chip.ngpio = nr_gpios * 2; +diff --git a/drivers/gpio/gpio-aspeed.c b/drivers/gpio/gpio-aspeed.c +index 3c8f20c57695f..318a7d95a1a8b 100644 +--- a/drivers/gpio/gpio-aspeed.c ++++ b/drivers/gpio/gpio-aspeed.c +@@ -53,7 +53,7 @@ struct aspeed_gpio_config { + struct aspeed_gpio { + struct gpio_chip chip; + struct irq_chip irqc; +- spinlock_t lock; ++ raw_spinlock_t lock; + void __iomem *base; + int irq; + const struct aspeed_gpio_config *config; +@@ -413,14 +413,14 @@ static void aspeed_gpio_set(struct gpio_chip *gc, unsigned int offset, + unsigned long flags; + bool copro; + +- spin_lock_irqsave(&gpio->lock, flags); ++ raw_spin_lock_irqsave(&gpio->lock, flags); + copro = aspeed_gpio_copro_request(gpio, offset); + + __aspeed_gpio_set(gc, offset, val); + + if (copro) + aspeed_gpio_copro_release(gpio, offset); +- spin_unlock_irqrestore(&gpio->lock, flags); ++ raw_spin_unlock_irqrestore(&gpio->lock, flags); + } + + static int aspeed_gpio_dir_in(struct gpio_chip *gc, unsigned int offset) +@@ -435,7 +435,7 @@ static int aspeed_gpio_dir_in(struct gpio_chip *gc, unsigned int offset) + if (!have_input(gpio, offset)) + return -ENOTSUPP; + +- spin_lock_irqsave(&gpio->lock, flags); ++ raw_spin_lock_irqsave(&gpio->lock, flags); + + reg = ioread32(addr); + reg &= ~GPIO_BIT(offset); +@@ -445,7 +445,7 @@ static int aspeed_gpio_dir_in(struct gpio_chip *gc, unsigned int offset) + if (copro) + aspeed_gpio_copro_release(gpio, offset); + +- spin_unlock_irqrestore(&gpio->lock, flags); ++ raw_spin_unlock_irqrestore(&gpio->lock, flags); + + return 0; + } +@@ -463,7 +463,7 @@ static int aspeed_gpio_dir_out(struct gpio_chip *gc, + if (!have_output(gpio, offset)) + return -ENOTSUPP; + +- spin_lock_irqsave(&gpio->lock, flags); ++ raw_spin_lock_irqsave(&gpio->lock, flags); + + reg = ioread32(addr); + reg |= GPIO_BIT(offset); +@@ -474,7 +474,7 @@ static int aspeed_gpio_dir_out(struct gpio_chip *gc, + + if (copro) + aspeed_gpio_copro_release(gpio, offset); +- spin_unlock_irqrestore(&gpio->lock, flags); ++ raw_spin_unlock_irqrestore(&gpio->lock, flags); + + return 0; + } +@@ -492,11 +492,11 @@ static int aspeed_gpio_get_direction(struct gpio_chip *gc, unsigned int offset) + if (!have_output(gpio, offset)) + return GPIO_LINE_DIRECTION_IN; + +- spin_lock_irqsave(&gpio->lock, flags); ++ raw_spin_lock_irqsave(&gpio->lock, flags); + + val = ioread32(bank_reg(gpio, bank, reg_dir)) & GPIO_BIT(offset); + +- spin_unlock_irqrestore(&gpio->lock, flags); ++ raw_spin_unlock_irqrestore(&gpio->lock, flags); + + return val ? GPIO_LINE_DIRECTION_OUT : GPIO_LINE_DIRECTION_IN; + } +@@ -539,14 +539,14 @@ static void aspeed_gpio_irq_ack(struct irq_data *d) + + status_addr = bank_reg(gpio, bank, reg_irq_status); + +- spin_lock_irqsave(&gpio->lock, flags); ++ raw_spin_lock_irqsave(&gpio->lock, flags); + copro = aspeed_gpio_copro_request(gpio, offset); + + iowrite32(bit, status_addr); + + if (copro) + aspeed_gpio_copro_release(gpio, offset); +- spin_unlock_irqrestore(&gpio->lock, flags); ++ raw_spin_unlock_irqrestore(&gpio->lock, flags); + } + + static void aspeed_gpio_irq_set_mask(struct irq_data *d, bool set) +@@ -565,7 +565,7 @@ static void aspeed_gpio_irq_set_mask(struct irq_data *d, bool set) + + addr = bank_reg(gpio, bank, reg_irq_enable); + +- spin_lock_irqsave(&gpio->lock, flags); ++ raw_spin_lock_irqsave(&gpio->lock, flags); + copro = aspeed_gpio_copro_request(gpio, offset); + + reg = ioread32(addr); +@@ -577,7 +577,7 @@ static void aspeed_gpio_irq_set_mask(struct irq_data *d, bool set) + + if (copro) + aspeed_gpio_copro_release(gpio, offset); +- spin_unlock_irqrestore(&gpio->lock, flags); ++ raw_spin_unlock_irqrestore(&gpio->lock, flags); + } + + static void aspeed_gpio_irq_mask(struct irq_data *d) +@@ -629,7 +629,7 @@ static int aspeed_gpio_set_type(struct irq_data *d, unsigned int type) + return -EINVAL; + } + +- spin_lock_irqsave(&gpio->lock, flags); ++ raw_spin_lock_irqsave(&gpio->lock, flags); + copro = aspeed_gpio_copro_request(gpio, offset); + + addr = bank_reg(gpio, bank, reg_irq_type0); +@@ -649,7 +649,7 @@ static int aspeed_gpio_set_type(struct irq_data *d, unsigned int type) + + if (copro) + aspeed_gpio_copro_release(gpio, offset); +- spin_unlock_irqrestore(&gpio->lock, flags); ++ raw_spin_unlock_irqrestore(&gpio->lock, flags); + + irq_set_handler_locked(d, handler); + +@@ -716,7 +716,7 @@ static int aspeed_gpio_reset_tolerance(struct gpio_chip *chip, + + treg = bank_reg(gpio, to_bank(offset), reg_tolerance); + +- spin_lock_irqsave(&gpio->lock, flags); ++ raw_spin_lock_irqsave(&gpio->lock, flags); + copro = aspeed_gpio_copro_request(gpio, offset); + + val = readl(treg); +@@ -730,7 +730,7 @@ static int aspeed_gpio_reset_tolerance(struct gpio_chip *chip, + + if (copro) + aspeed_gpio_copro_release(gpio, offset); +- spin_unlock_irqrestore(&gpio->lock, flags); ++ raw_spin_unlock_irqrestore(&gpio->lock, flags); + + return 0; + } +@@ -856,7 +856,7 @@ static int enable_debounce(struct gpio_chip *chip, unsigned int offset, + return rc; + } + +- spin_lock_irqsave(&gpio->lock, flags); ++ raw_spin_lock_irqsave(&gpio->lock, flags); + + if (timer_allocation_registered(gpio, offset)) { + rc = unregister_allocated_timer(gpio, offset); +@@ -916,7 +916,7 @@ static int enable_debounce(struct gpio_chip *chip, unsigned int offset, + configure_timer(gpio, offset, i); + + out: +- spin_unlock_irqrestore(&gpio->lock, flags); ++ raw_spin_unlock_irqrestore(&gpio->lock, flags); + + return rc; + } +@@ -927,13 +927,13 @@ static int disable_debounce(struct gpio_chip *chip, unsigned int offset) + unsigned long flags; + int rc; + +- spin_lock_irqsave(&gpio->lock, flags); ++ raw_spin_lock_irqsave(&gpio->lock, flags); + + rc = unregister_allocated_timer(gpio, offset); + if (!rc) + configure_timer(gpio, offset, 0); + +- spin_unlock_irqrestore(&gpio->lock, flags); ++ raw_spin_unlock_irqrestore(&gpio->lock, flags); + + return rc; + } +@@ -1015,7 +1015,7 @@ int aspeed_gpio_copro_grab_gpio(struct gpio_desc *desc, + return -EINVAL; + bindex = offset >> 3; + +- spin_lock_irqsave(&gpio->lock, flags); ++ raw_spin_lock_irqsave(&gpio->lock, flags); + + /* Sanity check, this shouldn't happen */ + if (gpio->cf_copro_bankmap[bindex] == 0xff) { +@@ -1036,7 +1036,7 @@ int aspeed_gpio_copro_grab_gpio(struct gpio_desc *desc, + if (bit) + *bit = GPIO_OFFSET(offset); + bail: +- spin_unlock_irqrestore(&gpio->lock, flags); ++ raw_spin_unlock_irqrestore(&gpio->lock, flags); + return rc; + } + EXPORT_SYMBOL_GPL(aspeed_gpio_copro_grab_gpio); +@@ -1060,7 +1060,7 @@ int aspeed_gpio_copro_release_gpio(struct gpio_desc *desc) + return -EINVAL; + bindex = offset >> 3; + +- spin_lock_irqsave(&gpio->lock, flags); ++ raw_spin_lock_irqsave(&gpio->lock, flags); + + /* Sanity check, this shouldn't happen */ + if (gpio->cf_copro_bankmap[bindex] == 0) { +@@ -1074,7 +1074,7 @@ int aspeed_gpio_copro_release_gpio(struct gpio_desc *desc) + aspeed_gpio_change_cmd_source(gpio, bank, bindex, + GPIO_CMDSRC_ARM); + bail: +- spin_unlock_irqrestore(&gpio->lock, flags); ++ raw_spin_unlock_irqrestore(&gpio->lock, flags); + return rc; + } + EXPORT_SYMBOL_GPL(aspeed_gpio_copro_release_gpio); +@@ -1148,7 +1148,7 @@ static int __init aspeed_gpio_probe(struct platform_device *pdev) + if (IS_ERR(gpio->base)) + return PTR_ERR(gpio->base); + +- spin_lock_init(&gpio->lock); ++ raw_spin_lock_init(&gpio->lock); + + gpio_id = of_match_node(aspeed_gpio_of_table, pdev->dev.of_node); + if (!gpio_id) +diff --git a/drivers/gpio/gpio-idt3243x.c b/drivers/gpio/gpio-idt3243x.c +index 50003ad2e5898..08493b05be2da 100644 +--- a/drivers/gpio/gpio-idt3243x.c ++++ b/drivers/gpio/gpio-idt3243x.c +@@ -164,8 +164,8 @@ static int idt_gpio_probe(struct platform_device *pdev) + return PTR_ERR(ctrl->pic); + + parent_irq = platform_get_irq(pdev, 0); +- if (!parent_irq) +- return -EINVAL; ++ if (parent_irq < 0) ++ return parent_irq; + + girq = &ctrl->gc.irq; + girq->chip = &idt_gpio_irqchip; +diff --git a/drivers/gpio/gpio-mpc8xxx.c b/drivers/gpio/gpio-mpc8xxx.c +index 70d6ae20b1da5..01634c8d27b38 100644 +--- a/drivers/gpio/gpio-mpc8xxx.c ++++ b/drivers/gpio/gpio-mpc8xxx.c +@@ -388,8 +388,8 @@ static int mpc8xxx_probe(struct platform_device *pdev) + } + + mpc8xxx_gc->irqn = platform_get_irq(pdev, 0); +- if (!mpc8xxx_gc->irqn) +- return 0; ++ if (mpc8xxx_gc->irqn < 0) ++ return mpc8xxx_gc->irqn; + + mpc8xxx_gc->irq = irq_domain_create_linear(fwnode, + MPC8XXX_GPIO_PINS, +diff --git a/drivers/gpio/gpiolib-acpi.c b/drivers/gpio/gpiolib-acpi.c +index 47712b6903b51..d040c72fea582 100644 +--- a/drivers/gpio/gpiolib-acpi.c ++++ b/drivers/gpio/gpiolib-acpi.c +@@ -1059,10 +1059,17 @@ int acpi_dev_gpio_irq_get_by(struct acpi_device *adev, const char *name, int ind + irq_flags = acpi_dev_get_irq_type(info.triggering, + info.polarity); + +- /* Set type if specified and different than the current one */ +- if (irq_flags != IRQ_TYPE_NONE && +- irq_flags != irq_get_trigger_type(irq)) +- irq_set_irq_type(irq, irq_flags); ++ /* ++ * If the IRQ is not already in use then set type ++ * if specified and different than the current one. ++ */ ++ if (can_request_irq(irq, irq_flags)) { ++ if (irq_flags != IRQ_TYPE_NONE && ++ irq_flags != irq_get_trigger_type(irq)) ++ irq_set_irq_type(irq, irq_flags); ++ } else { ++ dev_dbg(&adev->dev, "IRQ %d already in use\n", irq); ++ } + + return irq; + } +diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_connectors.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_connectors.c +index 0de66f59adb8a..df1f9b88a53f9 100644 +--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_connectors.c ++++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_connectors.c +@@ -387,6 +387,9 @@ amdgpu_connector_lcd_native_mode(struct drm_encoder *encoder) + native_mode->vdisplay != 0 && + native_mode->clock != 0) { + mode = drm_mode_duplicate(dev, native_mode); ++ if (!mode) ++ return NULL; ++ + mode->type = DRM_MODE_TYPE_PREFERRED | DRM_MODE_TYPE_DRIVER; + drm_mode_set_name(mode); + +@@ -401,6 +404,9 @@ amdgpu_connector_lcd_native_mode(struct drm_encoder *encoder) + * simpler. + */ + mode = drm_cvt_mode(dev, native_mode->hdisplay, native_mode->vdisplay, 60, true, false, false); ++ if (!mode) ++ return NULL; ++ + mode->type = DRM_MODE_TYPE_PREFERRED | DRM_MODE_TYPE_DRIVER; + DRM_DEBUG_KMS("Adding cvt approximation of native panel mode %s\n", mode->name); + } +diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_irq.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_irq.c +index cc2e0c9cfe0a1..4f3c62adccbde 100644 +--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_irq.c ++++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_irq.c +@@ -333,7 +333,6 @@ int amdgpu_irq_init(struct amdgpu_device *adev) + if (!amdgpu_device_has_dc_support(adev)) { + if (!adev->enable_virtual_display) + /* Disable vblank IRQs aggressively for power-saving */ +- /* XXX: can this be enabled for DC? */ + adev_to_drm(adev)->vblank_disable_immediate = true; + + r = drm_vblank_init(adev_to_drm(adev), adev->mode_info.num_crtc); +diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_psp.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_psp.c +index 9b41cb8c3de54..86e2090bbd6e0 100644 +--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_psp.c ++++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_psp.c +@@ -2207,12 +2207,16 @@ static int psp_hw_start(struct psp_context *psp) + return ret; + } + ++ if (amdgpu_sriov_vf(adev) && amdgpu_in_reset(adev)) ++ goto skip_pin_bo; ++ + ret = psp_tmr_init(psp); + if (ret) { + DRM_ERROR("PSP tmr init failed!\n"); + return ret; + } + ++skip_pin_bo: + /* + * For ASICs with DF Cstate management centralized + * to PMFW, TMR setup should be performed after PMFW +diff --git a/drivers/gpu/drm/amd/amdgpu/cik.c b/drivers/gpu/drm/amd/amdgpu/cik.c +index 54f28c075f214..f10ce740a29cc 100644 +--- a/drivers/gpu/drm/amd/amdgpu/cik.c ++++ b/drivers/gpu/drm/amd/amdgpu/cik.c +@@ -1428,6 +1428,10 @@ static int cik_asic_reset(struct amdgpu_device *adev) + { + int r; + ++ /* APUs don't have full asic reset */ ++ if (adev->flags & AMD_IS_APU) ++ return 0; ++ + if (cik_asic_reset_method(adev) == AMD_RESET_METHOD_BACO) { + dev_info(adev->dev, "BACO reset\n"); + r = amdgpu_dpm_baco_reset(adev); +diff --git a/drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c b/drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c +index e47104a1f5596..3c01be6610144 100644 +--- a/drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c ++++ b/drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c +@@ -1021,10 +1021,14 @@ static int gmc_v10_0_gart_enable(struct amdgpu_device *adev) + return -EINVAL; + } + ++ if (amdgpu_sriov_vf(adev) && amdgpu_in_reset(adev)) ++ goto skip_pin_bo; ++ + r = amdgpu_gart_table_vram_pin(adev); + if (r) + return r; + ++skip_pin_bo: + r = adev->gfxhub.funcs->gart_enable(adev); + if (r) + return r; +diff --git a/drivers/gpu/drm/amd/amdgpu/gmc_v8_0.c b/drivers/gpu/drm/amd/amdgpu/gmc_v8_0.c +index 492ebed2915be..63b890f1e8afb 100644 +--- a/drivers/gpu/drm/amd/amdgpu/gmc_v8_0.c ++++ b/drivers/gpu/drm/amd/amdgpu/gmc_v8_0.c +@@ -515,10 +515,10 @@ static void gmc_v8_0_mc_program(struct amdgpu_device *adev) + static int gmc_v8_0_mc_init(struct amdgpu_device *adev) + { + int r; ++ u32 tmp; + + adev->gmc.vram_width = amdgpu_atombios_get_vram_width(adev); + if (!adev->gmc.vram_width) { +- u32 tmp; + int chansize, numchan; + + /* Get VRAM informations */ +@@ -562,8 +562,15 @@ static int gmc_v8_0_mc_init(struct amdgpu_device *adev) + adev->gmc.vram_width = numchan * chansize; + } + /* size in MB on si */ +- adev->gmc.mc_vram_size = RREG32(mmCONFIG_MEMSIZE) * 1024ULL * 1024ULL; +- adev->gmc.real_vram_size = RREG32(mmCONFIG_MEMSIZE) * 1024ULL * 1024ULL; ++ tmp = RREG32(mmCONFIG_MEMSIZE); ++ /* some boards may have garbage in the upper 16 bits */ ++ if (tmp & 0xffff0000) { ++ DRM_INFO("Probable bad vram size: 0x%08x\n", tmp); ++ if (tmp & 0xffff) ++ tmp &= 0xffff; ++ } ++ adev->gmc.mc_vram_size = tmp * 1024ULL * 1024ULL; ++ adev->gmc.real_vram_size = adev->gmc.mc_vram_size; + + if (!(adev->flags & AMD_IS_APU)) { + r = amdgpu_device_resize_fb_bar(adev); +diff --git a/drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c b/drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c +index 5551359d5dfdc..b5d93247237b1 100644 +--- a/drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c ++++ b/drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c +@@ -1708,10 +1708,14 @@ static int gmc_v9_0_gart_enable(struct amdgpu_device *adev) + return -EINVAL; + } + ++ if (amdgpu_sriov_vf(adev) && amdgpu_in_reset(adev)) ++ goto skip_pin_bo; ++ + r = amdgpu_gart_table_vram_pin(adev); + if (r) + return r; + ++skip_pin_bo: + r = adev->gfxhub.funcs->gart_enable(adev); + if (r) + return r; +diff --git a/drivers/gpu/drm/amd/amdgpu/vi.c b/drivers/gpu/drm/amd/amdgpu/vi.c +index fe9a7cc8d9eb0..6645ebbd2696c 100644 +--- a/drivers/gpu/drm/amd/amdgpu/vi.c ++++ b/drivers/gpu/drm/amd/amdgpu/vi.c +@@ -956,6 +956,10 @@ static int vi_asic_reset(struct amdgpu_device *adev) + { + int r; + ++ /* APUs don't have full asic reset */ ++ if (adev->flags & AMD_IS_APU) ++ return 0; ++ + if (vi_asic_reset_method(adev) == AMD_RESET_METHOD_BACO) { + dev_info(adev->dev, "BACO reset\n"); + r = amdgpu_dpm_baco_reset(adev); +diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_svm.c b/drivers/gpu/drm/amd/amdkfd/kfd_svm.c +index 5a674235ae41a..830809b694dd9 100644 +--- a/drivers/gpu/drm/amd/amdkfd/kfd_svm.c ++++ b/drivers/gpu/drm/amd/amdkfd/kfd_svm.c +@@ -936,7 +936,7 @@ svm_range_split(struct svm_range *prange, uint64_t start, uint64_t last, + } + + static int +-svm_range_split_tail(struct svm_range *prange, struct svm_range *new, ++svm_range_split_tail(struct svm_range *prange, + uint64_t new_last, struct list_head *insert_list) + { + struct svm_range *tail; +@@ -948,7 +948,7 @@ svm_range_split_tail(struct svm_range *prange, struct svm_range *new, + } + + static int +-svm_range_split_head(struct svm_range *prange, struct svm_range *new, ++svm_range_split_head(struct svm_range *prange, + uint64_t new_start, struct list_head *insert_list) + { + struct svm_range *head; +@@ -1755,49 +1755,54 @@ static struct svm_range *svm_range_clone(struct svm_range *old) + } + + /** +- * svm_range_handle_overlap - split overlap ranges +- * @svms: svm range list header +- * @new: range added with this attributes +- * @start: range added start address, in pages +- * @last: range last address, in pages +- * @update_list: output, the ranges attributes are updated. For set_attr, this +- * will do validation and map to GPUs. For unmap, this will be +- * removed and unmap from GPUs +- * @insert_list: output, the ranges will be inserted into svms, attributes are +- * not changes. For set_attr, this will add into svms. +- * @remove_list:output, the ranges will be removed from svms +- * @left: the remaining range after overlap, For set_attr, this will be added +- * as new range. ++ * svm_range_add - add svm range and handle overlap ++ * @p: the range add to this process svms ++ * @start: page size aligned ++ * @size: page size aligned ++ * @nattr: number of attributes ++ * @attrs: array of attributes ++ * @update_list: output, the ranges need validate and update GPU mapping ++ * @insert_list: output, the ranges need insert to svms ++ * @remove_list: output, the ranges are replaced and need remove from svms + * +- * Total have 5 overlap cases. ++ * Check if the virtual address range has overlap with any existing ranges, ++ * split partly overlapping ranges and add new ranges in the gaps. All changes ++ * should be applied to the range_list and interval tree transactionally. If ++ * any range split or allocation fails, the entire update fails. Therefore any ++ * existing overlapping svm_ranges are cloned and the original svm_ranges left ++ * unchanged. + * +- * This function handles overlap of an address interval with existing +- * struct svm_ranges for applying new attributes. This may require +- * splitting existing struct svm_ranges. All changes should be applied to +- * the range_list and interval tree transactionally. If any split operation +- * fails, the entire update fails. Therefore the existing overlapping +- * svm_ranges are cloned and the original svm_ranges left unchanged. If the +- * transaction succeeds, the modified clones are added and the originals +- * freed. Otherwise the clones are removed and the old svm_ranges remain. ++ * If the transaction succeeds, the caller can update and insert clones and ++ * new ranges, then free the originals. + * +- * Context: The caller must hold svms->lock ++ * Otherwise the caller can free the clones and new ranges, while the old ++ * svm_ranges remain unchanged. ++ * ++ * Context: Process context, caller must hold svms->lock ++ * ++ * Return: ++ * 0 - OK, otherwise error code + */ + static int +-svm_range_handle_overlap(struct svm_range_list *svms, struct svm_range *new, +- unsigned long start, unsigned long last, +- struct list_head *update_list, +- struct list_head *insert_list, +- struct list_head *remove_list, +- unsigned long *left) ++svm_range_add(struct kfd_process *p, uint64_t start, uint64_t size, ++ uint32_t nattr, struct kfd_ioctl_svm_attribute *attrs, ++ struct list_head *update_list, struct list_head *insert_list, ++ struct list_head *remove_list) + { ++ unsigned long last = start + size - 1UL; ++ struct svm_range_list *svms = &p->svms; + struct interval_tree_node *node; ++ struct svm_range new = {0}; + struct svm_range *prange; + struct svm_range *tmp; + int r = 0; + ++ pr_debug("svms 0x%p [0x%llx 0x%lx]\n", &p->svms, start, last); ++ + INIT_LIST_HEAD(update_list); + INIT_LIST_HEAD(insert_list); + INIT_LIST_HEAD(remove_list); ++ svm_range_apply_attrs(p, &new, nattr, attrs); + + node = interval_tree_iter_first(&svms->objects, start, last); + while (node) { +@@ -1825,14 +1830,14 @@ svm_range_handle_overlap(struct svm_range_list *svms, struct svm_range *new, + + if (node->start < start) { + pr_debug("change old range start\n"); +- r = svm_range_split_head(prange, new, start, ++ r = svm_range_split_head(prange, start, + insert_list); + if (r) + goto out; + } + if (node->last > last) { + pr_debug("change old range last\n"); +- r = svm_range_split_tail(prange, new, last, ++ r = svm_range_split_tail(prange, last, + insert_list); + if (r) + goto out; +@@ -1844,7 +1849,7 @@ svm_range_handle_overlap(struct svm_range_list *svms, struct svm_range *new, + prange = old; + } + +- if (!svm_range_is_same_attrs(prange, new)) ++ if (!svm_range_is_same_attrs(prange, &new)) + list_add(&prange->update_list, update_list); + + /* insert a new node if needed */ +@@ -1864,8 +1869,16 @@ svm_range_handle_overlap(struct svm_range_list *svms, struct svm_range *new, + start = next_start; + } + +- if (left && start <= last) +- *left = last - start + 1; ++ /* add a final range at the end if needed */ ++ if (start <= last) { ++ prange = svm_range_new(svms, start, last); ++ if (!prange) { ++ r = -ENOMEM; ++ goto out; ++ } ++ list_add(&prange->insert_list, insert_list); ++ list_add(&prange->update_list, update_list); ++ } + + out: + if (r) +@@ -2693,59 +2706,6 @@ svm_range_is_valid(struct mm_struct *mm, uint64_t start, uint64_t size) + return true; + } + +-/** +- * svm_range_add - add svm range and handle overlap +- * @p: the range add to this process svms +- * @start: page size aligned +- * @size: page size aligned +- * @nattr: number of attributes +- * @attrs: array of attributes +- * @update_list: output, the ranges need validate and update GPU mapping +- * @insert_list: output, the ranges need insert to svms +- * @remove_list: output, the ranges are replaced and need remove from svms +- * +- * Check if the virtual address range has overlap with the registered ranges, +- * split the overlapped range, copy and adjust pages address and vram nodes in +- * old and new ranges. +- * +- * Context: Process context, caller must hold svms->lock +- * +- * Return: +- * 0 - OK, otherwise error code +- */ +-static int +-svm_range_add(struct kfd_process *p, uint64_t start, uint64_t size, +- uint32_t nattr, struct kfd_ioctl_svm_attribute *attrs, +- struct list_head *update_list, struct list_head *insert_list, +- struct list_head *remove_list) +-{ +- uint64_t last = start + size - 1UL; +- struct svm_range_list *svms; +- struct svm_range new = {0}; +- struct svm_range *prange; +- unsigned long left = 0; +- int r = 0; +- +- pr_debug("svms 0x%p [0x%llx 0x%llx]\n", &p->svms, start, last); +- +- svm_range_apply_attrs(p, &new, nattr, attrs); +- +- svms = &p->svms; +- +- r = svm_range_handle_overlap(svms, &new, start, last, update_list, +- insert_list, remove_list, &left); +- if (r) +- return r; +- +- if (left) { +- prange = svm_range_new(svms, last - left + 1, last); +- list_add(&prange->insert_list, insert_list); +- list_add(&prange->update_list, update_list); +- } +- +- return 0; +-} +- + /** + * svm_range_best_prefetch_location - decide the best prefetch location + * @prange: svm range structure +diff --git a/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c b/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c +index 2fbaf6f869bfb..16556ae892d4a 100644 +--- a/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c ++++ b/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c +@@ -1279,6 +1279,9 @@ static int amdgpu_dm_init(struct amdgpu_device *adev) + adev_to_drm(adev)->mode_config.cursor_width = adev->dm.dc->caps.max_cursor_size; + adev_to_drm(adev)->mode_config.cursor_height = adev->dm.dc->caps.max_cursor_size; + ++ /* Disable vblank IRQs aggressively for power-saving */ ++ adev_to_drm(adev)->vblank_disable_immediate = true; ++ + if (drm_vblank_init(adev_to_drm(adev), adev->dm.display_indexes_num)) { + DRM_ERROR( + "amdgpu: failed to initialize sw for display support.\n"); +diff --git a/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_debugfs.c b/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_debugfs.c +index de9ec5ddb6c72..e94ddd5e7b638 100644 +--- a/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_debugfs.c ++++ b/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_debugfs.c +@@ -2908,10 +2908,13 @@ static int crc_win_update_set(void *data, u64 val) + struct amdgpu_device *adev = drm_to_adev(new_crtc->dev); + struct crc_rd_work *crc_rd_wrk = adev->dm.crc_rd_wrk; + ++ if (!crc_rd_wrk) ++ return 0; ++ + if (val) { + spin_lock_irq(&adev_to_drm(adev)->event_lock); + spin_lock_irq(&crc_rd_wrk->crc_rd_work_lock); +- if (crc_rd_wrk && crc_rd_wrk->crtc) { ++ if (crc_rd_wrk->crtc) { + old_crtc = crc_rd_wrk->crtc; + old_acrtc = to_amdgpu_crtc(old_crtc); + } +diff --git a/drivers/gpu/drm/amd/display/dc/clk_mgr/clk_mgr.c b/drivers/gpu/drm/amd/display/dc/clk_mgr/clk_mgr.c +index bb31541f80723..6420527fe476c 100644 +--- a/drivers/gpu/drm/amd/display/dc/clk_mgr/clk_mgr.c ++++ b/drivers/gpu/drm/amd/display/dc/clk_mgr/clk_mgr.c +@@ -306,8 +306,7 @@ void dc_destroy_clk_mgr(struct clk_mgr *clk_mgr_base) + case FAMILY_NV: + if (ASICREV_IS_SIENNA_CICHLID_P(clk_mgr_base->ctx->asic_id.hw_internal_rev)) { + dcn3_clk_mgr_destroy(clk_mgr); +- } +- if (ASICREV_IS_DIMGREY_CAVEFISH_P(clk_mgr_base->ctx->asic_id.hw_internal_rev)) { ++ } else if (ASICREV_IS_DIMGREY_CAVEFISH_P(clk_mgr_base->ctx->asic_id.hw_internal_rev)) { + dcn3_clk_mgr_destroy(clk_mgr); + } + if (ASICREV_IS_BEIGE_GOBY_P(clk_mgr_base->ctx->asic_id.hw_internal_rev)) { +diff --git a/drivers/gpu/drm/amd/display/dc/core/dc.c b/drivers/gpu/drm/amd/display/dc/core/dc.c +index c798c65d42765..1860ccc3f4f2c 100644 +--- a/drivers/gpu/drm/amd/display/dc/core/dc.c ++++ b/drivers/gpu/drm/amd/display/dc/core/dc.c +@@ -2703,7 +2703,8 @@ static void commit_planes_for_stream(struct dc *dc, + #endif + + if ((update_type != UPDATE_TYPE_FAST) && stream->update_flags.bits.dsc_changed) +- if (top_pipe_to_program->stream_res.tg->funcs->lock_doublebuffer_enable) { ++ if (top_pipe_to_program && ++ top_pipe_to_program->stream_res.tg->funcs->lock_doublebuffer_enable) { + if (should_use_dmub_lock(stream->link)) { + union dmub_hw_lock_flags hw_locks = { 0 }; + struct dmub_hw_lock_inst_flags inst_flags = { 0 }; +diff --git a/drivers/gpu/drm/amd/display/dc/core/dc_link.c b/drivers/gpu/drm/amd/display/dc/core/dc_link.c +index 1e44b13c1c7de..3c4205248efc2 100644 +--- a/drivers/gpu/drm/amd/display/dc/core/dc_link.c ++++ b/drivers/gpu/drm/amd/display/dc/core/dc_link.c +@@ -1696,6 +1696,8 @@ static void enable_stream_features(struct pipe_ctx *pipe_ctx) + union down_spread_ctrl old_downspread; + union down_spread_ctrl new_downspread; + ++ memset(&old_downspread, 0, sizeof(old_downspread)); ++ + core_link_read_dpcd(link, DP_DOWNSPREAD_CTRL, + &old_downspread.raw, sizeof(old_downspread)); + +diff --git a/drivers/gpu/drm/amd/display/dc/dcn31/dcn31_resource.c b/drivers/gpu/drm/amd/display/dc/dcn31/dcn31_resource.c +index 0fe570717ba01..d4fe5352421fc 100644 +--- a/drivers/gpu/drm/amd/display/dc/dcn31/dcn31_resource.c ++++ b/drivers/gpu/drm/amd/display/dc/dcn31/dcn31_resource.c +@@ -470,7 +470,8 @@ static const struct dcn30_afmt_mask afmt_mask = { + SE_DCN3_REG_LIST(id)\ + } + +-static const struct dcn10_stream_enc_registers stream_enc_regs[] = { ++/* Some encoders won't be initialized here - but they're logical, not physical. */ ++static const struct dcn10_stream_enc_registers stream_enc_regs[ENGINE_ID_COUNT] = { + stream_enc_regs(0), + stream_enc_regs(1), + stream_enc_regs(2), +diff --git a/drivers/gpu/drm/amd/pm/amdgpu_pm.c b/drivers/gpu/drm/amd/pm/amdgpu_pm.c +index 249cb0aeb5ae4..32a0fd5e84b73 100644 +--- a/drivers/gpu/drm/amd/pm/amdgpu_pm.c ++++ b/drivers/gpu/drm/amd/pm/amdgpu_pm.c +@@ -2117,6 +2117,12 @@ static int default_attr_update(struct amdgpu_device *adev, struct amdgpu_device_ + } + } + ++ /* setting should not be allowed from VF */ ++ if (amdgpu_sriov_vf(adev)) { ++ dev_attr->attr.mode &= ~S_IWUGO; ++ dev_attr->store = NULL; ++ } ++ + #undef DEVICE_ATTR_IS + + return 0; +diff --git a/drivers/gpu/drm/bridge/analogix/analogix_dp_reg.c b/drivers/gpu/drm/bridge/analogix/analogix_dp_reg.c +index cab6c8b92efd4..6a4f20fccf841 100644 +--- a/drivers/gpu/drm/bridge/analogix/analogix_dp_reg.c ++++ b/drivers/gpu/drm/bridge/analogix/analogix_dp_reg.c +@@ -998,11 +998,21 @@ int analogix_dp_send_psr_spd(struct analogix_dp_device *dp, + if (!blocking) + return 0; + ++ /* ++ * db[1]!=0: entering PSR, wait for fully active remote frame buffer. ++ * db[1]==0: exiting PSR, wait for either ++ * (a) ACTIVE_RESYNC - the sink "must display the ++ * incoming active frames from the Source device with no visible ++ * glitches and/or artifacts", even though timings may still be ++ * re-synchronizing; or ++ * (b) INACTIVE - the transition is fully complete. ++ */ + ret = readx_poll_timeout(analogix_dp_get_psr_status, dp, psr_status, + psr_status >= 0 && + ((vsc->db[1] && psr_status == DP_PSR_SINK_ACTIVE_RFB) || +- (!vsc->db[1] && psr_status == DP_PSR_SINK_INACTIVE)), 1500, +- DP_TIMEOUT_PSR_LOOP_MS * 1000); ++ (!vsc->db[1] && (psr_status == DP_PSR_SINK_ACTIVE_RESYNC || ++ psr_status == DP_PSR_SINK_INACTIVE))), ++ 1500, DP_TIMEOUT_PSR_LOOP_MS * 1000); + if (ret) { + dev_warn(dp->dev, "Failed to apply PSR %d\n", ret); + return ret; +diff --git a/drivers/gpu/drm/bridge/display-connector.c b/drivers/gpu/drm/bridge/display-connector.c +index 05eb759da6fc6..847a0dce7f1d3 100644 +--- a/drivers/gpu/drm/bridge/display-connector.c ++++ b/drivers/gpu/drm/bridge/display-connector.c +@@ -107,7 +107,7 @@ static int display_connector_probe(struct platform_device *pdev) + { + struct display_connector *conn; + unsigned int type; +- const char *label; ++ const char *label = NULL; + int ret; + + conn = devm_kzalloc(&pdev->dev, sizeof(*conn), GFP_KERNEL); +diff --git a/drivers/gpu/drm/bridge/megachips-stdpxxxx-ge-b850v3-fw.c b/drivers/gpu/drm/bridge/megachips-stdpxxxx-ge-b850v3-fw.c +index d2808c4a6fb1c..cce98bf2a4e73 100644 +--- a/drivers/gpu/drm/bridge/megachips-stdpxxxx-ge-b850v3-fw.c ++++ b/drivers/gpu/drm/bridge/megachips-stdpxxxx-ge-b850v3-fw.c +@@ -306,19 +306,10 @@ out: + mutex_unlock(&ge_b850v3_lvds_dev_mutex); + } + +-static int stdp4028_ge_b850v3_fw_probe(struct i2c_client *stdp4028_i2c, +- const struct i2c_device_id *id) ++static int ge_b850v3_register(void) + { ++ struct i2c_client *stdp4028_i2c = ge_b850v3_lvds_ptr->stdp4028_i2c; + struct device *dev = &stdp4028_i2c->dev; +- int ret; +- +- ret = ge_b850v3_lvds_init(dev); +- +- if (ret) +- return ret; +- +- ge_b850v3_lvds_ptr->stdp4028_i2c = stdp4028_i2c; +- i2c_set_clientdata(stdp4028_i2c, ge_b850v3_lvds_ptr); + + /* drm bridge initialization */ + ge_b850v3_lvds_ptr->bridge.funcs = &ge_b850v3_lvds_funcs; +@@ -343,6 +334,27 @@ static int stdp4028_ge_b850v3_fw_probe(struct i2c_client *stdp4028_i2c, + "ge-b850v3-lvds-dp", ge_b850v3_lvds_ptr); + } + ++static int stdp4028_ge_b850v3_fw_probe(struct i2c_client *stdp4028_i2c, ++ const struct i2c_device_id *id) ++{ ++ struct device *dev = &stdp4028_i2c->dev; ++ int ret; ++ ++ ret = ge_b850v3_lvds_init(dev); ++ ++ if (ret) ++ return ret; ++ ++ ge_b850v3_lvds_ptr->stdp4028_i2c = stdp4028_i2c; ++ i2c_set_clientdata(stdp4028_i2c, ge_b850v3_lvds_ptr); ++ ++ /* Only register after both bridges are probed */ ++ if (!ge_b850v3_lvds_ptr->stdp2690_i2c) ++ return 0; ++ ++ return ge_b850v3_register(); ++} ++ + static int stdp4028_ge_b850v3_fw_remove(struct i2c_client *stdp4028_i2c) + { + ge_b850v3_lvds_remove(); +@@ -386,7 +398,11 @@ static int stdp2690_ge_b850v3_fw_probe(struct i2c_client *stdp2690_i2c, + ge_b850v3_lvds_ptr->stdp2690_i2c = stdp2690_i2c; + i2c_set_clientdata(stdp2690_i2c, ge_b850v3_lvds_ptr); + +- return 0; ++ /* Only register after both bridges are probed */ ++ if (!ge_b850v3_lvds_ptr->stdp4028_i2c) ++ return 0; ++ ++ return ge_b850v3_register(); + } + + static int stdp2690_ge_b850v3_fw_remove(struct i2c_client *stdp2690_i2c) +diff --git a/drivers/gpu/drm/bridge/synopsys/dw-hdmi-ahb-audio.c b/drivers/gpu/drm/bridge/synopsys/dw-hdmi-ahb-audio.c +index d0db1acf11d73..7d2ed0ed2fe26 100644 +--- a/drivers/gpu/drm/bridge/synopsys/dw-hdmi-ahb-audio.c ++++ b/drivers/gpu/drm/bridge/synopsys/dw-hdmi-ahb-audio.c +@@ -320,13 +320,17 @@ static int dw_hdmi_open(struct snd_pcm_substream *substream) + struct snd_pcm_runtime *runtime = substream->runtime; + struct snd_dw_hdmi *dw = substream->private_data; + void __iomem *base = dw->data.base; ++ u8 *eld; + int ret; + + runtime->hw = dw_hdmi_hw; + +- ret = snd_pcm_hw_constraint_eld(runtime, dw->data.eld); +- if (ret < 0) +- return ret; ++ eld = dw->data.get_eld(dw->data.hdmi); ++ if (eld) { ++ ret = snd_pcm_hw_constraint_eld(runtime, eld); ++ if (ret < 0) ++ return ret; ++ } + + ret = snd_pcm_limit_hw_rates(runtime); + if (ret < 0) +diff --git a/drivers/gpu/drm/bridge/synopsys/dw-hdmi-audio.h b/drivers/gpu/drm/bridge/synopsys/dw-hdmi-audio.h +index cb07dc0da5a70..f72d27208ebef 100644 +--- a/drivers/gpu/drm/bridge/synopsys/dw-hdmi-audio.h ++++ b/drivers/gpu/drm/bridge/synopsys/dw-hdmi-audio.h +@@ -9,15 +9,15 @@ struct dw_hdmi_audio_data { + void __iomem *base; + int irq; + struct dw_hdmi *hdmi; +- u8 *eld; ++ u8 *(*get_eld)(struct dw_hdmi *hdmi); + }; + + struct dw_hdmi_i2s_audio_data { + struct dw_hdmi *hdmi; +- u8 *eld; + + void (*write)(struct dw_hdmi *hdmi, u8 val, int offset); + u8 (*read)(struct dw_hdmi *hdmi, int offset); ++ u8 *(*get_eld)(struct dw_hdmi *hdmi); + }; + + #endif +diff --git a/drivers/gpu/drm/bridge/synopsys/dw-hdmi-i2s-audio.c b/drivers/gpu/drm/bridge/synopsys/dw-hdmi-i2s-audio.c +index feb04f127b550..f50b47ac11a82 100644 +--- a/drivers/gpu/drm/bridge/synopsys/dw-hdmi-i2s-audio.c ++++ b/drivers/gpu/drm/bridge/synopsys/dw-hdmi-i2s-audio.c +@@ -135,8 +135,15 @@ static int dw_hdmi_i2s_get_eld(struct device *dev, void *data, uint8_t *buf, + size_t len) + { + struct dw_hdmi_i2s_audio_data *audio = data; ++ u8 *eld; ++ ++ eld = audio->get_eld(audio->hdmi); ++ if (eld) ++ memcpy(buf, eld, min_t(size_t, MAX_ELD_BYTES, len)); ++ else ++ /* Pass en empty ELD if connector not available */ ++ memset(buf, 0, len); + +- memcpy(buf, audio->eld, min_t(size_t, MAX_ELD_BYTES, len)); + return 0; + } + +diff --git a/drivers/gpu/drm/bridge/synopsys/dw-hdmi.c b/drivers/gpu/drm/bridge/synopsys/dw-hdmi.c +index f08d0fded61f7..e1211a5b334ba 100644 +--- a/drivers/gpu/drm/bridge/synopsys/dw-hdmi.c ++++ b/drivers/gpu/drm/bridge/synopsys/dw-hdmi.c +@@ -757,6 +757,14 @@ static void hdmi_enable_audio_clk(struct dw_hdmi *hdmi, bool enable) + hdmi_writeb(hdmi, hdmi->mc_clkdis, HDMI_MC_CLKDIS); + } + ++static u8 *hdmi_audio_get_eld(struct dw_hdmi *hdmi) ++{ ++ if (!hdmi->curr_conn) ++ return NULL; ++ ++ return hdmi->curr_conn->eld; ++} ++ + static void dw_hdmi_ahb_audio_enable(struct dw_hdmi *hdmi) + { + hdmi_set_cts_n(hdmi, hdmi->audio_cts, hdmi->audio_n); +@@ -3431,7 +3439,7 @@ struct dw_hdmi *dw_hdmi_probe(struct platform_device *pdev, + audio.base = hdmi->regs; + audio.irq = irq; + audio.hdmi = hdmi; +- audio.eld = hdmi->connector.eld; ++ audio.get_eld = hdmi_audio_get_eld; + hdmi->enable_audio = dw_hdmi_ahb_audio_enable; + hdmi->disable_audio = dw_hdmi_ahb_audio_disable; + +@@ -3444,7 +3452,7 @@ struct dw_hdmi *dw_hdmi_probe(struct platform_device *pdev, + struct dw_hdmi_i2s_audio_data audio; + + audio.hdmi = hdmi; +- audio.eld = hdmi->connector.eld; ++ audio.get_eld = hdmi_audio_get_eld; + audio.write = hdmi_writeb; + audio.read = hdmi_readb; + hdmi->enable_audio = dw_hdmi_i2s_audio_enable; +diff --git a/drivers/gpu/drm/bridge/ti-sn65dsi86.c b/drivers/gpu/drm/bridge/ti-sn65dsi86.c +index 41d48a393e7f5..4d08246f930c3 100644 +--- a/drivers/gpu/drm/bridge/ti-sn65dsi86.c ++++ b/drivers/gpu/drm/bridge/ti-sn65dsi86.c +@@ -188,6 +188,7 @@ static const struct regmap_config ti_sn65dsi86_regmap_config = { + .val_bits = 8, + .volatile_table = &ti_sn_bridge_volatile_table, + .cache_type = REGCACHE_NONE, ++ .max_register = 0xFF, + }; + + static void ti_sn65dsi86_write_u16(struct ti_sn65dsi86 *pdata, +diff --git a/drivers/gpu/drm/drm_dp_helper.c b/drivers/gpu/drm/drm_dp_helper.c +index 6d0f2c447f3b9..7bb24523a7493 100644 +--- a/drivers/gpu/drm/drm_dp_helper.c ++++ b/drivers/gpu/drm/drm_dp_helper.c +@@ -3214,27 +3214,13 @@ int drm_edp_backlight_enable(struct drm_dp_aux *aux, const struct drm_edp_backli + const u16 level) + { + int ret; +- u8 dpcd_buf, new_dpcd_buf; ++ u8 dpcd_buf = DP_EDP_BACKLIGHT_CONTROL_MODE_DPCD; + +- ret = drm_dp_dpcd_readb(aux, DP_EDP_BACKLIGHT_MODE_SET_REGISTER, &dpcd_buf); +- if (ret != 1) { +- drm_dbg_kms(aux->drm_dev, +- "%s: Failed to read backlight mode: %d\n", aux->name, ret); +- return ret < 0 ? ret : -EIO; +- } +- +- new_dpcd_buf = dpcd_buf; +- +- if ((dpcd_buf & DP_EDP_BACKLIGHT_CONTROL_MODE_MASK) != DP_EDP_BACKLIGHT_CONTROL_MODE_DPCD) { +- new_dpcd_buf &= ~DP_EDP_BACKLIGHT_CONTROL_MODE_MASK; +- new_dpcd_buf |= DP_EDP_BACKLIGHT_CONTROL_MODE_DPCD; +- +- if (bl->pwmgen_bit_count) { +- ret = drm_dp_dpcd_writeb(aux, DP_EDP_PWMGEN_BIT_COUNT, bl->pwmgen_bit_count); +- if (ret != 1) +- drm_dbg_kms(aux->drm_dev, "%s: Failed to write aux pwmgen bit count: %d\n", +- aux->name, ret); +- } ++ if (bl->pwmgen_bit_count) { ++ ret = drm_dp_dpcd_writeb(aux, DP_EDP_PWMGEN_BIT_COUNT, bl->pwmgen_bit_count); ++ if (ret != 1) ++ drm_dbg_kms(aux->drm_dev, "%s: Failed to write aux pwmgen bit count: %d\n", ++ aux->name, ret); + } + + if (bl->pwm_freq_pre_divider) { +@@ -3244,16 +3230,14 @@ int drm_edp_backlight_enable(struct drm_dp_aux *aux, const struct drm_edp_backli + "%s: Failed to write aux backlight frequency: %d\n", + aux->name, ret); + else +- new_dpcd_buf |= DP_EDP_BACKLIGHT_FREQ_AUX_SET_ENABLE; ++ dpcd_buf |= DP_EDP_BACKLIGHT_FREQ_AUX_SET_ENABLE; + } + +- if (new_dpcd_buf != dpcd_buf) { +- ret = drm_dp_dpcd_writeb(aux, DP_EDP_BACKLIGHT_MODE_SET_REGISTER, new_dpcd_buf); +- if (ret != 1) { +- drm_dbg_kms(aux->drm_dev, "%s: Failed to write aux backlight mode: %d\n", +- aux->name, ret); +- return ret < 0 ? ret : -EIO; +- } ++ ret = drm_dp_dpcd_writeb(aux, DP_EDP_BACKLIGHT_MODE_SET_REGISTER, dpcd_buf); ++ if (ret != 1) { ++ drm_dbg_kms(aux->drm_dev, "%s: Failed to write aux backlight mode: %d\n", ++ aux->name, ret); ++ return ret < 0 ? ret : -EIO; + } + + ret = drm_edp_backlight_set_level(aux, bl, level); +diff --git a/drivers/gpu/drm/drm_drv.c b/drivers/gpu/drm/drm_drv.c +index 7a5097467ba5c..b3a1636d1b984 100644 +--- a/drivers/gpu/drm/drm_drv.c ++++ b/drivers/gpu/drm/drm_drv.c +@@ -581,6 +581,7 @@ static int drm_dev_init(struct drm_device *dev, + const struct drm_driver *driver, + struct device *parent) + { ++ struct inode *inode; + int ret; + + if (!drm_core_init_complete) { +@@ -617,13 +618,15 @@ static int drm_dev_init(struct drm_device *dev, + if (ret) + return ret; + +- dev->anon_inode = drm_fs_inode_new(); +- if (IS_ERR(dev->anon_inode)) { +- ret = PTR_ERR(dev->anon_inode); ++ inode = drm_fs_inode_new(); ++ if (IS_ERR(inode)) { ++ ret = PTR_ERR(inode); + DRM_ERROR("Cannot allocate anonymous inode: %d\n", ret); + goto err; + } + ++ dev->anon_inode = inode; ++ + if (drm_core_check_feature(dev, DRIVER_RENDER)) { + ret = drm_minor_alloc(dev, DRM_MINOR_RENDER); + if (ret) +diff --git a/drivers/gpu/drm/drm_panel_orientation_quirks.c b/drivers/gpu/drm/drm_panel_orientation_quirks.c +index a950d5db211c5..9d1bd8f491ad7 100644 +--- a/drivers/gpu/drm/drm_panel_orientation_quirks.c ++++ b/drivers/gpu/drm/drm_panel_orientation_quirks.c +@@ -248,6 +248,12 @@ static const struct dmi_system_id orientation_data[] = { + DMI_EXACT_MATCH(DMI_PRODUCT_VERSION, "Lenovo ideapad D330-10IGM"), + }, + .driver_data = (void *)&lcd1200x1920_rightside_up, ++ }, { /* Lenovo Yoga Book X90F / X91F / X91L */ ++ .matches = { ++ /* Non exact match to match all versions */ ++ DMI_MATCH(DMI_PRODUCT_NAME, "Lenovo YB1-X9"), ++ }, ++ .driver_data = (void *)&lcd1200x1920_rightside_up, + }, { /* OneGX1 Pro */ + .matches = { + DMI_EXACT_MATCH(DMI_SYS_VENDOR, "SYSTEM_MANUFACTURER"), +diff --git a/drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c b/drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c +index 486259e154aff..225fa5879ebd9 100644 +--- a/drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c ++++ b/drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c +@@ -469,6 +469,12 @@ int etnaviv_ioctl_gem_submit(struct drm_device *dev, void *data, + return -EINVAL; + } + ++ if (args->stream_size > SZ_64K || args->nr_relocs > SZ_64K || ++ args->nr_bos > SZ_64K || args->nr_pmrs > 128) { ++ DRM_ERROR("submit arguments out of size limits\n"); ++ return -EINVAL; ++ } ++ + /* + * Copy the command submission and bo array to kernel space in + * one go, and do this outside of any locks. +diff --git a/drivers/gpu/drm/etnaviv/etnaviv_gpu.h b/drivers/gpu/drm/etnaviv/etnaviv_gpu.h +index 1c75c8ed5bcea..85eddd492774d 100644 +--- a/drivers/gpu/drm/etnaviv/etnaviv_gpu.h ++++ b/drivers/gpu/drm/etnaviv/etnaviv_gpu.h +@@ -130,6 +130,7 @@ struct etnaviv_gpu { + + /* hang detection */ + u32 hangcheck_dma_addr; ++ u32 hangcheck_fence; + + void __iomem *mmio; + int irq; +diff --git a/drivers/gpu/drm/etnaviv/etnaviv_sched.c b/drivers/gpu/drm/etnaviv/etnaviv_sched.c +index feb6da1b6cebc..bbf391f48f949 100644 +--- a/drivers/gpu/drm/etnaviv/etnaviv_sched.c ++++ b/drivers/gpu/drm/etnaviv/etnaviv_sched.c +@@ -107,8 +107,10 @@ static enum drm_gpu_sched_stat etnaviv_sched_timedout_job(struct drm_sched_job + */ + dma_addr = gpu_read(gpu, VIVS_FE_DMA_ADDRESS); + change = dma_addr - gpu->hangcheck_dma_addr; +- if (change < 0 || change > 16) { ++ if (gpu->completed_fence != gpu->hangcheck_fence || ++ change < 0 || change > 16) { + gpu->hangcheck_dma_addr = dma_addr; ++ gpu->hangcheck_fence = gpu->completed_fence; + goto out_no_timeout; + } + +diff --git a/drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c b/drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c +index ba2c08f1a797c..876620455ed31 100644 +--- a/drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c ++++ b/drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c +@@ -476,14 +476,14 @@ static const struct intel_ddi_buf_trans icl_combo_phy_ddi_translations_hdmi = { + static const union intel_ddi_buf_trans_entry _ehl_combo_phy_ddi_translations_dp[] = { + /* NT mV Trans mV db */ + { .icl = { 0xA, 0x33, 0x3F, 0x00, 0x00 } }, /* 350 350 0.0 */ +- { .icl = { 0xA, 0x47, 0x36, 0x00, 0x09 } }, /* 350 500 3.1 */ +- { .icl = { 0xC, 0x64, 0x34, 0x00, 0x0B } }, /* 350 700 6.0 */ +- { .icl = { 0x6, 0x7F, 0x30, 0x00, 0x0F } }, /* 350 900 8.2 */ ++ { .icl = { 0xA, 0x47, 0x38, 0x00, 0x07 } }, /* 350 500 3.1 */ ++ { .icl = { 0xC, 0x64, 0x33, 0x00, 0x0C } }, /* 350 700 6.0 */ ++ { .icl = { 0x6, 0x7F, 0x2F, 0x00, 0x10 } }, /* 350 900 8.2 */ + { .icl = { 0xA, 0x46, 0x3F, 0x00, 0x00 } }, /* 500 500 0.0 */ +- { .icl = { 0xC, 0x64, 0x38, 0x00, 0x07 } }, /* 500 700 2.9 */ ++ { .icl = { 0xC, 0x64, 0x37, 0x00, 0x08 } }, /* 500 700 2.9 */ + { .icl = { 0x6, 0x7F, 0x32, 0x00, 0x0D } }, /* 500 900 5.1 */ + { .icl = { 0xC, 0x61, 0x3F, 0x00, 0x00 } }, /* 650 700 0.6 */ +- { .icl = { 0x6, 0x7F, 0x38, 0x00, 0x07 } }, /* 600 900 3.5 */ ++ { .icl = { 0x6, 0x7F, 0x37, 0x00, 0x08 } }, /* 600 900 3.5 */ + { .icl = { 0x6, 0x7F, 0x3F, 0x00, 0x00 } }, /* 900 900 0.0 */ + }; + +diff --git a/drivers/gpu/drm/lima/lima_device.c b/drivers/gpu/drm/lima/lima_device.c +index 65fdca366e41f..36c9905894278 100644 +--- a/drivers/gpu/drm/lima/lima_device.c ++++ b/drivers/gpu/drm/lima/lima_device.c +@@ -357,6 +357,7 @@ int lima_device_init(struct lima_device *ldev) + int err, i; + + dma_set_coherent_mask(ldev->dev, DMA_BIT_MASK(32)); ++ dma_set_max_seg_size(ldev->dev, UINT_MAX); + + err = lima_clk_init(ldev); + if (err) +diff --git a/drivers/gpu/drm/msm/Kconfig b/drivers/gpu/drm/msm/Kconfig +index 3ddf739a6f9b8..c49b239231190 100644 +--- a/drivers/gpu/drm/msm/Kconfig ++++ b/drivers/gpu/drm/msm/Kconfig +@@ -63,6 +63,7 @@ config DRM_MSM_HDMI_HDCP + config DRM_MSM_DP + bool "Enable DisplayPort support in MSM DRM driver" + depends on DRM_MSM ++ select RATIONAL + default y + help + Compile in support for DP driver in MSM DRM driver. DP external +diff --git a/drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c b/drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c +index ad247c06e198f..93d916858d5ad 100644 +--- a/drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c ++++ b/drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c +@@ -73,8 +73,8 @@ static int _dpu_danger_signal_status(struct seq_file *s, + &status); + } else { + seq_puts(s, "\nSafe signal status:\n"); +- if (kms->hw_mdp->ops.get_danger_status) +- kms->hw_mdp->ops.get_danger_status(kms->hw_mdp, ++ if (kms->hw_mdp->ops.get_safe_status) ++ kms->hw_mdp->ops.get_safe_status(kms->hw_mdp, + &status); + } + pm_runtime_put_sync(&kms->pdev->dev); +diff --git a/drivers/gpu/drm/msm/dsi/dsi.c b/drivers/gpu/drm/msm/dsi/dsi.c +index 75ae3008b68f4..fc280cc434943 100644 +--- a/drivers/gpu/drm/msm/dsi/dsi.c ++++ b/drivers/gpu/drm/msm/dsi/dsi.c +@@ -215,9 +215,13 @@ int msm_dsi_modeset_init(struct msm_dsi *msm_dsi, struct drm_device *dev, + goto fail; + } + +- if (!msm_dsi_manager_validate_current_config(msm_dsi->id)) { +- ret = -EINVAL; +- goto fail; ++ if (msm_dsi_is_bonded_dsi(msm_dsi) && ++ !msm_dsi_is_master_dsi(msm_dsi)) { ++ /* ++ * Do not return an eror here, ++ * Just skip creating encoder/connector for the slave-DSI. ++ */ ++ return 0; + } + + msm_dsi->encoder = encoder; +diff --git a/drivers/gpu/drm/msm/dsi/dsi.h b/drivers/gpu/drm/msm/dsi/dsi.h +index 569c8ff062ba4..a63666e59d19e 100644 +--- a/drivers/gpu/drm/msm/dsi/dsi.h ++++ b/drivers/gpu/drm/msm/dsi/dsi.h +@@ -82,7 +82,6 @@ int msm_dsi_manager_cmd_xfer(int id, const struct mipi_dsi_msg *msg); + bool msm_dsi_manager_cmd_xfer_trigger(int id, u32 dma_base, u32 len); + int msm_dsi_manager_register(struct msm_dsi *msm_dsi); + void msm_dsi_manager_unregister(struct msm_dsi *msm_dsi); +-bool msm_dsi_manager_validate_current_config(u8 id); + void msm_dsi_manager_tpg_enable(void); + + /* msm dsi */ +diff --git a/drivers/gpu/drm/msm/dsi/dsi_manager.c b/drivers/gpu/drm/msm/dsi/dsi_manager.c +index fb4ccffdcfe13..fa4c396df6a92 100644 +--- a/drivers/gpu/drm/msm/dsi/dsi_manager.c ++++ b/drivers/gpu/drm/msm/dsi/dsi_manager.c +@@ -647,23 +647,6 @@ fail: + return ERR_PTR(ret); + } + +-bool msm_dsi_manager_validate_current_config(u8 id) +-{ +- bool is_bonded_dsi = IS_BONDED_DSI(); +- +- /* +- * For bonded DSI, we only have one drm panel. For this +- * use case, we register only one bridge/connector. +- * Skip bridge/connector initialisation if it is +- * slave-DSI for bonded DSI configuration. +- */ +- if (is_bonded_dsi && !IS_MASTER_DSI_LINK(id)) { +- DBG("Skip bridge registration for slave DSI->id: %d\n", id); +- return false; +- } +- return true; +-} +- + /* initialize bridge */ + struct drm_bridge *msm_dsi_manager_bridge_init(u8 id) + { +diff --git a/drivers/gpu/drm/msm/msm_gem_submit.c b/drivers/gpu/drm/msm/msm_gem_submit.c +index d9aef97eb93ad..7fb7ff043bcd7 100644 +--- a/drivers/gpu/drm/msm/msm_gem_submit.c ++++ b/drivers/gpu/drm/msm/msm_gem_submit.c +@@ -887,7 +887,7 @@ int msm_ioctl_gem_submit(struct drm_device *dev, void *data, + * to the underlying fence. + */ + submit->fence_id = idr_alloc_cyclic(&queue->fence_idr, +- submit->user_fence, 0, INT_MAX, GFP_KERNEL); ++ submit->user_fence, 1, INT_MAX, GFP_KERNEL); + if (submit->fence_id < 0) { + ret = submit->fence_id = 0; + submit->fence_id = 0; +diff --git a/drivers/gpu/drm/nouveau/dispnv04/disp.c b/drivers/gpu/drm/nouveau/dispnv04/disp.c +index 7739f46470d3e..99fee4d8cd318 100644 +--- a/drivers/gpu/drm/nouveau/dispnv04/disp.c ++++ b/drivers/gpu/drm/nouveau/dispnv04/disp.c +@@ -205,7 +205,7 @@ nv04_display_destroy(struct drm_device *dev) + nvif_notify_dtor(&disp->flip); + + nouveau_display(dev)->priv = NULL; +- kfree(disp); ++ vfree(disp); + + nvif_object_unmap(&drm->client.device.object); + } +@@ -223,7 +223,7 @@ nv04_display_create(struct drm_device *dev) + struct nv04_display *disp; + int i, ret; + +- disp = kzalloc(sizeof(*disp), GFP_KERNEL); ++ disp = vzalloc(sizeof(*disp)); + if (!disp) + return -ENOMEM; + +diff --git a/drivers/gpu/drm/nouveau/nvkm/subdev/pmu/base.c b/drivers/gpu/drm/nouveau/nvkm/subdev/pmu/base.c +index 24382875fb4f3..455e95a89259f 100644 +--- a/drivers/gpu/drm/nouveau/nvkm/subdev/pmu/base.c ++++ b/drivers/gpu/drm/nouveau/nvkm/subdev/pmu/base.c +@@ -94,20 +94,13 @@ nvkm_pmu_fini(struct nvkm_subdev *subdev, bool suspend) + return 0; + } + +-static int ++static void + nvkm_pmu_reset(struct nvkm_pmu *pmu) + { + struct nvkm_device *device = pmu->subdev.device; + + if (!pmu->func->enabled(pmu)) +- return 0; +- +- /* Inhibit interrupts, and wait for idle. */ +- nvkm_wr32(device, 0x10a014, 0x0000ffff); +- nvkm_msec(device, 2000, +- if (!nvkm_rd32(device, 0x10a04c)) +- break; +- ); ++ return; + + /* Reset. */ + if (pmu->func->reset) +@@ -118,25 +111,37 @@ nvkm_pmu_reset(struct nvkm_pmu *pmu) + if (!(nvkm_rd32(device, 0x10a10c) & 0x00000006)) + break; + ); +- +- return 0; + } + + static int + nvkm_pmu_preinit(struct nvkm_subdev *subdev) + { + struct nvkm_pmu *pmu = nvkm_pmu(subdev); +- return nvkm_pmu_reset(pmu); ++ nvkm_pmu_reset(pmu); ++ return 0; + } + + static int + nvkm_pmu_init(struct nvkm_subdev *subdev) + { + struct nvkm_pmu *pmu = nvkm_pmu(subdev); +- int ret = nvkm_pmu_reset(pmu); +- if (ret == 0 && pmu->func->init) +- ret = pmu->func->init(pmu); +- return ret; ++ struct nvkm_device *device = pmu->subdev.device; ++ ++ if (!pmu->func->init) ++ return 0; ++ ++ if (pmu->func->enabled(pmu)) { ++ /* Inhibit interrupts, and wait for idle. */ ++ nvkm_wr32(device, 0x10a014, 0x0000ffff); ++ nvkm_msec(device, 2000, ++ if (!nvkm_rd32(device, 0x10a04c)) ++ break; ++ ); ++ ++ nvkm_pmu_reset(pmu); ++ } ++ ++ return pmu->func->init(pmu); + } + + static void * +diff --git a/drivers/gpu/drm/panel/panel-feiyang-fy07024di26a30d.c b/drivers/gpu/drm/panel/panel-feiyang-fy07024di26a30d.c +index 581661b506f81..f9c1f7bc8218c 100644 +--- a/drivers/gpu/drm/panel/panel-feiyang-fy07024di26a30d.c ++++ b/drivers/gpu/drm/panel/panel-feiyang-fy07024di26a30d.c +@@ -227,7 +227,13 @@ static int feiyang_dsi_probe(struct mipi_dsi_device *dsi) + dsi->format = MIPI_DSI_FMT_RGB888; + dsi->lanes = 4; + +- return mipi_dsi_attach(dsi); ++ ret = mipi_dsi_attach(dsi); ++ if (ret < 0) { ++ drm_panel_remove(&ctx->panel); ++ return ret; ++ } ++ ++ return 0; + } + + static int feiyang_dsi_remove(struct mipi_dsi_device *dsi) +diff --git a/drivers/gpu/drm/panel/panel-innolux-p079zca.c b/drivers/gpu/drm/panel/panel-innolux-p079zca.c +index aea3162253914..f194b62e290ca 100644 +--- a/drivers/gpu/drm/panel/panel-innolux-p079zca.c ++++ b/drivers/gpu/drm/panel/panel-innolux-p079zca.c +@@ -484,6 +484,7 @@ static void innolux_panel_del(struct innolux_panel *innolux) + static int innolux_panel_probe(struct mipi_dsi_device *dsi) + { + const struct panel_desc *desc; ++ struct innolux_panel *innolux; + int err; + + desc = of_device_get_match_data(&dsi->dev); +@@ -495,7 +496,14 @@ static int innolux_panel_probe(struct mipi_dsi_device *dsi) + if (err < 0) + return err; + +- return mipi_dsi_attach(dsi); ++ err = mipi_dsi_attach(dsi); ++ if (err < 0) { ++ innolux = mipi_dsi_get_drvdata(dsi); ++ innolux_panel_del(innolux); ++ return err; ++ } ++ ++ return 0; + } + + static int innolux_panel_remove(struct mipi_dsi_device *dsi) +diff --git a/drivers/gpu/drm/panel/panel-jdi-lt070me05000.c b/drivers/gpu/drm/panel/panel-jdi-lt070me05000.c +index 733010b5e4f53..3c86ad262d5e0 100644 +--- a/drivers/gpu/drm/panel/panel-jdi-lt070me05000.c ++++ b/drivers/gpu/drm/panel/panel-jdi-lt070me05000.c +@@ -473,7 +473,13 @@ static int jdi_panel_probe(struct mipi_dsi_device *dsi) + if (ret < 0) + return ret; + +- return mipi_dsi_attach(dsi); ++ ret = mipi_dsi_attach(dsi); ++ if (ret < 0) { ++ jdi_panel_del(jdi); ++ return ret; ++ } ++ ++ return 0; + } + + static int jdi_panel_remove(struct mipi_dsi_device *dsi) +diff --git a/drivers/gpu/drm/panel/panel-kingdisplay-kd097d04.c b/drivers/gpu/drm/panel/panel-kingdisplay-kd097d04.c +index 86e4213e8bb13..daccb1fd5fdad 100644 +--- a/drivers/gpu/drm/panel/panel-kingdisplay-kd097d04.c ++++ b/drivers/gpu/drm/panel/panel-kingdisplay-kd097d04.c +@@ -406,7 +406,13 @@ static int kingdisplay_panel_probe(struct mipi_dsi_device *dsi) + if (err < 0) + return err; + +- return mipi_dsi_attach(dsi); ++ err = mipi_dsi_attach(dsi); ++ if (err < 0) { ++ kingdisplay_panel_del(kingdisplay); ++ return err; ++ } ++ ++ return 0; + } + + static int kingdisplay_panel_remove(struct mipi_dsi_device *dsi) +diff --git a/drivers/gpu/drm/panel/panel-novatek-nt36672a.c b/drivers/gpu/drm/panel/panel-novatek-nt36672a.c +index 533cd3934b8b7..839b263fb3c0f 100644 +--- a/drivers/gpu/drm/panel/panel-novatek-nt36672a.c ++++ b/drivers/gpu/drm/panel/panel-novatek-nt36672a.c +@@ -656,7 +656,13 @@ static int nt36672a_panel_probe(struct mipi_dsi_device *dsi) + if (err < 0) + return err; + +- return mipi_dsi_attach(dsi); ++ err = mipi_dsi_attach(dsi); ++ if (err < 0) { ++ drm_panel_remove(&pinfo->base); ++ return err; ++ } ++ ++ return 0; + } + + static int nt36672a_panel_remove(struct mipi_dsi_device *dsi) +diff --git a/drivers/gpu/drm/panel/panel-panasonic-vvx10f034n00.c b/drivers/gpu/drm/panel/panel-panasonic-vvx10f034n00.c +index 3c20beeb17819..3991f5d950af4 100644 +--- a/drivers/gpu/drm/panel/panel-panasonic-vvx10f034n00.c ++++ b/drivers/gpu/drm/panel/panel-panasonic-vvx10f034n00.c +@@ -241,7 +241,13 @@ static int wuxga_nt_panel_probe(struct mipi_dsi_device *dsi) + if (ret < 0) + return ret; + +- return mipi_dsi_attach(dsi); ++ ret = mipi_dsi_attach(dsi); ++ if (ret < 0) { ++ wuxga_nt_panel_del(wuxga_nt); ++ return ret; ++ } ++ ++ return 0; + } + + static int wuxga_nt_panel_remove(struct mipi_dsi_device *dsi) +diff --git a/drivers/gpu/drm/panel/panel-ronbo-rb070d30.c b/drivers/gpu/drm/panel/panel-ronbo-rb070d30.c +index a3782830ae3c4..1fb579a574d9f 100644 +--- a/drivers/gpu/drm/panel/panel-ronbo-rb070d30.c ++++ b/drivers/gpu/drm/panel/panel-ronbo-rb070d30.c +@@ -199,7 +199,13 @@ static int rb070d30_panel_dsi_probe(struct mipi_dsi_device *dsi) + dsi->format = MIPI_DSI_FMT_RGB888; + dsi->lanes = 4; + +- return mipi_dsi_attach(dsi); ++ ret = mipi_dsi_attach(dsi); ++ if (ret < 0) { ++ drm_panel_remove(&ctx->panel); ++ return ret; ++ } ++ ++ return 0; + } + + static int rb070d30_panel_dsi_remove(struct mipi_dsi_device *dsi) +diff --git a/drivers/gpu/drm/panel/panel-samsung-s6e88a0-ams452ef01.c b/drivers/gpu/drm/panel/panel-samsung-s6e88a0-ams452ef01.c +index ea63799ff2a1e..29fde3823212b 100644 +--- a/drivers/gpu/drm/panel/panel-samsung-s6e88a0-ams452ef01.c ++++ b/drivers/gpu/drm/panel/panel-samsung-s6e88a0-ams452ef01.c +@@ -247,6 +247,7 @@ static int s6e88a0_ams452ef01_probe(struct mipi_dsi_device *dsi) + ret = mipi_dsi_attach(dsi); + if (ret < 0) { + dev_err(dev, "Failed to attach to DSI host: %d\n", ret); ++ drm_panel_remove(&ctx->panel); + return ret; + } + +diff --git a/drivers/gpu/drm/panel/panel-samsung-sofef00.c b/drivers/gpu/drm/panel/panel-samsung-sofef00.c +index 8cb1853574bb8..6d107e14fcc55 100644 +--- a/drivers/gpu/drm/panel/panel-samsung-sofef00.c ++++ b/drivers/gpu/drm/panel/panel-samsung-sofef00.c +@@ -302,6 +302,7 @@ static int sofef00_panel_probe(struct mipi_dsi_device *dsi) + ret = mipi_dsi_attach(dsi); + if (ret < 0) { + dev_err(dev, "Failed to attach to DSI host: %d\n", ret); ++ drm_panel_remove(&ctx->panel); + return ret; + } + +diff --git a/drivers/gpu/drm/panel/panel-sharp-ls043t1le01.c b/drivers/gpu/drm/panel/panel-sharp-ls043t1le01.c +index b937e24dac8e0..25829a0a8e801 100644 +--- a/drivers/gpu/drm/panel/panel-sharp-ls043t1le01.c ++++ b/drivers/gpu/drm/panel/panel-sharp-ls043t1le01.c +@@ -296,7 +296,13 @@ static int sharp_nt_panel_probe(struct mipi_dsi_device *dsi) + if (ret < 0) + return ret; + +- return mipi_dsi_attach(dsi); ++ ret = mipi_dsi_attach(dsi); ++ if (ret < 0) { ++ sharp_nt_panel_del(sharp_nt); ++ return ret; ++ } ++ ++ return 0; + } + + static int sharp_nt_panel_remove(struct mipi_dsi_device *dsi) +diff --git a/drivers/gpu/drm/radeon/radeon_kms.c b/drivers/gpu/drm/radeon/radeon_kms.c +index 482fb0ae6cb5d..0e14907f2043e 100644 +--- a/drivers/gpu/drm/radeon/radeon_kms.c ++++ b/drivers/gpu/drm/radeon/radeon_kms.c +@@ -648,6 +648,8 @@ void radeon_driver_lastclose_kms(struct drm_device *dev) + int radeon_driver_open_kms(struct drm_device *dev, struct drm_file *file_priv) + { + struct radeon_device *rdev = dev->dev_private; ++ struct radeon_fpriv *fpriv; ++ struct radeon_vm *vm; + int r; + + file_priv->driver_priv = NULL; +@@ -660,48 +662,52 @@ int radeon_driver_open_kms(struct drm_device *dev, struct drm_file *file_priv) + + /* new gpu have virtual address space support */ + if (rdev->family >= CHIP_CAYMAN) { +- struct radeon_fpriv *fpriv; +- struct radeon_vm *vm; + + fpriv = kzalloc(sizeof(*fpriv), GFP_KERNEL); + if (unlikely(!fpriv)) { + r = -ENOMEM; +- goto out_suspend; ++ goto err_suspend; + } + + if (rdev->accel_working) { + vm = &fpriv->vm; + r = radeon_vm_init(rdev, vm); +- if (r) { +- kfree(fpriv); +- goto out_suspend; +- } ++ if (r) ++ goto err_fpriv; + + r = radeon_bo_reserve(rdev->ring_tmp_bo.bo, false); +- if (r) { +- radeon_vm_fini(rdev, vm); +- kfree(fpriv); +- goto out_suspend; +- } ++ if (r) ++ goto err_vm_fini; + + /* map the ib pool buffer read only into + * virtual address space */ + vm->ib_bo_va = radeon_vm_bo_add(rdev, vm, + rdev->ring_tmp_bo.bo); ++ if (!vm->ib_bo_va) { ++ r = -ENOMEM; ++ goto err_vm_fini; ++ } ++ + r = radeon_vm_bo_set_addr(rdev, vm->ib_bo_va, + RADEON_VA_IB_OFFSET, + RADEON_VM_PAGE_READABLE | + RADEON_VM_PAGE_SNOOPED); +- if (r) { +- radeon_vm_fini(rdev, vm); +- kfree(fpriv); +- goto out_suspend; +- } ++ if (r) ++ goto err_vm_fini; + } + file_priv->driver_priv = fpriv; + } + +-out_suspend: ++ pm_runtime_mark_last_busy(dev->dev); ++ pm_runtime_put_autosuspend(dev->dev); ++ return 0; ++ ++err_vm_fini: ++ radeon_vm_fini(rdev, vm); ++err_fpriv: ++ kfree(fpriv); ++ ++err_suspend: + pm_runtime_mark_last_busy(dev->dev); + pm_runtime_put_autosuspend(dev->dev); + return r; +diff --git a/drivers/gpu/drm/rcar-du/rcar_du_crtc.c b/drivers/gpu/drm/rcar-du/rcar_du_crtc.c +index ea7e39d035457..ee7e375ee6724 100644 +--- a/drivers/gpu/drm/rcar-du/rcar_du_crtc.c ++++ b/drivers/gpu/drm/rcar-du/rcar_du_crtc.c +@@ -215,6 +215,7 @@ static void rcar_du_crtc_set_display_timing(struct rcar_du_crtc *rcrtc) + const struct drm_display_mode *mode = &rcrtc->crtc.state->adjusted_mode; + struct rcar_du_device *rcdu = rcrtc->dev; + unsigned long mode_clock = mode->clock * 1000; ++ unsigned int hdse_offset; + u32 dsmr; + u32 escr; + +@@ -298,10 +299,15 @@ static void rcar_du_crtc_set_display_timing(struct rcar_du_crtc *rcrtc) + | DSMR_DIPM_DISP | DSMR_CSPM; + rcar_du_crtc_write(rcrtc, DSMR, dsmr); + ++ hdse_offset = 19; ++ if (rcrtc->group->cmms_mask & BIT(rcrtc->index % 2)) ++ hdse_offset += 25; ++ + /* Display timings */ +- rcar_du_crtc_write(rcrtc, HDSR, mode->htotal - mode->hsync_start - 19); ++ rcar_du_crtc_write(rcrtc, HDSR, mode->htotal - mode->hsync_start - ++ hdse_offset); + rcar_du_crtc_write(rcrtc, HDER, mode->htotal - mode->hsync_start + +- mode->hdisplay - 19); ++ mode->hdisplay - hdse_offset); + rcar_du_crtc_write(rcrtc, HSWR, mode->hsync_end - + mode->hsync_start - 1); + rcar_du_crtc_write(rcrtc, HCR, mode->htotal - 1); +@@ -836,6 +842,7 @@ rcar_du_crtc_mode_valid(struct drm_crtc *crtc, + struct rcar_du_crtc *rcrtc = to_rcar_crtc(crtc); + struct rcar_du_device *rcdu = rcrtc->dev; + bool interlaced = mode->flags & DRM_MODE_FLAG_INTERLACE; ++ unsigned int min_sync_porch; + unsigned int vbp; + + if (interlaced && !rcar_du_has(rcdu, RCAR_DU_FEATURE_INTERLACED)) +@@ -843,9 +850,14 @@ rcar_du_crtc_mode_valid(struct drm_crtc *crtc, + + /* + * The hardware requires a minimum combined horizontal sync and back +- * porch of 20 pixels and a minimum vertical back porch of 3 lines. ++ * porch of 20 pixels (when CMM isn't used) or 45 pixels (when CMM is ++ * used), and a minimum vertical back porch of 3 lines. + */ +- if (mode->htotal - mode->hsync_start < 20) ++ min_sync_porch = 20; ++ if (rcrtc->group->cmms_mask & BIT(rcrtc->index % 2)) ++ min_sync_porch += 25; ++ ++ if (mode->htotal - mode->hsync_start < min_sync_porch) + return MODE_HBLANK_NARROW; + + vbp = (mode->vtotal - mode->vsync_end) / (interlaced ? 2 : 1); +diff --git a/drivers/gpu/drm/rockchip/dw-mipi-dsi-rockchip.c b/drivers/gpu/drm/rockchip/dw-mipi-dsi-rockchip.c +index a2262bee5aa47..59c3d8ef6bf9a 100644 +--- a/drivers/gpu/drm/rockchip/dw-mipi-dsi-rockchip.c ++++ b/drivers/gpu/drm/rockchip/dw-mipi-dsi-rockchip.c +@@ -268,6 +268,8 @@ struct dw_mipi_dsi_rockchip { + struct dw_mipi_dsi *dmd; + const struct rockchip_dw_dsi_chip_data *cdata; + struct dw_mipi_dsi_plat_data pdata; ++ ++ bool dsi_bound; + }; + + struct dphy_pll_parameter_map { +@@ -773,10 +775,6 @@ static void dw_mipi_dsi_encoder_enable(struct drm_encoder *encoder) + if (mux < 0) + return; + +- pm_runtime_get_sync(dsi->dev); +- if (dsi->slave) +- pm_runtime_get_sync(dsi->slave->dev); +- + /* + * For the RK3399, the clk of grf must be enabled before writing grf + * register. And for RK3288 or other soc, this grf_clk must be NULL, +@@ -795,20 +793,10 @@ static void dw_mipi_dsi_encoder_enable(struct drm_encoder *encoder) + clk_disable_unprepare(dsi->grf_clk); + } + +-static void dw_mipi_dsi_encoder_disable(struct drm_encoder *encoder) +-{ +- struct dw_mipi_dsi_rockchip *dsi = to_dsi(encoder); +- +- if (dsi->slave) +- pm_runtime_put(dsi->slave->dev); +- pm_runtime_put(dsi->dev); +-} +- + static const struct drm_encoder_helper_funcs + dw_mipi_dsi_encoder_helper_funcs = { + .atomic_check = dw_mipi_dsi_encoder_atomic_check, + .enable = dw_mipi_dsi_encoder_enable, +- .disable = dw_mipi_dsi_encoder_disable, + }; + + static int rockchip_dsi_drm_create_encoder(struct dw_mipi_dsi_rockchip *dsi, +@@ -938,10 +926,14 @@ static int dw_mipi_dsi_rockchip_bind(struct device *dev, + put_device(second); + } + ++ pm_runtime_get_sync(dsi->dev); ++ if (dsi->slave) ++ pm_runtime_get_sync(dsi->slave->dev); ++ + ret = clk_prepare_enable(dsi->pllref_clk); + if (ret) { + DRM_DEV_ERROR(dev, "Failed to enable pllref_clk: %d\n", ret); +- return ret; ++ goto out_pm_runtime; + } + + /* +@@ -953,7 +945,7 @@ static int dw_mipi_dsi_rockchip_bind(struct device *dev, + ret = clk_prepare_enable(dsi->grf_clk); + if (ret) { + DRM_DEV_ERROR(dsi->dev, "Failed to enable grf_clk: %d\n", ret); +- return ret; ++ goto out_pll_clk; + } + + dw_mipi_dsi_rockchip_config(dsi); +@@ -965,16 +957,27 @@ static int dw_mipi_dsi_rockchip_bind(struct device *dev, + ret = rockchip_dsi_drm_create_encoder(dsi, drm_dev); + if (ret) { + DRM_DEV_ERROR(dev, "Failed to create drm encoder\n"); +- return ret; ++ goto out_pll_clk; + } + + ret = dw_mipi_dsi_bind(dsi->dmd, &dsi->encoder); + if (ret) { + DRM_DEV_ERROR(dev, "Failed to bind: %d\n", ret); +- return ret; ++ goto out_pll_clk; + } + ++ dsi->dsi_bound = true; ++ + return 0; ++ ++out_pll_clk: ++ clk_disable_unprepare(dsi->pllref_clk); ++out_pm_runtime: ++ pm_runtime_put(dsi->dev); ++ if (dsi->slave) ++ pm_runtime_put(dsi->slave->dev); ++ ++ return ret; + } + + static void dw_mipi_dsi_rockchip_unbind(struct device *dev, +@@ -986,9 +989,15 @@ static void dw_mipi_dsi_rockchip_unbind(struct device *dev, + if (dsi->is_slave) + return; + ++ dsi->dsi_bound = false; ++ + dw_mipi_dsi_unbind(dsi->dmd); + + clk_disable_unprepare(dsi->pllref_clk); ++ ++ pm_runtime_put(dsi->dev); ++ if (dsi->slave) ++ pm_runtime_put(dsi->slave->dev); + } + + static const struct component_ops dw_mipi_dsi_rockchip_ops = { +@@ -1276,6 +1285,36 @@ static const struct phy_ops dw_mipi_dsi_dphy_ops = { + .exit = dw_mipi_dsi_dphy_exit, + }; + ++static int __maybe_unused dw_mipi_dsi_rockchip_resume(struct device *dev) ++{ ++ struct dw_mipi_dsi_rockchip *dsi = dev_get_drvdata(dev); ++ int ret; ++ ++ /* ++ * Re-configure DSI state, if we were previously initialized. We need ++ * to do this before rockchip_drm_drv tries to re-enable() any panels. ++ */ ++ if (dsi->dsi_bound) { ++ ret = clk_prepare_enable(dsi->grf_clk); ++ if (ret) { ++ DRM_DEV_ERROR(dsi->dev, "Failed to enable grf_clk: %d\n", ret); ++ return ret; ++ } ++ ++ dw_mipi_dsi_rockchip_config(dsi); ++ if (dsi->slave) ++ dw_mipi_dsi_rockchip_config(dsi->slave); ++ ++ clk_disable_unprepare(dsi->grf_clk); ++ } ++ ++ return 0; ++} ++ ++static const struct dev_pm_ops dw_mipi_dsi_rockchip_pm_ops = { ++ SET_LATE_SYSTEM_SLEEP_PM_OPS(NULL, dw_mipi_dsi_rockchip_resume) ++}; ++ + static int dw_mipi_dsi_rockchip_probe(struct platform_device *pdev) + { + struct device *dev = &pdev->dev; +@@ -1397,14 +1436,10 @@ static int dw_mipi_dsi_rockchip_probe(struct platform_device *pdev) + if (ret != -EPROBE_DEFER) + DRM_DEV_ERROR(dev, + "Failed to probe dw_mipi_dsi: %d\n", ret); +- goto err_clkdisable; ++ return ret; + } + + return 0; +- +-err_clkdisable: +- clk_disable_unprepare(dsi->pllref_clk); +- return ret; + } + + static int dw_mipi_dsi_rockchip_remove(struct platform_device *pdev) +@@ -1593,6 +1628,7 @@ struct platform_driver dw_mipi_dsi_rockchip_driver = { + .remove = dw_mipi_dsi_rockchip_remove, + .driver = { + .of_match_table = dw_mipi_dsi_rockchip_dt_ids, ++ .pm = &dw_mipi_dsi_rockchip_pm_ops, + .name = "dw-mipi-dsi-rockchip", + }, + }; +diff --git a/drivers/gpu/drm/tegra/drm.c b/drivers/gpu/drm/tegra/drm.c +index 8d37d6b00562a..611cd8dad46ed 100644 +--- a/drivers/gpu/drm/tegra/drm.c ++++ b/drivers/gpu/drm/tegra/drm.c +@@ -21,6 +21,10 @@ + #include <drm/drm_prime.h> + #include <drm/drm_vblank.h> + ++#if IS_ENABLED(CONFIG_ARM_DMA_USE_IOMMU) ++#include <asm/dma-iommu.h> ++#endif ++ + #include "dc.h" + #include "drm.h" + #include "gem.h" +@@ -936,6 +940,17 @@ int host1x_client_iommu_attach(struct host1x_client *client) + struct iommu_group *group = NULL; + int err; + ++#if IS_ENABLED(CONFIG_ARM_DMA_USE_IOMMU) ++ if (client->dev->archdata.mapping) { ++ struct dma_iommu_mapping *mapping = ++ to_dma_iommu_mapping(client->dev); ++ arm_iommu_detach_device(client->dev); ++ arm_iommu_release_mapping(mapping); ++ ++ domain = iommu_get_domain_for_dev(client->dev); ++ } ++#endif ++ + /* + * If the host1x client is already attached to an IOMMU domain that is + * not the shared IOMMU domain, don't try to attach it to a different +diff --git a/drivers/gpu/drm/tegra/gr2d.c b/drivers/gpu/drm/tegra/gr2d.c +index de288cba39055..ba3722f1b8651 100644 +--- a/drivers/gpu/drm/tegra/gr2d.c ++++ b/drivers/gpu/drm/tegra/gr2d.c +@@ -4,9 +4,11 @@ + */ + + #include <linux/clk.h> ++#include <linux/delay.h> + #include <linux/iommu.h> + #include <linux/module.h> + #include <linux/of_device.h> ++#include <linux/reset.h> + + #include "drm.h" + #include "gem.h" +@@ -19,6 +21,7 @@ struct gr2d_soc { + struct gr2d { + struct tegra_drm_client client; + struct host1x_channel *channel; ++ struct reset_control *rst; + struct clk *clk; + + const struct gr2d_soc *soc; +@@ -208,6 +211,12 @@ static int gr2d_probe(struct platform_device *pdev) + if (!syncpts) + return -ENOMEM; + ++ gr2d->rst = devm_reset_control_get(dev, NULL); ++ if (IS_ERR(gr2d->rst)) { ++ dev_err(dev, "cannot get reset\n"); ++ return PTR_ERR(gr2d->rst); ++ } ++ + gr2d->clk = devm_clk_get(dev, NULL); + if (IS_ERR(gr2d->clk)) { + dev_err(dev, "cannot get clock\n"); +@@ -220,6 +229,14 @@ static int gr2d_probe(struct platform_device *pdev) + return err; + } + ++ usleep_range(2000, 4000); ++ ++ err = reset_control_deassert(gr2d->rst); ++ if (err < 0) { ++ dev_err(dev, "failed to deassert reset: %d\n", err); ++ goto disable_clk; ++ } ++ + INIT_LIST_HEAD(&gr2d->client.base.list); + gr2d->client.base.ops = &gr2d_client_ops; + gr2d->client.base.dev = dev; +@@ -234,8 +251,7 @@ static int gr2d_probe(struct platform_device *pdev) + err = host1x_client_register(&gr2d->client.base); + if (err < 0) { + dev_err(dev, "failed to register host1x client: %d\n", err); +- clk_disable_unprepare(gr2d->clk); +- return err; ++ goto assert_rst; + } + + /* initialize address register map */ +@@ -245,6 +261,13 @@ static int gr2d_probe(struct platform_device *pdev) + platform_set_drvdata(pdev, gr2d); + + return 0; ++ ++assert_rst: ++ (void)reset_control_assert(gr2d->rst); ++disable_clk: ++ clk_disable_unprepare(gr2d->clk); ++ ++ return err; + } + + static int gr2d_remove(struct platform_device *pdev) +@@ -259,6 +282,12 @@ static int gr2d_remove(struct platform_device *pdev) + return err; + } + ++ err = reset_control_assert(gr2d->rst); ++ if (err < 0) ++ dev_err(&pdev->dev, "failed to assert reset: %d\n", err); ++ ++ usleep_range(2000, 4000); ++ + clk_disable_unprepare(gr2d->clk); + + return 0; +diff --git a/drivers/gpu/drm/tegra/submit.c b/drivers/gpu/drm/tegra/submit.c +index 776f825df52fa..aba9d0c9d9031 100644 +--- a/drivers/gpu/drm/tegra/submit.c ++++ b/drivers/gpu/drm/tegra/submit.c +@@ -475,8 +475,10 @@ static void release_job(struct host1x_job *job) + kfree(job_data->used_mappings); + kfree(job_data); + +- if (pm_runtime_enabled(client->base.dev)) ++ if (pm_runtime_enabled(client->base.dev)) { ++ pm_runtime_mark_last_busy(client->base.dev); + pm_runtime_put_autosuspend(client->base.dev); ++ } + } + + int tegra_drm_ioctl_channel_submit(struct drm_device *drm, void *data, +diff --git a/drivers/gpu/drm/tegra/vic.c b/drivers/gpu/drm/tegra/vic.c +index c02010ff2b7f2..da4af53719917 100644 +--- a/drivers/gpu/drm/tegra/vic.c ++++ b/drivers/gpu/drm/tegra/vic.c +@@ -5,6 +5,7 @@ + + #include <linux/clk.h> + #include <linux/delay.h> ++#include <linux/dma-mapping.h> + #include <linux/host1x.h> + #include <linux/iommu.h> + #include <linux/module.h> +@@ -232,10 +233,8 @@ static int vic_load_firmware(struct vic *vic) + + if (!client->group) { + virt = dma_alloc_coherent(vic->dev, size, &iova, GFP_KERNEL); +- +- err = dma_mapping_error(vic->dev, iova); +- if (err < 0) +- return err; ++ if (!virt) ++ return -ENOMEM; + } else { + virt = tegra_drm_alloc(tegra, size, &iova); + } +diff --git a/drivers/gpu/drm/ttm/ttm_bo.c b/drivers/gpu/drm/ttm/ttm_bo.c +index bb9e02c31946e..900edaf5d68ee 100644 +--- a/drivers/gpu/drm/ttm/ttm_bo.c ++++ b/drivers/gpu/drm/ttm/ttm_bo.c +@@ -724,6 +724,8 @@ int ttm_mem_evict_first(struct ttm_device *bdev, + ret = ttm_bo_evict(bo, ctx); + if (locked) + ttm_bo_unreserve(bo); ++ else ++ ttm_bo_move_to_lru_tail_unlocked(bo); + + ttm_bo_put(bo); + return ret; +diff --git a/drivers/gpu/drm/vboxvideo/vbox_main.c b/drivers/gpu/drm/vboxvideo/vbox_main.c +index f28779715ccda..c9e8b3a63c621 100644 +--- a/drivers/gpu/drm/vboxvideo/vbox_main.c ++++ b/drivers/gpu/drm/vboxvideo/vbox_main.c +@@ -127,8 +127,8 @@ int vbox_hw_init(struct vbox_private *vbox) + /* Create guest-heap mem-pool use 2^4 = 16 byte chunks */ + vbox->guest_pool = devm_gen_pool_create(vbox->ddev.dev, 4, -1, + "vboxvideo-accel"); +- if (!vbox->guest_pool) +- return -ENOMEM; ++ if (IS_ERR(vbox->guest_pool)) ++ return PTR_ERR(vbox->guest_pool); + + ret = gen_pool_add_virt(vbox->guest_pool, + (unsigned long)vbox->guest_heap, +diff --git a/drivers/gpu/drm/vc4/vc4_crtc.c b/drivers/gpu/drm/vc4/vc4_crtc.c +index 18f5009ce90e3..e3ed52d96f423 100644 +--- a/drivers/gpu/drm/vc4/vc4_crtc.c ++++ b/drivers/gpu/drm/vc4/vc4_crtc.c +@@ -32,6 +32,7 @@ + #include <linux/clk.h> + #include <linux/component.h> + #include <linux/of_device.h> ++#include <linux/pm_runtime.h> + + #include <drm/drm_atomic.h> + #include <drm/drm_atomic_helper.h> +@@ -42,6 +43,7 @@ + #include <drm/drm_vblank.h> + + #include "vc4_drv.h" ++#include "vc4_hdmi.h" + #include "vc4_regs.h" + + #define HVS_FIFO_LATENCY_PIX 6 +@@ -496,8 +498,10 @@ int vc4_crtc_disable_at_boot(struct drm_crtc *crtc) + enum vc4_encoder_type encoder_type; + const struct vc4_pv_data *pv_data; + struct drm_encoder *encoder; ++ struct vc4_hdmi *vc4_hdmi; + unsigned encoder_sel; + int channel; ++ int ret; + + if (!(of_device_is_compatible(vc4_crtc->pdev->dev.of_node, + "brcm,bcm2711-pixelvalve2") || +@@ -525,7 +529,20 @@ int vc4_crtc_disable_at_boot(struct drm_crtc *crtc) + if (WARN_ON(!encoder)) + return 0; + +- return vc4_crtc_disable(crtc, encoder, NULL, channel); ++ vc4_hdmi = encoder_to_vc4_hdmi(encoder); ++ ret = pm_runtime_resume_and_get(&vc4_hdmi->pdev->dev); ++ if (ret) ++ return ret; ++ ++ ret = vc4_crtc_disable(crtc, encoder, NULL, channel); ++ if (ret) ++ return ret; ++ ++ ret = pm_runtime_put(&vc4_hdmi->pdev->dev); ++ if (ret) ++ return ret; ++ ++ return 0; + } + + static void vc4_crtc_atomic_disable(struct drm_crtc *crtc, +@@ -691,14 +708,14 @@ static void vc4_crtc_handle_page_flip(struct vc4_crtc *vc4_crtc) + struct drm_crtc *crtc = &vc4_crtc->base; + struct drm_device *dev = crtc->dev; + struct vc4_dev *vc4 = to_vc4_dev(dev); +- struct vc4_crtc_state *vc4_state = to_vc4_crtc_state(crtc->state); +- u32 chan = vc4_state->assigned_channel; ++ u32 chan = vc4_crtc->current_hvs_channel; + unsigned long flags; + + spin_lock_irqsave(&dev->event_lock, flags); ++ spin_lock(&vc4_crtc->irq_lock); + if (vc4_crtc->event && +- (vc4_state->mm.start == HVS_READ(SCALER_DISPLACTX(chan)) || +- vc4_state->feed_txp)) { ++ (vc4_crtc->current_dlist == HVS_READ(SCALER_DISPLACTX(chan)) || ++ vc4_crtc->feeds_txp)) { + drm_crtc_send_vblank_event(crtc, vc4_crtc->event); + vc4_crtc->event = NULL; + drm_crtc_vblank_put(crtc); +@@ -711,6 +728,7 @@ static void vc4_crtc_handle_page_flip(struct vc4_crtc *vc4_crtc) + */ + vc4_hvs_unmask_underrun(dev, chan); + } ++ spin_unlock(&vc4_crtc->irq_lock); + spin_unlock_irqrestore(&dev->event_lock, flags); + } + +@@ -876,7 +894,6 @@ struct drm_crtc_state *vc4_crtc_duplicate_state(struct drm_crtc *crtc) + return NULL; + + old_vc4_state = to_vc4_crtc_state(crtc->state); +- vc4_state->feed_txp = old_vc4_state->feed_txp; + vc4_state->margins = old_vc4_state->margins; + vc4_state->assigned_channel = old_vc4_state->assigned_channel; + +@@ -937,6 +954,7 @@ static const struct drm_crtc_funcs vc4_crtc_funcs = { + static const struct drm_crtc_helper_funcs vc4_crtc_helper_funcs = { + .mode_valid = vc4_crtc_mode_valid, + .atomic_check = vc4_crtc_atomic_check, ++ .atomic_begin = vc4_hvs_atomic_begin, + .atomic_flush = vc4_hvs_atomic_flush, + .atomic_enable = vc4_crtc_atomic_enable, + .atomic_disable = vc4_crtc_atomic_disable, +@@ -1111,6 +1129,7 @@ int vc4_crtc_init(struct drm_device *drm, struct vc4_crtc *vc4_crtc, + return PTR_ERR(primary_plane); + } + ++ spin_lock_init(&vc4_crtc->irq_lock); + drm_crtc_init_with_planes(drm, crtc, primary_plane, NULL, + crtc_funcs, NULL); + drm_crtc_helper_add(crtc, crtc_helper_funcs); +diff --git a/drivers/gpu/drm/vc4/vc4_drv.h b/drivers/gpu/drm/vc4/vc4_drv.h +index ef73e0aaf7261..4b550ebd9572d 100644 +--- a/drivers/gpu/drm/vc4/vc4_drv.h ++++ b/drivers/gpu/drm/vc4/vc4_drv.h +@@ -495,6 +495,33 @@ struct vc4_crtc { + struct drm_pending_vblank_event *event; + + struct debugfs_regset32 regset; ++ ++ /** ++ * @feeds_txp: True if the CRTC feeds our writeback controller. ++ */ ++ bool feeds_txp; ++ ++ /** ++ * @irq_lock: Spinlock protecting the resources shared between ++ * the atomic code and our vblank handler. ++ */ ++ spinlock_t irq_lock; ++ ++ /** ++ * @current_dlist: Start offset of the display list currently ++ * set in the HVS for that CRTC. Protected by @irq_lock, and ++ * copied in vc4_hvs_update_dlist() for the CRTC interrupt ++ * handler to have access to that value. ++ */ ++ unsigned int current_dlist; ++ ++ /** ++ * @current_hvs_channel: HVS channel currently assigned to the ++ * CRTC. Protected by @irq_lock, and copied in ++ * vc4_hvs_atomic_begin() for the CRTC interrupt handler to have ++ * access to that value. ++ */ ++ unsigned int current_hvs_channel; + }; + + static inline struct vc4_crtc * +@@ -521,7 +548,6 @@ struct vc4_crtc_state { + struct drm_crtc_state base; + /* Dlist area for this CRTC configuration. */ + struct drm_mm_node mm; +- bool feed_txp; + bool txp_armed; + unsigned int assigned_channel; + +@@ -908,6 +934,7 @@ extern struct platform_driver vc4_hvs_driver; + void vc4_hvs_stop_channel(struct drm_device *dev, unsigned int output); + int vc4_hvs_get_fifo_from_output(struct drm_device *dev, unsigned int output); + int vc4_hvs_atomic_check(struct drm_crtc *crtc, struct drm_atomic_state *state); ++void vc4_hvs_atomic_begin(struct drm_crtc *crtc, struct drm_atomic_state *state); + void vc4_hvs_atomic_enable(struct drm_crtc *crtc, struct drm_atomic_state *state); + void vc4_hvs_atomic_disable(struct drm_crtc *crtc, struct drm_atomic_state *state); + void vc4_hvs_atomic_flush(struct drm_crtc *crtc, struct drm_atomic_state *state); +diff --git a/drivers/gpu/drm/vc4/vc4_hdmi.c b/drivers/gpu/drm/vc4/vc4_hdmi.c +index ed8a4b7f8b6e2..5580267fb3624 100644 +--- a/drivers/gpu/drm/vc4/vc4_hdmi.c ++++ b/drivers/gpu/drm/vc4/vc4_hdmi.c +@@ -94,6 +94,7 @@ + # define VC4_HD_M_SW_RST BIT(2) + # define VC4_HD_M_ENABLE BIT(0) + ++#define HSM_MIN_CLOCK_FREQ 120000000 + #define CEC_CLOCK_FREQ 40000 + + #define HDMI_14_MAX_TMDS_CLK (340 * 1000 * 1000) +@@ -161,12 +162,16 @@ static void vc4_hdmi_cec_update_clk_div(struct vc4_hdmi *vc4_hdmi) + static void vc4_hdmi_cec_update_clk_div(struct vc4_hdmi *vc4_hdmi) {} + #endif + ++static void vc4_hdmi_enable_scrambling(struct drm_encoder *encoder); ++ + static enum drm_connector_status + vc4_hdmi_connector_detect(struct drm_connector *connector, bool force) + { + struct vc4_hdmi *vc4_hdmi = connector_to_vc4_hdmi(connector); + bool connected = false; + ++ WARN_ON(pm_runtime_resume_and_get(&vc4_hdmi->pdev->dev)); ++ + if (vc4_hdmi->hpd_gpio && + gpiod_get_value_cansleep(vc4_hdmi->hpd_gpio)) { + connected = true; +@@ -187,10 +192,13 @@ vc4_hdmi_connector_detect(struct drm_connector *connector, bool force) + } + } + ++ vc4_hdmi_enable_scrambling(&vc4_hdmi->encoder.base.base); ++ pm_runtime_put(&vc4_hdmi->pdev->dev); + return connector_status_connected; + } + + cec_phys_addr_invalidate(vc4_hdmi->cec_adap); ++ pm_runtime_put(&vc4_hdmi->pdev->dev); + return connector_status_disconnected; + } + +@@ -627,7 +635,6 @@ static void vc4_hdmi_encoder_post_crtc_powerdown(struct drm_encoder *encoder, + vc4_hdmi->variant->phy_disable(vc4_hdmi); + + clk_disable_unprepare(vc4_hdmi->pixel_bvb_clock); +- clk_disable_unprepare(vc4_hdmi->hsm_clock); + clk_disable_unprepare(vc4_hdmi->pixel_clock); + + ret = pm_runtime_put(&vc4_hdmi->pdev->dev); +@@ -893,28 +900,10 @@ static void vc4_hdmi_encoder_pre_crtc_configure(struct drm_encoder *encoder, + conn_state_to_vc4_hdmi_conn_state(conn_state); + struct drm_display_mode *mode = &encoder->crtc->state->adjusted_mode; + struct vc4_hdmi *vc4_hdmi = encoder_to_vc4_hdmi(encoder); +- unsigned long bvb_rate, pixel_rate, hsm_rate; ++ unsigned long pixel_rate = vc4_conn_state->pixel_rate; ++ unsigned long bvb_rate, hsm_rate; + int ret; + +- ret = pm_runtime_resume_and_get(&vc4_hdmi->pdev->dev); +- if (ret < 0) { +- DRM_ERROR("Failed to retain power domain: %d\n", ret); +- return; +- } +- +- pixel_rate = vc4_conn_state->pixel_rate; +- ret = clk_set_rate(vc4_hdmi->pixel_clock, pixel_rate); +- if (ret) { +- DRM_ERROR("Failed to set pixel clock rate: %d\n", ret); +- return; +- } +- +- ret = clk_prepare_enable(vc4_hdmi->pixel_clock); +- if (ret) { +- DRM_ERROR("Failed to turn on pixel clock: %d\n", ret); +- return; +- } +- + /* + * As stated in RPi's vc4 firmware "HDMI state machine (HSM) clock must + * be faster than pixel clock, infinitesimally faster, tested in +@@ -938,13 +927,25 @@ static void vc4_hdmi_encoder_pre_crtc_configure(struct drm_encoder *encoder, + return; + } + +- ret = clk_prepare_enable(vc4_hdmi->hsm_clock); +- if (ret) { +- DRM_ERROR("Failed to turn on HSM clock: %d\n", ret); +- clk_disable_unprepare(vc4_hdmi->pixel_clock); ++ ret = pm_runtime_resume_and_get(&vc4_hdmi->pdev->dev); ++ if (ret < 0) { ++ DRM_ERROR("Failed to retain power domain: %d\n", ret); + return; + } + ++ ret = clk_set_rate(vc4_hdmi->pixel_clock, pixel_rate); ++ if (ret) { ++ DRM_ERROR("Failed to set pixel clock rate: %d\n", ret); ++ goto err_put_runtime_pm; ++ } ++ ++ ret = clk_prepare_enable(vc4_hdmi->pixel_clock); ++ if (ret) { ++ DRM_ERROR("Failed to turn on pixel clock: %d\n", ret); ++ goto err_put_runtime_pm; ++ } ++ ++ + vc4_hdmi_cec_update_clk_div(vc4_hdmi); + + if (pixel_rate > 297000000) +@@ -957,17 +958,13 @@ static void vc4_hdmi_encoder_pre_crtc_configure(struct drm_encoder *encoder, + ret = clk_set_min_rate(vc4_hdmi->pixel_bvb_clock, bvb_rate); + if (ret) { + DRM_ERROR("Failed to set pixel bvb clock rate: %d\n", ret); +- clk_disable_unprepare(vc4_hdmi->hsm_clock); +- clk_disable_unprepare(vc4_hdmi->pixel_clock); +- return; ++ goto err_disable_pixel_clock; + } + + ret = clk_prepare_enable(vc4_hdmi->pixel_bvb_clock); + if (ret) { + DRM_ERROR("Failed to turn on pixel bvb clock: %d\n", ret); +- clk_disable_unprepare(vc4_hdmi->hsm_clock); +- clk_disable_unprepare(vc4_hdmi->pixel_clock); +- return; ++ goto err_disable_pixel_clock; + } + + if (vc4_hdmi->variant->phy_init) +@@ -980,6 +977,15 @@ static void vc4_hdmi_encoder_pre_crtc_configure(struct drm_encoder *encoder, + + if (vc4_hdmi->variant->set_timings) + vc4_hdmi->variant->set_timings(vc4_hdmi, conn_state, mode); ++ ++ return; ++ ++err_disable_pixel_clock: ++ clk_disable_unprepare(vc4_hdmi->pixel_clock); ++err_put_runtime_pm: ++ pm_runtime_put(&vc4_hdmi->pdev->dev); ++ ++ return; + } + + static void vc4_hdmi_encoder_pre_crtc_enable(struct drm_encoder *encoder, +@@ -1729,8 +1735,14 @@ static int vc4_hdmi_cec_adap_enable(struct cec_adapter *adap, bool enable) + struct vc4_hdmi *vc4_hdmi = cec_get_drvdata(adap); + /* clock period in microseconds */ + const u32 usecs = 1000000 / CEC_CLOCK_FREQ; +- u32 val = HDMI_READ(HDMI_CEC_CNTRL_5); ++ u32 val; ++ int ret; ++ ++ ret = pm_runtime_resume_and_get(&vc4_hdmi->pdev->dev); ++ if (ret) ++ return ret; + ++ val = HDMI_READ(HDMI_CEC_CNTRL_5); + val &= ~(VC4_HDMI_CEC_TX_SW_RESET | VC4_HDMI_CEC_RX_SW_RESET | + VC4_HDMI_CEC_CNT_TO_4700_US_MASK | + VC4_HDMI_CEC_CNT_TO_4500_US_MASK); +@@ -1876,6 +1888,8 @@ static int vc4_hdmi_cec_init(struct vc4_hdmi *vc4_hdmi) + if (ret < 0) + goto err_remove_handlers; + ++ pm_runtime_put(&vc4_hdmi->pdev->dev); ++ + return 0; + + err_remove_handlers: +@@ -2098,6 +2112,27 @@ static int vc5_hdmi_init_resources(struct vc4_hdmi *vc4_hdmi) + return 0; + } + ++static int __maybe_unused vc4_hdmi_runtime_suspend(struct device *dev) ++{ ++ struct vc4_hdmi *vc4_hdmi = dev_get_drvdata(dev); ++ ++ clk_disable_unprepare(vc4_hdmi->hsm_clock); ++ ++ return 0; ++} ++ ++static int vc4_hdmi_runtime_resume(struct device *dev) ++{ ++ struct vc4_hdmi *vc4_hdmi = dev_get_drvdata(dev); ++ int ret; ++ ++ ret = clk_prepare_enable(vc4_hdmi->hsm_clock); ++ if (ret) ++ return ret; ++ ++ return 0; ++} ++ + static int vc4_hdmi_bind(struct device *dev, struct device *master, void *data) + { + const struct vc4_hdmi_variant *variant = of_device_get_match_data(dev); +@@ -2161,6 +2196,31 @@ static int vc4_hdmi_bind(struct device *dev, struct device *master, void *data) + vc4_hdmi->disable_4kp60 = true; + } + ++ /* ++ * If we boot without any cable connected to the HDMI connector, ++ * the firmware will skip the HSM initialization and leave it ++ * with a rate of 0, resulting in a bus lockup when we're ++ * accessing the registers even if it's enabled. ++ * ++ * Let's put a sensible default at runtime_resume so that we ++ * don't end up in this situation. ++ */ ++ ret = clk_set_min_rate(vc4_hdmi->hsm_clock, HSM_MIN_CLOCK_FREQ); ++ if (ret) ++ goto err_put_ddc; ++ ++ /* ++ * We need to have the device powered up at this point to call ++ * our reset hook and for the CEC init. ++ */ ++ ret = vc4_hdmi_runtime_resume(dev); ++ if (ret) ++ goto err_put_ddc; ++ ++ pm_runtime_get_noresume(dev); ++ pm_runtime_set_active(dev); ++ pm_runtime_enable(dev); ++ + if (vc4_hdmi->variant->reset) + vc4_hdmi->variant->reset(vc4_hdmi); + +@@ -2172,8 +2232,6 @@ static int vc4_hdmi_bind(struct device *dev, struct device *master, void *data) + clk_prepare_enable(vc4_hdmi->pixel_bvb_clock); + } + +- pm_runtime_enable(dev); +- + drm_simple_encoder_init(drm, encoder, DRM_MODE_ENCODER_TMDS); + drm_encoder_helper_add(encoder, &vc4_hdmi_encoder_helper_funcs); + +@@ -2197,6 +2255,8 @@ static int vc4_hdmi_bind(struct device *dev, struct device *master, void *data) + vc4_hdmi_debugfs_regs, + vc4_hdmi); + ++ pm_runtime_put_sync(dev); ++ + return 0; + + err_free_cec: +@@ -2207,6 +2267,7 @@ err_destroy_conn: + vc4_hdmi_connector_destroy(&vc4_hdmi->connector); + err_destroy_encoder: + drm_encoder_cleanup(encoder); ++ pm_runtime_put_sync(dev); + pm_runtime_disable(dev); + err_put_ddc: + put_device(&vc4_hdmi->ddc->dev); +@@ -2352,11 +2413,18 @@ static const struct of_device_id vc4_hdmi_dt_match[] = { + {} + }; + ++static const struct dev_pm_ops vc4_hdmi_pm_ops = { ++ SET_RUNTIME_PM_OPS(vc4_hdmi_runtime_suspend, ++ vc4_hdmi_runtime_resume, ++ NULL) ++}; ++ + struct platform_driver vc4_hdmi_driver = { + .probe = vc4_hdmi_dev_probe, + .remove = vc4_hdmi_dev_remove, + .driver = { + .name = "vc4_hdmi", + .of_match_table = vc4_hdmi_dt_match, ++ .pm = &vc4_hdmi_pm_ops, + }, + }; +diff --git a/drivers/gpu/drm/vc4/vc4_hvs.c b/drivers/gpu/drm/vc4/vc4_hvs.c +index c239045e05d6f..604933e20e6a2 100644 +--- a/drivers/gpu/drm/vc4/vc4_hvs.c ++++ b/drivers/gpu/drm/vc4/vc4_hvs.c +@@ -365,17 +365,16 @@ static void vc4_hvs_update_dlist(struct drm_crtc *crtc) + struct vc4_dev *vc4 = to_vc4_dev(dev); + struct vc4_crtc *vc4_crtc = to_vc4_crtc(crtc); + struct vc4_crtc_state *vc4_state = to_vc4_crtc_state(crtc->state); ++ unsigned long flags; + + if (crtc->state->event) { +- unsigned long flags; +- + crtc->state->event->pipe = drm_crtc_index(crtc); + + WARN_ON(drm_crtc_vblank_get(crtc) != 0); + + spin_lock_irqsave(&dev->event_lock, flags); + +- if (!vc4_state->feed_txp || vc4_state->txp_armed) { ++ if (!vc4_crtc->feeds_txp || vc4_state->txp_armed) { + vc4_crtc->event = crtc->state->event; + crtc->state->event = NULL; + } +@@ -388,6 +387,22 @@ static void vc4_hvs_update_dlist(struct drm_crtc *crtc) + HVS_WRITE(SCALER_DISPLISTX(vc4_state->assigned_channel), + vc4_state->mm.start); + } ++ ++ spin_lock_irqsave(&vc4_crtc->irq_lock, flags); ++ vc4_crtc->current_dlist = vc4_state->mm.start; ++ spin_unlock_irqrestore(&vc4_crtc->irq_lock, flags); ++} ++ ++void vc4_hvs_atomic_begin(struct drm_crtc *crtc, ++ struct drm_atomic_state *state) ++{ ++ struct vc4_crtc *vc4_crtc = to_vc4_crtc(crtc); ++ struct vc4_crtc_state *vc4_state = to_vc4_crtc_state(crtc->state); ++ unsigned long flags; ++ ++ spin_lock_irqsave(&vc4_crtc->irq_lock, flags); ++ vc4_crtc->current_hvs_channel = vc4_state->assigned_channel; ++ spin_unlock_irqrestore(&vc4_crtc->irq_lock, flags); + } + + void vc4_hvs_atomic_enable(struct drm_crtc *crtc, +@@ -395,10 +410,9 @@ void vc4_hvs_atomic_enable(struct drm_crtc *crtc, + { + struct drm_device *dev = crtc->dev; + struct vc4_dev *vc4 = to_vc4_dev(dev); +- struct drm_crtc_state *new_crtc_state = drm_atomic_get_new_crtc_state(state, crtc); +- struct vc4_crtc_state *vc4_state = to_vc4_crtc_state(new_crtc_state); + struct drm_display_mode *mode = &crtc->state->adjusted_mode; +- bool oneshot = vc4_state->feed_txp; ++ struct vc4_crtc *vc4_crtc = to_vc4_crtc(crtc); ++ bool oneshot = vc4_crtc->feeds_txp; + + vc4_hvs_update_dlist(crtc); + vc4_hvs_init_channel(vc4, crtc, mode, oneshot); +diff --git a/drivers/gpu/drm/vc4/vc4_kms.c b/drivers/gpu/drm/vc4/vc4_kms.c +index b61792d2aa657..6030d4a821555 100644 +--- a/drivers/gpu/drm/vc4/vc4_kms.c ++++ b/drivers/gpu/drm/vc4/vc4_kms.c +@@ -233,6 +233,7 @@ static void vc4_hvs_pv_muxing_commit(struct vc4_dev *vc4, + unsigned int i; + + for_each_new_crtc_in_state(state, crtc, crtc_state, i) { ++ struct vc4_crtc *vc4_crtc = to_vc4_crtc(crtc); + struct vc4_crtc_state *vc4_state = to_vc4_crtc_state(crtc_state); + u32 dispctrl; + u32 dsp3_mux; +@@ -253,7 +254,7 @@ static void vc4_hvs_pv_muxing_commit(struct vc4_dev *vc4, + * TXP IP, and we need to disable the FIFO2 -> pixelvalve1 + * route. + */ +- if (vc4_state->feed_txp) ++ if (vc4_crtc->feeds_txp) + dsp3_mux = VC4_SET_FIELD(3, SCALER_DISPCTRL_DSP3_MUX); + else + dsp3_mux = VC4_SET_FIELD(2, SCALER_DISPCTRL_DSP3_MUX); +diff --git a/drivers/gpu/drm/vc4/vc4_txp.c b/drivers/gpu/drm/vc4/vc4_txp.c +index 2fc7f4b5fa098..9809ca3e29451 100644 +--- a/drivers/gpu/drm/vc4/vc4_txp.c ++++ b/drivers/gpu/drm/vc4/vc4_txp.c +@@ -391,7 +391,6 @@ static int vc4_txp_atomic_check(struct drm_crtc *crtc, + { + struct drm_crtc_state *crtc_state = drm_atomic_get_new_crtc_state(state, + crtc); +- struct vc4_crtc_state *vc4_state = to_vc4_crtc_state(crtc_state); + int ret; + + ret = vc4_hvs_atomic_check(crtc, state); +@@ -399,7 +398,6 @@ static int vc4_txp_atomic_check(struct drm_crtc *crtc, + return ret; + + crtc_state->no_vblank = true; +- vc4_state->feed_txp = true; + + return 0; + } +@@ -437,6 +435,7 @@ static void vc4_txp_atomic_disable(struct drm_crtc *crtc, + + static const struct drm_crtc_helper_funcs vc4_txp_crtc_helper_funcs = { + .atomic_check = vc4_txp_atomic_check, ++ .atomic_begin = vc4_hvs_atomic_begin, + .atomic_flush = vc4_hvs_atomic_flush, + .atomic_enable = vc4_txp_atomic_enable, + .atomic_disable = vc4_txp_atomic_disable, +@@ -482,6 +481,7 @@ static int vc4_txp_bind(struct device *dev, struct device *master, void *data) + + vc4_crtc->pdev = pdev; + vc4_crtc->data = &vc4_txp_crtc_data; ++ vc4_crtc->feeds_txp = true; + + txp->pdev = pdev; + +diff --git a/drivers/gpu/drm/vmwgfx/Makefile b/drivers/gpu/drm/vmwgfx/Makefile +index bc323f7d40321..18edc7ca5b454 100644 +--- a/drivers/gpu/drm/vmwgfx/Makefile ++++ b/drivers/gpu/drm/vmwgfx/Makefile +@@ -9,9 +9,8 @@ vmwgfx-y := vmwgfx_execbuf.o vmwgfx_gmr.o vmwgfx_kms.o vmwgfx_drv.o \ + vmwgfx_cotable.o vmwgfx_so.o vmwgfx_binding.o vmwgfx_msg.o \ + vmwgfx_simple_resource.o vmwgfx_va.o vmwgfx_blit.o \ + vmwgfx_validation.o vmwgfx_page_dirty.o vmwgfx_streamoutput.o \ +- vmwgfx_devcaps.o ttm_object.o ttm_memory.o ++ vmwgfx_devcaps.o ttm_object.o ttm_memory.o vmwgfx_system_manager.o + + vmwgfx-$(CONFIG_DRM_FBDEV_EMULATION) += vmwgfx_fb.o +-vmwgfx-$(CONFIG_TRANSPARENT_HUGEPAGE) += vmwgfx_thp.o + + obj-$(CONFIG_DRM_VMWGFX) := vmwgfx.o +diff --git a/drivers/gpu/drm/vmwgfx/ttm_memory.c b/drivers/gpu/drm/vmwgfx/ttm_memory.c +index edd17c30d5a51..2ced4c06ca451 100644 +--- a/drivers/gpu/drm/vmwgfx/ttm_memory.c ++++ b/drivers/gpu/drm/vmwgfx/ttm_memory.c +@@ -34,7 +34,6 @@ + #include <linux/mm.h> + #include <linux/module.h> + #include <linux/slab.h> +-#include <linux/swap.h> + + #include <drm/drm_device.h> + #include <drm/drm_file.h> +@@ -173,69 +172,7 @@ static struct kobj_type ttm_mem_zone_kobj_type = { + .sysfs_ops = &ttm_mem_zone_ops, + .default_attrs = ttm_mem_zone_attrs, + }; +- +-static struct attribute ttm_mem_global_lower_mem_limit = { +- .name = "lower_mem_limit", +- .mode = S_IRUGO | S_IWUSR +-}; +- +-static ssize_t ttm_mem_global_show(struct kobject *kobj, +- struct attribute *attr, +- char *buffer) +-{ +- struct ttm_mem_global *glob = +- container_of(kobj, struct ttm_mem_global, kobj); +- uint64_t val = 0; +- +- spin_lock(&glob->lock); +- val = glob->lower_mem_limit; +- spin_unlock(&glob->lock); +- /* convert from number of pages to KB */ +- val <<= (PAGE_SHIFT - 10); +- return snprintf(buffer, PAGE_SIZE, "%llu\n", +- (unsigned long long) val); +-} +- +-static ssize_t ttm_mem_global_store(struct kobject *kobj, +- struct attribute *attr, +- const char *buffer, +- size_t size) +-{ +- int chars; +- uint64_t val64; +- unsigned long val; +- struct ttm_mem_global *glob = +- container_of(kobj, struct ttm_mem_global, kobj); +- +- chars = sscanf(buffer, "%lu", &val); +- if (chars == 0) +- return size; +- +- val64 = val; +- /* convert from KB to number of pages */ +- val64 >>= (PAGE_SHIFT - 10); +- +- spin_lock(&glob->lock); +- glob->lower_mem_limit = val64; +- spin_unlock(&glob->lock); +- +- return size; +-} +- +-static struct attribute *ttm_mem_global_attrs[] = { +- &ttm_mem_global_lower_mem_limit, +- NULL +-}; +- +-static const struct sysfs_ops ttm_mem_global_ops = { +- .show = &ttm_mem_global_show, +- .store = &ttm_mem_global_store, +-}; +- +-static struct kobj_type ttm_mem_glob_kobj_type = { +- .sysfs_ops = &ttm_mem_global_ops, +- .default_attrs = ttm_mem_global_attrs, +-}; ++static struct kobj_type ttm_mem_glob_kobj_type = {0}; + + static bool ttm_zones_above_swap_target(struct ttm_mem_global *glob, + bool from_wq, uint64_t extra) +@@ -435,11 +372,6 @@ int ttm_mem_global_init(struct ttm_mem_global *glob, struct device *dev) + + si_meminfo(&si); + +- spin_lock(&glob->lock); +- /* set it as 0 by default to keep original behavior of OOM */ +- glob->lower_mem_limit = 0; +- spin_unlock(&glob->lock); +- + ret = ttm_mem_init_kernel_zone(glob, &si); + if (unlikely(ret != 0)) + goto out_no_zone; +@@ -527,35 +459,6 @@ void ttm_mem_global_free(struct ttm_mem_global *glob, + } + EXPORT_SYMBOL(ttm_mem_global_free); + +-/* +- * check if the available mem is under lower memory limit +- * +- * a. if no swap disk at all or free swap space is under swap_mem_limit +- * but available system mem is bigger than sys_mem_limit, allow TTM +- * allocation; +- * +- * b. if the available system mem is less than sys_mem_limit but free +- * swap disk is bigger than swap_mem_limit, allow TTM allocation. +- */ +-bool +-ttm_check_under_lowerlimit(struct ttm_mem_global *glob, +- uint64_t num_pages, +- struct ttm_operation_ctx *ctx) +-{ +- int64_t available; +- +- /* We allow over commit during suspend */ +- if (ctx->force_alloc) +- return false; +- +- available = get_nr_swap_pages() + si_mem_available(); +- available -= num_pages; +- if (available < glob->lower_mem_limit) +- return true; +- +- return false; +-} +- + static int ttm_mem_global_reserve(struct ttm_mem_global *glob, + struct ttm_mem_zone *single_zone, + uint64_t amount, bool reserve) +diff --git a/drivers/gpu/drm/vmwgfx/ttm_memory.h b/drivers/gpu/drm/vmwgfx/ttm_memory.h +index c50dba7744854..7b0d617ebcb1e 100644 +--- a/drivers/gpu/drm/vmwgfx/ttm_memory.h ++++ b/drivers/gpu/drm/vmwgfx/ttm_memory.h +@@ -50,8 +50,6 @@ + * @work: The workqueue callback for the shrink queue. + * @lock: Lock to protect the @shrink - and the memory accounting members, + * that is, essentially the whole structure with some exceptions. +- * @lower_mem_limit: include lower limit of swap space and lower limit of +- * system memory. + * @zones: Array of pointers to accounting zones. + * @num_zones: Number of populated entries in the @zones array. + * @zone_kernel: Pointer to the kernel zone. +@@ -69,7 +67,6 @@ extern struct ttm_mem_global { + struct workqueue_struct *swap_queue; + struct work_struct work; + spinlock_t lock; +- uint64_t lower_mem_limit; + struct ttm_mem_zone *zones[TTM_MEM_MAX_ZONES]; + unsigned int num_zones; + struct ttm_mem_zone *zone_kernel; +@@ -91,6 +88,5 @@ int ttm_mem_global_alloc_page(struct ttm_mem_global *glob, + void ttm_mem_global_free_page(struct ttm_mem_global *glob, + struct page *page, uint64_t size); + size_t ttm_round_pot(size_t size); +-bool ttm_check_under_lowerlimit(struct ttm_mem_global *glob, uint64_t num_pages, +- struct ttm_operation_ctx *ctx); ++ + #endif +diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_cmd.c b/drivers/gpu/drm/vmwgfx/vmwgfx_cmd.c +index 67db472d3493c..a3bfbb6c3e14a 100644 +--- a/drivers/gpu/drm/vmwgfx/vmwgfx_cmd.c ++++ b/drivers/gpu/drm/vmwgfx/vmwgfx_cmd.c +@@ -145,6 +145,13 @@ struct vmw_fifo_state *vmw_fifo_create(struct vmw_private *dev_priv) + (unsigned int) max, + (unsigned int) min, + (unsigned int) fifo->capabilities); ++ ++ if (unlikely(min >= max)) { ++ drm_warn(&dev_priv->drm, ++ "FIFO memory is not usable. Driver failed to initialize."); ++ return ERR_PTR(-ENXIO); ++ } ++ + return fifo; + } + +diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_drv.c b/drivers/gpu/drm/vmwgfx/vmwgfx_drv.c +index ab9a1750e1dff..8449d09c06f7a 100644 +--- a/drivers/gpu/drm/vmwgfx/vmwgfx_drv.c ++++ b/drivers/gpu/drm/vmwgfx/vmwgfx_drv.c +@@ -707,23 +707,15 @@ static int vmw_dma_masks(struct vmw_private *dev_priv) + static int vmw_vram_manager_init(struct vmw_private *dev_priv) + { + int ret; +-#ifdef CONFIG_TRANSPARENT_HUGEPAGE +- ret = vmw_thp_init(dev_priv); +-#else + ret = ttm_range_man_init(&dev_priv->bdev, TTM_PL_VRAM, false, + dev_priv->vram_size >> PAGE_SHIFT); +-#endif + ttm_resource_manager_set_used(ttm_manager_type(&dev_priv->bdev, TTM_PL_VRAM), false); + return ret; + } + + static void vmw_vram_manager_fini(struct vmw_private *dev_priv) + { +-#ifdef CONFIG_TRANSPARENT_HUGEPAGE +- vmw_thp_fini(dev_priv); +-#else + ttm_range_man_fini(&dev_priv->bdev, TTM_PL_VRAM); +-#endif + } + + static int vmw_setup_pci_resources(struct vmw_private *dev, +@@ -1071,6 +1063,12 @@ static int vmw_driver_load(struct vmw_private *dev_priv, u32 pci_id) + "3D will be disabled.\n"); + dev_priv->has_mob = false; + } ++ if (vmw_sys_man_init(dev_priv) != 0) { ++ drm_info(&dev_priv->drm, ++ "No MOB page table memory available. " ++ "3D will be disabled.\n"); ++ dev_priv->has_mob = false; ++ } + } + + if (dev_priv->has_mob && (dev_priv->capabilities & SVGA_CAP_DX)) { +@@ -1121,8 +1119,10 @@ out_no_fifo: + vmw_overlay_close(dev_priv); + vmw_kms_close(dev_priv); + out_no_kms: +- if (dev_priv->has_mob) ++ if (dev_priv->has_mob) { + vmw_gmrid_man_fini(dev_priv, VMW_PL_MOB); ++ vmw_sys_man_fini(dev_priv); ++ } + if (dev_priv->has_gmr) + vmw_gmrid_man_fini(dev_priv, VMW_PL_GMR); + vmw_devcaps_destroy(dev_priv); +@@ -1172,8 +1172,10 @@ static void vmw_driver_unload(struct drm_device *dev) + vmw_gmrid_man_fini(dev_priv, VMW_PL_GMR); + + vmw_release_device_early(dev_priv); +- if (dev_priv->has_mob) ++ if (dev_priv->has_mob) { + vmw_gmrid_man_fini(dev_priv, VMW_PL_MOB); ++ vmw_sys_man_fini(dev_priv); ++ } + vmw_devcaps_destroy(dev_priv); + vmw_vram_manager_fini(dev_priv); + ttm_device_fini(&dev_priv->bdev); +@@ -1617,34 +1619,40 @@ static int vmw_probe(struct pci_dev *pdev, const struct pci_device_id *ent) + + ret = drm_aperture_remove_conflicting_pci_framebuffers(pdev, &driver); + if (ret) +- return ret; ++ goto out_error; + + ret = pcim_enable_device(pdev); + if (ret) +- return ret; ++ goto out_error; + + vmw = devm_drm_dev_alloc(&pdev->dev, &driver, + struct vmw_private, drm); +- if (IS_ERR(vmw)) +- return PTR_ERR(vmw); ++ if (IS_ERR(vmw)) { ++ ret = PTR_ERR(vmw); ++ goto out_error; ++ } + + pci_set_drvdata(pdev, &vmw->drm); + + ret = ttm_mem_global_init(&ttm_mem_glob, &pdev->dev); + if (ret) +- return ret; ++ goto out_error; + + ret = vmw_driver_load(vmw, ent->device); + if (ret) +- return ret; ++ goto out_release; + + ret = drm_dev_register(&vmw->drm, 0); +- if (ret) { +- vmw_driver_unload(&vmw->drm); +- return ret; +- } ++ if (ret) ++ goto out_unload; + + return 0; ++out_unload: ++ vmw_driver_unload(&vmw->drm); ++out_release: ++ ttm_mem_global_release(&ttm_mem_glob); ++out_error: ++ return ret; + } + + static int __init vmwgfx_init(void) +diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_drv.h b/drivers/gpu/drm/vmwgfx/vmwgfx_drv.h +index 858aff99a3fe5..2a7cec4cb8a89 100644 +--- a/drivers/gpu/drm/vmwgfx/vmwgfx_drv.h ++++ b/drivers/gpu/drm/vmwgfx/vmwgfx_drv.h +@@ -59,11 +59,8 @@ + #define VMWGFX_DRIVER_MINOR 19 + #define VMWGFX_DRIVER_PATCHLEVEL 0 + #define VMWGFX_FIFO_STATIC_SIZE (1024*1024) +-#define VMWGFX_MAX_RELOCATIONS 2048 +-#define VMWGFX_MAX_VALIDATIONS 2048 + #define VMWGFX_MAX_DISPLAYS 16 + #define VMWGFX_CMD_BOUNCE_INIT_SIZE 32768 +-#define VMWGFX_ENABLE_SCREEN_TARGET_OTABLE 1 + + #define VMWGFX_PCI_ID_SVGA2 0x0405 + #define VMWGFX_PCI_ID_SVGA3 0x0406 +@@ -82,8 +79,9 @@ + VMWGFX_NUM_GB_SURFACE +\ + VMWGFX_NUM_GB_SCREEN_TARGET) + +-#define VMW_PL_GMR (TTM_PL_PRIV + 0) +-#define VMW_PL_MOB (TTM_PL_PRIV + 1) ++#define VMW_PL_GMR (TTM_PL_PRIV + 0) ++#define VMW_PL_MOB (TTM_PL_PRIV + 1) ++#define VMW_PL_SYSTEM (TTM_PL_PRIV + 2) + + #define VMW_RES_CONTEXT ttm_driver_type0 + #define VMW_RES_SURFACE ttm_driver_type1 +@@ -1039,7 +1037,6 @@ extern struct ttm_placement vmw_vram_placement; + extern struct ttm_placement vmw_vram_sys_placement; + extern struct ttm_placement vmw_vram_gmr_placement; + extern struct ttm_placement vmw_sys_placement; +-extern struct ttm_placement vmw_evictable_placement; + extern struct ttm_placement vmw_srf_placement; + extern struct ttm_placement vmw_mob_placement; + extern struct ttm_placement vmw_nonfixed_placement; +@@ -1251,6 +1248,12 @@ int vmw_overlay_num_free_overlays(struct vmw_private *dev_priv); + int vmw_gmrid_man_init(struct vmw_private *dev_priv, int type); + void vmw_gmrid_man_fini(struct vmw_private *dev_priv, int type); + ++/** ++ * System memory manager ++ */ ++int vmw_sys_man_init(struct vmw_private *dev_priv); ++void vmw_sys_man_fini(struct vmw_private *dev_priv); ++ + /** + * Prime - vmwgfx_prime.c + */ +@@ -1551,11 +1554,6 @@ void vmw_bo_dirty_unmap(struct vmw_buffer_object *vbo, + vm_fault_t vmw_bo_vm_fault(struct vm_fault *vmf); + vm_fault_t vmw_bo_vm_mkwrite(struct vm_fault *vmf); + +-/* Transparent hugepage support - vmwgfx_thp.c */ +-#ifdef CONFIG_TRANSPARENT_HUGEPAGE +-extern int vmw_thp_init(struct vmw_private *dev_priv); +-void vmw_thp_fini(struct vmw_private *dev_priv); +-#endif + + /** + * VMW_DEBUG_KMS - Debug output for kernel mode-setting +diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_mob.c b/drivers/gpu/drm/vmwgfx/vmwgfx_mob.c +index f9394207dd3cc..632e587519722 100644 +--- a/drivers/gpu/drm/vmwgfx/vmwgfx_mob.c ++++ b/drivers/gpu/drm/vmwgfx/vmwgfx_mob.c +@@ -1,7 +1,7 @@ + // SPDX-License-Identifier: GPL-2.0 OR MIT + /************************************************************************** + * +- * Copyright 2012-2015 VMware, Inc., Palo Alto, CA., USA ++ * Copyright 2012-2021 VMware, Inc., Palo Alto, CA., USA + * + * Permission is hereby granted, free of charge, to any person obtaining a + * copy of this software and associated documentation files (the +@@ -29,12 +29,6 @@ + + #include "vmwgfx_drv.h" + +-/* +- * If we set up the screen target otable, screen objects stop working. +- */ +- +-#define VMW_OTABLE_SETUP_SUB ((VMWGFX_ENABLE_SCREEN_TARGET_OTABLE ? 0 : 1)) +- + #ifdef CONFIG_64BIT + #define VMW_PPN_SIZE 8 + #define VMW_MOBFMT_PTDEPTH_0 SVGA3D_MOBFMT_PT64_0 +@@ -75,7 +69,7 @@ static const struct vmw_otable pre_dx_tables[] = { + {VMWGFX_NUM_GB_CONTEXT * sizeof(SVGAOTableContextEntry), NULL, true}, + {VMWGFX_NUM_GB_SHADER * sizeof(SVGAOTableShaderEntry), NULL, true}, + {VMWGFX_NUM_GB_SCREEN_TARGET * sizeof(SVGAOTableScreenTargetEntry), +- NULL, VMWGFX_ENABLE_SCREEN_TARGET_OTABLE} ++ NULL, true} + }; + + static const struct vmw_otable dx_tables[] = { +@@ -84,7 +78,7 @@ static const struct vmw_otable dx_tables[] = { + {VMWGFX_NUM_GB_CONTEXT * sizeof(SVGAOTableContextEntry), NULL, true}, + {VMWGFX_NUM_GB_SHADER * sizeof(SVGAOTableShaderEntry), NULL, true}, + {VMWGFX_NUM_GB_SCREEN_TARGET * sizeof(SVGAOTableScreenTargetEntry), +- NULL, VMWGFX_ENABLE_SCREEN_TARGET_OTABLE}, ++ NULL, true}, + {VMWGFX_NUM_DXCONTEXT * sizeof(SVGAOTableDXContextEntry), NULL, true}, + }; + +diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c b/drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c +index d85310b2608dd..f5e90d0e2d0f8 100644 +--- a/drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c ++++ b/drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c +@@ -1872,8 +1872,8 @@ int vmw_kms_stdu_init_display(struct vmw_private *dev_priv) + int i, ret; + + +- /* Do nothing if Screen Target support is turned off */ +- if (!VMWGFX_ENABLE_SCREEN_TARGET_OTABLE || !dev_priv->has_mob) ++ /* Do nothing if there's no support for MOBs */ ++ if (!dev_priv->has_mob) + return -ENOSYS; + + if (!(dev_priv->capabilities & SVGA_CAP_GBOBJECTS)) +diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_system_manager.c b/drivers/gpu/drm/vmwgfx/vmwgfx_system_manager.c +new file mode 100644 +index 0000000000000..b0005b03a6174 +--- /dev/null ++++ b/drivers/gpu/drm/vmwgfx/vmwgfx_system_manager.c +@@ -0,0 +1,90 @@ ++/* SPDX-License-Identifier: GPL-2.0 OR MIT */ ++/* ++ * Copyright 2021 VMware, Inc. ++ * ++ * Permission is hereby granted, free of charge, to any person ++ * obtaining a copy of this software and associated documentation ++ * files (the "Software"), to deal in the Software without ++ * restriction, including without limitation the rights to use, copy, ++ * modify, merge, publish, distribute, sublicense, and/or sell copies ++ * of the Software, and to permit persons to whom the Software is ++ * furnished to do so, subject to the following conditions: ++ * ++ * The above copyright notice and this permission notice shall be ++ * included in all copies or substantial portions of the Software. ++ * ++ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, ++ * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF ++ * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND ++ * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS ++ * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ++ * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN ++ * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE ++ * SOFTWARE. ++ * ++ */ ++ ++#include "vmwgfx_drv.h" ++ ++#include <drm/ttm/ttm_bo_driver.h> ++#include <drm/ttm/ttm_device.h> ++#include <drm/ttm/ttm_placement.h> ++#include <drm/ttm/ttm_resource.h> ++#include <linux/slab.h> ++ ++ ++static int vmw_sys_man_alloc(struct ttm_resource_manager *man, ++ struct ttm_buffer_object *bo, ++ const struct ttm_place *place, ++ struct ttm_resource **res) ++{ ++ *res = kzalloc(sizeof(**res), GFP_KERNEL); ++ if (!*res) ++ return -ENOMEM; ++ ++ ttm_resource_init(bo, place, *res); ++ return 0; ++} ++ ++static void vmw_sys_man_free(struct ttm_resource_manager *man, ++ struct ttm_resource *res) ++{ ++ kfree(res); ++} ++ ++static const struct ttm_resource_manager_func vmw_sys_manager_func = { ++ .alloc = vmw_sys_man_alloc, ++ .free = vmw_sys_man_free, ++}; ++ ++int vmw_sys_man_init(struct vmw_private *dev_priv) ++{ ++ struct ttm_device *bdev = &dev_priv->bdev; ++ struct ttm_resource_manager *man = ++ kzalloc(sizeof(*man), GFP_KERNEL); ++ ++ if (!man) ++ return -ENOMEM; ++ ++ man->use_tt = true; ++ man->func = &vmw_sys_manager_func; ++ ++ ttm_resource_manager_init(man, 0); ++ ttm_set_driver_manager(bdev, VMW_PL_SYSTEM, man); ++ ttm_resource_manager_set_used(man, true); ++ return 0; ++} ++ ++void vmw_sys_man_fini(struct vmw_private *dev_priv) ++{ ++ struct ttm_resource_manager *man = ttm_manager_type(&dev_priv->bdev, ++ VMW_PL_SYSTEM); ++ ++ ttm_resource_manager_evict_all(&dev_priv->bdev, man); ++ ++ ttm_resource_manager_set_used(man, false); ++ ttm_resource_manager_cleanup(man); ++ ++ ttm_set_driver_manager(&dev_priv->bdev, VMW_PL_SYSTEM, NULL); ++ kfree(man); ++} +diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_thp.c b/drivers/gpu/drm/vmwgfx/vmwgfx_thp.c +deleted file mode 100644 +index 2a3d3468e4e0a..0000000000000 +--- a/drivers/gpu/drm/vmwgfx/vmwgfx_thp.c ++++ /dev/null +@@ -1,184 +0,0 @@ +-// SPDX-License-Identifier: GPL-2.0 OR MIT +-/* +- * Huge page-table-entry support for IO memory. +- * +- * Copyright (C) 2007-2019 Vmware, Inc. All rights reservedd. +- */ +-#include "vmwgfx_drv.h" +-#include <drm/ttm/ttm_bo_driver.h> +-#include <drm/ttm/ttm_placement.h> +-#include <drm/ttm/ttm_range_manager.h> +- +-/** +- * struct vmw_thp_manager - Range manager implementing huge page alignment +- * +- * @manager: TTM resource manager. +- * @mm: The underlying range manager. Protected by @lock. +- * @lock: Manager lock. +- */ +-struct vmw_thp_manager { +- struct ttm_resource_manager manager; +- struct drm_mm mm; +- spinlock_t lock; +-}; +- +-static struct vmw_thp_manager *to_thp_manager(struct ttm_resource_manager *man) +-{ +- return container_of(man, struct vmw_thp_manager, manager); +-} +- +-static const struct ttm_resource_manager_func vmw_thp_func; +- +-static int vmw_thp_insert_aligned(struct ttm_buffer_object *bo, +- struct drm_mm *mm, struct drm_mm_node *node, +- unsigned long align_pages, +- const struct ttm_place *place, +- struct ttm_resource *mem, +- unsigned long lpfn, +- enum drm_mm_insert_mode mode) +-{ +- if (align_pages >= bo->page_alignment && +- (!bo->page_alignment || align_pages % bo->page_alignment == 0)) { +- return drm_mm_insert_node_in_range(mm, node, +- mem->num_pages, +- align_pages, 0, +- place->fpfn, lpfn, mode); +- } +- +- return -ENOSPC; +-} +- +-static int vmw_thp_get_node(struct ttm_resource_manager *man, +- struct ttm_buffer_object *bo, +- const struct ttm_place *place, +- struct ttm_resource **res) +-{ +- struct vmw_thp_manager *rman = to_thp_manager(man); +- struct drm_mm *mm = &rman->mm; +- struct ttm_range_mgr_node *node; +- unsigned long align_pages; +- unsigned long lpfn; +- enum drm_mm_insert_mode mode = DRM_MM_INSERT_BEST; +- int ret; +- +- node = kzalloc(struct_size(node, mm_nodes, 1), GFP_KERNEL); +- if (!node) +- return -ENOMEM; +- +- ttm_resource_init(bo, place, &node->base); +- +- lpfn = place->lpfn; +- if (!lpfn) +- lpfn = man->size; +- +- mode = DRM_MM_INSERT_BEST; +- if (place->flags & TTM_PL_FLAG_TOPDOWN) +- mode = DRM_MM_INSERT_HIGH; +- +- spin_lock(&rman->lock); +- if (IS_ENABLED(CONFIG_HAVE_ARCH_TRANSPARENT_HUGEPAGE_PUD)) { +- align_pages = (HPAGE_PUD_SIZE >> PAGE_SHIFT); +- if (node->base.num_pages >= align_pages) { +- ret = vmw_thp_insert_aligned(bo, mm, &node->mm_nodes[0], +- align_pages, place, +- &node->base, lpfn, mode); +- if (!ret) +- goto found_unlock; +- } +- } +- +- align_pages = (HPAGE_PMD_SIZE >> PAGE_SHIFT); +- if (node->base.num_pages >= align_pages) { +- ret = vmw_thp_insert_aligned(bo, mm, &node->mm_nodes[0], +- align_pages, place, &node->base, +- lpfn, mode); +- if (!ret) +- goto found_unlock; +- } +- +- ret = drm_mm_insert_node_in_range(mm, &node->mm_nodes[0], +- node->base.num_pages, +- bo->page_alignment, 0, +- place->fpfn, lpfn, mode); +-found_unlock: +- spin_unlock(&rman->lock); +- +- if (unlikely(ret)) { +- kfree(node); +- } else { +- node->base.start = node->mm_nodes[0].start; +- *res = &node->base; +- } +- +- return ret; +-} +- +-static void vmw_thp_put_node(struct ttm_resource_manager *man, +- struct ttm_resource *res) +-{ +- struct ttm_range_mgr_node *node = to_ttm_range_mgr_node(res); +- struct vmw_thp_manager *rman = to_thp_manager(man); +- +- spin_lock(&rman->lock); +- drm_mm_remove_node(&node->mm_nodes[0]); +- spin_unlock(&rman->lock); +- +- kfree(node); +-} +- +-int vmw_thp_init(struct vmw_private *dev_priv) +-{ +- struct vmw_thp_manager *rman; +- +- rman = kzalloc(sizeof(*rman), GFP_KERNEL); +- if (!rman) +- return -ENOMEM; +- +- ttm_resource_manager_init(&rman->manager, +- dev_priv->vram_size >> PAGE_SHIFT); +- +- rman->manager.func = &vmw_thp_func; +- drm_mm_init(&rman->mm, 0, rman->manager.size); +- spin_lock_init(&rman->lock); +- +- ttm_set_driver_manager(&dev_priv->bdev, TTM_PL_VRAM, &rman->manager); +- ttm_resource_manager_set_used(&rman->manager, true); +- return 0; +-} +- +-void vmw_thp_fini(struct vmw_private *dev_priv) +-{ +- struct ttm_resource_manager *man = ttm_manager_type(&dev_priv->bdev, TTM_PL_VRAM); +- struct vmw_thp_manager *rman = to_thp_manager(man); +- struct drm_mm *mm = &rman->mm; +- int ret; +- +- ttm_resource_manager_set_used(man, false); +- +- ret = ttm_resource_manager_evict_all(&dev_priv->bdev, man); +- if (ret) +- return; +- spin_lock(&rman->lock); +- drm_mm_clean(mm); +- drm_mm_takedown(mm); +- spin_unlock(&rman->lock); +- ttm_resource_manager_cleanup(man); +- ttm_set_driver_manager(&dev_priv->bdev, TTM_PL_VRAM, NULL); +- kfree(rman); +-} +- +-static void vmw_thp_debug(struct ttm_resource_manager *man, +- struct drm_printer *printer) +-{ +- struct vmw_thp_manager *rman = to_thp_manager(man); +- +- spin_lock(&rman->lock); +- drm_mm_print(&rman->mm, printer); +- spin_unlock(&rman->lock); +-} +- +-static const struct ttm_resource_manager_func vmw_thp_func = { +- .alloc = vmw_thp_get_node, +- .free = vmw_thp_put_node, +- .debug = vmw_thp_debug +-}; +diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c b/drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c +index 8b8991e3ed2d0..450bb1e9626f7 100644 +--- a/drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c ++++ b/drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c +@@ -92,6 +92,13 @@ static const struct ttm_place gmr_vram_placement_flags[] = { + } + }; + ++static const struct ttm_place vmw_sys_placement_flags = { ++ .fpfn = 0, ++ .lpfn = 0, ++ .mem_type = VMW_PL_SYSTEM, ++ .flags = 0 ++}; ++ + struct ttm_placement vmw_vram_gmr_placement = { + .num_placement = 2, + .placement = vram_gmr_placement_flags, +@@ -113,28 +120,11 @@ struct ttm_placement vmw_sys_placement = { + .busy_placement = &sys_placement_flags + }; + +-static const struct ttm_place evictable_placement_flags[] = { +- { +- .fpfn = 0, +- .lpfn = 0, +- .mem_type = TTM_PL_SYSTEM, +- .flags = 0 +- }, { +- .fpfn = 0, +- .lpfn = 0, +- .mem_type = TTM_PL_VRAM, +- .flags = 0 +- }, { +- .fpfn = 0, +- .lpfn = 0, +- .mem_type = VMW_PL_GMR, +- .flags = 0 +- }, { +- .fpfn = 0, +- .lpfn = 0, +- .mem_type = VMW_PL_MOB, +- .flags = 0 +- } ++struct ttm_placement vmw_pt_sys_placement = { ++ .num_placement = 1, ++ .placement = &vmw_sys_placement_flags, ++ .num_busy_placement = 1, ++ .busy_placement = &vmw_sys_placement_flags + }; + + static const struct ttm_place nonfixed_placement_flags[] = { +@@ -156,13 +146,6 @@ static const struct ttm_place nonfixed_placement_flags[] = { + } + }; + +-struct ttm_placement vmw_evictable_placement = { +- .num_placement = 4, +- .placement = evictable_placement_flags, +- .num_busy_placement = 1, +- .busy_placement = &sys_placement_flags +-}; +- + struct ttm_placement vmw_srf_placement = { + .num_placement = 1, + .num_busy_placement = 2, +@@ -484,6 +467,9 @@ static int vmw_ttm_bind(struct ttm_device *bdev, + &vmw_be->vsgt, ttm->num_pages, + vmw_be->gmr_id); + break; ++ case VMW_PL_SYSTEM: ++ /* Nothing to be done for a system bind */ ++ break; + default: + BUG(); + } +@@ -507,6 +493,8 @@ static void vmw_ttm_unbind(struct ttm_device *bdev, + case VMW_PL_MOB: + vmw_mob_unbind(vmw_be->dev_priv, vmw_be->mob); + break; ++ case VMW_PL_SYSTEM: ++ break; + default: + BUG(); + } +@@ -628,6 +616,7 @@ static int vmw_ttm_io_mem_reserve(struct ttm_device *bdev, struct ttm_resource * + + switch (mem->mem_type) { + case TTM_PL_SYSTEM: ++ case VMW_PL_SYSTEM: + case VMW_PL_GMR: + case VMW_PL_MOB: + return 0; +@@ -674,6 +663,11 @@ static void vmw_swap_notify(struct ttm_buffer_object *bo) + (void) ttm_bo_wait(bo, false, false); + } + ++static bool vmw_memtype_is_system(uint32_t mem_type) ++{ ++ return mem_type == TTM_PL_SYSTEM || mem_type == VMW_PL_SYSTEM; ++} ++ + static int vmw_move(struct ttm_buffer_object *bo, + bool evict, + struct ttm_operation_ctx *ctx, +@@ -684,7 +678,7 @@ static int vmw_move(struct ttm_buffer_object *bo, + struct ttm_resource_manager *new_man = ttm_manager_type(bo->bdev, new_mem->mem_type); + int ret; + +- if (new_man->use_tt && new_mem->mem_type != TTM_PL_SYSTEM) { ++ if (new_man->use_tt && !vmw_memtype_is_system(new_mem->mem_type)) { + ret = vmw_ttm_bind(bo->bdev, bo->ttm, new_mem); + if (ret) + return ret; +@@ -693,7 +687,7 @@ static int vmw_move(struct ttm_buffer_object *bo, + vmw_move_notify(bo, bo->resource, new_mem); + + if (old_man->use_tt && new_man->use_tt) { +- if (bo->resource->mem_type == TTM_PL_SYSTEM) { ++ if (vmw_memtype_is_system(bo->resource->mem_type)) { + ttm_bo_move_null(bo, new_mem); + return 0; + } +@@ -740,7 +734,7 @@ int vmw_bo_create_and_populate(struct vmw_private *dev_priv, + int ret; + + ret = vmw_bo_create_kernel(dev_priv, bo_size, +- &vmw_sys_placement, ++ &vmw_pt_sys_placement, + &bo); + if (unlikely(ret != 0)) + return ret; +diff --git a/drivers/gpu/host1x/Kconfig b/drivers/gpu/host1x/Kconfig +index 6dab94adf25e5..6815b4db17c1b 100644 +--- a/drivers/gpu/host1x/Kconfig ++++ b/drivers/gpu/host1x/Kconfig +@@ -2,6 +2,7 @@ + config TEGRA_HOST1X + tristate "NVIDIA Tegra host1x driver" + depends on ARCH_TEGRA || (ARM && COMPILE_TEST) ++ select DMA_SHARED_BUFFER + select IOMMU_IOVA + help + Driver for the NVIDIA Tegra host1x hardware. +diff --git a/drivers/gpu/host1x/dev.c b/drivers/gpu/host1x/dev.c +index fbb6447b8659e..3872e4cd26989 100644 +--- a/drivers/gpu/host1x/dev.c ++++ b/drivers/gpu/host1x/dev.c +@@ -18,6 +18,10 @@ + #include <trace/events/host1x.h> + #undef CREATE_TRACE_POINTS + ++#if IS_ENABLED(CONFIG_ARM_DMA_USE_IOMMU) ++#include <asm/dma-iommu.h> ++#endif ++ + #include "bus.h" + #include "channel.h" + #include "debug.h" +@@ -238,6 +242,17 @@ static struct iommu_domain *host1x_iommu_attach(struct host1x *host) + struct iommu_domain *domain = iommu_get_domain_for_dev(host->dev); + int err; + ++#if IS_ENABLED(CONFIG_ARM_DMA_USE_IOMMU) ++ if (host->dev->archdata.mapping) { ++ struct dma_iommu_mapping *mapping = ++ to_dma_iommu_mapping(host->dev); ++ arm_iommu_detach_device(host->dev); ++ arm_iommu_release_mapping(mapping); ++ ++ domain = iommu_get_domain_for_dev(host->dev); ++ } ++#endif ++ + /* + * We may not always want to enable IOMMU support (for example if the + * host1x firewall is already enabled and we don't support addressing +diff --git a/drivers/hid/hid-apple.c b/drivers/hid/hid-apple.c +index 6ccfa0cb997ab..b683c0e8557d4 100644 +--- a/drivers/hid/hid-apple.c ++++ b/drivers/hid/hid-apple.c +@@ -429,7 +429,7 @@ static int apple_input_configured(struct hid_device *hdev, + + if ((asc->quirks & APPLE_HAS_FN) && !asc->fn_found) { + hid_info(hdev, "Fn key not found (Apple Wireless Keyboard clone?), disabling Fn key handling\n"); +- asc->quirks = 0; ++ asc->quirks &= ~APPLE_HAS_FN; + } + + return 0; +diff --git a/drivers/hid/hid-ids.h b/drivers/hid/hid-ids.h +index 70e65eb1b868d..bdedf594e2d1e 100644 +--- a/drivers/hid/hid-ids.h ++++ b/drivers/hid/hid-ids.h +@@ -394,6 +394,7 @@ + #define USB_DEVICE_ID_HP_X2 0x074d + #define USB_DEVICE_ID_HP_X2_10_COVER 0x0755 + #define I2C_DEVICE_ID_HP_SPECTRE_X360_15 0x2817 ++#define I2C_DEVICE_ID_HP_ENVY_X360_15T_DR100 0x29CF + #define USB_DEVICE_ID_ASUS_UX550VE_TOUCHSCREEN 0x2544 + #define USB_DEVICE_ID_ASUS_UX550_TOUCHSCREEN 0x2706 + #define I2C_DEVICE_ID_SURFACE_GO_TOUCHSCREEN 0x261A +diff --git a/drivers/hid/hid-input.c b/drivers/hid/hid-input.c +index 55017db98d896..3d33c0c06cbb3 100644 +--- a/drivers/hid/hid-input.c ++++ b/drivers/hid/hid-input.c +@@ -327,6 +327,8 @@ static const struct hid_device_id hid_battery_quirks[] = { + HID_BATTERY_QUIRK_IGNORE }, + { HID_USB_DEVICE(USB_VENDOR_ID_ELAN, USB_DEVICE_ID_ASUS_UX550VE_TOUCHSCREEN), + HID_BATTERY_QUIRK_IGNORE }, ++ { HID_I2C_DEVICE(USB_VENDOR_ID_ELAN, I2C_DEVICE_ID_HP_ENVY_X360_15T_DR100), ++ HID_BATTERY_QUIRK_IGNORE }, + { HID_I2C_DEVICE(USB_VENDOR_ID_ELAN, I2C_DEVICE_ID_HP_SPECTRE_X360_15), + HID_BATTERY_QUIRK_IGNORE }, + { HID_I2C_DEVICE(USB_VENDOR_ID_ELAN, I2C_DEVICE_ID_SURFACE_GO_TOUCHSCREEN), +@@ -1330,6 +1332,12 @@ void hidinput_hid_event(struct hid_device *hid, struct hid_field *field, struct + + input = field->hidinput->input; + ++ if (usage->type == EV_ABS && ++ (((*quirks & HID_QUIRK_X_INVERT) && usage->code == ABS_X) || ++ ((*quirks & HID_QUIRK_Y_INVERT) && usage->code == ABS_Y))) { ++ value = field->logical_maximum - value; ++ } ++ + if (usage->hat_min < usage->hat_max || usage->hat_dir) { + int hat_dir = usage->hat_dir; + if (!hat_dir) +diff --git a/drivers/hid/hid-magicmouse.c b/drivers/hid/hid-magicmouse.c +index d7687ce706144..b8b08f0a8c541 100644 +--- a/drivers/hid/hid-magicmouse.c ++++ b/drivers/hid/hid-magicmouse.c +@@ -57,6 +57,8 @@ MODULE_PARM_DESC(report_undeciphered, "Report undeciphered multi-touch state fie + #define MOUSE_REPORT_ID 0x29 + #define MOUSE2_REPORT_ID 0x12 + #define DOUBLE_REPORT_ID 0xf7 ++#define USB_BATTERY_TIMEOUT_MS 60000 ++ + /* These definitions are not precise, but they're close enough. (Bits + * 0x03 seem to indicate the aspect ratio of the touch, bits 0x70 seem + * to be some kind of bit mask -- 0x20 may be a near-field reading, +@@ -140,6 +142,7 @@ struct magicmouse_sc { + + struct hid_device *hdev; + struct delayed_work work; ++ struct timer_list battery_timer; + }; + + static int magicmouse_firm_touch(struct magicmouse_sc *msc) +@@ -738,6 +741,44 @@ static void magicmouse_enable_mt_work(struct work_struct *work) + hid_err(msc->hdev, "unable to request touch data (%d)\n", ret); + } + ++static int magicmouse_fetch_battery(struct hid_device *hdev) ++{ ++#ifdef CONFIG_HID_BATTERY_STRENGTH ++ struct hid_report_enum *report_enum; ++ struct hid_report *report; ++ ++ if (!hdev->battery || hdev->vendor != USB_VENDOR_ID_APPLE || ++ (hdev->product != USB_DEVICE_ID_APPLE_MAGICMOUSE2 && ++ hdev->product != USB_DEVICE_ID_APPLE_MAGICTRACKPAD2)) ++ return -1; ++ ++ report_enum = &hdev->report_enum[hdev->battery_report_type]; ++ report = report_enum->report_id_hash[hdev->battery_report_id]; ++ ++ if (!report || report->maxfield < 1) ++ return -1; ++ ++ if (hdev->battery_capacity == hdev->battery_max) ++ return -1; ++ ++ hid_hw_request(hdev, report, HID_REQ_GET_REPORT); ++ return 0; ++#else ++ return -1; ++#endif ++} ++ ++static void magicmouse_battery_timer_tick(struct timer_list *t) ++{ ++ struct magicmouse_sc *msc = from_timer(msc, t, battery_timer); ++ struct hid_device *hdev = msc->hdev; ++ ++ if (magicmouse_fetch_battery(hdev) == 0) { ++ mod_timer(&msc->battery_timer, ++ jiffies + msecs_to_jiffies(USB_BATTERY_TIMEOUT_MS)); ++ } ++} ++ + static int magicmouse_probe(struct hid_device *hdev, + const struct hid_device_id *id) + { +@@ -745,11 +786,6 @@ static int magicmouse_probe(struct hid_device *hdev, + struct hid_report *report; + int ret; + +- if (id->vendor == USB_VENDOR_ID_APPLE && +- id->product == USB_DEVICE_ID_APPLE_MAGICTRACKPAD2 && +- hdev->type != HID_TYPE_USBMOUSE) +- return -ENODEV; +- + msc = devm_kzalloc(&hdev->dev, sizeof(*msc), GFP_KERNEL); + if (msc == NULL) { + hid_err(hdev, "can't alloc magicmouse descriptor\n"); +@@ -775,6 +811,16 @@ static int magicmouse_probe(struct hid_device *hdev, + return ret; + } + ++ timer_setup(&msc->battery_timer, magicmouse_battery_timer_tick, 0); ++ mod_timer(&msc->battery_timer, ++ jiffies + msecs_to_jiffies(USB_BATTERY_TIMEOUT_MS)); ++ magicmouse_fetch_battery(hdev); ++ ++ if (id->vendor == USB_VENDOR_ID_APPLE && ++ (id->product == USB_DEVICE_ID_APPLE_MAGICMOUSE2 || ++ (id->product == USB_DEVICE_ID_APPLE_MAGICTRACKPAD2 && hdev->type != HID_TYPE_USBMOUSE))) ++ return 0; ++ + if (!msc->input) { + hid_err(hdev, "magicmouse input not registered\n"); + ret = -ENOMEM; +@@ -827,6 +873,7 @@ static int magicmouse_probe(struct hid_device *hdev, + + return 0; + err_stop_hw: ++ del_timer_sync(&msc->battery_timer); + hid_hw_stop(hdev); + return ret; + } +@@ -835,17 +882,52 @@ static void magicmouse_remove(struct hid_device *hdev) + { + struct magicmouse_sc *msc = hid_get_drvdata(hdev); + +- if (msc) ++ if (msc) { + cancel_delayed_work_sync(&msc->work); ++ del_timer_sync(&msc->battery_timer); ++ } + + hid_hw_stop(hdev); + } + ++static __u8 *magicmouse_report_fixup(struct hid_device *hdev, __u8 *rdesc, ++ unsigned int *rsize) ++{ ++ /* ++ * Change the usage from: ++ * 0x06, 0x00, 0xff, // Usage Page (Vendor Defined Page 1) 0 ++ * 0x09, 0x0b, // Usage (Vendor Usage 0x0b) 3 ++ * To: ++ * 0x05, 0x01, // Usage Page (Generic Desktop) 0 ++ * 0x09, 0x02, // Usage (Mouse) 2 ++ */ ++ if (hdev->vendor == USB_VENDOR_ID_APPLE && ++ (hdev->product == USB_DEVICE_ID_APPLE_MAGICMOUSE2 || ++ hdev->product == USB_DEVICE_ID_APPLE_MAGICTRACKPAD2) && ++ *rsize == 83 && rdesc[46] == 0x84 && rdesc[58] == 0x85) { ++ hid_info(hdev, ++ "fixing up magicmouse battery report descriptor\n"); ++ *rsize = *rsize - 1; ++ rdesc = kmemdup(rdesc + 1, *rsize, GFP_KERNEL); ++ if (!rdesc) ++ return NULL; ++ ++ rdesc[0] = 0x05; ++ rdesc[1] = 0x01; ++ rdesc[2] = 0x09; ++ rdesc[3] = 0x02; ++ } ++ ++ return rdesc; ++} ++ + static const struct hid_device_id magic_mice[] = { + { HID_BLUETOOTH_DEVICE(USB_VENDOR_ID_APPLE, + USB_DEVICE_ID_APPLE_MAGICMOUSE), .driver_data = 0 }, + { HID_BLUETOOTH_DEVICE(BT_VENDOR_ID_APPLE, + USB_DEVICE_ID_APPLE_MAGICMOUSE2), .driver_data = 0 }, ++ { HID_USB_DEVICE(USB_VENDOR_ID_APPLE, ++ USB_DEVICE_ID_APPLE_MAGICMOUSE2), .driver_data = 0 }, + { HID_BLUETOOTH_DEVICE(USB_VENDOR_ID_APPLE, + USB_DEVICE_ID_APPLE_MAGICTRACKPAD), .driver_data = 0 }, + { HID_BLUETOOTH_DEVICE(BT_VENDOR_ID_APPLE, +@@ -861,6 +943,7 @@ static struct hid_driver magicmouse_driver = { + .id_table = magic_mice, + .probe = magicmouse_probe, + .remove = magicmouse_remove, ++ .report_fixup = magicmouse_report_fixup, + .raw_event = magicmouse_raw_event, + .event = magicmouse_event, + .input_mapping = magicmouse_input_mapping, +diff --git a/drivers/hid/hid-uclogic-params.c b/drivers/hid/hid-uclogic-params.c +index adff1bd68d9f8..3e70f969fb849 100644 +--- a/drivers/hid/hid-uclogic-params.c ++++ b/drivers/hid/hid-uclogic-params.c +@@ -66,7 +66,7 @@ static int uclogic_params_get_str_desc(__u8 **pbuf, struct hid_device *hdev, + __u8 idx, size_t len) + { + int rc; +- struct usb_device *udev = hid_to_usb_dev(hdev); ++ struct usb_device *udev; + __u8 *buf = NULL; + + /* Check arguments */ +@@ -75,6 +75,8 @@ static int uclogic_params_get_str_desc(__u8 **pbuf, struct hid_device *hdev, + goto cleanup; + } + ++ udev = hid_to_usb_dev(hdev); ++ + buf = kmalloc(len, GFP_KERNEL); + if (buf == NULL) { + rc = -ENOMEM; +@@ -450,7 +452,7 @@ static int uclogic_params_frame_init_v1_buttonpad( + { + int rc; + bool found = false; +- struct usb_device *usb_dev = hid_to_usb_dev(hdev); ++ struct usb_device *usb_dev; + char *str_buf = NULL; + const size_t str_len = 16; + +@@ -460,6 +462,8 @@ static int uclogic_params_frame_init_v1_buttonpad( + goto cleanup; + } + ++ usb_dev = hid_to_usb_dev(hdev); ++ + /* + * Enable generic button mode + */ +@@ -707,9 +711,9 @@ static int uclogic_params_huion_init(struct uclogic_params *params, + struct hid_device *hdev) + { + int rc; +- struct usb_device *udev = hid_to_usb_dev(hdev); +- struct usb_interface *iface = to_usb_interface(hdev->dev.parent); +- __u8 bInterfaceNumber = iface->cur_altsetting->desc.bInterfaceNumber; ++ struct usb_device *udev; ++ struct usb_interface *iface; ++ __u8 bInterfaceNumber; + bool found; + /* The resulting parameters (noop) */ + struct uclogic_params p = {0, }; +@@ -723,6 +727,10 @@ static int uclogic_params_huion_init(struct uclogic_params *params, + goto cleanup; + } + ++ udev = hid_to_usb_dev(hdev); ++ iface = to_usb_interface(hdev->dev.parent); ++ bInterfaceNumber = iface->cur_altsetting->desc.bInterfaceNumber; ++ + /* If it's not a pen interface */ + if (bInterfaceNumber != 0) { + /* TODO: Consider marking the interface invalid */ +@@ -834,10 +842,10 @@ int uclogic_params_init(struct uclogic_params *params, + struct hid_device *hdev) + { + int rc; +- struct usb_device *udev = hid_to_usb_dev(hdev); +- __u8 bNumInterfaces = udev->config->desc.bNumInterfaces; +- struct usb_interface *iface = to_usb_interface(hdev->dev.parent); +- __u8 bInterfaceNumber = iface->cur_altsetting->desc.bInterfaceNumber; ++ struct usb_device *udev; ++ __u8 bNumInterfaces; ++ struct usb_interface *iface; ++ __u8 bInterfaceNumber; + bool found; + /* The resulting parameters (noop) */ + struct uclogic_params p = {0, }; +@@ -848,6 +856,11 @@ int uclogic_params_init(struct uclogic_params *params, + goto cleanup; + } + ++ udev = hid_to_usb_dev(hdev); ++ bNumInterfaces = udev->config->desc.bNumInterfaces; ++ iface = to_usb_interface(hdev->dev.parent); ++ bInterfaceNumber = iface->cur_altsetting->desc.bInterfaceNumber; ++ + /* + * Set replacement report descriptor if the original matches the + * specified size. Otherwise keep interface unchanged. +diff --git a/drivers/hid/hid-vivaldi.c b/drivers/hid/hid-vivaldi.c +index 72957a9f71170..576518e704ee6 100644 +--- a/drivers/hid/hid-vivaldi.c ++++ b/drivers/hid/hid-vivaldi.c +@@ -74,10 +74,11 @@ static void vivaldi_feature_mapping(struct hid_device *hdev, + struct hid_usage *usage) + { + struct vivaldi_data *drvdata = hid_get_drvdata(hdev); ++ struct hid_report *report = field->report; + int fn_key; + int ret; + u32 report_len; +- u8 *buf; ++ u8 *report_data, *buf; + + if (field->logical != HID_USAGE_FN_ROW_PHYSMAP || + (usage->hid & HID_USAGE_PAGE) != HID_UP_ORDINAL) +@@ -89,12 +90,24 @@ static void vivaldi_feature_mapping(struct hid_device *hdev, + if (fn_key > drvdata->max_function_row_key) + drvdata->max_function_row_key = fn_key; + +- buf = hid_alloc_report_buf(field->report, GFP_KERNEL); +- if (!buf) ++ report_data = buf = hid_alloc_report_buf(report, GFP_KERNEL); ++ if (!report_data) + return; + +- report_len = hid_report_len(field->report); +- ret = hid_hw_raw_request(hdev, field->report->id, buf, ++ report_len = hid_report_len(report); ++ if (!report->id) { ++ /* ++ * hid_hw_raw_request() will stuff report ID (which will be 0) ++ * into the first byte of the buffer even for unnumbered ++ * reports, so we need to account for this to avoid getting ++ * -EOVERFLOW in return. ++ * Note that hid_alloc_report_buf() adds 7 bytes to the size ++ * so we can safely say that we have space for an extra byte. ++ */ ++ report_len++; ++ } ++ ++ ret = hid_hw_raw_request(hdev, report->id, report_data, + report_len, HID_FEATURE_REPORT, + HID_REQ_GET_REPORT); + if (ret < 0) { +@@ -103,7 +116,16 @@ static void vivaldi_feature_mapping(struct hid_device *hdev, + goto out; + } + +- ret = hid_report_raw_event(hdev, HID_FEATURE_REPORT, buf, ++ if (!report->id) { ++ /* ++ * Undo the damage from hid_hw_raw_request() for unnumbered ++ * reports. ++ */ ++ report_data++; ++ report_len--; ++ } ++ ++ ret = hid_report_raw_event(hdev, HID_FEATURE_REPORT, report_data, + report_len, 0); + if (ret) { + dev_warn(&hdev->dev, "failed to report feature %d\n", +diff --git a/drivers/hid/i2c-hid/i2c-hid-acpi.c b/drivers/hid/i2c-hid/i2c-hid-acpi.c +index a6f0257a26de3..b96ae15e0ad91 100644 +--- a/drivers/hid/i2c-hid/i2c-hid-acpi.c ++++ b/drivers/hid/i2c-hid/i2c-hid-acpi.c +@@ -111,7 +111,7 @@ static int i2c_hid_acpi_probe(struct i2c_client *client) + } + + return i2c_hid_core_probe(client, &ihid_acpi->ops, +- hid_descriptor_address); ++ hid_descriptor_address, 0); + } + + static const struct acpi_device_id i2c_hid_acpi_match[] = { +diff --git a/drivers/hid/i2c-hid/i2c-hid-core.c b/drivers/hid/i2c-hid/i2c-hid-core.c +index 517141138b007..4804d71e5293a 100644 +--- a/drivers/hid/i2c-hid/i2c-hid-core.c ++++ b/drivers/hid/i2c-hid/i2c-hid-core.c +@@ -912,7 +912,7 @@ static void i2c_hid_core_shutdown_tail(struct i2c_hid *ihid) + } + + int i2c_hid_core_probe(struct i2c_client *client, struct i2chid_ops *ops, +- u16 hid_descriptor_address) ++ u16 hid_descriptor_address, u32 quirks) + { + int ret; + struct i2c_hid *ihid; +@@ -1009,6 +1009,8 @@ int i2c_hid_core_probe(struct i2c_client *client, struct i2chid_ops *ops, + goto err_mem_free; + } + ++ hid->quirks |= quirks; ++ + return 0; + + err_mem_free: +diff --git a/drivers/hid/i2c-hid/i2c-hid-of-goodix.c b/drivers/hid/i2c-hid/i2c-hid-of-goodix.c +index 52674149a2750..b4dad66fa954d 100644 +--- a/drivers/hid/i2c-hid/i2c-hid-of-goodix.c ++++ b/drivers/hid/i2c-hid/i2c-hid-of-goodix.c +@@ -150,7 +150,7 @@ static int i2c_hid_of_goodix_probe(struct i2c_client *client, + goodix_i2c_hid_deassert_reset(ihid_goodix, true); + mutex_unlock(&ihid_goodix->regulator_mutex); + +- return i2c_hid_core_probe(client, &ihid_goodix->ops, 0x0001); ++ return i2c_hid_core_probe(client, &ihid_goodix->ops, 0x0001, 0); + } + + static const struct goodix_i2c_hid_timing_data goodix_gt7375p_timing_data = { +diff --git a/drivers/hid/i2c-hid/i2c-hid-of.c b/drivers/hid/i2c-hid/i2c-hid-of.c +index 4bf7cea926379..97a27a803f58d 100644 +--- a/drivers/hid/i2c-hid/i2c-hid-of.c ++++ b/drivers/hid/i2c-hid/i2c-hid-of.c +@@ -21,6 +21,7 @@ + + #include <linux/delay.h> + #include <linux/device.h> ++#include <linux/hid.h> + #include <linux/i2c.h> + #include <linux/kernel.h> + #include <linux/module.h> +@@ -71,6 +72,7 @@ static int i2c_hid_of_probe(struct i2c_client *client, + struct device *dev = &client->dev; + struct i2c_hid_of *ihid_of; + u16 hid_descriptor_address; ++ u32 quirks = 0; + int ret; + u32 val; + +@@ -105,8 +107,14 @@ static int i2c_hid_of_probe(struct i2c_client *client, + if (ret) + return ret; + ++ if (device_property_read_bool(dev, "touchscreen-inverted-x")) ++ quirks |= HID_QUIRK_X_INVERT; ++ ++ if (device_property_read_bool(dev, "touchscreen-inverted-y")) ++ quirks |= HID_QUIRK_Y_INVERT; ++ + return i2c_hid_core_probe(client, &ihid_of->ops, +- hid_descriptor_address); ++ hid_descriptor_address, quirks); + } + + static const struct of_device_id i2c_hid_of_match[] = { +diff --git a/drivers/hid/i2c-hid/i2c-hid.h b/drivers/hid/i2c-hid/i2c-hid.h +index 05a7827d211af..236cc062d5ef8 100644 +--- a/drivers/hid/i2c-hid/i2c-hid.h ++++ b/drivers/hid/i2c-hid/i2c-hid.h +@@ -32,7 +32,7 @@ struct i2chid_ops { + }; + + int i2c_hid_core_probe(struct i2c_client *client, struct i2chid_ops *ops, +- u16 hid_descriptor_address); ++ u16 hid_descriptor_address, u32 quirks); + int i2c_hid_core_remove(struct i2c_client *client); + + void i2c_hid_core_shutdown(struct i2c_client *client); +diff --git a/drivers/hid/uhid.c b/drivers/hid/uhid.c +index 8fe3efcb83271..fc06d8bb42e0f 100644 +--- a/drivers/hid/uhid.c ++++ b/drivers/hid/uhid.c +@@ -28,11 +28,22 @@ + + struct uhid_device { + struct mutex devlock; ++ ++ /* This flag tracks whether the HID device is usable for commands from ++ * userspace. The flag is already set before hid_add_device(), which ++ * runs in workqueue context, to allow hid_add_device() to communicate ++ * with userspace. ++ * However, if hid_add_device() fails, the flag is cleared without ++ * holding devlock. ++ * We guarantee that if @running changes from true to false while you're ++ * holding @devlock, it's still fine to access @hid. ++ */ + bool running; + + __u8 *rd_data; + uint rd_size; + ++ /* When this is NULL, userspace may use UHID_CREATE/UHID_CREATE2. */ + struct hid_device *hid; + struct uhid_event input_buf; + +@@ -63,9 +74,18 @@ static void uhid_device_add_worker(struct work_struct *work) + if (ret) { + hid_err(uhid->hid, "Cannot register HID device: error %d\n", ret); + +- hid_destroy_device(uhid->hid); +- uhid->hid = NULL; ++ /* We used to call hid_destroy_device() here, but that's really ++ * messy to get right because we have to coordinate with ++ * concurrent writes from userspace that might be in the middle ++ * of using uhid->hid. ++ * Just leave uhid->hid as-is for now, and clean it up when ++ * userspace tries to close or reinitialize the uhid instance. ++ * ++ * However, we do have to clear the ->running flag and do a ++ * wakeup to make sure userspace knows that the device is gone. ++ */ + uhid->running = false; ++ wake_up_interruptible(&uhid->report_wait); + } + } + +@@ -474,7 +494,7 @@ static int uhid_dev_create2(struct uhid_device *uhid, + void *rd_data; + int ret; + +- if (uhid->running) ++ if (uhid->hid) + return -EALREADY; + + rd_size = ev->u.create2.rd_size; +@@ -556,7 +576,7 @@ static int uhid_dev_create(struct uhid_device *uhid, + + static int uhid_dev_destroy(struct uhid_device *uhid) + { +- if (!uhid->running) ++ if (!uhid->hid) + return -EINVAL; + + uhid->running = false; +@@ -565,6 +585,7 @@ static int uhid_dev_destroy(struct uhid_device *uhid) + cancel_work_sync(&uhid->worker); + + hid_destroy_device(uhid->hid); ++ uhid->hid = NULL; + kfree(uhid->rd_data); + + return 0; +diff --git a/drivers/hid/wacom_wac.c b/drivers/hid/wacom_wac.c +index 2a4cc39962e76..a7176fc0635dd 100644 +--- a/drivers/hid/wacom_wac.c ++++ b/drivers/hid/wacom_wac.c +@@ -2588,6 +2588,24 @@ static void wacom_wac_finger_slot(struct wacom_wac *wacom_wac, + } + } + ++static bool wacom_wac_slot_is_active(struct input_dev *dev, int key) ++{ ++ struct input_mt *mt = dev->mt; ++ struct input_mt_slot *s; ++ ++ if (!mt) ++ return false; ++ ++ for (s = mt->slots; s != mt->slots + mt->num_slots; s++) { ++ if (s->key == key && ++ input_mt_get_value(s, ABS_MT_TRACKING_ID) >= 0) { ++ return true; ++ } ++ } ++ ++ return false; ++} ++ + static void wacom_wac_finger_event(struct hid_device *hdev, + struct hid_field *field, struct hid_usage *usage, __s32 value) + { +@@ -2638,9 +2656,14 @@ static void wacom_wac_finger_event(struct hid_device *hdev, + } + + if (usage->usage_index + 1 == field->report_count) { +- if (equivalent_usage == wacom_wac->hid_data.last_slot_field && +- wacom_wac->hid_data.confidence) +- wacom_wac_finger_slot(wacom_wac, wacom_wac->touch_input); ++ if (equivalent_usage == wacom_wac->hid_data.last_slot_field) { ++ bool touch_removed = wacom_wac_slot_is_active(wacom_wac->touch_input, ++ wacom_wac->hid_data.id) && !wacom_wac->hid_data.tipswitch; ++ ++ if (wacom_wac->hid_data.confidence || touch_removed) { ++ wacom_wac_finger_slot(wacom_wac, wacom_wac->touch_input); ++ } ++ } + } + } + +@@ -2659,6 +2682,10 @@ static void wacom_wac_finger_pre_report(struct hid_device *hdev, + + hid_data->confidence = true; + ++ hid_data->cc_report = 0; ++ hid_data->cc_index = -1; ++ hid_data->cc_value_index = -1; ++ + for (i = 0; i < report->maxfield; i++) { + struct hid_field *field = report->field[i]; + int j; +@@ -2692,11 +2719,14 @@ static void wacom_wac_finger_pre_report(struct hid_device *hdev, + hid_data->cc_index >= 0) { + struct hid_field *field = report->field[hid_data->cc_index]; + int value = field->value[hid_data->cc_value_index]; +- if (value) ++ if (value) { + hid_data->num_expected = value; ++ hid_data->num_received = 0; ++ } + } + else { + hid_data->num_expected = wacom_wac->features.touch_max; ++ hid_data->num_received = 0; + } + } + +@@ -2724,6 +2754,7 @@ static void wacom_wac_finger_report(struct hid_device *hdev, + + input_sync(input); + wacom_wac->hid_data.num_received = 0; ++ wacom_wac->hid_data.num_expected = 0; + + /* keep touch state for pen event */ + wacom_wac->shared->touch_down = wacom_wac_finger_count_touches(wacom_wac); +diff --git a/drivers/hsi/hsi_core.c b/drivers/hsi/hsi_core.c +index ec90713564e32..884066109699c 100644 +--- a/drivers/hsi/hsi_core.c ++++ b/drivers/hsi/hsi_core.c +@@ -102,6 +102,7 @@ struct hsi_client *hsi_new_client(struct hsi_port *port, + if (device_register(&cl->device) < 0) { + pr_err("hsi: failed to register client: %s\n", info->name); + put_device(&cl->device); ++ goto err; + } + + return cl; +diff --git a/drivers/hwmon/mr75203.c b/drivers/hwmon/mr75203.c +index 868243dba1ee0..1ba1e31459690 100644 +--- a/drivers/hwmon/mr75203.c ++++ b/drivers/hwmon/mr75203.c +@@ -93,7 +93,7 @@ + #define VM_CH_REQ BIT(21) + + #define IP_TMR 0x05 +-#define POWER_DELAY_CYCLE_256 0x80 ++#define POWER_DELAY_CYCLE_256 0x100 + #define POWER_DELAY_CYCLE_64 0x40 + + #define PVT_POLL_DELAY_US 20 +diff --git a/drivers/i2c/busses/i2c-designware-pcidrv.c b/drivers/i2c/busses/i2c-designware-pcidrv.c +index 0f409a4c2da0d..5b45941bcbddc 100644 +--- a/drivers/i2c/busses/i2c-designware-pcidrv.c ++++ b/drivers/i2c/busses/i2c-designware-pcidrv.c +@@ -39,10 +39,10 @@ enum dw_pci_ctl_id_t { + }; + + struct dw_scl_sda_cfg { +- u32 ss_hcnt; +- u32 fs_hcnt; +- u32 ss_lcnt; +- u32 fs_lcnt; ++ u16 ss_hcnt; ++ u16 fs_hcnt; ++ u16 ss_lcnt; ++ u16 fs_lcnt; + u32 sda_hold; + }; + +diff --git a/drivers/i2c/busses/i2c-i801.c b/drivers/i2c/busses/i2c-i801.c +index 1f929e6c30bea..98e39a17fb830 100644 +--- a/drivers/i2c/busses/i2c-i801.c ++++ b/drivers/i2c/busses/i2c-i801.c +@@ -763,6 +763,11 @@ static int i801_block_transaction(struct i801_priv *priv, union i2c_smbus_data * + int result = 0; + unsigned char hostc; + ++ if (read_write == I2C_SMBUS_READ && command == I2C_SMBUS_BLOCK_DATA) ++ data->block[0] = I2C_SMBUS_BLOCK_MAX; ++ else if (data->block[0] < 1 || data->block[0] > I2C_SMBUS_BLOCK_MAX) ++ return -EPROTO; ++ + if (command == I2C_SMBUS_I2C_BLOCK_DATA) { + if (read_write == I2C_SMBUS_WRITE) { + /* set I2C_EN bit in configuration register */ +@@ -776,16 +781,6 @@ static int i801_block_transaction(struct i801_priv *priv, union i2c_smbus_data * + } + } + +- if (read_write == I2C_SMBUS_WRITE +- || command == I2C_SMBUS_I2C_BLOCK_DATA) { +- if (data->block[0] < 1) +- data->block[0] = 1; +- if (data->block[0] > I2C_SMBUS_BLOCK_MAX) +- data->block[0] = I2C_SMBUS_BLOCK_MAX; +- } else { +- data->block[0] = 32; /* max for SMBus block reads */ +- } +- + /* Experience has shown that the block buffer can only be used for + SMBus (not I2C) block transactions, even though the datasheet + doesn't mention this limitation. */ +diff --git a/drivers/i2c/busses/i2c-mpc.c b/drivers/i2c/busses/i2c-mpc.c +index db26cc36e13fe..6c698c10d3cdb 100644 +--- a/drivers/i2c/busses/i2c-mpc.c ++++ b/drivers/i2c/busses/i2c-mpc.c +@@ -119,23 +119,30 @@ static inline void writeccr(struct mpc_i2c *i2c, u32 x) + /* Sometimes 9th clock pulse isn't generated, and slave doesn't release + * the bus, because it wants to send ACK. + * Following sequence of enabling/disabling and sending start/stop generates +- * the 9 pulses, so it's all OK. ++ * the 9 pulses, each with a START then ending with STOP, so it's all OK. + */ + static void mpc_i2c_fixup(struct mpc_i2c *i2c) + { + int k; +- u32 delay_val = 1000000 / i2c->real_clk + 1; +- +- if (delay_val < 2) +- delay_val = 2; ++ unsigned long flags; + + for (k = 9; k; k--) { + writeccr(i2c, 0); +- writeccr(i2c, CCR_MSTA | CCR_MTX | CCR_MEN); ++ writeb(0, i2c->base + MPC_I2C_SR); /* clear any status bits */ ++ writeccr(i2c, CCR_MEN | CCR_MSTA); /* START */ ++ readb(i2c->base + MPC_I2C_DR); /* init xfer */ ++ udelay(15); /* let it hit the bus */ ++ local_irq_save(flags); /* should not be delayed further */ ++ writeccr(i2c, CCR_MEN | CCR_MSTA | CCR_RSTA); /* delay SDA */ + readb(i2c->base + MPC_I2C_DR); +- writeccr(i2c, CCR_MEN); +- udelay(delay_val << 1); ++ if (k != 1) ++ udelay(5); ++ local_irq_restore(flags); + } ++ writeccr(i2c, CCR_MEN); /* Initiate STOP */ ++ readb(i2c->base + MPC_I2C_DR); ++ udelay(15); /* Let STOP propagate */ ++ writeccr(i2c, 0); + } + + static int i2c_mpc_wait_sr(struct mpc_i2c *i2c, int mask) +diff --git a/drivers/iio/adc/ti-adc081c.c b/drivers/iio/adc/ti-adc081c.c +index 16fc608db36a5..bd48b073e7200 100644 +--- a/drivers/iio/adc/ti-adc081c.c ++++ b/drivers/iio/adc/ti-adc081c.c +@@ -19,6 +19,7 @@ + #include <linux/i2c.h> + #include <linux/module.h> + #include <linux/mod_devicetable.h> ++#include <linux/property.h> + + #include <linux/iio/iio.h> + #include <linux/iio/buffer.h> +@@ -156,13 +157,16 @@ static int adc081c_probe(struct i2c_client *client, + { + struct iio_dev *iio; + struct adc081c *adc; +- struct adcxx1c_model *model; ++ const struct adcxx1c_model *model; + int err; + + if (!i2c_check_functionality(client->adapter, I2C_FUNC_SMBUS_WORD_DATA)) + return -EOPNOTSUPP; + +- model = &adcxx1c_models[id->driver_data]; ++ if (dev_fwnode(&client->dev)) ++ model = device_get_match_data(&client->dev); ++ else ++ model = &adcxx1c_models[id->driver_data]; + + iio = devm_iio_device_alloc(&client->dev, sizeof(*adc)); + if (!iio) +@@ -210,10 +214,17 @@ static const struct i2c_device_id adc081c_id[] = { + }; + MODULE_DEVICE_TABLE(i2c, adc081c_id); + ++static const struct acpi_device_id adc081c_acpi_match[] = { ++ /* Used on some AAEON boards */ ++ { "ADC081C", (kernel_ulong_t)&adcxx1c_models[ADC081C] }, ++ { } ++}; ++MODULE_DEVICE_TABLE(acpi, adc081c_acpi_match); ++ + static const struct of_device_id adc081c_of_match[] = { +- { .compatible = "ti,adc081c" }, +- { .compatible = "ti,adc101c" }, +- { .compatible = "ti,adc121c" }, ++ { .compatible = "ti,adc081c", .data = &adcxx1c_models[ADC081C] }, ++ { .compatible = "ti,adc101c", .data = &adcxx1c_models[ADC101C] }, ++ { .compatible = "ti,adc121c", .data = &adcxx1c_models[ADC121C] }, + { } + }; + MODULE_DEVICE_TABLE(of, adc081c_of_match); +@@ -222,6 +233,7 @@ static struct i2c_driver adc081c_driver = { + .driver = { + .name = "adc081c", + .of_match_table = adc081c_of_match, ++ .acpi_match_table = adc081c_acpi_match, + }, + .probe = adc081c_probe, + .id_table = adc081c_id, +diff --git a/drivers/iio/industrialio-trigger.c b/drivers/iio/industrialio-trigger.c +index 93990ff1dfe39..f504ed351b3e2 100644 +--- a/drivers/iio/industrialio-trigger.c ++++ b/drivers/iio/industrialio-trigger.c +@@ -162,6 +162,39 @@ static struct iio_trigger *iio_trigger_acquire_by_name(const char *name) + return trig; + } + ++static void iio_reenable_work_fn(struct work_struct *work) ++{ ++ struct iio_trigger *trig = container_of(work, struct iio_trigger, ++ reenable_work); ++ ++ /* ++ * This 'might' occur after the trigger state is set to disabled - ++ * in that case the driver should skip reenabling. ++ */ ++ trig->ops->reenable(trig); ++} ++ ++/* ++ * In general, reenable callbacks may need to sleep and this path is ++ * not performance sensitive, so just queue up a work item ++ * to reneable the trigger for us. ++ * ++ * Races that can cause this. ++ * 1) A handler occurs entirely in interrupt context so the counter ++ * the final decrement is still in this interrupt. ++ * 2) The trigger has been removed, but one last interrupt gets through. ++ * ++ * For (1) we must call reenable, but not in atomic context. ++ * For (2) it should be safe to call reenanble, if drivers never blindly ++ * reenable after state is off. ++ */ ++static void iio_trigger_notify_done_atomic(struct iio_trigger *trig) ++{ ++ if (atomic_dec_and_test(&trig->use_count) && trig->ops && ++ trig->ops->reenable) ++ schedule_work(&trig->reenable_work); ++} ++ + void iio_trigger_poll(struct iio_trigger *trig) + { + int i; +@@ -173,7 +206,7 @@ void iio_trigger_poll(struct iio_trigger *trig) + if (trig->subirqs[i].enabled) + generic_handle_irq(trig->subirq_base + i); + else +- iio_trigger_notify_done(trig); ++ iio_trigger_notify_done_atomic(trig); + } + } + } +@@ -535,6 +568,7 @@ struct iio_trigger *viio_trigger_alloc(struct device *parent, + trig->dev.type = &iio_trig_type; + trig->dev.bus = &iio_bus_type; + device_initialize(&trig->dev); ++ INIT_WORK(&trig->reenable_work, iio_reenable_work_fn); + + mutex_init(&trig->pool_lock); + trig->subirq_base = irq_alloc_descs(-1, 0, +diff --git a/drivers/infiniband/core/cma.c b/drivers/infiniband/core/cma.c +index 704ce595542c5..3d419741e8916 100644 +--- a/drivers/infiniband/core/cma.c ++++ b/drivers/infiniband/core/cma.c +@@ -766,6 +766,7 @@ static int cma_resolve_ib_dev(struct rdma_id_private *id_priv) + unsigned int p; + u16 pkey, index; + enum ib_port_state port_state; ++ int ret; + int i; + + cma_dev = NULL; +@@ -784,9 +785,14 @@ static int cma_resolve_ib_dev(struct rdma_id_private *id_priv) + + if (ib_get_cached_port_state(cur_dev->device, p, &port_state)) + continue; +- for (i = 0; !rdma_query_gid(cur_dev->device, +- p, i, &gid); +- i++) { ++ ++ for (i = 0; i < cur_dev->device->port_data[p].immutable.gid_tbl_len; ++ ++i) { ++ ret = rdma_query_gid(cur_dev->device, p, i, ++ &gid); ++ if (ret) ++ continue; ++ + if (!memcmp(&gid, dgid, sizeof(gid))) { + cma_dev = cur_dev; + sgid = gid; +@@ -4031,8 +4037,7 @@ static int cma_resolve_ib_udp(struct rdma_id_private *id_priv, + + memset(&req, 0, sizeof req); + offset = cma_user_data_offset(id_priv); +- req.private_data_len = offset + conn_param->private_data_len; +- if (req.private_data_len < conn_param->private_data_len) ++ if (check_add_overflow(offset, conn_param->private_data_len, &req.private_data_len)) + return -EINVAL; + + if (req.private_data_len) { +@@ -4091,8 +4096,7 @@ static int cma_connect_ib(struct rdma_id_private *id_priv, + + memset(&req, 0, sizeof req); + offset = cma_user_data_offset(id_priv); +- req.private_data_len = offset + conn_param->private_data_len; +- if (req.private_data_len < conn_param->private_data_len) ++ if (check_add_overflow(offset, conn_param->private_data_len, &req.private_data_len)) + return -EINVAL; + + if (req.private_data_len) { +diff --git a/drivers/infiniband/core/device.c b/drivers/infiniband/core/device.c +index f4814bb7f082f..6ab46648af909 100644 +--- a/drivers/infiniband/core/device.c ++++ b/drivers/infiniband/core/device.c +@@ -2461,7 +2461,8 @@ int ib_find_gid(struct ib_device *device, union ib_gid *gid, + ++i) { + ret = rdma_query_gid(device, port, i, &tmp_gid); + if (ret) +- return ret; ++ continue; ++ + if (!memcmp(&tmp_gid, gid, sizeof *gid)) { + *port_num = port; + if (index) +diff --git a/drivers/infiniband/hw/bnxt_re/qplib_rcfw.c b/drivers/infiniband/hw/bnxt_re/qplib_rcfw.c +index 5d384def5e5fe..d2d39126f1852 100644 +--- a/drivers/infiniband/hw/bnxt_re/qplib_rcfw.c ++++ b/drivers/infiniband/hw/bnxt_re/qplib_rcfw.c +@@ -618,8 +618,6 @@ int bnxt_qplib_alloc_rcfw_channel(struct bnxt_qplib_res *res, + if (!cmdq->cmdq_bitmap) + goto fail; + +- cmdq->bmap_size = bmap_size; +- + /* Allocate one extra to hold the QP1 entries */ + rcfw->qp_tbl_size = qp_tbl_sz + 1; + rcfw->qp_tbl = kcalloc(rcfw->qp_tbl_size, sizeof(struct bnxt_qplib_qp_node), +@@ -667,8 +665,8 @@ void bnxt_qplib_disable_rcfw_channel(struct bnxt_qplib_rcfw *rcfw) + iounmap(cmdq->cmdq_mbox.reg.bar_reg); + iounmap(creq->creq_db.reg.bar_reg); + +- indx = find_first_bit(cmdq->cmdq_bitmap, cmdq->bmap_size); +- if (indx != cmdq->bmap_size) ++ indx = find_first_bit(cmdq->cmdq_bitmap, rcfw->cmdq_depth); ++ if (indx != rcfw->cmdq_depth) + dev_err(&rcfw->pdev->dev, + "disabling RCFW with pending cmd-bit %lx\n", indx); + +diff --git a/drivers/infiniband/hw/bnxt_re/qplib_rcfw.h b/drivers/infiniband/hw/bnxt_re/qplib_rcfw.h +index 9474c00465821..0c6d0b70ce890 100644 +--- a/drivers/infiniband/hw/bnxt_re/qplib_rcfw.h ++++ b/drivers/infiniband/hw/bnxt_re/qplib_rcfw.h +@@ -152,7 +152,6 @@ struct bnxt_qplib_cmdq_ctx { + wait_queue_head_t waitq; + unsigned long flags; + unsigned long *cmdq_bitmap; +- u32 bmap_size; + u32 seq_num; + }; + +diff --git a/drivers/infiniband/hw/cxgb4/qp.c b/drivers/infiniband/hw/cxgb4/qp.c +index d20b4ef2c853d..ffbd9a89981e7 100644 +--- a/drivers/infiniband/hw/cxgb4/qp.c ++++ b/drivers/infiniband/hw/cxgb4/qp.c +@@ -2460,6 +2460,7 @@ int c4iw_ib_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, + memset(attr, 0, sizeof(*attr)); + memset(init_attr, 0, sizeof(*init_attr)); + attr->qp_state = to_ib_qp_state(qhp->attr.state); ++ attr->cur_qp_state = to_ib_qp_state(qhp->attr.state); + init_attr->cap.max_send_wr = qhp->attr.sq_num_entries; + init_attr->cap.max_recv_wr = qhp->attr.rq_num_entries; + init_attr->cap.max_send_sge = qhp->attr.sq_max_sges; +diff --git a/drivers/infiniband/hw/hns/hns_roce_main.c b/drivers/infiniband/hw/hns/hns_roce_main.c +index 5d39bd08582af..1f2209de88122 100644 +--- a/drivers/infiniband/hw/hns/hns_roce_main.c ++++ b/drivers/infiniband/hw/hns/hns_roce_main.c +@@ -269,6 +269,9 @@ static enum rdma_link_layer hns_roce_get_link_layer(struct ib_device *device, + static int hns_roce_query_pkey(struct ib_device *ib_dev, u32 port, u16 index, + u16 *pkey) + { ++ if (index > 0) ++ return -EINVAL; ++ + *pkey = PKEY_ID; + + return 0; +@@ -349,7 +352,7 @@ static int hns_roce_mmap(struct ib_ucontext *context, + return rdma_user_mmap_io(context, vma, + to_hr_ucontext(context)->uar.pfn, + PAGE_SIZE, +- pgprot_noncached(vma->vm_page_prot), ++ pgprot_device(vma->vm_page_prot), + NULL); + + /* vm_pgoff: 1 -- TPTR */ +diff --git a/drivers/infiniband/hw/qedr/verbs.c b/drivers/infiniband/hw/qedr/verbs.c +index 3d4e4a766574a..f652d083ff20f 100644 +--- a/drivers/infiniband/hw/qedr/verbs.c ++++ b/drivers/infiniband/hw/qedr/verbs.c +@@ -1941,6 +1941,7 @@ static int qedr_create_user_qp(struct qedr_dev *dev, + /* db offset was calculated in copy_qp_uresp, now set in the user q */ + if (qedr_qp_has_sq(qp)) { + qp->usq.db_addr = ctx->dpi_addr + uresp.sq_db_offset; ++ qp->sq.max_wr = attrs->cap.max_send_wr; + rc = qedr_db_recovery_add(dev, qp->usq.db_addr, + &qp->usq.db_rec_data->db_data, + DB_REC_WIDTH_32B, +@@ -1951,6 +1952,7 @@ static int qedr_create_user_qp(struct qedr_dev *dev, + + if (qedr_qp_has_rq(qp)) { + qp->urq.db_addr = ctx->dpi_addr + uresp.rq_db_offset; ++ qp->rq.max_wr = attrs->cap.max_recv_wr; + rc = qedr_db_recovery_add(dev, qp->urq.db_addr, + &qp->urq.db_rec_data->db_data, + DB_REC_WIDTH_32B, +diff --git a/drivers/infiniband/sw/rxe/rxe_opcode.c b/drivers/infiniband/sw/rxe/rxe_opcode.c +index 3ef5a10a6efd8..47ebaac8f4754 100644 +--- a/drivers/infiniband/sw/rxe/rxe_opcode.c ++++ b/drivers/infiniband/sw/rxe/rxe_opcode.c +@@ -117,7 +117,7 @@ struct rxe_opcode_info rxe_opcode[RXE_NUM_OPCODE] = { + } + }, + [IB_OPCODE_RC_SEND_MIDDLE] = { +- .name = "IB_OPCODE_RC_SEND_MIDDLE]", ++ .name = "IB_OPCODE_RC_SEND_MIDDLE", + .mask = RXE_PAYLOAD_MASK | RXE_REQ_MASK | RXE_SEND_MASK + | RXE_MIDDLE_MASK, + .length = RXE_BTH_BYTES, +diff --git a/drivers/infiniband/ulp/rtrs/rtrs-clt.c b/drivers/infiniband/ulp/rtrs/rtrs-clt.c +index bc8824b4ee0d4..55ebe01ec9951 100644 +--- a/drivers/infiniband/ulp/rtrs/rtrs-clt.c ++++ b/drivers/infiniband/ulp/rtrs/rtrs-clt.c +@@ -867,7 +867,7 @@ static struct rtrs_clt_sess *get_next_path_min_latency(struct path_it *it) + struct rtrs_clt_sess *min_path = NULL; + struct rtrs_clt *clt = it->clt; + struct rtrs_clt_sess *sess; +- ktime_t min_latency = INT_MAX; ++ ktime_t min_latency = KTIME_MAX; + ktime_t latency; + + list_for_each_entry_rcu(sess, &clt->paths_list, s.entry) { +diff --git a/drivers/interconnect/qcom/icc-rpm.c b/drivers/interconnect/qcom/icc-rpm.c +index 54de49ca7808a..ddf1805ded0c0 100644 +--- a/drivers/interconnect/qcom/icc-rpm.c ++++ b/drivers/interconnect/qcom/icc-rpm.c +@@ -68,6 +68,7 @@ static int qcom_icc_set(struct icc_node *src, struct icc_node *dst) + rate = max(sum_bw, max_peak_bw); + + do_div(rate, qn->buswidth); ++ rate = min_t(u64, rate, LONG_MAX); + + if (qn->rate == rate) + return 0; +diff --git a/drivers/iommu/amd/amd_iommu_types.h b/drivers/iommu/amd/amd_iommu_types.h +index 8dbe61e2b3c15..8394c2787ff89 100644 +--- a/drivers/iommu/amd/amd_iommu_types.h ++++ b/drivers/iommu/amd/amd_iommu_types.h +@@ -643,8 +643,6 @@ struct amd_iommu { + /* DebugFS Info */ + struct dentry *debugfs; + #endif +- /* IRQ notifier for IntCapXT interrupt */ +- struct irq_affinity_notify intcapxt_notify; + }; + + static inline struct amd_iommu *dev_to_amd_iommu(struct device *dev) +diff --git a/drivers/iommu/amd/init.c b/drivers/iommu/amd/init.c +index 2a822b229bd05..50e5a728f12a1 100644 +--- a/drivers/iommu/amd/init.c ++++ b/drivers/iommu/amd/init.c +@@ -804,16 +804,27 @@ static int iommu_ga_log_enable(struct amd_iommu *iommu) + { + #ifdef CONFIG_IRQ_REMAP + u32 status, i; ++ u64 entry; + + if (!iommu->ga_log) + return -EINVAL; + +- status = readl(iommu->mmio_base + MMIO_STATUS_OFFSET); +- + /* Check if already running */ +- if (status & (MMIO_STATUS_GALOG_RUN_MASK)) ++ status = readl(iommu->mmio_base + MMIO_STATUS_OFFSET); ++ if (WARN_ON(status & (MMIO_STATUS_GALOG_RUN_MASK))) + return 0; + ++ entry = iommu_virt_to_phys(iommu->ga_log) | GA_LOG_SIZE_512; ++ memcpy_toio(iommu->mmio_base + MMIO_GA_LOG_BASE_OFFSET, ++ &entry, sizeof(entry)); ++ entry = (iommu_virt_to_phys(iommu->ga_log_tail) & ++ (BIT_ULL(52)-1)) & ~7ULL; ++ memcpy_toio(iommu->mmio_base + MMIO_GA_LOG_TAIL_OFFSET, ++ &entry, sizeof(entry)); ++ writel(0x00, iommu->mmio_base + MMIO_GA_HEAD_OFFSET); ++ writel(0x00, iommu->mmio_base + MMIO_GA_TAIL_OFFSET); ++ ++ + iommu_feature_enable(iommu, CONTROL_GAINT_EN); + iommu_feature_enable(iommu, CONTROL_GALOG_EN); + +@@ -823,7 +834,7 @@ static int iommu_ga_log_enable(struct amd_iommu *iommu) + break; + } + +- if (i >= LOOP_TIMEOUT) ++ if (WARN_ON(i >= LOOP_TIMEOUT)) + return -EINVAL; + #endif /* CONFIG_IRQ_REMAP */ + return 0; +@@ -832,8 +843,6 @@ static int iommu_ga_log_enable(struct amd_iommu *iommu) + static int iommu_init_ga_log(struct amd_iommu *iommu) + { + #ifdef CONFIG_IRQ_REMAP +- u64 entry; +- + if (!AMD_IOMMU_GUEST_IR_VAPIC(amd_iommu_guest_ir)) + return 0; + +@@ -847,16 +856,6 @@ static int iommu_init_ga_log(struct amd_iommu *iommu) + if (!iommu->ga_log_tail) + goto err_out; + +- entry = iommu_virt_to_phys(iommu->ga_log) | GA_LOG_SIZE_512; +- memcpy_toio(iommu->mmio_base + MMIO_GA_LOG_BASE_OFFSET, +- &entry, sizeof(entry)); +- entry = (iommu_virt_to_phys(iommu->ga_log_tail) & +- (BIT_ULL(52)-1)) & ~7ULL; +- memcpy_toio(iommu->mmio_base + MMIO_GA_LOG_TAIL_OFFSET, +- &entry, sizeof(entry)); +- writel(0x00, iommu->mmio_base + MMIO_GA_HEAD_OFFSET); +- writel(0x00, iommu->mmio_base + MMIO_GA_TAIL_OFFSET); +- + return 0; + err_out: + free_ga_log(iommu); +@@ -2013,48 +2012,18 @@ union intcapxt { + }; + } __attribute__ ((packed)); + +-/* +- * There isn't really any need to mask/unmask at the irqchip level because +- * the 64-bit INTCAPXT registers can be updated atomically without tearing +- * when the affinity is being updated. +- */ +-static void intcapxt_unmask_irq(struct irq_data *data) +-{ +-} +- +-static void intcapxt_mask_irq(struct irq_data *data) +-{ +-} + + static struct irq_chip intcapxt_controller; + + static int intcapxt_irqdomain_activate(struct irq_domain *domain, + struct irq_data *irqd, bool reserve) + { +- struct amd_iommu *iommu = irqd->chip_data; +- struct irq_cfg *cfg = irqd_cfg(irqd); +- union intcapxt xt; +- +- xt.capxt = 0ULL; +- xt.dest_mode_logical = apic->dest_mode_logical; +- xt.vector = cfg->vector; +- xt.destid_0_23 = cfg->dest_apicid & GENMASK(23, 0); +- xt.destid_24_31 = cfg->dest_apicid >> 24; +- +- /** +- * Current IOMMU implemtation uses the same IRQ for all +- * 3 IOMMU interrupts. +- */ +- writeq(xt.capxt, iommu->mmio_base + MMIO_INTCAPXT_EVT_OFFSET); +- writeq(xt.capxt, iommu->mmio_base + MMIO_INTCAPXT_PPR_OFFSET); +- writeq(xt.capxt, iommu->mmio_base + MMIO_INTCAPXT_GALOG_OFFSET); + return 0; + } + + static void intcapxt_irqdomain_deactivate(struct irq_domain *domain, + struct irq_data *irqd) + { +- intcapxt_mask_irq(irqd); + } + + +@@ -2088,6 +2057,38 @@ static void intcapxt_irqdomain_free(struct irq_domain *domain, unsigned int virq + irq_domain_free_irqs_top(domain, virq, nr_irqs); + } + ++ ++static void intcapxt_unmask_irq(struct irq_data *irqd) ++{ ++ struct amd_iommu *iommu = irqd->chip_data; ++ struct irq_cfg *cfg = irqd_cfg(irqd); ++ union intcapxt xt; ++ ++ xt.capxt = 0ULL; ++ xt.dest_mode_logical = apic->dest_mode_logical; ++ xt.vector = cfg->vector; ++ xt.destid_0_23 = cfg->dest_apicid & GENMASK(23, 0); ++ xt.destid_24_31 = cfg->dest_apicid >> 24; ++ ++ /** ++ * Current IOMMU implementation uses the same IRQ for all ++ * 3 IOMMU interrupts. ++ */ ++ writeq(xt.capxt, iommu->mmio_base + MMIO_INTCAPXT_EVT_OFFSET); ++ writeq(xt.capxt, iommu->mmio_base + MMIO_INTCAPXT_PPR_OFFSET); ++ writeq(xt.capxt, iommu->mmio_base + MMIO_INTCAPXT_GALOG_OFFSET); ++} ++ ++static void intcapxt_mask_irq(struct irq_data *irqd) ++{ ++ struct amd_iommu *iommu = irqd->chip_data; ++ ++ writeq(0, iommu->mmio_base + MMIO_INTCAPXT_EVT_OFFSET); ++ writeq(0, iommu->mmio_base + MMIO_INTCAPXT_PPR_OFFSET); ++ writeq(0, iommu->mmio_base + MMIO_INTCAPXT_GALOG_OFFSET); ++} ++ ++ + static int intcapxt_set_affinity(struct irq_data *irqd, + const struct cpumask *mask, bool force) + { +@@ -2097,8 +2098,12 @@ static int intcapxt_set_affinity(struct irq_data *irqd, + ret = parent->chip->irq_set_affinity(parent, mask, force); + if (ret < 0 || ret == IRQ_SET_MASK_OK_DONE) + return ret; ++ return 0; ++} + +- return intcapxt_irqdomain_activate(irqd->domain, irqd, false); ++static int intcapxt_set_wake(struct irq_data *irqd, unsigned int on) ++{ ++ return on ? -EOPNOTSUPP : 0; + } + + static struct irq_chip intcapxt_controller = { +@@ -2108,7 +2113,8 @@ static struct irq_chip intcapxt_controller = { + .irq_ack = irq_chip_ack_parent, + .irq_retrigger = irq_chip_retrigger_hierarchy, + .irq_set_affinity = intcapxt_set_affinity, +- .flags = IRQCHIP_SKIP_SET_WAKE, ++ .irq_set_wake = intcapxt_set_wake, ++ .flags = IRQCHIP_MASK_ON_SUSPEND, + }; + + static const struct irq_domain_ops intcapxt_domain_ops = { +@@ -2170,7 +2176,6 @@ static int iommu_setup_intcapxt(struct amd_iommu *iommu) + return ret; + } + +- iommu_feature_enable(iommu, CONTROL_INTCAPXT_EN); + return 0; + } + +@@ -2193,6 +2198,10 @@ static int iommu_init_irq(struct amd_iommu *iommu) + + iommu->int_enabled = true; + enable_faults: ++ ++ if (amd_iommu_xt_mode == IRQ_REMAP_X2APIC_MODE) ++ iommu_feature_enable(iommu, CONTROL_INTCAPXT_EN); ++ + iommu_feature_enable(iommu, CONTROL_EVT_INT_EN); + + if (iommu->ppr_log != NULL) +diff --git a/drivers/iommu/arm/arm-smmu/arm-smmu-qcom.c b/drivers/iommu/arm/arm-smmu/arm-smmu-qcom.c +index 55690af1b25d0..c998960495b4e 100644 +--- a/drivers/iommu/arm/arm-smmu/arm-smmu-qcom.c ++++ b/drivers/iommu/arm/arm-smmu/arm-smmu-qcom.c +@@ -51,7 +51,7 @@ static void qcom_adreno_smmu_get_fault_info(const void *cookie, + info->fsynr1 = arm_smmu_cb_read(smmu, cfg->cbndx, ARM_SMMU_CB_FSYNR1); + info->far = arm_smmu_cb_readq(smmu, cfg->cbndx, ARM_SMMU_CB_FAR); + info->cbfrsynra = arm_smmu_gr1_read(smmu, ARM_SMMU_GR1_CBFRSYNRA(cfg->cbndx)); +- info->ttbr0 = arm_smmu_cb_read(smmu, cfg->cbndx, ARM_SMMU_CB_TTBR0); ++ info->ttbr0 = arm_smmu_cb_readq(smmu, cfg->cbndx, ARM_SMMU_CB_TTBR0); + info->contextidr = arm_smmu_cb_read(smmu, cfg->cbndx, ARM_SMMU_CB_CONTEXTIDR); + } + +diff --git a/drivers/iommu/io-pgtable-arm-v7s.c b/drivers/iommu/io-pgtable-arm-v7s.c +index bfb6acb651e5f..be066c1503d37 100644 +--- a/drivers/iommu/io-pgtable-arm-v7s.c ++++ b/drivers/iommu/io-pgtable-arm-v7s.c +@@ -246,13 +246,17 @@ static void *__arm_v7s_alloc_table(int lvl, gfp_t gfp, + __GFP_ZERO | ARM_V7S_TABLE_GFP_DMA, get_order(size)); + else if (lvl == 2) + table = kmem_cache_zalloc(data->l2_tables, gfp); ++ ++ if (!table) ++ return NULL; ++ + phys = virt_to_phys(table); + if (phys != (arm_v7s_iopte)phys) { + /* Doesn't fit in PTE */ + dev_err(dev, "Page table does not fit in PTE: %pa", &phys); + goto out_free; + } +- if (table && !cfg->coherent_walk) { ++ if (!cfg->coherent_walk) { + dma = dma_map_single(dev, table, size, DMA_TO_DEVICE); + if (dma_mapping_error(dev, dma)) + goto out_free; +diff --git a/drivers/iommu/io-pgtable-arm.c b/drivers/iommu/io-pgtable-arm.c +index dd9e47189d0d9..94ff319ae8acc 100644 +--- a/drivers/iommu/io-pgtable-arm.c ++++ b/drivers/iommu/io-pgtable-arm.c +@@ -315,11 +315,12 @@ static int arm_lpae_init_pte(struct arm_lpae_io_pgtable *data, + static arm_lpae_iopte arm_lpae_install_table(arm_lpae_iopte *table, + arm_lpae_iopte *ptep, + arm_lpae_iopte curr, +- struct io_pgtable_cfg *cfg) ++ struct arm_lpae_io_pgtable *data) + { + arm_lpae_iopte old, new; ++ struct io_pgtable_cfg *cfg = &data->iop.cfg; + +- new = __pa(table) | ARM_LPAE_PTE_TYPE_TABLE; ++ new = paddr_to_iopte(__pa(table), data) | ARM_LPAE_PTE_TYPE_TABLE; + if (cfg->quirks & IO_PGTABLE_QUIRK_ARM_NS) + new |= ARM_LPAE_PTE_NSTABLE; + +@@ -380,7 +381,7 @@ static int __arm_lpae_map(struct arm_lpae_io_pgtable *data, unsigned long iova, + if (!cptep) + return -ENOMEM; + +- pte = arm_lpae_install_table(cptep, ptep, 0, cfg); ++ pte = arm_lpae_install_table(cptep, ptep, 0, data); + if (pte) + __arm_lpae_free_pages(cptep, tblsz, cfg); + } else if (!cfg->coherent_walk && !(pte & ARM_LPAE_PTE_SW_SYNC)) { +@@ -592,7 +593,7 @@ static size_t arm_lpae_split_blk_unmap(struct arm_lpae_io_pgtable *data, + __arm_lpae_init_pte(data, blk_paddr, pte, lvl, 1, &tablep[i]); + } + +- pte = arm_lpae_install_table(tablep, ptep, blk_pte, cfg); ++ pte = arm_lpae_install_table(tablep, ptep, blk_pte, data); + if (pte != blk_pte) { + __arm_lpae_free_pages(tablep, tablesz, cfg); + /* +diff --git a/drivers/iommu/iommu.c b/drivers/iommu/iommu.c +index 3303d707bab4b..f62fb6a58f109 100644 +--- a/drivers/iommu/iommu.c ++++ b/drivers/iommu/iommu.c +@@ -287,11 +287,11 @@ int iommu_probe_device(struct device *dev) + */ + mutex_lock(&group->mutex); + iommu_alloc_default_domain(group, dev); +- mutex_unlock(&group->mutex); + + if (group->default_domain) { + ret = __iommu_attach_device(group->default_domain, dev); + if (ret) { ++ mutex_unlock(&group->mutex); + iommu_group_put(group); + goto err_release; + } +@@ -299,6 +299,7 @@ int iommu_probe_device(struct device *dev) + + iommu_create_device_direct_mappings(group, dev); + ++ mutex_unlock(&group->mutex); + iommu_group_put(group); + + if (ops->probe_finalize) +diff --git a/drivers/iommu/iova.c b/drivers/iommu/iova.c +index 9e8bc802ac053..920fcc27c9a1e 100644 +--- a/drivers/iommu/iova.c ++++ b/drivers/iommu/iova.c +@@ -83,8 +83,7 @@ static void free_iova_flush_queue(struct iova_domain *iovad) + if (!has_iova_flush_queue(iovad)) + return; + +- if (timer_pending(&iovad->fq_timer)) +- del_timer(&iovad->fq_timer); ++ del_timer_sync(&iovad->fq_timer); + + fq_destroy_all_entries(iovad); + +diff --git a/drivers/irqchip/irq-gic-v3.c b/drivers/irqchip/irq-gic-v3.c +index fd4e9a37fea67..7bbccb13b896b 100644 +--- a/drivers/irqchip/irq-gic-v3.c ++++ b/drivers/irqchip/irq-gic-v3.c +@@ -920,6 +920,22 @@ static int __gic_update_rdist_properties(struct redist_region *region, + { + u64 typer = gic_read_typer(ptr + GICR_TYPER); + ++ /* Boot-time cleanip */ ++ if ((typer & GICR_TYPER_VLPIS) && (typer & GICR_TYPER_RVPEID)) { ++ u64 val; ++ ++ /* Deactivate any present vPE */ ++ val = gicr_read_vpendbaser(ptr + SZ_128K + GICR_VPENDBASER); ++ if (val & GICR_VPENDBASER_Valid) ++ gicr_write_vpendbaser(GICR_VPENDBASER_PendingLast, ++ ptr + SZ_128K + GICR_VPENDBASER); ++ ++ /* Mark the VPE table as invalid */ ++ val = gicr_read_vpropbaser(ptr + SZ_128K + GICR_VPROPBASER); ++ val &= ~GICR_VPROPBASER_4_1_VALID; ++ gicr_write_vpropbaser(val, ptr + SZ_128K + GICR_VPROPBASER); ++ } ++ + gic_data.rdists.has_vlpis &= !!(typer & GICR_TYPER_VLPIS); + + /* RVPEID implies some form of DirectLPI, no matter what the doc says... :-/ */ +diff --git a/drivers/leds/leds-lp55xx-common.c b/drivers/leds/leds-lp55xx-common.c +index d1657c46ee2f8..9fdfc1b9a1a0c 100644 +--- a/drivers/leds/leds-lp55xx-common.c ++++ b/drivers/leds/leds-lp55xx-common.c +@@ -439,6 +439,8 @@ int lp55xx_init_device(struct lp55xx_chip *chip) + return -EINVAL; + + if (pdata->enable_gpiod) { ++ gpiod_direction_output(pdata->enable_gpiod, 0); ++ + gpiod_set_consumer_name(pdata->enable_gpiod, "LP55xx enable"); + gpiod_set_value(pdata->enable_gpiod, 0); + usleep_range(1000, 2000); /* Keep enable down at least 1ms */ +@@ -694,7 +696,7 @@ struct lp55xx_platform_data *lp55xx_of_populate_pdata(struct device *dev, + of_property_read_u8(np, "clock-mode", &pdata->clock_mode); + + pdata->enable_gpiod = devm_gpiod_get_optional(dev, "enable", +- GPIOD_OUT_LOW); ++ GPIOD_ASIS); + if (IS_ERR(pdata->enable_gpiod)) + return ERR_CAST(pdata->enable_gpiod); + +diff --git a/drivers/mailbox/mailbox-mpfs.c b/drivers/mailbox/mailbox-mpfs.c +index 0d6e2231a2c75..4e34854d12389 100644 +--- a/drivers/mailbox/mailbox-mpfs.c ++++ b/drivers/mailbox/mailbox-mpfs.c +@@ -232,7 +232,7 @@ static int mpfs_mbox_probe(struct platform_device *pdev) + } + + static const struct of_device_id mpfs_mbox_of_match[] = { +- {.compatible = "microchip,polarfire-soc-mailbox", }, ++ {.compatible = "microchip,mpfs-mailbox", }, + {}, + }; + MODULE_DEVICE_TABLE(of, mpfs_mbox_of_match); +diff --git a/drivers/mailbox/mtk-cmdq-mailbox.c b/drivers/mailbox/mtk-cmdq-mailbox.c +index bb4793c7b38fd..3583c2aad0edc 100644 +--- a/drivers/mailbox/mtk-cmdq-mailbox.c ++++ b/drivers/mailbox/mtk-cmdq-mailbox.c +@@ -660,7 +660,7 @@ static const struct gce_plat gce_plat_v5 = { + .thread_nr = 24, + .shift = 3, + .control_by_sw = true, +- .gce_num = 2 ++ .gce_num = 1 + }; + + static const struct gce_plat gce_plat_v6 = { +diff --git a/drivers/md/dm.c b/drivers/md/dm.c +index 76d9da49fda75..671bb454f1649 100644 +--- a/drivers/md/dm.c ++++ b/drivers/md/dm.c +@@ -1794,8 +1794,10 @@ static struct mapped_device *alloc_dev(int minor) + if (IS_ENABLED(CONFIG_DAX_DRIVER)) { + md->dax_dev = alloc_dax(md, md->disk->disk_name, + &dm_dax_ops, 0); +- if (IS_ERR(md->dax_dev)) ++ if (IS_ERR(md->dax_dev)) { ++ md->dax_dev = NULL; + goto bad; ++ } + } + + format_dev_t(md->name, MKDEV(_major, minor)); +diff --git a/drivers/md/md.c b/drivers/md/md.c +index 44006b860d0a5..2d31a079be33f 100644 +--- a/drivers/md/md.c ++++ b/drivers/md/md.c +@@ -5872,13 +5872,6 @@ int md_run(struct mddev *mddev) + if (err) + goto exit_bio_set; + } +- if (mddev->level != 1 && mddev->level != 10 && +- !bioset_initialized(&mddev->io_acct_set)) { +- err = bioset_init(&mddev->io_acct_set, BIO_POOL_SIZE, +- offsetof(struct md_io_acct, bio_clone), 0); +- if (err) +- goto exit_sync_set; +- } + + spin_lock(&pers_lock); + pers = find_pers(mddev->level, mddev->clevel); +@@ -6055,9 +6048,6 @@ bitmap_abort: + module_put(pers->owner); + md_bitmap_destroy(mddev); + abort: +- if (mddev->level != 1 && mddev->level != 10) +- bioset_exit(&mddev->io_acct_set); +-exit_sync_set: + bioset_exit(&mddev->sync_set); + exit_bio_set: + bioset_exit(&mddev->bio_set); +@@ -8590,6 +8580,23 @@ void md_submit_discard_bio(struct mddev *mddev, struct md_rdev *rdev, + } + EXPORT_SYMBOL_GPL(md_submit_discard_bio); + ++int acct_bioset_init(struct mddev *mddev) ++{ ++ int err = 0; ++ ++ if (!bioset_initialized(&mddev->io_acct_set)) ++ err = bioset_init(&mddev->io_acct_set, BIO_POOL_SIZE, ++ offsetof(struct md_io_acct, bio_clone), 0); ++ return err; ++} ++EXPORT_SYMBOL_GPL(acct_bioset_init); ++ ++void acct_bioset_exit(struct mddev *mddev) ++{ ++ bioset_exit(&mddev->io_acct_set); ++} ++EXPORT_SYMBOL_GPL(acct_bioset_exit); ++ + static void md_end_io_acct(struct bio *bio) + { + struct md_io_acct *md_io_acct = bio->bi_private; +diff --git a/drivers/md/md.h b/drivers/md/md.h +index 4c96c36bd01a1..62852d7011457 100644 +--- a/drivers/md/md.h ++++ b/drivers/md/md.h +@@ -721,6 +721,8 @@ extern void md_error(struct mddev *mddev, struct md_rdev *rdev); + extern void md_finish_reshape(struct mddev *mddev); + void md_submit_discard_bio(struct mddev *mddev, struct md_rdev *rdev, + struct bio *bio, sector_t start, sector_t size); ++int acct_bioset_init(struct mddev *mddev); ++void acct_bioset_exit(struct mddev *mddev); + void md_account_bio(struct mddev *mddev, struct bio **bio); + + extern bool __must_check md_flush_request(struct mddev *mddev, struct bio *bio); +diff --git a/drivers/md/persistent-data/dm-btree.c b/drivers/md/persistent-data/dm-btree.c +index 0703ca7a7d9a4..5ce64e93aae74 100644 +--- a/drivers/md/persistent-data/dm-btree.c ++++ b/drivers/md/persistent-data/dm-btree.c +@@ -81,14 +81,16 @@ void inc_children(struct dm_transaction_manager *tm, struct btree_node *n, + } + + static int insert_at(size_t value_size, struct btree_node *node, unsigned index, +- uint64_t key, void *value) +- __dm_written_to_disk(value) ++ uint64_t key, void *value) ++ __dm_written_to_disk(value) + { + uint32_t nr_entries = le32_to_cpu(node->header.nr_entries); ++ uint32_t max_entries = le32_to_cpu(node->header.max_entries); + __le64 key_le = cpu_to_le64(key); + + if (index > nr_entries || +- index >= le32_to_cpu(node->header.max_entries)) { ++ index >= max_entries || ++ nr_entries >= max_entries) { + DMERR("too many entries in btree node for insert"); + __dm_unbless_for_disk(value); + return -ENOMEM; +diff --git a/drivers/md/persistent-data/dm-space-map-common.c b/drivers/md/persistent-data/dm-space-map-common.c +index 4a6a2a9b4eb49..bfbfa750e0160 100644 +--- a/drivers/md/persistent-data/dm-space-map-common.c ++++ b/drivers/md/persistent-data/dm-space-map-common.c +@@ -283,6 +283,11 @@ int sm_ll_lookup_bitmap(struct ll_disk *ll, dm_block_t b, uint32_t *result) + struct disk_index_entry ie_disk; + struct dm_block *blk; + ++ if (b >= ll->nr_blocks) { ++ DMERR_LIMIT("metadata block out of bounds"); ++ return -EINVAL; ++ } ++ + b = do_div(index, ll->entries_per_block); + r = ll->load_ie(ll, index, &ie_disk); + if (r < 0) +diff --git a/drivers/md/raid0.c b/drivers/md/raid0.c +index 62c8b6adac70e..b59a77b31b90d 100644 +--- a/drivers/md/raid0.c ++++ b/drivers/md/raid0.c +@@ -356,7 +356,21 @@ static sector_t raid0_size(struct mddev *mddev, sector_t sectors, int raid_disks + return array_sectors; + } + +-static void raid0_free(struct mddev *mddev, void *priv); ++static void free_conf(struct mddev *mddev, struct r0conf *conf) ++{ ++ kfree(conf->strip_zone); ++ kfree(conf->devlist); ++ kfree(conf); ++ mddev->private = NULL; ++} ++ ++static void raid0_free(struct mddev *mddev, void *priv) ++{ ++ struct r0conf *conf = priv; ++ ++ free_conf(mddev, conf); ++ acct_bioset_exit(mddev); ++} + + static int raid0_run(struct mddev *mddev) + { +@@ -370,11 +384,16 @@ static int raid0_run(struct mddev *mddev) + if (md_check_no_bitmap(mddev)) + return -EINVAL; + ++ if (acct_bioset_init(mddev)) { ++ pr_err("md/raid0:%s: alloc acct bioset failed.\n", mdname(mddev)); ++ return -ENOMEM; ++ } ++ + /* if private is not null, we are here after takeover */ + if (mddev->private == NULL) { + ret = create_strip_zones(mddev, &conf); + if (ret < 0) +- return ret; ++ goto exit_acct_set; + mddev->private = conf; + } + conf = mddev->private; +@@ -413,17 +432,16 @@ static int raid0_run(struct mddev *mddev) + dump_zones(mddev); + + ret = md_integrity_register(mddev); ++ if (ret) ++ goto free; + + return ret; +-} + +-static void raid0_free(struct mddev *mddev, void *priv) +-{ +- struct r0conf *conf = priv; +- +- kfree(conf->strip_zone); +- kfree(conf->devlist); +- kfree(conf); ++free: ++ free_conf(mddev, conf); ++exit_acct_set: ++ acct_bioset_exit(mddev); ++ return ret; + } + + static void raid0_handle_discard(struct mddev *mddev, struct bio *bio) +diff --git a/drivers/md/raid5.c b/drivers/md/raid5.c +index 02ed53b20654c..b9d062f0a02b2 100644 +--- a/drivers/md/raid5.c ++++ b/drivers/md/raid5.c +@@ -7446,12 +7446,19 @@ static int raid5_run(struct mddev *mddev) + struct md_rdev *rdev; + struct md_rdev *journal_dev = NULL; + sector_t reshape_offset = 0; +- int i; ++ int i, ret = 0; + long long min_offset_diff = 0; + int first = 1; + +- if (mddev_init_writes_pending(mddev) < 0) ++ if (acct_bioset_init(mddev)) { ++ pr_err("md/raid456:%s: alloc acct bioset failed.\n", mdname(mddev)); + return -ENOMEM; ++ } ++ ++ if (mddev_init_writes_pending(mddev) < 0) { ++ ret = -ENOMEM; ++ goto exit_acct_set; ++ } + + if (mddev->recovery_cp != MaxSector) + pr_notice("md/raid:%s: not clean -- starting background reconstruction\n", +@@ -7482,7 +7489,8 @@ static int raid5_run(struct mddev *mddev) + (mddev->bitmap_info.offset || mddev->bitmap_info.file)) { + pr_notice("md/raid:%s: array cannot have both journal and bitmap\n", + mdname(mddev)); +- return -EINVAL; ++ ret = -EINVAL; ++ goto exit_acct_set; + } + + if (mddev->reshape_position != MaxSector) { +@@ -7507,13 +7515,15 @@ static int raid5_run(struct mddev *mddev) + if (journal_dev) { + pr_warn("md/raid:%s: don't support reshape with journal - aborting.\n", + mdname(mddev)); +- return -EINVAL; ++ ret = -EINVAL; ++ goto exit_acct_set; + } + + if (mddev->new_level != mddev->level) { + pr_warn("md/raid:%s: unsupported reshape required - aborting.\n", + mdname(mddev)); +- return -EINVAL; ++ ret = -EINVAL; ++ goto exit_acct_set; + } + old_disks = mddev->raid_disks - mddev->delta_disks; + /* reshape_position must be on a new-stripe boundary, and one +@@ -7529,7 +7539,8 @@ static int raid5_run(struct mddev *mddev) + if (sector_div(here_new, chunk_sectors * new_data_disks)) { + pr_warn("md/raid:%s: reshape_position not on a stripe boundary\n", + mdname(mddev)); +- return -EINVAL; ++ ret = -EINVAL; ++ goto exit_acct_set; + } + reshape_offset = here_new * chunk_sectors; + /* here_new is the stripe we will write to */ +@@ -7551,7 +7562,8 @@ static int raid5_run(struct mddev *mddev) + else if (mddev->ro == 0) { + pr_warn("md/raid:%s: in-place reshape must be started in read-only mode - aborting\n", + mdname(mddev)); +- return -EINVAL; ++ ret = -EINVAL; ++ goto exit_acct_set; + } + } else if (mddev->reshape_backwards + ? (here_new * chunk_sectors + min_offset_diff <= +@@ -7561,7 +7573,8 @@ static int raid5_run(struct mddev *mddev) + /* Reading from the same stripe as writing to - bad */ + pr_warn("md/raid:%s: reshape_position too early for auto-recovery - aborting.\n", + mdname(mddev)); +- return -EINVAL; ++ ret = -EINVAL; ++ goto exit_acct_set; + } + pr_debug("md/raid:%s: reshape will continue\n", mdname(mddev)); + /* OK, we should be able to continue; */ +@@ -7585,8 +7598,10 @@ static int raid5_run(struct mddev *mddev) + else + conf = mddev->private; + +- if (IS_ERR(conf)) +- return PTR_ERR(conf); ++ if (IS_ERR(conf)) { ++ ret = PTR_ERR(conf); ++ goto exit_acct_set; ++ } + + if (test_bit(MD_HAS_JOURNAL, &mddev->flags)) { + if (!journal_dev) { +@@ -7786,7 +7801,10 @@ abort: + free_conf(conf); + mddev->private = NULL; + pr_warn("md/raid:%s: failed to run raid set.\n", mdname(mddev)); +- return -EIO; ++ ret = -EIO; ++exit_acct_set: ++ acct_bioset_exit(mddev); ++ return ret; + } + + static void raid5_free(struct mddev *mddev, void *priv) +@@ -7794,6 +7812,7 @@ static void raid5_free(struct mddev *mddev, void *priv) + struct r5conf *conf = priv; + + free_conf(conf); ++ acct_bioset_exit(mddev); + mddev->to_remove = &raid5_attrs_group; + } + +diff --git a/drivers/media/Kconfig b/drivers/media/Kconfig +index b07812657cee6..f3f24c63536b6 100644 +--- a/drivers/media/Kconfig ++++ b/drivers/media/Kconfig +@@ -141,10 +141,10 @@ config MEDIA_TEST_SUPPORT + prompt "Test drivers" if MEDIA_SUPPORT_FILTER + default y if !MEDIA_SUPPORT_FILTER + help +- Those drivers should not be used on production Kernels, but +- can be useful on debug ones. It enables several dummy drivers +- that simulate a real hardware. Very useful to test userspace +- applications and to validate if the subsystem core is doesn't ++ These drivers should not be used on production kernels, but ++ can be useful on debug ones. This option enables several dummy drivers ++ that simulate real hardware. Very useful to test userspace ++ applications and to validate if the subsystem core doesn't + have regressions. + + Say Y if you want to use some virtual test driver. +diff --git a/drivers/media/cec/core/cec-adap.c b/drivers/media/cec/core/cec-adap.c +index cd9cb354dc2c7..1f599e300e42e 100644 +--- a/drivers/media/cec/core/cec-adap.c ++++ b/drivers/media/cec/core/cec-adap.c +@@ -161,10 +161,10 @@ static void cec_queue_event(struct cec_adapter *adap, + u64 ts = ktime_get_ns(); + struct cec_fh *fh; + +- mutex_lock(&adap->devnode.lock); ++ mutex_lock(&adap->devnode.lock_fhs); + list_for_each_entry(fh, &adap->devnode.fhs, list) + cec_queue_event_fh(fh, ev, ts); +- mutex_unlock(&adap->devnode.lock); ++ mutex_unlock(&adap->devnode.lock_fhs); + } + + /* Notify userspace that the CEC pin changed state at the given time. */ +@@ -178,11 +178,12 @@ void cec_queue_pin_cec_event(struct cec_adapter *adap, bool is_high, + }; + struct cec_fh *fh; + +- mutex_lock(&adap->devnode.lock); +- list_for_each_entry(fh, &adap->devnode.fhs, list) ++ mutex_lock(&adap->devnode.lock_fhs); ++ list_for_each_entry(fh, &adap->devnode.fhs, list) { + if (fh->mode_follower == CEC_MODE_MONITOR_PIN) + cec_queue_event_fh(fh, &ev, ktime_to_ns(ts)); +- mutex_unlock(&adap->devnode.lock); ++ } ++ mutex_unlock(&adap->devnode.lock_fhs); + } + EXPORT_SYMBOL_GPL(cec_queue_pin_cec_event); + +@@ -195,10 +196,10 @@ void cec_queue_pin_hpd_event(struct cec_adapter *adap, bool is_high, ktime_t ts) + }; + struct cec_fh *fh; + +- mutex_lock(&adap->devnode.lock); ++ mutex_lock(&adap->devnode.lock_fhs); + list_for_each_entry(fh, &adap->devnode.fhs, list) + cec_queue_event_fh(fh, &ev, ktime_to_ns(ts)); +- mutex_unlock(&adap->devnode.lock); ++ mutex_unlock(&adap->devnode.lock_fhs); + } + EXPORT_SYMBOL_GPL(cec_queue_pin_hpd_event); + +@@ -211,10 +212,10 @@ void cec_queue_pin_5v_event(struct cec_adapter *adap, bool is_high, ktime_t ts) + }; + struct cec_fh *fh; + +- mutex_lock(&adap->devnode.lock); ++ mutex_lock(&adap->devnode.lock_fhs); + list_for_each_entry(fh, &adap->devnode.fhs, list) + cec_queue_event_fh(fh, &ev, ktime_to_ns(ts)); +- mutex_unlock(&adap->devnode.lock); ++ mutex_unlock(&adap->devnode.lock_fhs); + } + EXPORT_SYMBOL_GPL(cec_queue_pin_5v_event); + +@@ -286,12 +287,12 @@ static void cec_queue_msg_monitor(struct cec_adapter *adap, + u32 monitor_mode = valid_la ? CEC_MODE_MONITOR : + CEC_MODE_MONITOR_ALL; + +- mutex_lock(&adap->devnode.lock); ++ mutex_lock(&adap->devnode.lock_fhs); + list_for_each_entry(fh, &adap->devnode.fhs, list) { + if (fh->mode_follower >= monitor_mode) + cec_queue_msg_fh(fh, msg); + } +- mutex_unlock(&adap->devnode.lock); ++ mutex_unlock(&adap->devnode.lock_fhs); + } + + /* +@@ -302,12 +303,12 @@ static void cec_queue_msg_followers(struct cec_adapter *adap, + { + struct cec_fh *fh; + +- mutex_lock(&adap->devnode.lock); ++ mutex_lock(&adap->devnode.lock_fhs); + list_for_each_entry(fh, &adap->devnode.fhs, list) { + if (fh->mode_follower == CEC_MODE_FOLLOWER) + cec_queue_msg_fh(fh, msg); + } +- mutex_unlock(&adap->devnode.lock); ++ mutex_unlock(&adap->devnode.lock_fhs); + } + + /* Notify userspace of an adapter state change. */ +@@ -1573,6 +1574,7 @@ void __cec_s_phys_addr(struct cec_adapter *adap, u16 phys_addr, bool block) + /* Disabling monitor all mode should always succeed */ + if (adap->monitor_all_cnt) + WARN_ON(call_op(adap, adap_monitor_all_enable, false)); ++ /* serialize adap_enable */ + mutex_lock(&adap->devnode.lock); + if (adap->needs_hpd || list_empty(&adap->devnode.fhs)) { + WARN_ON(adap->ops->adap_enable(adap, false)); +@@ -1584,14 +1586,16 @@ void __cec_s_phys_addr(struct cec_adapter *adap, u16 phys_addr, bool block) + return; + } + ++ /* serialize adap_enable */ + mutex_lock(&adap->devnode.lock); + adap->last_initiator = 0xff; + adap->transmit_in_progress = false; + +- if ((adap->needs_hpd || list_empty(&adap->devnode.fhs)) && +- adap->ops->adap_enable(adap, true)) { +- mutex_unlock(&adap->devnode.lock); +- return; ++ if (adap->needs_hpd || list_empty(&adap->devnode.fhs)) { ++ if (adap->ops->adap_enable(adap, true)) { ++ mutex_unlock(&adap->devnode.lock); ++ return; ++ } + } + + if (adap->monitor_all_cnt && +diff --git a/drivers/media/cec/core/cec-api.c b/drivers/media/cec/core/cec-api.c +index 769e6b4cddce3..52c30e4e20055 100644 +--- a/drivers/media/cec/core/cec-api.c ++++ b/drivers/media/cec/core/cec-api.c +@@ -586,6 +586,7 @@ static int cec_open(struct inode *inode, struct file *filp) + return err; + } + ++ /* serialize adap_enable */ + mutex_lock(&devnode->lock); + if (list_empty(&devnode->fhs) && + !adap->needs_hpd && +@@ -624,7 +625,9 @@ static int cec_open(struct inode *inode, struct file *filp) + } + #endif + ++ mutex_lock(&devnode->lock_fhs); + list_add(&fh->list, &devnode->fhs); ++ mutex_unlock(&devnode->lock_fhs); + mutex_unlock(&devnode->lock); + + return 0; +@@ -653,8 +656,11 @@ static int cec_release(struct inode *inode, struct file *filp) + cec_monitor_all_cnt_dec(adap); + mutex_unlock(&adap->lock); + ++ /* serialize adap_enable */ + mutex_lock(&devnode->lock); ++ mutex_lock(&devnode->lock_fhs); + list_del(&fh->list); ++ mutex_unlock(&devnode->lock_fhs); + if (cec_is_registered(adap) && list_empty(&devnode->fhs) && + !adap->needs_hpd && adap->phys_addr == CEC_PHYS_ADDR_INVALID) { + WARN_ON(adap->ops->adap_enable(adap, false)); +diff --git a/drivers/media/cec/core/cec-core.c b/drivers/media/cec/core/cec-core.c +index 551689d371a71..ec67065d52021 100644 +--- a/drivers/media/cec/core/cec-core.c ++++ b/drivers/media/cec/core/cec-core.c +@@ -169,8 +169,10 @@ static void cec_devnode_unregister(struct cec_adapter *adap) + devnode->registered = false; + devnode->unregistered = true; + ++ mutex_lock(&devnode->lock_fhs); + list_for_each_entry(fh, &devnode->fhs, list) + wake_up_interruptible(&fh->wait); ++ mutex_unlock(&devnode->lock_fhs); + + mutex_unlock(&devnode->lock); + +@@ -272,6 +274,7 @@ struct cec_adapter *cec_allocate_adapter(const struct cec_adap_ops *ops, + + /* adap->devnode initialization */ + INIT_LIST_HEAD(&adap->devnode.fhs); ++ mutex_init(&adap->devnode.lock_fhs); + mutex_init(&adap->devnode.lock); + + adap->kthread = kthread_run(cec_thread_func, adap, "cec-%s", name); +diff --git a/drivers/media/cec/core/cec-pin.c b/drivers/media/cec/core/cec-pin.c +index 8c613aa649c6f..0eb90cc0ffb0f 100644 +--- a/drivers/media/cec/core/cec-pin.c ++++ b/drivers/media/cec/core/cec-pin.c +@@ -1033,6 +1033,7 @@ static int cec_pin_thread_func(void *_adap) + { + struct cec_adapter *adap = _adap; + struct cec_pin *pin = adap->pin; ++ bool irq_enabled = false; + + for (;;) { + wait_event_interruptible(pin->kthread_waitq, +@@ -1060,6 +1061,7 @@ static int cec_pin_thread_func(void *_adap) + ns_to_ktime(pin->work_rx_msg.rx_ts)); + msg->len = 0; + } ++ + if (pin->work_tx_status) { + unsigned int tx_status = pin->work_tx_status; + +@@ -1083,27 +1085,39 @@ static int cec_pin_thread_func(void *_adap) + switch (atomic_xchg(&pin->work_irq_change, + CEC_PIN_IRQ_UNCHANGED)) { + case CEC_PIN_IRQ_DISABLE: +- pin->ops->disable_irq(adap); ++ if (irq_enabled) { ++ pin->ops->disable_irq(adap); ++ irq_enabled = false; ++ } + cec_pin_high(pin); + cec_pin_to_idle(pin); + hrtimer_start(&pin->timer, ns_to_ktime(0), + HRTIMER_MODE_REL); + break; + case CEC_PIN_IRQ_ENABLE: ++ if (irq_enabled) ++ break; + pin->enable_irq_failed = !pin->ops->enable_irq(adap); + if (pin->enable_irq_failed) { + cec_pin_to_idle(pin); + hrtimer_start(&pin->timer, ns_to_ktime(0), + HRTIMER_MODE_REL); ++ } else { ++ irq_enabled = true; + } + break; + default: + break; + } +- + if (kthread_should_stop()) + break; + } ++ if (pin->ops->disable_irq && irq_enabled) ++ pin->ops->disable_irq(adap); ++ hrtimer_cancel(&pin->timer); ++ cec_pin_read(pin); ++ cec_pin_to_idle(pin); ++ pin->state = CEC_ST_OFF; + return 0; + } + +@@ -1130,13 +1144,7 @@ static int cec_pin_adap_enable(struct cec_adapter *adap, bool enable) + hrtimer_start(&pin->timer, ns_to_ktime(0), + HRTIMER_MODE_REL); + } else { +- if (pin->ops->disable_irq) +- pin->ops->disable_irq(adap); +- hrtimer_cancel(&pin->timer); + kthread_stop(pin->kthread); +- cec_pin_read(pin); +- cec_pin_to_idle(pin); +- pin->state = CEC_ST_OFF; + } + return 0; + } +@@ -1157,11 +1165,8 @@ void cec_pin_start_timer(struct cec_pin *pin) + if (pin->state != CEC_ST_RX_IRQ) + return; + +- atomic_set(&pin->work_irq_change, CEC_PIN_IRQ_UNCHANGED); +- pin->ops->disable_irq(pin->adap); +- cec_pin_high(pin); +- cec_pin_to_idle(pin); +- hrtimer_start(&pin->timer, ns_to_ktime(0), HRTIMER_MODE_REL); ++ atomic_set(&pin->work_irq_change, CEC_PIN_IRQ_DISABLE); ++ wake_up_interruptible(&pin->kthread_waitq); + } + + static int cec_pin_adap_transmit(struct cec_adapter *adap, u8 attempts, +diff --git a/drivers/media/common/saa7146/saa7146_fops.c b/drivers/media/common/saa7146/saa7146_fops.c +index baf5772c52a96..be32159777142 100644 +--- a/drivers/media/common/saa7146/saa7146_fops.c ++++ b/drivers/media/common/saa7146/saa7146_fops.c +@@ -521,7 +521,7 @@ int saa7146_vv_init(struct saa7146_dev* dev, struct saa7146_ext_vv *ext_vv) + ERR("out of memory. aborting.\n"); + kfree(vv); + v4l2_ctrl_handler_free(hdl); +- return -1; ++ return -ENOMEM; + } + + saa7146_video_uops.init(dev,vv); +diff --git a/drivers/media/common/videobuf2/videobuf2-dma-contig.c b/drivers/media/common/videobuf2/videobuf2-dma-contig.c +index be376f3011b68..f8c65b0401054 100644 +--- a/drivers/media/common/videobuf2/videobuf2-dma-contig.c ++++ b/drivers/media/common/videobuf2/videobuf2-dma-contig.c +@@ -157,7 +157,7 @@ static void *vb2_dc_alloc(struct vb2_buffer *vb, + GFP_KERNEL | vb->vb2_queue->gfp_flags, + buf->attrs); + if (!buf->cookie) { +- dev_err(dev, "dma_alloc_coherent of size %ld failed\n", size); ++ dev_err(dev, "dma_alloc_coherent of size %lu failed\n", size); + kfree(buf); + return ERR_PTR(-ENOMEM); + } +@@ -204,9 +204,9 @@ static int vb2_dc_mmap(void *buf_priv, struct vm_area_struct *vma) + + vma->vm_ops->open(vma); + +- pr_debug("%s: mapped dma addr 0x%08lx at 0x%08lx, size %ld\n", +- __func__, (unsigned long)buf->dma_addr, vma->vm_start, +- buf->size); ++ pr_debug("%s: mapped dma addr 0x%08lx at 0x%08lx, size %lu\n", ++ __func__, (unsigned long)buf->dma_addr, vma->vm_start, ++ buf->size); + + return 0; + } +diff --git a/drivers/media/dvb-core/dmxdev.c b/drivers/media/dvb-core/dmxdev.c +index 5d5a48475a54f..01f288fa37e0e 100644 +--- a/drivers/media/dvb-core/dmxdev.c ++++ b/drivers/media/dvb-core/dmxdev.c +@@ -1413,7 +1413,7 @@ static const struct dvb_device dvbdev_dvr = { + }; + int dvb_dmxdev_init(struct dmxdev *dmxdev, struct dvb_adapter *dvb_adapter) + { +- int i; ++ int i, ret; + + if (dmxdev->demux->open(dmxdev->demux) < 0) + return -EUSERS; +@@ -1432,14 +1432,26 @@ int dvb_dmxdev_init(struct dmxdev *dmxdev, struct dvb_adapter *dvb_adapter) + DMXDEV_STATE_FREE); + } + +- dvb_register_device(dvb_adapter, &dmxdev->dvbdev, &dvbdev_demux, dmxdev, ++ ret = dvb_register_device(dvb_adapter, &dmxdev->dvbdev, &dvbdev_demux, dmxdev, + DVB_DEVICE_DEMUX, dmxdev->filternum); +- dvb_register_device(dvb_adapter, &dmxdev->dvr_dvbdev, &dvbdev_dvr, ++ if (ret < 0) ++ goto err_register_dvbdev; ++ ++ ret = dvb_register_device(dvb_adapter, &dmxdev->dvr_dvbdev, &dvbdev_dvr, + dmxdev, DVB_DEVICE_DVR, dmxdev->filternum); ++ if (ret < 0) ++ goto err_register_dvr_dvbdev; + + dvb_ringbuffer_init(&dmxdev->dvr_buffer, NULL, 8192); + + return 0; ++ ++err_register_dvr_dvbdev: ++ dvb_unregister_device(dmxdev->dvbdev); ++err_register_dvbdev: ++ vfree(dmxdev->filter); ++ dmxdev->filter = NULL; ++ return ret; + } + + EXPORT_SYMBOL(dvb_dmxdev_init); +diff --git a/drivers/media/dvb-frontends/dib8000.c b/drivers/media/dvb-frontends/dib8000.c +index bb02354a48b81..d67f2dd997d06 100644 +--- a/drivers/media/dvb-frontends/dib8000.c ++++ b/drivers/media/dvb-frontends/dib8000.c +@@ -4473,8 +4473,10 @@ static struct dvb_frontend *dib8000_init(struct i2c_adapter *i2c_adap, u8 i2c_ad + + state->timf_default = cfg->pll->timf; + +- if (dib8000_identify(&state->i2c) == 0) ++ if (dib8000_identify(&state->i2c) == 0) { ++ kfree(fe); + goto error; ++ } + + dibx000_init_i2c_master(&state->i2c_master, DIB8000, state->i2c.adap, state->i2c.addr); + +diff --git a/drivers/media/i2c/imx274.c b/drivers/media/i2c/imx274.c +index 0dce92872176d..4d9b64c61f603 100644 +--- a/drivers/media/i2c/imx274.c ++++ b/drivers/media/i2c/imx274.c +@@ -1367,6 +1367,10 @@ static int imx274_s_frame_interval(struct v4l2_subdev *sd, + int min, max, def; + int ret; + ++ ret = pm_runtime_resume_and_get(&imx274->client->dev); ++ if (ret < 0) ++ return ret; ++ + mutex_lock(&imx274->lock); + ret = imx274_set_frame_interval(imx274, fi->interval); + +@@ -1398,6 +1402,7 @@ static int imx274_s_frame_interval(struct v4l2_subdev *sd, + + unlock: + mutex_unlock(&imx274->lock); ++ pm_runtime_put(&imx274->client->dev); + + return ret; + } +diff --git a/drivers/media/i2c/ov8865.c b/drivers/media/i2c/ov8865.c +index ce50f3ea87b8e..92f6c3a940cfb 100644 +--- a/drivers/media/i2c/ov8865.c ++++ b/drivers/media/i2c/ov8865.c +@@ -2330,27 +2330,27 @@ static int ov8865_sensor_power(struct ov8865_sensor *sensor, bool on) + if (ret) { + dev_err(sensor->dev, + "failed to enable DOVDD regulator\n"); +- goto disable; ++ return ret; + } + + ret = regulator_enable(sensor->avdd); + if (ret) { + dev_err(sensor->dev, + "failed to enable AVDD regulator\n"); +- goto disable; ++ goto disable_dovdd; + } + + ret = regulator_enable(sensor->dvdd); + if (ret) { + dev_err(sensor->dev, + "failed to enable DVDD regulator\n"); +- goto disable; ++ goto disable_avdd; + } + + ret = clk_prepare_enable(sensor->extclk); + if (ret) { + dev_err(sensor->dev, "failed to enable EXTCLK clock\n"); +- goto disable; ++ goto disable_dvdd; + } + + gpiod_set_value_cansleep(sensor->reset, 0); +@@ -2359,14 +2359,16 @@ static int ov8865_sensor_power(struct ov8865_sensor *sensor, bool on) + /* Time to enter streaming mode according to power timings. */ + usleep_range(10000, 12000); + } else { +-disable: + gpiod_set_value_cansleep(sensor->powerdown, 1); + gpiod_set_value_cansleep(sensor->reset, 1); + + clk_disable_unprepare(sensor->extclk); + ++disable_dvdd: + regulator_disable(sensor->dvdd); ++disable_avdd: + regulator_disable(sensor->avdd); ++disable_dovdd: + regulator_disable(sensor->dovdd); + } + +@@ -2891,14 +2893,16 @@ static int ov8865_probe(struct i2c_client *client) + if (ret) + goto error_mutex; + ++ mutex_lock(&sensor->mutex); + ret = ov8865_state_init(sensor); ++ mutex_unlock(&sensor->mutex); + if (ret) + goto error_ctrls; + + /* Runtime PM */ + +- pm_runtime_enable(sensor->dev); + pm_runtime_set_suspended(sensor->dev); ++ pm_runtime_enable(sensor->dev); + + /* V4L2 subdev register */ + +diff --git a/drivers/media/pci/b2c2/flexcop-pci.c b/drivers/media/pci/b2c2/flexcop-pci.c +index 6a4c7cb0ad0f9..486c8ec0fa60d 100644 +--- a/drivers/media/pci/b2c2/flexcop-pci.c ++++ b/drivers/media/pci/b2c2/flexcop-pci.c +@@ -185,6 +185,8 @@ static irqreturn_t flexcop_pci_isr(int irq, void *dev_id) + dma_addr_t cur_addr = + fc->read_ibi_reg(fc,dma1_008).dma_0x8.dma_cur_addr << 2; + u32 cur_pos = cur_addr - fc_pci->dma[0].dma_addr0; ++ if (cur_pos > fc_pci->dma[0].size * 2) ++ goto error; + + deb_irq("%u irq: %08x cur_addr: %llx: cur_pos: %08x, last_cur_pos: %08x ", + jiffies_to_usecs(jiffies - fc_pci->last_irq), +@@ -225,6 +227,7 @@ static irqreturn_t flexcop_pci_isr(int irq, void *dev_id) + ret = IRQ_NONE; + } + ++error: + spin_unlock_irqrestore(&fc_pci->irq_lock, flags); + return ret; + } +diff --git a/drivers/media/pci/saa7146/hexium_gemini.c b/drivers/media/pci/saa7146/hexium_gemini.c +index 2214c74bbbf15..3947701cd6c7e 100644 +--- a/drivers/media/pci/saa7146/hexium_gemini.c ++++ b/drivers/media/pci/saa7146/hexium_gemini.c +@@ -284,7 +284,12 @@ static int hexium_attach(struct saa7146_dev *dev, struct saa7146_pci_extension_d + hexium_set_input(hexium, 0); + hexium->cur_input = 0; + +- saa7146_vv_init(dev, &vv_data); ++ ret = saa7146_vv_init(dev, &vv_data); ++ if (ret) { ++ i2c_del_adapter(&hexium->i2c_adapter); ++ kfree(hexium); ++ return ret; ++ } + + vv_data.vid_ops.vidioc_enum_input = vidioc_enum_input; + vv_data.vid_ops.vidioc_g_input = vidioc_g_input; +diff --git a/drivers/media/pci/saa7146/hexium_orion.c b/drivers/media/pci/saa7146/hexium_orion.c +index 39d14c179d229..2eb4bee16b71f 100644 +--- a/drivers/media/pci/saa7146/hexium_orion.c ++++ b/drivers/media/pci/saa7146/hexium_orion.c +@@ -355,10 +355,16 @@ static struct saa7146_ext_vv vv_data; + static int hexium_attach(struct saa7146_dev *dev, struct saa7146_pci_extension_data *info) + { + struct hexium *hexium = (struct hexium *) dev->ext_priv; ++ int ret; + + DEB_EE("\n"); + +- saa7146_vv_init(dev, &vv_data); ++ ret = saa7146_vv_init(dev, &vv_data); ++ if (ret) { ++ pr_err("Error in saa7146_vv_init()\n"); ++ return ret; ++ } ++ + vv_data.vid_ops.vidioc_enum_input = vidioc_enum_input; + vv_data.vid_ops.vidioc_g_input = vidioc_g_input; + vv_data.vid_ops.vidioc_s_input = vidioc_s_input; +diff --git a/drivers/media/pci/saa7146/mxb.c b/drivers/media/pci/saa7146/mxb.c +index 73fc901ecf3db..bf0b9b0914cd5 100644 +--- a/drivers/media/pci/saa7146/mxb.c ++++ b/drivers/media/pci/saa7146/mxb.c +@@ -683,10 +683,16 @@ static struct saa7146_ext_vv vv_data; + static int mxb_attach(struct saa7146_dev *dev, struct saa7146_pci_extension_data *info) + { + struct mxb *mxb; ++ int ret; + + DEB_EE("dev:%p\n", dev); + +- saa7146_vv_init(dev, &vv_data); ++ ret = saa7146_vv_init(dev, &vv_data); ++ if (ret) { ++ ERR("Error in saa7146_vv_init()"); ++ return ret; ++ } ++ + if (mxb_probe(dev)) { + saa7146_vv_release(dev); + return -1; +diff --git a/drivers/media/platform/aspeed-video.c b/drivers/media/platform/aspeed-video.c +index 7bb6babdcade0..debc7509c173c 100644 +--- a/drivers/media/platform/aspeed-video.c ++++ b/drivers/media/platform/aspeed-video.c +@@ -500,6 +500,10 @@ static void aspeed_video_enable_mode_detect(struct aspeed_video *video) + aspeed_video_update(video, VE_INTERRUPT_CTRL, 0, + VE_INTERRUPT_MODE_DETECT); + ++ /* Disable mode detect in order to re-trigger */ ++ aspeed_video_update(video, VE_SEQ_CTRL, ++ VE_SEQ_CTRL_TRIG_MODE_DET, 0); ++ + /* Trigger mode detect */ + aspeed_video_update(video, VE_SEQ_CTRL, 0, VE_SEQ_CTRL_TRIG_MODE_DET); + } +@@ -552,6 +556,8 @@ static void aspeed_video_irq_res_change(struct aspeed_video *video, ulong delay) + set_bit(VIDEO_RES_CHANGE, &video->flags); + clear_bit(VIDEO_FRAME_INPRG, &video->flags); + ++ video->v4l2_input_status = V4L2_IN_ST_NO_SIGNAL; ++ + aspeed_video_off(video); + aspeed_video_bufs_done(video, VB2_BUF_STATE_ERROR); + +@@ -786,10 +792,6 @@ static void aspeed_video_get_resolution(struct aspeed_video *video) + return; + } + +- /* Disable mode detect in order to re-trigger */ +- aspeed_video_update(video, VE_SEQ_CTRL, +- VE_SEQ_CTRL_TRIG_MODE_DET, 0); +- + aspeed_video_check_and_set_polarity(video); + + aspeed_video_enable_mode_detect(video); +@@ -1337,7 +1339,6 @@ static void aspeed_video_resolution_work(struct work_struct *work) + struct delayed_work *dwork = to_delayed_work(work); + struct aspeed_video *video = container_of(dwork, struct aspeed_video, + res_work); +- u32 input_status = video->v4l2_input_status; + + aspeed_video_on(video); + +@@ -1350,8 +1351,7 @@ static void aspeed_video_resolution_work(struct work_struct *work) + aspeed_video_get_resolution(video); + + if (video->detected_timings.width != video->active_timings.width || +- video->detected_timings.height != video->active_timings.height || +- input_status != video->v4l2_input_status) { ++ video->detected_timings.height != video->active_timings.height) { + static const struct v4l2_event ev = { + .type = V4L2_EVENT_SOURCE_CHANGE, + .u.src_change.changes = V4L2_EVENT_SRC_CH_RESOLUTION, +diff --git a/drivers/media/platform/coda/coda-common.c b/drivers/media/platform/coda/coda-common.c +index 0e312b0842d7f..9a2640a9c75c6 100644 +--- a/drivers/media/platform/coda/coda-common.c ++++ b/drivers/media/platform/coda/coda-common.c +@@ -1537,11 +1537,13 @@ static void coda_pic_run_work(struct work_struct *work) + + if (!wait_for_completion_timeout(&ctx->completion, + msecs_to_jiffies(1000))) { +- dev_err(dev->dev, "CODA PIC_RUN timeout\n"); ++ if (ctx->use_bit) { ++ dev_err(dev->dev, "CODA PIC_RUN timeout\n"); + +- ctx->hold = true; ++ ctx->hold = true; + +- coda_hw_reset(ctx); ++ coda_hw_reset(ctx); ++ } + + if (ctx->ops->run_timeout) + ctx->ops->run_timeout(ctx); +diff --git a/drivers/media/platform/coda/coda-jpeg.c b/drivers/media/platform/coda/coda-jpeg.c +index b11cfbe166dd3..a72f4655e5ad5 100644 +--- a/drivers/media/platform/coda/coda-jpeg.c ++++ b/drivers/media/platform/coda/coda-jpeg.c +@@ -1127,7 +1127,8 @@ static int coda9_jpeg_prepare_encode(struct coda_ctx *ctx) + coda_write(dev, 0, CODA9_REG_JPEG_GBU_BT_PTR); + coda_write(dev, 0, CODA9_REG_JPEG_GBU_WD_PTR); + coda_write(dev, 0, CODA9_REG_JPEG_GBU_BBSR); +- coda_write(dev, 0, CODA9_REG_JPEG_BBC_STRM_CTRL); ++ coda_write(dev, BIT(31) | ((end_addr - start_addr - header_len) / 256), ++ CODA9_REG_JPEG_BBC_STRM_CTRL); + coda_write(dev, 0, CODA9_REG_JPEG_GBU_CTRL); + coda_write(dev, 0, CODA9_REG_JPEG_GBU_FF_RPTR); + coda_write(dev, 127, CODA9_REG_JPEG_GBU_BBER); +@@ -1257,6 +1258,23 @@ static void coda9_jpeg_finish_encode(struct coda_ctx *ctx) + coda_hw_reset(ctx); + } + ++static void coda9_jpeg_encode_timeout(struct coda_ctx *ctx) ++{ ++ struct coda_dev *dev = ctx->dev; ++ u32 end_addr, wr_ptr; ++ ++ /* Handle missing BBC overflow interrupt via timeout */ ++ end_addr = coda_read(dev, CODA9_REG_JPEG_BBC_END_ADDR); ++ wr_ptr = coda_read(dev, CODA9_REG_JPEG_BBC_WR_PTR); ++ if (wr_ptr >= end_addr - 256) { ++ v4l2_err(&dev->v4l2_dev, "JPEG too large for capture buffer\n"); ++ coda9_jpeg_finish_encode(ctx); ++ return; ++ } ++ ++ coda_hw_reset(ctx); ++} ++ + static void coda9_jpeg_release(struct coda_ctx *ctx) + { + int i; +@@ -1276,6 +1294,7 @@ const struct coda_context_ops coda9_jpeg_encode_ops = { + .start_streaming = coda9_jpeg_start_encoding, + .prepare_run = coda9_jpeg_prepare_encode, + .finish_run = coda9_jpeg_finish_encode, ++ .run_timeout = coda9_jpeg_encode_timeout, + .release = coda9_jpeg_release, + }; + +diff --git a/drivers/media/platform/coda/imx-vdoa.c b/drivers/media/platform/coda/imx-vdoa.c +index 8bc0d83718193..dd6e2e320264e 100644 +--- a/drivers/media/platform/coda/imx-vdoa.c ++++ b/drivers/media/platform/coda/imx-vdoa.c +@@ -287,7 +287,11 @@ static int vdoa_probe(struct platform_device *pdev) + struct resource *res; + int ret; + +- dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(32)); ++ ret = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(32)); ++ if (ret) { ++ dev_err(&pdev->dev, "DMA enable failed\n"); ++ return ret; ++ } + + vdoa = devm_kzalloc(&pdev->dev, sizeof(*vdoa), GFP_KERNEL); + if (!vdoa) +diff --git a/drivers/media/platform/imx-pxp.c b/drivers/media/platform/imx-pxp.c +index 4321edc0c23d9..8e9c6fee75a48 100644 +--- a/drivers/media/platform/imx-pxp.c ++++ b/drivers/media/platform/imx-pxp.c +@@ -1661,6 +1661,8 @@ static int pxp_probe(struct platform_device *pdev) + if (irq < 0) + return irq; + ++ spin_lock_init(&dev->irqlock); ++ + ret = devm_request_threaded_irq(&pdev->dev, irq, NULL, pxp_irq_handler, + IRQF_ONESHOT, dev_name(&pdev->dev), dev); + if (ret < 0) { +@@ -1678,8 +1680,6 @@ static int pxp_probe(struct platform_device *pdev) + goto err_clk; + } + +- spin_lock_init(&dev->irqlock); +- + ret = v4l2_device_register(&pdev->dev, &dev->v4l2_dev); + if (ret) + goto err_clk; +diff --git a/drivers/media/platform/mtk-vcodec/mtk_vcodec_enc_drv.c b/drivers/media/platform/mtk-vcodec/mtk_vcodec_enc_drv.c +index 45d1870c83dd7..4ced20ca647b1 100644 +--- a/drivers/media/platform/mtk-vcodec/mtk_vcodec_enc_drv.c ++++ b/drivers/media/platform/mtk-vcodec/mtk_vcodec_enc_drv.c +@@ -218,11 +218,11 @@ static int fops_vcodec_release(struct file *file) + mtk_v4l2_debug(1, "[%d] encoder", ctx->id); + mutex_lock(&dev->dev_mutex); + ++ v4l2_m2m_ctx_release(ctx->m2m_ctx); + mtk_vcodec_enc_release(ctx); + v4l2_fh_del(&ctx->fh); + v4l2_fh_exit(&ctx->fh); + v4l2_ctrl_handler_free(&ctx->ctrl_hdl); +- v4l2_m2m_ctx_release(ctx->m2m_ctx); + + list_del_init(&ctx->list); + kfree(ctx); +diff --git a/drivers/media/platform/qcom/venus/core.c b/drivers/media/platform/qcom/venus/core.c +index 91b15842c5558..1f0181b6353c9 100644 +--- a/drivers/media/platform/qcom/venus/core.c ++++ b/drivers/media/platform/qcom/venus/core.c +@@ -349,11 +349,11 @@ static int venus_probe(struct platform_device *pdev) + + ret = venus_firmware_init(core); + if (ret) +- goto err_runtime_disable; ++ goto err_of_depopulate; + + ret = venus_boot(core); + if (ret) +- goto err_runtime_disable; ++ goto err_firmware_deinit; + + ret = hfi_core_resume(core, true); + if (ret) +@@ -385,6 +385,10 @@ err_dev_unregister: + v4l2_device_unregister(&core->v4l2_dev); + err_venus_shutdown: + venus_shutdown(core); ++err_firmware_deinit: ++ venus_firmware_deinit(core); ++err_of_depopulate: ++ of_platform_depopulate(dev); + err_runtime_disable: + pm_runtime_put_noidle(dev); + pm_runtime_set_suspended(dev); +@@ -472,7 +476,8 @@ static __maybe_unused int venus_runtime_suspend(struct device *dev) + err_video_path: + icc_set_bw(core->cpucfg_path, kbps_to_icc(1000), 0); + err_cpucfg_path: +- pm_ops->core_power(core, POWER_ON); ++ if (pm_ops->core_power) ++ pm_ops->core_power(core, POWER_ON); + + return ret; + } +diff --git a/drivers/media/platform/qcom/venus/pm_helpers.c b/drivers/media/platform/qcom/venus/pm_helpers.c +index e031fd17f4e75..a591dd315ebcc 100644 +--- a/drivers/media/platform/qcom/venus/pm_helpers.c ++++ b/drivers/media/platform/qcom/venus/pm_helpers.c +@@ -163,14 +163,12 @@ static u32 load_per_type(struct venus_core *core, u32 session_type) + struct venus_inst *inst = NULL; + u32 mbs_per_sec = 0; + +- mutex_lock(&core->lock); + list_for_each_entry(inst, &core->instances, list) { + if (inst->session_type != session_type) + continue; + + mbs_per_sec += load_per_instance(inst); + } +- mutex_unlock(&core->lock); + + return mbs_per_sec; + } +@@ -219,14 +217,12 @@ static int load_scale_bw(struct venus_core *core) + struct venus_inst *inst = NULL; + u32 mbs_per_sec, avg, peak, total_avg = 0, total_peak = 0; + +- mutex_lock(&core->lock); + list_for_each_entry(inst, &core->instances, list) { + mbs_per_sec = load_per_instance(inst); + mbs_to_bw(inst, mbs_per_sec, &avg, &peak); + total_avg += avg; + total_peak += peak; + } +- mutex_unlock(&core->lock); + + /* + * keep minimum bandwidth vote for "video-mem" path, +@@ -253,8 +249,9 @@ static int load_scale_v1(struct venus_inst *inst) + struct device *dev = core->dev; + u32 mbs_per_sec; + unsigned int i; +- int ret; ++ int ret = 0; + ++ mutex_lock(&core->lock); + mbs_per_sec = load_per_type(core, VIDC_SESSION_TYPE_ENC) + + load_per_type(core, VIDC_SESSION_TYPE_DEC); + +@@ -279,17 +276,19 @@ set_freq: + if (ret) { + dev_err(dev, "failed to set clock rate %lu (%d)\n", + freq, ret); +- return ret; ++ goto exit; + } + + ret = load_scale_bw(core); + if (ret) { + dev_err(dev, "failed to set bandwidth (%d)\n", + ret); +- return ret; ++ goto exit; + } + +- return 0; ++exit: ++ mutex_unlock(&core->lock); ++ return ret; + } + + static int core_get_v1(struct venus_core *core) +@@ -587,8 +586,8 @@ min_loaded_core(struct venus_inst *inst, u32 *min_coreid, u32 *min_load, bool lo + if (inst->session_type == VIDC_SESSION_TYPE_DEC) + vpp_freq = inst_pos->clk_data.vpp_freq; + else if (inst->session_type == VIDC_SESSION_TYPE_ENC) +- vpp_freq = low_power ? inst_pos->clk_data.vpp_freq : +- inst_pos->clk_data.low_power_freq; ++ vpp_freq = low_power ? inst_pos->clk_data.low_power_freq : ++ inst_pos->clk_data.vpp_freq; + else + continue; + +@@ -1116,13 +1115,13 @@ static int load_scale_v4(struct venus_inst *inst) + struct device *dev = core->dev; + unsigned long freq = 0, freq_core1 = 0, freq_core2 = 0; + unsigned long filled_len = 0; +- int i, ret; ++ int i, ret = 0; + + for (i = 0; i < inst->num_input_bufs; i++) + filled_len = max(filled_len, inst->payloads[i]); + + if (inst->session_type == VIDC_SESSION_TYPE_DEC && !filled_len) +- return 0; ++ return ret; + + freq = calculate_inst_freq(inst, filled_len); + inst->clk_data.freq = freq; +@@ -1138,7 +1137,6 @@ static int load_scale_v4(struct venus_inst *inst) + freq_core2 += inst->clk_data.freq; + } + } +- mutex_unlock(&core->lock); + + freq = max(freq_core1, freq_core2); + +@@ -1162,17 +1160,19 @@ set_freq: + if (ret) { + dev_err(dev, "failed to set clock rate %lu (%d)\n", + freq, ret); +- return ret; ++ goto exit; + } + + ret = load_scale_bw(core); + if (ret) { + dev_err(dev, "failed to set bandwidth (%d)\n", + ret); +- return ret; ++ goto exit; + } + +- return 0; ++exit: ++ mutex_unlock(&core->lock); ++ return ret; + } + + static const struct venus_pm_ops pm_ops_v4 = { +diff --git a/drivers/media/platform/rcar-vin/rcar-csi2.c b/drivers/media/platform/rcar-vin/rcar-csi2.c +index ba4a380016cc4..0c5e2f7e04beb 100644 +--- a/drivers/media/platform/rcar-vin/rcar-csi2.c ++++ b/drivers/media/platform/rcar-vin/rcar-csi2.c +@@ -445,16 +445,23 @@ static int rcsi2_wait_phy_start(struct rcar_csi2 *priv, + static int rcsi2_set_phypll(struct rcar_csi2 *priv, unsigned int mbps) + { + const struct rcsi2_mbps_reg *hsfreq; ++ const struct rcsi2_mbps_reg *hsfreq_prev = NULL; + +- for (hsfreq = priv->info->hsfreqrange; hsfreq->mbps != 0; hsfreq++) ++ for (hsfreq = priv->info->hsfreqrange; hsfreq->mbps != 0; hsfreq++) { + if (hsfreq->mbps >= mbps) + break; ++ hsfreq_prev = hsfreq; ++ } + + if (!hsfreq->mbps) { + dev_err(priv->dev, "Unsupported PHY speed (%u Mbps)", mbps); + return -ERANGE; + } + ++ if (hsfreq_prev && ++ ((mbps - hsfreq_prev->mbps) <= (hsfreq->mbps - mbps))) ++ hsfreq = hsfreq_prev; ++ + rcsi2_write(priv, PHYPLL_REG, PHYPLL_HSFREQRANGE(hsfreq->reg)); + + return 0; +@@ -982,10 +989,17 @@ static int rcsi2_phtw_write_mbps(struct rcar_csi2 *priv, unsigned int mbps, + const struct rcsi2_mbps_reg *values, u16 code) + { + const struct rcsi2_mbps_reg *value; ++ const struct rcsi2_mbps_reg *prev_value = NULL; + +- for (value = values; value->mbps; value++) ++ for (value = values; value->mbps; value++) { + if (value->mbps >= mbps) + break; ++ prev_value = value; ++ } ++ ++ if (prev_value && ++ ((mbps - prev_value->mbps) <= (value->mbps - mbps))) ++ value = prev_value; + + if (!value->mbps) { + dev_err(priv->dev, "Unsupported PHY speed (%u Mbps)", mbps); +diff --git a/drivers/media/platform/rcar-vin/rcar-v4l2.c b/drivers/media/platform/rcar-vin/rcar-v4l2.c +index 0d141155f0e3e..eb8c79bac540f 100644 +--- a/drivers/media/platform/rcar-vin/rcar-v4l2.c ++++ b/drivers/media/platform/rcar-vin/rcar-v4l2.c +@@ -175,20 +175,27 @@ static void rvin_format_align(struct rvin_dev *vin, struct v4l2_pix_format *pix) + break; + } + +- /* HW limit width to a multiple of 32 (2^5) for NV12/16 else 2 (2^1) */ ++ /* Hardware limits width alignment based on format. */ + switch (pix->pixelformat) { ++ /* Multiple of 32 (2^5) for NV12/16. */ + case V4L2_PIX_FMT_NV12: + case V4L2_PIX_FMT_NV16: + walign = 5; + break; +- default: ++ /* Multiple of 2 (2^1) for YUV. */ ++ case V4L2_PIX_FMT_YUYV: ++ case V4L2_PIX_FMT_UYVY: + walign = 1; + break; ++ /* No multiple for RGB. */ ++ default: ++ walign = 0; ++ break; + } + + /* Limit to VIN capabilities */ +- v4l_bound_align_image(&pix->width, 2, vin->info->max_width, walign, +- &pix->height, 4, vin->info->max_height, 2, 0); ++ v4l_bound_align_image(&pix->width, 5, vin->info->max_width, walign, ++ &pix->height, 2, vin->info->max_height, 0, 0); + + pix->bytesperline = rvin_format_bytesperline(vin, pix); + pix->sizeimage = rvin_format_sizeimage(pix); +diff --git a/drivers/media/platform/rockchip/rkisp1/rkisp1-dev.c b/drivers/media/platform/rockchip/rkisp1/rkisp1-dev.c +index 7474150b94ed3..560f928c37520 100644 +--- a/drivers/media/platform/rockchip/rkisp1/rkisp1-dev.c ++++ b/drivers/media/platform/rockchip/rkisp1/rkisp1-dev.c +@@ -426,7 +426,7 @@ static void rkisp1_debug_init(struct rkisp1_device *rkisp1) + { + struct rkisp1_debug *debug = &rkisp1->debug; + +- debug->debugfs_dir = debugfs_create_dir(RKISP1_DRIVER_NAME, NULL); ++ debug->debugfs_dir = debugfs_create_dir(dev_name(rkisp1->dev), NULL); + debugfs_create_ulong("data_loss", 0444, debug->debugfs_dir, + &debug->data_loss); + debugfs_create_ulong("outform_size_err", 0444, debug->debugfs_dir, +diff --git a/drivers/media/radio/si470x/radio-si470x-i2c.c b/drivers/media/radio/si470x/radio-si470x-i2c.c +index a972c0705ac79..76d39e2e87706 100644 +--- a/drivers/media/radio/si470x/radio-si470x-i2c.c ++++ b/drivers/media/radio/si470x/radio-si470x-i2c.c +@@ -368,7 +368,7 @@ static int si470x_i2c_probe(struct i2c_client *client) + if (radio->hdl.error) { + retval = radio->hdl.error; + dev_err(&client->dev, "couldn't register control\n"); +- goto err_dev; ++ goto err_all; + } + + /* video device initialization */ +@@ -463,7 +463,6 @@ static int si470x_i2c_probe(struct i2c_client *client) + return 0; + err_all: + v4l2_ctrl_handler_free(&radio->hdl); +-err_dev: + v4l2_device_unregister(&radio->v4l2_dev); + err_initial: + return retval; +diff --git a/drivers/media/rc/igorplugusb.c b/drivers/media/rc/igorplugusb.c +index effaa5751d6c9..3e9988ee785f0 100644 +--- a/drivers/media/rc/igorplugusb.c ++++ b/drivers/media/rc/igorplugusb.c +@@ -64,9 +64,11 @@ static void igorplugusb_irdata(struct igorplugusb *ir, unsigned len) + if (start >= len) { + dev_err(ir->dev, "receive overflow invalid: %u", overflow); + } else { +- if (overflow > 0) ++ if (overflow > 0) { + dev_warn(ir->dev, "receive overflow, at least %u lost", + overflow); ++ ir_raw_event_reset(ir->rc); ++ } + + do { + rawir.duration = ir->buf_in[i] * 85; +diff --git a/drivers/media/rc/mceusb.c b/drivers/media/rc/mceusb.c +index 137a71954aabf..5f296f985b07a 100644 +--- a/drivers/media/rc/mceusb.c ++++ b/drivers/media/rc/mceusb.c +@@ -1430,7 +1430,7 @@ static void mceusb_gen1_init(struct mceusb_dev *ir) + */ + ret = usb_control_msg(ir->usbdev, usb_rcvctrlpipe(ir->usbdev, 0), + USB_REQ_SET_ADDRESS, USB_TYPE_VENDOR, 0, 0, +- data, USB_CTRL_MSG_SZ, HZ * 3); ++ data, USB_CTRL_MSG_SZ, 3000); + dev_dbg(dev, "set address - ret = %d", ret); + dev_dbg(dev, "set address - data[0] = %d, data[1] = %d", + data[0], data[1]); +@@ -1438,20 +1438,20 @@ static void mceusb_gen1_init(struct mceusb_dev *ir) + /* set feature: bit rate 38400 bps */ + ret = usb_control_msg(ir->usbdev, usb_sndctrlpipe(ir->usbdev, 0), + USB_REQ_SET_FEATURE, USB_TYPE_VENDOR, +- 0xc04e, 0x0000, NULL, 0, HZ * 3); ++ 0xc04e, 0x0000, NULL, 0, 3000); + + dev_dbg(dev, "set feature - ret = %d", ret); + + /* bRequest 4: set char length to 8 bits */ + ret = usb_control_msg(ir->usbdev, usb_sndctrlpipe(ir->usbdev, 0), + 4, USB_TYPE_VENDOR, +- 0x0808, 0x0000, NULL, 0, HZ * 3); ++ 0x0808, 0x0000, NULL, 0, 3000); + dev_dbg(dev, "set char length - retB = %d", ret); + + /* bRequest 2: set handshaking to use DTR/DSR */ + ret = usb_control_msg(ir->usbdev, usb_sndctrlpipe(ir->usbdev, 0), + 2, USB_TYPE_VENDOR, +- 0x0000, 0x0100, NULL, 0, HZ * 3); ++ 0x0000, 0x0100, NULL, 0, 3000); + dev_dbg(dev, "set handshake - retC = %d", ret); + + /* device resume */ +diff --git a/drivers/media/rc/redrat3.c b/drivers/media/rc/redrat3.c +index ac85464864b9d..cb22316b3f002 100644 +--- a/drivers/media/rc/redrat3.c ++++ b/drivers/media/rc/redrat3.c +@@ -404,7 +404,7 @@ static int redrat3_send_cmd(int cmd, struct redrat3_dev *rr3) + udev = rr3->udev; + res = usb_control_msg(udev, usb_rcvctrlpipe(udev, 0), cmd, + USB_TYPE_VENDOR | USB_RECIP_DEVICE | USB_DIR_IN, +- 0x0000, 0x0000, data, sizeof(u8), HZ * 10); ++ 0x0000, 0x0000, data, sizeof(u8), 10000); + + if (res < 0) { + dev_err(rr3->dev, "%s: Error sending rr3 cmd res %d, data %d", +@@ -480,7 +480,7 @@ static u32 redrat3_get_timeout(struct redrat3_dev *rr3) + pipe = usb_rcvctrlpipe(rr3->udev, 0); + ret = usb_control_msg(rr3->udev, pipe, RR3_GET_IR_PARAM, + USB_TYPE_VENDOR | USB_RECIP_DEVICE | USB_DIR_IN, +- RR3_IR_IO_SIG_TIMEOUT, 0, tmp, len, HZ * 5); ++ RR3_IR_IO_SIG_TIMEOUT, 0, tmp, len, 5000); + if (ret != len) + dev_warn(rr3->dev, "Failed to read timeout from hardware\n"); + else { +@@ -510,7 +510,7 @@ static int redrat3_set_timeout(struct rc_dev *rc_dev, unsigned int timeoutus) + ret = usb_control_msg(udev, usb_sndctrlpipe(udev, 0), RR3_SET_IR_PARAM, + USB_TYPE_VENDOR | USB_RECIP_DEVICE | USB_DIR_OUT, + RR3_IR_IO_SIG_TIMEOUT, 0, timeout, sizeof(*timeout), +- HZ * 25); ++ 25000); + dev_dbg(dev, "set ir parm timeout %d ret 0x%02x\n", + be32_to_cpu(*timeout), ret); + +@@ -542,32 +542,32 @@ static void redrat3_reset(struct redrat3_dev *rr3) + *val = 0x01; + rc = usb_control_msg(udev, rxpipe, RR3_RESET, + USB_TYPE_VENDOR | USB_RECIP_DEVICE | USB_DIR_IN, +- RR3_CPUCS_REG_ADDR, 0, val, len, HZ * 25); ++ RR3_CPUCS_REG_ADDR, 0, val, len, 25000); + dev_dbg(dev, "reset returned 0x%02x\n", rc); + + *val = length_fuzz; + rc = usb_control_msg(udev, txpipe, RR3_SET_IR_PARAM, + USB_TYPE_VENDOR | USB_RECIP_DEVICE | USB_DIR_OUT, +- RR3_IR_IO_LENGTH_FUZZ, 0, val, len, HZ * 25); ++ RR3_IR_IO_LENGTH_FUZZ, 0, val, len, 25000); + dev_dbg(dev, "set ir parm len fuzz %d rc 0x%02x\n", *val, rc); + + *val = (65536 - (minimum_pause * 2000)) / 256; + rc = usb_control_msg(udev, txpipe, RR3_SET_IR_PARAM, + USB_TYPE_VENDOR | USB_RECIP_DEVICE | USB_DIR_OUT, +- RR3_IR_IO_MIN_PAUSE, 0, val, len, HZ * 25); ++ RR3_IR_IO_MIN_PAUSE, 0, val, len, 25000); + dev_dbg(dev, "set ir parm min pause %d rc 0x%02x\n", *val, rc); + + *val = periods_measure_carrier; + rc = usb_control_msg(udev, txpipe, RR3_SET_IR_PARAM, + USB_TYPE_VENDOR | USB_RECIP_DEVICE | USB_DIR_OUT, +- RR3_IR_IO_PERIODS_MF, 0, val, len, HZ * 25); ++ RR3_IR_IO_PERIODS_MF, 0, val, len, 25000); + dev_dbg(dev, "set ir parm periods measure carrier %d rc 0x%02x", *val, + rc); + + *val = RR3_DRIVER_MAXLENS; + rc = usb_control_msg(udev, txpipe, RR3_SET_IR_PARAM, + USB_TYPE_VENDOR | USB_RECIP_DEVICE | USB_DIR_OUT, +- RR3_IR_IO_MAX_LENGTHS, 0, val, len, HZ * 25); ++ RR3_IR_IO_MAX_LENGTHS, 0, val, len, 25000); + dev_dbg(dev, "set ir parm max lens %d rc 0x%02x\n", *val, rc); + + kfree(val); +@@ -585,7 +585,7 @@ static void redrat3_get_firmware_rev(struct redrat3_dev *rr3) + rc = usb_control_msg(rr3->udev, usb_rcvctrlpipe(rr3->udev, 0), + RR3_FW_VERSION, + USB_TYPE_VENDOR | USB_RECIP_DEVICE | USB_DIR_IN, +- 0, 0, buffer, RR3_FW_VERSION_LEN, HZ * 5); ++ 0, 0, buffer, RR3_FW_VERSION_LEN, 5000); + + if (rc >= 0) + dev_info(rr3->dev, "Firmware rev: %s", buffer); +@@ -825,14 +825,14 @@ static int redrat3_transmit_ir(struct rc_dev *rcdev, unsigned *txbuf, + + pipe = usb_sndbulkpipe(rr3->udev, rr3->ep_out->bEndpointAddress); + ret = usb_bulk_msg(rr3->udev, pipe, irdata, +- sendbuf_len, &ret_len, 10 * HZ); ++ sendbuf_len, &ret_len, 10000); + dev_dbg(dev, "sent %d bytes, (ret %d)\n", ret_len, ret); + + /* now tell the hardware to transmit what we sent it */ + pipe = usb_rcvctrlpipe(rr3->udev, 0); + ret = usb_control_msg(rr3->udev, pipe, RR3_TX_SEND_SIGNAL, + USB_TYPE_VENDOR | USB_RECIP_DEVICE | USB_DIR_IN, +- 0, 0, irdata, 2, HZ * 10); ++ 0, 0, irdata, 2, 10000); + + if (ret < 0) + dev_err(dev, "Error: control msg send failed, rc %d\n", ret); +diff --git a/drivers/media/tuners/msi001.c b/drivers/media/tuners/msi001.c +index 78e6fd600d8ef..44247049a3190 100644 +--- a/drivers/media/tuners/msi001.c ++++ b/drivers/media/tuners/msi001.c +@@ -442,6 +442,13 @@ static int msi001_probe(struct spi_device *spi) + V4L2_CID_RF_TUNER_BANDWIDTH_AUTO, 0, 1, 1, 1); + dev->bandwidth = v4l2_ctrl_new_std(&dev->hdl, &msi001_ctrl_ops, + V4L2_CID_RF_TUNER_BANDWIDTH, 200000, 8000000, 1, 200000); ++ if (dev->hdl.error) { ++ ret = dev->hdl.error; ++ dev_err(&spi->dev, "Could not initialize controls\n"); ++ /* control init failed, free handler */ ++ goto err_ctrl_handler_free; ++ } ++ + v4l2_ctrl_auto_cluster(2, &dev->bandwidth_auto, 0, false); + dev->lna_gain = v4l2_ctrl_new_std(&dev->hdl, &msi001_ctrl_ops, + V4L2_CID_RF_TUNER_LNA_GAIN, 0, 1, 1, 1); +diff --git a/drivers/media/tuners/si2157.c b/drivers/media/tuners/si2157.c +index fefb2625f6558..75ddf7ed1faff 100644 +--- a/drivers/media/tuners/si2157.c ++++ b/drivers/media/tuners/si2157.c +@@ -90,7 +90,7 @@ static int si2157_init(struct dvb_frontend *fe) + dev_dbg(&client->dev, "\n"); + + /* Try to get Xtal trim property, to verify tuner still running */ +- memcpy(cmd.args, "\x15\x00\x04\x02", 4); ++ memcpy(cmd.args, "\x15\x00\x02\x04", 4); + cmd.wlen = 4; + cmd.rlen = 4; + ret = si2157_cmd_execute(client, &cmd); +diff --git a/drivers/media/usb/b2c2/flexcop-usb.c b/drivers/media/usb/b2c2/flexcop-usb.c +index 5d38171b7638c..bfeb92d93de39 100644 +--- a/drivers/media/usb/b2c2/flexcop-usb.c ++++ b/drivers/media/usb/b2c2/flexcop-usb.c +@@ -87,7 +87,7 @@ static int flexcop_usb_readwrite_dw(struct flexcop_device *fc, u16 wRegOffsPCI, + 0, + fc_usb->data, + sizeof(u32), +- B2C2_WAIT_FOR_OPERATION_RDW * HZ); ++ B2C2_WAIT_FOR_OPERATION_RDW); + + if (ret != sizeof(u32)) { + err("error while %s dword from %d (%d).", read ? "reading" : +@@ -155,7 +155,7 @@ static int flexcop_usb_v8_memory_req(struct flexcop_usb *fc_usb, + wIndex, + fc_usb->data, + buflen, +- nWaitTime * HZ); ++ nWaitTime); + if (ret != buflen) + ret = -EIO; + +@@ -248,13 +248,13 @@ static int flexcop_usb_i2c_req(struct flexcop_i2c_adapter *i2c, + /* DKT 020208 - add this to support special case of DiSEqC */ + case USB_FUNC_I2C_CHECKWRITE: + pipe = B2C2_USB_CTRL_PIPE_OUT; +- nWaitTime = 2; ++ nWaitTime = 2000; + request_type |= USB_DIR_OUT; + break; + case USB_FUNC_I2C_READ: + case USB_FUNC_I2C_REPEATREAD: + pipe = B2C2_USB_CTRL_PIPE_IN; +- nWaitTime = 2; ++ nWaitTime = 2000; + request_type |= USB_DIR_IN; + break; + default: +@@ -281,7 +281,7 @@ static int flexcop_usb_i2c_req(struct flexcop_i2c_adapter *i2c, + wIndex, + fc_usb->data, + buflen, +- nWaitTime * HZ); ++ nWaitTime); + + if (ret != buflen) + ret = -EIO; +diff --git a/drivers/media/usb/b2c2/flexcop-usb.h b/drivers/media/usb/b2c2/flexcop-usb.h +index 2f230bf72252b..c7cca1a5ee59d 100644 +--- a/drivers/media/usb/b2c2/flexcop-usb.h ++++ b/drivers/media/usb/b2c2/flexcop-usb.h +@@ -91,13 +91,13 @@ typedef enum { + UTILITY_SRAM_TESTVERIFY = 0x16, + } flexcop_usb_utility_function_t; + +-#define B2C2_WAIT_FOR_OPERATION_RW (1*HZ) +-#define B2C2_WAIT_FOR_OPERATION_RDW (3*HZ) +-#define B2C2_WAIT_FOR_OPERATION_WDW (1*HZ) ++#define B2C2_WAIT_FOR_OPERATION_RW 1000 ++#define B2C2_WAIT_FOR_OPERATION_RDW 3000 ++#define B2C2_WAIT_FOR_OPERATION_WDW 1000 + +-#define B2C2_WAIT_FOR_OPERATION_V8READ (3*HZ) +-#define B2C2_WAIT_FOR_OPERATION_V8WRITE (3*HZ) +-#define B2C2_WAIT_FOR_OPERATION_V8FLASH (3*HZ) ++#define B2C2_WAIT_FOR_OPERATION_V8READ 3000 ++#define B2C2_WAIT_FOR_OPERATION_V8WRITE 3000 ++#define B2C2_WAIT_FOR_OPERATION_V8FLASH 3000 + + typedef enum { + V8_MEMORY_PAGE_DVB_CI = 0x20, +diff --git a/drivers/media/usb/cpia2/cpia2_usb.c b/drivers/media/usb/cpia2/cpia2_usb.c +index 76aac06f9fb8e..cba03b2864738 100644 +--- a/drivers/media/usb/cpia2/cpia2_usb.c ++++ b/drivers/media/usb/cpia2/cpia2_usb.c +@@ -550,7 +550,7 @@ static int write_packet(struct usb_device *udev, + 0, /* index */ + buf, /* buffer */ + size, +- HZ); ++ 1000); + + kfree(buf); + return ret; +@@ -582,7 +582,7 @@ static int read_packet(struct usb_device *udev, + 0, /* index */ + buf, /* buffer */ + size, +- HZ); ++ 1000); + + if (ret >= 0) + memcpy(registers, buf, size); +diff --git a/drivers/media/usb/dvb-usb/dib0700_core.c b/drivers/media/usb/dvb-usb/dib0700_core.c +index 70219b3e85666..7ea8f68b0f458 100644 +--- a/drivers/media/usb/dvb-usb/dib0700_core.c ++++ b/drivers/media/usb/dvb-usb/dib0700_core.c +@@ -618,8 +618,6 @@ int dib0700_streaming_ctrl(struct dvb_usb_adapter *adap, int onoff) + deb_info("the endpoint number (%i) is not correct, use the adapter id instead", adap->fe_adap[0].stream.props.endpoint); + if (onoff) + st->channel_state |= 1 << (adap->id); +- else +- st->channel_state |= 1 << ~(adap->id); + } else { + if (onoff) + st->channel_state |= 1 << (adap->fe_adap[0].stream.props.endpoint-2); +diff --git a/drivers/media/usb/dvb-usb/dw2102.c b/drivers/media/usb/dvb-usb/dw2102.c +index f0e686b05dc63..ca75ebdc10b37 100644 +--- a/drivers/media/usb/dvb-usb/dw2102.c ++++ b/drivers/media/usb/dvb-usb/dw2102.c +@@ -2150,46 +2150,153 @@ static struct dvb_usb_device_properties s6x0_properties = { + } + }; + +-static const struct dvb_usb_device_description d1100 = { +- "Prof 1100 USB ", +- {&dw2102_table[PROF_1100], NULL}, +- {NULL}, +-}; ++static struct dvb_usb_device_properties p1100_properties = { ++ .caps = DVB_USB_IS_AN_I2C_ADAPTER, ++ .usb_ctrl = DEVICE_SPECIFIC, ++ .size_of_priv = sizeof(struct dw2102_state), ++ .firmware = P1100_FIRMWARE, ++ .no_reconnect = 1, + +-static const struct dvb_usb_device_description d660 = { +- "TeVii S660 USB", +- {&dw2102_table[TEVII_S660], NULL}, +- {NULL}, +-}; ++ .i2c_algo = &s6x0_i2c_algo, ++ .rc.core = { ++ .rc_interval = 150, ++ .rc_codes = RC_MAP_TBS_NEC, ++ .module_name = "dw2102", ++ .allowed_protos = RC_PROTO_BIT_NEC, ++ .rc_query = prof_rc_query, ++ }, + +-static const struct dvb_usb_device_description d480_1 = { +- "TeVii S480.1 USB", +- {&dw2102_table[TEVII_S480_1], NULL}, +- {NULL}, ++ .generic_bulk_ctrl_endpoint = 0x81, ++ .num_adapters = 1, ++ .download_firmware = dw2102_load_firmware, ++ .read_mac_address = s6x0_read_mac_address, ++ .adapter = { ++ { ++ .num_frontends = 1, ++ .fe = {{ ++ .frontend_attach = stv0288_frontend_attach, ++ .stream = { ++ .type = USB_BULK, ++ .count = 8, ++ .endpoint = 0x82, ++ .u = { ++ .bulk = { ++ .buffersize = 4096, ++ } ++ } ++ }, ++ } }, ++ } ++ }, ++ .num_device_descs = 1, ++ .devices = { ++ {"Prof 1100 USB ", ++ {&dw2102_table[PROF_1100], NULL}, ++ {NULL}, ++ }, ++ } + }; + +-static const struct dvb_usb_device_description d480_2 = { +- "TeVii S480.2 USB", +- {&dw2102_table[TEVII_S480_2], NULL}, +- {NULL}, +-}; ++static struct dvb_usb_device_properties s660_properties = { ++ .caps = DVB_USB_IS_AN_I2C_ADAPTER, ++ .usb_ctrl = DEVICE_SPECIFIC, ++ .size_of_priv = sizeof(struct dw2102_state), ++ .firmware = S660_FIRMWARE, ++ .no_reconnect = 1, + +-static const struct dvb_usb_device_description d7500 = { +- "Prof 7500 USB DVB-S2", +- {&dw2102_table[PROF_7500], NULL}, +- {NULL}, +-}; ++ .i2c_algo = &s6x0_i2c_algo, ++ .rc.core = { ++ .rc_interval = 150, ++ .rc_codes = RC_MAP_TEVII_NEC, ++ .module_name = "dw2102", ++ .allowed_protos = RC_PROTO_BIT_NEC, ++ .rc_query = dw2102_rc_query, ++ }, + +-static const struct dvb_usb_device_description d421 = { +- "TeVii S421 PCI", +- {&dw2102_table[TEVII_S421], NULL}, +- {NULL}, ++ .generic_bulk_ctrl_endpoint = 0x81, ++ .num_adapters = 1, ++ .download_firmware = dw2102_load_firmware, ++ .read_mac_address = s6x0_read_mac_address, ++ .adapter = { ++ { ++ .num_frontends = 1, ++ .fe = {{ ++ .frontend_attach = ds3000_frontend_attach, ++ .stream = { ++ .type = USB_BULK, ++ .count = 8, ++ .endpoint = 0x82, ++ .u = { ++ .bulk = { ++ .buffersize = 4096, ++ } ++ } ++ }, ++ } }, ++ } ++ }, ++ .num_device_descs = 3, ++ .devices = { ++ {"TeVii S660 USB", ++ {&dw2102_table[TEVII_S660], NULL}, ++ {NULL}, ++ }, ++ {"TeVii S480.1 USB", ++ {&dw2102_table[TEVII_S480_1], NULL}, ++ {NULL}, ++ }, ++ {"TeVii S480.2 USB", ++ {&dw2102_table[TEVII_S480_2], NULL}, ++ {NULL}, ++ }, ++ } + }; + +-static const struct dvb_usb_device_description d632 = { +- "TeVii S632 USB", +- {&dw2102_table[TEVII_S632], NULL}, +- {NULL}, ++static struct dvb_usb_device_properties p7500_properties = { ++ .caps = DVB_USB_IS_AN_I2C_ADAPTER, ++ .usb_ctrl = DEVICE_SPECIFIC, ++ .size_of_priv = sizeof(struct dw2102_state), ++ .firmware = P7500_FIRMWARE, ++ .no_reconnect = 1, ++ ++ .i2c_algo = &s6x0_i2c_algo, ++ .rc.core = { ++ .rc_interval = 150, ++ .rc_codes = RC_MAP_TBS_NEC, ++ .module_name = "dw2102", ++ .allowed_protos = RC_PROTO_BIT_NEC, ++ .rc_query = prof_rc_query, ++ }, ++ ++ .generic_bulk_ctrl_endpoint = 0x81, ++ .num_adapters = 1, ++ .download_firmware = dw2102_load_firmware, ++ .read_mac_address = s6x0_read_mac_address, ++ .adapter = { ++ { ++ .num_frontends = 1, ++ .fe = {{ ++ .frontend_attach = prof_7500_frontend_attach, ++ .stream = { ++ .type = USB_BULK, ++ .count = 8, ++ .endpoint = 0x82, ++ .u = { ++ .bulk = { ++ .buffersize = 4096, ++ } ++ } ++ }, ++ } }, ++ } ++ }, ++ .num_device_descs = 1, ++ .devices = { ++ {"Prof 7500 USB DVB-S2", ++ {&dw2102_table[PROF_7500], NULL}, ++ {NULL}, ++ }, ++ } + }; + + static struct dvb_usb_device_properties su3000_properties = { +@@ -2273,6 +2380,59 @@ static struct dvb_usb_device_properties su3000_properties = { + } + }; + ++static struct dvb_usb_device_properties s421_properties = { ++ .caps = DVB_USB_IS_AN_I2C_ADAPTER, ++ .usb_ctrl = DEVICE_SPECIFIC, ++ .size_of_priv = sizeof(struct dw2102_state), ++ .power_ctrl = su3000_power_ctrl, ++ .num_adapters = 1, ++ .identify_state = su3000_identify_state, ++ .i2c_algo = &su3000_i2c_algo, ++ ++ .rc.core = { ++ .rc_interval = 150, ++ .rc_codes = RC_MAP_SU3000, ++ .module_name = "dw2102", ++ .allowed_protos = RC_PROTO_BIT_RC5, ++ .rc_query = su3000_rc_query, ++ }, ++ ++ .read_mac_address = su3000_read_mac_address, ++ ++ .generic_bulk_ctrl_endpoint = 0x01, ++ ++ .adapter = { ++ { ++ .num_frontends = 1, ++ .fe = {{ ++ .streaming_ctrl = su3000_streaming_ctrl, ++ .frontend_attach = m88rs2000_frontend_attach, ++ .stream = { ++ .type = USB_BULK, ++ .count = 8, ++ .endpoint = 0x82, ++ .u = { ++ .bulk = { ++ .buffersize = 4096, ++ } ++ } ++ } ++ } }, ++ } ++ }, ++ .num_device_descs = 2, ++ .devices = { ++ { "TeVii S421 PCI", ++ { &dw2102_table[TEVII_S421], NULL }, ++ { NULL }, ++ }, ++ { "TeVii S632 USB", ++ { &dw2102_table[TEVII_S632], NULL }, ++ { NULL }, ++ }, ++ } ++}; ++ + static struct dvb_usb_device_properties t220_properties = { + .caps = DVB_USB_IS_AN_I2C_ADAPTER, + .usb_ctrl = DEVICE_SPECIFIC, +@@ -2390,101 +2550,33 @@ static struct dvb_usb_device_properties tt_s2_4600_properties = { + static int dw2102_probe(struct usb_interface *intf, + const struct usb_device_id *id) + { +- int retval = -ENOMEM; +- struct dvb_usb_device_properties *p1100; +- struct dvb_usb_device_properties *s660; +- struct dvb_usb_device_properties *p7500; +- struct dvb_usb_device_properties *s421; +- +- p1100 = kmemdup(&s6x0_properties, +- sizeof(struct dvb_usb_device_properties), GFP_KERNEL); +- if (!p1100) +- goto err0; +- +- /* copy default structure */ +- /* fill only different fields */ +- p1100->firmware = P1100_FIRMWARE; +- p1100->devices[0] = d1100; +- p1100->rc.core.rc_query = prof_rc_query; +- p1100->rc.core.rc_codes = RC_MAP_TBS_NEC; +- p1100->adapter->fe[0].frontend_attach = stv0288_frontend_attach; +- +- s660 = kmemdup(&s6x0_properties, +- sizeof(struct dvb_usb_device_properties), GFP_KERNEL); +- if (!s660) +- goto err1; +- +- s660->firmware = S660_FIRMWARE; +- s660->num_device_descs = 3; +- s660->devices[0] = d660; +- s660->devices[1] = d480_1; +- s660->devices[2] = d480_2; +- s660->adapter->fe[0].frontend_attach = ds3000_frontend_attach; +- +- p7500 = kmemdup(&s6x0_properties, +- sizeof(struct dvb_usb_device_properties), GFP_KERNEL); +- if (!p7500) +- goto err2; +- +- p7500->firmware = P7500_FIRMWARE; +- p7500->devices[0] = d7500; +- p7500->rc.core.rc_query = prof_rc_query; +- p7500->rc.core.rc_codes = RC_MAP_TBS_NEC; +- p7500->adapter->fe[0].frontend_attach = prof_7500_frontend_attach; +- +- +- s421 = kmemdup(&su3000_properties, +- sizeof(struct dvb_usb_device_properties), GFP_KERNEL); +- if (!s421) +- goto err3; +- +- s421->num_device_descs = 2; +- s421->devices[0] = d421; +- s421->devices[1] = d632; +- s421->adapter->fe[0].frontend_attach = m88rs2000_frontend_attach; +- +- if (0 == dvb_usb_device_init(intf, &dw2102_properties, +- THIS_MODULE, NULL, adapter_nr) || +- 0 == dvb_usb_device_init(intf, &dw2104_properties, +- THIS_MODULE, NULL, adapter_nr) || +- 0 == dvb_usb_device_init(intf, &dw3101_properties, +- THIS_MODULE, NULL, adapter_nr) || +- 0 == dvb_usb_device_init(intf, &s6x0_properties, +- THIS_MODULE, NULL, adapter_nr) || +- 0 == dvb_usb_device_init(intf, p1100, +- THIS_MODULE, NULL, adapter_nr) || +- 0 == dvb_usb_device_init(intf, s660, +- THIS_MODULE, NULL, adapter_nr) || +- 0 == dvb_usb_device_init(intf, p7500, +- THIS_MODULE, NULL, adapter_nr) || +- 0 == dvb_usb_device_init(intf, s421, +- THIS_MODULE, NULL, adapter_nr) || +- 0 == dvb_usb_device_init(intf, &su3000_properties, +- THIS_MODULE, NULL, adapter_nr) || +- 0 == dvb_usb_device_init(intf, &t220_properties, +- THIS_MODULE, NULL, adapter_nr) || +- 0 == dvb_usb_device_init(intf, &tt_s2_4600_properties, +- THIS_MODULE, NULL, adapter_nr)) { +- +- /* clean up copied properties */ +- kfree(s421); +- kfree(p7500); +- kfree(s660); +- kfree(p1100); ++ if (!(dvb_usb_device_init(intf, &dw2102_properties, ++ THIS_MODULE, NULL, adapter_nr) && ++ dvb_usb_device_init(intf, &dw2104_properties, ++ THIS_MODULE, NULL, adapter_nr) && ++ dvb_usb_device_init(intf, &dw3101_properties, ++ THIS_MODULE, NULL, adapter_nr) && ++ dvb_usb_device_init(intf, &s6x0_properties, ++ THIS_MODULE, NULL, adapter_nr) && ++ dvb_usb_device_init(intf, &p1100_properties, ++ THIS_MODULE, NULL, adapter_nr) && ++ dvb_usb_device_init(intf, &s660_properties, ++ THIS_MODULE, NULL, adapter_nr) && ++ dvb_usb_device_init(intf, &p7500_properties, ++ THIS_MODULE, NULL, adapter_nr) && ++ dvb_usb_device_init(intf, &s421_properties, ++ THIS_MODULE, NULL, adapter_nr) && ++ dvb_usb_device_init(intf, &su3000_properties, ++ THIS_MODULE, NULL, adapter_nr) && ++ dvb_usb_device_init(intf, &t220_properties, ++ THIS_MODULE, NULL, adapter_nr) && ++ dvb_usb_device_init(intf, &tt_s2_4600_properties, ++ THIS_MODULE, NULL, adapter_nr))) { + + return 0; + } + +- retval = -ENODEV; +- kfree(s421); +-err3: +- kfree(p7500); +-err2: +- kfree(s660); +-err1: +- kfree(p1100); +-err0: +- return retval; ++ return -ENODEV; + } + + static void dw2102_disconnect(struct usb_interface *intf) +diff --git a/drivers/media/usb/dvb-usb/m920x.c b/drivers/media/usb/dvb-usb/m920x.c +index 4bb5b82599a79..691e05833db19 100644 +--- a/drivers/media/usb/dvb-usb/m920x.c ++++ b/drivers/media/usb/dvb-usb/m920x.c +@@ -274,6 +274,13 @@ static int m920x_i2c_xfer(struct i2c_adapter *adap, struct i2c_msg msg[], int nu + /* Should check for ack here, if we knew how. */ + } + if (msg[i].flags & I2C_M_RD) { ++ char *read = kmalloc(1, GFP_KERNEL); ++ if (!read) { ++ ret = -ENOMEM; ++ kfree(read); ++ goto unlock; ++ } ++ + for (j = 0; j < msg[i].len; j++) { + /* Last byte of transaction? + * Send STOP, otherwise send ACK. */ +@@ -281,9 +288,12 @@ static int m920x_i2c_xfer(struct i2c_adapter *adap, struct i2c_msg msg[], int nu + + if ((ret = m920x_read(d->udev, M9206_I2C, 0x0, + 0x20 | stop, +- &msg[i].buf[j], 1)) != 0) ++ read, 1)) != 0) + goto unlock; ++ msg[i].buf[j] = read[0]; + } ++ ++ kfree(read); + } else { + for (j = 0; j < msg[i].len; j++) { + /* Last byte of transaction? Then send STOP. */ +diff --git a/drivers/media/usb/em28xx/em28xx-cards.c b/drivers/media/usb/em28xx/em28xx-cards.c +index 948e22e29b42a..ebc430b05f21c 100644 +--- a/drivers/media/usb/em28xx/em28xx-cards.c ++++ b/drivers/media/usb/em28xx/em28xx-cards.c +@@ -3625,8 +3625,10 @@ static int em28xx_init_dev(struct em28xx *dev, struct usb_device *udev, + + if (dev->is_audio_only) { + retval = em28xx_audio_setup(dev); +- if (retval) +- return -ENODEV; ++ if (retval) { ++ retval = -ENODEV; ++ goto err_deinit_media; ++ } + em28xx_init_extension(dev); + + return 0; +@@ -3645,7 +3647,7 @@ static int em28xx_init_dev(struct em28xx *dev, struct usb_device *udev, + dev_err(&dev->intf->dev, + "%s: em28xx_i2c_register bus 0 - error [%d]!\n", + __func__, retval); +- return retval; ++ goto err_deinit_media; + } + + /* register i2c bus 1 */ +@@ -3661,9 +3663,7 @@ static int em28xx_init_dev(struct em28xx *dev, struct usb_device *udev, + "%s: em28xx_i2c_register bus 1 - error [%d]!\n", + __func__, retval); + +- em28xx_i2c_unregister(dev, 0); +- +- return retval; ++ goto err_unreg_i2c; + } + } + +@@ -3671,6 +3671,12 @@ static int em28xx_init_dev(struct em28xx *dev, struct usb_device *udev, + em28xx_card_setup(dev); + + return 0; ++ ++err_unreg_i2c: ++ em28xx_i2c_unregister(dev, 0); ++err_deinit_media: ++ em28xx_unregister_media_device(dev); ++ return retval; + } + + static int em28xx_duplicate_dev(struct em28xx *dev) +diff --git a/drivers/media/usb/em28xx/em28xx-core.c b/drivers/media/usb/em28xx/em28xx-core.c +index acc0bf7dbe2b1..c837cc528a335 100644 +--- a/drivers/media/usb/em28xx/em28xx-core.c ++++ b/drivers/media/usb/em28xx/em28xx-core.c +@@ -89,7 +89,7 @@ int em28xx_read_reg_req_len(struct em28xx *dev, u8 req, u16 reg, + mutex_lock(&dev->ctrl_urb_lock); + ret = usb_control_msg(udev, pipe, req, + USB_DIR_IN | USB_TYPE_VENDOR | USB_RECIP_DEVICE, +- 0x0000, reg, dev->urb_buf, len, HZ); ++ 0x0000, reg, dev->urb_buf, len, 1000); + if (ret < 0) { + em28xx_regdbg("(pipe 0x%08x): IN: %02x %02x %02x %02x %02x %02x %02x %02x failed with error %i\n", + pipe, +@@ -158,7 +158,7 @@ int em28xx_write_regs_req(struct em28xx *dev, u8 req, u16 reg, char *buf, + memcpy(dev->urb_buf, buf, len); + ret = usb_control_msg(udev, pipe, req, + USB_DIR_OUT | USB_TYPE_VENDOR | USB_RECIP_DEVICE, +- 0x0000, reg, dev->urb_buf, len, HZ); ++ 0x0000, reg, dev->urb_buf, len, 1000); + mutex_unlock(&dev->ctrl_urb_lock); + + if (ret < 0) { +diff --git a/drivers/media/usb/pvrusb2/pvrusb2-hdw.c b/drivers/media/usb/pvrusb2/pvrusb2-hdw.c +index d38dee1792e41..3915d551d59e7 100644 +--- a/drivers/media/usb/pvrusb2/pvrusb2-hdw.c ++++ b/drivers/media/usb/pvrusb2/pvrusb2-hdw.c +@@ -1467,7 +1467,7 @@ static int pvr2_upload_firmware1(struct pvr2_hdw *hdw) + for (address = 0; address < fwsize; address += 0x800) { + memcpy(fw_ptr, fw_entry->data + address, 0x800); + ret += usb_control_msg(hdw->usb_dev, pipe, 0xa0, 0x40, address, +- 0, fw_ptr, 0x800, HZ); ++ 0, fw_ptr, 0x800, 1000); + } + + trace_firmware("Upload done, releasing device's CPU"); +@@ -1605,7 +1605,7 @@ int pvr2_upload_firmware2(struct pvr2_hdw *hdw) + ((u32 *)fw_ptr)[icnt] = swab32(((u32 *)fw_ptr)[icnt]); + + ret |= usb_bulk_msg(hdw->usb_dev, pipe, fw_ptr,bcnt, +- &actual_length, HZ); ++ &actual_length, 1000); + ret |= (actual_length != bcnt); + if (ret) break; + fw_done += bcnt; +@@ -3438,7 +3438,7 @@ void pvr2_hdw_cpufw_set_enabled(struct pvr2_hdw *hdw, + 0xa0,0xc0, + address,0, + hdw->fw_buffer+address, +- 0x800,HZ); ++ 0x800,1000); + if (ret < 0) break; + } + +@@ -3977,7 +3977,7 @@ void pvr2_hdw_cpureset_assert(struct pvr2_hdw *hdw,int val) + /* Write the CPUCS register on the 8051. The lsb of the register + is the reset bit; a 1 asserts reset while a 0 clears it. */ + pipe = usb_sndctrlpipe(hdw->usb_dev, 0); +- ret = usb_control_msg(hdw->usb_dev,pipe,0xa0,0x40,0xe600,0,da,1,HZ); ++ ret = usb_control_msg(hdw->usb_dev,pipe,0xa0,0x40,0xe600,0,da,1,1000); + if (ret < 0) { + pvr2_trace(PVR2_TRACE_ERROR_LEGS, + "cpureset_assert(%d) error=%d",val,ret); +diff --git a/drivers/media/usb/s2255/s2255drv.c b/drivers/media/usb/s2255/s2255drv.c +index 3b0e4ed75d99c..acf18e2251a52 100644 +--- a/drivers/media/usb/s2255/s2255drv.c ++++ b/drivers/media/usb/s2255/s2255drv.c +@@ -1882,7 +1882,7 @@ static long s2255_vendor_req(struct s2255_dev *dev, unsigned char Request, + USB_TYPE_VENDOR | USB_RECIP_DEVICE | + USB_DIR_IN, + Value, Index, buf, +- TransferBufferLength, HZ * 5); ++ TransferBufferLength, USB_CTRL_SET_TIMEOUT); + + if (r >= 0) + memcpy(TransferBuffer, buf, TransferBufferLength); +@@ -1891,7 +1891,7 @@ static long s2255_vendor_req(struct s2255_dev *dev, unsigned char Request, + r = usb_control_msg(dev->udev, usb_sndctrlpipe(dev->udev, 0), + Request, USB_TYPE_VENDOR | USB_RECIP_DEVICE, + Value, Index, buf, +- TransferBufferLength, HZ * 5); ++ TransferBufferLength, USB_CTRL_SET_TIMEOUT); + } + kfree(buf); + return r; +diff --git a/drivers/media/usb/stk1160/stk1160-core.c b/drivers/media/usb/stk1160/stk1160-core.c +index b4f8bc5db1389..4e1698f788187 100644 +--- a/drivers/media/usb/stk1160/stk1160-core.c ++++ b/drivers/media/usb/stk1160/stk1160-core.c +@@ -65,7 +65,7 @@ int stk1160_read_reg(struct stk1160 *dev, u16 reg, u8 *value) + return -ENOMEM; + ret = usb_control_msg(dev->udev, pipe, 0x00, + USB_DIR_IN | USB_TYPE_VENDOR | USB_RECIP_DEVICE, +- 0x00, reg, buf, sizeof(u8), HZ); ++ 0x00, reg, buf, sizeof(u8), 1000); + if (ret < 0) { + stk1160_err("read failed on reg 0x%x (%d)\n", + reg, ret); +@@ -85,7 +85,7 @@ int stk1160_write_reg(struct stk1160 *dev, u16 reg, u16 value) + + ret = usb_control_msg(dev->udev, pipe, 0x01, + USB_DIR_OUT | USB_TYPE_VENDOR | USB_RECIP_DEVICE, +- value, reg, NULL, 0, HZ); ++ value, reg, NULL, 0, 1000); + if (ret < 0) { + stk1160_err("write failed on reg 0x%x (%d)\n", + reg, ret); +diff --git a/drivers/media/usb/uvc/uvcvideo.h b/drivers/media/usb/uvc/uvcvideo.h +index cce5e38133cd3..c3ea6a53869f5 100644 +--- a/drivers/media/usb/uvc/uvcvideo.h ++++ b/drivers/media/usb/uvc/uvcvideo.h +@@ -189,7 +189,7 @@ + /* Maximum status buffer size in bytes of interrupt URB. */ + #define UVC_MAX_STATUS_SIZE 16 + +-#define UVC_CTRL_CONTROL_TIMEOUT 500 ++#define UVC_CTRL_CONTROL_TIMEOUT 5000 + #define UVC_CTRL_STREAMING_TIMEOUT 5000 + + /* Maximum allowed number of control mappings per device */ +diff --git a/drivers/media/v4l2-core/v4l2-ioctl.c b/drivers/media/v4l2-core/v4l2-ioctl.c +index f4f67b385d00a..c7308a2a80a0f 100644 +--- a/drivers/media/v4l2-core/v4l2-ioctl.c ++++ b/drivers/media/v4l2-core/v4l2-ioctl.c +@@ -2088,6 +2088,7 @@ static int v4l_prepare_buf(const struct v4l2_ioctl_ops *ops, + static int v4l_g_parm(const struct v4l2_ioctl_ops *ops, + struct file *file, void *fh, void *arg) + { ++ struct video_device *vfd = video_devdata(file); + struct v4l2_streamparm *p = arg; + v4l2_std_id std; + int ret = check_fmt(file, p->type); +@@ -2099,7 +2100,8 @@ static int v4l_g_parm(const struct v4l2_ioctl_ops *ops, + if (p->type != V4L2_BUF_TYPE_VIDEO_CAPTURE && + p->type != V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE) + return -EINVAL; +- p->parm.capture.readbuffers = 2; ++ if (vfd->device_caps & V4L2_CAP_READWRITE) ++ p->parm.capture.readbuffers = 2; + ret = ops->vidioc_g_std(file, fh, &std); + if (ret == 0) + v4l2_video_std_frame_period(std, &p->parm.capture.timeperframe); +diff --git a/drivers/memory/renesas-rpc-if.c b/drivers/memory/renesas-rpc-if.c +index 77a011d5ff8c1..8618702233002 100644 +--- a/drivers/memory/renesas-rpc-if.c ++++ b/drivers/memory/renesas-rpc-if.c +@@ -244,7 +244,7 @@ int rpcif_sw_init(struct rpcif *rpc, struct device *dev) + res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "dirmap"); + rpc->dirmap = devm_ioremap_resource(&pdev->dev, res); + if (IS_ERR(rpc->dirmap)) +- rpc->dirmap = NULL; ++ return PTR_ERR(rpc->dirmap); + rpc->size = resource_size(res); + + rpc->rstc = devm_reset_control_get_exclusive(&pdev->dev, NULL); +diff --git a/drivers/mfd/atmel-flexcom.c b/drivers/mfd/atmel-flexcom.c +index d2f5c073fdf31..559eb4d352b68 100644 +--- a/drivers/mfd/atmel-flexcom.c ++++ b/drivers/mfd/atmel-flexcom.c +@@ -87,8 +87,7 @@ static const struct of_device_id atmel_flexcom_of_match[] = { + }; + MODULE_DEVICE_TABLE(of, atmel_flexcom_of_match); + +-#ifdef CONFIG_PM_SLEEP +-static int atmel_flexcom_resume(struct device *dev) ++static int __maybe_unused atmel_flexcom_resume_noirq(struct device *dev) + { + struct atmel_flexcom *ddata = dev_get_drvdata(dev); + int err; +@@ -105,16 +104,16 @@ static int atmel_flexcom_resume(struct device *dev) + + return 0; + } +-#endif + +-static SIMPLE_DEV_PM_OPS(atmel_flexcom_pm_ops, NULL, +- atmel_flexcom_resume); ++static const struct dev_pm_ops atmel_flexcom_pm_ops = { ++ .resume_noirq = atmel_flexcom_resume_noirq, ++}; + + static struct platform_driver atmel_flexcom_driver = { + .probe = atmel_flexcom_probe, + .driver = { + .name = "atmel_flexcom", +- .pm = &atmel_flexcom_pm_ops, ++ .pm = pm_ptr(&atmel_flexcom_pm_ops), + .of_match_table = atmel_flexcom_of_match, + }, + }; +diff --git a/drivers/mfd/tps65910.c b/drivers/mfd/tps65910.c +index 6e105cca27d47..67e2707af4bce 100644 +--- a/drivers/mfd/tps65910.c ++++ b/drivers/mfd/tps65910.c +@@ -436,15 +436,6 @@ static void tps65910_power_off(void) + + tps65910 = dev_get_drvdata(&tps65910_i2c_client->dev); + +- /* +- * The PWR_OFF bit needs to be set separately, before transitioning +- * to the OFF state. It enables the "sequential" power-off mode on +- * TPS65911, it's a NO-OP on TPS65910. +- */ +- if (regmap_set_bits(tps65910->regmap, TPS65910_DEVCTRL, +- DEVCTRL_PWR_OFF_MASK) < 0) +- return; +- + regmap_update_bits(tps65910->regmap, TPS65910_DEVCTRL, + DEVCTRL_DEV_OFF_MASK | DEVCTRL_DEV_ON_MASK, + DEVCTRL_DEV_OFF_MASK); +@@ -504,6 +495,19 @@ static int tps65910_i2c_probe(struct i2c_client *i2c, + tps65910_sleepinit(tps65910, pmic_plat_data); + + if (pmic_plat_data->pm_off && !pm_power_off) { ++ /* ++ * The PWR_OFF bit needs to be set separately, before ++ * transitioning to the OFF state. It enables the "sequential" ++ * power-off mode on TPS65911, it's a NO-OP on TPS65910. ++ */ ++ ret = regmap_set_bits(tps65910->regmap, TPS65910_DEVCTRL, ++ DEVCTRL_PWR_OFF_MASK); ++ if (ret) { ++ dev_err(&i2c->dev, "failed to set power-off mode: %d\n", ++ ret); ++ return ret; ++ } ++ + tps65910_i2c_client = i2c; + pm_power_off = tps65910_power_off; + } +diff --git a/drivers/misc/eeprom/at25.c b/drivers/misc/eeprom/at25.c +index b38978a3b3ffa..9193b812bc07e 100644 +--- a/drivers/misc/eeprom/at25.c ++++ b/drivers/misc/eeprom/at25.c +@@ -17,8 +17,6 @@ + #include <linux/spi/spi.h> + #include <linux/spi/eeprom.h> + #include <linux/property.h> +-#include <linux/of.h> +-#include <linux/of_device.h> + #include <linux/math.h> + + /* +@@ -380,13 +378,14 @@ static int at25_probe(struct spi_device *spi) + int sr; + u8 id[FM25_ID_LEN]; + u8 sernum[FM25_SN_LEN]; ++ bool is_fram; + int i; +- const struct of_device_id *match; +- bool is_fram = 0; + +- match = of_match_device(of_match_ptr(at25_of_match), &spi->dev); +- if (match && !strcmp(match->compatible, "cypress,fm25")) +- is_fram = 1; ++ err = device_property_match_string(&spi->dev, "compatible", "cypress,fm25"); ++ if (err >= 0) ++ is_fram = true; ++ else ++ is_fram = false; + + at25 = devm_kzalloc(&spi->dev, sizeof(struct at25_data), GFP_KERNEL); + if (!at25) +diff --git a/drivers/misc/habanalabs/common/firmware_if.c b/drivers/misc/habanalabs/common/firmware_if.c +index 8d2568c63f19e..a8e683964ab03 100644 +--- a/drivers/misc/habanalabs/common/firmware_if.c ++++ b/drivers/misc/habanalabs/common/firmware_if.c +@@ -1703,6 +1703,9 @@ static int hl_fw_dynamic_validate_descriptor(struct hl_device *hdev, + return rc; + } + ++ /* here we can mark the descriptor as valid as the content has been validated */ ++ fw_loader->dynamic_loader.fw_desc_valid = true; ++ + return 0; + } + +@@ -1759,7 +1762,13 @@ static int hl_fw_dynamic_read_and_validate_descriptor(struct hl_device *hdev, + return rc; + } + +- /* extract address copy the descriptor from */ ++ /* ++ * extract address to copy the descriptor from ++ * in addition, as the descriptor value is going to be over-ridden by new data- we mark it ++ * as invalid. ++ * it will be marked again as valid once validated ++ */ ++ fw_loader->dynamic_loader.fw_desc_valid = false; + src = hdev->pcie_bar[region->bar_id] + region->offset_in_bar + + response->ram_offset; + memcpy_fromio(fw_desc, src, sizeof(struct lkd_fw_comms_desc)); +@@ -2239,6 +2248,9 @@ static int hl_fw_dynamic_init_cpu(struct hl_device *hdev, + dev_info(hdev->dev, + "Loading firmware to device, may take some time...\n"); + ++ /* initialize FW descriptor as invalid */ ++ fw_loader->dynamic_loader.fw_desc_valid = false; ++ + /* + * In this stage, "cpu_dyn_regs" contains only LKD's hard coded values! + * It will be updated from FW after hl_fw_dynamic_request_descriptor(). +@@ -2325,7 +2337,8 @@ static int hl_fw_dynamic_init_cpu(struct hl_device *hdev, + return 0; + + protocol_err: +- fw_read_errors(hdev, le32_to_cpu(dyn_regs->cpu_boot_err0), ++ if (fw_loader->dynamic_loader.fw_desc_valid) ++ fw_read_errors(hdev, le32_to_cpu(dyn_regs->cpu_boot_err0), + le32_to_cpu(dyn_regs->cpu_boot_err1), + le32_to_cpu(dyn_regs->cpu_boot_dev_sts0), + le32_to_cpu(dyn_regs->cpu_boot_dev_sts1)); +diff --git a/drivers/misc/habanalabs/common/habanalabs.h b/drivers/misc/habanalabs/common/habanalabs.h +index bebebcb163ee8..dfcd87b98ca08 100644 +--- a/drivers/misc/habanalabs/common/habanalabs.h ++++ b/drivers/misc/habanalabs/common/habanalabs.h +@@ -992,6 +992,7 @@ struct fw_response { + * @image_region: region to copy the FW image to + * @fw_image_size: size of FW image to load + * @wait_for_bl_timeout: timeout for waiting for boot loader to respond ++ * @fw_desc_valid: true if FW descriptor has been validated and hence the data can be used + */ + struct dynamic_fw_load_mgr { + struct fw_response response; +@@ -999,6 +1000,7 @@ struct dynamic_fw_load_mgr { + struct pci_mem_region *image_region; + size_t fw_image_size; + u32 wait_for_bl_timeout; ++ bool fw_desc_valid; + }; + + /** +diff --git a/drivers/misc/lattice-ecp3-config.c b/drivers/misc/lattice-ecp3-config.c +index 0f54730c7ed56..98828030b5a4d 100644 +--- a/drivers/misc/lattice-ecp3-config.c ++++ b/drivers/misc/lattice-ecp3-config.c +@@ -76,12 +76,12 @@ static void firmware_load(const struct firmware *fw, void *context) + + if (fw == NULL) { + dev_err(&spi->dev, "Cannot load firmware, aborting\n"); +- return; ++ goto out; + } + + if (fw->size == 0) { + dev_err(&spi->dev, "Error: Firmware size is 0!\n"); +- return; ++ goto out; + } + + /* Fill dummy data (24 stuffing bits for commands) */ +@@ -103,7 +103,7 @@ static void firmware_load(const struct firmware *fw, void *context) + dev_err(&spi->dev, + "Error: No supported FPGA detected (JEDEC_ID=%08x)!\n", + jedec_id); +- return; ++ goto out; + } + + dev_info(&spi->dev, "FPGA %s detected\n", ecp3_dev[i].name); +@@ -116,7 +116,7 @@ static void firmware_load(const struct firmware *fw, void *context) + buffer = kzalloc(fw->size + 8, GFP_KERNEL); + if (!buffer) { + dev_err(&spi->dev, "Error: Can't allocate memory!\n"); +- return; ++ goto out; + } + + /* +@@ -155,7 +155,7 @@ static void firmware_load(const struct firmware *fw, void *context) + "Error: Timeout waiting for FPGA to clear (status=%08x)!\n", + status); + kfree(buffer); +- return; ++ goto out; + } + + dev_info(&spi->dev, "Configuring the FPGA...\n"); +@@ -181,7 +181,7 @@ static void firmware_load(const struct firmware *fw, void *context) + release_firmware(fw); + + kfree(buffer); +- ++out: + complete(&data->fw_loaded); + } + +diff --git a/drivers/misc/lkdtm/Makefile b/drivers/misc/lkdtm/Makefile +index aa12097668d33..e2984ce51fe4d 100644 +--- a/drivers/misc/lkdtm/Makefile ++++ b/drivers/misc/lkdtm/Makefile +@@ -20,7 +20,7 @@ CFLAGS_REMOVE_rodata.o += $(CC_FLAGS_LTO) + + OBJCOPYFLAGS := + OBJCOPYFLAGS_rodata_objcopy.o := \ +- --rename-section .noinstr.text=.rodata,alloc,readonly,load ++ --rename-section .noinstr.text=.rodata,alloc,readonly,load,contents + targets += rodata.o rodata_objcopy.o + $(obj)/rodata_objcopy.o: $(obj)/rodata.o FORCE + $(call if_changed,objcopy) +diff --git a/drivers/misc/mei/hbm.c b/drivers/misc/mei/hbm.c +index be41843df75bc..cebcca6d6d3ef 100644 +--- a/drivers/misc/mei/hbm.c ++++ b/drivers/misc/mei/hbm.c +@@ -672,10 +672,14 @@ static void mei_hbm_cl_dma_map_res(struct mei_device *dev, + if (!cl) + return; + +- dev_dbg(dev->dev, "cl dma map result = %d\n", res->status); +- cl->status = res->status; +- if (!cl->status) ++ if (res->status) { ++ dev_err(dev->dev, "cl dma map failed %d\n", res->status); ++ cl->status = -EFAULT; ++ } else { ++ dev_dbg(dev->dev, "cl dma map succeeded\n"); + cl->dma_mapped = 1; ++ cl->status = 0; ++ } + wake_up(&cl->wait); + } + +@@ -698,10 +702,14 @@ static void mei_hbm_cl_dma_unmap_res(struct mei_device *dev, + if (!cl) + return; + +- dev_dbg(dev->dev, "cl dma unmap result = %d\n", res->status); +- cl->status = res->status; +- if (!cl->status) ++ if (res->status) { ++ dev_err(dev->dev, "cl dma unmap failed %d\n", res->status); ++ cl->status = -EFAULT; ++ } else { ++ dev_dbg(dev->dev, "cl dma unmap succeeded\n"); + cl->dma_mapped = 0; ++ cl->status = 0; ++ } + wake_up(&cl->wait); + } + +diff --git a/drivers/mmc/core/sdio.c b/drivers/mmc/core/sdio.c +index 68edf7a615be5..5447c47157aa5 100644 +--- a/drivers/mmc/core/sdio.c ++++ b/drivers/mmc/core/sdio.c +@@ -708,6 +708,8 @@ try_again: + if (host->ops->init_card) + host->ops->init_card(host, card); + ++ card->ocr = ocr_card; ++ + /* + * If the host and card support UHS-I mode request the card + * to switch to 1.8V signaling level. No 1.8v signalling if +@@ -820,7 +822,7 @@ try_again: + goto mismatch; + } + } +- card->ocr = ocr_card; ++ + mmc_fixup_device(card, sdio_fixup_methods); + + if (card->type == MMC_TYPE_SD_COMBO) { +diff --git a/drivers/mmc/host/meson-mx-sdhc-mmc.c b/drivers/mmc/host/meson-mx-sdhc-mmc.c +index 8fdd0bbbfa21f..28aa78aa08f3f 100644 +--- a/drivers/mmc/host/meson-mx-sdhc-mmc.c ++++ b/drivers/mmc/host/meson-mx-sdhc-mmc.c +@@ -854,6 +854,11 @@ static int meson_mx_sdhc_probe(struct platform_device *pdev) + goto err_disable_pclk; + + irq = platform_get_irq(pdev, 0); ++ if (irq < 0) { ++ ret = irq; ++ goto err_disable_pclk; ++ } ++ + ret = devm_request_threaded_irq(dev, irq, meson_mx_sdhc_irq, + meson_mx_sdhc_irq_thread, IRQF_ONESHOT, + NULL, host); +diff --git a/drivers/mmc/host/meson-mx-sdio.c b/drivers/mmc/host/meson-mx-sdio.c +index d4a48916bfb67..3a19a05ef55a7 100644 +--- a/drivers/mmc/host/meson-mx-sdio.c ++++ b/drivers/mmc/host/meson-mx-sdio.c +@@ -662,6 +662,11 @@ static int meson_mx_mmc_probe(struct platform_device *pdev) + } + + irq = platform_get_irq(pdev, 0); ++ if (irq < 0) { ++ ret = irq; ++ goto error_free_mmc; ++ } ++ + ret = devm_request_threaded_irq(host->controller_dev, irq, + meson_mx_mmc_irq, + meson_mx_mmc_irq_thread, IRQF_ONESHOT, +diff --git a/drivers/mmc/host/mtk-sd.c b/drivers/mmc/host/mtk-sd.c +index 9e6dab7e34242..1ac92015992ed 100644 +--- a/drivers/mmc/host/mtk-sd.c ++++ b/drivers/mmc/host/mtk-sd.c +@@ -628,12 +628,11 @@ static void msdc_reset_hw(struct msdc_host *host) + u32 val; + + sdr_set_bits(host->base + MSDC_CFG, MSDC_CFG_RST); +- while (readl(host->base + MSDC_CFG) & MSDC_CFG_RST) +- cpu_relax(); ++ readl_poll_timeout(host->base + MSDC_CFG, val, !(val & MSDC_CFG_RST), 0, 0); + + sdr_set_bits(host->base + MSDC_FIFOCS, MSDC_FIFOCS_CLR); +- while (readl(host->base + MSDC_FIFOCS) & MSDC_FIFOCS_CLR) +- cpu_relax(); ++ readl_poll_timeout(host->base + MSDC_FIFOCS, val, ++ !(val & MSDC_FIFOCS_CLR), 0, 0); + + val = readl(host->base + MSDC_INT); + writel(val, host->base + MSDC_INT); +@@ -806,8 +805,9 @@ static void msdc_gate_clock(struct msdc_host *host) + clk_disable_unprepare(host->h_clk); + } + +-static void msdc_ungate_clock(struct msdc_host *host) ++static int msdc_ungate_clock(struct msdc_host *host) + { ++ u32 val; + int ret; + + clk_prepare_enable(host->h_clk); +@@ -817,11 +817,11 @@ static void msdc_ungate_clock(struct msdc_host *host) + ret = clk_bulk_prepare_enable(MSDC_NR_CLOCKS, host->bulk_clks); + if (ret) { + dev_err(host->dev, "Cannot enable pclk/axi/ahb clock gates\n"); +- return; ++ return ret; + } + +- while (!(readl(host->base + MSDC_CFG) & MSDC_CFG_CKSTB)) +- cpu_relax(); ++ return readl_poll_timeout(host->base + MSDC_CFG, val, ++ (val & MSDC_CFG_CKSTB), 1, 20000); + } + + static void msdc_set_mclk(struct msdc_host *host, unsigned char timing, u32 hz) +@@ -832,6 +832,7 @@ static void msdc_set_mclk(struct msdc_host *host, unsigned char timing, u32 hz) + u32 div; + u32 sclk; + u32 tune_reg = host->dev_comp->pad_tune_reg; ++ u32 val; + + if (!hz) { + dev_dbg(host->dev, "set mclk to 0\n"); +@@ -912,8 +913,7 @@ static void msdc_set_mclk(struct msdc_host *host, unsigned char timing, u32 hz) + else + clk_prepare_enable(clk_get_parent(host->src_clk)); + +- while (!(readl(host->base + MSDC_CFG) & MSDC_CFG_CKSTB)) +- cpu_relax(); ++ readl_poll_timeout(host->base + MSDC_CFG, val, (val & MSDC_CFG_CKSTB), 0, 0); + sdr_set_bits(host->base + MSDC_CFG, MSDC_CFG_CKPDN); + mmc->actual_clock = sclk; + host->mclk = hz; +@@ -1223,13 +1223,13 @@ static bool msdc_cmd_done(struct msdc_host *host, int events, + static inline bool msdc_cmd_is_ready(struct msdc_host *host, + struct mmc_request *mrq, struct mmc_command *cmd) + { +- /* The max busy time we can endure is 20ms */ +- unsigned long tmo = jiffies + msecs_to_jiffies(20); ++ u32 val; ++ int ret; + +- while ((readl(host->base + SDC_STS) & SDC_STS_CMDBUSY) && +- time_before(jiffies, tmo)) +- cpu_relax(); +- if (readl(host->base + SDC_STS) & SDC_STS_CMDBUSY) { ++ /* The max busy time we can endure is 20ms */ ++ ret = readl_poll_timeout_atomic(host->base + SDC_STS, val, ++ !(val & SDC_STS_CMDBUSY), 1, 20000); ++ if (ret) { + dev_err(host->dev, "CMD bus busy detected\n"); + host->error |= REQ_CMD_BUSY; + msdc_cmd_done(host, MSDC_INT_CMDTMO, mrq, cmd); +@@ -1237,12 +1237,10 @@ static inline bool msdc_cmd_is_ready(struct msdc_host *host, + } + + if (mmc_resp_type(cmd) == MMC_RSP_R1B || cmd->data) { +- tmo = jiffies + msecs_to_jiffies(20); + /* R1B or with data, should check SDCBUSY */ +- while ((readl(host->base + SDC_STS) & SDC_STS_SDCBUSY) && +- time_before(jiffies, tmo)) +- cpu_relax(); +- if (readl(host->base + SDC_STS) & SDC_STS_SDCBUSY) { ++ ret = readl_poll_timeout_atomic(host->base + SDC_STS, val, ++ !(val & SDC_STS_SDCBUSY), 1, 20000); ++ if (ret) { + dev_err(host->dev, "Controller busy detected\n"); + host->error |= REQ_CMD_BUSY; + msdc_cmd_done(host, MSDC_INT_CMDTMO, mrq, cmd); +@@ -1367,6 +1365,8 @@ static bool msdc_data_xfer_done(struct msdc_host *host, u32 events, + (MSDC_INT_XFER_COMPL | MSDC_INT_DATCRCERR | MSDC_INT_DATTMO + | MSDC_INT_DMA_BDCSERR | MSDC_INT_DMA_GPDCSERR + | MSDC_INT_DMA_PROTECT); ++ u32 val; ++ int ret; + + spin_lock_irqsave(&host->lock, flags); + done = !host->data; +@@ -1383,8 +1383,14 @@ static bool msdc_data_xfer_done(struct msdc_host *host, u32 events, + readl(host->base + MSDC_DMA_CFG)); + sdr_set_field(host->base + MSDC_DMA_CTRL, MSDC_DMA_CTRL_STOP, + 1); +- while (readl(host->base + MSDC_DMA_CFG) & MSDC_DMA_CFG_STS) +- cpu_relax(); ++ ++ ret = readl_poll_timeout_atomic(host->base + MSDC_DMA_CFG, val, ++ !(val & MSDC_DMA_CFG_STS), 1, 20000); ++ if (ret) { ++ dev_dbg(host->dev, "DMA stop timed out\n"); ++ return false; ++ } ++ + sdr_clr_bits(host->base + MSDC_INTEN, data_ints_mask); + dev_dbg(host->dev, "DMA stop\n"); + +@@ -2598,7 +2604,11 @@ static int msdc_drv_probe(struct platform_device *pdev) + spin_lock_init(&host->lock); + + platform_set_drvdata(pdev, mmc); +- msdc_ungate_clock(host); ++ ret = msdc_ungate_clock(host); ++ if (ret) { ++ dev_err(&pdev->dev, "Cannot ungate clocks!\n"); ++ goto release_mem; ++ } + msdc_init_hw(host); + + if (mmc->caps2 & MMC_CAP2_CQE) { +@@ -2757,8 +2767,12 @@ static int __maybe_unused msdc_runtime_resume(struct device *dev) + { + struct mmc_host *mmc = dev_get_drvdata(dev); + struct msdc_host *host = mmc_priv(mmc); ++ int ret; ++ ++ ret = msdc_ungate_clock(host); ++ if (ret) ++ return ret; + +- msdc_ungate_clock(host); + msdc_restore_reg(host); + return 0; + } +diff --git a/drivers/mmc/host/sdhci-pci-gli.c b/drivers/mmc/host/sdhci-pci-gli.c +index 4fd99c1e82ba3..ad50f16658fe2 100644 +--- a/drivers/mmc/host/sdhci-pci-gli.c ++++ b/drivers/mmc/host/sdhci-pci-gli.c +@@ -12,6 +12,7 @@ + #include <linux/pci.h> + #include <linux/mmc/mmc.h> + #include <linux/delay.h> ++#include <linux/of.h> + #include "sdhci.h" + #include "sdhci-pci.h" + #include "cqhci.h" +@@ -116,6 +117,8 @@ + #define PCI_GLI_9755_PECONF 0x44 + #define PCI_GLI_9755_LFCLK GENMASK(14, 12) + #define PCI_GLI_9755_DMACLK BIT(29) ++#define PCI_GLI_9755_INVERT_CD BIT(30) ++#define PCI_GLI_9755_INVERT_WP BIT(31) + + #define PCI_GLI_9755_CFG2 0x48 + #define PCI_GLI_9755_CFG2_L1DLY GENMASK(28, 24) +@@ -570,6 +573,14 @@ static void gl9755_hw_setting(struct sdhci_pci_slot *slot) + gl9755_wt_on(pdev); + + pci_read_config_dword(pdev, PCI_GLI_9755_PECONF, &value); ++ /* ++ * Apple ARM64 platforms using these chips may have ++ * inverted CD/WP detection. ++ */ ++ if (of_property_read_bool(pdev->dev.of_node, "cd-inverted")) ++ value |= PCI_GLI_9755_INVERT_CD; ++ if (of_property_read_bool(pdev->dev.of_node, "wp-inverted")) ++ value |= PCI_GLI_9755_INVERT_WP; + value &= ~PCI_GLI_9755_LFCLK; + value &= ~PCI_GLI_9755_DMACLK; + pci_write_config_dword(pdev, PCI_GLI_9755_PECONF, value); +diff --git a/drivers/mmc/host/tmio_mmc_core.c b/drivers/mmc/host/tmio_mmc_core.c +index e2affa52ef469..a5850d83908be 100644 +--- a/drivers/mmc/host/tmio_mmc_core.c ++++ b/drivers/mmc/host/tmio_mmc_core.c +@@ -960,14 +960,8 @@ static void tmio_mmc_set_ios(struct mmc_host *mmc, struct mmc_ios *ios) + case MMC_POWER_OFF: + tmio_mmc_power_off(host); + /* For R-Car Gen2+, we need to reset SDHI specific SCC */ +- if (host->pdata->flags & TMIO_MMC_MIN_RCAR2) { +- host->reset(host); +- +- if (host->native_hotplug) +- tmio_mmc_enable_mmc_irqs(host, +- TMIO_STAT_CARD_REMOVE | +- TMIO_STAT_CARD_INSERT); +- } ++ if (host->pdata->flags & TMIO_MMC_MIN_RCAR2) ++ tmio_mmc_reset(host); + + host->set_clock(host, 0); + break; +@@ -1175,6 +1169,7 @@ int tmio_mmc_host_probe(struct tmio_mmc_host *_host) + if (mmc_can_gpio_cd(mmc)) + _host->ops.get_cd = mmc_gpio_get_cd; + ++ /* must be set before tmio_mmc_reset() */ + _host->native_hotplug = !(mmc_can_gpio_cd(mmc) || + mmc->caps & MMC_CAP_NEEDS_POLL || + !mmc_card_is_removable(mmc)); +@@ -1295,10 +1290,6 @@ int tmio_mmc_host_runtime_resume(struct device *dev) + if (host->clk_cache) + host->set_clock(host, host->clk_cache); + +- if (host->native_hotplug) +- tmio_mmc_enable_mmc_irqs(host, +- TMIO_STAT_CARD_REMOVE | TMIO_STAT_CARD_INSERT); +- + tmio_mmc_enable_dma(host, true); + + return 0; +diff --git a/drivers/mtd/hyperbus/rpc-if.c b/drivers/mtd/hyperbus/rpc-if.c +index ecb050ba95cdf..dc164c18f8429 100644 +--- a/drivers/mtd/hyperbus/rpc-if.c ++++ b/drivers/mtd/hyperbus/rpc-if.c +@@ -124,7 +124,9 @@ static int rpcif_hb_probe(struct platform_device *pdev) + if (!hyperbus) + return -ENOMEM; + +- rpcif_sw_init(&hyperbus->rpc, pdev->dev.parent); ++ error = rpcif_sw_init(&hyperbus->rpc, pdev->dev.parent); ++ if (error) ++ return error; + + platform_set_drvdata(pdev, hyperbus); + +@@ -150,9 +152,9 @@ static int rpcif_hb_remove(struct platform_device *pdev) + { + struct rpcif_hyperbus *hyperbus = platform_get_drvdata(pdev); + int error = hyperbus_unregister_device(&hyperbus->hbdev); +- struct rpcif *rpc = dev_get_drvdata(pdev->dev.parent); + +- rpcif_disable_rpm(rpc); ++ rpcif_disable_rpm(&hyperbus->rpc); ++ + return error; + } + +diff --git a/drivers/mtd/mtdcore.c b/drivers/mtd/mtdcore.c +index 1532291989471..54df9cfd588ea 100644 +--- a/drivers/mtd/mtdcore.c ++++ b/drivers/mtd/mtdcore.c +@@ -825,8 +825,7 @@ static struct nvmem_device *mtd_otp_nvmem_register(struct mtd_info *mtd, + + /* OTP nvmem will be registered on the physical device */ + config.dev = mtd->dev.parent; +- /* just reuse the compatible as name */ +- config.name = compatible; ++ config.name = kasprintf(GFP_KERNEL, "%s-%s", dev_name(&mtd->dev), compatible); + config.id = NVMEM_DEVID_NONE; + config.owner = THIS_MODULE; + config.type = NVMEM_TYPE_OTP; +@@ -842,6 +841,7 @@ static struct nvmem_device *mtd_otp_nvmem_register(struct mtd_info *mtd, + nvmem = NULL; + + of_node_put(np); ++ kfree(config.name); + + return nvmem; + } +diff --git a/drivers/mtd/mtdpart.c b/drivers/mtd/mtdpart.c +index 04af12b66110c..357661b62c94d 100644 +--- a/drivers/mtd/mtdpart.c ++++ b/drivers/mtd/mtdpart.c +@@ -312,7 +312,7 @@ static int __mtd_del_partition(struct mtd_info *mtd) + if (err) + return err; + +- list_del(&child->part.node); ++ list_del(&mtd->part.node); + free_partition(mtd); + + return 0; +diff --git a/drivers/mtd/nand/raw/davinci_nand.c b/drivers/mtd/nand/raw/davinci_nand.c +index 118da9944e3bc..45fec8c192aba 100644 +--- a/drivers/mtd/nand/raw/davinci_nand.c ++++ b/drivers/mtd/nand/raw/davinci_nand.c +@@ -371,77 +371,6 @@ correct: + return corrected; + } + +-/** +- * nand_read_page_hwecc_oob_first - hw ecc, read oob first +- * @chip: nand chip info structure +- * @buf: buffer to store read data +- * @oob_required: caller requires OOB data read to chip->oob_poi +- * @page: page number to read +- * +- * Hardware ECC for large page chips, require OOB to be read first. For this +- * ECC mode, the write_page method is re-used from ECC_HW. These methods +- * read/write ECC from the OOB area, unlike the ECC_HW_SYNDROME support with +- * multiple ECC steps, follows the "infix ECC" scheme and reads/writes ECC from +- * the data area, by overwriting the NAND manufacturer bad block markings. +- */ +-static int nand_davinci_read_page_hwecc_oob_first(struct nand_chip *chip, +- uint8_t *buf, +- int oob_required, int page) +-{ +- struct mtd_info *mtd = nand_to_mtd(chip); +- int i, eccsize = chip->ecc.size, ret; +- int eccbytes = chip->ecc.bytes; +- int eccsteps = chip->ecc.steps; +- uint8_t *p = buf; +- uint8_t *ecc_code = chip->ecc.code_buf; +- uint8_t *ecc_calc = chip->ecc.calc_buf; +- unsigned int max_bitflips = 0; +- +- /* Read the OOB area first */ +- ret = nand_read_oob_op(chip, page, 0, chip->oob_poi, mtd->oobsize); +- if (ret) +- return ret; +- +- ret = nand_read_page_op(chip, page, 0, NULL, 0); +- if (ret) +- return ret; +- +- ret = mtd_ooblayout_get_eccbytes(mtd, ecc_code, chip->oob_poi, 0, +- chip->ecc.total); +- if (ret) +- return ret; +- +- for (i = 0; eccsteps; eccsteps--, i += eccbytes, p += eccsize) { +- int stat; +- +- chip->ecc.hwctl(chip, NAND_ECC_READ); +- +- ret = nand_read_data_op(chip, p, eccsize, false, false); +- if (ret) +- return ret; +- +- chip->ecc.calculate(chip, p, &ecc_calc[i]); +- +- stat = chip->ecc.correct(chip, p, &ecc_code[i], NULL); +- if (stat == -EBADMSG && +- (chip->ecc.options & NAND_ECC_GENERIC_ERASED_CHECK)) { +- /* check for empty pages with bitflips */ +- stat = nand_check_erased_ecc_chunk(p, eccsize, +- &ecc_code[i], +- eccbytes, NULL, 0, +- chip->ecc.strength); +- } +- +- if (stat < 0) { +- mtd->ecc_stats.failed++; +- } else { +- mtd->ecc_stats.corrected += stat; +- max_bitflips = max_t(unsigned int, max_bitflips, stat); +- } +- } +- return max_bitflips; +-} +- + /*----------------------------------------------------------------------*/ + + /* An ECC layout for using 4-bit ECC with small-page flash, storing +@@ -651,7 +580,7 @@ static int davinci_nand_attach_chip(struct nand_chip *chip) + } else if (chunks == 4 || chunks == 8) { + mtd_set_ooblayout(mtd, + nand_get_large_page_ooblayout()); +- chip->ecc.read_page = nand_davinci_read_page_hwecc_oob_first; ++ chip->ecc.read_page = nand_read_page_hwecc_oob_first; + } else { + return -EIO; + } +diff --git a/drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c b/drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c +index 4d08e4ab5c1b6..6e9f7d80ef8b8 100644 +--- a/drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c ++++ b/drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c +@@ -713,14 +713,32 @@ static void gpmi_nfc_compute_timings(struct gpmi_nand_data *this, + (use_half_period ? BM_GPMI_CTRL1_HALF_PERIOD : 0); + } + +-static void gpmi_nfc_apply_timings(struct gpmi_nand_data *this) ++static int gpmi_nfc_apply_timings(struct gpmi_nand_data *this) + { + struct gpmi_nfc_hardware_timing *hw = &this->hw; + struct resources *r = &this->resources; + void __iomem *gpmi_regs = r->gpmi_regs; + unsigned int dll_wait_time_us; ++ int ret; ++ ++ /* Clock dividers do NOT guarantee a clean clock signal on its output ++ * during the change of the divide factor on i.MX6Q/UL/SX. On i.MX7/8, ++ * all clock dividers provide these guarantee. ++ */ ++ if (GPMI_IS_MX6Q(this) || GPMI_IS_MX6SX(this)) ++ clk_disable_unprepare(r->clock[0]); + +- clk_set_rate(r->clock[0], hw->clk_rate); ++ ret = clk_set_rate(r->clock[0], hw->clk_rate); ++ if (ret) { ++ dev_err(this->dev, "cannot set clock rate to %lu Hz: %d\n", hw->clk_rate, ret); ++ return ret; ++ } ++ ++ if (GPMI_IS_MX6Q(this) || GPMI_IS_MX6SX(this)) { ++ ret = clk_prepare_enable(r->clock[0]); ++ if (ret) ++ return ret; ++ } + + writel(hw->timing0, gpmi_regs + HW_GPMI_TIMING0); + writel(hw->timing1, gpmi_regs + HW_GPMI_TIMING1); +@@ -739,6 +757,8 @@ static void gpmi_nfc_apply_timings(struct gpmi_nand_data *this) + + /* Wait for the DLL to settle. */ + udelay(dll_wait_time_us); ++ ++ return 0; + } + + static int gpmi_setup_interface(struct nand_chip *chip, int chipnr, +@@ -1034,15 +1054,6 @@ static int gpmi_get_clks(struct gpmi_nand_data *this) + r->clock[i] = clk; + } + +- if (GPMI_IS_MX6(this)) +- /* +- * Set the default value for the gpmi clock. +- * +- * If you want to use the ONFI nand which is in the +- * Synchronous Mode, you should change the clock as you need. +- */ +- clk_set_rate(r->clock[0], 22000000); +- + return 0; + + err_clock: +@@ -2280,7 +2291,9 @@ static int gpmi_nfc_exec_op(struct nand_chip *chip, + */ + if (this->hw.must_apply_timings) { + this->hw.must_apply_timings = false; +- gpmi_nfc_apply_timings(this); ++ ret = gpmi_nfc_apply_timings(this); ++ if (ret) ++ return ret; + } + + dev_dbg(this->dev, "%s: %d instructions\n", __func__, op->ninstrs); +diff --git a/drivers/mtd/nand/raw/ingenic/ingenic_nand_drv.c b/drivers/mtd/nand/raw/ingenic/ingenic_nand_drv.c +index 0e9d426fe4f2b..b18861bdcdc88 100644 +--- a/drivers/mtd/nand/raw/ingenic/ingenic_nand_drv.c ++++ b/drivers/mtd/nand/raw/ingenic/ingenic_nand_drv.c +@@ -32,6 +32,7 @@ struct jz_soc_info { + unsigned long addr_offset; + unsigned long cmd_offset; + const struct mtd_ooblayout_ops *oob_layout; ++ bool oob_first; + }; + + struct ingenic_nand_cs { +@@ -240,6 +241,9 @@ static int ingenic_nand_attach_chip(struct nand_chip *chip) + if (chip->bbt_options & NAND_BBT_USE_FLASH) + chip->bbt_options |= NAND_BBT_NO_OOB; + ++ if (nfc->soc_info->oob_first) ++ chip->ecc.read_page = nand_read_page_hwecc_oob_first; ++ + /* For legacy reasons we use a different layout on the qi,lb60 board. */ + if (of_machine_is_compatible("qi,lb60")) + mtd_set_ooblayout(mtd, &qi_lb60_ooblayout_ops); +@@ -534,6 +538,7 @@ static const struct jz_soc_info jz4740_soc_info = { + .data_offset = 0x00000000, + .cmd_offset = 0x00008000, + .addr_offset = 0x00010000, ++ .oob_first = true, + }; + + static const struct jz_soc_info jz4725b_soc_info = { +diff --git a/drivers/mtd/nand/raw/nand_base.c b/drivers/mtd/nand/raw/nand_base.c +index a130320de4128..d5a2110eb38ed 100644 +--- a/drivers/mtd/nand/raw/nand_base.c ++++ b/drivers/mtd/nand/raw/nand_base.c +@@ -3160,6 +3160,73 @@ static int nand_read_page_hwecc(struct nand_chip *chip, uint8_t *buf, + return max_bitflips; + } + ++/** ++ * nand_read_page_hwecc_oob_first - Hardware ECC page read with ECC ++ * data read from OOB area ++ * @chip: nand chip info structure ++ * @buf: buffer to store read data ++ * @oob_required: caller requires OOB data read to chip->oob_poi ++ * @page: page number to read ++ * ++ * Hardware ECC for large page chips, which requires the ECC data to be ++ * extracted from the OOB before the actual data is read. ++ */ ++int nand_read_page_hwecc_oob_first(struct nand_chip *chip, uint8_t *buf, ++ int oob_required, int page) ++{ ++ struct mtd_info *mtd = nand_to_mtd(chip); ++ int i, eccsize = chip->ecc.size, ret; ++ int eccbytes = chip->ecc.bytes; ++ int eccsteps = chip->ecc.steps; ++ uint8_t *p = buf; ++ uint8_t *ecc_code = chip->ecc.code_buf; ++ unsigned int max_bitflips = 0; ++ ++ /* Read the OOB area first */ ++ ret = nand_read_oob_op(chip, page, 0, chip->oob_poi, mtd->oobsize); ++ if (ret) ++ return ret; ++ ++ /* Move read cursor to start of page */ ++ ret = nand_change_read_column_op(chip, 0, NULL, 0, false); ++ if (ret) ++ return ret; ++ ++ ret = mtd_ooblayout_get_eccbytes(mtd, ecc_code, chip->oob_poi, 0, ++ chip->ecc.total); ++ if (ret) ++ return ret; ++ ++ for (i = 0; eccsteps; eccsteps--, i += eccbytes, p += eccsize) { ++ int stat; ++ ++ chip->ecc.hwctl(chip, NAND_ECC_READ); ++ ++ ret = nand_read_data_op(chip, p, eccsize, false, false); ++ if (ret) ++ return ret; ++ ++ stat = chip->ecc.correct(chip, p, &ecc_code[i], NULL); ++ if (stat == -EBADMSG && ++ (chip->ecc.options & NAND_ECC_GENERIC_ERASED_CHECK)) { ++ /* check for empty pages with bitflips */ ++ stat = nand_check_erased_ecc_chunk(p, eccsize, ++ &ecc_code[i], ++ eccbytes, NULL, 0, ++ chip->ecc.strength); ++ } ++ ++ if (stat < 0) { ++ mtd->ecc_stats.failed++; ++ } else { ++ mtd->ecc_stats.corrected += stat; ++ max_bitflips = max_t(unsigned int, max_bitflips, stat); ++ } ++ } ++ return max_bitflips; ++} ++EXPORT_SYMBOL_GPL(nand_read_page_hwecc_oob_first); ++ + /** + * nand_read_page_syndrome - [REPLACEABLE] hardware ECC syndrome based page read + * @chip: nand chip info structure +diff --git a/drivers/net/bonding/bond_main.c b/drivers/net/bonding/bond_main.c +index 77dc79a7f5748..83cdaabd7b69d 100644 +--- a/drivers/net/bonding/bond_main.c ++++ b/drivers/net/bonding/bond_main.c +@@ -1096,9 +1096,6 @@ static bool bond_should_notify_peers(struct bonding *bond) + slave = rcu_dereference(bond->curr_active_slave); + rcu_read_unlock(); + +- netdev_dbg(bond->dev, "bond_should_notify_peers: slave %s\n", +- slave ? slave->dev->name : "NULL"); +- + if (!slave || !bond->send_peer_notif || + bond->send_peer_notif % + max(1, bond->params.peer_notif_delay) != 0 || +@@ -1106,6 +1103,9 @@ static bool bond_should_notify_peers(struct bonding *bond) + test_bit(__LINK_STATE_LINKWATCH_PENDING, &slave->dev->state)) + return false; + ++ netdev_dbg(bond->dev, "bond_should_notify_peers: slave %s\n", ++ slave ? slave->dev->name : "NULL"); ++ + return true; + } + +@@ -3872,8 +3872,8 @@ u32 bond_xmit_hash(struct bonding *bond, struct sk_buff *skb) + skb->l4_hash) + return skb->hash; + +- return __bond_xmit_hash(bond, skb, skb->head, skb->protocol, +- skb->mac_header, skb->network_header, ++ return __bond_xmit_hash(bond, skb, skb->data, skb->protocol, ++ skb_mac_offset(skb), skb_network_offset(skb), + skb_headlen(skb)); + } + +@@ -4843,25 +4843,39 @@ static netdev_tx_t bond_xmit_broadcast(struct sk_buff *skb, + struct bonding *bond = netdev_priv(bond_dev); + struct slave *slave = NULL; + struct list_head *iter; ++ bool xmit_suc = false; ++ bool skb_used = false; + + bond_for_each_slave_rcu(bond, slave, iter) { +- if (bond_is_last_slave(bond, slave)) +- break; +- if (bond_slave_is_up(slave) && slave->link == BOND_LINK_UP) { +- struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC); ++ struct sk_buff *skb2; ++ ++ if (!(bond_slave_is_up(slave) && slave->link == BOND_LINK_UP)) ++ continue; + ++ if (bond_is_last_slave(bond, slave)) { ++ skb2 = skb; ++ skb_used = true; ++ } else { ++ skb2 = skb_clone(skb, GFP_ATOMIC); + if (!skb2) { + net_err_ratelimited("%s: Error: %s: skb_clone() failed\n", + bond_dev->name, __func__); + continue; + } +- bond_dev_queue_xmit(bond, skb2, slave->dev); + } ++ ++ if (bond_dev_queue_xmit(bond, skb2, slave->dev) == NETDEV_TX_OK) ++ xmit_suc = true; + } +- if (slave && bond_slave_is_up(slave) && slave->link == BOND_LINK_UP) +- return bond_dev_queue_xmit(bond, skb, slave->dev); + +- return bond_tx_drop(bond_dev, skb); ++ if (!skb_used) ++ dev_kfree_skb_any(skb); ++ ++ if (xmit_suc) ++ return NETDEV_TX_OK; ++ ++ atomic_long_inc(&bond_dev->tx_dropped); ++ return NET_XMIT_DROP; + } + + /*------------------------- Device initialization ---------------------------*/ +diff --git a/drivers/net/can/flexcan.c b/drivers/net/can/flexcan.c +index 7734229aa0788..18d7bb99ec1bd 100644 +--- a/drivers/net/can/flexcan.c ++++ b/drivers/net/can/flexcan.c +@@ -173,9 +173,9 @@ + + /* FLEXCAN interrupt flag register (IFLAG) bits */ + /* Errata ERR005829 step7: Reserve first valid MB */ +-#define FLEXCAN_TX_MB_RESERVED_OFF_FIFO 8 +-#define FLEXCAN_TX_MB_RESERVED_OFF_TIMESTAMP 0 +-#define FLEXCAN_RX_MB_OFF_TIMESTAMP_FIRST (FLEXCAN_TX_MB_RESERVED_OFF_TIMESTAMP + 1) ++#define FLEXCAN_TX_MB_RESERVED_RX_FIFO 8 ++#define FLEXCAN_TX_MB_RESERVED_RX_MAILBOX 0 ++#define FLEXCAN_RX_MB_RX_MAILBOX_FIRST (FLEXCAN_TX_MB_RESERVED_RX_MAILBOX + 1) + #define FLEXCAN_IFLAG_MB(x) BIT_ULL(x) + #define FLEXCAN_IFLAG_RX_FIFO_OVERFLOW BIT(7) + #define FLEXCAN_IFLAG_RX_FIFO_WARN BIT(6) +@@ -234,8 +234,8 @@ + #define FLEXCAN_QUIRK_ENABLE_EACEN_RRS BIT(3) + /* Disable non-correctable errors interrupt and freeze mode */ + #define FLEXCAN_QUIRK_DISABLE_MECR BIT(4) +-/* Use timestamp based offloading */ +-#define FLEXCAN_QUIRK_USE_OFF_TIMESTAMP BIT(5) ++/* Use mailboxes (not FIFO) for RX path */ ++#define FLEXCAN_QUIRK_USE_RX_MAILBOX BIT(5) + /* No interrupt for error passive */ + #define FLEXCAN_QUIRK_BROKEN_PERR_STATE BIT(6) + /* default to BE register access */ +@@ -252,6 +252,12 @@ + #define FLEXCAN_QUIRK_NR_IRQ_3 BIT(12) + /* Setup 16 mailboxes */ + #define FLEXCAN_QUIRK_NR_MB_16 BIT(13) ++/* Device supports RX via mailboxes */ ++#define FLEXCAN_QUIRK_SUPPPORT_RX_MAILBOX BIT(14) ++/* Device supports RTR reception via mailboxes */ ++#define FLEXCAN_QUIRK_SUPPPORT_RX_MAILBOX_RTR BIT(15) ++/* Device supports RX via FIFO */ ++#define FLEXCAN_QUIRK_SUPPPORT_RX_FIFO BIT(16) + + /* Structure of the message buffer */ + struct flexcan_mb { +@@ -365,7 +371,7 @@ struct flexcan_priv { + + struct clk *clk_ipg; + struct clk *clk_per; +- const struct flexcan_devtype_data *devtype_data; ++ struct flexcan_devtype_data devtype_data; + struct regulator *reg_xceiver; + struct flexcan_stop_mode stm; + +@@ -382,59 +388,78 @@ struct flexcan_priv { + + static const struct flexcan_devtype_data fsl_mcf5441x_devtype_data = { + .quirks = FLEXCAN_QUIRK_BROKEN_PERR_STATE | +- FLEXCAN_QUIRK_NR_IRQ_3 | FLEXCAN_QUIRK_NR_MB_16, ++ FLEXCAN_QUIRK_NR_IRQ_3 | FLEXCAN_QUIRK_NR_MB_16 | ++ FLEXCAN_QUIRK_SUPPPORT_RX_FIFO, + }; + + static const struct flexcan_devtype_data fsl_p1010_devtype_data = { + .quirks = FLEXCAN_QUIRK_BROKEN_WERR_STATE | + FLEXCAN_QUIRK_BROKEN_PERR_STATE | +- FLEXCAN_QUIRK_DEFAULT_BIG_ENDIAN, ++ FLEXCAN_QUIRK_DEFAULT_BIG_ENDIAN | ++ FLEXCAN_QUIRK_SUPPPORT_RX_MAILBOX | ++ FLEXCAN_QUIRK_SUPPPORT_RX_FIFO, + }; + + static const struct flexcan_devtype_data fsl_imx25_devtype_data = { + .quirks = FLEXCAN_QUIRK_BROKEN_WERR_STATE | +- FLEXCAN_QUIRK_BROKEN_PERR_STATE, ++ FLEXCAN_QUIRK_BROKEN_PERR_STATE | ++ FLEXCAN_QUIRK_SUPPPORT_RX_MAILBOX | ++ FLEXCAN_QUIRK_SUPPPORT_RX_FIFO, + }; + + static const struct flexcan_devtype_data fsl_imx28_devtype_data = { +- .quirks = FLEXCAN_QUIRK_BROKEN_PERR_STATE, ++ .quirks = FLEXCAN_QUIRK_BROKEN_PERR_STATE | ++ FLEXCAN_QUIRK_SUPPPORT_RX_MAILBOX | ++ FLEXCAN_QUIRK_SUPPPORT_RX_FIFO, + }; + + static const struct flexcan_devtype_data fsl_imx6q_devtype_data = { + .quirks = FLEXCAN_QUIRK_DISABLE_RXFG | FLEXCAN_QUIRK_ENABLE_EACEN_RRS | +- FLEXCAN_QUIRK_USE_OFF_TIMESTAMP | FLEXCAN_QUIRK_BROKEN_PERR_STATE | +- FLEXCAN_QUIRK_SETUP_STOP_MODE_GPR, ++ FLEXCAN_QUIRK_USE_RX_MAILBOX | FLEXCAN_QUIRK_BROKEN_PERR_STATE | ++ FLEXCAN_QUIRK_SETUP_STOP_MODE_GPR | ++ FLEXCAN_QUIRK_SUPPPORT_RX_MAILBOX | ++ FLEXCAN_QUIRK_SUPPPORT_RX_MAILBOX_RTR, + }; + + static const struct flexcan_devtype_data fsl_imx8qm_devtype_data = { + .quirks = FLEXCAN_QUIRK_DISABLE_RXFG | FLEXCAN_QUIRK_ENABLE_EACEN_RRS | +- FLEXCAN_QUIRK_USE_OFF_TIMESTAMP | FLEXCAN_QUIRK_BROKEN_PERR_STATE | +- FLEXCAN_QUIRK_SUPPORT_FD | FLEXCAN_QUIRK_SETUP_STOP_MODE_SCFW, ++ FLEXCAN_QUIRK_USE_RX_MAILBOX | FLEXCAN_QUIRK_BROKEN_PERR_STATE | ++ FLEXCAN_QUIRK_SUPPORT_FD | FLEXCAN_QUIRK_SETUP_STOP_MODE_SCFW | ++ FLEXCAN_QUIRK_SUPPPORT_RX_MAILBOX | ++ FLEXCAN_QUIRK_SUPPPORT_RX_MAILBOX_RTR, + }; + + static struct flexcan_devtype_data fsl_imx8mp_devtype_data = { + .quirks = FLEXCAN_QUIRK_DISABLE_RXFG | FLEXCAN_QUIRK_ENABLE_EACEN_RRS | +- FLEXCAN_QUIRK_DISABLE_MECR | FLEXCAN_QUIRK_USE_OFF_TIMESTAMP | ++ FLEXCAN_QUIRK_DISABLE_MECR | FLEXCAN_QUIRK_USE_RX_MAILBOX | + FLEXCAN_QUIRK_BROKEN_PERR_STATE | FLEXCAN_QUIRK_SETUP_STOP_MODE_GPR | +- FLEXCAN_QUIRK_SUPPORT_FD | FLEXCAN_QUIRK_SUPPORT_ECC, ++ FLEXCAN_QUIRK_SUPPORT_FD | FLEXCAN_QUIRK_SUPPORT_ECC | ++ FLEXCAN_QUIRK_SUPPPORT_RX_MAILBOX | ++ FLEXCAN_QUIRK_SUPPPORT_RX_MAILBOX_RTR, + }; + + static const struct flexcan_devtype_data fsl_vf610_devtype_data = { + .quirks = FLEXCAN_QUIRK_DISABLE_RXFG | FLEXCAN_QUIRK_ENABLE_EACEN_RRS | +- FLEXCAN_QUIRK_DISABLE_MECR | FLEXCAN_QUIRK_USE_OFF_TIMESTAMP | +- FLEXCAN_QUIRK_BROKEN_PERR_STATE | FLEXCAN_QUIRK_SUPPORT_ECC, ++ FLEXCAN_QUIRK_DISABLE_MECR | FLEXCAN_QUIRK_USE_RX_MAILBOX | ++ FLEXCAN_QUIRK_BROKEN_PERR_STATE | FLEXCAN_QUIRK_SUPPORT_ECC | ++ FLEXCAN_QUIRK_SUPPPORT_RX_MAILBOX | ++ FLEXCAN_QUIRK_SUPPPORT_RX_MAILBOX_RTR, + }; + + static const struct flexcan_devtype_data fsl_ls1021a_r2_devtype_data = { + .quirks = FLEXCAN_QUIRK_DISABLE_RXFG | FLEXCAN_QUIRK_ENABLE_EACEN_RRS | +- FLEXCAN_QUIRK_BROKEN_PERR_STATE | FLEXCAN_QUIRK_USE_OFF_TIMESTAMP, ++ FLEXCAN_QUIRK_BROKEN_PERR_STATE | FLEXCAN_QUIRK_USE_RX_MAILBOX | ++ FLEXCAN_QUIRK_SUPPPORT_RX_MAILBOX | ++ FLEXCAN_QUIRK_SUPPPORT_RX_MAILBOX_RTR, + }; + + static const struct flexcan_devtype_data fsl_lx2160a_r1_devtype_data = { + .quirks = FLEXCAN_QUIRK_DISABLE_RXFG | FLEXCAN_QUIRK_ENABLE_EACEN_RRS | + FLEXCAN_QUIRK_DISABLE_MECR | FLEXCAN_QUIRK_BROKEN_PERR_STATE | +- FLEXCAN_QUIRK_USE_OFF_TIMESTAMP | FLEXCAN_QUIRK_SUPPORT_FD | +- FLEXCAN_QUIRK_SUPPORT_ECC, ++ FLEXCAN_QUIRK_USE_RX_MAILBOX | FLEXCAN_QUIRK_SUPPORT_FD | ++ FLEXCAN_QUIRK_SUPPORT_ECC | ++ FLEXCAN_QUIRK_SUPPPORT_RX_MAILBOX | ++ FLEXCAN_QUIRK_SUPPPORT_RX_MAILBOX_RTR, + }; + + static const struct can_bittiming_const flexcan_bittiming_const = { +@@ -596,7 +621,7 @@ static inline int flexcan_enter_stop_mode(struct flexcan_priv *priv) + priv->write(reg_mcr, ®s->mcr); + + /* enable stop request */ +- if (priv->devtype_data->quirks & FLEXCAN_QUIRK_SETUP_STOP_MODE_SCFW) { ++ if (priv->devtype_data.quirks & FLEXCAN_QUIRK_SETUP_STOP_MODE_SCFW) { + ret = flexcan_stop_mode_enable_scfw(priv, true); + if (ret < 0) + return ret; +@@ -615,7 +640,7 @@ static inline int flexcan_exit_stop_mode(struct flexcan_priv *priv) + int ret; + + /* remove stop request */ +- if (priv->devtype_data->quirks & FLEXCAN_QUIRK_SETUP_STOP_MODE_SCFW) { ++ if (priv->devtype_data.quirks & FLEXCAN_QUIRK_SETUP_STOP_MODE_SCFW) { + ret = flexcan_stop_mode_enable_scfw(priv, false); + if (ret < 0) + return ret; +@@ -1018,7 +1043,7 @@ static struct sk_buff *flexcan_mailbox_read(struct can_rx_offload *offload, + + mb = flexcan_get_mb(priv, n); + +- if (priv->devtype_data->quirks & FLEXCAN_QUIRK_USE_OFF_TIMESTAMP) { ++ if (priv->devtype_data.quirks & FLEXCAN_QUIRK_USE_RX_MAILBOX) { + u32 code; + + do { +@@ -1083,7 +1108,7 @@ static struct sk_buff *flexcan_mailbox_read(struct can_rx_offload *offload, + } + + mark_as_read: +- if (priv->devtype_data->quirks & FLEXCAN_QUIRK_USE_OFF_TIMESTAMP) ++ if (priv->devtype_data.quirks & FLEXCAN_QUIRK_USE_RX_MAILBOX) + flexcan_write64(priv, FLEXCAN_IFLAG_MB(n), ®s->iflag1); + else + priv->write(FLEXCAN_IFLAG_RX_FIFO_AVAILABLE, ®s->iflag1); +@@ -1109,7 +1134,7 @@ static irqreturn_t flexcan_irq(int irq, void *dev_id) + enum can_state last_state = priv->can.state; + + /* reception interrupt */ +- if (priv->devtype_data->quirks & FLEXCAN_QUIRK_USE_OFF_TIMESTAMP) { ++ if (priv->devtype_data.quirks & FLEXCAN_QUIRK_USE_RX_MAILBOX) { + u64 reg_iflag_rx; + int ret; + +@@ -1169,7 +1194,7 @@ static irqreturn_t flexcan_irq(int irq, void *dev_id) + + /* state change interrupt or broken error state quirk fix is enabled */ + if ((reg_esr & FLEXCAN_ESR_ERR_STATE) || +- (priv->devtype_data->quirks & (FLEXCAN_QUIRK_BROKEN_WERR_STATE | ++ (priv->devtype_data.quirks & (FLEXCAN_QUIRK_BROKEN_WERR_STATE | + FLEXCAN_QUIRK_BROKEN_PERR_STATE))) + flexcan_irq_state(dev, reg_esr); + +@@ -1191,11 +1216,11 @@ static irqreturn_t flexcan_irq(int irq, void *dev_id) + * (1): enabled if FLEXCAN_QUIRK_BROKEN_WERR_STATE is enabled + */ + if ((last_state != priv->can.state) && +- (priv->devtype_data->quirks & FLEXCAN_QUIRK_BROKEN_PERR_STATE) && ++ (priv->devtype_data.quirks & FLEXCAN_QUIRK_BROKEN_PERR_STATE) && + !(priv->can.ctrlmode & CAN_CTRLMODE_BERR_REPORTING)) { + switch (priv->can.state) { + case CAN_STATE_ERROR_ACTIVE: +- if (priv->devtype_data->quirks & ++ if (priv->devtype_data.quirks & + FLEXCAN_QUIRK_BROKEN_WERR_STATE) + flexcan_error_irq_enable(priv); + else +@@ -1423,26 +1448,26 @@ static int flexcan_rx_offload_setup(struct net_device *dev) + else + priv->mb_size = sizeof(struct flexcan_mb) + CAN_MAX_DLEN; + +- if (priv->devtype_data->quirks & FLEXCAN_QUIRK_NR_MB_16) ++ if (priv->devtype_data.quirks & FLEXCAN_QUIRK_NR_MB_16) + priv->mb_count = 16; + else + priv->mb_count = (sizeof(priv->regs->mb[0]) / priv->mb_size) + + (sizeof(priv->regs->mb[1]) / priv->mb_size); + +- if (priv->devtype_data->quirks & FLEXCAN_QUIRK_USE_OFF_TIMESTAMP) ++ if (priv->devtype_data.quirks & FLEXCAN_QUIRK_USE_RX_MAILBOX) + priv->tx_mb_reserved = +- flexcan_get_mb(priv, FLEXCAN_TX_MB_RESERVED_OFF_TIMESTAMP); ++ flexcan_get_mb(priv, FLEXCAN_TX_MB_RESERVED_RX_MAILBOX); + else + priv->tx_mb_reserved = +- flexcan_get_mb(priv, FLEXCAN_TX_MB_RESERVED_OFF_FIFO); ++ flexcan_get_mb(priv, FLEXCAN_TX_MB_RESERVED_RX_FIFO); + priv->tx_mb_idx = priv->mb_count - 1; + priv->tx_mb = flexcan_get_mb(priv, priv->tx_mb_idx); + priv->tx_mask = FLEXCAN_IFLAG_MB(priv->tx_mb_idx); + + priv->offload.mailbox_read = flexcan_mailbox_read; + +- if (priv->devtype_data->quirks & FLEXCAN_QUIRK_USE_OFF_TIMESTAMP) { +- priv->offload.mb_first = FLEXCAN_RX_MB_OFF_TIMESTAMP_FIRST; ++ if (priv->devtype_data.quirks & FLEXCAN_QUIRK_USE_RX_MAILBOX) { ++ priv->offload.mb_first = FLEXCAN_RX_MB_RX_MAILBOX_FIRST; + priv->offload.mb_last = priv->mb_count - 2; + + priv->rx_mask = GENMASK_ULL(priv->offload.mb_last, +@@ -1506,7 +1531,7 @@ static int flexcan_chip_start(struct net_device *dev) + if (err) + goto out_chip_disable; + +- if (priv->devtype_data->quirks & FLEXCAN_QUIRK_SUPPORT_ECC) ++ if (priv->devtype_data.quirks & FLEXCAN_QUIRK_SUPPORT_ECC) + flexcan_ram_init(dev); + + flexcan_set_bittiming(dev); +@@ -1532,10 +1557,10 @@ static int flexcan_chip_start(struct net_device *dev) + /* MCR + * + * FIFO: +- * - disable for timestamp mode ++ * - disable for mailbox mode + * - enable for FIFO mode + */ +- if (priv->devtype_data->quirks & FLEXCAN_QUIRK_USE_OFF_TIMESTAMP) ++ if (priv->devtype_data.quirks & FLEXCAN_QUIRK_USE_RX_MAILBOX) + reg_mcr &= ~FLEXCAN_MCR_FEN; + else + reg_mcr |= FLEXCAN_MCR_FEN; +@@ -1586,7 +1611,7 @@ static int flexcan_chip_start(struct net_device *dev) + * on most Flexcan cores, too. Otherwise we don't get + * any error warning or passive interrupts. + */ +- if (priv->devtype_data->quirks & FLEXCAN_QUIRK_BROKEN_WERR_STATE || ++ if (priv->devtype_data.quirks & FLEXCAN_QUIRK_BROKEN_WERR_STATE || + priv->can.ctrlmode & CAN_CTRLMODE_BERR_REPORTING) + reg_ctrl |= FLEXCAN_CTRL_ERR_MSK; + else +@@ -1599,7 +1624,7 @@ static int flexcan_chip_start(struct net_device *dev) + netdev_dbg(dev, "%s: writing ctrl=0x%08x", __func__, reg_ctrl); + priv->write(reg_ctrl, ®s->ctrl); + +- if ((priv->devtype_data->quirks & FLEXCAN_QUIRK_ENABLE_EACEN_RRS)) { ++ if ((priv->devtype_data.quirks & FLEXCAN_QUIRK_ENABLE_EACEN_RRS)) { + reg_ctrl2 = priv->read(®s->ctrl2); + reg_ctrl2 |= FLEXCAN_CTRL2_EACEN | FLEXCAN_CTRL2_RRS; + priv->write(reg_ctrl2, ®s->ctrl2); +@@ -1631,7 +1656,7 @@ static int flexcan_chip_start(struct net_device *dev) + priv->write(reg_fdctrl, ®s->fdctrl); + } + +- if (priv->devtype_data->quirks & FLEXCAN_QUIRK_USE_OFF_TIMESTAMP) { ++ if (priv->devtype_data.quirks & FLEXCAN_QUIRK_USE_RX_MAILBOX) { + for (i = priv->offload.mb_first; i <= priv->offload.mb_last; i++) { + mb = flexcan_get_mb(priv, i); + priv->write(FLEXCAN_MB_CODE_RX_EMPTY, +@@ -1639,7 +1664,7 @@ static int flexcan_chip_start(struct net_device *dev) + } + } else { + /* clear and invalidate unused mailboxes first */ +- for (i = FLEXCAN_TX_MB_RESERVED_OFF_FIFO; i < priv->mb_count; i++) { ++ for (i = FLEXCAN_TX_MB_RESERVED_RX_FIFO; i < priv->mb_count; i++) { + mb = flexcan_get_mb(priv, i); + priv->write(FLEXCAN_MB_CODE_RX_INACTIVE, + &mb->can_ctrl); +@@ -1659,7 +1684,7 @@ static int flexcan_chip_start(struct net_device *dev) + priv->write(0x0, ®s->rx14mask); + priv->write(0x0, ®s->rx15mask); + +- if (priv->devtype_data->quirks & FLEXCAN_QUIRK_DISABLE_RXFG) ++ if (priv->devtype_data.quirks & FLEXCAN_QUIRK_DISABLE_RXFG) + priv->write(0x0, ®s->rxfgmask); + + /* clear acceptance filters */ +@@ -1673,7 +1698,7 @@ static int flexcan_chip_start(struct net_device *dev) + * This also works around errata e5295 which generates false + * positive memory errors and put the device in freeze mode. + */ +- if (priv->devtype_data->quirks & FLEXCAN_QUIRK_DISABLE_MECR) { ++ if (priv->devtype_data.quirks & FLEXCAN_QUIRK_DISABLE_MECR) { + /* Follow the protocol as described in "Detection + * and Correction of Memory Errors" to write to + * MECR register (step 1 - 5) +@@ -1799,7 +1824,7 @@ static int flexcan_open(struct net_device *dev) + if (err) + goto out_can_rx_offload_disable; + +- if (priv->devtype_data->quirks & FLEXCAN_QUIRK_NR_IRQ_3) { ++ if (priv->devtype_data.quirks & FLEXCAN_QUIRK_NR_IRQ_3) { + err = request_irq(priv->irq_boff, + flexcan_irq, IRQF_SHARED, dev->name, dev); + if (err) +@@ -1845,7 +1870,7 @@ static int flexcan_close(struct net_device *dev) + netif_stop_queue(dev); + flexcan_chip_interrupts_disable(dev); + +- if (priv->devtype_data->quirks & FLEXCAN_QUIRK_NR_IRQ_3) { ++ if (priv->devtype_data.quirks & FLEXCAN_QUIRK_NR_IRQ_3) { + free_irq(priv->irq_err, dev); + free_irq(priv->irq_boff, dev); + } +@@ -2051,9 +2076,9 @@ static int flexcan_setup_stop_mode(struct platform_device *pdev) + + priv = netdev_priv(dev); + +- if (priv->devtype_data->quirks & FLEXCAN_QUIRK_SETUP_STOP_MODE_SCFW) ++ if (priv->devtype_data.quirks & FLEXCAN_QUIRK_SETUP_STOP_MODE_SCFW) + ret = flexcan_setup_stop_mode_scfw(pdev); +- else if (priv->devtype_data->quirks & FLEXCAN_QUIRK_SETUP_STOP_MODE_GPR) ++ else if (priv->devtype_data.quirks & FLEXCAN_QUIRK_SETUP_STOP_MODE_GPR) + ret = flexcan_setup_stop_mode_gpr(pdev); + else + /* return 0 directly if doesn't support stop mode feature */ +@@ -2164,8 +2189,25 @@ static int flexcan_probe(struct platform_device *pdev) + return -ENODEV; + + if ((devtype_data->quirks & FLEXCAN_QUIRK_SUPPORT_FD) && +- !(devtype_data->quirks & FLEXCAN_QUIRK_USE_OFF_TIMESTAMP)) { +- dev_err(&pdev->dev, "CAN-FD mode doesn't work with FIFO mode!\n"); ++ !((devtype_data->quirks & ++ (FLEXCAN_QUIRK_USE_RX_MAILBOX | ++ FLEXCAN_QUIRK_SUPPPORT_RX_MAILBOX | ++ FLEXCAN_QUIRK_SUPPPORT_RX_MAILBOX_RTR | ++ FLEXCAN_QUIRK_SUPPPORT_RX_FIFO)) == ++ (FLEXCAN_QUIRK_USE_RX_MAILBOX | ++ FLEXCAN_QUIRK_SUPPPORT_RX_MAILBOX | ++ FLEXCAN_QUIRK_SUPPPORT_RX_MAILBOX_RTR))) { ++ dev_err(&pdev->dev, "CAN-FD mode doesn't work in RX-FIFO mode!\n"); ++ return -EINVAL; ++ } ++ ++ if ((devtype_data->quirks & ++ (FLEXCAN_QUIRK_SUPPPORT_RX_MAILBOX | ++ FLEXCAN_QUIRK_SUPPPORT_RX_MAILBOX_RTR)) == ++ FLEXCAN_QUIRK_SUPPPORT_RX_MAILBOX_RTR) { ++ dev_err(&pdev->dev, ++ "Quirks (0x%08x) inconsistent: RX_MAILBOX_RX supported but not RX_MAILBOX\n", ++ devtype_data->quirks); + return -EINVAL; + } + +@@ -2181,9 +2223,10 @@ static int flexcan_probe(struct platform_device *pdev) + dev->flags |= IFF_ECHO; + + priv = netdev_priv(dev); ++ priv->devtype_data = *devtype_data; + + if (of_property_read_bool(pdev->dev.of_node, "big-endian") || +- devtype_data->quirks & FLEXCAN_QUIRK_DEFAULT_BIG_ENDIAN) { ++ priv->devtype_data.quirks & FLEXCAN_QUIRK_DEFAULT_BIG_ENDIAN) { + priv->read = flexcan_read_be; + priv->write = flexcan_write_be; + } else { +@@ -2202,10 +2245,9 @@ static int flexcan_probe(struct platform_device *pdev) + priv->clk_ipg = clk_ipg; + priv->clk_per = clk_per; + priv->clk_src = clk_src; +- priv->devtype_data = devtype_data; + priv->reg_xceiver = reg_xceiver; + +- if (devtype_data->quirks & FLEXCAN_QUIRK_NR_IRQ_3) { ++ if (priv->devtype_data.quirks & FLEXCAN_QUIRK_NR_IRQ_3) { + priv->irq_boff = platform_get_irq(pdev, 1); + if (priv->irq_boff <= 0) { + err = -ENODEV; +@@ -2218,7 +2260,7 @@ static int flexcan_probe(struct platform_device *pdev) + } + } + +- if (priv->devtype_data->quirks & FLEXCAN_QUIRK_SUPPORT_FD) { ++ if (priv->devtype_data.quirks & FLEXCAN_QUIRK_SUPPORT_FD) { + priv->can.ctrlmode_supported |= CAN_CTRLMODE_FD | + CAN_CTRLMODE_FD_NON_ISO; + priv->can.bittiming_const = &flexcan_fd_bittiming_const; +diff --git a/drivers/net/can/rcar/rcar_canfd.c b/drivers/net/can/rcar/rcar_canfd.c +index ff9d0f5ae0dd2..388521e70837f 100644 +--- a/drivers/net/can/rcar/rcar_canfd.c ++++ b/drivers/net/can/rcar/rcar_canfd.c +@@ -1640,8 +1640,7 @@ static int rcar_canfd_channel_probe(struct rcar_canfd_global *gpriv, u32 ch, + ndev = alloc_candev(sizeof(*priv), RCANFD_FIFO_DEPTH); + if (!ndev) { + dev_err(&pdev->dev, "alloc_candev() failed\n"); +- err = -ENOMEM; +- goto fail; ++ return -ENOMEM; + } + priv = netdev_priv(ndev); + +@@ -1735,8 +1734,8 @@ static int rcar_canfd_channel_probe(struct rcar_canfd_global *gpriv, u32 ch, + + fail_candev: + netif_napi_del(&priv->napi); +- free_candev(ndev); + fail: ++ free_candev(ndev); + return err; + } + +diff --git a/drivers/net/can/softing/softing_cs.c b/drivers/net/can/softing/softing_cs.c +index 2e93ee7923739..e5c939b63fa65 100644 +--- a/drivers/net/can/softing/softing_cs.c ++++ b/drivers/net/can/softing/softing_cs.c +@@ -293,7 +293,7 @@ static int softingcs_probe(struct pcmcia_device *pcmcia) + return 0; + + platform_failed: +- kfree(dev); ++ platform_device_put(pdev); + mem_failed: + pcmcia_bad: + pcmcia_failed: +diff --git a/drivers/net/can/softing/softing_fw.c b/drivers/net/can/softing/softing_fw.c +index 7e15368779931..32286f861a195 100644 +--- a/drivers/net/can/softing/softing_fw.c ++++ b/drivers/net/can/softing/softing_fw.c +@@ -565,18 +565,19 @@ int softing_startstop(struct net_device *dev, int up) + if (ret < 0) + goto failed; + } +- /* enable_error_frame */ +- /* ++ ++ /* enable_error_frame ++ * + * Error reporting is switched off at the moment since + * the receiving of them is not yet 100% verified + * This should be enabled sooner or later +- * +- if (error_reporting) { ++ */ ++ if (0 && error_reporting) { + ret = softing_fct_cmd(card, 51, "enable_error_frame"); + if (ret < 0) + goto failed; + } +- */ ++ + /* initialize interface */ + iowrite16(1, &card->dpram[DPRAM_FCT_PARAM + 2]); + iowrite16(1, &card->dpram[DPRAM_FCT_PARAM + 4]); +diff --git a/drivers/net/can/spi/mcp251xfd/mcp251xfd-core.c b/drivers/net/can/spi/mcp251xfd/mcp251xfd-core.c +index e16dc482f3270..9a4791d88683c 100644 +--- a/drivers/net/can/spi/mcp251xfd/mcp251xfd-core.c ++++ b/drivers/net/can/spi/mcp251xfd/mcp251xfd-core.c +@@ -1336,7 +1336,7 @@ mcp251xfd_tef_obj_read(const struct mcp251xfd_priv *priv, + len > tx_ring->obj_num || + offset + len > tx_ring->obj_num)) { + netdev_err(priv->ndev, +- "Trying to read to many TEF objects (max=%d, offset=%d, len=%d).\n", ++ "Trying to read too many TEF objects (max=%d, offset=%d, len=%d).\n", + tx_ring->obj_num, offset, len); + return -ERANGE; + } +@@ -2625,7 +2625,7 @@ static int mcp251xfd_register_chip_detect(struct mcp251xfd_priv *priv) + if (!mcp251xfd_is_251X(priv) && + priv->devtype_data.model != devtype_data->model) { + netdev_info(ndev, +- "Detected %s, but firmware specifies a %s. Fixing up.", ++ "Detected %s, but firmware specifies a %s. Fixing up.\n", + __mcp251xfd_get_model_str(devtype_data->model), + mcp251xfd_get_model_str(priv)); + } +@@ -2662,7 +2662,7 @@ static int mcp251xfd_register_check_rx_int(struct mcp251xfd_priv *priv) + return 0; + + netdev_info(priv->ndev, +- "RX_INT active after softreset, disabling RX_INT support."); ++ "RX_INT active after softreset, disabling RX_INT support.\n"); + devm_gpiod_put(&priv->spi->dev, priv->rx_int); + priv->rx_int = NULL; + +diff --git a/drivers/net/can/xilinx_can.c b/drivers/net/can/xilinx_can.c +index 3b883e607d8ba..a579b9b791ede 100644 +--- a/drivers/net/can/xilinx_can.c ++++ b/drivers/net/can/xilinx_can.c +@@ -1762,7 +1762,12 @@ static int xcan_probe(struct platform_device *pdev) + spin_lock_init(&priv->tx_lock); + + /* Get IRQ for the device */ +- ndev->irq = platform_get_irq(pdev, 0); ++ ret = platform_get_irq(pdev, 0); ++ if (ret < 0) ++ goto err_free; ++ ++ ndev->irq = ret; ++ + ndev->flags |= IFF_ECHO; /* We support local echo */ + + platform_set_drvdata(pdev, ndev); +diff --git a/drivers/net/dsa/hirschmann/hellcreek.c b/drivers/net/dsa/hirschmann/hellcreek.c +index 354655f9ed003..950a54ec4b59b 100644 +--- a/drivers/net/dsa/hirschmann/hellcreek.c ++++ b/drivers/net/dsa/hirschmann/hellcreek.c +@@ -710,8 +710,9 @@ static int __hellcreek_fdb_add(struct hellcreek *hellcreek, + u16 meta = 0; + + dev_dbg(hellcreek->dev, "Add static FDB entry: MAC=%pM, MASK=0x%02x, " +- "OBT=%d, REPRIO_EN=%d, PRIO=%d\n", entry->mac, entry->portmask, +- entry->is_obt, entry->reprio_en, entry->reprio_tc); ++ "OBT=%d, PASS_BLOCKED=%d, REPRIO_EN=%d, PRIO=%d\n", entry->mac, ++ entry->portmask, entry->is_obt, entry->pass_blocked, ++ entry->reprio_en, entry->reprio_tc); + + /* Add mac address */ + hellcreek_write(hellcreek, entry->mac[1] | (entry->mac[0] << 8), HR_FDBWDH); +@@ -722,6 +723,8 @@ static int __hellcreek_fdb_add(struct hellcreek *hellcreek, + meta |= entry->portmask << HR_FDBWRM0_PORTMASK_SHIFT; + if (entry->is_obt) + meta |= HR_FDBWRM0_OBT; ++ if (entry->pass_blocked) ++ meta |= HR_FDBWRM0_PASS_BLOCKED; + if (entry->reprio_en) { + meta |= HR_FDBWRM0_REPRIO_EN; + meta |= entry->reprio_tc << HR_FDBWRM0_REPRIO_TC_SHIFT; +@@ -1049,7 +1052,7 @@ static void hellcreek_setup_tc_identity_mapping(struct hellcreek *hellcreek) + + static int hellcreek_setup_fdb(struct hellcreek *hellcreek) + { +- static struct hellcreek_fdb_entry ptp = { ++ static struct hellcreek_fdb_entry l2_ptp = { + /* MAC: 01-1B-19-00-00-00 */ + .mac = { 0x01, 0x1b, 0x19, 0x00, 0x00, 0x00 }, + .portmask = 0x03, /* Management ports */ +@@ -1060,24 +1063,94 @@ static int hellcreek_setup_fdb(struct hellcreek *hellcreek) + .reprio_tc = 6, /* TC: 6 as per IEEE 802.1AS */ + .reprio_en = 1, + }; +- static struct hellcreek_fdb_entry p2p = { ++ static struct hellcreek_fdb_entry udp4_ptp = { ++ /* MAC: 01-00-5E-00-01-81 */ ++ .mac = { 0x01, 0x00, 0x5e, 0x00, 0x01, 0x81 }, ++ .portmask = 0x03, /* Management ports */ ++ .age = 0, ++ .is_obt = 0, ++ .pass_blocked = 0, ++ .is_static = 1, ++ .reprio_tc = 6, ++ .reprio_en = 1, ++ }; ++ static struct hellcreek_fdb_entry udp6_ptp = { ++ /* MAC: 33-33-00-00-01-81 */ ++ .mac = { 0x33, 0x33, 0x00, 0x00, 0x01, 0x81 }, ++ .portmask = 0x03, /* Management ports */ ++ .age = 0, ++ .is_obt = 0, ++ .pass_blocked = 0, ++ .is_static = 1, ++ .reprio_tc = 6, ++ .reprio_en = 1, ++ }; ++ static struct hellcreek_fdb_entry l2_p2p = { + /* MAC: 01-80-C2-00-00-0E */ + .mac = { 0x01, 0x80, 0xc2, 0x00, 0x00, 0x0e }, + .portmask = 0x03, /* Management ports */ + .age = 0, + .is_obt = 0, +- .pass_blocked = 0, ++ .pass_blocked = 1, + .is_static = 1, + .reprio_tc = 6, /* TC: 6 as per IEEE 802.1AS */ + .reprio_en = 1, + }; ++ static struct hellcreek_fdb_entry udp4_p2p = { ++ /* MAC: 01-00-5E-00-00-6B */ ++ .mac = { 0x01, 0x00, 0x5e, 0x00, 0x00, 0x6b }, ++ .portmask = 0x03, /* Management ports */ ++ .age = 0, ++ .is_obt = 0, ++ .pass_blocked = 1, ++ .is_static = 1, ++ .reprio_tc = 6, ++ .reprio_en = 1, ++ }; ++ static struct hellcreek_fdb_entry udp6_p2p = { ++ /* MAC: 33-33-00-00-00-6B */ ++ .mac = { 0x33, 0x33, 0x00, 0x00, 0x00, 0x6b }, ++ .portmask = 0x03, /* Management ports */ ++ .age = 0, ++ .is_obt = 0, ++ .pass_blocked = 1, ++ .is_static = 1, ++ .reprio_tc = 6, ++ .reprio_en = 1, ++ }; ++ static struct hellcreek_fdb_entry stp = { ++ /* MAC: 01-80-C2-00-00-00 */ ++ .mac = { 0x01, 0x80, 0xc2, 0x00, 0x00, 0x00 }, ++ .portmask = 0x03, /* Management ports */ ++ .age = 0, ++ .is_obt = 0, ++ .pass_blocked = 1, ++ .is_static = 1, ++ .reprio_tc = 6, ++ .reprio_en = 1, ++ }; + int ret; + + mutex_lock(&hellcreek->reg_lock); +- ret = __hellcreek_fdb_add(hellcreek, &ptp); ++ ret = __hellcreek_fdb_add(hellcreek, &l2_ptp); ++ if (ret) ++ goto out; ++ ret = __hellcreek_fdb_add(hellcreek, &udp4_ptp); ++ if (ret) ++ goto out; ++ ret = __hellcreek_fdb_add(hellcreek, &udp6_ptp); ++ if (ret) ++ goto out; ++ ret = __hellcreek_fdb_add(hellcreek, &l2_p2p); ++ if (ret) ++ goto out; ++ ret = __hellcreek_fdb_add(hellcreek, &udp4_p2p); ++ if (ret) ++ goto out; ++ ret = __hellcreek_fdb_add(hellcreek, &udp6_p2p); + if (ret) + goto out; +- ret = __hellcreek_fdb_add(hellcreek, &p2p); ++ ret = __hellcreek_fdb_add(hellcreek, &stp); + out: + mutex_unlock(&hellcreek->reg_lock); + +diff --git a/drivers/net/ethernet/broadcom/bnxt/Makefile b/drivers/net/ethernet/broadcom/bnxt/Makefile +index c6ef7ec2c1151..2bc2b707d6eee 100644 +--- a/drivers/net/ethernet/broadcom/bnxt/Makefile ++++ b/drivers/net/ethernet/broadcom/bnxt/Makefile +@@ -1,6 +1,6 @@ + # SPDX-License-Identifier: GPL-2.0-only + obj-$(CONFIG_BNXT) += bnxt_en.o + +-bnxt_en-y := bnxt.o bnxt_hwrm.o bnxt_sriov.o bnxt_ethtool.o bnxt_dcb.o bnxt_ulp.o bnxt_xdp.o bnxt_ptp.o bnxt_vfr.o bnxt_devlink.o bnxt_dim.o ++bnxt_en-y := bnxt.o bnxt_hwrm.o bnxt_sriov.o bnxt_ethtool.o bnxt_dcb.o bnxt_ulp.o bnxt_xdp.o bnxt_ptp.o bnxt_vfr.o bnxt_devlink.o bnxt_dim.o bnxt_coredump.o + bnxt_en-$(CONFIG_BNXT_FLOWER_OFFLOAD) += bnxt_tc.o + bnxt_en-$(CONFIG_DEBUG_FS) += bnxt_debugfs.o +diff --git a/drivers/net/ethernet/broadcom/bnxt/bnxt.c b/drivers/net/ethernet/broadcom/bnxt/bnxt.c +index 0fba01db336cc..a8855a200a3c5 100644 +--- a/drivers/net/ethernet/broadcom/bnxt/bnxt.c ++++ b/drivers/net/ethernet/broadcom/bnxt/bnxt.c +@@ -8004,6 +8004,12 @@ static int bnxt_hwrm_ver_get(struct bnxt *bp) + bp->hwrm_cmd_timeout = le16_to_cpu(resp->def_req_timeout); + if (!bp->hwrm_cmd_timeout) + bp->hwrm_cmd_timeout = DFLT_HWRM_CMD_TIMEOUT; ++ bp->hwrm_cmd_max_timeout = le16_to_cpu(resp->max_req_timeout) * 1000; ++ if (!bp->hwrm_cmd_max_timeout) ++ bp->hwrm_cmd_max_timeout = HWRM_CMD_MAX_TIMEOUT; ++ else if (bp->hwrm_cmd_max_timeout > HWRM_CMD_MAX_TIMEOUT) ++ netdev_warn(bp->dev, "Device requests max timeout of %d seconds, may trigger hung task watchdog\n", ++ bp->hwrm_cmd_max_timeout / 1000); + + if (resp->hwrm_intf_maj_8b >= 1) { + bp->hwrm_max_req_len = le16_to_cpu(resp->max_req_win_len); +diff --git a/drivers/net/ethernet/broadcom/bnxt/bnxt.h b/drivers/net/ethernet/broadcom/bnxt/bnxt.h +index 19fe6478e9b4b..0a5137c1f6d4e 100644 +--- a/drivers/net/ethernet/broadcom/bnxt/bnxt.h ++++ b/drivers/net/ethernet/broadcom/bnxt/bnxt.h +@@ -1901,7 +1901,8 @@ struct bnxt { + + u16 hwrm_max_req_len; + u16 hwrm_max_ext_req_len; +- int hwrm_cmd_timeout; ++ unsigned int hwrm_cmd_timeout; ++ unsigned int hwrm_cmd_max_timeout; + struct mutex hwrm_cmd_lock; /* serialize hwrm messages */ + struct hwrm_ver_get_output ver_resp; + #define FW_VER_STR_LEN 32 +diff --git a/drivers/net/ethernet/broadcom/bnxt/bnxt_coredump.c b/drivers/net/ethernet/broadcom/bnxt/bnxt_coredump.c +new file mode 100644 +index 0000000000000..156f76bcea7eb +--- /dev/null ++++ b/drivers/net/ethernet/broadcom/bnxt/bnxt_coredump.c +@@ -0,0 +1,372 @@ ++/* Broadcom NetXtreme-C/E network driver. ++ * ++ * Copyright (c) 2021 Broadcom Limited ++ * ++ * This program is free software; you can redistribute it and/or modify ++ * it under the terms of the GNU General Public License as published by ++ * the Free Software Foundation. ++ */ ++ ++#include <linux/types.h> ++#include <linux/errno.h> ++#include <linux/pci.h> ++#include "bnxt_hsi.h" ++#include "bnxt.h" ++#include "bnxt_hwrm.h" ++#include "bnxt_coredump.h" ++ ++static int bnxt_hwrm_dbg_dma_data(struct bnxt *bp, void *msg, ++ struct bnxt_hwrm_dbg_dma_info *info) ++{ ++ struct hwrm_dbg_cmn_input *cmn_req = msg; ++ __le16 *seq_ptr = msg + info->seq_off; ++ struct hwrm_dbg_cmn_output *cmn_resp; ++ u16 seq = 0, len, segs_off; ++ dma_addr_t dma_handle; ++ void *dma_buf, *resp; ++ int rc, off = 0; ++ ++ dma_buf = hwrm_req_dma_slice(bp, msg, info->dma_len, &dma_handle); ++ if (!dma_buf) { ++ hwrm_req_drop(bp, msg); ++ return -ENOMEM; ++ } ++ ++ hwrm_req_timeout(bp, msg, bp->hwrm_cmd_max_timeout); ++ cmn_resp = hwrm_req_hold(bp, msg); ++ resp = cmn_resp; ++ ++ segs_off = offsetof(struct hwrm_dbg_coredump_list_output, ++ total_segments); ++ cmn_req->host_dest_addr = cpu_to_le64(dma_handle); ++ cmn_req->host_buf_len = cpu_to_le32(info->dma_len); ++ while (1) { ++ *seq_ptr = cpu_to_le16(seq); ++ rc = hwrm_req_send(bp, msg); ++ if (rc) ++ break; ++ ++ len = le16_to_cpu(*((__le16 *)(resp + info->data_len_off))); ++ if (!seq && ++ cmn_req->req_type == cpu_to_le16(HWRM_DBG_COREDUMP_LIST)) { ++ info->segs = le16_to_cpu(*((__le16 *)(resp + ++ segs_off))); ++ if (!info->segs) { ++ rc = -EIO; ++ break; ++ } ++ ++ info->dest_buf_size = info->segs * ++ sizeof(struct coredump_segment_record); ++ info->dest_buf = kmalloc(info->dest_buf_size, ++ GFP_KERNEL); ++ if (!info->dest_buf) { ++ rc = -ENOMEM; ++ break; ++ } ++ } ++ ++ if (info->dest_buf) { ++ if ((info->seg_start + off + len) <= ++ BNXT_COREDUMP_BUF_LEN(info->buf_len)) { ++ memcpy(info->dest_buf + off, dma_buf, len); ++ } else { ++ rc = -ENOBUFS; ++ break; ++ } ++ } ++ ++ if (cmn_req->req_type == ++ cpu_to_le16(HWRM_DBG_COREDUMP_RETRIEVE)) ++ info->dest_buf_size += len; ++ ++ if (!(cmn_resp->flags & HWRM_DBG_CMN_FLAGS_MORE)) ++ break; ++ ++ seq++; ++ off += len; ++ } ++ hwrm_req_drop(bp, msg); ++ return rc; ++} ++ ++static int bnxt_hwrm_dbg_coredump_list(struct bnxt *bp, ++ struct bnxt_coredump *coredump) ++{ ++ struct bnxt_hwrm_dbg_dma_info info = {NULL}; ++ struct hwrm_dbg_coredump_list_input *req; ++ int rc; ++ ++ rc = hwrm_req_init(bp, req, HWRM_DBG_COREDUMP_LIST); ++ if (rc) ++ return rc; ++ ++ info.dma_len = COREDUMP_LIST_BUF_LEN; ++ info.seq_off = offsetof(struct hwrm_dbg_coredump_list_input, seq_no); ++ info.data_len_off = offsetof(struct hwrm_dbg_coredump_list_output, ++ data_len); ++ ++ rc = bnxt_hwrm_dbg_dma_data(bp, req, &info); ++ if (!rc) { ++ coredump->data = info.dest_buf; ++ coredump->data_size = info.dest_buf_size; ++ coredump->total_segs = info.segs; ++ } ++ return rc; ++} ++ ++static int bnxt_hwrm_dbg_coredump_initiate(struct bnxt *bp, u16 component_id, ++ u16 segment_id) ++{ ++ struct hwrm_dbg_coredump_initiate_input *req; ++ int rc; ++ ++ rc = hwrm_req_init(bp, req, HWRM_DBG_COREDUMP_INITIATE); ++ if (rc) ++ return rc; ++ ++ hwrm_req_timeout(bp, req, bp->hwrm_cmd_max_timeout); ++ req->component_id = cpu_to_le16(component_id); ++ req->segment_id = cpu_to_le16(segment_id); ++ ++ return hwrm_req_send(bp, req); ++} ++ ++static int bnxt_hwrm_dbg_coredump_retrieve(struct bnxt *bp, u16 component_id, ++ u16 segment_id, u32 *seg_len, ++ void *buf, u32 buf_len, u32 offset) ++{ ++ struct hwrm_dbg_coredump_retrieve_input *req; ++ struct bnxt_hwrm_dbg_dma_info info = {NULL}; ++ int rc; ++ ++ rc = hwrm_req_init(bp, req, HWRM_DBG_COREDUMP_RETRIEVE); ++ if (rc) ++ return rc; ++ ++ req->component_id = cpu_to_le16(component_id); ++ req->segment_id = cpu_to_le16(segment_id); ++ ++ info.dma_len = COREDUMP_RETRIEVE_BUF_LEN; ++ info.seq_off = offsetof(struct hwrm_dbg_coredump_retrieve_input, ++ seq_no); ++ info.data_len_off = offsetof(struct hwrm_dbg_coredump_retrieve_output, ++ data_len); ++ if (buf) { ++ info.dest_buf = buf + offset; ++ info.buf_len = buf_len; ++ info.seg_start = offset; ++ } ++ ++ rc = bnxt_hwrm_dbg_dma_data(bp, req, &info); ++ if (!rc) ++ *seg_len = info.dest_buf_size; ++ ++ return rc; ++} ++ ++static void ++bnxt_fill_coredump_seg_hdr(struct bnxt *bp, ++ struct bnxt_coredump_segment_hdr *seg_hdr, ++ struct coredump_segment_record *seg_rec, u32 seg_len, ++ int status, u32 duration, u32 instance) ++{ ++ memset(seg_hdr, 0, sizeof(*seg_hdr)); ++ memcpy(seg_hdr->signature, "sEgM", 4); ++ if (seg_rec) { ++ seg_hdr->component_id = (__force __le32)seg_rec->component_id; ++ seg_hdr->segment_id = (__force __le32)seg_rec->segment_id; ++ seg_hdr->low_version = seg_rec->version_low; ++ seg_hdr->high_version = seg_rec->version_hi; ++ } else { ++ /* For hwrm_ver_get response Component id = 2 ++ * and Segment id = 0 ++ */ ++ seg_hdr->component_id = cpu_to_le32(2); ++ seg_hdr->segment_id = 0; ++ } ++ seg_hdr->function_id = cpu_to_le16(bp->pdev->devfn); ++ seg_hdr->length = cpu_to_le32(seg_len); ++ seg_hdr->status = cpu_to_le32(status); ++ seg_hdr->duration = cpu_to_le32(duration); ++ seg_hdr->data_offset = cpu_to_le32(sizeof(*seg_hdr)); ++ seg_hdr->instance = cpu_to_le32(instance); ++} ++ ++static void ++bnxt_fill_coredump_record(struct bnxt *bp, struct bnxt_coredump_record *record, ++ time64_t start, s16 start_utc, u16 total_segs, ++ int status) ++{ ++ time64_t end = ktime_get_real_seconds(); ++ u32 os_ver_major = 0, os_ver_minor = 0; ++ struct tm tm; ++ ++ time64_to_tm(start, 0, &tm); ++ memset(record, 0, sizeof(*record)); ++ memcpy(record->signature, "cOrE", 4); ++ record->flags = 0; ++ record->low_version = 0; ++ record->high_version = 1; ++ record->asic_state = 0; ++ strscpy(record->system_name, utsname()->nodename, ++ sizeof(record->system_name)); ++ record->year = cpu_to_le16(tm.tm_year + 1900); ++ record->month = cpu_to_le16(tm.tm_mon + 1); ++ record->day = cpu_to_le16(tm.tm_mday); ++ record->hour = cpu_to_le16(tm.tm_hour); ++ record->minute = cpu_to_le16(tm.tm_min); ++ record->second = cpu_to_le16(tm.tm_sec); ++ record->utc_bias = cpu_to_le16(start_utc); ++ strcpy(record->commandline, "ethtool -w"); ++ record->total_segments = cpu_to_le32(total_segs); ++ ++ if (sscanf(utsname()->release, "%u.%u", &os_ver_major, &os_ver_minor) != 2) ++ netdev_warn(bp->dev, "Unknown OS release in coredump\n"); ++ record->os_ver_major = cpu_to_le32(os_ver_major); ++ record->os_ver_minor = cpu_to_le32(os_ver_minor); ++ ++ strscpy(record->os_name, utsname()->sysname, sizeof(record->os_name)); ++ time64_to_tm(end, 0, &tm); ++ record->end_year = cpu_to_le16(tm.tm_year + 1900); ++ record->end_month = cpu_to_le16(tm.tm_mon + 1); ++ record->end_day = cpu_to_le16(tm.tm_mday); ++ record->end_hour = cpu_to_le16(tm.tm_hour); ++ record->end_minute = cpu_to_le16(tm.tm_min); ++ record->end_second = cpu_to_le16(tm.tm_sec); ++ record->end_utc_bias = cpu_to_le16(sys_tz.tz_minuteswest * 60); ++ record->asic_id1 = cpu_to_le32(bp->chip_num << 16 | ++ bp->ver_resp.chip_rev << 8 | ++ bp->ver_resp.chip_metal); ++ record->asic_id2 = 0; ++ record->coredump_status = cpu_to_le32(status); ++ record->ioctl_low_version = 0; ++ record->ioctl_high_version = 0; ++} ++ ++static int __bnxt_get_coredump(struct bnxt *bp, void *buf, u32 *dump_len) ++{ ++ u32 ver_get_resp_len = sizeof(struct hwrm_ver_get_output); ++ u32 offset = 0, seg_hdr_len, seg_record_len, buf_len = 0; ++ struct coredump_segment_record *seg_record = NULL; ++ struct bnxt_coredump_segment_hdr seg_hdr; ++ struct bnxt_coredump coredump = {NULL}; ++ time64_t start_time; ++ u16 start_utc; ++ int rc = 0, i; ++ ++ if (buf) ++ buf_len = *dump_len; ++ ++ start_time = ktime_get_real_seconds(); ++ start_utc = sys_tz.tz_minuteswest * 60; ++ seg_hdr_len = sizeof(seg_hdr); ++ ++ /* First segment should be hwrm_ver_get response */ ++ *dump_len = seg_hdr_len + ver_get_resp_len; ++ if (buf) { ++ bnxt_fill_coredump_seg_hdr(bp, &seg_hdr, NULL, ver_get_resp_len, ++ 0, 0, 0); ++ memcpy(buf + offset, &seg_hdr, seg_hdr_len); ++ offset += seg_hdr_len; ++ memcpy(buf + offset, &bp->ver_resp, ver_get_resp_len); ++ offset += ver_get_resp_len; ++ } ++ ++ rc = bnxt_hwrm_dbg_coredump_list(bp, &coredump); ++ if (rc) { ++ netdev_err(bp->dev, "Failed to get coredump segment list\n"); ++ goto err; ++ } ++ ++ *dump_len += seg_hdr_len * coredump.total_segs; ++ ++ seg_record = (struct coredump_segment_record *)coredump.data; ++ seg_record_len = sizeof(*seg_record); ++ ++ for (i = 0; i < coredump.total_segs; i++) { ++ u16 comp_id = le16_to_cpu(seg_record->component_id); ++ u16 seg_id = le16_to_cpu(seg_record->segment_id); ++ u32 duration = 0, seg_len = 0; ++ unsigned long start, end; ++ ++ if (buf && ((offset + seg_hdr_len) > ++ BNXT_COREDUMP_BUF_LEN(buf_len))) { ++ rc = -ENOBUFS; ++ goto err; ++ } ++ ++ start = jiffies; ++ ++ rc = bnxt_hwrm_dbg_coredump_initiate(bp, comp_id, seg_id); ++ if (rc) { ++ netdev_err(bp->dev, ++ "Failed to initiate coredump for seg = %d\n", ++ seg_record->segment_id); ++ goto next_seg; ++ } ++ ++ /* Write segment data into the buffer */ ++ rc = bnxt_hwrm_dbg_coredump_retrieve(bp, comp_id, seg_id, ++ &seg_len, buf, buf_len, ++ offset + seg_hdr_len); ++ if (rc && rc == -ENOBUFS) ++ goto err; ++ else if (rc) ++ netdev_err(bp->dev, ++ "Failed to retrieve coredump for seg = %d\n", ++ seg_record->segment_id); ++ ++next_seg: ++ end = jiffies; ++ duration = jiffies_to_msecs(end - start); ++ bnxt_fill_coredump_seg_hdr(bp, &seg_hdr, seg_record, seg_len, ++ rc, duration, 0); ++ ++ if (buf) { ++ /* Write segment header into the buffer */ ++ memcpy(buf + offset, &seg_hdr, seg_hdr_len); ++ offset += seg_hdr_len + seg_len; ++ } ++ ++ *dump_len += seg_len; ++ seg_record = ++ (struct coredump_segment_record *)((u8 *)seg_record + ++ seg_record_len); ++ } ++ ++err: ++ if (buf) ++ bnxt_fill_coredump_record(bp, buf + offset, start_time, ++ start_utc, coredump.total_segs + 1, ++ rc); ++ kfree(coredump.data); ++ *dump_len += sizeof(struct bnxt_coredump_record); ++ if (rc == -ENOBUFS) ++ netdev_err(bp->dev, "Firmware returned large coredump buffer\n"); ++ return rc; ++} ++ ++int bnxt_get_coredump(struct bnxt *bp, u16 dump_type, void *buf, u32 *dump_len) ++{ ++ if (dump_type == BNXT_DUMP_CRASH) { ++#ifdef CONFIG_TEE_BNXT_FW ++ return tee_bnxt_copy_coredump(buf, 0, *dump_len); ++#else ++ return -EOPNOTSUPP; ++#endif ++ } else { ++ return __bnxt_get_coredump(bp, buf, dump_len); ++ } ++} ++ ++u32 bnxt_get_coredump_length(struct bnxt *bp, u16 dump_type) ++{ ++ u32 len = 0; ++ ++ if (dump_type == BNXT_DUMP_CRASH) ++ len = BNXT_CRASH_DUMP_LEN; ++ else ++ __bnxt_get_coredump(bp, NULL, &len); ++ return len; ++} +diff --git a/drivers/net/ethernet/broadcom/bnxt/bnxt_coredump.h b/drivers/net/ethernet/broadcom/bnxt/bnxt_coredump.h +index 09c22f8fe3991..b1a1b2fffb194 100644 +--- a/drivers/net/ethernet/broadcom/bnxt/bnxt_coredump.h ++++ b/drivers/net/ethernet/broadcom/bnxt/bnxt_coredump.h +@@ -10,6 +10,10 @@ + #ifndef BNXT_COREDUMP_H + #define BNXT_COREDUMP_H + ++#include <linux/utsname.h> ++#include <linux/time.h> ++#include <linux/rtc.h> ++ + struct bnxt_coredump_segment_hdr { + __u8 signature[4]; + __le32 component_id; +@@ -63,4 +67,51 @@ struct bnxt_coredump_record { + __u8 ioctl_high_version; + __le16 rsvd3[313]; + }; ++ ++#define BNXT_CRASH_DUMP_LEN (8 << 20) ++ ++#define COREDUMP_LIST_BUF_LEN 2048 ++#define COREDUMP_RETRIEVE_BUF_LEN 4096 ++ ++struct bnxt_coredump { ++ void *data; ++ int data_size; ++ u16 total_segs; ++}; ++ ++#define BNXT_COREDUMP_BUF_LEN(len) ((len) - sizeof(struct bnxt_coredump_record)) ++ ++struct bnxt_hwrm_dbg_dma_info { ++ void *dest_buf; ++ int dest_buf_size; ++ u16 dma_len; ++ u16 seq_off; ++ u16 data_len_off; ++ u16 segs; ++ u32 seg_start; ++ u32 buf_len; ++}; ++ ++struct hwrm_dbg_cmn_input { ++ __le16 req_type; ++ __le16 cmpl_ring; ++ __le16 seq_id; ++ __le16 target_id; ++ __le64 resp_addr; ++ __le64 host_dest_addr; ++ __le32 host_buf_len; ++}; ++ ++struct hwrm_dbg_cmn_output { ++ __le16 error_code; ++ __le16 req_type; ++ __le16 seq_id; ++ __le16 resp_len; ++ u8 flags; ++ #define HWRM_DBG_CMN_FLAGS_MORE 1 ++}; ++ ++int bnxt_get_coredump(struct bnxt *bp, u16 dump_type, void *buf, u32 *dump_len); ++u32 bnxt_get_coredump_length(struct bnxt *bp, u16 dump_type); ++ + #endif +diff --git a/drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c b/drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c +index 7260910e75fb2..2497925105215 100644 +--- a/drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c ++++ b/drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c +@@ -31,9 +31,6 @@ + #include "bnxt_nvm_defs.h" /* NVRAM content constant and structure defs */ + #include "bnxt_fw_hdr.h" /* Firmware hdr constant and structure defs */ + #include "bnxt_coredump.h" +-#define FLASH_NVRAM_TIMEOUT ((HWRM_CMD_TIMEOUT) * 100) +-#define FLASH_PACKAGE_TIMEOUT ((HWRM_CMD_TIMEOUT) * 200) +-#define INSTALL_PACKAGE_TIMEOUT ((HWRM_CMD_TIMEOUT) * 200) + + static u32 bnxt_get_msglevel(struct net_device *dev) + { +@@ -2167,7 +2164,7 @@ static int bnxt_flash_nvram(struct net_device *dev, u16 dir_type, + req->host_src_addr = cpu_to_le64(dma_handle); + } + +- hwrm_req_timeout(bp, req, FLASH_NVRAM_TIMEOUT); ++ hwrm_req_timeout(bp, req, bp->hwrm_cmd_max_timeout); + req->dir_type = cpu_to_le16(dir_type); + req->dir_ordinal = cpu_to_le16(dir_ordinal); + req->dir_ext = cpu_to_le16(dir_ext); +@@ -2508,8 +2505,8 @@ int bnxt_flash_package_from_fw_obj(struct net_device *dev, const struct firmware + return rc; + } + +- hwrm_req_timeout(bp, modify, FLASH_PACKAGE_TIMEOUT); +- hwrm_req_timeout(bp, install, INSTALL_PACKAGE_TIMEOUT); ++ hwrm_req_timeout(bp, modify, bp->hwrm_cmd_max_timeout); ++ hwrm_req_timeout(bp, install, bp->hwrm_cmd_max_timeout); + + hwrm_req_hold(bp, modify); + modify->host_src_addr = cpu_to_le64(dma_handle); +@@ -3609,337 +3606,6 @@ static int bnxt_reset(struct net_device *dev, u32 *flags) + return 0; + } + +-static int bnxt_hwrm_dbg_dma_data(struct bnxt *bp, void *msg, +- struct bnxt_hwrm_dbg_dma_info *info) +-{ +- struct hwrm_dbg_cmn_input *cmn_req = msg; +- __le16 *seq_ptr = msg + info->seq_off; +- struct hwrm_dbg_cmn_output *cmn_resp; +- u16 seq = 0, len, segs_off; +- dma_addr_t dma_handle; +- void *dma_buf, *resp; +- int rc, off = 0; +- +- dma_buf = hwrm_req_dma_slice(bp, msg, info->dma_len, &dma_handle); +- if (!dma_buf) { +- hwrm_req_drop(bp, msg); +- return -ENOMEM; +- } +- +- hwrm_req_timeout(bp, msg, HWRM_COREDUMP_TIMEOUT); +- cmn_resp = hwrm_req_hold(bp, msg); +- resp = cmn_resp; +- +- segs_off = offsetof(struct hwrm_dbg_coredump_list_output, +- total_segments); +- cmn_req->host_dest_addr = cpu_to_le64(dma_handle); +- cmn_req->host_buf_len = cpu_to_le32(info->dma_len); +- while (1) { +- *seq_ptr = cpu_to_le16(seq); +- rc = hwrm_req_send(bp, msg); +- if (rc) +- break; +- +- len = le16_to_cpu(*((__le16 *)(resp + info->data_len_off))); +- if (!seq && +- cmn_req->req_type == cpu_to_le16(HWRM_DBG_COREDUMP_LIST)) { +- info->segs = le16_to_cpu(*((__le16 *)(resp + +- segs_off))); +- if (!info->segs) { +- rc = -EIO; +- break; +- } +- +- info->dest_buf_size = info->segs * +- sizeof(struct coredump_segment_record); +- info->dest_buf = kmalloc(info->dest_buf_size, +- GFP_KERNEL); +- if (!info->dest_buf) { +- rc = -ENOMEM; +- break; +- } +- } +- +- if (info->dest_buf) { +- if ((info->seg_start + off + len) <= +- BNXT_COREDUMP_BUF_LEN(info->buf_len)) { +- memcpy(info->dest_buf + off, dma_buf, len); +- } else { +- rc = -ENOBUFS; +- break; +- } +- } +- +- if (cmn_req->req_type == +- cpu_to_le16(HWRM_DBG_COREDUMP_RETRIEVE)) +- info->dest_buf_size += len; +- +- if (!(cmn_resp->flags & HWRM_DBG_CMN_FLAGS_MORE)) +- break; +- +- seq++; +- off += len; +- } +- hwrm_req_drop(bp, msg); +- return rc; +-} +- +-static int bnxt_hwrm_dbg_coredump_list(struct bnxt *bp, +- struct bnxt_coredump *coredump) +-{ +- struct bnxt_hwrm_dbg_dma_info info = {NULL}; +- struct hwrm_dbg_coredump_list_input *req; +- int rc; +- +- rc = hwrm_req_init(bp, req, HWRM_DBG_COREDUMP_LIST); +- if (rc) +- return rc; +- +- info.dma_len = COREDUMP_LIST_BUF_LEN; +- info.seq_off = offsetof(struct hwrm_dbg_coredump_list_input, seq_no); +- info.data_len_off = offsetof(struct hwrm_dbg_coredump_list_output, +- data_len); +- +- rc = bnxt_hwrm_dbg_dma_data(bp, req, &info); +- if (!rc) { +- coredump->data = info.dest_buf; +- coredump->data_size = info.dest_buf_size; +- coredump->total_segs = info.segs; +- } +- return rc; +-} +- +-static int bnxt_hwrm_dbg_coredump_initiate(struct bnxt *bp, u16 component_id, +- u16 segment_id) +-{ +- struct hwrm_dbg_coredump_initiate_input *req; +- int rc; +- +- rc = hwrm_req_init(bp, req, HWRM_DBG_COREDUMP_INITIATE); +- if (rc) +- return rc; +- +- hwrm_req_timeout(bp, req, HWRM_COREDUMP_TIMEOUT); +- req->component_id = cpu_to_le16(component_id); +- req->segment_id = cpu_to_le16(segment_id); +- +- return hwrm_req_send(bp, req); +-} +- +-static int bnxt_hwrm_dbg_coredump_retrieve(struct bnxt *bp, u16 component_id, +- u16 segment_id, u32 *seg_len, +- void *buf, u32 buf_len, u32 offset) +-{ +- struct hwrm_dbg_coredump_retrieve_input *req; +- struct bnxt_hwrm_dbg_dma_info info = {NULL}; +- int rc; +- +- rc = hwrm_req_init(bp, req, HWRM_DBG_COREDUMP_RETRIEVE); +- if (rc) +- return rc; +- +- req->component_id = cpu_to_le16(component_id); +- req->segment_id = cpu_to_le16(segment_id); +- +- info.dma_len = COREDUMP_RETRIEVE_BUF_LEN; +- info.seq_off = offsetof(struct hwrm_dbg_coredump_retrieve_input, +- seq_no); +- info.data_len_off = offsetof(struct hwrm_dbg_coredump_retrieve_output, +- data_len); +- if (buf) { +- info.dest_buf = buf + offset; +- info.buf_len = buf_len; +- info.seg_start = offset; +- } +- +- rc = bnxt_hwrm_dbg_dma_data(bp, req, &info); +- if (!rc) +- *seg_len = info.dest_buf_size; +- +- return rc; +-} +- +-static void +-bnxt_fill_coredump_seg_hdr(struct bnxt *bp, +- struct bnxt_coredump_segment_hdr *seg_hdr, +- struct coredump_segment_record *seg_rec, u32 seg_len, +- int status, u32 duration, u32 instance) +-{ +- memset(seg_hdr, 0, sizeof(*seg_hdr)); +- memcpy(seg_hdr->signature, "sEgM", 4); +- if (seg_rec) { +- seg_hdr->component_id = (__force __le32)seg_rec->component_id; +- seg_hdr->segment_id = (__force __le32)seg_rec->segment_id; +- seg_hdr->low_version = seg_rec->version_low; +- seg_hdr->high_version = seg_rec->version_hi; +- } else { +- /* For hwrm_ver_get response Component id = 2 +- * and Segment id = 0 +- */ +- seg_hdr->component_id = cpu_to_le32(2); +- seg_hdr->segment_id = 0; +- } +- seg_hdr->function_id = cpu_to_le16(bp->pdev->devfn); +- seg_hdr->length = cpu_to_le32(seg_len); +- seg_hdr->status = cpu_to_le32(status); +- seg_hdr->duration = cpu_to_le32(duration); +- seg_hdr->data_offset = cpu_to_le32(sizeof(*seg_hdr)); +- seg_hdr->instance = cpu_to_le32(instance); +-} +- +-static void +-bnxt_fill_coredump_record(struct bnxt *bp, struct bnxt_coredump_record *record, +- time64_t start, s16 start_utc, u16 total_segs, +- int status) +-{ +- time64_t end = ktime_get_real_seconds(); +- u32 os_ver_major = 0, os_ver_minor = 0; +- struct tm tm; +- +- time64_to_tm(start, 0, &tm); +- memset(record, 0, sizeof(*record)); +- memcpy(record->signature, "cOrE", 4); +- record->flags = 0; +- record->low_version = 0; +- record->high_version = 1; +- record->asic_state = 0; +- strlcpy(record->system_name, utsname()->nodename, +- sizeof(record->system_name)); +- record->year = cpu_to_le16(tm.tm_year + 1900); +- record->month = cpu_to_le16(tm.tm_mon + 1); +- record->day = cpu_to_le16(tm.tm_mday); +- record->hour = cpu_to_le16(tm.tm_hour); +- record->minute = cpu_to_le16(tm.tm_min); +- record->second = cpu_to_le16(tm.tm_sec); +- record->utc_bias = cpu_to_le16(start_utc); +- strcpy(record->commandline, "ethtool -w"); +- record->total_segments = cpu_to_le32(total_segs); +- +- sscanf(utsname()->release, "%u.%u", &os_ver_major, &os_ver_minor); +- record->os_ver_major = cpu_to_le32(os_ver_major); +- record->os_ver_minor = cpu_to_le32(os_ver_minor); +- +- strlcpy(record->os_name, utsname()->sysname, 32); +- time64_to_tm(end, 0, &tm); +- record->end_year = cpu_to_le16(tm.tm_year + 1900); +- record->end_month = cpu_to_le16(tm.tm_mon + 1); +- record->end_day = cpu_to_le16(tm.tm_mday); +- record->end_hour = cpu_to_le16(tm.tm_hour); +- record->end_minute = cpu_to_le16(tm.tm_min); +- record->end_second = cpu_to_le16(tm.tm_sec); +- record->end_utc_bias = cpu_to_le16(sys_tz.tz_minuteswest * 60); +- record->asic_id1 = cpu_to_le32(bp->chip_num << 16 | +- bp->ver_resp.chip_rev << 8 | +- bp->ver_resp.chip_metal); +- record->asic_id2 = 0; +- record->coredump_status = cpu_to_le32(status); +- record->ioctl_low_version = 0; +- record->ioctl_high_version = 0; +-} +- +-static int bnxt_get_coredump(struct bnxt *bp, void *buf, u32 *dump_len) +-{ +- u32 ver_get_resp_len = sizeof(struct hwrm_ver_get_output); +- u32 offset = 0, seg_hdr_len, seg_record_len, buf_len = 0; +- struct coredump_segment_record *seg_record = NULL; +- struct bnxt_coredump_segment_hdr seg_hdr; +- struct bnxt_coredump coredump = {NULL}; +- time64_t start_time; +- u16 start_utc; +- int rc = 0, i; +- +- if (buf) +- buf_len = *dump_len; +- +- start_time = ktime_get_real_seconds(); +- start_utc = sys_tz.tz_minuteswest * 60; +- seg_hdr_len = sizeof(seg_hdr); +- +- /* First segment should be hwrm_ver_get response */ +- *dump_len = seg_hdr_len + ver_get_resp_len; +- if (buf) { +- bnxt_fill_coredump_seg_hdr(bp, &seg_hdr, NULL, ver_get_resp_len, +- 0, 0, 0); +- memcpy(buf + offset, &seg_hdr, seg_hdr_len); +- offset += seg_hdr_len; +- memcpy(buf + offset, &bp->ver_resp, ver_get_resp_len); +- offset += ver_get_resp_len; +- } +- +- rc = bnxt_hwrm_dbg_coredump_list(bp, &coredump); +- if (rc) { +- netdev_err(bp->dev, "Failed to get coredump segment list\n"); +- goto err; +- } +- +- *dump_len += seg_hdr_len * coredump.total_segs; +- +- seg_record = (struct coredump_segment_record *)coredump.data; +- seg_record_len = sizeof(*seg_record); +- +- for (i = 0; i < coredump.total_segs; i++) { +- u16 comp_id = le16_to_cpu(seg_record->component_id); +- u16 seg_id = le16_to_cpu(seg_record->segment_id); +- u32 duration = 0, seg_len = 0; +- unsigned long start, end; +- +- if (buf && ((offset + seg_hdr_len) > +- BNXT_COREDUMP_BUF_LEN(buf_len))) { +- rc = -ENOBUFS; +- goto err; +- } +- +- start = jiffies; +- +- rc = bnxt_hwrm_dbg_coredump_initiate(bp, comp_id, seg_id); +- if (rc) { +- netdev_err(bp->dev, +- "Failed to initiate coredump for seg = %d\n", +- seg_record->segment_id); +- goto next_seg; +- } +- +- /* Write segment data into the buffer */ +- rc = bnxt_hwrm_dbg_coredump_retrieve(bp, comp_id, seg_id, +- &seg_len, buf, buf_len, +- offset + seg_hdr_len); +- if (rc && rc == -ENOBUFS) +- goto err; +- else if (rc) +- netdev_err(bp->dev, +- "Failed to retrieve coredump for seg = %d\n", +- seg_record->segment_id); +- +-next_seg: +- end = jiffies; +- duration = jiffies_to_msecs(end - start); +- bnxt_fill_coredump_seg_hdr(bp, &seg_hdr, seg_record, seg_len, +- rc, duration, 0); +- +- if (buf) { +- /* Write segment header into the buffer */ +- memcpy(buf + offset, &seg_hdr, seg_hdr_len); +- offset += seg_hdr_len + seg_len; +- } +- +- *dump_len += seg_len; +- seg_record = +- (struct coredump_segment_record *)((u8 *)seg_record + +- seg_record_len); +- } +- +-err: +- if (buf) +- bnxt_fill_coredump_record(bp, buf + offset, start_time, +- start_utc, coredump.total_segs + 1, +- rc); +- kfree(coredump.data); +- *dump_len += sizeof(struct bnxt_coredump_record); +- if (rc == -ENOBUFS) +- netdev_err(bp->dev, "Firmware returned large coredump buffer\n"); +- return rc; +-} +- + static int bnxt_set_dump(struct net_device *dev, struct ethtool_dump *dump) + { + struct bnxt *bp = netdev_priv(dev); +@@ -3971,10 +3637,7 @@ static int bnxt_get_dump_flag(struct net_device *dev, struct ethtool_dump *dump) + bp->ver_resp.hwrm_fw_rsvd_8b; + + dump->flag = bp->dump_flag; +- if (bp->dump_flag == BNXT_DUMP_CRASH) +- dump->len = BNXT_CRASH_DUMP_LEN; +- else +- bnxt_get_coredump(bp, NULL, &dump->len); ++ dump->len = bnxt_get_coredump_length(bp, bp->dump_flag); + return 0; + } + +@@ -3989,15 +3652,7 @@ static int bnxt_get_dump_data(struct net_device *dev, struct ethtool_dump *dump, + memset(buf, 0, dump->len); + + dump->flag = bp->dump_flag; +- if (dump->flag == BNXT_DUMP_CRASH) { +-#ifdef CONFIG_TEE_BNXT_FW +- return tee_bnxt_copy_coredump(buf, 0, dump->len); +-#endif +- } else { +- return bnxt_get_coredump(bp, buf, &dump->len); +- } +- +- return 0; ++ return bnxt_get_coredump(bp, dump->flag, buf, &dump->len); + } + + static int bnxt_get_ts_info(struct net_device *dev, +diff --git a/drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.h b/drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.h +index 0a57cb6a4a4bf..11a719f98defd 100644 +--- a/drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.h ++++ b/drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.h +@@ -22,49 +22,6 @@ struct bnxt_led_cfg { + u8 rsvd; + }; + +-#define COREDUMP_LIST_BUF_LEN 2048 +-#define COREDUMP_RETRIEVE_BUF_LEN 4096 +- +-struct bnxt_coredump { +- void *data; +- int data_size; +- u16 total_segs; +-}; +- +-#define BNXT_COREDUMP_BUF_LEN(len) ((len) - sizeof(struct bnxt_coredump_record)) +- +-struct bnxt_hwrm_dbg_dma_info { +- void *dest_buf; +- int dest_buf_size; +- u16 dma_len; +- u16 seq_off; +- u16 data_len_off; +- u16 segs; +- u32 seg_start; +- u32 buf_len; +-}; +- +-struct hwrm_dbg_cmn_input { +- __le16 req_type; +- __le16 cmpl_ring; +- __le16 seq_id; +- __le16 target_id; +- __le64 resp_addr; +- __le64 host_dest_addr; +- __le32 host_buf_len; +-}; +- +-struct hwrm_dbg_cmn_output { +- __le16 error_code; +- __le16 req_type; +- __le16 seq_id; +- __le16 resp_len; +- u8 flags; +- #define HWRM_DBG_CMN_FLAGS_MORE 1 +-}; +- +-#define BNXT_CRASH_DUMP_LEN (8 << 20) +- + #define BNXT_LED_DFLT_ENA \ + (PORT_LED_CFG_REQ_ENABLES_LED0_ID | \ + PORT_LED_CFG_REQ_ENABLES_LED0_STATE | \ +diff --git a/drivers/net/ethernet/broadcom/bnxt/bnxt_hwrm.c b/drivers/net/ethernet/broadcom/bnxt/bnxt_hwrm.c +index bb7327b82d0b2..8171f4912fa01 100644 +--- a/drivers/net/ethernet/broadcom/bnxt/bnxt_hwrm.c ++++ b/drivers/net/ethernet/broadcom/bnxt/bnxt_hwrm.c +@@ -496,7 +496,7 @@ static int __hwrm_send(struct bnxt *bp, struct bnxt_hwrm_ctx *ctx) + } + + /* Limit timeout to an upper limit */ +- timeout = min_t(uint, ctx->timeout, HWRM_CMD_MAX_TIMEOUT); ++ timeout = min(ctx->timeout, bp->hwrm_cmd_max_timeout ?: HWRM_CMD_MAX_TIMEOUT); + /* convert timeout to usec */ + timeout *= 1000; + +diff --git a/drivers/net/ethernet/broadcom/bnxt/bnxt_hwrm.h b/drivers/net/ethernet/broadcom/bnxt/bnxt_hwrm.h +index 4d17f0d5363bb..9a9fc4e8041b6 100644 +--- a/drivers/net/ethernet/broadcom/bnxt/bnxt_hwrm.h ++++ b/drivers/net/ethernet/broadcom/bnxt/bnxt_hwrm.h +@@ -58,11 +58,10 @@ void hwrm_update_token(struct bnxt *bp, u16 seq, enum bnxt_hwrm_wait_state s); + + #define BNXT_HWRM_MAX_REQ_LEN (bp->hwrm_max_req_len) + #define BNXT_HWRM_SHORT_REQ_LEN sizeof(struct hwrm_short_input) +-#define HWRM_CMD_MAX_TIMEOUT 40000 ++#define HWRM_CMD_MAX_TIMEOUT 40000U + #define SHORT_HWRM_CMD_TIMEOUT 20 + #define HWRM_CMD_TIMEOUT (bp->hwrm_cmd_timeout) + #define HWRM_RESET_TIMEOUT ((HWRM_CMD_TIMEOUT) * 4) +-#define HWRM_COREDUMP_TIMEOUT ((HWRM_CMD_TIMEOUT) * 12) + #define BNXT_HWRM_TARGET 0xffff + #define BNXT_HWRM_NO_CMPL_RING -1 + #define BNXT_HWRM_REQ_MAX_SIZE 128 +diff --git a/drivers/net/ethernet/broadcom/genet/bcmgenet.c b/drivers/net/ethernet/broadcom/genet/bcmgenet.c +index 23c7595d2a1d3..9a67e24f46b27 100644 +--- a/drivers/net/ethernet/broadcom/genet/bcmgenet.c ++++ b/drivers/net/ethernet/broadcom/genet/bcmgenet.c +@@ -3966,10 +3966,12 @@ static int bcmgenet_probe(struct platform_device *pdev) + + /* Request the WOL interrupt and advertise suspend if available */ + priv->wol_irq_disabled = true; +- err = devm_request_irq(&pdev->dev, priv->wol_irq, bcmgenet_wol_isr, 0, +- dev->name, priv); +- if (!err) +- device_set_wakeup_capable(&pdev->dev, 1); ++ if (priv->wol_irq > 0) { ++ err = devm_request_irq(&pdev->dev, priv->wol_irq, ++ bcmgenet_wol_isr, 0, dev->name, priv); ++ if (!err) ++ device_set_wakeup_capable(&pdev->dev, 1); ++ } + + /* Set the needed headroom to account for any possible + * features enabling/disabling at runtime +diff --git a/drivers/net/ethernet/chelsio/libcxgb/libcxgb_cm.c b/drivers/net/ethernet/chelsio/libcxgb/libcxgb_cm.c +index d04a6c1634452..da8d10475a08e 100644 +--- a/drivers/net/ethernet/chelsio/libcxgb/libcxgb_cm.c ++++ b/drivers/net/ethernet/chelsio/libcxgb/libcxgb_cm.c +@@ -32,6 +32,7 @@ + + #include <linux/tcp.h> + #include <linux/ipv6.h> ++#include <net/inet_ecn.h> + #include <net/route.h> + #include <net/ip6_route.h> + +@@ -99,7 +100,7 @@ cxgb_find_route(struct cxgb4_lld_info *lldi, + + rt = ip_route_output_ports(&init_net, &fl4, NULL, peer_ip, local_ip, + peer_port, local_port, IPPROTO_TCP, +- tos, 0); ++ tos & ~INET_ECN_MASK, 0); + if (IS_ERR(rt)) + return NULL; + n = dst_neigh_lookup(&rt->dst, &peer_ip); +diff --git a/drivers/net/ethernet/cortina/gemini.c b/drivers/net/ethernet/cortina/gemini.c +index 6e745ca4c4333..012ca11a38cc1 100644 +--- a/drivers/net/ethernet/cortina/gemini.c ++++ b/drivers/net/ethernet/cortina/gemini.c +@@ -305,21 +305,21 @@ static void gmac_speed_set(struct net_device *netdev) + switch (phydev->speed) { + case 1000: + status.bits.speed = GMAC_SPEED_1000; +- if (phydev->interface == PHY_INTERFACE_MODE_RGMII) ++ if (phy_interface_mode_is_rgmii(phydev->interface)) + status.bits.mii_rmii = GMAC_PHY_RGMII_1000; + netdev_dbg(netdev, "connect %s to RGMII @ 1Gbit\n", + phydev_name(phydev)); + break; + case 100: + status.bits.speed = GMAC_SPEED_100; +- if (phydev->interface == PHY_INTERFACE_MODE_RGMII) ++ if (phy_interface_mode_is_rgmii(phydev->interface)) + status.bits.mii_rmii = GMAC_PHY_RGMII_100_10; + netdev_dbg(netdev, "connect %s to RGMII @ 100 Mbit\n", + phydev_name(phydev)); + break; + case 10: + status.bits.speed = GMAC_SPEED_10; +- if (phydev->interface == PHY_INTERFACE_MODE_RGMII) ++ if (phy_interface_mode_is_rgmii(phydev->interface)) + status.bits.mii_rmii = GMAC_PHY_RGMII_100_10; + netdev_dbg(netdev, "connect %s to RGMII @ 10 Mbit\n", + phydev_name(phydev)); +@@ -389,6 +389,9 @@ static int gmac_setup_phy(struct net_device *netdev) + status.bits.mii_rmii = GMAC_PHY_GMII; + break; + case PHY_INTERFACE_MODE_RGMII: ++ case PHY_INTERFACE_MODE_RGMII_ID: ++ case PHY_INTERFACE_MODE_RGMII_TXID: ++ case PHY_INTERFACE_MODE_RGMII_RXID: + netdev_dbg(netdev, + "RGMII: set GMAC0 and GMAC1 to MII/RGMII mode\n"); + status.bits.mii_rmii = GMAC_PHY_RGMII_100_10; +diff --git a/drivers/net/ethernet/freescale/fman/mac.c b/drivers/net/ethernet/freescale/fman/mac.c +index d9fc5c456bf3e..39ae965cd4f64 100644 +--- a/drivers/net/ethernet/freescale/fman/mac.c ++++ b/drivers/net/ethernet/freescale/fman/mac.c +@@ -94,14 +94,17 @@ static void mac_exception(void *handle, enum fman_mac_exceptions ex) + __func__, ex); + } + +-static void set_fman_mac_params(struct mac_device *mac_dev, +- struct fman_mac_params *params) ++static int set_fman_mac_params(struct mac_device *mac_dev, ++ struct fman_mac_params *params) + { + struct mac_priv_s *priv = mac_dev->priv; + + params->base_addr = (typeof(params->base_addr)) + devm_ioremap(priv->dev, mac_dev->res->start, + resource_size(mac_dev->res)); ++ if (!params->base_addr) ++ return -ENOMEM; ++ + memcpy(¶ms->addr, mac_dev->addr, sizeof(mac_dev->addr)); + params->max_speed = priv->max_speed; + params->phy_if = mac_dev->phy_if; +@@ -112,6 +115,8 @@ static void set_fman_mac_params(struct mac_device *mac_dev, + params->event_cb = mac_exception; + params->dev_id = mac_dev; + params->internal_phy_node = priv->internal_phy_node; ++ ++ return 0; + } + + static int tgec_initialization(struct mac_device *mac_dev) +@@ -123,7 +128,9 @@ static int tgec_initialization(struct mac_device *mac_dev) + + priv = mac_dev->priv; + +- set_fman_mac_params(mac_dev, ¶ms); ++ err = set_fman_mac_params(mac_dev, ¶ms); ++ if (err) ++ goto _return; + + mac_dev->fman_mac = tgec_config(¶ms); + if (!mac_dev->fman_mac) { +@@ -169,7 +176,9 @@ static int dtsec_initialization(struct mac_device *mac_dev) + + priv = mac_dev->priv; + +- set_fman_mac_params(mac_dev, ¶ms); ++ err = set_fman_mac_params(mac_dev, ¶ms); ++ if (err) ++ goto _return; + + mac_dev->fman_mac = dtsec_config(¶ms); + if (!mac_dev->fman_mac) { +@@ -218,7 +227,9 @@ static int memac_initialization(struct mac_device *mac_dev) + + priv = mac_dev->priv; + +- set_fman_mac_params(mac_dev, ¶ms); ++ err = set_fman_mac_params(mac_dev, ¶ms); ++ if (err) ++ goto _return; + + if (priv->max_speed == SPEED_10000) + params.phy_if = PHY_INTERFACE_MODE_XGMII; +diff --git a/drivers/net/ethernet/freescale/xgmac_mdio.c b/drivers/net/ethernet/freescale/xgmac_mdio.c +index 0b68852379da5..27d07468765f0 100644 +--- a/drivers/net/ethernet/freescale/xgmac_mdio.c ++++ b/drivers/net/ethernet/freescale/xgmac_mdio.c +@@ -52,6 +52,7 @@ struct tgec_mdio_controller { + struct mdio_fsl_priv { + struct tgec_mdio_controller __iomem *mdio_base; + bool is_little_endian; ++ bool has_a009885; + bool has_a011043; + }; + +@@ -187,10 +188,10 @@ static int xgmac_mdio_read(struct mii_bus *bus, int phy_id, int regnum) + { + struct mdio_fsl_priv *priv = (struct mdio_fsl_priv *)bus->priv; + struct tgec_mdio_controller __iomem *regs = priv->mdio_base; ++ unsigned long flags; + uint16_t dev_addr; + uint32_t mdio_stat; + uint32_t mdio_ctl; +- uint16_t value; + int ret; + bool endian = priv->is_little_endian; + +@@ -222,12 +223,18 @@ static int xgmac_mdio_read(struct mii_bus *bus, int phy_id, int regnum) + return ret; + } + ++ if (priv->has_a009885) ++ /* Once the operation completes, i.e. MDIO_STAT_BSY clears, we ++ * must read back the data register within 16 MDC cycles. ++ */ ++ local_irq_save(flags); ++ + /* Initiate the read */ + xgmac_write32(mdio_ctl | MDIO_CTL_READ, ®s->mdio_ctl, endian); + + ret = xgmac_wait_until_done(&bus->dev, regs, endian); + if (ret) +- return ret; ++ goto irq_restore; + + /* Return all Fs if nothing was there */ + if ((xgmac_read32(®s->mdio_stat, endian) & MDIO_STAT_RD_ER) && +@@ -235,13 +242,17 @@ static int xgmac_mdio_read(struct mii_bus *bus, int phy_id, int regnum) + dev_dbg(&bus->dev, + "Error while reading PHY%d reg at %d.%hhu\n", + phy_id, dev_addr, regnum); +- return 0xffff; ++ ret = 0xffff; ++ } else { ++ ret = xgmac_read32(®s->mdio_data, endian) & 0xffff; ++ dev_dbg(&bus->dev, "read %04x\n", ret); + } + +- value = xgmac_read32(®s->mdio_data, endian) & 0xffff; +- dev_dbg(&bus->dev, "read %04x\n", value); ++irq_restore: ++ if (priv->has_a009885) ++ local_irq_restore(flags); + +- return value; ++ return ret; + } + + static int xgmac_mdio_probe(struct platform_device *pdev) +@@ -288,6 +299,8 @@ static int xgmac_mdio_probe(struct platform_device *pdev) + priv->is_little_endian = device_property_read_bool(&pdev->dev, + "little-endian"); + ++ priv->has_a009885 = device_property_read_bool(&pdev->dev, ++ "fsl,erratum-a009885"); + priv->has_a011043 = device_property_read_bool(&pdev->dev, + "fsl,erratum-a011043"); + +@@ -319,9 +332,10 @@ err_ioremap: + static int xgmac_mdio_remove(struct platform_device *pdev) + { + struct mii_bus *bus = platform_get_drvdata(pdev); ++ struct mdio_fsl_priv *priv = bus->priv; + + mdiobus_unregister(bus); +- iounmap(bus->priv); ++ iounmap(priv->mdio_base); + mdiobus_free(bus); + + return 0; +diff --git a/drivers/net/ethernet/i825xx/sni_82596.c b/drivers/net/ethernet/i825xx/sni_82596.c +index 27937c5d79567..daec9ce04531b 100644 +--- a/drivers/net/ethernet/i825xx/sni_82596.c ++++ b/drivers/net/ethernet/i825xx/sni_82596.c +@@ -117,9 +117,10 @@ static int sni_82596_probe(struct platform_device *dev) + netdevice->dev_addr[5] = readb(eth_addr + 0x06); + iounmap(eth_addr); + +- if (!netdevice->irq) { ++ if (netdevice->irq < 0) { + printk(KERN_ERR "%s: IRQ not found for i82596 at 0x%lx\n", + __FILE__, netdevice->base_addr); ++ retval = netdevice->irq; + goto probe_failed; + } + +diff --git a/drivers/net/ethernet/intel/igc/igc_main.c b/drivers/net/ethernet/intel/igc/igc_main.c +index 0a96627391a8c..c7fa978cdf02e 100644 +--- a/drivers/net/ethernet/intel/igc/igc_main.c ++++ b/drivers/net/ethernet/intel/igc/igc_main.c +@@ -2447,8 +2447,10 @@ static struct sk_buff *igc_construct_skb_zc(struct igc_ring *ring, + + skb_reserve(skb, xdp->data_meta - xdp->data_hard_start); + memcpy(__skb_put(skb, totalsize), xdp->data_meta, totalsize); +- if (metasize) ++ if (metasize) { + skb_metadata_set(skb, metasize); ++ __skb_pull(skb, metasize); ++ } + + return skb; + } +diff --git a/drivers/net/ethernet/marvell/octeontx2/af/ptp.c b/drivers/net/ethernet/marvell/octeontx2/af/ptp.c +index 9b8e59f4c206d..77cb52b80c60f 100644 +--- a/drivers/net/ethernet/marvell/octeontx2/af/ptp.c ++++ b/drivers/net/ethernet/marvell/octeontx2/af/ptp.c +@@ -85,6 +85,8 @@ struct ptp *ptp_get(void) + /* Check driver is bound to PTP block */ + if (!ptp) + ptp = ERR_PTR(-EPROBE_DEFER); ++ else ++ pci_dev_get(ptp->pdev); + + return ptp; + } +diff --git a/drivers/net/ethernet/mediatek/mtk_eth_soc.c b/drivers/net/ethernet/mediatek/mtk_eth_soc.c +index 398c23cec8151..ee1fd472e9252 100644 +--- a/drivers/net/ethernet/mediatek/mtk_eth_soc.c ++++ b/drivers/net/ethernet/mediatek/mtk_eth_soc.c +@@ -91,46 +91,53 @@ static int mtk_mdio_busy_wait(struct mtk_eth *eth) + } + + dev_err(eth->dev, "mdio: MDIO timeout\n"); +- return -1; ++ return -ETIMEDOUT; + } + +-static u32 _mtk_mdio_write(struct mtk_eth *eth, u32 phy_addr, +- u32 phy_register, u32 write_data) ++static int _mtk_mdio_write(struct mtk_eth *eth, u32 phy_addr, u32 phy_reg, ++ u32 write_data) + { +- if (mtk_mdio_busy_wait(eth)) +- return -1; ++ int ret; + +- write_data &= 0xffff; ++ ret = mtk_mdio_busy_wait(eth); ++ if (ret < 0) ++ return ret; + +- mtk_w32(eth, PHY_IAC_ACCESS | PHY_IAC_START | PHY_IAC_WRITE | +- (phy_register << PHY_IAC_REG_SHIFT) | +- (phy_addr << PHY_IAC_ADDR_SHIFT) | write_data, ++ mtk_w32(eth, PHY_IAC_ACCESS | ++ PHY_IAC_START_C22 | ++ PHY_IAC_CMD_WRITE | ++ PHY_IAC_REG(phy_reg) | ++ PHY_IAC_ADDR(phy_addr) | ++ PHY_IAC_DATA(write_data), + MTK_PHY_IAC); + +- if (mtk_mdio_busy_wait(eth)) +- return -1; ++ ret = mtk_mdio_busy_wait(eth); ++ if (ret < 0) ++ return ret; + + return 0; + } + +-static u32 _mtk_mdio_read(struct mtk_eth *eth, int phy_addr, int phy_reg) ++static int _mtk_mdio_read(struct mtk_eth *eth, u32 phy_addr, u32 phy_reg) + { +- u32 d; ++ int ret; + +- if (mtk_mdio_busy_wait(eth)) +- return 0xffff; ++ ret = mtk_mdio_busy_wait(eth); ++ if (ret < 0) ++ return ret; + +- mtk_w32(eth, PHY_IAC_ACCESS | PHY_IAC_START | PHY_IAC_READ | +- (phy_reg << PHY_IAC_REG_SHIFT) | +- (phy_addr << PHY_IAC_ADDR_SHIFT), ++ mtk_w32(eth, PHY_IAC_ACCESS | ++ PHY_IAC_START_C22 | ++ PHY_IAC_CMD_C22_READ | ++ PHY_IAC_REG(phy_reg) | ++ PHY_IAC_ADDR(phy_addr), + MTK_PHY_IAC); + +- if (mtk_mdio_busy_wait(eth)) +- return 0xffff; +- +- d = mtk_r32(eth, MTK_PHY_IAC) & 0xffff; ++ ret = mtk_mdio_busy_wait(eth); ++ if (ret < 0) ++ return ret; + +- return d; ++ return mtk_r32(eth, MTK_PHY_IAC) & PHY_IAC_DATA_MASK; + } + + static int mtk_mdio_write(struct mii_bus *bus, int phy_addr, +@@ -217,7 +224,7 @@ static void mtk_mac_config(struct phylink_config *config, unsigned int mode, + phylink_config); + struct mtk_eth *eth = mac->hw; + u32 mcr_cur, mcr_new, sid, i; +- int val, ge_mode, err; ++ int val, ge_mode, err = 0; + + /* MT76x8 has no hardware settings between for the MAC */ + if (!MTK_HAS_CAPS(eth->soc->caps, MTK_SOC_MT7628) && +diff --git a/drivers/net/ethernet/mediatek/mtk_eth_soc.h b/drivers/net/ethernet/mediatek/mtk_eth_soc.h +index 5ef70dd8b49c6..f2d90639d7ed1 100644 +--- a/drivers/net/ethernet/mediatek/mtk_eth_soc.h ++++ b/drivers/net/ethernet/mediatek/mtk_eth_soc.h +@@ -341,11 +341,17 @@ + /* PHY Indirect Access Control registers */ + #define MTK_PHY_IAC 0x10004 + #define PHY_IAC_ACCESS BIT(31) +-#define PHY_IAC_READ BIT(19) +-#define PHY_IAC_WRITE BIT(18) +-#define PHY_IAC_START BIT(16) +-#define PHY_IAC_ADDR_SHIFT 20 +-#define PHY_IAC_REG_SHIFT 25 ++#define PHY_IAC_REG_MASK GENMASK(29, 25) ++#define PHY_IAC_REG(x) FIELD_PREP(PHY_IAC_REG_MASK, (x)) ++#define PHY_IAC_ADDR_MASK GENMASK(24, 20) ++#define PHY_IAC_ADDR(x) FIELD_PREP(PHY_IAC_ADDR_MASK, (x)) ++#define PHY_IAC_CMD_MASK GENMASK(19, 18) ++#define PHY_IAC_CMD_WRITE FIELD_PREP(PHY_IAC_CMD_MASK, 1) ++#define PHY_IAC_CMD_C22_READ FIELD_PREP(PHY_IAC_CMD_MASK, 2) ++#define PHY_IAC_START_MASK GENMASK(17, 16) ++#define PHY_IAC_START_C22 FIELD_PREP(PHY_IAC_START_MASK, 1) ++#define PHY_IAC_DATA_MASK GENMASK(15, 0) ++#define PHY_IAC_DATA(x) FIELD_PREP(PHY_IAC_DATA_MASK, (x)) + #define PHY_IAC_TIMEOUT HZ + + #define MTK_MAC_MISC 0x1000c +diff --git a/drivers/net/ethernet/mellanox/mlx5/core/cmd.c b/drivers/net/ethernet/mellanox/mlx5/core/cmd.c +index bea35530c2d0b..00f63fbfe9b48 100644 +--- a/drivers/net/ethernet/mellanox/mlx5/core/cmd.c ++++ b/drivers/net/ethernet/mellanox/mlx5/core/cmd.c +@@ -147,8 +147,12 @@ static void cmd_ent_put(struct mlx5_cmd_work_ent *ent) + if (!refcount_dec_and_test(&ent->refcnt)) + return; + +- if (ent->idx >= 0) +- cmd_free_index(ent->cmd, ent->idx); ++ if (ent->idx >= 0) { ++ struct mlx5_cmd *cmd = ent->cmd; ++ ++ cmd_free_index(cmd, ent->idx); ++ up(ent->page_queue ? &cmd->pages_sem : &cmd->sem); ++ } + + cmd_free_ent(ent); + } +@@ -895,25 +899,6 @@ static bool opcode_allowed(struct mlx5_cmd *cmd, u16 opcode) + return cmd->allowed_opcode == opcode; + } + +-static int cmd_alloc_index_retry(struct mlx5_cmd *cmd) +-{ +- unsigned long alloc_end = jiffies + msecs_to_jiffies(1000); +- int idx; +- +-retry: +- idx = cmd_alloc_index(cmd); +- if (idx < 0 && time_before(jiffies, alloc_end)) { +- /* Index allocation can fail on heavy load of commands. This is a temporary +- * situation as the current command already holds the semaphore, meaning that +- * another command completion is being handled and it is expected to release +- * the entry index soon. +- */ +- cpu_relax(); +- goto retry; +- } +- return idx; +-} +- + bool mlx5_cmd_is_down(struct mlx5_core_dev *dev) + { + return pci_channel_offline(dev->pdev) || +@@ -938,7 +923,7 @@ static void cmd_work_handler(struct work_struct *work) + sem = ent->page_queue ? &cmd->pages_sem : &cmd->sem; + down(sem); + if (!ent->page_queue) { +- alloc_ret = cmd_alloc_index_retry(cmd); ++ alloc_ret = cmd_alloc_index(cmd); + if (alloc_ret < 0) { + mlx5_core_err_rl(dev, "failed to allocate command entry\n"); + if (ent->callback) { +@@ -1594,8 +1579,6 @@ static void mlx5_cmd_comp_handler(struct mlx5_core_dev *dev, u64 vec, bool force + vector = vec & 0xffffffff; + for (i = 0; i < (1 << cmd->log_sz); i++) { + if (test_bit(i, &vector)) { +- struct semaphore *sem; +- + ent = cmd->ent_arr[i]; + + /* if we already completed the command, ignore it */ +@@ -1618,10 +1601,6 @@ static void mlx5_cmd_comp_handler(struct mlx5_core_dev *dev, u64 vec, bool force + dev->state == MLX5_DEVICE_STATE_INTERNAL_ERROR) + cmd_ent_put(ent); + +- if (ent->page_queue) +- sem = &cmd->pages_sem; +- else +- sem = &cmd->sem; + ent->ts2 = ktime_get_ns(); + memcpy(ent->out->first.data, ent->lay->out, sizeof(ent->lay->out)); + dump_command(dev, ent, 0); +@@ -1675,7 +1654,6 @@ static void mlx5_cmd_comp_handler(struct mlx5_core_dev *dev, u64 vec, bool force + */ + complete(&ent->done); + } +- up(sem); + } + } + } +diff --git a/drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c b/drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c +index 4a13ef561587d..cf03297c84710 100644 +--- a/drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c ++++ b/drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c +@@ -1,6 +1,7 @@ + /* SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB */ + /* Copyright (c) 2018 Mellanox Technologies. */ + ++#include <net/inet_ecn.h> + #include <net/vxlan.h> + #include <net/gre.h> + #include <net/geneve.h> +@@ -229,7 +230,7 @@ int mlx5e_tc_tun_create_header_ipv4(struct mlx5e_priv *priv, + int err; + + /* add the IP fields */ +- attr.fl.fl4.flowi4_tos = tun_key->tos; ++ attr.fl.fl4.flowi4_tos = tun_key->tos & ~INET_ECN_MASK; + attr.fl.fl4.daddr = tun_key->u.ipv4.dst; + attr.fl.fl4.saddr = tun_key->u.ipv4.src; + attr.ttl = tun_key->ttl; +@@ -344,7 +345,7 @@ int mlx5e_tc_tun_update_header_ipv4(struct mlx5e_priv *priv, + int err; + + /* add the IP fields */ +- attr.fl.fl4.flowi4_tos = tun_key->tos; ++ attr.fl.fl4.flowi4_tos = tun_key->tos & ~INET_ECN_MASK; + attr.fl.fl4.daddr = tun_key->u.ipv4.dst; + attr.fl.fl4.saddr = tun_key->u.ipv4.src; + attr.ttl = tun_key->ttl; +diff --git a/drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c b/drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c +index ec0163d75dd25..700c463ea3679 100644 +--- a/drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c ++++ b/drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c +@@ -1544,6 +1544,8 @@ mlx5e_init_fib_work_ipv4(struct mlx5e_priv *priv, + struct net_device *fib_dev; + + fen_info = container_of(info, struct fib_entry_notifier_info, info); ++ if (fen_info->fi->nh) ++ return NULL; + fib_dev = fib_info_nh(fen_info->fi, 0)->fib_nh_dev; + if (!fib_dev || fib_dev->netdev_ops != &mlx5e_netdev_ops || + fen_info->dst_len != 32) +diff --git a/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/pool.c b/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/pool.c +index 7b562d2c8a196..279cd8f4e79f7 100644 +--- a/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/pool.c ++++ b/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/pool.c +@@ -11,13 +11,13 @@ static int mlx5e_xsk_map_pool(struct mlx5e_priv *priv, + { + struct device *dev = mlx5_core_dma_dev(priv->mdev); + +- return xsk_pool_dma_map(pool, dev, 0); ++ return xsk_pool_dma_map(pool, dev, DMA_ATTR_SKIP_CPU_SYNC); + } + + static void mlx5e_xsk_unmap_pool(struct mlx5e_priv *priv, + struct xsk_buff_pool *pool) + { +- return xsk_pool_dma_unmap(pool, 0); ++ return xsk_pool_dma_unmap(pool, DMA_ATTR_SKIP_CPU_SYNC); + } + + static int mlx5e_xsk_get_pools(struct mlx5e_xsk *xsk) +diff --git a/drivers/net/ethernet/mellanox/mlx5/core/en_main.c b/drivers/net/ethernet/mellanox/mlx5/core/en_main.c +index baa0d7d48fc0c..f075bb8ccd00d 100644 +--- a/drivers/net/ethernet/mellanox/mlx5/core/en_main.c ++++ b/drivers/net/ethernet/mellanox/mlx5/core/en_main.c +@@ -4485,15 +4485,22 @@ static void mlx5e_build_nic_netdev(struct net_device *netdev) + } + + if (mlx5_vxlan_allowed(mdev->vxlan) || mlx5_geneve_tx_allowed(mdev)) { +- netdev->hw_features |= NETIF_F_GSO_UDP_TUNNEL; +- netdev->hw_enc_features |= NETIF_F_GSO_UDP_TUNNEL; +- netdev->vlan_features |= NETIF_F_GSO_UDP_TUNNEL; ++ netdev->hw_features |= NETIF_F_GSO_UDP_TUNNEL | ++ NETIF_F_GSO_UDP_TUNNEL_CSUM; ++ netdev->hw_enc_features |= NETIF_F_GSO_UDP_TUNNEL | ++ NETIF_F_GSO_UDP_TUNNEL_CSUM; ++ netdev->gso_partial_features = NETIF_F_GSO_UDP_TUNNEL_CSUM; ++ netdev->vlan_features |= NETIF_F_GSO_UDP_TUNNEL | ++ NETIF_F_GSO_UDP_TUNNEL_CSUM; + } + + if (mlx5e_tunnel_proto_supported_tx(mdev, IPPROTO_GRE)) { +- netdev->hw_features |= NETIF_F_GSO_GRE; +- netdev->hw_enc_features |= NETIF_F_GSO_GRE; +- netdev->gso_partial_features |= NETIF_F_GSO_GRE; ++ netdev->hw_features |= NETIF_F_GSO_GRE | ++ NETIF_F_GSO_GRE_CSUM; ++ netdev->hw_enc_features |= NETIF_F_GSO_GRE | ++ NETIF_F_GSO_GRE_CSUM; ++ netdev->gso_partial_features |= NETIF_F_GSO_GRE | ++ NETIF_F_GSO_GRE_CSUM; + } + + if (mlx5e_tunnel_proto_supported_tx(mdev, IPPROTO_IPIP)) { +diff --git a/drivers/net/ethernet/mellanox/mlx5/core/en_rep.c b/drivers/net/ethernet/mellanox/mlx5/core/en_rep.c +index edecd149dcab3..161b60e1139b3 100644 +--- a/drivers/net/ethernet/mellanox/mlx5/core/en_rep.c ++++ b/drivers/net/ethernet/mellanox/mlx5/core/en_rep.c +@@ -50,6 +50,7 @@ + #include "fs_core.h" + #include "lib/mlx5.h" + #include "lib/devcom.h" ++#include "lib/vxlan.h" + #define CREATE_TRACE_POINTS + #include "diag/en_rep_tracepoint.h" + #include "en_accel/ipsec.h" +@@ -1016,6 +1017,7 @@ static void mlx5e_uplink_rep_enable(struct mlx5e_priv *priv) + rtnl_lock(); + if (netif_running(netdev)) + mlx5e_open(netdev); ++ udp_tunnel_nic_reset_ntf(priv->netdev); + netif_device_attach(netdev); + rtnl_unlock(); + } +@@ -1037,6 +1039,7 @@ static void mlx5e_uplink_rep_disable(struct mlx5e_priv *priv) + mlx5_notifier_unregister(mdev, &priv->events_nb); + mlx5e_rep_tc_disable(priv); + mlx5_lag_remove_netdev(mdev, priv->netdev); ++ mlx5_vxlan_reset_to_default(mdev->vxlan); + } + + static MLX5E_DEFINE_STATS_GRP(sw_rep, 0); +diff --git a/drivers/net/ethernet/mellanox/mlx5/core/en_rx.c b/drivers/net/ethernet/mellanox/mlx5/core/en_rx.c +index 29a6586ef28dc..0015545d5235b 100644 +--- a/drivers/net/ethernet/mellanox/mlx5/core/en_rx.c ++++ b/drivers/net/ethernet/mellanox/mlx5/core/en_rx.c +@@ -271,8 +271,8 @@ static inline int mlx5e_page_alloc_pool(struct mlx5e_rq *rq, + if (unlikely(!dma_info->page)) + return -ENOMEM; + +- dma_info->addr = dma_map_page(rq->pdev, dma_info->page, 0, +- PAGE_SIZE, rq->buff.map_dir); ++ dma_info->addr = dma_map_page_attrs(rq->pdev, dma_info->page, 0, PAGE_SIZE, ++ rq->buff.map_dir, DMA_ATTR_SKIP_CPU_SYNC); + if (unlikely(dma_mapping_error(rq->pdev, dma_info->addr))) { + page_pool_recycle_direct(rq->page_pool, dma_info->page); + dma_info->page = NULL; +@@ -293,7 +293,8 @@ static inline int mlx5e_page_alloc(struct mlx5e_rq *rq, + + void mlx5e_page_dma_unmap(struct mlx5e_rq *rq, struct mlx5e_dma_info *dma_info) + { +- dma_unmap_page(rq->pdev, dma_info->addr, PAGE_SIZE, rq->buff.map_dir); ++ dma_unmap_page_attrs(rq->pdev, dma_info->addr, PAGE_SIZE, rq->buff.map_dir, ++ DMA_ATTR_SKIP_CPU_SYNC); + } + + void mlx5e_page_release_dynamic(struct mlx5e_rq *rq, +diff --git a/drivers/net/ethernet/mellanox/mlx5/core/en_tc.c b/drivers/net/ethernet/mellanox/mlx5/core/en_tc.c +index fa461bc57baee..8b041deb25e5f 100644 +--- a/drivers/net/ethernet/mellanox/mlx5/core/en_tc.c ++++ b/drivers/net/ethernet/mellanox/mlx5/core/en_tc.c +@@ -1874,6 +1874,111 @@ u8 mlx5e_tc_get_ip_version(struct mlx5_flow_spec *spec, bool outer) + return ip_version; + } + ++/* Tunnel device follows RFC 6040, see include/net/inet_ecn.h. ++ * And changes inner ip_ecn depending on inner and outer ip_ecn as follows: ++ * +---------+----------------------------------------+ ++ * |Arriving | Arriving Outer Header | ++ * | Inner +---------+---------+---------+----------+ ++ * | Header | Not-ECT | ECT(0) | ECT(1) | CE | ++ * +---------+---------+---------+---------+----------+ ++ * | Not-ECT | Not-ECT | Not-ECT | Not-ECT | <drop> | ++ * | ECT(0) | ECT(0) | ECT(0) | ECT(1) | CE* | ++ * | ECT(1) | ECT(1) | ECT(1) | ECT(1)* | CE* | ++ * | CE | CE | CE | CE | CE | ++ * +---------+---------+---------+---------+----------+ ++ * ++ * Tc matches on inner after decapsulation on tunnel device, but hw offload matches ++ * the inner ip_ecn value before hardware decap action. ++ * ++ * Cells marked are changed from original inner packet ip_ecn value during decap, and ++ * so matching those values on inner ip_ecn before decap will fail. ++ * ++ * The following helper allows offload when inner ip_ecn won't be changed by outer ip_ecn, ++ * except for the outer ip_ecn = CE, where in all cases inner ip_ecn will be changed to CE, ++ * and such we can drop the inner ip_ecn=CE match. ++ */ ++ ++static int mlx5e_tc_verify_tunnel_ecn(struct mlx5e_priv *priv, ++ struct flow_cls_offload *f, ++ bool *match_inner_ecn) ++{ ++ u8 outer_ecn_mask = 0, outer_ecn_key = 0, inner_ecn_mask = 0, inner_ecn_key = 0; ++ struct flow_rule *rule = flow_cls_offload_flow_rule(f); ++ struct netlink_ext_ack *extack = f->common.extack; ++ struct flow_match_ip match; ++ ++ *match_inner_ecn = true; ++ ++ if (flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_ENC_IP)) { ++ flow_rule_match_enc_ip(rule, &match); ++ outer_ecn_key = match.key->tos & INET_ECN_MASK; ++ outer_ecn_mask = match.mask->tos & INET_ECN_MASK; ++ } ++ ++ if (flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_IP)) { ++ flow_rule_match_ip(rule, &match); ++ inner_ecn_key = match.key->tos & INET_ECN_MASK; ++ inner_ecn_mask = match.mask->tos & INET_ECN_MASK; ++ } ++ ++ if (outer_ecn_mask != 0 && outer_ecn_mask != INET_ECN_MASK) { ++ NL_SET_ERR_MSG_MOD(extack, "Partial match on enc_tos ecn bits isn't supported"); ++ netdev_warn(priv->netdev, "Partial match on enc_tos ecn bits isn't supported"); ++ return -EOPNOTSUPP; ++ } ++ ++ if (!outer_ecn_mask) { ++ if (!inner_ecn_mask) ++ return 0; ++ ++ NL_SET_ERR_MSG_MOD(extack, ++ "Matching on tos ecn bits without also matching enc_tos ecn bits isn't supported"); ++ netdev_warn(priv->netdev, ++ "Matching on tos ecn bits without also matching enc_tos ecn bits isn't supported"); ++ return -EOPNOTSUPP; ++ } ++ ++ if (inner_ecn_mask && inner_ecn_mask != INET_ECN_MASK) { ++ NL_SET_ERR_MSG_MOD(extack, ++ "Partial match on tos ecn bits with match on enc_tos ecn bits isn't supported"); ++ netdev_warn(priv->netdev, ++ "Partial match on tos ecn bits with match on enc_tos ecn bits isn't supported"); ++ return -EOPNOTSUPP; ++ } ++ ++ if (!inner_ecn_mask) ++ return 0; ++ ++ /* Both inner and outer have full mask on ecn */ ++ ++ if (outer_ecn_key == INET_ECN_ECT_1) { ++ /* inner ecn might change by DECAP action */ ++ ++ NL_SET_ERR_MSG_MOD(extack, "Match on enc_tos ecn = ECT(1) isn't supported"); ++ netdev_warn(priv->netdev, "Match on enc_tos ecn = ECT(1) isn't supported"); ++ return -EOPNOTSUPP; ++ } ++ ++ if (outer_ecn_key != INET_ECN_CE) ++ return 0; ++ ++ if (inner_ecn_key != INET_ECN_CE) { ++ /* Can't happen in software, as packet ecn will be changed to CE after decap */ ++ NL_SET_ERR_MSG_MOD(extack, ++ "Match on tos enc_tos ecn = CE while match on tos ecn != CE isn't supported"); ++ netdev_warn(priv->netdev, ++ "Match on tos enc_tos ecn = CE while match on tos ecn != CE isn't supported"); ++ return -EOPNOTSUPP; ++ } ++ ++ /* outer ecn = CE, inner ecn = CE, as decap will change inner ecn to CE in anycase, ++ * drop match on inner ecn ++ */ ++ *match_inner_ecn = false; ++ ++ return 0; ++} ++ + static int parse_tunnel_attr(struct mlx5e_priv *priv, + struct mlx5e_tc_flow *flow, + struct mlx5_flow_spec *spec, +@@ -2067,6 +2172,7 @@ static int __parse_cls_flower(struct mlx5e_priv *priv, + struct flow_rule *rule = flow_cls_offload_flow_rule(f); + struct flow_dissector *dissector = rule->match.dissector; + enum fs_flow_table_type fs_type; ++ bool match_inner_ecn = true; + u16 addr_type = 0; + u8 ip_proto = 0; + u8 *match_level; +@@ -2120,6 +2226,10 @@ static int __parse_cls_flower(struct mlx5e_priv *priv, + headers_c = get_match_inner_headers_criteria(spec); + headers_v = get_match_inner_headers_value(spec); + } ++ ++ err = mlx5e_tc_verify_tunnel_ecn(priv, f, &match_inner_ecn); ++ if (err) ++ return err; + } + + err = mlx5e_flower_parse_meta(filter_dev, f); +@@ -2341,10 +2451,12 @@ static int __parse_cls_flower(struct mlx5e_priv *priv, + struct flow_match_ip match; + + flow_rule_match_ip(rule, &match); +- MLX5_SET(fte_match_set_lyr_2_4, headers_c, ip_ecn, +- match.mask->tos & 0x3); +- MLX5_SET(fte_match_set_lyr_2_4, headers_v, ip_ecn, +- match.key->tos & 0x3); ++ if (match_inner_ecn) { ++ MLX5_SET(fte_match_set_lyr_2_4, headers_c, ip_ecn, ++ match.mask->tos & 0x3); ++ MLX5_SET(fte_match_set_lyr_2_4, headers_v, ip_ecn, ++ match.key->tos & 0x3); ++ } + + MLX5_SET(fte_match_set_lyr_2_4, headers_c, ip_dscp, + match.mask->tos >> 2); +diff --git a/drivers/net/ethernet/mellanox/mlx5/core/esw/legacy.c b/drivers/net/ethernet/mellanox/mlx5/core/esw/legacy.c +index df277a6cddc0b..0c4c743ca31e1 100644 +--- a/drivers/net/ethernet/mellanox/mlx5/core/esw/legacy.c ++++ b/drivers/net/ethernet/mellanox/mlx5/core/esw/legacy.c +@@ -431,7 +431,7 @@ int mlx5_eswitch_set_vport_vlan(struct mlx5_eswitch *esw, + int err = 0; + + if (!mlx5_esw_allowed(esw)) +- return -EPERM; ++ return vlan ? -EPERM : 0; + + if (vlan || qos) + set_flags = SET_VLAN_STRIP | SET_VLAN_INSERT; +diff --git a/drivers/net/ethernet/mellanox/mlx5/core/lag_mp.c b/drivers/net/ethernet/mellanox/mlx5/core/lag_mp.c +index 21fdaf708f1fe..30282d86e6b96 100644 +--- a/drivers/net/ethernet/mellanox/mlx5/core/lag_mp.c ++++ b/drivers/net/ethernet/mellanox/mlx5/core/lag_mp.c +@@ -268,10 +268,8 @@ static int mlx5_lag_fib_event(struct notifier_block *nb, + fen_info = container_of(info, struct fib_entry_notifier_info, + info); + fi = fen_info->fi; +- if (fi->nh) { +- NL_SET_ERR_MSG_MOD(info->extack, "IPv4 route with nexthop objects is not supported"); +- return notifier_from_errno(-EINVAL); +- } ++ if (fi->nh) ++ return NOTIFY_DONE; + fib_dev = fib_info_nh(fen_info->fi, 0)->fib_nh_dev; + if (fib_dev != ldev->pf[MLX5_LAG_P1].netdev && + fib_dev != ldev->pf[MLX5_LAG_P2].netdev) { +diff --git a/drivers/net/ethernet/mellanox/mlx5/core/main.c b/drivers/net/ethernet/mellanox/mlx5/core/main.c +index 92b01858d7f3e..29b7297a836a5 100644 +--- a/drivers/net/ethernet/mellanox/mlx5/core/main.c ++++ b/drivers/net/ethernet/mellanox/mlx5/core/main.c +@@ -97,6 +97,8 @@ enum { + MLX5_ATOMIC_REQ_MODE_HOST_ENDIANNESS = 0x1, + }; + ++#define LOG_MAX_SUPPORTED_QPS 0xff ++ + static struct mlx5_profile profile[] = { + [0] = { + .mask = 0, +@@ -108,7 +110,7 @@ static struct mlx5_profile profile[] = { + [2] = { + .mask = MLX5_PROF_MASK_QP_SIZE | + MLX5_PROF_MASK_MR_CACHE, +- .log_max_qp = 18, ++ .log_max_qp = LOG_MAX_SUPPORTED_QPS, + .mr_cache[0] = { + .size = 500, + .limit = 250 +@@ -513,7 +515,9 @@ static int handle_hca_cap(struct mlx5_core_dev *dev, void *set_ctx) + to_fw_pkey_sz(dev, 128)); + + /* Check log_max_qp from HCA caps to set in current profile */ +- if (MLX5_CAP_GEN_MAX(dev, log_max_qp) < prof->log_max_qp) { ++ if (prof->log_max_qp == LOG_MAX_SUPPORTED_QPS) { ++ prof->log_max_qp = MLX5_CAP_GEN_MAX(dev, log_max_qp); ++ } else if (MLX5_CAP_GEN_MAX(dev, log_max_qp) < prof->log_max_qp) { + mlx5_core_warn(dev, "log_max_qp value in current profile is %d, changing it to HCA capability limit (%d)\n", + prof->log_max_qp, + MLX5_CAP_GEN_MAX(dev, log_max_qp)); +diff --git a/drivers/net/ethernet/mellanox/mlx5/core/sf/dev/dev.c b/drivers/net/ethernet/mellanox/mlx5/core/sf/dev/dev.c +index 871c2fbe18d39..64bbc18332d56 100644 +--- a/drivers/net/ethernet/mellanox/mlx5/core/sf/dev/dev.c ++++ b/drivers/net/ethernet/mellanox/mlx5/core/sf/dev/dev.c +@@ -28,10 +28,7 @@ bool mlx5_sf_dev_allocated(const struct mlx5_core_dev *dev) + { + struct mlx5_sf_dev_table *table = dev->priv.sf_dev_table; + +- if (!mlx5_sf_dev_supported(dev)) +- return false; +- +- return !xa_empty(&table->devices); ++ return table && !xa_empty(&table->devices); + } + + static ssize_t sfnum_show(struct device *dev, struct device_attribute *attr, char *buf) +diff --git a/drivers/net/ethernet/mellanox/mlxsw/cmd.h b/drivers/net/ethernet/mellanox/mlxsw/cmd.h +index 392ce3cb27f72..51b260d54237e 100644 +--- a/drivers/net/ethernet/mellanox/mlxsw/cmd.h ++++ b/drivers/net/ethernet/mellanox/mlxsw/cmd.h +@@ -935,6 +935,18 @@ static inline int mlxsw_cmd_sw2hw_rdq(struct mlxsw_core *mlxsw_core, + */ + MLXSW_ITEM32(cmd_mbox, sw2hw_dq, cq, 0x00, 24, 8); + ++enum mlxsw_cmd_mbox_sw2hw_dq_sdq_lp { ++ MLXSW_CMD_MBOX_SW2HW_DQ_SDQ_LP_WQE, ++ MLXSW_CMD_MBOX_SW2HW_DQ_SDQ_LP_IGNORE_WQE, ++}; ++ ++/* cmd_mbox_sw2hw_dq_sdq_lp ++ * SDQ local Processing ++ * 0: local processing by wqe.lp ++ * 1: local processing (ignoring wqe.lp) ++ */ ++MLXSW_ITEM32(cmd_mbox, sw2hw_dq, sdq_lp, 0x00, 23, 1); ++ + /* cmd_mbox_sw2hw_dq_sdq_tclass + * SDQ: CPU Egress TClass + * RDQ: Reserved +diff --git a/drivers/net/ethernet/mellanox/mlxsw/pci.c b/drivers/net/ethernet/mellanox/mlxsw/pci.c +index fcace73eae40f..d9f9cbba62465 100644 +--- a/drivers/net/ethernet/mellanox/mlxsw/pci.c ++++ b/drivers/net/ethernet/mellanox/mlxsw/pci.c +@@ -285,6 +285,7 @@ static int mlxsw_pci_sdq_init(struct mlxsw_pci *mlxsw_pci, char *mbox, + struct mlxsw_pci_queue *q) + { + int tclass; ++ int lp; + int i; + int err; + +@@ -292,9 +293,12 @@ static int mlxsw_pci_sdq_init(struct mlxsw_pci *mlxsw_pci, char *mbox, + q->consumer_counter = 0; + tclass = q->num == MLXSW_PCI_SDQ_EMAD_INDEX ? MLXSW_PCI_SDQ_EMAD_TC : + MLXSW_PCI_SDQ_CTL_TC; ++ lp = q->num == MLXSW_PCI_SDQ_EMAD_INDEX ? MLXSW_CMD_MBOX_SW2HW_DQ_SDQ_LP_IGNORE_WQE : ++ MLXSW_CMD_MBOX_SW2HW_DQ_SDQ_LP_WQE; + + /* Set CQ of same number of this SDQ. */ + mlxsw_cmd_mbox_sw2hw_dq_cq_set(mbox, q->num); ++ mlxsw_cmd_mbox_sw2hw_dq_sdq_lp_set(mbox, lp); + mlxsw_cmd_mbox_sw2hw_dq_sdq_tclass_set(mbox, tclass); + mlxsw_cmd_mbox_sw2hw_dq_log2_dq_sz_set(mbox, 3); /* 8 pages */ + for (i = 0; i < MLXSW_PCI_AQ_PAGES; i++) { +@@ -1678,7 +1682,7 @@ static int mlxsw_pci_skb_transmit(void *bus_priv, struct sk_buff *skb, + + wqe = elem_info->elem; + mlxsw_pci_wqe_c_set(wqe, 1); /* always report completion */ +- mlxsw_pci_wqe_lp_set(wqe, !!tx_info->is_emad); ++ mlxsw_pci_wqe_lp_set(wqe, 0); + mlxsw_pci_wqe_type_set(wqe, MLXSW_PCI_WQE_TYPE_ETHERNET); + + err = mlxsw_pci_wqe_frag_map(mlxsw_pci, wqe, 0, skb->data, +@@ -1973,6 +1977,7 @@ int mlxsw_pci_driver_register(struct pci_driver *pci_driver) + { + pci_driver->probe = mlxsw_pci_probe; + pci_driver->remove = mlxsw_pci_remove; ++ pci_driver->shutdown = mlxsw_pci_remove; + return pci_register_driver(pci_driver); + } + EXPORT_SYMBOL(mlxsw_pci_driver_register); +diff --git a/drivers/net/ethernet/mscc/ocelot.c b/drivers/net/ethernet/mscc/ocelot.c +index 00b5e6860bf69..be2dd69bfee8b 100644 +--- a/drivers/net/ethernet/mscc/ocelot.c ++++ b/drivers/net/ethernet/mscc/ocelot.c +@@ -555,7 +555,10 @@ void ocelot_phylink_mac_link_up(struct ocelot *ocelot, int port, + + ocelot_write_rix(ocelot, 0, ANA_POL_FLOWC, port); + +- ocelot_fields_write(ocelot, port, SYS_PAUSE_CFG_PAUSE_ENA, tx_pause); ++ /* Don't attempt to send PAUSE frames on the NPI port, it's broken */ ++ if (port != ocelot->npi) ++ ocelot_fields_write(ocelot, port, SYS_PAUSE_CFG_PAUSE_ENA, ++ tx_pause); + + /* Undo the effects of ocelot_phylink_mac_link_down: + * enable MAC module +@@ -1302,8 +1305,7 @@ int ocelot_get_ts_info(struct ocelot *ocelot, int port, + } + EXPORT_SYMBOL(ocelot_get_ts_info); + +-static u32 ocelot_get_bond_mask(struct ocelot *ocelot, struct net_device *bond, +- bool only_active_ports) ++static u32 ocelot_get_bond_mask(struct ocelot *ocelot, struct net_device *bond) + { + u32 mask = 0; + int port; +@@ -1314,12 +1316,8 @@ static u32 ocelot_get_bond_mask(struct ocelot *ocelot, struct net_device *bond, + if (!ocelot_port) + continue; + +- if (ocelot_port->bond == bond) { +- if (only_active_ports && !ocelot_port->lag_tx_active) +- continue; +- ++ if (ocelot_port->bond == bond) + mask |= BIT(port); +- } + } + + return mask; +@@ -1406,10 +1404,8 @@ void ocelot_apply_bridge_fwd_mask(struct ocelot *ocelot) + mask = ocelot_get_bridge_fwd_mask(ocelot, port, bridge); + mask |= cpu_fwd_mask; + mask &= ~BIT(port); +- if (bond) { +- mask &= ~ocelot_get_bond_mask(ocelot, bond, +- false); +- } ++ if (bond) ++ mask &= ~ocelot_get_bond_mask(ocelot, bond); + } else { + /* Standalone ports forward only to DSA tag_8021q CPU + * ports (if those exist), or to the hardware CPU port +@@ -1727,13 +1723,17 @@ static void ocelot_set_aggr_pgids(struct ocelot *ocelot) + if (!bond || (visited & BIT(lag))) + continue; + +- bond_mask = ocelot_get_bond_mask(ocelot, bond, true); ++ bond_mask = ocelot_get_bond_mask(ocelot, bond); + + for_each_set_bit(port, &bond_mask, ocelot->num_phys_ports) { ++ struct ocelot_port *ocelot_port = ocelot->ports[port]; ++ + // Destination mask + ocelot_write_rix(ocelot, bond_mask, + ANA_PGID_PGID, port); +- aggr_idx[num_active_ports++] = port; ++ ++ if (ocelot_port->lag_tx_active) ++ aggr_idx[num_active_ports++] = port; + } + + for_each_aggr_pgid(ocelot, i) { +@@ -1782,8 +1782,7 @@ static void ocelot_setup_logical_port_ids(struct ocelot *ocelot) + + bond = ocelot_port->bond; + if (bond) { +- int lag = __ffs(ocelot_get_bond_mask(ocelot, bond, +- false)); ++ int lag = __ffs(ocelot_get_bond_mask(ocelot, bond)); + + ocelot_rmw_gix(ocelot, + ANA_PORT_PORT_CFG_PORTID_VAL(lag), +diff --git a/drivers/net/ethernet/mscc/ocelot_flower.c b/drivers/net/ethernet/mscc/ocelot_flower.c +index 8b843d3c9189a..afa062c5f82d3 100644 +--- a/drivers/net/ethernet/mscc/ocelot_flower.c ++++ b/drivers/net/ethernet/mscc/ocelot_flower.c +@@ -467,13 +467,6 @@ ocelot_flower_parse_key(struct ocelot *ocelot, int port, bool ingress, + return -EOPNOTSUPP; + } + +- if (filter->block_id == VCAP_IS1 && +- !is_zero_ether_addr(match.mask->dst)) { +- NL_SET_ERR_MSG_MOD(extack, +- "Key type S1_NORMAL cannot match on destination MAC"); +- return -EOPNOTSUPP; +- } +- + /* The hw support mac matches only for MAC_ETYPE key, + * therefore if other matches(port, tcp flags, etc) are added + * then just bail out +@@ -488,6 +481,14 @@ ocelot_flower_parse_key(struct ocelot *ocelot, int port, bool ingress, + return -EOPNOTSUPP; + + flow_rule_match_eth_addrs(rule, &match); ++ ++ if (filter->block_id == VCAP_IS1 && ++ !is_zero_ether_addr(match.mask->dst)) { ++ NL_SET_ERR_MSG_MOD(extack, ++ "Key type S1_NORMAL cannot match on destination MAC"); ++ return -EOPNOTSUPP; ++ } ++ + filter->key_type = OCELOT_VCAP_KEY_ETYPE; + ether_addr_copy(filter->key.etype.dmac.value, + match.key->dst); +diff --git a/drivers/net/ethernet/mscc/ocelot_net.c b/drivers/net/ethernet/mscc/ocelot_net.c +index 2545727fd5b2f..c08c56e07b1d3 100644 +--- a/drivers/net/ethernet/mscc/ocelot_net.c ++++ b/drivers/net/ethernet/mscc/ocelot_net.c +@@ -1168,7 +1168,7 @@ static int ocelot_netdevice_bridge_join(struct net_device *dev, + ocelot_port_bridge_join(ocelot, port, bridge); + + err = switchdev_bridge_port_offload(brport_dev, dev, priv, +- &ocelot_netdevice_nb, ++ &ocelot_switchdev_nb, + &ocelot_switchdev_blocking_nb, + false, extack); + if (err) +@@ -1182,7 +1182,7 @@ static int ocelot_netdevice_bridge_join(struct net_device *dev, + + err_switchdev_sync: + switchdev_bridge_port_unoffload(brport_dev, priv, +- &ocelot_netdevice_nb, ++ &ocelot_switchdev_nb, + &ocelot_switchdev_blocking_nb); + err_switchdev_offload: + ocelot_port_bridge_leave(ocelot, port, bridge); +@@ -1195,7 +1195,7 @@ static void ocelot_netdevice_pre_bridge_leave(struct net_device *dev, + struct ocelot_port_private *priv = netdev_priv(dev); + + switchdev_bridge_port_unoffload(brport_dev, priv, +- &ocelot_netdevice_nb, ++ &ocelot_switchdev_nb, + &ocelot_switchdev_blocking_nb); + } + +diff --git a/drivers/net/ethernet/renesas/ravb_main.c b/drivers/net/ethernet/renesas/ravb_main.c +index 0f85f2d97b18d..4e08b7219403c 100644 +--- a/drivers/net/ethernet/renesas/ravb_main.c ++++ b/drivers/net/ethernet/renesas/ravb_main.c +@@ -30,8 +30,7 @@ + #include <linux/spinlock.h> + #include <linux/sys_soc.h> + #include <linux/reset.h> +- +-#include <asm/div64.h> ++#include <linux/math64.h> + + #include "ravb.h" + +@@ -2061,8 +2060,7 @@ static int ravb_set_gti(struct net_device *ndev) + if (!rate) + return -EINVAL; + +- inc = 1000000000ULL << 20; +- do_div(inc, rate); ++ inc = div64_ul(1000000000ULL << 20, rate); + + if (inc < GTI_TIV_MIN || inc > GTI_TIV_MAX) { + dev_err(dev, "gti.tiv increment 0x%llx is outside the range 0x%x - 0x%x\n", +diff --git a/drivers/net/ethernet/rocker/rocker_ofdpa.c b/drivers/net/ethernet/rocker/rocker_ofdpa.c +index 3e1ca7a8d0295..bc70c6abd6a5b 100644 +--- a/drivers/net/ethernet/rocker/rocker_ofdpa.c ++++ b/drivers/net/ethernet/rocker/rocker_ofdpa.c +@@ -2783,7 +2783,8 @@ static void ofdpa_fib4_abort(struct rocker *rocker) + if (!ofdpa_port) + continue; + nh->fib_nh_flags &= ~RTNH_F_OFFLOAD; +- ofdpa_flow_tbl_del(ofdpa_port, OFDPA_OP_FLAG_REMOVE, ++ ofdpa_flow_tbl_del(ofdpa_port, ++ OFDPA_OP_FLAG_REMOVE | OFDPA_OP_FLAG_NOWAIT, + flow_entry); + } + spin_unlock_irqrestore(&ofdpa->flow_tbl_lock, flags); +diff --git a/drivers/net/ethernet/stmicro/stmmac/dwmac-qcom-ethqos.c b/drivers/net/ethernet/stmicro/stmmac/dwmac-qcom-ethqos.c +index 5c74b6279d690..6b1d9e8879f46 100644 +--- a/drivers/net/ethernet/stmicro/stmmac/dwmac-qcom-ethqos.c ++++ b/drivers/net/ethernet/stmicro/stmmac/dwmac-qcom-ethqos.c +@@ -113,8 +113,10 @@ static void rgmii_updatel(struct qcom_ethqos *ethqos, + rgmii_writel(ethqos, temp, offset); + } + +-static void rgmii_dump(struct qcom_ethqos *ethqos) ++static void rgmii_dump(void *priv) + { ++ struct qcom_ethqos *ethqos = priv; ++ + dev_dbg(ðqos->pdev->dev, "Rgmii register dump\n"); + dev_dbg(ðqos->pdev->dev, "RGMII_IO_MACRO_CONFIG: %x\n", + rgmii_readl(ethqos, RGMII_IO_MACRO_CONFIG)); +@@ -499,6 +501,7 @@ static int qcom_ethqos_probe(struct platform_device *pdev) + + plat_dat->bsp_priv = ethqos; + plat_dat->fix_mac_speed = ethqos_fix_mac_speed; ++ plat_dat->dump_debug_regs = rgmii_dump; + plat_dat->has_gmac4 = 1; + plat_dat->pmt = 1; + plat_dat->tso_en = of_property_read_bool(np, "snps,tso"); +@@ -507,8 +510,6 @@ static int qcom_ethqos_probe(struct platform_device *pdev) + if (ret) + goto err_clk; + +- rgmii_dump(ethqos); +- + return ret; + + err_clk: +diff --git a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c b/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c +index 3422f0746d825..06e5431cf51df 100644 +--- a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c ++++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c +@@ -7077,6 +7077,9 @@ int stmmac_dvr_probe(struct device *device, + stmmac_init_fs(ndev); + #endif + ++ if (priv->plat->dump_debug_regs) ++ priv->plat->dump_debug_regs(priv->plat->bsp_priv); ++ + /* Let pm_runtime_put() disable the clocks. + * If CONFIG_PM is not enabled, the clocks will stay powered. + */ +diff --git a/drivers/net/ethernet/ti/cpsw.c b/drivers/net/ethernet/ti/cpsw.c +index 66f7ddd9b1f99..e226ecd95a2cc 100644 +--- a/drivers/net/ethernet/ti/cpsw.c ++++ b/drivers/net/ethernet/ti/cpsw.c +@@ -349,7 +349,7 @@ static void cpsw_rx_handler(void *token, int len, int status) + struct cpsw_common *cpsw = ndev_to_cpsw(xmeta->ndev); + int pkt_size = cpsw->rx_packet_max; + int ret = 0, port, ch = xmeta->ch; +- int headroom = CPSW_HEADROOM; ++ int headroom = CPSW_HEADROOM_NA; + struct net_device *ndev = xmeta->ndev; + struct cpsw_priv *priv; + struct page_pool *pool; +@@ -392,7 +392,7 @@ static void cpsw_rx_handler(void *token, int len, int status) + } + + if (priv->xdp_prog) { +- int headroom = CPSW_HEADROOM, size = len; ++ int size = len; + + xdp_init_buff(&xdp, PAGE_SIZE, &priv->xdp_rxq[ch]); + if (status & CPDMA_RX_VLAN_ENCAP) { +@@ -442,7 +442,7 @@ requeue: + xmeta->ndev = ndev; + xmeta->ch = ch; + +- dma = page_pool_get_dma_addr(new_page) + CPSW_HEADROOM; ++ dma = page_pool_get_dma_addr(new_page) + CPSW_HEADROOM_NA; + ret = cpdma_chan_submit_mapped(cpsw->rxv[ch].ch, new_page, dma, + pkt_size, 0); + if (ret < 0) { +diff --git a/drivers/net/ethernet/ti/cpsw_new.c b/drivers/net/ethernet/ti/cpsw_new.c +index 7968f24d99c85..9e16afbdbdc1d 100644 +--- a/drivers/net/ethernet/ti/cpsw_new.c ++++ b/drivers/net/ethernet/ti/cpsw_new.c +@@ -283,7 +283,7 @@ static void cpsw_rx_handler(void *token, int len, int status) + { + struct page *new_page, *page = token; + void *pa = page_address(page); +- int headroom = CPSW_HEADROOM; ++ int headroom = CPSW_HEADROOM_NA; + struct cpsw_meta_xdp *xmeta; + struct cpsw_common *cpsw; + struct net_device *ndev; +@@ -336,7 +336,7 @@ static void cpsw_rx_handler(void *token, int len, int status) + } + + if (priv->xdp_prog) { +- int headroom = CPSW_HEADROOM, size = len; ++ int size = len; + + xdp_init_buff(&xdp, PAGE_SIZE, &priv->xdp_rxq[ch]); + if (status & CPDMA_RX_VLAN_ENCAP) { +@@ -386,7 +386,7 @@ requeue: + xmeta->ndev = ndev; + xmeta->ch = ch; + +- dma = page_pool_get_dma_addr(new_page) + CPSW_HEADROOM; ++ dma = page_pool_get_dma_addr(new_page) + CPSW_HEADROOM_NA; + ret = cpdma_chan_submit_mapped(cpsw->rxv[ch].ch, new_page, dma, + pkt_size, 0); + if (ret < 0) { +diff --git a/drivers/net/ethernet/ti/cpsw_priv.c b/drivers/net/ethernet/ti/cpsw_priv.c +index ecc2a6b7e28f2..6bb5ac51d23c3 100644 +--- a/drivers/net/ethernet/ti/cpsw_priv.c ++++ b/drivers/net/ethernet/ti/cpsw_priv.c +@@ -1120,7 +1120,7 @@ int cpsw_fill_rx_channels(struct cpsw_priv *priv) + xmeta->ndev = priv->ndev; + xmeta->ch = ch; + +- dma = page_pool_get_dma_addr(page) + CPSW_HEADROOM; ++ dma = page_pool_get_dma_addr(page) + CPSW_HEADROOM_NA; + ret = cpdma_chan_idle_submit_mapped(cpsw->rxv[ch].ch, + page, dma, + cpsw->rx_packet_max, +diff --git a/drivers/net/ethernet/xilinx/xilinx_axienet_main.c b/drivers/net/ethernet/xilinx/xilinx_axienet_main.c +index 871b5ec3183d6..2169417210c2b 100644 +--- a/drivers/net/ethernet/xilinx/xilinx_axienet_main.c ++++ b/drivers/net/ethernet/xilinx/xilinx_axienet_main.c +@@ -41,8 +41,9 @@ + #include "xilinx_axienet.h" + + /* Descriptors defines for Tx and Rx DMA */ +-#define TX_BD_NUM_DEFAULT 64 ++#define TX_BD_NUM_DEFAULT 128 + #define RX_BD_NUM_DEFAULT 1024 ++#define TX_BD_NUM_MIN (MAX_SKB_FRAGS + 1) + #define TX_BD_NUM_MAX 4096 + #define RX_BD_NUM_MAX 4096 + +@@ -496,7 +497,8 @@ static void axienet_setoptions(struct net_device *ndev, u32 options) + + static int __axienet_device_reset(struct axienet_local *lp) + { +- u32 timeout; ++ u32 value; ++ int ret; + + /* Reset Axi DMA. This would reset Axi Ethernet core as well. The reset + * process of Axi DMA takes a while to complete as all pending +@@ -506,15 +508,23 @@ static int __axienet_device_reset(struct axienet_local *lp) + * they both reset the entire DMA core, so only one needs to be used. + */ + axienet_dma_out32(lp, XAXIDMA_TX_CR_OFFSET, XAXIDMA_CR_RESET_MASK); +- timeout = DELAY_OF_ONE_MILLISEC; +- while (axienet_dma_in32(lp, XAXIDMA_TX_CR_OFFSET) & +- XAXIDMA_CR_RESET_MASK) { +- udelay(1); +- if (--timeout == 0) { +- netdev_err(lp->ndev, "%s: DMA reset timeout!\n", +- __func__); +- return -ETIMEDOUT; +- } ++ ret = read_poll_timeout(axienet_dma_in32, value, ++ !(value & XAXIDMA_CR_RESET_MASK), ++ DELAY_OF_ONE_MILLISEC, 50000, false, lp, ++ XAXIDMA_TX_CR_OFFSET); ++ if (ret) { ++ dev_err(lp->dev, "%s: DMA reset timeout!\n", __func__); ++ return ret; ++ } ++ ++ /* Wait for PhyRstCmplt bit to be set, indicating the PHY reset has finished */ ++ ret = read_poll_timeout(axienet_ior, value, ++ value & XAE_INT_PHYRSTCMPLT_MASK, ++ DELAY_OF_ONE_MILLISEC, 50000, false, lp, ++ XAE_IS_OFFSET); ++ if (ret) { ++ dev_err(lp->dev, "%s: timeout waiting for PhyRstCmplt\n", __func__); ++ return ret; + } + + return 0; +@@ -623,6 +633,8 @@ static int axienet_free_tx_chain(struct net_device *ndev, u32 first_bd, + if (nr_bds == -1 && !(status & XAXIDMA_BD_STS_COMPLETE_MASK)) + break; + ++ /* Ensure we see complete descriptor update */ ++ dma_rmb(); + phys = desc_get_phys_addr(lp, cur_p); + dma_unmap_single(ndev->dev.parent, phys, + (cur_p->cntrl & XAXIDMA_BD_CTRL_LENGTH_MASK), +@@ -631,13 +643,15 @@ static int axienet_free_tx_chain(struct net_device *ndev, u32 first_bd, + if (cur_p->skb && (status & XAXIDMA_BD_STS_COMPLETE_MASK)) + dev_consume_skb_irq(cur_p->skb); + +- cur_p->cntrl = 0; + cur_p->app0 = 0; + cur_p->app1 = 0; + cur_p->app2 = 0; + cur_p->app4 = 0; +- cur_p->status = 0; + cur_p->skb = NULL; ++ /* ensure our transmit path and device don't prematurely see status cleared */ ++ wmb(); ++ cur_p->cntrl = 0; ++ cur_p->status = 0; + + if (sizep) + *sizep += status & XAXIDMA_BD_STS_ACTUAL_LEN_MASK; +@@ -646,6 +660,32 @@ static int axienet_free_tx_chain(struct net_device *ndev, u32 first_bd, + return i; + } + ++/** ++ * axienet_check_tx_bd_space - Checks if a BD/group of BDs are currently busy ++ * @lp: Pointer to the axienet_local structure ++ * @num_frag: The number of BDs to check for ++ * ++ * Return: 0, on success ++ * NETDEV_TX_BUSY, if any of the descriptors are not free ++ * ++ * This function is invoked before BDs are allocated and transmission starts. ++ * This function returns 0 if a BD or group of BDs can be allocated for ++ * transmission. If the BD or any of the BDs are not free the function ++ * returns a busy status. This is invoked from axienet_start_xmit. ++ */ ++static inline int axienet_check_tx_bd_space(struct axienet_local *lp, ++ int num_frag) ++{ ++ struct axidma_bd *cur_p; ++ ++ /* Ensure we see all descriptor updates from device or TX IRQ path */ ++ rmb(); ++ cur_p = &lp->tx_bd_v[(lp->tx_bd_tail + num_frag) % lp->tx_bd_num]; ++ if (cur_p->cntrl) ++ return NETDEV_TX_BUSY; ++ return 0; ++} ++ + /** + * axienet_start_xmit_done - Invoked once a transmit is completed by the + * Axi DMA Tx channel. +@@ -675,30 +715,8 @@ static void axienet_start_xmit_done(struct net_device *ndev) + /* Matches barrier in axienet_start_xmit */ + smp_mb(); + +- netif_wake_queue(ndev); +-} +- +-/** +- * axienet_check_tx_bd_space - Checks if a BD/group of BDs are currently busy +- * @lp: Pointer to the axienet_local structure +- * @num_frag: The number of BDs to check for +- * +- * Return: 0, on success +- * NETDEV_TX_BUSY, if any of the descriptors are not free +- * +- * This function is invoked before BDs are allocated and transmission starts. +- * This function returns 0 if a BD or group of BDs can be allocated for +- * transmission. If the BD or any of the BDs are not free the function +- * returns a busy status. This is invoked from axienet_start_xmit. +- */ +-static inline int axienet_check_tx_bd_space(struct axienet_local *lp, +- int num_frag) +-{ +- struct axidma_bd *cur_p; +- cur_p = &lp->tx_bd_v[(lp->tx_bd_tail + num_frag) % lp->tx_bd_num]; +- if (cur_p->status & XAXIDMA_BD_STS_ALL_MASK) +- return NETDEV_TX_BUSY; +- return 0; ++ if (!axienet_check_tx_bd_space(lp, MAX_SKB_FRAGS + 1)) ++ netif_wake_queue(ndev); + } + + /** +@@ -730,20 +748,15 @@ axienet_start_xmit(struct sk_buff *skb, struct net_device *ndev) + num_frag = skb_shinfo(skb)->nr_frags; + cur_p = &lp->tx_bd_v[lp->tx_bd_tail]; + +- if (axienet_check_tx_bd_space(lp, num_frag)) { +- if (netif_queue_stopped(ndev)) +- return NETDEV_TX_BUSY; +- ++ if (axienet_check_tx_bd_space(lp, num_frag + 1)) { ++ /* Should not happen as last start_xmit call should have ++ * checked for sufficient space and queue should only be ++ * woken when sufficient space is available. ++ */ + netif_stop_queue(ndev); +- +- /* Matches barrier in axienet_start_xmit_done */ +- smp_mb(); +- +- /* Space might have just been freed - check again */ +- if (axienet_check_tx_bd_space(lp, num_frag)) +- return NETDEV_TX_BUSY; +- +- netif_wake_queue(ndev); ++ if (net_ratelimit()) ++ netdev_warn(ndev, "TX ring unexpectedly full\n"); ++ return NETDEV_TX_BUSY; + } + + if (skb->ip_summed == CHECKSUM_PARTIAL) { +@@ -804,6 +817,18 @@ axienet_start_xmit(struct sk_buff *skb, struct net_device *ndev) + if (++lp->tx_bd_tail >= lp->tx_bd_num) + lp->tx_bd_tail = 0; + ++ /* Stop queue if next transmit may not have space */ ++ if (axienet_check_tx_bd_space(lp, MAX_SKB_FRAGS + 1)) { ++ netif_stop_queue(ndev); ++ ++ /* Matches barrier in axienet_start_xmit_done */ ++ smp_mb(); ++ ++ /* Space might have just been freed - check again */ ++ if (!axienet_check_tx_bd_space(lp, MAX_SKB_FRAGS + 1)) ++ netif_wake_queue(ndev); ++ } ++ + return NETDEV_TX_OK; + } + +@@ -834,6 +859,8 @@ static void axienet_recv(struct net_device *ndev) + + tail_p = lp->rx_bd_p + sizeof(*lp->rx_bd_v) * lp->rx_bd_ci; + ++ /* Ensure we see complete descriptor update */ ++ dma_rmb(); + phys = desc_get_phys_addr(lp, cur_p); + dma_unmap_single(ndev->dev.parent, phys, lp->max_frm_size, + DMA_FROM_DEVICE); +@@ -1346,7 +1373,8 @@ static int axienet_ethtools_set_ringparam(struct net_device *ndev, + if (ering->rx_pending > RX_BD_NUM_MAX || + ering->rx_mini_pending || + ering->rx_jumbo_pending || +- ering->rx_pending > TX_BD_NUM_MAX) ++ ering->tx_pending < TX_BD_NUM_MIN || ++ ering->tx_pending > TX_BD_NUM_MAX) + return -EINVAL; + + if (netif_running(ndev)) +@@ -2082,6 +2110,11 @@ static int axienet_probe(struct platform_device *pdev) + lp->coalesce_count_rx = XAXIDMA_DFT_RX_THRESHOLD; + lp->coalesce_count_tx = XAXIDMA_DFT_TX_THRESHOLD; + ++ /* Reset core now that clocks are enabled, prior to accessing MDIO */ ++ ret = __axienet_device_reset(lp); ++ if (ret) ++ goto cleanup_clk; ++ + lp->phy_node = of_parse_phandle(pdev->dev.of_node, "phy-handle", 0); + if (lp->phy_node) { + ret = axienet_mdio_setup(lp); +diff --git a/drivers/net/ipa/ipa_endpoint.c b/drivers/net/ipa/ipa_endpoint.c +index 03a1709934208..c8f90cb1ee8f3 100644 +--- a/drivers/net/ipa/ipa_endpoint.c ++++ b/drivers/net/ipa/ipa_endpoint.c +@@ -1067,6 +1067,7 @@ static void ipa_endpoint_replenish(struct ipa_endpoint *endpoint, bool add_one) + { + struct gsi *gsi; + u32 backlog; ++ int delta; + + if (!endpoint->replenish_enabled) { + if (add_one) +@@ -1084,10 +1085,8 @@ static void ipa_endpoint_replenish(struct ipa_endpoint *endpoint, bool add_one) + + try_again_later: + /* The last one didn't succeed, so fix the backlog */ +- backlog = atomic_inc_return(&endpoint->replenish_backlog); +- +- if (add_one) +- atomic_inc(&endpoint->replenish_backlog); ++ delta = add_one ? 2 : 1; ++ backlog = atomic_add_return(delta, &endpoint->replenish_backlog); + + /* Whenever a receive buffer transaction completes we'll try to + * replenish again. It's unlikely, but if we fail to supply even +diff --git a/drivers/net/phy/marvell.c b/drivers/net/phy/marvell.c +index 4fcfca4e17021..c10cc2cd53b66 100644 +--- a/drivers/net/phy/marvell.c ++++ b/drivers/net/phy/marvell.c +@@ -189,6 +189,8 @@ + #define MII_88E1510_GEN_CTRL_REG_1_MODE_RGMII_SGMII 0x4 + #define MII_88E1510_GEN_CTRL_REG_1_RESET 0x8000 /* Soft reset */ + ++#define MII_88E1510_MSCR_2 0x15 ++ + #define MII_VCT5_TX_RX_MDI0_COUPLING 0x10 + #define MII_VCT5_TX_RX_MDI1_COUPLING 0x11 + #define MII_VCT5_TX_RX_MDI2_COUPLING 0x12 +@@ -1242,6 +1244,12 @@ static int m88e1118_config_init(struct phy_device *phydev) + if (err < 0) + return err; + ++ if (phy_interface_is_rgmii(phydev)) { ++ err = m88e1121_config_aneg_rgmii_delays(phydev); ++ if (err < 0) ++ return err; ++ } ++ + /* Adjust LED Control */ + if (phydev->dev_flags & MARVELL_PHY_M1118_DNS323_LEDS) + err = phy_write(phydev, 0x10, 0x1100); +@@ -1932,6 +1940,58 @@ static void marvell_get_stats(struct phy_device *phydev, + data[i] = marvell_get_stat(phydev, i); + } + ++static int m88e1510_loopback(struct phy_device *phydev, bool enable) ++{ ++ int err; ++ ++ if (enable) { ++ u16 bmcr_ctl = 0, mscr2_ctl = 0; ++ ++ if (phydev->speed == SPEED_1000) ++ bmcr_ctl = BMCR_SPEED1000; ++ else if (phydev->speed == SPEED_100) ++ bmcr_ctl = BMCR_SPEED100; ++ ++ if (phydev->duplex == DUPLEX_FULL) ++ bmcr_ctl |= BMCR_FULLDPLX; ++ ++ err = phy_write(phydev, MII_BMCR, bmcr_ctl); ++ if (err < 0) ++ return err; ++ ++ if (phydev->speed == SPEED_1000) ++ mscr2_ctl = BMCR_SPEED1000; ++ else if (phydev->speed == SPEED_100) ++ mscr2_ctl = BMCR_SPEED100; ++ ++ err = phy_modify_paged(phydev, MII_MARVELL_MSCR_PAGE, ++ MII_88E1510_MSCR_2, BMCR_SPEED1000 | ++ BMCR_SPEED100, mscr2_ctl); ++ if (err < 0) ++ return err; ++ ++ /* Need soft reset to have speed configuration takes effect */ ++ err = genphy_soft_reset(phydev); ++ if (err < 0) ++ return err; ++ ++ /* FIXME: Based on trial and error test, it seem 1G need to have ++ * delay between soft reset and loopback enablement. ++ */ ++ if (phydev->speed == SPEED_1000) ++ msleep(1000); ++ ++ return phy_modify(phydev, MII_BMCR, BMCR_LOOPBACK, ++ BMCR_LOOPBACK); ++ } else { ++ err = phy_modify(phydev, MII_BMCR, BMCR_LOOPBACK, 0); ++ if (err < 0) ++ return err; ++ ++ return phy_config_aneg(phydev); ++ } ++} ++ + static int marvell_vct5_wait_complete(struct phy_device *phydev) + { + int i; +@@ -3078,7 +3138,7 @@ static struct phy_driver marvell_drivers[] = { + .get_sset_count = marvell_get_sset_count, + .get_strings = marvell_get_strings, + .get_stats = marvell_get_stats, +- .set_loopback = genphy_loopback, ++ .set_loopback = m88e1510_loopback, + .get_tunable = m88e1011_get_tunable, + .set_tunable = m88e1011_set_tunable, + .cable_test_start = marvell_vct7_cable_test_start, +diff --git a/drivers/net/phy/mdio_bus.c b/drivers/net/phy/mdio_bus.c +index 6865d9319197f..8dc6e6269c65e 100644 +--- a/drivers/net/phy/mdio_bus.c ++++ b/drivers/net/phy/mdio_bus.c +@@ -591,7 +591,7 @@ int __mdiobus_register(struct mii_bus *bus, struct module *owner) + mdiobus_setup_mdiodev_from_board_info(bus, mdiobus_create_device); + + bus->state = MDIOBUS_REGISTERED; +- pr_info("%s: probed\n", bus->name); ++ dev_dbg(&bus->dev, "probed\n"); + return 0; + + error: +diff --git a/drivers/net/phy/micrel.c b/drivers/net/phy/micrel.c +index aec0fcefdccd6..e2ac61f44c945 100644 +--- a/drivers/net/phy/micrel.c ++++ b/drivers/net/phy/micrel.c +@@ -1547,8 +1547,8 @@ static struct phy_driver ksphy_driver[] = { + .config_init = kszphy_config_init, + .config_intr = kszphy_config_intr, + .handle_interrupt = kszphy_handle_interrupt, +- .suspend = genphy_suspend, +- .resume = genphy_resume, ++ .suspend = kszphy_suspend, ++ .resume = kszphy_resume, + }, { + .phy_id = PHY_ID_KSZ8021, + .phy_id_mask = 0x00ffffff, +@@ -1562,8 +1562,8 @@ static struct phy_driver ksphy_driver[] = { + .get_sset_count = kszphy_get_sset_count, + .get_strings = kszphy_get_strings, + .get_stats = kszphy_get_stats, +- .suspend = genphy_suspend, +- .resume = genphy_resume, ++ .suspend = kszphy_suspend, ++ .resume = kszphy_resume, + }, { + .phy_id = PHY_ID_KSZ8031, + .phy_id_mask = 0x00ffffff, +@@ -1577,8 +1577,8 @@ static struct phy_driver ksphy_driver[] = { + .get_sset_count = kszphy_get_sset_count, + .get_strings = kszphy_get_strings, + .get_stats = kszphy_get_stats, +- .suspend = genphy_suspend, +- .resume = genphy_resume, ++ .suspend = kszphy_suspend, ++ .resume = kszphy_resume, + }, { + .phy_id = PHY_ID_KSZ8041, + .phy_id_mask = MICREL_PHY_ID_MASK, +@@ -1609,8 +1609,8 @@ static struct phy_driver ksphy_driver[] = { + .get_sset_count = kszphy_get_sset_count, + .get_strings = kszphy_get_strings, + .get_stats = kszphy_get_stats, +- .suspend = genphy_suspend, +- .resume = genphy_resume, ++ .suspend = kszphy_suspend, ++ .resume = kszphy_resume, + }, { + .name = "Micrel KSZ8051", + /* PHY_BASIC_FEATURES */ +@@ -1623,8 +1623,8 @@ static struct phy_driver ksphy_driver[] = { + .get_strings = kszphy_get_strings, + .get_stats = kszphy_get_stats, + .match_phy_device = ksz8051_match_phy_device, +- .suspend = genphy_suspend, +- .resume = genphy_resume, ++ .suspend = kszphy_suspend, ++ .resume = kszphy_resume, + }, { + .phy_id = PHY_ID_KSZ8001, + .name = "Micrel KSZ8001 or KS8721", +@@ -1638,8 +1638,8 @@ static struct phy_driver ksphy_driver[] = { + .get_sset_count = kszphy_get_sset_count, + .get_strings = kszphy_get_strings, + .get_stats = kszphy_get_stats, +- .suspend = genphy_suspend, +- .resume = genphy_resume, ++ .suspend = kszphy_suspend, ++ .resume = kszphy_resume, + }, { + .phy_id = PHY_ID_KSZ8081, + .name = "Micrel KSZ8081 or KSZ8091", +@@ -1669,8 +1669,8 @@ static struct phy_driver ksphy_driver[] = { + .config_init = ksz8061_config_init, + .config_intr = kszphy_config_intr, + .handle_interrupt = kszphy_handle_interrupt, +- .suspend = genphy_suspend, +- .resume = genphy_resume, ++ .suspend = kszphy_suspend, ++ .resume = kszphy_resume, + }, { + .phy_id = PHY_ID_KSZ9021, + .phy_id_mask = 0x000ffffe, +@@ -1685,8 +1685,8 @@ static struct phy_driver ksphy_driver[] = { + .get_sset_count = kszphy_get_sset_count, + .get_strings = kszphy_get_strings, + .get_stats = kszphy_get_stats, +- .suspend = genphy_suspend, +- .resume = genphy_resume, ++ .suspend = kszphy_suspend, ++ .resume = kszphy_resume, + .read_mmd = genphy_read_mmd_unsupported, + .write_mmd = genphy_write_mmd_unsupported, + }, { +@@ -1704,7 +1704,7 @@ static struct phy_driver ksphy_driver[] = { + .get_sset_count = kszphy_get_sset_count, + .get_strings = kszphy_get_strings, + .get_stats = kszphy_get_stats, +- .suspend = genphy_suspend, ++ .suspend = kszphy_suspend, + .resume = kszphy_resume, + }, { + .phy_id = PHY_ID_LAN8814, +@@ -1732,7 +1732,7 @@ static struct phy_driver ksphy_driver[] = { + .get_sset_count = kszphy_get_sset_count, + .get_strings = kszphy_get_strings, + .get_stats = kszphy_get_stats, +- .suspend = genphy_suspend, ++ .suspend = kszphy_suspend, + .resume = kszphy_resume, + }, { + .phy_id = PHY_ID_KSZ8873MLL, +diff --git a/drivers/net/phy/phy-core.c b/drivers/net/phy/phy-core.c +index 2870c33b8975d..271fc01f7f7fd 100644 +--- a/drivers/net/phy/phy-core.c ++++ b/drivers/net/phy/phy-core.c +@@ -162,11 +162,11 @@ static const struct phy_setting settings[] = { + PHY_SETTING( 2500, FULL, 2500baseT_Full ), + PHY_SETTING( 2500, FULL, 2500baseX_Full ), + /* 1G */ +- PHY_SETTING( 1000, FULL, 1000baseKX_Full ), + PHY_SETTING( 1000, FULL, 1000baseT_Full ), + PHY_SETTING( 1000, HALF, 1000baseT_Half ), + PHY_SETTING( 1000, FULL, 1000baseT1_Full ), + PHY_SETTING( 1000, FULL, 1000baseX_Full ), ++ PHY_SETTING( 1000, FULL, 1000baseKX_Full ), + /* 100M */ + PHY_SETTING( 100, FULL, 100baseT_Full ), + PHY_SETTING( 100, FULL, 100baseT1_Full ), +diff --git a/drivers/net/phy/sfp.c b/drivers/net/phy/sfp.c +index ab77a9f439ef9..4720b24ca51b5 100644 +--- a/drivers/net/phy/sfp.c ++++ b/drivers/net/phy/sfp.c +@@ -1641,17 +1641,20 @@ static int sfp_sm_probe_for_phy(struct sfp *sfp) + static int sfp_module_parse_power(struct sfp *sfp) + { + u32 power_mW = 1000; ++ bool supports_a2; + + if (sfp->id.ext.options & cpu_to_be16(SFP_OPTIONS_POWER_DECL)) + power_mW = 1500; + if (sfp->id.ext.options & cpu_to_be16(SFP_OPTIONS_HIGH_POWER_LEVEL)) + power_mW = 2000; + ++ supports_a2 = sfp->id.ext.sff8472_compliance != ++ SFP_SFF8472_COMPLIANCE_NONE || ++ sfp->id.ext.diagmon & SFP_DIAGMON_DDM; ++ + if (power_mW > sfp->max_power_mW) { + /* Module power specification exceeds the allowed maximum. */ +- if (sfp->id.ext.sff8472_compliance == +- SFP_SFF8472_COMPLIANCE_NONE && +- !(sfp->id.ext.diagmon & SFP_DIAGMON_DDM)) { ++ if (!supports_a2) { + /* The module appears not to implement bus address + * 0xa2, so assume that the module powers up in the + * indicated mode. +@@ -1668,11 +1671,25 @@ static int sfp_module_parse_power(struct sfp *sfp) + } + } + ++ if (power_mW <= 1000) { ++ /* Modules below 1W do not require a power change sequence */ ++ sfp->module_power_mW = power_mW; ++ return 0; ++ } ++ ++ if (!supports_a2) { ++ /* The module power level is below the host maximum and the ++ * module appears not to implement bus address 0xa2, so assume ++ * that the module powers up in the indicated mode. ++ */ ++ return 0; ++ } ++ + /* If the module requires a higher power mode, but also requires + * an address change sequence, warn the user that the module may + * not be functional. + */ +- if (sfp->id.ext.diagmon & SFP_DIAGMON_ADDRMODE && power_mW > 1000) { ++ if (sfp->id.ext.diagmon & SFP_DIAGMON_ADDRMODE) { + dev_warn(sfp->dev, + "Address Change Sequence not supported but module requires %u.%uW, module may not be functional\n", + power_mW / 1000, (power_mW / 100) % 10); +diff --git a/drivers/net/ppp/ppp_generic.c b/drivers/net/ppp/ppp_generic.c +index fb52cd175b45d..829d6ada1704c 100644 +--- a/drivers/net/ppp/ppp_generic.c ++++ b/drivers/net/ppp/ppp_generic.c +@@ -69,6 +69,8 @@ + #define MPHDRLEN 6 /* multilink protocol header length */ + #define MPHDRLEN_SSN 4 /* ditto with short sequence numbers */ + ++#define PPP_PROTO_LEN 2 ++ + /* + * An instance of /dev/ppp can be associated with either a ppp + * interface unit or a ppp channel. In both cases, file->private_data +@@ -497,6 +499,9 @@ static ssize_t ppp_write(struct file *file, const char __user *buf, + + if (!pf) + return -ENXIO; ++ /* All PPP packets should start with the 2-byte protocol */ ++ if (count < PPP_PROTO_LEN) ++ return -EINVAL; + ret = -ENOMEM; + skb = alloc_skb(count + pf->hdrlen, GFP_KERNEL); + if (!skb) +@@ -1764,7 +1769,7 @@ ppp_send_frame(struct ppp *ppp, struct sk_buff *skb) + } + + ++ppp->stats64.tx_packets; +- ppp->stats64.tx_bytes += skb->len - 2; ++ ppp->stats64.tx_bytes += skb->len - PPP_PROTO_LEN; + + switch (proto) { + case PPP_IP: +diff --git a/drivers/net/usb/mcs7830.c b/drivers/net/usb/mcs7830.c +index 66866bef25df7..a31a3b9cbd58d 100644 +--- a/drivers/net/usb/mcs7830.c ++++ b/drivers/net/usb/mcs7830.c +@@ -108,8 +108,16 @@ static const char driver_name[] = "MOSCHIP usb-ethernet driver"; + + static int mcs7830_get_reg(struct usbnet *dev, u16 index, u16 size, void *data) + { +- return usbnet_read_cmd(dev, MCS7830_RD_BREQ, MCS7830_RD_BMREQ, +- 0x0000, index, data, size); ++ int ret; ++ ++ ret = usbnet_read_cmd(dev, MCS7830_RD_BREQ, MCS7830_RD_BMREQ, ++ 0x0000, index, data, size); ++ if (ret < 0) ++ return ret; ++ else if (ret < size) ++ return -ENODATA; ++ ++ return ret; + } + + static int mcs7830_set_reg(struct usbnet *dev, u16 index, u16 size, const void *data) +diff --git a/drivers/net/usb/smsc95xx.c b/drivers/net/usb/smsc95xx.c +index f91dabd65ecd8..026e7487c45b5 100644 +--- a/drivers/net/usb/smsc95xx.c ++++ b/drivers/net/usb/smsc95xx.c +@@ -1961,7 +1961,8 @@ static const struct driver_info smsc95xx_info = { + .bind = smsc95xx_bind, + .unbind = smsc95xx_unbind, + .link_reset = smsc95xx_link_reset, +- .reset = smsc95xx_start_phy, ++ .reset = smsc95xx_reset, ++ .check_connect = smsc95xx_start_phy, + .stop = smsc95xx_stop, + .rx_fixup = smsc95xx_rx_fixup, + .tx_fixup = smsc95xx_tx_fixup, +diff --git a/drivers/net/wireless/ath/ar5523/ar5523.c b/drivers/net/wireless/ath/ar5523/ar5523.c +index 49cc4b7ed5163..1baec4b412c8d 100644 +--- a/drivers/net/wireless/ath/ar5523/ar5523.c ++++ b/drivers/net/wireless/ath/ar5523/ar5523.c +@@ -153,6 +153,10 @@ static void ar5523_cmd_rx_cb(struct urb *urb) + ar5523_err(ar, "Invalid reply to WDCMSG_TARGET_START"); + return; + } ++ if (!cmd->odata) { ++ ar5523_err(ar, "Unexpected WDCMSG_TARGET_START reply"); ++ return; ++ } + memcpy(cmd->odata, hdr + 1, sizeof(u32)); + cmd->olen = sizeof(u32); + cmd->res = 0; +diff --git a/drivers/net/wireless/ath/ath10k/core.c b/drivers/net/wireless/ath/ath10k/core.c +index 64c7145b51a2e..58e86e662ab83 100644 +--- a/drivers/net/wireless/ath/ath10k/core.c ++++ b/drivers/net/wireless/ath/ath10k/core.c +@@ -89,6 +89,7 @@ static const struct ath10k_hw_params ath10k_hw_params_list[] = { + .rri_on_ddr = false, + .hw_filter_reset_required = true, + .fw_diag_ce_download = false, ++ .credit_size_workaround = false, + .tx_stats_over_pktlog = true, + .dynamic_sar_support = false, + }, +@@ -124,6 +125,7 @@ static const struct ath10k_hw_params ath10k_hw_params_list[] = { + .rri_on_ddr = false, + .hw_filter_reset_required = true, + .fw_diag_ce_download = false, ++ .credit_size_workaround = false, + .tx_stats_over_pktlog = true, + .dynamic_sar_support = false, + }, +@@ -160,6 +162,7 @@ static const struct ath10k_hw_params ath10k_hw_params_list[] = { + .rri_on_ddr = false, + .hw_filter_reset_required = true, + .fw_diag_ce_download = false, ++ .credit_size_workaround = false, + .tx_stats_over_pktlog = false, + .dynamic_sar_support = false, + }, +@@ -190,6 +193,7 @@ static const struct ath10k_hw_params ath10k_hw_params_list[] = { + .num_wds_entries = 0x20, + .uart_pin_workaround = true, + .tx_stats_over_pktlog = false, ++ .credit_size_workaround = false, + .bmi_large_size_download = true, + .supports_peer_stats_info = true, + .dynamic_sar_support = true, +@@ -226,6 +230,7 @@ static const struct ath10k_hw_params ath10k_hw_params_list[] = { + .rri_on_ddr = false, + .hw_filter_reset_required = true, + .fw_diag_ce_download = false, ++ .credit_size_workaround = false, + .tx_stats_over_pktlog = false, + .dynamic_sar_support = false, + }, +@@ -261,6 +266,7 @@ static const struct ath10k_hw_params ath10k_hw_params_list[] = { + .rri_on_ddr = false, + .hw_filter_reset_required = true, + .fw_diag_ce_download = false, ++ .credit_size_workaround = false, + .tx_stats_over_pktlog = false, + .dynamic_sar_support = false, + }, +@@ -296,6 +302,7 @@ static const struct ath10k_hw_params ath10k_hw_params_list[] = { + .rri_on_ddr = false, + .hw_filter_reset_required = true, + .fw_diag_ce_download = false, ++ .credit_size_workaround = false, + .tx_stats_over_pktlog = false, + .dynamic_sar_support = false, + }, +@@ -334,6 +341,7 @@ static const struct ath10k_hw_params ath10k_hw_params_list[] = { + .rri_on_ddr = false, + .hw_filter_reset_required = true, + .fw_diag_ce_download = true, ++ .credit_size_workaround = false, + .tx_stats_over_pktlog = false, + .supports_peer_stats_info = true, + .dynamic_sar_support = true, +@@ -376,6 +384,7 @@ static const struct ath10k_hw_params ath10k_hw_params_list[] = { + .rri_on_ddr = false, + .hw_filter_reset_required = true, + .fw_diag_ce_download = false, ++ .credit_size_workaround = false, + .tx_stats_over_pktlog = false, + .dynamic_sar_support = false, + }, +@@ -424,6 +433,7 @@ static const struct ath10k_hw_params ath10k_hw_params_list[] = { + .rri_on_ddr = false, + .hw_filter_reset_required = true, + .fw_diag_ce_download = false, ++ .credit_size_workaround = false, + .tx_stats_over_pktlog = false, + .dynamic_sar_support = false, + }, +@@ -469,6 +479,7 @@ static const struct ath10k_hw_params ath10k_hw_params_list[] = { + .rri_on_ddr = false, + .hw_filter_reset_required = true, + .fw_diag_ce_download = false, ++ .credit_size_workaround = false, + .tx_stats_over_pktlog = false, + .dynamic_sar_support = false, + }, +@@ -504,6 +515,7 @@ static const struct ath10k_hw_params ath10k_hw_params_list[] = { + .rri_on_ddr = false, + .hw_filter_reset_required = true, + .fw_diag_ce_download = false, ++ .credit_size_workaround = false, + .tx_stats_over_pktlog = false, + .dynamic_sar_support = false, + }, +@@ -541,6 +553,7 @@ static const struct ath10k_hw_params ath10k_hw_params_list[] = { + .rri_on_ddr = false, + .hw_filter_reset_required = true, + .fw_diag_ce_download = true, ++ .credit_size_workaround = false, + .tx_stats_over_pktlog = false, + .dynamic_sar_support = false, + }, +@@ -570,6 +583,7 @@ static const struct ath10k_hw_params ath10k_hw_params_list[] = { + .ast_skid_limit = 0x10, + .num_wds_entries = 0x20, + .uart_pin_workaround = true, ++ .credit_size_workaround = true, + .dynamic_sar_support = false, + }, + { +@@ -611,6 +625,7 @@ static const struct ath10k_hw_params ath10k_hw_params_list[] = { + .rri_on_ddr = false, + .hw_filter_reset_required = true, + .fw_diag_ce_download = false, ++ .credit_size_workaround = false, + .tx_stats_over_pktlog = false, + .dynamic_sar_support = false, + }, +@@ -639,6 +654,7 @@ static const struct ath10k_hw_params ath10k_hw_params_list[] = { + .rri_on_ddr = true, + .hw_filter_reset_required = false, + .fw_diag_ce_download = false, ++ .credit_size_workaround = false, + .tx_stats_over_pktlog = false, + .dynamic_sar_support = true, + }, +@@ -714,6 +730,7 @@ static void ath10k_send_suspend_complete(struct ath10k *ar) + + static int ath10k_init_sdio(struct ath10k *ar, enum ath10k_firmware_mode mode) + { ++ bool mtu_workaround = ar->hw_params.credit_size_workaround; + int ret; + u32 param = 0; + +@@ -731,7 +748,7 @@ static int ath10k_init_sdio(struct ath10k *ar, enum ath10k_firmware_mode mode) + + param |= HI_ACS_FLAGS_SDIO_REDUCE_TX_COMPL_SET; + +- if (mode == ATH10K_FIRMWARE_MODE_NORMAL) ++ if (mode == ATH10K_FIRMWARE_MODE_NORMAL && !mtu_workaround) + param |= HI_ACS_FLAGS_ALT_DATA_CREDIT_SIZE; + else + param &= ~HI_ACS_FLAGS_ALT_DATA_CREDIT_SIZE; +diff --git a/drivers/net/wireless/ath/ath10k/htt_tx.c b/drivers/net/wireless/ath/ath10k/htt_tx.c +index d6b8bdcef4160..b793eac2cfac8 100644 +--- a/drivers/net/wireless/ath/ath10k/htt_tx.c ++++ b/drivers/net/wireless/ath/ath10k/htt_tx.c +@@ -147,6 +147,9 @@ void ath10k_htt_tx_dec_pending(struct ath10k_htt *htt) + htt->num_pending_tx--; + if (htt->num_pending_tx == htt->max_num_pending_tx - 1) + ath10k_mac_tx_unlock(htt->ar, ATH10K_TX_PAUSE_Q_FULL); ++ ++ if (htt->num_pending_tx == 0) ++ wake_up(&htt->empty_tx_wq); + } + + int ath10k_htt_tx_inc_pending(struct ath10k_htt *htt) +diff --git a/drivers/net/wireless/ath/ath10k/hw.h b/drivers/net/wireless/ath/ath10k/hw.h +index 6b03c7787e36a..591ef7416b613 100644 +--- a/drivers/net/wireless/ath/ath10k/hw.h ++++ b/drivers/net/wireless/ath/ath10k/hw.h +@@ -618,6 +618,9 @@ struct ath10k_hw_params { + */ + bool uart_pin_workaround; + ++ /* Workaround for the credit size calculation */ ++ bool credit_size_workaround; ++ + /* tx stats support over pktlog */ + bool tx_stats_over_pktlog; + +diff --git a/drivers/net/wireless/ath/ath10k/txrx.c b/drivers/net/wireless/ath/ath10k/txrx.c +index 7c9ea0c073d8b..6f8b642188941 100644 +--- a/drivers/net/wireless/ath/ath10k/txrx.c ++++ b/drivers/net/wireless/ath/ath10k/txrx.c +@@ -82,8 +82,6 @@ int ath10k_txrx_tx_unref(struct ath10k_htt *htt, + flags = skb_cb->flags; + ath10k_htt_tx_free_msdu_id(htt, tx_done->msdu_id); + ath10k_htt_tx_dec_pending(htt); +- if (htt->num_pending_tx == 0) +- wake_up(&htt->empty_tx_wq); + spin_unlock_bh(&htt->tx_lock); + + rcu_read_lock(); +diff --git a/drivers/net/wireless/ath/ath11k/ahb.c b/drivers/net/wireless/ath/ath11k/ahb.c +index 8c9c781afc3e5..3fb0aa0008259 100644 +--- a/drivers/net/wireless/ath/ath11k/ahb.c ++++ b/drivers/net/wireless/ath/ath11k/ahb.c +@@ -175,8 +175,11 @@ static void __ath11k_ahb_ext_irq_disable(struct ath11k_base *ab) + + ath11k_ahb_ext_grp_disable(irq_grp); + +- napi_synchronize(&irq_grp->napi); +- napi_disable(&irq_grp->napi); ++ if (irq_grp->napi_enabled) { ++ napi_synchronize(&irq_grp->napi); ++ napi_disable(&irq_grp->napi); ++ irq_grp->napi_enabled = false; ++ } + } + } + +@@ -206,13 +209,13 @@ static void ath11k_ahb_clearbit32(struct ath11k_base *ab, u8 bit, u32 offset) + + static void ath11k_ahb_ce_irq_enable(struct ath11k_base *ab, u16 ce_id) + { +- const struct ce_pipe_config *ce_config; ++ const struct ce_attr *ce_attr; + +- ce_config = &ab->hw_params.target_ce_config[ce_id]; +- if (__le32_to_cpu(ce_config->pipedir) & PIPEDIR_OUT) ++ ce_attr = &ab->hw_params.host_ce_config[ce_id]; ++ if (ce_attr->src_nentries) + ath11k_ahb_setbit32(ab, ce_id, CE_HOST_IE_ADDRESS); + +- if (__le32_to_cpu(ce_config->pipedir) & PIPEDIR_IN) { ++ if (ce_attr->dest_nentries) { + ath11k_ahb_setbit32(ab, ce_id, CE_HOST_IE_2_ADDRESS); + ath11k_ahb_setbit32(ab, ce_id + CE_HOST_IE_3_SHIFT, + CE_HOST_IE_3_ADDRESS); +@@ -221,13 +224,13 @@ static void ath11k_ahb_ce_irq_enable(struct ath11k_base *ab, u16 ce_id) + + static void ath11k_ahb_ce_irq_disable(struct ath11k_base *ab, u16 ce_id) + { +- const struct ce_pipe_config *ce_config; ++ const struct ce_attr *ce_attr; + +- ce_config = &ab->hw_params.target_ce_config[ce_id]; +- if (__le32_to_cpu(ce_config->pipedir) & PIPEDIR_OUT) ++ ce_attr = &ab->hw_params.host_ce_config[ce_id]; ++ if (ce_attr->src_nentries) + ath11k_ahb_clearbit32(ab, ce_id, CE_HOST_IE_ADDRESS); + +- if (__le32_to_cpu(ce_config->pipedir) & PIPEDIR_IN) { ++ if (ce_attr->dest_nentries) { + ath11k_ahb_clearbit32(ab, ce_id, CE_HOST_IE_2_ADDRESS); + ath11k_ahb_clearbit32(ab, ce_id + CE_HOST_IE_3_SHIFT, + CE_HOST_IE_3_ADDRESS); +@@ -300,7 +303,10 @@ static void ath11k_ahb_ext_irq_enable(struct ath11k_base *ab) + for (i = 0; i < ATH11K_EXT_IRQ_GRP_NUM_MAX; i++) { + struct ath11k_ext_irq_grp *irq_grp = &ab->ext_irq_grp[i]; + +- napi_enable(&irq_grp->napi); ++ if (!irq_grp->napi_enabled) { ++ napi_enable(&irq_grp->napi); ++ irq_grp->napi_enabled = true; ++ } + ath11k_ahb_ext_grp_enable(irq_grp); + } + } +diff --git a/drivers/net/wireless/ath/ath11k/core.c b/drivers/net/wireless/ath/ath11k/core.c +index 969bf1a590d99..7dcf6b13f7949 100644 +--- a/drivers/net/wireless/ath/ath11k/core.c ++++ b/drivers/net/wireless/ath/ath11k/core.c +@@ -347,11 +347,26 @@ static int ath11k_core_create_board_name(struct ath11k_base *ab, char *name, + scnprintf(variant, sizeof(variant), ",variant=%s", + ab->qmi.target.bdf_ext); + +- scnprintf(name, name_len, +- "bus=%s,qmi-chip-id=%d,qmi-board-id=%d%s", +- ath11k_bus_str(ab->hif.bus), +- ab->qmi.target.chip_id, +- ab->qmi.target.board_id, variant); ++ switch (ab->id.bdf_search) { ++ case ATH11K_BDF_SEARCH_BUS_AND_BOARD: ++ scnprintf(name, name_len, ++ "bus=%s,vendor=%04x,device=%04x,subsystem-vendor=%04x,subsystem-device=%04x,qmi-chip-id=%d,qmi-board-id=%d%s", ++ ath11k_bus_str(ab->hif.bus), ++ ab->id.vendor, ab->id.device, ++ ab->id.subsystem_vendor, ++ ab->id.subsystem_device, ++ ab->qmi.target.chip_id, ++ ab->qmi.target.board_id, ++ variant); ++ break; ++ default: ++ scnprintf(name, name_len, ++ "bus=%s,qmi-chip-id=%d,qmi-board-id=%d%s", ++ ath11k_bus_str(ab->hif.bus), ++ ab->qmi.target.chip_id, ++ ab->qmi.target.board_id, variant); ++ break; ++ } + + ath11k_dbg(ab, ATH11K_DBG_BOOT, "boot using board name '%s'\n", name); + +@@ -588,7 +603,7 @@ static int ath11k_core_fetch_board_data_api_1(struct ath11k_base *ab, + return 0; + } + +-#define BOARD_NAME_SIZE 100 ++#define BOARD_NAME_SIZE 200 + int ath11k_core_fetch_bdf(struct ath11k_base *ab, struct ath11k_board_data *bd) + { + char boardname[BOARD_NAME_SIZE]; +diff --git a/drivers/net/wireless/ath/ath11k/core.h b/drivers/net/wireless/ath/ath11k/core.h +index 018fb2385f2a3..caa8f6eba0097 100644 +--- a/drivers/net/wireless/ath/ath11k/core.h ++++ b/drivers/net/wireless/ath/ath11k/core.h +@@ -47,6 +47,11 @@ enum ath11k_supported_bw { + ATH11K_BW_160 = 3, + }; + ++enum ath11k_bdf_search { ++ ATH11K_BDF_SEARCH_DEFAULT, ++ ATH11K_BDF_SEARCH_BUS_AND_BOARD, ++}; ++ + enum wme_ac { + WME_AC_BE, + WME_AC_BK, +@@ -132,6 +137,7 @@ struct ath11k_ext_irq_grp { + u32 num_irq; + u32 grp_id; + u64 timestamp; ++ bool napi_enabled; + struct napi_struct napi; + struct net_device napi_ndev; + }; +@@ -701,7 +707,6 @@ struct ath11k_base { + u32 wlan_init_status; + int irq_num[ATH11K_IRQ_NUM_MAX]; + struct ath11k_ext_irq_grp ext_irq_grp[ATH11K_EXT_IRQ_GRP_NUM_MAX]; +- struct napi_struct *napi; + struct ath11k_targ_cap target_caps; + u32 ext_service_bitmap[WMI_SERVICE_EXT_BM_SIZE]; + bool pdevs_macaddr_valid; +@@ -747,6 +752,14 @@ struct ath11k_base { + + struct completion htc_suspend; + ++ struct { ++ enum ath11k_bdf_search bdf_search; ++ u32 vendor; ++ u32 device; ++ u32 subsystem_vendor; ++ u32 subsystem_device; ++ } id; ++ + /* must be last */ + u8 drv_priv[0] __aligned(sizeof(void *)); + }; +diff --git a/drivers/net/wireless/ath/ath11k/dp.h b/drivers/net/wireless/ath/ath11k/dp.h +index ee768ccce46e1..d3e50e34f23dd 100644 +--- a/drivers/net/wireless/ath/ath11k/dp.h ++++ b/drivers/net/wireless/ath/ath11k/dp.h +@@ -515,7 +515,8 @@ struct htt_ppdu_stats_cfg_cmd { + } __packed; + + #define HTT_PPDU_STATS_CFG_MSG_TYPE GENMASK(7, 0) +-#define HTT_PPDU_STATS_CFG_PDEV_ID GENMASK(15, 8) ++#define HTT_PPDU_STATS_CFG_SOC_STATS BIT(8) ++#define HTT_PPDU_STATS_CFG_PDEV_ID GENMASK(15, 9) + #define HTT_PPDU_STATS_CFG_TLV_TYPE_BITMASK GENMASK(31, 16) + + enum htt_ppdu_stats_tag_type { +diff --git a/drivers/net/wireless/ath/ath11k/dp_tx.c b/drivers/net/wireless/ath/ath11k/dp_tx.c +index 8bba5234f81fc..bb8744ccfa00c 100644 +--- a/drivers/net/wireless/ath/ath11k/dp_tx.c ++++ b/drivers/net/wireless/ath/ath11k/dp_tx.c +@@ -895,7 +895,7 @@ int ath11k_dp_tx_htt_h2t_ppdu_stats_req(struct ath11k *ar, u32 mask) + cmd->msg = FIELD_PREP(HTT_PPDU_STATS_CFG_MSG_TYPE, + HTT_H2T_MSG_TYPE_PPDU_STATS_CFG); + +- pdev_mask = 1 << (i + 1); ++ pdev_mask = 1 << (ar->pdev_idx + i); + cmd->msg |= FIELD_PREP(HTT_PPDU_STATS_CFG_PDEV_ID, pdev_mask); + cmd->msg |= FIELD_PREP(HTT_PPDU_STATS_CFG_TLV_TYPE_BITMASK, mask); + +diff --git a/drivers/net/wireless/ath/ath11k/hal.c b/drivers/net/wireless/ath/ath11k/hal.c +index eaa0edca55761..5dbf5596c9e8e 100644 +--- a/drivers/net/wireless/ath/ath11k/hal.c ++++ b/drivers/net/wireless/ath/ath11k/hal.c +@@ -947,6 +947,7 @@ int ath11k_hal_srng_setup(struct ath11k_base *ab, enum hal_ring_type type, + srng->msi_data = params->msi_data; + srng->initialized = 1; + spin_lock_init(&srng->lock); ++ lockdep_set_class(&srng->lock, hal->srng_key + ring_id); + + for (i = 0; i < HAL_SRNG_NUM_REG_GRP; i++) { + srng->hwreg_base[i] = srng_config->reg_start[i] + +@@ -1233,6 +1234,24 @@ static int ath11k_hal_srng_create_config(struct ath11k_base *ab) + return 0; + } + ++static void ath11k_hal_register_srng_key(struct ath11k_base *ab) ++{ ++ struct ath11k_hal *hal = &ab->hal; ++ u32 ring_id; ++ ++ for (ring_id = 0; ring_id < HAL_SRNG_RING_ID_MAX; ring_id++) ++ lockdep_register_key(hal->srng_key + ring_id); ++} ++ ++static void ath11k_hal_unregister_srng_key(struct ath11k_base *ab) ++{ ++ struct ath11k_hal *hal = &ab->hal; ++ u32 ring_id; ++ ++ for (ring_id = 0; ring_id < HAL_SRNG_RING_ID_MAX; ring_id++) ++ lockdep_unregister_key(hal->srng_key + ring_id); ++} ++ + int ath11k_hal_srng_init(struct ath11k_base *ab) + { + struct ath11k_hal *hal = &ab->hal; +@@ -1252,6 +1271,8 @@ int ath11k_hal_srng_init(struct ath11k_base *ab) + if (ret) + goto err_free_cont_rdp; + ++ ath11k_hal_register_srng_key(ab); ++ + return 0; + + err_free_cont_rdp: +@@ -1266,6 +1287,7 @@ void ath11k_hal_srng_deinit(struct ath11k_base *ab) + { + struct ath11k_hal *hal = &ab->hal; + ++ ath11k_hal_unregister_srng_key(ab); + ath11k_hal_free_cont_rdp(ab); + ath11k_hal_free_cont_wrp(ab); + kfree(hal->srng_config); +diff --git a/drivers/net/wireless/ath/ath11k/hal.h b/drivers/net/wireless/ath/ath11k/hal.h +index 35ed3a14e200a..7fdcd8bbf7e98 100644 +--- a/drivers/net/wireless/ath/ath11k/hal.h ++++ b/drivers/net/wireless/ath/ath11k/hal.h +@@ -901,6 +901,8 @@ struct ath11k_hal { + /* shadow register configuration */ + u32 shadow_reg_addr[HAL_SHADOW_NUM_REGS]; + int num_shadow_reg_configured; ++ ++ struct lock_class_key srng_key[HAL_SRNG_RING_ID_MAX]; + }; + + u32 ath11k_hal_reo_qdesc_size(u32 ba_window_size, u8 tid); +diff --git a/drivers/net/wireless/ath/ath11k/hw.c b/drivers/net/wireless/ath/ath11k/hw.c +index d9596903b0a58..3e92cc7cfe4c9 100644 +--- a/drivers/net/wireless/ath/ath11k/hw.c ++++ b/drivers/net/wireless/ath/ath11k/hw.c +@@ -1015,8 +1015,6 @@ const struct ath11k_hw_ring_mask ath11k_hw_ring_mask_ipq8074 = { + const struct ath11k_hw_ring_mask ath11k_hw_ring_mask_qca6390 = { + .tx = { + ATH11K_TX_RING_MASK_0, +- ATH11K_TX_RING_MASK_1, +- ATH11K_TX_RING_MASK_2, + }, + .rx_mon_status = { + 0, 0, 0, 0, +diff --git a/drivers/net/wireless/ath/ath11k/mac.c b/drivers/net/wireless/ath/ath11k/mac.c +index 89a64ebd620f3..3834be1587057 100644 +--- a/drivers/net/wireless/ath/ath11k/mac.c ++++ b/drivers/net/wireless/ath/ath11k/mac.c +@@ -1,6 +1,7 @@ + // SPDX-License-Identifier: BSD-3-Clause-Clear + /* + * Copyright (c) 2018-2019 The Linux Foundation. All rights reserved. ++ * Copyright (c) 2021 Qualcomm Innovation Center, Inc. All rights reserved. + */ + + #include <net/mac80211.h> +@@ -767,11 +768,15 @@ static int ath11k_mac_setup_bcn_tmpl(struct ath11k_vif *arvif) + + if (cfg80211_find_ie(WLAN_EID_RSN, ies, (skb_tail_pointer(bcn) - ies))) + arvif->rsnie_present = true; ++ else ++ arvif->rsnie_present = false; + + if (cfg80211_find_vendor_ie(WLAN_OUI_MICROSOFT, + WLAN_OUI_TYPE_MICROSOFT_WPA, + ies, (skb_tail_pointer(bcn) - ies))) + arvif->wpaie_present = true; ++ else ++ arvif->wpaie_present = false; + + ret = ath11k_wmi_bcn_tmpl(ar, arvif->vdev_id, &offs, bcn); + +@@ -2576,9 +2581,12 @@ static int ath11k_mac_op_hw_scan(struct ieee80211_hw *hw, + arg.scan_id = ATH11K_SCAN_ID; + + if (req->ie_len) { ++ arg.extraie.ptr = kmemdup(req->ie, req->ie_len, GFP_KERNEL); ++ if (!arg.extraie.ptr) { ++ ret = -ENOMEM; ++ goto exit; ++ } + arg.extraie.len = req->ie_len; +- arg.extraie.ptr = kzalloc(req->ie_len, GFP_KERNEL); +- memcpy(arg.extraie.ptr, req->ie, req->ie_len); + } + + if (req->n_ssids) { +@@ -2655,9 +2663,7 @@ static int ath11k_install_key(struct ath11k_vif *arvif, + return 0; + + if (cmd == DISABLE_KEY) { +- /* TODO: Check if FW expects value other than NONE for del */ +- /* arg.key_cipher = WMI_CIPHER_NONE; */ +- arg.key_len = 0; ++ arg.key_cipher = WMI_CIPHER_NONE; + arg.key_data = NULL; + goto install; + } +@@ -2789,7 +2795,7 @@ static int ath11k_mac_op_set_key(struct ieee80211_hw *hw, enum set_key_cmd cmd, + /* flush the fragments cache during key (re)install to + * ensure all frags in the new frag list belong to the same key. + */ +- if (peer && cmd == SET_KEY) ++ if (peer && sta && cmd == SET_KEY) + ath11k_peer_frags_flush(ar, peer); + spin_unlock_bh(&ab->base_lock); + +@@ -4131,23 +4137,32 @@ static int __ath11k_set_antenna(struct ath11k *ar, u32 tx_ant, u32 rx_ant) + return 0; + } + +-int ath11k_mac_tx_mgmt_pending_free(int buf_id, void *skb, void *ctx) ++static void ath11k_mac_tx_mgmt_free(struct ath11k *ar, int buf_id) + { +- struct sk_buff *msdu = skb; ++ struct sk_buff *msdu; + struct ieee80211_tx_info *info; +- struct ath11k *ar = ctx; +- struct ath11k_base *ab = ar->ab; + + spin_lock_bh(&ar->txmgmt_idr_lock); +- idr_remove(&ar->txmgmt_idr, buf_id); ++ msdu = idr_remove(&ar->txmgmt_idr, buf_id); + spin_unlock_bh(&ar->txmgmt_idr_lock); +- dma_unmap_single(ab->dev, ATH11K_SKB_CB(msdu)->paddr, msdu->len, ++ ++ if (!msdu) ++ return; ++ ++ dma_unmap_single(ar->ab->dev, ATH11K_SKB_CB(msdu)->paddr, msdu->len, + DMA_TO_DEVICE); + + info = IEEE80211_SKB_CB(msdu); + memset(&info->status, 0, sizeof(info->status)); + + ieee80211_free_txskb(ar->hw, msdu); ++} ++ ++int ath11k_mac_tx_mgmt_pending_free(int buf_id, void *skb, void *ctx) ++{ ++ struct ath11k *ar = ctx; ++ ++ ath11k_mac_tx_mgmt_free(ar, buf_id); + + return 0; + } +@@ -4156,17 +4171,10 @@ static int ath11k_mac_vif_txmgmt_idr_remove(int buf_id, void *skb, void *ctx) + { + struct ieee80211_vif *vif = ctx; + struct ath11k_skb_cb *skb_cb = ATH11K_SKB_CB((struct sk_buff *)skb); +- struct sk_buff *msdu = skb; + struct ath11k *ar = skb_cb->ar; +- struct ath11k_base *ab = ar->ab; + +- if (skb_cb->vif == vif) { +- spin_lock_bh(&ar->txmgmt_idr_lock); +- idr_remove(&ar->txmgmt_idr, buf_id); +- spin_unlock_bh(&ar->txmgmt_idr_lock); +- dma_unmap_single(ab->dev, skb_cb->paddr, msdu->len, +- DMA_TO_DEVICE); +- } ++ if (skb_cb->vif == vif) ++ ath11k_mac_tx_mgmt_free(ar, buf_id); + + return 0; + } +@@ -4181,6 +4189,8 @@ static int ath11k_mac_mgmt_tx_wmi(struct ath11k *ar, struct ath11k_vif *arvif, + int buf_id; + int ret; + ++ ATH11K_SKB_CB(skb)->ar = ar; ++ + spin_lock_bh(&ar->txmgmt_idr_lock); + buf_id = idr_alloc(&ar->txmgmt_idr, skb, 0, + ATH11K_TX_MGMT_NUM_PENDING_MAX, GFP_ATOMIC); +diff --git a/drivers/net/wireless/ath/ath11k/pci.c b/drivers/net/wireless/ath/ath11k/pci.c +index 5abb38cc3b55f..54ce08f1c6e0c 100644 +--- a/drivers/net/wireless/ath/ath11k/pci.c ++++ b/drivers/net/wireless/ath/ath11k/pci.c +@@ -632,8 +632,11 @@ static void __ath11k_pci_ext_irq_disable(struct ath11k_base *sc) + + ath11k_pci_ext_grp_disable(irq_grp); + +- napi_synchronize(&irq_grp->napi); +- napi_disable(&irq_grp->napi); ++ if (irq_grp->napi_enabled) { ++ napi_synchronize(&irq_grp->napi); ++ napi_disable(&irq_grp->napi); ++ irq_grp->napi_enabled = false; ++ } + } + } + +@@ -652,7 +655,10 @@ static void ath11k_pci_ext_irq_enable(struct ath11k_base *ab) + for (i = 0; i < ATH11K_EXT_IRQ_GRP_NUM_MAX; i++) { + struct ath11k_ext_irq_grp *irq_grp = &ab->ext_irq_grp[i]; + +- napi_enable(&irq_grp->napi); ++ if (!irq_grp->napi_enabled) { ++ napi_enable(&irq_grp->napi); ++ irq_grp->napi_enabled = true; ++ } + ath11k_pci_ext_grp_enable(irq_grp); + } + } +@@ -1218,6 +1224,15 @@ static int ath11k_pci_probe(struct pci_dev *pdev, + goto err_free_core; + } + ++ ath11k_dbg(ab, ATH11K_DBG_BOOT, "pci probe %04x:%04x %04x:%04x\n", ++ pdev->vendor, pdev->device, ++ pdev->subsystem_vendor, pdev->subsystem_device); ++ ++ ab->id.vendor = pdev->vendor; ++ ab->id.device = pdev->device; ++ ab->id.subsystem_vendor = pdev->subsystem_vendor; ++ ab->id.subsystem_device = pdev->subsystem_device; ++ + switch (pci_dev->device) { + case QCA6390_DEVICE_ID: + ath11k_pci_read_hw_version(ab, &soc_hw_version_major, +@@ -1240,6 +1255,7 @@ static int ath11k_pci_probe(struct pci_dev *pdev, + ab->hw_rev = ATH11K_HW_QCN9074_HW10; + break; + case WCN6855_DEVICE_ID: ++ ab->id.bdf_search = ATH11K_BDF_SEARCH_BUS_AND_BOARD; + ath11k_pci_read_hw_version(ab, &soc_hw_version_major, + &soc_hw_version_minor); + switch (soc_hw_version_major) { +diff --git a/drivers/net/wireless/ath/ath11k/qmi.c b/drivers/net/wireless/ath/ath11k/qmi.c +index 4c5071b7d11dc..e4a65513a1bfd 100644 +--- a/drivers/net/wireless/ath/ath11k/qmi.c ++++ b/drivers/net/wireless/ath/ath11k/qmi.c +@@ -1770,7 +1770,7 @@ static int ath11k_qmi_alloc_target_mem_chunk(struct ath11k_base *ab) + chunk->vaddr = dma_alloc_coherent(ab->dev, + chunk->size, + &chunk->paddr, +- GFP_KERNEL); ++ GFP_KERNEL | __GFP_NOWARN); + if (!chunk->vaddr) { + if (ab->qmi.mem_seg_count <= ATH11K_QMI_FW_MEM_REQ_SEGMENT_CNT) { + ath11k_dbg(ab, ATH11K_DBG_QMI, +diff --git a/drivers/net/wireless/ath/ath11k/reg.c b/drivers/net/wireless/ath/ath11k/reg.c +index 92c59009a8ac2..f793324ad0b73 100644 +--- a/drivers/net/wireless/ath/ath11k/reg.c ++++ b/drivers/net/wireless/ath/ath11k/reg.c +@@ -459,6 +459,9 @@ ath11k_reg_adjust_bw(u16 start_freq, u16 end_freq, u16 max_bw) + { + u16 bw; + ++ if (end_freq <= start_freq) ++ return 0; ++ + bw = end_freq - start_freq; + bw = min_t(u16, bw, max_bw); + +@@ -466,8 +469,10 @@ ath11k_reg_adjust_bw(u16 start_freq, u16 end_freq, u16 max_bw) + bw = 80; + else if (bw >= 40 && bw < 80) + bw = 40; +- else if (bw < 40) ++ else if (bw >= 20 && bw < 40) + bw = 20; ++ else ++ bw = 0; + + return bw; + } +@@ -491,73 +496,77 @@ ath11k_reg_update_weather_radar_band(struct ath11k_base *ab, + struct cur_reg_rule *reg_rule, + u8 *rule_idx, u32 flags, u16 max_bw) + { ++ u32 start_freq; + u32 end_freq; + u16 bw; + u8 i; + + i = *rule_idx; + ++ /* there might be situations when even the input rule must be dropped */ ++ i--; ++ ++ /* frequencies below weather radar */ + bw = ath11k_reg_adjust_bw(reg_rule->start_freq, + ETSI_WEATHER_RADAR_BAND_LOW, max_bw); ++ if (bw > 0) { ++ i++; + +- ath11k_reg_update_rule(regd->reg_rules + i, reg_rule->start_freq, +- ETSI_WEATHER_RADAR_BAND_LOW, bw, +- reg_rule->ant_gain, reg_rule->reg_power, +- flags); ++ ath11k_reg_update_rule(regd->reg_rules + i, ++ reg_rule->start_freq, ++ ETSI_WEATHER_RADAR_BAND_LOW, bw, ++ reg_rule->ant_gain, reg_rule->reg_power, ++ flags); + +- ath11k_dbg(ab, ATH11K_DBG_REG, +- "\t%d. (%d - %d @ %d) (%d, %d) (%d ms) (FLAGS %d)\n", +- i + 1, reg_rule->start_freq, ETSI_WEATHER_RADAR_BAND_LOW, +- bw, reg_rule->ant_gain, reg_rule->reg_power, +- regd->reg_rules[i].dfs_cac_ms, +- flags); +- +- if (reg_rule->end_freq > ETSI_WEATHER_RADAR_BAND_HIGH) +- end_freq = ETSI_WEATHER_RADAR_BAND_HIGH; +- else +- end_freq = reg_rule->end_freq; ++ ath11k_dbg(ab, ATH11K_DBG_REG, ++ "\t%d. (%d - %d @ %d) (%d, %d) (%d ms) (FLAGS %d)\n", ++ i + 1, reg_rule->start_freq, ++ ETSI_WEATHER_RADAR_BAND_LOW, bw, reg_rule->ant_gain, ++ reg_rule->reg_power, regd->reg_rules[i].dfs_cac_ms, ++ flags); ++ } + +- bw = ath11k_reg_adjust_bw(ETSI_WEATHER_RADAR_BAND_LOW, end_freq, +- max_bw); ++ /* weather radar frequencies */ ++ start_freq = max_t(u32, reg_rule->start_freq, ++ ETSI_WEATHER_RADAR_BAND_LOW); ++ end_freq = min_t(u32, reg_rule->end_freq, ETSI_WEATHER_RADAR_BAND_HIGH); + +- i++; ++ bw = ath11k_reg_adjust_bw(start_freq, end_freq, max_bw); ++ if (bw > 0) { ++ i++; + +- ath11k_reg_update_rule(regd->reg_rules + i, +- ETSI_WEATHER_RADAR_BAND_LOW, end_freq, bw, +- reg_rule->ant_gain, reg_rule->reg_power, +- flags); ++ ath11k_reg_update_rule(regd->reg_rules + i, start_freq, ++ end_freq, bw, reg_rule->ant_gain, ++ reg_rule->reg_power, flags); + +- regd->reg_rules[i].dfs_cac_ms = ETSI_WEATHER_RADAR_BAND_CAC_TIMEOUT; ++ regd->reg_rules[i].dfs_cac_ms = ETSI_WEATHER_RADAR_BAND_CAC_TIMEOUT; + +- ath11k_dbg(ab, ATH11K_DBG_REG, +- "\t%d. (%d - %d @ %d) (%d, %d) (%d ms) (FLAGS %d)\n", +- i + 1, ETSI_WEATHER_RADAR_BAND_LOW, end_freq, +- bw, reg_rule->ant_gain, reg_rule->reg_power, +- regd->reg_rules[i].dfs_cac_ms, +- flags); +- +- if (end_freq == reg_rule->end_freq) { +- regd->n_reg_rules--; +- *rule_idx = i; +- return; ++ ath11k_dbg(ab, ATH11K_DBG_REG, ++ "\t%d. (%d - %d @ %d) (%d, %d) (%d ms) (FLAGS %d)\n", ++ i + 1, start_freq, end_freq, bw, ++ reg_rule->ant_gain, reg_rule->reg_power, ++ regd->reg_rules[i].dfs_cac_ms, flags); + } + ++ /* frequencies above weather radar */ + bw = ath11k_reg_adjust_bw(ETSI_WEATHER_RADAR_BAND_HIGH, + reg_rule->end_freq, max_bw); ++ if (bw > 0) { ++ i++; + +- i++; +- +- ath11k_reg_update_rule(regd->reg_rules + i, ETSI_WEATHER_RADAR_BAND_HIGH, +- reg_rule->end_freq, bw, +- reg_rule->ant_gain, reg_rule->reg_power, +- flags); ++ ath11k_reg_update_rule(regd->reg_rules + i, ++ ETSI_WEATHER_RADAR_BAND_HIGH, ++ reg_rule->end_freq, bw, ++ reg_rule->ant_gain, reg_rule->reg_power, ++ flags); + +- ath11k_dbg(ab, ATH11K_DBG_REG, +- "\t%d. (%d - %d @ %d) (%d, %d) (%d ms) (FLAGS %d)\n", +- i + 1, ETSI_WEATHER_RADAR_BAND_HIGH, reg_rule->end_freq, +- bw, reg_rule->ant_gain, reg_rule->reg_power, +- regd->reg_rules[i].dfs_cac_ms, +- flags); ++ ath11k_dbg(ab, ATH11K_DBG_REG, ++ "\t%d. (%d - %d @ %d) (%d, %d) (%d ms) (FLAGS %d)\n", ++ i + 1, ETSI_WEATHER_RADAR_BAND_HIGH, ++ reg_rule->end_freq, bw, reg_rule->ant_gain, ++ reg_rule->reg_power, regd->reg_rules[i].dfs_cac_ms, ++ flags); ++ } + + *rule_idx = i; + } +diff --git a/drivers/net/wireless/ath/ath11k/wmi.c b/drivers/net/wireless/ath/ath11k/wmi.c +index c22ec921b2e97..b11070cf159cc 100644 +--- a/drivers/net/wireless/ath/ath11k/wmi.c ++++ b/drivers/net/wireless/ath/ath11k/wmi.c +@@ -1671,7 +1671,8 @@ int ath11k_wmi_vdev_install_key(struct ath11k *ar, + tlv = (struct wmi_tlv *)(skb->data + sizeof(*cmd)); + tlv->header = FIELD_PREP(WMI_TLV_TAG, WMI_TAG_ARRAY_BYTE) | + FIELD_PREP(WMI_TLV_LEN, key_len_aligned); +- memcpy(tlv->value, (u8 *)arg->key_data, key_len_aligned); ++ if (arg->key_data) ++ memcpy(tlv->value, (u8 *)arg->key_data, key_len_aligned); + + ret = ath11k_wmi_cmd_send(wmi, skb, WMI_VDEV_INSTALL_KEY_CMDID); + if (ret) { +@@ -5852,7 +5853,7 @@ static int ath11k_reg_chan_list_event(struct ath11k_base *ab, struct sk_buff *sk + ar = ab->pdevs[pdev_idx].ar; + kfree(ab->new_regd[pdev_idx]); + ab->new_regd[pdev_idx] = regd; +- ieee80211_queue_work(ar->hw, &ar->regd_update_work); ++ queue_work(ab->workqueue, &ar->regd_update_work); + } else { + /* This regd would be applied during mac registration and is + * held constant throughout for regd intersection purpose +diff --git a/drivers/net/wireless/ath/ath9k/hif_usb.c b/drivers/net/wireless/ath/ath9k/hif_usb.c +index 860da13bfb6ac..f06eec99de688 100644 +--- a/drivers/net/wireless/ath/ath9k/hif_usb.c ++++ b/drivers/net/wireless/ath/ath9k/hif_usb.c +@@ -590,6 +590,13 @@ static void ath9k_hif_usb_rx_stream(struct hif_device_usb *hif_dev, + return; + } + ++ if (pkt_len > 2 * MAX_RX_BUF_SIZE) { ++ dev_err(&hif_dev->udev->dev, ++ "ath9k_htc: invalid pkt_len (%x)\n", pkt_len); ++ RX_STAT_INC(skb_dropped); ++ return; ++ } ++ + pad_len = 4 - (pkt_len & 0x3); + if (pad_len == 4) + pad_len = 0; +diff --git a/drivers/net/wireless/ath/ath9k/htc.h b/drivers/net/wireless/ath/ath9k/htc.h +index 0a1634238e673..6b45e63fae4ba 100644 +--- a/drivers/net/wireless/ath/ath9k/htc.h ++++ b/drivers/net/wireless/ath/ath9k/htc.h +@@ -281,6 +281,7 @@ struct ath9k_htc_rxbuf { + struct ath9k_htc_rx { + struct list_head rxbuf; + spinlock_t rxbuflock; ++ bool initialized; + }; + + #define ATH9K_HTC_TX_CLEANUP_INTERVAL 50 /* ms */ +@@ -305,6 +306,7 @@ struct ath9k_htc_tx { + DECLARE_BITMAP(tx_slot, MAX_TX_BUF_NUM); + struct timer_list cleanup_timer; + spinlock_t tx_lock; ++ bool initialized; + }; + + struct ath9k_htc_tx_ctl { +diff --git a/drivers/net/wireless/ath/ath9k/htc_drv_txrx.c b/drivers/net/wireless/ath/ath9k/htc_drv_txrx.c +index 8e69e8989f6d3..6a850a0bfa8ad 100644 +--- a/drivers/net/wireless/ath/ath9k/htc_drv_txrx.c ++++ b/drivers/net/wireless/ath/ath9k/htc_drv_txrx.c +@@ -813,6 +813,11 @@ int ath9k_tx_init(struct ath9k_htc_priv *priv) + skb_queue_head_init(&priv->tx.data_vi_queue); + skb_queue_head_init(&priv->tx.data_vo_queue); + skb_queue_head_init(&priv->tx.tx_failed); ++ ++ /* Allow ath9k_wmi_event_tasklet(WMI_TXSTATUS_EVENTID) to operate. */ ++ smp_wmb(); ++ priv->tx.initialized = true; ++ + return 0; + } + +@@ -1130,6 +1135,10 @@ void ath9k_htc_rxep(void *drv_priv, struct sk_buff *skb, + struct ath9k_htc_rxbuf *rxbuf = NULL, *tmp_buf = NULL; + unsigned long flags; + ++ /* Check if ath9k_rx_init() completed. */ ++ if (!data_race(priv->rx.initialized)) ++ goto err; ++ + spin_lock_irqsave(&priv->rx.rxbuflock, flags); + list_for_each_entry(tmp_buf, &priv->rx.rxbuf, list) { + if (!tmp_buf->in_process) { +@@ -1185,6 +1194,10 @@ int ath9k_rx_init(struct ath9k_htc_priv *priv) + list_add_tail(&rxbuf->list, &priv->rx.rxbuf); + } + ++ /* Allow ath9k_htc_rxep() to operate. */ ++ smp_wmb(); ++ priv->rx.initialized = true; ++ + return 0; + + err: +diff --git a/drivers/net/wireless/ath/ath9k/wmi.c b/drivers/net/wireless/ath/ath9k/wmi.c +index fe29ad4b9023c..f315c54bd3ac0 100644 +--- a/drivers/net/wireless/ath/ath9k/wmi.c ++++ b/drivers/net/wireless/ath/ath9k/wmi.c +@@ -169,6 +169,10 @@ void ath9k_wmi_event_tasklet(struct tasklet_struct *t) + &wmi->drv_priv->fatal_work); + break; + case WMI_TXSTATUS_EVENTID: ++ /* Check if ath9k_tx_init() completed. */ ++ if (!data_race(priv->tx.initialized)) ++ break; ++ + spin_lock_bh(&priv->tx.tx_lock); + if (priv->tx.flags & ATH9K_HTC_OP_TX_DRAIN) { + spin_unlock_bh(&priv->tx.tx_lock); +diff --git a/drivers/net/wireless/ath/wcn36xx/dxe.c b/drivers/net/wireless/ath/wcn36xx/dxe.c +index aff04ef662663..e1a35c2eadb6c 100644 +--- a/drivers/net/wireless/ath/wcn36xx/dxe.c ++++ b/drivers/net/wireless/ath/wcn36xx/dxe.c +@@ -272,6 +272,21 @@ static int wcn36xx_dxe_enable_ch_int(struct wcn36xx *wcn, u16 wcn_ch) + return 0; + } + ++static void wcn36xx_dxe_disable_ch_int(struct wcn36xx *wcn, u16 wcn_ch) ++{ ++ int reg_data = 0; ++ ++ wcn36xx_dxe_read_register(wcn, ++ WCN36XX_DXE_INT_MASK_REG, ++ ®_data); ++ ++ reg_data &= ~wcn_ch; ++ ++ wcn36xx_dxe_write_register(wcn, ++ WCN36XX_DXE_INT_MASK_REG, ++ (int)reg_data); ++} ++ + static int wcn36xx_dxe_fill_skb(struct device *dev, + struct wcn36xx_dxe_ctl *ctl, + gfp_t gfp) +@@ -869,7 +884,6 @@ int wcn36xx_dxe_init(struct wcn36xx *wcn) + WCN36XX_DXE_WQ_TX_L); + + wcn36xx_dxe_read_register(wcn, WCN36XX_DXE_REG_CH_EN, ®_data); +- wcn36xx_dxe_enable_ch_int(wcn, WCN36XX_INT_MASK_CHAN_TX_L); + + /***************************************/ + /* Init descriptors for TX HIGH channel */ +@@ -893,9 +907,6 @@ int wcn36xx_dxe_init(struct wcn36xx *wcn) + + wcn36xx_dxe_read_register(wcn, WCN36XX_DXE_REG_CH_EN, ®_data); + +- /* Enable channel interrupts */ +- wcn36xx_dxe_enable_ch_int(wcn, WCN36XX_INT_MASK_CHAN_TX_H); +- + /***************************************/ + /* Init descriptors for RX LOW channel */ + /***************************************/ +@@ -905,7 +916,6 @@ int wcn36xx_dxe_init(struct wcn36xx *wcn) + goto out_err_rxl_ch; + } + +- + /* For RX we need to preallocated buffers */ + wcn36xx_dxe_ch_alloc_skb(wcn, &wcn->dxe_rx_l_ch); + +@@ -928,9 +938,6 @@ int wcn36xx_dxe_init(struct wcn36xx *wcn) + WCN36XX_DXE_REG_CTL_RX_L, + WCN36XX_DXE_CH_DEFAULT_CTL_RX_L); + +- /* Enable channel interrupts */ +- wcn36xx_dxe_enable_ch_int(wcn, WCN36XX_INT_MASK_CHAN_RX_L); +- + /***************************************/ + /* Init descriptors for RX HIGH channel */ + /***************************************/ +@@ -962,15 +969,18 @@ int wcn36xx_dxe_init(struct wcn36xx *wcn) + WCN36XX_DXE_REG_CTL_RX_H, + WCN36XX_DXE_CH_DEFAULT_CTL_RX_H); + +- /* Enable channel interrupts */ +- wcn36xx_dxe_enable_ch_int(wcn, WCN36XX_INT_MASK_CHAN_RX_H); +- + ret = wcn36xx_dxe_request_irqs(wcn); + if (ret < 0) + goto out_err_irq; + + timer_setup(&wcn->tx_ack_timer, wcn36xx_dxe_tx_timer, 0); + ++ /* Enable channel interrupts */ ++ wcn36xx_dxe_enable_ch_int(wcn, WCN36XX_INT_MASK_CHAN_TX_L); ++ wcn36xx_dxe_enable_ch_int(wcn, WCN36XX_INT_MASK_CHAN_TX_H); ++ wcn36xx_dxe_enable_ch_int(wcn, WCN36XX_INT_MASK_CHAN_RX_L); ++ wcn36xx_dxe_enable_ch_int(wcn, WCN36XX_INT_MASK_CHAN_RX_H); ++ + return 0; + + out_err_irq: +@@ -987,6 +997,14 @@ out_err_txh_ch: + + void wcn36xx_dxe_deinit(struct wcn36xx *wcn) + { ++ int reg_data = 0; ++ ++ /* Disable channel interrupts */ ++ wcn36xx_dxe_disable_ch_int(wcn, WCN36XX_INT_MASK_CHAN_RX_H); ++ wcn36xx_dxe_disable_ch_int(wcn, WCN36XX_INT_MASK_CHAN_RX_L); ++ wcn36xx_dxe_disable_ch_int(wcn, WCN36XX_INT_MASK_CHAN_TX_H); ++ wcn36xx_dxe_disable_ch_int(wcn, WCN36XX_INT_MASK_CHAN_TX_L); ++ + free_irq(wcn->tx_irq, wcn); + free_irq(wcn->rx_irq, wcn); + del_timer(&wcn->tx_ack_timer); +@@ -996,6 +1014,15 @@ void wcn36xx_dxe_deinit(struct wcn36xx *wcn) + wcn->tx_ack_skb = NULL; + } + ++ /* Put the DXE block into reset before freeing memory */ ++ reg_data = WCN36XX_DXE_REG_RESET; ++ wcn36xx_dxe_write_register(wcn, WCN36XX_DXE_REG_CSR_RESET, reg_data); ++ + wcn36xx_dxe_ch_free_skbs(wcn, &wcn->dxe_rx_l_ch); + wcn36xx_dxe_ch_free_skbs(wcn, &wcn->dxe_rx_h_ch); ++ ++ wcn36xx_dxe_deinit_descs(wcn->dev, &wcn->dxe_tx_l_ch); ++ wcn36xx_dxe_deinit_descs(wcn->dev, &wcn->dxe_tx_h_ch); ++ wcn36xx_dxe_deinit_descs(wcn->dev, &wcn->dxe_rx_l_ch); ++ wcn36xx_dxe_deinit_descs(wcn->dev, &wcn->dxe_rx_h_ch); + } +diff --git a/drivers/net/wireless/ath/wcn36xx/main.c b/drivers/net/wireless/ath/wcn36xx/main.c +index 5d82aca370a72..cf9e1396bd046 100644 +--- a/drivers/net/wireless/ath/wcn36xx/main.c ++++ b/drivers/net/wireless/ath/wcn36xx/main.c +@@ -400,6 +400,7 @@ static void wcn36xx_change_opchannel(struct wcn36xx *wcn, int ch) + static int wcn36xx_config(struct ieee80211_hw *hw, u32 changed) + { + struct wcn36xx *wcn = hw->priv; ++ int ret; + + wcn36xx_dbg(WCN36XX_DBG_MAC, "mac config changed 0x%08x\n", changed); + +@@ -415,17 +416,31 @@ static int wcn36xx_config(struct ieee80211_hw *hw, u32 changed) + * want to receive/transmit regular data packets, then + * simply stop the scan session and exit PS mode. + */ +- wcn36xx_smd_finish_scan(wcn, HAL_SYS_MODE_SCAN, +- wcn->sw_scan_vif); +- wcn->sw_scan_channel = 0; ++ if (wcn->sw_scan_channel) ++ wcn36xx_smd_end_scan(wcn, wcn->sw_scan_channel); ++ if (wcn->sw_scan_init) { ++ wcn36xx_smd_finish_scan(wcn, HAL_SYS_MODE_SCAN, ++ wcn->sw_scan_vif); ++ } + } else if (wcn->sw_scan) { + /* A scan is ongoing, do not change the operating + * channel, but start a scan session on the channel. + */ +- wcn36xx_smd_init_scan(wcn, HAL_SYS_MODE_SCAN, +- wcn->sw_scan_vif); ++ if (wcn->sw_scan_channel) ++ wcn36xx_smd_end_scan(wcn, wcn->sw_scan_channel); ++ if (!wcn->sw_scan_init) { ++ /* This can fail if we are unable to notify the ++ * operating channel. ++ */ ++ ret = wcn36xx_smd_init_scan(wcn, ++ HAL_SYS_MODE_SCAN, ++ wcn->sw_scan_vif); ++ if (ret) { ++ mutex_unlock(&wcn->conf_mutex); ++ return -EIO; ++ } ++ } + wcn36xx_smd_start_scan(wcn, ch); +- wcn->sw_scan_channel = ch; + } else { + wcn36xx_change_opchannel(wcn, ch); + } +@@ -713,7 +728,12 @@ static void wcn36xx_sw_scan_complete(struct ieee80211_hw *hw, + struct wcn36xx *wcn = hw->priv; + + /* ensure that any scan session is finished */ +- wcn36xx_smd_finish_scan(wcn, HAL_SYS_MODE_SCAN, wcn->sw_scan_vif); ++ if (wcn->sw_scan_channel) ++ wcn36xx_smd_end_scan(wcn, wcn->sw_scan_channel); ++ if (wcn->sw_scan_init) { ++ wcn36xx_smd_finish_scan(wcn, HAL_SYS_MODE_SCAN, ++ wcn->sw_scan_vif); ++ } + wcn->sw_scan = false; + wcn->sw_scan_opchannel = 0; + } +diff --git a/drivers/net/wireless/ath/wcn36xx/smd.c b/drivers/net/wireless/ath/wcn36xx/smd.c +index 70bffe3d87a12..c056fae1d6418 100644 +--- a/drivers/net/wireless/ath/wcn36xx/smd.c ++++ b/drivers/net/wireless/ath/wcn36xx/smd.c +@@ -721,6 +721,7 @@ int wcn36xx_smd_init_scan(struct wcn36xx *wcn, enum wcn36xx_hal_sys_mode mode, + wcn36xx_err("hal_init_scan response failed err=%d\n", ret); + goto out; + } ++ wcn->sw_scan_init = true; + out: + mutex_unlock(&wcn->hal_mutex); + return ret; +@@ -751,6 +752,7 @@ int wcn36xx_smd_start_scan(struct wcn36xx *wcn, u8 scan_channel) + wcn36xx_err("hal_start_scan response failed err=%d\n", ret); + goto out; + } ++ wcn->sw_scan_channel = scan_channel; + out: + mutex_unlock(&wcn->hal_mutex); + return ret; +@@ -781,6 +783,7 @@ int wcn36xx_smd_end_scan(struct wcn36xx *wcn, u8 scan_channel) + wcn36xx_err("hal_end_scan response failed err=%d\n", ret); + goto out; + } ++ wcn->sw_scan_channel = 0; + out: + mutex_unlock(&wcn->hal_mutex); + return ret; +@@ -822,6 +825,7 @@ int wcn36xx_smd_finish_scan(struct wcn36xx *wcn, + wcn36xx_err("hal_finish_scan response failed err=%d\n", ret); + goto out; + } ++ wcn->sw_scan_init = false; + out: + mutex_unlock(&wcn->hal_mutex); + return ret; +@@ -939,7 +943,7 @@ int wcn36xx_smd_update_channel_list(struct wcn36xx *wcn, struct cfg80211_scan_re + + INIT_HAL_MSG((*msg_body), WCN36XX_HAL_UPDATE_CHANNEL_LIST_REQ); + +- msg_body->num_channel = min_t(u8, req->n_channels, sizeof(msg_body->channels)); ++ msg_body->num_channel = min_t(u8, req->n_channels, ARRAY_SIZE(msg_body->channels)); + for (i = 0; i < msg_body->num_channel; i++) { + struct wcn36xx_hal_channel_param *param = &msg_body->channels[i]; + u32 min_power = WCN36XX_HAL_DEFAULT_MIN_POWER; +@@ -2675,7 +2679,7 @@ static int wcn36xx_smd_missed_beacon_ind(struct wcn36xx *wcn, + wcn36xx_dbg(WCN36XX_DBG_HAL, "beacon missed bss_index %d\n", + tmp->bss_index); + vif = wcn36xx_priv_to_vif(tmp); +- ieee80211_connection_loss(vif); ++ ieee80211_beacon_loss(vif); + } + return 0; + } +@@ -2690,7 +2694,7 @@ static int wcn36xx_smd_missed_beacon_ind(struct wcn36xx *wcn, + wcn36xx_dbg(WCN36XX_DBG_HAL, "beacon missed bss_index %d\n", + rsp->bss_index); + vif = wcn36xx_priv_to_vif(tmp); +- ieee80211_connection_loss(vif); ++ ieee80211_beacon_loss(vif); + return 0; + } + } +diff --git a/drivers/net/wireless/ath/wcn36xx/txrx.c b/drivers/net/wireless/ath/wcn36xx/txrx.c +index bbd7194c82e27..f33e7228a1010 100644 +--- a/drivers/net/wireless/ath/wcn36xx/txrx.c ++++ b/drivers/net/wireless/ath/wcn36xx/txrx.c +@@ -237,7 +237,6 @@ int wcn36xx_rx_skb(struct wcn36xx *wcn, struct sk_buff *skb) + const struct wcn36xx_rate *rate; + struct ieee80211_hdr *hdr; + struct wcn36xx_rx_bd *bd; +- struct ieee80211_supported_band *sband; + u16 fc, sn; + + /* +@@ -259,8 +258,6 @@ int wcn36xx_rx_skb(struct wcn36xx *wcn, struct sk_buff *skb) + fc = __le16_to_cpu(hdr->frame_control); + sn = IEEE80211_SEQ_TO_SN(__le16_to_cpu(hdr->seq_ctrl)); + +- status.freq = WCN36XX_CENTER_FREQ(wcn); +- status.band = WCN36XX_BAND(wcn); + status.mactime = 10; + status.signal = -get_rssi0(bd); + status.antenna = 1; +@@ -272,18 +269,36 @@ int wcn36xx_rx_skb(struct wcn36xx *wcn, struct sk_buff *skb) + + wcn36xx_dbg(WCN36XX_DBG_RX, "status.flags=%x\n", status.flag); + ++ if (bd->scan_learn) { ++ /* If packet originate from hardware scanning, extract the ++ * band/channel from bd descriptor. ++ */ ++ u8 hwch = (bd->reserved0 << 4) + bd->rx_ch; ++ ++ if (bd->rf_band != 1 && hwch <= sizeof(ab_rx_ch_map) && hwch >= 1) { ++ status.band = NL80211_BAND_5GHZ; ++ status.freq = ieee80211_channel_to_frequency(ab_rx_ch_map[hwch - 1], ++ status.band); ++ } else { ++ status.band = NL80211_BAND_2GHZ; ++ status.freq = ieee80211_channel_to_frequency(hwch, status.band); ++ } ++ } else { ++ status.band = WCN36XX_BAND(wcn); ++ status.freq = WCN36XX_CENTER_FREQ(wcn); ++ } ++ + if (bd->rate_id < ARRAY_SIZE(wcn36xx_rate_table)) { + rate = &wcn36xx_rate_table[bd->rate_id]; + status.encoding = rate->encoding; + status.enc_flags = rate->encoding_flags; + status.bw = rate->bw; + status.rate_idx = rate->mcs_or_legacy_index; +- sband = wcn->hw->wiphy->bands[status.band]; + status.nss = 1; + + if (status.band == NL80211_BAND_5GHZ && + status.encoding == RX_ENC_LEGACY && +- status.rate_idx >= sband->n_bitrates) { ++ status.rate_idx >= 4) { + /* no dsss rates in 5Ghz rates table */ + status.rate_idx -= 4; + } +@@ -298,22 +313,6 @@ int wcn36xx_rx_skb(struct wcn36xx *wcn, struct sk_buff *skb) + ieee80211_is_probe_resp(hdr->frame_control)) + status.boottime_ns = ktime_get_boottime_ns(); + +- if (bd->scan_learn) { +- /* If packet originates from hardware scanning, extract the +- * band/channel from bd descriptor. +- */ +- u8 hwch = (bd->reserved0 << 4) + bd->rx_ch; +- +- if (bd->rf_band != 1 && hwch <= sizeof(ab_rx_ch_map) && hwch >= 1) { +- status.band = NL80211_BAND_5GHZ; +- status.freq = ieee80211_channel_to_frequency(ab_rx_ch_map[hwch - 1], +- status.band); +- } else { +- status.band = NL80211_BAND_2GHZ; +- status.freq = ieee80211_channel_to_frequency(hwch, status.band); +- } +- } +- + memcpy(IEEE80211_SKB_RXCB(skb), &status, sizeof(status)); + + if (ieee80211_is_beacon(hdr->frame_control)) { +diff --git a/drivers/net/wireless/ath/wcn36xx/wcn36xx.h b/drivers/net/wireless/ath/wcn36xx/wcn36xx.h +index e9560f35e9bcf..428546a6047f0 100644 +--- a/drivers/net/wireless/ath/wcn36xx/wcn36xx.h ++++ b/drivers/net/wireless/ath/wcn36xx/wcn36xx.h +@@ -246,6 +246,7 @@ struct wcn36xx { + struct cfg80211_scan_request *scan_req; + bool sw_scan; + u8 sw_scan_opchannel; ++ bool sw_scan_init; + u8 sw_scan_channel; + struct ieee80211_vif *sw_scan_vif; + struct mutex scan_lock; +diff --git a/drivers/net/wireless/intel/iwlwifi/iwl-csr.h b/drivers/net/wireless/intel/iwlwifi/iwl-csr.h +index cf796403c45c0..845a09d0dabaf 100644 +--- a/drivers/net/wireless/intel/iwlwifi/iwl-csr.h ++++ b/drivers/net/wireless/intel/iwlwifi/iwl-csr.h +@@ -104,9 +104,10 @@ + /* GIO Chicken Bits (PCI Express bus link power management) */ + #define CSR_GIO_CHICKEN_BITS (CSR_BASE+0x100) + +-/* Doorbell NMI (since Bz) */ ++/* Doorbell - since Bz ++ * connected to UREG_DOORBELL_TO_ISR6 (lower 16 bits only) ++ */ + #define CSR_DOORBELL_VECTOR (CSR_BASE + 0x130) +-#define CSR_DOORBELL_VECTOR_NMI BIT(1) + + /* host chicken bits */ + #define CSR_HOST_CHICKEN (CSR_BASE + 0x204) +diff --git a/drivers/net/wireless/intel/iwlwifi/iwl-drv.c b/drivers/net/wireless/intel/iwlwifi/iwl-drv.c +index 94553f272d377..b7f7b9c5b670c 100644 +--- a/drivers/net/wireless/intel/iwlwifi/iwl-drv.c ++++ b/drivers/net/wireless/intel/iwlwifi/iwl-drv.c +@@ -131,6 +131,9 @@ static void iwl_dealloc_ucode(struct iwl_drv *drv) + + for (i = 0; i < IWL_UCODE_TYPE_MAX; i++) + iwl_free_fw_img(drv, drv->fw.img + i); ++ ++ /* clear the data for the aborted load case */ ++ memset(&drv->fw, 0, sizeof(drv->fw)); + } + + static int iwl_alloc_fw_desc(struct iwl_drv *drv, struct fw_desc *desc, +@@ -1333,6 +1336,7 @@ static void iwl_req_fw_callback(const struct firmware *ucode_raw, void *context) + int i; + bool load_module = false; + bool usniffer_images = false; ++ bool failure = true; + + fw->ucode_capa.max_probe_length = IWL_DEFAULT_MAX_PROBE_LENGTH; + fw->ucode_capa.standard_phy_calibration_size = +@@ -1593,15 +1597,9 @@ static void iwl_req_fw_callback(const struct firmware *ucode_raw, void *context) + * else from proceeding if the module fails to load + * or hangs loading. + */ +- if (load_module) { ++ if (load_module) + request_module("%s", op->name); +-#ifdef CONFIG_IWLWIFI_OPMODE_MODULAR +- if (err) +- IWL_ERR(drv, +- "failed to load module %s (error %d), is dynamic loading enabled?\n", +- op->name, err); +-#endif +- } ++ failure = false; + goto free; + + try_again: +@@ -1617,6 +1615,9 @@ static void iwl_req_fw_callback(const struct firmware *ucode_raw, void *context) + complete(&drv->request_firmware_complete); + device_release_driver(drv->trans->dev); + free: ++ if (failure) ++ iwl_dealloc_ucode(drv); ++ + if (pieces) { + for (i = 0; i < ARRAY_SIZE(pieces->img); i++) + kfree(pieces->img[i].sec); +diff --git a/drivers/net/wireless/intel/iwlwifi/iwl-io.c b/drivers/net/wireless/intel/iwlwifi/iwl-io.c +index 2517c4ae07ab3..5e76ab6c8ad0a 100644 +--- a/drivers/net/wireless/intel/iwlwifi/iwl-io.c ++++ b/drivers/net/wireless/intel/iwlwifi/iwl-io.c +@@ -218,7 +218,7 @@ void iwl_force_nmi(struct iwl_trans *trans) + UREG_DOORBELL_TO_ISR6_NMI_BIT); + else + iwl_write32(trans, CSR_DOORBELL_VECTOR, +- CSR_DOORBELL_VECTOR_NMI); ++ UREG_DOORBELL_TO_ISR6_NMI_BIT); + } + IWL_EXPORT_SYMBOL(iwl_force_nmi); + +diff --git a/drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c b/drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c +index 03e5bf5cb9094..bb5fff8174435 100644 +--- a/drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c ++++ b/drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c +@@ -499,7 +499,7 @@ iwl_mvm_ftm_put_target(struct iwl_mvm *mvm, struct ieee80211_vif *vif, + rcu_read_lock(); + + sta = rcu_dereference(mvm->fw_id_to_mac_id[mvmvif->ap_sta_id]); +- if (sta->mfp) ++ if (sta->mfp && (peer->ftm.trigger_based || peer->ftm.non_trigger_based)) + FTM_PUT_FLAG(PMF); + + rcu_read_unlock(); +@@ -1054,7 +1054,7 @@ static void iwl_mvm_ftm_rtt_smoothing(struct iwl_mvm *mvm, + overshoot = IWL_MVM_FTM_INITIATOR_SMOOTH_OVERSHOOT; + alpha = IWL_MVM_FTM_INITIATOR_SMOOTH_ALPHA; + +- rtt_avg = (alpha * rtt + (100 - alpha) * resp->rtt_avg) / 100; ++ rtt_avg = div_s64(alpha * rtt + (100 - alpha) * resp->rtt_avg, 100); + + IWL_DEBUG_INFO(mvm, + "%pM: prev rtt_avg=%lld, new rtt_avg=%lld, rtt=%lld\n", +diff --git a/drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c b/drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c +index 7e5ad943b20cb..750217393f480 100644 +--- a/drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c ++++ b/drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c +@@ -1687,6 +1687,7 @@ static void iwl_mvm_recalc_multicast(struct iwl_mvm *mvm) + struct iwl_mvm_mc_iter_data iter_data = { + .mvm = mvm, + }; ++ int ret; + + lockdep_assert_held(&mvm->mutex); + +@@ -1696,6 +1697,22 @@ static void iwl_mvm_recalc_multicast(struct iwl_mvm *mvm) + ieee80211_iterate_active_interfaces_atomic( + mvm->hw, IEEE80211_IFACE_ITER_NORMAL, + iwl_mvm_mc_iface_iterator, &iter_data); ++ ++ /* ++ * Send a (synchronous) ech command so that we wait for the ++ * multiple asynchronous MCAST_FILTER_CMD commands sent by ++ * the interface iterator. Otherwise, we might get here over ++ * and over again (by userspace just sending a lot of these) ++ * and the CPU can send them faster than the firmware can ++ * process them. ++ * Note that the CPU is still faster - but with this we'll ++ * actually send fewer commands overall because the CPU will ++ * not schedule the work in mac80211 as frequently if it's ++ * still running when rescheduled (possibly multiple times). ++ */ ++ ret = iwl_mvm_send_cmd_pdu(mvm, ECHO_CMD, 0, 0, NULL); ++ if (ret) ++ IWL_ERR(mvm, "Failed to synchronize multicast groups update\n"); + } + + static u64 iwl_mvm_prepare_multicast(struct ieee80211_hw *hw, +diff --git a/drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c b/drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c +index c12f303cf652c..efccdd3f33773 100644 +--- a/drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c ++++ b/drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c +@@ -121,12 +121,39 @@ static int iwl_mvm_create_skb(struct iwl_mvm *mvm, struct sk_buff *skb, + struct iwl_rx_mpdu_desc *desc = (void *)pkt->data; + unsigned int headlen, fraglen, pad_len = 0; + unsigned int hdrlen = ieee80211_hdrlen(hdr->frame_control); ++ u8 mic_crc_len = u8_get_bits(desc->mac_flags1, ++ IWL_RX_MPDU_MFLG1_MIC_CRC_LEN_MASK) << 1; + + if (desc->mac_flags2 & IWL_RX_MPDU_MFLG2_PAD) { + len -= 2; + pad_len = 2; + } + ++ /* ++ * For non monitor interface strip the bytes the RADA might not have ++ * removed. As monitor interface cannot exist with other interfaces ++ * this removal is safe. ++ */ ++ if (mic_crc_len && !ieee80211_hw_check(mvm->hw, RX_INCLUDES_FCS)) { ++ u32 pkt_flags = le32_to_cpu(pkt->len_n_flags); ++ ++ /* ++ * If RADA was not enabled then decryption was not performed so ++ * the MIC cannot be removed. ++ */ ++ if (!(pkt_flags & FH_RSCSR_RADA_EN)) { ++ if (WARN_ON(crypt_len > mic_crc_len)) ++ return -EINVAL; ++ ++ mic_crc_len -= crypt_len; ++ } ++ ++ if (WARN_ON(mic_crc_len > len)) ++ return -EINVAL; ++ ++ len -= mic_crc_len; ++ } ++ + /* If frame is small enough to fit in skb->head, pull it completely. + * If not, only pull ieee80211_hdr (including crypto if present, and + * an additional 8 bytes for SNAP/ethertype, see below) so that +diff --git a/drivers/net/wireless/intel/iwlwifi/mvm/scan.c b/drivers/net/wireless/intel/iwlwifi/mvm/scan.c +index d78e436fa8b53..5461bf3999593 100644 +--- a/drivers/net/wireless/intel/iwlwifi/mvm/scan.c ++++ b/drivers/net/wireless/intel/iwlwifi/mvm/scan.c +@@ -1924,22 +1924,19 @@ static void iwl_mvm_scan_6ghz_passive_scan(struct iwl_mvm *mvm, + } + + /* +- * 6GHz passive scan is allowed while associated in a defined time +- * interval following HW reset or resume flow ++ * 6GHz passive scan is allowed in a defined time interval following HW ++ * reset or resume flow, or while not associated and a large interval ++ * has passed since the last 6GHz passive scan. + */ +- if (vif->bss_conf.assoc && ++ if ((vif->bss_conf.assoc || ++ time_after(mvm->last_6ghz_passive_scan_jiffies + ++ (IWL_MVM_6GHZ_PASSIVE_SCAN_TIMEOUT * HZ), jiffies)) && + (time_before(mvm->last_reset_or_resume_time_jiffies + + (IWL_MVM_6GHZ_PASSIVE_SCAN_ASSOC_TIMEOUT * HZ), + jiffies))) { +- IWL_DEBUG_SCAN(mvm, "6GHz passive scan: associated\n"); +- return; +- } +- +- /* No need for 6GHz passive scan if not enough time elapsed */ +- if (time_after(mvm->last_6ghz_passive_scan_jiffies + +- (IWL_MVM_6GHZ_PASSIVE_SCAN_TIMEOUT * HZ), jiffies)) { +- IWL_DEBUG_SCAN(mvm, +- "6GHz passive scan: timeout did not expire\n"); ++ IWL_DEBUG_SCAN(mvm, "6GHz passive scan: %s\n", ++ vif->bss_conf.assoc ? "associated" : ++ "timeout did not expire"); + return; + } + +@@ -2490,7 +2487,7 @@ static int iwl_mvm_check_running_scans(struct iwl_mvm *mvm, int type) + return -EIO; + } + +-#define SCAN_TIMEOUT 20000 ++#define SCAN_TIMEOUT 30000 + + void iwl_mvm_scan_timeout_wk(struct work_struct *work) + { +diff --git a/drivers/net/wireless/intel/iwlwifi/mvm/time-event.c b/drivers/net/wireless/intel/iwlwifi/mvm/time-event.c +index e91f8e889df70..ab06dcda1462a 100644 +--- a/drivers/net/wireless/intel/iwlwifi/mvm/time-event.c ++++ b/drivers/net/wireless/intel/iwlwifi/mvm/time-event.c +@@ -49,14 +49,13 @@ void iwl_mvm_roc_done_wk(struct work_struct *wk) + struct iwl_mvm *mvm = container_of(wk, struct iwl_mvm, roc_done_wk); + + /* +- * Clear the ROC_RUNNING /ROC_AUX_RUNNING status bit. ++ * Clear the ROC_RUNNING status bit. + * This will cause the TX path to drop offchannel transmissions. + * That would also be done by mac80211, but it is racy, in particular + * in the case that the time event actually completed in the firmware + * (which is handled in iwl_mvm_te_handle_notif). + */ + clear_bit(IWL_MVM_STATUS_ROC_RUNNING, &mvm->status); +- clear_bit(IWL_MVM_STATUS_ROC_AUX_RUNNING, &mvm->status); + + synchronize_net(); + +@@ -82,9 +81,19 @@ void iwl_mvm_roc_done_wk(struct work_struct *wk) + mvmvif = iwl_mvm_vif_from_mac80211(mvm->p2p_device_vif); + iwl_mvm_flush_sta(mvm, &mvmvif->bcast_sta, true); + } +- } else { ++ } ++ ++ /* ++ * Clear the ROC_AUX_RUNNING status bit. ++ * This will cause the TX path to drop offchannel transmissions. ++ * That would also be done by mac80211, but it is racy, in particular ++ * in the case that the time event actually completed in the firmware ++ * (which is handled in iwl_mvm_te_handle_notif). ++ */ ++ if (test_and_clear_bit(IWL_MVM_STATUS_ROC_AUX_RUNNING, &mvm->status)) { + /* do the same in case of hot spot 2.0 */ + iwl_mvm_flush_sta(mvm, &mvm->aux_sta, true); ++ + /* In newer version of this command an aux station is added only + * in cases of dedicated tx queue and need to be removed in end + * of use */ +@@ -687,11 +696,14 @@ static bool __iwl_mvm_remove_time_event(struct iwl_mvm *mvm, + iwl_mvm_te_clear_data(mvm, te_data); + spin_unlock_bh(&mvm->time_event_lock); + +- /* When session protection is supported, the te_data->id field ++ /* When session protection is used, the te_data->id field + * is reused to save session protection's configuration. ++ * For AUX ROC, HOT_SPOT_CMD is used and the te_data->id field is set ++ * to HOT_SPOT_CMD. + */ + if (fw_has_capa(&mvm->fw->ucode_capa, +- IWL_UCODE_TLV_CAPA_SESSION_PROT_CMD)) { ++ IWL_UCODE_TLV_CAPA_SESSION_PROT_CMD) && ++ id != HOT_SPOT_CMD) { + if (mvmvif && id < SESSION_PROTECT_CONF_MAX_ID) { + /* Session protection is still ongoing. Cancel it */ + iwl_mvm_cancel_session_protection(mvm, mvmvif, id); +@@ -1027,7 +1039,7 @@ void iwl_mvm_stop_roc(struct iwl_mvm *mvm, struct ieee80211_vif *vif) + iwl_mvm_p2p_roc_finished(mvm); + } else { + iwl_mvm_remove_aux_roc_te(mvm, mvmvif, +- &mvmvif->time_event_data); ++ &mvmvif->hs_time_event_data); + iwl_mvm_roc_finished(mvm); + } + +@@ -1158,15 +1170,10 @@ void iwl_mvm_schedule_session_protection(struct iwl_mvm *mvm, + cpu_to_le32(FW_CMD_ID_AND_COLOR(mvmvif->id, + mvmvif->color)), + .action = cpu_to_le32(FW_CTXT_ACTION_ADD), ++ .conf_id = cpu_to_le32(SESSION_PROTECT_CONF_ASSOC), + .duration_tu = cpu_to_le32(MSEC_TO_TU(duration)), + }; + +- /* The time_event_data.id field is reused to save session +- * protection's configuration. +- */ +- mvmvif->time_event_data.id = SESSION_PROTECT_CONF_ASSOC; +- cmd.conf_id = cpu_to_le32(mvmvif->time_event_data.id); +- + lockdep_assert_held(&mvm->mutex); + + spin_lock_bh(&mvm->time_event_lock); +@@ -1180,6 +1187,11 @@ void iwl_mvm_schedule_session_protection(struct iwl_mvm *mvm, + } + + iwl_mvm_te_clear_data(mvm, te_data); ++ /* ++ * The time_event_data.id field is reused to save session ++ * protection's configuration. ++ */ ++ te_data->id = le32_to_cpu(cmd.conf_id); + te_data->duration = le32_to_cpu(cmd.duration_tu); + te_data->vif = vif; + spin_unlock_bh(&mvm->time_event_lock); +diff --git a/drivers/net/wireless/intel/iwlwifi/pcie/rx.c b/drivers/net/wireless/intel/iwlwifi/pcie/rx.c +index 8e45eb38304b2..fea89330f692c 100644 +--- a/drivers/net/wireless/intel/iwlwifi/pcie/rx.c ++++ b/drivers/net/wireless/intel/iwlwifi/pcie/rx.c +@@ -2261,7 +2261,12 @@ irqreturn_t iwl_pcie_irq_msix_handler(int irq, void *dev_id) + } + } + +- if (inta_hw & MSIX_HW_INT_CAUSES_REG_WAKEUP) { ++ /* ++ * In some rare cases when the HW is in a bad state, we may ++ * get this interrupt too early, when prph_info is still NULL. ++ * So make sure that it's not NULL to prevent crashing. ++ */ ++ if (inta_hw & MSIX_HW_INT_CAUSES_REG_WAKEUP && trans_pcie->prph_info) { + u32 sleep_notif = + le32_to_cpu(trans_pcie->prph_info->sleep_notif); + if (sleep_notif == IWL_D3_SLEEP_STATUS_SUSPEND || +diff --git a/drivers/net/wireless/intel/iwlwifi/queue/tx.c b/drivers/net/wireless/intel/iwlwifi/queue/tx.c +index 451b060693501..0f3526b0c5b00 100644 +--- a/drivers/net/wireless/intel/iwlwifi/queue/tx.c ++++ b/drivers/net/wireless/intel/iwlwifi/queue/tx.c +@@ -1072,6 +1072,7 @@ int iwl_txq_alloc(struct iwl_trans *trans, struct iwl_txq *txq, int slots_num, + return 0; + err_free_tfds: + dma_free_coherent(trans->dev, tfd_sz, txq->tfds, txq->dma_addr); ++ txq->tfds = NULL; + error: + if (txq->entries && cmd_queue) + for (i = 0; i < slots_num; i++) +diff --git a/drivers/net/wireless/marvell/mwifiex/sta_event.c b/drivers/net/wireless/marvell/mwifiex/sta_event.c +index 68c63268e2e6b..2b2e6e0166e14 100644 +--- a/drivers/net/wireless/marvell/mwifiex/sta_event.c ++++ b/drivers/net/wireless/marvell/mwifiex/sta_event.c +@@ -365,10 +365,12 @@ static void mwifiex_process_uap_tx_pause(struct mwifiex_private *priv, + sta_ptr = mwifiex_get_sta_entry(priv, tp->peermac); + if (sta_ptr && sta_ptr->tx_pause != tp->tx_pause) { + sta_ptr->tx_pause = tp->tx_pause; ++ spin_unlock_bh(&priv->sta_list_spinlock); + mwifiex_update_ralist_tx_pause(priv, tp->peermac, + tp->tx_pause); ++ } else { ++ spin_unlock_bh(&priv->sta_list_spinlock); + } +- spin_unlock_bh(&priv->sta_list_spinlock); + } + } + +@@ -400,11 +402,13 @@ static void mwifiex_process_sta_tx_pause(struct mwifiex_private *priv, + sta_ptr = mwifiex_get_sta_entry(priv, tp->peermac); + if (sta_ptr && sta_ptr->tx_pause != tp->tx_pause) { + sta_ptr->tx_pause = tp->tx_pause; ++ spin_unlock_bh(&priv->sta_list_spinlock); + mwifiex_update_ralist_tx_pause(priv, + tp->peermac, + tp->tx_pause); ++ } else { ++ spin_unlock_bh(&priv->sta_list_spinlock); + } +- spin_unlock_bh(&priv->sta_list_spinlock); + } + } + } +diff --git a/drivers/net/wireless/marvell/mwifiex/usb.c b/drivers/net/wireless/marvell/mwifiex/usb.c +index 9736aa0ab7fd4..8f01fcbe93961 100644 +--- a/drivers/net/wireless/marvell/mwifiex/usb.c ++++ b/drivers/net/wireless/marvell/mwifiex/usb.c +@@ -130,7 +130,8 @@ static int mwifiex_usb_recv(struct mwifiex_adapter *adapter, + default: + mwifiex_dbg(adapter, ERROR, + "unknown recv_type %#x\n", recv_type); +- return -1; ++ ret = -1; ++ goto exit_restore_skb; + } + break; + case MWIFIEX_USB_EP_DATA: +diff --git a/drivers/net/wireless/mediatek/mt76/mt7603/mac.c b/drivers/net/wireless/mediatek/mt76/mt7603/mac.c +index 3972c56136a20..65f1f2bb80835 100644 +--- a/drivers/net/wireless/mediatek/mt76/mt7603/mac.c ++++ b/drivers/net/wireless/mediatek/mt76/mt7603/mac.c +@@ -525,6 +525,10 @@ mt7603_mac_fill_rx(struct mt7603_dev *dev, struct sk_buff *skb) + if (rxd2 & MT_RXD2_NORMAL_TKIP_MIC_ERR) + status->flag |= RX_FLAG_MMIC_ERROR; + ++ /* ICV error or CCMP/BIP/WPI MIC error */ ++ if (rxd2 & MT_RXD2_NORMAL_ICV_ERR) ++ status->flag |= RX_FLAG_ONLY_MONITOR; ++ + if (FIELD_GET(MT_RXD2_NORMAL_SEC_MODE, rxd2) != 0 && + !(rxd2 & (MT_RXD2_NORMAL_CLM | MT_RXD2_NORMAL_CM))) { + status->flag |= RX_FLAG_DECRYPTED; +diff --git a/drivers/net/wireless/mediatek/mt76/mt7615/mac.c b/drivers/net/wireless/mediatek/mt76/mt7615/mac.c +index 5455231f51881..f2704149834a0 100644 +--- a/drivers/net/wireless/mediatek/mt76/mt7615/mac.c ++++ b/drivers/net/wireless/mediatek/mt76/mt7615/mac.c +@@ -286,9 +286,16 @@ static int mt7615_mac_fill_rx(struct mt7615_dev *dev, struct sk_buff *skb) + if (rxd2 & MT_RXD2_NORMAL_AMSDU_ERR) + return -EINVAL; + ++ hdr_trans = rxd1 & MT_RXD1_NORMAL_HDR_TRANS; ++ if (hdr_trans && (rxd2 & MT_RXD2_NORMAL_CM)) ++ return -EINVAL; ++ ++ /* ICV error or CCMP/BIP/WPI MIC error */ ++ if (rxd2 & MT_RXD2_NORMAL_ICV_ERR) ++ status->flag |= RX_FLAG_ONLY_MONITOR; ++ + unicast = (rxd1 & MT_RXD1_NORMAL_ADDR_TYPE) == MT_RXD1_NORMAL_U2M; + idx = FIELD_GET(MT_RXD2_NORMAL_WLAN_IDX, rxd2); +- hdr_trans = rxd1 & MT_RXD1_NORMAL_HDR_TRANS; + status->wcid = mt7615_rx_get_wcid(dev, idx, unicast); + + if (status->wcid) { +diff --git a/drivers/net/wireless/mediatek/mt76/mt7615/main.c b/drivers/net/wireless/mediatek/mt76/mt7615/main.c +index 51260a669d166..fc266da54fe7b 100644 +--- a/drivers/net/wireless/mediatek/mt76/mt7615/main.c ++++ b/drivers/net/wireless/mediatek/mt76/mt7615/main.c +@@ -211,11 +211,9 @@ static int mt7615_add_interface(struct ieee80211_hw *hw, + mvif->mt76.omac_idx = idx; + + mvif->mt76.band_idx = ext_phy; +- if (mt7615_ext_phy(dev)) +- mvif->mt76.wmm_idx = ext_phy * (MT7615_MAX_WMM_SETS / 2) + +- mvif->mt76.idx % (MT7615_MAX_WMM_SETS / 2); +- else +- mvif->mt76.wmm_idx = mvif->mt76.idx % MT7615_MAX_WMM_SETS; ++ mvif->mt76.wmm_idx = vif->type != NL80211_IFTYPE_AP; ++ if (ext_phy) ++ mvif->mt76.wmm_idx += 2; + + dev->mt76.vif_mask |= BIT(mvif->mt76.idx); + dev->omac_mask |= BIT_ULL(mvif->mt76.omac_idx); +diff --git a/drivers/net/wireless/mediatek/mt76/mt7615/pci_init.c b/drivers/net/wireless/mediatek/mt76/mt7615/pci_init.c +index a2465b49ecd0c..87b4aa52ee0f9 100644 +--- a/drivers/net/wireless/mediatek/mt76/mt7615/pci_init.c ++++ b/drivers/net/wireless/mediatek/mt76/mt7615/pci_init.c +@@ -28,8 +28,6 @@ static void mt7615_pci_init_work(struct work_struct *work) + return; + + mt7615_init_work(dev); +- if (dev->dbdc_support) +- mt7615_register_ext_phy(dev); + } + + static int mt7615_init_hardware(struct mt7615_dev *dev) +@@ -160,6 +158,12 @@ int mt7615_register_device(struct mt7615_dev *dev) + mt7615_init_txpower(dev, &dev->mphy.sband_2g.sband); + mt7615_init_txpower(dev, &dev->mphy.sband_5g.sband); + ++ if (dev->dbdc_support) { ++ ret = mt7615_register_ext_phy(dev); ++ if (ret) ++ return ret; ++ } ++ + return mt7615_init_debugfs(dev); + } + +diff --git a/drivers/net/wireless/mediatek/mt76/mt7915/mac.c b/drivers/net/wireless/mediatek/mt76/mt7915/mac.c +index bbc996f86b5c3..ff613d7056119 100644 +--- a/drivers/net/wireless/mediatek/mt76/mt7915/mac.c ++++ b/drivers/net/wireless/mediatek/mt76/mt7915/mac.c +@@ -349,9 +349,16 @@ mt7915_mac_fill_rx(struct mt7915_dev *dev, struct sk_buff *skb) + if (rxd2 & MT_RXD2_NORMAL_AMSDU_ERR) + return -EINVAL; + ++ hdr_trans = rxd2 & MT_RXD2_NORMAL_HDR_TRANS; ++ if (hdr_trans && (rxd1 & MT_RXD1_NORMAL_CM)) ++ return -EINVAL; ++ ++ /* ICV error or CCMP/BIP/WPI MIC error */ ++ if (rxd1 & MT_RXD1_NORMAL_ICV_ERR) ++ status->flag |= RX_FLAG_ONLY_MONITOR; ++ + unicast = FIELD_GET(MT_RXD3_NORMAL_ADDR_TYPE, rxd3) == MT_RXD3_NORMAL_U2M; + idx = FIELD_GET(MT_RXD1_NORMAL_WLAN_IDX, rxd1); +- hdr_trans = rxd2 & MT_RXD2_NORMAL_HDR_TRANS; + status->wcid = mt7915_rx_get_wcid(dev, idx, unicast); + + if (status->wcid) { +diff --git a/drivers/net/wireless/mediatek/mt76/mt7921/mac.c b/drivers/net/wireless/mediatek/mt76/mt7921/mac.c +index 8a16f3f4d5253..04a288029c98e 100644 +--- a/drivers/net/wireless/mediatek/mt76/mt7921/mac.c ++++ b/drivers/net/wireless/mediatek/mt76/mt7921/mac.c +@@ -383,10 +383,17 @@ int mt7921_mac_fill_rx(struct mt7921_dev *dev, struct sk_buff *skb) + if (rxd2 & MT_RXD2_NORMAL_AMSDU_ERR) + return -EINVAL; + ++ hdr_trans = rxd2 & MT_RXD2_NORMAL_HDR_TRANS; ++ if (hdr_trans && (rxd1 & MT_RXD1_NORMAL_CM)) ++ return -EINVAL; ++ ++ /* ICV error or CCMP/BIP/WPI MIC error */ ++ if (rxd1 & MT_RXD1_NORMAL_ICV_ERR) ++ status->flag |= RX_FLAG_ONLY_MONITOR; ++ + chfreq = FIELD_GET(MT_RXD3_NORMAL_CH_FREQ, rxd3); + unicast = FIELD_GET(MT_RXD3_NORMAL_ADDR_TYPE, rxd3) == MT_RXD3_NORMAL_U2M; + idx = FIELD_GET(MT_RXD1_NORMAL_WLAN_IDX, rxd1); +- hdr_trans = rxd2 & MT_RXD2_NORMAL_HDR_TRANS; + status->wcid = mt7921_rx_get_wcid(dev, idx, unicast); + + if (status->wcid) { +diff --git a/drivers/net/wireless/mediatek/mt76/mt7921/main.c b/drivers/net/wireless/mediatek/mt76/mt7921/main.c +index 63ec140c9c372..9eb90e6f01031 100644 +--- a/drivers/net/wireless/mediatek/mt76/mt7921/main.c ++++ b/drivers/net/wireless/mediatek/mt76/mt7921/main.c +@@ -285,12 +285,6 @@ static int mt7921_add_interface(struct ieee80211_hw *hw, + mtxq->wcid = &mvif->sta.wcid; + } + +- if (vif->type != NL80211_IFTYPE_AP && +- (!mvif->mt76.omac_idx || mvif->mt76.omac_idx > 3)) +- vif->offload_flags = 0; +- +- vif->offload_flags |= IEEE80211_OFFLOAD_ENCAP_4ADDR; +- + out: + mt7921_mutex_release(dev); + +diff --git a/drivers/net/wireless/microchip/wilc1000/netdev.c b/drivers/net/wireless/microchip/wilc1000/netdev.c +index 7e4d9235251cb..9dfb1a285e6a4 100644 +--- a/drivers/net/wireless/microchip/wilc1000/netdev.c ++++ b/drivers/net/wireless/microchip/wilc1000/netdev.c +@@ -901,7 +901,6 @@ void wilc_netdev_cleanup(struct wilc *wilc) + + wilc_wlan_cfg_deinit(wilc); + wlan_deinit_locks(wilc); +- kfree(wilc->bus_data); + wiphy_unregister(wilc->wiphy); + wiphy_free(wilc->wiphy); + } +diff --git a/drivers/net/wireless/microchip/wilc1000/sdio.c b/drivers/net/wireless/microchip/wilc1000/sdio.c +index 42e03a701ae16..8b3b735231085 100644 +--- a/drivers/net/wireless/microchip/wilc1000/sdio.c ++++ b/drivers/net/wireless/microchip/wilc1000/sdio.c +@@ -167,9 +167,11 @@ free: + static void wilc_sdio_remove(struct sdio_func *func) + { + struct wilc *wilc = sdio_get_drvdata(func); ++ struct wilc_sdio *sdio_priv = wilc->bus_data; + + clk_disable_unprepare(wilc->rtc_clk); + wilc_netdev_cleanup(wilc); ++ kfree(sdio_priv); + } + + static int wilc_sdio_reset(struct wilc *wilc) +diff --git a/drivers/net/wireless/microchip/wilc1000/spi.c b/drivers/net/wireless/microchip/wilc1000/spi.c +index dd481dc0b5ce0..c98c0999a6b67 100644 +--- a/drivers/net/wireless/microchip/wilc1000/spi.c ++++ b/drivers/net/wireless/microchip/wilc1000/spi.c +@@ -182,9 +182,11 @@ free: + static int wilc_bus_remove(struct spi_device *spi) + { + struct wilc *wilc = spi_get_drvdata(spi); ++ struct wilc_spi *spi_priv = wilc->bus_data; + + clk_disable_unprepare(wilc->rtc_clk); + wilc_netdev_cleanup(wilc); ++ kfree(spi_priv); + + return 0; + } +diff --git a/drivers/net/wireless/realtek/rtw88/main.c b/drivers/net/wireless/realtek/rtw88/main.c +index 6bb55e663fc36..69512856bb462 100644 +--- a/drivers/net/wireless/realtek/rtw88/main.c ++++ b/drivers/net/wireless/realtek/rtw88/main.c +@@ -1859,7 +1859,7 @@ int rtw_core_init(struct rtw_dev *rtwdev) + + /* default rx filter setting */ + rtwdev->hal.rcr = BIT_APP_FCS | BIT_APP_MIC | BIT_APP_ICV | +- BIT_HTC_LOC_CTRL | BIT_APP_PHYSTS | ++ BIT_PKTCTL_DLEN | BIT_HTC_LOC_CTRL | BIT_APP_PHYSTS | + BIT_AB | BIT_AM | BIT_APM; + + ret = rtw_load_firmware(rtwdev, RTW_NORMAL_FW); +diff --git a/drivers/net/wireless/realtek/rtw88/pci.c b/drivers/net/wireless/realtek/rtw88/pci.c +index a7a6ebfaa203c..08cf66141889b 100644 +--- a/drivers/net/wireless/realtek/rtw88/pci.c ++++ b/drivers/net/wireless/realtek/rtw88/pci.c +@@ -2,7 +2,6 @@ + /* Copyright(c) 2018-2019 Realtek Corporation + */ + +-#include <linux/dmi.h> + #include <linux/module.h> + #include <linux/pci.h> + #include "main.h" +@@ -1409,7 +1408,11 @@ static void rtw_pci_link_ps(struct rtw_dev *rtwdev, bool enter) + * throughput. This is probably because the ASPM behavior slightly + * varies from different SOC. + */ +- if (rtwpci->link_ctrl & PCI_EXP_LNKCTL_ASPM_L1) ++ if (!(rtwpci->link_ctrl & PCI_EXP_LNKCTL_ASPM_L1)) ++ return; ++ ++ if ((enter && atomic_dec_if_positive(&rtwpci->link_usage) == 0) || ++ (!enter && atomic_inc_return(&rtwpci->link_usage) == 1)) + rtw_pci_aspm_set(rtwdev, enter); + } + +@@ -1658,6 +1661,9 @@ static int rtw_pci_napi_poll(struct napi_struct *napi, int budget) + priv); + int work_done = 0; + ++ if (rtwpci->rx_no_aspm) ++ rtw_pci_link_ps(rtwdev, false); ++ + while (work_done < budget) { + u32 work_done_once; + +@@ -1681,6 +1687,8 @@ static int rtw_pci_napi_poll(struct napi_struct *napi, int budget) + if (rtw_pci_get_hw_rx_ring_nr(rtwdev, rtwpci)) + napi_schedule(napi); + } ++ if (rtwpci->rx_no_aspm) ++ rtw_pci_link_ps(rtwdev, true); + + return work_done; + } +@@ -1702,50 +1710,13 @@ static void rtw_pci_napi_deinit(struct rtw_dev *rtwdev) + netif_napi_del(&rtwpci->napi); + } + +-enum rtw88_quirk_dis_pci_caps { +- QUIRK_DIS_PCI_CAP_MSI, +- QUIRK_DIS_PCI_CAP_ASPM, +-}; +- +-static int disable_pci_caps(const struct dmi_system_id *dmi) +-{ +- uintptr_t dis_caps = (uintptr_t)dmi->driver_data; +- +- if (dis_caps & BIT(QUIRK_DIS_PCI_CAP_MSI)) +- rtw_disable_msi = true; +- if (dis_caps & BIT(QUIRK_DIS_PCI_CAP_ASPM)) +- rtw_pci_disable_aspm = true; +- +- return 1; +-} +- +-static const struct dmi_system_id rtw88_pci_quirks[] = { +- { +- .callback = disable_pci_caps, +- .ident = "Protempo Ltd L116HTN6SPW", +- .matches = { +- DMI_MATCH(DMI_SYS_VENDOR, "Protempo Ltd"), +- DMI_MATCH(DMI_PRODUCT_NAME, "L116HTN6SPW"), +- }, +- .driver_data = (void *)BIT(QUIRK_DIS_PCI_CAP_ASPM), +- }, +- { +- .callback = disable_pci_caps, +- .ident = "HP HP Pavilion Laptop 14-ce0xxx", +- .matches = { +- DMI_MATCH(DMI_SYS_VENDOR, "HP"), +- DMI_MATCH(DMI_PRODUCT_NAME, "HP Pavilion Laptop 14-ce0xxx"), +- }, +- .driver_data = (void *)BIT(QUIRK_DIS_PCI_CAP_ASPM), +- }, +- {} +-}; +- + int rtw_pci_probe(struct pci_dev *pdev, + const struct pci_device_id *id) + { ++ struct pci_dev *bridge = pci_upstream_bridge(pdev); + struct ieee80211_hw *hw; + struct rtw_dev *rtwdev; ++ struct rtw_pci *rtwpci; + int drv_data_size; + int ret; + +@@ -1763,6 +1734,9 @@ int rtw_pci_probe(struct pci_dev *pdev, + rtwdev->hci.ops = &rtw_pci_ops; + rtwdev->hci.type = RTW_HCI_TYPE_PCIE; + ++ rtwpci = (struct rtw_pci *)rtwdev->priv; ++ atomic_set(&rtwpci->link_usage, 1); ++ + ret = rtw_core_init(rtwdev); + if (ret) + goto err_release_hw; +@@ -1791,7 +1765,10 @@ int rtw_pci_probe(struct pci_dev *pdev, + goto err_destroy_pci; + } + +- dmi_check_system(rtw88_pci_quirks); ++ /* Disable PCIe ASPM L1 while doing NAPI poll for 8821CE */ ++ if (pdev->device == 0xc821 && bridge->vendor == PCI_VENDOR_ID_INTEL) ++ rtwpci->rx_no_aspm = true; ++ + rtw_pci_phy_cfg(rtwdev); + + ret = rtw_register_hw(rtwdev, hw); +diff --git a/drivers/net/wireless/realtek/rtw88/pci.h b/drivers/net/wireless/realtek/rtw88/pci.h +index 66f78eb7757c5..0c37efd8c66fa 100644 +--- a/drivers/net/wireless/realtek/rtw88/pci.h ++++ b/drivers/net/wireless/realtek/rtw88/pci.h +@@ -223,6 +223,8 @@ struct rtw_pci { + struct rtw_pci_tx_ring tx_rings[RTK_MAX_TX_QUEUE_NUM]; + struct rtw_pci_rx_ring rx_rings[RTK_MAX_RX_QUEUE_NUM]; + u16 link_ctrl; ++ atomic_t link_usage; ++ bool rx_no_aspm; + DECLARE_BITMAP(flags, NUM_OF_RTW_PCI_FLAGS); + + void __iomem *mmap; +diff --git a/drivers/net/wireless/realtek/rtw88/rtw8821c.h b/drivers/net/wireless/realtek/rtw88/rtw8821c.h +index 112faa60f653e..d9fbddd7b0f35 100644 +--- a/drivers/net/wireless/realtek/rtw88/rtw8821c.h ++++ b/drivers/net/wireless/realtek/rtw88/rtw8821c.h +@@ -131,7 +131,7 @@ _rtw_write32s_mask(struct rtw_dev *rtwdev, u32 addr, u32 mask, u32 data) + #define WLAN_TX_FUNC_CFG2 0x30 + #define WLAN_MAC_OPT_NORM_FUNC1 0x98 + #define WLAN_MAC_OPT_LB_FUNC1 0x80 +-#define WLAN_MAC_OPT_FUNC2 0x30810041 ++#define WLAN_MAC_OPT_FUNC2 0xb0810041 + + #define WLAN_SIFS_CFG (WLAN_SIFS_CCK_CONT_TX | \ + (WLAN_SIFS_OFDM_CONT_TX << BIT_SHIFT_SIFS_OFDM_CTX) | \ +diff --git a/drivers/net/wireless/realtek/rtw88/rtw8822b.c b/drivers/net/wireless/realtek/rtw88/rtw8822b.c +index f1789155e9016..247f26e3e8192 100644 +--- a/drivers/net/wireless/realtek/rtw88/rtw8822b.c ++++ b/drivers/net/wireless/realtek/rtw88/rtw8822b.c +@@ -204,7 +204,7 @@ static void rtw8822b_phy_set_param(struct rtw_dev *rtwdev) + #define WLAN_TX_FUNC_CFG2 0x30 + #define WLAN_MAC_OPT_NORM_FUNC1 0x98 + #define WLAN_MAC_OPT_LB_FUNC1 0x80 +-#define WLAN_MAC_OPT_FUNC2 0x30810041 ++#define WLAN_MAC_OPT_FUNC2 0xb0810041 + + #define WLAN_SIFS_CFG (WLAN_SIFS_CCK_CONT_TX | \ + (WLAN_SIFS_OFDM_CONT_TX << BIT_SHIFT_SIFS_OFDM_CTX) | \ +diff --git a/drivers/net/wireless/realtek/rtw88/rtw8822c.c b/drivers/net/wireless/realtek/rtw88/rtw8822c.c +index f3ad079967a68..bc87e3cb9cdce 100644 +--- a/drivers/net/wireless/realtek/rtw88/rtw8822c.c ++++ b/drivers/net/wireless/realtek/rtw88/rtw8822c.c +@@ -1962,7 +1962,7 @@ static void rtw8822c_phy_set_param(struct rtw_dev *rtwdev) + #define WLAN_TX_FUNC_CFG2 0x30 + #define WLAN_MAC_OPT_NORM_FUNC1 0x98 + #define WLAN_MAC_OPT_LB_FUNC1 0x80 +-#define WLAN_MAC_OPT_FUNC2 0x30810041 ++#define WLAN_MAC_OPT_FUNC2 0xb0810041 + #define WLAN_MAC_INT_MIG_CFG 0x33330000 + + #define WLAN_SIFS_CFG (WLAN_SIFS_CCK_CONT_TX | \ +diff --git a/drivers/net/wireless/rsi/rsi_91x_main.c b/drivers/net/wireless/rsi/rsi_91x_main.c +index f1bf71e6c6081..5d1490fc32db4 100644 +--- a/drivers/net/wireless/rsi/rsi_91x_main.c ++++ b/drivers/net/wireless/rsi/rsi_91x_main.c +@@ -23,6 +23,7 @@ + #include "rsi_common.h" + #include "rsi_coex.h" + #include "rsi_hal.h" ++#include "rsi_usb.h" + + u32 rsi_zone_enabled = /* INFO_ZONE | + INIT_ZONE | +@@ -168,6 +169,9 @@ int rsi_read_pkt(struct rsi_common *common, u8 *rx_pkt, s32 rcv_pkt_len) + frame_desc = &rx_pkt[index]; + actual_length = *(u16 *)&frame_desc[0]; + offset = *(u16 *)&frame_desc[2]; ++ if (!rcv_pkt_len && offset > ++ RSI_MAX_RX_USB_PKT_SIZE - FRAME_DESC_SZ) ++ goto fail; + + queueno = rsi_get_queueno(frame_desc, offset); + length = rsi_get_length(frame_desc, offset); +diff --git a/drivers/net/wireless/rsi/rsi_91x_usb.c b/drivers/net/wireless/rsi/rsi_91x_usb.c +index 6821ea9918956..66fe386ec9cc6 100644 +--- a/drivers/net/wireless/rsi/rsi_91x_usb.c ++++ b/drivers/net/wireless/rsi/rsi_91x_usb.c +@@ -269,8 +269,12 @@ static void rsi_rx_done_handler(struct urb *urb) + struct rsi_91x_usbdev *dev = (struct rsi_91x_usbdev *)rx_cb->data; + int status = -EINVAL; + ++ if (!rx_cb->rx_skb) ++ return; ++ + if (urb->status) { + dev_kfree_skb(rx_cb->rx_skb); ++ rx_cb->rx_skb = NULL; + return; + } + +@@ -294,8 +298,10 @@ out: + if (rsi_rx_urb_submit(dev->priv, rx_cb->ep_num, GFP_ATOMIC)) + rsi_dbg(ERR_ZONE, "%s: Failed in urb submission", __func__); + +- if (status) ++ if (status) { + dev_kfree_skb(rx_cb->rx_skb); ++ rx_cb->rx_skb = NULL; ++ } + } + + static void rsi_rx_urb_kill(struct rsi_hw *adapter, u8 ep_num) +@@ -324,7 +330,6 @@ static int rsi_rx_urb_submit(struct rsi_hw *adapter, u8 ep_num, gfp_t mem_flags) + struct sk_buff *skb; + u8 dword_align_bytes = 0; + +-#define RSI_MAX_RX_USB_PKT_SIZE 3000 + skb = dev_alloc_skb(RSI_MAX_RX_USB_PKT_SIZE); + if (!skb) + return -ENOMEM; +diff --git a/drivers/net/wireless/rsi/rsi_usb.h b/drivers/net/wireless/rsi/rsi_usb.h +index 254d19b664123..961851748bc4c 100644 +--- a/drivers/net/wireless/rsi/rsi_usb.h ++++ b/drivers/net/wireless/rsi/rsi_usb.h +@@ -44,6 +44,8 @@ + #define RSI_USB_BUF_SIZE 4096 + #define RSI_USB_CTRL_BUF_SIZE 0x04 + ++#define RSI_MAX_RX_USB_PKT_SIZE 3000 ++ + struct rx_usb_ctrl_block { + u8 *data; + struct urb *rx_urb; +diff --git a/drivers/net/wwan/mhi_wwan_mbim.c b/drivers/net/wwan/mhi_wwan_mbim.c +index 71bf9b4f769f5..6872782e8dd89 100644 +--- a/drivers/net/wwan/mhi_wwan_mbim.c ++++ b/drivers/net/wwan/mhi_wwan_mbim.c +@@ -385,13 +385,13 @@ static void mhi_net_rx_refill_work(struct work_struct *work) + int err; + + while (!mhi_queue_is_full(mdev, DMA_FROM_DEVICE)) { +- struct sk_buff *skb = alloc_skb(MHI_DEFAULT_MRU, GFP_KERNEL); ++ struct sk_buff *skb = alloc_skb(mbim->mru, GFP_KERNEL); + + if (unlikely(!skb)) + break; + + err = mhi_queue_skb(mdev, DMA_FROM_DEVICE, skb, +- MHI_DEFAULT_MRU, MHI_EOT); ++ mbim->mru, MHI_EOT); + if (unlikely(err)) { + kfree_skb(skb); + break; +diff --git a/drivers/nvmem/core.c b/drivers/nvmem/core.c +index 8976da38b375a..9aecb83021a2d 100644 +--- a/drivers/nvmem/core.c ++++ b/drivers/nvmem/core.c +@@ -307,6 +307,8 @@ static umode_t nvmem_bin_attr_is_visible(struct kobject *kobj, + struct device *dev = kobj_to_dev(kobj); + struct nvmem_device *nvmem = to_nvmem_device(dev); + ++ attr->size = nvmem->size; ++ + return nvmem_bin_attr_get_umode(nvmem); + } + +diff --git a/drivers/of/base.c b/drivers/of/base.c +index 0ac17256258d5..54719f8156ed1 100644 +--- a/drivers/of/base.c ++++ b/drivers/of/base.c +@@ -1327,9 +1327,14 @@ int of_phandle_iterator_next(struct of_phandle_iterator *it) + * property data length + */ + if (it->cur + count > it->list_end) { +- pr_err("%pOF: %s = %d found %d\n", +- it->parent, it->cells_name, +- count, it->cell_count); ++ if (it->cells_name) ++ pr_err("%pOF: %s = %d found %td\n", ++ it->parent, it->cells_name, ++ count, it->list_end - it->cur); ++ else ++ pr_err("%pOF: phandle %s needs %d, found %td\n", ++ it->parent, of_node_full_name(it->node), ++ count, it->list_end - it->cur); + goto err; + } + } +diff --git a/drivers/of/fdt.c b/drivers/of/fdt.c +index 4546572af24bb..59a7a9ee58ef7 100644 +--- a/drivers/of/fdt.c ++++ b/drivers/of/fdt.c +@@ -482,9 +482,11 @@ static int __init early_init_dt_reserve_memory_arch(phys_addr_t base, + if (nomap) { + /* + * If the memory is already reserved (by another region), we +- * should not allow it to be marked nomap. ++ * should not allow it to be marked nomap, but don't worry ++ * if the region isn't memory as it won't be mapped. + */ +- if (memblock_is_region_reserved(base, size)) ++ if (memblock_overlaps_region(&memblock.memory, base, size) && ++ memblock_is_region_reserved(base, size)) + return -EBUSY; + + return memblock_mark_nomap(base, size); +@@ -969,18 +971,22 @@ static void __init early_init_dt_check_for_elfcorehdr(unsigned long node) + elfcorehdr_addr, elfcorehdr_size); + } + +-static phys_addr_t cap_mem_addr; +-static phys_addr_t cap_mem_size; ++static unsigned long chosen_node_offset = -FDT_ERR_NOTFOUND; + + /** + * early_init_dt_check_for_usable_mem_range - Decode usable memory range + * location from flat tree +- * @node: reference to node containing usable memory range location ('chosen') + */ +-static void __init early_init_dt_check_for_usable_mem_range(unsigned long node) ++void __init early_init_dt_check_for_usable_mem_range(void) + { + const __be32 *prop; + int len; ++ phys_addr_t cap_mem_addr; ++ phys_addr_t cap_mem_size; ++ unsigned long node = chosen_node_offset; ++ ++ if ((long)node < 0) ++ return; + + pr_debug("Looking for usable-memory-range property... "); + +@@ -993,6 +999,8 @@ static void __init early_init_dt_check_for_usable_mem_range(unsigned long node) + + pr_debug("cap_mem_start=%pa cap_mem_size=%pa\n", &cap_mem_addr, + &cap_mem_size); ++ ++ memblock_cap_memory_range(cap_mem_addr, cap_mem_size); + } + + #ifdef CONFIG_SERIAL_EARLYCON +@@ -1141,9 +1149,10 @@ int __init early_init_dt_scan_chosen(unsigned long node, const char *uname, + (strcmp(uname, "chosen") != 0 && strcmp(uname, "chosen@0") != 0)) + return 0; + ++ chosen_node_offset = node; ++ + early_init_dt_check_for_initrd(node); + early_init_dt_check_for_elfcorehdr(node); +- early_init_dt_check_for_usable_mem_range(node); + + /* Retrieve command line */ + p = of_get_flat_dt_prop(node, "bootargs", &l); +@@ -1279,7 +1288,7 @@ void __init early_init_dt_scan_nodes(void) + of_scan_flat_dt(early_init_dt_scan_memory, NULL); + + /* Handle linux,usable-memory-range property */ +- memblock_cap_memory_range(cap_mem_addr, cap_mem_size); ++ early_init_dt_check_for_usable_mem_range(); + } + + bool __init early_init_dt_scan(void *params) +diff --git a/drivers/of/unittest.c b/drivers/of/unittest.c +index 5b85a2a3792ae..2bee1d992408f 100644 +--- a/drivers/of/unittest.c ++++ b/drivers/of/unittest.c +@@ -911,11 +911,18 @@ static void __init of_unittest_dma_ranges_one(const char *path, + if (!rc) { + phys_addr_t paddr; + dma_addr_t dma_addr; +- struct device dev_bogus; ++ struct device *dev_bogus; + +- dev_bogus.dma_range_map = map; +- paddr = dma_to_phys(&dev_bogus, expect_dma_addr); +- dma_addr = phys_to_dma(&dev_bogus, expect_paddr); ++ dev_bogus = kzalloc(sizeof(struct device), GFP_KERNEL); ++ if (!dev_bogus) { ++ unittest(0, "kzalloc() failed\n"); ++ kfree(map); ++ return; ++ } ++ ++ dev_bogus->dma_range_map = map; ++ paddr = dma_to_phys(dev_bogus, expect_dma_addr); ++ dma_addr = phys_to_dma(dev_bogus, expect_paddr); + + unittest(paddr == expect_paddr, + "of_dma_get_range: wrong phys addr %pap (expecting %llx) on node %pOF\n", +@@ -925,6 +932,7 @@ static void __init of_unittest_dma_ranges_one(const char *path, + &dma_addr, expect_dma_addr, np); + + kfree(map); ++ kfree(dev_bogus); + } + of_node_put(np); + #endif +@@ -934,8 +942,9 @@ static void __init of_unittest_parse_dma_ranges(void) + { + of_unittest_dma_ranges_one("/testcase-data/address-tests/device@70000000", + 0x0, 0x20000000); +- of_unittest_dma_ranges_one("/testcase-data/address-tests/bus@80000000/device@1000", +- 0x100000000, 0x20000000); ++ if (IS_ENABLED(CONFIG_ARCH_DMA_ADDR_T_64BIT)) ++ of_unittest_dma_ranges_one("/testcase-data/address-tests/bus@80000000/device@1000", ++ 0x100000000, 0x20000000); + of_unittest_dma_ranges_one("/testcase-data/address-tests/pci@90000000", + 0x80000000, 0x20000000); + } +diff --git a/drivers/parisc/pdc_stable.c b/drivers/parisc/pdc_stable.c +index e090978518f1a..4760f82def6ec 100644 +--- a/drivers/parisc/pdc_stable.c ++++ b/drivers/parisc/pdc_stable.c +@@ -979,8 +979,10 @@ pdcs_register_pathentries(void) + entry->kobj.kset = paths_kset; + err = kobject_init_and_add(&entry->kobj, &ktype_pdcspath, NULL, + "%s", entry->name); +- if (err) ++ if (err) { ++ kobject_put(&entry->kobj); + return err; ++ } + + /* kobject is now registered */ + write_lock(&entry->rw_lock); +diff --git a/drivers/pci/controller/pci-aardvark.c b/drivers/pci/controller/pci-aardvark.c +index c3b725afa11fd..85323cbc4888a 100644 +--- a/drivers/pci/controller/pci-aardvark.c ++++ b/drivers/pci/controller/pci-aardvark.c +@@ -872,7 +872,6 @@ advk_pci_bridge_emul_pcie_conf_read(struct pci_bridge_emul *bridge, + return PCI_BRIDGE_EMUL_HANDLED; + } + +- case PCI_CAP_LIST_ID: + case PCI_EXP_DEVCAP: + case PCI_EXP_DEVCTL: + *value = advk_readl(pcie, PCIE_CORE_PCIEXP_CAP + reg); +@@ -953,6 +952,9 @@ static int advk_sw_pci_bridge_init(struct advk_pcie *pcie) + /* Support interrupt A for MSI feature */ + bridge->conf.intpin = PCIE_CORE_INT_A_ASSERT_ENABLE; + ++ /* Aardvark HW provides PCIe Capability structure in version 2 */ ++ bridge->pcie_conf.cap = cpu_to_le16(2); ++ + /* Indicates supports for Completion Retry Status */ + bridge->pcie_conf.rootcap = cpu_to_le16(PCI_EXP_RTCAP_CRSVIS); + +diff --git a/drivers/pci/controller/pci-mvebu.c b/drivers/pci/controller/pci-mvebu.c +index ed13e81cd691d..2dc6890dbcaa2 100644 +--- a/drivers/pci/controller/pci-mvebu.c ++++ b/drivers/pci/controller/pci-mvebu.c +@@ -573,6 +573,8 @@ static struct pci_bridge_emul_ops mvebu_pci_bridge_emul_ops = { + static void mvebu_pci_bridge_emul_init(struct mvebu_pcie_port *port) + { + struct pci_bridge_emul *bridge = &port->bridge; ++ u32 pcie_cap = mvebu_readl(port, PCIE_CAP_PCIEXP); ++ u8 pcie_cap_ver = ((pcie_cap >> 16) & PCI_EXP_FLAGS_VERS); + + bridge->conf.vendor = PCI_VENDOR_ID_MARVELL; + bridge->conf.device = mvebu_readl(port, PCIE_DEV_ID_OFF) >> 16; +@@ -585,6 +587,12 @@ static void mvebu_pci_bridge_emul_init(struct mvebu_pcie_port *port) + bridge->conf.iolimit = PCI_IO_RANGE_TYPE_32; + } + ++ /* ++ * Older mvebu hardware provides PCIe Capability structure only in ++ * version 1. New hardware provides it in version 2. ++ */ ++ bridge->pcie_conf.cap = cpu_to_le16(pcie_cap_ver); ++ + bridge->has_pcie = true; + bridge->data = port; + bridge->ops = &mvebu_pci_bridge_emul_ops; +diff --git a/drivers/pci/controller/pci-xgene.c b/drivers/pci/controller/pci-xgene.c +index e64536047b651..7d7d8970fdc25 100644 +--- a/drivers/pci/controller/pci-xgene.c ++++ b/drivers/pci/controller/pci-xgene.c +@@ -466,7 +466,7 @@ static int xgene_pcie_select_ib_reg(u8 *ib_reg_mask, u64 size) + return 1; + } + +- if ((size > SZ_1K) && (size < SZ_1T) && !(*ib_reg_mask & (1 << 0))) { ++ if ((size > SZ_1K) && (size < SZ_4G) && !(*ib_reg_mask & (1 << 0))) { + *ib_reg_mask |= (1 << 0); + return 0; + } +diff --git a/drivers/pci/hotplug/pciehp.h b/drivers/pci/hotplug/pciehp.h +index 69fd401691be6..10d7e7e1b5530 100644 +--- a/drivers/pci/hotplug/pciehp.h ++++ b/drivers/pci/hotplug/pciehp.h +@@ -75,6 +75,8 @@ extern int pciehp_poll_time; + * @reset_lock: prevents access to the Data Link Layer Link Active bit in the + * Link Status register and to the Presence Detect State bit in the Slot + * Status register during a slot reset which may cause them to flap ++ * @depth: Number of additional hotplug ports in the path to the root bus, ++ * used as lock subclass for @reset_lock + * @ist_running: flag to keep user request waiting while IRQ thread is running + * @request_result: result of last user request submitted to the IRQ thread + * @requester: wait queue to wake up on completion of user request, +@@ -106,6 +108,7 @@ struct controller { + + struct hotplug_slot hotplug_slot; /* hotplug core interface */ + struct rw_semaphore reset_lock; ++ unsigned int depth; + unsigned int ist_running; + int request_result; + wait_queue_head_t requester; +diff --git a/drivers/pci/hotplug/pciehp_core.c b/drivers/pci/hotplug/pciehp_core.c +index ad3393930ecb4..e7fe4b42f0394 100644 +--- a/drivers/pci/hotplug/pciehp_core.c ++++ b/drivers/pci/hotplug/pciehp_core.c +@@ -166,7 +166,7 @@ static void pciehp_check_presence(struct controller *ctrl) + { + int occupied; + +- down_read(&ctrl->reset_lock); ++ down_read_nested(&ctrl->reset_lock, ctrl->depth); + mutex_lock(&ctrl->state_lock); + + occupied = pciehp_card_present_or_link_active(ctrl); +diff --git a/drivers/pci/hotplug/pciehp_hpc.c b/drivers/pci/hotplug/pciehp_hpc.c +index 3024d7e85e6a7..3a46227e2c73e 100644 +--- a/drivers/pci/hotplug/pciehp_hpc.c ++++ b/drivers/pci/hotplug/pciehp_hpc.c +@@ -583,7 +583,7 @@ static void pciehp_ignore_dpc_link_change(struct controller *ctrl, + * the corresponding link change may have been ignored above. + * Synthesize it to ensure that it is acted on. + */ +- down_read(&ctrl->reset_lock); ++ down_read_nested(&ctrl->reset_lock, ctrl->depth); + if (!pciehp_check_link_active(ctrl)) + pciehp_request(ctrl, PCI_EXP_SLTSTA_DLLSC); + up_read(&ctrl->reset_lock); +@@ -746,7 +746,7 @@ static irqreturn_t pciehp_ist(int irq, void *dev_id) + * Disable requests have higher priority than Presence Detect Changed + * or Data Link Layer State Changed events. + */ +- down_read(&ctrl->reset_lock); ++ down_read_nested(&ctrl->reset_lock, ctrl->depth); + if (events & DISABLE_SLOT) + pciehp_handle_disable_request(ctrl); + else if (events & (PCI_EXP_SLTSTA_PDC | PCI_EXP_SLTSTA_DLLSC)) +@@ -880,7 +880,7 @@ int pciehp_reset_slot(struct hotplug_slot *hotplug_slot, bool probe) + if (probe) + return 0; + +- down_write(&ctrl->reset_lock); ++ down_write_nested(&ctrl->reset_lock, ctrl->depth); + + if (!ATTN_BUTTN(ctrl)) { + ctrl_mask |= PCI_EXP_SLTCTL_PDCE; +@@ -936,6 +936,20 @@ static inline void dbg_ctrl(struct controller *ctrl) + + #define FLAG(x, y) (((x) & (y)) ? '+' : '-') + ++static inline int pcie_hotplug_depth(struct pci_dev *dev) ++{ ++ struct pci_bus *bus = dev->bus; ++ int depth = 0; ++ ++ while (bus->parent) { ++ bus = bus->parent; ++ if (bus->self && bus->self->is_hotplug_bridge) ++ depth++; ++ } ++ ++ return depth; ++} ++ + struct controller *pcie_init(struct pcie_device *dev) + { + struct controller *ctrl; +@@ -949,6 +963,7 @@ struct controller *pcie_init(struct pcie_device *dev) + return NULL; + + ctrl->pcie = dev; ++ ctrl->depth = pcie_hotplug_depth(dev->port); + pcie_capability_read_dword(pdev, PCI_EXP_SLTCAP, &slot_cap); + + if (pdev->hotplug_user_indicators) +diff --git a/drivers/pci/msi.c b/drivers/pci/msi.c +index e11530cb05699..cc4c2b8a5efd7 100644 +--- a/drivers/pci/msi.c ++++ b/drivers/pci/msi.c +@@ -1193,19 +1193,24 @@ EXPORT_SYMBOL(pci_free_irq_vectors); + + /** + * pci_irq_vector - return Linux IRQ number of a device vector +- * @dev: PCI device to operate on +- * @nr: device-relative interrupt vector index (0-based). ++ * @dev: PCI device to operate on ++ * @nr: Interrupt vector index (0-based) ++ * ++ * @nr has the following meanings depending on the interrupt mode: ++ * MSI-X: The index in the MSI-X vector table ++ * MSI: The index of the enabled MSI vectors ++ * INTx: Must be 0 ++ * ++ * Return: The Linux interrupt number or -EINVAl if @nr is out of range. + */ + int pci_irq_vector(struct pci_dev *dev, unsigned int nr) + { + if (dev->msix_enabled) { + struct msi_desc *entry; +- int i = 0; + + for_each_pci_msi_entry(entry, dev) { +- if (i == nr) ++ if (entry->msi_attrib.entry_nr == nr) + return entry->irq; +- i++; + } + WARN_ON_ONCE(1); + return -EINVAL; +@@ -1229,17 +1234,22 @@ EXPORT_SYMBOL(pci_irq_vector); + * pci_irq_get_affinity - return the affinity of a particular MSI vector + * @dev: PCI device to operate on + * @nr: device-relative interrupt vector index (0-based). ++ * ++ * @nr has the following meanings depending on the interrupt mode: ++ * MSI-X: The index in the MSI-X vector table ++ * MSI: The index of the enabled MSI vectors ++ * INTx: Must be 0 ++ * ++ * Return: A cpumask pointer or NULL if @nr is out of range + */ + const struct cpumask *pci_irq_get_affinity(struct pci_dev *dev, int nr) + { + if (dev->msix_enabled) { + struct msi_desc *entry; +- int i = 0; + + for_each_pci_msi_entry(entry, dev) { +- if (i == nr) ++ if (entry->msi_attrib.entry_nr == nr) + return &entry->affinity->mask; +- i++; + } + WARN_ON_ONCE(1); + return NULL; +diff --git a/drivers/pci/pci-bridge-emul.c b/drivers/pci/pci-bridge-emul.c +index db97cddfc85e1..37504c2cce9b8 100644 +--- a/drivers/pci/pci-bridge-emul.c ++++ b/drivers/pci/pci-bridge-emul.c +@@ -139,8 +139,13 @@ struct pci_bridge_reg_behavior pci_regs_behavior[PCI_STD_HEADER_SIZEOF / 4] = { + .ro = GENMASK(7, 0), + }, + ++ /* ++ * If expansion ROM is unsupported then ROM Base Address register must ++ * be implemented as read-only register that return 0 when read, same ++ * as for unused Base Address registers. ++ */ + [PCI_ROM_ADDRESS1 / 4] = { +- .rw = GENMASK(31, 11) | BIT(0), ++ .ro = ~0, + }, + + /* +@@ -171,41 +176,55 @@ struct pci_bridge_reg_behavior pcie_cap_regs_behavior[PCI_CAP_PCIE_SIZEOF / 4] = + [PCI_CAP_LIST_ID / 4] = { + /* + * Capability ID, Next Capability Pointer and +- * Capabilities register are all read-only. ++ * bits [14:0] of Capabilities register are all read-only. ++ * Bit 15 of Capabilities register is reserved. + */ +- .ro = ~0, ++ .ro = GENMASK(30, 0), + }, + + [PCI_EXP_DEVCAP / 4] = { +- .ro = ~0, ++ /* ++ * Bits [31:29] and [17:16] are reserved. ++ * Bits [27:18] are reserved for non-upstream ports. ++ * Bits 28 and [14:6] are reserved for non-endpoint devices. ++ * Other bits are read-only. ++ */ ++ .ro = BIT(15) | GENMASK(5, 0), + }, + + [PCI_EXP_DEVCTL / 4] = { +- /* Device control register is RW */ +- .rw = GENMASK(15, 0), ++ /* ++ * Device control register is RW, except bit 15 which is ++ * reserved for non-endpoints or non-PCIe-to-PCI/X bridges. ++ */ ++ .rw = GENMASK(14, 0), + + /* + * Device status register has bits 6 and [3:0] W1C, [5:4] RO, +- * the rest is reserved ++ * the rest is reserved. Also bit 6 is reserved for non-upstream ++ * ports. + */ +- .w1c = (BIT(6) | GENMASK(3, 0)) << 16, ++ .w1c = GENMASK(3, 0) << 16, + .ro = GENMASK(5, 4) << 16, + }, + + [PCI_EXP_LNKCAP / 4] = { +- /* All bits are RO, except bit 23 which is reserved */ +- .ro = lower_32_bits(~BIT(23)), ++ /* ++ * All bits are RO, except bit 23 which is reserved and ++ * bit 18 which is reserved for non-upstream ports. ++ */ ++ .ro = lower_32_bits(~(BIT(23) | PCI_EXP_LNKCAP_CLKPM)), + }, + + [PCI_EXP_LNKCTL / 4] = { + /* + * Link control has bits [15:14], [11:3] and [1:0] RW, the +- * rest is reserved. ++ * rest is reserved. Bit 8 is reserved for non-upstream ports. + * + * Link status has bits [13:0] RO, and bits [15:14] + * W1C. + */ +- .rw = GENMASK(15, 14) | GENMASK(11, 3) | GENMASK(1, 0), ++ .rw = GENMASK(15, 14) | GENMASK(11, 9) | GENMASK(7, 3) | GENMASK(1, 0), + .ro = GENMASK(13, 0) << 16, + .w1c = GENMASK(15, 14) << 16, + }, +@@ -277,11 +296,9 @@ int pci_bridge_emul_init(struct pci_bridge_emul *bridge, + + if (bridge->has_pcie) { + bridge->conf.capabilities_pointer = PCI_CAP_PCIE_START; ++ bridge->conf.status |= cpu_to_le16(PCI_STATUS_CAP_LIST); + bridge->pcie_conf.cap_id = PCI_CAP_ID_EXP; +- /* Set PCIe v2, root port, slot support */ +- bridge->pcie_conf.cap = +- cpu_to_le16(PCI_EXP_TYPE_ROOT_PORT << 4 | 2 | +- PCI_EXP_FLAGS_SLOT); ++ bridge->pcie_conf.cap |= cpu_to_le16(PCI_EXP_TYPE_ROOT_PORT << 4); + bridge->pcie_cap_regs_behavior = + kmemdup(pcie_cap_regs_behavior, + sizeof(pcie_cap_regs_behavior), +@@ -290,6 +307,27 @@ int pci_bridge_emul_init(struct pci_bridge_emul *bridge, + kfree(bridge->pci_regs_behavior); + return -ENOMEM; + } ++ /* These bits are applicable only for PCI and reserved on PCIe */ ++ bridge->pci_regs_behavior[PCI_CACHE_LINE_SIZE / 4].ro &= ++ ~GENMASK(15, 8); ++ bridge->pci_regs_behavior[PCI_COMMAND / 4].ro &= ++ ~((PCI_COMMAND_SPECIAL | PCI_COMMAND_INVALIDATE | ++ PCI_COMMAND_VGA_PALETTE | PCI_COMMAND_WAIT | ++ PCI_COMMAND_FAST_BACK) | ++ (PCI_STATUS_66MHZ | PCI_STATUS_FAST_BACK | ++ PCI_STATUS_DEVSEL_MASK) << 16); ++ bridge->pci_regs_behavior[PCI_PRIMARY_BUS / 4].ro &= ++ ~GENMASK(31, 24); ++ bridge->pci_regs_behavior[PCI_IO_BASE / 4].ro &= ++ ~((PCI_STATUS_66MHZ | PCI_STATUS_FAST_BACK | ++ PCI_STATUS_DEVSEL_MASK) << 16); ++ bridge->pci_regs_behavior[PCI_INTERRUPT_LINE / 4].rw &= ++ ~((PCI_BRIDGE_CTL_MASTER_ABORT | ++ BIT(8) | BIT(9) | BIT(11)) << 16); ++ bridge->pci_regs_behavior[PCI_INTERRUPT_LINE / 4].ro &= ++ ~((PCI_BRIDGE_CTL_FAST_BACK) << 16); ++ bridge->pci_regs_behavior[PCI_INTERRUPT_LINE / 4].w1c &= ++ ~(BIT(10) << 16); + } + + if (flags & PCI_BRIDGE_EMUL_NO_PREFETCHABLE_BAR) { +diff --git a/drivers/pci/quirks.c b/drivers/pci/quirks.c +index 208fa03acdda0..0663762ea69db 100644 +--- a/drivers/pci/quirks.c ++++ b/drivers/pci/quirks.c +@@ -4103,6 +4103,9 @@ DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_MARVELL_EXT, 0x9120, + quirk_dma_func1_alias); + DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_MARVELL_EXT, 0x9123, + quirk_dma_func1_alias); ++/* https://bugzilla.kernel.org/show_bug.cgi?id=42679#c136 */ ++DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_MARVELL_EXT, 0x9125, ++ quirk_dma_func1_alias); + DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_MARVELL_EXT, 0x9128, + quirk_dma_func1_alias); + /* https://bugzilla.kernel.org/show_bug.cgi?id=42679#c14 */ +diff --git a/drivers/pcmcia/cs.c b/drivers/pcmcia/cs.c +index e211e2619680c..f70197154a362 100644 +--- a/drivers/pcmcia/cs.c ++++ b/drivers/pcmcia/cs.c +@@ -666,18 +666,16 @@ static int pccardd(void *__skt) + if (events || sysfs_events) + continue; + ++ set_current_state(TASK_INTERRUPTIBLE); + if (kthread_should_stop()) + break; + +- set_current_state(TASK_INTERRUPTIBLE); +- + schedule(); + +- /* make sure we are running */ +- __set_current_state(TASK_RUNNING); +- + try_to_freeze(); + } ++ /* make sure we are running before we exit */ ++ __set_current_state(TASK_RUNNING); + + /* shut down socket, if a device is still present */ + if (skt->state & SOCKET_PRESENT) { +diff --git a/drivers/pcmcia/rsrc_nonstatic.c b/drivers/pcmcia/rsrc_nonstatic.c +index bb15a8bdbaab5..1cac528707111 100644 +--- a/drivers/pcmcia/rsrc_nonstatic.c ++++ b/drivers/pcmcia/rsrc_nonstatic.c +@@ -690,6 +690,9 @@ static struct resource *__nonstatic_find_io_region(struct pcmcia_socket *s, + unsigned long min = base; + int ret; + ++ if (!res) ++ return NULL; ++ + data.mask = align - 1; + data.offset = base & data.mask; + data.map = &s_data->io_db; +@@ -809,6 +812,9 @@ static struct resource *nonstatic_find_mem_region(u_long base, u_long num, + unsigned long min, max; + int ret, i, j; + ++ if (!res) ++ return NULL; ++ + low = low || !(s->features & SS_CAP_PAGE_REGS); + + data.mask = align - 1; +diff --git a/drivers/perf/arm-cmn.c b/drivers/perf/arm-cmn.c +index bc3cba5f8c5dc..400eb7f579dce 100644 +--- a/drivers/perf/arm-cmn.c ++++ b/drivers/perf/arm-cmn.c +@@ -1561,7 +1561,8 @@ static int arm_cmn_probe(struct platform_device *pdev) + + err = perf_pmu_register(&cmn->pmu, name, -1); + if (err) +- cpuhp_state_remove_instance(arm_cmn_hp_state, &cmn->cpuhp_node); ++ cpuhp_state_remove_instance_nocalls(arm_cmn_hp_state, &cmn->cpuhp_node); ++ + return err; + } + +@@ -1572,7 +1573,7 @@ static int arm_cmn_remove(struct platform_device *pdev) + writel_relaxed(0, cmn->dtc[0].base + CMN_DT_DTC_CTL); + + perf_pmu_unregister(&cmn->pmu); +- cpuhp_state_remove_instance(arm_cmn_hp_state, &cmn->cpuhp_node); ++ cpuhp_state_remove_instance_nocalls(arm_cmn_hp_state, &cmn->cpuhp_node); + return 0; + } + +diff --git a/drivers/phy/cadence/phy-cadence-sierra.c b/drivers/phy/cadence/phy-cadence-sierra.c +index e93818e3991fd..3e2d096d54fd7 100644 +--- a/drivers/phy/cadence/phy-cadence-sierra.c ++++ b/drivers/phy/cadence/phy-cadence-sierra.c +@@ -215,7 +215,10 @@ static const int pll_mux_parent_index[][SIERRA_NUM_CMN_PLLC_PARENTS] = { + [CMN_PLLLC1] = { PLL1_REFCLK, PLL0_REFCLK }, + }; + +-static u32 cdns_sierra_pll_mux_table[] = { 0, 1 }; ++static u32 cdns_sierra_pll_mux_table[][SIERRA_NUM_CMN_PLLC_PARENTS] = { ++ [CMN_PLLLC] = { 0, 1 }, ++ [CMN_PLLLC1] = { 1, 0 }, ++}; + + struct cdns_sierra_inst { + struct phy *phy; +@@ -436,11 +439,25 @@ static const struct phy_ops ops = { + static u8 cdns_sierra_pll_mux_get_parent(struct clk_hw *hw) + { + struct cdns_sierra_pll_mux *mux = to_cdns_sierra_pll_mux(hw); ++ struct regmap_field *plllc1en_field = mux->plllc1en_field; ++ struct regmap_field *termen_field = mux->termen_field; + struct regmap_field *field = mux->pfdclk_sel_preg; + unsigned int val; ++ int index; + + regmap_field_read(field, &val); +- return clk_mux_val_to_index(hw, cdns_sierra_pll_mux_table, 0, val); ++ ++ if (strstr(clk_hw_get_name(hw), clk_names[CDNS_SIERRA_PLL_CMNLC1])) { ++ index = clk_mux_val_to_index(hw, cdns_sierra_pll_mux_table[CMN_PLLLC1], 0, val); ++ if (index == 1) { ++ regmap_field_write(plllc1en_field, 1); ++ regmap_field_write(termen_field, 1); ++ } ++ } else { ++ index = clk_mux_val_to_index(hw, cdns_sierra_pll_mux_table[CMN_PLLLC], 0, val); ++ } ++ ++ return index; + } + + static int cdns_sierra_pll_mux_set_parent(struct clk_hw *hw, u8 index) +@@ -458,7 +475,11 @@ static int cdns_sierra_pll_mux_set_parent(struct clk_hw *hw, u8 index) + ret |= regmap_field_write(termen_field, 1); + } + +- val = cdns_sierra_pll_mux_table[index]; ++ if (strstr(clk_hw_get_name(hw), clk_names[CDNS_SIERRA_PLL_CMNLC1])) ++ val = cdns_sierra_pll_mux_table[CMN_PLLLC1][index]; ++ else ++ val = cdns_sierra_pll_mux_table[CMN_PLLLC][index]; ++ + ret |= regmap_field_write(field, val); + + return ret; +@@ -496,8 +517,8 @@ static int cdns_sierra_pll_mux_register(struct cdns_sierra_phy *sp, + for (i = 0; i < num_parents; i++) { + clk = sp->input_clks[pll_mux_parent_index[clk_index][i]]; + if (IS_ERR_OR_NULL(clk)) { +- dev_err(dev, "No parent clock for derived_refclk\n"); +- return PTR_ERR(clk); ++ dev_err(dev, "No parent clock for PLL mux clocks\n"); ++ return IS_ERR(clk) ? PTR_ERR(clk) : -ENOENT; + } + parent_names[i] = __clk_get_name(clk); + } +diff --git a/drivers/phy/mediatek/phy-mtk-mipi-dsi.c b/drivers/phy/mediatek/phy-mtk-mipi-dsi.c +index 28ad9403c4414..67b005d5b9e35 100644 +--- a/drivers/phy/mediatek/phy-mtk-mipi-dsi.c ++++ b/drivers/phy/mediatek/phy-mtk-mipi-dsi.c +@@ -146,6 +146,8 @@ static int mtk_mipi_tx_probe(struct platform_device *pdev) + return -ENOMEM; + + mipi_tx->driver_data = of_device_get_match_data(dev); ++ if (!mipi_tx->driver_data) ++ return -ENODEV; + + mipi_tx->regs = devm_platform_ioremap_resource(pdev, 0); + if (IS_ERR(mipi_tx->regs)) +diff --git a/drivers/phy/mediatek/phy-mtk-tphy.c b/drivers/phy/mediatek/phy-mtk-tphy.c +index cdcef865fe9e5..98a942c607a67 100644 +--- a/drivers/phy/mediatek/phy-mtk-tphy.c ++++ b/drivers/phy/mediatek/phy-mtk-tphy.c +@@ -12,6 +12,7 @@ + #include <linux/iopoll.h> + #include <linux/mfd/syscon.h> + #include <linux/module.h> ++#include <linux/nvmem-consumer.h> + #include <linux/of_address.h> + #include <linux/of_device.h> + #include <linux/phy/phy.h> +@@ -41,6 +42,9 @@ + #define SSUSB_SIFSLV_V2_U3PHYD 0x200 + #define SSUSB_SIFSLV_V2_U3PHYA 0x400 + ++#define U3P_MISC_REG1 0x04 ++#define MR1_EFUSE_AUTO_LOAD_DIS BIT(6) ++ + #define U3P_USBPHYACR0 0x000 + #define PA0_RG_U2PLL_FORCE_ON BIT(15) + #define PA0_USB20_PLL_PREDIV GENMASK(7, 6) +@@ -133,6 +137,8 @@ + #define P3C_RG_SWRST_U3_PHYD_FORCE_EN BIT(24) + + #define U3P_U3_PHYA_REG0 0x000 ++#define P3A_RG_IEXT_INTR GENMASK(15, 10) ++#define P3A_RG_IEXT_INTR_VAL(x) ((0x3f & (x)) << 10) + #define P3A_RG_CLKDRV_OFF GENMASK(3, 2) + #define P3A_RG_CLKDRV_OFF_VAL(x) ((0x3 & (x)) << 2) + +@@ -187,6 +193,19 @@ + #define P3D_RG_FWAKE_TH GENMASK(21, 16) + #define P3D_RG_FWAKE_TH_VAL(x) ((0x3f & (x)) << 16) + ++#define U3P_U3_PHYD_IMPCAL0 0x010 ++#define P3D_RG_FORCE_TX_IMPEL BIT(31) ++#define P3D_RG_TX_IMPEL GENMASK(28, 24) ++#define P3D_RG_TX_IMPEL_VAL(x) ((0x1f & (x)) << 24) ++ ++#define U3P_U3_PHYD_IMPCAL1 0x014 ++#define P3D_RG_FORCE_RX_IMPEL BIT(31) ++#define P3D_RG_RX_IMPEL GENMASK(28, 24) ++#define P3D_RG_RX_IMPEL_VAL(x) ((0x1f & (x)) << 24) ++ ++#define U3P_U3_PHYD_RSV 0x054 ++#define P3D_RG_EFUSE_AUTO_LOAD_DIS BIT(12) ++ + #define U3P_U3_PHYD_CDR1 0x05c + #define P3D_RG_CDR_BIR_LTD1 GENMASK(28, 24) + #define P3D_RG_CDR_BIR_LTD1_VAL(x) ((0x1f & (x)) << 24) +@@ -307,6 +326,11 @@ struct mtk_phy_pdata { + * 48M PLL, fix it by switching PLL to 26M from default 48M + */ + bool sw_pll_48m_to_26m; ++ /* ++ * Some SoCs (e.g. mt8195) drop a bit when use auto load efuse, ++ * support sw way, also support it for v2/v3 optionally. ++ */ ++ bool sw_efuse_supported; + enum mtk_phy_version version; + }; + +@@ -336,6 +360,10 @@ struct mtk_phy_instance { + struct regmap *type_sw; + u32 type_sw_reg; + u32 type_sw_index; ++ u32 efuse_sw_en; ++ u32 efuse_intr; ++ u32 efuse_tx_imp; ++ u32 efuse_rx_imp; + int eye_src; + int eye_vrt; + int eye_term; +@@ -1040,6 +1068,130 @@ static int phy_type_set(struct mtk_phy_instance *instance) + return 0; + } + ++static int phy_efuse_get(struct mtk_tphy *tphy, struct mtk_phy_instance *instance) ++{ ++ struct device *dev = &instance->phy->dev; ++ int ret = 0; ++ ++ /* tphy v1 doesn't support sw efuse, skip it */ ++ if (!tphy->pdata->sw_efuse_supported) { ++ instance->efuse_sw_en = 0; ++ return 0; ++ } ++ ++ /* software efuse is optional */ ++ instance->efuse_sw_en = device_property_read_bool(dev, "nvmem-cells"); ++ if (!instance->efuse_sw_en) ++ return 0; ++ ++ switch (instance->type) { ++ case PHY_TYPE_USB2: ++ ret = nvmem_cell_read_variable_le_u32(dev, "intr", &instance->efuse_intr); ++ if (ret) { ++ dev_err(dev, "fail to get u2 intr efuse, %d\n", ret); ++ break; ++ } ++ ++ /* no efuse, ignore it */ ++ if (!instance->efuse_intr) { ++ dev_warn(dev, "no u2 intr efuse, but dts enable it\n"); ++ instance->efuse_sw_en = 0; ++ break; ++ } ++ ++ dev_dbg(dev, "u2 efuse - intr %x\n", instance->efuse_intr); ++ break; ++ ++ case PHY_TYPE_USB3: ++ case PHY_TYPE_PCIE: ++ ret = nvmem_cell_read_variable_le_u32(dev, "intr", &instance->efuse_intr); ++ if (ret) { ++ dev_err(dev, "fail to get u3 intr efuse, %d\n", ret); ++ break; ++ } ++ ++ ret = nvmem_cell_read_variable_le_u32(dev, "rx_imp", &instance->efuse_rx_imp); ++ if (ret) { ++ dev_err(dev, "fail to get u3 rx_imp efuse, %d\n", ret); ++ break; ++ } ++ ++ ret = nvmem_cell_read_variable_le_u32(dev, "tx_imp", &instance->efuse_tx_imp); ++ if (ret) { ++ dev_err(dev, "fail to get u3 tx_imp efuse, %d\n", ret); ++ break; ++ } ++ ++ /* no efuse, ignore it */ ++ if (!instance->efuse_intr && ++ !instance->efuse_rx_imp && ++ !instance->efuse_rx_imp) { ++ dev_warn(dev, "no u3 intr efuse, but dts enable it\n"); ++ instance->efuse_sw_en = 0; ++ break; ++ } ++ ++ dev_dbg(dev, "u3 efuse - intr %x, rx_imp %x, tx_imp %x\n", ++ instance->efuse_intr, instance->efuse_rx_imp,instance->efuse_tx_imp); ++ break; ++ default: ++ dev_err(dev, "no sw efuse for type %d\n", instance->type); ++ ret = -EINVAL; ++ } ++ ++ return ret; ++} ++ ++static void phy_efuse_set(struct mtk_phy_instance *instance) ++{ ++ struct device *dev = &instance->phy->dev; ++ struct u2phy_banks *u2_banks = &instance->u2_banks; ++ struct u3phy_banks *u3_banks = &instance->u3_banks; ++ u32 tmp; ++ ++ if (!instance->efuse_sw_en) ++ return; ++ ++ switch (instance->type) { ++ case PHY_TYPE_USB2: ++ tmp = readl(u2_banks->misc + U3P_MISC_REG1); ++ tmp |= MR1_EFUSE_AUTO_LOAD_DIS; ++ writel(tmp, u2_banks->misc + U3P_MISC_REG1); ++ ++ tmp = readl(u2_banks->com + U3P_USBPHYACR1); ++ tmp &= ~PA1_RG_INTR_CAL; ++ tmp |= PA1_RG_INTR_CAL_VAL(instance->efuse_intr); ++ writel(tmp, u2_banks->com + U3P_USBPHYACR1); ++ break; ++ case PHY_TYPE_USB3: ++ case PHY_TYPE_PCIE: ++ tmp = readl(u3_banks->phyd + U3P_U3_PHYD_RSV); ++ tmp |= P3D_RG_EFUSE_AUTO_LOAD_DIS; ++ writel(tmp, u3_banks->phyd + U3P_U3_PHYD_RSV); ++ ++ tmp = readl(u3_banks->phyd + U3P_U3_PHYD_IMPCAL0); ++ tmp &= ~P3D_RG_TX_IMPEL; ++ tmp |= P3D_RG_TX_IMPEL_VAL(instance->efuse_tx_imp); ++ tmp |= P3D_RG_FORCE_TX_IMPEL; ++ writel(tmp, u3_banks->phyd + U3P_U3_PHYD_IMPCAL0); ++ ++ tmp = readl(u3_banks->phyd + U3P_U3_PHYD_IMPCAL1); ++ tmp &= ~P3D_RG_RX_IMPEL; ++ tmp |= P3D_RG_RX_IMPEL_VAL(instance->efuse_rx_imp); ++ tmp |= P3D_RG_FORCE_RX_IMPEL; ++ writel(tmp, u3_banks->phyd + U3P_U3_PHYD_IMPCAL1); ++ ++ tmp = readl(u3_banks->phya + U3P_U3_PHYA_REG0); ++ tmp &= ~P3A_RG_IEXT_INTR; ++ tmp |= P3A_RG_IEXT_INTR_VAL(instance->efuse_intr); ++ writel(tmp, u3_banks->phya + U3P_U3_PHYA_REG0); ++ break; ++ default: ++ dev_warn(dev, "no sw efuse for type %d\n", instance->type); ++ break; ++ } ++} ++ + static int mtk_phy_init(struct phy *phy) + { + struct mtk_phy_instance *instance = phy_get_drvdata(phy); +@@ -1050,6 +1202,8 @@ static int mtk_phy_init(struct phy *phy) + if (ret) + return ret; + ++ phy_efuse_set(instance); ++ + switch (instance->type) { + case PHY_TYPE_USB2: + u2_phy_instance_init(tphy, instance); +@@ -1134,6 +1288,7 @@ static struct phy *mtk_phy_xlate(struct device *dev, + struct mtk_phy_instance *instance = NULL; + struct device_node *phy_np = args->np; + int index; ++ int ret; + + if (args->args_count != 1) { + dev_err(dev, "invalid number of cells in 'phy' property\n"); +@@ -1174,6 +1329,10 @@ static struct phy *mtk_phy_xlate(struct device *dev, + return ERR_PTR(-EINVAL); + } + ++ ret = phy_efuse_get(tphy, instance); ++ if (ret) ++ return ERR_PTR(ret); ++ + phy_parse_property(tphy, instance); + phy_type_set(instance); + +@@ -1196,10 +1355,12 @@ static const struct mtk_phy_pdata tphy_v1_pdata = { + + static const struct mtk_phy_pdata tphy_v2_pdata = { + .avoid_rx_sen_degradation = false, ++ .sw_efuse_supported = true, + .version = MTK_PHY_V2, + }; + + static const struct mtk_phy_pdata tphy_v3_pdata = { ++ .sw_efuse_supported = true, + .version = MTK_PHY_V3, + }; + +@@ -1210,6 +1371,7 @@ static const struct mtk_phy_pdata mt8173_pdata = { + + static const struct mtk_phy_pdata mt8195_pdata = { + .sw_pll_48m_to_26m = true, ++ .sw_efuse_supported = true, + .version = MTK_PHY_V3, + }; + +diff --git a/drivers/phy/socionext/phy-uniphier-usb3ss.c b/drivers/phy/socionext/phy-uniphier-usb3ss.c +index 6700645bcbe6b..3b5ffc16a6947 100644 +--- a/drivers/phy/socionext/phy-uniphier-usb3ss.c ++++ b/drivers/phy/socionext/phy-uniphier-usb3ss.c +@@ -22,11 +22,13 @@ + #include <linux/reset.h> + + #define SSPHY_TESTI 0x0 +-#define SSPHY_TESTO 0x4 + #define TESTI_DAT_MASK GENMASK(13, 6) + #define TESTI_ADR_MASK GENMASK(5, 1) + #define TESTI_WR_EN BIT(0) + ++#define SSPHY_TESTO 0x4 ++#define TESTO_DAT_MASK GENMASK(7, 0) ++ + #define PHY_F(regno, msb, lsb) { (regno), (msb), (lsb) } + + #define CDR_CPD_TRIM PHY_F(7, 3, 0) /* RxPLL charge pump current */ +@@ -84,12 +86,12 @@ static void uniphier_u3ssphy_set_param(struct uniphier_u3ssphy_priv *priv, + val = FIELD_PREP(TESTI_DAT_MASK, 1); + val |= FIELD_PREP(TESTI_ADR_MASK, p->field.reg_no); + uniphier_u3ssphy_testio_write(priv, val); +- val = readl(priv->base + SSPHY_TESTO); ++ val = readl(priv->base + SSPHY_TESTO) & TESTO_DAT_MASK; + + /* update value */ +- val &= ~FIELD_PREP(TESTI_DAT_MASK, field_mask); ++ val &= ~field_mask; + data = field_mask & (p->value << p->field.lsb); +- val = FIELD_PREP(TESTI_DAT_MASK, data); ++ val = FIELD_PREP(TESTI_DAT_MASK, data | val); + val |= FIELD_PREP(TESTI_ADR_MASK, p->field.reg_no); + uniphier_u3ssphy_testio_write(priv, val); + uniphier_u3ssphy_testio_write(priv, val | TESTI_WR_EN); +diff --git a/drivers/pinctrl/pinctrl-rockchip.c b/drivers/pinctrl/pinctrl-rockchip.c +index 5ce260f152ce5..dc52da94af0b9 100644 +--- a/drivers/pinctrl/pinctrl-rockchip.c ++++ b/drivers/pinctrl/pinctrl-rockchip.c +@@ -2748,7 +2748,7 @@ static int rockchip_pinctrl_probe(struct platform_device *pdev) + + platform_set_drvdata(pdev, info); + +- ret = of_platform_populate(np, rockchip_bank_match, NULL, NULL); ++ ret = of_platform_populate(np, NULL, NULL, &pdev->dev); + if (ret) { + dev_err(&pdev->dev, "failed to register gpio device\n"); + return ret; +diff --git a/drivers/power/reset/mt6323-poweroff.c b/drivers/power/reset/mt6323-poweroff.c +index 0532803e6cbc4..d90e76fcb9383 100644 +--- a/drivers/power/reset/mt6323-poweroff.c ++++ b/drivers/power/reset/mt6323-poweroff.c +@@ -57,6 +57,9 @@ static int mt6323_pwrc_probe(struct platform_device *pdev) + return -ENOMEM; + + res = platform_get_resource(pdev, IORESOURCE_MEM, 0); ++ if (!res) ++ return -EINVAL; ++ + pwrc->base = res->start; + pwrc->regmap = mt6397_chip->regmap; + pwrc->dev = &pdev->dev; +diff --git a/drivers/ptp/ptp_vclock.c b/drivers/ptp/ptp_vclock.c +index baee0379482bc..ab1d233173e13 100644 +--- a/drivers/ptp/ptp_vclock.c ++++ b/drivers/ptp/ptp_vclock.c +@@ -185,8 +185,8 @@ out: + } + EXPORT_SYMBOL(ptp_get_vclocks_index); + +-void ptp_convert_timestamp(struct skb_shared_hwtstamps *hwtstamps, +- int vclock_index) ++ktime_t ptp_convert_timestamp(const struct skb_shared_hwtstamps *hwtstamps, ++ int vclock_index) + { + char name[PTP_CLOCK_NAME_LEN] = ""; + struct ptp_vclock *vclock; +@@ -198,12 +198,12 @@ void ptp_convert_timestamp(struct skb_shared_hwtstamps *hwtstamps, + snprintf(name, PTP_CLOCK_NAME_LEN, "ptp%d", vclock_index); + dev = class_find_device_by_name(ptp_class, name); + if (!dev) +- return; ++ return 0; + + ptp = dev_get_drvdata(dev); + if (!ptp->is_virtual_clock) { + put_device(dev); +- return; ++ return 0; + } + + vclock = info_to_vclock(ptp->info); +@@ -215,7 +215,7 @@ void ptp_convert_timestamp(struct skb_shared_hwtstamps *hwtstamps, + spin_unlock_irqrestore(&vclock->lock, flags); + + put_device(dev); +- hwtstamps->hwtstamp = ns_to_ktime(ns); ++ return ns_to_ktime(ns); + } + EXPORT_SYMBOL(ptp_convert_timestamp); + #endif +diff --git a/drivers/regulator/da9121-regulator.c b/drivers/regulator/da9121-regulator.c +index e669250902580..0a4fd449c27d1 100644 +--- a/drivers/regulator/da9121-regulator.c ++++ b/drivers/regulator/da9121-regulator.c +@@ -253,6 +253,11 @@ static int da9121_set_current_limit(struct regulator_dev *rdev, + goto error; + } + ++ if (rdev->desc->ops->is_enabled(rdev)) { ++ ret = -EBUSY; ++ goto error; ++ } ++ + ret = da9121_ceiling_selector(rdev, min_ua, max_ua, &sel); + if (ret < 0) + goto error; +diff --git a/drivers/regulator/qcom-labibb-regulator.c b/drivers/regulator/qcom-labibb-regulator.c +index b3da0dc58782f..639b71eb41ffe 100644 +--- a/drivers/regulator/qcom-labibb-regulator.c ++++ b/drivers/regulator/qcom-labibb-regulator.c +@@ -260,7 +260,7 @@ static irqreturn_t qcom_labibb_ocp_isr(int irq, void *chip) + + /* If the regulator is not enabled, this is a fake event */ + if (!ops->is_enabled(vreg->rdev)) +- return 0; ++ return IRQ_HANDLED; + + /* If we tried to recover for too many times it's not getting better */ + if (vreg->ocp_irq_count > LABIBB_MAX_OCP_COUNT) +diff --git a/drivers/regulator/qcom_smd-regulator.c b/drivers/regulator/qcom_smd-regulator.c +index 198fcc6551f6d..8e077792bddd9 100644 +--- a/drivers/regulator/qcom_smd-regulator.c ++++ b/drivers/regulator/qcom_smd-regulator.c +@@ -9,6 +9,7 @@ + #include <linux/of_device.h> + #include <linux/platform_device.h> + #include <linux/regulator/driver.h> ++#include <linux/regulator/of_regulator.h> + #include <linux/soc/qcom/smd-rpm.h> + + struct qcom_rpm_reg { +@@ -1190,52 +1191,91 @@ static const struct of_device_id rpm_of_match[] = { + }; + MODULE_DEVICE_TABLE(of, rpm_of_match); + +-static int rpm_reg_probe(struct platform_device *pdev) ++/** ++ * rpm_regulator_init_vreg() - initialize all attributes of a qcom_smd-regulator ++ * @vreg: Pointer to the individual qcom_smd-regulator resource ++ * @dev: Pointer to the top level qcom_smd-regulator PMIC device ++ * @node: Pointer to the individual qcom_smd-regulator resource ++ * device node ++ * @rpm: Pointer to the rpm bus node ++ * @pmic_rpm_data: Pointer to a null-terminated array of qcom_smd-regulator ++ * resources defined for the top level PMIC device ++ * ++ * Return: 0 on success, errno on failure ++ */ ++static int rpm_regulator_init_vreg(struct qcom_rpm_reg *vreg, struct device *dev, ++ struct device_node *node, struct qcom_smd_rpm *rpm, ++ const struct rpm_regulator_data *pmic_rpm_data) + { +- const struct rpm_regulator_data *reg; +- const struct of_device_id *match; +- struct regulator_config config = { }; ++ struct regulator_config config = {}; ++ const struct rpm_regulator_data *rpm_data; + struct regulator_dev *rdev; ++ int ret; ++ ++ for (rpm_data = pmic_rpm_data; rpm_data->name; rpm_data++) ++ if (of_node_name_eq(node, rpm_data->name)) ++ break; ++ ++ if (!rpm_data->name) { ++ dev_err(dev, "Unknown regulator %pOFn\n", node); ++ return -EINVAL; ++ } ++ ++ vreg->dev = dev; ++ vreg->rpm = rpm; ++ vreg->type = rpm_data->type; ++ vreg->id = rpm_data->id; ++ ++ memcpy(&vreg->desc, rpm_data->desc, sizeof(vreg->desc)); ++ vreg->desc.name = rpm_data->name; ++ vreg->desc.supply_name = rpm_data->supply; ++ vreg->desc.owner = THIS_MODULE; ++ vreg->desc.type = REGULATOR_VOLTAGE; ++ vreg->desc.of_match = rpm_data->name; ++ ++ config.dev = dev; ++ config.of_node = node; ++ config.driver_data = vreg; ++ ++ rdev = devm_regulator_register(dev, &vreg->desc, &config); ++ if (IS_ERR(rdev)) { ++ ret = PTR_ERR(rdev); ++ dev_err(dev, "%pOFn: devm_regulator_register() failed, ret=%d\n", node, ret); ++ return ret; ++ } ++ ++ return 0; ++} ++ ++static int rpm_reg_probe(struct platform_device *pdev) ++{ ++ struct device *dev = &pdev->dev; ++ const struct rpm_regulator_data *vreg_data; ++ struct device_node *node; + struct qcom_rpm_reg *vreg; + struct qcom_smd_rpm *rpm; ++ int ret; + + rpm = dev_get_drvdata(pdev->dev.parent); + if (!rpm) { +- dev_err(&pdev->dev, "unable to retrieve handle to rpm\n"); ++ dev_err(&pdev->dev, "Unable to retrieve handle to rpm\n"); + return -ENODEV; + } + +- match = of_match_device(rpm_of_match, &pdev->dev); +- if (!match) { +- dev_err(&pdev->dev, "failed to match device\n"); ++ vreg_data = of_device_get_match_data(dev); ++ if (!vreg_data) + return -ENODEV; +- } + +- for (reg = match->data; reg->name; reg++) { ++ for_each_available_child_of_node(dev->of_node, node) { + vreg = devm_kzalloc(&pdev->dev, sizeof(*vreg), GFP_KERNEL); + if (!vreg) + return -ENOMEM; + +- vreg->dev = &pdev->dev; +- vreg->type = reg->type; +- vreg->id = reg->id; +- vreg->rpm = rpm; +- +- memcpy(&vreg->desc, reg->desc, sizeof(vreg->desc)); +- +- vreg->desc.id = -1; +- vreg->desc.owner = THIS_MODULE; +- vreg->desc.type = REGULATOR_VOLTAGE; +- vreg->desc.name = reg->name; +- vreg->desc.supply_name = reg->supply; +- vreg->desc.of_match = reg->name; +- +- config.dev = &pdev->dev; +- config.driver_data = vreg; +- rdev = devm_regulator_register(&pdev->dev, &vreg->desc, &config); +- if (IS_ERR(rdev)) { +- dev_err(&pdev->dev, "failed to register %s\n", reg->name); +- return PTR_ERR(rdev); ++ ret = rpm_regulator_init_vreg(vreg, dev, node, rpm, vreg_data); ++ ++ if (ret < 0) { ++ of_node_put(node); ++ return ret; + } + } + +diff --git a/drivers/remoteproc/imx_rproc.c b/drivers/remoteproc/imx_rproc.c +index ff620688fad94..05c39e1c56b49 100644 +--- a/drivers/remoteproc/imx_rproc.c ++++ b/drivers/remoteproc/imx_rproc.c +@@ -830,6 +830,7 @@ static int imx_rproc_remove(struct platform_device *pdev) + clk_disable_unprepare(priv->clk); + rproc_del(rproc); + imx_rproc_free_mbox(rproc); ++ destroy_workqueue(priv->workqueue); + rproc_free(rproc); + + return 0; +diff --git a/drivers/rpmsg/rpmsg_core.c b/drivers/rpmsg/rpmsg_core.c +index 9151836190ce3..a71de08acc7b9 100644 +--- a/drivers/rpmsg/rpmsg_core.c ++++ b/drivers/rpmsg/rpmsg_core.c +@@ -519,13 +519,25 @@ static int rpmsg_dev_probe(struct device *dev) + err = rpdrv->probe(rpdev); + if (err) { + dev_err(dev, "%s: failed: %d\n", __func__, err); +- if (ept) +- rpmsg_destroy_ept(ept); +- goto out; ++ goto destroy_ept; + } + +- if (ept && rpdev->ops->announce_create) ++ if (ept && rpdev->ops->announce_create) { + err = rpdev->ops->announce_create(rpdev); ++ if (err) { ++ dev_err(dev, "failed to announce creation\n"); ++ goto remove_rpdev; ++ } ++ } ++ ++ return 0; ++ ++remove_rpdev: ++ if (rpdrv->remove) ++ rpdrv->remove(rpdev); ++destroy_ept: ++ if (ept) ++ rpmsg_destroy_ept(ept); + out: + return err; + } +diff --git a/drivers/rtc/rtc-cmos.c b/drivers/rtc/rtc-cmos.c +index 4eb53412b8085..dc3f8b0dde989 100644 +--- a/drivers/rtc/rtc-cmos.c ++++ b/drivers/rtc/rtc-cmos.c +@@ -457,7 +457,10 @@ static int cmos_set_alarm(struct device *dev, struct rtc_wkalrm *t) + min = t->time.tm_min; + sec = t->time.tm_sec; + ++ spin_lock_irq(&rtc_lock); + rtc_control = CMOS_READ(RTC_CONTROL); ++ spin_unlock_irq(&rtc_lock); ++ + if (!(rtc_control & RTC_DM_BINARY) || RTC_ALWAYS_BCD) { + /* Writing 0xff means "don't care" or "match all". */ + mon = (mon <= 12) ? bin2bcd(mon) : 0xff; +diff --git a/drivers/rtc/rtc-pxa.c b/drivers/rtc/rtc-pxa.c +index d2f1d8f754bf3..cf8119b6d3204 100644 +--- a/drivers/rtc/rtc-pxa.c ++++ b/drivers/rtc/rtc-pxa.c +@@ -330,6 +330,10 @@ static int __init pxa_rtc_probe(struct platform_device *pdev) + if (sa1100_rtc->irq_alarm < 0) + return -ENXIO; + ++ sa1100_rtc->rtc = devm_rtc_allocate_device(&pdev->dev); ++ if (IS_ERR(sa1100_rtc->rtc)) ++ return PTR_ERR(sa1100_rtc->rtc); ++ + pxa_rtc->base = devm_ioremap(dev, pxa_rtc->ress->start, + resource_size(pxa_rtc->ress)); + if (!pxa_rtc->base) { +diff --git a/drivers/scsi/lpfc/lpfc.h b/drivers/scsi/lpfc/lpfc.h +index befeb7c342903..19fd9d263f47f 100644 +--- a/drivers/scsi/lpfc/lpfc.h ++++ b/drivers/scsi/lpfc/lpfc.h +@@ -1022,7 +1022,6 @@ struct lpfc_hba { + #define HBA_DEVLOSS_TMO 0x2000 /* HBA in devloss timeout */ + #define HBA_RRQ_ACTIVE 0x4000 /* process the rrq active list */ + #define HBA_IOQ_FLUSH 0x8000 /* FCP/NVME I/O queues being flushed */ +-#define HBA_FW_DUMP_OP 0x10000 /* Skips fn reset before FW dump */ + #define HBA_RECOVERABLE_UE 0x20000 /* Firmware supports recoverable UE */ + #define HBA_FORCED_LINK_SPEED 0x40000 /* + * Firmware supports Forced Link Speed +@@ -1038,6 +1037,7 @@ struct lpfc_hba { + #define HBA_HBEAT_TMO 0x8000000 /* HBEAT initiated after timeout */ + #define HBA_FLOGI_OUTSTANDING 0x10000000 /* FLOGI is outstanding */ + ++ struct completion *fw_dump_cmpl; /* cmpl event tracker for fw_dump */ + uint32_t fcp_ring_in_use; /* When polling test if intr-hndlr active*/ + struct lpfc_dmabuf slim2p; + +diff --git a/drivers/scsi/lpfc/lpfc_attr.c b/drivers/scsi/lpfc/lpfc_attr.c +index ebe417921dac0..f20c4fe1fb8b9 100644 +--- a/drivers/scsi/lpfc/lpfc_attr.c ++++ b/drivers/scsi/lpfc/lpfc_attr.c +@@ -1709,25 +1709,25 @@ lpfc_sli4_pdev_reg_request(struct lpfc_hba *phba, uint32_t opcode) + before_fc_flag = phba->pport->fc_flag; + sriov_nr_virtfn = phba->cfg_sriov_nr_virtfn; + +- /* Disable SR-IOV virtual functions if enabled */ +- if (phba->cfg_sriov_nr_virtfn) { +- pci_disable_sriov(pdev); +- phba->cfg_sriov_nr_virtfn = 0; +- } ++ if (opcode == LPFC_FW_DUMP) { ++ init_completion(&online_compl); ++ phba->fw_dump_cmpl = &online_compl; ++ } else { ++ /* Disable SR-IOV virtual functions if enabled */ ++ if (phba->cfg_sriov_nr_virtfn) { ++ pci_disable_sriov(pdev); ++ phba->cfg_sriov_nr_virtfn = 0; ++ } + +- if (opcode == LPFC_FW_DUMP) +- phba->hba_flag |= HBA_FW_DUMP_OP; ++ status = lpfc_do_offline(phba, LPFC_EVT_OFFLINE); + +- status = lpfc_do_offline(phba, LPFC_EVT_OFFLINE); ++ if (status != 0) ++ return status; + +- if (status != 0) { +- phba->hba_flag &= ~HBA_FW_DUMP_OP; +- return status; ++ /* wait for the device to be quiesced before firmware reset */ ++ msleep(100); + } + +- /* wait for the device to be quiesced before firmware reset */ +- msleep(100); +- + reg_val = readl(phba->sli4_hba.conf_regs_memmap_p + + LPFC_CTL_PDEV_CTL_OFFSET); + +@@ -1756,24 +1756,42 @@ lpfc_sli4_pdev_reg_request(struct lpfc_hba *phba, uint32_t opcode) + lpfc_printf_log(phba, KERN_ERR, LOG_SLI, + "3153 Fail to perform the requested " + "access: x%x\n", reg_val); ++ if (phba->fw_dump_cmpl) ++ phba->fw_dump_cmpl = NULL; + return rc; + } + + /* keep the original port state */ +- if (before_fc_flag & FC_OFFLINE_MODE) +- goto out; +- +- init_completion(&online_compl); +- job_posted = lpfc_workq_post_event(phba, &status, &online_compl, +- LPFC_EVT_ONLINE); +- if (!job_posted) ++ if (before_fc_flag & FC_OFFLINE_MODE) { ++ if (phba->fw_dump_cmpl) ++ phba->fw_dump_cmpl = NULL; + goto out; ++ } + +- wait_for_completion(&online_compl); ++ /* Firmware dump will trigger an HA_ERATT event, and ++ * lpfc_handle_eratt_s4 routine already handles bringing the port back ++ * online. ++ */ ++ if (opcode == LPFC_FW_DUMP) { ++ wait_for_completion(phba->fw_dump_cmpl); ++ } else { ++ init_completion(&online_compl); ++ job_posted = lpfc_workq_post_event(phba, &status, &online_compl, ++ LPFC_EVT_ONLINE); ++ if (!job_posted) ++ goto out; + ++ wait_for_completion(&online_compl); ++ } + out: + /* in any case, restore the virtual functions enabled as before */ + if (sriov_nr_virtfn) { ++ /* If fw_dump was performed, first disable to clean up */ ++ if (opcode == LPFC_FW_DUMP) { ++ pci_disable_sriov(pdev); ++ phba->cfg_sriov_nr_virtfn = 0; ++ } ++ + sriov_err = + lpfc_sli_probe_sriov_nr_virtfn(phba, sriov_nr_virtfn); + if (!sriov_err) +diff --git a/drivers/scsi/lpfc/lpfc_els.c b/drivers/scsi/lpfc/lpfc_els.c +index f7197b7161d52..f08ab8269f441 100644 +--- a/drivers/scsi/lpfc/lpfc_els.c ++++ b/drivers/scsi/lpfc/lpfc_els.c +@@ -3531,11 +3531,6 @@ lpfc_issue_els_rscn(struct lpfc_vport *vport, uint8_t retry) + return 1; + } + +- /* This will cause the callback-function lpfc_cmpl_els_cmd to +- * trigger the release of node. +- */ +- if (!(vport->fc_flag & FC_PT2PT)) +- lpfc_nlp_put(ndlp); + return 0; + } + +@@ -6877,6 +6872,7 @@ static int + lpfc_get_rdp_info(struct lpfc_hba *phba, struct lpfc_rdp_context *rdp_context) + { + LPFC_MBOXQ_t *mbox = NULL; ++ struct lpfc_dmabuf *mp; + int rc; + + mbox = mempool_alloc(phba->mbox_mem_pool, GFP_KERNEL); +@@ -6892,8 +6888,11 @@ lpfc_get_rdp_info(struct lpfc_hba *phba, struct lpfc_rdp_context *rdp_context) + mbox->mbox_cmpl = lpfc_mbx_cmpl_rdp_page_a0; + mbox->ctx_ndlp = (struct lpfc_rdp_context *)rdp_context; + rc = lpfc_sli_issue_mbox(phba, mbox, MBX_NOWAIT); +- if (rc == MBX_NOT_FINISHED) ++ if (rc == MBX_NOT_FINISHED) { ++ mp = (struct lpfc_dmabuf *)mbox->ctx_buf; ++ lpfc_mbuf_free(phba, mp->virt, mp->phys); + goto issue_mbox_fail; ++ } + + return 0; + +diff --git a/drivers/scsi/lpfc/lpfc_hbadisc.c b/drivers/scsi/lpfc/lpfc_hbadisc.c +index 9ccb904e35fcf..3bb7c2aa949f7 100644 +--- a/drivers/scsi/lpfc/lpfc_hbadisc.c ++++ b/drivers/scsi/lpfc/lpfc_hbadisc.c +@@ -869,10 +869,16 @@ lpfc_work_done(struct lpfc_hba *phba) + if (phba->pci_dev_grp == LPFC_PCI_DEV_OC) + lpfc_sli4_post_async_mbox(phba); + +- if (ha_copy & HA_ERATT) ++ if (ha_copy & HA_ERATT) { + /* Handle the error attention event */ + lpfc_handle_eratt(phba); + ++ if (phba->fw_dump_cmpl) { ++ complete(phba->fw_dump_cmpl); ++ phba->fw_dump_cmpl = NULL; ++ } ++ } ++ + if (ha_copy & HA_MBATT) + lpfc_sli_handle_mb_event(phba); + +diff --git a/drivers/scsi/lpfc/lpfc_init.c b/drivers/scsi/lpfc/lpfc_init.c +index 0fee8d590b0c4..2bbd1be6cc5d4 100644 +--- a/drivers/scsi/lpfc/lpfc_init.c ++++ b/drivers/scsi/lpfc/lpfc_init.c +@@ -5314,8 +5314,10 @@ lpfc_sli4_async_link_evt(struct lpfc_hba *phba, + */ + if (!(phba->hba_flag & HBA_FCOE_MODE)) { + rc = lpfc_sli_issue_mbox(phba, pmb, MBX_NOWAIT); +- if (rc == MBX_NOT_FINISHED) ++ if (rc == MBX_NOT_FINISHED) { ++ lpfc_mbuf_free(phba, mp->virt, mp->phys); + goto out_free_dmabuf; ++ } + return; + } + /* +@@ -6266,8 +6268,10 @@ lpfc_sli4_async_fc_evt(struct lpfc_hba *phba, struct lpfc_acqe_fc_la *acqe_fc) + } + + rc = lpfc_sli_issue_mbox(phba, pmb, MBX_NOWAIT); +- if (rc == MBX_NOT_FINISHED) ++ if (rc == MBX_NOT_FINISHED) { ++ lpfc_mbuf_free(phba, mp->virt, mp->phys); + goto out_free_dmabuf; ++ } + return; + + out_free_dmabuf: +diff --git a/drivers/scsi/lpfc/lpfc_nportdisc.c b/drivers/scsi/lpfc/lpfc_nportdisc.c +index 27263f02ab9f6..7d717a4ac14d1 100644 +--- a/drivers/scsi/lpfc/lpfc_nportdisc.c ++++ b/drivers/scsi/lpfc/lpfc_nportdisc.c +@@ -322,6 +322,7 @@ lpfc_rcv_plogi(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, + { + struct lpfc_hba *phba = vport->phba; + struct lpfc_dmabuf *pcmd; ++ struct lpfc_dmabuf *mp; + uint64_t nlp_portwwn = 0; + uint32_t *lp; + IOCB_t *icmd; +@@ -571,6 +572,11 @@ lpfc_rcv_plogi(struct lpfc_vport *vport, struct lpfc_nodelist *ndlp, + * a default RPI. + */ + if (phba->sli_rev == LPFC_SLI_REV4) { ++ mp = (struct lpfc_dmabuf *)login_mbox->ctx_buf; ++ if (mp) { ++ lpfc_mbuf_free(phba, mp->virt, mp->phys); ++ kfree(mp); ++ } + mempool_free(login_mbox, phba->mbox_mem_pool); + login_mbox = NULL; + } else { +diff --git a/drivers/scsi/lpfc/lpfc_sli.c b/drivers/scsi/lpfc/lpfc_sli.c +index 9c1f485952ef7..e5009f21d97e1 100644 +--- a/drivers/scsi/lpfc/lpfc_sli.c ++++ b/drivers/scsi/lpfc/lpfc_sli.c +@@ -5043,12 +5043,6 @@ lpfc_sli4_brdreset(struct lpfc_hba *phba) + phba->fcf.fcf_flag = 0; + spin_unlock_irq(&phba->hbalock); + +- /* SLI4 INTF 2: if FW dump is being taken skip INIT_PORT */ +- if (phba->hba_flag & HBA_FW_DUMP_OP) { +- phba->hba_flag &= ~HBA_FW_DUMP_OP; +- return rc; +- } +- + /* Now physically reset the device */ + lpfc_printf_log(phba, KERN_INFO, LOG_INIT, + "0389 Performing PCI function reset!\n"); +diff --git a/drivers/scsi/mpi3mr/mpi3mr.h b/drivers/scsi/mpi3mr/mpi3mr.h +index 9787b53a2b598..2cc42432bd0c0 100644 +--- a/drivers/scsi/mpi3mr/mpi3mr.h ++++ b/drivers/scsi/mpi3mr/mpi3mr.h +@@ -79,7 +79,8 @@ extern int prot_mask; + + /* Operational queue management definitions */ + #define MPI3MR_OP_REQ_Q_QD 512 +-#define MPI3MR_OP_REP_Q_QD 4096 ++#define MPI3MR_OP_REP_Q_QD 1024 ++#define MPI3MR_OP_REP_Q_QD4K 4096 + #define MPI3MR_OP_REQ_Q_SEG_SIZE 4096 + #define MPI3MR_OP_REP_Q_SEG_SIZE 4096 + #define MPI3MR_MAX_SEG_LIST_SIZE 4096 +diff --git a/drivers/scsi/mpi3mr/mpi3mr_fw.c b/drivers/scsi/mpi3mr/mpi3mr_fw.c +index 4a8316c6bd41a..5af36c54cb596 100644 +--- a/drivers/scsi/mpi3mr/mpi3mr_fw.c ++++ b/drivers/scsi/mpi3mr/mpi3mr_fw.c +@@ -1278,7 +1278,7 @@ static void mpi3mr_free_op_req_q_segments(struct mpi3mr_ioc *mrioc, u16 q_idx) + mrioc->op_reply_qinfo[q_idx].q_segment_list = NULL; + } + } else +- size = mrioc->req_qinfo[q_idx].num_requests * ++ size = mrioc->req_qinfo[q_idx].segment_qd * + mrioc->facts.op_req_sz; + + for (j = 0; j < mrioc->req_qinfo[q_idx].num_segments; j++) { +@@ -1565,6 +1565,8 @@ static int mpi3mr_create_op_reply_q(struct mpi3mr_ioc *mrioc, u16 qidx) + + reply_qid = qidx + 1; + op_reply_q->num_replies = MPI3MR_OP_REP_Q_QD; ++ if (!mrioc->pdev->revision) ++ op_reply_q->num_replies = MPI3MR_OP_REP_Q_QD4K; + op_reply_q->ci = 0; + op_reply_q->ephase = 1; + atomic_set(&op_reply_q->pend_ios, 0); +diff --git a/drivers/scsi/pm8001/pm8001_hwi.c b/drivers/scsi/pm8001/pm8001_hwi.c +index 639b7e38a1947..880e1f356defc 100644 +--- a/drivers/scsi/pm8001/pm8001_hwi.c ++++ b/drivers/scsi/pm8001/pm8001_hwi.c +@@ -1325,7 +1325,9 @@ int pm8001_mpi_build_cmd(struct pm8001_hba_info *pm8001_ha, + int q_index = circularQ - pm8001_ha->inbnd_q_tbl; + int rv; + +- WARN_ON(q_index >= PM8001_MAX_INB_NUM); ++ if (WARN_ON(q_index >= pm8001_ha->max_q_num)) ++ return -EINVAL; ++ + spin_lock_irqsave(&circularQ->iq_lock, flags); + rv = pm8001_mpi_msg_free_get(circularQ, pm8001_ha->iomb_size, + &pMessage); +diff --git a/drivers/scsi/scsi.c b/drivers/scsi/scsi.c +index 291ecc33b1fe6..4fc9466d820a7 100644 +--- a/drivers/scsi/scsi.c ++++ b/drivers/scsi/scsi.c +@@ -209,11 +209,11 @@ void scsi_finish_command(struct scsi_cmnd *cmd) + + + /* +- * 1024 is big enough for saturating the fast scsi LUN now ++ * 1024 is big enough for saturating fast SCSI LUNs. + */ + int scsi_device_max_queue_depth(struct scsi_device *sdev) + { +- return max_t(int, sdev->host->can_queue, 1024); ++ return min_t(int, sdev->host->can_queue, 1024); + } + + /** +diff --git a/drivers/scsi/scsi_debugfs.c b/drivers/scsi/scsi_debugfs.c +index d9109771f274d..db8517f1a485a 100644 +--- a/drivers/scsi/scsi_debugfs.c ++++ b/drivers/scsi/scsi_debugfs.c +@@ -9,6 +9,7 @@ + static const char *const scsi_cmd_flags[] = { + SCSI_CMD_FLAG_NAME(TAGGED), + SCSI_CMD_FLAG_NAME(INITIALIZED), ++ SCSI_CMD_FLAG_NAME(LAST), + }; + #undef SCSI_CMD_FLAG_NAME + +diff --git a/drivers/scsi/scsi_pm.c b/drivers/scsi/scsi_pm.c +index 3717eea37ecb3..e91a0a5bc7a3e 100644 +--- a/drivers/scsi/scsi_pm.c ++++ b/drivers/scsi/scsi_pm.c +@@ -262,7 +262,7 @@ static int sdev_runtime_resume(struct device *dev) + blk_pre_runtime_resume(sdev->request_queue); + if (pm && pm->runtime_resume) + err = pm->runtime_resume(dev); +- blk_post_runtime_resume(sdev->request_queue, err); ++ blk_post_runtime_resume(sdev->request_queue); + + return err; + } +diff --git a/drivers/scsi/sr.c b/drivers/scsi/sr.c +index 8b17b35283aa5..1203374828b97 100644 +--- a/drivers/scsi/sr.c ++++ b/drivers/scsi/sr.c +@@ -851,7 +851,7 @@ static void get_capabilities(struct scsi_cd *cd) + + + /* allocate transfer buffer */ +- buffer = kmalloc(512, GFP_KERNEL | GFP_DMA); ++ buffer = kmalloc(512, GFP_KERNEL); + if (!buffer) { + sr_printk(KERN_ERR, cd, "out of memory.\n"); + return; +diff --git a/drivers/scsi/sr_vendor.c b/drivers/scsi/sr_vendor.c +index 1f988a1b9166f..a61635326ae0a 100644 +--- a/drivers/scsi/sr_vendor.c ++++ b/drivers/scsi/sr_vendor.c +@@ -131,7 +131,7 @@ int sr_set_blocklength(Scsi_CD *cd, int blocklength) + if (cd->vendor == VENDOR_TOSHIBA) + density = (blocklength > 2048) ? 0x81 : 0x83; + +- buffer = kmalloc(512, GFP_KERNEL | GFP_DMA); ++ buffer = kmalloc(512, GFP_KERNEL); + if (!buffer) + return -ENOMEM; + +@@ -179,7 +179,7 @@ int sr_cd_check(struct cdrom_device_info *cdi) + if (cd->cdi.mask & CDC_MULTI_SESSION) + return 0; + +- buffer = kmalloc(512, GFP_KERNEL | GFP_DMA); ++ buffer = kmalloc(512, GFP_KERNEL); + if (!buffer) + return -ENOMEM; + +diff --git a/drivers/scsi/ufs/tc-dwc-g210-pci.c b/drivers/scsi/ufs/tc-dwc-g210-pci.c +index 679289e1a78e6..7b08e2e07cc5f 100644 +--- a/drivers/scsi/ufs/tc-dwc-g210-pci.c ++++ b/drivers/scsi/ufs/tc-dwc-g210-pci.c +@@ -110,7 +110,6 @@ tc_dwc_g210_pci_probe(struct pci_dev *pdev, const struct pci_device_id *id) + return err; + } + +- pci_set_drvdata(pdev, hba); + pm_runtime_put_noidle(&pdev->dev); + pm_runtime_allow(&pdev->dev); + +diff --git a/drivers/scsi/ufs/ufs-mediatek.c b/drivers/scsi/ufs/ufs-mediatek.c +index 80b3545dd17d6..4e53857605de8 100644 +--- a/drivers/scsi/ufs/ufs-mediatek.c ++++ b/drivers/scsi/ufs/ufs-mediatek.c +@@ -501,7 +501,7 @@ static void ufs_mtk_init_va09_pwr_ctrl(struct ufs_hba *hba) + struct ufs_mtk_host *host = ufshcd_get_variant(hba); + + host->reg_va09 = regulator_get(hba->dev, "va09"); +- if (!host->reg_va09) ++ if (IS_ERR(host->reg_va09)) + dev_info(hba->dev, "failed to get va09"); + else + host->caps |= UFS_MTK_CAP_VA09_PWR_CTRL; +diff --git a/drivers/scsi/ufs/ufshcd-pci.c b/drivers/scsi/ufs/ufshcd-pci.c +index f725248ba57f4..f76692053ca17 100644 +--- a/drivers/scsi/ufs/ufshcd-pci.c ++++ b/drivers/scsi/ufs/ufshcd-pci.c +@@ -538,8 +538,6 @@ ufshcd_pci_probe(struct pci_dev *pdev, const struct pci_device_id *id) + return err; + } + +- pci_set_drvdata(pdev, hba); +- + hba->vops = (struct ufs_hba_variant_ops *)id->driver_data; + + err = ufshcd_init(hba, mmio_base, pdev->irq); +diff --git a/drivers/scsi/ufs/ufshcd-pltfrm.c b/drivers/scsi/ufs/ufshcd-pltfrm.c +index eaeae83b999fd..8b16bbbcb806c 100644 +--- a/drivers/scsi/ufs/ufshcd-pltfrm.c ++++ b/drivers/scsi/ufs/ufshcd-pltfrm.c +@@ -361,8 +361,6 @@ int ufshcd_pltfrm_init(struct platform_device *pdev, + goto dealloc_host; + } + +- platform_set_drvdata(pdev, hba); +- + pm_runtime_set_active(&pdev->dev); + pm_runtime_enable(&pdev->dev); + +diff --git a/drivers/scsi/ufs/ufshcd.c b/drivers/scsi/ufs/ufshcd.c +index 31adf25e57b0d..ae7bdd8703198 100644 +--- a/drivers/scsi/ufs/ufshcd.c ++++ b/drivers/scsi/ufs/ufshcd.c +@@ -1657,7 +1657,8 @@ int ufshcd_hold(struct ufs_hba *hba, bool async) + bool flush_result; + unsigned long flags; + +- if (!ufshcd_is_clkgating_allowed(hba)) ++ if (!ufshcd_is_clkgating_allowed(hba) || ++ !hba->clk_gating.is_initialized) + goto out; + spin_lock_irqsave(hba->host->host_lock, flags); + hba->clk_gating.active_reqs++; +@@ -1817,7 +1818,7 @@ static void __ufshcd_release(struct ufs_hba *hba) + + if (hba->clk_gating.active_reqs || hba->clk_gating.is_suspended || + hba->ufshcd_state != UFSHCD_STATE_OPERATIONAL || +- hba->outstanding_tasks || ++ hba->outstanding_tasks || !hba->clk_gating.is_initialized || + hba->active_uic_cmd || hba->uic_async_done || + hba->clk_gating.state == CLKS_OFF) + return; +@@ -1952,11 +1953,15 @@ static void ufshcd_exit_clk_gating(struct ufs_hba *hba) + { + if (!hba->clk_gating.is_initialized) + return; ++ + ufshcd_remove_clk_gating_sysfs(hba); +- cancel_work_sync(&hba->clk_gating.ungate_work); +- cancel_delayed_work_sync(&hba->clk_gating.gate_work); +- destroy_workqueue(hba->clk_gating.clk_gating_workq); ++ ++ /* Ungate the clock if necessary. */ ++ ufshcd_hold(hba, false); + hba->clk_gating.is_initialized = false; ++ ufshcd_release(hba); ++ ++ destroy_workqueue(hba->clk_gating.clk_gating_workq); + } + + /* Must be called with host lock acquired */ +@@ -9366,6 +9371,13 @@ int ufshcd_init(struct ufs_hba *hba, void __iomem *mmio_base, unsigned int irq) + struct device *dev = hba->dev; + char eh_wq_name[sizeof("ufs_eh_wq_00")]; + ++ /* ++ * dev_set_drvdata() must be called before any callbacks are registered ++ * that use dev_get_drvdata() (frequency scaling, clock scaling, hwmon, ++ * sysfs). ++ */ ++ dev_set_drvdata(dev, hba); ++ + if (!mmio_base) { + dev_err(hba->dev, + "Invalid memory reference for mmio_base is NULL\n"); +diff --git a/drivers/soc/imx/gpcv2.c b/drivers/soc/imx/gpcv2.c +index 8b7a01773aec2..b4aa28420f2a8 100644 +--- a/drivers/soc/imx/gpcv2.c ++++ b/drivers/soc/imx/gpcv2.c +@@ -369,7 +369,7 @@ static int imx_pgc_power_down(struct generic_pm_domain *genpd) + } + } + +- pm_runtime_put(domain->dev); ++ pm_runtime_put_sync_suspend(domain->dev); + + return 0; + +diff --git a/drivers/soc/mediatek/mtk-scpsys.c b/drivers/soc/mediatek/mtk-scpsys.c +index ca75b14931ec9..670cc82d17dc2 100644 +--- a/drivers/soc/mediatek/mtk-scpsys.c ++++ b/drivers/soc/mediatek/mtk-scpsys.c +@@ -411,12 +411,17 @@ out: + return ret; + } + +-static void init_clks(struct platform_device *pdev, struct clk **clk) ++static int init_clks(struct platform_device *pdev, struct clk **clk) + { + int i; + +- for (i = CLK_NONE + 1; i < CLK_MAX; i++) ++ for (i = CLK_NONE + 1; i < CLK_MAX; i++) { + clk[i] = devm_clk_get(&pdev->dev, clk_names[i]); ++ if (IS_ERR(clk[i])) ++ return PTR_ERR(clk[i]); ++ } ++ ++ return 0; + } + + static struct scp *init_scp(struct platform_device *pdev, +@@ -426,7 +431,7 @@ static struct scp *init_scp(struct platform_device *pdev, + { + struct genpd_onecell_data *pd_data; + struct resource *res; +- int i, j; ++ int i, j, ret; + struct scp *scp; + struct clk *clk[CLK_MAX]; + +@@ -481,7 +486,9 @@ static struct scp *init_scp(struct platform_device *pdev, + + pd_data->num_domains = num; + +- init_clks(pdev, clk); ++ ret = init_clks(pdev, clk); ++ if (ret) ++ return ERR_PTR(ret); + + for (i = 0; i < num; i++) { + struct scp_domain *scpd = &scp->domains[i]; +diff --git a/drivers/soc/qcom/cpr.c b/drivers/soc/qcom/cpr.c +index 4ce8e816154f9..84dd93472a252 100644 +--- a/drivers/soc/qcom/cpr.c ++++ b/drivers/soc/qcom/cpr.c +@@ -1010,7 +1010,7 @@ static int cpr_interpolate(const struct corner *corner, int step_volt, + return corner->uV; + + temp = f_diff * (uV_high - uV_low); +- do_div(temp, f_high - f_low); ++ temp = div64_ul(temp, f_high - f_low); + + /* + * max_volt_scale has units of uV/MHz while freq values +diff --git a/drivers/soc/ti/pruss.c b/drivers/soc/ti/pruss.c +index 49da387d77494..b36779309e49b 100644 +--- a/drivers/soc/ti/pruss.c ++++ b/drivers/soc/ti/pruss.c +@@ -129,7 +129,7 @@ static int pruss_clk_init(struct pruss *pruss, struct device_node *cfg_node) + + clks_np = of_get_child_by_name(cfg_node, "clocks"); + if (!clks_np) { +- dev_err(dev, "%pOF is missing its 'clocks' node\n", clks_np); ++ dev_err(dev, "%pOF is missing its 'clocks' node\n", cfg_node); + return -ENODEV; + } + +diff --git a/drivers/spi/spi-hisi-kunpeng.c b/drivers/spi/spi-hisi-kunpeng.c +index 58b823a16fc4d..525cc0143a305 100644 +--- a/drivers/spi/spi-hisi-kunpeng.c ++++ b/drivers/spi/spi-hisi-kunpeng.c +@@ -127,7 +127,6 @@ struct hisi_spi { + void __iomem *regs; + int irq; + u32 fifo_len; /* depth of the FIFO buffer */ +- u16 bus_num; + + /* Current message transfer state info */ + const void *tx; +@@ -165,7 +164,10 @@ static int hisi_spi_debugfs_init(struct hisi_spi *hs) + { + char name[32]; + +- snprintf(name, 32, "hisi_spi%d", hs->bus_num); ++ struct spi_controller *master; ++ ++ master = container_of(hs->dev, struct spi_controller, dev); ++ snprintf(name, 32, "hisi_spi%d", master->bus_num); + hs->debugfs = debugfs_create_dir(name, NULL); + if (!hs->debugfs) + return -ENOMEM; +@@ -467,7 +469,6 @@ static int hisi_spi_probe(struct platform_device *pdev) + hs = spi_controller_get_devdata(master); + hs->dev = dev; + hs->irq = irq; +- hs->bus_num = pdev->id; + + hs->regs = devm_platform_ioremap_resource(pdev, 0); + if (IS_ERR(hs->regs)) +@@ -490,7 +491,7 @@ static int hisi_spi_probe(struct platform_device *pdev) + master->use_gpio_descriptors = true; + master->mode_bits = SPI_CPOL | SPI_CPHA | SPI_CS_HIGH | SPI_LOOP; + master->bits_per_word_mask = SPI_BPW_RANGE_MASK(4, 32); +- master->bus_num = hs->bus_num; ++ master->bus_num = pdev->id; + master->setup = hisi_spi_setup; + master->cleanup = hisi_spi_cleanup; + master->transfer_one = hisi_spi_transfer_one; +@@ -506,15 +507,15 @@ static int hisi_spi_probe(struct platform_device *pdev) + return ret; + } + +- if (hisi_spi_debugfs_init(hs)) +- dev_info(dev, "failed to create debugfs dir\n"); +- + ret = spi_register_controller(master); + if (ret) { + dev_err(dev, "failed to register spi master, ret=%d\n", ret); + return ret; + } + ++ if (hisi_spi_debugfs_init(hs)) ++ dev_info(dev, "failed to create debugfs dir\n"); ++ + dev_info(dev, "hw version:0x%x max-freq:%u kHz\n", + readl(hs->regs + HISI_SPI_VERSION), + master->max_speed_hz / 1000); +diff --git a/drivers/spi/spi-meson-spifc.c b/drivers/spi/spi-meson-spifc.c +index 8eca6f24cb799..c8ed7815c4ba6 100644 +--- a/drivers/spi/spi-meson-spifc.c ++++ b/drivers/spi/spi-meson-spifc.c +@@ -349,6 +349,7 @@ static int meson_spifc_probe(struct platform_device *pdev) + return 0; + out_clk: + clk_disable_unprepare(spifc->clk); ++ pm_runtime_disable(spifc->dev); + out_err: + spi_master_put(master); + return ret; +diff --git a/drivers/spi/spi-uniphier.c b/drivers/spi/spi-uniphier.c +index 8900e51e1a1cc..342ee8d2c4761 100644 +--- a/drivers/spi/spi-uniphier.c ++++ b/drivers/spi/spi-uniphier.c +@@ -767,12 +767,13 @@ out_master_put: + + static int uniphier_spi_remove(struct platform_device *pdev) + { +- struct uniphier_spi_priv *priv = platform_get_drvdata(pdev); ++ struct spi_master *master = platform_get_drvdata(pdev); ++ struct uniphier_spi_priv *priv = spi_master_get_devdata(master); + +- if (priv->master->dma_tx) +- dma_release_channel(priv->master->dma_tx); +- if (priv->master->dma_rx) +- dma_release_channel(priv->master->dma_rx); ++ if (master->dma_tx) ++ dma_release_channel(master->dma_tx); ++ if (master->dma_rx) ++ dma_release_channel(master->dma_rx); + + clk_disable_unprepare(priv->clk); + +diff --git a/drivers/spi/spi.c b/drivers/spi/spi.c +index 97b5a811bd7f7..a42b9e8521ce0 100644 +--- a/drivers/spi/spi.c ++++ b/drivers/spi/spi.c +@@ -868,12 +868,9 @@ static void spi_set_cs(struct spi_device *spi, bool enable, bool force) + spi->controller->last_cs_enable = enable; + spi->controller->last_cs_mode_high = spi->mode & SPI_CS_HIGH; + +- if (spi->cs_gpiod || gpio_is_valid(spi->cs_gpio) || +- !spi->controller->set_cs_timing) { +- if (activate) +- spi_delay_exec(&spi->cs_setup, NULL); +- else +- spi_delay_exec(&spi->cs_hold, NULL); ++ if ((spi->cs_gpiod || gpio_is_valid(spi->cs_gpio) || ++ !spi->controller->set_cs_timing) && !activate) { ++ spi_delay_exec(&spi->cs_hold, NULL); + } + + if (spi->mode & SPI_CS_HIGH) +@@ -915,7 +912,9 @@ static void spi_set_cs(struct spi_device *spi, bool enable, bool force) + + if (spi->cs_gpiod || gpio_is_valid(spi->cs_gpio) || + !spi->controller->set_cs_timing) { +- if (!activate) ++ if (activate) ++ spi_delay_exec(&spi->cs_setup, NULL); ++ else + spi_delay_exec(&spi->cs_inactive, NULL); + } + } +diff --git a/drivers/staging/greybus/audio_topology.c b/drivers/staging/greybus/audio_topology.c +index 7f7d558b76d04..62d7674852bec 100644 +--- a/drivers/staging/greybus/audio_topology.c ++++ b/drivers/staging/greybus/audio_topology.c +@@ -147,6 +147,9 @@ static const char **gb_generate_enum_strings(struct gbaudio_module_info *gb, + + items = le32_to_cpu(gbenum->items); + strings = devm_kcalloc(gb->dev, items, sizeof(char *), GFP_KERNEL); ++ if (!strings) ++ return NULL; ++ + data = gbenum->names; + + for (i = 0; i < items; i++) { +@@ -655,6 +658,8 @@ static int gbaudio_tplg_create_enum_kctl(struct gbaudio_module_info *gb, + /* since count=1, and reg is dummy */ + gbe->items = le32_to_cpu(gb_enum->items); + gbe->texts = gb_generate_enum_strings(gb, gb_enum); ++ if (!gbe->texts) ++ return -ENOMEM; + + /* debug enum info */ + dev_dbg(gb->dev, "Max:%d, name_length:%d\n", gbe->items, +@@ -862,6 +867,8 @@ static int gbaudio_tplg_create_enum_ctl(struct gbaudio_module_info *gb, + /* since count=1, and reg is dummy */ + gbe->items = le32_to_cpu(gb_enum->items); + gbe->texts = gb_generate_enum_strings(gb, gb_enum); ++ if (!gbe->texts) ++ return -ENOMEM; + + /* debug enum info */ + dev_dbg(gb->dev, "Max:%d, name_length:%d\n", gbe->items, +@@ -1072,6 +1079,10 @@ static int gbaudio_tplg_create_widget(struct gbaudio_module_info *module, + csize += le16_to_cpu(gbenum->names_length); + control->texts = (const char * const *) + gb_generate_enum_strings(module, gbenum); ++ if (!control->texts) { ++ ret = -ENOMEM; ++ goto error; ++ } + control->items = le32_to_cpu(gbenum->items); + } else { + csize = sizeof(struct gb_audio_control); +@@ -1181,6 +1192,10 @@ static int gbaudio_tplg_process_kcontrols(struct gbaudio_module_info *module, + csize += le16_to_cpu(gbenum->names_length); + control->texts = (const char * const *) + gb_generate_enum_strings(module, gbenum); ++ if (!control->texts) { ++ ret = -ENOMEM; ++ goto error; ++ } + control->items = le32_to_cpu(gbenum->items); + } else { + csize = sizeof(struct gb_audio_control); +diff --git a/drivers/staging/media/atomisp/i2c/ov2680.h b/drivers/staging/media/atomisp/i2c/ov2680.h +index 874115f35fcad..798b28e134b64 100644 +--- a/drivers/staging/media/atomisp/i2c/ov2680.h ++++ b/drivers/staging/media/atomisp/i2c/ov2680.h +@@ -289,8 +289,6 @@ static struct ov2680_reg const ov2680_global_setting[] = { + */ + static struct ov2680_reg const ov2680_QCIF_30fps[] = { + {0x3086, 0x01}, +- {0x3501, 0x24}, +- {0x3502, 0x40}, + {0x370a, 0x23}, + {0x3801, 0xa0}, + {0x3802, 0x00}, +@@ -334,8 +332,6 @@ static struct ov2680_reg const ov2680_QCIF_30fps[] = { + */ + static struct ov2680_reg const ov2680_CIF_30fps[] = { + {0x3086, 0x01}, +- {0x3501, 0x24}, +- {0x3502, 0x40}, + {0x370a, 0x23}, + {0x3801, 0xa0}, + {0x3802, 0x00}, +@@ -377,8 +373,6 @@ static struct ov2680_reg const ov2680_CIF_30fps[] = { + */ + static struct ov2680_reg const ov2680_QVGA_30fps[] = { + {0x3086, 0x01}, +- {0x3501, 0x24}, +- {0x3502, 0x40}, + {0x370a, 0x23}, + {0x3801, 0xa0}, + {0x3802, 0x00}, +@@ -420,8 +414,6 @@ static struct ov2680_reg const ov2680_QVGA_30fps[] = { + */ + static struct ov2680_reg const ov2680_656x496_30fps[] = { + {0x3086, 0x01}, +- {0x3501, 0x24}, +- {0x3502, 0x40}, + {0x370a, 0x23}, + {0x3801, 0xa0}, + {0x3802, 0x00}, +@@ -463,8 +455,6 @@ static struct ov2680_reg const ov2680_656x496_30fps[] = { + */ + static struct ov2680_reg const ov2680_720x592_30fps[] = { + {0x3086, 0x01}, +- {0x3501, 0x26}, +- {0x3502, 0x40}, + {0x370a, 0x23}, + {0x3801, 0x00}, // X_ADDR_START; + {0x3802, 0x00}, +@@ -508,8 +498,6 @@ static struct ov2680_reg const ov2680_720x592_30fps[] = { + */ + static struct ov2680_reg const ov2680_800x600_30fps[] = { + {0x3086, 0x01}, +- {0x3501, 0x26}, +- {0x3502, 0x40}, + {0x370a, 0x23}, + {0x3801, 0x00}, + {0x3802, 0x00}, +@@ -551,8 +539,6 @@ static struct ov2680_reg const ov2680_800x600_30fps[] = { + */ + static struct ov2680_reg const ov2680_720p_30fps[] = { + {0x3086, 0x00}, +- {0x3501, 0x48}, +- {0x3502, 0xe0}, + {0x370a, 0x21}, + {0x3801, 0xa0}, + {0x3802, 0x00}, +@@ -594,8 +580,6 @@ static struct ov2680_reg const ov2680_720p_30fps[] = { + */ + static struct ov2680_reg const ov2680_1296x976_30fps[] = { + {0x3086, 0x00}, +- {0x3501, 0x48}, +- {0x3502, 0xe0}, + {0x370a, 0x21}, + {0x3801, 0xa0}, + {0x3802, 0x00}, +@@ -637,8 +621,6 @@ static struct ov2680_reg const ov2680_1296x976_30fps[] = { + */ + static struct ov2680_reg const ov2680_1456x1096_30fps[] = { + {0x3086, 0x00}, +- {0x3501, 0x48}, +- {0x3502, 0xe0}, + {0x370a, 0x21}, + {0x3801, 0x90}, + {0x3802, 0x00}, +@@ -682,8 +664,6 @@ static struct ov2680_reg const ov2680_1456x1096_30fps[] = { + + static struct ov2680_reg const ov2680_1616x916_30fps[] = { + {0x3086, 0x00}, +- {0x3501, 0x48}, +- {0x3502, 0xe0}, + {0x370a, 0x21}, + {0x3801, 0x00}, + {0x3802, 0x00}, +@@ -726,8 +706,6 @@ static struct ov2680_reg const ov2680_1616x916_30fps[] = { + #if 0 + static struct ov2680_reg const ov2680_1616x1082_30fps[] = { + {0x3086, 0x00}, +- {0x3501, 0x48}, +- {0x3502, 0xe0}, + {0x370a, 0x21}, + {0x3801, 0x00}, + {0x3802, 0x00}, +@@ -769,8 +747,6 @@ static struct ov2680_reg const ov2680_1616x1082_30fps[] = { + */ + static struct ov2680_reg const ov2680_1616x1216_30fps[] = { + {0x3086, 0x00}, +- {0x3501, 0x48}, +- {0x3502, 0xe0}, + {0x370a, 0x21}, + {0x3801, 0x00}, + {0x3802, 0x00}, +diff --git a/drivers/staging/media/atomisp/pci/atomisp_cmd.c b/drivers/staging/media/atomisp/pci/atomisp_cmd.c +index 366161cff5602..ef0b0963cf930 100644 +--- a/drivers/staging/media/atomisp/pci/atomisp_cmd.c ++++ b/drivers/staging/media/atomisp/pci/atomisp_cmd.c +@@ -1715,6 +1715,12 @@ void atomisp_wdt_refresh_pipe(struct atomisp_video_pipe *pipe, + { + unsigned long next; + ++ if (!pipe->asd) { ++ dev_err(pipe->isp->dev, "%s(): asd is NULL, device is %s\n", ++ __func__, pipe->vdev.name); ++ return; ++ } ++ + if (delay != ATOMISP_WDT_KEEP_CURRENT_DELAY) + pipe->wdt_duration = delay; + +@@ -1777,6 +1783,12 @@ void atomisp_wdt_refresh(struct atomisp_sub_device *asd, unsigned int delay) + /* ISP2401 */ + void atomisp_wdt_stop_pipe(struct atomisp_video_pipe *pipe, bool sync) + { ++ if (!pipe->asd) { ++ dev_err(pipe->isp->dev, "%s(): asd is NULL, device is %s\n", ++ __func__, pipe->vdev.name); ++ return; ++ } ++ + if (!atomisp_is_wdt_running(pipe)) + return; + +@@ -4109,6 +4121,12 @@ void atomisp_handle_parameter_and_buffer(struct atomisp_video_pipe *pipe) + unsigned long irqflags; + bool need_to_enqueue_buffer = false; + ++ if (!asd) { ++ dev_err(pipe->isp->dev, "%s(): asd is NULL, device is %s\n", ++ __func__, pipe->vdev.name); ++ return; ++ } ++ + if (atomisp_is_vf_pipe(pipe)) + return; + +@@ -4196,6 +4214,12 @@ int atomisp_set_parameters(struct video_device *vdev, + struct atomisp_css_params *css_param = &asd->params.css_param; + int ret; + ++ if (!asd) { ++ dev_err(pipe->isp->dev, "%s(): asd is NULL, device is %s\n", ++ __func__, vdev->name); ++ return -EINVAL; ++ } ++ + if (!asd->stream_env[ATOMISP_INPUT_STREAM_GENERAL].stream) { + dev_err(asd->isp->dev, "%s: internal error!\n", __func__); + return -EINVAL; +@@ -4857,6 +4881,12 @@ int atomisp_try_fmt(struct video_device *vdev, struct v4l2_pix_format *f, + int source_pad = atomisp_subdev_source_pad(vdev); + int ret; + ++ if (!asd) { ++ dev_err(isp->dev, "%s(): asd is NULL, device is %s\n", ++ __func__, vdev->name); ++ return -EINVAL; ++ } ++ + if (!isp->inputs[asd->input_curr].camera) + return -EINVAL; + +@@ -5194,10 +5224,17 @@ static int atomisp_set_fmt_to_isp(struct video_device *vdev, + int (*configure_pp_input)(struct atomisp_sub_device *asd, + unsigned int width, unsigned int height) = + configure_pp_input_nop; +- u16 stream_index = atomisp_source_pad_to_stream_id(asd, source_pad); ++ u16 stream_index; + const struct atomisp_in_fmt_conv *fc; + int ret, i; + ++ if (!asd) { ++ dev_err(isp->dev, "%s(): asd is NULL, device is %s\n", ++ __func__, vdev->name); ++ return -EINVAL; ++ } ++ stream_index = atomisp_source_pad_to_stream_id(asd, source_pad); ++ + v4l2_fh_init(&fh.vfh, vdev); + + isp_sink_crop = atomisp_subdev_get_rect( +@@ -5493,7 +5530,8 @@ static int atomisp_set_fmt_to_snr(struct video_device *vdev, + unsigned int padding_w, unsigned int padding_h, + unsigned int dvs_env_w, unsigned int dvs_env_h) + { +- struct atomisp_sub_device *asd = atomisp_to_video_pipe(vdev)->asd; ++ struct atomisp_video_pipe *pipe = atomisp_to_video_pipe(vdev); ++ struct atomisp_sub_device *asd = pipe->asd; + const struct atomisp_format_bridge *format; + struct v4l2_subdev_pad_config pad_cfg; + struct v4l2_subdev_state pad_state = { +@@ -5504,7 +5542,7 @@ static int atomisp_set_fmt_to_snr(struct video_device *vdev, + }; + struct v4l2_mbus_framefmt *ffmt = &vformat.format; + struct v4l2_mbus_framefmt *req_ffmt; +- struct atomisp_device *isp = asd->isp; ++ struct atomisp_device *isp; + struct atomisp_input_stream_info *stream_info = + (struct atomisp_input_stream_info *)ffmt->reserved; + u16 stream_index = ATOMISP_INPUT_STREAM_GENERAL; +@@ -5512,6 +5550,14 @@ static int atomisp_set_fmt_to_snr(struct video_device *vdev, + struct v4l2_subdev_fh fh; + int ret; + ++ if (!asd) { ++ dev_err(pipe->isp->dev, "%s(): asd is NULL, device is %s\n", ++ __func__, vdev->name); ++ return -EINVAL; ++ } ++ ++ isp = asd->isp; ++ + v4l2_fh_init(&fh.vfh, vdev); + + stream_index = atomisp_source_pad_to_stream_id(asd, source_pad); +@@ -5602,6 +5648,12 @@ int atomisp_set_fmt(struct video_device *vdev, struct v4l2_format *f) + struct v4l2_subdev_fh fh; + int ret; + ++ if (!asd) { ++ dev_err(isp->dev, "%s(): asd is NULL, device is %s\n", ++ __func__, vdev->name); ++ return -EINVAL; ++ } ++ + if (source_pad >= ATOMISP_SUBDEV_PADS_NUM) + return -EINVAL; + +@@ -6034,6 +6086,12 @@ int atomisp_set_fmt_file(struct video_device *vdev, struct v4l2_format *f) + struct v4l2_subdev_fh fh; + int ret; + ++ if (!asd) { ++ dev_err(isp->dev, "%s(): asd is NULL, device is %s\n", ++ __func__, vdev->name); ++ return -EINVAL; ++ } ++ + v4l2_fh_init(&fh.vfh, vdev); + + dev_dbg(isp->dev, "setting fmt %ux%u 0x%x for file inject\n", +@@ -6359,6 +6417,12 @@ bool atomisp_is_vf_pipe(struct atomisp_video_pipe *pipe) + { + struct atomisp_sub_device *asd = pipe->asd; + ++ if (!asd) { ++ dev_err(pipe->isp->dev, "%s(): asd is NULL, device is %s\n", ++ __func__, pipe->vdev.name); ++ return false; ++ } ++ + if (pipe == &asd->video_out_vf) + return true; + +@@ -6572,6 +6636,12 @@ static int atomisp_get_pipe_id(struct atomisp_video_pipe *pipe) + { + struct atomisp_sub_device *asd = pipe->asd; + ++ if (!asd) { ++ dev_err(pipe->isp->dev, "%s(): asd is NULL, device is %s\n", ++ __func__, pipe->vdev.name); ++ return -EINVAL; ++ } ++ + if (ATOMISP_USE_YUVPP(asd)) { + return IA_CSS_PIPE_ID_YUVPP; + } else if (asd->vfpp->val == ATOMISP_VFPP_DISABLE_SCALER) { +@@ -6609,6 +6679,12 @@ int atomisp_get_invalid_frame_num(struct video_device *vdev, + struct ia_css_pipe_info p_info; + int ret; + ++ if (!asd) { ++ dev_err(pipe->isp->dev, "%s(): asd is NULL, device is %s\n", ++ __func__, vdev->name); ++ return -EINVAL; ++ } ++ + if (asd->isp->inputs[asd->input_curr].camera_caps-> + sensor[asd->sensor_curr].stream_num > 1) { + /* External ISP */ +diff --git a/drivers/staging/media/atomisp/pci/atomisp_fops.c b/drivers/staging/media/atomisp/pci/atomisp_fops.c +index f82bf082aa796..18fff47bd25d2 100644 +--- a/drivers/staging/media/atomisp/pci/atomisp_fops.c ++++ b/drivers/staging/media/atomisp/pci/atomisp_fops.c +@@ -877,6 +877,11 @@ done: + else + pipe->users++; + rt_mutex_unlock(&isp->mutex); ++ ++ /* Ensure that a mode is set */ ++ if (asd) ++ v4l2_ctrl_s_ctrl(asd->run_mode, pipe->default_run_mode); ++ + return 0; + + css_error: +@@ -1171,6 +1176,12 @@ static int atomisp_mmap(struct file *file, struct vm_area_struct *vma) + u32 origin_size, new_size; + int ret; + ++ if (!asd) { ++ dev_err(isp->dev, "%s(): asd is NULL, device is %s\n", ++ __func__, vdev->name); ++ return -EINVAL; ++ } ++ + if (!(vma->vm_flags & (VM_WRITE | VM_READ))) + return -EACCES; + +diff --git a/drivers/staging/media/atomisp/pci/atomisp_gmin_platform.c b/drivers/staging/media/atomisp/pci/atomisp_gmin_platform.c +index d8c9e31314b2e..62dc06e224765 100644 +--- a/drivers/staging/media/atomisp/pci/atomisp_gmin_platform.c ++++ b/drivers/staging/media/atomisp/pci/atomisp_gmin_platform.c +@@ -481,7 +481,7 @@ fail: + + static u8 gmin_get_pmic_id_and_addr(struct device *dev) + { +- struct i2c_client *power; ++ struct i2c_client *power = NULL; + static u8 pmic_i2c_addr; + + if (pmic_id) +diff --git a/drivers/staging/media/atomisp/pci/atomisp_ioctl.c b/drivers/staging/media/atomisp/pci/atomisp_ioctl.c +index c8a625667e81e..b7dda4b96d49c 100644 +--- a/drivers/staging/media/atomisp/pci/atomisp_ioctl.c ++++ b/drivers/staging/media/atomisp/pci/atomisp_ioctl.c +@@ -646,6 +646,12 @@ static int atomisp_g_input(struct file *file, void *fh, unsigned int *input) + struct atomisp_device *isp = video_get_drvdata(vdev); + struct atomisp_sub_device *asd = atomisp_to_video_pipe(vdev)->asd; + ++ if (!asd) { ++ dev_err(isp->dev, "%s(): asd is NULL, device is %s\n", ++ __func__, vdev->name); ++ return -EINVAL; ++ } ++ + rt_mutex_lock(&isp->mutex); + *input = asd->input_curr; + rt_mutex_unlock(&isp->mutex); +@@ -665,6 +671,12 @@ static int atomisp_s_input(struct file *file, void *fh, unsigned int input) + struct v4l2_subdev *motor; + int ret; + ++ if (!asd) { ++ dev_err(isp->dev, "%s(): asd is NULL, device is %s\n", ++ __func__, vdev->name); ++ return -EINVAL; ++ } ++ + rt_mutex_lock(&isp->mutex); + if (input >= ATOM_ISP_MAX_INPUTS || input >= isp->input_cnt) { + dev_dbg(isp->dev, "input_cnt: %d\n", isp->input_cnt); +@@ -761,18 +773,33 @@ static int atomisp_enum_fmt_cap(struct file *file, void *fh, + struct video_device *vdev = video_devdata(file); + struct atomisp_device *isp = video_get_drvdata(vdev); + struct atomisp_sub_device *asd = atomisp_to_video_pipe(vdev)->asd; +- struct v4l2_subdev_mbus_code_enum code = { 0 }; ++ struct v4l2_subdev_mbus_code_enum code = { ++ .which = V4L2_SUBDEV_FORMAT_ACTIVE, ++ }; ++ struct v4l2_subdev *camera; + unsigned int i, fi = 0; + int rval; + ++ if (!asd) { ++ dev_err(isp->dev, "%s(): asd is NULL, device is %s\n", ++ __func__, vdev->name); ++ return -EINVAL; ++ } ++ ++ camera = isp->inputs[asd->input_curr].camera; ++ if(!camera) { ++ dev_err(isp->dev, "%s(): camera is NULL, device is %s\n", ++ __func__, vdev->name); ++ return -EINVAL; ++ } ++ + rt_mutex_lock(&isp->mutex); +- rval = v4l2_subdev_call(isp->inputs[asd->input_curr].camera, pad, +- enum_mbus_code, NULL, &code); ++ ++ rval = v4l2_subdev_call(camera, pad, enum_mbus_code, NULL, &code); + if (rval == -ENOIOCTLCMD) { + dev_warn(isp->dev, +- "enum_mbus_code pad op not supported. Please fix your sensor driver!\n"); +- // rval = v4l2_subdev_call(isp->inputs[asd->input_curr].camera, +- // video, enum_mbus_fmt, 0, &code.code); ++ "enum_mbus_code pad op not supported by %s. Please fix your sensor driver!\n", ++ camera->name); + } + rt_mutex_unlock(&isp->mutex); + +@@ -802,6 +829,8 @@ static int atomisp_enum_fmt_cap(struct file *file, void *fh, + f->pixelformat = format->pixelformat; + return 0; + } ++ dev_err(isp->dev, "%s(): format for code %x not found.\n", ++ __func__, code.code); + + return -EINVAL; + } +@@ -834,6 +863,72 @@ static int atomisp_g_fmt_file(struct file *file, void *fh, + return 0; + } + ++static int atomisp_adjust_fmt(struct v4l2_format *f) ++{ ++ const struct atomisp_format_bridge *format_bridge; ++ u32 padded_width; ++ ++ format_bridge = atomisp_get_format_bridge(f->fmt.pix.pixelformat); ++ ++ padded_width = f->fmt.pix.width + pad_w; ++ ++ if (format_bridge->planar) { ++ f->fmt.pix.bytesperline = padded_width; ++ f->fmt.pix.sizeimage = PAGE_ALIGN(f->fmt.pix.height * ++ DIV_ROUND_UP(format_bridge->depth * ++ padded_width, 8)); ++ } else { ++ f->fmt.pix.bytesperline = DIV_ROUND_UP(format_bridge->depth * ++ padded_width, 8); ++ f->fmt.pix.sizeimage = PAGE_ALIGN(f->fmt.pix.height * f->fmt.pix.bytesperline); ++ } ++ ++ if (f->fmt.pix.field == V4L2_FIELD_ANY) ++ f->fmt.pix.field = V4L2_FIELD_NONE; ++ ++ format_bridge = atomisp_get_format_bridge(f->fmt.pix.pixelformat); ++ if (!format_bridge) ++ return -EINVAL; ++ ++ /* Currently, raw formats are broken!!! */ ++ if (format_bridge->sh_fmt == IA_CSS_FRAME_FORMAT_RAW) { ++ f->fmt.pix.pixelformat = V4L2_PIX_FMT_YUV420; ++ ++ format_bridge = atomisp_get_format_bridge(f->fmt.pix.pixelformat); ++ if (!format_bridge) ++ return -EINVAL; ++ } ++ ++ padded_width = f->fmt.pix.width + pad_w; ++ ++ if (format_bridge->planar) { ++ f->fmt.pix.bytesperline = padded_width; ++ f->fmt.pix.sizeimage = PAGE_ALIGN(f->fmt.pix.height * ++ DIV_ROUND_UP(format_bridge->depth * ++ padded_width, 8)); ++ } else { ++ f->fmt.pix.bytesperline = DIV_ROUND_UP(format_bridge->depth * ++ padded_width, 8); ++ f->fmt.pix.sizeimage = PAGE_ALIGN(f->fmt.pix.height * f->fmt.pix.bytesperline); ++ } ++ ++ if (f->fmt.pix.field == V4L2_FIELD_ANY) ++ f->fmt.pix.field = V4L2_FIELD_NONE; ++ ++ /* ++ * FIXME: do we need to setup this differently, depending on the ++ * sensor or the pipeline? ++ */ ++ f->fmt.pix.colorspace = V4L2_COLORSPACE_REC709; ++ f->fmt.pix.ycbcr_enc = V4L2_YCBCR_ENC_709; ++ f->fmt.pix.xfer_func = V4L2_XFER_FUNC_709; ++ ++ f->fmt.pix.width -= pad_w; ++ f->fmt.pix.height -= pad_h; ++ ++ return 0; ++} ++ + /* This function looks up the closest available resolution. */ + static int atomisp_try_fmt_cap(struct file *file, void *fh, + struct v4l2_format *f) +@@ -845,7 +940,11 @@ static int atomisp_try_fmt_cap(struct file *file, void *fh, + rt_mutex_lock(&isp->mutex); + ret = atomisp_try_fmt(vdev, &f->fmt.pix, NULL); + rt_mutex_unlock(&isp->mutex); +- return ret; ++ ++ if (ret) ++ return ret; ++ ++ return atomisp_adjust_fmt(f); + } + + static int atomisp_s_fmt_cap(struct file *file, void *fh, +@@ -1024,9 +1123,16 @@ int __atomisp_reqbufs(struct file *file, void *fh, + struct ia_css_frame *frame; + struct videobuf_vmalloc_memory *vm_mem; + u16 source_pad = atomisp_subdev_source_pad(vdev); +- u16 stream_id = atomisp_source_pad_to_stream_id(asd, source_pad); ++ u16 stream_id; + int ret = 0, i = 0; + ++ if (!asd) { ++ dev_err(pipe->isp->dev, "%s(): asd is NULL, device is %s\n", ++ __func__, vdev->name); ++ return -EINVAL; ++ } ++ stream_id = atomisp_source_pad_to_stream_id(asd, source_pad); ++ + if (req->count == 0) { + mutex_lock(&pipe->capq.vb_lock); + if (!list_empty(&pipe->capq.stream)) +@@ -1154,6 +1260,12 @@ static int atomisp_qbuf(struct file *file, void *fh, struct v4l2_buffer *buf) + u32 pgnr; + int ret = 0; + ++ if (!asd) { ++ dev_err(isp->dev, "%s(): asd is NULL, device is %s\n", ++ __func__, vdev->name); ++ return -EINVAL; ++ } ++ + rt_mutex_lock(&isp->mutex); + if (isp->isp_fatal_error) { + ret = -EIO; +@@ -1389,6 +1501,12 @@ static int atomisp_dqbuf(struct file *file, void *fh, struct v4l2_buffer *buf) + struct atomisp_device *isp = video_get_drvdata(vdev); + int ret = 0; + ++ if (!asd) { ++ dev_err(isp->dev, "%s(): asd is NULL, device is %s\n", ++ __func__, vdev->name); ++ return -EINVAL; ++ } ++ + rt_mutex_lock(&isp->mutex); + + if (isp->isp_fatal_error) { +@@ -1640,6 +1758,12 @@ static int atomisp_streamon(struct file *file, void *fh, + int ret = 0; + unsigned long irqflags; + ++ if (!asd) { ++ dev_err(isp->dev, "%s(): asd is NULL, device is %s\n", ++ __func__, vdev->name); ++ return -EINVAL; ++ } ++ + dev_dbg(isp->dev, "Start stream on pad %d for asd%d\n", + atomisp_subdev_source_pad(vdev), asd->index); + +@@ -1901,6 +2025,12 @@ int __atomisp_streamoff(struct file *file, void *fh, enum v4l2_buf_type type) + unsigned long flags; + bool first_streamoff = false; + ++ if (!asd) { ++ dev_err(isp->dev, "%s(): asd is NULL, device is %s\n", ++ __func__, vdev->name); ++ return -EINVAL; ++ } ++ + dev_dbg(isp->dev, "Stop stream on pad %d for asd%d\n", + atomisp_subdev_source_pad(vdev), asd->index); + +@@ -2150,6 +2280,12 @@ static int atomisp_g_ctrl(struct file *file, void *fh, + struct atomisp_device *isp = video_get_drvdata(vdev); + int i, ret = -EINVAL; + ++ if (!asd) { ++ dev_err(isp->dev, "%s(): asd is NULL, device is %s\n", ++ __func__, vdev->name); ++ return -EINVAL; ++ } ++ + for (i = 0; i < ctrls_num; i++) { + if (ci_v4l2_controls[i].id == control->id) { + ret = 0; +@@ -2229,6 +2365,12 @@ static int atomisp_s_ctrl(struct file *file, void *fh, + struct atomisp_device *isp = video_get_drvdata(vdev); + int i, ret = -EINVAL; + ++ if (!asd) { ++ dev_err(isp->dev, "%s(): asd is NULL, device is %s\n", ++ __func__, vdev->name); ++ return -EINVAL; ++ } ++ + for (i = 0; i < ctrls_num; i++) { + if (ci_v4l2_controls[i].id == control->id) { + ret = 0; +@@ -2310,6 +2452,12 @@ static int atomisp_queryctl(struct file *file, void *fh, + struct atomisp_sub_device *asd = atomisp_to_video_pipe(vdev)->asd; + struct atomisp_device *isp = video_get_drvdata(vdev); + ++ if (!asd) { ++ dev_err(isp->dev, "%s(): asd is NULL, device is %s\n", ++ __func__, vdev->name); ++ return -EINVAL; ++ } ++ + switch (qc->id) { + case V4L2_CID_FOCUS_ABSOLUTE: + case V4L2_CID_FOCUS_RELATIVE: +@@ -2355,6 +2503,12 @@ static int atomisp_camera_g_ext_ctrls(struct file *file, void *fh, + int i; + int ret = 0; + ++ if (!asd) { ++ dev_err(isp->dev, "%s(): asd is NULL, device is %s\n", ++ __func__, vdev->name); ++ return -EINVAL; ++ } ++ + if (!IS_ISP2401) + motor = isp->inputs[asd->input_curr].motor; + else +@@ -2466,6 +2620,12 @@ static int atomisp_camera_s_ext_ctrls(struct file *file, void *fh, + int i; + int ret = 0; + ++ if (!asd) { ++ dev_err(isp->dev, "%s(): asd is NULL, device is %s\n", ++ __func__, vdev->name); ++ return -EINVAL; ++ } ++ + if (!IS_ISP2401) + motor = isp->inputs[asd->input_curr].motor; + else +@@ -2591,6 +2751,12 @@ static int atomisp_g_parm(struct file *file, void *fh, + struct atomisp_sub_device *asd = atomisp_to_video_pipe(vdev)->asd; + struct atomisp_device *isp = video_get_drvdata(vdev); + ++ if (!asd) { ++ dev_err(isp->dev, "%s(): asd is NULL, device is %s\n", ++ __func__, vdev->name); ++ return -EINVAL; ++ } ++ + if (parm->type != V4L2_BUF_TYPE_VIDEO_CAPTURE) { + dev_err(isp->dev, "unsupported v4l2 buf type\n"); + return -EINVAL; +@@ -2613,6 +2779,12 @@ static int atomisp_s_parm(struct file *file, void *fh, + int rval; + int fps; + ++ if (!asd) { ++ dev_err(isp->dev, "%s(): asd is NULL, device is %s\n", ++ __func__, vdev->name); ++ return -EINVAL; ++ } ++ + if (parm->type != V4L2_BUF_TYPE_VIDEO_CAPTURE) { + dev_err(isp->dev, "unsupported v4l2 buf type\n"); + return -EINVAL; +diff --git a/drivers/staging/media/atomisp/pci/atomisp_subdev.c b/drivers/staging/media/atomisp/pci/atomisp_subdev.c +index 12f22ad007c73..ffaf11e0b0ad8 100644 +--- a/drivers/staging/media/atomisp/pci/atomisp_subdev.c ++++ b/drivers/staging/media/atomisp/pci/atomisp_subdev.c +@@ -1164,23 +1164,28 @@ static int isp_subdev_init_entities(struct atomisp_sub_device *asd) + + atomisp_init_acc_pipe(asd, &asd->video_acc); + +- ret = atomisp_video_init(&asd->video_in, "MEMORY"); ++ ret = atomisp_video_init(&asd->video_in, "MEMORY", ++ ATOMISP_RUN_MODE_SDV); + if (ret < 0) + return ret; + +- ret = atomisp_video_init(&asd->video_out_capture, "CAPTURE"); ++ ret = atomisp_video_init(&asd->video_out_capture, "CAPTURE", ++ ATOMISP_RUN_MODE_STILL_CAPTURE); + if (ret < 0) + return ret; + +- ret = atomisp_video_init(&asd->video_out_vf, "VIEWFINDER"); ++ ret = atomisp_video_init(&asd->video_out_vf, "VIEWFINDER", ++ ATOMISP_RUN_MODE_CONTINUOUS_CAPTURE); + if (ret < 0) + return ret; + +- ret = atomisp_video_init(&asd->video_out_preview, "PREVIEW"); ++ ret = atomisp_video_init(&asd->video_out_preview, "PREVIEW", ++ ATOMISP_RUN_MODE_PREVIEW); + if (ret < 0) + return ret; + +- ret = atomisp_video_init(&asd->video_out_video_capture, "VIDEO"); ++ ret = atomisp_video_init(&asd->video_out_video_capture, "VIDEO", ++ ATOMISP_RUN_MODE_VIDEO); + if (ret < 0) + return ret; + +diff --git a/drivers/staging/media/atomisp/pci/atomisp_subdev.h b/drivers/staging/media/atomisp/pci/atomisp_subdev.h +index d6fcfab6352d7..a8d210ea5f8be 100644 +--- a/drivers/staging/media/atomisp/pci/atomisp_subdev.h ++++ b/drivers/staging/media/atomisp/pci/atomisp_subdev.h +@@ -81,6 +81,9 @@ struct atomisp_video_pipe { + /* the link list to store per_frame parameters */ + struct list_head per_frame_params; + ++ /* Store here the initial run mode */ ++ unsigned int default_run_mode; ++ + unsigned int buffers_in_css; + + /* irq_lock is used to protect video buffer state change operations and +diff --git a/drivers/staging/media/atomisp/pci/atomisp_v4l2.c b/drivers/staging/media/atomisp/pci/atomisp_v4l2.c +index 1e324f1f656e5..14c39b8987c95 100644 +--- a/drivers/staging/media/atomisp/pci/atomisp_v4l2.c ++++ b/drivers/staging/media/atomisp/pci/atomisp_v4l2.c +@@ -447,7 +447,8 @@ const struct atomisp_dfs_config dfs_config_cht_soc = { + .dfs_table_size = ARRAY_SIZE(dfs_rules_cht_soc), + }; + +-int atomisp_video_init(struct atomisp_video_pipe *video, const char *name) ++int atomisp_video_init(struct atomisp_video_pipe *video, const char *name, ++ unsigned int run_mode) + { + int ret; + const char *direction; +@@ -478,6 +479,7 @@ int atomisp_video_init(struct atomisp_video_pipe *video, const char *name) + "ATOMISP ISP %s %s", name, direction); + video->vdev.release = video_device_release_empty; + video_set_drvdata(&video->vdev, video->isp); ++ video->default_run_mode = run_mode; + + return 0; + } +@@ -711,15 +713,15 @@ static int atomisp_mrfld_power(struct atomisp_device *isp, bool enable) + + dev_dbg(isp->dev, "IUNIT power-%s.\n", enable ? "on" : "off"); + +- /*WA:Enable DVFS*/ ++ /* WA for P-Unit, if DVFS enabled, ISP timeout observed */ + if (IS_CHT && enable) +- punit_ddr_dvfs_enable(true); ++ punit_ddr_dvfs_enable(false); + + /* + * FIXME:WA for ECS28A, with this sleep, CTS + * android.hardware.camera2.cts.CameraDeviceTest#testCameraDeviceAbort + * PASS, no impact on other platforms +- */ ++ */ + if (IS_BYT && enable) + msleep(10); + +@@ -727,7 +729,7 @@ static int atomisp_mrfld_power(struct atomisp_device *isp, bool enable) + iosf_mbi_modify(BT_MBI_UNIT_PMC, MBI_REG_READ, MRFLD_ISPSSPM0, + val, MRFLD_ISPSSPM0_ISPSSC_MASK); + +- /*WA:Enable DVFS*/ ++ /* WA:Enable DVFS */ + if (IS_CHT && !enable) + punit_ddr_dvfs_enable(true); + +@@ -1182,6 +1184,7 @@ static void atomisp_unregister_entities(struct atomisp_device *isp) + + v4l2_device_unregister(&isp->v4l2_dev); + media_device_unregister(&isp->media_dev); ++ media_device_cleanup(&isp->media_dev); + } + + static int atomisp_register_entities(struct atomisp_device *isp) +diff --git a/drivers/staging/media/atomisp/pci/atomisp_v4l2.h b/drivers/staging/media/atomisp/pci/atomisp_v4l2.h +index 81bb356b81720..72611b8286a4a 100644 +--- a/drivers/staging/media/atomisp/pci/atomisp_v4l2.h ++++ b/drivers/staging/media/atomisp/pci/atomisp_v4l2.h +@@ -27,7 +27,8 @@ struct v4l2_device; + struct atomisp_device; + struct firmware; + +-int atomisp_video_init(struct atomisp_video_pipe *video, const char *name); ++int atomisp_video_init(struct atomisp_video_pipe *video, const char *name, ++ unsigned int run_mode); + void atomisp_acc_init(struct atomisp_acc_pipe *video, const char *name); + void atomisp_video_unregister(struct atomisp_video_pipe *video); + void atomisp_acc_unregister(struct atomisp_acc_pipe *video); +diff --git a/drivers/staging/media/atomisp/pci/sh_css.c b/drivers/staging/media/atomisp/pci/sh_css.c +index c4b35cbab3737..ba25d0da8b811 100644 +--- a/drivers/staging/media/atomisp/pci/sh_css.c ++++ b/drivers/staging/media/atomisp/pci/sh_css.c +@@ -522,6 +522,7 @@ ia_css_stream_input_format_bits_per_pixel(struct ia_css_stream *stream) + return bpp; + } + ++/* TODO: move define to proper file in tools */ + #define GP_ISEL_TPG_MODE 0x90058 + + #if !defined(ISP2401) +@@ -573,12 +574,8 @@ sh_css_config_input_network(struct ia_css_stream *stream) + vblank_cycles = vblank_lines * (width + hblank_cycles); + sh_css_sp_configure_sync_gen(width, height, hblank_cycles, + vblank_cycles); +- if (!IS_ISP2401) { +- if (pipe->stream->config.mode == IA_CSS_INPUT_MODE_TPG) { +- /* TODO: move define to proper file in tools */ +- ia_css_device_store_uint32(GP_ISEL_TPG_MODE, 0); +- } +- } ++ if (pipe->stream->config.mode == IA_CSS_INPUT_MODE_TPG) ++ ia_css_device_store_uint32(GP_ISEL_TPG_MODE, 0); + } + ia_css_debug_dtrace(IA_CSS_DEBUG_TRACE_PRIVATE, + "sh_css_config_input_network() leave:\n"); +@@ -1009,16 +1006,14 @@ static bool sh_css_translate_stream_cfg_to_isys_stream_descr( + * ia_css_isys_stream_capture_indication() instead of + * ia_css_pipeline_sp_wait_for_isys_stream_N() as isp processing of + * capture takes longer than getting an ISYS frame +- * +- * Only 2401 relevant ?? + */ +-#if 0 // FIXME: NOT USED on Yocto Aero +- isys_stream_descr->polling_mode +- = early_polling ? INPUT_SYSTEM_POLL_ON_CAPTURE_REQUEST +- : INPUT_SYSTEM_POLL_ON_WAIT_FOR_FRAME; +- ia_css_debug_dtrace(IA_CSS_DEBUG_TRACE_PRIVATE, +- "sh_css_translate_stream_cfg_to_isys_stream_descr() leave:\n"); +-#endif ++ if (IS_ISP2401) { ++ isys_stream_descr->polling_mode ++ = early_polling ? INPUT_SYSTEM_POLL_ON_CAPTURE_REQUEST ++ : INPUT_SYSTEM_POLL_ON_WAIT_FOR_FRAME; ++ ia_css_debug_dtrace(IA_CSS_DEBUG_TRACE_PRIVATE, ++ "sh_css_translate_stream_cfg_to_isys_stream_descr() leave:\n"); ++ } + + return rc; + } +@@ -1433,7 +1428,7 @@ static void start_pipe( + + assert(me); /* all callers are in this file and call with non null argument */ + +- if (!IS_ISP2401) { ++ if (IS_ISP2401) { + coord = &me->config.internal_frame_origin_bqs_on_sctbl; + params = me->stream->isp_params_configs; + } +diff --git a/drivers/staging/media/atomisp/pci/sh_css_mipi.c b/drivers/staging/media/atomisp/pci/sh_css_mipi.c +index 75489f7d75eec..c1f2f6151c5f8 100644 +--- a/drivers/staging/media/atomisp/pci/sh_css_mipi.c ++++ b/drivers/staging/media/atomisp/pci/sh_css_mipi.c +@@ -374,17 +374,17 @@ static bool buffers_needed(struct ia_css_pipe *pipe) + { + if (!IS_ISP2401) { + if (pipe->stream->config.mode == IA_CSS_INPUT_MODE_BUFFERED_SENSOR) +- return false; +- else + return true; ++ else ++ return false; + } + + if (pipe->stream->config.mode == IA_CSS_INPUT_MODE_BUFFERED_SENSOR || + pipe->stream->config.mode == IA_CSS_INPUT_MODE_TPG || + pipe->stream->config.mode == IA_CSS_INPUT_MODE_PRBS) +- return false; ++ return true; + +- return true; ++ return false; + } + + int +@@ -423,14 +423,17 @@ allocate_mipi_frames(struct ia_css_pipe *pipe, + return 0; /* AM TODO: Check */ + } + +- if (!IS_ISP2401) ++ if (!IS_ISP2401) { + port = (unsigned int)pipe->stream->config.source.port.port; +- else +- err = ia_css_mipi_is_source_port_valid(pipe, &port); ++ } else { ++ /* Returns true if port is valid. So, invert it */ ++ err = !ia_css_mipi_is_source_port_valid(pipe, &port); ++ } + + assert(port < N_CSI_PORTS); + +- if (port >= N_CSI_PORTS || err) { ++ if ((!IS_ISP2401 && port >= N_CSI_PORTS) || ++ (IS_ISP2401 && err)) { + ia_css_debug_dtrace(IA_CSS_DEBUG_TRACE_PRIVATE, + "allocate_mipi_frames(%p) exit: error: port is not correct (port=%d).\n", + pipe, port); +@@ -552,14 +555,17 @@ free_mipi_frames(struct ia_css_pipe *pipe) + return err; + } + +- if (!IS_ISP2401) ++ if (!IS_ISP2401) { + port = (unsigned int)pipe->stream->config.source.port.port; +- else +- err = ia_css_mipi_is_source_port_valid(pipe, &port); ++ } else { ++ /* Returns true if port is valid. So, invert it */ ++ err = !ia_css_mipi_is_source_port_valid(pipe, &port); ++ } + + assert(port < N_CSI_PORTS); + +- if (port >= N_CSI_PORTS || err) { ++ if ((!IS_ISP2401 && port >= N_CSI_PORTS) || ++ (IS_ISP2401 && err)) { + ia_css_debug_dtrace(IA_CSS_DEBUG_TRACE_PRIVATE, + "free_mipi_frames(%p, %d) exit: error: pipe port is not correct.\n", + pipe, port); +@@ -663,14 +669,17 @@ send_mipi_frames(struct ia_css_pipe *pipe) + /* TODO: AM: maybe this should be returning an error. */ + } + +- if (!IS_ISP2401) ++ if (!IS_ISP2401) { + port = (unsigned int)pipe->stream->config.source.port.port; +- else +- err = ia_css_mipi_is_source_port_valid(pipe, &port); ++ } else { ++ /* Returns true if port is valid. So, invert it */ ++ err = !ia_css_mipi_is_source_port_valid(pipe, &port); ++ } + + assert(port < N_CSI_PORTS); + +- if (port >= N_CSI_PORTS || err) { ++ if ((!IS_ISP2401 && port >= N_CSI_PORTS) || ++ (IS_ISP2401 && err)) { + IA_CSS_ERROR("send_mipi_frames(%p) exit: invalid port specified (port=%d).\n", + pipe, port); + return err; +diff --git a/drivers/staging/media/atomisp/pci/sh_css_params.c b/drivers/staging/media/atomisp/pci/sh_css_params.c +index dbd3bfe3d343c..ccc0078795648 100644 +--- a/drivers/staging/media/atomisp/pci/sh_css_params.c ++++ b/drivers/staging/media/atomisp/pci/sh_css_params.c +@@ -2431,7 +2431,7 @@ sh_css_create_isp_params(struct ia_css_stream *stream, + unsigned int i; + struct sh_css_ddr_address_map *ddr_ptrs; + struct sh_css_ddr_address_map_size *ddr_ptrs_size; +- int err = 0; ++ int err; + size_t params_size; + struct ia_css_isp_parameters *params = + kvmalloc(sizeof(struct ia_css_isp_parameters), GFP_KERNEL); +@@ -2473,7 +2473,11 @@ sh_css_create_isp_params(struct ia_css_stream *stream, + succ &= (ddr_ptrs->macc_tbl != mmgr_NULL); + + *isp_params_out = params; +- return err; ++ ++ if (!succ) ++ return -ENOMEM; ++ ++ return 0; + } + + static bool +diff --git a/drivers/staging/media/hantro/hantro_drv.c b/drivers/staging/media/hantro/hantro_drv.c +index 20e5081588719..281aa585e3375 100644 +--- a/drivers/staging/media/hantro/hantro_drv.c ++++ b/drivers/staging/media/hantro/hantro_drv.c +@@ -958,7 +958,7 @@ static int hantro_probe(struct platform_device *pdev) + ret = clk_bulk_prepare(vpu->variant->num_clocks, vpu->clocks); + if (ret) { + dev_err(&pdev->dev, "Failed to prepare clocks\n"); +- return ret; ++ goto err_pm_disable; + } + + ret = v4l2_device_register(&pdev->dev, &vpu->v4l2_dev); +@@ -1014,6 +1014,7 @@ err_v4l2_unreg: + v4l2_device_unregister(&vpu->v4l2_dev); + err_clk_unprepare: + clk_bulk_unprepare(vpu->variant->num_clocks, vpu->clocks); ++err_pm_disable: + pm_runtime_dont_use_autosuspend(vpu->dev); + pm_runtime_disable(vpu->dev); + return ret; +diff --git a/drivers/staging/media/hantro/hantro_h1_jpeg_enc.c b/drivers/staging/media/hantro/hantro_h1_jpeg_enc.c +index 56cf261a8e958..9cd713c02a455 100644 +--- a/drivers/staging/media/hantro/hantro_h1_jpeg_enc.c ++++ b/drivers/staging/media/hantro/hantro_h1_jpeg_enc.c +@@ -140,7 +140,7 @@ int hantro_h1_jpeg_enc_run(struct hantro_ctx *ctx) + return 0; + } + +-void hantro_jpeg_enc_done(struct hantro_ctx *ctx) ++void hantro_h1_jpeg_enc_done(struct hantro_ctx *ctx) + { + struct hantro_dev *vpu = ctx->dev; + u32 bytesused = vepu_read(vpu, H1_REG_STR_BUF_LIMIT) / 8; +diff --git a/drivers/staging/media/hantro/hantro_hw.h b/drivers/staging/media/hantro/hantro_hw.h +index df7b5e3a57b9b..fd738653a5735 100644 +--- a/drivers/staging/media/hantro/hantro_hw.h ++++ b/drivers/staging/media/hantro/hantro_hw.h +@@ -235,7 +235,8 @@ int hantro_h1_jpeg_enc_run(struct hantro_ctx *ctx); + int rockchip_vpu2_jpeg_enc_run(struct hantro_ctx *ctx); + int hantro_jpeg_enc_init(struct hantro_ctx *ctx); + void hantro_jpeg_enc_exit(struct hantro_ctx *ctx); +-void hantro_jpeg_enc_done(struct hantro_ctx *ctx); ++void hantro_h1_jpeg_enc_done(struct hantro_ctx *ctx); ++void rockchip_vpu2_jpeg_enc_done(struct hantro_ctx *ctx); + + dma_addr_t hantro_h264_get_ref_buf(struct hantro_ctx *ctx, + unsigned int dpb_idx); +diff --git a/drivers/staging/media/hantro/rockchip_vpu2_hw_jpeg_enc.c b/drivers/staging/media/hantro/rockchip_vpu2_hw_jpeg_enc.c +index 991213ce16108..5d9ff420f0b5f 100644 +--- a/drivers/staging/media/hantro/rockchip_vpu2_hw_jpeg_enc.c ++++ b/drivers/staging/media/hantro/rockchip_vpu2_hw_jpeg_enc.c +@@ -171,3 +171,20 @@ int rockchip_vpu2_jpeg_enc_run(struct hantro_ctx *ctx) + + return 0; + } ++ ++void rockchip_vpu2_jpeg_enc_done(struct hantro_ctx *ctx) ++{ ++ struct hantro_dev *vpu = ctx->dev; ++ u32 bytesused = vepu_read(vpu, VEPU_REG_STR_BUF_LIMIT) / 8; ++ struct vb2_v4l2_buffer *dst_buf = hantro_get_dst_buf(ctx); ++ ++ /* ++ * TODO: Rework the JPEG encoder to eliminate the need ++ * for a bounce buffer. ++ */ ++ memcpy(vb2_plane_vaddr(&dst_buf->vb2_buf, 0) + ++ ctx->vpu_dst_fmt->header_size, ++ ctx->jpeg_enc.bounce_buffer.cpu, bytesused); ++ vb2_set_plane_payload(&dst_buf->vb2_buf, 0, ++ ctx->vpu_dst_fmt->header_size + bytesused); ++} +diff --git a/drivers/staging/media/hantro/rockchip_vpu_hw.c b/drivers/staging/media/hantro/rockchip_vpu_hw.c +index d4f52957cc534..0c22039162a00 100644 +--- a/drivers/staging/media/hantro/rockchip_vpu_hw.c ++++ b/drivers/staging/media/hantro/rockchip_vpu_hw.c +@@ -343,7 +343,7 @@ static const struct hantro_codec_ops rk3066_vpu_codec_ops[] = { + .run = hantro_h1_jpeg_enc_run, + .reset = rockchip_vpu1_enc_reset, + .init = hantro_jpeg_enc_init, +- .done = hantro_jpeg_enc_done, ++ .done = hantro_h1_jpeg_enc_done, + .exit = hantro_jpeg_enc_exit, + }, + [HANTRO_MODE_H264_DEC] = { +@@ -371,7 +371,7 @@ static const struct hantro_codec_ops rk3288_vpu_codec_ops[] = { + .run = hantro_h1_jpeg_enc_run, + .reset = rockchip_vpu1_enc_reset, + .init = hantro_jpeg_enc_init, +- .done = hantro_jpeg_enc_done, ++ .done = hantro_h1_jpeg_enc_done, + .exit = hantro_jpeg_enc_exit, + }, + [HANTRO_MODE_H264_DEC] = { +@@ -399,6 +399,7 @@ static const struct hantro_codec_ops rk3399_vpu_codec_ops[] = { + .run = rockchip_vpu2_jpeg_enc_run, + .reset = rockchip_vpu2_enc_reset, + .init = hantro_jpeg_enc_init, ++ .done = rockchip_vpu2_jpeg_enc_done, + .exit = hantro_jpeg_enc_exit, + }, + [HANTRO_MODE_H264_DEC] = { +diff --git a/drivers/staging/rtl8192e/rtllib.h b/drivers/staging/rtl8192e/rtllib.h +index c6f8b772335c1..c985e4ebc545a 100644 +--- a/drivers/staging/rtl8192e/rtllib.h ++++ b/drivers/staging/rtl8192e/rtllib.h +@@ -1980,7 +1980,7 @@ void SendDisassociation(struct rtllib_device *ieee, bool deauth, u16 asRsn); + void rtllib_softmac_xmit(struct rtllib_txb *txb, struct rtllib_device *ieee); + + void rtllib_start_ibss(struct rtllib_device *ieee); +-void rtllib_softmac_init(struct rtllib_device *ieee); ++int rtllib_softmac_init(struct rtllib_device *ieee); + void rtllib_softmac_free(struct rtllib_device *ieee); + void rtllib_disassociate(struct rtllib_device *ieee); + void rtllib_stop_scan(struct rtllib_device *ieee); +diff --git a/drivers/staging/rtl8192e/rtllib_module.c b/drivers/staging/rtl8192e/rtllib_module.c +index 64d9feee1f392..f00ac94b2639b 100644 +--- a/drivers/staging/rtl8192e/rtllib_module.c ++++ b/drivers/staging/rtl8192e/rtllib_module.c +@@ -88,7 +88,7 @@ struct net_device *alloc_rtllib(int sizeof_priv) + err = rtllib_networks_allocate(ieee); + if (err) { + pr_err("Unable to allocate beacon storage: %d\n", err); +- goto failed; ++ goto free_netdev; + } + rtllib_networks_initialize(ieee); + +@@ -121,11 +121,13 @@ struct net_device *alloc_rtllib(int sizeof_priv) + ieee->hwsec_active = 0; + + memset(ieee->swcamtable, 0, sizeof(struct sw_cam_table) * 32); +- rtllib_softmac_init(ieee); ++ err = rtllib_softmac_init(ieee); ++ if (err) ++ goto free_crypt_info; + + ieee->pHTInfo = kzalloc(sizeof(struct rt_hi_throughput), GFP_KERNEL); + if (!ieee->pHTInfo) +- return NULL; ++ goto free_softmac; + + HTUpdateDefaultSetting(ieee); + HTInitializeHTInfo(ieee); +@@ -141,8 +143,14 @@ struct net_device *alloc_rtllib(int sizeof_priv) + + return dev; + +- failed: ++free_softmac: ++ rtllib_softmac_free(ieee); ++free_crypt_info: ++ lib80211_crypt_info_free(&ieee->crypt_info); ++ rtllib_networks_free(ieee); ++free_netdev: + free_netdev(dev); ++ + return NULL; + } + EXPORT_SYMBOL(alloc_rtllib); +diff --git a/drivers/staging/rtl8192e/rtllib_softmac.c b/drivers/staging/rtl8192e/rtllib_softmac.c +index d2726d01c7573..503d33be71d99 100644 +--- a/drivers/staging/rtl8192e/rtllib_softmac.c ++++ b/drivers/staging/rtl8192e/rtllib_softmac.c +@@ -2952,7 +2952,7 @@ void rtllib_start_protocol(struct rtllib_device *ieee) + } + } + +-void rtllib_softmac_init(struct rtllib_device *ieee) ++int rtllib_softmac_init(struct rtllib_device *ieee) + { + int i; + +@@ -2963,7 +2963,8 @@ void rtllib_softmac_init(struct rtllib_device *ieee) + ieee->seq_ctrl[i] = 0; + ieee->dot11d_info = kzalloc(sizeof(struct rt_dot11d_info), GFP_ATOMIC); + if (!ieee->dot11d_info) +- netdev_err(ieee->dev, "Can't alloc memory for DOT11D\n"); ++ return -ENOMEM; ++ + ieee->LinkDetectInfo.SlotIndex = 0; + ieee->LinkDetectInfo.SlotNum = 2; + ieee->LinkDetectInfo.NumRecvBcnInPeriod = 0; +@@ -3029,6 +3030,7 @@ void rtllib_softmac_init(struct rtllib_device *ieee) + + tasklet_setup(&ieee->ps_task, rtllib_sta_ps); + ++ return 0; + } + + void rtllib_softmac_free(struct rtllib_device *ieee) +diff --git a/drivers/tee/tee_core.c b/drivers/tee/tee_core.c +index 2b37bc408fc3d..85102d12d7169 100644 +--- a/drivers/tee/tee_core.c ++++ b/drivers/tee/tee_core.c +@@ -98,8 +98,10 @@ void teedev_ctx_put(struct tee_context *ctx) + + static void teedev_close_context(struct tee_context *ctx) + { +- tee_device_put(ctx->teedev); ++ struct tee_device *teedev = ctx->teedev; ++ + teedev_ctx_put(ctx); ++ tee_device_put(teedev); + } + + static int tee_open(struct inode *inode, struct file *filp) +diff --git a/drivers/thermal/imx8mm_thermal.c b/drivers/thermal/imx8mm_thermal.c +index 7442e013738f8..af666bd9e8d4d 100644 +--- a/drivers/thermal/imx8mm_thermal.c ++++ b/drivers/thermal/imx8mm_thermal.c +@@ -21,6 +21,7 @@ + #define TPS 0x4 + #define TRITSR 0x20 /* TMU immediate temp */ + ++#define TER_ADC_PD BIT(30) + #define TER_EN BIT(31) + #define TRITSR_TEMP0_VAL_MASK 0xff + #define TRITSR_TEMP1_VAL_MASK 0xff0000 +@@ -113,6 +114,8 @@ static void imx8mm_tmu_enable(struct imx8mm_tmu *tmu, bool enable) + + val = readl_relaxed(tmu->base + TER); + val = enable ? (val | TER_EN) : (val & ~TER_EN); ++ if (tmu->socdata->version == TMU_VER2) ++ val = enable ? (val & ~TER_ADC_PD) : (val | TER_ADC_PD); + writel_relaxed(val, tmu->base + TER); + } + +diff --git a/drivers/thermal/imx_thermal.c b/drivers/thermal/imx_thermal.c +index 2c7473d86a59b..16663373b6829 100644 +--- a/drivers/thermal/imx_thermal.c ++++ b/drivers/thermal/imx_thermal.c +@@ -15,6 +15,7 @@ + #include <linux/regmap.h> + #include <linux/thermal.h> + #include <linux/nvmem-consumer.h> ++#include <linux/pm_runtime.h> + + #define REG_SET 0x4 + #define REG_CLR 0x8 +@@ -194,6 +195,7 @@ static struct thermal_soc_data thermal_imx7d_data = { + }; + + struct imx_thermal_data { ++ struct device *dev; + struct cpufreq_policy *policy; + struct thermal_zone_device *tz; + struct thermal_cooling_device *cdev; +@@ -252,44 +254,15 @@ static int imx_get_temp(struct thermal_zone_device *tz, int *temp) + const struct thermal_soc_data *soc_data = data->socdata; + struct regmap *map = data->tempmon; + unsigned int n_meas; +- bool wait, run_measurement; + u32 val; ++ int ret; + +- run_measurement = !data->irq_enabled; +- if (!run_measurement) { +- /* Check if a measurement is currently in progress */ +- regmap_read(map, soc_data->temp_data, &val); +- wait = !(val & soc_data->temp_valid_mask); +- } else { +- /* +- * Every time we measure the temperature, we will power on the +- * temperature sensor, enable measurements, take a reading, +- * disable measurements, power off the temperature sensor. +- */ +- regmap_write(map, soc_data->sensor_ctrl + REG_CLR, +- soc_data->power_down_mask); +- regmap_write(map, soc_data->sensor_ctrl + REG_SET, +- soc_data->measure_temp_mask); +- +- wait = true; +- } +- +- /* +- * According to the temp sensor designers, it may require up to ~17us +- * to complete a measurement. +- */ +- if (wait) +- usleep_range(20, 50); ++ ret = pm_runtime_resume_and_get(data->dev); ++ if (ret < 0) ++ return ret; + + regmap_read(map, soc_data->temp_data, &val); + +- if (run_measurement) { +- regmap_write(map, soc_data->sensor_ctrl + REG_CLR, +- soc_data->measure_temp_mask); +- regmap_write(map, soc_data->sensor_ctrl + REG_SET, +- soc_data->power_down_mask); +- } +- + if ((val & soc_data->temp_valid_mask) == 0) { + dev_dbg(&tz->device, "temp measurement never finished\n"); + return -EAGAIN; +@@ -328,6 +301,8 @@ static int imx_get_temp(struct thermal_zone_device *tz, int *temp) + enable_irq(data->irq); + } + ++ pm_runtime_put(data->dev); ++ + return 0; + } + +@@ -335,24 +310,16 @@ static int imx_change_mode(struct thermal_zone_device *tz, + enum thermal_device_mode mode) + { + struct imx_thermal_data *data = tz->devdata; +- struct regmap *map = data->tempmon; +- const struct thermal_soc_data *soc_data = data->socdata; + + if (mode == THERMAL_DEVICE_ENABLED) { +- regmap_write(map, soc_data->sensor_ctrl + REG_CLR, +- soc_data->power_down_mask); +- regmap_write(map, soc_data->sensor_ctrl + REG_SET, +- soc_data->measure_temp_mask); ++ pm_runtime_get(data->dev); + + if (!data->irq_enabled) { + data->irq_enabled = true; + enable_irq(data->irq); + } + } else { +- regmap_write(map, soc_data->sensor_ctrl + REG_CLR, +- soc_data->measure_temp_mask); +- regmap_write(map, soc_data->sensor_ctrl + REG_SET, +- soc_data->power_down_mask); ++ pm_runtime_put(data->dev); + + if (data->irq_enabled) { + disable_irq(data->irq); +@@ -393,6 +360,11 @@ static int imx_set_trip_temp(struct thermal_zone_device *tz, int trip, + int temp) + { + struct imx_thermal_data *data = tz->devdata; ++ int ret; ++ ++ ret = pm_runtime_resume_and_get(data->dev); ++ if (ret < 0) ++ return ret; + + /* do not allow changing critical threshold */ + if (trip == IMX_TRIP_CRITICAL) +@@ -406,6 +378,8 @@ static int imx_set_trip_temp(struct thermal_zone_device *tz, int trip, + + imx_set_alarm_temp(data, temp); + ++ pm_runtime_put(data->dev); ++ + return 0; + } + +@@ -681,6 +655,8 @@ static int imx_thermal_probe(struct platform_device *pdev) + if (!data) + return -ENOMEM; + ++ data->dev = &pdev->dev; ++ + map = syscon_regmap_lookup_by_phandle(pdev->dev.of_node, "fsl,tempmon"); + if (IS_ERR(map)) { + ret = PTR_ERR(map); +@@ -800,6 +776,16 @@ static int imx_thermal_probe(struct platform_device *pdev) + data->socdata->power_down_mask); + regmap_write(map, data->socdata->sensor_ctrl + REG_SET, + data->socdata->measure_temp_mask); ++ /* After power up, we need a delay before first access can be done. */ ++ usleep_range(20, 50); ++ ++ /* the core was configured and enabled just before */ ++ pm_runtime_set_active(&pdev->dev); ++ pm_runtime_enable(data->dev); ++ ++ ret = pm_runtime_resume_and_get(data->dev); ++ if (ret < 0) ++ goto disable_runtime_pm; + + data->irq_enabled = true; + ret = thermal_zone_device_enable(data->tz); +@@ -814,10 +800,15 @@ static int imx_thermal_probe(struct platform_device *pdev) + goto thermal_zone_unregister; + } + ++ pm_runtime_put(data->dev); ++ + return 0; + + thermal_zone_unregister: + thermal_zone_device_unregister(data->tz); ++disable_runtime_pm: ++ pm_runtime_put_noidle(data->dev); ++ pm_runtime_disable(data->dev); + clk_disable: + clk_disable_unprepare(data->thermal_clk); + legacy_cleanup: +@@ -829,13 +820,9 @@ legacy_cleanup: + static int imx_thermal_remove(struct platform_device *pdev) + { + struct imx_thermal_data *data = platform_get_drvdata(pdev); +- struct regmap *map = data->tempmon; + +- /* Disable measurements */ +- regmap_write(map, data->socdata->sensor_ctrl + REG_SET, +- data->socdata->power_down_mask); +- if (!IS_ERR(data->thermal_clk)) +- clk_disable_unprepare(data->thermal_clk); ++ pm_runtime_put_noidle(data->dev); ++ pm_runtime_disable(data->dev); + + thermal_zone_device_unregister(data->tz); + imx_thermal_unregister_legacy_cooling(data); +@@ -858,29 +845,79 @@ static int __maybe_unused imx_thermal_suspend(struct device *dev) + ret = thermal_zone_device_disable(data->tz); + if (ret) + return ret; ++ ++ return pm_runtime_force_suspend(data->dev); ++} ++ ++static int __maybe_unused imx_thermal_resume(struct device *dev) ++{ ++ struct imx_thermal_data *data = dev_get_drvdata(dev); ++ int ret; ++ ++ ret = pm_runtime_force_resume(data->dev); ++ if (ret) ++ return ret; ++ /* Enabled thermal sensor after resume */ ++ return thermal_zone_device_enable(data->tz); ++} ++ ++static int __maybe_unused imx_thermal_runtime_suspend(struct device *dev) ++{ ++ struct imx_thermal_data *data = dev_get_drvdata(dev); ++ const struct thermal_soc_data *socdata = data->socdata; ++ struct regmap *map = data->tempmon; ++ int ret; ++ ++ ret = regmap_write(map, socdata->sensor_ctrl + REG_CLR, ++ socdata->measure_temp_mask); ++ if (ret) ++ return ret; ++ ++ ret = regmap_write(map, socdata->sensor_ctrl + REG_SET, ++ socdata->power_down_mask); ++ if (ret) ++ return ret; ++ + clk_disable_unprepare(data->thermal_clk); + + return 0; + } + +-static int __maybe_unused imx_thermal_resume(struct device *dev) ++static int __maybe_unused imx_thermal_runtime_resume(struct device *dev) + { + struct imx_thermal_data *data = dev_get_drvdata(dev); ++ const struct thermal_soc_data *socdata = data->socdata; ++ struct regmap *map = data->tempmon; + int ret; + + ret = clk_prepare_enable(data->thermal_clk); + if (ret) + return ret; +- /* Enabled thermal sensor after resume */ +- ret = thermal_zone_device_enable(data->tz); ++ ++ ret = regmap_write(map, socdata->sensor_ctrl + REG_CLR, ++ socdata->power_down_mask); ++ if (ret) ++ return ret; ++ ++ ret = regmap_write(map, socdata->sensor_ctrl + REG_SET, ++ socdata->measure_temp_mask); + if (ret) + return ret; + ++ /* ++ * According to the temp sensor designers, it may require up to ~17us ++ * to complete a measurement. ++ */ ++ usleep_range(20, 50); ++ + return 0; + } + +-static SIMPLE_DEV_PM_OPS(imx_thermal_pm_ops, +- imx_thermal_suspend, imx_thermal_resume); ++static const struct dev_pm_ops imx_thermal_pm_ops = { ++ SET_SYSTEM_SLEEP_PM_OPS(imx_thermal_suspend, imx_thermal_resume) ++ SET_RUNTIME_PM_OPS(imx_thermal_runtime_suspend, ++ imx_thermal_runtime_resume, NULL) ++}; + + static struct platform_driver imx_thermal = { + .driver = { +diff --git a/drivers/thunderbolt/acpi.c b/drivers/thunderbolt/acpi.c +index b67e72d5644b3..7c9597a339295 100644 +--- a/drivers/thunderbolt/acpi.c ++++ b/drivers/thunderbolt/acpi.c +@@ -7,6 +7,7 @@ + */ + + #include <linux/acpi.h> ++#include <linux/pm_runtime.h> + + #include "tb.h" + +@@ -74,8 +75,18 @@ static acpi_status tb_acpi_add_link(acpi_handle handle, u32 level, void *data, + pci_pcie_type(pdev) == PCI_EXP_TYPE_DOWNSTREAM))) { + const struct device_link *link; + ++ /* ++ * Make them both active first to make sure the NHI does ++ * not runtime suspend before the consumer. The ++ * pm_runtime_put() below then allows the consumer to ++ * runtime suspend again (which then allows NHI runtime ++ * suspend too now that the device link is established). ++ */ ++ pm_runtime_get_sync(&pdev->dev); ++ + link = device_link_add(&pdev->dev, &nhi->pdev->dev, + DL_FLAG_AUTOREMOVE_SUPPLIER | ++ DL_FLAG_RPM_ACTIVE | + DL_FLAG_PM_RUNTIME); + if (link) { + dev_dbg(&nhi->pdev->dev, "created link from %s\n", +@@ -84,6 +95,8 @@ static acpi_status tb_acpi_add_link(acpi_handle handle, u32 level, void *data, + dev_warn(&nhi->pdev->dev, "device link creation from %s failed\n", + dev_name(&pdev->dev)); + } ++ ++ pm_runtime_put(&pdev->dev); + } + + out_put: +diff --git a/drivers/tty/mxser.c b/drivers/tty/mxser.c +index 1216f3985e18e..da375851af4e6 100644 +--- a/drivers/tty/mxser.c ++++ b/drivers/tty/mxser.c +@@ -261,7 +261,6 @@ struct mxser_port { + unsigned int xmit_head; + unsigned int xmit_tail; + unsigned int xmit_cnt; +- int closing; + + spinlock_t slock; + }; +@@ -923,7 +922,6 @@ static void mxser_close(struct tty_struct *tty, struct file *filp) + return; + if (tty_port_close_start(port, tty, filp) == 0) + return; +- info->closing = 1; + mutex_lock(&port->mutex); + mxser_close_port(port); + mxser_flush_buffer(tty); +@@ -932,7 +930,6 @@ static void mxser_close(struct tty_struct *tty, struct file *filp) + mxser_shutdown_port(port); + tty_port_set_initialized(port, 0); + mutex_unlock(&port->mutex); +- info->closing = 0; + /* Right now the tty_port set is done outside of the close_end helper + as we don't yet have everyone using refcounts */ + tty_port_close_end(port, tty); +@@ -1693,7 +1690,7 @@ static bool mxser_port_isr(struct mxser_port *port) + + iir &= MOXA_MUST_IIR_MASK; + tty = tty_port_tty_get(&port->port); +- if (!tty || port->closing || !tty_port_initialized(&port->port)) { ++ if (!tty) { + status = inb(port->ioaddr + UART_LSR); + outb(MOXA_MUST_FCR_GDA_MODE_ENABLE | UART_FCR_ENABLE_FIFO | + UART_FCR_CLEAR_RCVR | UART_FCR_CLEAR_XMIT, +diff --git a/drivers/tty/serial/8250/8250_bcm7271.c b/drivers/tty/serial/8250/8250_bcm7271.c +index 5163d60756b73..0877cf24f7de0 100644 +--- a/drivers/tty/serial/8250/8250_bcm7271.c ++++ b/drivers/tty/serial/8250/8250_bcm7271.c +@@ -1076,14 +1076,18 @@ static int brcmuart_probe(struct platform_device *pdev) + priv->rx_bufs = dma_alloc_coherent(dev, + priv->rx_size, + &priv->rx_addr, GFP_KERNEL); +- if (!priv->rx_bufs) ++ if (!priv->rx_bufs) { ++ ret = -EINVAL; + goto err; ++ } + priv->tx_size = UART_XMIT_SIZE; + priv->tx_buf = dma_alloc_coherent(dev, + priv->tx_size, + &priv->tx_addr, GFP_KERNEL); +- if (!priv->tx_buf) ++ if (!priv->tx_buf) { ++ ret = -EINVAL; + goto err; ++ } + } + + ret = serial8250_register_8250_port(&up); +@@ -1097,6 +1101,7 @@ static int brcmuart_probe(struct platform_device *pdev) + if (priv->dma_enabled) { + dma_irq = platform_get_irq_byname(pdev, "dma"); + if (dma_irq < 0) { ++ ret = dma_irq; + dev_err(dev, "no IRQ resource info\n"); + goto err1; + } +@@ -1116,7 +1121,7 @@ err1: + err: + brcmuart_free_bufs(dev, priv); + brcmuart_arbitration(priv, 0); +- return -ENODEV; ++ return ret; + } + + static int brcmuart_remove(struct platform_device *pdev) +diff --git a/drivers/tty/serial/amba-pl010.c b/drivers/tty/serial/amba-pl010.c +index e744b953ca346..47654073123d6 100644 +--- a/drivers/tty/serial/amba-pl010.c ++++ b/drivers/tty/serial/amba-pl010.c +@@ -446,14 +446,11 @@ pl010_set_termios(struct uart_port *port, struct ktermios *termios, + if ((termios->c_cflag & CREAD) == 0) + uap->port.ignore_status_mask |= UART_DUMMY_RSR_RX; + +- /* first, disable everything */ + old_cr = readb(uap->port.membase + UART010_CR) & ~UART010_CR_MSIE; + + if (UART_ENABLE_MS(port, termios->c_cflag)) + old_cr |= UART010_CR_MSIE; + +- writel(0, uap->port.membase + UART010_CR); +- + /* Set baud rate */ + quot -= 1; + writel((quot & 0xf00) >> 8, uap->port.membase + UART010_LCRM); +diff --git a/drivers/tty/serial/amba-pl011.c b/drivers/tty/serial/amba-pl011.c +index 52518a606c06a..6ec34260d6b18 100644 +--- a/drivers/tty/serial/amba-pl011.c ++++ b/drivers/tty/serial/amba-pl011.c +@@ -2105,9 +2105,7 @@ pl011_set_termios(struct uart_port *port, struct ktermios *termios, + if (port->rs485.flags & SER_RS485_ENABLED) + termios->c_cflag &= ~CRTSCTS; + +- /* first, disable everything */ + old_cr = pl011_read(uap, REG_CR); +- pl011_write(0, uap, REG_CR); + + if (termios->c_cflag & CRTSCTS) { + if (old_cr & UART011_CR_RTS) +@@ -2183,32 +2181,13 @@ static const char *pl011_type(struct uart_port *port) + return uap->port.type == PORT_AMBA ? uap->type : NULL; + } + +-/* +- * Release the memory region(s) being used by 'port' +- */ +-static void pl011_release_port(struct uart_port *port) +-{ +- release_mem_region(port->mapbase, SZ_4K); +-} +- +-/* +- * Request the memory region(s) being used by 'port' +- */ +-static int pl011_request_port(struct uart_port *port) +-{ +- return request_mem_region(port->mapbase, SZ_4K, "uart-pl011") +- != NULL ? 0 : -EBUSY; +-} +- + /* + * Configure/autoconfigure the port. + */ + static void pl011_config_port(struct uart_port *port, int flags) + { +- if (flags & UART_CONFIG_TYPE) { ++ if (flags & UART_CONFIG_TYPE) + port->type = PORT_AMBA; +- pl011_request_port(port); +- } + } + + /* +@@ -2223,6 +2202,8 @@ static int pl011_verify_port(struct uart_port *port, struct serial_struct *ser) + ret = -EINVAL; + if (ser->baud_base < 9600) + ret = -EINVAL; ++ if (port->mapbase != (unsigned long) ser->iomem_base) ++ ret = -EINVAL; + return ret; + } + +@@ -2275,8 +2256,6 @@ static const struct uart_ops amba_pl011_pops = { + .flush_buffer = pl011_dma_flush_buffer, + .set_termios = pl011_set_termios, + .type = pl011_type, +- .release_port = pl011_release_port, +- .request_port = pl011_request_port, + .config_port = pl011_config_port, + .verify_port = pl011_verify_port, + #ifdef CONFIG_CONSOLE_POLL +@@ -2306,8 +2285,6 @@ static const struct uart_ops sbsa_uart_pops = { + .shutdown = sbsa_uart_shutdown, + .set_termios = sbsa_uart_set_termios, + .type = pl011_type, +- .release_port = pl011_release_port, +- .request_port = pl011_request_port, + .config_port = pl011_config_port, + .verify_port = pl011_verify_port, + #ifdef CONFIG_CONSOLE_POLL +diff --git a/drivers/tty/serial/atmel_serial.c b/drivers/tty/serial/atmel_serial.c +index 249ea35088d27..dd350c5908804 100644 +--- a/drivers/tty/serial/atmel_serial.c ++++ b/drivers/tty/serial/atmel_serial.c +@@ -1004,6 +1004,13 @@ static void atmel_tx_dma(struct uart_port *port) + desc->callback = atmel_complete_tx_dma; + desc->callback_param = atmel_port; + atmel_port->cookie_tx = dmaengine_submit(desc); ++ if (dma_submit_error(atmel_port->cookie_tx)) { ++ dev_err(port->dev, "dma_submit_error %d\n", ++ atmel_port->cookie_tx); ++ return; ++ } ++ ++ dma_async_issue_pending(chan); + } + + if (uart_circ_chars_pending(xmit) < WAKEUP_CHARS) +@@ -1258,6 +1265,13 @@ static int atmel_prepare_rx_dma(struct uart_port *port) + desc->callback_param = port; + atmel_port->desc_rx = desc; + atmel_port->cookie_rx = dmaengine_submit(desc); ++ if (dma_submit_error(atmel_port->cookie_rx)) { ++ dev_err(port->dev, "dma_submit_error %d\n", ++ atmel_port->cookie_rx); ++ goto chan_err; ++ } ++ ++ dma_async_issue_pending(atmel_port->chan_rx); + + return 0; + +diff --git a/drivers/tty/serial/imx.c b/drivers/tty/serial/imx.c +index 51a9f9423b1a6..7820049aba5af 100644 +--- a/drivers/tty/serial/imx.c ++++ b/drivers/tty/serial/imx.c +@@ -486,18 +486,21 @@ static void imx_uart_stop_tx(struct uart_port *port) + static void imx_uart_stop_rx(struct uart_port *port) + { + struct imx_port *sport = (struct imx_port *)port; +- u32 ucr1, ucr2; ++ u32 ucr1, ucr2, ucr4; + + ucr1 = imx_uart_readl(sport, UCR1); + ucr2 = imx_uart_readl(sport, UCR2); ++ ucr4 = imx_uart_readl(sport, UCR4); + + if (sport->dma_is_enabled) { + ucr1 &= ~(UCR1_RXDMAEN | UCR1_ATDMAEN); + } else { + ucr1 &= ~UCR1_RRDYEN; + ucr2 &= ~UCR2_ATEN; ++ ucr4 &= ~UCR4_OREN; + } + imx_uart_writel(sport, ucr1, UCR1); ++ imx_uart_writel(sport, ucr4, UCR4); + + ucr2 &= ~UCR2_RXEN; + imx_uart_writel(sport, ucr2, UCR2); +@@ -1544,7 +1547,7 @@ static void imx_uart_shutdown(struct uart_port *port) + imx_uart_writel(sport, ucr1, UCR1); + + ucr4 = imx_uart_readl(sport, UCR4); +- ucr4 &= ~(UCR4_OREN | UCR4_TCEN); ++ ucr4 &= ~UCR4_TCEN; + imx_uart_writel(sport, ucr4, UCR4); + + spin_unlock_irqrestore(&sport->port.lock, flags); +diff --git a/drivers/tty/serial/liteuart.c b/drivers/tty/serial/liteuart.c +index 2941659e52747..7f74bf7bdcff8 100644 +--- a/drivers/tty/serial/liteuart.c ++++ b/drivers/tty/serial/liteuart.c +@@ -436,4 +436,4 @@ module_exit(liteuart_exit); + MODULE_AUTHOR("Antmicro <www.antmicro.com>"); + MODULE_DESCRIPTION("LiteUART serial driver"); + MODULE_LICENSE("GPL v2"); +-MODULE_ALIAS("platform: liteuart"); ++MODULE_ALIAS("platform:liteuart"); +diff --git a/drivers/tty/serial/serial_core.c b/drivers/tty/serial/serial_core.c +index 61e3dd0222af1..dc6129ddef85d 100644 +--- a/drivers/tty/serial/serial_core.c ++++ b/drivers/tty/serial/serial_core.c +@@ -162,7 +162,7 @@ static void uart_port_dtr_rts(struct uart_port *uport, int raise) + int RTS_after_send = !!(uport->rs485.flags & SER_RS485_RTS_AFTER_SEND); + + if (raise) { +- if (rs485_on && !RTS_after_send) { ++ if (rs485_on && RTS_after_send) { + uart_set_mctrl(uport, TIOCM_DTR); + uart_clear_mctrl(uport, TIOCM_RTS); + } else { +@@ -171,7 +171,7 @@ static void uart_port_dtr_rts(struct uart_port *uport, int raise) + } else { + unsigned int clear = TIOCM_DTR; + +- clear |= (!rs485_on || !RTS_after_send) ? TIOCM_RTS : 0; ++ clear |= (!rs485_on || RTS_after_send) ? TIOCM_RTS : 0; + uart_clear_mctrl(uport, clear); + } + } +@@ -2393,7 +2393,8 @@ uart_configure_port(struct uart_driver *drv, struct uart_state *state, + * We probably don't need a spinlock around this, but + */ + spin_lock_irqsave(&port->lock, flags); +- port->ops->set_mctrl(port, port->mctrl & TIOCM_DTR); ++ port->mctrl &= TIOCM_DTR; ++ port->ops->set_mctrl(port, port->mctrl); + spin_unlock_irqrestore(&port->lock, flags); + + /* +diff --git a/drivers/tty/serial/stm32-usart.c b/drivers/tty/serial/stm32-usart.c +index 8f032e77b954a..3366914dad7a8 100644 +--- a/drivers/tty/serial/stm32-usart.c ++++ b/drivers/tty/serial/stm32-usart.c +@@ -691,6 +691,11 @@ static void stm32_usart_shutdown(struct uart_port *port) + u32 val, isr; + int ret; + ++ if (stm32_port->tx_dma_busy) { ++ dmaengine_terminate_async(stm32_port->tx_ch); ++ stm32_usart_clr_bits(port, ofs->cr3, USART_CR3_DMAT); ++ } ++ + /* Disable modem control interrupts */ + stm32_usart_disable_ms(port); + +@@ -1385,7 +1390,6 @@ static int stm32_usart_serial_remove(struct platform_device *pdev) + stm32_usart_clr_bits(port, ofs->cr3, USART_CR3_DMAR); + + if (stm32_port->tx_ch) { +- dmaengine_terminate_async(stm32_port->tx_ch); + stm32_usart_of_dma_tx_remove(stm32_port, pdev); + dma_release_channel(stm32_port->tx_ch); + } +diff --git a/drivers/tty/serial/uartlite.c b/drivers/tty/serial/uartlite.c +index dfc1ba4e15724..36871cebd6a0f 100644 +--- a/drivers/tty/serial/uartlite.c ++++ b/drivers/tty/serial/uartlite.c +@@ -612,7 +612,7 @@ static struct uart_driver ulite_uart_driver = { + * + * Returns: 0 on success, <0 otherwise + */ +-static int ulite_assign(struct device *dev, int id, u32 base, int irq, ++static int ulite_assign(struct device *dev, int id, phys_addr_t base, int irq, + struct uartlite_data *pdata) + { + struct uart_port *port; +diff --git a/drivers/usb/core/hub.c b/drivers/usb/core/hub.c +index 3bc4a86c3d0a5..ac6c5ccfe1cb7 100644 +--- a/drivers/usb/core/hub.c ++++ b/drivers/usb/core/hub.c +@@ -1110,7 +1110,10 @@ static void hub_activate(struct usb_hub *hub, enum hub_activation_type type) + } else { + hub_power_on(hub, true); + } +- } ++ /* Give some time on remote wakeup to let links to transit to U0 */ ++ } else if (hub_is_superspeed(hub->hdev)) ++ msleep(20); ++ + init2: + + /* +diff --git a/drivers/usb/dwc2/gadget.c b/drivers/usb/dwc2/gadget.c +index 2190225bf3da2..ea0d2d6139a68 100644 +--- a/drivers/usb/dwc2/gadget.c ++++ b/drivers/usb/dwc2/gadget.c +@@ -4974,7 +4974,18 @@ int dwc2_gadget_init(struct dwc2_hsotg *hsotg) + hsotg->params.g_np_tx_fifo_size); + dev_dbg(dev, "RXFIFO size: %d\n", hsotg->params.g_rx_fifo_size); + +- hsotg->gadget.max_speed = USB_SPEED_HIGH; ++ switch (hsotg->params.speed) { ++ case DWC2_SPEED_PARAM_LOW: ++ hsotg->gadget.max_speed = USB_SPEED_LOW; ++ break; ++ case DWC2_SPEED_PARAM_FULL: ++ hsotg->gadget.max_speed = USB_SPEED_FULL; ++ break; ++ default: ++ hsotg->gadget.max_speed = USB_SPEED_HIGH; ++ break; ++ } ++ + hsotg->gadget.ops = &dwc2_hsotg_gadget_ops; + hsotg->gadget.name = dev_name(dev); + hsotg->remote_wakeup_allowed = 0; +diff --git a/drivers/usb/dwc2/hcd.c b/drivers/usb/dwc2/hcd.c +index a215ec9e172e6..657dbd50faf11 100644 +--- a/drivers/usb/dwc2/hcd.c ++++ b/drivers/usb/dwc2/hcd.c +@@ -4403,11 +4403,12 @@ static int _dwc2_hcd_suspend(struct usb_hcd *hcd) + * If not hibernation nor partial power down are supported, + * clock gating is used to save power. + */ +- if (!hsotg->params.no_clock_gating) ++ if (!hsotg->params.no_clock_gating) { + dwc2_host_enter_clock_gating(hsotg); + +- /* After entering suspend, hardware is not accessible */ +- clear_bit(HCD_FLAG_HW_ACCESSIBLE, &hcd->flags); ++ /* After entering suspend, hardware is not accessible */ ++ clear_bit(HCD_FLAG_HW_ACCESSIBLE, &hcd->flags); ++ } + break; + default: + goto skip_power_saving; +diff --git a/drivers/usb/dwc3/dwc3-meson-g12a.c b/drivers/usb/dwc3/dwc3-meson-g12a.c +index d0f9b7c296b0d..bd814df3bf8b8 100644 +--- a/drivers/usb/dwc3/dwc3-meson-g12a.c ++++ b/drivers/usb/dwc3/dwc3-meson-g12a.c +@@ -755,16 +755,16 @@ static int dwc3_meson_g12a_probe(struct platform_device *pdev) + + ret = dwc3_meson_g12a_get_phys(priv); + if (ret) +- goto err_disable_clks; ++ goto err_rearm; + + ret = priv->drvdata->setup_regmaps(priv, base); + if (ret) +- goto err_disable_clks; ++ goto err_rearm; + + if (priv->vbus) { + ret = regulator_enable(priv->vbus); + if (ret) +- goto err_disable_clks; ++ goto err_rearm; + } + + /* Get dr_mode */ +@@ -825,6 +825,9 @@ err_disable_regulator: + if (priv->vbus) + regulator_disable(priv->vbus); + ++err_rearm: ++ reset_control_rearm(priv->reset); ++ + err_disable_clks: + clk_bulk_disable_unprepare(priv->drvdata->num_clks, + priv->drvdata->clks); +@@ -852,6 +855,8 @@ static int dwc3_meson_g12a_remove(struct platform_device *pdev) + pm_runtime_put_noidle(dev); + pm_runtime_set_suspended(dev); + ++ reset_control_rearm(priv->reset); ++ + clk_bulk_disable_unprepare(priv->drvdata->num_clks, + priv->drvdata->clks); + +@@ -892,7 +897,7 @@ static int __maybe_unused dwc3_meson_g12a_suspend(struct device *dev) + phy_exit(priv->phys[i]); + } + +- reset_control_assert(priv->reset); ++ reset_control_rearm(priv->reset); + + return 0; + } +@@ -902,7 +907,9 @@ static int __maybe_unused dwc3_meson_g12a_resume(struct device *dev) + struct dwc3_meson_g12a *priv = dev_get_drvdata(dev); + int i, ret; + +- reset_control_deassert(priv->reset); ++ ret = reset_control_reset(priv->reset); ++ if (ret) ++ return ret; + + ret = priv->drvdata->usb_init(priv); + if (ret) +diff --git a/drivers/usb/dwc3/dwc3-qcom.c b/drivers/usb/dwc3/dwc3-qcom.c +index 3cb01cdd02c29..b81a9e1c13153 100644 +--- a/drivers/usb/dwc3/dwc3-qcom.c ++++ b/drivers/usb/dwc3/dwc3-qcom.c +@@ -769,9 +769,12 @@ static int dwc3_qcom_probe(struct platform_device *pdev) + + if (qcom->acpi_pdata->is_urs) { + qcom->urs_usb = dwc3_qcom_create_urs_usb_platdev(dev); +- if (!qcom->urs_usb) { ++ if (IS_ERR_OR_NULL(qcom->urs_usb)) { + dev_err(dev, "failed to create URS USB platdev\n"); +- return -ENODEV; ++ if (!qcom->urs_usb) ++ return -ENODEV; ++ else ++ return PTR_ERR(qcom->urs_usb); + } + } + } +diff --git a/drivers/usb/gadget/function/f_fs.c b/drivers/usb/gadget/function/f_fs.c +index aac343f7d7d3d..782d67c2c6e0d 100644 +--- a/drivers/usb/gadget/function/f_fs.c ++++ b/drivers/usb/gadget/function/f_fs.c +@@ -614,7 +614,7 @@ static int ffs_ep0_open(struct inode *inode, struct file *file) + file->private_data = ffs; + ffs_data_opened(ffs); + +- return 0; ++ return stream_open(inode, file); + } + + static int ffs_ep0_release(struct inode *inode, struct file *file) +@@ -1154,7 +1154,7 @@ ffs_epfile_open(struct inode *inode, struct file *file) + file->private_data = epfile; + ffs_data_opened(epfile->ffs); + +- return 0; ++ return stream_open(inode, file); + } + + static int ffs_aio_cancel(struct kiocb *kiocb) +diff --git a/drivers/usb/gadget/function/u_audio.c b/drivers/usb/gadget/function/u_audio.c +index ad16163b5ff80..d22ac23c94b0f 100644 +--- a/drivers/usb/gadget/function/u_audio.c ++++ b/drivers/usb/gadget/function/u_audio.c +@@ -1097,7 +1097,7 @@ int g_audio_setup(struct g_audio *g_audio, const char *pcm_name, + } + + kctl->id.device = pcm->device; +- kctl->id.subdevice = i; ++ kctl->id.subdevice = 0; + + err = snd_ctl_add(card, kctl); + if (err < 0) +@@ -1120,7 +1120,7 @@ int g_audio_setup(struct g_audio *g_audio, const char *pcm_name, + } + + kctl->id.device = pcm->device; +- kctl->id.subdevice = i; ++ kctl->id.subdevice = 0; + + + kctl->tlv.c = u_audio_volume_tlv; +diff --git a/drivers/usb/host/ehci-brcm.c b/drivers/usb/host/ehci-brcm.c +index d3626bfa966b4..6a0f64c9e5e88 100644 +--- a/drivers/usb/host/ehci-brcm.c ++++ b/drivers/usb/host/ehci-brcm.c +@@ -62,8 +62,12 @@ static int ehci_brcm_hub_control( + u32 __iomem *status_reg; + unsigned long flags; + int retval, irq_disabled = 0; ++ u32 temp; + +- status_reg = &ehci->regs->port_status[(wIndex & 0xff) - 1]; ++ temp = (wIndex & 0xff) - 1; ++ if (temp >= HCS_N_PORTS_MAX) /* Avoid index-out-of-bounds warning */ ++ temp = 0; ++ status_reg = &ehci->regs->port_status[temp]; + + /* + * RESUME is cleared when GetPortStatus() is called 20ms after start +diff --git a/drivers/usb/host/uhci-platform.c b/drivers/usb/host/uhci-platform.c +index 70dbd95c3f063..be9e9db7cad10 100644 +--- a/drivers/usb/host/uhci-platform.c ++++ b/drivers/usb/host/uhci-platform.c +@@ -113,7 +113,8 @@ static int uhci_hcd_platform_probe(struct platform_device *pdev) + num_ports); + } + if (of_device_is_compatible(np, "aspeed,ast2400-uhci") || +- of_device_is_compatible(np, "aspeed,ast2500-uhci")) { ++ of_device_is_compatible(np, "aspeed,ast2500-uhci") || ++ of_device_is_compatible(np, "aspeed,ast2600-uhci")) { + uhci->is_aspeed = 1; + dev_info(&pdev->dev, + "Enabled Aspeed implementation workarounds\n"); +diff --git a/drivers/usb/misc/ftdi-elan.c b/drivers/usb/misc/ftdi-elan.c +index e5a8fcdbb78e7..6c38c62d29b26 100644 +--- a/drivers/usb/misc/ftdi-elan.c ++++ b/drivers/usb/misc/ftdi-elan.c +@@ -202,6 +202,7 @@ static void ftdi_elan_delete(struct kref *kref) + mutex_unlock(&ftdi_module_lock); + kfree(ftdi->bulk_in_buffer); + ftdi->bulk_in_buffer = NULL; ++ kfree(ftdi); + } + + static void ftdi_elan_put_kref(struct usb_ftdi *ftdi) +diff --git a/drivers/vdpa/ifcvf/ifcvf_base.c b/drivers/vdpa/ifcvf/ifcvf_base.c +index 2808f1ba9f7b8..7d41dfe48adee 100644 +--- a/drivers/vdpa/ifcvf/ifcvf_base.c ++++ b/drivers/vdpa/ifcvf/ifcvf_base.c +@@ -143,8 +143,8 @@ int ifcvf_init_hw(struct ifcvf_hw *hw, struct pci_dev *pdev) + IFCVF_DBG(pdev, "hw->isr = %p\n", hw->isr); + break; + case VIRTIO_PCI_CAP_DEVICE_CFG: +- hw->net_cfg = get_cap_addr(hw, &cap); +- IFCVF_DBG(pdev, "hw->net_cfg = %p\n", hw->net_cfg); ++ hw->dev_cfg = get_cap_addr(hw, &cap); ++ IFCVF_DBG(pdev, "hw->dev_cfg = %p\n", hw->dev_cfg); + break; + } + +@@ -153,7 +153,7 @@ next: + } + + if (hw->common_cfg == NULL || hw->notify_base == NULL || +- hw->isr == NULL || hw->net_cfg == NULL) { ++ hw->isr == NULL || hw->dev_cfg == NULL) { + IFCVF_ERR(pdev, "Incomplete PCI capabilities\n"); + return -EIO; + } +@@ -174,7 +174,7 @@ next: + IFCVF_DBG(pdev, + "PCI capability mapping: common cfg: %p, notify base: %p\n, isr cfg: %p, device cfg: %p, multiplier: %u\n", + hw->common_cfg, hw->notify_base, hw->isr, +- hw->net_cfg, hw->notify_off_multiplier); ++ hw->dev_cfg, hw->notify_off_multiplier); + + return 0; + } +@@ -242,33 +242,54 @@ int ifcvf_verify_min_features(struct ifcvf_hw *hw, u64 features) + return 0; + } + +-void ifcvf_read_net_config(struct ifcvf_hw *hw, u64 offset, ++u32 ifcvf_get_config_size(struct ifcvf_hw *hw) ++{ ++ struct ifcvf_adapter *adapter; ++ u32 config_size; ++ ++ adapter = vf_to_adapter(hw); ++ switch (hw->dev_type) { ++ case VIRTIO_ID_NET: ++ config_size = sizeof(struct virtio_net_config); ++ break; ++ case VIRTIO_ID_BLOCK: ++ config_size = sizeof(struct virtio_blk_config); ++ break; ++ default: ++ config_size = 0; ++ IFCVF_ERR(adapter->pdev, "VIRTIO ID %u not supported\n", hw->dev_type); ++ } ++ ++ return config_size; ++} ++ ++void ifcvf_read_dev_config(struct ifcvf_hw *hw, u64 offset, + void *dst, int length) + { + u8 old_gen, new_gen, *p; + int i; + +- WARN_ON(offset + length > sizeof(struct virtio_net_config)); ++ WARN_ON(offset + length > hw->config_size); + do { + old_gen = ifc_ioread8(&hw->common_cfg->config_generation); + p = dst; + for (i = 0; i < length; i++) +- *p++ = ifc_ioread8(hw->net_cfg + offset + i); ++ *p++ = ifc_ioread8(hw->dev_cfg + offset + i); + + new_gen = ifc_ioread8(&hw->common_cfg->config_generation); + } while (old_gen != new_gen); + } + +-void ifcvf_write_net_config(struct ifcvf_hw *hw, u64 offset, ++void ifcvf_write_dev_config(struct ifcvf_hw *hw, u64 offset, + const void *src, int length) + { + const u8 *p; + int i; + + p = src; +- WARN_ON(offset + length > sizeof(struct virtio_net_config)); ++ WARN_ON(offset + length > hw->config_size); + for (i = 0; i < length; i++) +- ifc_iowrite8(*p++, hw->net_cfg + offset + i); ++ ifc_iowrite8(*p++, hw->dev_cfg + offset + i); + } + + static void ifcvf_set_features(struct ifcvf_hw *hw, u64 features) +diff --git a/drivers/vdpa/ifcvf/ifcvf_base.h b/drivers/vdpa/ifcvf/ifcvf_base.h +index 09918af3ecf82..c486873f370a8 100644 +--- a/drivers/vdpa/ifcvf/ifcvf_base.h ++++ b/drivers/vdpa/ifcvf/ifcvf_base.h +@@ -71,12 +71,14 @@ struct ifcvf_hw { + u64 hw_features; + u32 dev_type; + struct virtio_pci_common_cfg __iomem *common_cfg; +- void __iomem *net_cfg; ++ void __iomem *dev_cfg; + struct vring_info vring[IFCVF_MAX_QUEUES]; + void __iomem * const *base; + char config_msix_name[256]; + struct vdpa_callback config_cb; + unsigned int config_irq; ++ /* virtio-net or virtio-blk device config size */ ++ u32 config_size; + }; + + struct ifcvf_adapter { +@@ -105,9 +107,9 @@ int ifcvf_init_hw(struct ifcvf_hw *hw, struct pci_dev *dev); + int ifcvf_start_hw(struct ifcvf_hw *hw); + void ifcvf_stop_hw(struct ifcvf_hw *hw); + void ifcvf_notify_queue(struct ifcvf_hw *hw, u16 qid); +-void ifcvf_read_net_config(struct ifcvf_hw *hw, u64 offset, ++void ifcvf_read_dev_config(struct ifcvf_hw *hw, u64 offset, + void *dst, int length); +-void ifcvf_write_net_config(struct ifcvf_hw *hw, u64 offset, ++void ifcvf_write_dev_config(struct ifcvf_hw *hw, u64 offset, + const void *src, int length); + u8 ifcvf_get_status(struct ifcvf_hw *hw); + void ifcvf_set_status(struct ifcvf_hw *hw, u8 status); +@@ -120,4 +122,5 @@ u16 ifcvf_get_vq_state(struct ifcvf_hw *hw, u16 qid); + int ifcvf_set_vq_state(struct ifcvf_hw *hw, u16 qid, u16 num); + struct ifcvf_adapter *vf_to_adapter(struct ifcvf_hw *hw); + int ifcvf_probed_virtio_net(struct ifcvf_hw *hw); ++u32 ifcvf_get_config_size(struct ifcvf_hw *hw); + #endif /* _IFCVF_H_ */ +diff --git a/drivers/vdpa/ifcvf/ifcvf_main.c b/drivers/vdpa/ifcvf/ifcvf_main.c +index dcd648e1f7e7e..003530b19b4ee 100644 +--- a/drivers/vdpa/ifcvf/ifcvf_main.c ++++ b/drivers/vdpa/ifcvf/ifcvf_main.c +@@ -366,24 +366,9 @@ static u32 ifcvf_vdpa_get_vq_align(struct vdpa_device *vdpa_dev) + + static size_t ifcvf_vdpa_get_config_size(struct vdpa_device *vdpa_dev) + { +- struct ifcvf_adapter *adapter = vdpa_to_adapter(vdpa_dev); + struct ifcvf_hw *vf = vdpa_to_vf(vdpa_dev); +- struct pci_dev *pdev = adapter->pdev; +- size_t size; +- +- switch (vf->dev_type) { +- case VIRTIO_ID_NET: +- size = sizeof(struct virtio_net_config); +- break; +- case VIRTIO_ID_BLOCK: +- size = sizeof(struct virtio_blk_config); +- break; +- default: +- size = 0; +- IFCVF_ERR(pdev, "VIRTIO ID %u not supported\n", vf->dev_type); +- } + +- return size; ++ return vf->config_size; + } + + static void ifcvf_vdpa_get_config(struct vdpa_device *vdpa_dev, +@@ -392,8 +377,7 @@ static void ifcvf_vdpa_get_config(struct vdpa_device *vdpa_dev, + { + struct ifcvf_hw *vf = vdpa_to_vf(vdpa_dev); + +- WARN_ON(offset + len > sizeof(struct virtio_net_config)); +- ifcvf_read_net_config(vf, offset, buf, len); ++ ifcvf_read_dev_config(vf, offset, buf, len); + } + + static void ifcvf_vdpa_set_config(struct vdpa_device *vdpa_dev, +@@ -402,8 +386,7 @@ static void ifcvf_vdpa_set_config(struct vdpa_device *vdpa_dev, + { + struct ifcvf_hw *vf = vdpa_to_vf(vdpa_dev); + +- WARN_ON(offset + len > sizeof(struct virtio_net_config)); +- ifcvf_write_net_config(vf, offset, buf, len); ++ ifcvf_write_dev_config(vf, offset, buf, len); + } + + static void ifcvf_vdpa_set_config_cb(struct vdpa_device *vdpa_dev, +@@ -541,6 +524,7 @@ static int ifcvf_vdpa_dev_add(struct vdpa_mgmt_dev *mdev, const char *name) + vf->vring[i].irq = -EINVAL; + + vf->hw_features = ifcvf_get_hw_features(vf); ++ vf->config_size = ifcvf_get_config_size(vf); + + adapter->vdpa.mdev = &ifcvf_mgmt_dev->mdev; + ret = _vdpa_register_device(&adapter->vdpa, vf->nr_vring); +diff --git a/drivers/vdpa/mlx5/net/mlx5_vnet.c b/drivers/vdpa/mlx5/net/mlx5_vnet.c +index ae85d2dd6eb76..1afbda216df52 100644 +--- a/drivers/vdpa/mlx5/net/mlx5_vnet.c ++++ b/drivers/vdpa/mlx5/net/mlx5_vnet.c +@@ -873,8 +873,6 @@ static int create_virtqueue(struct mlx5_vdpa_net *ndev, struct mlx5_vdpa_virtque + MLX5_SET(virtio_q, vq_ctx, umem_3_id, mvq->umem3.id); + MLX5_SET(virtio_q, vq_ctx, umem_3_size, mvq->umem3.size); + MLX5_SET(virtio_q, vq_ctx, pd, ndev->mvdev.res.pdn); +- if (MLX5_CAP_DEV_VDPA_EMULATION(ndev->mvdev.mdev, eth_frame_offload_type)) +- MLX5_SET(virtio_q, vq_ctx, virtio_version_1_0, 1); + + err = mlx5_cmd_exec(ndev->mvdev.mdev, in, inlen, out, sizeof(out)); + if (err) +@@ -1512,9 +1510,11 @@ static int change_num_qps(struct mlx5_vdpa_dev *mvdev, int newqps) + return 0; + + clean_added: +- for (--i; i >= cur_qps; --i) ++ for (--i; i >= 2 * cur_qps; --i) + teardown_vq(ndev, &ndev->vqs[i]); + ++ ndev->cur_num_vqs = 2 * cur_qps; ++ + return err; + } + +diff --git a/drivers/video/backlight/qcom-wled.c b/drivers/video/backlight/qcom-wled.c +index d094299c2a485..f12c76d6e61de 100644 +--- a/drivers/video/backlight/qcom-wled.c ++++ b/drivers/video/backlight/qcom-wled.c +@@ -231,14 +231,14 @@ struct wled { + static int wled3_set_brightness(struct wled *wled, u16 brightness) + { + int rc, i; +- u8 v[2]; ++ __le16 v; + +- v[0] = brightness & 0xff; +- v[1] = (brightness >> 8) & 0xf; ++ v = cpu_to_le16(brightness & WLED3_SINK_REG_BRIGHT_MAX); + + for (i = 0; i < wled->cfg.num_strings; ++i) { + rc = regmap_bulk_write(wled->regmap, wled->ctrl_addr + +- WLED3_SINK_REG_BRIGHT(i), v, 2); ++ WLED3_SINK_REG_BRIGHT(wled->cfg.enabled_strings[i]), ++ &v, sizeof(v)); + if (rc < 0) + return rc; + } +@@ -250,18 +250,18 @@ static int wled4_set_brightness(struct wled *wled, u16 brightness) + { + int rc, i; + u16 low_limit = wled->max_brightness * 4 / 1000; +- u8 v[2]; ++ __le16 v; + + /* WLED4's lower limit of operation is 0.4% */ + if (brightness > 0 && brightness < low_limit) + brightness = low_limit; + +- v[0] = brightness & 0xff; +- v[1] = (brightness >> 8) & 0xf; ++ v = cpu_to_le16(brightness & WLED3_SINK_REG_BRIGHT_MAX); + + for (i = 0; i < wled->cfg.num_strings; ++i) { + rc = regmap_bulk_write(wled->regmap, wled->sink_addr + +- WLED4_SINK_REG_BRIGHT(i), v, 2); ++ WLED4_SINK_REG_BRIGHT(wled->cfg.enabled_strings[i]), ++ &v, sizeof(v)); + if (rc < 0) + return rc; + } +@@ -273,21 +273,20 @@ static int wled5_set_brightness(struct wled *wled, u16 brightness) + { + int rc, offset; + u16 low_limit = wled->max_brightness * 1 / 1000; +- u8 v[2]; ++ __le16 v; + + /* WLED5's lower limit is 0.1% */ + if (brightness < low_limit) + brightness = low_limit; + +- v[0] = brightness & 0xff; +- v[1] = (brightness >> 8) & 0x7f; ++ v = cpu_to_le16(brightness & WLED5_SINK_REG_BRIGHT_MAX_15B); + + offset = (wled->cfg.mod_sel == MOD_A) ? + WLED5_SINK_REG_MOD_A_BRIGHTNESS_LSB : + WLED5_SINK_REG_MOD_B_BRIGHTNESS_LSB; + + rc = regmap_bulk_write(wled->regmap, wled->sink_addr + offset, +- v, 2); ++ &v, sizeof(v)); + return rc; + } + +@@ -572,7 +571,7 @@ unlock_mutex: + + static void wled_auto_string_detection(struct wled *wled) + { +- int rc = 0, i, delay_time_us; ++ int rc = 0, i, j, delay_time_us; + u32 sink_config = 0; + u8 sink_test = 0, sink_valid = 0, val; + bool fault_set; +@@ -619,14 +618,15 @@ static void wled_auto_string_detection(struct wled *wled) + + /* Iterate through the strings one by one */ + for (i = 0; i < wled->cfg.num_strings; i++) { +- sink_test = BIT((WLED4_SINK_REG_CURR_SINK_SHFT + i)); ++ j = wled->cfg.enabled_strings[i]; ++ sink_test = BIT((WLED4_SINK_REG_CURR_SINK_SHFT + j)); + + /* Enable feedback control */ + rc = regmap_write(wled->regmap, wled->ctrl_addr + +- WLED3_CTRL_REG_FEEDBACK_CONTROL, i + 1); ++ WLED3_CTRL_REG_FEEDBACK_CONTROL, j + 1); + if (rc < 0) { + dev_err(wled->dev, "Failed to enable feedback for SINK %d rc = %d\n", +- i + 1, rc); ++ j + 1, rc); + goto failed_detect; + } + +@@ -635,7 +635,7 @@ static void wled_auto_string_detection(struct wled *wled) + WLED4_SINK_REG_CURR_SINK, sink_test); + if (rc < 0) { + dev_err(wled->dev, "Failed to configure SINK %d rc=%d\n", +- i + 1, rc); ++ j + 1, rc); + goto failed_detect; + } + +@@ -662,7 +662,7 @@ static void wled_auto_string_detection(struct wled *wled) + + if (fault_set) + dev_dbg(wled->dev, "WLED OVP fault detected with SINK %d\n", +- i + 1); ++ j + 1); + else + sink_valid |= sink_test; + +@@ -702,15 +702,16 @@ static void wled_auto_string_detection(struct wled *wled) + /* Enable valid sinks */ + if (wled->version == 4) { + for (i = 0; i < wled->cfg.num_strings; i++) { ++ j = wled->cfg.enabled_strings[i]; + if (sink_config & +- BIT(WLED4_SINK_REG_CURR_SINK_SHFT + i)) ++ BIT(WLED4_SINK_REG_CURR_SINK_SHFT + j)) + val = WLED4_SINK_REG_STR_MOD_MASK; + else + /* Disable modulator_en for unused sink */ + val = 0; + + rc = regmap_write(wled->regmap, wled->sink_addr + +- WLED4_SINK_REG_STR_MOD_EN(i), val); ++ WLED4_SINK_REG_STR_MOD_EN(j), val); + if (rc < 0) { + dev_err(wled->dev, "Failed to configure MODULATOR_EN rc=%d\n", + rc); +@@ -1256,21 +1257,6 @@ static const struct wled_var_cfg wled5_ovp_cfg = { + .size = 16, + }; + +-static u32 wled3_num_strings_values_fn(u32 idx) +-{ +- return idx + 1; +-} +- +-static const struct wled_var_cfg wled3_num_strings_cfg = { +- .fn = wled3_num_strings_values_fn, +- .size = 3, +-}; +- +-static const struct wled_var_cfg wled4_num_strings_cfg = { +- .fn = wled3_num_strings_values_fn, +- .size = 4, +-}; +- + static u32 wled3_switch_freq_values_fn(u32 idx) + { + return 19200 / (2 * (1 + idx)); +@@ -1344,11 +1330,6 @@ static int wled_configure(struct wled *wled) + .val_ptr = &cfg->switch_freq, + .cfg = &wled3_switch_freq_cfg, + }, +- { +- .name = "qcom,num-strings", +- .val_ptr = &cfg->num_strings, +- .cfg = &wled3_num_strings_cfg, +- }, + }; + + const struct wled_u32_opts wled4_opts[] = { +@@ -1372,11 +1353,6 @@ static int wled_configure(struct wled *wled) + .val_ptr = &cfg->switch_freq, + .cfg = &wled3_switch_freq_cfg, + }, +- { +- .name = "qcom,num-strings", +- .val_ptr = &cfg->num_strings, +- .cfg = &wled4_num_strings_cfg, +- }, + }; + + const struct wled_u32_opts wled5_opts[] = { +@@ -1400,11 +1376,6 @@ static int wled_configure(struct wled *wled) + .val_ptr = &cfg->switch_freq, + .cfg = &wled3_switch_freq_cfg, + }, +- { +- .name = "qcom,num-strings", +- .val_ptr = &cfg->num_strings, +- .cfg = &wled4_num_strings_cfg, +- }, + { + .name = "qcom,modulator-sel", + .val_ptr = &cfg->mod_sel, +@@ -1523,16 +1494,57 @@ static int wled_configure(struct wled *wled) + *bool_opts[i].val_ptr = true; + } + +- cfg->num_strings = cfg->num_strings + 1; +- + string_len = of_property_count_elems_of_size(dev->of_node, + "qcom,enabled-strings", + sizeof(u32)); +- if (string_len > 0) +- of_property_read_u32_array(dev->of_node, ++ if (string_len > 0) { ++ if (string_len > wled->max_string_count) { ++ dev_err(dev, "Cannot have more than %d strings\n", ++ wled->max_string_count); ++ return -EINVAL; ++ } ++ ++ rc = of_property_read_u32_array(dev->of_node, + "qcom,enabled-strings", + wled->cfg.enabled_strings, +- sizeof(u32)); ++ string_len); ++ if (rc) { ++ dev_err(dev, "Failed to read %d elements from qcom,enabled-strings: %d\n", ++ string_len, rc); ++ return rc; ++ } ++ ++ for (i = 0; i < string_len; ++i) { ++ if (wled->cfg.enabled_strings[i] >= wled->max_string_count) { ++ dev_err(dev, ++ "qcom,enabled-strings index %d at %d is out of bounds\n", ++ wled->cfg.enabled_strings[i], i); ++ return -EINVAL; ++ } ++ } ++ ++ cfg->num_strings = string_len; ++ } ++ ++ rc = of_property_read_u32(dev->of_node, "qcom,num-strings", &val); ++ if (!rc) { ++ if (val < 1 || val > wled->max_string_count) { ++ dev_err(dev, "qcom,num-strings must be between 1 and %d\n", ++ wled->max_string_count); ++ return -EINVAL; ++ } ++ ++ if (string_len > 0) { ++ dev_warn(dev, "Only one of qcom,num-strings or qcom,enabled-strings" ++ " should be set\n"); ++ if (val > string_len) { ++ dev_err(dev, "qcom,num-strings exceeds qcom,enabled-strings\n"); ++ return -EINVAL; ++ } ++ } ++ ++ cfg->num_strings = val; ++ } + + return 0; + } +diff --git a/drivers/virtio/virtio_mem.c b/drivers/virtio/virtio_mem.c +index bef8ad6bf4661..4624a2c3d0553 100644 +--- a/drivers/virtio/virtio_mem.c ++++ b/drivers/virtio/virtio_mem.c +@@ -577,7 +577,7 @@ static int virtio_mem_sbm_sb_states_prepare_next_mb(struct virtio_mem *vm) + return -ENOMEM; + + mutex_lock(&vm->hotplug_mutex); +- if (new_bitmap) ++ if (vm->sbm.sb_states) + memcpy(new_bitmap, vm->sbm.sb_states, old_pages * PAGE_SIZE); + + old_bitmap = vm->sbm.sb_states; +diff --git a/drivers/virtio/virtio_ring.c b/drivers/virtio/virtio_ring.c +index 9919407973cd5..603a6f4345efd 100644 +--- a/drivers/virtio/virtio_ring.c ++++ b/drivers/virtio/virtio_ring.c +@@ -1197,8 +1197,10 @@ static inline int virtqueue_add_packed(struct virtqueue *_vq, + if (virtqueue_use_indirect(_vq, total_sg)) { + err = virtqueue_add_indirect_packed(vq, sgs, total_sg, out_sgs, + in_sgs, data, gfp); +- if (err != -ENOMEM) ++ if (err != -ENOMEM) { ++ END_USE(vq); + return err; ++ } + + /* fall back on direct */ + } +diff --git a/drivers/w1/slaves/w1_ds28e04.c b/drivers/w1/slaves/w1_ds28e04.c +index e4f336111edc6..6cef6e2edb892 100644 +--- a/drivers/w1/slaves/w1_ds28e04.c ++++ b/drivers/w1/slaves/w1_ds28e04.c +@@ -32,7 +32,7 @@ static int w1_strong_pullup = 1; + module_param_named(strong_pullup, w1_strong_pullup, int, 0); + + /* enable/disable CRC checking on DS28E04-100 memory accesses */ +-static char w1_enable_crccheck = 1; ++static bool w1_enable_crccheck = true; + + #define W1_EEPROM_SIZE 512 + #define W1_PAGE_COUNT 16 +@@ -339,32 +339,18 @@ static BIN_ATTR_RW(pio, 1); + static ssize_t crccheck_show(struct device *dev, struct device_attribute *attr, + char *buf) + { +- if (put_user(w1_enable_crccheck + 0x30, buf)) +- return -EFAULT; +- +- return sizeof(w1_enable_crccheck); ++ return sysfs_emit(buf, "%d\n", w1_enable_crccheck); + } + + static ssize_t crccheck_store(struct device *dev, struct device_attribute *attr, + const char *buf, size_t count) + { +- char val; +- +- if (count != 1 || !buf) +- return -EINVAL; ++ int err = kstrtobool(buf, &w1_enable_crccheck); + +- if (get_user(val, buf)) +- return -EFAULT; ++ if (err) ++ return err; + +- /* convert to decimal */ +- val = val - 0x30; +- if (val != 0 && val != 1) +- return -EINVAL; +- +- /* set the new value */ +- w1_enable_crccheck = val; +- +- return sizeof(w1_enable_crccheck); ++ return count; + } + + static DEVICE_ATTR_RW(crccheck); +diff --git a/drivers/xen/gntdev.c b/drivers/xen/gntdev.c +index fec1b65371665..59ffea8000791 100644 +--- a/drivers/xen/gntdev.c ++++ b/drivers/xen/gntdev.c +@@ -250,13 +250,13 @@ void gntdev_put_map(struct gntdev_priv *priv, struct gntdev_grant_map *map) + if (!refcount_dec_and_test(&map->users)) + return; + ++ if (map->pages && !use_ptemod) ++ unmap_grant_pages(map, 0, map->count); ++ + if (map->notify.flags & UNMAP_NOTIFY_SEND_EVENT) { + notify_remote_via_evtchn(map->notify.event); + evtchn_put(map->notify.event); + } +- +- if (map->pages && !use_ptemod) +- unmap_grant_pages(map, 0, map->count); + gntdev_free_map(map); + } + +diff --git a/fs/btrfs/backref.c b/fs/btrfs/backref.c +index f735b8798ba12..8b090c40daf77 100644 +--- a/fs/btrfs/backref.c ++++ b/fs/btrfs/backref.c +@@ -1214,7 +1214,12 @@ again: + ret = btrfs_search_slot(NULL, fs_info->extent_root, &key, path, 0, 0); + if (ret < 0) + goto out; +- BUG_ON(ret == 0); ++ if (ret == 0) { ++ /* This shouldn't happen, indicates a bug or fs corruption. */ ++ ASSERT(ret != 0); ++ ret = -EUCLEAN; ++ goto out; ++ } + + #ifdef CONFIG_BTRFS_FS_RUN_SANITY_TESTS + if (trans && likely(trans->type != __TRANS_DUMMY) && +@@ -1360,10 +1365,18 @@ again: + goto out; + if (!ret && extent_item_pos) { + /* +- * we've recorded that parent, so we must extend +- * its inode list here ++ * We've recorded that parent, so we must extend ++ * its inode list here. ++ * ++ * However if there was corruption we may not ++ * have found an eie, return an error in this ++ * case. + */ +- BUG_ON(!eie); ++ ASSERT(eie); ++ if (!eie) { ++ ret = -EUCLEAN; ++ goto out; ++ } + while (eie->next) + eie = eie->next; + eie->next = ref->inode_list; +diff --git a/fs/btrfs/ctree.c b/fs/btrfs/ctree.c +index 84627cbd5b5b5..95a6a63caf047 100644 +--- a/fs/btrfs/ctree.c ++++ b/fs/btrfs/ctree.c +@@ -1568,12 +1568,9 @@ static struct extent_buffer *btrfs_search_slot_get_root(struct btrfs_root *root, + { + struct btrfs_fs_info *fs_info = root->fs_info; + struct extent_buffer *b; +- int root_lock; ++ int root_lock = 0; + int level = 0; + +- /* We try very hard to do read locks on the root */ +- root_lock = BTRFS_READ_LOCK; +- + if (p->search_commit_root) { + /* + * The commit roots are read only so we always do read locks, +@@ -1611,6 +1608,9 @@ static struct extent_buffer *btrfs_search_slot_get_root(struct btrfs_root *root, + goto out; + } + ++ /* We try very hard to do read locks on the root */ ++ root_lock = BTRFS_READ_LOCK; ++ + /* + * If the level is set to maximum, we can skip trying to get the read + * lock. +@@ -1637,6 +1637,17 @@ static struct extent_buffer *btrfs_search_slot_get_root(struct btrfs_root *root, + level = btrfs_header_level(b); + + out: ++ /* ++ * The root may have failed to write out at some point, and thus is no ++ * longer valid, return an error in this case. ++ */ ++ if (!extent_buffer_uptodate(b)) { ++ if (root_lock) ++ btrfs_tree_unlock_rw(b, root_lock); ++ free_extent_buffer(b); ++ return ERR_PTR(-EIO); ++ } ++ + p->nodes[level] = b; + if (!p->skip_locking) + p->locks[level] = root_lock; +diff --git a/fs/btrfs/inode.c b/fs/btrfs/inode.c +index 4af74b62e7d97..e92f0b0afe9ec 100644 +--- a/fs/btrfs/inode.c ++++ b/fs/btrfs/inode.c +@@ -10586,9 +10586,19 @@ static int btrfs_add_swap_extent(struct swap_info_struct *sis, + struct btrfs_swap_info *bsi) + { + unsigned long nr_pages; ++ unsigned long max_pages; + u64 first_ppage, first_ppage_reported, next_ppage; + int ret; + ++ /* ++ * Our swapfile may have had its size extended after the swap header was ++ * written. In that case activating the swapfile should not go beyond ++ * the max size set in the swap header. ++ */ ++ if (bsi->nr_pages >= sis->max) ++ return 0; ++ ++ max_pages = sis->max - bsi->nr_pages; + first_ppage = ALIGN(bsi->block_start, PAGE_SIZE) >> PAGE_SHIFT; + next_ppage = ALIGN_DOWN(bsi->block_start + bsi->block_len, + PAGE_SIZE) >> PAGE_SHIFT; +@@ -10596,6 +10606,7 @@ static int btrfs_add_swap_extent(struct swap_info_struct *sis, + if (first_ppage >= next_ppage) + return 0; + nr_pages = next_ppage - first_ppage; ++ nr_pages = min(nr_pages, max_pages); + + first_ppage_reported = first_ppage; + if (bsi->start == 0) +diff --git a/fs/btrfs/qgroup.c b/fs/btrfs/qgroup.c +index db680f5be745a..7f6a05e670f57 100644 +--- a/fs/btrfs/qgroup.c ++++ b/fs/btrfs/qgroup.c +@@ -940,6 +940,14 @@ int btrfs_quota_enable(struct btrfs_fs_info *fs_info) + int ret = 0; + int slot; + ++ /* ++ * We need to have subvol_sem write locked, to prevent races between ++ * concurrent tasks trying to enable quotas, because we will unlock ++ * and relock qgroup_ioctl_lock before setting fs_info->quota_root ++ * and before setting BTRFS_FS_QUOTA_ENABLED. ++ */ ++ lockdep_assert_held_write(&fs_info->subvol_sem); ++ + mutex_lock(&fs_info->qgroup_ioctl_lock); + if (fs_info->quota_root) + goto out; +@@ -1117,8 +1125,19 @@ out_add_root: + goto out_free_path; + } + ++ mutex_unlock(&fs_info->qgroup_ioctl_lock); ++ /* ++ * Commit the transaction while not holding qgroup_ioctl_lock, to avoid ++ * a deadlock with tasks concurrently doing other qgroup operations, such ++ * adding/removing qgroups or adding/deleting qgroup relations for example, ++ * because all qgroup operations first start or join a transaction and then ++ * lock the qgroup_ioctl_lock mutex. ++ * We are safe from a concurrent task trying to enable quotas, by calling ++ * this function, since we are serialized by fs_info->subvol_sem. ++ */ + ret = btrfs_commit_transaction(trans); + trans = NULL; ++ mutex_lock(&fs_info->qgroup_ioctl_lock); + if (ret) + goto out_free_path; + +diff --git a/fs/debugfs/file.c b/fs/debugfs/file.c +index 7d162b0efbf03..950c63fa4d0b2 100644 +--- a/fs/debugfs/file.c ++++ b/fs/debugfs/file.c +@@ -147,7 +147,7 @@ static int debugfs_locked_down(struct inode *inode, + struct file *filp, + const struct file_operations *real_fops) + { +- if ((inode->i_mode & 07777) == 0444 && ++ if ((inode->i_mode & 07777 & ~0444) == 0 && + !(filp->f_mode & FMODE_WRITE) && + !real_fops->unlocked_ioctl && + !real_fops->compat_ioctl && +diff --git a/fs/dlm/lock.c b/fs/dlm/lock.c +index c502c065d0075..28d1f35b11a4d 100644 +--- a/fs/dlm/lock.c ++++ b/fs/dlm/lock.c +@@ -3973,6 +3973,14 @@ static int validate_message(struct dlm_lkb *lkb, struct dlm_message *ms) + int from = ms->m_header.h_nodeid; + int error = 0; + ++ /* currently mixing of user/kernel locks are not supported */ ++ if (ms->m_flags & DLM_IFL_USER && ~lkb->lkb_flags & DLM_IFL_USER) { ++ log_error(lkb->lkb_resource->res_ls, ++ "got user dlm message for a kernel lock"); ++ error = -EINVAL; ++ goto out; ++ } ++ + switch (ms->m_type) { + case DLM_MSG_CONVERT: + case DLM_MSG_UNLOCK: +@@ -4001,6 +4009,7 @@ static int validate_message(struct dlm_lkb *lkb, struct dlm_message *ms) + error = -EINVAL; + } + ++out: + if (error) + log_error(lkb->lkb_resource->res_ls, + "ignore invalid message %d from %d %x %x %x %d", +diff --git a/fs/dlm/lowcomms.c b/fs/dlm/lowcomms.c +index 8f715c620e1f8..7a8efce1c343e 100644 +--- a/fs/dlm/lowcomms.c ++++ b/fs/dlm/lowcomms.c +@@ -592,8 +592,8 @@ int dlm_lowcomms_nodes_set_mark(int nodeid, unsigned int mark) + static void lowcomms_error_report(struct sock *sk) + { + struct connection *con; +- struct sockaddr_storage saddr; + void (*orig_report)(struct sock *) = NULL; ++ struct inet_sock *inet; + + read_lock_bh(&sk->sk_callback_lock); + con = sock2con(sk); +@@ -601,33 +601,33 @@ static void lowcomms_error_report(struct sock *sk) + goto out; + + orig_report = listen_sock.sk_error_report; +- if (kernel_getpeername(sk->sk_socket, (struct sockaddr *)&saddr) < 0) { +- printk_ratelimited(KERN_ERR "dlm: node %d: socket error " +- "sending to node %d, port %d, " +- "sk_err=%d/%d\n", dlm_our_nodeid(), +- con->nodeid, dlm_config.ci_tcp_port, +- sk->sk_err, sk->sk_err_soft); +- } else if (saddr.ss_family == AF_INET) { +- struct sockaddr_in *sin4 = (struct sockaddr_in *)&saddr; + ++ inet = inet_sk(sk); ++ switch (sk->sk_family) { ++ case AF_INET: + printk_ratelimited(KERN_ERR "dlm: node %d: socket error " +- "sending to node %d at %pI4, port %d, " ++ "sending to node %d at %pI4, dport %d, " + "sk_err=%d/%d\n", dlm_our_nodeid(), +- con->nodeid, &sin4->sin_addr.s_addr, +- dlm_config.ci_tcp_port, sk->sk_err, ++ con->nodeid, &inet->inet_daddr, ++ ntohs(inet->inet_dport), sk->sk_err, + sk->sk_err_soft); +- } else { +- struct sockaddr_in6 *sin6 = (struct sockaddr_in6 *)&saddr; +- ++ break; ++#if IS_ENABLED(CONFIG_IPV6) ++ case AF_INET6: + printk_ratelimited(KERN_ERR "dlm: node %d: socket error " +- "sending to node %d at %u.%u.%u.%u, " +- "port %d, sk_err=%d/%d\n", dlm_our_nodeid(), +- con->nodeid, sin6->sin6_addr.s6_addr32[0], +- sin6->sin6_addr.s6_addr32[1], +- sin6->sin6_addr.s6_addr32[2], +- sin6->sin6_addr.s6_addr32[3], +- dlm_config.ci_tcp_port, sk->sk_err, ++ "sending to node %d at %pI6c, " ++ "dport %d, sk_err=%d/%d\n", dlm_our_nodeid(), ++ con->nodeid, &sk->sk_v6_daddr, ++ ntohs(inet->inet_dport), sk->sk_err, + sk->sk_err_soft); ++ break; ++#endif ++ default: ++ printk_ratelimited(KERN_ERR "dlm: node %d: socket error " ++ "invalid socket family %d set, " ++ "sk_err=%d/%d\n", dlm_our_nodeid(), ++ sk->sk_family, sk->sk_err, sk->sk_err_soft); ++ goto out; + } + + /* below sendcon only handling */ +diff --git a/fs/ext4/ext4.h b/fs/ext4/ext4.h +index 3825195539d74..c97860ef322db 100644 +--- a/fs/ext4/ext4.h ++++ b/fs/ext4/ext4.h +@@ -2934,6 +2934,7 @@ bool ext4_fc_replay_check_excluded(struct super_block *sb, ext4_fsblk_t block); + void ext4_fc_replay_cleanup(struct super_block *sb); + int ext4_fc_commit(journal_t *journal, tid_t commit_tid); + int __init ext4_fc_init_dentry_cache(void); ++void ext4_fc_destroy_dentry_cache(void); + + /* mballoc.c */ + extern const struct seq_operations ext4_mb_seq_groups_ops; +diff --git a/fs/ext4/ext4_jbd2.c b/fs/ext4/ext4_jbd2.c +index 6def7339056db..3477a16d08aee 100644 +--- a/fs/ext4/ext4_jbd2.c ++++ b/fs/ext4/ext4_jbd2.c +@@ -162,6 +162,8 @@ int __ext4_journal_ensure_credits(handle_t *handle, int check_cred, + { + if (!ext4_handle_valid(handle)) + return 0; ++ if (is_handle_aborted(handle)) ++ return -EROFS; + if (jbd2_handle_buffer_credits(handle) >= check_cred && + handle->h_revoke_credits >= revoke_cred) + return 0; +diff --git a/fs/ext4/extents.c b/fs/ext4/extents.c +index daa83525749b3..f9bd3231a1415 100644 +--- a/fs/ext4/extents.c ++++ b/fs/ext4/extents.c +@@ -4645,8 +4645,6 @@ static long ext4_zero_range(struct file *file, loff_t offset, + ret = ext4_mark_inode_dirty(handle, inode); + if (unlikely(ret)) + goto out_handle; +- ext4_fc_track_range(handle, inode, offset >> inode->i_sb->s_blocksize_bits, +- (offset + len - 1) >> inode->i_sb->s_blocksize_bits); + /* Zero out partial block at the edges of the range */ + ret = ext4_zero_partial_blocks(handle, inode, offset, len); + if (ret >= 0) +diff --git a/fs/ext4/fast_commit.c b/fs/ext4/fast_commit.c +index 8ea5a81e65548..e0d393425caca 100644 +--- a/fs/ext4/fast_commit.c ++++ b/fs/ext4/fast_commit.c +@@ -1809,11 +1809,14 @@ ext4_fc_replay_del_range(struct super_block *sb, struct ext4_fc_tl *tl, + } + } + +- ret = ext4_punch_hole(inode, +- le32_to_cpu(lrange.fc_lblk) << sb->s_blocksize_bits, +- le32_to_cpu(lrange.fc_len) << sb->s_blocksize_bits); +- if (ret) +- jbd_debug(1, "ext4_punch_hole returned %d", ret); ++ down_write(&EXT4_I(inode)->i_data_sem); ++ ret = ext4_ext_remove_space(inode, lrange.fc_lblk, ++ lrange.fc_lblk + lrange.fc_len - 1); ++ up_write(&EXT4_I(inode)->i_data_sem); ++ if (ret) { ++ iput(inode); ++ return 0; ++ } + ext4_ext_replay_shrink_inode(inode, + i_size_read(inode) >> sb->s_blocksize_bits); + ext4_mark_inode_dirty(NULL, inode); +@@ -2185,3 +2188,8 @@ int __init ext4_fc_init_dentry_cache(void) + + return 0; + } ++ ++void ext4_fc_destroy_dentry_cache(void) ++{ ++ kmem_cache_destroy(ext4_fc_dentry_cachep); ++} +diff --git a/fs/ext4/inode.c b/fs/ext4/inode.c +index 9097fccdc6889..b6746cc86cee3 100644 +--- a/fs/ext4/inode.c ++++ b/fs/ext4/inode.c +@@ -741,10 +741,11 @@ out_sem: + if (ret) + return ret; + } +- ext4_fc_track_range(handle, inode, map->m_lblk, +- map->m_lblk + map->m_len - 1); + } +- ++ if (retval > 0 && (map->m_flags & EXT4_MAP_UNWRITTEN || ++ map->m_flags & EXT4_MAP_MAPPED)) ++ ext4_fc_track_range(handle, inode, map->m_lblk, ++ map->m_lblk + map->m_len - 1); + if (retval < 0) + ext_debug(inode, "failed with err %d\n", retval); + return retval; +@@ -1844,30 +1845,16 @@ int ext4_da_get_block_prep(struct inode *inode, sector_t iblock, + return 0; + } + +-static int bget_one(handle_t *handle, struct inode *inode, +- struct buffer_head *bh) +-{ +- get_bh(bh); +- return 0; +-} +- +-static int bput_one(handle_t *handle, struct inode *inode, +- struct buffer_head *bh) +-{ +- put_bh(bh); +- return 0; +-} +- + static int __ext4_journalled_writepage(struct page *page, + unsigned int len) + { + struct address_space *mapping = page->mapping; + struct inode *inode = mapping->host; +- struct buffer_head *page_bufs = NULL; + handle_t *handle = NULL; + int ret = 0, err = 0; + int inline_data = ext4_has_inline_data(inode); + struct buffer_head *inode_bh = NULL; ++ loff_t size; + + ClearPageChecked(page); + +@@ -1877,14 +1864,6 @@ static int __ext4_journalled_writepage(struct page *page, + inode_bh = ext4_journalled_write_inline_data(inode, len, page); + if (inode_bh == NULL) + goto out; +- } else { +- page_bufs = page_buffers(page); +- if (!page_bufs) { +- BUG(); +- goto out; +- } +- ext4_walk_page_buffers(handle, inode, page_bufs, 0, len, +- NULL, bget_one); + } + /* + * We need to release the page lock before we start the +@@ -1905,7 +1884,8 @@ static int __ext4_journalled_writepage(struct page *page, + + lock_page(page); + put_page(page); +- if (page->mapping != mapping) { ++ size = i_size_read(inode); ++ if (page->mapping != mapping || page_offset(page) > size) { + /* The page got truncated from under us */ + ext4_journal_stop(handle); + ret = 0; +@@ -1915,6 +1895,13 @@ static int __ext4_journalled_writepage(struct page *page, + if (inline_data) { + ret = ext4_mark_inode_dirty(handle, inode); + } else { ++ struct buffer_head *page_bufs = page_buffers(page); ++ ++ if (page->index == size >> PAGE_SHIFT) ++ len = size & ~PAGE_MASK; ++ else ++ len = PAGE_SIZE; ++ + ret = ext4_walk_page_buffers(handle, inode, page_bufs, 0, len, + NULL, do_journal_get_write_access); + +@@ -1935,9 +1922,6 @@ static int __ext4_journalled_writepage(struct page *page, + out: + unlock_page(page); + out_no_pagelock: +- if (!inline_data && page_bufs) +- ext4_walk_page_buffers(NULL, inode, page_bufs, 0, len, +- NULL, bput_one); + brelse(inode_bh); + return ret; + } +@@ -4371,7 +4355,7 @@ has_buffer: + static int __ext4_get_inode_loc_noinmem(struct inode *inode, + struct ext4_iloc *iloc) + { +- ext4_fsblk_t err_blk; ++ ext4_fsblk_t err_blk = 0; + int ret; + + ret = __ext4_get_inode_loc(inode->i_sb, inode->i_ino, iloc, 0, +@@ -4386,7 +4370,7 @@ static int __ext4_get_inode_loc_noinmem(struct inode *inode, + + int ext4_get_inode_loc(struct inode *inode, struct ext4_iloc *iloc) + { +- ext4_fsblk_t err_blk; ++ ext4_fsblk_t err_blk = 0; + int ret; + + /* We have all inode data except xattrs in memory here. */ +@@ -5413,8 +5397,7 @@ int ext4_setattr(struct user_namespace *mnt_userns, struct dentry *dentry, + ext4_fc_track_range(handle, inode, + (attr->ia_size > 0 ? attr->ia_size - 1 : 0) >> + inode->i_sb->s_blocksize_bits, +- (oldsize > 0 ? oldsize - 1 : 0) >> +- inode->i_sb->s_blocksize_bits); ++ EXT_MAX_BLOCKS - 1); + else + ext4_fc_track_range( + handle, inode, +diff --git a/fs/ext4/ioctl.c b/fs/ext4/ioctl.c +index 606dee9e08a32..220a4c8178b5e 100644 +--- a/fs/ext4/ioctl.c ++++ b/fs/ext4/ioctl.c +@@ -1117,8 +1117,6 @@ resizefs_out: + sizeof(range))) + return -EFAULT; + +- range.minlen = max((unsigned int)range.minlen, +- q->limits.discard_granularity); + ret = ext4_trim_fs(sb, &range); + if (ret < 0) + return ret; +diff --git a/fs/ext4/mballoc.c b/fs/ext4/mballoc.c +index 72bfac2d6dce9..b05a83be88715 100644 +--- a/fs/ext4/mballoc.c ++++ b/fs/ext4/mballoc.c +@@ -4814,7 +4814,7 @@ ext4_mb_release_group_pa(struct ext4_buddy *e4b, + */ + static noinline_for_stack int + ext4_mb_discard_group_preallocations(struct super_block *sb, +- ext4_group_t group, int needed) ++ ext4_group_t group, int *busy) + { + struct ext4_group_info *grp = ext4_get_group_info(sb, group); + struct buffer_head *bitmap_bh = NULL; +@@ -4822,8 +4822,7 @@ ext4_mb_discard_group_preallocations(struct super_block *sb, + struct list_head list; + struct ext4_buddy e4b; + int err; +- int busy = 0; +- int free, free_total = 0; ++ int free = 0; + + mb_debug(sb, "discard preallocation for group %u\n", group); + if (list_empty(&grp->bb_prealloc_list)) +@@ -4846,19 +4845,14 @@ ext4_mb_discard_group_preallocations(struct super_block *sb, + goto out_dbg; + } + +- if (needed == 0) +- needed = EXT4_CLUSTERS_PER_GROUP(sb) + 1; +- + INIT_LIST_HEAD(&list); +-repeat: +- free = 0; + ext4_lock_group(sb, group); + list_for_each_entry_safe(pa, tmp, + &grp->bb_prealloc_list, pa_group_list) { + spin_lock(&pa->pa_lock); + if (atomic_read(&pa->pa_count)) { + spin_unlock(&pa->pa_lock); +- busy = 1; ++ *busy = 1; + continue; + } + if (pa->pa_deleted) { +@@ -4898,22 +4892,13 @@ repeat: + call_rcu(&(pa)->u.pa_rcu, ext4_mb_pa_callback); + } + +- free_total += free; +- +- /* if we still need more blocks and some PAs were used, try again */ +- if (free_total < needed && busy) { +- ext4_unlock_group(sb, group); +- cond_resched(); +- busy = 0; +- goto repeat; +- } + ext4_unlock_group(sb, group); + ext4_mb_unload_buddy(&e4b); + put_bh(bitmap_bh); + out_dbg: + mb_debug(sb, "discarded (%d) blocks preallocated for group %u bb_free (%d)\n", +- free_total, group, grp->bb_free); +- return free_total; ++ free, group, grp->bb_free); ++ return free; + } + + /* +@@ -5455,13 +5440,24 @@ static int ext4_mb_discard_preallocations(struct super_block *sb, int needed) + { + ext4_group_t i, ngroups = ext4_get_groups_count(sb); + int ret; +- int freed = 0; ++ int freed = 0, busy = 0; ++ int retry = 0; + + trace_ext4_mb_discard_preallocations(sb, needed); ++ ++ if (needed == 0) ++ needed = EXT4_CLUSTERS_PER_GROUP(sb) + 1; ++ repeat: + for (i = 0; i < ngroups && needed > 0; i++) { +- ret = ext4_mb_discard_group_preallocations(sb, i, needed); ++ ret = ext4_mb_discard_group_preallocations(sb, i, &busy); + freed += ret; + needed -= ret; ++ cond_resched(); ++ } ++ ++ if (needed > 0 && busy && ++retry < 3) { ++ busy = 0; ++ goto repeat; + } + + return freed; +@@ -6405,6 +6401,7 @@ ext4_trim_all_free(struct super_block *sb, ext4_group_t group, + */ + int ext4_trim_fs(struct super_block *sb, struct fstrim_range *range) + { ++ struct request_queue *q = bdev_get_queue(sb->s_bdev); + struct ext4_group_info *grp; + ext4_group_t group, first_group, last_group; + ext4_grpblk_t cnt = 0, first_cluster, last_cluster; +@@ -6423,6 +6420,13 @@ int ext4_trim_fs(struct super_block *sb, struct fstrim_range *range) + start >= max_blks || + range->len < sb->s_blocksize) + return -EINVAL; ++ /* No point to try to trim less than discard granularity */ ++ if (range->minlen < q->limits.discard_granularity) { ++ minlen = EXT4_NUM_B2C(EXT4_SB(sb), ++ q->limits.discard_granularity >> sb->s_blocksize_bits); ++ if (minlen > EXT4_CLUSTERS_PER_GROUP(sb)) ++ goto out; ++ } + if (end >= max_blks) + end = max_blks - 1; + if (end <= first_data_blk) +diff --git a/fs/ext4/migrate.c b/fs/ext4/migrate.c +index 7e0b4f81c6c06..ff8916e1d38e9 100644 +--- a/fs/ext4/migrate.c ++++ b/fs/ext4/migrate.c +@@ -437,12 +437,12 @@ int ext4_ext_migrate(struct inode *inode) + percpu_down_write(&sbi->s_writepages_rwsem); + + /* +- * Worst case we can touch the allocation bitmaps, a bgd +- * block, and a block to link in the orphan list. We do need +- * need to worry about credits for modifying the quota inode. ++ * Worst case we can touch the allocation bitmaps and a block ++ * group descriptor block. We do need need to worry about ++ * credits for modifying the quota inode. + */ + handle = ext4_journal_start(inode, EXT4_HT_MIGRATE, +- 4 + EXT4_MAXQUOTAS_TRANS_BLOCKS(inode->i_sb)); ++ 3 + EXT4_MAXQUOTAS_TRANS_BLOCKS(inode->i_sb)); + + if (IS_ERR(handle)) { + retval = PTR_ERR(handle); +@@ -459,6 +459,13 @@ int ext4_ext_migrate(struct inode *inode) + ext4_journal_stop(handle); + goto out_unlock; + } ++ /* ++ * Use the correct seed for checksum (i.e. the seed from 'inode'). This ++ * is so that the metadata blocks will have the correct checksum after ++ * the migration. ++ */ ++ ei = EXT4_I(inode); ++ EXT4_I(tmp_inode)->i_csum_seed = ei->i_csum_seed; + i_size_write(tmp_inode, i_size_read(inode)); + /* + * Set the i_nlink to zero so it will be deleted later +@@ -467,7 +474,6 @@ int ext4_ext_migrate(struct inode *inode) + clear_nlink(tmp_inode); + + ext4_ext_tree_init(handle, tmp_inode); +- ext4_orphan_add(handle, tmp_inode); + ext4_journal_stop(handle); + + /* +@@ -492,17 +498,10 @@ int ext4_ext_migrate(struct inode *inode) + + handle = ext4_journal_start(inode, EXT4_HT_MIGRATE, 1); + if (IS_ERR(handle)) { +- /* +- * It is impossible to update on-disk structures without +- * a handle, so just rollback in-core changes and live other +- * work to orphan_list_cleanup() +- */ +- ext4_orphan_del(NULL, tmp_inode); + retval = PTR_ERR(handle); + goto out_tmp_inode; + } + +- ei = EXT4_I(inode); + i_data = ei->i_data; + memset(&lb, 0, sizeof(lb)); + +diff --git a/fs/ext4/super.c b/fs/ext4/super.c +index 8a67e5f3f5763..877c5c17e61f0 100644 +--- a/fs/ext4/super.c ++++ b/fs/ext4/super.c +@@ -6266,10 +6266,7 @@ static int ext4_quota_on(struct super_block *sb, int type, int format_id, + + lockdep_set_quota_inode(path->dentry->d_inode, I_DATA_SEM_QUOTA); + err = dquot_quota_on(sb, type, format_id, path); +- if (err) { +- lockdep_set_quota_inode(path->dentry->d_inode, +- I_DATA_SEM_NORMAL); +- } else { ++ if (!err) { + struct inode *inode = d_inode(path->dentry); + handle_t *handle; + +@@ -6289,7 +6286,12 @@ static int ext4_quota_on(struct super_block *sb, int type, int format_id, + ext4_journal_stop(handle); + unlock_inode: + inode_unlock(inode); ++ if (err) ++ dquot_quota_off(sb, type); + } ++ if (err) ++ lockdep_set_quota_inode(path->dentry->d_inode, ++ I_DATA_SEM_NORMAL); + return err; + } + +@@ -6352,8 +6354,19 @@ int ext4_enable_quotas(struct super_block *sb) + "Failed to enable quota tracking " + "(type=%d, err=%d). Please run " + "e2fsck to fix.", type, err); +- for (type--; type >= 0; type--) ++ for (type--; type >= 0; type--) { ++ struct inode *inode; ++ ++ inode = sb_dqopt(sb)->files[type]; ++ if (inode) ++ inode = igrab(inode); + dquot_quota_off(sb, type); ++ if (inode) { ++ lockdep_set_quota_inode(inode, ++ I_DATA_SEM_NORMAL); ++ iput(inode); ++ } ++ } + + return err; + } +@@ -6457,7 +6470,7 @@ static ssize_t ext4_quota_write(struct super_block *sb, int type, + struct buffer_head *bh; + handle_t *handle = journal_current_handle(); + +- if (EXT4_SB(sb)->s_journal && !handle) { ++ if (!handle) { + ext4_msg(sb, KERN_WARNING, "Quota write (off=%llu, len=%llu)" + " cancelled because transaction is not started", + (unsigned long long)off, (unsigned long long)len); +@@ -6640,6 +6653,7 @@ static int __init ext4_init_fs(void) + out: + unregister_as_ext2(); + unregister_as_ext3(); ++ ext4_fc_destroy_dentry_cache(); + out05: + destroy_inodecache(); + out1: +@@ -6666,6 +6680,7 @@ static void __exit ext4_exit_fs(void) + unregister_as_ext2(); + unregister_as_ext3(); + unregister_filesystem(&ext4_fs_type); ++ ext4_fc_destroy_dentry_cache(); + destroy_inodecache(); + ext4_exit_mballoc(); + ext4_exit_sysfs(); +diff --git a/fs/f2fs/checkpoint.c b/fs/f2fs/checkpoint.c +index 7b02827242312..99fced979718e 100644 +--- a/fs/f2fs/checkpoint.c ++++ b/fs/f2fs/checkpoint.c +@@ -1305,8 +1305,8 @@ static void update_ckpt_flags(struct f2fs_sb_info *sbi, struct cp_control *cpc) + unsigned long flags; + + if (cpc->reason & CP_UMOUNT) { +- if (le32_to_cpu(ckpt->cp_pack_total_block_count) > +- sbi->blocks_per_seg - NM_I(sbi)->nat_bits_blocks) { ++ if (le32_to_cpu(ckpt->cp_pack_total_block_count) + ++ NM_I(sbi)->nat_bits_blocks > sbi->blocks_per_seg) { + clear_ckpt_flags(sbi, CP_NAT_BITS_FLAG); + f2fs_notice(sbi, "Disable nat_bits due to no space"); + } else if (!is_set_ckpt_flags(sbi, CP_NAT_BITS_FLAG) && +diff --git a/fs/f2fs/compress.c b/fs/f2fs/compress.c +index 9b663eaf48057..58d255d3a518a 100644 +--- a/fs/f2fs/compress.c ++++ b/fs/f2fs/compress.c +@@ -1448,25 +1448,38 @@ static int f2fs_write_raw_pages(struct compress_ctx *cc, + enum iostat_type io_type) + { + struct address_space *mapping = cc->inode->i_mapping; +- int _submitted, compr_blocks, ret; +- int i = -1, err = 0; ++ int _submitted, compr_blocks, ret, i; + + compr_blocks = f2fs_compressed_blocks(cc); +- if (compr_blocks < 0) { +- err = compr_blocks; +- goto out_err; ++ ++ for (i = 0; i < cc->cluster_size; i++) { ++ if (!cc->rpages[i]) ++ continue; ++ ++ redirty_page_for_writepage(wbc, cc->rpages[i]); ++ unlock_page(cc->rpages[i]); + } + ++ if (compr_blocks < 0) ++ return compr_blocks; ++ + for (i = 0; i < cc->cluster_size; i++) { + if (!cc->rpages[i]) + continue; + retry_write: ++ lock_page(cc->rpages[i]); ++ + if (cc->rpages[i]->mapping != mapping) { ++continue_unlock: + unlock_page(cc->rpages[i]); + continue; + } + +- BUG_ON(!PageLocked(cc->rpages[i])); ++ if (!PageDirty(cc->rpages[i])) ++ goto continue_unlock; ++ ++ if (!clear_page_dirty_for_io(cc->rpages[i])) ++ goto continue_unlock; + + ret = f2fs_write_single_data_page(cc->rpages[i], &_submitted, + NULL, NULL, wbc, io_type, +@@ -1481,26 +1494,15 @@ retry_write: + * avoid deadlock caused by cluster update race + * from foreground operation. + */ +- if (IS_NOQUOTA(cc->inode)) { +- err = 0; +- goto out_err; +- } ++ if (IS_NOQUOTA(cc->inode)) ++ return 0; + ret = 0; + cond_resched(); + congestion_wait(BLK_RW_ASYNC, + DEFAULT_IO_TIMEOUT); +- lock_page(cc->rpages[i]); +- +- if (!PageDirty(cc->rpages[i])) { +- unlock_page(cc->rpages[i]); +- continue; +- } +- +- clear_page_dirty_for_io(cc->rpages[i]); + goto retry_write; + } +- err = ret; +- goto out_err; ++ return ret; + } + + *submitted += _submitted; +@@ -1509,14 +1511,6 @@ retry_write: + f2fs_balance_fs(F2FS_M_SB(mapping), true); + + return 0; +-out_err: +- for (++i; i < cc->cluster_size; i++) { +- if (!cc->rpages[i]) +- continue; +- redirty_page_for_writepage(wbc, cc->rpages[i]); +- unlock_page(cc->rpages[i]); +- } +- return err; + } + + int f2fs_write_multi_pages(struct compress_ctx *cc, +diff --git a/fs/f2fs/data.c b/fs/f2fs/data.c +index f4fd6c246c9a9..e662355cf8c9b 100644 +--- a/fs/f2fs/data.c ++++ b/fs/f2fs/data.c +@@ -2564,6 +2564,11 @@ bool f2fs_should_update_outplace(struct inode *inode, struct f2fs_io_info *fio) + { + struct f2fs_sb_info *sbi = F2FS_I_SB(inode); + ++ /* The below cases were checked when setting it. */ ++ if (f2fs_is_pinned_file(inode)) ++ return false; ++ if (fio && is_sbi_flag_set(sbi, SBI_NEED_FSCK)) ++ return true; + if (f2fs_lfs_mode(sbi)) + return true; + if (S_ISDIR(inode->i_mode)) +@@ -2572,8 +2577,6 @@ bool f2fs_should_update_outplace(struct inode *inode, struct f2fs_io_info *fio) + return true; + if (f2fs_is_atomic_file(inode)) + return true; +- if (is_sbi_flag_set(sbi, SBI_NEED_FSCK)) +- return true; + + /* swap file is migrating in aligned write mode */ + if (is_inode_flag_set(inode, FI_ALIGNED_WRITE)) +diff --git a/fs/f2fs/f2fs.h b/fs/f2fs/f2fs.h +index c242274e3479b..c68817d83a53b 100644 +--- a/fs/f2fs/f2fs.h ++++ b/fs/f2fs/f2fs.h +@@ -1012,6 +1012,7 @@ struct f2fs_sm_info { + unsigned int segment_count; /* total # of segments */ + unsigned int main_segments; /* # of segments in main area */ + unsigned int reserved_segments; /* # of reserved segments */ ++ unsigned int additional_reserved_segments;/* reserved segs for IO align feature */ + unsigned int ovp_segments; /* # of overprovision segments */ + + /* a threshold to reclaim prefree segments */ +@@ -2184,6 +2185,11 @@ static inline int inc_valid_block_count(struct f2fs_sb_info *sbi, + + if (!__allow_reserved_blocks(sbi, inode, true)) + avail_user_block_count -= F2FS_OPTION(sbi).root_reserved_blocks; ++ ++ if (F2FS_IO_ALIGNED(sbi)) ++ avail_user_block_count -= sbi->blocks_per_seg * ++ SM_I(sbi)->additional_reserved_segments; ++ + if (unlikely(is_sbi_flag_set(sbi, SBI_CP_DISABLED))) { + if (avail_user_block_count > sbi->unusable_block_count) + avail_user_block_count -= sbi->unusable_block_count; +@@ -2430,6 +2436,11 @@ static inline int inc_valid_node_count(struct f2fs_sb_info *sbi, + + if (!__allow_reserved_blocks(sbi, inode, false)) + valid_block_count += F2FS_OPTION(sbi).root_reserved_blocks; ++ ++ if (F2FS_IO_ALIGNED(sbi)) ++ valid_block_count += sbi->blocks_per_seg * ++ SM_I(sbi)->additional_reserved_segments; ++ + user_block_count = sbi->user_block_count; + if (unlikely(is_sbi_flag_set(sbi, SBI_CP_DISABLED))) + user_block_count -= sbi->unusable_block_count; +diff --git a/fs/f2fs/file.c b/fs/f2fs/file.c +index 9c8ef33bd8d32..7ed44752c7580 100644 +--- a/fs/f2fs/file.c ++++ b/fs/f2fs/file.c +@@ -3143,17 +3143,17 @@ static int f2fs_ioc_set_pin_file(struct file *filp, unsigned long arg) + + inode_lock(inode); + +- if (f2fs_should_update_outplace(inode, NULL)) { +- ret = -EINVAL; +- goto out; +- } +- + if (!pin) { + clear_inode_flag(inode, FI_PIN_FILE); + f2fs_i_gc_failures_write(inode, 0); + goto done; + } + ++ if (f2fs_should_update_outplace(inode, NULL)) { ++ ret = -EINVAL; ++ goto out; ++ } ++ + if (f2fs_pin_file_control(inode, false)) { + ret = -EAGAIN; + goto out; +diff --git a/fs/f2fs/gc.c b/fs/f2fs/gc.c +index 77391e3b7d68f..264821df0add1 100644 +--- a/fs/f2fs/gc.c ++++ b/fs/f2fs/gc.c +@@ -1023,6 +1023,9 @@ static bool is_alive(struct f2fs_sb_info *sbi, struct f2fs_summary *sum, + set_sbi_flag(sbi, SBI_NEED_FSCK); + } + ++ if (f2fs_check_nid_range(sbi, dni->ino)) ++ return false; ++ + *nofs = ofs_of_node(node_page); + source_blkaddr = data_blkaddr(NULL, node_page, ofs_in_node); + f2fs_put_page(node_page, 1); +@@ -1036,7 +1039,7 @@ static bool is_alive(struct f2fs_sb_info *sbi, struct f2fs_summary *sum, + if (!test_and_set_bit(segno, SIT_I(sbi)->invalid_segmap)) { + f2fs_err(sbi, "mismatched blkaddr %u (source_blkaddr %u) in seg %u", + blkaddr, source_blkaddr, segno); +- f2fs_bug_on(sbi, 1); ++ set_sbi_flag(sbi, SBI_NEED_FSCK); + } + } + #endif +@@ -1454,7 +1457,8 @@ next_step: + + if (phase == 3) { + inode = f2fs_iget(sb, dni.ino); +- if (IS_ERR(inode) || is_bad_inode(inode)) ++ if (IS_ERR(inode) || is_bad_inode(inode) || ++ special_file(inode->i_mode)) + continue; + + if (!down_write_trylock( +diff --git a/fs/f2fs/inode.c b/fs/f2fs/inode.c +index 1213f15ffd68c..4557de37a9119 100644 +--- a/fs/f2fs/inode.c ++++ b/fs/f2fs/inode.c +@@ -516,6 +516,11 @@ make_now: + } else if (ino == F2FS_COMPRESS_INO(sbi)) { + #ifdef CONFIG_F2FS_FS_COMPRESSION + inode->i_mapping->a_ops = &f2fs_compress_aops; ++ /* ++ * generic_error_remove_page only truncates pages of regular ++ * inode ++ */ ++ inode->i_mode |= S_IFREG; + #endif + mapping_set_gfp_mask(inode->i_mapping, + GFP_NOFS | __GFP_HIGHMEM | __GFP_MOVABLE); +diff --git a/fs/f2fs/segment.h b/fs/f2fs/segment.h +index 89fff258727d1..b2d82c9f5a16d 100644 +--- a/fs/f2fs/segment.h ++++ b/fs/f2fs/segment.h +@@ -537,7 +537,8 @@ static inline unsigned int free_segments(struct f2fs_sb_info *sbi) + + static inline unsigned int reserved_segments(struct f2fs_sb_info *sbi) + { +- return SM_I(sbi)->reserved_segments; ++ return SM_I(sbi)->reserved_segments + ++ SM_I(sbi)->additional_reserved_segments; + } + + static inline unsigned int free_sections(struct f2fs_sb_info *sbi) +diff --git a/fs/f2fs/super.c b/fs/f2fs/super.c +index 8795a5a8d4e89..6174c4f4cee7f 100644 +--- a/fs/f2fs/super.c ++++ b/fs/f2fs/super.c +@@ -327,6 +327,46 @@ static inline void limit_reserve_root(struct f2fs_sb_info *sbi) + F2FS_OPTION(sbi).s_resgid)); + } + ++static inline int adjust_reserved_segment(struct f2fs_sb_info *sbi) ++{ ++ unsigned int sec_blks = sbi->blocks_per_seg * sbi->segs_per_sec; ++ unsigned int avg_vblocks; ++ unsigned int wanted_reserved_segments; ++ block_t avail_user_block_count; ++ ++ if (!F2FS_IO_ALIGNED(sbi)) ++ return 0; ++ ++ /* average valid block count in section in worst case */ ++ avg_vblocks = sec_blks / F2FS_IO_SIZE(sbi); ++ ++ /* ++ * we need enough free space when migrating one section in worst case ++ */ ++ wanted_reserved_segments = (F2FS_IO_SIZE(sbi) / avg_vblocks) * ++ reserved_segments(sbi); ++ wanted_reserved_segments -= reserved_segments(sbi); ++ ++ avail_user_block_count = sbi->user_block_count - ++ sbi->current_reserved_blocks - ++ F2FS_OPTION(sbi).root_reserved_blocks; ++ ++ if (wanted_reserved_segments * sbi->blocks_per_seg > ++ avail_user_block_count) { ++ f2fs_err(sbi, "IO align feature can't grab additional reserved segment: %u, available segments: %u", ++ wanted_reserved_segments, ++ avail_user_block_count >> sbi->log_blocks_per_seg); ++ return -ENOSPC; ++ } ++ ++ SM_I(sbi)->additional_reserved_segments = wanted_reserved_segments; ++ ++ f2fs_info(sbi, "IO align feature needs additional reserved segment: %u", ++ wanted_reserved_segments); ++ ++ return 0; ++} ++ + static inline void adjust_unusable_cap_perc(struct f2fs_sb_info *sbi) + { + if (!F2FS_OPTION(sbi).unusable_cap_perc) +@@ -4148,6 +4188,10 @@ try_onemore: + goto free_nm; + } + ++ err = adjust_reserved_segment(sbi); ++ if (err) ++ goto free_nm; ++ + /* For write statistics */ + sbi->sectors_written_start = f2fs_get_sectors_written(sbi); + +diff --git a/fs/f2fs/sysfs.c b/fs/f2fs/sysfs.c +index a32fe31c33b8e..abc4344fba394 100644 +--- a/fs/f2fs/sysfs.c ++++ b/fs/f2fs/sysfs.c +@@ -415,7 +415,9 @@ out: + if (a->struct_type == RESERVED_BLOCKS) { + spin_lock(&sbi->stat_lock); + if (t > (unsigned long)(sbi->user_block_count - +- F2FS_OPTION(sbi).root_reserved_blocks)) { ++ F2FS_OPTION(sbi).root_reserved_blocks - ++ sbi->blocks_per_seg * ++ SM_I(sbi)->additional_reserved_segments)) { + spin_unlock(&sbi->stat_lock); + return -EINVAL; + } +diff --git a/fs/fuse/file.c b/fs/fuse/file.c +index 5c5ed58d91a73..2004d362361e1 100644 +--- a/fs/fuse/file.c ++++ b/fs/fuse/file.c +@@ -2913,7 +2913,7 @@ fuse_direct_IO(struct kiocb *iocb, struct iov_iter *iter) + + static int fuse_writeback_range(struct inode *inode, loff_t start, loff_t end) + { +- int err = filemap_write_and_wait_range(inode->i_mapping, start, -1); ++ int err = filemap_write_and_wait_range(inode->i_mapping, start, LLONG_MAX); + + if (!err) + fuse_sync_writes(inode); +diff --git a/fs/io_uring.c b/fs/io_uring.c +index 0006fc7479ca3..ecffeddf90c68 100644 +--- a/fs/io_uring.c ++++ b/fs/io_uring.c +@@ -5925,6 +5925,7 @@ static int io_poll_update(struct io_kiocb *req, unsigned int issue_flags) + * update those. For multishot, if we're racing with completion, just + * let completion re-add it. + */ ++ io_poll_remove_double(preq); + completing = !__io_poll_remove_one(preq, &preq->poll, false); + if (completing && (preq->poll.events & EPOLLONESHOT)) { + ret = -EALREADY; +diff --git a/fs/jffs2/file.c b/fs/jffs2/file.c +index 4fc8cd698d1a4..bd7d58d27bfc6 100644 +--- a/fs/jffs2/file.c ++++ b/fs/jffs2/file.c +@@ -136,20 +136,15 @@ static int jffs2_write_begin(struct file *filp, struct address_space *mapping, + struct page *pg; + struct inode *inode = mapping->host; + struct jffs2_inode_info *f = JFFS2_INODE_INFO(inode); ++ struct jffs2_sb_info *c = JFFS2_SB_INFO(inode->i_sb); + pgoff_t index = pos >> PAGE_SHIFT; + uint32_t pageofs = index << PAGE_SHIFT; + int ret = 0; + +- pg = grab_cache_page_write_begin(mapping, index, flags); +- if (!pg) +- return -ENOMEM; +- *pagep = pg; +- + jffs2_dbg(1, "%s()\n", __func__); + + if (pageofs > inode->i_size) { + /* Make new hole frag from old EOF to new page */ +- struct jffs2_sb_info *c = JFFS2_SB_INFO(inode->i_sb); + struct jffs2_raw_inode ri; + struct jffs2_full_dnode *fn; + uint32_t alloc_len; +@@ -160,7 +155,7 @@ static int jffs2_write_begin(struct file *filp, struct address_space *mapping, + ret = jffs2_reserve_space(c, sizeof(ri), &alloc_len, + ALLOC_NORMAL, JFFS2_SUMMARY_INODE_SIZE); + if (ret) +- goto out_page; ++ goto out_err; + + mutex_lock(&f->sem); + memset(&ri, 0, sizeof(ri)); +@@ -190,7 +185,7 @@ static int jffs2_write_begin(struct file *filp, struct address_space *mapping, + ret = PTR_ERR(fn); + jffs2_complete_reservation(c); + mutex_unlock(&f->sem); +- goto out_page; ++ goto out_err; + } + ret = jffs2_add_full_dnode_to_inode(c, f, fn); + if (f->metadata) { +@@ -205,13 +200,26 @@ static int jffs2_write_begin(struct file *filp, struct address_space *mapping, + jffs2_free_full_dnode(fn); + jffs2_complete_reservation(c); + mutex_unlock(&f->sem); +- goto out_page; ++ goto out_err; + } + jffs2_complete_reservation(c); + inode->i_size = pageofs; + mutex_unlock(&f->sem); + } + ++ /* ++ * While getting a page and reading data in, lock c->alloc_sem until ++ * the page is Uptodate. Otherwise GC task may attempt to read the same ++ * page in read_cache_page(), which causes a deadlock. ++ */ ++ mutex_lock(&c->alloc_sem); ++ pg = grab_cache_page_write_begin(mapping, index, flags); ++ if (!pg) { ++ ret = -ENOMEM; ++ goto release_sem; ++ } ++ *pagep = pg; ++ + /* + * Read in the page if it wasn't already present. Cannot optimize away + * the whole page write case until jffs2_write_end can handle the +@@ -221,15 +229,17 @@ static int jffs2_write_begin(struct file *filp, struct address_space *mapping, + mutex_lock(&f->sem); + ret = jffs2_do_readpage_nolock(inode, pg); + mutex_unlock(&f->sem); +- if (ret) +- goto out_page; ++ if (ret) { ++ unlock_page(pg); ++ put_page(pg); ++ goto release_sem; ++ } + } + jffs2_dbg(1, "end write_begin(). pg->flags %lx\n", pg->flags); +- return ret; + +-out_page: +- unlock_page(pg); +- put_page(pg); ++release_sem: ++ mutex_unlock(&c->alloc_sem); ++out_err: + return ret; + } + +diff --git a/fs/ksmbd/connection.c b/fs/ksmbd/connection.c +index b57a0d8a392ff..f7d5e8b7bef77 100644 +--- a/fs/ksmbd/connection.c ++++ b/fs/ksmbd/connection.c +@@ -62,6 +62,7 @@ struct ksmbd_conn *ksmbd_conn_alloc(void) + atomic_set(&conn->req_running, 0); + atomic_set(&conn->r_count, 0); + conn->total_credits = 1; ++ conn->outstanding_credits = 1; + + init_waitqueue_head(&conn->req_running_q); + INIT_LIST_HEAD(&conn->conns_list); +diff --git a/fs/ksmbd/connection.h b/fs/ksmbd/connection.h +index e5403c587a58c..8694aef482c1a 100644 +--- a/fs/ksmbd/connection.h ++++ b/fs/ksmbd/connection.h +@@ -61,8 +61,8 @@ struct ksmbd_conn { + atomic_t req_running; + /* References which are made for this Server object*/ + atomic_t r_count; +- unsigned short total_credits; +- unsigned short max_credits; ++ unsigned int total_credits; ++ unsigned int outstanding_credits; + spinlock_t credits_lock; + wait_queue_head_t req_running_q; + /* Lock to protect requests list*/ +diff --git a/fs/ksmbd/ksmbd_netlink.h b/fs/ksmbd/ksmbd_netlink.h +index c6718a05d347f..71bfb7de44725 100644 +--- a/fs/ksmbd/ksmbd_netlink.h ++++ b/fs/ksmbd/ksmbd_netlink.h +@@ -103,6 +103,8 @@ struct ksmbd_startup_request { + * we set the SPARSE_FILES bit (0x40). + */ + __u32 sub_auth[3]; /* Subauth value for Security ID */ ++ __u32 smb2_max_credits; /* MAX credits */ ++ __u32 reserved[128]; /* Reserved room */ + __u32 ifc_list_sz; /* interfaces list size */ + __s8 ____payload[]; + }; +@@ -113,7 +115,7 @@ struct ksmbd_startup_request { + * IPC request to shutdown ksmbd server. + */ + struct ksmbd_shutdown_request { +- __s32 reserved; ++ __s32 reserved[16]; + }; + + /* +@@ -122,6 +124,7 @@ struct ksmbd_shutdown_request { + struct ksmbd_login_request { + __u32 handle; + __s8 account[KSMBD_REQ_MAX_ACCOUNT_NAME_SZ]; /* user account name */ ++ __u32 reserved[16]; /* Reserved room */ + }; + + /* +@@ -135,6 +138,7 @@ struct ksmbd_login_response { + __u16 status; + __u16 hash_sz; /* hash size */ + __s8 hash[KSMBD_REQ_MAX_HASH_SZ]; /* password hash */ ++ __u32 reserved[16]; /* Reserved room */ + }; + + /* +@@ -143,6 +147,7 @@ struct ksmbd_login_response { + struct ksmbd_share_config_request { + __u32 handle; + __s8 share_name[KSMBD_REQ_MAX_SHARE_NAME]; /* share name */ ++ __u32 reserved[16]; /* Reserved room */ + }; + + /* +@@ -157,6 +162,7 @@ struct ksmbd_share_config_response { + __u16 force_directory_mode; + __u16 force_uid; + __u16 force_gid; ++ __u32 reserved[128]; /* Reserved room */ + __u32 veto_list_sz; + __s8 ____payload[]; + }; +@@ -187,6 +193,7 @@ struct ksmbd_tree_connect_request { + __s8 account[KSMBD_REQ_MAX_ACCOUNT_NAME_SZ]; + __s8 share[KSMBD_REQ_MAX_SHARE_NAME]; + __s8 peer_addr[64]; ++ __u32 reserved[16]; /* Reserved room */ + }; + + /* +@@ -196,6 +203,7 @@ struct ksmbd_tree_connect_response { + __u32 handle; + __u16 status; + __u16 connection_flags; ++ __u32 reserved[16]; /* Reserved room */ + }; + + /* +@@ -204,6 +212,7 @@ struct ksmbd_tree_connect_response { + struct ksmbd_tree_disconnect_request { + __u64 session_id; /* session id */ + __u64 connect_id; /* tree connection id */ ++ __u32 reserved[16]; /* Reserved room */ + }; + + /* +@@ -212,6 +221,7 @@ struct ksmbd_tree_disconnect_request { + struct ksmbd_logout_request { + __s8 account[KSMBD_REQ_MAX_ACCOUNT_NAME_SZ]; /* user account name */ + __u32 account_flags; ++ __u32 reserved[16]; /* Reserved room */ + }; + + /* +diff --git a/fs/ksmbd/smb2misc.c b/fs/ksmbd/smb2misc.c +index 9f516f73bd1b6..cc1c38686ecd7 100644 +--- a/fs/ksmbd/smb2misc.c ++++ b/fs/ksmbd/smb2misc.c +@@ -290,7 +290,7 @@ static int smb2_validate_credit_charge(struct ksmbd_conn *conn, + unsigned int req_len = 0, expect_resp_len = 0, calc_credit_num, max_len; + unsigned short credit_charge = le16_to_cpu(hdr->CreditCharge); + void *__hdr = hdr; +- int ret; ++ int ret = 0; + + switch (hdr->Command) { + case SMB2_QUERY_INFO: +@@ -327,21 +327,27 @@ static int smb2_validate_credit_charge(struct ksmbd_conn *conn, + ksmbd_debug(SMB, "Insufficient credit charge, given: %d, needed: %d\n", + credit_charge, calc_credit_num); + return 1; +- } else if (credit_charge > conn->max_credits) { ++ } else if (credit_charge > conn->vals->max_credits) { + ksmbd_debug(SMB, "Too large credit charge: %d\n", credit_charge); + return 1; + } + + spin_lock(&conn->credits_lock); +- if (credit_charge <= conn->total_credits) { +- conn->total_credits -= credit_charge; +- ret = 0; +- } else { ++ if (credit_charge > conn->total_credits) { + ksmbd_debug(SMB, "Insufficient credits granted, given: %u, granted: %u\n", + credit_charge, conn->total_credits); + ret = 1; + } ++ ++ if ((u64)conn->outstanding_credits + credit_charge > conn->vals->max_credits) { ++ ksmbd_debug(SMB, "Limits exceeding the maximum allowable outstanding requests, given : %u, pending : %u\n", ++ credit_charge, conn->outstanding_credits); ++ ret = 1; ++ } else ++ conn->outstanding_credits += credit_charge; ++ + spin_unlock(&conn->credits_lock); ++ + return ret; + } + +diff --git a/fs/ksmbd/smb2ops.c b/fs/ksmbd/smb2ops.c +index 2a6205103df2c..f0a5b704f301c 100644 +--- a/fs/ksmbd/smb2ops.c ++++ b/fs/ksmbd/smb2ops.c +@@ -20,6 +20,7 @@ static struct smb_version_values smb21_server_values = { + .max_read_size = SMB21_DEFAULT_IOSIZE, + .max_write_size = SMB21_DEFAULT_IOSIZE, + .max_trans_size = SMB21_DEFAULT_IOSIZE, ++ .max_credits = SMB2_MAX_CREDITS, + .large_lock_type = 0, + .exclusive_lock_type = SMB2_LOCKFLAG_EXCLUSIVE, + .shared_lock_type = SMB2_LOCKFLAG_SHARED, +@@ -45,6 +46,7 @@ static struct smb_version_values smb30_server_values = { + .max_read_size = SMB3_DEFAULT_IOSIZE, + .max_write_size = SMB3_DEFAULT_IOSIZE, + .max_trans_size = SMB3_DEFAULT_TRANS_SIZE, ++ .max_credits = SMB2_MAX_CREDITS, + .large_lock_type = 0, + .exclusive_lock_type = SMB2_LOCKFLAG_EXCLUSIVE, + .shared_lock_type = SMB2_LOCKFLAG_SHARED, +@@ -71,6 +73,7 @@ static struct smb_version_values smb302_server_values = { + .max_read_size = SMB3_DEFAULT_IOSIZE, + .max_write_size = SMB3_DEFAULT_IOSIZE, + .max_trans_size = SMB3_DEFAULT_TRANS_SIZE, ++ .max_credits = SMB2_MAX_CREDITS, + .large_lock_type = 0, + .exclusive_lock_type = SMB2_LOCKFLAG_EXCLUSIVE, + .shared_lock_type = SMB2_LOCKFLAG_SHARED, +@@ -97,6 +100,7 @@ static struct smb_version_values smb311_server_values = { + .max_read_size = SMB3_DEFAULT_IOSIZE, + .max_write_size = SMB3_DEFAULT_IOSIZE, + .max_trans_size = SMB3_DEFAULT_TRANS_SIZE, ++ .max_credits = SMB2_MAX_CREDITS, + .large_lock_type = 0, + .exclusive_lock_type = SMB2_LOCKFLAG_EXCLUSIVE, + .shared_lock_type = SMB2_LOCKFLAG_SHARED, +@@ -198,7 +202,6 @@ void init_smb2_1_server(struct ksmbd_conn *conn) + conn->ops = &smb2_0_server_ops; + conn->cmds = smb2_0_server_cmds; + conn->max_cmds = ARRAY_SIZE(smb2_0_server_cmds); +- conn->max_credits = SMB2_MAX_CREDITS; + conn->signing_algorithm = SIGNING_ALG_HMAC_SHA256; + + if (server_conf.flags & KSMBD_GLOBAL_FLAG_SMB2_LEASES) +@@ -216,7 +219,6 @@ void init_smb3_0_server(struct ksmbd_conn *conn) + conn->ops = &smb3_0_server_ops; + conn->cmds = smb2_0_server_cmds; + conn->max_cmds = ARRAY_SIZE(smb2_0_server_cmds); +- conn->max_credits = SMB2_MAX_CREDITS; + conn->signing_algorithm = SIGNING_ALG_AES_CMAC; + + if (server_conf.flags & KSMBD_GLOBAL_FLAG_SMB2_LEASES) +@@ -241,7 +243,6 @@ void init_smb3_02_server(struct ksmbd_conn *conn) + conn->ops = &smb3_0_server_ops; + conn->cmds = smb2_0_server_cmds; + conn->max_cmds = ARRAY_SIZE(smb2_0_server_cmds); +- conn->max_credits = SMB2_MAX_CREDITS; + conn->signing_algorithm = SIGNING_ALG_AES_CMAC; + + if (server_conf.flags & KSMBD_GLOBAL_FLAG_SMB2_LEASES) +@@ -266,7 +267,6 @@ int init_smb3_11_server(struct ksmbd_conn *conn) + conn->ops = &smb3_11_server_ops; + conn->cmds = smb2_0_server_cmds; + conn->max_cmds = ARRAY_SIZE(smb2_0_server_cmds); +- conn->max_credits = SMB2_MAX_CREDITS; + conn->signing_algorithm = SIGNING_ALG_AES_CMAC; + + if (server_conf.flags & KSMBD_GLOBAL_FLAG_SMB2_LEASES) +@@ -305,3 +305,11 @@ void init_smb2_max_trans_size(unsigned int sz) + smb302_server_values.max_trans_size = sz; + smb311_server_values.max_trans_size = sz; + } ++ ++void init_smb2_max_credits(unsigned int sz) ++{ ++ smb21_server_values.max_credits = sz; ++ smb30_server_values.max_credits = sz; ++ smb302_server_values.max_credits = sz; ++ smb311_server_values.max_credits = sz; ++} +diff --git a/fs/ksmbd/smb2pdu.c b/fs/ksmbd/smb2pdu.c +index cb71cbfc5fc9b..f694ee10a0bc8 100644 +--- a/fs/ksmbd/smb2pdu.c ++++ b/fs/ksmbd/smb2pdu.c +@@ -301,16 +301,15 @@ int smb2_set_rsp_credits(struct ksmbd_work *work) + struct smb2_hdr *req_hdr = ksmbd_req_buf_next(work); + struct smb2_hdr *hdr = ksmbd_resp_buf_next(work); + struct ksmbd_conn *conn = work->conn; +- unsigned short credits_requested; ++ unsigned short credits_requested, aux_max; + unsigned short credit_charge, credits_granted = 0; +- unsigned short aux_max, aux_credits; + + if (work->send_no_response) + return 0; + + hdr->CreditCharge = req_hdr->CreditCharge; + +- if (conn->total_credits > conn->max_credits) { ++ if (conn->total_credits > conn->vals->max_credits) { + hdr->CreditRequest = 0; + pr_err("Total credits overflow: %d\n", conn->total_credits); + return -EINVAL; +@@ -318,6 +317,14 @@ int smb2_set_rsp_credits(struct ksmbd_work *work) + + credit_charge = max_t(unsigned short, + le16_to_cpu(req_hdr->CreditCharge), 1); ++ if (credit_charge > conn->total_credits) { ++ ksmbd_debug(SMB, "Insufficient credits granted, given: %u, granted: %u\n", ++ credit_charge, conn->total_credits); ++ return -EINVAL; ++ } ++ ++ conn->total_credits -= credit_charge; ++ conn->outstanding_credits -= credit_charge; + credits_requested = max_t(unsigned short, + le16_to_cpu(req_hdr->CreditRequest), 1); + +@@ -327,16 +334,14 @@ int smb2_set_rsp_credits(struct ksmbd_work *work) + * TODO: Need to adjuct CreditRequest value according to + * current cpu load + */ +- aux_credits = credits_requested - 1; + if (hdr->Command == SMB2_NEGOTIATE) +- aux_max = 0; ++ aux_max = 1; + else +- aux_max = conn->max_credits - credit_charge; +- aux_credits = min_t(unsigned short, aux_credits, aux_max); +- credits_granted = credit_charge + aux_credits; ++ aux_max = conn->vals->max_credits - credit_charge; ++ credits_granted = min_t(unsigned short, credits_requested, aux_max); + +- if (conn->max_credits - conn->total_credits < credits_granted) +- credits_granted = conn->max_credits - ++ if (conn->vals->max_credits - conn->total_credits < credits_granted) ++ credits_granted = conn->vals->max_credits - + conn->total_credits; + + conn->total_credits += credits_granted; +@@ -1457,11 +1462,6 @@ static int ntlm_authenticate(struct ksmbd_work *work) + + sess->user = user; + if (user_guest(sess->user)) { +- if (conn->sign) { +- ksmbd_debug(SMB, "Guest login not allowed when signing enabled\n"); +- return -EPERM; +- } +- + rsp->SessionFlags = SMB2_SESSION_FLAG_IS_GUEST_LE; + } else { + struct authenticate_message *authblob; +@@ -1474,38 +1474,39 @@ static int ntlm_authenticate(struct ksmbd_work *work) + ksmbd_debug(SMB, "authentication failed\n"); + return -EPERM; + } ++ } + +- /* +- * If session state is SMB2_SESSION_VALID, We can assume +- * that it is reauthentication. And the user/password +- * has been verified, so return it here. +- */ +- if (sess->state == SMB2_SESSION_VALID) { +- if (conn->binding) +- goto binding_session; +- return 0; +- } ++ /* ++ * If session state is SMB2_SESSION_VALID, We can assume ++ * that it is reauthentication. And the user/password ++ * has been verified, so return it here. ++ */ ++ if (sess->state == SMB2_SESSION_VALID) { ++ if (conn->binding) ++ goto binding_session; ++ return 0; ++ } + +- if ((conn->sign || server_conf.enforced_signing) || +- (req->SecurityMode & SMB2_NEGOTIATE_SIGNING_REQUIRED)) +- sess->sign = true; ++ if ((rsp->SessionFlags != SMB2_SESSION_FLAG_IS_GUEST_LE && ++ (conn->sign || server_conf.enforced_signing)) || ++ (req->SecurityMode & SMB2_NEGOTIATE_SIGNING_REQUIRED)) ++ sess->sign = true; + +- if (smb3_encryption_negotiated(conn) && +- !(req->Flags & SMB2_SESSION_REQ_FLAG_BINDING)) { +- rc = conn->ops->generate_encryptionkey(sess); +- if (rc) { +- ksmbd_debug(SMB, +- "SMB3 encryption key generation failed\n"); +- return -EINVAL; +- } +- sess->enc = true; +- rsp->SessionFlags = SMB2_SESSION_FLAG_ENCRYPT_DATA_LE; +- /* +- * signing is disable if encryption is enable +- * on this session +- */ +- sess->sign = false; ++ if (smb3_encryption_negotiated(conn) && ++ !(req->Flags & SMB2_SESSION_REQ_FLAG_BINDING)) { ++ rc = conn->ops->generate_encryptionkey(sess); ++ if (rc) { ++ ksmbd_debug(SMB, ++ "SMB3 encryption key generation failed\n"); ++ return -EINVAL; + } ++ sess->enc = true; ++ rsp->SessionFlags = SMB2_SESSION_FLAG_ENCRYPT_DATA_LE; ++ /* ++ * signing is disable if encryption is enable ++ * on this session ++ */ ++ sess->sign = false; + } + + binding_session: +diff --git a/fs/ksmbd/smb2pdu.h b/fs/ksmbd/smb2pdu.h +index ff5a2f01d34ae..4f8574944ac19 100644 +--- a/fs/ksmbd/smb2pdu.h ++++ b/fs/ksmbd/smb2pdu.h +@@ -1647,6 +1647,7 @@ int init_smb3_11_server(struct ksmbd_conn *conn); + void init_smb2_max_read_size(unsigned int sz); + void init_smb2_max_write_size(unsigned int sz); + void init_smb2_max_trans_size(unsigned int sz); ++void init_smb2_max_credits(unsigned int sz); + + bool is_smb2_neg_cmd(struct ksmbd_work *work); + bool is_smb2_rsp(struct ksmbd_work *work); +diff --git a/fs/ksmbd/smb_common.h b/fs/ksmbd/smb_common.h +index 6e79e7577f6b7..1eba8dabaf317 100644 +--- a/fs/ksmbd/smb_common.h ++++ b/fs/ksmbd/smb_common.h +@@ -412,6 +412,7 @@ struct smb_version_values { + __u32 max_read_size; + __u32 max_write_size; + __u32 max_trans_size; ++ __u32 max_credits; + __u32 large_lock_type; + __u32 exclusive_lock_type; + __u32 shared_lock_type; +diff --git a/fs/ksmbd/transport_ipc.c b/fs/ksmbd/transport_ipc.c +index 1acf1892a466c..3ad6881e0f7ed 100644 +--- a/fs/ksmbd/transport_ipc.c ++++ b/fs/ksmbd/transport_ipc.c +@@ -301,6 +301,8 @@ static int ipc_server_config_on_startup(struct ksmbd_startup_request *req) + init_smb2_max_write_size(req->smb2_max_write); + if (req->smb2_max_trans) + init_smb2_max_trans_size(req->smb2_max_trans); ++ if (req->smb2_max_credits) ++ init_smb2_max_credits(req->smb2_max_credits); + + ret = ksmbd_set_netbios_name(req->netbios_name); + ret |= ksmbd_set_server_string(req->server_string); +diff --git a/fs/ksmbd/transport_tcp.c b/fs/ksmbd/transport_tcp.c +index c14320e03b698..82a1429bbe127 100644 +--- a/fs/ksmbd/transport_tcp.c ++++ b/fs/ksmbd/transport_tcp.c +@@ -404,7 +404,7 @@ static int create_socket(struct interface *iface) + &ksmbd_socket); + if (ret) { + pr_err("Can't create socket for ipv4: %d\n", ret); +- goto out_error; ++ goto out_clear; + } + + sin.sin_family = PF_INET; +@@ -462,6 +462,7 @@ static int create_socket(struct interface *iface) + + out_error: + tcp_destroy_socket(ksmbd_socket); ++out_clear: + iface->ksmbd_socket = NULL; + return ret; + } +diff --git a/fs/ubifs/super.c b/fs/ubifs/super.c +index f0fb25727d961..eb05038b71911 100644 +--- a/fs/ubifs/super.c ++++ b/fs/ubifs/super.c +@@ -1853,7 +1853,6 @@ out: + kthread_stop(c->bgt); + c->bgt = NULL; + } +- free_wbufs(c); + kfree(c->write_reserve_buf); + c->write_reserve_buf = NULL; + vfree(c->ileb_buf); +diff --git a/fs/udf/ialloc.c b/fs/udf/ialloc.c +index 2ecf0e87660e3..b5d611cee749c 100644 +--- a/fs/udf/ialloc.c ++++ b/fs/udf/ialloc.c +@@ -77,6 +77,7 @@ struct inode *udf_new_inode(struct inode *dir, umode_t mode) + GFP_KERNEL); + } + if (!iinfo->i_data) { ++ make_bad_inode(inode); + iput(inode); + return ERR_PTR(-ENOMEM); + } +@@ -86,6 +87,7 @@ struct inode *udf_new_inode(struct inode *dir, umode_t mode) + dinfo->i_location.partitionReferenceNum, + start, &err); + if (err) { ++ make_bad_inode(inode); + iput(inode); + return ERR_PTR(err); + } +diff --git a/include/acpi/acpi_bus.h b/include/acpi/acpi_bus.h +index 13d93371790ec..e9c7d7b270e73 100644 +--- a/include/acpi/acpi_bus.h ++++ b/include/acpi/acpi_bus.h +@@ -613,9 +613,10 @@ int acpi_enable_wakeup_device_power(struct acpi_device *dev, int state); + int acpi_disable_wakeup_device_power(struct acpi_device *dev); + + #ifdef CONFIG_X86 +-bool acpi_device_always_present(struct acpi_device *adev); ++bool acpi_device_override_status(struct acpi_device *adev, unsigned long long *status); + #else +-static inline bool acpi_device_always_present(struct acpi_device *adev) ++static inline bool acpi_device_override_status(struct acpi_device *adev, ++ unsigned long long *status) + { + return false; + } +diff --git a/include/acpi/actypes.h b/include/acpi/actypes.h +index 92c71dfce0d5d..cefbb7ad253e0 100644 +--- a/include/acpi/actypes.h ++++ b/include/acpi/actypes.h +@@ -536,8 +536,14 @@ typedef u64 acpi_integer; + * Can be used with access_width of struct acpi_generic_address and access_size of + * struct acpi_resource_generic_register. + */ +-#define ACPI_ACCESS_BIT_WIDTH(size) (1 << ((size) + 2)) +-#define ACPI_ACCESS_BYTE_WIDTH(size) (1 << ((size) - 1)) ++#define ACPI_ACCESS_BIT_SHIFT 2 ++#define ACPI_ACCESS_BYTE_SHIFT -1 ++#define ACPI_ACCESS_BIT_MAX (31 - ACPI_ACCESS_BIT_SHIFT) ++#define ACPI_ACCESS_BYTE_MAX (31 - ACPI_ACCESS_BYTE_SHIFT) ++#define ACPI_ACCESS_BIT_DEFAULT (8 - ACPI_ACCESS_BIT_SHIFT) ++#define ACPI_ACCESS_BYTE_DEFAULT (8 - ACPI_ACCESS_BYTE_SHIFT) ++#define ACPI_ACCESS_BIT_WIDTH(size) (1 << ((size) + ACPI_ACCESS_BIT_SHIFT)) ++#define ACPI_ACCESS_BYTE_WIDTH(size) (1 << ((size) + ACPI_ACCESS_BYTE_SHIFT)) + + /******************************************************************************* + * +diff --git a/include/asm-generic/bitops/find.h b/include/asm-generic/bitops/find.h +index 0d132ee2a2913..835f959a25f25 100644 +--- a/include/asm-generic/bitops/find.h ++++ b/include/asm-generic/bitops/find.h +@@ -97,6 +97,7 @@ unsigned long find_next_zero_bit(const unsigned long *addr, unsigned long size, + + #ifdef CONFIG_GENERIC_FIND_FIRST_BIT + ++#ifndef find_first_bit + /** + * find_first_bit - find the first set bit in a memory region + * @addr: The address to start the search at +@@ -116,7 +117,9 @@ unsigned long find_first_bit(const unsigned long *addr, unsigned long size) + + return _find_first_bit(addr, size); + } ++#endif + ++#ifndef find_first_zero_bit + /** + * find_first_zero_bit - find the first cleared bit in a memory region + * @addr: The address to start the search at +@@ -136,6 +139,8 @@ unsigned long find_first_zero_bit(const unsigned long *addr, unsigned long size) + + return _find_first_zero_bit(addr, size); + } ++#endif ++ + #else /* CONFIG_GENERIC_FIND_FIRST_BIT */ + + #ifndef find_first_bit +diff --git a/include/linux/blk-pm.h b/include/linux/blk-pm.h +index b80c65aba2493..2580e05a8ab67 100644 +--- a/include/linux/blk-pm.h ++++ b/include/linux/blk-pm.h +@@ -14,7 +14,7 @@ extern void blk_pm_runtime_init(struct request_queue *q, struct device *dev); + extern int blk_pre_runtime_suspend(struct request_queue *q); + extern void blk_post_runtime_suspend(struct request_queue *q, int err); + extern void blk_pre_runtime_resume(struct request_queue *q); +-extern void blk_post_runtime_resume(struct request_queue *q, int err); ++extern void blk_post_runtime_resume(struct request_queue *q); + extern void blk_set_runtime_active(struct request_queue *q); + #else + static inline void blk_pm_runtime_init(struct request_queue *q, +diff --git a/include/linux/bpf_verifier.h b/include/linux/bpf_verifier.h +index 5424124dbe365..364550dd19c4a 100644 +--- a/include/linux/bpf_verifier.h ++++ b/include/linux/bpf_verifier.h +@@ -396,6 +396,13 @@ static inline bool bpf_verifier_log_needed(const struct bpf_verifier_log *log) + log->level == BPF_LOG_KERNEL); + } + ++static inline bool ++bpf_verifier_log_attr_valid(const struct bpf_verifier_log *log) ++{ ++ return log->len_total >= 128 && log->len_total <= UINT_MAX >> 2 && ++ log->level && log->ubuf && !(log->level & ~BPF_LOG_MASK); ++} ++ + #define BPF_MAX_SUBPROGS 256 + + struct bpf_subprog_info { +diff --git a/include/linux/hid.h b/include/linux/hid.h +index f453be385bd47..26742ca14609a 100644 +--- a/include/linux/hid.h ++++ b/include/linux/hid.h +@@ -349,6 +349,8 @@ struct hid_item { + /* BIT(9) reserved for backward compatibility, was NO_INIT_INPUT_REPORTS */ + #define HID_QUIRK_ALWAYS_POLL BIT(10) + #define HID_QUIRK_INPUT_PER_APP BIT(11) ++#define HID_QUIRK_X_INVERT BIT(12) ++#define HID_QUIRK_Y_INVERT BIT(13) + #define HID_QUIRK_SKIP_OUTPUT_REPORTS BIT(16) + #define HID_QUIRK_SKIP_OUTPUT_REPORT_ID BIT(17) + #define HID_QUIRK_NO_OUTPUT_REPORTS_ON_INTR_EP BIT(18) +diff --git a/include/linux/iio/trigger.h b/include/linux/iio/trigger.h +index 096f68dd2e0ca..4c69b144677b1 100644 +--- a/include/linux/iio/trigger.h ++++ b/include/linux/iio/trigger.h +@@ -55,6 +55,7 @@ struct iio_trigger_ops { + * @attached_own_device:[INTERN] if we are using our own device as trigger, + * i.e. if we registered a poll function to the same + * device as the one providing the trigger. ++ * @reenable_work: [INTERN] work item used to ensure reenable can sleep. + **/ + struct iio_trigger { + const struct iio_trigger_ops *ops; +@@ -74,6 +75,7 @@ struct iio_trigger { + unsigned long pool[BITS_TO_LONGS(CONFIG_IIO_CONSUMERS_PER_TRIGGER)]; + struct mutex pool_lock; + bool attached_own_device; ++ struct work_struct reenable_work; + }; + + +diff --git a/include/linux/ipv6.h b/include/linux/ipv6.h +index c383630d3f065..07cba0b3496d5 100644 +--- a/include/linux/ipv6.h ++++ b/include/linux/ipv6.h +@@ -132,6 +132,7 @@ struct inet6_skb_parm { + __u16 dsthao; + #endif + __u16 frag_max_size; ++ __u16 srhoff; + + #define IP6SKB_XFRM_TRANSFORMED 1 + #define IP6SKB_FORWARDED 2 +@@ -141,6 +142,7 @@ struct inet6_skb_parm { + #define IP6SKB_HOPBYHOP 32 + #define IP6SKB_L3SLAVE 64 + #define IP6SKB_JUMBOGRAM 128 ++#define IP6SKB_SEG6 256 + }; + + #if defined(CONFIG_NET_L3_MASTER_DEV) +diff --git a/include/linux/mmzone.h b/include/linux/mmzone.h +index 6a1d79d84675a..fa1ef98614bcb 100644 +--- a/include/linux/mmzone.h ++++ b/include/linux/mmzone.h +@@ -1031,6 +1031,15 @@ static inline int is_highmem_idx(enum zone_type idx) + #endif + } + ++#ifdef CONFIG_ZONE_DMA ++bool has_managed_dma(void); ++#else ++static inline bool has_managed_dma(void) ++{ ++ return false; ++} ++#endif ++ + /** + * is_highmem - helper function to quickly check if a struct zone is a + * highmem zone or not. This is an attempt to keep references +diff --git a/include/linux/mtd/rawnand.h b/include/linux/mtd/rawnand.h +index b2f9dd3cbd695..5b88cd51fadb5 100644 +--- a/include/linux/mtd/rawnand.h ++++ b/include/linux/mtd/rawnand.h +@@ -1539,6 +1539,8 @@ int nand_read_data_op(struct nand_chip *chip, void *buf, unsigned int len, + bool force_8bit, bool check_only); + int nand_write_data_op(struct nand_chip *chip, const void *buf, + unsigned int len, bool force_8bit); ++int nand_read_page_hwecc_oob_first(struct nand_chip *chip, uint8_t *buf, ++ int oob_required, int page); + + /* Scan and identify a NAND device */ + int nand_scan_with_ids(struct nand_chip *chip, unsigned int max_chips, +diff --git a/include/linux/of_fdt.h b/include/linux/of_fdt.h +index cf6a65b94d40e..6508b97dbf1d2 100644 +--- a/include/linux/of_fdt.h ++++ b/include/linux/of_fdt.h +@@ -62,6 +62,7 @@ extern int early_init_dt_scan_chosen(unsigned long node, const char *uname, + int depth, void *data); + extern int early_init_dt_scan_memory(unsigned long node, const char *uname, + int depth, void *data); ++extern void early_init_dt_check_for_usable_mem_range(void); + extern int early_init_dt_scan_chosen_stdout(void); + extern void early_init_fdt_scan_reserved_mem(void); + extern void early_init_fdt_reserve_self(void); +@@ -87,6 +88,7 @@ extern void unflatten_and_copy_device_tree(void); + extern void early_init_devtree(void *); + extern void early_get_first_memblock_info(void *, phys_addr_t *); + #else /* CONFIG_OF_EARLY_FLATTREE */ ++static inline void early_init_dt_check_for_usable_mem_range(void) {} + static inline int early_init_dt_scan_chosen_stdout(void) { return -ENODEV; } + static inline void early_init_fdt_scan_reserved_mem(void) {} + static inline void early_init_fdt_reserve_self(void) {} +diff --git a/include/linux/pm_runtime.h b/include/linux/pm_runtime.h +index eddd66d426caf..016de5776b6db 100644 +--- a/include/linux/pm_runtime.h ++++ b/include/linux/pm_runtime.h +@@ -58,6 +58,7 @@ extern void pm_runtime_get_suppliers(struct device *dev); + extern void pm_runtime_put_suppliers(struct device *dev); + extern void pm_runtime_new_link(struct device *dev); + extern void pm_runtime_drop_link(struct device_link *link); ++extern void pm_runtime_release_supplier(struct device_link *link, bool check_idle); + + extern int devm_pm_runtime_enable(struct device *dev); + +@@ -283,6 +284,8 @@ static inline void pm_runtime_get_suppliers(struct device *dev) {} + static inline void pm_runtime_put_suppliers(struct device *dev) {} + static inline void pm_runtime_new_link(struct device *dev) {} + static inline void pm_runtime_drop_link(struct device_link *link) {} ++static inline void pm_runtime_release_supplier(struct device_link *link, ++ bool check_idle) {} + + #endif /* !CONFIG_PM */ + +diff --git a/include/linux/psi_types.h b/include/linux/psi_types.h +index 0a23300d49af7..0819c82dba920 100644 +--- a/include/linux/psi_types.h ++++ b/include/linux/psi_types.h +@@ -21,7 +21,17 @@ enum psi_task_count { + * don't have to special case any state tracking for it. + */ + NR_ONCPU, +- NR_PSI_TASK_COUNTS = 4, ++ /* ++ * For IO and CPU stalls the presence of running/oncpu tasks ++ * in the domain means a partial rather than a full stall. ++ * For memory it's not so simple because of page reclaimers: ++ * they are running/oncpu while representing a stall. To tell ++ * whether a domain has productivity left or not, we need to ++ * distinguish between regular running (i.e. productive) ++ * threads and memstall ones. ++ */ ++ NR_MEMSTALL_RUNNING, ++ NR_PSI_TASK_COUNTS = 5, + }; + + /* Task state bitmasks */ +@@ -29,6 +39,7 @@ enum psi_task_count { + #define TSK_MEMSTALL (1 << NR_MEMSTALL) + #define TSK_RUNNING (1 << NR_RUNNING) + #define TSK_ONCPU (1 << NR_ONCPU) ++#define TSK_MEMSTALL_RUNNING (1 << NR_MEMSTALL_RUNNING) + + /* Resources that workloads could be stalled on */ + enum psi_res { +diff --git a/include/linux/ptp_clock_kernel.h b/include/linux/ptp_clock_kernel.h +index 2e5565067355b..554454cb86931 100644 +--- a/include/linux/ptp_clock_kernel.h ++++ b/include/linux/ptp_clock_kernel.h +@@ -351,15 +351,17 @@ int ptp_get_vclocks_index(int pclock_index, int **vclock_index); + * + * @hwtstamps: skb_shared_hwtstamps structure pointer + * @vclock_index: phc index of ptp vclock. ++ * ++ * Returns converted timestamp, or 0 on error. + */ +-void ptp_convert_timestamp(struct skb_shared_hwtstamps *hwtstamps, +- int vclock_index); ++ktime_t ptp_convert_timestamp(const struct skb_shared_hwtstamps *hwtstamps, ++ int vclock_index); + #else + static inline int ptp_get_vclocks_index(int pclock_index, int **vclock_index) + { return 0; } +-static inline void ptp_convert_timestamp(struct skb_shared_hwtstamps *hwtstamps, +- int vclock_index) +-{ } ++static inline ktime_t ptp_convert_timestamp(const struct skb_shared_hwtstamps *hwtstamps, ++ int vclock_index) ++{ return 0; } + + #endif + +diff --git a/include/linux/skbuff.h b/include/linux/skbuff.h +index b8c273af2910c..532f5d402f060 100644 +--- a/include/linux/skbuff.h ++++ b/include/linux/skbuff.h +@@ -286,7 +286,10 @@ struct nf_bridge_info { + struct tc_skb_ext { + __u32 chain; + __u16 mru; +- bool post_ct; ++ __u16 zone; ++ u8 post_ct:1; ++ u8 post_ct_snat:1; ++ u8 post_ct_dnat:1; + }; + #endif + +@@ -1370,7 +1373,7 @@ skb_flow_dissect_ct(const struct sk_buff *skb, + struct flow_dissector *flow_dissector, + void *target_container, + u16 *ctinfo_map, size_t mapsize, +- bool post_ct); ++ bool post_ct, u16 zone); + void + skb_flow_dissect_tunnel_info(const struct sk_buff *skb, + struct flow_dissector *flow_dissector, +diff --git a/include/linux/stmmac.h b/include/linux/stmmac.h +index a6f03b36fc4f7..1450397fc0bcd 100644 +--- a/include/linux/stmmac.h ++++ b/include/linux/stmmac.h +@@ -233,6 +233,7 @@ struct plat_stmmacenet_data { + int (*clks_config)(void *priv, bool enabled); + int (*crosststamp)(ktime_t *device, struct system_counterval_t *system, + void *ctx); ++ void (*dump_debug_regs)(void *priv); + void *bsp_priv; + struct clk *stmmac_clk; + struct clk *pclk; +diff --git a/include/media/cec.h b/include/media/cec.h +index 208c9613c07eb..77346f757036d 100644 +--- a/include/media/cec.h ++++ b/include/media/cec.h +@@ -26,13 +26,17 @@ + * @dev: cec device + * @cdev: cec character device + * @minor: device node minor number ++ * @lock: lock to serialize open/release and registration + * @registered: the device was correctly registered + * @unregistered: the device was unregistered ++ * @lock_fhs: lock to control access to @fhs + * @fhs: the list of open filehandles (cec_fh) +- * @lock: lock to control access to this structure + * + * This structure represents a cec-related device node. + * ++ * To add or remove filehandles from @fhs the @lock must be taken first, ++ * followed by @lock_fhs. It is safe to access @fhs if either lock is held. ++ * + * The @parent is a physical device. It must be set by core or device drivers + * before registering the node. + */ +@@ -43,10 +47,13 @@ struct cec_devnode { + + /* device info */ + int minor; ++ /* serialize open/release and registration */ ++ struct mutex lock; + bool registered; + bool unregistered; ++ /* protect access to fhs */ ++ struct mutex lock_fhs; + struct list_head fhs; +- struct mutex lock; + }; + + struct cec_adapter; +diff --git a/include/net/inet_frag.h b/include/net/inet_frag.h +index 48cc5795ceda6..63540be0fc34a 100644 +--- a/include/net/inet_frag.h ++++ b/include/net/inet_frag.h +@@ -117,8 +117,15 @@ int fqdir_init(struct fqdir **fqdirp, struct inet_frags *f, struct net *net); + + static inline void fqdir_pre_exit(struct fqdir *fqdir) + { +- fqdir->high_thresh = 0; /* prevent creation of new frags */ +- fqdir->dead = true; ++ /* Prevent creation of new frags. ++ * Pairs with READ_ONCE() in inet_frag_find(). ++ */ ++ WRITE_ONCE(fqdir->high_thresh, 0); ++ ++ /* Pairs with READ_ONCE() in inet_frag_kill(), ip_expire() ++ * and ip6frag_expire_frag_queue(). ++ */ ++ WRITE_ONCE(fqdir->dead, true); + } + void fqdir_exit(struct fqdir *fqdir); + +diff --git a/include/net/ipv6_frag.h b/include/net/ipv6_frag.h +index 851029ecff13c..0a4779175a523 100644 +--- a/include/net/ipv6_frag.h ++++ b/include/net/ipv6_frag.h +@@ -67,7 +67,8 @@ ip6frag_expire_frag_queue(struct net *net, struct frag_queue *fq) + struct sk_buff *head; + + rcu_read_lock(); +- if (fq->q.fqdir->dead) ++ /* Paired with the WRITE_ONCE() in fqdir_pre_exit(). */ ++ if (READ_ONCE(fq->q.fqdir->dead)) + goto out_rcu_unlock; + spin_lock(&fq->q.lock); + +diff --git a/include/net/pkt_sched.h b/include/net/pkt_sched.h +index 05f18e81f3e87..9e7b21c0b3a6d 100644 +--- a/include/net/pkt_sched.h ++++ b/include/net/pkt_sched.h +@@ -197,7 +197,10 @@ struct tc_skb_cb { + struct qdisc_skb_cb qdisc_cb; + + u16 mru; +- bool post_ct; ++ u8 post_ct:1; ++ u8 post_ct_snat:1; ++ u8 post_ct_dnat:1; ++ u16 zone; /* Only valid if post_ct = true */ + }; + + static inline struct tc_skb_cb *tc_skb_cb(const struct sk_buff *skb) +diff --git a/include/net/sch_generic.h b/include/net/sch_generic.h +index 6e7cd00333577..4121ffd0faf8f 100644 +--- a/include/net/sch_generic.h ++++ b/include/net/sch_generic.h +@@ -1260,6 +1260,7 @@ struct psched_ratecfg { + u64 rate_bytes_ps; /* bytes per second */ + u32 mult; + u16 overhead; ++ u16 mpu; + u8 linklayer; + u8 shift; + }; +@@ -1269,6 +1270,9 @@ static inline u64 psched_l2t_ns(const struct psched_ratecfg *r, + { + len += r->overhead; + ++ if (len < r->mpu) ++ len = r->mpu; ++ + if (unlikely(r->linklayer == TC_LINKLAYER_ATM)) + return ((u64)(DIV_ROUND_UP(len,48)*53) * r->mult) >> r->shift; + +@@ -1291,6 +1295,7 @@ static inline void psched_ratecfg_getrate(struct tc_ratespec *res, + res->rate = min_t(u64, r->rate_bytes_ps, ~0U); + + res->overhead = r->overhead; ++ res->mpu = r->mpu; + res->linklayer = (r->linklayer & TC_LINKLAYER_MASK); + } + +diff --git a/include/net/seg6.h b/include/net/seg6.h +index 9d19c15e8545c..af668f17b3988 100644 +--- a/include/net/seg6.h ++++ b/include/net/seg6.h +@@ -58,9 +58,30 @@ extern int seg6_local_init(void); + extern void seg6_local_exit(void); + + extern bool seg6_validate_srh(struct ipv6_sr_hdr *srh, int len, bool reduced); ++extern struct ipv6_sr_hdr *seg6_get_srh(struct sk_buff *skb, int flags); ++extern void seg6_icmp_srh(struct sk_buff *skb, struct inet6_skb_parm *opt); + extern int seg6_do_srh_encap(struct sk_buff *skb, struct ipv6_sr_hdr *osrh, + int proto); + extern int seg6_do_srh_inline(struct sk_buff *skb, struct ipv6_sr_hdr *osrh); + extern int seg6_lookup_nexthop(struct sk_buff *skb, struct in6_addr *nhaddr, + u32 tbl_id); ++ ++/* If the packet which invoked an ICMP error contains an SRH return ++ * the true destination address from within the SRH, otherwise use the ++ * destination address in the IP header. ++ */ ++static inline const struct in6_addr *seg6_get_daddr(struct sk_buff *skb, ++ struct inet6_skb_parm *opt) ++{ ++ struct ipv6_sr_hdr *srh; ++ ++ if (opt->flags & IP6SKB_SEG6) { ++ srh = (struct ipv6_sr_hdr *)(skb->data + opt->srhoff); ++ return &srh->segments[0]; ++ } ++ ++ return NULL; ++} ++ ++ + #endif +diff --git a/include/net/xfrm.h b/include/net/xfrm.h +index 2308210793a01..2b1ce8534993c 100644 +--- a/include/net/xfrm.h ++++ b/include/net/xfrm.h +@@ -200,6 +200,11 @@ struct xfrm_state { + struct xfrm_algo_aead *aead; + const char *geniv; + ++ /* mapping change rate limiting */ ++ __be16 new_mapping_sport; ++ u32 new_mapping; /* seconds */ ++ u32 mapping_maxage; /* seconds for input SA */ ++ + /* Data for encapsulator */ + struct xfrm_encap_tmpl *encap; + struct sock __rcu *encap_sk; +@@ -1162,7 +1167,7 @@ static inline int xfrm_route_forward(struct sk_buff *skb, unsigned short family) + { + struct net *net = dev_net(skb->dev); + +- if (xfrm_default_allow(net, XFRM_POLICY_FWD)) ++ if (xfrm_default_allow(net, XFRM_POLICY_OUT)) + return !net->xfrm.policy_count[XFRM_POLICY_OUT] || + (skb_dst(skb)->flags & DST_NOXFRM) || + __xfrm_route_forward(skb, family); +diff --git a/include/sound/hda_codec.h b/include/sound/hda_codec.h +index 0e45963bb767f..82d9daa178517 100644 +--- a/include/sound/hda_codec.h ++++ b/include/sound/hda_codec.h +@@ -8,7 +8,7 @@ + #ifndef __SOUND_HDA_CODEC_H + #define __SOUND_HDA_CODEC_H + +-#include <linux/kref.h> ++#include <linux/refcount.h> + #include <linux/mod_devicetable.h> + #include <sound/info.h> + #include <sound/control.h> +@@ -166,8 +166,8 @@ struct hda_pcm { + bool own_chmap; /* codec driver provides own channel maps */ + /* private: */ + struct hda_codec *codec; +- struct kref kref; + struct list_head list; ++ unsigned int disconnected:1; + }; + + /* codec information */ +@@ -187,6 +187,8 @@ struct hda_codec { + + /* PCM to create, set by patch_ops.build_pcms callback */ + struct list_head pcm_list_head; ++ refcount_t pcm_ref; ++ wait_queue_head_t remove_sleep; + + /* codec specific info */ + void *spec; +@@ -420,7 +422,7 @@ void snd_hda_codec_cleanup_for_unbind(struct hda_codec *codec); + + static inline void snd_hda_codec_pcm_get(struct hda_pcm *pcm) + { +- kref_get(&pcm->kref); ++ refcount_inc(&pcm->codec->pcm_ref); + } + void snd_hda_codec_pcm_put(struct hda_pcm *pcm); + +diff --git a/include/trace/events/cgroup.h b/include/trace/events/cgroup.h +index 7f42a3de59e6b..dd7d7c9efecdf 100644 +--- a/include/trace/events/cgroup.h ++++ b/include/trace/events/cgroup.h +@@ -59,8 +59,8 @@ DECLARE_EVENT_CLASS(cgroup, + + TP_STRUCT__entry( + __field( int, root ) +- __field( int, id ) + __field( int, level ) ++ __field( u64, id ) + __string( path, path ) + ), + +@@ -71,7 +71,7 @@ DECLARE_EVENT_CLASS(cgroup, + __assign_str(path, path); + ), + +- TP_printk("root=%d id=%d level=%d path=%s", ++ TP_printk("root=%d id=%llu level=%d path=%s", + __entry->root, __entry->id, __entry->level, __get_str(path)) + ); + +@@ -126,8 +126,8 @@ DECLARE_EVENT_CLASS(cgroup_migrate, + + TP_STRUCT__entry( + __field( int, dst_root ) +- __field( int, dst_id ) + __field( int, dst_level ) ++ __field( u64, dst_id ) + __field( int, pid ) + __string( dst_path, path ) + __string( comm, task->comm ) +@@ -142,7 +142,7 @@ DECLARE_EVENT_CLASS(cgroup_migrate, + __assign_str(comm, task->comm); + ), + +- TP_printk("dst_root=%d dst_id=%d dst_level=%d dst_path=%s pid=%d comm=%s", ++ TP_printk("dst_root=%d dst_id=%llu dst_level=%d dst_path=%s pid=%d comm=%s", + __entry->dst_root, __entry->dst_id, __entry->dst_level, + __get_str(dst_path), __entry->pid, __get_str(comm)) + ); +@@ -171,8 +171,8 @@ DECLARE_EVENT_CLASS(cgroup_event, + + TP_STRUCT__entry( + __field( int, root ) +- __field( int, id ) + __field( int, level ) ++ __field( u64, id ) + __string( path, path ) + __field( int, val ) + ), +@@ -185,7 +185,7 @@ DECLARE_EVENT_CLASS(cgroup_event, + __entry->val = val; + ), + +- TP_printk("root=%d id=%d level=%d path=%s val=%d", ++ TP_printk("root=%d id=%llu level=%d path=%s val=%d", + __entry->root, __entry->id, __entry->level, __get_str(path), + __entry->val) + ); +diff --git a/include/uapi/linux/xfrm.h b/include/uapi/linux/xfrm.h +index eda0426ec4c2b..4e29d78518902 100644 +--- a/include/uapi/linux/xfrm.h ++++ b/include/uapi/linux/xfrm.h +@@ -313,6 +313,7 @@ enum xfrm_attr_type_t { + XFRMA_SET_MARK, /* __u32 */ + XFRMA_SET_MARK_MASK, /* __u32 */ + XFRMA_IF_ID, /* __u32 */ ++ XFRMA_MTIMER_THRESH, /* __u32 in seconds for input SA */ + __XFRMA_MAX + + #define XFRMA_OUTPUT_MARK XFRMA_SET_MARK /* Compatibility */ +diff --git a/kernel/audit.c b/kernel/audit.c +index 4cebadb5f30db..eab7282668ab9 100644 +--- a/kernel/audit.c ++++ b/kernel/audit.c +@@ -1540,6 +1540,20 @@ static void audit_receive(struct sk_buff *skb) + nlh = nlmsg_next(nlh, &len); + } + audit_ctl_unlock(); ++ ++ /* can't block with the ctrl lock, so penalize the sender now */ ++ if (audit_backlog_limit && ++ (skb_queue_len(&audit_queue) > audit_backlog_limit)) { ++ DECLARE_WAITQUEUE(wait, current); ++ ++ /* wake kauditd to try and flush the queue */ ++ wake_up_interruptible(&kauditd_wait); ++ ++ add_wait_queue_exclusive(&audit_backlog_wait, &wait); ++ set_current_state(TASK_UNINTERRUPTIBLE); ++ schedule_timeout(audit_backlog_wait_time); ++ remove_wait_queue(&audit_backlog_wait, &wait); ++ } + } + + /* Log information about who is connecting to the audit multicast socket */ +@@ -1824,7 +1838,9 @@ struct audit_buffer *audit_log_start(struct audit_context *ctx, gfp_t gfp_mask, + * task_tgid_vnr() since auditd_pid is set in audit_receive_msg() + * using a PID anchored in the caller's namespace + * 2. generator holding the audit_cmd_mutex - we don't want to block +- * while holding the mutex */ ++ * while holding the mutex, although we do penalize the sender ++ * later in audit_receive() when it is safe to block ++ */ + if (!(auditd_test_task(current) || audit_ctl_owner_current())) { + long stime = audit_backlog_wait_time; + +diff --git a/kernel/bpf/btf.c b/kernel/bpf/btf.c +index dfe61df4f974d..b8ed4da63bc8c 100644 +--- a/kernel/bpf/btf.c ++++ b/kernel/bpf/btf.c +@@ -4332,8 +4332,7 @@ static struct btf *btf_parse(bpfptr_t btf_data, u32 btf_data_size, + log->len_total = log_size; + + /* log attributes have to be sane */ +- if (log->len_total < 128 || log->len_total > UINT_MAX >> 8 || +- !log->level || !log->ubuf) { ++ if (!bpf_verifier_log_attr_valid(log)) { + err = -EINVAL; + goto errout; + } +diff --git a/kernel/bpf/inode.c b/kernel/bpf/inode.c +index 80da1db47c686..5a8d9f7467bf4 100644 +--- a/kernel/bpf/inode.c ++++ b/kernel/bpf/inode.c +@@ -648,12 +648,22 @@ static int bpf_parse_param(struct fs_context *fc, struct fs_parameter *param) + int opt; + + opt = fs_parse(fc, bpf_fs_parameters, param, &result); +- if (opt < 0) ++ if (opt < 0) { + /* We might like to report bad mount options here, but + * traditionally we've ignored all mount options, so we'd + * better continue to ignore non-existing options for bpf. + */ +- return opt == -ENOPARAM ? 0 : opt; ++ if (opt == -ENOPARAM) { ++ opt = vfs_parse_fs_param_source(fc, param); ++ if (opt != -ENOPARAM) ++ return opt; ++ ++ return 0; ++ } ++ ++ if (opt < 0) ++ return opt; ++ } + + switch (opt) { + case OPT_MODE: +diff --git a/kernel/bpf/verifier.c b/kernel/bpf/verifier.c +index b84e63d62b8af..670721e39c0e8 100644 +--- a/kernel/bpf/verifier.c ++++ b/kernel/bpf/verifier.c +@@ -5785,6 +5785,7 @@ static int __check_func_call(struct bpf_verifier_env *env, struct bpf_insn *insn + } + + if (insn->code == (BPF_JMP | BPF_CALL) && ++ insn->src_reg == 0 && + insn->imm == BPF_FUNC_timer_set_callback) { + struct bpf_verifier_state *async_cb; + +@@ -8771,15 +8772,15 @@ static void mark_ptr_or_null_reg(struct bpf_func_state *state, + { + if (reg_type_may_be_null(reg->type) && reg->id == id && + !WARN_ON_ONCE(!reg->id)) { +- /* Old offset (both fixed and variable parts) should +- * have been known-zero, because we don't allow pointer +- * arithmetic on pointers that might be NULL. +- */ + if (WARN_ON_ONCE(reg->smin_value || reg->smax_value || + !tnum_equals_const(reg->var_off, 0) || + reg->off)) { +- __mark_reg_known_zero(reg); +- reg->off = 0; ++ /* Old offset (both fixed and variable parts) should ++ * have been known-zero, because we don't allow pointer ++ * arithmetic on pointers that might be NULL. If we ++ * see this happening, don't convert the register. ++ */ ++ return; + } + if (is_null) { + reg->type = SCALAR_VALUE; +@@ -9198,9 +9199,13 @@ static int check_ld_imm(struct bpf_verifier_env *env, struct bpf_insn *insn) + return 0; + } + +- if (insn->src_reg == BPF_PSEUDO_BTF_ID) { +- mark_reg_known_zero(env, regs, insn->dst_reg); ++ /* All special src_reg cases are listed below. From this point onwards ++ * we either succeed and assign a corresponding dst_reg->type after ++ * zeroing the offset, or fail and reject the program. ++ */ ++ mark_reg_known_zero(env, regs, insn->dst_reg); + ++ if (insn->src_reg == BPF_PSEUDO_BTF_ID) { + dst_reg->type = aux->btf_var.reg_type; + switch (dst_reg->type) { + case PTR_TO_MEM: +@@ -9237,7 +9242,6 @@ static int check_ld_imm(struct bpf_verifier_env *env, struct bpf_insn *insn) + } + + map = env->used_maps[aux->map_index]; +- mark_reg_known_zero(env, regs, insn->dst_reg); + dst_reg->map_ptr = map; + + if (insn->src_reg == BPF_PSEUDO_MAP_VALUE || +@@ -13759,11 +13763,11 @@ int bpf_check(struct bpf_prog **prog, union bpf_attr *attr, bpfptr_t uattr) + log->ubuf = (char __user *) (unsigned long) attr->log_buf; + log->len_total = attr->log_size; + +- ret = -EINVAL; + /* log attributes have to be sane */ +- if (log->len_total < 128 || log->len_total > UINT_MAX >> 2 || +- !log->level || !log->ubuf || log->level & ~BPF_LOG_MASK) ++ if (!bpf_verifier_log_attr_valid(log)) { ++ ret = -EINVAL; + goto err_unlock; ++ } + } + + if (IS_ERR(btf_vmlinux)) { +diff --git a/kernel/dma/pool.c b/kernel/dma/pool.c +index 5f84e6cdb78ea..4d40dcce7604b 100644 +--- a/kernel/dma/pool.c ++++ b/kernel/dma/pool.c +@@ -203,7 +203,7 @@ static int __init dma_atomic_pool_init(void) + GFP_KERNEL); + if (!atomic_pool_kernel) + ret = -ENOMEM; +- if (IS_ENABLED(CONFIG_ZONE_DMA)) { ++ if (has_managed_dma()) { + atomic_pool_dma = __dma_atomic_pool_init(atomic_pool_size, + GFP_KERNEL | GFP_DMA); + if (!atomic_pool_dma) +@@ -226,7 +226,7 @@ static inline struct gen_pool *dma_guess_pool(struct gen_pool *prev, gfp_t gfp) + if (prev == NULL) { + if (IS_ENABLED(CONFIG_ZONE_DMA32) && (gfp & GFP_DMA32)) + return atomic_pool_dma32; +- if (IS_ENABLED(CONFIG_ZONE_DMA) && (gfp & GFP_DMA)) ++ if (atomic_pool_dma && (gfp & GFP_DMA)) + return atomic_pool_dma; + return atomic_pool_kernel; + } +diff --git a/kernel/rcu/rcutorture.c b/kernel/rcu/rcutorture.c +index 968696ace8f3f..f922937eb39ad 100644 +--- a/kernel/rcu/rcutorture.c ++++ b/kernel/rcu/rcutorture.c +@@ -46,6 +46,7 @@ + #include <linux/oom.h> + #include <linux/tick.h> + #include <linux/rcupdate_trace.h> ++#include <linux/nmi.h> + + #include "rcu.h" + +@@ -109,6 +110,8 @@ torture_param(int, shutdown_secs, 0, "Shutdown time (s), <= zero to disable."); + torture_param(int, stall_cpu, 0, "Stall duration (s), zero to disable."); + torture_param(int, stall_cpu_holdoff, 10, + "Time to wait before starting stall (s)."); ++torture_param(bool, stall_no_softlockup, false, ++ "Avoid softlockup warning during cpu stall."); + torture_param(int, stall_cpu_irqsoff, 0, "Disable interrupts while stalling."); + torture_param(int, stall_cpu_block, 0, "Sleep while stalling."); + torture_param(int, stall_gp_kthread, 0, +@@ -2052,6 +2055,8 @@ static int rcu_torture_stall(void *args) + #else + schedule_timeout_uninterruptible(HZ); + #endif ++ } else if (stall_no_softlockup) { ++ touch_softlockup_watchdog(); + } + if (stall_cpu_irqsoff) + local_irq_enable(); +diff --git a/kernel/rcu/tree_exp.h b/kernel/rcu/tree_exp.h +index 454b516ea566e..16f94118ca34b 100644 +--- a/kernel/rcu/tree_exp.h ++++ b/kernel/rcu/tree_exp.h +@@ -387,6 +387,7 @@ retry_ipi: + continue; + } + if (get_cpu() == cpu) { ++ mask_ofl_test |= mask; + put_cpu(); + continue; + } +diff --git a/kernel/sched/cpuacct.c b/kernel/sched/cpuacct.c +index 893eece65bfda..ab67d97a84428 100644 +--- a/kernel/sched/cpuacct.c ++++ b/kernel/sched/cpuacct.c +@@ -21,15 +21,11 @@ static const char * const cpuacct_stat_desc[] = { + [CPUACCT_STAT_SYSTEM] = "system", + }; + +-struct cpuacct_usage { +- u64 usages[CPUACCT_STAT_NSTATS]; +-}; +- + /* track CPU usage of a group of tasks and its child groups */ + struct cpuacct { + struct cgroup_subsys_state css; + /* cpuusage holds pointer to a u64-type object on every CPU */ +- struct cpuacct_usage __percpu *cpuusage; ++ u64 __percpu *cpuusage; + struct kernel_cpustat __percpu *cpustat; + }; + +@@ -49,7 +45,7 @@ static inline struct cpuacct *parent_ca(struct cpuacct *ca) + return css_ca(ca->css.parent); + } + +-static DEFINE_PER_CPU(struct cpuacct_usage, root_cpuacct_cpuusage); ++static DEFINE_PER_CPU(u64, root_cpuacct_cpuusage); + static struct cpuacct root_cpuacct = { + .cpustat = &kernel_cpustat, + .cpuusage = &root_cpuacct_cpuusage, +@@ -68,7 +64,7 @@ cpuacct_css_alloc(struct cgroup_subsys_state *parent_css) + if (!ca) + goto out; + +- ca->cpuusage = alloc_percpu(struct cpuacct_usage); ++ ca->cpuusage = alloc_percpu(u64); + if (!ca->cpuusage) + goto out_free_ca; + +@@ -99,7 +95,8 @@ static void cpuacct_css_free(struct cgroup_subsys_state *css) + static u64 cpuacct_cpuusage_read(struct cpuacct *ca, int cpu, + enum cpuacct_stat_index index) + { +- struct cpuacct_usage *cpuusage = per_cpu_ptr(ca->cpuusage, cpu); ++ u64 *cpuusage = per_cpu_ptr(ca->cpuusage, cpu); ++ u64 *cpustat = per_cpu_ptr(ca->cpustat, cpu)->cpustat; + u64 data; + + /* +@@ -115,14 +112,17 @@ static u64 cpuacct_cpuusage_read(struct cpuacct *ca, int cpu, + raw_spin_rq_lock_irq(cpu_rq(cpu)); + #endif + +- if (index == CPUACCT_STAT_NSTATS) { +- int i = 0; +- +- data = 0; +- for (i = 0; i < CPUACCT_STAT_NSTATS; i++) +- data += cpuusage->usages[i]; +- } else { +- data = cpuusage->usages[index]; ++ switch (index) { ++ case CPUACCT_STAT_USER: ++ data = cpustat[CPUTIME_USER] + cpustat[CPUTIME_NICE]; ++ break; ++ case CPUACCT_STAT_SYSTEM: ++ data = cpustat[CPUTIME_SYSTEM] + cpustat[CPUTIME_IRQ] + ++ cpustat[CPUTIME_SOFTIRQ]; ++ break; ++ case CPUACCT_STAT_NSTATS: ++ data = *cpuusage; ++ break; + } + + #ifndef CONFIG_64BIT +@@ -132,10 +132,14 @@ static u64 cpuacct_cpuusage_read(struct cpuacct *ca, int cpu, + return data; + } + +-static void cpuacct_cpuusage_write(struct cpuacct *ca, int cpu, u64 val) ++static void cpuacct_cpuusage_write(struct cpuacct *ca, int cpu) + { +- struct cpuacct_usage *cpuusage = per_cpu_ptr(ca->cpuusage, cpu); +- int i; ++ u64 *cpuusage = per_cpu_ptr(ca->cpuusage, cpu); ++ u64 *cpustat = per_cpu_ptr(ca->cpustat, cpu)->cpustat; ++ ++ /* Don't allow to reset global kernel_cpustat */ ++ if (ca == &root_cpuacct) ++ return; + + #ifndef CONFIG_64BIT + /* +@@ -143,9 +147,10 @@ static void cpuacct_cpuusage_write(struct cpuacct *ca, int cpu, u64 val) + */ + raw_spin_rq_lock_irq(cpu_rq(cpu)); + #endif +- +- for (i = 0; i < CPUACCT_STAT_NSTATS; i++) +- cpuusage->usages[i] = val; ++ *cpuusage = 0; ++ cpustat[CPUTIME_USER] = cpustat[CPUTIME_NICE] = 0; ++ cpustat[CPUTIME_SYSTEM] = cpustat[CPUTIME_IRQ] = 0; ++ cpustat[CPUTIME_SOFTIRQ] = 0; + + #ifndef CONFIG_64BIT + raw_spin_rq_unlock_irq(cpu_rq(cpu)); +@@ -196,7 +201,7 @@ static int cpuusage_write(struct cgroup_subsys_state *css, struct cftype *cft, + return -EINVAL; + + for_each_possible_cpu(cpu) +- cpuacct_cpuusage_write(ca, cpu, 0); ++ cpuacct_cpuusage_write(ca, cpu); + + return 0; + } +@@ -243,25 +248,10 @@ static int cpuacct_all_seq_show(struct seq_file *m, void *V) + seq_puts(m, "\n"); + + for_each_possible_cpu(cpu) { +- struct cpuacct_usage *cpuusage = per_cpu_ptr(ca->cpuusage, cpu); +- + seq_printf(m, "%d", cpu); +- +- for (index = 0; index < CPUACCT_STAT_NSTATS; index++) { +-#ifndef CONFIG_64BIT +- /* +- * Take rq->lock to make 64-bit read safe on 32-bit +- * platforms. +- */ +- raw_spin_rq_lock_irq(cpu_rq(cpu)); +-#endif +- +- seq_printf(m, " %llu", cpuusage->usages[index]); +- +-#ifndef CONFIG_64BIT +- raw_spin_rq_unlock_irq(cpu_rq(cpu)); +-#endif +- } ++ for (index = 0; index < CPUACCT_STAT_NSTATS; index++) ++ seq_printf(m, " %llu", ++ cpuacct_cpuusage_read(ca, cpu, index)); + seq_puts(m, "\n"); + } + return 0; +@@ -339,16 +329,11 @@ static struct cftype files[] = { + void cpuacct_charge(struct task_struct *tsk, u64 cputime) + { + struct cpuacct *ca; +- int index = CPUACCT_STAT_SYSTEM; +- struct pt_regs *regs = get_irq_regs() ? : task_pt_regs(tsk); +- +- if (regs && user_mode(regs)) +- index = CPUACCT_STAT_USER; + + rcu_read_lock(); + + for (ca = task_ca(tsk); ca; ca = parent_ca(ca)) +- __this_cpu_add(ca->cpuusage->usages[index], cputime); ++ __this_cpu_add(*ca->cpuusage, cputime); + + rcu_read_unlock(); + } +diff --git a/kernel/sched/cputime.c b/kernel/sched/cputime.c +index 872e481d5098c..042a6dbce8f32 100644 +--- a/kernel/sched/cputime.c ++++ b/kernel/sched/cputime.c +@@ -148,10 +148,10 @@ void account_guest_time(struct task_struct *p, u64 cputime) + + /* Add guest time to cpustat. */ + if (task_nice(p) > 0) { +- cpustat[CPUTIME_NICE] += cputime; ++ task_group_account_field(p, CPUTIME_NICE, cputime); + cpustat[CPUTIME_GUEST_NICE] += cputime; + } else { +- cpustat[CPUTIME_USER] += cputime; ++ task_group_account_field(p, CPUTIME_USER, cputime); + cpustat[CPUTIME_GUEST] += cputime; + } + } +diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c +index 6f16dfb742462..d41f966f5866a 100644 +--- a/kernel/sched/fair.c ++++ b/kernel/sched/fair.c +@@ -6429,8 +6429,10 @@ static int select_idle_sibling(struct task_struct *p, int prev, int target) + * pattern is IO completions. + */ + if (is_per_cpu_kthread(current) && ++ in_task() && + prev == smp_processor_id() && +- this_rq()->nr_running <= 1) { ++ this_rq()->nr_running <= 1 && ++ asym_fits_capacity(task_util, prev)) { + return prev; + } + +diff --git a/kernel/sched/psi.c b/kernel/sched/psi.c +index 1652f2bb54b79..69b19d3af690f 100644 +--- a/kernel/sched/psi.c ++++ b/kernel/sched/psi.c +@@ -34,13 +34,19 @@ + * delayed on that resource such that nobody is advancing and the CPU + * goes idle. This leaves both workload and CPU unproductive. + * +- * Naturally, the FULL state doesn't exist for the CPU resource at the +- * system level, but exist at the cgroup level, means all non-idle tasks +- * in a cgroup are delayed on the CPU resource which used by others outside +- * of the cgroup or throttled by the cgroup cpu.max configuration. +- * + * SOME = nr_delayed_tasks != 0 +- * FULL = nr_delayed_tasks != 0 && nr_running_tasks == 0 ++ * FULL = nr_delayed_tasks != 0 && nr_productive_tasks == 0 ++ * ++ * What it means for a task to be productive is defined differently ++ * for each resource. For IO, productive means a running task. For ++ * memory, productive means a running task that isn't a reclaimer. For ++ * CPU, productive means an oncpu task. ++ * ++ * Naturally, the FULL state doesn't exist for the CPU resource at the ++ * system level, but exist at the cgroup level. At the cgroup level, ++ * FULL means all non-idle tasks in the cgroup are delayed on the CPU ++ * resource which is being used by others outside of the cgroup or ++ * throttled by the cgroup cpu.max configuration. + * + * The percentage of wallclock time spent in those compound stall + * states gives pressure numbers between 0 and 100 for each resource, +@@ -81,13 +87,13 @@ + * + * threads = min(nr_nonidle_tasks, nr_cpus) + * SOME = min(nr_delayed_tasks / threads, 1) +- * FULL = (threads - min(nr_running_tasks, threads)) / threads ++ * FULL = (threads - min(nr_productive_tasks, threads)) / threads + * + * For the 257 number crunchers on 256 CPUs, this yields: + * + * threads = min(257, 256) + * SOME = min(1 / 256, 1) = 0.4% +- * FULL = (256 - min(257, 256)) / 256 = 0% ++ * FULL = (256 - min(256, 256)) / 256 = 0% + * + * For the 1 out of 4 memory-delayed tasks, this yields: + * +@@ -112,7 +118,7 @@ + * For each runqueue, we track: + * + * tSOME[cpu] = time(nr_delayed_tasks[cpu] != 0) +- * tFULL[cpu] = time(nr_delayed_tasks[cpu] && !nr_running_tasks[cpu]) ++ * tFULL[cpu] = time(nr_delayed_tasks[cpu] && !nr_productive_tasks[cpu]) + * tNONIDLE[cpu] = time(nr_nonidle_tasks[cpu] != 0) + * + * and then periodically aggregate: +@@ -233,7 +239,8 @@ static bool test_state(unsigned int *tasks, enum psi_states state) + case PSI_MEM_SOME: + return unlikely(tasks[NR_MEMSTALL]); + case PSI_MEM_FULL: +- return unlikely(tasks[NR_MEMSTALL] && !tasks[NR_RUNNING]); ++ return unlikely(tasks[NR_MEMSTALL] && ++ tasks[NR_RUNNING] == tasks[NR_MEMSTALL_RUNNING]); + case PSI_CPU_SOME: + return unlikely(tasks[NR_RUNNING] > tasks[NR_ONCPU]); + case PSI_CPU_FULL: +@@ -710,10 +717,11 @@ static void psi_group_change(struct psi_group *group, int cpu, + if (groupc->tasks[t]) { + groupc->tasks[t]--; + } else if (!psi_bug) { +- printk_deferred(KERN_ERR "psi: task underflow! cpu=%d t=%d tasks=[%u %u %u %u] clear=%x set=%x\n", ++ printk_deferred(KERN_ERR "psi: task underflow! cpu=%d t=%d tasks=[%u %u %u %u %u] clear=%x set=%x\n", + cpu, t, groupc->tasks[0], + groupc->tasks[1], groupc->tasks[2], +- groupc->tasks[3], clear, set); ++ groupc->tasks[3], groupc->tasks[4], ++ clear, set); + psi_bug = 1; + } + } +@@ -854,12 +862,15 @@ void psi_task_switch(struct task_struct *prev, struct task_struct *next, + int clear = TSK_ONCPU, set = 0; + + /* +- * When we're going to sleep, psi_dequeue() lets us handle +- * TSK_RUNNING and TSK_IOWAIT here, where we can combine it +- * with TSK_ONCPU and save walking common ancestors twice. ++ * When we're going to sleep, psi_dequeue() lets us ++ * handle TSK_RUNNING, TSK_MEMSTALL_RUNNING and ++ * TSK_IOWAIT here, where we can combine it with ++ * TSK_ONCPU and save walking common ancestors twice. + */ + if (sleep) { + clear |= TSK_RUNNING; ++ if (prev->in_memstall) ++ clear |= TSK_MEMSTALL_RUNNING; + if (prev->in_iowait) + set |= TSK_IOWAIT; + } +@@ -908,7 +919,7 @@ void psi_memstall_enter(unsigned long *flags) + rq = this_rq_lock_irq(&rf); + + current->in_memstall = 1; +- psi_task_change(current, 0, TSK_MEMSTALL); ++ psi_task_change(current, 0, TSK_MEMSTALL | TSK_MEMSTALL_RUNNING); + + rq_unlock_irq(rq, &rf); + } +@@ -937,7 +948,7 @@ void psi_memstall_leave(unsigned long *flags) + rq = this_rq_lock_irq(&rf); + + current->in_memstall = 0; +- psi_task_change(current, TSK_MEMSTALL, 0); ++ psi_task_change(current, TSK_MEMSTALL | TSK_MEMSTALL_RUNNING, 0); + + rq_unlock_irq(rq, &rf); + } +diff --git a/kernel/sched/rt.c b/kernel/sched/rt.c +index bfef3f39b5552..54f9bb3f15605 100644 +--- a/kernel/sched/rt.c ++++ b/kernel/sched/rt.c +@@ -52,11 +52,8 @@ void init_rt_bandwidth(struct rt_bandwidth *rt_b, u64 period, u64 runtime) + rt_b->rt_period_timer.function = sched_rt_period_timer; + } + +-static void start_rt_bandwidth(struct rt_bandwidth *rt_b) ++static inline void do_start_rt_bandwidth(struct rt_bandwidth *rt_b) + { +- if (!rt_bandwidth_enabled() || rt_b->rt_runtime == RUNTIME_INF) +- return; +- + raw_spin_lock(&rt_b->rt_runtime_lock); + if (!rt_b->rt_period_active) { + rt_b->rt_period_active = 1; +@@ -75,6 +72,14 @@ static void start_rt_bandwidth(struct rt_bandwidth *rt_b) + raw_spin_unlock(&rt_b->rt_runtime_lock); + } + ++static void start_rt_bandwidth(struct rt_bandwidth *rt_b) ++{ ++ if (!rt_bandwidth_enabled() || rt_b->rt_runtime == RUNTIME_INF) ++ return; ++ ++ do_start_rt_bandwidth(rt_b); ++} ++ + void init_rt_rq(struct rt_rq *rt_rq) + { + struct rt_prio_array *array; +@@ -1029,13 +1034,17 @@ static void update_curr_rt(struct rq *rq) + + for_each_sched_rt_entity(rt_se) { + struct rt_rq *rt_rq = rt_rq_of_se(rt_se); ++ int exceeded; + + if (sched_rt_runtime(rt_rq) != RUNTIME_INF) { + raw_spin_lock(&rt_rq->rt_runtime_lock); + rt_rq->rt_time += delta_exec; +- if (sched_rt_runtime_exceeded(rt_rq)) ++ exceeded = sched_rt_runtime_exceeded(rt_rq); ++ if (exceeded) + resched_curr(rq); + raw_spin_unlock(&rt_rq->rt_runtime_lock); ++ if (exceeded) ++ do_start_rt_bandwidth(sched_rt_bandwidth(rt_rq)); + } + } + } +@@ -2785,8 +2794,12 @@ static int sched_rt_global_validate(void) + + static void sched_rt_do_global(void) + { ++ unsigned long flags; ++ ++ raw_spin_lock_irqsave(&def_rt_bandwidth.rt_runtime_lock, flags); + def_rt_bandwidth.rt_runtime = global_rt_runtime(); + def_rt_bandwidth.rt_period = ns_to_ktime(global_rt_period()); ++ raw_spin_unlock_irqrestore(&def_rt_bandwidth.rt_runtime_lock, flags); + } + + int sched_rt_handler(struct ctl_table *table, int write, void *buffer, +diff --git a/kernel/sched/stats.h b/kernel/sched/stats.h +index d8f8eb0c655ba..606a3982d13a5 100644 +--- a/kernel/sched/stats.h ++++ b/kernel/sched/stats.h +@@ -69,6 +69,9 @@ static inline void psi_enqueue(struct task_struct *p, bool wakeup) + if (static_branch_likely(&psi_disabled)) + return; + ++ if (p->in_memstall) ++ set |= TSK_MEMSTALL_RUNNING; ++ + if (!wakeup || p->sched_psi_wake_requeue) { + if (p->in_memstall) + set |= TSK_MEMSTALL; +@@ -99,7 +102,7 @@ static inline void psi_dequeue(struct task_struct *p, bool sleep) + return; + + if (p->in_memstall) +- clear |= TSK_MEMSTALL; ++ clear |= (TSK_MEMSTALL | TSK_MEMSTALL_RUNNING); + + psi_task_change(p, clear, 0); + } +diff --git a/kernel/time/clocksource.c b/kernel/time/clocksource.c +index b8a14d2fb5ba6..bcad1a1e5dcf1 100644 +--- a/kernel/time/clocksource.c ++++ b/kernel/time/clocksource.c +@@ -107,7 +107,7 @@ static u64 suspend_start; + * This delay could be due to SMIs, NMIs, or to VCPU preemptions. Used as + * a lower bound for cs->uncertainty_margin values when registering clocks. + */ +-#define WATCHDOG_MAX_SKEW (50 * NSEC_PER_USEC) ++#define WATCHDOG_MAX_SKEW (100 * NSEC_PER_USEC) + + #ifdef CONFIG_CLOCKSOURCE_WATCHDOG + static void clocksource_watchdog_work(struct work_struct *work); +@@ -205,17 +205,24 @@ EXPORT_SYMBOL_GPL(max_cswd_read_retries); + static int verify_n_cpus = 8; + module_param(verify_n_cpus, int, 0644); + +-static bool cs_watchdog_read(struct clocksource *cs, u64 *csnow, u64 *wdnow) ++enum wd_read_status { ++ WD_READ_SUCCESS, ++ WD_READ_UNSTABLE, ++ WD_READ_SKIP ++}; ++ ++static enum wd_read_status cs_watchdog_read(struct clocksource *cs, u64 *csnow, u64 *wdnow) + { + unsigned int nretries; +- u64 wd_end, wd_delta; +- int64_t wd_delay; ++ u64 wd_end, wd_end2, wd_delta; ++ int64_t wd_delay, wd_seq_delay; + + for (nretries = 0; nretries <= max_cswd_read_retries; nretries++) { + local_irq_disable(); + *wdnow = watchdog->read(watchdog); + *csnow = cs->read(cs); + wd_end = watchdog->read(watchdog); ++ wd_end2 = watchdog->read(watchdog); + local_irq_enable(); + + wd_delta = clocksource_delta(wd_end, *wdnow, watchdog->mask); +@@ -226,13 +233,34 @@ static bool cs_watchdog_read(struct clocksource *cs, u64 *csnow, u64 *wdnow) + pr_warn("timekeeping watchdog on CPU%d: %s retried %d times before success\n", + smp_processor_id(), watchdog->name, nretries); + } +- return true; ++ return WD_READ_SUCCESS; + } ++ ++ /* ++ * Now compute delay in consecutive watchdog read to see if ++ * there is too much external interferences that cause ++ * significant delay in reading both clocksource and watchdog. ++ * ++ * If consecutive WD read-back delay > WATCHDOG_MAX_SKEW/2, ++ * report system busy, reinit the watchdog and skip the current ++ * watchdog test. ++ */ ++ wd_delta = clocksource_delta(wd_end2, wd_end, watchdog->mask); ++ wd_seq_delay = clocksource_cyc2ns(wd_delta, watchdog->mult, watchdog->shift); ++ if (wd_seq_delay > WATCHDOG_MAX_SKEW/2) ++ goto skip_test; + } + + pr_warn("timekeeping watchdog on CPU%d: %s read-back delay of %lldns, attempt %d, marking unstable\n", + smp_processor_id(), watchdog->name, wd_delay, nretries); +- return false; ++ return WD_READ_UNSTABLE; ++ ++skip_test: ++ pr_info("timekeeping watchdog on CPU%d: %s wd-wd read-back delay of %lldns\n", ++ smp_processor_id(), watchdog->name, wd_seq_delay); ++ pr_info("wd-%s-wd read-back delay of %lldns, clock-skew test skipped!\n", ++ cs->name, wd_delay); ++ return WD_READ_SKIP; + } + + static u64 csnow_mid; +@@ -356,6 +384,7 @@ static void clocksource_watchdog(struct timer_list *unused) + int next_cpu, reset_pending; + int64_t wd_nsec, cs_nsec; + struct clocksource *cs; ++ enum wd_read_status read_ret; + u32 md; + + spin_lock(&watchdog_lock); +@@ -373,9 +402,12 @@ static void clocksource_watchdog(struct timer_list *unused) + continue; + } + +- if (!cs_watchdog_read(cs, &csnow, &wdnow)) { +- /* Clock readout unreliable, so give it up. */ +- __clocksource_unstable(cs); ++ read_ret = cs_watchdog_read(cs, &csnow, &wdnow); ++ ++ if (read_ret != WD_READ_SUCCESS) { ++ if (read_ret == WD_READ_UNSTABLE) ++ /* Clock readout unreliable, so give it up. */ ++ __clocksource_unstable(cs); + continue; + } + +diff --git a/kernel/trace/bpf_trace.c b/kernel/trace/bpf_trace.c +index 6c1038526d1fc..5a18b861fcf75 100644 +--- a/kernel/trace/bpf_trace.c ++++ b/kernel/trace/bpf_trace.c +@@ -1322,9 +1322,6 @@ static const struct bpf_func_proto bpf_perf_prog_read_value_proto = { + BPF_CALL_4(bpf_read_branch_records, struct bpf_perf_event_data_kern *, ctx, + void *, buf, u32, size, u64, flags) + { +-#ifndef CONFIG_X86 +- return -ENOENT; +-#else + static const u32 br_entry_size = sizeof(struct perf_branch_entry); + struct perf_branch_stack *br_stack = ctx->data->br_stack; + u32 to_copy; +@@ -1333,7 +1330,7 @@ BPF_CALL_4(bpf_read_branch_records, struct bpf_perf_event_data_kern *, ctx, + return -EINVAL; + + if (unlikely(!br_stack)) +- return -EINVAL; ++ return -ENOENT; + + if (flags & BPF_F_GET_BRANCH_RECORDS_SIZE) + return br_stack->nr * br_entry_size; +@@ -1345,7 +1342,6 @@ BPF_CALL_4(bpf_read_branch_records, struct bpf_perf_event_data_kern *, ctx, + memcpy(buf, br_stack->entries, to_copy); + + return to_copy; +-#endif + } + + static const struct bpf_func_proto bpf_read_branch_records_proto = { +diff --git a/kernel/trace/trace_kprobe.c b/kernel/trace/trace_kprobe.c +index 92caef33b68c2..385981c293baf 100644 +--- a/kernel/trace/trace_kprobe.c ++++ b/kernel/trace/trace_kprobe.c +@@ -1175,15 +1175,18 @@ static int probes_profile_seq_show(struct seq_file *m, void *v) + { + struct dyn_event *ev = v; + struct trace_kprobe *tk; ++ unsigned long nmissed; + + if (!is_trace_kprobe(ev)) + return 0; + + tk = to_trace_kprobe(ev); ++ nmissed = trace_kprobe_is_return(tk) ? ++ tk->rp.kp.nmissed + tk->rp.nmissed : tk->rp.kp.nmissed; + seq_printf(m, " %-44s %15lu %15lu\n", + trace_probe_name(&tk->tp), + trace_kprobe_nhit(tk), +- tk->rp.kp.nmissed); ++ nmissed); + + return 0; + } +diff --git a/kernel/trace/trace_osnoise.c b/kernel/trace/trace_osnoise.c +index c4f14fb98aaac..65a518649997b 100644 +--- a/kernel/trace/trace_osnoise.c ++++ b/kernel/trace/trace_osnoise.c +@@ -1932,6 +1932,13 @@ out_unhook_irq: + return -EINVAL; + } + ++static void osnoise_unhook_events(void) ++{ ++ unhook_thread_events(); ++ unhook_softirq_events(); ++ unhook_irq_events(); ++} ++ + static int __osnoise_tracer_start(struct trace_array *tr) + { + int retval; +@@ -1949,7 +1956,14 @@ static int __osnoise_tracer_start(struct trace_array *tr) + + retval = start_per_cpu_kthreads(tr); + if (retval) { +- unhook_irq_events(); ++ trace_osnoise_callback_enabled = false; ++ /* ++ * Make sure that ftrace_nmi_enter/exit() see ++ * trace_osnoise_callback_enabled as false before continuing. ++ */ ++ barrier(); ++ ++ osnoise_unhook_events(); + return retval; + } + +@@ -1981,9 +1995,7 @@ static void osnoise_tracer_stop(struct trace_array *tr) + + stop_per_cpu_kthreads(); + +- unhook_irq_events(); +- unhook_softirq_events(); +- unhook_thread_events(); ++ osnoise_unhook_events(); + + osnoise_busy = false; + } +diff --git a/kernel/trace/trace_syscalls.c b/kernel/trace/trace_syscalls.c +index 8bfcd3b094226..f755bde42fd07 100644 +--- a/kernel/trace/trace_syscalls.c ++++ b/kernel/trace/trace_syscalls.c +@@ -323,8 +323,7 @@ static void ftrace_syscall_enter(void *data, struct pt_regs *regs, long id) + + trace_ctx = tracing_gen_ctx(); + +- buffer = tr->array_buffer.buffer; +- event = trace_buffer_lock_reserve(buffer, ++ event = trace_event_buffer_lock_reserve(&buffer, trace_file, + sys_data->enter_event->event.type, size, trace_ctx); + if (!event) + return; +@@ -367,8 +366,7 @@ static void ftrace_syscall_exit(void *data, struct pt_regs *regs, long ret) + + trace_ctx = tracing_gen_ctx(); + +- buffer = tr->array_buffer.buffer; +- event = trace_buffer_lock_reserve(buffer, ++ event = trace_event_buffer_lock_reserve(&buffer, trace_file, + sys_data->exit_event->event.type, sizeof(*entry), + trace_ctx); + if (!event) +diff --git a/kernel/tsacct.c b/kernel/tsacct.c +index 257ffb993ea23..fd2f7a052fdd9 100644 +--- a/kernel/tsacct.c ++++ b/kernel/tsacct.c +@@ -38,11 +38,10 @@ void bacct_add_tsk(struct user_namespace *user_ns, + stats->ac_btime = clamp_t(time64_t, btime, 0, U32_MAX); + stats->ac_btime64 = btime; + +- if (thread_group_leader(tsk)) { ++ if (tsk->flags & PF_EXITING) + stats->ac_exitcode = tsk->exit_code; +- if (tsk->flags & PF_FORKNOEXEC) +- stats->ac_flag |= AFORK; +- } ++ if (thread_group_leader(tsk) && (tsk->flags & PF_FORKNOEXEC)) ++ stats->ac_flag |= AFORK; + if (tsk->flags & PF_SUPERPRIV) + stats->ac_flag |= ASU; + if (tsk->flags & PF_DUMPCORE) +diff --git a/lib/kunit/test.c b/lib/kunit/test.c +index f246b847024e3..9aef816e573c1 100644 +--- a/lib/kunit/test.c ++++ b/lib/kunit/test.c +@@ -504,16 +504,18 @@ int kunit_run_tests(struct kunit_suite *suite) + struct kunit_result_stats param_stats = { 0 }; + test_case->status = KUNIT_SKIPPED; + +- if (test_case->generate_params) { ++ if (!test_case->generate_params) { ++ /* Non-parameterised test. */ ++ kunit_run_case_catch_errors(suite, test_case, &test); ++ kunit_update_stats(¶m_stats, test.status); ++ } else { + /* Get initial param. */ + param_desc[0] = '\0'; + test.param_value = test_case->generate_params(NULL, param_desc); +- } + +- do { +- kunit_run_case_catch_errors(suite, test_case, &test); ++ while (test.param_value) { ++ kunit_run_case_catch_errors(suite, test_case, &test); + +- if (test_case->generate_params) { + if (param_desc[0] == '\0') { + snprintf(param_desc, sizeof(param_desc), + "param-%d", test.param_index); +@@ -530,11 +532,11 @@ int kunit_run_tests(struct kunit_suite *suite) + param_desc[0] = '\0'; + test.param_value = test_case->generate_params(test.param_value, param_desc); + test.param_index++; +- } + +- kunit_update_stats(¶m_stats, test.status); ++ kunit_update_stats(¶m_stats, test.status); ++ } ++ } + +- } while (test.param_value); + + kunit_print_test_stats(&test, param_stats); + +diff --git a/lib/logic_iomem.c b/lib/logic_iomem.c +index 9bdfde0c0f86d..549b22d4bcde1 100644 +--- a/lib/logic_iomem.c ++++ b/lib/logic_iomem.c +@@ -21,15 +21,15 @@ struct logic_iomem_area { + + #define AREA_SHIFT 24 + #define MAX_AREA_SIZE (1 << AREA_SHIFT) +-#define MAX_AREAS ((1ULL<<32) / MAX_AREA_SIZE) ++#define MAX_AREAS ((1U << 31) / MAX_AREA_SIZE) + #define AREA_BITS ((MAX_AREAS - 1) << AREA_SHIFT) + #define AREA_MASK (MAX_AREA_SIZE - 1) + #ifdef CONFIG_64BIT + #define IOREMAP_BIAS 0xDEAD000000000000UL + #define IOREMAP_MASK 0xFFFFFFFF00000000UL + #else +-#define IOREMAP_BIAS 0 +-#define IOREMAP_MASK 0 ++#define IOREMAP_BIAS 0x80000000UL ++#define IOREMAP_MASK 0x80000000UL + #endif + + static DEFINE_MUTEX(regions_mtx); +@@ -79,7 +79,7 @@ static void __iomem *real_ioremap(phys_addr_t offset, size_t size) + static void real_iounmap(void __iomem *addr) + { + WARN(1, "invalid iounmap for addr 0x%llx\n", +- (unsigned long long __force)addr); ++ (unsigned long long)(uintptr_t __force)addr); + } + #endif /* CONFIG_LOGIC_IOMEM_FALLBACK */ + +@@ -173,7 +173,7 @@ EXPORT_SYMBOL(iounmap); + static u##sz real_raw_read ## op(const volatile void __iomem *addr) \ + { \ + WARN(1, "Invalid read" #op " at address %llx\n", \ +- (unsigned long long __force)addr); \ ++ (unsigned long long)(uintptr_t __force)addr); \ + return (u ## sz)~0ULL; \ + } \ + \ +@@ -181,7 +181,8 @@ static void real_raw_write ## op(u ## sz val, \ + volatile void __iomem *addr) \ + { \ + WARN(1, "Invalid writeq" #op " of 0x%llx at address %llx\n", \ +- (unsigned long long)val, (unsigned long long __force)addr);\ ++ (unsigned long long)val, \ ++ (unsigned long long)(uintptr_t __force)addr);\ + } \ + + MAKE_FALLBACK(b, 8); +@@ -194,14 +195,14 @@ MAKE_FALLBACK(q, 64); + static void real_memset_io(volatile void __iomem *addr, int value, size_t size) + { + WARN(1, "Invalid memset_io at address 0x%llx\n", +- (unsigned long long __force)addr); ++ (unsigned long long)(uintptr_t __force)addr); + } + + static void real_memcpy_fromio(void *buffer, const volatile void __iomem *addr, + size_t size) + { + WARN(1, "Invalid memcpy_fromio at address 0x%llx\n", +- (unsigned long long __force)addr); ++ (unsigned long long)(uintptr_t __force)addr); + + memset(buffer, 0xff, size); + } +@@ -210,7 +211,7 @@ static void real_memcpy_toio(volatile void __iomem *addr, const void *buffer, + size_t size) + { + WARN(1, "Invalid memcpy_toio at address 0x%llx\n", +- (unsigned long long __force)addr); ++ (unsigned long long)(uintptr_t __force)addr); + } + #endif /* CONFIG_LOGIC_IOMEM_FALLBACK */ + +diff --git a/lib/mpi/mpi-mod.c b/lib/mpi/mpi-mod.c +index 47bc59edd4ff9..54fcc01564d9d 100644 +--- a/lib/mpi/mpi-mod.c ++++ b/lib/mpi/mpi-mod.c +@@ -40,6 +40,8 @@ mpi_barrett_t mpi_barrett_init(MPI m, int copy) + + mpi_normalize(m); + ctx = kcalloc(1, sizeof(*ctx), GFP_KERNEL); ++ if (!ctx) ++ return NULL; + + if (copy) { + ctx->m = mpi_copy(m); +diff --git a/lib/test_hmm.c b/lib/test_hmm.c +index c259842f6d443..ac794e3540693 100644 +--- a/lib/test_hmm.c ++++ b/lib/test_hmm.c +@@ -1087,9 +1087,33 @@ static long dmirror_fops_unlocked_ioctl(struct file *filp, + return 0; + } + ++static int dmirror_fops_mmap(struct file *file, struct vm_area_struct *vma) ++{ ++ unsigned long addr; ++ ++ for (addr = vma->vm_start; addr < vma->vm_end; addr += PAGE_SIZE) { ++ struct page *page; ++ int ret; ++ ++ page = alloc_page(GFP_KERNEL | __GFP_ZERO); ++ if (!page) ++ return -ENOMEM; ++ ++ ret = vm_insert_page(vma, addr, page); ++ if (ret) { ++ __free_page(page); ++ return ret; ++ } ++ put_page(page); ++ } ++ ++ return 0; ++} ++ + static const struct file_operations dmirror_fops = { + .open = dmirror_fops_open, + .release = dmirror_fops_release, ++ .mmap = dmirror_fops_mmap, + .unlocked_ioctl = dmirror_fops_unlocked_ioctl, + .llseek = default_llseek, + .owner = THIS_MODULE, +diff --git a/lib/test_meminit.c b/lib/test_meminit.c +index e4f706a404b3a..3ca717f113977 100644 +--- a/lib/test_meminit.c ++++ b/lib/test_meminit.c +@@ -337,6 +337,7 @@ static int __init do_kmem_cache_size_bulk(int size, int *total_failures) + if (num) + kmem_cache_free_bulk(c, num, objects); + } ++ kmem_cache_destroy(c); + *total_failures += fail; + return 1; + } +diff --git a/mm/hmm.c b/mm/hmm.c +index 842e265992380..bd56641c79d4e 100644 +--- a/mm/hmm.c ++++ b/mm/hmm.c +@@ -300,7 +300,8 @@ static int hmm_vma_handle_pte(struct mm_walk *walk, unsigned long addr, + * Since each architecture defines a struct page for the zero page, just + * fall through and treat it like a normal page. + */ +- if (pte_special(pte) && !pte_devmap(pte) && ++ if (!vm_normal_page(walk->vma, addr, pte) && ++ !pte_devmap(pte) && + !is_zero_pfn(pte_pfn(pte))) { + if (hmm_pte_need_fault(hmm_vma_walk, pfn_req_flags, 0)) { + pte_unmap(ptep); +@@ -518,7 +519,7 @@ static int hmm_vma_walk_test(unsigned long start, unsigned long end, + struct hmm_range *range = hmm_vma_walk->range; + struct vm_area_struct *vma = walk->vma; + +- if (!(vma->vm_flags & (VM_IO | VM_PFNMAP | VM_MIXEDMAP)) && ++ if (!(vma->vm_flags & (VM_IO | VM_PFNMAP)) && + vma->vm_flags & VM_READ) + return 0; + +diff --git a/mm/page_alloc.c b/mm/page_alloc.c +index 23d3339ac4e8e..7773bae3b6edb 100644 +--- a/mm/page_alloc.c ++++ b/mm/page_alloc.c +@@ -4210,7 +4210,9 @@ void warn_alloc(gfp_t gfp_mask, nodemask_t *nodemask, const char *fmt, ...) + va_list args; + static DEFINE_RATELIMIT_STATE(nopage_rs, 10*HZ, 1); + +- if ((gfp_mask & __GFP_NOWARN) || !__ratelimit(&nopage_rs)) ++ if ((gfp_mask & __GFP_NOWARN) || ++ !__ratelimit(&nopage_rs) || ++ ((gfp_mask & __GFP_DMA) && !has_managed_dma())) + return; + + va_start(args, fmt); +@@ -9449,3 +9451,18 @@ bool take_page_off_buddy(struct page *page) + return ret; + } + #endif ++ ++#ifdef CONFIG_ZONE_DMA ++bool has_managed_dma(void) ++{ ++ struct pglist_data *pgdat; ++ ++ for_each_online_pgdat(pgdat) { ++ struct zone *zone = &pgdat->node_zones[ZONE_DMA]; ++ ++ if (managed_zone(zone)) ++ return true; ++ } ++ return false; ++} ++#endif /* CONFIG_ZONE_DMA */ +diff --git a/mm/shmem.c b/mm/shmem.c +index b5860f4a2738e..1609a8daba26e 100644 +--- a/mm/shmem.c ++++ b/mm/shmem.c +@@ -555,7 +555,7 @@ static unsigned long shmem_unused_huge_shrink(struct shmem_sb_info *sbinfo, + struct shmem_inode_info *info; + struct page *page; + unsigned long batch = sc ? sc->nr_to_scan : 128; +- int removed = 0, split = 0; ++ int split = 0; + + if (list_empty(&sbinfo->shrinklist)) + return SHRINK_STOP; +@@ -570,7 +570,6 @@ static unsigned long shmem_unused_huge_shrink(struct shmem_sb_info *sbinfo, + /* inode is about to be evicted */ + if (!inode) { + list_del_init(&info->shrinklist); +- removed++; + goto next; + } + +@@ -578,12 +577,12 @@ static unsigned long shmem_unused_huge_shrink(struct shmem_sb_info *sbinfo, + if (round_up(inode->i_size, PAGE_SIZE) == + round_up(inode->i_size, HPAGE_PMD_SIZE)) { + list_move(&info->shrinklist, &to_remove); +- removed++; + goto next; + } + + list_move(&info->shrinklist, &list); + next: ++ sbinfo->shrinklist_len--; + if (!--batch) + break; + } +@@ -603,7 +602,7 @@ next: + inode = &info->vfs_inode; + + if (nr_to_split && split >= nr_to_split) +- goto leave; ++ goto move_back; + + page = find_get_page(inode->i_mapping, + (inode->i_size & HPAGE_PMD_MASK) >> PAGE_SHIFT); +@@ -617,38 +616,44 @@ next: + } + + /* +- * Leave the inode on the list if we failed to lock +- * the page at this time. ++ * Move the inode on the list back to shrinklist if we failed ++ * to lock the page at this time. + * + * Waiting for the lock may lead to deadlock in the + * reclaim path. + */ + if (!trylock_page(page)) { + put_page(page); +- goto leave; ++ goto move_back; + } + + ret = split_huge_page(page); + unlock_page(page); + put_page(page); + +- /* If split failed leave the inode on the list */ ++ /* If split failed move the inode on the list back to shrinklist */ + if (ret) +- goto leave; ++ goto move_back; + + split++; + drop: + list_del_init(&info->shrinklist); +- removed++; +-leave: ++ goto put; ++move_back: ++ /* ++ * Make sure the inode is either on the global list or deleted ++ * from any local list before iput() since it could be deleted ++ * in another thread once we put the inode (then the local list ++ * is corrupted). ++ */ ++ spin_lock(&sbinfo->shrinklist_lock); ++ list_move(&info->shrinklist, &sbinfo->shrinklist); ++ sbinfo->shrinklist_len++; ++ spin_unlock(&sbinfo->shrinklist_lock); ++put: + iput(inode); + } + +- spin_lock(&sbinfo->shrinklist_lock); +- list_splice_tail(&list, &sbinfo->shrinklist); +- sbinfo->shrinklist_len -= removed; +- spin_unlock(&sbinfo->shrinklist_lock); +- + return split; + } + +diff --git a/net/ax25/af_ax25.c b/net/ax25/af_ax25.c +index aa7785687e757..7473e0cc6d469 100644 +--- a/net/ax25/af_ax25.c ++++ b/net/ax25/af_ax25.c +@@ -536,7 +536,7 @@ static int ax25_setsockopt(struct socket *sock, int level, int optname, + ax25_cb *ax25; + struct net_device *dev; + char devname[IFNAMSIZ]; +- unsigned long opt; ++ unsigned int opt; + int res = 0; + + if (level != SOL_AX25) +@@ -568,7 +568,7 @@ static int ax25_setsockopt(struct socket *sock, int level, int optname, + break; + + case AX25_T1: +- if (opt < 1 || opt > ULONG_MAX / HZ) { ++ if (opt < 1 || opt > UINT_MAX / HZ) { + res = -EINVAL; + break; + } +@@ -577,7 +577,7 @@ static int ax25_setsockopt(struct socket *sock, int level, int optname, + break; + + case AX25_T2: +- if (opt < 1 || opt > ULONG_MAX / HZ) { ++ if (opt < 1 || opt > UINT_MAX / HZ) { + res = -EINVAL; + break; + } +@@ -593,7 +593,7 @@ static int ax25_setsockopt(struct socket *sock, int level, int optname, + break; + + case AX25_T3: +- if (opt < 1 || opt > ULONG_MAX / HZ) { ++ if (opt < 1 || opt > UINT_MAX / HZ) { + res = -EINVAL; + break; + } +@@ -601,7 +601,7 @@ static int ax25_setsockopt(struct socket *sock, int level, int optname, + break; + + case AX25_IDLE: +- if (opt > ULONG_MAX / (60 * HZ)) { ++ if (opt > UINT_MAX / (60 * HZ)) { + res = -EINVAL; + break; + } +diff --git a/net/batman-adv/netlink.c b/net/batman-adv/netlink.c +index 29276284d281c..00875e1d8c44c 100644 +--- a/net/batman-adv/netlink.c ++++ b/net/batman-adv/netlink.c +@@ -1368,21 +1368,21 @@ static const struct genl_small_ops batadv_netlink_ops[] = { + { + .cmd = BATADV_CMD_TP_METER, + .validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP, +- .flags = GENL_ADMIN_PERM, ++ .flags = GENL_UNS_ADMIN_PERM, + .doit = batadv_netlink_tp_meter_start, + .internal_flags = BATADV_FLAG_NEED_MESH, + }, + { + .cmd = BATADV_CMD_TP_METER_CANCEL, + .validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP, +- .flags = GENL_ADMIN_PERM, ++ .flags = GENL_UNS_ADMIN_PERM, + .doit = batadv_netlink_tp_meter_cancel, + .internal_flags = BATADV_FLAG_NEED_MESH, + }, + { + .cmd = BATADV_CMD_GET_ROUTING_ALGOS, + .validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP, +- .flags = GENL_ADMIN_PERM, ++ .flags = GENL_UNS_ADMIN_PERM, + .dumpit = batadv_algo_dump, + }, + { +@@ -1397,68 +1397,68 @@ static const struct genl_small_ops batadv_netlink_ops[] = { + { + .cmd = BATADV_CMD_GET_TRANSTABLE_LOCAL, + .validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP, +- .flags = GENL_ADMIN_PERM, ++ .flags = GENL_UNS_ADMIN_PERM, + .dumpit = batadv_tt_local_dump, + }, + { + .cmd = BATADV_CMD_GET_TRANSTABLE_GLOBAL, + .validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP, +- .flags = GENL_ADMIN_PERM, ++ .flags = GENL_UNS_ADMIN_PERM, + .dumpit = batadv_tt_global_dump, + }, + { + .cmd = BATADV_CMD_GET_ORIGINATORS, + .validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP, +- .flags = GENL_ADMIN_PERM, ++ .flags = GENL_UNS_ADMIN_PERM, + .dumpit = batadv_orig_dump, + }, + { + .cmd = BATADV_CMD_GET_NEIGHBORS, + .validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP, +- .flags = GENL_ADMIN_PERM, ++ .flags = GENL_UNS_ADMIN_PERM, + .dumpit = batadv_hardif_neigh_dump, + }, + { + .cmd = BATADV_CMD_GET_GATEWAYS, + .validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP, +- .flags = GENL_ADMIN_PERM, ++ .flags = GENL_UNS_ADMIN_PERM, + .dumpit = batadv_gw_dump, + }, + { + .cmd = BATADV_CMD_GET_BLA_CLAIM, + .validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP, +- .flags = GENL_ADMIN_PERM, ++ .flags = GENL_UNS_ADMIN_PERM, + .dumpit = batadv_bla_claim_dump, + }, + { + .cmd = BATADV_CMD_GET_BLA_BACKBONE, + .validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP, +- .flags = GENL_ADMIN_PERM, ++ .flags = GENL_UNS_ADMIN_PERM, + .dumpit = batadv_bla_backbone_dump, + }, + { + .cmd = BATADV_CMD_GET_DAT_CACHE, + .validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP, +- .flags = GENL_ADMIN_PERM, ++ .flags = GENL_UNS_ADMIN_PERM, + .dumpit = batadv_dat_cache_dump, + }, + { + .cmd = BATADV_CMD_GET_MCAST_FLAGS, + .validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP, +- .flags = GENL_ADMIN_PERM, ++ .flags = GENL_UNS_ADMIN_PERM, + .dumpit = batadv_mcast_flags_dump, + }, + { + .cmd = BATADV_CMD_SET_MESH, + .validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP, +- .flags = GENL_ADMIN_PERM, ++ .flags = GENL_UNS_ADMIN_PERM, + .doit = batadv_netlink_set_mesh, + .internal_flags = BATADV_FLAG_NEED_MESH, + }, + { + .cmd = BATADV_CMD_SET_HARDIF, + .validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP, +- .flags = GENL_ADMIN_PERM, ++ .flags = GENL_UNS_ADMIN_PERM, + .doit = batadv_netlink_set_hardif, + .internal_flags = BATADV_FLAG_NEED_MESH | + BATADV_FLAG_NEED_HARDIF, +@@ -1474,7 +1474,7 @@ static const struct genl_small_ops batadv_netlink_ops[] = { + { + .cmd = BATADV_CMD_SET_VLAN, + .validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP, +- .flags = GENL_ADMIN_PERM, ++ .flags = GENL_UNS_ADMIN_PERM, + .doit = batadv_netlink_set_vlan, + .internal_flags = BATADV_FLAG_NEED_MESH | + BATADV_FLAG_NEED_VLAN, +diff --git a/net/bluetooth/cmtp/core.c b/net/bluetooth/cmtp/core.c +index 0a2d78e811cf5..83eb84e8e688f 100644 +--- a/net/bluetooth/cmtp/core.c ++++ b/net/bluetooth/cmtp/core.c +@@ -501,9 +501,7 @@ static int __init cmtp_init(void) + { + BT_INFO("CMTP (CAPI Emulation) ver %s", VERSION); + +- cmtp_init_sockets(); +- +- return 0; ++ return cmtp_init_sockets(); + } + + static void __exit cmtp_exit(void) +diff --git a/net/bluetooth/hci_core.c b/net/bluetooth/hci_core.c +index 325db9c4c6109..53f1b08017aab 100644 +--- a/net/bluetooth/hci_core.c ++++ b/net/bluetooth/hci_core.c +@@ -3999,6 +3999,7 @@ int hci_register_dev(struct hci_dev *hdev) + return id; + + err_wqueue: ++ debugfs_remove_recursive(hdev->debugfs); + destroy_workqueue(hdev->workqueue); + destroy_workqueue(hdev->req_workqueue); + err: +diff --git a/net/bluetooth/hci_event.c b/net/bluetooth/hci_event.c +index 0bca035bf2dcc..20e36126bbdae 100644 +--- a/net/bluetooth/hci_event.c ++++ b/net/bluetooth/hci_event.c +@@ -1325,8 +1325,10 @@ static void hci_cc_le_set_ext_adv_enable(struct hci_dev *hdev, + &conn->le_conn_timeout, + conn->conn_timeout); + } else { +- if (adv) { +- adv->enabled = false; ++ if (cp->num_of_sets) { ++ if (adv) ++ adv->enabled = false; ++ + /* If just one instance was disabled check if there are + * any other instance enabled before clearing HCI_LE_ADV + */ +@@ -5780,7 +5782,8 @@ static void hci_le_adv_report_evt(struct hci_dev *hdev, struct sk_buff *skb) + struct hci_ev_le_advertising_info *ev = ptr; + s8 rssi; + +- if (ev->length <= HCI_MAX_AD_LENGTH) { ++ if (ev->length <= HCI_MAX_AD_LENGTH && ++ ev->data + ev->length <= skb_tail_pointer(skb)) { + rssi = ev->data[ev->length]; + process_adv_report(hdev, ev->evt_type, &ev->bdaddr, + ev->bdaddr_type, NULL, 0, rssi, +@@ -5790,6 +5793,11 @@ static void hci_le_adv_report_evt(struct hci_dev *hdev, struct sk_buff *skb) + } + + ptr += sizeof(*ev) + ev->length + 1; ++ ++ if (ptr > (void *) skb_tail_pointer(skb) - sizeof(*ev)) { ++ bt_dev_err(hdev, "Malicious advertising data. Stopping processing"); ++ break; ++ } + } + + hci_dev_unlock(hdev); +diff --git a/net/bluetooth/hci_request.c b/net/bluetooth/hci_request.c +index f15626607b2d6..1d34d330afd34 100644 +--- a/net/bluetooth/hci_request.c ++++ b/net/bluetooth/hci_request.c +@@ -2318,7 +2318,7 @@ int __hci_req_enable_ext_advertising(struct hci_request *req, u8 instance) + /* Set duration per instance since controller is responsible for + * scheduling it. + */ +- if (adv_instance && adv_instance->duration) { ++ if (adv_instance && adv_instance->timeout) { + u16 duration = adv_instance->timeout * MSEC_PER_SEC; + + /* Time = N * 10 ms */ +diff --git a/net/bluetooth/hci_sysfs.c b/net/bluetooth/hci_sysfs.c +index 7827639ecf5c3..4e3e0451b08c1 100644 +--- a/net/bluetooth/hci_sysfs.c ++++ b/net/bluetooth/hci_sysfs.c +@@ -86,6 +86,8 @@ static void bt_host_release(struct device *dev) + + if (hci_dev_test_flag(hdev, HCI_UNREGISTER)) + hci_release_dev(hdev); ++ else ++ kfree(hdev); + module_put(THIS_MODULE); + } + +diff --git a/net/bluetooth/l2cap_sock.c b/net/bluetooth/l2cap_sock.c +index 160c016a5dfb9..d2c6785205992 100644 +--- a/net/bluetooth/l2cap_sock.c ++++ b/net/bluetooth/l2cap_sock.c +@@ -161,7 +161,11 @@ static int l2cap_sock_bind(struct socket *sock, struct sockaddr *addr, int alen) + break; + } + +- if (chan->psm && bdaddr_type_is_le(chan->src_type)) ++ /* Use L2CAP_MODE_LE_FLOWCTL (CoC) in case of LE address and ++ * L2CAP_MODE_EXT_FLOWCTL (ECRED) has not been set. ++ */ ++ if (chan->psm && bdaddr_type_is_le(chan->src_type) && ++ chan->mode != L2CAP_MODE_EXT_FLOWCTL) + chan->mode = L2CAP_MODE_LE_FLOWCTL; + + chan->state = BT_BOUND; +@@ -172,6 +176,21 @@ done: + return err; + } + ++static void l2cap_sock_init_pid(struct sock *sk) ++{ ++ struct l2cap_chan *chan = l2cap_pi(sk)->chan; ++ ++ /* Only L2CAP_MODE_EXT_FLOWCTL ever need to access the PID in order to ++ * group the channels being requested. ++ */ ++ if (chan->mode != L2CAP_MODE_EXT_FLOWCTL) ++ return; ++ ++ spin_lock(&sk->sk_peer_lock); ++ sk->sk_peer_pid = get_pid(task_tgid(current)); ++ spin_unlock(&sk->sk_peer_lock); ++} ++ + static int l2cap_sock_connect(struct socket *sock, struct sockaddr *addr, + int alen, int flags) + { +@@ -240,9 +259,15 @@ static int l2cap_sock_connect(struct socket *sock, struct sockaddr *addr, + return -EINVAL; + } + +- if (chan->psm && bdaddr_type_is_le(chan->src_type) && !chan->mode) ++ /* Use L2CAP_MODE_LE_FLOWCTL (CoC) in case of LE address and ++ * L2CAP_MODE_EXT_FLOWCTL (ECRED) has not been set. ++ */ ++ if (chan->psm && bdaddr_type_is_le(chan->src_type) && ++ chan->mode != L2CAP_MODE_EXT_FLOWCTL) + chan->mode = L2CAP_MODE_LE_FLOWCTL; + ++ l2cap_sock_init_pid(sk); ++ + err = l2cap_chan_connect(chan, la.l2_psm, __le16_to_cpu(la.l2_cid), + &la.l2_bdaddr, la.l2_bdaddr_type); + if (err) +@@ -298,6 +323,8 @@ static int l2cap_sock_listen(struct socket *sock, int backlog) + goto done; + } + ++ l2cap_sock_init_pid(sk); ++ + sk->sk_max_ack_backlog = backlog; + sk->sk_ack_backlog = 0; + +@@ -876,6 +903,8 @@ static int l2cap_sock_setsockopt(struct socket *sock, int level, int optname, + struct l2cap_conn *conn; + int len, err = 0; + u32 opt; ++ u16 mtu; ++ u8 mode; + + BT_DBG("sk %p", sk); + +@@ -1058,16 +1087,16 @@ static int l2cap_sock_setsockopt(struct socket *sock, int level, int optname, + break; + } + +- if (copy_from_sockptr(&opt, optval, sizeof(u16))) { ++ if (copy_from_sockptr(&mtu, optval, sizeof(u16))) { + err = -EFAULT; + break; + } + + if (chan->mode == L2CAP_MODE_EXT_FLOWCTL && + sk->sk_state == BT_CONNECTED) +- err = l2cap_chan_reconfigure(chan, opt); ++ err = l2cap_chan_reconfigure(chan, mtu); + else +- chan->imtu = opt; ++ chan->imtu = mtu; + + break; + +@@ -1089,14 +1118,14 @@ static int l2cap_sock_setsockopt(struct socket *sock, int level, int optname, + break; + } + +- if (copy_from_sockptr(&opt, optval, sizeof(u8))) { ++ if (copy_from_sockptr(&mode, optval, sizeof(u8))) { + err = -EFAULT; + break; + } + +- BT_DBG("opt %u", opt); ++ BT_DBG("mode %u", mode); + +- err = l2cap_set_mode(chan, opt); ++ err = l2cap_set_mode(chan, mode); + if (err) + break; + +diff --git a/net/bluetooth/mgmt.c b/net/bluetooth/mgmt.c +index cea01e275f1ea..f09f0a78eb7be 100644 +--- a/net/bluetooth/mgmt.c ++++ b/net/bluetooth/mgmt.c +@@ -3806,7 +3806,7 @@ static const u8 rpa_resolution_uuid[16] = { + static int read_exp_features_info(struct sock *sk, struct hci_dev *hdev, + void *data, u16 data_len) + { +- char buf[62]; /* Enough space for 3 features */ ++ char buf[62]; /* Enough space for 3 features */ + struct mgmt_rp_read_exp_features_info *rp = (void *)buf; + u16 idx = 0; + u32 flags; +@@ -3892,150 +3892,186 @@ static int exp_debug_feature_changed(bool enabled, struct sock *skip) + } + #endif + +-static int set_exp_feature(struct sock *sk, struct hci_dev *hdev, +- void *data, u16 data_len) ++#define EXP_FEAT(_uuid, _set_func) \ ++{ \ ++ .uuid = _uuid, \ ++ .set_func = _set_func, \ ++} ++ ++/* The zero key uuid is special. Multiple exp features are set through it. */ ++static int set_zero_key_func(struct sock *sk, struct hci_dev *hdev, ++ struct mgmt_cp_set_exp_feature *cp, u16 data_len) + { +- struct mgmt_cp_set_exp_feature *cp = data; + struct mgmt_rp_set_exp_feature rp; + +- bt_dev_dbg(hdev, "sock %p", sk); +- +- if (!memcmp(cp->uuid, ZERO_KEY, 16)) { +- memset(rp.uuid, 0, 16); +- rp.flags = cpu_to_le32(0); ++ memset(rp.uuid, 0, 16); ++ rp.flags = cpu_to_le32(0); + + #ifdef CONFIG_BT_FEATURE_DEBUG +- if (!hdev) { +- bool changed = bt_dbg_get(); ++ if (!hdev) { ++ bool changed = bt_dbg_get(); + +- bt_dbg_set(false); ++ bt_dbg_set(false); + +- if (changed) +- exp_debug_feature_changed(false, sk); +- } ++ if (changed) ++ exp_debug_feature_changed(false, sk); ++ } + #endif + +- if (hdev && use_ll_privacy(hdev) && !hdev_is_powered(hdev)) { +- bool changed = hci_dev_test_flag(hdev, +- HCI_ENABLE_LL_PRIVACY); +- +- hci_dev_clear_flag(hdev, HCI_ENABLE_LL_PRIVACY); ++ if (hdev && use_ll_privacy(hdev) && !hdev_is_powered(hdev)) { ++ bool changed; + +- if (changed) +- exp_ll_privacy_feature_changed(false, hdev, sk); +- } ++ changed = hci_dev_test_and_clear_flag(hdev, ++ HCI_ENABLE_LL_PRIVACY); ++ if (changed) ++ exp_ll_privacy_feature_changed(false, hdev, sk); ++ } + +- hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS); ++ hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS); + +- return mgmt_cmd_complete(sk, hdev ? hdev->id : MGMT_INDEX_NONE, +- MGMT_OP_SET_EXP_FEATURE, 0, +- &rp, sizeof(rp)); +- } ++ return mgmt_cmd_complete(sk, hdev ? hdev->id : MGMT_INDEX_NONE, ++ MGMT_OP_SET_EXP_FEATURE, 0, ++ &rp, sizeof(rp)); ++} + + #ifdef CONFIG_BT_FEATURE_DEBUG +- if (!memcmp(cp->uuid, debug_uuid, 16)) { +- bool val, changed; +- int err; ++static int set_debug_func(struct sock *sk, struct hci_dev *hdev, ++ struct mgmt_cp_set_exp_feature *cp, u16 data_len) ++{ ++ struct mgmt_rp_set_exp_feature rp; + +- /* Command requires to use the non-controller index */ +- if (hdev) +- return mgmt_cmd_status(sk, hdev->id, +- MGMT_OP_SET_EXP_FEATURE, +- MGMT_STATUS_INVALID_INDEX); ++ bool val, changed; ++ int err; + +- /* Parameters are limited to a single octet */ +- if (data_len != MGMT_SET_EXP_FEATURE_SIZE + 1) +- return mgmt_cmd_status(sk, MGMT_INDEX_NONE, +- MGMT_OP_SET_EXP_FEATURE, +- MGMT_STATUS_INVALID_PARAMS); ++ /* Command requires to use the non-controller index */ ++ if (hdev) ++ return mgmt_cmd_status(sk, hdev->id, ++ MGMT_OP_SET_EXP_FEATURE, ++ MGMT_STATUS_INVALID_INDEX); + +- /* Only boolean on/off is supported */ +- if (cp->param[0] != 0x00 && cp->param[0] != 0x01) +- return mgmt_cmd_status(sk, MGMT_INDEX_NONE, +- MGMT_OP_SET_EXP_FEATURE, +- MGMT_STATUS_INVALID_PARAMS); ++ /* Parameters are limited to a single octet */ ++ if (data_len != MGMT_SET_EXP_FEATURE_SIZE + 1) ++ return mgmt_cmd_status(sk, MGMT_INDEX_NONE, ++ MGMT_OP_SET_EXP_FEATURE, ++ MGMT_STATUS_INVALID_PARAMS); + +- val = !!cp->param[0]; +- changed = val ? !bt_dbg_get() : bt_dbg_get(); +- bt_dbg_set(val); ++ /* Only boolean on/off is supported */ ++ if (cp->param[0] != 0x00 && cp->param[0] != 0x01) ++ return mgmt_cmd_status(sk, MGMT_INDEX_NONE, ++ MGMT_OP_SET_EXP_FEATURE, ++ MGMT_STATUS_INVALID_PARAMS); + +- memcpy(rp.uuid, debug_uuid, 16); +- rp.flags = cpu_to_le32(val ? BIT(0) : 0); ++ val = !!cp->param[0]; ++ changed = val ? !bt_dbg_get() : bt_dbg_get(); ++ bt_dbg_set(val); + +- hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS); ++ memcpy(rp.uuid, debug_uuid, 16); ++ rp.flags = cpu_to_le32(val ? BIT(0) : 0); + +- err = mgmt_cmd_complete(sk, MGMT_INDEX_NONE, +- MGMT_OP_SET_EXP_FEATURE, 0, +- &rp, sizeof(rp)); ++ hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS); + +- if (changed) +- exp_debug_feature_changed(val, sk); ++ err = mgmt_cmd_complete(sk, MGMT_INDEX_NONE, ++ MGMT_OP_SET_EXP_FEATURE, 0, ++ &rp, sizeof(rp)); + +- return err; +- } ++ if (changed) ++ exp_debug_feature_changed(val, sk); ++ ++ return err; ++} + #endif + +- if (!memcmp(cp->uuid, rpa_resolution_uuid, 16)) { +- bool val, changed; +- int err; +- u32 flags; ++static int set_rpa_resolution_func(struct sock *sk, struct hci_dev *hdev, ++ struct mgmt_cp_set_exp_feature *cp, ++ u16 data_len) ++{ ++ struct mgmt_rp_set_exp_feature rp; ++ bool val, changed; ++ int err; ++ u32 flags; + +- /* Command requires to use the controller index */ +- if (!hdev) +- return mgmt_cmd_status(sk, MGMT_INDEX_NONE, +- MGMT_OP_SET_EXP_FEATURE, +- MGMT_STATUS_INVALID_INDEX); ++ /* Command requires to use the controller index */ ++ if (!hdev) ++ return mgmt_cmd_status(sk, MGMT_INDEX_NONE, ++ MGMT_OP_SET_EXP_FEATURE, ++ MGMT_STATUS_INVALID_INDEX); + +- /* Changes can only be made when controller is powered down */ +- if (hdev_is_powered(hdev)) +- return mgmt_cmd_status(sk, hdev->id, +- MGMT_OP_SET_EXP_FEATURE, +- MGMT_STATUS_REJECTED); ++ /* Changes can only be made when controller is powered down */ ++ if (hdev_is_powered(hdev)) ++ return mgmt_cmd_status(sk, hdev->id, ++ MGMT_OP_SET_EXP_FEATURE, ++ MGMT_STATUS_REJECTED); + +- /* Parameters are limited to a single octet */ +- if (data_len != MGMT_SET_EXP_FEATURE_SIZE + 1) +- return mgmt_cmd_status(sk, hdev->id, +- MGMT_OP_SET_EXP_FEATURE, +- MGMT_STATUS_INVALID_PARAMS); ++ /* Parameters are limited to a single octet */ ++ if (data_len != MGMT_SET_EXP_FEATURE_SIZE + 1) ++ return mgmt_cmd_status(sk, hdev->id, ++ MGMT_OP_SET_EXP_FEATURE, ++ MGMT_STATUS_INVALID_PARAMS); + +- /* Only boolean on/off is supported */ +- if (cp->param[0] != 0x00 && cp->param[0] != 0x01) +- return mgmt_cmd_status(sk, hdev->id, +- MGMT_OP_SET_EXP_FEATURE, +- MGMT_STATUS_INVALID_PARAMS); ++ /* Only boolean on/off is supported */ ++ if (cp->param[0] != 0x00 && cp->param[0] != 0x01) ++ return mgmt_cmd_status(sk, hdev->id, ++ MGMT_OP_SET_EXP_FEATURE, ++ MGMT_STATUS_INVALID_PARAMS); + +- val = !!cp->param[0]; ++ val = !!cp->param[0]; + +- if (val) { +- changed = !hci_dev_test_flag(hdev, ++ if (val) { ++ changed = !hci_dev_test_and_set_flag(hdev, + HCI_ENABLE_LL_PRIVACY); +- hci_dev_set_flag(hdev, HCI_ENABLE_LL_PRIVACY); +- hci_dev_clear_flag(hdev, HCI_ADVERTISING); ++ hci_dev_clear_flag(hdev, HCI_ADVERTISING); + +- /* Enable LL privacy + supported settings changed */ +- flags = BIT(0) | BIT(1); +- } else { +- changed = hci_dev_test_flag(hdev, +- HCI_ENABLE_LL_PRIVACY); +- hci_dev_clear_flag(hdev, HCI_ENABLE_LL_PRIVACY); ++ /* Enable LL privacy + supported settings changed */ ++ flags = BIT(0) | BIT(1); ++ } else { ++ changed = hci_dev_test_and_clear_flag(hdev, ++ HCI_ENABLE_LL_PRIVACY); + +- /* Disable LL privacy + supported settings changed */ +- flags = BIT(1); +- } ++ /* Disable LL privacy + supported settings changed */ ++ flags = BIT(1); ++ } + +- memcpy(rp.uuid, rpa_resolution_uuid, 16); +- rp.flags = cpu_to_le32(flags); ++ memcpy(rp.uuid, rpa_resolution_uuid, 16); ++ rp.flags = cpu_to_le32(flags); + +- hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS); ++ hci_sock_set_flag(sk, HCI_MGMT_EXP_FEATURE_EVENTS); + +- err = mgmt_cmd_complete(sk, hdev->id, +- MGMT_OP_SET_EXP_FEATURE, 0, +- &rp, sizeof(rp)); ++ err = mgmt_cmd_complete(sk, hdev->id, ++ MGMT_OP_SET_EXP_FEATURE, 0, ++ &rp, sizeof(rp)); + +- if (changed) +- exp_ll_privacy_feature_changed(val, hdev, sk); ++ if (changed) ++ exp_ll_privacy_feature_changed(val, hdev, sk); + +- return err; ++ return err; ++} ++ ++static const struct mgmt_exp_feature { ++ const u8 *uuid; ++ int (*set_func)(struct sock *sk, struct hci_dev *hdev, ++ struct mgmt_cp_set_exp_feature *cp, u16 data_len); ++} exp_features[] = { ++ EXP_FEAT(ZERO_KEY, set_zero_key_func), ++#ifdef CONFIG_BT_FEATURE_DEBUG ++ EXP_FEAT(debug_uuid, set_debug_func), ++#endif ++ EXP_FEAT(rpa_resolution_uuid, set_rpa_resolution_func), ++ ++ /* end with a null feature */ ++ EXP_FEAT(NULL, NULL) ++}; ++ ++static int set_exp_feature(struct sock *sk, struct hci_dev *hdev, ++ void *data, u16 data_len) ++{ ++ struct mgmt_cp_set_exp_feature *cp = data; ++ size_t i = 0; ++ ++ bt_dev_dbg(hdev, "sock %p", sk); ++ ++ for (i = 0; exp_features[i].uuid; i++) { ++ if (!memcmp(cp->uuid, exp_features[i].uuid, 16)) ++ return exp_features[i].set_func(sk, hdev, cp, data_len); + } + + return mgmt_cmd_status(sk, hdev ? hdev->id : MGMT_INDEX_NONE, +diff --git a/net/bridge/br_netfilter_hooks.c b/net/bridge/br_netfilter_hooks.c +index 8edfb98ae1d58..68c0d0f928908 100644 +--- a/net/bridge/br_netfilter_hooks.c ++++ b/net/bridge/br_netfilter_hooks.c +@@ -743,6 +743,9 @@ static int br_nf_dev_queue_xmit(struct net *net, struct sock *sk, struct sk_buff + if (nf_bridge->frag_max_size && nf_bridge->frag_max_size < mtu) + mtu = nf_bridge->frag_max_size; + ++ nf_bridge_update_protocol(skb); ++ nf_bridge_push_encap_header(skb); ++ + if (skb_is_gso(skb) || skb->len + mtu_reserved <= mtu) { + nf_bridge_info_free(skb); + return br_dev_queue_push_xmit(net, sk, skb); +@@ -760,8 +763,6 @@ static int br_nf_dev_queue_xmit(struct net *net, struct sock *sk, struct sk_buff + + IPCB(skb)->frag_max_size = nf_bridge->frag_max_size; + +- nf_bridge_update_protocol(skb); +- + data = this_cpu_ptr(&brnf_frag_data_storage); + + if (skb_vlan_tag_present(skb)) { +@@ -789,8 +790,6 @@ static int br_nf_dev_queue_xmit(struct net *net, struct sock *sk, struct sk_buff + + IP6CB(skb)->frag_max_size = nf_bridge->frag_max_size; + +- nf_bridge_update_protocol(skb); +- + data = this_cpu_ptr(&brnf_frag_data_storage); + data->encap_size = nf_bridge_encap_header_len(skb); + data->size = ETH_HLEN + data->encap_size; +diff --git a/net/core/dev.c b/net/core/dev.c +index e0878a500aa92..33dc2a3ff7d78 100644 +--- a/net/core/dev.c ++++ b/net/core/dev.c +@@ -9636,6 +9636,12 @@ static int bpf_xdp_link_update(struct bpf_link *link, struct bpf_prog *new_prog, + goto out_unlock; + } + old_prog = link->prog; ++ if (old_prog->type != new_prog->type || ++ old_prog->expected_attach_type != new_prog->expected_attach_type) { ++ err = -EINVAL; ++ goto out_unlock; ++ } ++ + if (old_prog == new_prog) { + /* no-op, don't disturb drivers */ + bpf_prog_put(new_prog); +diff --git a/net/core/devlink.c b/net/core/devlink.c +index 6931713e363fd..db76c55e1a6d7 100644 +--- a/net/core/devlink.c ++++ b/net/core/devlink.c +@@ -8795,8 +8795,6 @@ static const struct genl_small_ops devlink_nl_ops[] = { + GENL_DONT_VALIDATE_DUMP_STRICT, + .dumpit = devlink_nl_cmd_health_reporter_dump_get_dumpit, + .flags = GENL_ADMIN_PERM, +- .internal_flags = DEVLINK_NL_FLAG_NEED_DEVLINK_OR_PORT | +- DEVLINK_NL_FLAG_NO_LOCK, + }, + { + .cmd = DEVLINK_CMD_HEALTH_REPORTER_DUMP_CLEAR, +diff --git a/net/core/filter.c b/net/core/filter.c +index 1e6831880d1fd..f207e4782bd0e 100644 +--- a/net/core/filter.c ++++ b/net/core/filter.c +@@ -4742,12 +4742,14 @@ static int _bpf_setsockopt(struct sock *sk, int level, int optname, + switch (optname) { + case SO_RCVBUF: + val = min_t(u32, val, sysctl_rmem_max); ++ val = min_t(int, val, INT_MAX / 2); + sk->sk_userlocks |= SOCK_RCVBUF_LOCK; + WRITE_ONCE(sk->sk_rcvbuf, + max_t(int, val * 2, SOCK_MIN_RCVBUF)); + break; + case SO_SNDBUF: + val = min_t(u32, val, sysctl_wmem_max); ++ val = min_t(int, val, INT_MAX / 2); + sk->sk_userlocks |= SOCK_SNDBUF_LOCK; + WRITE_ONCE(sk->sk_sndbuf, + max_t(int, val * 2, SOCK_MIN_SNDBUF)); +@@ -8176,9 +8178,9 @@ void bpf_warn_invalid_xdp_action(u32 act) + { + const u32 act_max = XDP_REDIRECT; + +- WARN_ONCE(1, "%s XDP return value %u, expect packet loss!\n", +- act > act_max ? "Illegal" : "Driver unsupported", +- act); ++ pr_warn_once("%s XDP return value %u, expect packet loss!\n", ++ act > act_max ? "Illegal" : "Driver unsupported", ++ act); + } + EXPORT_SYMBOL_GPL(bpf_warn_invalid_xdp_action); + +diff --git a/net/core/flow_dissector.c b/net/core/flow_dissector.c +index bac0184cf3de7..edffdaa875f1f 100644 +--- a/net/core/flow_dissector.c ++++ b/net/core/flow_dissector.c +@@ -238,7 +238,7 @@ void + skb_flow_dissect_ct(const struct sk_buff *skb, + struct flow_dissector *flow_dissector, + void *target_container, u16 *ctinfo_map, +- size_t mapsize, bool post_ct) ++ size_t mapsize, bool post_ct, u16 zone) + { + #if IS_ENABLED(CONFIG_NF_CONNTRACK) + struct flow_dissector_key_ct *key; +@@ -260,6 +260,7 @@ skb_flow_dissect_ct(const struct sk_buff *skb, + if (!ct) { + key->ct_state = TCA_FLOWER_KEY_CT_FLAGS_TRACKED | + TCA_FLOWER_KEY_CT_FLAGS_INVALID; ++ key->ct_zone = zone; + return; + } + +diff --git a/net/core/net-sysfs.c b/net/core/net-sysfs.c +index dfa5ecff7f738..a4ae652633844 100644 +--- a/net/core/net-sysfs.c ++++ b/net/core/net-sysfs.c +@@ -1820,6 +1820,9 @@ static void remove_queue_kobjects(struct net_device *dev) + + net_rx_queue_update_kobjects(dev, real_rx, 0); + netdev_queue_update_kobjects(dev, real_tx, 0); ++ ++ dev->real_num_rx_queues = 0; ++ dev->real_num_tx_queues = 0; + #ifdef CONFIG_SYSFS + kset_unregister(dev->queues_kset); + #endif +diff --git a/net/core/net_namespace.c b/net/core/net_namespace.c +index 202fa5eacd0f9..9702d2b0d9207 100644 +--- a/net/core/net_namespace.c ++++ b/net/core/net_namespace.c +@@ -164,8 +164,10 @@ static void ops_exit_list(const struct pernet_operations *ops, + { + struct net *net; + if (ops->exit) { +- list_for_each_entry(net, net_exit_list, exit_list) ++ list_for_each_entry(net, net_exit_list, exit_list) { + ops->exit(net); ++ cond_resched(); ++ } + } + if (ops->exit_batch) + ops->exit_batch(net_exit_list); +diff --git a/net/core/sock.c b/net/core/sock.c +index 1b31e10181629..6ea317f84edd2 100644 +--- a/net/core/sock.c ++++ b/net/core/sock.c +@@ -830,6 +830,8 @@ static int sock_timestamping_bind_phc(struct sock *sk, int phc_index) + } + + num = ethtool_get_phc_vclocks(dev, &vclock_index); ++ dev_put(dev); ++ + for (i = 0; i < num; i++) { + if (*(vclock_index + i) == phc_index) { + match = true; +diff --git a/net/core/sock_map.c b/net/core/sock_map.c +index c89f527411e84..8288b5382f08d 100644 +--- a/net/core/sock_map.c ++++ b/net/core/sock_map.c +@@ -292,15 +292,23 @@ static int sock_map_link(struct bpf_map *map, struct sock *sk) + if (skb_verdict) + psock_set_prog(&psock->progs.skb_verdict, skb_verdict); + ++ /* msg_* and stream_* programs references tracked in psock after this ++ * point. Reference dec and cleanup will occur through psock destructor ++ */ + ret = sock_map_init_proto(sk, psock); +- if (ret < 0) +- goto out_drop; ++ if (ret < 0) { ++ sk_psock_put(sk, psock); ++ goto out; ++ } + + write_lock_bh(&sk->sk_callback_lock); + if (stream_parser && stream_verdict && !psock->saved_data_ready) { + ret = sk_psock_init_strp(sk, psock); +- if (ret) +- goto out_unlock_drop; ++ if (ret) { ++ write_unlock_bh(&sk->sk_callback_lock); ++ sk_psock_put(sk, psock); ++ goto out; ++ } + sk_psock_start_strp(sk, psock); + } else if (!stream_parser && stream_verdict && !psock->saved_data_ready) { + sk_psock_start_verdict(sk,psock); +@@ -309,10 +317,6 @@ static int sock_map_link(struct bpf_map *map, struct sock *sk) + } + write_unlock_bh(&sk->sk_callback_lock); + return 0; +-out_unlock_drop: +- write_unlock_bh(&sk->sk_callback_lock); +-out_drop: +- sk_psock_put(sk, psock); + out_progs: + if (skb_verdict) + bpf_prog_put(skb_verdict); +@@ -325,6 +329,7 @@ out_put_stream_parser: + out_put_stream_verdict: + if (stream_verdict) + bpf_prog_put(stream_verdict); ++out: + return ret; + } + +diff --git a/net/dsa/switch.c b/net/dsa/switch.c +index 44558fbdc65b3..fb69f2f14234e 100644 +--- a/net/dsa/switch.c ++++ b/net/dsa/switch.c +@@ -644,7 +644,7 @@ static int + dsa_switch_mrp_add_ring_role(struct dsa_switch *ds, + struct dsa_notifier_mrp_ring_role_info *info) + { +- if (!ds->ops->port_mrp_add) ++ if (!ds->ops->port_mrp_add_ring_role) + return -EOPNOTSUPP; + + if (ds->index == info->sw_index) +@@ -658,7 +658,7 @@ static int + dsa_switch_mrp_del_ring_role(struct dsa_switch *ds, + struct dsa_notifier_mrp_ring_role_info *info) + { +- if (!ds->ops->port_mrp_del) ++ if (!ds->ops->port_mrp_del_ring_role) + return -EOPNOTSUPP; + + if (ds->index == info->sw_index) +diff --git a/net/ipv4/fib_semantics.c b/net/ipv4/fib_semantics.c +index 92c29ab3d0428..5dfb94abe7b10 100644 +--- a/net/ipv4/fib_semantics.c ++++ b/net/ipv4/fib_semantics.c +@@ -29,6 +29,7 @@ + #include <linux/init.h> + #include <linux/slab.h> + #include <linux/netlink.h> ++#include <linux/hash.h> + + #include <net/arp.h> + #include <net/ip.h> +@@ -249,7 +250,6 @@ void free_fib_info(struct fib_info *fi) + pr_warn("Freeing alive fib_info %p\n", fi); + return; + } +- fib_info_cnt--; + + call_rcu(&fi->rcu, free_fib_info_rcu); + } +@@ -260,6 +260,10 @@ void fib_release_info(struct fib_info *fi) + spin_lock_bh(&fib_info_lock); + if (fi && refcount_dec_and_test(&fi->fib_treeref)) { + hlist_del(&fi->fib_hash); ++ ++ /* Paired with READ_ONCE() in fib_create_info(). */ ++ WRITE_ONCE(fib_info_cnt, fib_info_cnt - 1); ++ + if (fi->fib_prefsrc) + hlist_del(&fi->fib_lhash); + if (fi->nh) { +@@ -316,11 +320,15 @@ static inline int nh_comp(struct fib_info *fi, struct fib_info *ofi) + + static inline unsigned int fib_devindex_hashfn(unsigned int val) + { +- unsigned int mask = DEVINDEX_HASHSIZE - 1; ++ return hash_32(val, DEVINDEX_HASHBITS); ++} ++ ++static struct hlist_head * ++fib_info_devhash_bucket(const struct net_device *dev) ++{ ++ u32 val = net_hash_mix(dev_net(dev)) ^ dev->ifindex; + +- return (val ^ +- (val >> DEVINDEX_HASHBITS) ^ +- (val >> (DEVINDEX_HASHBITS * 2))) & mask; ++ return &fib_info_devhash[fib_devindex_hashfn(val)]; + } + + static unsigned int fib_info_hashfn_1(int init_val, u8 protocol, u8 scope, +@@ -430,12 +438,11 @@ int ip_fib_check_default(__be32 gw, struct net_device *dev) + { + struct hlist_head *head; + struct fib_nh *nh; +- unsigned int hash; + + spin_lock(&fib_info_lock); + +- hash = fib_devindex_hashfn(dev->ifindex); +- head = &fib_info_devhash[hash]; ++ head = fib_info_devhash_bucket(dev); ++ + hlist_for_each_entry(nh, head, nh_hash) { + if (nh->fib_nh_dev == dev && + nh->fib_nh_gw4 == gw && +@@ -1430,7 +1437,9 @@ struct fib_info *fib_create_info(struct fib_config *cfg, + #endif + + err = -ENOBUFS; +- if (fib_info_cnt >= fib_info_hash_size) { ++ ++ /* Paired with WRITE_ONCE() in fib_release_info() */ ++ if (READ_ONCE(fib_info_cnt) >= fib_info_hash_size) { + unsigned int new_size = fib_info_hash_size << 1; + struct hlist_head *new_info_hash; + struct hlist_head *new_laddrhash; +@@ -1462,7 +1471,6 @@ struct fib_info *fib_create_info(struct fib_config *cfg, + return ERR_PTR(err); + } + +- fib_info_cnt++; + fi->fib_net = net; + fi->fib_protocol = cfg->fc_protocol; + fi->fib_scope = cfg->fc_scope; +@@ -1589,6 +1597,7 @@ link_it: + refcount_set(&fi->fib_treeref, 1); + refcount_set(&fi->fib_clntref, 1); + spin_lock_bh(&fib_info_lock); ++ fib_info_cnt++; + hlist_add_head(&fi->fib_hash, + &fib_info_hash[fib_info_hashfn(fi)]); + if (fi->fib_prefsrc) { +@@ -1602,12 +1611,10 @@ link_it: + } else { + change_nexthops(fi) { + struct hlist_head *head; +- unsigned int hash; + + if (!nexthop_nh->fib_nh_dev) + continue; +- hash = fib_devindex_hashfn(nexthop_nh->fib_nh_dev->ifindex); +- head = &fib_info_devhash[hash]; ++ head = fib_info_devhash_bucket(nexthop_nh->fib_nh_dev); + hlist_add_head(&nexthop_nh->nh_hash, head); + } endfor_nexthops(fi) + } +@@ -1959,8 +1966,7 @@ void fib_nhc_update_mtu(struct fib_nh_common *nhc, u32 new, u32 orig) + + void fib_sync_mtu(struct net_device *dev, u32 orig_mtu) + { +- unsigned int hash = fib_devindex_hashfn(dev->ifindex); +- struct hlist_head *head = &fib_info_devhash[hash]; ++ struct hlist_head *head = fib_info_devhash_bucket(dev); + struct fib_nh *nh; + + hlist_for_each_entry(nh, head, nh_hash) { +@@ -1979,12 +1985,11 @@ void fib_sync_mtu(struct net_device *dev, u32 orig_mtu) + */ + int fib_sync_down_dev(struct net_device *dev, unsigned long event, bool force) + { +- int ret = 0; +- int scope = RT_SCOPE_NOWHERE; ++ struct hlist_head *head = fib_info_devhash_bucket(dev); + struct fib_info *prev_fi = NULL; +- unsigned int hash = fib_devindex_hashfn(dev->ifindex); +- struct hlist_head *head = &fib_info_devhash[hash]; ++ int scope = RT_SCOPE_NOWHERE; + struct fib_nh *nh; ++ int ret = 0; + + if (force) + scope = -1; +@@ -2129,7 +2134,6 @@ out: + int fib_sync_up(struct net_device *dev, unsigned char nh_flags) + { + struct fib_info *prev_fi; +- unsigned int hash; + struct hlist_head *head; + struct fib_nh *nh; + int ret; +@@ -2145,8 +2149,7 @@ int fib_sync_up(struct net_device *dev, unsigned char nh_flags) + } + + prev_fi = NULL; +- hash = fib_devindex_hashfn(dev->ifindex); +- head = &fib_info_devhash[hash]; ++ head = fib_info_devhash_bucket(dev); + ret = 0; + + hlist_for_each_entry(nh, head, nh_hash) { +diff --git a/net/ipv4/inet_fragment.c b/net/ipv4/inet_fragment.c +index 05cd198d7a6ba..341096807100c 100644 +--- a/net/ipv4/inet_fragment.c ++++ b/net/ipv4/inet_fragment.c +@@ -235,9 +235,9 @@ void inet_frag_kill(struct inet_frag_queue *fq) + /* The RCU read lock provides a memory barrier + * guaranteeing that if fqdir->dead is false then + * the hash table destruction will not start until +- * after we unlock. Paired with inet_frags_exit_net(). ++ * after we unlock. Paired with fqdir_pre_exit(). + */ +- if (!fqdir->dead) { ++ if (!READ_ONCE(fqdir->dead)) { + rhashtable_remove_fast(&fqdir->rhashtable, &fq->node, + fqdir->f->rhash_params); + refcount_dec(&fq->refcnt); +@@ -352,9 +352,11 @@ static struct inet_frag_queue *inet_frag_create(struct fqdir *fqdir, + /* TODO : call from rcu_read_lock() and no longer use refcount_inc_not_zero() */ + struct inet_frag_queue *inet_frag_find(struct fqdir *fqdir, void *key) + { ++ /* This pairs with WRITE_ONCE() in fqdir_pre_exit(). */ ++ long high_thresh = READ_ONCE(fqdir->high_thresh); + struct inet_frag_queue *fq = NULL, *prev; + +- if (!fqdir->high_thresh || frag_mem_limit(fqdir) > fqdir->high_thresh) ++ if (!high_thresh || frag_mem_limit(fqdir) > high_thresh) + return NULL; + + rcu_read_lock(); +diff --git a/net/ipv4/ip_fragment.c b/net/ipv4/ip_fragment.c +index cfeb8890f94ee..fad803d2d711e 100644 +--- a/net/ipv4/ip_fragment.c ++++ b/net/ipv4/ip_fragment.c +@@ -144,7 +144,8 @@ static void ip_expire(struct timer_list *t) + + rcu_read_lock(); + +- if (qp->q.fqdir->dead) ++ /* Paired with WRITE_ONCE() in fqdir_pre_exit(). */ ++ if (READ_ONCE(qp->q.fqdir->dead)) + goto out_rcu_unlock; + + spin_lock(&qp->q.lock); +diff --git a/net/ipv4/ip_gre.c b/net/ipv4/ip_gre.c +index 0fe6c936dc54a..e7f3e37e4aa83 100644 +--- a/net/ipv4/ip_gre.c ++++ b/net/ipv4/ip_gre.c +@@ -604,8 +604,9 @@ static int gre_fill_metadata_dst(struct net_device *dev, struct sk_buff *skb) + + key = &info->key; + ip_tunnel_init_flow(&fl4, IPPROTO_GRE, key->u.ipv4.dst, key->u.ipv4.src, +- tunnel_id_to_key32(key->tun_id), key->tos, 0, +- skb->mark, skb_get_hash(skb)); ++ tunnel_id_to_key32(key->tun_id), ++ key->tos & ~INET_ECN_MASK, 0, skb->mark, ++ skb_get_hash(skb)); + rt = ip_route_output_key(dev_net(dev), &fl4); + if (IS_ERR(rt)) + return PTR_ERR(rt); +diff --git a/net/ipv4/netfilter/ipt_CLUSTERIP.c b/net/ipv4/netfilter/ipt_CLUSTERIP.c +index 8fd1aba8af31c..b518f20c9a244 100644 +--- a/net/ipv4/netfilter/ipt_CLUSTERIP.c ++++ b/net/ipv4/netfilter/ipt_CLUSTERIP.c +@@ -520,8 +520,11 @@ static int clusterip_tg_check(const struct xt_tgchk_param *par) + if (IS_ERR(config)) + return PTR_ERR(config); + } +- } else if (memcmp(&config->clustermac, &cipinfo->clustermac, ETH_ALEN)) ++ } else if (memcmp(&config->clustermac, &cipinfo->clustermac, ETH_ALEN)) { ++ clusterip_config_entry_put(config); ++ clusterip_config_put(config); + return -EINVAL; ++ } + + ret = nf_ct_netns_get(par->net, par->family); + if (ret < 0) { +diff --git a/net/ipv4/tcp_bpf.c b/net/ipv4/tcp_bpf.c +index f70aa0932bd6c..9b9b02052fd36 100644 +--- a/net/ipv4/tcp_bpf.c ++++ b/net/ipv4/tcp_bpf.c +@@ -196,12 +196,39 @@ msg_bytes_ready: + long timeo; + int data; + ++ if (sock_flag(sk, SOCK_DONE)) ++ goto out; ++ ++ if (sk->sk_err) { ++ copied = sock_error(sk); ++ goto out; ++ } ++ ++ if (sk->sk_shutdown & RCV_SHUTDOWN) ++ goto out; ++ ++ if (sk->sk_state == TCP_CLOSE) { ++ copied = -ENOTCONN; ++ goto out; ++ } ++ + timeo = sock_rcvtimeo(sk, nonblock); ++ if (!timeo) { ++ copied = -EAGAIN; ++ goto out; ++ } ++ ++ if (signal_pending(current)) { ++ copied = sock_intr_errno(timeo); ++ goto out; ++ } ++ + data = tcp_msg_wait_data(sk, psock, timeo); + if (data && !sk_psock_queue_empty(psock)) + goto msg_bytes_ready; + copied = -EAGAIN; + } ++out: + release_sock(sk); + sk_psock_put(sk, psock); + return copied; +diff --git a/net/ipv6/icmp.c b/net/ipv6/icmp.c +index a7c31ab67c5d6..96c5cc0f30ceb 100644 +--- a/net/ipv6/icmp.c ++++ b/net/ipv6/icmp.c +@@ -57,6 +57,7 @@ + #include <net/protocol.h> + #include <net/raw.h> + #include <net/rawv6.h> ++#include <net/seg6.h> + #include <net/transp_v6.h> + #include <net/ip6_route.h> + #include <net/addrconf.h> +@@ -820,6 +821,7 @@ out_bh_enable: + + void icmpv6_notify(struct sk_buff *skb, u8 type, u8 code, __be32 info) + { ++ struct inet6_skb_parm *opt = IP6CB(skb); + const struct inet6_protocol *ipprot; + int inner_offset; + __be16 frag_off; +@@ -829,6 +831,8 @@ void icmpv6_notify(struct sk_buff *skb, u8 type, u8 code, __be32 info) + if (!pskb_may_pull(skb, sizeof(struct ipv6hdr))) + goto out; + ++ seg6_icmp_srh(skb, opt); ++ + nexthdr = ((struct ipv6hdr *)skb->data)->nexthdr; + if (ipv6_ext_hdr(nexthdr)) { + /* now skip over extension headers */ +@@ -853,7 +857,7 @@ void icmpv6_notify(struct sk_buff *skb, u8 type, u8 code, __be32 info) + + ipprot = rcu_dereference(inet6_protos[nexthdr]); + if (ipprot && ipprot->err_handler) +- ipprot->err_handler(skb, NULL, type, code, inner_offset, info); ++ ipprot->err_handler(skb, opt, type, code, inner_offset, info); + + raw6_icmp_error(skb, nexthdr, type, code, inner_offset, info); + return; +diff --git a/net/ipv6/ip6_gre.c b/net/ipv6/ip6_gre.c +index 3ad201d372d88..466a5610e3ca9 100644 +--- a/net/ipv6/ip6_gre.c ++++ b/net/ipv6/ip6_gre.c +@@ -755,6 +755,7 @@ static netdev_tx_t __gre6_xmit(struct sk_buff *skb, + fl6->daddr = key->u.ipv6.dst; + fl6->flowlabel = key->label; + fl6->flowi6_uid = sock_net_uid(dev_net(dev), NULL); ++ fl6->fl6_gre_key = tunnel_id_to_key32(key->tun_id); + + dsfield = key->tos; + flags = key->tun_flags & +@@ -990,6 +991,7 @@ static netdev_tx_t ip6erspan_tunnel_xmit(struct sk_buff *skb, + fl6.daddr = key->u.ipv6.dst; + fl6.flowlabel = key->label; + fl6.flowi6_uid = sock_net_uid(dev_net(dev), NULL); ++ fl6.fl6_gre_key = tunnel_id_to_key32(key->tun_id); + + dsfield = key->tos; + if (!(tun_info->key.tun_flags & TUNNEL_ERSPAN_OPT)) +@@ -1098,6 +1100,7 @@ static void ip6gre_tnl_link_config_common(struct ip6_tnl *t) + fl6->flowi6_oif = p->link; + fl6->flowlabel = 0; + fl6->flowi6_proto = IPPROTO_GRE; ++ fl6->fl6_gre_key = t->parms.o_key; + + if (!(p->flags&IP6_TNL_F_USE_ORIG_TCLASS)) + fl6->flowlabel |= IPV6_TCLASS_MASK & p->flowinfo; +@@ -1544,7 +1547,7 @@ static void ip6gre_fb_tunnel_init(struct net_device *dev) + static struct inet6_protocol ip6gre_protocol __read_mostly = { + .handler = gre_rcv, + .err_handler = ip6gre_err, +- .flags = INET6_PROTO_NOPOLICY|INET6_PROTO_FINAL, ++ .flags = INET6_PROTO_FINAL, + }; + + static void ip6gre_destroy_tunnels(struct net *net, struct list_head *head) +diff --git a/net/ipv6/seg6.c b/net/ipv6/seg6.c +index e412817fba2f3..fa6b64c95d3ae 100644 +--- a/net/ipv6/seg6.c ++++ b/net/ipv6/seg6.c +@@ -75,6 +75,65 @@ bool seg6_validate_srh(struct ipv6_sr_hdr *srh, int len, bool reduced) + return true; + } + ++struct ipv6_sr_hdr *seg6_get_srh(struct sk_buff *skb, int flags) ++{ ++ struct ipv6_sr_hdr *srh; ++ int len, srhoff = 0; ++ ++ if (ipv6_find_hdr(skb, &srhoff, IPPROTO_ROUTING, NULL, &flags) < 0) ++ return NULL; ++ ++ if (!pskb_may_pull(skb, srhoff + sizeof(*srh))) ++ return NULL; ++ ++ srh = (struct ipv6_sr_hdr *)(skb->data + srhoff); ++ ++ len = (srh->hdrlen + 1) << 3; ++ ++ if (!pskb_may_pull(skb, srhoff + len)) ++ return NULL; ++ ++ /* note that pskb_may_pull may change pointers in header; ++ * for this reason it is necessary to reload them when needed. ++ */ ++ srh = (struct ipv6_sr_hdr *)(skb->data + srhoff); ++ ++ if (!seg6_validate_srh(srh, len, true)) ++ return NULL; ++ ++ return srh; ++} ++ ++/* Determine if an ICMP invoking packet contains a segment routing ++ * header. If it does, extract the offset to the true destination ++ * address, which is in the first segment address. ++ */ ++void seg6_icmp_srh(struct sk_buff *skb, struct inet6_skb_parm *opt) ++{ ++ __u16 network_header = skb->network_header; ++ struct ipv6_sr_hdr *srh; ++ ++ /* Update network header to point to the invoking packet ++ * inside the ICMP packet, so we can use the seg6_get_srh() ++ * helper. ++ */ ++ skb_reset_network_header(skb); ++ ++ srh = seg6_get_srh(skb, 0); ++ if (!srh) ++ goto out; ++ ++ if (srh->type != IPV6_SRCRT_TYPE_4) ++ goto out; ++ ++ opt->flags |= IP6SKB_SEG6; ++ opt->srhoff = (unsigned char *)srh - skb->data; ++ ++out: ++ /* Restore the network header back to the ICMP packet */ ++ skb->network_header = network_header; ++} ++ + static struct genl_family seg6_genl_family; + + static const struct nla_policy seg6_genl_policy[SEG6_ATTR_MAX + 1] = { +diff --git a/net/ipv6/seg6_local.c b/net/ipv6/seg6_local.c +index 2dc40b3f373ef..ef88489c71f52 100644 +--- a/net/ipv6/seg6_local.c ++++ b/net/ipv6/seg6_local.c +@@ -150,40 +150,11 @@ static struct seg6_local_lwt *seg6_local_lwtunnel(struct lwtunnel_state *lwt) + return (struct seg6_local_lwt *)lwt->data; + } + +-static struct ipv6_sr_hdr *get_srh(struct sk_buff *skb, int flags) +-{ +- struct ipv6_sr_hdr *srh; +- int len, srhoff = 0; +- +- if (ipv6_find_hdr(skb, &srhoff, IPPROTO_ROUTING, NULL, &flags) < 0) +- return NULL; +- +- if (!pskb_may_pull(skb, srhoff + sizeof(*srh))) +- return NULL; +- +- srh = (struct ipv6_sr_hdr *)(skb->data + srhoff); +- +- len = (srh->hdrlen + 1) << 3; +- +- if (!pskb_may_pull(skb, srhoff + len)) +- return NULL; +- +- /* note that pskb_may_pull may change pointers in header; +- * for this reason it is necessary to reload them when needed. +- */ +- srh = (struct ipv6_sr_hdr *)(skb->data + srhoff); +- +- if (!seg6_validate_srh(srh, len, true)) +- return NULL; +- +- return srh; +-} +- + static struct ipv6_sr_hdr *get_and_validate_srh(struct sk_buff *skb) + { + struct ipv6_sr_hdr *srh; + +- srh = get_srh(skb, IP6_FH_F_SKIP_RH); ++ srh = seg6_get_srh(skb, IP6_FH_F_SKIP_RH); + if (!srh) + return NULL; + +@@ -200,7 +171,7 @@ static bool decap_and_validate(struct sk_buff *skb, int proto) + struct ipv6_sr_hdr *srh; + unsigned int off = 0; + +- srh = get_srh(skb, 0); ++ srh = seg6_get_srh(skb, 0); + if (srh && srh->segments_left > 0) + return false; + +diff --git a/net/ipv6/udp.c b/net/ipv6/udp.c +index 8cd8c0bce0986..932c6f2a54946 100644 +--- a/net/ipv6/udp.c ++++ b/net/ipv6/udp.c +@@ -40,6 +40,7 @@ + #include <net/transp_v6.h> + #include <net/ip6_route.h> + #include <net/raw.h> ++#include <net/seg6.h> + #include <net/tcp_states.h> + #include <net/ip6_checksum.h> + #include <net/ip6_tunnel.h> +@@ -561,7 +562,7 @@ int __udp6_lib_err(struct sk_buff *skb, struct inet6_skb_parm *opt, + struct ipv6_pinfo *np; + const struct ipv6hdr *hdr = (const struct ipv6hdr *)skb->data; + const struct in6_addr *saddr = &hdr->saddr; +- const struct in6_addr *daddr = &hdr->daddr; ++ const struct in6_addr *daddr = seg6_get_daddr(skb, opt) ? : &hdr->daddr; + struct udphdr *uh = (struct udphdr *)(skb->data+offset); + bool tunnel = false; + struct sock *sk; +diff --git a/net/mac80211/rx.c b/net/mac80211/rx.c +index 1958e4d59b524..92ce173dd0c13 100644 +--- a/net/mac80211/rx.c ++++ b/net/mac80211/rx.c +@@ -4933,7 +4933,7 @@ void ieee80211_rx_list(struct ieee80211_hw *hw, struct ieee80211_sta *pubsta, + goto drop; + break; + case RX_ENC_VHT: +- if (WARN_ONCE(status->rate_idx > 9 || ++ if (WARN_ONCE(status->rate_idx > 11 || + !status->nss || + status->nss > 8, + "Rate marked as a VHT rate but data is invalid: MCS: %d, NSS: %d\n", +diff --git a/net/mptcp/options.c b/net/mptcp/options.c +index 0966855a7c251..e515ba9ccb5d8 100644 +--- a/net/mptcp/options.c ++++ b/net/mptcp/options.c +@@ -823,10 +823,13 @@ bool mptcp_established_options(struct sock *sk, struct sk_buff *skb, + if (mptcp_established_options_mp(sk, skb, snd_data_fin, &opt_size, remaining, opts)) + ret = true; + else if (mptcp_established_options_dss(sk, skb, snd_data_fin, &opt_size, remaining, opts)) { ++ unsigned int mp_fail_size; ++ + ret = true; +- if (mptcp_established_options_mp_fail(sk, &opt_size, remaining, opts)) { +- *size += opt_size; +- remaining -= opt_size; ++ if (mptcp_established_options_mp_fail(sk, &mp_fail_size, ++ remaining - opt_size, opts)) { ++ *size += opt_size + mp_fail_size; ++ remaining -= opt_size - mp_fail_size; + return true; + } + } +@@ -1318,6 +1321,7 @@ void mptcp_write_options(__be32 *ptr, const struct tcp_sock *tp, + put_unaligned_be32(mpext->data_len << 16 | + TCPOPT_NOP << 8 | TCPOPT_NOP, ptr); + } ++ ptr += 1; + } + } else if ((OPTION_MPTCP_MPC_SYN | OPTION_MPTCP_MPC_SYNACK | + OPTION_MPTCP_MPC_ACK) & opts->suboptions) { +diff --git a/net/mptcp/pm_netlink.c b/net/mptcp/pm_netlink.c +index b79251a36dcbc..d96860053816a 100644 +--- a/net/mptcp/pm_netlink.c ++++ b/net/mptcp/pm_netlink.c +@@ -710,6 +710,8 @@ static void mptcp_pm_nl_rm_addr_or_subflow(struct mptcp_sock *msk, + return; + + for (i = 0; i < rm_list->nr; i++) { ++ bool removed = false; ++ + list_for_each_entry_safe(subflow, tmp, &msk->conn_list, node) { + struct sock *ssk = mptcp_subflow_tcp_sock(subflow); + int how = RCV_SHUTDOWN | SEND_SHUTDOWN; +@@ -729,15 +731,19 @@ static void mptcp_pm_nl_rm_addr_or_subflow(struct mptcp_sock *msk, + mptcp_close_ssk(sk, ssk, subflow); + spin_lock_bh(&msk->pm.lock); + +- if (rm_type == MPTCP_MIB_RMADDR) { +- msk->pm.add_addr_accepted--; +- WRITE_ONCE(msk->pm.accept_addr, true); +- } else if (rm_type == MPTCP_MIB_RMSUBFLOW) { +- msk->pm.local_addr_used--; +- } ++ removed = true; + msk->pm.subflows--; + __MPTCP_INC_STATS(sock_net(sk), rm_type); + } ++ if (!removed) ++ continue; ++ ++ if (rm_type == MPTCP_MIB_RMADDR) { ++ msk->pm.add_addr_accepted--; ++ WRITE_ONCE(msk->pm.accept_addr, true); ++ } else if (rm_type == MPTCP_MIB_RMSUBFLOW) { ++ msk->pm.local_addr_used--; ++ } + } + } + +diff --git a/net/netfilter/nft_payload.c b/net/netfilter/nft_payload.c +index a44b14f6c0dc0..132875cd7fff2 100644 +--- a/net/netfilter/nft_payload.c ++++ b/net/netfilter/nft_payload.c +@@ -502,6 +502,9 @@ static int nft_payload_l4csum_offset(const struct nft_pktinfo *pkt, + struct sk_buff *skb, + unsigned int *l4csum_offset) + { ++ if (pkt->fragoff) ++ return -1; ++ + switch (pkt->tprot) { + case IPPROTO_TCP: + *l4csum_offset = offsetof(struct tcphdr, check); +diff --git a/net/netfilter/nft_set_pipapo.c b/net/netfilter/nft_set_pipapo.c +index dce866d93feed..2c8051d8cca69 100644 +--- a/net/netfilter/nft_set_pipapo.c ++++ b/net/netfilter/nft_set_pipapo.c +@@ -1290,6 +1290,11 @@ static struct nft_pipapo_match *pipapo_clone(struct nft_pipapo_match *old) + if (!new->scratch_aligned) + goto out_scratch; + #endif ++ for_each_possible_cpu(i) ++ *per_cpu_ptr(new->scratch, i) = NULL; ++ ++ if (pipapo_realloc_scratch(new, old->bsize_max)) ++ goto out_scratch_realloc; + + rcu_head_init(&new->rcu); + +@@ -1334,6 +1339,9 @@ out_lt: + kvfree(dst->lt); + dst--; + } ++out_scratch_realloc: ++ for_each_possible_cpu(i) ++ kfree(*per_cpu_ptr(new->scratch, i)); + #ifdef NFT_PIPAPO_ALIGN + free_percpu(new->scratch_aligned); + #endif +diff --git a/net/netrom/af_netrom.c b/net/netrom/af_netrom.c +index eef0e3f2f25b0..e5c8a295e6406 100644 +--- a/net/netrom/af_netrom.c ++++ b/net/netrom/af_netrom.c +@@ -298,7 +298,7 @@ static int nr_setsockopt(struct socket *sock, int level, int optname, + { + struct sock *sk = sock->sk; + struct nr_sock *nr = nr_sk(sk); +- unsigned long opt; ++ unsigned int opt; + + if (level != SOL_NETROM) + return -ENOPROTOOPT; +@@ -306,18 +306,18 @@ static int nr_setsockopt(struct socket *sock, int level, int optname, + if (optlen < sizeof(unsigned int)) + return -EINVAL; + +- if (copy_from_sockptr(&opt, optval, sizeof(unsigned long))) ++ if (copy_from_sockptr(&opt, optval, sizeof(opt))) + return -EFAULT; + + switch (optname) { + case NETROM_T1: +- if (opt < 1 || opt > ULONG_MAX / HZ) ++ if (opt < 1 || opt > UINT_MAX / HZ) + return -EINVAL; + nr->t1 = opt * HZ; + return 0; + + case NETROM_T2: +- if (opt < 1 || opt > ULONG_MAX / HZ) ++ if (opt < 1 || opt > UINT_MAX / HZ) + return -EINVAL; + nr->t2 = opt * HZ; + return 0; +@@ -329,13 +329,13 @@ static int nr_setsockopt(struct socket *sock, int level, int optname, + return 0; + + case NETROM_T4: +- if (opt < 1 || opt > ULONG_MAX / HZ) ++ if (opt < 1 || opt > UINT_MAX / HZ) + return -EINVAL; + nr->t4 = opt * HZ; + return 0; + + case NETROM_IDLE: +- if (opt > ULONG_MAX / (60 * HZ)) ++ if (opt > UINT_MAX / (60 * HZ)) + return -EINVAL; + nr->idle = opt * 60 * HZ; + return 0; +diff --git a/net/nfc/llcp_sock.c b/net/nfc/llcp_sock.c +index 6cfd30fc07985..0b93a17b9f11f 100644 +--- a/net/nfc/llcp_sock.c ++++ b/net/nfc/llcp_sock.c +@@ -789,6 +789,11 @@ static int llcp_sock_sendmsg(struct socket *sock, struct msghdr *msg, + + lock_sock(sk); + ++ if (!llcp_sock->local) { ++ release_sock(sk); ++ return -ENODEV; ++ } ++ + if (sk->sk_type == SOCK_DGRAM) { + DECLARE_SOCKADDR(struct sockaddr_nfc_llcp *, addr, + msg->msg_name); +diff --git a/net/openvswitch/flow.c b/net/openvswitch/flow.c +index 9713035b89e3a..02096f2ec6784 100644 +--- a/net/openvswitch/flow.c ++++ b/net/openvswitch/flow.c +@@ -34,6 +34,7 @@ + #include <net/mpls.h> + #include <net/ndisc.h> + #include <net/nsh.h> ++#include <net/netfilter/nf_conntrack_zones.h> + + #include "conntrack.h" + #include "datapath.h" +@@ -858,8 +859,9 @@ int ovs_flow_key_extract(const struct ip_tunnel_info *tun_info, + #if IS_ENABLED(CONFIG_NET_TC_SKB_EXT) + struct tc_skb_ext *tc_ext; + #endif +- bool post_ct = false; ++ bool post_ct = false, post_ct_snat = false, post_ct_dnat = false; + int res, err; ++ u16 zone = 0; + + /* Extract metadata from packet. */ + if (tun_info) { +@@ -898,6 +900,9 @@ int ovs_flow_key_extract(const struct ip_tunnel_info *tun_info, + key->recirc_id = tc_ext ? tc_ext->chain : 0; + OVS_CB(skb)->mru = tc_ext ? tc_ext->mru : 0; + post_ct = tc_ext ? tc_ext->post_ct : false; ++ post_ct_snat = post_ct ? tc_ext->post_ct_snat : false; ++ post_ct_dnat = post_ct ? tc_ext->post_ct_dnat : false; ++ zone = post_ct ? tc_ext->zone : 0; + } else { + key->recirc_id = 0; + } +@@ -906,8 +911,19 @@ int ovs_flow_key_extract(const struct ip_tunnel_info *tun_info, + #endif + + err = key_extract(skb, key); +- if (!err) ++ if (!err) { + ovs_ct_fill_key(skb, key, post_ct); /* Must be after key_extract(). */ ++ if (post_ct) { ++ if (!skb_get_nfct(skb)) { ++ key->ct_zone = zone; ++ } else { ++ if (!post_ct_dnat) ++ key->ct_state &= ~OVS_CS_F_DST_NAT; ++ if (!post_ct_snat) ++ key->ct_state &= ~OVS_CS_F_SRC_NAT; ++ } ++ } ++ } + return err; + } + +diff --git a/net/sched/act_ct.c b/net/sched/act_ct.c +index 98e248b9c0b17..2a17eb77c9049 100644 +--- a/net/sched/act_ct.c ++++ b/net/sched/act_ct.c +@@ -839,6 +839,12 @@ static int ct_nat_execute(struct sk_buff *skb, struct nf_conn *ct, + } + + err = nf_nat_packet(ct, ctinfo, hooknum, skb); ++ if (err == NF_ACCEPT) { ++ if (maniptype == NF_NAT_MANIP_SRC) ++ tc_skb_cb(skb)->post_ct_snat = 1; ++ if (maniptype == NF_NAT_MANIP_DST) ++ tc_skb_cb(skb)->post_ct_dnat = 1; ++ } + out: + return err; + } +@@ -1049,6 +1055,7 @@ out_push: + skb_push_rcsum(skb, nh_ofs); + + tc_skb_cb(skb)->post_ct = true; ++ tc_skb_cb(skb)->zone = p->zone; + out_clear: + if (defrag) + qdisc_skb_cb(skb)->pkt_len = skb->len; +diff --git a/net/sched/cls_api.c b/net/sched/cls_api.c +index ff8a9383bf1c4..cc9409aa755eb 100644 +--- a/net/sched/cls_api.c ++++ b/net/sched/cls_api.c +@@ -1625,6 +1625,9 @@ int tcf_classify(struct sk_buff *skb, + ext->chain = last_executed_chain; + ext->mru = cb->mru; + ext->post_ct = cb->post_ct; ++ ext->post_ct_snat = cb->post_ct_snat; ++ ext->post_ct_dnat = cb->post_ct_dnat; ++ ext->zone = cb->zone; + } + + return ret; +diff --git a/net/sched/cls_flower.c b/net/sched/cls_flower.c +index 161bd91c8c6b0..709348262410c 100644 +--- a/net/sched/cls_flower.c ++++ b/net/sched/cls_flower.c +@@ -311,6 +311,7 @@ static int fl_classify(struct sk_buff *skb, const struct tcf_proto *tp, + { + struct cls_fl_head *head = rcu_dereference_bh(tp->root); + bool post_ct = tc_skb_cb(skb)->post_ct; ++ u16 zone = tc_skb_cb(skb)->zone; + struct fl_flow_key skb_key; + struct fl_flow_mask *mask; + struct cls_fl_filter *f; +@@ -328,7 +329,7 @@ static int fl_classify(struct sk_buff *skb, const struct tcf_proto *tp, + skb_flow_dissect_ct(skb, &mask->dissector, &skb_key, + fl_ct_info_to_flower_map, + ARRAY_SIZE(fl_ct_info_to_flower_map), +- post_ct); ++ post_ct, zone); + skb_flow_dissect_hash(skb, &mask->dissector, &skb_key); + skb_flow_dissect(skb, &mask->dissector, &skb_key, 0); + +diff --git a/net/sched/sch_api.c b/net/sched/sch_api.c +index 12f39a2dffd47..4bbfd26223274 100644 +--- a/net/sched/sch_api.c ++++ b/net/sched/sch_api.c +@@ -1062,7 +1062,7 @@ static int qdisc_graft(struct net_device *dev, struct Qdisc *parent, + + qdisc_offload_graft_root(dev, new, old, extack); + +- if (new && new->ops->attach) ++ if (new && new->ops->attach && !ingress) + goto skip; + + for (i = 0; i < num_q; i++) { +diff --git a/net/sched/sch_generic.c b/net/sched/sch_generic.c +index 66d2fbe9ef501..47ca76ba7ffa8 100644 +--- a/net/sched/sch_generic.c ++++ b/net/sched/sch_generic.c +@@ -1455,6 +1455,7 @@ void psched_ratecfg_precompute(struct psched_ratecfg *r, + { + memset(r, 0, sizeof(*r)); + r->overhead = conf->overhead; ++ r->mpu = conf->mpu; + r->rate_bytes_ps = max_t(u64, conf->rate, rate64); + r->linklayer = (conf->linklayer & TC_LINKLAYER_MASK); + psched_ratecfg_precompute__(r->rate_bytes_ps, &r->mult, &r->shift); +diff --git a/net/smc/af_smc.c b/net/smc/af_smc.c +index eea6d4a854e90..07ff719f39077 100644 +--- a/net/smc/af_smc.c ++++ b/net/smc/af_smc.c +@@ -613,10 +613,12 @@ static int smc_connect_decline_fallback(struct smc_sock *smc, int reason_code, + + static void smc_conn_abort(struct smc_sock *smc, int local_first) + { ++ struct smc_connection *conn = &smc->conn; ++ struct smc_link_group *lgr = conn->lgr; ++ ++ smc_conn_free(conn); + if (local_first) +- smc_lgr_cleanup_early(&smc->conn); +- else +- smc_conn_free(&smc->conn); ++ smc_lgr_cleanup_early(lgr); + } + + /* check if there is a rdma device available for this connection. */ +diff --git a/net/smc/smc_core.c b/net/smc/smc_core.c +index 506b8498623b0..36e93a3f284d0 100644 +--- a/net/smc/smc_core.c ++++ b/net/smc/smc_core.c +@@ -170,8 +170,10 @@ static int smc_lgr_register_conn(struct smc_connection *conn, bool first) + + if (!conn->lgr->is_smcd) { + rc = smcr_lgr_conn_assign_link(conn, first); +- if (rc) ++ if (rc) { ++ conn->lgr = NULL; + return rc; ++ } + } + /* find a new alert_token_local value not yet used by some connection + * in this link group +@@ -579,15 +581,13 @@ int smcd_nl_get_lgr(struct sk_buff *skb, struct netlink_callback *cb) + return skb->len; + } + +-void smc_lgr_cleanup_early(struct smc_connection *conn) ++void smc_lgr_cleanup_early(struct smc_link_group *lgr) + { +- struct smc_link_group *lgr = conn->lgr; + spinlock_t *lgr_lock; + + if (!lgr) + return; + +- smc_conn_free(conn); + smc_lgr_list_head(lgr, &lgr_lock); + spin_lock_bh(lgr_lock); + /* do not use this link group for new connections */ +@@ -1387,16 +1387,11 @@ void smc_smcd_terminate_all(struct smcd_dev *smcd) + /* Called when an SMCR device is removed or the smc module is unloaded. + * If smcibdev is given, all SMCR link groups using this device are terminated. + * If smcibdev is NULL, all SMCR link groups are terminated. +- * +- * We must wait here for QPs been destroyed before we destroy the CQs, +- * or we won't received any CQEs and cdc_pend_tx_wr cannot reach 0 thus +- * smc_sock cannot be released. + */ + void smc_smcr_terminate_all(struct smc_ib_device *smcibdev) + { + struct smc_link_group *lgr, *lg; + LIST_HEAD(lgr_free_list); +- LIST_HEAD(lgr_linkdown_list); + int i; + + spin_lock_bh(&smc_lgr_list.lock); +@@ -1408,7 +1403,7 @@ void smc_smcr_terminate_all(struct smc_ib_device *smcibdev) + list_for_each_entry_safe(lgr, lg, &smc_lgr_list.list, list) { + for (i = 0; i < SMC_LINKS_PER_LGR_MAX; i++) { + if (lgr->lnk[i].smcibdev == smcibdev) +- list_move_tail(&lgr->list, &lgr_linkdown_list); ++ smcr_link_down_cond_sched(&lgr->lnk[i]); + } + } + } +@@ -1420,16 +1415,6 @@ void smc_smcr_terminate_all(struct smc_ib_device *smcibdev) + __smc_lgr_terminate(lgr, false); + } + +- list_for_each_entry_safe(lgr, lg, &lgr_linkdown_list, list) { +- for (i = 0; i < SMC_LINKS_PER_LGR_MAX; i++) { +- if (lgr->lnk[i].smcibdev == smcibdev) { +- mutex_lock(&lgr->llc_conf_mutex); +- smcr_link_down_cond(&lgr->lnk[i]); +- mutex_unlock(&lgr->llc_conf_mutex); +- } +- } +- } +- + if (smcibdev) { + if (atomic_read(&smcibdev->lnk_cnt)) + wait_event(smcibdev->lnks_deleted, +@@ -1750,8 +1735,10 @@ create: + write_lock_bh(&lgr->conns_lock); + rc = smc_lgr_register_conn(conn, true); + write_unlock_bh(&lgr->conns_lock); +- if (rc) ++ if (rc) { ++ smc_lgr_cleanup_early(lgr); + goto out; ++ } + } + conn->local_tx_ctrl.common.type = SMC_CDC_MSG_TYPE; + conn->local_tx_ctrl.len = SMC_WR_TX_SIZE; +diff --git a/net/smc/smc_core.h b/net/smc/smc_core.h +index 51a3e8248ade2..9a0523f4c7ba6 100644 +--- a/net/smc/smc_core.h ++++ b/net/smc/smc_core.h +@@ -419,7 +419,7 @@ static inline void smc_set_pci_values(struct pci_dev *pci_dev, + struct smc_sock; + struct smc_clc_msg_accept_confirm; + +-void smc_lgr_cleanup_early(struct smc_connection *conn); ++void smc_lgr_cleanup_early(struct smc_link_group *lgr); + void smc_lgr_terminate_sched(struct smc_link_group *lgr); + void smcr_port_add(struct smc_ib_device *smcibdev, u8 ibport); + void smcr_port_err(struct smc_ib_device *smcibdev, u8 ibport); +diff --git a/net/socket.c b/net/socket.c +index 7f64a6eccf63f..5053eb0100e48 100644 +--- a/net/socket.c ++++ b/net/socket.c +@@ -829,6 +829,7 @@ void __sock_recv_timestamp(struct msghdr *msg, struct sock *sk, + int empty = 1, false_tstamp = 0; + struct skb_shared_hwtstamps *shhwtstamps = + skb_hwtstamps(skb); ++ ktime_t hwtstamp; + + /* Race occurred between timestamp enabling and packet + receiving. Fill in the current time for now. */ +@@ -877,10 +878,12 @@ void __sock_recv_timestamp(struct msghdr *msg, struct sock *sk, + (sk->sk_tsflags & SOF_TIMESTAMPING_RAW_HARDWARE) && + !skb_is_swtx_tstamp(skb, false_tstamp)) { + if (sk->sk_tsflags & SOF_TIMESTAMPING_BIND_PHC) +- ptp_convert_timestamp(shhwtstamps, sk->sk_bind_phc); ++ hwtstamp = ptp_convert_timestamp(shhwtstamps, ++ sk->sk_bind_phc); ++ else ++ hwtstamp = shhwtstamps->hwtstamp; + +- if (ktime_to_timespec64_cond(shhwtstamps->hwtstamp, +- tss.ts + 2)) { ++ if (ktime_to_timespec64_cond(hwtstamp, tss.ts + 2)) { + empty = 0; + + if ((sk->sk_tsflags & SOF_TIMESTAMPING_OPT_PKTINFO) && +diff --git a/net/unix/garbage.c b/net/unix/garbage.c +index 12e2ddaf887f2..d45d5366115a7 100644 +--- a/net/unix/garbage.c ++++ b/net/unix/garbage.c +@@ -192,8 +192,11 @@ void wait_for_unix_gc(void) + { + /* If number of inflight sockets is insane, + * force a garbage collect right now. ++ * Paired with the WRITE_ONCE() in unix_inflight(), ++ * unix_notinflight() and gc_in_progress(). + */ +- if (unix_tot_inflight > UNIX_INFLIGHT_TRIGGER_GC && !gc_in_progress) ++ if (READ_ONCE(unix_tot_inflight) > UNIX_INFLIGHT_TRIGGER_GC && ++ !READ_ONCE(gc_in_progress)) + unix_gc(); + wait_event(unix_gc_wait, gc_in_progress == false); + } +@@ -213,7 +216,9 @@ void unix_gc(void) + if (gc_in_progress) + goto out; + +- gc_in_progress = true; ++ /* Paired with READ_ONCE() in wait_for_unix_gc(). */ ++ WRITE_ONCE(gc_in_progress, true); ++ + /* First, select candidates for garbage collection. Only + * in-flight sockets are considered, and from those only ones + * which don't have any external reference. +@@ -299,7 +304,10 @@ void unix_gc(void) + + /* All candidates should have been detached by now. */ + BUG_ON(!list_empty(&gc_candidates)); +- gc_in_progress = false; ++ ++ /* Paired with READ_ONCE() in wait_for_unix_gc(). */ ++ WRITE_ONCE(gc_in_progress, false); ++ + wake_up(&unix_gc_wait); + + out: +diff --git a/net/unix/scm.c b/net/unix/scm.c +index 052ae709ce289..aa27a02478dc1 100644 +--- a/net/unix/scm.c ++++ b/net/unix/scm.c +@@ -60,7 +60,8 @@ void unix_inflight(struct user_struct *user, struct file *fp) + } else { + BUG_ON(list_empty(&u->link)); + } +- unix_tot_inflight++; ++ /* Paired with READ_ONCE() in wait_for_unix_gc() */ ++ WRITE_ONCE(unix_tot_inflight, unix_tot_inflight + 1); + } + user->unix_inflight++; + spin_unlock(&unix_gc_lock); +@@ -80,7 +81,8 @@ void unix_notinflight(struct user_struct *user, struct file *fp) + + if (atomic_long_dec_and_test(&u->inflight)) + list_del_init(&u->link); +- unix_tot_inflight--; ++ /* Paired with READ_ONCE() in wait_for_unix_gc() */ ++ WRITE_ONCE(unix_tot_inflight, unix_tot_inflight - 1); + } + user->unix_inflight--; + spin_unlock(&unix_gc_lock); +diff --git a/net/xfrm/xfrm_compat.c b/net/xfrm/xfrm_compat.c +index 2bf2693901631..a0f62fa02e06e 100644 +--- a/net/xfrm/xfrm_compat.c ++++ b/net/xfrm/xfrm_compat.c +@@ -127,6 +127,7 @@ static const struct nla_policy compat_policy[XFRMA_MAX+1] = { + [XFRMA_SET_MARK] = { .type = NLA_U32 }, + [XFRMA_SET_MARK_MASK] = { .type = NLA_U32 }, + [XFRMA_IF_ID] = { .type = NLA_U32 }, ++ [XFRMA_MTIMER_THRESH] = { .type = NLA_U32 }, + }; + + static struct nlmsghdr *xfrm_nlmsg_put_compat(struct sk_buff *skb, +@@ -274,9 +275,10 @@ static int xfrm_xlate64_attr(struct sk_buff *dst, const struct nlattr *src) + case XFRMA_SET_MARK: + case XFRMA_SET_MARK_MASK: + case XFRMA_IF_ID: ++ case XFRMA_MTIMER_THRESH: + return xfrm_nla_cpy(dst, src, nla_len(src)); + default: +- BUILD_BUG_ON(XFRMA_MAX != XFRMA_IF_ID); ++ BUILD_BUG_ON(XFRMA_MAX != XFRMA_MTIMER_THRESH); + pr_warn_once("unsupported nla_type %d\n", src->nla_type); + return -EOPNOTSUPP; + } +@@ -431,7 +433,7 @@ static int xfrm_xlate32_attr(void *dst, const struct nlattr *nla, + int err; + + if (type > XFRMA_MAX) { +- BUILD_BUG_ON(XFRMA_MAX != XFRMA_IF_ID); ++ BUILD_BUG_ON(XFRMA_MAX != XFRMA_MTIMER_THRESH); + NL_SET_ERR_MSG(extack, "Bad attribute"); + return -EOPNOTSUPP; + } +diff --git a/net/xfrm/xfrm_interface.c b/net/xfrm/xfrm_interface.c +index 41de46b5ffa94..57448fc519fcd 100644 +--- a/net/xfrm/xfrm_interface.c ++++ b/net/xfrm/xfrm_interface.c +@@ -637,11 +637,16 @@ static int xfrmi_newlink(struct net *src_net, struct net_device *dev, + struct netlink_ext_ack *extack) + { + struct net *net = dev_net(dev); +- struct xfrm_if_parms p; ++ struct xfrm_if_parms p = {}; + struct xfrm_if *xi; + int err; + + xfrmi_netlink_parms(data, &p); ++ if (!p.if_id) { ++ NL_SET_ERR_MSG(extack, "if_id must be non zero"); ++ return -EINVAL; ++ } ++ + xi = xfrmi_locate(net, &p); + if (xi) + return -EEXIST; +@@ -666,7 +671,12 @@ static int xfrmi_changelink(struct net_device *dev, struct nlattr *tb[], + { + struct xfrm_if *xi = netdev_priv(dev); + struct net *net = xi->net; +- struct xfrm_if_parms p; ++ struct xfrm_if_parms p = {}; ++ ++ if (!p.if_id) { ++ NL_SET_ERR_MSG(extack, "if_id must be non zero"); ++ return -EINVAL; ++ } + + xfrmi_netlink_parms(data, &p); + xi = xfrmi_locate(net, &p); +diff --git a/net/xfrm/xfrm_output.c b/net/xfrm/xfrm_output.c +index 229544bc70c21..4dc4a7bbe51cf 100644 +--- a/net/xfrm/xfrm_output.c ++++ b/net/xfrm/xfrm_output.c +@@ -647,10 +647,12 @@ static int xfrm_output_gso(struct net *net, struct sock *sk, struct sk_buff *skb + * This requires hardware to know the inner packet type to calculate + * the inner header checksum. Save inner ip protocol here to avoid + * traversing the packet in the vendor's xmit code. +- * If the encap type is IPIP, just save skb->inner_ipproto. Otherwise, +- * get the ip protocol from the IP header. ++ * For IPsec tunnel mode save the ip protocol from the IP header of the ++ * plain text packet. Otherwise If the encap type is IPIP, just save ++ * skb->inner_ipproto in any other case get the ip protocol from the IP ++ * header. + */ +-static void xfrm_get_inner_ipproto(struct sk_buff *skb) ++static void xfrm_get_inner_ipproto(struct sk_buff *skb, struct xfrm_state *x) + { + struct xfrm_offload *xo = xfrm_offload(skb); + const struct ethhdr *eth; +@@ -658,6 +660,25 @@ static void xfrm_get_inner_ipproto(struct sk_buff *skb) + if (!xo) + return; + ++ if (x->outer_mode.encap == XFRM_MODE_TUNNEL) { ++ switch (x->outer_mode.family) { ++ case AF_INET: ++ xo->inner_ipproto = ip_hdr(skb)->protocol; ++ break; ++ case AF_INET6: ++ xo->inner_ipproto = ipv6_hdr(skb)->nexthdr; ++ break; ++ default: ++ break; ++ } ++ ++ return; ++ } ++ ++ /* non-Tunnel Mode */ ++ if (!skb->encapsulation) ++ return; ++ + if (skb->inner_protocol_type == ENCAP_TYPE_IPPROTO) { + xo->inner_ipproto = skb->inner_ipproto; + return; +@@ -712,8 +733,7 @@ int xfrm_output(struct sock *sk, struct sk_buff *skb) + sp->xvec[sp->len++] = x; + xfrm_state_hold(x); + +- if (skb->encapsulation) +- xfrm_get_inner_ipproto(skb); ++ xfrm_get_inner_ipproto(skb, x); + skb->encapsulation = 1; + + if (skb_is_gso(skb)) { +diff --git a/net/xfrm/xfrm_policy.c b/net/xfrm/xfrm_policy.c +index 37d17a79617c9..37b149f632624 100644 +--- a/net/xfrm/xfrm_policy.c ++++ b/net/xfrm/xfrm_policy.c +@@ -31,8 +31,10 @@ + #include <linux/if_tunnel.h> + #include <net/dst.h> + #include <net/flow.h> ++#include <net/inet_ecn.h> + #include <net/xfrm.h> + #include <net/ip.h> ++#include <net/gre.h> + #if IS_ENABLED(CONFIG_IPV6_MIP6) + #include <net/mip6.h> + #endif +@@ -3296,7 +3298,7 @@ decode_session4(struct sk_buff *skb, struct flowi *fl, bool reverse) + fl4->flowi4_proto = iph->protocol; + fl4->daddr = reverse ? iph->saddr : iph->daddr; + fl4->saddr = reverse ? iph->daddr : iph->saddr; +- fl4->flowi4_tos = iph->tos; ++ fl4->flowi4_tos = iph->tos & ~INET_ECN_MASK; + + if (!ip_is_fragment(iph)) { + switch (iph->protocol) { +@@ -3424,6 +3426,26 @@ decode_session6(struct sk_buff *skb, struct flowi *fl, bool reverse) + } + fl6->flowi6_proto = nexthdr; + return; ++ case IPPROTO_GRE: ++ if (!onlyproto && ++ (nh + offset + 12 < skb->data || ++ pskb_may_pull(skb, nh + offset + 12 - skb->data))) { ++ struct gre_base_hdr *gre_hdr; ++ __be32 *gre_key; ++ ++ nh = skb_network_header(skb); ++ gre_hdr = (struct gre_base_hdr *)(nh + offset); ++ gre_key = (__be32 *)(gre_hdr + 1); ++ ++ if (gre_hdr->flags & GRE_KEY) { ++ if (gre_hdr->flags & GRE_CSUM) ++ gre_key++; ++ fl6->fl6_gre_key = *gre_key; ++ } ++ } ++ fl6->flowi6_proto = nexthdr; ++ return; ++ + #if IS_ENABLED(CONFIG_IPV6_MIP6) + case IPPROTO_MH: + offset += ipv6_optlen(exthdr); +diff --git a/net/xfrm/xfrm_state.c b/net/xfrm/xfrm_state.c +index a2f4001221d16..78d51399a0f4b 100644 +--- a/net/xfrm/xfrm_state.c ++++ b/net/xfrm/xfrm_state.c +@@ -1593,6 +1593,9 @@ static struct xfrm_state *xfrm_state_clone(struct xfrm_state *orig, + x->km.seq = orig->km.seq; + x->replay = orig->replay; + x->preplay = orig->preplay; ++ x->mapping_maxage = orig->mapping_maxage; ++ x->new_mapping = 0; ++ x->new_mapping_sport = 0; + + return x; + +@@ -2242,7 +2245,7 @@ int km_query(struct xfrm_state *x, struct xfrm_tmpl *t, struct xfrm_policy *pol) + } + EXPORT_SYMBOL(km_query); + +-int km_new_mapping(struct xfrm_state *x, xfrm_address_t *ipaddr, __be16 sport) ++static int __km_new_mapping(struct xfrm_state *x, xfrm_address_t *ipaddr, __be16 sport) + { + int err = -EINVAL; + struct xfrm_mgr *km; +@@ -2257,6 +2260,24 @@ int km_new_mapping(struct xfrm_state *x, xfrm_address_t *ipaddr, __be16 sport) + rcu_read_unlock(); + return err; + } ++ ++int km_new_mapping(struct xfrm_state *x, xfrm_address_t *ipaddr, __be16 sport) ++{ ++ int ret = 0; ++ ++ if (x->mapping_maxage) { ++ if ((jiffies / HZ - x->new_mapping) > x->mapping_maxage || ++ x->new_mapping_sport != sport) { ++ x->new_mapping_sport = sport; ++ x->new_mapping = jiffies / HZ; ++ ret = __km_new_mapping(x, ipaddr, sport); ++ } ++ } else { ++ ret = __km_new_mapping(x, ipaddr, sport); ++ } ++ ++ return ret; ++} + EXPORT_SYMBOL(km_new_mapping); + + void km_policy_expired(struct xfrm_policy *pol, int dir, int hard, u32 portid) +diff --git a/net/xfrm/xfrm_user.c b/net/xfrm/xfrm_user.c +index 3a3cb09eec122..b10f88822c0df 100644 +--- a/net/xfrm/xfrm_user.c ++++ b/net/xfrm/xfrm_user.c +@@ -282,6 +282,10 @@ static int verify_newsa_info(struct xfrm_usersa_info *p, + + err = 0; + ++ if (attrs[XFRMA_MTIMER_THRESH]) ++ if (!attrs[XFRMA_ENCAP]) ++ err = -EINVAL; ++ + out: + return err; + } +@@ -521,6 +525,7 @@ static void xfrm_update_ae_params(struct xfrm_state *x, struct nlattr **attrs, + struct nlattr *lt = attrs[XFRMA_LTIME_VAL]; + struct nlattr *et = attrs[XFRMA_ETIMER_THRESH]; + struct nlattr *rt = attrs[XFRMA_REPLAY_THRESH]; ++ struct nlattr *mt = attrs[XFRMA_MTIMER_THRESH]; + + if (re) { + struct xfrm_replay_state_esn *replay_esn; +@@ -552,6 +557,9 @@ static void xfrm_update_ae_params(struct xfrm_state *x, struct nlattr **attrs, + + if (rt) + x->replay_maxdiff = nla_get_u32(rt); ++ ++ if (mt) ++ x->mapping_maxage = nla_get_u32(mt); + } + + static void xfrm_smark_init(struct nlattr **attrs, struct xfrm_mark *m) +@@ -621,8 +629,13 @@ static struct xfrm_state *xfrm_state_construct(struct net *net, + + xfrm_smark_init(attrs, &x->props.smark); + +- if (attrs[XFRMA_IF_ID]) ++ if (attrs[XFRMA_IF_ID]) { + x->if_id = nla_get_u32(attrs[XFRMA_IF_ID]); ++ if (!x->if_id) { ++ err = -EINVAL; ++ goto error; ++ } ++ } + + err = __xfrm_init_state(x, false, attrs[XFRMA_OFFLOAD_DEV]); + if (err) +@@ -1024,8 +1037,13 @@ static int copy_to_user_state_extra(struct xfrm_state *x, + if (ret) + goto out; + } +- if (x->security) ++ if (x->security) { + ret = copy_sec_ctx(x->security, skb); ++ if (ret) ++ goto out; ++ } ++ if (x->mapping_maxage) ++ ret = nla_put_u32(skb, XFRMA_MTIMER_THRESH, x->mapping_maxage); + out: + return ret; + } +@@ -1413,8 +1431,13 @@ static int xfrm_alloc_userspi(struct sk_buff *skb, struct nlmsghdr *nlh, + + mark = xfrm_mark_get(attrs, &m); + +- if (attrs[XFRMA_IF_ID]) ++ if (attrs[XFRMA_IF_ID]) { + if_id = nla_get_u32(attrs[XFRMA_IF_ID]); ++ if (!if_id) { ++ err = -EINVAL; ++ goto out_noput; ++ } ++ } + + if (p->info.seq) { + x = xfrm_find_acq_byseq(net, mark, p->info.seq); +@@ -1727,8 +1750,13 @@ static struct xfrm_policy *xfrm_policy_construct(struct net *net, struct xfrm_us + + xfrm_mark_get(attrs, &xp->mark); + +- if (attrs[XFRMA_IF_ID]) ++ if (attrs[XFRMA_IF_ID]) { + xp->if_id = nla_get_u32(attrs[XFRMA_IF_ID]); ++ if (!xp->if_id) { ++ err = -EINVAL; ++ goto error; ++ } ++ } + + return xp; + error: +@@ -3058,7 +3086,7 @@ static inline unsigned int xfrm_sa_len(struct xfrm_state *x) + if (x->props.extra_flags) + l += nla_total_size(sizeof(x->props.extra_flags)); + if (x->xso.dev) +- l += nla_total_size(sizeof(x->xso)); ++ l += nla_total_size(sizeof(struct xfrm_user_offload)); + if (x->props.smark.v | x->props.smark.m) { + l += nla_total_size(sizeof(x->props.smark.v)); + l += nla_total_size(sizeof(x->props.smark.m)); +@@ -3069,6 +3097,9 @@ static inline unsigned int xfrm_sa_len(struct xfrm_state *x) + /* Must count x->lastused as it may become non-zero behind our back. */ + l += nla_total_size_64bit(sizeof(u64)); + ++ if (x->mapping_maxage) ++ l += nla_total_size(sizeof(x->mapping_maxage)); ++ + return l; + } + +diff --git a/samples/bpf/Makefile b/samples/bpf/Makefile +index 5fd48a8d4f10a..c6e38e43c3fd0 100644 +--- a/samples/bpf/Makefile ++++ b/samples/bpf/Makefile +@@ -59,7 +59,11 @@ tprogs-y += xdp_redirect + tprogs-y += xdp_monitor + + # Libbpf dependencies +-LIBBPF = $(TOOLS_PATH)/lib/bpf/libbpf.a ++LIBBPF_SRC = $(TOOLS_PATH)/lib/bpf ++LIBBPF_OUTPUT = $(abspath $(BPF_SAMPLES_PATH))/libbpf ++LIBBPF_DESTDIR = $(LIBBPF_OUTPUT) ++LIBBPF_INCLUDE = $(LIBBPF_DESTDIR)/include ++LIBBPF = $(LIBBPF_OUTPUT)/libbpf.a + + CGROUP_HELPERS := ../../tools/testing/selftests/bpf/cgroup_helpers.o + TRACE_HELPERS := ../../tools/testing/selftests/bpf/trace_helpers.o +@@ -198,7 +202,7 @@ TPROGS_CFLAGS += -Wstrict-prototypes + + TPROGS_CFLAGS += -I$(objtree)/usr/include + TPROGS_CFLAGS += -I$(srctree)/tools/testing/selftests/bpf/ +-TPROGS_CFLAGS += -I$(srctree)/tools/lib/ ++TPROGS_CFLAGS += -I$(LIBBPF_INCLUDE) + TPROGS_CFLAGS += -I$(srctree)/tools/include + TPROGS_CFLAGS += -I$(srctree)/tools/perf + TPROGS_CFLAGS += -DHAVE_ATTR_TEST=0 +@@ -209,6 +213,11 @@ TPROGS_LDFLAGS := -L$(SYSROOT)/usr/lib + endif + + TPROGS_LDLIBS += $(LIBBPF) -lelf -lz ++TPROGLDLIBS_xdp_monitor += -lm ++TPROGLDLIBS_xdp_redirect += -lm ++TPROGLDLIBS_xdp_redirect_cpu += -lm ++TPROGLDLIBS_xdp_redirect_map += -lm ++TPROGLDLIBS_xdp_redirect_map_multi += -lm + TPROGLDLIBS_tracex4 += -lrt + TPROGLDLIBS_trace_output += -lrt + TPROGLDLIBS_map_perf_test += -lrt +@@ -268,16 +277,27 @@ all: + clean: + $(MAKE) -C ../../ M=$(CURDIR) clean + @find $(CURDIR) -type f -name '*~' -delete ++ @$(RM) -r $(CURDIR)/libbpf $(CURDIR)/bpftool + +-$(LIBBPF): FORCE ++$(LIBBPF): FORCE | $(LIBBPF_OUTPUT) + # Fix up variables inherited from Kbuild that tools/ build system won't like +- $(MAKE) -C $(dir $@) RM='rm -rf' EXTRA_CFLAGS="$(TPROGS_CFLAGS)" \ +- LDFLAGS=$(TPROGS_LDFLAGS) srctree=$(BPF_SAMPLES_PATH)/../../ O= ++ $(MAKE) -C $(LIBBPF_SRC) RM='rm -rf' EXTRA_CFLAGS="$(TPROGS_CFLAGS)" \ ++ LDFLAGS=$(TPROGS_LDFLAGS) srctree=$(BPF_SAMPLES_PATH)/../../ \ ++ O= OUTPUT=$(LIBBPF_OUTPUT)/ DESTDIR=$(LIBBPF_DESTDIR) prefix= \ ++ $@ install_headers + + BPFTOOLDIR := $(TOOLS_PATH)/bpf/bpftool +-BPFTOOL := $(BPFTOOLDIR)/bpftool +-$(BPFTOOL): $(wildcard $(BPFTOOLDIR)/*.[ch] $(BPFTOOLDIR)/Makefile) +- $(MAKE) -C $(BPFTOOLDIR) srctree=$(BPF_SAMPLES_PATH)/../../ ++BPFTOOL_OUTPUT := $(abspath $(BPF_SAMPLES_PATH))/bpftool ++BPFTOOL := $(BPFTOOL_OUTPUT)/bpftool ++$(BPFTOOL): $(LIBBPF) $(wildcard $(BPFTOOLDIR)/*.[ch] $(BPFTOOLDIR)/Makefile) | $(BPFTOOL_OUTPUT) ++ $(MAKE) -C $(BPFTOOLDIR) srctree=$(BPF_SAMPLES_PATH)/../../ \ ++ OUTPUT=$(BPFTOOL_OUTPUT)/ \ ++ LIBBPF_OUTPUT=$(LIBBPF_OUTPUT)/ \ ++ LIBBPF_DESTDIR=$(LIBBPF_DESTDIR)/ ++ ++$(LIBBPF_OUTPUT) $(BPFTOOL_OUTPUT): ++ $(call msg,MKDIR,$@) ++ $(Q)mkdir -p $@ + + $(obj)/syscall_nrs.h: $(obj)/syscall_nrs.s FORCE + $(call filechk,offsets,__SYSCALL_NRS_H__) +@@ -309,6 +329,11 @@ verify_target_bpf: verify_cmds + $(BPF_SAMPLES_PATH)/*.c: verify_target_bpf $(LIBBPF) + $(src)/*.c: verify_target_bpf $(LIBBPF) + ++libbpf_hdrs: $(LIBBPF) ++$(obj)/$(TRACE_HELPERS) $(obj)/$(CGROUP_HELPERS) $(obj)/$(XDP_SAMPLE): | libbpf_hdrs ++ ++.PHONY: libbpf_hdrs ++ + $(obj)/xdp_redirect_cpu_user.o: $(obj)/xdp_redirect_cpu.skel.h + $(obj)/xdp_redirect_map_multi_user.o: $(obj)/xdp_redirect_map_multi.skel.h + $(obj)/xdp_redirect_map_user.o: $(obj)/xdp_redirect_map.skel.h +@@ -320,6 +345,17 @@ $(obj)/hbm_out_kern.o: $(src)/hbm.h $(src)/hbm_kern.h + $(obj)/hbm.o: $(src)/hbm.h + $(obj)/hbm_edt_kern.o: $(src)/hbm.h $(src)/hbm_kern.h + ++# Override includes for xdp_sample_user.o because $(srctree)/usr/include in ++# TPROGS_CFLAGS causes conflicts ++XDP_SAMPLE_CFLAGS += -Wall -O2 \ ++ -I$(src)/../../tools/include \ ++ -I$(src)/../../tools/include/uapi \ ++ -I$(LIBBPF_INCLUDE) \ ++ -I$(src)/../../tools/testing/selftests/bpf ++ ++$(obj)/$(XDP_SAMPLE): TPROGS_CFLAGS = $(XDP_SAMPLE_CFLAGS) ++$(obj)/$(XDP_SAMPLE): $(src)/xdp_sample_user.h $(src)/xdp_sample_shared.h ++ + -include $(BPF_SAMPLES_PATH)/Makefile.target + + VMLINUX_BTF_PATHS ?= $(abspath $(if $(O),$(O)/vmlinux)) \ +@@ -366,7 +402,7 @@ $(obj)/%.bpf.o: $(src)/%.bpf.c $(obj)/vmlinux.h $(src)/xdp_sample.bpf.h $(src)/x + $(Q)$(CLANG) -g -O2 -target bpf -D__TARGET_ARCH_$(SRCARCH) \ + -Wno-compare-distinct-pointer-types -I$(srctree)/include \ + -I$(srctree)/samples/bpf -I$(srctree)/tools/include \ +- -I$(srctree)/tools/lib $(CLANG_SYS_INCLUDES) \ ++ -I$(LIBBPF_INCLUDE) $(CLANG_SYS_INCLUDES) \ + -c $(filter %.bpf.c,$^) -o $@ + + LINKED_SKELS := xdp_redirect_cpu.skel.h xdp_redirect_map_multi.skel.h \ +@@ -403,7 +439,7 @@ $(obj)/%.o: $(src)/%.c + @echo " CLANG-bpf " $@ + $(Q)$(CLANG) $(NOSTDINC_FLAGS) $(LINUXINCLUDE) $(BPF_EXTRA_CFLAGS) \ + -I$(obj) -I$(srctree)/tools/testing/selftests/bpf/ \ +- -I$(srctree)/tools/lib/ \ ++ -I$(LIBBPF_INCLUDE) \ + -D__KERNEL__ -D__BPF_TRACING__ -Wno-unused-value -Wno-pointer-sign \ + -D__TARGET_ARCH_$(SRCARCH) -Wno-compare-distinct-pointer-types \ + -Wno-gnu-variable-sized-type-not-at-end \ +diff --git a/samples/bpf/Makefile.target b/samples/bpf/Makefile.target +index 5a368affa0386..7621f55e2947d 100644 +--- a/samples/bpf/Makefile.target ++++ b/samples/bpf/Makefile.target +@@ -73,14 +73,3 @@ quiet_cmd_tprog-cobjs = CC $@ + cmd_tprog-cobjs = $(CC) $(tprogc_flags) -c -o $@ $< + $(tprog-cobjs): $(obj)/%.o: $(src)/%.c FORCE + $(call if_changed_dep,tprog-cobjs) +- +-# Override includes for xdp_sample_user.o because $(srctree)/usr/include in +-# TPROGS_CFLAGS causes conflicts +-XDP_SAMPLE_CFLAGS += -Wall -O2 -lm \ +- -I./tools/include \ +- -I./tools/include/uapi \ +- -I./tools/lib \ +- -I./tools/testing/selftests/bpf +-$(obj)/xdp_sample_user.o: $(src)/xdp_sample_user.c \ +- $(src)/xdp_sample_user.h $(src)/xdp_sample_shared.h +- $(CC) $(XDP_SAMPLE_CFLAGS) -c -o $@ $< +diff --git a/samples/bpf/hbm_kern.h b/samples/bpf/hbm_kern.h +index 722b3fadb4675..1752a46a2b056 100644 +--- a/samples/bpf/hbm_kern.h ++++ b/samples/bpf/hbm_kern.h +@@ -9,8 +9,6 @@ + * Include file for sample Host Bandwidth Manager (HBM) BPF programs + */ + #define KBUILD_MODNAME "foo" +-#include <stddef.h> +-#include <stdbool.h> + #include <uapi/linux/bpf.h> + #include <uapi/linux/if_ether.h> + #include <uapi/linux/if_packet.h> +diff --git a/samples/bpf/lwt_len_hist_kern.c b/samples/bpf/lwt_len_hist_kern.c +index 9ed63e10e1709..1fa14c54963a1 100644 +--- a/samples/bpf/lwt_len_hist_kern.c ++++ b/samples/bpf/lwt_len_hist_kern.c +@@ -16,13 +16,6 @@ + #include <uapi/linux/in.h> + #include <bpf/bpf_helpers.h> + +-# define printk(fmt, ...) \ +- ({ \ +- char ____fmt[] = fmt; \ +- bpf_trace_printk(____fmt, sizeof(____fmt), \ +- ##__VA_ARGS__); \ +- }) +- + struct bpf_elf_map { + __u32 type; + __u32 size_key; +diff --git a/samples/bpf/xdp_sample_user.h b/samples/bpf/xdp_sample_user.h +index d97465ff8c62c..5f44b877ecf5f 100644 +--- a/samples/bpf/xdp_sample_user.h ++++ b/samples/bpf/xdp_sample_user.h +@@ -45,7 +45,9 @@ const char *get_driver_name(int ifindex); + int get_mac_addr(int ifindex, void *mac_addr); + + #pragma GCC diagnostic push ++#ifndef __clang__ + #pragma GCC diagnostic ignored "-Wstringop-truncation" ++#endif + __attribute__((unused)) + static inline char *safe_strncpy(char *dst, const char *src, size_t size) + { +diff --git a/scripts/dtc/dtx_diff b/scripts/dtc/dtx_diff +index d3422ee15e300..f2bbde4bba86b 100755 +--- a/scripts/dtc/dtx_diff ++++ b/scripts/dtc/dtx_diff +@@ -59,12 +59,8 @@ Otherwise DTx is treated as a dts source file (aka .dts). + or '/include/' to be processed. + + If DTx_1 and DTx_2 are in different architectures, then this script +- may not work since \${ARCH} is part of the include path. Two possible +- workarounds: +- +- `basename $0` \\ +- <(ARCH=arch_of_dtx_1 `basename $0` DTx_1) \\ +- <(ARCH=arch_of_dtx_2 `basename $0` DTx_2) ++ may not work since \${ARCH} is part of the include path. The following ++ workaround can be used: + + `basename $0` ARCH=arch_of_dtx_1 DTx_1 >tmp_dtx_1.dts + `basename $0` ARCH=arch_of_dtx_2 DTx_2 >tmp_dtx_2.dts +diff --git a/scripts/sphinx-pre-install b/scripts/sphinx-pre-install +index 288e86a9d1e58..f126ecbb0494d 100755 +--- a/scripts/sphinx-pre-install ++++ b/scripts/sphinx-pre-install +@@ -78,6 +78,7 @@ my %texlive = ( + 'ucs.sty' => 'texlive-ucs', + 'upquote.sty' => 'texlive-upquote', + 'wrapfig.sty' => 'texlive-wrapfig', ++ 'ctexhook.sty' => 'texlive-ctex', + ); + + # +@@ -369,6 +370,9 @@ sub give_debian_hints() + ); + + if ($pdf) { ++ check_missing_file(["/usr/share/texlive/texmf-dist/tex/latex/ctex/ctexhook.sty"], ++ "texlive-lang-chinese", 2); ++ + check_missing_file(["/usr/share/fonts/truetype/dejavu/DejaVuSans.ttf"], + "fonts-dejavu", 2); + +diff --git a/security/selinux/hooks.c b/security/selinux/hooks.c +index 9309e62d46eda..baa12d1007c7c 100644 +--- a/security/selinux/hooks.c ++++ b/security/selinux/hooks.c +@@ -987,18 +987,22 @@ out: + static int selinux_add_opt(int token, const char *s, void **mnt_opts) + { + struct selinux_mnt_opts *opts = *mnt_opts; ++ bool is_alloc_opts = false; + + if (token == Opt_seclabel) /* eaten and completely ignored */ + return 0; + ++ if (!s) ++ return -ENOMEM; ++ + if (!opts) { + opts = kzalloc(sizeof(struct selinux_mnt_opts), GFP_KERNEL); + if (!opts) + return -ENOMEM; + *mnt_opts = opts; ++ is_alloc_opts = true; + } +- if (!s) +- return -ENOMEM; ++ + switch (token) { + case Opt_context: + if (opts->context || opts->defcontext) +@@ -1023,6 +1027,10 @@ static int selinux_add_opt(int token, const char *s, void **mnt_opts) + } + return 0; + Einval: ++ if (is_alloc_opts) { ++ kfree(opts); ++ *mnt_opts = NULL; ++ } + pr_warn(SEL_MOUNT_FAIL_MSG); + return -EINVAL; + } +diff --git a/sound/core/jack.c b/sound/core/jack.c +index 537df1e98f8ac..d1e3055f2b6a5 100644 +--- a/sound/core/jack.c ++++ b/sound/core/jack.c +@@ -62,10 +62,13 @@ static int snd_jack_dev_free(struct snd_device *device) + struct snd_card *card = device->card; + struct snd_jack_kctl *jack_kctl, *tmp_jack_kctl; + ++ down_write(&card->controls_rwsem); + list_for_each_entry_safe(jack_kctl, tmp_jack_kctl, &jack->kctl_list, list) { + list_del_init(&jack_kctl->list); + snd_ctl_remove(card, jack_kctl->kctl); + } ++ up_write(&card->controls_rwsem); ++ + if (jack->private_free) + jack->private_free(jack); + +diff --git a/sound/core/misc.c b/sound/core/misc.c +index 3579dd7a161f7..50e4aaa6270d1 100644 +--- a/sound/core/misc.c ++++ b/sound/core/misc.c +@@ -112,7 +112,7 @@ snd_pci_quirk_lookup_id(u16 vendor, u16 device, + { + const struct snd_pci_quirk *q; + +- for (q = list; q->subvendor; q++) { ++ for (q = list; q->subvendor || q->subdevice; q++) { + if (q->subvendor != vendor) + continue; + if (!q->subdevice || +diff --git a/sound/core/oss/pcm_oss.c b/sound/core/oss/pcm_oss.c +index 20a0a4771b9a8..3ee9edf858156 100644 +--- a/sound/core/oss/pcm_oss.c ++++ b/sound/core/oss/pcm_oss.c +@@ -2065,7 +2065,7 @@ static int snd_pcm_oss_set_trigger(struct snd_pcm_oss_file *pcm_oss_file, int tr + int err, cmd; + + #ifdef OSS_DEBUG +- pcm_dbg(substream->pcm, "pcm_oss: trigger = 0x%x\n", trigger); ++ pr_debug("pcm_oss: trigger = 0x%x\n", trigger); + #endif + + psubstream = pcm_oss_file->streams[SNDRV_PCM_STREAM_PLAYBACK]; +diff --git a/sound/core/pcm.c b/sound/core/pcm.c +index 6fd3677685d70..ba4a987ed1c62 100644 +--- a/sound/core/pcm.c ++++ b/sound/core/pcm.c +@@ -810,7 +810,11 @@ EXPORT_SYMBOL(snd_pcm_new_internal); + static void free_chmap(struct snd_pcm_str *pstr) + { + if (pstr->chmap_kctl) { +- snd_ctl_remove(pstr->pcm->card, pstr->chmap_kctl); ++ struct snd_card *card = pstr->pcm->card; ++ ++ down_write(&card->controls_rwsem); ++ snd_ctl_remove(card, pstr->chmap_kctl); ++ up_write(&card->controls_rwsem); + pstr->chmap_kctl = NULL; + } + } +diff --git a/sound/core/seq/seq_queue.c b/sound/core/seq/seq_queue.c +index d6c02dea976c8..bc933104c3eea 100644 +--- a/sound/core/seq/seq_queue.c ++++ b/sound/core/seq/seq_queue.c +@@ -235,12 +235,15 @@ struct snd_seq_queue *snd_seq_queue_find_name(char *name) + + /* -------------------------------------------------------- */ + ++#define MAX_CELL_PROCESSES_IN_QUEUE 1000 ++ + void snd_seq_check_queue(struct snd_seq_queue *q, int atomic, int hop) + { + unsigned long flags; + struct snd_seq_event_cell *cell; + snd_seq_tick_time_t cur_tick; + snd_seq_real_time_t cur_time; ++ int processed = 0; + + if (q == NULL) + return; +@@ -263,6 +266,8 @@ void snd_seq_check_queue(struct snd_seq_queue *q, int atomic, int hop) + if (!cell) + break; + snd_seq_dispatch_event(cell, atomic, hop); ++ if (++processed >= MAX_CELL_PROCESSES_IN_QUEUE) ++ goto out; /* the rest processed at the next batch */ + } + + /* Process time queue... */ +@@ -272,14 +277,19 @@ void snd_seq_check_queue(struct snd_seq_queue *q, int atomic, int hop) + if (!cell) + break; + snd_seq_dispatch_event(cell, atomic, hop); ++ if (++processed >= MAX_CELL_PROCESSES_IN_QUEUE) ++ goto out; /* the rest processed at the next batch */ + } + ++ out: + /* free lock */ + spin_lock_irqsave(&q->check_lock, flags); + if (q->check_again) { + q->check_again = 0; +- spin_unlock_irqrestore(&q->check_lock, flags); +- goto __again; ++ if (processed < MAX_CELL_PROCESSES_IN_QUEUE) { ++ spin_unlock_irqrestore(&q->check_lock, flags); ++ goto __again; ++ } + } + q->check_blocked = 0; + spin_unlock_irqrestore(&q->check_lock, flags); +diff --git a/sound/hda/hdac_stream.c b/sound/hda/hdac_stream.c +index 9867555883c34..aa7955fdf68a0 100644 +--- a/sound/hda/hdac_stream.c ++++ b/sound/hda/hdac_stream.c +@@ -534,17 +534,11 @@ static void azx_timecounter_init(struct hdac_stream *azx_dev, + cc->mask = CLOCKSOURCE_MASK(32); + + /* +- * Converting from 24 MHz to ns means applying a 125/3 factor. +- * To avoid any saturation issues in intermediate operations, +- * the 125 factor is applied first. The division is applied +- * last after reading the timecounter value. +- * Applying the 1/3 factor as part of the multiplication +- * requires at least 20 bits for a decent precision, however +- * overflows occur after about 4 hours or less, not a option. ++ * Calculate the optimal mult/shift values. The counter wraps ++ * around after ~178.9 seconds. + */ +- +- cc->mult = 125; /* saturation after 195 years */ +- cc->shift = 0; ++ clocks_calc_mult_shift(&cc->mult, &cc->shift, 24000000, ++ NSEC_PER_SEC, 178); + + nsec = 0; /* audio time is elapsed time since trigger */ + timecounter_init(tc, cc, nsec); +diff --git a/sound/pci/hda/hda_bind.c b/sound/pci/hda/hda_bind.c +index 1c8bffc3eec6e..7153bd53e1893 100644 +--- a/sound/pci/hda/hda_bind.c ++++ b/sound/pci/hda/hda_bind.c +@@ -156,6 +156,11 @@ static int hda_codec_driver_remove(struct device *dev) + return codec->bus->core.ext_ops->hdev_detach(&codec->core); + } + ++ refcount_dec(&codec->pcm_ref); ++ snd_hda_codec_disconnect_pcms(codec); ++ wait_event(codec->remove_sleep, !refcount_read(&codec->pcm_ref)); ++ snd_power_sync_ref(codec->bus->card); ++ + if (codec->patch_ops.free) + codec->patch_ops.free(codec); + snd_hda_codec_cleanup_for_unbind(codec); +diff --git a/sound/pci/hda/hda_codec.c b/sound/pci/hda/hda_codec.c +index 0c4a337c9fc0d..7016b48227bf2 100644 +--- a/sound/pci/hda/hda_codec.c ++++ b/sound/pci/hda/hda_codec.c +@@ -703,20 +703,10 @@ get_hda_cvt_setup(struct hda_codec *codec, hda_nid_t nid) + /* + * PCM device + */ +-static void release_pcm(struct kref *kref) +-{ +- struct hda_pcm *pcm = container_of(kref, struct hda_pcm, kref); +- +- if (pcm->pcm) +- snd_device_free(pcm->codec->card, pcm->pcm); +- clear_bit(pcm->device, pcm->codec->bus->pcm_dev_bits); +- kfree(pcm->name); +- kfree(pcm); +-} +- + void snd_hda_codec_pcm_put(struct hda_pcm *pcm) + { +- kref_put(&pcm->kref, release_pcm); ++ if (refcount_dec_and_test(&pcm->codec->pcm_ref)) ++ wake_up(&pcm->codec->remove_sleep); + } + EXPORT_SYMBOL_GPL(snd_hda_codec_pcm_put); + +@@ -731,7 +721,6 @@ struct hda_pcm *snd_hda_codec_pcm_new(struct hda_codec *codec, + return NULL; + + pcm->codec = codec; +- kref_init(&pcm->kref); + va_start(args, fmt); + pcm->name = kvasprintf(GFP_KERNEL, fmt, args); + va_end(args); +@@ -741,6 +730,7 @@ struct hda_pcm *snd_hda_codec_pcm_new(struct hda_codec *codec, + } + + list_add_tail(&pcm->list, &codec->pcm_list_head); ++ refcount_inc(&codec->pcm_ref); + return pcm; + } + EXPORT_SYMBOL_GPL(snd_hda_codec_pcm_new); +@@ -748,15 +738,31 @@ EXPORT_SYMBOL_GPL(snd_hda_codec_pcm_new); + /* + * codec destructor + */ ++void snd_hda_codec_disconnect_pcms(struct hda_codec *codec) ++{ ++ struct hda_pcm *pcm; ++ ++ list_for_each_entry(pcm, &codec->pcm_list_head, list) { ++ if (pcm->disconnected) ++ continue; ++ if (pcm->pcm) ++ snd_device_disconnect(codec->card, pcm->pcm); ++ snd_hda_codec_pcm_put(pcm); ++ pcm->disconnected = 1; ++ } ++} ++ + static void codec_release_pcms(struct hda_codec *codec) + { + struct hda_pcm *pcm, *n; + + list_for_each_entry_safe(pcm, n, &codec->pcm_list_head, list) { +- list_del_init(&pcm->list); ++ list_del(&pcm->list); + if (pcm->pcm) +- snd_device_disconnect(codec->card, pcm->pcm); +- snd_hda_codec_pcm_put(pcm); ++ snd_device_free(pcm->codec->card, pcm->pcm); ++ clear_bit(pcm->device, pcm->codec->bus->pcm_dev_bits); ++ kfree(pcm->name); ++ kfree(pcm); + } + } + +@@ -769,6 +775,7 @@ void snd_hda_codec_cleanup_for_unbind(struct hda_codec *codec) + codec->registered = 0; + } + ++ snd_hda_codec_disconnect_pcms(codec); + cancel_delayed_work_sync(&codec->jackpoll_work); + if (!codec->in_freeing) + snd_hda_ctls_clear(codec); +@@ -792,6 +799,7 @@ void snd_hda_codec_cleanup_for_unbind(struct hda_codec *codec) + remove_conn_list(codec); + snd_hdac_regmap_exit(&codec->core); + codec->configured = 0; ++ refcount_set(&codec->pcm_ref, 1); /* reset refcount */ + } + EXPORT_SYMBOL_GPL(snd_hda_codec_cleanup_for_unbind); + +@@ -958,6 +966,8 @@ int snd_hda_codec_device_new(struct hda_bus *bus, struct snd_card *card, + snd_array_init(&codec->verbs, sizeof(struct hda_verb *), 8); + INIT_LIST_HEAD(&codec->conn_list); + INIT_LIST_HEAD(&codec->pcm_list_head); ++ refcount_set(&codec->pcm_ref, 1); ++ init_waitqueue_head(&codec->remove_sleep); + + INIT_DELAYED_WORK(&codec->jackpoll_work, hda_jackpoll_work); + codec->depop_delay = -1; +@@ -1727,8 +1737,11 @@ void snd_hda_ctls_clear(struct hda_codec *codec) + { + int i; + struct hda_nid_item *items = codec->mixers.list; ++ ++ down_write(&codec->card->controls_rwsem); + for (i = 0; i < codec->mixers.used; i++) + snd_ctl_remove(codec->card, items[i].kctl); ++ up_write(&codec->card->controls_rwsem); + snd_array_free(&codec->mixers); + snd_array_free(&codec->nids); + } +diff --git a/sound/pci/hda/hda_controller.c b/sound/pci/hda/hda_controller.c +index 930ae4002a818..75dcb14ff20ad 100644 +--- a/sound/pci/hda/hda_controller.c ++++ b/sound/pci/hda/hda_controller.c +@@ -504,7 +504,6 @@ static int azx_get_time_info(struct snd_pcm_substream *substream, + snd_pcm_gettime(substream->runtime, system_ts); + + nsec = timecounter_read(&azx_dev->core.tc); +- nsec = div_u64(nsec, 3); /* can be optimized */ + if (audio_tstamp_config->report_delay) + nsec = azx_adjust_codec_delay(substream, nsec); + +diff --git a/sound/pci/hda/hda_local.h b/sound/pci/hda/hda_local.h +index d22c96eb2f8fb..8621f576446b8 100644 +--- a/sound/pci/hda/hda_local.h ++++ b/sound/pci/hda/hda_local.h +@@ -137,6 +137,7 @@ int __snd_hda_add_vmaster(struct hda_codec *codec, char *name, + int snd_hda_codec_reset(struct hda_codec *codec); + void snd_hda_codec_register(struct hda_codec *codec); + void snd_hda_codec_cleanup_for_unbind(struct hda_codec *codec); ++void snd_hda_codec_disconnect_pcms(struct hda_codec *codec); + + #define snd_hda_regmap_sync(codec) snd_hdac_regmap_sync(&(codec)->core) + +diff --git a/sound/pci/hda/patch_cs8409-tables.c b/sound/pci/hda/patch_cs8409-tables.c +index 0fb0a428428b4..df0b4522babf7 100644 +--- a/sound/pci/hda/patch_cs8409-tables.c ++++ b/sound/pci/hda/patch_cs8409-tables.c +@@ -252,6 +252,7 @@ struct sub_codec cs8409_cs42l42_codec = { + .init_seq_num = ARRAY_SIZE(cs42l42_init_reg_seq), + .hp_jack_in = 0, + .mic_jack_in = 0, ++ .force_status_change = 1, + .paged = 1, + .suspended = 1, + .no_type_dect = 0, +@@ -443,6 +444,7 @@ struct sub_codec dolphin_cs42l42_0 = { + .init_seq_num = ARRAY_SIZE(dolphin_c0_init_reg_seq), + .hp_jack_in = 0, + .mic_jack_in = 0, ++ .force_status_change = 1, + .paged = 1, + .suspended = 1, + .no_type_dect = 0, +@@ -456,6 +458,7 @@ struct sub_codec dolphin_cs42l42_1 = { + .init_seq_num = ARRAY_SIZE(dolphin_c1_init_reg_seq), + .hp_jack_in = 0, + .mic_jack_in = 0, ++ .force_status_change = 1, + .paged = 1, + .suspended = 1, + .no_type_dect = 1, +diff --git a/sound/pci/hda/patch_cs8409.c b/sound/pci/hda/patch_cs8409.c +index 039b9f2f8e947..aff2b5abb81ea 100644 +--- a/sound/pci/hda/patch_cs8409.c ++++ b/sound/pci/hda/patch_cs8409.c +@@ -628,15 +628,17 @@ static void cs42l42_run_jack_detect(struct sub_codec *cs42l42) + cs8409_i2c_write(cs42l42, 0x1b74, 0x07); + cs8409_i2c_write(cs42l42, 0x131b, 0xFD); + cs8409_i2c_write(cs42l42, 0x1120, 0x80); +- /* Wait ~100us*/ +- usleep_range(100, 200); ++ /* Wait ~20ms*/ ++ usleep_range(20000, 25000); + cs8409_i2c_write(cs42l42, 0x111f, 0x77); + cs8409_i2c_write(cs42l42, 0x1120, 0xc0); + } + + static int cs42l42_handle_tip_sense(struct sub_codec *cs42l42, unsigned int reg_ts_status) + { +- int status_changed = 0; ++ int status_changed = cs42l42->force_status_change; ++ ++ cs42l42->force_status_change = 0; + + /* TIP_SENSE INSERT/REMOVE */ + switch (reg_ts_status) { +@@ -791,6 +793,7 @@ static void cs42l42_suspend(struct sub_codec *cs42l42) + cs42l42->last_page = 0; + cs42l42->hp_jack_in = 0; + cs42l42->mic_jack_in = 0; ++ cs42l42->force_status_change = 1; + + /* Put CS42L42 into Reset */ + gpio_data = snd_hda_codec_read(codec, CS8409_PIN_AFG, 0, AC_VERB_GET_GPIO_DATA, 0); +diff --git a/sound/pci/hda/patch_cs8409.h b/sound/pci/hda/patch_cs8409.h +index ade2b838590cf..d0b725c7285b6 100644 +--- a/sound/pci/hda/patch_cs8409.h ++++ b/sound/pci/hda/patch_cs8409.h +@@ -305,6 +305,7 @@ struct sub_codec { + + unsigned int hp_jack_in:1; + unsigned int mic_jack_in:1; ++ unsigned int force_status_change:1; + unsigned int suspended:1; + unsigned int paged:1; + unsigned int last_page; +diff --git a/sound/soc/codecs/Kconfig b/sound/soc/codecs/Kconfig +index 216cea04ad704..f12c9b9426788 100644 +--- a/sound/soc/codecs/Kconfig ++++ b/sound/soc/codecs/Kconfig +@@ -235,8 +235,7 @@ config SND_SOC_ALL_CODECS + imply SND_SOC_UDA1380 + imply SND_SOC_WCD9335 + imply SND_SOC_WCD934X +- imply SND_SOC_WCD937X +- imply SND_SOC_WCD938X ++ imply SND_SOC_WCD938X_SDW + imply SND_SOC_LPASS_RX_MACRO + imply SND_SOC_LPASS_TX_MACRO + imply SND_SOC_WL1273 +diff --git a/sound/soc/codecs/rt5663.c b/sound/soc/codecs/rt5663.c +index be9fc58ff6812..ee09ccd448dcd 100644 +--- a/sound/soc/codecs/rt5663.c ++++ b/sound/soc/codecs/rt5663.c +@@ -3461,6 +3461,7 @@ static void rt5663_calibrate(struct rt5663_priv *rt5663) + static int rt5663_parse_dp(struct rt5663_priv *rt5663, struct device *dev) + { + int table_size; ++ int ret; + + device_property_read_u32(dev, "realtek,dc_offset_l_manual", + &rt5663->pdata.dc_offset_l_manual); +@@ -3477,9 +3478,11 @@ static int rt5663_parse_dp(struct rt5663_priv *rt5663, struct device *dev) + table_size = sizeof(struct impedance_mapping_table) * + rt5663->pdata.impedance_sensing_num; + rt5663->imp_table = devm_kzalloc(dev, table_size, GFP_KERNEL); +- device_property_read_u32_array(dev, ++ ret = device_property_read_u32_array(dev, + "realtek,impedance_sensing_table", + (u32 *)rt5663->imp_table, table_size); ++ if (ret) ++ return ret; + } + + return 0; +@@ -3504,8 +3507,11 @@ static int rt5663_i2c_probe(struct i2c_client *i2c, + + if (pdata) + rt5663->pdata = *pdata; +- else +- rt5663_parse_dp(rt5663, &i2c->dev); ++ else { ++ ret = rt5663_parse_dp(rt5663, &i2c->dev); ++ if (ret) ++ return ret; ++ } + + for (i = 0; i < ARRAY_SIZE(rt5663->supplies); i++) + rt5663->supplies[i].supply = rt5663_supply_names[i]; +diff --git a/sound/soc/fsl/fsl_asrc.c b/sound/soc/fsl/fsl_asrc.c +index 24b41881a68f8..d7d1536a4f377 100644 +--- a/sound/soc/fsl/fsl_asrc.c ++++ b/sound/soc/fsl/fsl_asrc.c +@@ -19,6 +19,7 @@ + #include "fsl_asrc.h" + + #define IDEAL_RATIO_DECIMAL_DEPTH 26 ++#define DIVIDER_NUM 64 + + #define pair_err(fmt, ...) \ + dev_err(&asrc->pdev->dev, "Pair %c: " fmt, 'A' + index, ##__VA_ARGS__) +@@ -101,6 +102,55 @@ static unsigned char clk_map_imx8qxp[2][ASRC_CLK_MAP_LEN] = { + }, + }; + ++/* ++ * According to RM, the divider range is 1 ~ 8, ++ * prescaler is power of 2 from 1 ~ 128. ++ */ ++static int asrc_clk_divider[DIVIDER_NUM] = { ++ 1, 2, 4, 8, 16, 32, 64, 128, /* divider = 1 */ ++ 2, 4, 8, 16, 32, 64, 128, 256, /* divider = 2 */ ++ 3, 6, 12, 24, 48, 96, 192, 384, /* divider = 3 */ ++ 4, 8, 16, 32, 64, 128, 256, 512, /* divider = 4 */ ++ 5, 10, 20, 40, 80, 160, 320, 640, /* divider = 5 */ ++ 6, 12, 24, 48, 96, 192, 384, 768, /* divider = 6 */ ++ 7, 14, 28, 56, 112, 224, 448, 896, /* divider = 7 */ ++ 8, 16, 32, 64, 128, 256, 512, 1024, /* divider = 8 */ ++}; ++ ++/* ++ * Check if the divider is available for internal ratio mode ++ */ ++static bool fsl_asrc_divider_avail(int clk_rate, int rate, int *div) ++{ ++ u32 rem, i; ++ u64 n; ++ ++ if (div) ++ *div = 0; ++ ++ if (clk_rate == 0 || rate == 0) ++ return false; ++ ++ n = clk_rate; ++ rem = do_div(n, rate); ++ ++ if (div) ++ *div = n; ++ ++ if (rem != 0) ++ return false; ++ ++ for (i = 0; i < DIVIDER_NUM; i++) { ++ if (n == asrc_clk_divider[i]) ++ break; ++ } ++ ++ if (i == DIVIDER_NUM) ++ return false; ++ ++ return true; ++} ++ + /** + * fsl_asrc_sel_proc - Select the pre-processing and post-processing options + * @inrate: input sample rate +@@ -330,12 +380,12 @@ static int fsl_asrc_config_pair(struct fsl_asrc_pair *pair, bool use_ideal_rate) + enum asrc_word_width input_word_width; + enum asrc_word_width output_word_width; + u32 inrate, outrate, indiv, outdiv; +- u32 clk_index[2], div[2], rem[2]; ++ u32 clk_index[2], div[2]; + u64 clk_rate; + int in, out, channels; + int pre_proc, post_proc; + struct clk *clk; +- bool ideal; ++ bool ideal, div_avail; + + if (!config) { + pair_err("invalid pair config\n"); +@@ -415,8 +465,7 @@ static int fsl_asrc_config_pair(struct fsl_asrc_pair *pair, bool use_ideal_rate) + clk = asrc_priv->asrck_clk[clk_index[ideal ? OUT : IN]]; + + clk_rate = clk_get_rate(clk); +- rem[IN] = do_div(clk_rate, inrate); +- div[IN] = (u32)clk_rate; ++ div_avail = fsl_asrc_divider_avail(clk_rate, inrate, &div[IN]); + + /* + * The divider range is [1, 1024], defined by the hardware. For non- +@@ -425,7 +474,7 @@ static int fsl_asrc_config_pair(struct fsl_asrc_pair *pair, bool use_ideal_rate) + * only result in different converting speeds. So remainder does not + * matter, as long as we keep the divider within its valid range. + */ +- if (div[IN] == 0 || (!ideal && (div[IN] > 1024 || rem[IN] != 0))) { ++ if (div[IN] == 0 || (!ideal && !div_avail)) { + pair_err("failed to support input sample rate %dHz by asrck_%x\n", + inrate, clk_index[ideal ? OUT : IN]); + return -EINVAL; +@@ -436,13 +485,12 @@ static int fsl_asrc_config_pair(struct fsl_asrc_pair *pair, bool use_ideal_rate) + clk = asrc_priv->asrck_clk[clk_index[OUT]]; + clk_rate = clk_get_rate(clk); + if (ideal && use_ideal_rate) +- rem[OUT] = do_div(clk_rate, IDEAL_RATIO_RATE); ++ div_avail = fsl_asrc_divider_avail(clk_rate, IDEAL_RATIO_RATE, &div[OUT]); + else +- rem[OUT] = do_div(clk_rate, outrate); +- div[OUT] = clk_rate; ++ div_avail = fsl_asrc_divider_avail(clk_rate, outrate, &div[OUT]); + + /* Output divider has the same limitation as the input one */ +- if (div[OUT] == 0 || (!ideal && (div[OUT] > 1024 || rem[OUT] != 0))) { ++ if (div[OUT] == 0 || (!ideal && !div_avail)) { + pair_err("failed to support output sample rate %dHz by asrck_%x\n", + outrate, clk_index[OUT]); + return -EINVAL; +@@ -621,8 +669,7 @@ static void fsl_asrc_select_clk(struct fsl_asrc_priv *asrc_priv, + clk_index = asrc_priv->clk_map[j][i]; + clk_rate = clk_get_rate(asrc_priv->asrck_clk[clk_index]); + /* Only match a perfect clock source with no remainder */ +- if (clk_rate != 0 && (clk_rate / rate[j]) <= 1024 && +- (clk_rate % rate[j]) == 0) ++ if (fsl_asrc_divider_avail(clk_rate, rate[j], NULL)) + break; + } + +diff --git a/sound/soc/fsl/fsl_mqs.c b/sound/soc/fsl/fsl_mqs.c +index 69aeb0e71844d..0d4efbed41dab 100644 +--- a/sound/soc/fsl/fsl_mqs.c ++++ b/sound/soc/fsl/fsl_mqs.c +@@ -337,4 +337,4 @@ module_platform_driver(fsl_mqs_driver); + MODULE_AUTHOR("Shengjiu Wang <Shengjiu.Wang@nxp.com>"); + MODULE_DESCRIPTION("MQS codec driver"); + MODULE_LICENSE("GPL v2"); +-MODULE_ALIAS("platform: fsl-mqs"); ++MODULE_ALIAS("platform:fsl-mqs"); +diff --git a/sound/soc/fsl/imx-card.c b/sound/soc/fsl/imx-card.c +index 58fd0639a0698..db947180617a6 100644 +--- a/sound/soc/fsl/imx-card.c ++++ b/sound/soc/fsl/imx-card.c +@@ -120,7 +120,7 @@ struct imx_card_data { + + static struct imx_akcodec_fs_mul ak4458_fs_mul[] = { + /* Normal, < 32kHz */ +- { .rmin = 8000, .rmax = 24000, .wmin = 1024, .wmax = 1024, }, ++ { .rmin = 8000, .rmax = 24000, .wmin = 256, .wmax = 1024, }, + /* Normal, 32kHz */ + { .rmin = 32000, .rmax = 32000, .wmin = 256, .wmax = 1024, }, + /* Normal */ +@@ -151,8 +151,8 @@ static struct imx_akcodec_fs_mul ak4497_fs_mul[] = { + * Table 7 - mapping multiplier and speed mode + * Tables 8 & 9 - mapping speed mode and LRCK fs + */ +- { .rmin = 8000, .rmax = 32000, .wmin = 1024, .wmax = 1024, }, /* Normal, <= 32kHz */ +- { .rmin = 44100, .rmax = 48000, .wmin = 512, .wmax = 512, }, /* Normal */ ++ { .rmin = 8000, .rmax = 32000, .wmin = 256, .wmax = 1024, }, /* Normal, <= 32kHz */ ++ { .rmin = 44100, .rmax = 48000, .wmin = 256, .wmax = 512, }, /* Normal */ + { .rmin = 88200, .rmax = 96000, .wmin = 256, .wmax = 256, }, /* Double */ + { .rmin = 176400, .rmax = 192000, .wmin = 128, .wmax = 128, }, /* Quad */ + { .rmin = 352800, .rmax = 384000, .wmin = 128, .wmax = 128, }, /* Oct */ +@@ -164,7 +164,7 @@ static struct imx_akcodec_fs_mul ak4497_fs_mul[] = { + * (Table 4 from datasheet) + */ + static struct imx_akcodec_fs_mul ak5558_fs_mul[] = { +- { .rmin = 8000, .rmax = 32000, .wmin = 1024, .wmax = 1024, }, ++ { .rmin = 8000, .rmax = 32000, .wmin = 512, .wmax = 1024, }, + { .rmin = 44100, .rmax = 48000, .wmin = 512, .wmax = 512, }, + { .rmin = 88200, .rmax = 96000, .wmin = 256, .wmax = 256, }, + { .rmin = 176400, .rmax = 192000, .wmin = 128, .wmax = 128, }, +@@ -247,13 +247,14 @@ static bool codec_is_akcodec(unsigned int type) + } + + static unsigned long akcodec_get_mclk_rate(struct snd_pcm_substream *substream, +- struct snd_pcm_hw_params *params) ++ struct snd_pcm_hw_params *params, ++ int slots, int slot_width) + { + struct snd_soc_pcm_runtime *rtd = substream->private_data; + struct imx_card_data *data = snd_soc_card_get_drvdata(rtd->card); + const struct imx_card_plat_data *plat_data = data->plat_data; + struct dai_link_data *link_data = &data->link_data[rtd->num]; +- unsigned int width = link_data->slots * link_data->slot_width; ++ unsigned int width = slots * slot_width; + unsigned int rate = params_rate(params); + int i; + +@@ -349,7 +350,7 @@ static int imx_aif_hw_params(struct snd_pcm_substream *substream, + + /* Set MCLK freq */ + if (codec_is_akcodec(plat_data->type)) +- mclk_freq = akcodec_get_mclk_rate(substream, params); ++ mclk_freq = akcodec_get_mclk_rate(substream, params, slots, slot_width); + else + mclk_freq = params_rate(params) * slots * slot_width; + /* Use the maximum freq from DSD512 (512*44100 = 22579200) */ +@@ -553,8 +554,23 @@ static int imx_card_parse_of(struct imx_card_data *data) + link_data->cpu_sysclk_id = FSL_SAI_CLK_MAST1; + + /* sai may support mclk/bclk = 1 */ +- if (of_find_property(np, "fsl,mclk-equal-bclk", NULL)) ++ if (of_find_property(np, "fsl,mclk-equal-bclk", NULL)) { + link_data->one2one_ratio = true; ++ } else { ++ int i; ++ ++ /* ++ * i.MX8MQ don't support one2one ratio, then ++ * with ak4497 only 16bit case is supported. ++ */ ++ for (i = 0; i < ARRAY_SIZE(ak4497_fs_mul); i++) { ++ if (ak4497_fs_mul[i].rmin == 705600 && ++ ak4497_fs_mul[i].rmax == 768000) { ++ ak4497_fs_mul[i].wmin = 32; ++ ak4497_fs_mul[i].wmax = 32; ++ } ++ } ++ } + } + + link->cpus->of_node = args.np; +diff --git a/sound/soc/fsl/imx-hdmi.c b/sound/soc/fsl/imx-hdmi.c +index 34a0dceae6216..ef8d7a65ebc61 100644 +--- a/sound/soc/fsl/imx-hdmi.c ++++ b/sound/soc/fsl/imx-hdmi.c +@@ -145,6 +145,8 @@ static int imx_hdmi_probe(struct platform_device *pdev) + data->dai.capture_only = false; + data->dai.init = imx_hdmi_init; + ++ put_device(&cpu_pdev->dev); ++ + if (of_node_name_eq(cpu_np, "sai")) { + data->cpu_priv.sysclk_id[1] = FSL_SAI_CLK_MAST1; + data->cpu_priv.sysclk_id[0] = FSL_SAI_CLK_MAST1; +diff --git a/sound/soc/intel/boards/sof_sdw.c b/sound/soc/intel/boards/sof_sdw.c +index f10496206ceed..76759b2099064 100644 +--- a/sound/soc/intel/boards/sof_sdw.c ++++ b/sound/soc/intel/boards/sof_sdw.c +@@ -188,7 +188,7 @@ static const struct dmi_system_id sof_sdw_quirk_table[] = { + }, + .driver_data = (void *)(SOF_SDW_TGL_HDMI | + SOF_SDW_PCH_DMIC | +- RT711_JD2), ++ RT711_JD1), + }, + { + /* NUC15 'Bishop County' LAPBC510 and LAPBC710 skews */ +diff --git a/sound/soc/intel/catpt/dsp.c b/sound/soc/intel/catpt/dsp.c +index 9c5fd18f2600f..346bec0003066 100644 +--- a/sound/soc/intel/catpt/dsp.c ++++ b/sound/soc/intel/catpt/dsp.c +@@ -65,6 +65,7 @@ static int catpt_dma_memcpy(struct catpt_dev *cdev, struct dma_chan *chan, + { + struct dma_async_tx_descriptor *desc; + enum dma_status status; ++ int ret; + + desc = dmaengine_prep_dma_memcpy(chan, dst_addr, src_addr, size, + DMA_CTRL_ACK); +@@ -77,13 +78,22 @@ static int catpt_dma_memcpy(struct catpt_dev *cdev, struct dma_chan *chan, + catpt_updatel_shim(cdev, HMDC, + CATPT_HMDC_HDDA(CATPT_DMA_DEVID, chan->chan_id), + CATPT_HMDC_HDDA(CATPT_DMA_DEVID, chan->chan_id)); +- dmaengine_submit(desc); ++ ++ ret = dma_submit_error(dmaengine_submit(desc)); ++ if (ret) { ++ dev_err(cdev->dev, "submit tx failed: %d\n", ret); ++ goto clear_hdda; ++ } ++ + status = dma_wait_for_async_tx(desc); ++ ret = (status == DMA_COMPLETE) ? 0 : -EPROTO; ++ ++clear_hdda: + /* regardless of status, disable access to HOST memory in demand mode */ + catpt_updatel_shim(cdev, HMDC, + CATPT_HMDC_HDDA(CATPT_DMA_DEVID, chan->chan_id), 0); + +- return (status == DMA_COMPLETE) ? 0 : -EPROTO; ++ return ret; + } + + int catpt_dma_memcpy_todsp(struct catpt_dev *cdev, struct dma_chan *chan, +diff --git a/sound/soc/intel/skylake/skl-pcm.c b/sound/soc/intel/skylake/skl-pcm.c +index 9ecaf6a1e8475..e4aa366d356eb 100644 +--- a/sound/soc/intel/skylake/skl-pcm.c ++++ b/sound/soc/intel/skylake/skl-pcm.c +@@ -1251,7 +1251,6 @@ static int skl_platform_soc_get_time_info( + snd_pcm_gettime(substream->runtime, system_ts); + + nsec = timecounter_read(&hstr->tc); +- nsec = div_u64(nsec, 3); /* can be optimized */ + if (audio_tstamp_config->report_delay) + nsec = skl_adjust_codec_delay(substream, nsec); + +diff --git a/sound/soc/mediatek/mt8173/mt8173-max98090.c b/sound/soc/mediatek/mt8173/mt8173-max98090.c +index fc94314bfc02f..3bdd4931316cd 100644 +--- a/sound/soc/mediatek/mt8173/mt8173-max98090.c ++++ b/sound/soc/mediatek/mt8173/mt8173-max98090.c +@@ -180,6 +180,9 @@ static int mt8173_max98090_dev_probe(struct platform_device *pdev) + if (ret) + dev_err(&pdev->dev, "%s snd_soc_register_card fail %d\n", + __func__, ret); ++ ++ of_node_put(codec_node); ++ of_node_put(platform_node); + return ret; + } + +diff --git a/sound/soc/mediatek/mt8173/mt8173-rt5650-rt5514.c b/sound/soc/mediatek/mt8173/mt8173-rt5650-rt5514.c +index 0f28dc2217c09..390da5bf727eb 100644 +--- a/sound/soc/mediatek/mt8173/mt8173-rt5650-rt5514.c ++++ b/sound/soc/mediatek/mt8173/mt8173-rt5650-rt5514.c +@@ -218,6 +218,8 @@ static int mt8173_rt5650_rt5514_dev_probe(struct platform_device *pdev) + if (ret) + dev_err(&pdev->dev, "%s snd_soc_register_card fail %d\n", + __func__, ret); ++ ++ of_node_put(platform_node); + return ret; + } + +diff --git a/sound/soc/mediatek/mt8173/mt8173-rt5650-rt5676.c b/sound/soc/mediatek/mt8173/mt8173-rt5650-rt5676.c +index 077c6ee067806..c8e4e85e10575 100644 +--- a/sound/soc/mediatek/mt8173/mt8173-rt5650-rt5676.c ++++ b/sound/soc/mediatek/mt8173/mt8173-rt5650-rt5676.c +@@ -285,6 +285,8 @@ static int mt8173_rt5650_rt5676_dev_probe(struct platform_device *pdev) + if (ret) + dev_err(&pdev->dev, "%s snd_soc_register_card fail %d\n", + __func__, ret); ++ ++ of_node_put(platform_node); + return ret; + } + +diff --git a/sound/soc/mediatek/mt8173/mt8173-rt5650.c b/sound/soc/mediatek/mt8173/mt8173-rt5650.c +index c28ebf891cb05..e168d31f44459 100644 +--- a/sound/soc/mediatek/mt8173/mt8173-rt5650.c ++++ b/sound/soc/mediatek/mt8173/mt8173-rt5650.c +@@ -323,6 +323,8 @@ static int mt8173_rt5650_dev_probe(struct platform_device *pdev) + if (ret) + dev_err(&pdev->dev, "%s snd_soc_register_card fail %d\n", + __func__, ret); ++ ++ of_node_put(platform_node); + return ret; + } + +diff --git a/sound/soc/mediatek/mt8183/mt8183-da7219-max98357.c b/sound/soc/mediatek/mt8183/mt8183-da7219-max98357.c +index a4d26a6fc8492..bda103211e0bd 100644 +--- a/sound/soc/mediatek/mt8183/mt8183-da7219-max98357.c ++++ b/sound/soc/mediatek/mt8183/mt8183-da7219-max98357.c +@@ -781,7 +781,11 @@ static int mt8183_da7219_max98357_dev_probe(struct platform_device *pdev) + return ret; + } + +- return devm_snd_soc_register_card(&pdev->dev, card); ++ ret = devm_snd_soc_register_card(&pdev->dev, card); ++ ++ of_node_put(platform_node); ++ of_node_put(hdmi_codec); ++ return ret; + } + + #ifdef CONFIG_OF +diff --git a/sound/soc/mediatek/mt8183/mt8183-mt6358-ts3a227-max98357.c b/sound/soc/mediatek/mt8183/mt8183-mt6358-ts3a227-max98357.c +index 94dcbd36c8697..c7b10c48c6c22 100644 +--- a/sound/soc/mediatek/mt8183/mt8183-mt6358-ts3a227-max98357.c ++++ b/sound/soc/mediatek/mt8183/mt8183-mt6358-ts3a227-max98357.c +@@ -780,7 +780,12 @@ mt8183_mt6358_ts3a227_max98357_dev_probe(struct platform_device *pdev) + __func__, ret); + } + +- return devm_snd_soc_register_card(&pdev->dev, card); ++ ret = devm_snd_soc_register_card(&pdev->dev, card); ++ ++ of_node_put(platform_node); ++ of_node_put(ec_codec); ++ of_node_put(hdmi_codec); ++ return ret; + } + + #ifdef CONFIG_OF +diff --git a/sound/soc/mediatek/mt8192/mt8192-mt6359-rt1015-rt5682.c b/sound/soc/mediatek/mt8192/mt8192-mt6359-rt1015-rt5682.c +index a606133951b70..24a5d0adec1ba 100644 +--- a/sound/soc/mediatek/mt8192/mt8192-mt6359-rt1015-rt5682.c ++++ b/sound/soc/mediatek/mt8192/mt8192-mt6359-rt1015-rt5682.c +@@ -1172,7 +1172,11 @@ static int mt8192_mt6359_dev_probe(struct platform_device *pdev) + return ret; + } + +- return devm_snd_soc_register_card(&pdev->dev, card); ++ ret = devm_snd_soc_register_card(&pdev->dev, card); ++ ++ of_node_put(platform_node); ++ of_node_put(hdmi_codec); ++ return ret; + } + + #ifdef CONFIG_OF +diff --git a/sound/soc/mediatek/mt8195/mt8195-afe-pcm.c b/sound/soc/mediatek/mt8195/mt8195-afe-pcm.c +index 6635c3f72eccc..2edb40fe27ccb 100644 +--- a/sound/soc/mediatek/mt8195/mt8195-afe-pcm.c ++++ b/sound/soc/mediatek/mt8195/mt8195-afe-pcm.c +@@ -3028,7 +3028,7 @@ static const struct reg_sequence mt8195_afe_reg_defaults[] = { + + static const struct reg_sequence mt8195_cg_patch[] = { + { AUDIO_TOP_CON0, 0xfffffffb }, +- { AUDIO_TOP_CON1, 0xfffffffa }, ++ { AUDIO_TOP_CON1, 0xfffffff8 }, + }; + + static int mt8195_afe_init_registers(struct mtk_base_afe *afe) +diff --git a/sound/soc/mediatek/mt8195/mt8195-dai-pcm.c b/sound/soc/mediatek/mt8195/mt8195-dai-pcm.c +index 5d10d2c4c991c..151914c873acd 100644 +--- a/sound/soc/mediatek/mt8195/mt8195-dai-pcm.c ++++ b/sound/soc/mediatek/mt8195/mt8195-dai-pcm.c +@@ -80,8 +80,15 @@ static const struct snd_soc_dapm_widget mtk_dai_pcm_widgets[] = { + mtk_dai_pcm_o001_mix, + ARRAY_SIZE(mtk_dai_pcm_o001_mix)), + ++ SND_SOC_DAPM_SUPPLY("PCM_EN", PCM_INTF_CON1, ++ PCM_INTF_CON1_PCM_EN_SHIFT, 0, NULL, 0), ++ + SND_SOC_DAPM_INPUT("PCM1_INPUT"), + SND_SOC_DAPM_OUTPUT("PCM1_OUTPUT"), ++ ++ SND_SOC_DAPM_CLOCK_SUPPLY("aud_asrc11"), ++ SND_SOC_DAPM_CLOCK_SUPPLY("aud_asrc12"), ++ SND_SOC_DAPM_CLOCK_SUPPLY("aud_pcmif"), + }; + + static const struct snd_soc_dapm_route mtk_dai_pcm_routes[] = { +@@ -97,22 +104,18 @@ static const struct snd_soc_dapm_route mtk_dai_pcm_routes[] = { + {"PCM1 Playback", NULL, "O000"}, + {"PCM1 Playback", NULL, "O001"}, + ++ {"PCM1 Playback", NULL, "PCM_EN"}, ++ {"PCM1 Playback", NULL, "aud_asrc12"}, ++ {"PCM1 Playback", NULL, "aud_pcmif"}, ++ ++ {"PCM1 Capture", NULL, "PCM_EN"}, ++ {"PCM1 Capture", NULL, "aud_asrc11"}, ++ {"PCM1 Capture", NULL, "aud_pcmif"}, ++ + {"PCM1_OUTPUT", NULL, "PCM1 Playback"}, + {"PCM1 Capture", NULL, "PCM1_INPUT"}, + }; + +-static void mtk_dai_pcm_enable(struct mtk_base_afe *afe) +-{ +- regmap_update_bits(afe->regmap, PCM_INTF_CON1, +- PCM_INTF_CON1_PCM_EN, PCM_INTF_CON1_PCM_EN); +-} +- +-static void mtk_dai_pcm_disable(struct mtk_base_afe *afe) +-{ +- regmap_update_bits(afe->regmap, PCM_INTF_CON1, +- PCM_INTF_CON1_PCM_EN, 0x0); +-} +- + static int mtk_dai_pcm_configure(struct snd_pcm_substream *substream, + struct snd_soc_dai *dai) + { +@@ -207,54 +210,22 @@ static int mtk_dai_pcm_configure(struct snd_pcm_substream *substream, + } + + /* dai ops */ +-static int mtk_dai_pcm_startup(struct snd_pcm_substream *substream, +- struct snd_soc_dai *dai) +-{ +- struct mtk_base_afe *afe = snd_soc_dai_get_drvdata(dai); +- struct mt8195_afe_private *afe_priv = afe->platform_priv; +- +- if (dai->component->active) +- return 0; +- +- mt8195_afe_enable_clk(afe, afe_priv->clk[MT8195_CLK_AUD_ASRC11]); +- mt8195_afe_enable_clk(afe, afe_priv->clk[MT8195_CLK_AUD_ASRC12]); +- mt8195_afe_enable_clk(afe, afe_priv->clk[MT8195_CLK_AUD_PCMIF]); +- +- return 0; +-} +- +-static void mtk_dai_pcm_shutdown(struct snd_pcm_substream *substream, +- struct snd_soc_dai *dai) +-{ +- struct mtk_base_afe *afe = snd_soc_dai_get_drvdata(dai); +- struct mt8195_afe_private *afe_priv = afe->platform_priv; +- +- if (dai->component->active) +- return; +- +- mtk_dai_pcm_disable(afe); +- +- mt8195_afe_disable_clk(afe, afe_priv->clk[MT8195_CLK_AUD_PCMIF]); +- mt8195_afe_disable_clk(afe, afe_priv->clk[MT8195_CLK_AUD_ASRC12]); +- mt8195_afe_disable_clk(afe, afe_priv->clk[MT8195_CLK_AUD_ASRC11]); +-} +- + static int mtk_dai_pcm_prepare(struct snd_pcm_substream *substream, + struct snd_soc_dai *dai) + { +- struct mtk_base_afe *afe = snd_soc_dai_get_drvdata(dai); +- int ret = 0; ++ int ret; + +- if (snd_soc_dai_stream_active(dai, SNDRV_PCM_STREAM_PLAYBACK) && +- snd_soc_dai_stream_active(dai, SNDRV_PCM_STREAM_CAPTURE)) ++ dev_dbg(dai->dev, "%s(), id %d, stream %d, widget active p %d, c %d\n", ++ __func__, dai->id, substream->stream, ++ dai->playback_widget->active, dai->capture_widget->active); ++ ++ if (dai->playback_widget->active || dai->capture_widget->active) + return 0; + + ret = mtk_dai_pcm_configure(substream, dai); + if (ret) + return ret; + +- mtk_dai_pcm_enable(afe); +- + return 0; + } + +@@ -316,8 +287,6 @@ static int mtk_dai_pcm_set_fmt(struct snd_soc_dai *dai, unsigned int fmt) + } + + static const struct snd_soc_dai_ops mtk_dai_pcm_ops = { +- .startup = mtk_dai_pcm_startup, +- .shutdown = mtk_dai_pcm_shutdown, + .prepare = mtk_dai_pcm_prepare, + .set_fmt = mtk_dai_pcm_set_fmt, + }; +diff --git a/sound/soc/mediatek/mt8195/mt8195-reg.h b/sound/soc/mediatek/mt8195/mt8195-reg.h +index d06f9cf85a4ec..d3871353db415 100644 +--- a/sound/soc/mediatek/mt8195/mt8195-reg.h ++++ b/sound/soc/mediatek/mt8195/mt8195-reg.h +@@ -2550,6 +2550,7 @@ + #define PCM_INTF_CON1_PCM_FMT(x) (((x) & 0x3) << 1) + #define PCM_INTF_CON1_PCM_FMT_MASK (0x3 << 1) + #define PCM_INTF_CON1_PCM_EN BIT(0) ++#define PCM_INTF_CON1_PCM_EN_SHIFT 0 + + /* PCM_INTF_CON2 */ + #define PCM_INTF_CON2_CLK_DOMAIN_SEL(x) (((x) & 0x3) << 23) +diff --git a/sound/soc/samsung/idma.c b/sound/soc/samsung/idma.c +index 66bcc2f97544b..c3f1b054e2389 100644 +--- a/sound/soc/samsung/idma.c ++++ b/sound/soc/samsung/idma.c +@@ -360,6 +360,8 @@ static int preallocate_idma_buffer(struct snd_pcm *pcm, int stream) + buf->addr = idma.lp_tx_addr; + buf->bytes = idma_hardware.buffer_bytes_max; + buf->area = (unsigned char * __force)ioremap(buf->addr, buf->bytes); ++ if (!buf->area) ++ return -ENOMEM; + + return 0; + } +diff --git a/sound/soc/uniphier/Kconfig b/sound/soc/uniphier/Kconfig +index aa3592ee1358b..ddfa6424c656b 100644 +--- a/sound/soc/uniphier/Kconfig ++++ b/sound/soc/uniphier/Kconfig +@@ -23,7 +23,6 @@ config SND_SOC_UNIPHIER_LD11 + tristate "UniPhier LD11/LD20 Device Driver" + depends on SND_SOC_UNIPHIER + select SND_SOC_UNIPHIER_AIO +- select SND_SOC_UNIPHIER_AIO_DMA + help + This adds ASoC driver for Socionext UniPhier LD11/LD20 + input and output that can be used with other codecs. +@@ -34,7 +33,6 @@ config SND_SOC_UNIPHIER_PXS2 + tristate "UniPhier PXs2 Device Driver" + depends on SND_SOC_UNIPHIER + select SND_SOC_UNIPHIER_AIO +- select SND_SOC_UNIPHIER_AIO_DMA + help + This adds ASoC driver for Socionext UniPhier PXs2 + input and output that can be used with other codecs. +diff --git a/sound/usb/format.c b/sound/usb/format.c +index f5e676a51b30d..405dc0bf6678c 100644 +--- a/sound/usb/format.c ++++ b/sound/usb/format.c +@@ -375,7 +375,7 @@ static int parse_uac2_sample_rate_range(struct snd_usb_audio *chip, + for (rate = min; rate <= max; rate += res) { + + /* Filter out invalid rates on Presonus Studio 1810c */ +- if (chip->usb_id == USB_ID(0x0194f, 0x010c) && ++ if (chip->usb_id == USB_ID(0x194f, 0x010c) && + !s1810c_valid_sample_rate(fp, rate)) + goto skip_rate; + +diff --git a/sound/usb/mixer_quirks.c b/sound/usb/mixer_quirks.c +index 823b6b8de942d..d48729e6a3b0a 100644 +--- a/sound/usb/mixer_quirks.c ++++ b/sound/usb/mixer_quirks.c +@@ -3254,7 +3254,7 @@ int snd_usb_mixer_apply_create_quirk(struct usb_mixer_interface *mixer) + err = snd_rme_controls_create(mixer); + break; + +- case USB_ID(0x0194f, 0x010c): /* Presonus Studio 1810c */ ++ case USB_ID(0x194f, 0x010c): /* Presonus Studio 1810c */ + err = snd_sc1810_init_mixer(mixer); + break; + case USB_ID(0x2a39, 0x3fb0): /* RME Babyface Pro FS */ +diff --git a/sound/usb/quirks.c b/sound/usb/quirks.c +index 64e1c20311ed4..ab9f3da49941f 100644 +--- a/sound/usb/quirks.c ++++ b/sound/usb/quirks.c +@@ -1290,7 +1290,7 @@ int snd_usb_apply_interface_quirk(struct snd_usb_audio *chip, + if (chip->usb_id == USB_ID(0x0763, 0x2012)) + return fasttrackpro_skip_setting_quirk(chip, iface, altno); + /* presonus studio 1810c: skip altsets incompatible with device_setup */ +- if (chip->usb_id == USB_ID(0x0194f, 0x010c)) ++ if (chip->usb_id == USB_ID(0x194f, 0x010c)) + return s1810c_skip_setting_quirk(chip, iface, altno); + + +diff --git a/tools/bpf/bpftool/Documentation/Makefile b/tools/bpf/bpftool/Documentation/Makefile +index c49487905cebe..f89929c7038d5 100644 +--- a/tools/bpf/bpftool/Documentation/Makefile ++++ b/tools/bpf/bpftool/Documentation/Makefile +@@ -1,6 +1,5 @@ + # SPDX-License-Identifier: GPL-2.0-only + include ../../../scripts/Makefile.include +-include ../../../scripts/utilities.mak + + INSTALL ?= install + RM ?= rm -f +diff --git a/tools/bpf/bpftool/Documentation/bpftool-btf.rst b/tools/bpf/bpftool/Documentation/bpftool-btf.rst +index 88b28aa7431f6..4425d942dd39a 100644 +--- a/tools/bpf/bpftool/Documentation/bpftool-btf.rst ++++ b/tools/bpf/bpftool/Documentation/bpftool-btf.rst +@@ -13,7 +13,7 @@ SYNOPSIS + **bpftool** [*OPTIONS*] **btf** *COMMAND* + + *OPTIONS* := { { **-j** | **--json** } [{ **-p** | **--pretty** }] | {**-d** | **--debug** } | +- { **-B** | **--base-btf** } } ++ { **-B** | **--base-btf** } } + + *COMMANDS* := { **dump** | **help** } + +diff --git a/tools/bpf/bpftool/Documentation/bpftool-cgroup.rst b/tools/bpf/bpftool/Documentation/bpftool-cgroup.rst +index 3e4395eede4f7..13a217a2503d8 100644 +--- a/tools/bpf/bpftool/Documentation/bpftool-cgroup.rst ++++ b/tools/bpf/bpftool/Documentation/bpftool-cgroup.rst +@@ -13,7 +13,7 @@ SYNOPSIS + **bpftool** [*OPTIONS*] **cgroup** *COMMAND* + + *OPTIONS* := { { **-j** | **--json** } [{ **-p** | **--pretty** }] | { **-d** | **--debug** } | +- { **-f** | **--bpffs** } } ++ { **-f** | **--bpffs** } } + + *COMMANDS* := + { **show** | **list** | **tree** | **attach** | **detach** | **help** } +diff --git a/tools/bpf/bpftool/Documentation/bpftool-gen.rst b/tools/bpf/bpftool/Documentation/bpftool-gen.rst +index 2ef2f2df02799..2a137f8a4cea0 100644 +--- a/tools/bpf/bpftool/Documentation/bpftool-gen.rst ++++ b/tools/bpf/bpftool/Documentation/bpftool-gen.rst +@@ -13,7 +13,7 @@ SYNOPSIS + **bpftool** [*OPTIONS*] **gen** *COMMAND* + + *OPTIONS* := { { **-j** | **--json** } [{ **-p** | **--pretty** }] | { **-d** | **--debug** } | +- { **-L** | **--use-loader** } } ++ { **-L** | **--use-loader** } } + + *COMMAND* := { **object** | **skeleton** | **help** } + +diff --git a/tools/bpf/bpftool/Documentation/bpftool-link.rst b/tools/bpf/bpftool/Documentation/bpftool-link.rst +index 0de90f086238c..9434349636a5e 100644 +--- a/tools/bpf/bpftool/Documentation/bpftool-link.rst ++++ b/tools/bpf/bpftool/Documentation/bpftool-link.rst +@@ -13,7 +13,7 @@ SYNOPSIS + **bpftool** [*OPTIONS*] **link** *COMMAND* + + *OPTIONS* := { { **-j** | **--json** } [{ **-p** | **--pretty** }] | { **-d** | **--debug** } | +- { **-f** | **--bpffs** } | { **-n** | **--nomount** } } ++ { **-f** | **--bpffs** } | { **-n** | **--nomount** } } + + *COMMANDS* := { **show** | **list** | **pin** | **help** } + +diff --git a/tools/bpf/bpftool/Documentation/bpftool-map.rst b/tools/bpf/bpftool/Documentation/bpftool-map.rst +index d0c4abe08abab..1445cadc15d4c 100644 +--- a/tools/bpf/bpftool/Documentation/bpftool-map.rst ++++ b/tools/bpf/bpftool/Documentation/bpftool-map.rst +@@ -13,11 +13,11 @@ SYNOPSIS + **bpftool** [*OPTIONS*] **map** *COMMAND* + + *OPTIONS* := { { **-j** | **--json** } [{ **-p** | **--pretty** }] | { **-d** | **--debug** } | +- { **-f** | **--bpffs** } | { **-n** | **--nomount** } } ++ { **-f** | **--bpffs** } | { **-n** | **--nomount** } } + + *COMMANDS* := +- { **show** | **list** | **create** | **dump** | **update** | **lookup** | **getnext** +- | **delete** | **pin** | **help** } ++ { **show** | **list** | **create** | **dump** | **update** | **lookup** | **getnext** | ++ **delete** | **pin** | **help** } + + MAP COMMANDS + ============= +diff --git a/tools/bpf/bpftool/Documentation/bpftool-prog.rst b/tools/bpf/bpftool/Documentation/bpftool-prog.rst +index 91608cb7e44a0..f27265bd589b4 100644 +--- a/tools/bpf/bpftool/Documentation/bpftool-prog.rst ++++ b/tools/bpf/bpftool/Documentation/bpftool-prog.rst +@@ -13,12 +13,12 @@ SYNOPSIS + **bpftool** [*OPTIONS*] **prog** *COMMAND* + + *OPTIONS* := { { **-j** | **--json** } [{ **-p** | **--pretty** }] | { **-d** | **--debug** } | +- { **-f** | **--bpffs** } | { **-m** | **--mapcompat** } | { **-n** | **--nomount** } | +- { **-L** | **--use-loader** } } ++ { **-f** | **--bpffs** } | { **-m** | **--mapcompat** } | { **-n** | **--nomount** } | ++ { **-L** | **--use-loader** } } + + *COMMANDS* := +- { **show** | **list** | **dump xlated** | **dump jited** | **pin** | **load** +- | **loadall** | **help** } ++ { **show** | **list** | **dump xlated** | **dump jited** | **pin** | **load** | ++ **loadall** | **help** } + + PROG COMMANDS + ============= +diff --git a/tools/bpf/bpftool/Documentation/bpftool.rst b/tools/bpf/bpftool/Documentation/bpftool.rst +index bb23f55bb05ad..8ac86565c501e 100644 +--- a/tools/bpf/bpftool/Documentation/bpftool.rst ++++ b/tools/bpf/bpftool/Documentation/bpftool.rst +@@ -19,14 +19,14 @@ SYNOPSIS + *OBJECT* := { **map** | **program** | **cgroup** | **perf** | **net** | **feature** } + + *OPTIONS* := { { **-V** | **--version** } | +- { **-j** | **--json** } [{ **-p** | **--pretty** }] | { **-d** | **--debug** } } ++ { **-j** | **--json** } [{ **-p** | **--pretty** }] | { **-d** | **--debug** } } + + *MAP-COMMANDS* := + { **show** | **list** | **create** | **dump** | **update** | **lookup** | **getnext** | +- **delete** | **pin** | **event_pipe** | **help** } ++ **delete** | **pin** | **event_pipe** | **help** } + + *PROG-COMMANDS* := { **show** | **list** | **dump jited** | **dump xlated** | **pin** | +- **load** | **attach** | **detach** | **help** } ++ **load** | **attach** | **detach** | **help** } + + *CGROUP-COMMANDS* := { **show** | **list** | **attach** | **detach** | **help** } + +diff --git a/tools/bpf/bpftool/Makefile b/tools/bpf/bpftool/Makefile +index d73232be1e991..cce52df3be064 100644 +--- a/tools/bpf/bpftool/Makefile ++++ b/tools/bpf/bpftool/Makefile +@@ -1,6 +1,5 @@ + # SPDX-License-Identifier: GPL-2.0-only + include ../../scripts/Makefile.include +-include ../../scripts/utilities.mak + + ifeq ($(srctree),) + srctree := $(patsubst %/,%,$(dir $(CURDIR))) +diff --git a/tools/bpf/bpftool/main.c b/tools/bpf/bpftool/main.c +index 02eaaf065f651..d27ec4f852bbb 100644 +--- a/tools/bpf/bpftool/main.c ++++ b/tools/bpf/bpftool/main.c +@@ -402,6 +402,8 @@ int main(int argc, char **argv) + }; + int opt, ret; + ++ setlinebuf(stdout); ++ + last_do_help = do_help; + pretty_output = false; + json_output = false; +diff --git a/tools/bpf/bpftool/prog.c b/tools/bpf/bpftool/prog.c +index fe59404e87046..f8755beb3d9eb 100644 +--- a/tools/bpf/bpftool/prog.c ++++ b/tools/bpf/bpftool/prog.c +@@ -629,8 +629,8 @@ prog_dump(struct bpf_prog_info *info, enum dump_mode mode, + char func_sig[1024]; + unsigned char *buf; + __u32 member_len; ++ int fd, err = -1; + ssize_t n; +- int fd; + + if (mode == DUMP_JITED) { + if (info->jited_prog_len == 0 || !info->jited_prog_insns) { +@@ -669,7 +669,7 @@ prog_dump(struct bpf_prog_info *info, enum dump_mode mode, + if (fd < 0) { + p_err("can't open file %s: %s", filepath, + strerror(errno)); +- return -1; ++ goto exit_free; + } + + n = write(fd, buf, member_len); +@@ -677,7 +677,7 @@ prog_dump(struct bpf_prog_info *info, enum dump_mode mode, + if (n != (ssize_t)member_len) { + p_err("error writing output file: %s", + n < 0 ? strerror(errno) : "short write"); +- return -1; ++ goto exit_free; + } + + if (json_output) +@@ -691,7 +691,7 @@ prog_dump(struct bpf_prog_info *info, enum dump_mode mode, + info->netns_ino, + &disasm_opt); + if (!name) +- return -1; ++ goto exit_free; + } + + if (info->nr_jited_func_lens && info->jited_func_lens) { +@@ -786,9 +786,12 @@ prog_dump(struct bpf_prog_info *info, enum dump_mode mode, + kernel_syms_destroy(&dd); + } + +- btf__free(btf); ++ err = 0; + +- return 0; ++exit_free: ++ btf__free(btf); ++ bpf_prog_linfo__free(prog_linfo); ++ return err; + } + + static int do_dump(int argc, char **argv) +diff --git a/tools/include/nolibc/nolibc.h b/tools/include/nolibc/nolibc.h +index 3430667b0d241..3e2c6f2ed587f 100644 +--- a/tools/include/nolibc/nolibc.h ++++ b/tools/include/nolibc/nolibc.h +@@ -399,16 +399,22 @@ struct stat { + }) + + /* startup code */ ++/* ++ * x86-64 System V ABI mandates: ++ * 1) %rsp must be 16-byte aligned right before the function call. ++ * 2) The deepest stack frame should be zero (the %rbp). ++ * ++ */ + asm(".section .text\n" + ".global _start\n" + "_start:\n" + "pop %rdi\n" // argc (first arg, %rdi) + "mov %rsp, %rsi\n" // argv[] (second arg, %rsi) + "lea 8(%rsi,%rdi,8),%rdx\n" // then a NULL then envp (third arg, %rdx) +- "and $-16, %rsp\n" // x86 ABI : esp must be 16-byte aligned when +- "sub $8, %rsp\n" // entering the callee ++ "xor %ebp, %ebp\n" // zero the stack frame ++ "and $-16, %rsp\n" // x86 ABI : esp must be 16-byte aligned before call + "call main\n" // main() returns the status code, we'll exit with it. +- "movzb %al, %rdi\n" // retrieve exit code from 8 lower bits ++ "mov %eax, %edi\n" // retrieve exit code (32 bit) + "mov $60, %rax\n" // NR_exit == 60 + "syscall\n" // really exit + "hlt\n" // ensure it does not return +@@ -577,20 +583,28 @@ struct sys_stat_struct { + }) + + /* startup code */ ++/* ++ * i386 System V ABI mandates: ++ * 1) last pushed argument must be 16-byte aligned. ++ * 2) The deepest stack frame should be set to zero ++ * ++ */ + asm(".section .text\n" + ".global _start\n" + "_start:\n" + "pop %eax\n" // argc (first arg, %eax) + "mov %esp, %ebx\n" // argv[] (second arg, %ebx) + "lea 4(%ebx,%eax,4),%ecx\n" // then a NULL then envp (third arg, %ecx) +- "and $-16, %esp\n" // x86 ABI : esp must be 16-byte aligned when ++ "xor %ebp, %ebp\n" // zero the stack frame ++ "and $-16, %esp\n" // x86 ABI : esp must be 16-byte aligned before ++ "sub $4, %esp\n" // the call instruction (args are aligned) + "push %ecx\n" // push all registers on the stack so that we + "push %ebx\n" // support both regparm and plain stack modes + "push %eax\n" + "call main\n" // main() returns the status code in %eax +- "movzbl %al, %ebx\n" // retrieve exit code from lower 8 bits +- "movl $1, %eax\n" // NR_exit == 1 +- "int $0x80\n" // exit now ++ "mov %eax, %ebx\n" // retrieve exit code (32-bit int) ++ "movl $1, %eax\n" // NR_exit == 1 ++ "int $0x80\n" // exit now + "hlt\n" // ensure it does not + ""); + +@@ -774,7 +788,6 @@ asm(".section .text\n" + "and %r3, %r1, $-8\n" // AAPCS : sp must be 8-byte aligned in the + "mov %sp, %r3\n" // callee, an bl doesn't push (lr=pc) + "bl main\n" // main() returns the status code, we'll exit with it. +- "and %r0, %r0, $0xff\n" // limit exit code to 8 bits + "movs r7, $1\n" // NR_exit == 1 + "svc $0x00\n" + ""); +@@ -971,7 +984,6 @@ asm(".section .text\n" + "add x2, x2, x1\n" // + argv + "and sp, x1, -16\n" // sp must be 16-byte aligned in the callee + "bl main\n" // main() returns the status code, we'll exit with it. +- "and x0, x0, 0xff\n" // limit exit code to 8 bits + "mov x8, 93\n" // NR_exit == 93 + "svc #0\n" + ""); +@@ -1176,7 +1188,7 @@ asm(".section .text\n" + "addiu $sp,$sp,-16\n" // the callee expects to save a0..a3 there! + "jal main\n" // main() returns the status code, we'll exit with it. + "nop\n" // delayed slot +- "and $a0, $v0, 0xff\n" // limit exit code to 8 bits ++ "move $a0, $v0\n" // retrieve 32-bit exit code from v0 + "li $v0, 4001\n" // NR_exit == 4001 + "syscall\n" + ".end __start\n" +@@ -1374,7 +1386,6 @@ asm(".section .text\n" + "add a2,a2,a1\n" // + argv + "andi sp,a1,-16\n" // sp must be 16-byte aligned + "call main\n" // main() returns the status code, we'll exit with it. +- "andi a0, a0, 0xff\n" // limit exit code to 8 bits + "li a7, 93\n" // NR_exit == 93 + "ecall\n" + ""); +diff --git a/tools/lib/bpf/btf.c b/tools/lib/bpf/btf.c +index 1b9341ef638b0..5f3d20ae66d56 100644 +--- a/tools/lib/bpf/btf.c ++++ b/tools/lib/bpf/btf.c +@@ -2626,15 +2626,11 @@ void btf_ext__free(struct btf_ext *btf_ext) + free(btf_ext); + } + +-struct btf_ext *btf_ext__new(__u8 *data, __u32 size) ++struct btf_ext *btf_ext__new(const __u8 *data, __u32 size) + { + struct btf_ext *btf_ext; + int err; + +- err = btf_ext_parse_hdr(data, size); +- if (err) +- return libbpf_err_ptr(err); +- + btf_ext = calloc(1, sizeof(struct btf_ext)); + if (!btf_ext) + return libbpf_err_ptr(-ENOMEM); +@@ -2647,6 +2643,10 @@ struct btf_ext *btf_ext__new(__u8 *data, __u32 size) + } + memcpy(btf_ext->data, data, size); + ++ err = btf_ext_parse_hdr(btf_ext->data, size); ++ if (err) ++ goto done; ++ + if (btf_ext->hdr->hdr_len < offsetofend(struct btf_ext_header, line_info_len)) { + err = -EINVAL; + goto done; +@@ -3358,8 +3358,8 @@ static long btf_hash_struct(struct btf_type *t) + } + + /* +- * Check structural compatibility of two FUNC_PROTOs, ignoring referenced type +- * IDs. This check is performed during type graph equivalence check and ++ * Check structural compatibility of two STRUCTs/UNIONs, ignoring referenced ++ * type IDs. This check is performed during type graph equivalence check and + * referenced types equivalence is checked separately. + */ + static bool btf_shallow_equal_struct(struct btf_type *t1, struct btf_type *t2) +@@ -3730,6 +3730,31 @@ static int btf_dedup_identical_arrays(struct btf_dedup *d, __u32 id1, __u32 id2) + return btf_equal_array(t1, t2); + } + ++/* Check if given two types are identical STRUCT/UNION definitions */ ++static bool btf_dedup_identical_structs(struct btf_dedup *d, __u32 id1, __u32 id2) ++{ ++ const struct btf_member *m1, *m2; ++ struct btf_type *t1, *t2; ++ int n, i; ++ ++ t1 = btf_type_by_id(d->btf, id1); ++ t2 = btf_type_by_id(d->btf, id2); ++ ++ if (!btf_is_composite(t1) || btf_kind(t1) != btf_kind(t2)) ++ return false; ++ ++ if (!btf_shallow_equal_struct(t1, t2)) ++ return false; ++ ++ m1 = btf_members(t1); ++ m2 = btf_members(t2); ++ for (i = 0, n = btf_vlen(t1); i < n; i++, m1++, m2++) { ++ if (m1->type != m2->type) ++ return false; ++ } ++ return true; ++} ++ + /* + * Check equivalence of BTF type graph formed by candidate struct/union (we'll + * call it "candidate graph" in this description for brevity) to a type graph +@@ -3841,6 +3866,8 @@ static int btf_dedup_is_equiv(struct btf_dedup *d, __u32 cand_id, + + hypot_type_id = d->hypot_map[canon_id]; + if (hypot_type_id <= BTF_MAX_NR_TYPES) { ++ if (hypot_type_id == cand_id) ++ return 1; + /* In some cases compiler will generate different DWARF types + * for *identical* array type definitions and use them for + * different fields within the *same* struct. This breaks type +@@ -3849,8 +3876,18 @@ static int btf_dedup_is_equiv(struct btf_dedup *d, __u32 cand_id, + * types within a single CU. So work around that by explicitly + * allowing identical array types here. + */ +- return hypot_type_id == cand_id || +- btf_dedup_identical_arrays(d, hypot_type_id, cand_id); ++ if (btf_dedup_identical_arrays(d, hypot_type_id, cand_id)) ++ return 1; ++ /* It turns out that similar situation can happen with ++ * struct/union sometimes, sigh... Handle the case where ++ * structs/unions are exactly the same, down to the referenced ++ * type IDs. Anything more complicated (e.g., if referenced ++ * types are different, but equivalent) is *way more* ++ * complicated and requires a many-to-many equivalence mapping. ++ */ ++ if (btf_dedup_identical_structs(d, hypot_type_id, cand_id)) ++ return 1; ++ return 0; + } + + if (btf_dedup_hypot_map_add(d, canon_id, cand_id)) +diff --git a/tools/lib/bpf/btf.h b/tools/lib/bpf/btf.h +index 4a711f990904b..b0ee338a0cc87 100644 +--- a/tools/lib/bpf/btf.h ++++ b/tools/lib/bpf/btf.h +@@ -80,7 +80,7 @@ LIBBPF_API int btf__get_map_kv_tids(const struct btf *btf, const char *map_name, + __u32 expected_value_size, + __u32 *key_type_id, __u32 *value_type_id); + +-LIBBPF_API struct btf_ext *btf_ext__new(__u8 *data, __u32 size); ++LIBBPF_API struct btf_ext *btf_ext__new(const __u8 *data, __u32 size); + LIBBPF_API void btf_ext__free(struct btf_ext *btf_ext); + LIBBPF_API const void *btf_ext__get_raw_data(const struct btf_ext *btf_ext, + __u32 *size); +diff --git a/tools/lib/bpf/btf_dump.c b/tools/lib/bpf/btf_dump.c +index e4b483f15fb99..8c9325802793b 100644 +--- a/tools/lib/bpf/btf_dump.c ++++ b/tools/lib/bpf/btf_dump.c +@@ -2186,7 +2186,7 @@ static int btf_dump_dump_type_data(struct btf_dump *d, + __u8 bits_offset, + __u8 bit_sz) + { +- int size, err; ++ int size, err = 0; + + size = btf_dump_type_data_check_overflow(d, t, id, data, bits_offset); + if (size < 0) +diff --git a/tools/lib/bpf/gen_loader.c b/tools/lib/bpf/gen_loader.c +index 8df718a6b142d..33c19590ee434 100644 +--- a/tools/lib/bpf/gen_loader.c ++++ b/tools/lib/bpf/gen_loader.c +@@ -663,9 +663,11 @@ void bpf_gen__prog_load(struct bpf_gen *gen, + debug_ret(gen, "prog_load %s insn_cnt %d", attr.prog_name, attr.insn_cnt); + /* successful or not, close btf module FDs used in extern ksyms and attach_btf_obj_fd */ + cleanup_relos(gen, insns); +- if (gen->attach_kind) ++ if (gen->attach_kind) { + emit_sys_close_blob(gen, + attr_field(prog_load_attr, attach_btf_obj_fd)); ++ gen->attach_kind = 0; ++ } + emit_check_err(gen); + /* remember prog_fd in the stack, if successful */ + emit(gen, BPF_STX_MEM(BPF_W, BPF_REG_10, BPF_REG_7, +diff --git a/tools/lib/bpf/libbpf.c b/tools/lib/bpf/libbpf.c +index 7145463a4a562..0ad29203cbfbf 100644 +--- a/tools/lib/bpf/libbpf.c ++++ b/tools/lib/bpf/libbpf.c +@@ -8676,7 +8676,10 @@ int bpf_map__set_inner_map_fd(struct bpf_map *map, int fd) + pr_warn("error: inner_map_fd already specified\n"); + return libbpf_err(-EINVAL); + } +- zfree(&map->inner_map); ++ if (map->inner_map) { ++ bpf_map__destroy(map->inner_map); ++ zfree(&map->inner_map); ++ } + map->inner_map_fd = fd; + return 0; + } +diff --git a/tools/lib/bpf/linker.c b/tools/lib/bpf/linker.c +index 2df880cefdaee..6b2f59ddb6918 100644 +--- a/tools/lib/bpf/linker.c ++++ b/tools/lib/bpf/linker.c +@@ -211,6 +211,7 @@ void bpf_linker__free(struct bpf_linker *linker) + } + free(linker->secs); + ++ free(linker->glob_syms); + free(linker); + } + +@@ -2000,7 +2001,7 @@ add_sym: + static int linker_append_elf_relos(struct bpf_linker *linker, struct src_obj *obj) + { + struct src_sec *src_symtab = &obj->secs[obj->symtab_sec_idx]; +- struct dst_sec *dst_symtab = &linker->secs[linker->symtab_sec_idx]; ++ struct dst_sec *dst_symtab; + int i, err; + + for (i = 1; i < obj->sec_cnt; i++) { +@@ -2033,6 +2034,9 @@ static int linker_append_elf_relos(struct bpf_linker *linker, struct src_obj *ob + return -1; + } + ++ /* add_dst_sec() above could have invalidated linker->secs */ ++ dst_symtab = &linker->secs[linker->symtab_sec_idx]; ++ + /* shdr->sh_link points to SYMTAB */ + dst_sec->shdr->sh_link = linker->symtab_sec_idx; + +diff --git a/tools/perf/Makefile.config b/tools/perf/Makefile.config +index 3c077f61d676d..71772b20ea737 100644 +--- a/tools/perf/Makefile.config ++++ b/tools/perf/Makefile.config +@@ -143,7 +143,10 @@ FEATURE_CHECK_LDFLAGS-libcrypto = -lcrypto + ifdef CSINCLUDES + LIBOPENCSD_CFLAGS := -I$(CSINCLUDES) + endif +-OPENCSDLIBS := -lopencsd_c_api -lopencsd -lstdc++ ++OPENCSDLIBS := -lopencsd_c_api ++ifeq ($(findstring -static,${LDFLAGS}),-static) ++ OPENCSDLIBS += -lopencsd -lstdc++ ++endif + ifdef CSLIBS + LIBOPENCSD_LDFLAGS := -L$(CSLIBS) + endif +diff --git a/tools/perf/util/debug.c b/tools/perf/util/debug.c +index 2c06abf6dcd26..65e6c22f38e4f 100644 +--- a/tools/perf/util/debug.c ++++ b/tools/perf/util/debug.c +@@ -179,7 +179,7 @@ static int trace_event_printer(enum binary_printer_ops op, + break; + case BINARY_PRINT_CHAR_DATA: + printed += color_fprintf(fp, color, "%c", +- isprint(ch) ? ch : '.'); ++ isprint(ch) && isascii(ch) ? ch : '.'); + break; + case BINARY_PRINT_CHAR_PAD: + printed += color_fprintf(fp, color, " "); +diff --git a/tools/perf/util/evsel.c b/tools/perf/util/evsel.c +index dbfeceb2546c6..c87f9974c0c10 100644 +--- a/tools/perf/util/evsel.c ++++ b/tools/perf/util/evsel.c +@@ -1047,6 +1047,17 @@ void __weak arch_evsel__set_sample_weight(struct evsel *evsel) + evsel__set_sample_bit(evsel, WEIGHT); + } + ++static void evsel__set_default_freq_period(struct record_opts *opts, ++ struct perf_event_attr *attr) ++{ ++ if (opts->freq) { ++ attr->freq = 1; ++ attr->sample_freq = opts->freq; ++ } else { ++ attr->sample_period = opts->default_interval; ++ } ++} ++ + /* + * The enable_on_exec/disabled value strategy: + * +@@ -1113,14 +1124,12 @@ void evsel__config(struct evsel *evsel, struct record_opts *opts, + * We default some events to have a default interval. But keep + * it a weak assumption overridable by the user. + */ +- if (!attr->sample_period) { +- if (opts->freq) { +- attr->freq = 1; +- attr->sample_freq = opts->freq; +- } else { +- attr->sample_period = opts->default_interval; +- } +- } ++ if ((evsel->is_libpfm_event && !attr->sample_period) || ++ (!evsel->is_libpfm_event && (!attr->sample_period || ++ opts->user_freq != UINT_MAX || ++ opts->user_interval != ULLONG_MAX))) ++ evsel__set_default_freq_period(opts, attr); ++ + /* + * If attr->freq was set (here or earlier), ask for period + * to be sampled. +diff --git a/tools/perf/util/probe-event.c b/tools/perf/util/probe-event.c +index b2a02c9ab8ea9..a834918a0a0d3 100644 +--- a/tools/perf/util/probe-event.c ++++ b/tools/perf/util/probe-event.c +@@ -3083,6 +3083,9 @@ static int find_probe_trace_events_from_map(struct perf_probe_event *pev, + for (j = 0; j < num_matched_functions; j++) { + sym = syms[j]; + ++ if (sym->type != STT_FUNC) ++ continue; ++ + /* There can be duplicated symbols in the map */ + for (i = 0; i < j; i++) + if (sym->start == syms[i]->start) { +diff --git a/tools/testing/selftests/bpf/btf_helpers.c b/tools/testing/selftests/bpf/btf_helpers.c +index b692e6ead9b55..0a4ad7cb2c200 100644 +--- a/tools/testing/selftests/bpf/btf_helpers.c ++++ b/tools/testing/selftests/bpf/btf_helpers.c +@@ -246,18 +246,23 @@ const char *btf_type_c_dump(const struct btf *btf) + d = btf_dump__new(btf, NULL, &opts, btf_dump_printf); + if (libbpf_get_error(d)) { + fprintf(stderr, "Failed to create btf_dump instance: %ld\n", libbpf_get_error(d)); +- return NULL; ++ goto err_out; + } + + for (i = 1; i <= btf__get_nr_types(btf); i++) { + err = btf_dump__dump_type(d, i); + if (err) { + fprintf(stderr, "Failed to dump type [%d]: %d\n", i, err); +- return NULL; ++ goto err_out; + } + } + ++ btf_dump__free(d); + fflush(buf_file); + fclose(buf_file); + return buf; ++err_out: ++ btf_dump__free(d); ++ fclose(buf_file); ++ return NULL; + } +diff --git a/tools/testing/selftests/bpf/prog_tests/bpf_iter.c b/tools/testing/selftests/bpf/prog_tests/bpf_iter.c +index 77ac24b191d4c..dc18e5ae0febc 100644 +--- a/tools/testing/selftests/bpf/prog_tests/bpf_iter.c ++++ b/tools/testing/selftests/bpf/prog_tests/bpf_iter.c +@@ -1208,13 +1208,14 @@ static void test_task_vma(void) + goto out; + + /* Read CMP_BUFFER_SIZE (1kB) from bpf_iter. Read in small chunks +- * to trigger seq_file corner cases. The expected output is much +- * longer than 1kB, so the while loop will terminate. ++ * to trigger seq_file corner cases. + */ + len = 0; + while (len < CMP_BUFFER_SIZE) { + err = read_fd_into_buffer(iter_fd, task_vma_output + len, + min(read_size, CMP_BUFFER_SIZE - len)); ++ if (!err) ++ break; + if (CHECK(err < 0, "read_iter_fd", "read_iter_fd failed\n")) + goto out; + len += err; +diff --git a/tools/testing/selftests/bpf/prog_tests/migrate_reuseport.c b/tools/testing/selftests/bpf/prog_tests/migrate_reuseport.c +index 59adb4715394f..3c85247f96f95 100644 +--- a/tools/testing/selftests/bpf/prog_tests/migrate_reuseport.c ++++ b/tools/testing/selftests/bpf/prog_tests/migrate_reuseport.c +@@ -204,8 +204,8 @@ static int pass_ack(struct migrate_reuseport_test_case *test_case) + { + int err; + +- err = bpf_link__detach(test_case->link); +- if (!ASSERT_OK(err, "bpf_link__detach")) ++ err = bpf_link__destroy(test_case->link); ++ if (!ASSERT_OK(err, "bpf_link__destroy")) + return -1; + + test_case->link = NULL; +diff --git a/tools/testing/selftests/bpf/prog_tests/skb_ctx.c b/tools/testing/selftests/bpf/prog_tests/skb_ctx.c +index fafeddaad6a99..23915be6172d6 100644 +--- a/tools/testing/selftests/bpf/prog_tests/skb_ctx.c ++++ b/tools/testing/selftests/bpf/prog_tests/skb_ctx.c +@@ -105,4 +105,6 @@ void test_skb_ctx(void) + "ctx_out_mark", + "skb->mark == %u, expected %d\n", + skb.mark, 10); ++ ++ bpf_object__close(obj); + } +diff --git a/tools/testing/selftests/bpf/prog_tests/tc_redirect.c b/tools/testing/selftests/bpf/prog_tests/tc_redirect.c +index e7201ba29ccd6..47e3159729d21 100644 +--- a/tools/testing/selftests/bpf/prog_tests/tc_redirect.c ++++ b/tools/testing/selftests/bpf/prog_tests/tc_redirect.c +@@ -105,6 +105,13 @@ static int setns_by_fd(int nsfd) + if (!ASSERT_OK(err, "unshare")) + return err; + ++ /* Make our /sys mount private, so the following umount won't ++ * trigger the global umount in case it's shared. ++ */ ++ err = mount("none", "/sys", NULL, MS_PRIVATE, NULL); ++ if (!ASSERT_OK(err, "remount private /sys")) ++ return err; ++ + err = umount2("/sys", MNT_DETACH); + if (!ASSERT_OK(err, "umount2 /sys")) + return err; +diff --git a/tools/testing/selftests/clone3/clone3.c b/tools/testing/selftests/clone3/clone3.c +index 42be3b9258301..076cf4325f783 100644 +--- a/tools/testing/selftests/clone3/clone3.c ++++ b/tools/testing/selftests/clone3/clone3.c +@@ -52,6 +52,12 @@ static int call_clone3(uint64_t flags, size_t size, enum test_mode test_mode) + size = sizeof(struct __clone_args); + + switch (test_mode) { ++ case CLONE3_ARGS_NO_TEST: ++ /* ++ * Uses default 'flags' and 'SIGCHLD' ++ * assignment. ++ */ ++ break; + case CLONE3_ARGS_ALL_0: + args.flags = 0; + args.exit_signal = 0; +diff --git a/tools/testing/selftests/ftrace/test.d/kprobe/profile.tc b/tools/testing/selftests/ftrace/test.d/kprobe/profile.tc +index 98166fa3eb91c..34fb89b0c61fa 100644 +--- a/tools/testing/selftests/ftrace/test.d/kprobe/profile.tc ++++ b/tools/testing/selftests/ftrace/test.d/kprobe/profile.tc +@@ -1,6 +1,6 @@ + #!/bin/sh + # SPDX-License-Identifier: GPL-2.0 +-# description: Kprobe dynamic event - adding and removing ++# description: Kprobe profile + # requires: kprobe_events + + ! grep -q 'myevent' kprobe_profile +diff --git a/tools/testing/selftests/kselftest_harness.h b/tools/testing/selftests/kselftest_harness.h +index ae0f0f33b2a6e..79a182cfa43ad 100644 +--- a/tools/testing/selftests/kselftest_harness.h ++++ b/tools/testing/selftests/kselftest_harness.h +@@ -969,7 +969,7 @@ void __run_test(struct __fixture_metadata *f, + t->passed = 1; + t->skip = 0; + t->trigger = 0; +- t->step = 0; ++ t->step = 1; + t->no_print = 0; + memset(t->results->reason, 0, sizeof(t->results->reason)); + +diff --git a/tools/testing/selftests/powerpc/security/spectre_v2.c b/tools/testing/selftests/powerpc/security/spectre_v2.c +index adc2b7294e5fd..83647b8277e7d 100644 +--- a/tools/testing/selftests/powerpc/security/spectre_v2.c ++++ b/tools/testing/selftests/powerpc/security/spectre_v2.c +@@ -193,7 +193,7 @@ int spectre_v2_test(void) + * We are not vulnerable and reporting otherwise, so + * missing such a mismatch is safe. + */ +- if (state == VULNERABLE) ++ if (miss_percent > 95) + return 4; + + return 1; +diff --git a/tools/testing/selftests/powerpc/signal/.gitignore b/tools/testing/selftests/powerpc/signal/.gitignore +index ce3375cd8e73e..8f6c816099a48 100644 +--- a/tools/testing/selftests/powerpc/signal/.gitignore ++++ b/tools/testing/selftests/powerpc/signal/.gitignore +@@ -4,3 +4,4 @@ signal_tm + sigfuz + sigreturn_vdso + sig_sc_double_restart ++sigreturn_kernel +diff --git a/tools/testing/selftests/powerpc/signal/Makefile b/tools/testing/selftests/powerpc/signal/Makefile +index d6ae54663aed7..84e201572466d 100644 +--- a/tools/testing/selftests/powerpc/signal/Makefile ++++ b/tools/testing/selftests/powerpc/signal/Makefile +@@ -1,5 +1,6 @@ + # SPDX-License-Identifier: GPL-2.0 + TEST_GEN_PROGS := signal signal_tm sigfuz sigreturn_vdso sig_sc_double_restart ++TEST_GEN_PROGS += sigreturn_kernel + + CFLAGS += -maltivec + $(OUTPUT)/signal_tm: CFLAGS += -mhtm +diff --git a/tools/testing/selftests/powerpc/signal/sigreturn_kernel.c b/tools/testing/selftests/powerpc/signal/sigreturn_kernel.c +new file mode 100644 +index 0000000000000..0a1b6e591eeed +--- /dev/null ++++ b/tools/testing/selftests/powerpc/signal/sigreturn_kernel.c +@@ -0,0 +1,132 @@ ++// SPDX-License-Identifier: GPL-2.0 ++/* ++ * Test that we can't sigreturn to kernel addresses, or to kernel mode. ++ */ ++ ++#define _GNU_SOURCE ++ ++#include <stdio.h> ++#include <signal.h> ++#include <stdlib.h> ++#include <sys/types.h> ++#include <sys/wait.h> ++#include <unistd.h> ++ ++#include "utils.h" ++ ++#define MSR_PR (1ul << 14) ++ ++static volatile unsigned long long sigreturn_addr; ++static volatile unsigned long long sigreturn_msr_mask; ++ ++static void sigusr1_handler(int signo, siginfo_t *si, void *uc_ptr) ++{ ++ ucontext_t *uc = (ucontext_t *)uc_ptr; ++ ++ if (sigreturn_addr) ++ UCONTEXT_NIA(uc) = sigreturn_addr; ++ ++ if (sigreturn_msr_mask) ++ UCONTEXT_MSR(uc) &= sigreturn_msr_mask; ++} ++ ++static pid_t fork_child(void) ++{ ++ pid_t pid; ++ ++ pid = fork(); ++ if (pid == 0) { ++ raise(SIGUSR1); ++ exit(0); ++ } ++ ++ return pid; ++} ++ ++static int expect_segv(pid_t pid) ++{ ++ int child_ret; ++ ++ waitpid(pid, &child_ret, 0); ++ FAIL_IF(WIFEXITED(child_ret)); ++ FAIL_IF(!WIFSIGNALED(child_ret)); ++ FAIL_IF(WTERMSIG(child_ret) != 11); ++ ++ return 0; ++} ++ ++int test_sigreturn_kernel(void) ++{ ++ struct sigaction act; ++ int child_ret, i; ++ pid_t pid; ++ ++ act.sa_sigaction = sigusr1_handler; ++ act.sa_flags = SA_SIGINFO; ++ sigemptyset(&act.sa_mask); ++ ++ FAIL_IF(sigaction(SIGUSR1, &act, NULL)); ++ ++ for (i = 0; i < 2; i++) { ++ // Return to kernel ++ sigreturn_addr = 0xcull << 60; ++ pid = fork_child(); ++ expect_segv(pid); ++ ++ // Return to kernel virtual ++ sigreturn_addr = 0xc008ull << 48; ++ pid = fork_child(); ++ expect_segv(pid); ++ ++ // Return out of range ++ sigreturn_addr = 0xc010ull << 48; ++ pid = fork_child(); ++ expect_segv(pid); ++ ++ // Return to no-man's land, just below PAGE_OFFSET ++ sigreturn_addr = (0xcull << 60) - (64 * 1024); ++ pid = fork_child(); ++ expect_segv(pid); ++ ++ // Return to no-man's land, above TASK_SIZE_4PB ++ sigreturn_addr = 0x1ull << 52; ++ pid = fork_child(); ++ expect_segv(pid); ++ ++ // Return to 0xd space ++ sigreturn_addr = 0xdull << 60; ++ pid = fork_child(); ++ expect_segv(pid); ++ ++ // Return to 0xe space ++ sigreturn_addr = 0xeull << 60; ++ pid = fork_child(); ++ expect_segv(pid); ++ ++ // Return to 0xf space ++ sigreturn_addr = 0xfull << 60; ++ pid = fork_child(); ++ expect_segv(pid); ++ ++ // Attempt to set PR=0 for 2nd loop (should be blocked by kernel) ++ sigreturn_msr_mask = ~MSR_PR; ++ } ++ ++ printf("All children killed as expected\n"); ++ ++ // Don't change address, just MSR, should return to user as normal ++ sigreturn_addr = 0; ++ sigreturn_msr_mask = ~MSR_PR; ++ pid = fork_child(); ++ waitpid(pid, &child_ret, 0); ++ FAIL_IF(!WIFEXITED(child_ret)); ++ FAIL_IF(WIFSIGNALED(child_ret)); ++ FAIL_IF(WEXITSTATUS(child_ret) != 0); ++ ++ return 0; ++} ++ ++int main(void) ++{ ++ return test_harness(test_sigreturn_kernel, "sigreturn_kernel"); ++} +diff --git a/tools/testing/selftests/vm/hmm-tests.c b/tools/testing/selftests/vm/hmm-tests.c +index 864f126ffd78f..203323967b507 100644 +--- a/tools/testing/selftests/vm/hmm-tests.c ++++ b/tools/testing/selftests/vm/hmm-tests.c +@@ -1248,6 +1248,48 @@ TEST_F(hmm, anon_teardown) + } + } + ++/* ++ * Test memory snapshot without faulting in pages accessed by the device. ++ */ ++TEST_F(hmm, mixedmap) ++{ ++ struct hmm_buffer *buffer; ++ unsigned long npages; ++ unsigned long size; ++ unsigned char *m; ++ int ret; ++ ++ npages = 1; ++ size = npages << self->page_shift; ++ ++ buffer = malloc(sizeof(*buffer)); ++ ASSERT_NE(buffer, NULL); ++ ++ buffer->fd = -1; ++ buffer->size = size; ++ buffer->mirror = malloc(npages); ++ ASSERT_NE(buffer->mirror, NULL); ++ ++ ++ /* Reserve a range of addresses. */ ++ buffer->ptr = mmap(NULL, size, ++ PROT_READ | PROT_WRITE, ++ MAP_PRIVATE, ++ self->fd, 0); ++ ASSERT_NE(buffer->ptr, MAP_FAILED); ++ ++ /* Simulate a device snapshotting CPU pagetables. */ ++ ret = hmm_dmirror_cmd(self->fd, HMM_DMIRROR_SNAPSHOT, buffer, npages); ++ ASSERT_EQ(ret, 0); ++ ASSERT_EQ(buffer->cpages, npages); ++ ++ /* Check what the device saw. */ ++ m = buffer->mirror; ++ ASSERT_EQ(m[0], HMM_DMIRROR_PROT_READ); ++ ++ hmm_buffer_free(buffer); ++} ++ + /* + * Test memory snapshot without faulting in pages accessed by the device. + */ |