3065 lines
90 KiB
Diff
3065 lines
90 KiB
Diff
diff --git a/Makefile b/Makefile
|
|
index ffcdc36c56f54..df9b1d07ca097 100644
|
|
--- a/Makefile
|
|
+++ b/Makefile
|
|
@@ -1,7 +1,7 @@
|
|
# SPDX-License-Identifier: GPL-2.0
|
|
VERSION = 5
|
|
PATCHLEVEL = 4
|
|
-SUBLEVEL = 152
|
|
+SUBLEVEL = 153
|
|
EXTRAVERSION =
|
|
NAME = Kleptomaniac Octopus
|
|
|
|
diff --git a/arch/arm/boot/dts/imx53-m53menlo.dts b/arch/arm/boot/dts/imx53-m53menlo.dts
|
|
index 64faf5b46d92f..03c43c1912a7e 100644
|
|
--- a/arch/arm/boot/dts/imx53-m53menlo.dts
|
|
+++ b/arch/arm/boot/dts/imx53-m53menlo.dts
|
|
@@ -56,6 +56,7 @@
|
|
panel {
|
|
compatible = "edt,etm0700g0dh6";
|
|
pinctrl-0 = <&pinctrl_display_gpio>;
|
|
+ pinctrl-names = "default";
|
|
enable-gpios = <&gpio6 0 GPIO_ACTIVE_HIGH>;
|
|
|
|
port {
|
|
@@ -76,8 +77,7 @@
|
|
regulator-name = "vbus";
|
|
regulator-min-microvolt = <5000000>;
|
|
regulator-max-microvolt = <5000000>;
|
|
- gpio = <&gpio1 2 GPIO_ACTIVE_HIGH>;
|
|
- enable-active-high;
|
|
+ gpio = <&gpio1 2 0>;
|
|
};
|
|
};
|
|
|
|
diff --git a/arch/arm/boot/dts/omap3430-sdp.dts b/arch/arm/boot/dts/omap3430-sdp.dts
|
|
index 0abd61108a539..ec16979825378 100644
|
|
--- a/arch/arm/boot/dts/omap3430-sdp.dts
|
|
+++ b/arch/arm/boot/dts/omap3430-sdp.dts
|
|
@@ -101,7 +101,7 @@
|
|
|
|
nand@1,0 {
|
|
compatible = "ti,omap2-nand";
|
|
- reg = <0 0 4>; /* CS0, offset 0, IO size 4 */
|
|
+ reg = <1 0 4>; /* CS1, offset 0, IO size 4 */
|
|
interrupt-parent = <&gpmc>;
|
|
interrupts = <0 IRQ_TYPE_NONE>, /* fifoevent */
|
|
<1 IRQ_TYPE_NONE>; /* termcount */
|
|
diff --git a/arch/arm/boot/dts/qcom-apq8064.dtsi b/arch/arm/boot/dts/qcom-apq8064.dtsi
|
|
index 2b075e287610f..764984c95c686 100644
|
|
--- a/arch/arm/boot/dts/qcom-apq8064.dtsi
|
|
+++ b/arch/arm/boot/dts/qcom-apq8064.dtsi
|
|
@@ -198,7 +198,7 @@
|
|
clock-frequency = <19200000>;
|
|
};
|
|
|
|
- pxo_board {
|
|
+ pxo_board: pxo_board {
|
|
compatible = "fixed-clock";
|
|
#clock-cells = <0>;
|
|
clock-frequency = <27000000>;
|
|
@@ -1147,7 +1147,7 @@
|
|
};
|
|
|
|
gpu: adreno-3xx@4300000 {
|
|
- compatible = "qcom,adreno-3xx";
|
|
+ compatible = "qcom,adreno-320.2", "qcom,adreno";
|
|
reg = <0x04300000 0x20000>;
|
|
reg-names = "kgsl_3d0_reg_memory";
|
|
interrupts = <GIC_SPI 80 IRQ_TYPE_LEVEL_HIGH>;
|
|
@@ -1162,7 +1162,6 @@
|
|
<&mmcc GFX3D_AHB_CLK>,
|
|
<&mmcc GFX3D_AXI_CLK>,
|
|
<&mmcc MMSS_IMEM_AHB_CLK>;
|
|
- qcom,chipid = <0x03020002>;
|
|
|
|
iommus = <&gfx3d 0
|
|
&gfx3d 1
|
|
@@ -1305,7 +1304,7 @@
|
|
reg-names = "dsi_pll", "dsi_phy", "dsi_phy_regulator";
|
|
clock-names = "iface_clk", "ref";
|
|
clocks = <&mmcc DSI_M_AHB_CLK>,
|
|
- <&cxo_board>;
|
|
+ <&pxo_board>;
|
|
};
|
|
|
|
|
|
diff --git a/arch/arm/mach-imx/pm-imx6.c b/arch/arm/mach-imx/pm-imx6.c
|
|
index baf3b47601af0..1b73e4e76310c 100644
|
|
--- a/arch/arm/mach-imx/pm-imx6.c
|
|
+++ b/arch/arm/mach-imx/pm-imx6.c
|
|
@@ -9,6 +9,7 @@
|
|
#include <linux/io.h>
|
|
#include <linux/irq.h>
|
|
#include <linux/genalloc.h>
|
|
+#include <linux/irqchip/arm-gic.h>
|
|
#include <linux/mfd/syscon.h>
|
|
#include <linux/mfd/syscon/imx6q-iomuxc-gpr.h>
|
|
#include <linux/of.h>
|
|
@@ -618,6 +619,7 @@ static void __init imx6_pm_common_init(const struct imx6_pm_socdata
|
|
|
|
static void imx6_pm_stby_poweroff(void)
|
|
{
|
|
+ gic_cpu_if_down(0);
|
|
imx6_set_lpm(STOP_POWER_OFF);
|
|
imx6q_suspend_finish(0);
|
|
|
|
diff --git a/arch/arm/mach-omap2/omap_hwmod.c b/arch/arm/mach-omap2/omap_hwmod.c
|
|
index eb74aa1826614..6289b288d60a6 100644
|
|
--- a/arch/arm/mach-omap2/omap_hwmod.c
|
|
+++ b/arch/arm/mach-omap2/omap_hwmod.c
|
|
@@ -3656,6 +3656,8 @@ int omap_hwmod_init_module(struct device *dev,
|
|
oh->flags |= HWMOD_SWSUP_SIDLE_ACT;
|
|
if (data->cfg->quirks & SYSC_QUIRK_SWSUP_MSTANDBY)
|
|
oh->flags |= HWMOD_SWSUP_MSTANDBY;
|
|
+ if (data->cfg->quirks & SYSC_QUIRK_CLKDM_NOAUTO)
|
|
+ oh->flags |= HWMOD_CLKDM_NOAUTO;
|
|
|
|
error = omap_hwmod_check_module(dev, oh, data, sysc_fields,
|
|
rev_offs, sysc_offs, syss_offs,
|
|
diff --git a/arch/arm/net/bpf_jit_32.c b/arch/arm/net/bpf_jit_32.c
|
|
index b51a8c7b01114..1c6e57f1dbc48 100644
|
|
--- a/arch/arm/net/bpf_jit_32.c
|
|
+++ b/arch/arm/net/bpf_jit_32.c
|
|
@@ -36,6 +36,10 @@
|
|
* +-----+
|
|
* |RSVD | JIT scratchpad
|
|
* current ARM_SP => +-----+ <= (BPF_FP - STACK_SIZE + SCRATCH_SIZE)
|
|
+ * | ... | caller-saved registers
|
|
+ * +-----+
|
|
+ * | ... | arguments passed on stack
|
|
+ * ARM_SP during call => +-----|
|
|
* | |
|
|
* | ... | Function call stack
|
|
* | |
|
|
@@ -63,6 +67,12 @@
|
|
*
|
|
* When popping registers off the stack at the end of a BPF function, we
|
|
* reference them via the current ARM_FP register.
|
|
+ *
|
|
+ * Some eBPF operations are implemented via a call to a helper function.
|
|
+ * Such calls are "invisible" in the eBPF code, so it is up to the calling
|
|
+ * program to preserve any caller-saved ARM registers during the call. The
|
|
+ * JIT emits code to push and pop those registers onto the stack, immediately
|
|
+ * above the callee stack frame.
|
|
*/
|
|
#define CALLEE_MASK (1 << ARM_R4 | 1 << ARM_R5 | 1 << ARM_R6 | \
|
|
1 << ARM_R7 | 1 << ARM_R8 | 1 << ARM_R9 | \
|
|
@@ -70,6 +80,8 @@
|
|
#define CALLEE_PUSH_MASK (CALLEE_MASK | 1 << ARM_LR)
|
|
#define CALLEE_POP_MASK (CALLEE_MASK | 1 << ARM_PC)
|
|
|
|
+#define CALLER_MASK (1 << ARM_R0 | 1 << ARM_R1 | 1 << ARM_R2 | 1 << ARM_R3)
|
|
+
|
|
enum {
|
|
/* Stack layout - these are offsets from (top of stack - 4) */
|
|
BPF_R2_HI,
|
|
@@ -464,6 +476,7 @@ static inline int epilogue_offset(const struct jit_ctx *ctx)
|
|
|
|
static inline void emit_udivmod(u8 rd, u8 rm, u8 rn, struct jit_ctx *ctx, u8 op)
|
|
{
|
|
+ const int exclude_mask = BIT(ARM_R0) | BIT(ARM_R1);
|
|
const s8 *tmp = bpf2a32[TMP_REG_1];
|
|
|
|
#if __LINUX_ARM_ARCH__ == 7
|
|
@@ -495,11 +508,17 @@ static inline void emit_udivmod(u8 rd, u8 rm, u8 rn, struct jit_ctx *ctx, u8 op)
|
|
emit(ARM_MOV_R(ARM_R0, rm), ctx);
|
|
}
|
|
|
|
+ /* Push caller-saved registers on stack */
|
|
+ emit(ARM_PUSH(CALLER_MASK & ~exclude_mask), ctx);
|
|
+
|
|
/* Call appropriate function */
|
|
emit_mov_i(ARM_IP, op == BPF_DIV ?
|
|
(u32)jit_udiv32 : (u32)jit_mod32, ctx);
|
|
emit_blx_r(ARM_IP, ctx);
|
|
|
|
+ /* Restore caller-saved registers from stack */
|
|
+ emit(ARM_POP(CALLER_MASK & ~exclude_mask), ctx);
|
|
+
|
|
/* Save return value */
|
|
if (rd != ARM_R0)
|
|
emit(ARM_MOV_R(rd, ARM_R0), ctx);
|
|
diff --git a/arch/arm64/boot/dts/freescale/fsl-ls1028a.dtsi b/arch/arm64/boot/dts/freescale/fsl-ls1028a.dtsi
|
|
index 5716ac20bddd1..02ae6bfff5658 100644
|
|
--- a/arch/arm64/boot/dts/freescale/fsl-ls1028a.dtsi
|
|
+++ b/arch/arm64/boot/dts/freescale/fsl-ls1028a.dtsi
|
|
@@ -287,6 +287,24 @@
|
|
status = "disabled";
|
|
};
|
|
|
|
+ can0: can@2180000 {
|
|
+ compatible = "fsl,ls1028ar1-flexcan", "fsl,lx2160ar1-flexcan";
|
|
+ reg = <0x0 0x2180000 0x0 0x10000>;
|
|
+ interrupts = <GIC_SPI 21 IRQ_TYPE_LEVEL_HIGH>;
|
|
+ clocks = <&sysclk>, <&clockgen 4 1>;
|
|
+ clock-names = "ipg", "per";
|
|
+ status = "disabled";
|
|
+ };
|
|
+
|
|
+ can1: can@2190000 {
|
|
+ compatible = "fsl,ls1028ar1-flexcan", "fsl,lx2160ar1-flexcan";
|
|
+ reg = <0x0 0x2190000 0x0 0x10000>;
|
|
+ interrupts = <GIC_SPI 22 IRQ_TYPE_LEVEL_HIGH>;
|
|
+ clocks = <&sysclk>, <&clockgen 4 1>;
|
|
+ clock-names = "ipg", "per";
|
|
+ status = "disabled";
|
|
+ };
|
|
+
|
|
duart0: serial@21c0500 {
|
|
compatible = "fsl,ns16550", "ns16550a";
|
|
reg = <0x00 0x21c0500 0x0 0x100>;
|
|
@@ -496,14 +514,14 @@
|
|
compatible = "arm,sp805", "arm,primecell";
|
|
reg = <0x0 0xc000000 0x0 0x1000>;
|
|
clocks = <&clockgen 4 15>, <&clockgen 4 15>;
|
|
- clock-names = "apb_pclk", "wdog_clk";
|
|
+ clock-names = "wdog_clk", "apb_pclk";
|
|
};
|
|
|
|
cluster1_core1_watchdog: watchdog@c010000 {
|
|
compatible = "arm,sp805", "arm,primecell";
|
|
reg = <0x0 0xc010000 0x0 0x1000>;
|
|
clocks = <&clockgen 4 15>, <&clockgen 4 15>;
|
|
- clock-names = "apb_pclk", "wdog_clk";
|
|
+ clock-names = "wdog_clk", "apb_pclk";
|
|
};
|
|
|
|
sai1: audio-controller@f100000 {
|
|
diff --git a/arch/arm64/boot/dts/freescale/fsl-ls1088a.dtsi b/arch/arm64/boot/dts/freescale/fsl-ls1088a.dtsi
|
|
index c676d0771762f..407ebdb35cd2e 100644
|
|
--- a/arch/arm64/boot/dts/freescale/fsl-ls1088a.dtsi
|
|
+++ b/arch/arm64/boot/dts/freescale/fsl-ls1088a.dtsi
|
|
@@ -640,56 +640,56 @@
|
|
compatible = "arm,sp805-wdt", "arm,primecell";
|
|
reg = <0x0 0xc000000 0x0 0x1000>;
|
|
clocks = <&clockgen 4 3>, <&clockgen 4 3>;
|
|
- clock-names = "apb_pclk", "wdog_clk";
|
|
+ clock-names = "wdog_clk", "apb_pclk";
|
|
};
|
|
|
|
cluster1_core1_watchdog: wdt@c010000 {
|
|
compatible = "arm,sp805-wdt", "arm,primecell";
|
|
reg = <0x0 0xc010000 0x0 0x1000>;
|
|
clocks = <&clockgen 4 3>, <&clockgen 4 3>;
|
|
- clock-names = "apb_pclk", "wdog_clk";
|
|
+ clock-names = "wdog_clk", "apb_pclk";
|
|
};
|
|
|
|
cluster1_core2_watchdog: wdt@c020000 {
|
|
compatible = "arm,sp805-wdt", "arm,primecell";
|
|
reg = <0x0 0xc020000 0x0 0x1000>;
|
|
clocks = <&clockgen 4 3>, <&clockgen 4 3>;
|
|
- clock-names = "apb_pclk", "wdog_clk";
|
|
+ clock-names = "wdog_clk", "apb_pclk";
|
|
};
|
|
|
|
cluster1_core3_watchdog: wdt@c030000 {
|
|
compatible = "arm,sp805-wdt", "arm,primecell";
|
|
reg = <0x0 0xc030000 0x0 0x1000>;
|
|
clocks = <&clockgen 4 3>, <&clockgen 4 3>;
|
|
- clock-names = "apb_pclk", "wdog_clk";
|
|
+ clock-names = "wdog_clk", "apb_pclk";
|
|
};
|
|
|
|
cluster2_core0_watchdog: wdt@c100000 {
|
|
compatible = "arm,sp805-wdt", "arm,primecell";
|
|
reg = <0x0 0xc100000 0x0 0x1000>;
|
|
clocks = <&clockgen 4 3>, <&clockgen 4 3>;
|
|
- clock-names = "apb_pclk", "wdog_clk";
|
|
+ clock-names = "wdog_clk", "apb_pclk";
|
|
};
|
|
|
|
cluster2_core1_watchdog: wdt@c110000 {
|
|
compatible = "arm,sp805-wdt", "arm,primecell";
|
|
reg = <0x0 0xc110000 0x0 0x1000>;
|
|
clocks = <&clockgen 4 3>, <&clockgen 4 3>;
|
|
- clock-names = "apb_pclk", "wdog_clk";
|
|
+ clock-names = "wdog_clk", "apb_pclk";
|
|
};
|
|
|
|
cluster2_core2_watchdog: wdt@c120000 {
|
|
compatible = "arm,sp805-wdt", "arm,primecell";
|
|
reg = <0x0 0xc120000 0x0 0x1000>;
|
|
clocks = <&clockgen 4 3>, <&clockgen 4 3>;
|
|
- clock-names = "apb_pclk", "wdog_clk";
|
|
+ clock-names = "wdog_clk", "apb_pclk";
|
|
};
|
|
|
|
cluster2_core3_watchdog: wdt@c130000 {
|
|
compatible = "arm,sp805-wdt", "arm,primecell";
|
|
reg = <0x0 0xc130000 0x0 0x1000>;
|
|
clocks = <&clockgen 4 3>, <&clockgen 4 3>;
|
|
- clock-names = "apb_pclk", "wdog_clk";
|
|
+ clock-names = "wdog_clk", "apb_pclk";
|
|
};
|
|
|
|
fsl_mc: fsl-mc@80c000000 {
|
|
diff --git a/arch/arm64/boot/dts/freescale/fsl-ls208xa.dtsi b/arch/arm64/boot/dts/freescale/fsl-ls208xa.dtsi
|
|
index cdb2fa47637da..82f0fe6acbfb7 100644
|
|
--- a/arch/arm64/boot/dts/freescale/fsl-ls208xa.dtsi
|
|
+++ b/arch/arm64/boot/dts/freescale/fsl-ls208xa.dtsi
|
|
@@ -230,56 +230,56 @@
|
|
compatible = "arm,sp805-wdt", "arm,primecell";
|
|
reg = <0x0 0xc000000 0x0 0x1000>;
|
|
clocks = <&clockgen 4 3>, <&clockgen 4 3>;
|
|
- clock-names = "apb_pclk", "wdog_clk";
|
|
+ clock-names = "wdog_clk", "apb_pclk";
|
|
};
|
|
|
|
cluster1_core1_watchdog: wdt@c010000 {
|
|
compatible = "arm,sp805-wdt", "arm,primecell";
|
|
reg = <0x0 0xc010000 0x0 0x1000>;
|
|
clocks = <&clockgen 4 3>, <&clockgen 4 3>;
|
|
- clock-names = "apb_pclk", "wdog_clk";
|
|
+ clock-names = "wdog_clk", "apb_pclk";
|
|
};
|
|
|
|
cluster2_core0_watchdog: wdt@c100000 {
|
|
compatible = "arm,sp805-wdt", "arm,primecell";
|
|
reg = <0x0 0xc100000 0x0 0x1000>;
|
|
clocks = <&clockgen 4 3>, <&clockgen 4 3>;
|
|
- clock-names = "apb_pclk", "wdog_clk";
|
|
+ clock-names = "wdog_clk", "apb_pclk";
|
|
};
|
|
|
|
cluster2_core1_watchdog: wdt@c110000 {
|
|
compatible = "arm,sp805-wdt", "arm,primecell";
|
|
reg = <0x0 0xc110000 0x0 0x1000>;
|
|
clocks = <&clockgen 4 3>, <&clockgen 4 3>;
|
|
- clock-names = "apb_pclk", "wdog_clk";
|
|
+ clock-names = "wdog_clk", "apb_pclk";
|
|
};
|
|
|
|
cluster3_core0_watchdog: wdt@c200000 {
|
|
compatible = "arm,sp805-wdt", "arm,primecell";
|
|
reg = <0x0 0xc200000 0x0 0x1000>;
|
|
clocks = <&clockgen 4 3>, <&clockgen 4 3>;
|
|
- clock-names = "apb_pclk", "wdog_clk";
|
|
+ clock-names = "wdog_clk", "apb_pclk";
|
|
};
|
|
|
|
cluster3_core1_watchdog: wdt@c210000 {
|
|
compatible = "arm,sp805-wdt", "arm,primecell";
|
|
reg = <0x0 0xc210000 0x0 0x1000>;
|
|
clocks = <&clockgen 4 3>, <&clockgen 4 3>;
|
|
- clock-names = "apb_pclk", "wdog_clk";
|
|
+ clock-names = "wdog_clk", "apb_pclk";
|
|
};
|
|
|
|
cluster4_core0_watchdog: wdt@c300000 {
|
|
compatible = "arm,sp805-wdt", "arm,primecell";
|
|
reg = <0x0 0xc300000 0x0 0x1000>;
|
|
clocks = <&clockgen 4 3>, <&clockgen 4 3>;
|
|
- clock-names = "apb_pclk", "wdog_clk";
|
|
+ clock-names = "wdog_clk", "apb_pclk";
|
|
};
|
|
|
|
cluster4_core1_watchdog: wdt@c310000 {
|
|
compatible = "arm,sp805-wdt", "arm,primecell";
|
|
reg = <0x0 0xc310000 0x0 0x1000>;
|
|
clocks = <&clockgen 4 3>, <&clockgen 4 3>;
|
|
- clock-names = "apb_pclk", "wdog_clk";
|
|
+ clock-names = "wdog_clk", "apb_pclk";
|
|
};
|
|
|
|
crypto: crypto@8000000 {
|
|
diff --git a/arch/arm64/boot/dts/qcom/pm8150.dtsi b/arch/arm64/boot/dts/qcom/pm8150.dtsi
|
|
index c0b197458665d..6f7dfcb8c0421 100644
|
|
--- a/arch/arm64/boot/dts/qcom/pm8150.dtsi
|
|
+++ b/arch/arm64/boot/dts/qcom/pm8150.dtsi
|
|
@@ -17,7 +17,7 @@
|
|
#size-cells = <0>;
|
|
|
|
pon: power-on@800 {
|
|
- compatible = "qcom,pm8916-pon";
|
|
+ compatible = "qcom,pm8998-pon";
|
|
reg = <0x0800>;
|
|
pwrkey {
|
|
compatible = "qcom,pm8941-pwrkey";
|
|
diff --git a/arch/mips/Kconfig b/arch/mips/Kconfig
|
|
index 6ecdc690f7336..2bfef67d52c63 100644
|
|
--- a/arch/mips/Kconfig
|
|
+++ b/arch/mips/Kconfig
|
|
@@ -46,6 +46,7 @@ config MIPS
|
|
select HAVE_ARCH_TRACEHOOK
|
|
select HAVE_ARCH_TRANSPARENT_HUGEPAGE if CPU_SUPPORTS_HUGEPAGES
|
|
select HAVE_ASM_MODVERSIONS
|
|
+ select HAVE_CBPF_JIT if !64BIT && !CPU_MICROMIPS
|
|
select HAVE_EBPF_JIT if 64BIT && !CPU_MICROMIPS && TARGET_ISA_REV >= 2
|
|
select HAVE_CONTEXT_TRACKING
|
|
select HAVE_COPY_THREAD_TLS
|
|
diff --git a/arch/mips/net/Makefile b/arch/mips/net/Makefile
|
|
index 2d03af7d6b19d..d55912349039c 100644
|
|
--- a/arch/mips/net/Makefile
|
|
+++ b/arch/mips/net/Makefile
|
|
@@ -1,4 +1,5 @@
|
|
# SPDX-License-Identifier: GPL-2.0-only
|
|
# MIPS networking code
|
|
|
|
+obj-$(CONFIG_MIPS_CBPF_JIT) += bpf_jit.o bpf_jit_asm.o
|
|
obj-$(CONFIG_MIPS_EBPF_JIT) += ebpf_jit.o
|
|
diff --git a/arch/mips/net/bpf_jit.c b/arch/mips/net/bpf_jit.c
|
|
new file mode 100644
|
|
index 0000000000000..29a288ff4f183
|
|
--- /dev/null
|
|
+++ b/arch/mips/net/bpf_jit.c
|
|
@@ -0,0 +1,1299 @@
|
|
+/*
|
|
+ * Just-In-Time compiler for BPF filters on MIPS
|
|
+ *
|
|
+ * Copyright (c) 2014 Imagination Technologies Ltd.
|
|
+ * Author: Markos Chandras <markos.chandras@imgtec.com>
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify it
|
|
+ * under the terms of the GNU General Public License as published by the
|
|
+ * Free Software Foundation; version 2 of the License.
|
|
+ */
|
|
+
|
|
+#include <linux/bitops.h>
|
|
+#include <linux/compiler.h>
|
|
+#include <linux/errno.h>
|
|
+#include <linux/filter.h>
|
|
+#include <linux/if_vlan.h>
|
|
+#include <linux/moduleloader.h>
|
|
+#include <linux/netdevice.h>
|
|
+#include <linux/string.h>
|
|
+#include <linux/slab.h>
|
|
+#include <linux/types.h>
|
|
+#include <asm/asm.h>
|
|
+#include <asm/bitops.h>
|
|
+#include <asm/cacheflush.h>
|
|
+#include <asm/cpu-features.h>
|
|
+#include <asm/uasm.h>
|
|
+
|
|
+#include "bpf_jit.h"
|
|
+
|
|
+/* ABI
|
|
+ * r_skb_hl SKB header length
|
|
+ * r_data SKB data pointer
|
|
+ * r_off Offset
|
|
+ * r_A BPF register A
|
|
+ * r_X BPF register X
|
|
+ * r_skb *skb
|
|
+ * r_M *scratch memory
|
|
+ * r_skb_len SKB length
|
|
+ *
|
|
+ * On entry (*bpf_func)(*skb, *filter)
|
|
+ * a0 = MIPS_R_A0 = skb;
|
|
+ * a1 = MIPS_R_A1 = filter;
|
|
+ *
|
|
+ * Stack
|
|
+ * ...
|
|
+ * M[15]
|
|
+ * M[14]
|
|
+ * M[13]
|
|
+ * ...
|
|
+ * M[0] <-- r_M
|
|
+ * saved reg k-1
|
|
+ * saved reg k-2
|
|
+ * ...
|
|
+ * saved reg 0 <-- r_sp
|
|
+ * <no argument area>
|
|
+ *
|
|
+ * Packet layout
|
|
+ *
|
|
+ * <--------------------- len ------------------------>
|
|
+ * <--skb-len(r_skb_hl)-->< ----- skb->data_len ------>
|
|
+ * ----------------------------------------------------
|
|
+ * | skb->data |
|
|
+ * ----------------------------------------------------
|
|
+ */
|
|
+
|
|
+#define ptr typeof(unsigned long)
|
|
+
|
|
+#define SCRATCH_OFF(k) (4 * (k))
|
|
+
|
|
+/* JIT flags */
|
|
+#define SEEN_CALL (1 << BPF_MEMWORDS)
|
|
+#define SEEN_SREG_SFT (BPF_MEMWORDS + 1)
|
|
+#define SEEN_SREG_BASE (1 << SEEN_SREG_SFT)
|
|
+#define SEEN_SREG(x) (SEEN_SREG_BASE << (x))
|
|
+#define SEEN_OFF SEEN_SREG(2)
|
|
+#define SEEN_A SEEN_SREG(3)
|
|
+#define SEEN_X SEEN_SREG(4)
|
|
+#define SEEN_SKB SEEN_SREG(5)
|
|
+#define SEEN_MEM SEEN_SREG(6)
|
|
+/* SEEN_SK_DATA also implies skb_hl an skb_len */
|
|
+#define SEEN_SKB_DATA (SEEN_SREG(7) | SEEN_SREG(1) | SEEN_SREG(0))
|
|
+
|
|
+/* Arguments used by JIT */
|
|
+#define ARGS_USED_BY_JIT 2 /* only applicable to 64-bit */
|
|
+
|
|
+#define SBIT(x) (1 << (x)) /* Signed version of BIT() */
|
|
+
|
|
+/**
|
|
+ * struct jit_ctx - JIT context
|
|
+ * @skf: The sk_filter
|
|
+ * @prologue_bytes: Number of bytes for prologue
|
|
+ * @idx: Instruction index
|
|
+ * @flags: JIT flags
|
|
+ * @offsets: Instruction offsets
|
|
+ * @target: Memory location for the compiled filter
|
|
+ */
|
|
+struct jit_ctx {
|
|
+ const struct bpf_prog *skf;
|
|
+ unsigned int prologue_bytes;
|
|
+ u32 idx;
|
|
+ u32 flags;
|
|
+ u32 *offsets;
|
|
+ u32 *target;
|
|
+};
|
|
+
|
|
+
|
|
+static inline int optimize_div(u32 *k)
|
|
+{
|
|
+ /* power of 2 divides can be implemented with right shift */
|
|
+ if (!(*k & (*k-1))) {
|
|
+ *k = ilog2(*k);
|
|
+ return 1;
|
|
+ }
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+static inline void emit_jit_reg_move(ptr dst, ptr src, struct jit_ctx *ctx);
|
|
+
|
|
+/* Simply emit the instruction if the JIT memory space has been allocated */
|
|
+#define emit_instr(ctx, func, ...) \
|
|
+do { \
|
|
+ if ((ctx)->target != NULL) { \
|
|
+ u32 *p = &(ctx)->target[ctx->idx]; \
|
|
+ uasm_i_##func(&p, ##__VA_ARGS__); \
|
|
+ } \
|
|
+ (ctx)->idx++; \
|
|
+} while (0)
|
|
+
|
|
+/*
|
|
+ * Similar to emit_instr but it must be used when we need to emit
|
|
+ * 32-bit or 64-bit instructions
|
|
+ */
|
|
+#define emit_long_instr(ctx, func, ...) \
|
|
+do { \
|
|
+ if ((ctx)->target != NULL) { \
|
|
+ u32 *p = &(ctx)->target[ctx->idx]; \
|
|
+ UASM_i_##func(&p, ##__VA_ARGS__); \
|
|
+ } \
|
|
+ (ctx)->idx++; \
|
|
+} while (0)
|
|
+
|
|
+/* Determine if immediate is within the 16-bit signed range */
|
|
+static inline bool is_range16(s32 imm)
|
|
+{
|
|
+ return !(imm >= SBIT(15) || imm < -SBIT(15));
|
|
+}
|
|
+
|
|
+static inline void emit_addu(unsigned int dst, unsigned int src1,
|
|
+ unsigned int src2, struct jit_ctx *ctx)
|
|
+{
|
|
+ emit_instr(ctx, addu, dst, src1, src2);
|
|
+}
|
|
+
|
|
+static inline void emit_nop(struct jit_ctx *ctx)
|
|
+{
|
|
+ emit_instr(ctx, nop);
|
|
+}
|
|
+
|
|
+/* Load a u32 immediate to a register */
|
|
+static inline void emit_load_imm(unsigned int dst, u32 imm, struct jit_ctx *ctx)
|
|
+{
|
|
+ if (ctx->target != NULL) {
|
|
+ /* addiu can only handle s16 */
|
|
+ if (!is_range16(imm)) {
|
|
+ u32 *p = &ctx->target[ctx->idx];
|
|
+ uasm_i_lui(&p, r_tmp_imm, (s32)imm >> 16);
|
|
+ p = &ctx->target[ctx->idx + 1];
|
|
+ uasm_i_ori(&p, dst, r_tmp_imm, imm & 0xffff);
|
|
+ } else {
|
|
+ u32 *p = &ctx->target[ctx->idx];
|
|
+ uasm_i_addiu(&p, dst, r_zero, imm);
|
|
+ }
|
|
+ }
|
|
+ ctx->idx++;
|
|
+
|
|
+ if (!is_range16(imm))
|
|
+ ctx->idx++;
|
|
+}
|
|
+
|
|
+static inline void emit_or(unsigned int dst, unsigned int src1,
|
|
+ unsigned int src2, struct jit_ctx *ctx)
|
|
+{
|
|
+ emit_instr(ctx, or, dst, src1, src2);
|
|
+}
|
|
+
|
|
+static inline void emit_ori(unsigned int dst, unsigned src, u32 imm,
|
|
+ struct jit_ctx *ctx)
|
|
+{
|
|
+ if (imm >= BIT(16)) {
|
|
+ emit_load_imm(r_tmp, imm, ctx);
|
|
+ emit_or(dst, src, r_tmp, ctx);
|
|
+ } else {
|
|
+ emit_instr(ctx, ori, dst, src, imm);
|
|
+ }
|
|
+}
|
|
+
|
|
+static inline void emit_daddiu(unsigned int dst, unsigned int src,
|
|
+ int imm, struct jit_ctx *ctx)
|
|
+{
|
|
+ /*
|
|
+ * Only used for stack, so the imm is relatively small
|
|
+ * and it fits in 15-bits
|
|
+ */
|
|
+ emit_instr(ctx, daddiu, dst, src, imm);
|
|
+}
|
|
+
|
|
+static inline void emit_addiu(unsigned int dst, unsigned int src,
|
|
+ u32 imm, struct jit_ctx *ctx)
|
|
+{
|
|
+ if (!is_range16(imm)) {
|
|
+ emit_load_imm(r_tmp, imm, ctx);
|
|
+ emit_addu(dst, r_tmp, src, ctx);
|
|
+ } else {
|
|
+ emit_instr(ctx, addiu, dst, src, imm);
|
|
+ }
|
|
+}
|
|
+
|
|
+static inline void emit_and(unsigned int dst, unsigned int src1,
|
|
+ unsigned int src2, struct jit_ctx *ctx)
|
|
+{
|
|
+ emit_instr(ctx, and, dst, src1, src2);
|
|
+}
|
|
+
|
|
+static inline void emit_andi(unsigned int dst, unsigned int src,
|
|
+ u32 imm, struct jit_ctx *ctx)
|
|
+{
|
|
+ /* If imm does not fit in u16 then load it to register */
|
|
+ if (imm >= BIT(16)) {
|
|
+ emit_load_imm(r_tmp, imm, ctx);
|
|
+ emit_and(dst, src, r_tmp, ctx);
|
|
+ } else {
|
|
+ emit_instr(ctx, andi, dst, src, imm);
|
|
+ }
|
|
+}
|
|
+
|
|
+static inline void emit_xor(unsigned int dst, unsigned int src1,
|
|
+ unsigned int src2, struct jit_ctx *ctx)
|
|
+{
|
|
+ emit_instr(ctx, xor, dst, src1, src2);
|
|
+}
|
|
+
|
|
+static inline void emit_xori(ptr dst, ptr src, u32 imm, struct jit_ctx *ctx)
|
|
+{
|
|
+ /* If imm does not fit in u16 then load it to register */
|
|
+ if (imm >= BIT(16)) {
|
|
+ emit_load_imm(r_tmp, imm, ctx);
|
|
+ emit_xor(dst, src, r_tmp, ctx);
|
|
+ } else {
|
|
+ emit_instr(ctx, xori, dst, src, imm);
|
|
+ }
|
|
+}
|
|
+
|
|
+static inline void emit_stack_offset(int offset, struct jit_ctx *ctx)
|
|
+{
|
|
+ emit_long_instr(ctx, ADDIU, r_sp, r_sp, offset);
|
|
+}
|
|
+
|
|
+static inline void emit_subu(unsigned int dst, unsigned int src1,
|
|
+ unsigned int src2, struct jit_ctx *ctx)
|
|
+{
|
|
+ emit_instr(ctx, subu, dst, src1, src2);
|
|
+}
|
|
+
|
|
+static inline void emit_neg(unsigned int reg, struct jit_ctx *ctx)
|
|
+{
|
|
+ emit_subu(reg, r_zero, reg, ctx);
|
|
+}
|
|
+
|
|
+static inline void emit_sllv(unsigned int dst, unsigned int src,
|
|
+ unsigned int sa, struct jit_ctx *ctx)
|
|
+{
|
|
+ emit_instr(ctx, sllv, dst, src, sa);
|
|
+}
|
|
+
|
|
+static inline void emit_sll(unsigned int dst, unsigned int src,
|
|
+ unsigned int sa, struct jit_ctx *ctx)
|
|
+{
|
|
+ /* sa is 5-bits long */
|
|
+ if (sa >= BIT(5))
|
|
+ /* Shifting >= 32 results in zero */
|
|
+ emit_jit_reg_move(dst, r_zero, ctx);
|
|
+ else
|
|
+ emit_instr(ctx, sll, dst, src, sa);
|
|
+}
|
|
+
|
|
+static inline void emit_srlv(unsigned int dst, unsigned int src,
|
|
+ unsigned int sa, struct jit_ctx *ctx)
|
|
+{
|
|
+ emit_instr(ctx, srlv, dst, src, sa);
|
|
+}
|
|
+
|
|
+static inline void emit_srl(unsigned int dst, unsigned int src,
|
|
+ unsigned int sa, struct jit_ctx *ctx)
|
|
+{
|
|
+ /* sa is 5-bits long */
|
|
+ if (sa >= BIT(5))
|
|
+ /* Shifting >= 32 results in zero */
|
|
+ emit_jit_reg_move(dst, r_zero, ctx);
|
|
+ else
|
|
+ emit_instr(ctx, srl, dst, src, sa);
|
|
+}
|
|
+
|
|
+static inline void emit_slt(unsigned int dst, unsigned int src1,
|
|
+ unsigned int src2, struct jit_ctx *ctx)
|
|
+{
|
|
+ emit_instr(ctx, slt, dst, src1, src2);
|
|
+}
|
|
+
|
|
+static inline void emit_sltu(unsigned int dst, unsigned int src1,
|
|
+ unsigned int src2, struct jit_ctx *ctx)
|
|
+{
|
|
+ emit_instr(ctx, sltu, dst, src1, src2);
|
|
+}
|
|
+
|
|
+static inline void emit_sltiu(unsigned dst, unsigned int src,
|
|
+ unsigned int imm, struct jit_ctx *ctx)
|
|
+{
|
|
+ /* 16 bit immediate */
|
|
+ if (!is_range16((s32)imm)) {
|
|
+ emit_load_imm(r_tmp, imm, ctx);
|
|
+ emit_sltu(dst, src, r_tmp, ctx);
|
|
+ } else {
|
|
+ emit_instr(ctx, sltiu, dst, src, imm);
|
|
+ }
|
|
+
|
|
+}
|
|
+
|
|
+/* Store register on the stack */
|
|
+static inline void emit_store_stack_reg(ptr reg, ptr base,
|
|
+ unsigned int offset,
|
|
+ struct jit_ctx *ctx)
|
|
+{
|
|
+ emit_long_instr(ctx, SW, reg, offset, base);
|
|
+}
|
|
+
|
|
+static inline void emit_store(ptr reg, ptr base, unsigned int offset,
|
|
+ struct jit_ctx *ctx)
|
|
+{
|
|
+ emit_instr(ctx, sw, reg, offset, base);
|
|
+}
|
|
+
|
|
+static inline void emit_load_stack_reg(ptr reg, ptr base,
|
|
+ unsigned int offset,
|
|
+ struct jit_ctx *ctx)
|
|
+{
|
|
+ emit_long_instr(ctx, LW, reg, offset, base);
|
|
+}
|
|
+
|
|
+static inline void emit_load(unsigned int reg, unsigned int base,
|
|
+ unsigned int offset, struct jit_ctx *ctx)
|
|
+{
|
|
+ emit_instr(ctx, lw, reg, offset, base);
|
|
+}
|
|
+
|
|
+static inline void emit_load_byte(unsigned int reg, unsigned int base,
|
|
+ unsigned int offset, struct jit_ctx *ctx)
|
|
+{
|
|
+ emit_instr(ctx, lb, reg, offset, base);
|
|
+}
|
|
+
|
|
+static inline void emit_half_load(unsigned int reg, unsigned int base,
|
|
+ unsigned int offset, struct jit_ctx *ctx)
|
|
+{
|
|
+ emit_instr(ctx, lh, reg, offset, base);
|
|
+}
|
|
+
|
|
+static inline void emit_half_load_unsigned(unsigned int reg, unsigned int base,
|
|
+ unsigned int offset, struct jit_ctx *ctx)
|
|
+{
|
|
+ emit_instr(ctx, lhu, reg, offset, base);
|
|
+}
|
|
+
|
|
+static inline void emit_mul(unsigned int dst, unsigned int src1,
|
|
+ unsigned int src2, struct jit_ctx *ctx)
|
|
+{
|
|
+ emit_instr(ctx, mul, dst, src1, src2);
|
|
+}
|
|
+
|
|
+static inline void emit_div(unsigned int dst, unsigned int src,
|
|
+ struct jit_ctx *ctx)
|
|
+{
|
|
+ if (ctx->target != NULL) {
|
|
+ u32 *p = &ctx->target[ctx->idx];
|
|
+ uasm_i_divu(&p, dst, src);
|
|
+ p = &ctx->target[ctx->idx + 1];
|
|
+ uasm_i_mflo(&p, dst);
|
|
+ }
|
|
+ ctx->idx += 2; /* 2 insts */
|
|
+}
|
|
+
|
|
+static inline void emit_mod(unsigned int dst, unsigned int src,
|
|
+ struct jit_ctx *ctx)
|
|
+{
|
|
+ if (ctx->target != NULL) {
|
|
+ u32 *p = &ctx->target[ctx->idx];
|
|
+ uasm_i_divu(&p, dst, src);
|
|
+ p = &ctx->target[ctx->idx + 1];
|
|
+ uasm_i_mfhi(&p, dst);
|
|
+ }
|
|
+ ctx->idx += 2; /* 2 insts */
|
|
+}
|
|
+
|
|
+static inline void emit_dsll(unsigned int dst, unsigned int src,
|
|
+ unsigned int sa, struct jit_ctx *ctx)
|
|
+{
|
|
+ emit_instr(ctx, dsll, dst, src, sa);
|
|
+}
|
|
+
|
|
+static inline void emit_dsrl32(unsigned int dst, unsigned int src,
|
|
+ unsigned int sa, struct jit_ctx *ctx)
|
|
+{
|
|
+ emit_instr(ctx, dsrl32, dst, src, sa);
|
|
+}
|
|
+
|
|
+static inline void emit_wsbh(unsigned int dst, unsigned int src,
|
|
+ struct jit_ctx *ctx)
|
|
+{
|
|
+ emit_instr(ctx, wsbh, dst, src);
|
|
+}
|
|
+
|
|
+/* load pointer to register */
|
|
+static inline void emit_load_ptr(unsigned int dst, unsigned int src,
|
|
+ int imm, struct jit_ctx *ctx)
|
|
+{
|
|
+ /* src contains the base addr of the 32/64-pointer */
|
|
+ emit_long_instr(ctx, LW, dst, imm, src);
|
|
+}
|
|
+
|
|
+/* load a function pointer to register */
|
|
+static inline void emit_load_func(unsigned int reg, ptr imm,
|
|
+ struct jit_ctx *ctx)
|
|
+{
|
|
+ if (IS_ENABLED(CONFIG_64BIT)) {
|
|
+ /* At this point imm is always 64-bit */
|
|
+ emit_load_imm(r_tmp, (u64)imm >> 32, ctx);
|
|
+ emit_dsll(r_tmp_imm, r_tmp, 16, ctx); /* left shift by 16 */
|
|
+ emit_ori(r_tmp, r_tmp_imm, (imm >> 16) & 0xffff, ctx);
|
|
+ emit_dsll(r_tmp_imm, r_tmp, 16, ctx); /* left shift by 16 */
|
|
+ emit_ori(reg, r_tmp_imm, imm & 0xffff, ctx);
|
|
+ } else {
|
|
+ emit_load_imm(reg, imm, ctx);
|
|
+ }
|
|
+}
|
|
+
|
|
+/* Move to real MIPS register */
|
|
+static inline void emit_reg_move(ptr dst, ptr src, struct jit_ctx *ctx)
|
|
+{
|
|
+ emit_long_instr(ctx, ADDU, dst, src, r_zero);
|
|
+}
|
|
+
|
|
+/* Move to JIT (32-bit) register */
|
|
+static inline void emit_jit_reg_move(ptr dst, ptr src, struct jit_ctx *ctx)
|
|
+{
|
|
+ emit_addu(dst, src, r_zero, ctx);
|
|
+}
|
|
+
|
|
+/* Compute the immediate value for PC-relative branches. */
|
|
+static inline u32 b_imm(unsigned int tgt, struct jit_ctx *ctx)
|
|
+{
|
|
+ if (ctx->target == NULL)
|
|
+ return 0;
|
|
+
|
|
+ /*
|
|
+ * We want a pc-relative branch. We only do forward branches
|
|
+ * so tgt is always after pc. tgt is the instruction offset
|
|
+ * we want to jump to.
|
|
+
|
|
+ * Branch on MIPS:
|
|
+ * I: target_offset <- sign_extend(offset)
|
|
+ * I+1: PC += target_offset (delay slot)
|
|
+ *
|
|
+ * ctx->idx currently points to the branch instruction
|
|
+ * but the offset is added to the delay slot so we need
|
|
+ * to subtract 4.
|
|
+ */
|
|
+ return ctx->offsets[tgt] -
|
|
+ (ctx->idx * 4 - ctx->prologue_bytes) - 4;
|
|
+}
|
|
+
|
|
+static inline void emit_bcond(int cond, unsigned int reg1, unsigned int reg2,
|
|
+ unsigned int imm, struct jit_ctx *ctx)
|
|
+{
|
|
+ if (ctx->target != NULL) {
|
|
+ u32 *p = &ctx->target[ctx->idx];
|
|
+
|
|
+ switch (cond) {
|
|
+ case MIPS_COND_EQ:
|
|
+ uasm_i_beq(&p, reg1, reg2, imm);
|
|
+ break;
|
|
+ case MIPS_COND_NE:
|
|
+ uasm_i_bne(&p, reg1, reg2, imm);
|
|
+ break;
|
|
+ case MIPS_COND_ALL:
|
|
+ uasm_i_b(&p, imm);
|
|
+ break;
|
|
+ default:
|
|
+ pr_warn("%s: Unhandled branch conditional: %d\n",
|
|
+ __func__, cond);
|
|
+ }
|
|
+ }
|
|
+ ctx->idx++;
|
|
+}
|
|
+
|
|
+static inline void emit_b(unsigned int imm, struct jit_ctx *ctx)
|
|
+{
|
|
+ emit_bcond(MIPS_COND_ALL, r_zero, r_zero, imm, ctx);
|
|
+}
|
|
+
|
|
+static inline void emit_jalr(unsigned int link, unsigned int reg,
|
|
+ struct jit_ctx *ctx)
|
|
+{
|
|
+ emit_instr(ctx, jalr, link, reg);
|
|
+}
|
|
+
|
|
+static inline void emit_jr(unsigned int reg, struct jit_ctx *ctx)
|
|
+{
|
|
+ emit_instr(ctx, jr, reg);
|
|
+}
|
|
+
|
|
+static inline u16 align_sp(unsigned int num)
|
|
+{
|
|
+ /* Double word alignment for 32-bit, quadword for 64-bit */
|
|
+ unsigned int align = IS_ENABLED(CONFIG_64BIT) ? 16 : 8;
|
|
+ num = (num + (align - 1)) & -align;
|
|
+ return num;
|
|
+}
|
|
+
|
|
+static void save_bpf_jit_regs(struct jit_ctx *ctx, unsigned offset)
|
|
+{
|
|
+ int i = 0, real_off = 0;
|
|
+ u32 sflags, tmp_flags;
|
|
+
|
|
+ /* Adjust the stack pointer */
|
|
+ if (offset)
|
|
+ emit_stack_offset(-align_sp(offset), ctx);
|
|
+
|
|
+ tmp_flags = sflags = ctx->flags >> SEEN_SREG_SFT;
|
|
+ /* sflags is essentially a bitmap */
|
|
+ while (tmp_flags) {
|
|
+ if ((sflags >> i) & 0x1) {
|
|
+ emit_store_stack_reg(MIPS_R_S0 + i, r_sp, real_off,
|
|
+ ctx);
|
|
+ real_off += SZREG;
|
|
+ }
|
|
+ i++;
|
|
+ tmp_flags >>= 1;
|
|
+ }
|
|
+
|
|
+ /* save return address */
|
|
+ if (ctx->flags & SEEN_CALL) {
|
|
+ emit_store_stack_reg(r_ra, r_sp, real_off, ctx);
|
|
+ real_off += SZREG;
|
|
+ }
|
|
+
|
|
+ /* Setup r_M leaving the alignment gap if necessary */
|
|
+ if (ctx->flags & SEEN_MEM) {
|
|
+ if (real_off % (SZREG * 2))
|
|
+ real_off += SZREG;
|
|
+ emit_long_instr(ctx, ADDIU, r_M, r_sp, real_off);
|
|
+ }
|
|
+}
|
|
+
|
|
+static void restore_bpf_jit_regs(struct jit_ctx *ctx,
|
|
+ unsigned int offset)
|
|
+{
|
|
+ int i, real_off = 0;
|
|
+ u32 sflags, tmp_flags;
|
|
+
|
|
+ tmp_flags = sflags = ctx->flags >> SEEN_SREG_SFT;
|
|
+ /* sflags is a bitmap */
|
|
+ i = 0;
|
|
+ while (tmp_flags) {
|
|
+ if ((sflags >> i) & 0x1) {
|
|
+ emit_load_stack_reg(MIPS_R_S0 + i, r_sp, real_off,
|
|
+ ctx);
|
|
+ real_off += SZREG;
|
|
+ }
|
|
+ i++;
|
|
+ tmp_flags >>= 1;
|
|
+ }
|
|
+
|
|
+ /* restore return address */
|
|
+ if (ctx->flags & SEEN_CALL)
|
|
+ emit_load_stack_reg(r_ra, r_sp, real_off, ctx);
|
|
+
|
|
+ /* Restore the sp and discard the scrach memory */
|
|
+ if (offset)
|
|
+ emit_stack_offset(align_sp(offset), ctx);
|
|
+}
|
|
+
|
|
+static unsigned int get_stack_depth(struct jit_ctx *ctx)
|
|
+{
|
|
+ int sp_off = 0;
|
|
+
|
|
+
|
|
+ /* How may s* regs do we need to preserved? */
|
|
+ sp_off += hweight32(ctx->flags >> SEEN_SREG_SFT) * SZREG;
|
|
+
|
|
+ if (ctx->flags & SEEN_MEM)
|
|
+ sp_off += 4 * BPF_MEMWORDS; /* BPF_MEMWORDS are 32-bit */
|
|
+
|
|
+ if (ctx->flags & SEEN_CALL)
|
|
+ sp_off += SZREG; /* Space for our ra register */
|
|
+
|
|
+ return sp_off;
|
|
+}
|
|
+
|
|
+static void build_prologue(struct jit_ctx *ctx)
|
|
+{
|
|
+ int sp_off;
|
|
+
|
|
+ /* Calculate the total offset for the stack pointer */
|
|
+ sp_off = get_stack_depth(ctx);
|
|
+ save_bpf_jit_regs(ctx, sp_off);
|
|
+
|
|
+ if (ctx->flags & SEEN_SKB)
|
|
+ emit_reg_move(r_skb, MIPS_R_A0, ctx);
|
|
+
|
|
+ if (ctx->flags & SEEN_SKB_DATA) {
|
|
+ /* Load packet length */
|
|
+ emit_load(r_skb_len, r_skb, offsetof(struct sk_buff, len),
|
|
+ ctx);
|
|
+ emit_load(r_tmp, r_skb, offsetof(struct sk_buff, data_len),
|
|
+ ctx);
|
|
+ /* Load the data pointer */
|
|
+ emit_load_ptr(r_skb_data, r_skb,
|
|
+ offsetof(struct sk_buff, data), ctx);
|
|
+ /* Load the header length */
|
|
+ emit_subu(r_skb_hl, r_skb_len, r_tmp, ctx);
|
|
+ }
|
|
+
|
|
+ if (ctx->flags & SEEN_X)
|
|
+ emit_jit_reg_move(r_X, r_zero, ctx);
|
|
+
|
|
+ /*
|
|
+ * Do not leak kernel data to userspace, we only need to clear
|
|
+ * r_A if it is ever used. In fact if it is never used, we
|
|
+ * will not save/restore it, so clearing it in this case would
|
|
+ * corrupt the state of the caller.
|
|
+ */
|
|
+ if (bpf_needs_clear_a(&ctx->skf->insns[0]) &&
|
|
+ (ctx->flags & SEEN_A))
|
|
+ emit_jit_reg_move(r_A, r_zero, ctx);
|
|
+}
|
|
+
|
|
+static void build_epilogue(struct jit_ctx *ctx)
|
|
+{
|
|
+ unsigned int sp_off;
|
|
+
|
|
+ /* Calculate the total offset for the stack pointer */
|
|
+
|
|
+ sp_off = get_stack_depth(ctx);
|
|
+ restore_bpf_jit_regs(ctx, sp_off);
|
|
+
|
|
+ /* Return */
|
|
+ emit_jr(r_ra, ctx);
|
|
+ emit_nop(ctx);
|
|
+}
|
|
+
|
|
+#define CHOOSE_LOAD_FUNC(K, func) \
|
|
+ ((int)K < 0 ? ((int)K >= SKF_LL_OFF ? func##_negative : func) : \
|
|
+ func##_positive)
|
|
+
|
|
+static bool is_bad_offset(int b_off)
|
|
+{
|
|
+ return b_off > 0x1ffff || b_off < -0x20000;
|
|
+}
|
|
+
|
|
+static int build_body(struct jit_ctx *ctx)
|
|
+{
|
|
+ const struct bpf_prog *prog = ctx->skf;
|
|
+ const struct sock_filter *inst;
|
|
+ unsigned int i, off, condt;
|
|
+ u32 k, b_off __maybe_unused;
|
|
+ u8 (*sk_load_func)(unsigned long *skb, int offset);
|
|
+
|
|
+ for (i = 0; i < prog->len; i++) {
|
|
+ u16 code;
|
|
+
|
|
+ inst = &(prog->insns[i]);
|
|
+ pr_debug("%s: code->0x%02x, jt->0x%x, jf->0x%x, k->0x%x\n",
|
|
+ __func__, inst->code, inst->jt, inst->jf, inst->k);
|
|
+ k = inst->k;
|
|
+ code = bpf_anc_helper(inst);
|
|
+
|
|
+ if (ctx->target == NULL)
|
|
+ ctx->offsets[i] = ctx->idx * 4;
|
|
+
|
|
+ switch (code) {
|
|
+ case BPF_LD | BPF_IMM:
|
|
+ /* A <- k ==> li r_A, k */
|
|
+ ctx->flags |= SEEN_A;
|
|
+ emit_load_imm(r_A, k, ctx);
|
|
+ break;
|
|
+ case BPF_LD | BPF_W | BPF_LEN:
|
|
+ BUILD_BUG_ON(FIELD_SIZEOF(struct sk_buff, len) != 4);
|
|
+ /* A <- len ==> lw r_A, offset(skb) */
|
|
+ ctx->flags |= SEEN_SKB | SEEN_A;
|
|
+ off = offsetof(struct sk_buff, len);
|
|
+ emit_load(r_A, r_skb, off, ctx);
|
|
+ break;
|
|
+ case BPF_LD | BPF_MEM:
|
|
+ /* A <- M[k] ==> lw r_A, offset(M) */
|
|
+ ctx->flags |= SEEN_MEM | SEEN_A;
|
|
+ emit_load(r_A, r_M, SCRATCH_OFF(k), ctx);
|
|
+ break;
|
|
+ case BPF_LD | BPF_W | BPF_ABS:
|
|
+ /* A <- P[k:4] */
|
|
+ sk_load_func = CHOOSE_LOAD_FUNC(k, sk_load_word);
|
|
+ goto load;
|
|
+ case BPF_LD | BPF_H | BPF_ABS:
|
|
+ /* A <- P[k:2] */
|
|
+ sk_load_func = CHOOSE_LOAD_FUNC(k, sk_load_half);
|
|
+ goto load;
|
|
+ case BPF_LD | BPF_B | BPF_ABS:
|
|
+ /* A <- P[k:1] */
|
|
+ sk_load_func = CHOOSE_LOAD_FUNC(k, sk_load_byte);
|
|
+load:
|
|
+ emit_load_imm(r_off, k, ctx);
|
|
+load_common:
|
|
+ ctx->flags |= SEEN_CALL | SEEN_OFF |
|
|
+ SEEN_SKB | SEEN_A | SEEN_SKB_DATA;
|
|
+
|
|
+ emit_load_func(r_s0, (ptr)sk_load_func, ctx);
|
|
+ emit_reg_move(MIPS_R_A0, r_skb, ctx);
|
|
+ emit_jalr(MIPS_R_RA, r_s0, ctx);
|
|
+ /* Load second argument to delay slot */
|
|
+ emit_reg_move(MIPS_R_A1, r_off, ctx);
|
|
+ /* Check the error value */
|
|
+ emit_bcond(MIPS_COND_EQ, r_ret, 0, b_imm(i + 1, ctx),
|
|
+ ctx);
|
|
+ /* Load return register on DS for failures */
|
|
+ emit_reg_move(r_ret, r_zero, ctx);
|
|
+ /* Return with error */
|
|
+ b_off = b_imm(prog->len, ctx);
|
|
+ if (is_bad_offset(b_off))
|
|
+ return -E2BIG;
|
|
+ emit_b(b_off, ctx);
|
|
+ emit_nop(ctx);
|
|
+ break;
|
|
+ case BPF_LD | BPF_W | BPF_IND:
|
|
+ /* A <- P[X + k:4] */
|
|
+ sk_load_func = sk_load_word;
|
|
+ goto load_ind;
|
|
+ case BPF_LD | BPF_H | BPF_IND:
|
|
+ /* A <- P[X + k:2] */
|
|
+ sk_load_func = sk_load_half;
|
|
+ goto load_ind;
|
|
+ case BPF_LD | BPF_B | BPF_IND:
|
|
+ /* A <- P[X + k:1] */
|
|
+ sk_load_func = sk_load_byte;
|
|
+load_ind:
|
|
+ ctx->flags |= SEEN_OFF | SEEN_X;
|
|
+ emit_addiu(r_off, r_X, k, ctx);
|
|
+ goto load_common;
|
|
+ case BPF_LDX | BPF_IMM:
|
|
+ /* X <- k */
|
|
+ ctx->flags |= SEEN_X;
|
|
+ emit_load_imm(r_X, k, ctx);
|
|
+ break;
|
|
+ case BPF_LDX | BPF_MEM:
|
|
+ /* X <- M[k] */
|
|
+ ctx->flags |= SEEN_X | SEEN_MEM;
|
|
+ emit_load(r_X, r_M, SCRATCH_OFF(k), ctx);
|
|
+ break;
|
|
+ case BPF_LDX | BPF_W | BPF_LEN:
|
|
+ /* X <- len */
|
|
+ ctx->flags |= SEEN_X | SEEN_SKB;
|
|
+ off = offsetof(struct sk_buff, len);
|
|
+ emit_load(r_X, r_skb, off, ctx);
|
|
+ break;
|
|
+ case BPF_LDX | BPF_B | BPF_MSH:
|
|
+ /* X <- 4 * (P[k:1] & 0xf) */
|
|
+ ctx->flags |= SEEN_X | SEEN_CALL | SEEN_SKB;
|
|
+ /* Load offset to a1 */
|
|
+ emit_load_func(r_s0, (ptr)sk_load_byte, ctx);
|
|
+ /*
|
|
+ * This may emit two instructions so it may not fit
|
|
+ * in the delay slot. So use a0 in the delay slot.
|
|
+ */
|
|
+ emit_load_imm(MIPS_R_A1, k, ctx);
|
|
+ emit_jalr(MIPS_R_RA, r_s0, ctx);
|
|
+ emit_reg_move(MIPS_R_A0, r_skb, ctx); /* delay slot */
|
|
+ /* Check the error value */
|
|
+ b_off = b_imm(prog->len, ctx);
|
|
+ if (is_bad_offset(b_off))
|
|
+ return -E2BIG;
|
|
+ emit_bcond(MIPS_COND_NE, r_ret, 0, b_off, ctx);
|
|
+ emit_reg_move(r_ret, r_zero, ctx);
|
|
+ /* We are good */
|
|
+ /* X <- P[1:K] & 0xf */
|
|
+ emit_andi(r_X, r_A, 0xf, ctx);
|
|
+ /* X << 2 */
|
|
+ emit_b(b_imm(i + 1, ctx), ctx);
|
|
+ emit_sll(r_X, r_X, 2, ctx); /* delay slot */
|
|
+ break;
|
|
+ case BPF_ST:
|
|
+ /* M[k] <- A */
|
|
+ ctx->flags |= SEEN_MEM | SEEN_A;
|
|
+ emit_store(r_A, r_M, SCRATCH_OFF(k), ctx);
|
|
+ break;
|
|
+ case BPF_STX:
|
|
+ /* M[k] <- X */
|
|
+ ctx->flags |= SEEN_MEM | SEEN_X;
|
|
+ emit_store(r_X, r_M, SCRATCH_OFF(k), ctx);
|
|
+ break;
|
|
+ case BPF_ALU | BPF_ADD | BPF_K:
|
|
+ /* A += K */
|
|
+ ctx->flags |= SEEN_A;
|
|
+ emit_addiu(r_A, r_A, k, ctx);
|
|
+ break;
|
|
+ case BPF_ALU | BPF_ADD | BPF_X:
|
|
+ /* A += X */
|
|
+ ctx->flags |= SEEN_A | SEEN_X;
|
|
+ emit_addu(r_A, r_A, r_X, ctx);
|
|
+ break;
|
|
+ case BPF_ALU | BPF_SUB | BPF_K:
|
|
+ /* A -= K */
|
|
+ ctx->flags |= SEEN_A;
|
|
+ emit_addiu(r_A, r_A, -k, ctx);
|
|
+ break;
|
|
+ case BPF_ALU | BPF_SUB | BPF_X:
|
|
+ /* A -= X */
|
|
+ ctx->flags |= SEEN_A | SEEN_X;
|
|
+ emit_subu(r_A, r_A, r_X, ctx);
|
|
+ break;
|
|
+ case BPF_ALU | BPF_MUL | BPF_K:
|
|
+ /* A *= K */
|
|
+ /* Load K to scratch register before MUL */
|
|
+ ctx->flags |= SEEN_A;
|
|
+ emit_load_imm(r_s0, k, ctx);
|
|
+ emit_mul(r_A, r_A, r_s0, ctx);
|
|
+ break;
|
|
+ case BPF_ALU | BPF_MUL | BPF_X:
|
|
+ /* A *= X */
|
|
+ ctx->flags |= SEEN_A | SEEN_X;
|
|
+ emit_mul(r_A, r_A, r_X, ctx);
|
|
+ break;
|
|
+ case BPF_ALU | BPF_DIV | BPF_K:
|
|
+ /* A /= k */
|
|
+ if (k == 1)
|
|
+ break;
|
|
+ if (optimize_div(&k)) {
|
|
+ ctx->flags |= SEEN_A;
|
|
+ emit_srl(r_A, r_A, k, ctx);
|
|
+ break;
|
|
+ }
|
|
+ ctx->flags |= SEEN_A;
|
|
+ emit_load_imm(r_s0, k, ctx);
|
|
+ emit_div(r_A, r_s0, ctx);
|
|
+ break;
|
|
+ case BPF_ALU | BPF_MOD | BPF_K:
|
|
+ /* A %= k */
|
|
+ if (k == 1) {
|
|
+ ctx->flags |= SEEN_A;
|
|
+ emit_jit_reg_move(r_A, r_zero, ctx);
|
|
+ } else {
|
|
+ ctx->flags |= SEEN_A;
|
|
+ emit_load_imm(r_s0, k, ctx);
|
|
+ emit_mod(r_A, r_s0, ctx);
|
|
+ }
|
|
+ break;
|
|
+ case BPF_ALU | BPF_DIV | BPF_X:
|
|
+ /* A /= X */
|
|
+ ctx->flags |= SEEN_X | SEEN_A;
|
|
+ /* Check if r_X is zero */
|
|
+ b_off = b_imm(prog->len, ctx);
|
|
+ if (is_bad_offset(b_off))
|
|
+ return -E2BIG;
|
|
+ emit_bcond(MIPS_COND_EQ, r_X, r_zero, b_off, ctx);
|
|
+ emit_load_imm(r_ret, 0, ctx); /* delay slot */
|
|
+ emit_div(r_A, r_X, ctx);
|
|
+ break;
|
|
+ case BPF_ALU | BPF_MOD | BPF_X:
|
|
+ /* A %= X */
|
|
+ ctx->flags |= SEEN_X | SEEN_A;
|
|
+ /* Check if r_X is zero */
|
|
+ b_off = b_imm(prog->len, ctx);
|
|
+ if (is_bad_offset(b_off))
|
|
+ return -E2BIG;
|
|
+ emit_bcond(MIPS_COND_EQ, r_X, r_zero, b_off, ctx);
|
|
+ emit_load_imm(r_ret, 0, ctx); /* delay slot */
|
|
+ emit_mod(r_A, r_X, ctx);
|
|
+ break;
|
|
+ case BPF_ALU | BPF_OR | BPF_K:
|
|
+ /* A |= K */
|
|
+ ctx->flags |= SEEN_A;
|
|
+ emit_ori(r_A, r_A, k, ctx);
|
|
+ break;
|
|
+ case BPF_ALU | BPF_OR | BPF_X:
|
|
+ /* A |= X */
|
|
+ ctx->flags |= SEEN_A;
|
|
+ emit_ori(r_A, r_A, r_X, ctx);
|
|
+ break;
|
|
+ case BPF_ALU | BPF_XOR | BPF_K:
|
|
+ /* A ^= k */
|
|
+ ctx->flags |= SEEN_A;
|
|
+ emit_xori(r_A, r_A, k, ctx);
|
|
+ break;
|
|
+ case BPF_ANC | SKF_AD_ALU_XOR_X:
|
|
+ case BPF_ALU | BPF_XOR | BPF_X:
|
|
+ /* A ^= X */
|
|
+ ctx->flags |= SEEN_A;
|
|
+ emit_xor(r_A, r_A, r_X, ctx);
|
|
+ break;
|
|
+ case BPF_ALU | BPF_AND | BPF_K:
|
|
+ /* A &= K */
|
|
+ ctx->flags |= SEEN_A;
|
|
+ emit_andi(r_A, r_A, k, ctx);
|
|
+ break;
|
|
+ case BPF_ALU | BPF_AND | BPF_X:
|
|
+ /* A &= X */
|
|
+ ctx->flags |= SEEN_A | SEEN_X;
|
|
+ emit_and(r_A, r_A, r_X, ctx);
|
|
+ break;
|
|
+ case BPF_ALU | BPF_LSH | BPF_K:
|
|
+ /* A <<= K */
|
|
+ ctx->flags |= SEEN_A;
|
|
+ emit_sll(r_A, r_A, k, ctx);
|
|
+ break;
|
|
+ case BPF_ALU | BPF_LSH | BPF_X:
|
|
+ /* A <<= X */
|
|
+ ctx->flags |= SEEN_A | SEEN_X;
|
|
+ emit_sllv(r_A, r_A, r_X, ctx);
|
|
+ break;
|
|
+ case BPF_ALU | BPF_RSH | BPF_K:
|
|
+ /* A >>= K */
|
|
+ ctx->flags |= SEEN_A;
|
|
+ emit_srl(r_A, r_A, k, ctx);
|
|
+ break;
|
|
+ case BPF_ALU | BPF_RSH | BPF_X:
|
|
+ ctx->flags |= SEEN_A | SEEN_X;
|
|
+ emit_srlv(r_A, r_A, r_X, ctx);
|
|
+ break;
|
|
+ case BPF_ALU | BPF_NEG:
|
|
+ /* A = -A */
|
|
+ ctx->flags |= SEEN_A;
|
|
+ emit_neg(r_A, ctx);
|
|
+ break;
|
|
+ case BPF_JMP | BPF_JA:
|
|
+ /* pc += K */
|
|
+ b_off = b_imm(i + k + 1, ctx);
|
|
+ if (is_bad_offset(b_off))
|
|
+ return -E2BIG;
|
|
+ emit_b(b_off, ctx);
|
|
+ emit_nop(ctx);
|
|
+ break;
|
|
+ case BPF_JMP | BPF_JEQ | BPF_K:
|
|
+ /* pc += ( A == K ) ? pc->jt : pc->jf */
|
|
+ condt = MIPS_COND_EQ | MIPS_COND_K;
|
|
+ goto jmp_cmp;
|
|
+ case BPF_JMP | BPF_JEQ | BPF_X:
|
|
+ ctx->flags |= SEEN_X;
|
|
+ /* pc += ( A == X ) ? pc->jt : pc->jf */
|
|
+ condt = MIPS_COND_EQ | MIPS_COND_X;
|
|
+ goto jmp_cmp;
|
|
+ case BPF_JMP | BPF_JGE | BPF_K:
|
|
+ /* pc += ( A >= K ) ? pc->jt : pc->jf */
|
|
+ condt = MIPS_COND_GE | MIPS_COND_K;
|
|
+ goto jmp_cmp;
|
|
+ case BPF_JMP | BPF_JGE | BPF_X:
|
|
+ ctx->flags |= SEEN_X;
|
|
+ /* pc += ( A >= X ) ? pc->jt : pc->jf */
|
|
+ condt = MIPS_COND_GE | MIPS_COND_X;
|
|
+ goto jmp_cmp;
|
|
+ case BPF_JMP | BPF_JGT | BPF_K:
|
|
+ /* pc += ( A > K ) ? pc->jt : pc->jf */
|
|
+ condt = MIPS_COND_GT | MIPS_COND_K;
|
|
+ goto jmp_cmp;
|
|
+ case BPF_JMP | BPF_JGT | BPF_X:
|
|
+ ctx->flags |= SEEN_X;
|
|
+ /* pc += ( A > X ) ? pc->jt : pc->jf */
|
|
+ condt = MIPS_COND_GT | MIPS_COND_X;
|
|
+jmp_cmp:
|
|
+ /* Greater or Equal */
|
|
+ if ((condt & MIPS_COND_GE) ||
|
|
+ (condt & MIPS_COND_GT)) {
|
|
+ if (condt & MIPS_COND_K) { /* K */
|
|
+ ctx->flags |= SEEN_A;
|
|
+ emit_sltiu(r_s0, r_A, k, ctx);
|
|
+ } else { /* X */
|
|
+ ctx->flags |= SEEN_A |
|
|
+ SEEN_X;
|
|
+ emit_sltu(r_s0, r_A, r_X, ctx);
|
|
+ }
|
|
+ /* A < (K|X) ? r_scrach = 1 */
|
|
+ b_off = b_imm(i + inst->jf + 1, ctx);
|
|
+ emit_bcond(MIPS_COND_NE, r_s0, r_zero, b_off,
|
|
+ ctx);
|
|
+ emit_nop(ctx);
|
|
+ /* A > (K|X) ? scratch = 0 */
|
|
+ if (condt & MIPS_COND_GT) {
|
|
+ /* Checking for equality */
|
|
+ ctx->flags |= SEEN_A | SEEN_X;
|
|
+ if (condt & MIPS_COND_K)
|
|
+ emit_load_imm(r_s0, k, ctx);
|
|
+ else
|
|
+ emit_jit_reg_move(r_s0, r_X,
|
|
+ ctx);
|
|
+ b_off = b_imm(i + inst->jf + 1, ctx);
|
|
+ emit_bcond(MIPS_COND_EQ, r_A, r_s0,
|
|
+ b_off, ctx);
|
|
+ emit_nop(ctx);
|
|
+ /* Finally, A > K|X */
|
|
+ b_off = b_imm(i + inst->jt + 1, ctx);
|
|
+ emit_b(b_off, ctx);
|
|
+ emit_nop(ctx);
|
|
+ } else {
|
|
+ /* A >= (K|X) so jump */
|
|
+ b_off = b_imm(i + inst->jt + 1, ctx);
|
|
+ emit_b(b_off, ctx);
|
|
+ emit_nop(ctx);
|
|
+ }
|
|
+ } else {
|
|
+ /* A == K|X */
|
|
+ if (condt & MIPS_COND_K) { /* K */
|
|
+ ctx->flags |= SEEN_A;
|
|
+ emit_load_imm(r_s0, k, ctx);
|
|
+ /* jump true */
|
|
+ b_off = b_imm(i + inst->jt + 1, ctx);
|
|
+ emit_bcond(MIPS_COND_EQ, r_A, r_s0,
|
|
+ b_off, ctx);
|
|
+ emit_nop(ctx);
|
|
+ /* jump false */
|
|
+ b_off = b_imm(i + inst->jf + 1,
|
|
+ ctx);
|
|
+ emit_bcond(MIPS_COND_NE, r_A, r_s0,
|
|
+ b_off, ctx);
|
|
+ emit_nop(ctx);
|
|
+ } else { /* X */
|
|
+ /* jump true */
|
|
+ ctx->flags |= SEEN_A | SEEN_X;
|
|
+ b_off = b_imm(i + inst->jt + 1,
|
|
+ ctx);
|
|
+ emit_bcond(MIPS_COND_EQ, r_A, r_X,
|
|
+ b_off, ctx);
|
|
+ emit_nop(ctx);
|
|
+ /* jump false */
|
|
+ b_off = b_imm(i + inst->jf + 1, ctx);
|
|
+ emit_bcond(MIPS_COND_NE, r_A, r_X,
|
|
+ b_off, ctx);
|
|
+ emit_nop(ctx);
|
|
+ }
|
|
+ }
|
|
+ break;
|
|
+ case BPF_JMP | BPF_JSET | BPF_K:
|
|
+ ctx->flags |= SEEN_A;
|
|
+ /* pc += (A & K) ? pc -> jt : pc -> jf */
|
|
+ emit_load_imm(r_s1, k, ctx);
|
|
+ emit_and(r_s0, r_A, r_s1, ctx);
|
|
+ /* jump true */
|
|
+ b_off = b_imm(i + inst->jt + 1, ctx);
|
|
+ emit_bcond(MIPS_COND_NE, r_s0, r_zero, b_off, ctx);
|
|
+ emit_nop(ctx);
|
|
+ /* jump false */
|
|
+ b_off = b_imm(i + inst->jf + 1, ctx);
|
|
+ emit_b(b_off, ctx);
|
|
+ emit_nop(ctx);
|
|
+ break;
|
|
+ case BPF_JMP | BPF_JSET | BPF_X:
|
|
+ ctx->flags |= SEEN_X | SEEN_A;
|
|
+ /* pc += (A & X) ? pc -> jt : pc -> jf */
|
|
+ emit_and(r_s0, r_A, r_X, ctx);
|
|
+ /* jump true */
|
|
+ b_off = b_imm(i + inst->jt + 1, ctx);
|
|
+ emit_bcond(MIPS_COND_NE, r_s0, r_zero, b_off, ctx);
|
|
+ emit_nop(ctx);
|
|
+ /* jump false */
|
|
+ b_off = b_imm(i + inst->jf + 1, ctx);
|
|
+ emit_b(b_off, ctx);
|
|
+ emit_nop(ctx);
|
|
+ break;
|
|
+ case BPF_RET | BPF_A:
|
|
+ ctx->flags |= SEEN_A;
|
|
+ if (i != prog->len - 1) {
|
|
+ /*
|
|
+ * If this is not the last instruction
|
|
+ * then jump to the epilogue
|
|
+ */
|
|
+ b_off = b_imm(prog->len, ctx);
|
|
+ if (is_bad_offset(b_off))
|
|
+ return -E2BIG;
|
|
+ emit_b(b_off, ctx);
|
|
+ }
|
|
+ emit_reg_move(r_ret, r_A, ctx); /* delay slot */
|
|
+ break;
|
|
+ case BPF_RET | BPF_K:
|
|
+ /*
|
|
+ * It can emit two instructions so it does not fit on
|
|
+ * the delay slot.
|
|
+ */
|
|
+ emit_load_imm(r_ret, k, ctx);
|
|
+ if (i != prog->len - 1) {
|
|
+ /*
|
|
+ * If this is not the last instruction
|
|
+ * then jump to the epilogue
|
|
+ */
|
|
+ b_off = b_imm(prog->len, ctx);
|
|
+ if (is_bad_offset(b_off))
|
|
+ return -E2BIG;
|
|
+ emit_b(b_off, ctx);
|
|
+ emit_nop(ctx);
|
|
+ }
|
|
+ break;
|
|
+ case BPF_MISC | BPF_TAX:
|
|
+ /* X = A */
|
|
+ ctx->flags |= SEEN_X | SEEN_A;
|
|
+ emit_jit_reg_move(r_X, r_A, ctx);
|
|
+ break;
|
|
+ case BPF_MISC | BPF_TXA:
|
|
+ /* A = X */
|
|
+ ctx->flags |= SEEN_A | SEEN_X;
|
|
+ emit_jit_reg_move(r_A, r_X, ctx);
|
|
+ break;
|
|
+ /* AUX */
|
|
+ case BPF_ANC | SKF_AD_PROTOCOL:
|
|
+ /* A = ntohs(skb->protocol */
|
|
+ ctx->flags |= SEEN_SKB | SEEN_OFF | SEEN_A;
|
|
+ BUILD_BUG_ON(FIELD_SIZEOF(struct sk_buff,
|
|
+ protocol) != 2);
|
|
+ off = offsetof(struct sk_buff, protocol);
|
|
+ emit_half_load(r_A, r_skb, off, ctx);
|
|
+#ifdef CONFIG_CPU_LITTLE_ENDIAN
|
|
+ /* This needs little endian fixup */
|
|
+ if (cpu_has_wsbh) {
|
|
+ /* R2 and later have the wsbh instruction */
|
|
+ emit_wsbh(r_A, r_A, ctx);
|
|
+ } else {
|
|
+ /* Get first byte */
|
|
+ emit_andi(r_tmp_imm, r_A, 0xff, ctx);
|
|
+ /* Shift it */
|
|
+ emit_sll(r_tmp, r_tmp_imm, 8, ctx);
|
|
+ /* Get second byte */
|
|
+ emit_srl(r_tmp_imm, r_A, 8, ctx);
|
|
+ emit_andi(r_tmp_imm, r_tmp_imm, 0xff, ctx);
|
|
+ /* Put everyting together in r_A */
|
|
+ emit_or(r_A, r_tmp, r_tmp_imm, ctx);
|
|
+ }
|
|
+#endif
|
|
+ break;
|
|
+ case BPF_ANC | SKF_AD_CPU:
|
|
+ ctx->flags |= SEEN_A | SEEN_OFF;
|
|
+ /* A = current_thread_info()->cpu */
|
|
+ BUILD_BUG_ON(FIELD_SIZEOF(struct thread_info,
|
|
+ cpu) != 4);
|
|
+ off = offsetof(struct thread_info, cpu);
|
|
+ /* $28/gp points to the thread_info struct */
|
|
+ emit_load(r_A, 28, off, ctx);
|
|
+ break;
|
|
+ case BPF_ANC | SKF_AD_IFINDEX:
|
|
+ /* A = skb->dev->ifindex */
|
|
+ case BPF_ANC | SKF_AD_HATYPE:
|
|
+ /* A = skb->dev->type */
|
|
+ ctx->flags |= SEEN_SKB | SEEN_A;
|
|
+ off = offsetof(struct sk_buff, dev);
|
|
+ /* Load *dev pointer */
|
|
+ emit_load_ptr(r_s0, r_skb, off, ctx);
|
|
+ /* error (0) in the delay slot */
|
|
+ b_off = b_imm(prog->len, ctx);
|
|
+ if (is_bad_offset(b_off))
|
|
+ return -E2BIG;
|
|
+ emit_bcond(MIPS_COND_EQ, r_s0, r_zero, b_off, ctx);
|
|
+ emit_reg_move(r_ret, r_zero, ctx);
|
|
+ if (code == (BPF_ANC | SKF_AD_IFINDEX)) {
|
|
+ BUILD_BUG_ON(FIELD_SIZEOF(struct net_device, ifindex) != 4);
|
|
+ off = offsetof(struct net_device, ifindex);
|
|
+ emit_load(r_A, r_s0, off, ctx);
|
|
+ } else { /* (code == (BPF_ANC | SKF_AD_HATYPE) */
|
|
+ BUILD_BUG_ON(FIELD_SIZEOF(struct net_device, type) != 2);
|
|
+ off = offsetof(struct net_device, type);
|
|
+ emit_half_load_unsigned(r_A, r_s0, off, ctx);
|
|
+ }
|
|
+ break;
|
|
+ case BPF_ANC | SKF_AD_MARK:
|
|
+ ctx->flags |= SEEN_SKB | SEEN_A;
|
|
+ BUILD_BUG_ON(FIELD_SIZEOF(struct sk_buff, mark) != 4);
|
|
+ off = offsetof(struct sk_buff, mark);
|
|
+ emit_load(r_A, r_skb, off, ctx);
|
|
+ break;
|
|
+ case BPF_ANC | SKF_AD_RXHASH:
|
|
+ ctx->flags |= SEEN_SKB | SEEN_A;
|
|
+ BUILD_BUG_ON(FIELD_SIZEOF(struct sk_buff, hash) != 4);
|
|
+ off = offsetof(struct sk_buff, hash);
|
|
+ emit_load(r_A, r_skb, off, ctx);
|
|
+ break;
|
|
+ case BPF_ANC | SKF_AD_VLAN_TAG:
|
|
+ ctx->flags |= SEEN_SKB | SEEN_A;
|
|
+ BUILD_BUG_ON(FIELD_SIZEOF(struct sk_buff,
|
|
+ vlan_tci) != 2);
|
|
+ off = offsetof(struct sk_buff, vlan_tci);
|
|
+ emit_half_load_unsigned(r_A, r_skb, off, ctx);
|
|
+ break;
|
|
+ case BPF_ANC | SKF_AD_VLAN_TAG_PRESENT:
|
|
+ ctx->flags |= SEEN_SKB | SEEN_A;
|
|
+ emit_load_byte(r_A, r_skb, PKT_VLAN_PRESENT_OFFSET(), ctx);
|
|
+ if (PKT_VLAN_PRESENT_BIT)
|
|
+ emit_srl(r_A, r_A, PKT_VLAN_PRESENT_BIT, ctx);
|
|
+ if (PKT_VLAN_PRESENT_BIT < 7)
|
|
+ emit_andi(r_A, r_A, 1, ctx);
|
|
+ break;
|
|
+ case BPF_ANC | SKF_AD_PKTTYPE:
|
|
+ ctx->flags |= SEEN_SKB;
|
|
+
|
|
+ emit_load_byte(r_tmp, r_skb, PKT_TYPE_OFFSET(), ctx);
|
|
+ /* Keep only the last 3 bits */
|
|
+ emit_andi(r_A, r_tmp, PKT_TYPE_MAX, ctx);
|
|
+#ifdef __BIG_ENDIAN_BITFIELD
|
|
+ /* Get the actual packet type to the lower 3 bits */
|
|
+ emit_srl(r_A, r_A, 5, ctx);
|
|
+#endif
|
|
+ break;
|
|
+ case BPF_ANC | SKF_AD_QUEUE:
|
|
+ ctx->flags |= SEEN_SKB | SEEN_A;
|
|
+ BUILD_BUG_ON(FIELD_SIZEOF(struct sk_buff,
|
|
+ queue_mapping) != 2);
|
|
+ BUILD_BUG_ON(offsetof(struct sk_buff,
|
|
+ queue_mapping) > 0xff);
|
|
+ off = offsetof(struct sk_buff, queue_mapping);
|
|
+ emit_half_load_unsigned(r_A, r_skb, off, ctx);
|
|
+ break;
|
|
+ default:
|
|
+ pr_debug("%s: Unhandled opcode: 0x%02x\n", __FILE__,
|
|
+ inst->code);
|
|
+ return -1;
|
|
+ }
|
|
+ }
|
|
+
|
|
+ /* compute offsets only during the first pass */
|
|
+ if (ctx->target == NULL)
|
|
+ ctx->offsets[i] = ctx->idx * 4;
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+void bpf_jit_compile(struct bpf_prog *fp)
|
|
+{
|
|
+ struct jit_ctx ctx;
|
|
+ unsigned int alloc_size, tmp_idx;
|
|
+
|
|
+ if (!bpf_jit_enable)
|
|
+ return;
|
|
+
|
|
+ memset(&ctx, 0, sizeof(ctx));
|
|
+
|
|
+ ctx.offsets = kcalloc(fp->len + 1, sizeof(*ctx.offsets), GFP_KERNEL);
|
|
+ if (ctx.offsets == NULL)
|
|
+ return;
|
|
+
|
|
+ ctx.skf = fp;
|
|
+
|
|
+ if (build_body(&ctx))
|
|
+ goto out;
|
|
+
|
|
+ tmp_idx = ctx.idx;
|
|
+ build_prologue(&ctx);
|
|
+ ctx.prologue_bytes = (ctx.idx - tmp_idx) * 4;
|
|
+ /* just to complete the ctx.idx count */
|
|
+ build_epilogue(&ctx);
|
|
+
|
|
+ alloc_size = 4 * ctx.idx;
|
|
+ ctx.target = module_alloc(alloc_size);
|
|
+ if (ctx.target == NULL)
|
|
+ goto out;
|
|
+
|
|
+ /* Clean it */
|
|
+ memset(ctx.target, 0, alloc_size);
|
|
+
|
|
+ ctx.idx = 0;
|
|
+
|
|
+ /* Generate the actual JIT code */
|
|
+ build_prologue(&ctx);
|
|
+ if (build_body(&ctx)) {
|
|
+ module_memfree(ctx.target);
|
|
+ goto out;
|
|
+ }
|
|
+ build_epilogue(&ctx);
|
|
+
|
|
+ /* Update the icache */
|
|
+ flush_icache_range((ptr)ctx.target, (ptr)(ctx.target + ctx.idx));
|
|
+
|
|
+ if (bpf_jit_enable > 1)
|
|
+ /* Dump JIT code */
|
|
+ bpf_jit_dump(fp->len, alloc_size, 2, ctx.target);
|
|
+
|
|
+ fp->bpf_func = (void *)ctx.target;
|
|
+ fp->jited = 1;
|
|
+
|
|
+out:
|
|
+ kfree(ctx.offsets);
|
|
+}
|
|
+
|
|
+void bpf_jit_free(struct bpf_prog *fp)
|
|
+{
|
|
+ if (fp->jited)
|
|
+ module_memfree(fp->bpf_func);
|
|
+
|
|
+ bpf_prog_unlock_free(fp);
|
|
+}
|
|
diff --git a/arch/mips/net/bpf_jit_asm.S b/arch/mips/net/bpf_jit_asm.S
|
|
new file mode 100644
|
|
index 0000000000000..57154c5883b6f
|
|
--- /dev/null
|
|
+++ b/arch/mips/net/bpf_jit_asm.S
|
|
@@ -0,0 +1,285 @@
|
|
+/*
|
|
+ * bpf_jib_asm.S: Packet/header access helper functions for MIPS/MIPS64 BPF
|
|
+ * compiler.
|
|
+ *
|
|
+ * Copyright (C) 2015 Imagination Technologies Ltd.
|
|
+ * Author: Markos Chandras <markos.chandras@imgtec.com>
|
|
+ *
|
|
+ * This program is free software; you can redistribute it and/or modify it
|
|
+ * under the terms of the GNU General Public License as published by the
|
|
+ * Free Software Foundation; version 2 of the License.
|
|
+ */
|
|
+
|
|
+#include <asm/asm.h>
|
|
+#include <asm/isa-rev.h>
|
|
+#include <asm/regdef.h>
|
|
+#include "bpf_jit.h"
|
|
+
|
|
+/* ABI
|
|
+ *
|
|
+ * r_skb_hl skb header length
|
|
+ * r_skb_data skb data
|
|
+ * r_off(a1) offset register
|
|
+ * r_A BPF register A
|
|
+ * r_X PF register X
|
|
+ * r_skb(a0) *skb
|
|
+ * r_M *scratch memory
|
|
+ * r_skb_le skb length
|
|
+ * r_s0 Scratch register 0
|
|
+ * r_s1 Scratch register 1
|
|
+ *
|
|
+ * On entry:
|
|
+ * a0: *skb
|
|
+ * a1: offset (imm or imm + X)
|
|
+ *
|
|
+ * All non-BPF-ABI registers are free for use. On return, we only
|
|
+ * care about r_ret. The BPF-ABI registers are assumed to remain
|
|
+ * unmodified during the entire filter operation.
|
|
+ */
|
|
+
|
|
+#define skb a0
|
|
+#define offset a1
|
|
+#define SKF_LL_OFF (-0x200000) /* Can't include linux/filter.h in assembly */
|
|
+
|
|
+ /* We know better :) so prevent assembler reordering etc */
|
|
+ .set noreorder
|
|
+
|
|
+#define is_offset_negative(TYPE) \
|
|
+ /* If offset is negative we have more work to do */ \
|
|
+ slti t0, offset, 0; \
|
|
+ bgtz t0, bpf_slow_path_##TYPE##_neg; \
|
|
+ /* Be careful what follows in DS. */
|
|
+
|
|
+#define is_offset_in_header(SIZE, TYPE) \
|
|
+ /* Reading from header? */ \
|
|
+ addiu $r_s0, $r_skb_hl, -SIZE; \
|
|
+ slt t0, $r_s0, offset; \
|
|
+ bgtz t0, bpf_slow_path_##TYPE; \
|
|
+
|
|
+LEAF(sk_load_word)
|
|
+ is_offset_negative(word)
|
|
+FEXPORT(sk_load_word_positive)
|
|
+ is_offset_in_header(4, word)
|
|
+ /* Offset within header boundaries */
|
|
+ PTR_ADDU t1, $r_skb_data, offset
|
|
+ .set reorder
|
|
+ lw $r_A, 0(t1)
|
|
+ .set noreorder
|
|
+#ifdef CONFIG_CPU_LITTLE_ENDIAN
|
|
+# if MIPS_ISA_REV >= 2
|
|
+ wsbh t0, $r_A
|
|
+ rotr $r_A, t0, 16
|
|
+# else
|
|
+ sll t0, $r_A, 24
|
|
+ srl t1, $r_A, 24
|
|
+ srl t2, $r_A, 8
|
|
+ or t0, t0, t1
|
|
+ andi t2, t2, 0xff00
|
|
+ andi t1, $r_A, 0xff00
|
|
+ or t0, t0, t2
|
|
+ sll t1, t1, 8
|
|
+ or $r_A, t0, t1
|
|
+# endif
|
|
+#endif
|
|
+ jr $r_ra
|
|
+ move $r_ret, zero
|
|
+ END(sk_load_word)
|
|
+
|
|
+LEAF(sk_load_half)
|
|
+ is_offset_negative(half)
|
|
+FEXPORT(sk_load_half_positive)
|
|
+ is_offset_in_header(2, half)
|
|
+ /* Offset within header boundaries */
|
|
+ PTR_ADDU t1, $r_skb_data, offset
|
|
+ lhu $r_A, 0(t1)
|
|
+#ifdef CONFIG_CPU_LITTLE_ENDIAN
|
|
+# if MIPS_ISA_REV >= 2
|
|
+ wsbh $r_A, $r_A
|
|
+# else
|
|
+ sll t0, $r_A, 8
|
|
+ srl t1, $r_A, 8
|
|
+ andi t0, t0, 0xff00
|
|
+ or $r_A, t0, t1
|
|
+# endif
|
|
+#endif
|
|
+ jr $r_ra
|
|
+ move $r_ret, zero
|
|
+ END(sk_load_half)
|
|
+
|
|
+LEAF(sk_load_byte)
|
|
+ is_offset_negative(byte)
|
|
+FEXPORT(sk_load_byte_positive)
|
|
+ is_offset_in_header(1, byte)
|
|
+ /* Offset within header boundaries */
|
|
+ PTR_ADDU t1, $r_skb_data, offset
|
|
+ lbu $r_A, 0(t1)
|
|
+ jr $r_ra
|
|
+ move $r_ret, zero
|
|
+ END(sk_load_byte)
|
|
+
|
|
+/*
|
|
+ * call skb_copy_bits:
|
|
+ * (prototype in linux/skbuff.h)
|
|
+ *
|
|
+ * int skb_copy_bits(sk_buff *skb, int offset, void *to, int len)
|
|
+ *
|
|
+ * o32 mandates we leave 4 spaces for argument registers in case
|
|
+ * the callee needs to use them. Even though we don't care about
|
|
+ * the argument registers ourselves, we need to allocate that space
|
|
+ * to remain ABI compliant since the callee may want to use that space.
|
|
+ * We also allocate 2 more spaces for $r_ra and our return register (*to).
|
|
+ *
|
|
+ * n64 is a bit different. The *caller* will allocate the space to preserve
|
|
+ * the arguments. So in 64-bit kernels, we allocate the 4-arg space for no
|
|
+ * good reason but it does not matter that much really.
|
|
+ *
|
|
+ * (void *to) is returned in r_s0
|
|
+ *
|
|
+ */
|
|
+#ifdef CONFIG_CPU_LITTLE_ENDIAN
|
|
+#define DS_OFFSET(SIZE) (4 * SZREG)
|
|
+#else
|
|
+#define DS_OFFSET(SIZE) ((4 * SZREG) + (4 - SIZE))
|
|
+#endif
|
|
+#define bpf_slow_path_common(SIZE) \
|
|
+ /* Quick check. Are we within reasonable boundaries? */ \
|
|
+ LONG_ADDIU $r_s1, $r_skb_len, -SIZE; \
|
|
+ sltu $r_s0, offset, $r_s1; \
|
|
+ beqz $r_s0, fault; \
|
|
+ /* Load 4th argument in DS */ \
|
|
+ LONG_ADDIU a3, zero, SIZE; \
|
|
+ PTR_ADDIU $r_sp, $r_sp, -(6 * SZREG); \
|
|
+ PTR_LA t0, skb_copy_bits; \
|
|
+ PTR_S $r_ra, (5 * SZREG)($r_sp); \
|
|
+ /* Assign low slot to a2 */ \
|
|
+ PTR_ADDIU a2, $r_sp, DS_OFFSET(SIZE); \
|
|
+ jalr t0; \
|
|
+ /* Reset our destination slot (DS but it's ok) */ \
|
|
+ INT_S zero, (4 * SZREG)($r_sp); \
|
|
+ /* \
|
|
+ * skb_copy_bits returns 0 on success and -EFAULT \
|
|
+ * on error. Our data live in a2. Do not bother with \
|
|
+ * our data if an error has been returned. \
|
|
+ */ \
|
|
+ /* Restore our frame */ \
|
|
+ PTR_L $r_ra, (5 * SZREG)($r_sp); \
|
|
+ INT_L $r_s0, (4 * SZREG)($r_sp); \
|
|
+ bltz v0, fault; \
|
|
+ PTR_ADDIU $r_sp, $r_sp, 6 * SZREG; \
|
|
+ move $r_ret, zero; \
|
|
+
|
|
+NESTED(bpf_slow_path_word, (6 * SZREG), $r_sp)
|
|
+ bpf_slow_path_common(4)
|
|
+#ifdef CONFIG_CPU_LITTLE_ENDIAN
|
|
+# if MIPS_ISA_REV >= 2
|
|
+ wsbh t0, $r_s0
|
|
+ jr $r_ra
|
|
+ rotr $r_A, t0, 16
|
|
+# else
|
|
+ sll t0, $r_s0, 24
|
|
+ srl t1, $r_s0, 24
|
|
+ srl t2, $r_s0, 8
|
|
+ or t0, t0, t1
|
|
+ andi t2, t2, 0xff00
|
|
+ andi t1, $r_s0, 0xff00
|
|
+ or t0, t0, t2
|
|
+ sll t1, t1, 8
|
|
+ jr $r_ra
|
|
+ or $r_A, t0, t1
|
|
+# endif
|
|
+#else
|
|
+ jr $r_ra
|
|
+ move $r_A, $r_s0
|
|
+#endif
|
|
+
|
|
+ END(bpf_slow_path_word)
|
|
+
|
|
+NESTED(bpf_slow_path_half, (6 * SZREG), $r_sp)
|
|
+ bpf_slow_path_common(2)
|
|
+#ifdef CONFIG_CPU_LITTLE_ENDIAN
|
|
+# if MIPS_ISA_REV >= 2
|
|
+ jr $r_ra
|
|
+ wsbh $r_A, $r_s0
|
|
+# else
|
|
+ sll t0, $r_s0, 8
|
|
+ andi t1, $r_s0, 0xff00
|
|
+ andi t0, t0, 0xff00
|
|
+ srl t1, t1, 8
|
|
+ jr $r_ra
|
|
+ or $r_A, t0, t1
|
|
+# endif
|
|
+#else
|
|
+ jr $r_ra
|
|
+ move $r_A, $r_s0
|
|
+#endif
|
|
+
|
|
+ END(bpf_slow_path_half)
|
|
+
|
|
+NESTED(bpf_slow_path_byte, (6 * SZREG), $r_sp)
|
|
+ bpf_slow_path_common(1)
|
|
+ jr $r_ra
|
|
+ move $r_A, $r_s0
|
|
+
|
|
+ END(bpf_slow_path_byte)
|
|
+
|
|
+/*
|
|
+ * Negative entry points
|
|
+ */
|
|
+ .macro bpf_is_end_of_data
|
|
+ li t0, SKF_LL_OFF
|
|
+ /* Reading link layer data? */
|
|
+ slt t1, offset, t0
|
|
+ bgtz t1, fault
|
|
+ /* Be careful what follows in DS. */
|
|
+ .endm
|
|
+/*
|
|
+ * call skb_copy_bits:
|
|
+ * (prototype in linux/filter.h)
|
|
+ *
|
|
+ * void *bpf_internal_load_pointer_neg_helper(const struct sk_buff *skb,
|
|
+ * int k, unsigned int size)
|
|
+ *
|
|
+ * see above (bpf_slow_path_common) for ABI restrictions
|
|
+ */
|
|
+#define bpf_negative_common(SIZE) \
|
|
+ PTR_ADDIU $r_sp, $r_sp, -(6 * SZREG); \
|
|
+ PTR_LA t0, bpf_internal_load_pointer_neg_helper; \
|
|
+ PTR_S $r_ra, (5 * SZREG)($r_sp); \
|
|
+ jalr t0; \
|
|
+ li a2, SIZE; \
|
|
+ PTR_L $r_ra, (5 * SZREG)($r_sp); \
|
|
+ /* Check return pointer */ \
|
|
+ beqz v0, fault; \
|
|
+ PTR_ADDIU $r_sp, $r_sp, 6 * SZREG; \
|
|
+ /* Preserve our pointer */ \
|
|
+ move $r_s0, v0; \
|
|
+ /* Set return value */ \
|
|
+ move $r_ret, zero; \
|
|
+
|
|
+bpf_slow_path_word_neg:
|
|
+ bpf_is_end_of_data
|
|
+NESTED(sk_load_word_negative, (6 * SZREG), $r_sp)
|
|
+ bpf_negative_common(4)
|
|
+ jr $r_ra
|
|
+ lw $r_A, 0($r_s0)
|
|
+ END(sk_load_word_negative)
|
|
+
|
|
+bpf_slow_path_half_neg:
|
|
+ bpf_is_end_of_data
|
|
+NESTED(sk_load_half_negative, (6 * SZREG), $r_sp)
|
|
+ bpf_negative_common(2)
|
|
+ jr $r_ra
|
|
+ lhu $r_A, 0($r_s0)
|
|
+ END(sk_load_half_negative)
|
|
+
|
|
+bpf_slow_path_byte_neg:
|
|
+ bpf_is_end_of_data
|
|
+NESTED(sk_load_byte_negative, (6 * SZREG), $r_sp)
|
|
+ bpf_negative_common(1)
|
|
+ jr $r_ra
|
|
+ lbu $r_A, 0($r_s0)
|
|
+ END(sk_load_byte_negative)
|
|
+
|
|
+fault:
|
|
+ jr $r_ra
|
|
+ addiu $r_ret, zero, 1
|
|
diff --git a/arch/powerpc/boot/dts/fsl/t1023rdb.dts b/arch/powerpc/boot/dts/fsl/t1023rdb.dts
|
|
index 5ba6fbfca2742..f82f85c65964c 100644
|
|
--- a/arch/powerpc/boot/dts/fsl/t1023rdb.dts
|
|
+++ b/arch/powerpc/boot/dts/fsl/t1023rdb.dts
|
|
@@ -154,7 +154,7 @@
|
|
|
|
fm1mac3: ethernet@e4000 {
|
|
phy-handle = <&sgmii_aqr_phy3>;
|
|
- phy-connection-type = "sgmii-2500";
|
|
+ phy-connection-type = "2500base-x";
|
|
sleep = <&rcpm 0x20000000>;
|
|
};
|
|
|
|
diff --git a/arch/riscv/include/uapi/asm/unistd.h b/arch/riscv/include/uapi/asm/unistd.h
|
|
index 13ce76cc5affe..80dff2c2bf677 100644
|
|
--- a/arch/riscv/include/uapi/asm/unistd.h
|
|
+++ b/arch/riscv/include/uapi/asm/unistd.h
|
|
@@ -18,9 +18,10 @@
|
|
#ifdef __LP64__
|
|
#define __ARCH_WANT_NEW_STAT
|
|
#define __ARCH_WANT_SET_GET_RLIMIT
|
|
-#define __ARCH_WANT_SYS_CLONE3
|
|
#endif /* __LP64__ */
|
|
|
|
+#define __ARCH_WANT_SYS_CLONE3
|
|
+
|
|
#include <asm-generic/unistd.h>
|
|
|
|
/*
|
|
diff --git a/arch/s390/net/bpf_jit_comp.c b/arch/s390/net/bpf_jit_comp.c
|
|
index 2d29966276296..f63e4cb6c9b31 100644
|
|
--- a/arch/s390/net/bpf_jit_comp.c
|
|
+++ b/arch/s390/net/bpf_jit_comp.c
|
|
@@ -1385,7 +1385,7 @@ struct bpf_prog *bpf_int_jit_compile(struct bpf_prog *fp)
|
|
jit.addrs = kvcalloc(fp->len + 1, sizeof(*jit.addrs), GFP_KERNEL);
|
|
if (jit.addrs == NULL) {
|
|
fp = orig_fp;
|
|
- goto out;
|
|
+ goto free_addrs;
|
|
}
|
|
/*
|
|
* Three initial passes:
|
|
diff --git a/arch/x86/Kconfig b/arch/x86/Kconfig
|
|
index 36a28b9e46cbd..8c1590432e866 100644
|
|
--- a/arch/x86/Kconfig
|
|
+++ b/arch/x86/Kconfig
|
|
@@ -1425,7 +1425,7 @@ config HIGHMEM4G
|
|
|
|
config HIGHMEM64G
|
|
bool "64GB"
|
|
- depends on !M486 && !M586 && !M586TSC && !M586MMX && !MGEODE_LX && !MGEODEGX1 && !MCYRIXIII && !MELAN && !MWINCHIPC6 && !WINCHIP3D && !MK6
|
|
+ depends on !M486 && !M586 && !M586TSC && !M586MMX && !MGEODE_LX && !MGEODEGX1 && !MCYRIXIII && !MELAN && !MWINCHIPC6 && !MWINCHIP3D && !MK6
|
|
select X86_PAE
|
|
---help---
|
|
Select this if you have a 32-bit processor and more than 4
|
|
diff --git a/arch/x86/kernel/early-quirks.c b/arch/x86/kernel/early-quirks.c
|
|
index 2f9ec14be3b11..6f6b1d04dadf9 100644
|
|
--- a/arch/x86/kernel/early-quirks.c
|
|
+++ b/arch/x86/kernel/early-quirks.c
|
|
@@ -710,12 +710,6 @@ static struct chipset early_qrk[] __initdata = {
|
|
*/
|
|
{ PCI_VENDOR_ID_INTEL, 0x0f00,
|
|
PCI_CLASS_BRIDGE_HOST, PCI_ANY_ID, 0, force_disable_hpet},
|
|
- { PCI_VENDOR_ID_INTEL, 0x3e20,
|
|
- PCI_CLASS_BRIDGE_HOST, PCI_ANY_ID, 0, force_disable_hpet},
|
|
- { PCI_VENDOR_ID_INTEL, 0x3ec4,
|
|
- PCI_CLASS_BRIDGE_HOST, PCI_ANY_ID, 0, force_disable_hpet},
|
|
- { PCI_VENDOR_ID_INTEL, 0x8a12,
|
|
- PCI_CLASS_BRIDGE_HOST, PCI_ANY_ID, 0, force_disable_hpet},
|
|
{ PCI_VENDOR_ID_BROADCOM, 0x4331,
|
|
PCI_CLASS_NETWORK_OTHER, PCI_ANY_ID, 0, apple_airport_reset},
|
|
{}
|
|
diff --git a/arch/x86/kernel/hpet.c b/arch/x86/kernel/hpet.c
|
|
index c6f791bc481eb..9834d221e390f 100644
|
|
--- a/arch/x86/kernel/hpet.c
|
|
+++ b/arch/x86/kernel/hpet.c
|
|
@@ -9,6 +9,7 @@
|
|
|
|
#include <asm/hpet.h>
|
|
#include <asm/time.h>
|
|
+#include <asm/mwait.h>
|
|
|
|
#undef pr_fmt
|
|
#define pr_fmt(fmt) "hpet: " fmt
|
|
@@ -806,6 +807,83 @@ static bool __init hpet_counting(void)
|
|
return false;
|
|
}
|
|
|
|
+static bool __init mwait_pc10_supported(void)
|
|
+{
|
|
+ unsigned int eax, ebx, ecx, mwait_substates;
|
|
+
|
|
+ if (boot_cpu_data.x86_vendor != X86_VENDOR_INTEL)
|
|
+ return false;
|
|
+
|
|
+ if (!cpu_feature_enabled(X86_FEATURE_MWAIT))
|
|
+ return false;
|
|
+
|
|
+ if (boot_cpu_data.cpuid_level < CPUID_MWAIT_LEAF)
|
|
+ return false;
|
|
+
|
|
+ cpuid(CPUID_MWAIT_LEAF, &eax, &ebx, &ecx, &mwait_substates);
|
|
+
|
|
+ return (ecx & CPUID5_ECX_EXTENSIONS_SUPPORTED) &&
|
|
+ (ecx & CPUID5_ECX_INTERRUPT_BREAK) &&
|
|
+ (mwait_substates & (0xF << 28));
|
|
+}
|
|
+
|
|
+/*
|
|
+ * Check whether the system supports PC10. If so force disable HPET as that
|
|
+ * stops counting in PC10. This check is overbroad as it does not take any
|
|
+ * of the following into account:
|
|
+ *
|
|
+ * - ACPI tables
|
|
+ * - Enablement of intel_idle
|
|
+ * - Command line arguments which limit intel_idle C-state support
|
|
+ *
|
|
+ * That's perfectly fine. HPET is a piece of hardware designed by committee
|
|
+ * and the only reasons why it is still in use on modern systems is the
|
|
+ * fact that it is impossible to reliably query TSC and CPU frequency via
|
|
+ * CPUID or firmware.
|
|
+ *
|
|
+ * If HPET is functional it is useful for calibrating TSC, but this can be
|
|
+ * done via PMTIMER as well which seems to be the last remaining timer on
|
|
+ * X86/INTEL platforms that has not been completely wreckaged by feature
|
|
+ * creep.
|
|
+ *
|
|
+ * In theory HPET support should be removed altogether, but there are older
|
|
+ * systems out there which depend on it because TSC and APIC timer are
|
|
+ * dysfunctional in deeper C-states.
|
|
+ *
|
|
+ * It's only 20 years now that hardware people have been asked to provide
|
|
+ * reliable and discoverable facilities which can be used for timekeeping
|
|
+ * and per CPU timer interrupts.
|
|
+ *
|
|
+ * The probability that this problem is going to be solved in the
|
|
+ * forseeable future is close to zero, so the kernel has to be cluttered
|
|
+ * with heuristics to keep up with the ever growing amount of hardware and
|
|
+ * firmware trainwrecks. Hopefully some day hardware people will understand
|
|
+ * that the approach of "This can be fixed in software" is not sustainable.
|
|
+ * Hope dies last...
|
|
+ */
|
|
+static bool __init hpet_is_pc10_damaged(void)
|
|
+{
|
|
+ unsigned long long pcfg;
|
|
+
|
|
+ /* Check whether PC10 substates are supported */
|
|
+ if (!mwait_pc10_supported())
|
|
+ return false;
|
|
+
|
|
+ /* Check whether PC10 is enabled in PKG C-state limit */
|
|
+ rdmsrl(MSR_PKG_CST_CONFIG_CONTROL, pcfg);
|
|
+ if ((pcfg & 0xF) < 8)
|
|
+ return false;
|
|
+
|
|
+ if (hpet_force_user) {
|
|
+ pr_warn("HPET force enabled via command line, but dysfunctional in PC10.\n");
|
|
+ return false;
|
|
+ }
|
|
+
|
|
+ pr_info("HPET dysfunctional in PC10. Force disabled.\n");
|
|
+ boot_hpet_disable = true;
|
|
+ return true;
|
|
+}
|
|
+
|
|
/**
|
|
* hpet_enable - Try to setup the HPET timer. Returns 1 on success.
|
|
*/
|
|
@@ -819,6 +897,9 @@ int __init hpet_enable(void)
|
|
if (!is_hpet_capable())
|
|
return 0;
|
|
|
|
+ if (hpet_is_pc10_damaged())
|
|
+ return 0;
|
|
+
|
|
hpet_set_mapping();
|
|
if (!hpet_virt_address)
|
|
return 0;
|
|
diff --git a/arch/x86/platform/olpc/olpc.c b/arch/x86/platform/olpc/olpc.c
|
|
index ee2beda590d0d..1d4a00e767ece 100644
|
|
--- a/arch/x86/platform/olpc/olpc.c
|
|
+++ b/arch/x86/platform/olpc/olpc.c
|
|
@@ -274,7 +274,7 @@ static struct olpc_ec_driver ec_xo1_driver = {
|
|
|
|
static struct olpc_ec_driver ec_xo1_5_driver = {
|
|
.ec_cmd = olpc_xo1_ec_cmd,
|
|
-#ifdef CONFIG_OLPC_XO1_5_SCI
|
|
+#ifdef CONFIG_OLPC_XO15_SCI
|
|
/*
|
|
* XO-1.5 EC wakeups are available when olpc-xo15-sci driver is
|
|
* compiled in
|
|
diff --git a/arch/xtensa/include/asm/kmem_layout.h b/arch/xtensa/include/asm/kmem_layout.h
|
|
index 9c12babc016cd..6fc05cba61a27 100644
|
|
--- a/arch/xtensa/include/asm/kmem_layout.h
|
|
+++ b/arch/xtensa/include/asm/kmem_layout.h
|
|
@@ -11,6 +11,7 @@
|
|
#ifndef _XTENSA_KMEM_LAYOUT_H
|
|
#define _XTENSA_KMEM_LAYOUT_H
|
|
|
|
+#include <asm/core.h>
|
|
#include <asm/types.h>
|
|
|
|
#ifdef CONFIG_MMU
|
|
@@ -65,6 +66,34 @@
|
|
|
|
#endif
|
|
|
|
+/* KIO definition */
|
|
+
|
|
+#if XCHAL_HAVE_PTP_MMU
|
|
+#define XCHAL_KIO_CACHED_VADDR 0xe0000000
|
|
+#define XCHAL_KIO_BYPASS_VADDR 0xf0000000
|
|
+#define XCHAL_KIO_DEFAULT_PADDR 0xf0000000
|
|
+#else
|
|
+#define XCHAL_KIO_BYPASS_VADDR XCHAL_KIO_PADDR
|
|
+#define XCHAL_KIO_DEFAULT_PADDR 0x90000000
|
|
+#endif
|
|
+#define XCHAL_KIO_SIZE 0x10000000
|
|
+
|
|
+#if (!XCHAL_HAVE_PTP_MMU || XCHAL_HAVE_SPANNING_WAY) && defined(CONFIG_USE_OF)
|
|
+#define XCHAL_KIO_PADDR xtensa_get_kio_paddr()
|
|
+#ifndef __ASSEMBLY__
|
|
+extern unsigned long xtensa_kio_paddr;
|
|
+
|
|
+static inline unsigned long xtensa_get_kio_paddr(void)
|
|
+{
|
|
+ return xtensa_kio_paddr;
|
|
+}
|
|
+#endif
|
|
+#else
|
|
+#define XCHAL_KIO_PADDR XCHAL_KIO_DEFAULT_PADDR
|
|
+#endif
|
|
+
|
|
+/* KERNEL_STACK definition */
|
|
+
|
|
#ifndef CONFIG_KASAN
|
|
#define KERNEL_STACK_SHIFT 13
|
|
#else
|
|
diff --git a/arch/xtensa/include/asm/vectors.h b/arch/xtensa/include/asm/vectors.h
|
|
index 79fe3007919eb..4220c6dac44f4 100644
|
|
--- a/arch/xtensa/include/asm/vectors.h
|
|
+++ b/arch/xtensa/include/asm/vectors.h
|
|
@@ -21,50 +21,14 @@
|
|
#include <asm/core.h>
|
|
#include <asm/kmem_layout.h>
|
|
|
|
-#if XCHAL_HAVE_PTP_MMU
|
|
-#define XCHAL_KIO_CACHED_VADDR 0xe0000000
|
|
-#define XCHAL_KIO_BYPASS_VADDR 0xf0000000
|
|
-#define XCHAL_KIO_DEFAULT_PADDR 0xf0000000
|
|
-#else
|
|
-#define XCHAL_KIO_BYPASS_VADDR XCHAL_KIO_PADDR
|
|
-#define XCHAL_KIO_DEFAULT_PADDR 0x90000000
|
|
-#endif
|
|
-#define XCHAL_KIO_SIZE 0x10000000
|
|
-
|
|
-#if (!XCHAL_HAVE_PTP_MMU || XCHAL_HAVE_SPANNING_WAY) && defined(CONFIG_OF)
|
|
-#define XCHAL_KIO_PADDR xtensa_get_kio_paddr()
|
|
-#ifndef __ASSEMBLY__
|
|
-extern unsigned long xtensa_kio_paddr;
|
|
-
|
|
-static inline unsigned long xtensa_get_kio_paddr(void)
|
|
-{
|
|
- return xtensa_kio_paddr;
|
|
-}
|
|
-#endif
|
|
-#else
|
|
-#define XCHAL_KIO_PADDR XCHAL_KIO_DEFAULT_PADDR
|
|
-#endif
|
|
-
|
|
-#if defined(CONFIG_MMU)
|
|
-
|
|
-#if XCHAL_HAVE_PTP_MMU && XCHAL_HAVE_SPANNING_WAY
|
|
-/* Image Virtual Start Address */
|
|
-#define KERNELOFFSET (XCHAL_KSEG_CACHED_VADDR + \
|
|
- CONFIG_KERNEL_LOAD_ADDRESS - \
|
|
+#if defined(CONFIG_MMU) && XCHAL_HAVE_PTP_MMU && XCHAL_HAVE_SPANNING_WAY
|
|
+#define KERNELOFFSET (CONFIG_KERNEL_LOAD_ADDRESS + \
|
|
+ XCHAL_KSEG_CACHED_VADDR - \
|
|
XCHAL_KSEG_PADDR)
|
|
#else
|
|
#define KERNELOFFSET CONFIG_KERNEL_LOAD_ADDRESS
|
|
#endif
|
|
|
|
-#else /* !defined(CONFIG_MMU) */
|
|
- /* MMU Not being used - Virtual == Physical */
|
|
-
|
|
-/* Location of the start of the kernel text, _start */
|
|
-#define KERNELOFFSET CONFIG_KERNEL_LOAD_ADDRESS
|
|
-
|
|
-
|
|
-#endif /* CONFIG_MMU */
|
|
-
|
|
#define RESET_VECTOR1_VADDR (XCHAL_RESET_VECTOR1_VADDR)
|
|
#ifdef CONFIG_VECTORS_OFFSET
|
|
#define VECBASE_VADDR (KERNELOFFSET - CONFIG_VECTORS_OFFSET)
|
|
diff --git a/arch/xtensa/kernel/irq.c b/arch/xtensa/kernel/irq.c
|
|
index a48bf2d10ac2d..80cc9770a8d2d 100644
|
|
--- a/arch/xtensa/kernel/irq.c
|
|
+++ b/arch/xtensa/kernel/irq.c
|
|
@@ -145,7 +145,7 @@ unsigned xtensa_get_ext_irq_no(unsigned irq)
|
|
|
|
void __init init_IRQ(void)
|
|
{
|
|
-#ifdef CONFIG_OF
|
|
+#ifdef CONFIG_USE_OF
|
|
irqchip_init();
|
|
#else
|
|
#ifdef CONFIG_HAVE_SMP
|
|
diff --git a/arch/xtensa/kernel/setup.c b/arch/xtensa/kernel/setup.c
|
|
index d08172138369b..5a25bc2b80521 100644
|
|
--- a/arch/xtensa/kernel/setup.c
|
|
+++ b/arch/xtensa/kernel/setup.c
|
|
@@ -64,7 +64,7 @@ extern unsigned long initrd_end;
|
|
extern int initrd_below_start_ok;
|
|
#endif
|
|
|
|
-#ifdef CONFIG_OF
|
|
+#ifdef CONFIG_USE_OF
|
|
void *dtb_start = __dtb_start;
|
|
#endif
|
|
|
|
@@ -126,7 +126,7 @@ __tagtable(BP_TAG_INITRD, parse_tag_initrd);
|
|
|
|
#endif /* CONFIG_BLK_DEV_INITRD */
|
|
|
|
-#ifdef CONFIG_OF
|
|
+#ifdef CONFIG_USE_OF
|
|
|
|
static int __init parse_tag_fdt(const bp_tag_t *tag)
|
|
{
|
|
@@ -136,7 +136,7 @@ static int __init parse_tag_fdt(const bp_tag_t *tag)
|
|
|
|
__tagtable(BP_TAG_FDT, parse_tag_fdt);
|
|
|
|
-#endif /* CONFIG_OF */
|
|
+#endif /* CONFIG_USE_OF */
|
|
|
|
static int __init parse_tag_cmdline(const bp_tag_t* tag)
|
|
{
|
|
@@ -184,7 +184,7 @@ static int __init parse_bootparam(const bp_tag_t *tag)
|
|
}
|
|
#endif
|
|
|
|
-#ifdef CONFIG_OF
|
|
+#ifdef CONFIG_USE_OF
|
|
|
|
#if !XCHAL_HAVE_PTP_MMU || XCHAL_HAVE_SPANNING_WAY
|
|
unsigned long xtensa_kio_paddr = XCHAL_KIO_DEFAULT_PADDR;
|
|
@@ -233,7 +233,7 @@ void __init early_init_devtree(void *params)
|
|
strlcpy(command_line, boot_command_line, COMMAND_LINE_SIZE);
|
|
}
|
|
|
|
-#endif /* CONFIG_OF */
|
|
+#endif /* CONFIG_USE_OF */
|
|
|
|
/*
|
|
* Initialize architecture. (Early stage)
|
|
@@ -254,7 +254,7 @@ void __init init_arch(bp_tag_t *bp_start)
|
|
if (bp_start)
|
|
parse_bootparam(bp_start);
|
|
|
|
-#ifdef CONFIG_OF
|
|
+#ifdef CONFIG_USE_OF
|
|
early_init_devtree(dtb_start);
|
|
#endif
|
|
|
|
diff --git a/arch/xtensa/mm/mmu.c b/arch/xtensa/mm/mmu.c
|
|
index 03678c4afc39b..bc858a7f98ba4 100644
|
|
--- a/arch/xtensa/mm/mmu.c
|
|
+++ b/arch/xtensa/mm/mmu.c
|
|
@@ -101,7 +101,7 @@ void init_mmu(void)
|
|
|
|
void init_kio(void)
|
|
{
|
|
-#if XCHAL_HAVE_PTP_MMU && XCHAL_HAVE_SPANNING_WAY && defined(CONFIG_OF)
|
|
+#if XCHAL_HAVE_PTP_MMU && XCHAL_HAVE_SPANNING_WAY && defined(CONFIG_USE_OF)
|
|
/*
|
|
* Update the IO area mapping in case xtensa_kio_paddr has changed
|
|
*/
|
|
diff --git a/drivers/bus/ti-sysc.c b/drivers/bus/ti-sysc.c
|
|
index 90053c4a8290d..469ca73de4ce7 100644
|
|
--- a/drivers/bus/ti-sysc.c
|
|
+++ b/drivers/bus/ti-sysc.c
|
|
@@ -1388,6 +1388,9 @@ static const struct sysc_revision_quirk sysc_revision_quirks[] = {
|
|
/* Quirks that need to be set based on detected module */
|
|
SYSC_QUIRK("aess", 0, 0, 0x10, -ENODEV, 0x40000000, 0xffffffff,
|
|
SYSC_MODULE_QUIRK_AESS),
|
|
+ /* Errata i893 handling for dra7 dcan1 and 2 */
|
|
+ SYSC_QUIRK("dcan", 0x4ae3c000, 0x20, -ENODEV, -ENODEV, 0xa3170504, 0xffffffff,
|
|
+ SYSC_QUIRK_CLKDM_NOAUTO),
|
|
SYSC_QUIRK("dcan", 0x48480000, 0x20, -ENODEV, -ENODEV, 0xa3170504, 0xffffffff,
|
|
SYSC_QUIRK_CLKDM_NOAUTO),
|
|
SYSC_QUIRK("dss", 0x4832a000, 0, 0x10, 0x14, 0x00000020, 0xffffffff,
|
|
diff --git a/drivers/gpu/drm/nouveau/nouveau_debugfs.c b/drivers/gpu/drm/nouveau/nouveau_debugfs.c
|
|
index 3b13feca970f7..3c54d61e4fa94 100644
|
|
--- a/drivers/gpu/drm/nouveau/nouveau_debugfs.c
|
|
+++ b/drivers/gpu/drm/nouveau/nouveau_debugfs.c
|
|
@@ -207,6 +207,7 @@ static const struct file_operations nouveau_pstate_fops = {
|
|
.open = nouveau_debugfs_pstate_open,
|
|
.read = seq_read,
|
|
.write = nouveau_debugfs_pstate_set,
|
|
+ .release = single_release,
|
|
};
|
|
|
|
static struct drm_info_list nouveau_debugfs_list[] = {
|
|
diff --git a/drivers/i2c/i2c-core-acpi.c b/drivers/i2c/i2c-core-acpi.c
|
|
index c70983780ae79..fe466ee4c49bf 100644
|
|
--- a/drivers/i2c/i2c-core-acpi.c
|
|
+++ b/drivers/i2c/i2c-core-acpi.c
|
|
@@ -436,6 +436,7 @@ static int i2c_acpi_notify(struct notifier_block *nb, unsigned long value,
|
|
break;
|
|
|
|
i2c_acpi_register_device(adapter, adev, &info);
|
|
+ put_device(&adapter->dev);
|
|
break;
|
|
case ACPI_RECONFIG_DEVICE_REMOVE:
|
|
if (!acpi_device_enumerated(adev))
|
|
diff --git a/drivers/mmc/host/meson-gx-mmc.c b/drivers/mmc/host/meson-gx-mmc.c
|
|
index a3e3b274f0ea3..cdd57ce55b2fa 100644
|
|
--- a/drivers/mmc/host/meson-gx-mmc.c
|
|
+++ b/drivers/mmc/host/meson-gx-mmc.c
|
|
@@ -738,7 +738,7 @@ static void meson_mmc_desc_chain_transfer(struct mmc_host *mmc, u32 cmd_cfg)
|
|
writel(start, host->regs + SD_EMMC_START);
|
|
}
|
|
|
|
-/* local sg copy to buffer version with _to/fromio usage for dram_access_quirk */
|
|
+/* local sg copy for dram_access_quirk */
|
|
static void meson_mmc_copy_buffer(struct meson_host *host, struct mmc_data *data,
|
|
size_t buflen, bool to_buffer)
|
|
{
|
|
@@ -756,21 +756,27 @@ static void meson_mmc_copy_buffer(struct meson_host *host, struct mmc_data *data
|
|
sg_miter_start(&miter, sgl, nents, sg_flags);
|
|
|
|
while ((offset < buflen) && sg_miter_next(&miter)) {
|
|
- unsigned int len;
|
|
+ unsigned int buf_offset = 0;
|
|
+ unsigned int len, left;
|
|
+ u32 *buf = miter.addr;
|
|
|
|
len = min(miter.length, buflen - offset);
|
|
+ left = len;
|
|
|
|
- /* When dram_access_quirk, the bounce buffer is a iomem mapping */
|
|
- if (host->dram_access_quirk) {
|
|
- if (to_buffer)
|
|
- memcpy_toio(host->bounce_iomem_buf + offset, miter.addr, len);
|
|
- else
|
|
- memcpy_fromio(miter.addr, host->bounce_iomem_buf + offset, len);
|
|
+ if (to_buffer) {
|
|
+ do {
|
|
+ writel(*buf++, host->bounce_iomem_buf + offset + buf_offset);
|
|
+
|
|
+ buf_offset += 4;
|
|
+ left -= 4;
|
|
+ } while (left);
|
|
} else {
|
|
- if (to_buffer)
|
|
- memcpy(host->bounce_buf + offset, miter.addr, len);
|
|
- else
|
|
- memcpy(miter.addr, host->bounce_buf + offset, len);
|
|
+ do {
|
|
+ *buf++ = readl(host->bounce_iomem_buf + offset + buf_offset);
|
|
+
|
|
+ buf_offset += 4;
|
|
+ left -= 4;
|
|
+ } while (left);
|
|
}
|
|
|
|
offset += len;
|
|
@@ -822,7 +828,11 @@ static void meson_mmc_start_cmd(struct mmc_host *mmc, struct mmc_command *cmd)
|
|
if (data->flags & MMC_DATA_WRITE) {
|
|
cmd_cfg |= CMD_CFG_DATA_WR;
|
|
WARN_ON(xfer_bytes > host->bounce_buf_size);
|
|
- meson_mmc_copy_buffer(host, data, xfer_bytes, true);
|
|
+ if (host->dram_access_quirk)
|
|
+ meson_mmc_copy_buffer(host, data, xfer_bytes, true);
|
|
+ else
|
|
+ sg_copy_to_buffer(data->sg, data->sg_len,
|
|
+ host->bounce_buf, xfer_bytes);
|
|
dma_wmb();
|
|
}
|
|
|
|
@@ -841,12 +851,43 @@ static void meson_mmc_start_cmd(struct mmc_host *mmc, struct mmc_command *cmd)
|
|
writel(cmd->arg, host->regs + SD_EMMC_CMD_ARG);
|
|
}
|
|
|
|
+static int meson_mmc_validate_dram_access(struct mmc_host *mmc, struct mmc_data *data)
|
|
+{
|
|
+ struct scatterlist *sg;
|
|
+ int i;
|
|
+
|
|
+ /* Reject request if any element offset or size is not 32bit aligned */
|
|
+ for_each_sg(data->sg, sg, data->sg_len, i) {
|
|
+ if (!IS_ALIGNED(sg->offset, sizeof(u32)) ||
|
|
+ !IS_ALIGNED(sg->length, sizeof(u32))) {
|
|
+ dev_err(mmc_dev(mmc), "unaligned sg offset %u len %u\n",
|
|
+ data->sg->offset, data->sg->length);
|
|
+ return -EINVAL;
|
|
+ }
|
|
+ }
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
static void meson_mmc_request(struct mmc_host *mmc, struct mmc_request *mrq)
|
|
{
|
|
struct meson_host *host = mmc_priv(mmc);
|
|
bool needs_pre_post_req = mrq->data &&
|
|
!(mrq->data->host_cookie & SD_EMMC_PRE_REQ_DONE);
|
|
|
|
+ /*
|
|
+ * The memory at the end of the controller used as bounce buffer for
|
|
+ * the dram_access_quirk only accepts 32bit read/write access,
|
|
+ * check the aligment and length of the data before starting the request.
|
|
+ */
|
|
+ if (host->dram_access_quirk && mrq->data) {
|
|
+ mrq->cmd->error = meson_mmc_validate_dram_access(mmc, mrq->data);
|
|
+ if (mrq->cmd->error) {
|
|
+ mmc_request_done(mmc, mrq);
|
|
+ return;
|
|
+ }
|
|
+ }
|
|
+
|
|
if (needs_pre_post_req) {
|
|
meson_mmc_get_transfer_mode(mmc, mrq);
|
|
if (!meson_mmc_desc_chain_mode(mrq->data))
|
|
@@ -991,7 +1032,11 @@ static irqreturn_t meson_mmc_irq_thread(int irq, void *dev_id)
|
|
if (meson_mmc_bounce_buf_read(data)) {
|
|
xfer_bytes = data->blksz * data->blocks;
|
|
WARN_ON(xfer_bytes > host->bounce_buf_size);
|
|
- meson_mmc_copy_buffer(host, data, xfer_bytes, false);
|
|
+ if (host->dram_access_quirk)
|
|
+ meson_mmc_copy_buffer(host, data, xfer_bytes, false);
|
|
+ else
|
|
+ sg_copy_from_buffer(data->sg, data->sg_len,
|
|
+ host->bounce_buf, xfer_bytes);
|
|
}
|
|
|
|
next_cmd = meson_mmc_get_next_command(cmd);
|
|
diff --git a/drivers/net/ethernet/google/gve/gve.h b/drivers/net/ethernet/google/gve/gve.h
|
|
index ebc37e2569221..f19edd4c6c5bb 100644
|
|
--- a/drivers/net/ethernet/google/gve/gve.h
|
|
+++ b/drivers/net/ethernet/google/gve/gve.h
|
|
@@ -391,7 +391,7 @@ struct gve_queue_page_list *gve_assign_rx_qpl(struct gve_priv *priv)
|
|
gve_num_tx_qpls(priv));
|
|
|
|
/* we are out of rx qpls */
|
|
- if (id == priv->qpl_cfg.qpl_map_size)
|
|
+ if (id == gve_num_tx_qpls(priv) + gve_num_rx_qpls(priv))
|
|
return NULL;
|
|
|
|
set_bit(id, priv->qpl_cfg.qpl_id_map);
|
|
diff --git a/drivers/net/ethernet/google/gve/gve_main.c b/drivers/net/ethernet/google/gve/gve_main.c
|
|
index f8dfa7501f65a..5b450c6100add 100644
|
|
--- a/drivers/net/ethernet/google/gve/gve_main.c
|
|
+++ b/drivers/net/ethernet/google/gve/gve_main.c
|
|
@@ -30,6 +30,7 @@ static void gve_get_stats(struct net_device *dev, struct rtnl_link_stats64 *s)
|
|
{
|
|
struct gve_priv *priv = netdev_priv(dev);
|
|
unsigned int start;
|
|
+ u64 packets, bytes;
|
|
int ring;
|
|
|
|
if (priv->rx) {
|
|
@@ -37,10 +38,12 @@ static void gve_get_stats(struct net_device *dev, struct rtnl_link_stats64 *s)
|
|
do {
|
|
start =
|
|
u64_stats_fetch_begin(&priv->rx[ring].statss);
|
|
- s->rx_packets += priv->rx[ring].rpackets;
|
|
- s->rx_bytes += priv->rx[ring].rbytes;
|
|
+ packets = priv->rx[ring].rpackets;
|
|
+ bytes = priv->rx[ring].rbytes;
|
|
} while (u64_stats_fetch_retry(&priv->rx[ring].statss,
|
|
start));
|
|
+ s->rx_packets += packets;
|
|
+ s->rx_bytes += bytes;
|
|
}
|
|
}
|
|
if (priv->tx) {
|
|
@@ -48,10 +51,12 @@ static void gve_get_stats(struct net_device *dev, struct rtnl_link_stats64 *s)
|
|
do {
|
|
start =
|
|
u64_stats_fetch_begin(&priv->tx[ring].statss);
|
|
- s->tx_packets += priv->tx[ring].pkt_done;
|
|
- s->tx_bytes += priv->tx[ring].bytes_done;
|
|
+ packets = priv->tx[ring].pkt_done;
|
|
+ bytes = priv->tx[ring].bytes_done;
|
|
} while (u64_stats_fetch_retry(&priv->tx[ring].statss,
|
|
start));
|
|
+ s->tx_packets += packets;
|
|
+ s->tx_bytes += bytes;
|
|
}
|
|
}
|
|
}
|
|
diff --git a/drivers/net/ethernet/intel/i40e/i40e_main.c b/drivers/net/ethernet/intel/i40e/i40e_main.c
|
|
index 21ab7d2caddf5..917be10a5cf5c 100644
|
|
--- a/drivers/net/ethernet/intel/i40e/i40e_main.c
|
|
+++ b/drivers/net/ethernet/intel/i40e/i40e_main.c
|
|
@@ -4817,7 +4817,8 @@ static void i40e_clear_interrupt_scheme(struct i40e_pf *pf)
|
|
{
|
|
int i;
|
|
|
|
- i40e_free_misc_vector(pf);
|
|
+ if (test_bit(__I40E_MISC_IRQ_REQUESTED, pf->state))
|
|
+ i40e_free_misc_vector(pf);
|
|
|
|
i40e_put_lump(pf->irq_pile, pf->iwarp_base_vector,
|
|
I40E_IWARP_IRQ_PILE_ID);
|
|
@@ -9616,7 +9617,7 @@ static int i40e_get_capabilities(struct i40e_pf *pf,
|
|
if (pf->hw.aq.asq_last_status == I40E_AQ_RC_ENOMEM) {
|
|
/* retry with a larger buffer */
|
|
buf_len = data_size;
|
|
- } else if (pf->hw.aq.asq_last_status != I40E_AQ_RC_OK) {
|
|
+ } else if (pf->hw.aq.asq_last_status != I40E_AQ_RC_OK || err) {
|
|
dev_info(&pf->pdev->dev,
|
|
"capability discovery failed, err %s aq_err %s\n",
|
|
i40e_stat_str(&pf->hw, err),
|
|
diff --git a/drivers/net/phy/mdio_bus.c b/drivers/net/phy/mdio_bus.c
|
|
index 5bf06eac04ba3..bec73f0640d03 100644
|
|
--- a/drivers/net/phy/mdio_bus.c
|
|
+++ b/drivers/net/phy/mdio_bus.c
|
|
@@ -385,6 +385,13 @@ int __mdiobus_register(struct mii_bus *bus, struct module *owner)
|
|
bus->dev.groups = NULL;
|
|
dev_set_name(&bus->dev, "%s", bus->id);
|
|
|
|
+ /* We need to set state to MDIOBUS_UNREGISTERED to correctly release
|
|
+ * the device in mdiobus_free()
|
|
+ *
|
|
+ * State will be updated later in this function in case of success
|
|
+ */
|
|
+ bus->state = MDIOBUS_UNREGISTERED;
|
|
+
|
|
err = device_register(&bus->dev);
|
|
if (err) {
|
|
pr_err("mii_bus %s failed to register\n", bus->id);
|
|
diff --git a/drivers/net/phy/sfp.c b/drivers/net/phy/sfp.c
|
|
index 27b67f12ec455..5657c604602e8 100644
|
|
--- a/drivers/net/phy/sfp.c
|
|
+++ b/drivers/net/phy/sfp.c
|
|
@@ -115,7 +115,7 @@ static const char * const sm_state_strings[] = {
|
|
[SFP_S_LINK_UP] = "link_up",
|
|
[SFP_S_TX_FAULT] = "tx_fault",
|
|
[SFP_S_REINIT] = "reinit",
|
|
- [SFP_S_TX_DISABLE] = "rx_disable",
|
|
+ [SFP_S_TX_DISABLE] = "tx_disable",
|
|
};
|
|
|
|
static const char *sm_state_to_str(unsigned short sm_state)
|
|
diff --git a/drivers/ptp/ptp_pch.c b/drivers/ptp/ptp_pch.c
|
|
index dcd6e00c80467..a50656632df93 100644
|
|
--- a/drivers/ptp/ptp_pch.c
|
|
+++ b/drivers/ptp/ptp_pch.c
|
|
@@ -683,6 +683,7 @@ static const struct pci_device_id pch_ieee1588_pcidev_id[] = {
|
|
},
|
|
{0}
|
|
};
|
|
+MODULE_DEVICE_TABLE(pci, pch_ieee1588_pcidev_id);
|
|
|
|
static struct pci_driver pch_driver = {
|
|
.name = KBUILD_MODNAME,
|
|
diff --git a/drivers/soc/qcom/mdt_loader.c b/drivers/soc/qcom/mdt_loader.c
|
|
index eba7f76f9d61a..6034cd8992b0e 100644
|
|
--- a/drivers/soc/qcom/mdt_loader.c
|
|
+++ b/drivers/soc/qcom/mdt_loader.c
|
|
@@ -98,7 +98,7 @@ void *qcom_mdt_read_metadata(const struct firmware *fw, size_t *data_len)
|
|
if (ehdr->e_phnum < 2)
|
|
return ERR_PTR(-EINVAL);
|
|
|
|
- if (phdrs[0].p_type == PT_LOAD || phdrs[1].p_type == PT_LOAD)
|
|
+ if (phdrs[0].p_type == PT_LOAD)
|
|
return ERR_PTR(-EINVAL);
|
|
|
|
if ((phdrs[1].p_flags & QCOM_MDT_TYPE_MASK) != QCOM_MDT_TYPE_HASH)
|
|
diff --git a/drivers/soc/qcom/socinfo.c b/drivers/soc/qcom/socinfo.c
|
|
index 176696f8f38d1..3303bcaf67154 100644
|
|
--- a/drivers/soc/qcom/socinfo.c
|
|
+++ b/drivers/soc/qcom/socinfo.c
|
|
@@ -447,7 +447,7 @@ static int qcom_socinfo_probe(struct platform_device *pdev)
|
|
/* Feed the soc specific unique data into entropy pool */
|
|
add_device_randomness(info, item_size);
|
|
|
|
- platform_set_drvdata(pdev, qs->soc_dev);
|
|
+ platform_set_drvdata(pdev, qs);
|
|
|
|
return 0;
|
|
}
|
|
diff --git a/drivers/usb/class/cdc-acm.c b/drivers/usb/class/cdc-acm.c
|
|
index e139cda35f639..5dc8827ede7e8 100644
|
|
--- a/drivers/usb/class/cdc-acm.c
|
|
+++ b/drivers/usb/class/cdc-acm.c
|
|
@@ -339,6 +339,9 @@ static void acm_process_notification(struct acm *acm, unsigned char *buf)
|
|
acm->iocount.overrun++;
|
|
spin_unlock_irqrestore(&acm->read_lock, flags);
|
|
|
|
+ if (newctrl & ACM_CTRL_BRK)
|
|
+ tty_flip_buffer_push(&acm->port);
|
|
+
|
|
if (difference)
|
|
wake_up_all(&acm->wioctl);
|
|
|
|
@@ -474,11 +477,16 @@ static int acm_submit_read_urbs(struct acm *acm, gfp_t mem_flags)
|
|
|
|
static void acm_process_read_urb(struct acm *acm, struct urb *urb)
|
|
{
|
|
+ unsigned long flags;
|
|
+
|
|
if (!urb->actual_length)
|
|
return;
|
|
|
|
+ spin_lock_irqsave(&acm->read_lock, flags);
|
|
tty_insert_flip_string(&acm->port, urb->transfer_buffer,
|
|
urb->actual_length);
|
|
+ spin_unlock_irqrestore(&acm->read_lock, flags);
|
|
+
|
|
tty_flip_buffer_push(&acm->port);
|
|
}
|
|
|
|
diff --git a/drivers/usb/common/Kconfig b/drivers/usb/common/Kconfig
|
|
index d611477aae414..196f4a3975871 100644
|
|
--- a/drivers/usb/common/Kconfig
|
|
+++ b/drivers/usb/common/Kconfig
|
|
@@ -6,8 +6,7 @@ config USB_COMMON
|
|
|
|
config USB_LED_TRIG
|
|
bool "USB LED Triggers"
|
|
- depends on LEDS_CLASS && LEDS_TRIGGERS
|
|
- select USB_COMMON
|
|
+ depends on LEDS_CLASS && USB_COMMON && LEDS_TRIGGERS
|
|
help
|
|
This option adds LED triggers for USB host and/or gadget activity.
|
|
|
|
diff --git a/drivers/usb/typec/tcpm/tcpm.c b/drivers/usb/typec/tcpm/tcpm.c
|
|
index b40db48f8874d..89391939630bd 100644
|
|
--- a/drivers/usb/typec/tcpm/tcpm.c
|
|
+++ b/drivers/usb/typec/tcpm/tcpm.c
|
|
@@ -3679,6 +3679,7 @@ static void _tcpm_cc_change(struct tcpm_port *port, enum typec_cc_status cc1,
|
|
tcpm_set_state(port, SRC_ATTACH_WAIT, 0);
|
|
break;
|
|
case SRC_ATTACHED:
|
|
+ case SRC_STARTUP:
|
|
case SRC_SEND_CAPABILITIES:
|
|
case SRC_READY:
|
|
if (tcpm_port_is_disconnected(port) ||
|
|
diff --git a/drivers/video/fbdev/gbefb.c b/drivers/video/fbdev/gbefb.c
|
|
index b9f6a82a04953..6fdc6ab3ceb87 100644
|
|
--- a/drivers/video/fbdev/gbefb.c
|
|
+++ b/drivers/video/fbdev/gbefb.c
|
|
@@ -1269,7 +1269,7 @@ static struct platform_device *gbefb_device;
|
|
static int __init gbefb_init(void)
|
|
{
|
|
int ret = platform_driver_register(&gbefb_driver);
|
|
- if (!ret) {
|
|
+ if (IS_ENABLED(CONFIG_SGI_IP32) && !ret) {
|
|
gbefb_device = platform_device_alloc("gbefb", 0);
|
|
if (gbefb_device) {
|
|
ret = platform_device_add(gbefb_device);
|
|
diff --git a/drivers/xen/balloon.c b/drivers/xen/balloon.c
|
|
index be31c296eed4c..07f362c63ae90 100644
|
|
--- a/drivers/xen/balloon.c
|
|
+++ b/drivers/xen/balloon.c
|
|
@@ -508,12 +508,12 @@ static enum bp_state decrease_reservation(unsigned long nr_pages, gfp_t gfp)
|
|
}
|
|
|
|
/*
|
|
- * Stop waiting if either state is not BP_EAGAIN and ballooning action is
|
|
- * needed, or if the credit has changed while state is BP_EAGAIN.
|
|
+ * Stop waiting if either state is BP_DONE and ballooning action is
|
|
+ * needed, or if the credit has changed while state is not BP_DONE.
|
|
*/
|
|
static bool balloon_thread_cond(enum bp_state state, long credit)
|
|
{
|
|
- if (state != BP_EAGAIN)
|
|
+ if (state == BP_DONE)
|
|
credit = 0;
|
|
|
|
return current_credit() != credit || kthread_should_stop();
|
|
@@ -533,10 +533,19 @@ static int balloon_thread(void *unused)
|
|
|
|
set_freezable();
|
|
for (;;) {
|
|
- if (state == BP_EAGAIN)
|
|
- timeout = balloon_stats.schedule_delay * HZ;
|
|
- else
|
|
+ switch (state) {
|
|
+ case BP_DONE:
|
|
+ case BP_ECANCELED:
|
|
timeout = 3600 * HZ;
|
|
+ break;
|
|
+ case BP_EAGAIN:
|
|
+ timeout = balloon_stats.schedule_delay * HZ;
|
|
+ break;
|
|
+ case BP_WAIT:
|
|
+ timeout = HZ;
|
|
+ break;
|
|
+ }
|
|
+
|
|
credit = current_credit();
|
|
|
|
wait_event_freezable_timeout(balloon_thread_wq,
|
|
diff --git a/drivers/xen/privcmd.c b/drivers/xen/privcmd.c
|
|
index 9c9422e9fac4d..d4ff944cd16e1 100644
|
|
--- a/drivers/xen/privcmd.c
|
|
+++ b/drivers/xen/privcmd.c
|
|
@@ -810,11 +810,12 @@ static long privcmd_ioctl_mmap_resource(struct file *file,
|
|
unsigned int domid =
|
|
(xdata.flags & XENMEM_rsrc_acq_caller_owned) ?
|
|
DOMID_SELF : kdata.dom;
|
|
- int num;
|
|
+ int num, *errs = (int *)pfns;
|
|
|
|
+ BUILD_BUG_ON(sizeof(*errs) > sizeof(*pfns));
|
|
num = xen_remap_domain_mfn_array(vma,
|
|
kdata.addr & PAGE_MASK,
|
|
- pfns, kdata.num, (int *)pfns,
|
|
+ pfns, kdata.num, errs,
|
|
vma->vm_page_prot,
|
|
domid,
|
|
vma->vm_private_data);
|
|
@@ -824,7 +825,7 @@ static long privcmd_ioctl_mmap_resource(struct file *file,
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < num; i++) {
|
|
- rc = pfns[i];
|
|
+ rc = errs[i];
|
|
if (rc < 0)
|
|
break;
|
|
}
|
|
diff --git a/fs/nfsd/nfs4xdr.c b/fs/nfsd/nfs4xdr.c
|
|
index d6f244559e759..e61d9c4359573 100644
|
|
--- a/fs/nfsd/nfs4xdr.c
|
|
+++ b/fs/nfsd/nfs4xdr.c
|
|
@@ -3131,15 +3131,18 @@ nfsd4_encode_dirent(void *ccdv, const char *name, int namlen,
|
|
goto fail;
|
|
cd->rd_maxcount -= entry_bytes;
|
|
/*
|
|
- * RFC 3530 14.2.24 describes rd_dircount as only a "hint", so
|
|
- * let's always let through the first entry, at least:
|
|
+ * RFC 3530 14.2.24 describes rd_dircount as only a "hint", and
|
|
+ * notes that it could be zero. If it is zero, then the server
|
|
+ * should enforce only the rd_maxcount value.
|
|
*/
|
|
- if (!cd->rd_dircount)
|
|
- goto fail;
|
|
- name_and_cookie = 4 + 4 * XDR_QUADLEN(namlen) + 8;
|
|
- if (name_and_cookie > cd->rd_dircount && cd->cookie_offset)
|
|
- goto fail;
|
|
- cd->rd_dircount -= min(cd->rd_dircount, name_and_cookie);
|
|
+ if (cd->rd_dircount) {
|
|
+ name_and_cookie = 4 + 4 * XDR_QUADLEN(namlen) + 8;
|
|
+ if (name_and_cookie > cd->rd_dircount && cd->cookie_offset)
|
|
+ goto fail;
|
|
+ cd->rd_dircount -= min(cd->rd_dircount, name_and_cookie);
|
|
+ if (!cd->rd_dircount)
|
|
+ cd->rd_maxcount = 0;
|
|
+ }
|
|
|
|
cd->cookie_offset = cookie_offset;
|
|
skip_entry:
|
|
diff --git a/fs/nfsd/nfsctl.c b/fs/nfsd/nfsctl.c
|
|
index 7f39d6091dfa0..8e03d6c25097a 100644
|
|
--- a/fs/nfsd/nfsctl.c
|
|
+++ b/fs/nfsd/nfsctl.c
|
|
@@ -1549,7 +1549,7 @@ static int __init init_nfsd(void)
|
|
goto out_free_all;
|
|
return 0;
|
|
out_free_all:
|
|
- unregister_pernet_subsys(&nfsd_net_ops);
|
|
+ unregister_filesystem(&nfsd_fs_type);
|
|
out_free_exports:
|
|
remove_proc_entry("fs/nfs/exports", NULL);
|
|
remove_proc_entry("fs/nfs", NULL);
|
|
diff --git a/fs/overlayfs/dir.c b/fs/overlayfs/dir.c
|
|
index 073be36b0686c..876de87f604cd 100644
|
|
--- a/fs/overlayfs/dir.c
|
|
+++ b/fs/overlayfs/dir.c
|
|
@@ -1162,9 +1162,13 @@ static int ovl_rename(struct inode *olddir, struct dentry *old,
|
|
goto out_dput;
|
|
}
|
|
} else {
|
|
- if (!d_is_negative(newdentry) &&
|
|
- (!new_opaque || !ovl_is_whiteout(newdentry)))
|
|
- goto out_dput;
|
|
+ if (!d_is_negative(newdentry)) {
|
|
+ if (!new_opaque || !ovl_is_whiteout(newdentry))
|
|
+ goto out_dput;
|
|
+ } else {
|
|
+ if (flags & RENAME_EXCHANGE)
|
|
+ goto out_dput;
|
|
+ }
|
|
}
|
|
|
|
if (olddentry == trap)
|
|
diff --git a/kernel/bpf/stackmap.c b/kernel/bpf/stackmap.c
|
|
index fba2ade28fb3a..49c7a09d688d7 100644
|
|
--- a/kernel/bpf/stackmap.c
|
|
+++ b/kernel/bpf/stackmap.c
|
|
@@ -60,7 +60,8 @@ static inline int stack_map_data_size(struct bpf_map *map)
|
|
|
|
static int prealloc_elems_and_freelist(struct bpf_stack_map *smap)
|
|
{
|
|
- u32 elem_size = sizeof(struct stack_map_bucket) + smap->map.value_size;
|
|
+ u64 elem_size = sizeof(struct stack_map_bucket) +
|
|
+ (u64)smap->map.value_size;
|
|
int err;
|
|
|
|
smap->elems = bpf_map_area_alloc(elem_size * smap->map.max_entries,
|
|
diff --git a/net/bridge/br_netlink.c b/net/bridge/br_netlink.c
|
|
index 8a664148f57aa..cbcbc19efcb34 100644
|
|
--- a/net/bridge/br_netlink.c
|
|
+++ b/net/bridge/br_netlink.c
|
|
@@ -1536,7 +1536,7 @@ static size_t br_get_linkxstats_size(const struct net_device *dev, int attr)
|
|
}
|
|
|
|
return numvls * nla_total_size(sizeof(struct bridge_vlan_xstats)) +
|
|
- nla_total_size(sizeof(struct br_mcast_stats)) +
|
|
+ nla_total_size_64bit(sizeof(struct br_mcast_stats)) +
|
|
nla_total_size(0);
|
|
}
|
|
|
|
diff --git a/net/core/rtnetlink.c b/net/core/rtnetlink.c
|
|
index 6fbc9cb09dc0e..a53b101ce41ae 100644
|
|
--- a/net/core/rtnetlink.c
|
|
+++ b/net/core/rtnetlink.c
|
|
@@ -4950,7 +4950,7 @@ nla_put_failure:
|
|
static size_t if_nlmsg_stats_size(const struct net_device *dev,
|
|
u32 filter_mask)
|
|
{
|
|
- size_t size = 0;
|
|
+ size_t size = NLMSG_ALIGN(sizeof(struct if_stats_msg));
|
|
|
|
if (stats_attr_valid(filter_mask, IFLA_STATS_LINK_64, 0))
|
|
size += nla_total_size_64bit(sizeof(struct rtnl_link_stats64));
|
|
diff --git a/net/ipv4/inet_hashtables.c b/net/ipv4/inet_hashtables.c
|
|
index 006a34b185378..72fdf1fcbcaa9 100644
|
|
--- a/net/ipv4/inet_hashtables.c
|
|
+++ b/net/ipv4/inet_hashtables.c
|
|
@@ -239,8 +239,10 @@ static inline int compute_score(struct sock *sk, struct net *net,
|
|
|
|
if (!inet_sk_bound_dev_eq(net, sk->sk_bound_dev_if, dif, sdif))
|
|
return -1;
|
|
+ score = sk->sk_bound_dev_if ? 2 : 1;
|
|
|
|
- score = sk->sk_family == PF_INET ? 2 : 1;
|
|
+ if (sk->sk_family == PF_INET)
|
|
+ score++;
|
|
if (READ_ONCE(sk->sk_incoming_cpu) == raw_smp_processor_id())
|
|
score++;
|
|
}
|
|
diff --git a/net/ipv4/udp.c b/net/ipv4/udp.c
|
|
index de04d99418850..fdbd56ee1300c 100644
|
|
--- a/net/ipv4/udp.c
|
|
+++ b/net/ipv4/udp.c
|
|
@@ -386,7 +386,8 @@ static int compute_score(struct sock *sk, struct net *net,
|
|
dif, sdif);
|
|
if (!dev_match)
|
|
return -1;
|
|
- score += 4;
|
|
+ if (sk->sk_bound_dev_if)
|
|
+ score += 4;
|
|
|
|
if (READ_ONCE(sk->sk_incoming_cpu) == raw_smp_processor_id())
|
|
score++;
|
|
diff --git a/net/ipv6/inet6_hashtables.c b/net/ipv6/inet6_hashtables.c
|
|
index fbe9d4295eac3..ab12e00f6bfff 100644
|
|
--- a/net/ipv6/inet6_hashtables.c
|
|
+++ b/net/ipv6/inet6_hashtables.c
|
|
@@ -104,7 +104,7 @@ static inline int compute_score(struct sock *sk, struct net *net,
|
|
if (!inet_sk_bound_dev_eq(net, sk->sk_bound_dev_if, dif, sdif))
|
|
return -1;
|
|
|
|
- score = 1;
|
|
+ score = sk->sk_bound_dev_if ? 2 : 1;
|
|
if (READ_ONCE(sk->sk_incoming_cpu) == raw_smp_processor_id())
|
|
score++;
|
|
}
|
|
diff --git a/net/ipv6/udp.c b/net/ipv6/udp.c
|
|
index 5b8266f3e47f0..0f57c682afdd8 100644
|
|
--- a/net/ipv6/udp.c
|
|
+++ b/net/ipv6/udp.c
|
|
@@ -133,7 +133,8 @@ static int compute_score(struct sock *sk, struct net *net,
|
|
dev_match = udp_sk_bound_dev_eq(net, sk->sk_bound_dev_if, dif, sdif);
|
|
if (!dev_match)
|
|
return -1;
|
|
- score++;
|
|
+ if (sk->sk_bound_dev_if)
|
|
+ score++;
|
|
|
|
if (READ_ONCE(sk->sk_incoming_cpu) == raw_smp_processor_id())
|
|
score++;
|
|
diff --git a/net/netlink/af_netlink.c b/net/netlink/af_netlink.c
|
|
index acc76a738cfd8..cb35680db9b29 100644
|
|
--- a/net/netlink/af_netlink.c
|
|
+++ b/net/netlink/af_netlink.c
|
|
@@ -585,7 +585,10 @@ static int netlink_insert(struct sock *sk, u32 portid)
|
|
|
|
/* We need to ensure that the socket is hashed and visible. */
|
|
smp_wmb();
|
|
- nlk_sk(sk)->bound = portid;
|
|
+ /* Paired with lockless reads from netlink_bind(),
|
|
+ * netlink_connect() and netlink_sendmsg().
|
|
+ */
|
|
+ WRITE_ONCE(nlk_sk(sk)->bound, portid);
|
|
|
|
err:
|
|
release_sock(sk);
|
|
@@ -1003,7 +1006,8 @@ static int netlink_bind(struct socket *sock, struct sockaddr *addr,
|
|
if (nlk->ngroups < BITS_PER_LONG)
|
|
groups &= (1UL << nlk->ngroups) - 1;
|
|
|
|
- bound = nlk->bound;
|
|
+ /* Paired with WRITE_ONCE() in netlink_insert() */
|
|
+ bound = READ_ONCE(nlk->bound);
|
|
if (bound) {
|
|
/* Ensure nlk->portid is up-to-date. */
|
|
smp_rmb();
|
|
@@ -1089,8 +1093,9 @@ static int netlink_connect(struct socket *sock, struct sockaddr *addr,
|
|
|
|
/* No need for barriers here as we return to user-space without
|
|
* using any of the bound attributes.
|
|
+ * Paired with WRITE_ONCE() in netlink_insert().
|
|
*/
|
|
- if (!nlk->bound)
|
|
+ if (!READ_ONCE(nlk->bound))
|
|
err = netlink_autobind(sock);
|
|
|
|
if (err == 0) {
|
|
@@ -1879,7 +1884,8 @@ static int netlink_sendmsg(struct socket *sock, struct msghdr *msg, size_t len)
|
|
dst_group = nlk->dst_group;
|
|
}
|
|
|
|
- if (!nlk->bound) {
|
|
+ /* Paired with WRITE_ONCE() in netlink_insert() */
|
|
+ if (!READ_ONCE(nlk->bound)) {
|
|
err = netlink_autobind(sock);
|
|
if (err)
|
|
goto out;
|
|
diff --git a/net/sched/sch_fifo.c b/net/sched/sch_fifo.c
|
|
index 37c8aa75d70c5..56f4c1621e444 100644
|
|
--- a/net/sched/sch_fifo.c
|
|
+++ b/net/sched/sch_fifo.c
|
|
@@ -148,6 +148,9 @@ int fifo_set_limit(struct Qdisc *q, unsigned int limit)
|
|
if (strncmp(q->ops->id + 1, "fifo", 4) != 0)
|
|
return 0;
|
|
|
|
+ if (!q->ops->change)
|
|
+ return 0;
|
|
+
|
|
nla = kmalloc(nla_attr_size(sizeof(struct tc_fifo_qopt)), GFP_KERNEL);
|
|
if (nla) {
|
|
nla->nla_type = RTM_NEWQDISC;
|
|
diff --git a/net/sched/sch_taprio.c b/net/sched/sch_taprio.c
|
|
index da9ed0613eb7b..e14a66ce4884d 100644
|
|
--- a/net/sched/sch_taprio.c
|
|
+++ b/net/sched/sch_taprio.c
|
|
@@ -1630,6 +1630,10 @@ static void taprio_destroy(struct Qdisc *sch)
|
|
list_del(&q->taprio_list);
|
|
spin_unlock(&taprio_list_lock);
|
|
|
|
+ /* Note that taprio_reset() might not be called if an error
|
|
+ * happens in qdisc_create(), after taprio_init() has been called.
|
|
+ */
|
|
+ hrtimer_cancel(&q->advance_timer);
|
|
|
|
taprio_disable_offload(dev, q, NULL);
|
|
|