diff options
author | Mike Pagano <mpagano@gentoo.org> | 2016-07-11 15:59:39 -0400 |
---|---|---|
committer | Mike Pagano <mpagano@gentoo.org> | 2016-07-11 15:59:39 -0400 |
commit | c9aae51312f49294f6c20c309824b39d38803b9a (patch) | |
tree | 21f7ad53972be520146bb0db8de955da2f47de7e /1014_linux-4.4.15.patch | |
parent | Select SYSVIPC when GENTOO_LINUX_PORTAGE is selected. Dependency of IPC_NS. S... (diff) | |
download | linux-patches-c9aae51312f49294f6c20c309824b39d38803b9a.tar.gz linux-patches-c9aae51312f49294f6c20c309824b39d38803b9a.tar.bz2 linux-patches-c9aae51312f49294f6c20c309824b39d38803b9a.zip |
Linux patch 4.4.154.4-17
Diffstat (limited to '1014_linux-4.4.15.patch')
-rw-r--r-- | 1014_linux-4.4.15.patch | 1204 |
1 files changed, 1204 insertions, 0 deletions
diff --git a/1014_linux-4.4.15.patch b/1014_linux-4.4.15.patch new file mode 100644 index 00000000..e0685ae3 --- /dev/null +++ b/1014_linux-4.4.15.patch @@ -0,0 +1,1204 @@ +diff --git a/Makefile b/Makefile +index fadbb9d73c6d..979088079338 100644 +--- a/Makefile ++++ b/Makefile +@@ -1,6 +1,6 @@ + VERSION = 4 + PATCHLEVEL = 4 +-SUBLEVEL = 14 ++SUBLEVEL = 15 + EXTRAVERSION = + NAME = Blurry Fish Butt + +diff --git a/crypto/crypto_user.c b/crypto/crypto_user.c +index 43fe85f20d57..7097a3395b25 100644 +--- a/crypto/crypto_user.c ++++ b/crypto/crypto_user.c +@@ -455,6 +455,7 @@ static const int crypto_msg_min[CRYPTO_NR_MSGTYPES] = { + [CRYPTO_MSG_NEWALG - CRYPTO_MSG_BASE] = MSGSIZE(crypto_user_alg), + [CRYPTO_MSG_DELALG - CRYPTO_MSG_BASE] = MSGSIZE(crypto_user_alg), + [CRYPTO_MSG_UPDATEALG - CRYPTO_MSG_BASE] = MSGSIZE(crypto_user_alg), ++ [CRYPTO_MSG_GETALG - CRYPTO_MSG_BASE] = MSGSIZE(crypto_user_alg), + [CRYPTO_MSG_DELRNG - CRYPTO_MSG_BASE] = 0, + }; + +diff --git a/drivers/crypto/ux500/hash/hash_core.c b/drivers/crypto/ux500/hash/hash_core.c +index 66b1c3313e2e..cd4398498495 100644 +--- a/drivers/crypto/ux500/hash/hash_core.c ++++ b/drivers/crypto/ux500/hash/hash_core.c +@@ -797,7 +797,7 @@ static int hash_process_data(struct hash_device_data *device_data, + &device_data->state); + memmove(req_ctx->state.buffer, + device_data->state.buffer, +- HASH_BLOCK_SIZE / sizeof(u32)); ++ HASH_BLOCK_SIZE); + if (ret) { + dev_err(device_data->dev, + "%s: hash_resume_state() failed!\n", +@@ -848,7 +848,7 @@ static int hash_process_data(struct hash_device_data *device_data, + + memmove(device_data->state.buffer, + req_ctx->state.buffer, +- HASH_BLOCK_SIZE / sizeof(u32)); ++ HASH_BLOCK_SIZE); + if (ret) { + dev_err(device_data->dev, "%s: hash_save_state() failed!\n", + __func__); +diff --git a/drivers/crypto/vmx/aes_cbc.c b/drivers/crypto/vmx/aes_cbc.c +index 0b8fe2ec5315..f3801b983f42 100644 +--- a/drivers/crypto/vmx/aes_cbc.c ++++ b/drivers/crypto/vmx/aes_cbc.c +@@ -182,7 +182,7 @@ struct crypto_alg p8_aes_cbc_alg = { + .cra_name = "cbc(aes)", + .cra_driver_name = "p8_aes_cbc", + .cra_module = THIS_MODULE, +- .cra_priority = 1000, ++ .cra_priority = 2000, + .cra_type = &crypto_blkcipher_type, + .cra_flags = CRYPTO_ALG_TYPE_BLKCIPHER | CRYPTO_ALG_NEED_FALLBACK, + .cra_alignmask = 0, +diff --git a/drivers/crypto/vmx/aes_ctr.c b/drivers/crypto/vmx/aes_ctr.c +index ee1306cd8f59..404a1b69a3ab 100644 +--- a/drivers/crypto/vmx/aes_ctr.c ++++ b/drivers/crypto/vmx/aes_ctr.c +@@ -166,7 +166,7 @@ struct crypto_alg p8_aes_ctr_alg = { + .cra_name = "ctr(aes)", + .cra_driver_name = "p8_aes_ctr", + .cra_module = THIS_MODULE, +- .cra_priority = 1000, ++ .cra_priority = 2000, + .cra_type = &crypto_blkcipher_type, + .cra_flags = CRYPTO_ALG_TYPE_BLKCIPHER | CRYPTO_ALG_NEED_FALLBACK, + .cra_alignmask = 0, +diff --git a/drivers/net/ethernet/atheros/alx/main.c b/drivers/net/ethernet/atheros/alx/main.c +index bd377a6b067d..df54475d163b 100644 +--- a/drivers/net/ethernet/atheros/alx/main.c ++++ b/drivers/net/ethernet/atheros/alx/main.c +@@ -86,9 +86,14 @@ static int alx_refill_rx_ring(struct alx_priv *alx, gfp_t gfp) + while (!cur_buf->skb && next != rxq->read_idx) { + struct alx_rfd *rfd = &rxq->rfd[cur]; + +- skb = __netdev_alloc_skb(alx->dev, alx->rxbuf_size, gfp); ++ skb = __netdev_alloc_skb(alx->dev, alx->rxbuf_size + 64, gfp); + if (!skb) + break; ++ ++ /* Workround for the HW RX DMA overflow issue */ ++ if (((unsigned long)skb->data & 0xfff) == 0xfc0) ++ skb_reserve(skb, 64); ++ + dma = dma_map_single(&alx->hw.pdev->dev, + skb->data, alx->rxbuf_size, + DMA_FROM_DEVICE); +diff --git a/drivers/net/ethernet/cadence/macb.c b/drivers/net/ethernet/cadence/macb.c +index 169059c92f80..8d54e7b41bbf 100644 +--- a/drivers/net/ethernet/cadence/macb.c ++++ b/drivers/net/ethernet/cadence/macb.c +@@ -2405,9 +2405,9 @@ static int macb_init(struct platform_device *pdev) + if (bp->phy_interface == PHY_INTERFACE_MODE_RGMII) + val = GEM_BIT(RGMII); + else if (bp->phy_interface == PHY_INTERFACE_MODE_RMII && +- (bp->caps & MACB_CAPS_USRIO_DEFAULT_IS_MII)) ++ (bp->caps & MACB_CAPS_USRIO_DEFAULT_IS_MII_GMII)) + val = MACB_BIT(RMII); +- else if (!(bp->caps & MACB_CAPS_USRIO_DEFAULT_IS_MII)) ++ else if (!(bp->caps & MACB_CAPS_USRIO_DEFAULT_IS_MII_GMII)) + val = MACB_BIT(MII); + + if (bp->caps & MACB_CAPS_USRIO_HAS_CLKEN) +@@ -2738,7 +2738,7 @@ static int at91ether_init(struct platform_device *pdev) + } + + static const struct macb_config at91sam9260_config = { +- .caps = MACB_CAPS_USRIO_HAS_CLKEN | MACB_CAPS_USRIO_DEFAULT_IS_MII, ++ .caps = MACB_CAPS_USRIO_HAS_CLKEN | MACB_CAPS_USRIO_DEFAULT_IS_MII_GMII, + .clk_init = macb_clk_init, + .init = macb_init, + }; +@@ -2751,21 +2751,22 @@ static const struct macb_config pc302gem_config = { + }; + + static const struct macb_config sama5d2_config = { +- .caps = 0, ++ .caps = MACB_CAPS_USRIO_DEFAULT_IS_MII_GMII, + .dma_burst_length = 16, + .clk_init = macb_clk_init, + .init = macb_init, + }; + + static const struct macb_config sama5d3_config = { +- .caps = MACB_CAPS_SG_DISABLED | MACB_CAPS_GIGABIT_MODE_AVAILABLE, ++ .caps = MACB_CAPS_SG_DISABLED | MACB_CAPS_GIGABIT_MODE_AVAILABLE ++ | MACB_CAPS_USRIO_DEFAULT_IS_MII_GMII, + .dma_burst_length = 16, + .clk_init = macb_clk_init, + .init = macb_init, + }; + + static const struct macb_config sama5d4_config = { +- .caps = 0, ++ .caps = MACB_CAPS_USRIO_DEFAULT_IS_MII_GMII, + .dma_burst_length = 4, + .clk_init = macb_clk_init, + .init = macb_init, +diff --git a/drivers/net/ethernet/cadence/macb.h b/drivers/net/ethernet/cadence/macb.h +index d83b0db77821..3f385ab94988 100644 +--- a/drivers/net/ethernet/cadence/macb.h ++++ b/drivers/net/ethernet/cadence/macb.h +@@ -398,7 +398,7 @@ + /* Capability mask bits */ + #define MACB_CAPS_ISR_CLEAR_ON_WRITE 0x00000001 + #define MACB_CAPS_USRIO_HAS_CLKEN 0x00000002 +-#define MACB_CAPS_USRIO_DEFAULT_IS_MII 0x00000004 ++#define MACB_CAPS_USRIO_DEFAULT_IS_MII_GMII 0x00000004 + #define MACB_CAPS_NO_GIGABIT_HALF 0x00000008 + #define MACB_CAPS_FIFO_MODE 0x10000000 + #define MACB_CAPS_GIGABIT_MODE_AVAILABLE 0x20000000 +diff --git a/drivers/usb/core/quirks.c b/drivers/usb/core/quirks.c +index 6dc810bce295..944a6dca0fcb 100644 +--- a/drivers/usb/core/quirks.c ++++ b/drivers/usb/core/quirks.c +@@ -44,6 +44,9 @@ static const struct usb_device_id usb_quirk_list[] = { + /* Creative SB Audigy 2 NX */ + { USB_DEVICE(0x041e, 0x3020), .driver_info = USB_QUIRK_RESET_RESUME }, + ++ /* USB3503 */ ++ { USB_DEVICE(0x0424, 0x3503), .driver_info = USB_QUIRK_RESET_RESUME }, ++ + /* Microsoft Wireless Laser Mouse 6000 Receiver */ + { USB_DEVICE(0x045e, 0x00e1), .driver_info = USB_QUIRK_RESET_RESUME }, + +@@ -173,6 +176,10 @@ static const struct usb_device_id usb_quirk_list[] = { + /* MAYA44USB sound device */ + { USB_DEVICE(0x0a92, 0x0091), .driver_info = USB_QUIRK_RESET_RESUME }, + ++ /* ASUS Base Station(T100) */ ++ { USB_DEVICE(0x0b05, 0x17e0), .driver_info = ++ USB_QUIRK_IGNORE_REMOTE_WAKEUP }, ++ + /* Action Semiconductor flash disk */ + { USB_DEVICE(0x10d6, 0x2200), .driver_info = + USB_QUIRK_STRING_FETCH_255 }, +@@ -188,26 +195,22 @@ static const struct usb_device_id usb_quirk_list[] = { + { USB_DEVICE(0x1908, 0x1315), .driver_info = + USB_QUIRK_HONOR_BNUMINTERFACES }, + +- /* INTEL VALUE SSD */ +- { USB_DEVICE(0x8086, 0xf1a5), .driver_info = USB_QUIRK_RESET_RESUME }, +- +- /* USB3503 */ +- { USB_DEVICE(0x0424, 0x3503), .driver_info = USB_QUIRK_RESET_RESUME }, +- +- /* ASUS Base Station(T100) */ +- { USB_DEVICE(0x0b05, 0x17e0), .driver_info = +- USB_QUIRK_IGNORE_REMOTE_WAKEUP }, +- + /* Protocol and OTG Electrical Test Device */ + { USB_DEVICE(0x1a0a, 0x0200), .driver_info = + USB_QUIRK_LINEAR_UFRAME_INTR_BINTERVAL }, + ++ /* Acer C120 LED Projector */ ++ { USB_DEVICE(0x1de1, 0xc102), .driver_info = USB_QUIRK_NO_LPM }, ++ + /* Blackmagic Design Intensity Shuttle */ + { USB_DEVICE(0x1edb, 0xbd3b), .driver_info = USB_QUIRK_NO_LPM }, + + /* Blackmagic Design UltraStudio SDI */ + { USB_DEVICE(0x1edb, 0xbd4f), .driver_info = USB_QUIRK_NO_LPM }, + ++ /* INTEL VALUE SSD */ ++ { USB_DEVICE(0x8086, 0xf1a5), .driver_info = USB_QUIRK_RESET_RESUME }, ++ + { } /* terminating entry must be last */ + }; + +diff --git a/drivers/usb/dwc3/dwc3-exynos.c b/drivers/usb/dwc3/dwc3-exynos.c +index dd5cb5577dca..2f1fb7e7aa54 100644 +--- a/drivers/usb/dwc3/dwc3-exynos.c ++++ b/drivers/usb/dwc3/dwc3-exynos.c +@@ -128,12 +128,6 @@ static int dwc3_exynos_probe(struct platform_device *pdev) + + platform_set_drvdata(pdev, exynos); + +- ret = dwc3_exynos_register_phys(exynos); +- if (ret) { +- dev_err(dev, "couldn't register PHYs\n"); +- return ret; +- } +- + exynos->dev = dev; + + exynos->clk = devm_clk_get(dev, "usbdrd30"); +@@ -183,20 +177,29 @@ static int dwc3_exynos_probe(struct platform_device *pdev) + goto err3; + } + ++ ret = dwc3_exynos_register_phys(exynos); ++ if (ret) { ++ dev_err(dev, "couldn't register PHYs\n"); ++ goto err4; ++ } ++ + if (node) { + ret = of_platform_populate(node, NULL, NULL, dev); + if (ret) { + dev_err(dev, "failed to add dwc3 core\n"); +- goto err4; ++ goto err5; + } + } else { + dev_err(dev, "no device node, failed to add dwc3 core\n"); + ret = -ENODEV; +- goto err4; ++ goto err5; + } + + return 0; + ++err5: ++ platform_device_unregister(exynos->usb2_phy); ++ platform_device_unregister(exynos->usb3_phy); + err4: + regulator_disable(exynos->vdd10); + err3: +diff --git a/drivers/usb/gadget/legacy/inode.c b/drivers/usb/gadget/legacy/inode.c +index f454c7af489c..55386619a0f1 100644 +--- a/drivers/usb/gadget/legacy/inode.c ++++ b/drivers/usb/gadget/legacy/inode.c +@@ -937,8 +937,11 @@ ep0_read (struct file *fd, char __user *buf, size_t len, loff_t *ptr) + struct usb_ep *ep = dev->gadget->ep0; + struct usb_request *req = dev->req; + +- if ((retval = setup_req (ep, req, 0)) == 0) +- retval = usb_ep_queue (ep, req, GFP_ATOMIC); ++ if ((retval = setup_req (ep, req, 0)) == 0) { ++ spin_unlock_irq (&dev->lock); ++ retval = usb_ep_queue (ep, req, GFP_KERNEL); ++ spin_lock_irq (&dev->lock); ++ } + dev->state = STATE_DEV_CONNECTED; + + /* assume that was SET_CONFIGURATION */ +@@ -1456,8 +1459,11 @@ delegate: + w_length); + if (value < 0) + break; ++ ++ spin_unlock (&dev->lock); + value = usb_ep_queue (gadget->ep0, dev->req, +- GFP_ATOMIC); ++ GFP_KERNEL); ++ spin_lock (&dev->lock); + if (value < 0) { + clean_req (gadget->ep0, dev->req); + break; +@@ -1480,11 +1486,14 @@ delegate: + if (value >= 0 && dev->state != STATE_DEV_SETUP) { + req->length = value; + req->zero = value < w_length; +- value = usb_ep_queue (gadget->ep0, req, GFP_ATOMIC); ++ ++ spin_unlock (&dev->lock); ++ value = usb_ep_queue (gadget->ep0, req, GFP_KERNEL); + if (value < 0) { + DBG (dev, "ep_queue --> %d\n", value); + req->status = 0; + } ++ return value; + } + + /* device stalls when value < 0 */ +diff --git a/drivers/usb/host/ehci-tegra.c b/drivers/usb/host/ehci-tegra.c +index 4031b372008e..c1c1024a054c 100644 +--- a/drivers/usb/host/ehci-tegra.c ++++ b/drivers/usb/host/ehci-tegra.c +@@ -89,7 +89,7 @@ static int tegra_reset_usb_controller(struct platform_device *pdev) + if (!usb1_reset_attempted) { + struct reset_control *usb1_reset; + +- usb1_reset = of_reset_control_get(phy_np, "usb"); ++ usb1_reset = of_reset_control_get(phy_np, "utmi-pads"); + if (IS_ERR(usb1_reset)) { + dev_warn(&pdev->dev, + "can't get utmi-pads reset from the PHY\n"); +diff --git a/drivers/usb/host/xhci-pci.c b/drivers/usb/host/xhci-pci.c +index ea4fb4b0cd44..de644e56aa3b 100644 +--- a/drivers/usb/host/xhci-pci.c ++++ b/drivers/usb/host/xhci-pci.c +@@ -37,6 +37,7 @@ + /* Device for a quirk */ + #define PCI_VENDOR_ID_FRESCO_LOGIC 0x1b73 + #define PCI_DEVICE_ID_FRESCO_LOGIC_PDK 0x1000 ++#define PCI_DEVICE_ID_FRESCO_LOGIC_FL1009 0x1009 + #define PCI_DEVICE_ID_FRESCO_LOGIC_FL1400 0x1400 + + #define PCI_VENDOR_ID_ETRON 0x1b6f +@@ -115,6 +116,10 @@ static void xhci_pci_quirks(struct device *dev, struct xhci_hcd *xhci) + xhci->quirks |= XHCI_TRUST_TX_LENGTH; + } + ++ if (pdev->vendor == PCI_VENDOR_ID_FRESCO_LOGIC && ++ pdev->device == PCI_DEVICE_ID_FRESCO_LOGIC_FL1009) ++ xhci->quirks |= XHCI_BROKEN_STREAMS; ++ + if (pdev->vendor == PCI_VENDOR_ID_NEC) + xhci->quirks |= XHCI_NEC_HOST; + +diff --git a/drivers/usb/host/xhci-plat.c b/drivers/usb/host/xhci-plat.c +index 05647e6753cd..4fe7c9b56bc0 100644 +--- a/drivers/usb/host/xhci-plat.c ++++ b/drivers/usb/host/xhci-plat.c +@@ -132,6 +132,9 @@ static int xhci_plat_probe(struct platform_device *pdev) + ret = clk_prepare_enable(clk); + if (ret) + goto put_hcd; ++ } else if (PTR_ERR(clk) == -EPROBE_DEFER) { ++ ret = -EPROBE_DEFER; ++ goto put_hcd; + } + + if (of_device_is_compatible(pdev->dev.of_node, +diff --git a/drivers/usb/host/xhci-ring.c b/drivers/usb/host/xhci-ring.c +index 2b63969c2bbf..34cd23724bed 100644 +--- a/drivers/usb/host/xhci-ring.c ++++ b/drivers/usb/host/xhci-ring.c +@@ -289,6 +289,14 @@ static int xhci_abort_cmd_ring(struct xhci_hcd *xhci) + + temp_64 = xhci_read_64(xhci, &xhci->op_regs->cmd_ring); + xhci->cmd_ring_state = CMD_RING_STATE_ABORTED; ++ ++ /* ++ * Writing the CMD_RING_ABORT bit should cause a cmd completion event, ++ * however on some host hw the CMD_RING_RUNNING bit is correctly cleared ++ * but the completion event in never sent. Use the cmd timeout timer to ++ * handle those cases. Use twice the time to cover the bit polling retry ++ */ ++ mod_timer(&xhci->cmd_timer, jiffies + (2 * XHCI_CMD_DEFAULT_TIMEOUT)); + xhci_write_64(xhci, temp_64 | CMD_RING_ABORT, + &xhci->op_regs->cmd_ring); + +@@ -313,6 +321,7 @@ static int xhci_abort_cmd_ring(struct xhci_hcd *xhci) + + xhci_err(xhci, "Stopped the command ring failed, " + "maybe the host is dead\n"); ++ del_timer(&xhci->cmd_timer); + xhci->xhc_state |= XHCI_STATE_DYING; + xhci_quiesce(xhci); + xhci_halt(xhci); +@@ -1252,22 +1261,21 @@ void xhci_handle_command_timeout(unsigned long data) + int ret; + unsigned long flags; + u64 hw_ring_state; +- struct xhci_command *cur_cmd = NULL; ++ bool second_timeout = false; + xhci = (struct xhci_hcd *) data; + + /* mark this command to be cancelled */ + spin_lock_irqsave(&xhci->lock, flags); + if (xhci->current_cmd) { +- cur_cmd = xhci->current_cmd; +- cur_cmd->status = COMP_CMD_ABORT; ++ if (xhci->current_cmd->status == COMP_CMD_ABORT) ++ second_timeout = true; ++ xhci->current_cmd->status = COMP_CMD_ABORT; + } + +- + /* Make sure command ring is running before aborting it */ + hw_ring_state = xhci_read_64(xhci, &xhci->op_regs->cmd_ring); + if ((xhci->cmd_ring_state & CMD_RING_STATE_RUNNING) && + (hw_ring_state & CMD_RING_RUNNING)) { +- + spin_unlock_irqrestore(&xhci->lock, flags); + xhci_dbg(xhci, "Command timeout\n"); + ret = xhci_abort_cmd_ring(xhci); +@@ -1279,6 +1287,15 @@ void xhci_handle_command_timeout(unsigned long data) + } + return; + } ++ ++ /* command ring failed to restart, or host removed. Bail out */ ++ if (second_timeout || xhci->xhc_state & XHCI_STATE_REMOVING) { ++ spin_unlock_irqrestore(&xhci->lock, flags); ++ xhci_dbg(xhci, "command timed out twice, ring start fail?\n"); ++ xhci_cleanup_command_queue(xhci); ++ return; ++ } ++ + /* command timeout on stopped ring, ring can't be aborted */ + xhci_dbg(xhci, "Command timeout on stopped ring\n"); + xhci_handle_stopped_cmd_ring(xhci, xhci->current_cmd); +@@ -2727,7 +2744,8 @@ hw_died: + writel(irq_pending, &xhci->ir_set->irq_pending); + } + +- if (xhci->xhc_state & XHCI_STATE_DYING) { ++ if (xhci->xhc_state & XHCI_STATE_DYING || ++ xhci->xhc_state & XHCI_STATE_HALTED) { + xhci_dbg(xhci, "xHCI dying, ignoring interrupt. " + "Shouldn't IRQs be disabled?\n"); + /* Clear the event handler busy flag (RW1C); +diff --git a/drivers/usb/host/xhci.c b/drivers/usb/host/xhci.c +index ec9e758d5fcd..6fe0174da226 100644 +--- a/drivers/usb/host/xhci.c ++++ b/drivers/usb/host/xhci.c +@@ -680,20 +680,23 @@ void xhci_stop(struct usb_hcd *hcd) + u32 temp; + struct xhci_hcd *xhci = hcd_to_xhci(hcd); + +- if (xhci->xhc_state & XHCI_STATE_HALTED) +- return; +- + mutex_lock(&xhci->mutex); +- spin_lock_irq(&xhci->lock); +- xhci->xhc_state |= XHCI_STATE_HALTED; +- xhci->cmd_ring_state = CMD_RING_STATE_STOPPED; + +- /* Make sure the xHC is halted for a USB3 roothub +- * (xhci_stop() could be called as part of failed init). +- */ +- xhci_halt(xhci); +- xhci_reset(xhci); +- spin_unlock_irq(&xhci->lock); ++ if (!(xhci->xhc_state & XHCI_STATE_HALTED)) { ++ spin_lock_irq(&xhci->lock); ++ ++ xhci->xhc_state |= XHCI_STATE_HALTED; ++ xhci->cmd_ring_state = CMD_RING_STATE_STOPPED; ++ xhci_halt(xhci); ++ xhci_reset(xhci); ++ ++ spin_unlock_irq(&xhci->lock); ++ } ++ ++ if (!usb_hcd_is_primary_hcd(hcd)) { ++ mutex_unlock(&xhci->mutex); ++ return; ++ } + + xhci_cleanup_msix(xhci); + +diff --git a/drivers/usb/musb/musb_core.c b/drivers/usb/musb/musb_core.c +index ee9ff7028b92..00eed5d66fda 100644 +--- a/drivers/usb/musb/musb_core.c ++++ b/drivers/usb/musb/musb_core.c +@@ -2401,7 +2401,8 @@ static void musb_restore_context(struct musb *musb) + musb_writew(musb_base, MUSB_INTRTXE, musb->intrtxe); + musb_writew(musb_base, MUSB_INTRRXE, musb->intrrxe); + musb_writeb(musb_base, MUSB_INTRUSBE, musb->context.intrusbe); +- musb_writeb(musb_base, MUSB_DEVCTL, musb->context.devctl); ++ if (musb->context.devctl & MUSB_DEVCTL_SESSION) ++ musb_writeb(musb_base, MUSB_DEVCTL, musb->context.devctl); + + for (i = 0; i < musb->config->num_eps; ++i) { + struct musb_hw_ep *hw_ep; +diff --git a/drivers/usb/musb/musb_host.c b/drivers/usb/musb/musb_host.c +index 795a45b1b25b..59a63a0b7985 100644 +--- a/drivers/usb/musb/musb_host.c ++++ b/drivers/usb/musb/musb_host.c +@@ -594,14 +594,13 @@ musb_rx_reinit(struct musb *musb, struct musb_qh *qh, u8 epnum) + musb_writew(ep->regs, MUSB_TXCSR, 0); + + /* scrub all previous state, clearing toggle */ +- } else { +- csr = musb_readw(ep->regs, MUSB_RXCSR); +- if (csr & MUSB_RXCSR_RXPKTRDY) +- WARNING("rx%d, packet/%d ready?\n", ep->epnum, +- musb_readw(ep->regs, MUSB_RXCOUNT)); +- +- musb_h_flush_rxfifo(ep, MUSB_RXCSR_CLRDATATOG); + } ++ csr = musb_readw(ep->regs, MUSB_RXCSR); ++ if (csr & MUSB_RXCSR_RXPKTRDY) ++ WARNING("rx%d, packet/%d ready?\n", ep->epnum, ++ musb_readw(ep->regs, MUSB_RXCOUNT)); ++ ++ musb_h_flush_rxfifo(ep, MUSB_RXCSR_CLRDATATOG); + + /* target addr and (for multipoint) hub addr/port */ + if (musb->is_multipoint) { +@@ -995,9 +994,15 @@ static void musb_bulk_nak_timeout(struct musb *musb, struct musb_hw_ep *ep, + if (is_in) { + dma = is_dma_capable() ? ep->rx_channel : NULL; + +- /* clear nak timeout bit */ ++ /* ++ * Need to stop the transaction by clearing REQPKT first ++ * then the NAK Timeout bit ref MUSBMHDRC USB 2.0 HIGH-SPEED ++ * DUAL-ROLE CONTROLLER Programmer's Guide, section 9.2.2 ++ */ + rx_csr = musb_readw(epio, MUSB_RXCSR); + rx_csr |= MUSB_RXCSR_H_WZC_BITS; ++ rx_csr &= ~MUSB_RXCSR_H_REQPKT; ++ musb_writew(epio, MUSB_RXCSR, rx_csr); + rx_csr &= ~MUSB_RXCSR_DATAERROR; + musb_writew(epio, MUSB_RXCSR, rx_csr); + +@@ -1551,7 +1556,7 @@ static int musb_rx_dma_iso_cppi41(struct dma_controller *dma, + struct urb *urb, + size_t len) + { +- struct dma_channel *channel = hw_ep->tx_channel; ++ struct dma_channel *channel = hw_ep->rx_channel; + void __iomem *epio = hw_ep->regs; + dma_addr_t *buf; + u32 length, res; +diff --git a/drivers/usb/serial/mos7720.c b/drivers/usb/serial/mos7720.c +index 78b4f64c6b00..06c7dbc1c802 100644 +--- a/drivers/usb/serial/mos7720.c ++++ b/drivers/usb/serial/mos7720.c +@@ -2007,6 +2007,7 @@ static void mos7720_release(struct usb_serial *serial) + urblist_entry) + usb_unlink_urb(urbtrack->urb); + spin_unlock_irqrestore(&mos_parport->listlock, flags); ++ parport_del_port(mos_parport->pp); + + kref_put(&mos_parport->ref_count, destroy_mos_parport); + } +diff --git a/drivers/usb/storage/uas.c b/drivers/usb/storage/uas.c +index 9baf081174ce..e26e32169a36 100644 +--- a/drivers/usb/storage/uas.c ++++ b/drivers/usb/storage/uas.c +@@ -811,6 +811,7 @@ static int uas_slave_configure(struct scsi_device *sdev) + if (devinfo->flags & US_FL_BROKEN_FUA) + sdev->broken_fua = 1; + ++ scsi_change_queue_depth(sdev, devinfo->qdepth - 2); + return 0; + } + +diff --git a/include/linux/bpf.h b/include/linux/bpf.h +index 67bc2da5d233..4f6d29c8e3d8 100644 +--- a/include/linux/bpf.h ++++ b/include/linux/bpf.h +@@ -198,6 +198,10 @@ static inline struct bpf_prog *bpf_prog_get(u32 ufd) + static inline void bpf_prog_put(struct bpf_prog *prog) + { + } ++ ++static inline void bpf_prog_put_rcu(struct bpf_prog *prog) ++{ ++} + #endif /* CONFIG_BPF_SYSCALL */ + + /* verifier prototypes for helper functions called from eBPF programs */ +diff --git a/include/linux/net.h b/include/linux/net.h +index 25ef630f1bd6..c00b8d182226 100644 +--- a/include/linux/net.h ++++ b/include/linux/net.h +@@ -251,7 +251,8 @@ do { \ + DEFINE_DYNAMIC_DEBUG_METADATA(descriptor, fmt); \ + if (unlikely(descriptor.flags & _DPRINTK_FLAGS_PRINT) && \ + net_ratelimit()) \ +- __dynamic_pr_debug(&descriptor, fmt, ##__VA_ARGS__); \ ++ __dynamic_pr_debug(&descriptor, pr_fmt(fmt), \ ++ ##__VA_ARGS__); \ + } while (0) + #elif defined(DEBUG) + #define net_dbg_ratelimited(fmt, ...) \ +diff --git a/include/linux/skbuff.h b/include/linux/skbuff.h +index 4fde61804191..1716f9395010 100644 +--- a/include/linux/skbuff.h ++++ b/include/linux/skbuff.h +@@ -2564,6 +2564,13 @@ static inline int skb_clone_writable(const struct sk_buff *skb, unsigned int len + skb_headroom(skb) + len <= skb->hdr_len; + } + ++static inline int skb_try_make_writable(struct sk_buff *skb, ++ unsigned int write_len) ++{ ++ return skb_cloned(skb) && !skb_clone_writable(skb, write_len) && ++ pskb_expand_head(skb, 0, 0, GFP_ATOMIC); ++} ++ + static inline int __skb_cow(struct sk_buff *skb, unsigned int headroom, + int cloned) + { +diff --git a/include/linux/sock_diag.h b/include/linux/sock_diag.h +index fddebc617469..8ff34ed1ae8a 100644 +--- a/include/linux/sock_diag.h ++++ b/include/linux/sock_diag.h +@@ -35,6 +35,9 @@ enum sknetlink_groups sock_diag_destroy_group(const struct sock *sk) + { + switch (sk->sk_family) { + case AF_INET: ++ if (sk->sk_type == SOCK_RAW) ++ return SKNLGRP_NONE; ++ + switch (sk->sk_protocol) { + case IPPROTO_TCP: + return SKNLGRP_INET_TCP_DESTROY; +@@ -44,6 +47,9 @@ enum sknetlink_groups sock_diag_destroy_group(const struct sock *sk) + return SKNLGRP_NONE; + } + case AF_INET6: ++ if (sk->sk_type == SOCK_RAW) ++ return SKNLGRP_NONE; ++ + switch (sk->sk_protocol) { + case IPPROTO_TCP: + return SKNLGRP_INET6_TCP_DESTROY; +diff --git a/kernel/events/core.c b/kernel/events/core.c +index 95e47d2f2c67..12ecd4f0329f 100644 +--- a/kernel/events/core.c ++++ b/kernel/events/core.c +@@ -7101,7 +7101,7 @@ static void perf_event_free_bpf_prog(struct perf_event *event) + prog = event->tp_event->prog; + if (prog) { + event->tp_event->prog = NULL; +- bpf_prog_put(prog); ++ bpf_prog_put_rcu(prog); + } + } + +diff --git a/net/ax25/af_ax25.c b/net/ax25/af_ax25.c +index fbd0acf80b13..2fdebabbfacd 100644 +--- a/net/ax25/af_ax25.c ++++ b/net/ax25/af_ax25.c +@@ -976,7 +976,8 @@ static int ax25_release(struct socket *sock) + release_sock(sk); + ax25_disconnect(ax25, 0); + lock_sock(sk); +- ax25_destroy_socket(ax25); ++ if (!sock_flag(ax25->sk, SOCK_DESTROY)) ++ ax25_destroy_socket(ax25); + break; + + case AX25_STATE_3: +diff --git a/net/ax25/ax25_ds_timer.c b/net/ax25/ax25_ds_timer.c +index 951cd57bb07d..5237dff6941d 100644 +--- a/net/ax25/ax25_ds_timer.c ++++ b/net/ax25/ax25_ds_timer.c +@@ -102,6 +102,7 @@ void ax25_ds_heartbeat_expiry(ax25_cb *ax25) + switch (ax25->state) { + + case AX25_STATE_0: ++ case AX25_STATE_2: + /* Magic here: If we listen() and a new link dies before it + is accepted() it isn't 'dead' so doesn't get removed. */ + if (!sk || sock_flag(sk, SOCK_DESTROY) || +@@ -111,6 +112,7 @@ void ax25_ds_heartbeat_expiry(ax25_cb *ax25) + sock_hold(sk); + ax25_destroy_socket(ax25); + bh_unlock_sock(sk); ++ /* Ungrab socket and destroy it */ + sock_put(sk); + } else + ax25_destroy_socket(ax25); +@@ -213,7 +215,8 @@ void ax25_ds_t1_timeout(ax25_cb *ax25) + case AX25_STATE_2: + if (ax25->n2count == ax25->n2) { + ax25_send_control(ax25, AX25_DISC, AX25_POLLON, AX25_COMMAND); +- ax25_disconnect(ax25, ETIMEDOUT); ++ if (!sock_flag(ax25->sk, SOCK_DESTROY)) ++ ax25_disconnect(ax25, ETIMEDOUT); + return; + } else { + ax25->n2count++; +diff --git a/net/ax25/ax25_std_timer.c b/net/ax25/ax25_std_timer.c +index 004467c9e6e1..2c0d6ef66f9d 100644 +--- a/net/ax25/ax25_std_timer.c ++++ b/net/ax25/ax25_std_timer.c +@@ -38,6 +38,7 @@ void ax25_std_heartbeat_expiry(ax25_cb *ax25) + + switch (ax25->state) { + case AX25_STATE_0: ++ case AX25_STATE_2: + /* Magic here: If we listen() and a new link dies before it + is accepted() it isn't 'dead' so doesn't get removed. */ + if (!sk || sock_flag(sk, SOCK_DESTROY) || +@@ -47,6 +48,7 @@ void ax25_std_heartbeat_expiry(ax25_cb *ax25) + sock_hold(sk); + ax25_destroy_socket(ax25); + bh_unlock_sock(sk); ++ /* Ungrab socket and destroy it */ + sock_put(sk); + } else + ax25_destroy_socket(ax25); +@@ -144,7 +146,8 @@ void ax25_std_t1timer_expiry(ax25_cb *ax25) + case AX25_STATE_2: + if (ax25->n2count == ax25->n2) { + ax25_send_control(ax25, AX25_DISC, AX25_POLLON, AX25_COMMAND); +- ax25_disconnect(ax25, ETIMEDOUT); ++ if (!sock_flag(ax25->sk, SOCK_DESTROY)) ++ ax25_disconnect(ax25, ETIMEDOUT); + return; + } else { + ax25->n2count++; +diff --git a/net/ax25/ax25_subr.c b/net/ax25/ax25_subr.c +index 3b78e8473a01..655a7d4c96e1 100644 +--- a/net/ax25/ax25_subr.c ++++ b/net/ax25/ax25_subr.c +@@ -264,7 +264,8 @@ void ax25_disconnect(ax25_cb *ax25, int reason) + { + ax25_clear_queues(ax25); + +- ax25_stop_heartbeat(ax25); ++ if (!sock_flag(ax25->sk, SOCK_DESTROY)) ++ ax25_stop_heartbeat(ax25); + ax25_stop_t1timer(ax25); + ax25_stop_t2timer(ax25); + ax25_stop_t3timer(ax25); +diff --git a/net/bridge/br_multicast.c b/net/bridge/br_multicast.c +index ea9893743a0f..7173a685309a 100644 +--- a/net/bridge/br_multicast.c ++++ b/net/bridge/br_multicast.c +@@ -464,8 +464,11 @@ static struct sk_buff *br_ip6_multicast_alloc_query(struct net_bridge *br, + if (ipv6_dev_get_saddr(dev_net(br->dev), br->dev, &ip6h->daddr, 0, + &ip6h->saddr)) { + kfree_skb(skb); ++ br->has_ipv6_addr = 0; + return NULL; + } ++ ++ br->has_ipv6_addr = 1; + ipv6_eth_mc_map(&ip6h->daddr, eth->h_dest); + + hopopt = (u8 *)(ip6h + 1); +@@ -1736,6 +1739,7 @@ void br_multicast_init(struct net_bridge *br) + br->ip6_other_query.delay_time = 0; + br->ip6_querier.port = NULL; + #endif ++ br->has_ipv6_addr = 1; + + spin_lock_init(&br->multicast_lock); + setup_timer(&br->multicast_router_timer, +diff --git a/net/bridge/br_private.h b/net/bridge/br_private.h +index 216018c76018..1001a1b7df9b 100644 +--- a/net/bridge/br_private.h ++++ b/net/bridge/br_private.h +@@ -301,6 +301,7 @@ struct net_bridge + u8 multicast_disabled:1; + u8 multicast_querier:1; + u8 multicast_query_use_ifaddr:1; ++ u8 has_ipv6_addr:1; + + u32 hash_elasticity; + u32 hash_max; +@@ -574,10 +575,22 @@ static inline bool br_multicast_is_router(struct net_bridge *br) + + static inline bool + __br_multicast_querier_exists(struct net_bridge *br, +- struct bridge_mcast_other_query *querier) ++ struct bridge_mcast_other_query *querier, ++ const bool is_ipv6) + { ++ bool own_querier_enabled; ++ ++ if (br->multicast_querier) { ++ if (is_ipv6 && !br->has_ipv6_addr) ++ own_querier_enabled = false; ++ else ++ own_querier_enabled = true; ++ } else { ++ own_querier_enabled = false; ++ } ++ + return time_is_before_jiffies(querier->delay_time) && +- (br->multicast_querier || timer_pending(&querier->timer)); ++ (own_querier_enabled || timer_pending(&querier->timer)); + } + + static inline bool br_multicast_querier_exists(struct net_bridge *br, +@@ -585,10 +598,12 @@ static inline bool br_multicast_querier_exists(struct net_bridge *br, + { + switch (eth->h_proto) { + case (htons(ETH_P_IP)): +- return __br_multicast_querier_exists(br, &br->ip4_other_query); ++ return __br_multicast_querier_exists(br, ++ &br->ip4_other_query, false); + #if IS_ENABLED(CONFIG_IPV6) + case (htons(ETH_P_IPV6)): +- return __br_multicast_querier_exists(br, &br->ip6_other_query); ++ return __br_multicast_querier_exists(br, ++ &br->ip6_other_query, true); + #endif + default: + return false; +diff --git a/net/core/filter.c b/net/core/filter.c +index f393a22b9d50..75e9b2b2336d 100644 +--- a/net/core/filter.c ++++ b/net/core/filter.c +@@ -1275,9 +1275,7 @@ static u64 bpf_skb_store_bytes(u64 r1, u64 r2, u64 r3, u64 r4, u64 flags) + */ + if (unlikely((u32) offset > 0xffff || len > sizeof(buf))) + return -EFAULT; +- +- if (unlikely(skb_cloned(skb) && +- !skb_clone_writable(skb, offset + len))) ++ if (unlikely(skb_try_make_writable(skb, offset + len))) + return -EFAULT; + + ptr = skb_header_pointer(skb, offset, len, buf); +@@ -1321,8 +1319,7 @@ static u64 bpf_l3_csum_replace(u64 r1, u64 r2, u64 from, u64 to, u64 flags) + if (unlikely((u32) offset > 0xffff)) + return -EFAULT; + +- if (unlikely(skb_cloned(skb) && +- !skb_clone_writable(skb, offset + sizeof(sum)))) ++ if (unlikely(skb_try_make_writable(skb, offset + sizeof(sum)))) + return -EFAULT; + + ptr = skb_header_pointer(skb, offset, sizeof(sum), &sum); +@@ -1367,9 +1364,7 @@ static u64 bpf_l4_csum_replace(u64 r1, u64 r2, u64 from, u64 to, u64 flags) + + if (unlikely((u32) offset > 0xffff)) + return -EFAULT; +- +- if (unlikely(skb_cloned(skb) && +- !skb_clone_writable(skb, offset + sizeof(sum)))) ++ if (unlikely(skb_try_make_writable(skb, offset + sizeof(sum)))) + return -EFAULT; + + ptr = skb_header_pointer(skb, offset, sizeof(sum), &sum); +@@ -1554,6 +1549,13 @@ bool bpf_helper_changes_skb_data(void *func) + return true; + if (func == bpf_skb_vlan_pop) + return true; ++ if (func == bpf_skb_store_bytes) ++ return true; ++ if (func == bpf_l3_csum_replace) ++ return true; ++ if (func == bpf_l4_csum_replace) ++ return true; ++ + return false; + } + +diff --git a/net/core/neighbour.c b/net/core/neighbour.c +index f18ae91b652e..769cece9b00b 100644 +--- a/net/core/neighbour.c ++++ b/net/core/neighbour.c +@@ -2467,13 +2467,17 @@ int neigh_xmit(int index, struct net_device *dev, + tbl = neigh_tables[index]; + if (!tbl) + goto out; ++ rcu_read_lock_bh(); + neigh = __neigh_lookup_noref(tbl, addr, dev); + if (!neigh) + neigh = __neigh_create(tbl, addr, dev, false); + err = PTR_ERR(neigh); +- if (IS_ERR(neigh)) ++ if (IS_ERR(neigh)) { ++ rcu_read_unlock_bh(); + goto out_kfree_skb; ++ } + err = neigh->output(neigh, skb); ++ rcu_read_unlock_bh(); + } + else if (index == NEIGH_LINK_TABLE) { + err = dev_hard_header(skb, dev, ntohs(skb->protocol), +diff --git a/net/ipv4/esp4.c b/net/ipv4/esp4.c +index 477937465a20..d95631d09248 100644 +--- a/net/ipv4/esp4.c ++++ b/net/ipv4/esp4.c +@@ -23,6 +23,11 @@ struct esp_skb_cb { + void *tmp; + }; + ++struct esp_output_extra { ++ __be32 seqhi; ++ u32 esphoff; ++}; ++ + #define ESP_SKB_CB(__skb) ((struct esp_skb_cb *)&((__skb)->cb[0])) + + static u32 esp4_get_mtu(struct xfrm_state *x, int mtu); +@@ -35,11 +40,11 @@ static u32 esp4_get_mtu(struct xfrm_state *x, int mtu); + * + * TODO: Use spare space in skb for this where possible. + */ +-static void *esp_alloc_tmp(struct crypto_aead *aead, int nfrags, int seqhilen) ++static void *esp_alloc_tmp(struct crypto_aead *aead, int nfrags, int extralen) + { + unsigned int len; + +- len = seqhilen; ++ len = extralen; + + len += crypto_aead_ivsize(aead); + +@@ -57,15 +62,16 @@ static void *esp_alloc_tmp(struct crypto_aead *aead, int nfrags, int seqhilen) + return kmalloc(len, GFP_ATOMIC); + } + +-static inline __be32 *esp_tmp_seqhi(void *tmp) ++static inline void *esp_tmp_extra(void *tmp) + { +- return PTR_ALIGN((__be32 *)tmp, __alignof__(__be32)); ++ return PTR_ALIGN(tmp, __alignof__(struct esp_output_extra)); + } +-static inline u8 *esp_tmp_iv(struct crypto_aead *aead, void *tmp, int seqhilen) ++ ++static inline u8 *esp_tmp_iv(struct crypto_aead *aead, void *tmp, int extralen) + { + return crypto_aead_ivsize(aead) ? +- PTR_ALIGN((u8 *)tmp + seqhilen, +- crypto_aead_alignmask(aead) + 1) : tmp + seqhilen; ++ PTR_ALIGN((u8 *)tmp + extralen, ++ crypto_aead_alignmask(aead) + 1) : tmp + extralen; + } + + static inline struct aead_request *esp_tmp_req(struct crypto_aead *aead, u8 *iv) +@@ -99,7 +105,7 @@ static void esp_restore_header(struct sk_buff *skb, unsigned int offset) + { + struct ip_esp_hdr *esph = (void *)(skb->data + offset); + void *tmp = ESP_SKB_CB(skb)->tmp; +- __be32 *seqhi = esp_tmp_seqhi(tmp); ++ __be32 *seqhi = esp_tmp_extra(tmp); + + esph->seq_no = esph->spi; + esph->spi = *seqhi; +@@ -107,7 +113,11 @@ static void esp_restore_header(struct sk_buff *skb, unsigned int offset) + + static void esp_output_restore_header(struct sk_buff *skb) + { +- esp_restore_header(skb, skb_transport_offset(skb) - sizeof(__be32)); ++ void *tmp = ESP_SKB_CB(skb)->tmp; ++ struct esp_output_extra *extra = esp_tmp_extra(tmp); ++ ++ esp_restore_header(skb, skb_transport_offset(skb) + extra->esphoff - ++ sizeof(__be32)); + } + + static void esp_output_done_esn(struct crypto_async_request *base, int err) +@@ -121,6 +131,7 @@ static void esp_output_done_esn(struct crypto_async_request *base, int err) + static int esp_output(struct xfrm_state *x, struct sk_buff *skb) + { + int err; ++ struct esp_output_extra *extra; + struct ip_esp_hdr *esph; + struct crypto_aead *aead; + struct aead_request *req; +@@ -137,8 +148,7 @@ static int esp_output(struct xfrm_state *x, struct sk_buff *skb) + int tfclen; + int nfrags; + int assoclen; +- int seqhilen; +- __be32 *seqhi; ++ int extralen; + __be64 seqno; + + /* skb is pure payload to encrypt */ +@@ -166,21 +176,21 @@ static int esp_output(struct xfrm_state *x, struct sk_buff *skb) + nfrags = err; + + assoclen = sizeof(*esph); +- seqhilen = 0; ++ extralen = 0; + + if (x->props.flags & XFRM_STATE_ESN) { +- seqhilen += sizeof(__be32); +- assoclen += seqhilen; ++ extralen += sizeof(*extra); ++ assoclen += sizeof(__be32); + } + +- tmp = esp_alloc_tmp(aead, nfrags, seqhilen); ++ tmp = esp_alloc_tmp(aead, nfrags, extralen); + if (!tmp) { + err = -ENOMEM; + goto error; + } + +- seqhi = esp_tmp_seqhi(tmp); +- iv = esp_tmp_iv(aead, tmp, seqhilen); ++ extra = esp_tmp_extra(tmp); ++ iv = esp_tmp_iv(aead, tmp, extralen); + req = esp_tmp_req(aead, iv); + sg = esp_req_sg(aead, req); + +@@ -247,8 +257,10 @@ static int esp_output(struct xfrm_state *x, struct sk_buff *skb) + * encryption. + */ + if ((x->props.flags & XFRM_STATE_ESN)) { +- esph = (void *)(skb_transport_header(skb) - sizeof(__be32)); +- *seqhi = esph->spi; ++ extra->esphoff = (unsigned char *)esph - ++ skb_transport_header(skb); ++ esph = (struct ip_esp_hdr *)((unsigned char *)esph - 4); ++ extra->seqhi = esph->spi; + esph->seq_no = htonl(XFRM_SKB_CB(skb)->seq.output.hi); + aead_request_set_callback(req, 0, esp_output_done_esn, skb); + } +@@ -445,7 +457,7 @@ static int esp_input(struct xfrm_state *x, struct sk_buff *skb) + goto out; + + ESP_SKB_CB(skb)->tmp = tmp; +- seqhi = esp_tmp_seqhi(tmp); ++ seqhi = esp_tmp_extra(tmp); + iv = esp_tmp_iv(aead, tmp, seqhilen); + req = esp_tmp_req(aead, iv); + sg = esp_req_sg(aead, req); +diff --git a/net/ipv4/ipmr.c b/net/ipv4/ipmr.c +index c3a38353f5dc..9d1e555496e3 100644 +--- a/net/ipv4/ipmr.c ++++ b/net/ipv4/ipmr.c +@@ -882,8 +882,10 @@ static struct mfc_cache *ipmr_cache_alloc(void) + { + struct mfc_cache *c = kmem_cache_zalloc(mrt_cachep, GFP_KERNEL); + +- if (c) ++ if (c) { ++ c->mfc_un.res.last_assert = jiffies - MFC_ASSERT_THRESH - 1; + c->mfc_un.res.minvif = MAXVIFS; ++ } + return c; + } + +diff --git a/net/ipv6/ip6mr.c b/net/ipv6/ip6mr.c +index a10e77103c88..e207cb2468da 100644 +--- a/net/ipv6/ip6mr.c ++++ b/net/ipv6/ip6mr.c +@@ -1074,6 +1074,7 @@ static struct mfc6_cache *ip6mr_cache_alloc(void) + struct mfc6_cache *c = kmem_cache_zalloc(mrt_cachep, GFP_KERNEL); + if (!c) + return NULL; ++ c->mfc_un.res.last_assert = jiffies - MFC_ASSERT_THRESH - 1; + c->mfc_un.res.minvif = MAXMIFS; + return c; + } +diff --git a/net/ipv6/sit.c b/net/ipv6/sit.c +index dcccae86190f..ba3d2f3d66d2 100644 +--- a/net/ipv6/sit.c ++++ b/net/ipv6/sit.c +@@ -560,13 +560,13 @@ static int ipip6_err(struct sk_buff *skb, u32 info) + + if (type == ICMP_DEST_UNREACH && code == ICMP_FRAG_NEEDED) { + ipv4_update_pmtu(skb, dev_net(skb->dev), info, +- t->parms.link, 0, IPPROTO_IPV6, 0); ++ t->parms.link, 0, iph->protocol, 0); + err = 0; + goto out; + } + if (type == ICMP_REDIRECT) { + ipv4_redirect(skb, dev_net(skb->dev), t->parms.link, 0, +- IPPROTO_IPV6, 0); ++ iph->protocol, 0); + err = 0; + goto out; + } +diff --git a/net/sched/act_csum.c b/net/sched/act_csum.c +index b07c535ba8e7..eeb3eb3ea9eb 100644 +--- a/net/sched/act_csum.c ++++ b/net/sched/act_csum.c +@@ -105,9 +105,7 @@ static void *tcf_csum_skb_nextlayer(struct sk_buff *skb, + int hl = ihl + jhl; + + if (!pskb_may_pull(skb, ipl + ntkoff) || (ipl < hl) || +- (skb_cloned(skb) && +- !skb_clone_writable(skb, hl + ntkoff) && +- pskb_expand_head(skb, 0, 0, GFP_ATOMIC))) ++ skb_try_make_writable(skb, hl + ntkoff)) + return NULL; + else + return (void *)(skb_network_header(skb) + ihl); +@@ -365,9 +363,7 @@ static int tcf_csum_ipv4(struct sk_buff *skb, u32 update_flags) + } + + if (update_flags & TCA_CSUM_UPDATE_FLAG_IPV4HDR) { +- if (skb_cloned(skb) && +- !skb_clone_writable(skb, sizeof(*iph) + ntkoff) && +- pskb_expand_head(skb, 0, 0, GFP_ATOMIC)) ++ if (skb_try_make_writable(skb, sizeof(*iph) + ntkoff)) + goto fail; + + ip_send_check(ip_hdr(skb)); +diff --git a/net/sched/act_nat.c b/net/sched/act_nat.c +index b7c4ead8b5a8..27607b863aba 100644 +--- a/net/sched/act_nat.c ++++ b/net/sched/act_nat.c +@@ -126,9 +126,7 @@ static int tcf_nat(struct sk_buff *skb, const struct tc_action *a, + addr = iph->daddr; + + if (!((old_addr ^ addr) & mask)) { +- if (skb_cloned(skb) && +- !skb_clone_writable(skb, sizeof(*iph) + noff) && +- pskb_expand_head(skb, 0, 0, GFP_ATOMIC)) ++ if (skb_try_make_writable(skb, sizeof(*iph) + noff)) + goto drop; + + new_addr &= mask; +@@ -156,9 +154,7 @@ static int tcf_nat(struct sk_buff *skb, const struct tc_action *a, + struct tcphdr *tcph; + + if (!pskb_may_pull(skb, ihl + sizeof(*tcph) + noff) || +- (skb_cloned(skb) && +- !skb_clone_writable(skb, ihl + sizeof(*tcph) + noff) && +- pskb_expand_head(skb, 0, 0, GFP_ATOMIC))) ++ skb_try_make_writable(skb, ihl + sizeof(*tcph) + noff)) + goto drop; + + tcph = (void *)(skb_network_header(skb) + ihl); +@@ -171,9 +167,7 @@ static int tcf_nat(struct sk_buff *skb, const struct tc_action *a, + struct udphdr *udph; + + if (!pskb_may_pull(skb, ihl + sizeof(*udph) + noff) || +- (skb_cloned(skb) && +- !skb_clone_writable(skb, ihl + sizeof(*udph) + noff) && +- pskb_expand_head(skb, 0, 0, GFP_ATOMIC))) ++ skb_try_make_writable(skb, ihl + sizeof(*udph) + noff)) + goto drop; + + udph = (void *)(skb_network_header(skb) + ihl); +@@ -213,10 +207,8 @@ static int tcf_nat(struct sk_buff *skb, const struct tc_action *a, + if ((old_addr ^ addr) & mask) + break; + +- if (skb_cloned(skb) && +- !skb_clone_writable(skb, ihl + sizeof(*icmph) + +- sizeof(*iph) + noff) && +- pskb_expand_head(skb, 0, 0, GFP_ATOMIC)) ++ if (skb_try_make_writable(skb, ihl + sizeof(*icmph) + ++ sizeof(*iph) + noff)) + goto drop; + + icmph = (void *)(skb_network_header(skb) + ihl); +diff --git a/net/sched/sch_fifo.c b/net/sched/sch_fifo.c +index 2177eac0a61e..2e4bd2c0a50c 100644 +--- a/net/sched/sch_fifo.c ++++ b/net/sched/sch_fifo.c +@@ -37,14 +37,18 @@ static int pfifo_enqueue(struct sk_buff *skb, struct Qdisc *sch) + + static int pfifo_tail_enqueue(struct sk_buff *skb, struct Qdisc *sch) + { ++ unsigned int prev_backlog; ++ + if (likely(skb_queue_len(&sch->q) < sch->limit)) + return qdisc_enqueue_tail(skb, sch); + ++ prev_backlog = sch->qstats.backlog; + /* queue full, remove one skb to fulfill the limit */ + __qdisc_queue_drop_head(sch, &sch->q); + qdisc_qstats_drop(sch); + qdisc_enqueue_tail(skb, sch); + ++ qdisc_tree_reduce_backlog(sch, 0, prev_backlog - sch->qstats.backlog); + return NET_XMIT_CN; + } + +diff --git a/net/sched/sch_netem.c b/net/sched/sch_netem.c +index 4befe97a9034..b7c29d5b6f04 100644 +--- a/net/sched/sch_netem.c ++++ b/net/sched/sch_netem.c +@@ -650,14 +650,14 @@ deliver: + #endif + + if (q->qdisc) { ++ unsigned int pkt_len = qdisc_pkt_len(skb); + int err = qdisc_enqueue(skb, q->qdisc); + +- if (unlikely(err != NET_XMIT_SUCCESS)) { +- if (net_xmit_drop_count(err)) { +- qdisc_qstats_drop(sch); +- qdisc_tree_reduce_backlog(sch, 1, +- qdisc_pkt_len(skb)); +- } ++ if (err != NET_XMIT_SUCCESS && ++ net_xmit_drop_count(err)) { ++ qdisc_qstats_drop(sch); ++ qdisc_tree_reduce_backlog(sch, 1, ++ pkt_len); + } + goto tfifo_dequeue; + } |