--- patch-4.9.104-105 2018-05-30 22:28:26.000000000 +0200 +++ 61_patch-4.9.105.quilt 2018-08-13 20:20:55.578906087 +0200 @@ -1,24 +1,24 @@ -diff --git a/Makefile b/Makefile -index 780dcc8033b2..7d06dba3fde8 100644 ---- a/Makefile -+++ b/Makefile -@@ -1,6 +1,6 @@ - VERSION = 4 - PATCHLEVEL = 9 --SUBLEVEL = 104 -+SUBLEVEL = 105 - EXTRAVERSION = - NAME = Roaring Lionus - -diff --git a/net/ipv4/ip_vti.c b/net/ipv4/ip_vti.c -index 1ac55b116d5a..cbff0d6ff1ac 100644 ---- a/net/ipv4/ip_vti.c -+++ b/net/ipv4/ip_vti.c -@@ -396,6 +396,7 @@ static int vti_tunnel_init(struct net_device *dev) - memcpy(dev->dev_addr, &iph->saddr, 4); - memcpy(dev->broadcast, &iph->daddr, 4); - -+ dev->mtu = ETH_DATA_LEN; - dev->flags = IFF_NOARP; - dev->addr_len = 4; - dev->features |= NETIF_F_LLTX; +diff --git a/Makefile b/Makefile +index 780dcc8033b2..7d06dba3fde8 100644 +--- a/Makefile ++++ b/Makefile +@@ -1,6 +1,6 @@ + VERSION = 4 + PATCHLEVEL = 9 +-SUBLEVEL = 104 ++SUBLEVEL = 105 + EXTRAVERSION = + NAME = Roaring Lionus + +diff --git a/net/ipv4/ip_vti.c b/net/ipv4/ip_vti.c +index 1ac55b116d5a..cbff0d6ff1ac 100644 +--- a/net/ipv4/ip_vti.c ++++ b/net/ipv4/ip_vti.c +@@ -396,6 +396,7 @@ static int vti_tunnel_init(struct net_device *dev) + memcpy(dev->dev_addr, &iph->saddr, 4); + memcpy(dev->broadcast, &iph->daddr, 4); + ++ dev->mtu = ETH_DATA_LEN; + dev->flags = IFF_NOARP; + dev->addr_len = 4; + dev->features |= NETIF_F_LLTX; 105 --- patch-4.9.110-111 2018-07-03 21:07:09.000000000 +0200 +++ 61_patch-4.9.111.quilt 2018-08-13 20:20:55.574906046 +0200 @@ -1622,7 +1622,7 @@ + * Processor Family I/O for U Quad Core Platforms Specification Update, + * August 2017, Revision 002, Document#: 334660-002)[6] + * Device IDs from I/O datasheet: (7th Generation Intel Processor Family I/O -+ * for U/Y Platforms and 8th Generation Intel ® Processor Family I/O for U ++ * for U/Y Platforms and 8th Generation Intel ® Processor Family I/O for U + * Quad Core Platforms, Vol 1 of 2, August 2017, Document#: 334658-003)[7] + * + * 0x9d10-0x9d1b PCI Express Root port #{1-12} 111 --- patch-4.9.115-116 2018-07-28 07:56:57.000000000 +0200 +++ 61_patch-4.9.116.quilt 2018-08-13 20:20:55.578906087 +0200 @@ -48,7 +48,7 @@ index d76cd97a98b6..ee25a69630c3 100644 --- a/drivers/base/dd.c +++ b/drivers/base/dd.c -@@ -363,14 +363,6 @@ re_probe: +@@ -363,14 +363,6 @@ static int really_probe(struct device *dev, struct device_driver *drv) goto probe_failed; } @@ -857,7 +857,7 @@ index b8c43be24751..fcd8642ef2d2 100644 --- a/fs/exec.c +++ b/fs/exec.c -@@ -1228,15 +1228,14 @@ killed: +@@ -1228,15 +1228,14 @@ static int de_thread(struct task_struct *tsk) return -EAGAIN; } @@ -950,7 +950,7 @@ index 8cae7aa4a4ec..84c731aef0d8 100644 --- a/net/core/skbuff.c +++ b/net/core/skbuff.c -@@ -3253,6 +3253,7 @@ normal: +@@ -3253,6 +3253,7 @@ struct sk_buff *skb_segment(struct sk_buff *head_skb, net_warn_ratelimited( "skb_segment: too many frags: %u %u\n", pos, mss); @@ -958,7 +958,7 @@ goto err; } -@@ -3289,11 +3290,10 @@ skip_fraglist: +@@ -3289,11 +3290,10 @@ struct sk_buff *skb_segment(struct sk_buff *head_skb, perform_csum_check: if (!csum) { @@ -1114,7 +1114,7 @@ /* Send ACKs quickly, if "quick" count is not exhausted * and the session is not interactive. -@@ -4516,7 +4517,7 @@ coalesce_done: +@@ -4516,7 +4517,7 @@ static void tcp_data_queue_ofo(struct sock *sk, struct sk_buff *skb) /* All the bits are present. Drop. */ NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPOFOMERGE); @@ -1123,7 +1123,7 @@ skb = NULL; tcp_dsack_set(sk, seq, end_seq); goto add_sack; -@@ -4535,7 +4536,7 @@ coalesce_done: +@@ -4535,7 +4536,7 @@ static void tcp_data_queue_ofo(struct sock *sk, struct sk_buff *skb) TCP_SKB_CB(skb1)->end_seq); NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPOFOMERGE); @@ -1132,7 +1132,7 @@ goto merge_right; } } else if (tcp_try_coalesce(sk, skb1, skb, &fragstolen)) { -@@ -4917,6 +4918,7 @@ end: +@@ -4917,6 +4918,7 @@ tcp_collapse(struct sock *sk, struct sk_buff_head *list, struct rb_root *root, static void tcp_collapse_ofo_queue(struct sock *sk) { struct tcp_sock *tp = tcp_sk(sk); @@ -1140,7 +1140,7 @@ struct sk_buff *skb, *head; struct rb_node *p; u32 start, end; -@@ -4935,6 +4937,7 @@ new_range: +@@ -4935,6 +4937,7 @@ static void tcp_collapse_ofo_queue(struct sock *sk) } start = TCP_SKB_CB(skb)->seq; end = TCP_SKB_CB(skb)->end_seq; @@ -1148,7 +1148,7 @@ for (head = skb;;) { skb = tcp_skb_next(skb, NULL); -@@ -4945,11 +4948,20 @@ new_range: +@@ -4945,11 +4948,20 @@ static void tcp_collapse_ofo_queue(struct sock *sk) if (!skb || after(TCP_SKB_CB(skb)->seq, end) || before(TCP_SKB_CB(skb)->end_seq, start)) { @@ -1171,7 +1171,7 @@ if (unlikely(before(TCP_SKB_CB(skb)->seq, start))) start = TCP_SKB_CB(skb)->seq; if (after(TCP_SKB_CB(skb)->end_seq, end)) -@@ -4964,6 +4976,7 @@ new_range: +@@ -4964,6 +4976,7 @@ static void tcp_collapse_ofo_queue(struct sock *sk) * 2) not add too big latencies if thousands of packets sit there. * (But if application shrinks SO_RCVBUF, we could still end up * freeing whole queue here) @@ -1239,7 +1239,7 @@ tcp_dec_quickack_mode(sk, pkts); inet_csk_clear_xmit_timer(sk, ICSK_TIME_DACK); } -@@ -905,8 +910,8 @@ out: +@@ -905,8 +910,8 @@ void tcp_wfree(struct sk_buff *skb) * We are working here with either a clone of the original * SKB, or a fresh unique copy made by the retransmit engine. */ 116 --- patch-4.9.116-117 2018-08-03 08:03:57.000000000 +0200 +++ 61_patch-4.9.117.quilt 2018-08-13 20:20:55.574906046 +0200 @@ -814,7 +814,7 @@ struct drm_display_mode *mode) { struct drm_device *dev = connector->dev; -@@ -1798,7 +1798,7 @@ out: +@@ -1798,7 +1798,7 @@ radeon_dp_detect(struct drm_connector *connector, bool force) return ret; } @@ -1160,7 +1160,7 @@ index 0321d84addc7..15a86bb4e61c 100644 --- a/drivers/media/platform/omap3isp/isp.c +++ b/drivers/media/platform/omap3isp/isp.c -@@ -1941,6 +1941,7 @@ error_csiphy: +@@ -1941,6 +1941,7 @@ static int isp_initialize_modules(struct isp_device *isp) static void isp_detach_iommu(struct isp_device *isp) { @@ -2188,7 +2188,7 @@ lock->l_req_mode = mode; lock->l_ast_data = data; -@@ -1533,6 +1535,8 @@ out: +@@ -1533,6 +1535,8 @@ struct ldlm_lock *ldlm_lock_create(struct ldlm_namespace *ns, return ERR_PTR(rc); } @@ -2437,7 +2437,7 @@ btrfs_item_key_to_cpu(path->nodes[0], &found, btrfs_header_nritems(path->nodes[0]) - 1); -@@ -2323,6 +2340,8 @@ out: +@@ -2323,6 +2340,8 @@ qgroup_rescan_leaf(struct btrfs_fs_info *fs_info, struct btrfs_path *path, } btrfs_put_tree_mod_seq(fs_info, &tree_mod_seq_elem); @@ -2450,7 +2450,7 @@ index 44d34923de9c..44966fd00790 100644 --- a/fs/btrfs/tree-log.c +++ b/fs/btrfs/tree-log.c -@@ -2979,8 +2979,11 @@ out_wake_log_root: +@@ -2979,8 +2979,11 @@ int btrfs_sync_log(struct btrfs_trans_handle *trans, mutex_unlock(&log_root_tree->log_mutex); /* @@ -2463,7 +2463,7 @@ if (waitqueue_active(&log_root_tree->log_commit_wait[index2])) wake_up(&log_root_tree->log_commit_wait[index2]); out: -@@ -2991,8 +2994,11 @@ out: +@@ -2991,8 +2994,11 @@ int btrfs_sync_log(struct btrfs_trans_handle *trans, mutex_unlock(&root->log_mutex); /* @@ -2573,7 +2573,7 @@ out: return copied; } -@@ -895,7 +900,6 @@ retry_journal: +@@ -895,7 +900,6 @@ int ext4_da_write_inline_data_begin(struct address_space *mapping, goto out; } @@ -2581,7 +2581,7 @@ page = grab_cache_page_write_begin(mapping, 0, flags); if (!page) { ret = -ENOMEM; -@@ -913,6 +917,9 @@ retry_journal: +@@ -913,6 +917,9 @@ int ext4_da_write_inline_data_begin(struct address_space *mapping, if (ret < 0) goto out_release_page; } @@ -2735,7 +2735,7 @@ index 17ab23f64bba..ad4dfd29d923 100644 --- a/fs/f2fs/gc.c +++ b/fs/f2fs/gc.c -@@ -688,9 +688,14 @@ retry: +@@ -688,9 +688,14 @@ static void move_data_page(struct inode *inode, block_t bidx, int gc_type) set_cold_data(page); err = do_write_data_page(&fio); @@ -3022,7 +3022,7 @@ */ map = bpf_map_inc(map, false); if (IS_ERR(map)) { -@@ -3629,7 +3629,7 @@ free_log_buf: +@@ -3629,7 +3629,7 @@ int bpf_check(struct bpf_prog **prog, union bpf_attr *attr) vfree(log_buf); if (!env->prog->aux->used_maps) /* if we didn't copy map pointers into bpf_prog_info, release @@ -3105,7 +3105,7 @@ err = -ENOENT; if (!stopper1->enabled || !stopper2->enabled) -@@ -255,8 +255,8 @@ retry: +@@ -255,8 +255,8 @@ static int cpu_stop_queue_two_works(int cpu1, struct cpu_stop_work *work1, __cpu_stop_queue_work(stopper1, work1); __cpu_stop_queue_work(stopper2, work2); unlock: @@ -3373,7 +3373,7 @@ index 4f1f69be1865..8c778fa33031 100644 --- a/sound/pci/emu10k1/memory.c +++ b/sound/pci/emu10k1/memory.c -@@ -237,13 +237,13 @@ __found_pages: +@@ -237,13 +237,13 @@ search_empty(struct snd_emu10k1 *emu, int size) static int is_valid_page(struct snd_emu10k1 *emu, dma_addr_t addr) { if (addr & ~emu->dma_mask) { 117 --- patch-4.9.117-118 2018-08-06 16:30:50.000000000 +0200 +++ 61_patch-4.9.118.quilt 2018-08-13 20:20:55.574906046 +0200 @@ -279,7 +279,7 @@ index 0a5f62e0efcc..487bf5b8f545 100644 --- a/drivers/net/phy/mdio-mux-bcm-iproc.c +++ b/drivers/net/phy/mdio-mux-bcm-iproc.c -@@ -218,7 +218,7 @@ out: +@@ -218,7 +218,7 @@ static int mdio_mux_iproc_probe(struct platform_device *pdev) static int mdio_mux_iproc_remove(struct platform_device *pdev) { @@ -445,7 +445,7 @@ TRACE("Number of ids %d\n", le16_to_cpu(sblk->no_ids)); TRACE("sblk->inode_table_start %llx\n", msblk->inode_table); TRACE("sblk->directory_table_start %llx\n", msblk->directory_table); -@@ -272,7 +273,7 @@ allocate_id_index_table: +@@ -272,7 +273,7 @@ static int squashfs_fill_super(struct super_block *sb, void *data, int silent) sb->s_export_op = &squashfs_export_ops; handle_fragments: @@ -573,7 +573,7 @@ index 4bf3b8af0257..752711cd4834 100644 --- a/net/ipv4/ip_fragment.c +++ b/net/ipv4/ip_fragment.c -@@ -446,11 +446,16 @@ found: +@@ -446,11 +446,16 @@ static int ip_frag_queue(struct ipq *qp, struct sk_buff *skb) int i = end - FRAG_CB(next)->offset; /* overlap is 'i' bytes */ if (i < next->len) { @@ -755,7 +755,7 @@ if (unlikely(tcp_try_rmem_schedule(sk, skb, skb->truesize))) { NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPOFODROP); -@@ -4734,7 +4738,7 @@ queue_and_out: +@@ -4734,7 +4738,7 @@ static void tcp_data_queue(struct sock *sk, struct sk_buff *skb) tcp_dsack_set(sk, TCP_SKB_CB(skb)->seq, TCP_SKB_CB(skb)->end_seq); out_of_window: @@ -764,7 +764,7 @@ inet_csk_schedule_ack(sk); drop: tcp_drop(sk, skb); -@@ -4745,8 +4749,6 @@ drop: +@@ -4745,8 +4749,6 @@ static void tcp_data_queue(struct sock *sk, struct sk_buff *skb) if (!before(TCP_SKB_CB(skb)->seq, tp->rcv_nxt + tcp_receive_window(tp))) goto out_of_window; 118 --- patch-4.9.118-119 2018-08-09 12:25:59.000000000 +0200 +++ 61_patch-4.9.119.quilt 2018-08-13 20:20:55.574906046 +0200 @@ -420,7 +420,7 @@ coalesce_done: tcp_grow_window(sk, skb); kfree_skb_partial(skb, fragstolen); -@@ -4543,7 +4561,8 @@ coalesce_done: +@@ -4543,7 +4561,8 @@ static void tcp_data_queue_ofo(struct sock *sk, struct sk_buff *skb) tcp_drop(sk, skb1); goto merge_right; } 119 --- patch-4.9.119-120 2018-08-15 18:25:08.000000000 +0200 +++ 61_patch-4.9.120.quilt 2018-08-21 18:03:28.309839659 +0200 @@ -2027,7 +2027,7 @@ static void __init spectre_v2_select_mitigation(void) { enum spectre_v2_mitigation_cmd cmd = spectre_v2_parse_cmdline(); -@@ -387,22 +381,15 @@ retpoline_auto: +@@ -387,22 +381,15 @@ static void __init spectre_v2_select_mitigation(void) pr_info("%s\n", spectre_v2_strings[mode]); /* @@ -2700,7 +2700,7 @@ /* representing HT siblings of each logical CPU */ DEFINE_PER_CPU_READ_MOSTLY(cpumask_var_t, cpu_sibling_map); -@@ -295,6 +289,23 @@ found: +@@ -295,6 +289,23 @@ int topology_update_package_map(unsigned int pkg, unsigned int cpu) return 0; } @@ -3298,7 +3298,7 @@ asm( /* Store host registers */ "push %%" _ASM_DX "; push %%" _ASM_BP ";" -@@ -9298,6 +9549,37 @@ free_vcpu: +@@ -9298,6 +9549,37 @@ static struct kvm_vcpu *vmx_create_vcpu(struct kvm *kvm, unsigned int id) return ERR_PTR(err); } @@ -4123,7 +4123,7 @@ index ae41623e0f13..0d4878efd643 100644 --- a/drivers/infiniband/hw/mlx4/mr.c +++ b/drivers/infiniband/hw/mlx4/mr.c -@@ -131,6 +131,40 @@ out: +@@ -131,6 +131,40 @@ int mlx4_ib_umem_write_mtt(struct mlx4_ib_dev *dev, struct mlx4_mtt *mtt, return err; } @@ -4630,7 +4630,7 @@ inode->i_mtime = inode->i_atime = inode->i_ctime = current_time(inode); inode->i_mode = table->mode; -@@ -466,6 +506,15 @@ out: +@@ -466,6 +506,15 @@ static struct inode *proc_sys_make_inode(struct super_block *sb, return inode; } @@ -4936,7 +4936,7 @@ /* * Prevent irq alloc/free while the dying cpu reorganizes the -@@ -956,20 +1045,19 @@ out: +@@ -956,20 +1045,19 @@ static int __ref _cpu_down(unsigned int cpu, int tasks_frozen, return ret; } 120