]> git.ipfire.org Git - thirdparty/kernel/stable-queue.git/commitdiff
Fixes for 6.6
authorSasha Levin <sashal@kernel.org>
Sat, 4 Jan 2025 14:21:12 +0000 (09:21 -0500)
committerSasha Levin <sashal@kernel.org>
Sat, 4 Jan 2025 14:21:12 +0000 (09:21 -0500)
Signed-off-by: Sasha Levin <sashal@kernel.org>
52 files changed:
queue-6.6/af_packet-fix-vlan_get_protocol_dgram-vs-msg_peek.patch [new file with mode: 0644]
queue-6.6/af_packet-fix-vlan_get_tci-vs-msg_peek.patch [new file with mode: 0644]
queue-6.6/alsa-usb-audio-us16x08-initialize-array-before-use.patch [new file with mode: 0644]
queue-6.6/drm-bridge-adv7511_audio-update-audio-infoframe-prop.patch [new file with mode: 0644]
queue-6.6/drm-i915-dg1-fix-power-gate-sequence.patch [new file with mode: 0644]
queue-6.6/eth-bcmsysport-fix-call-balance-of-priv-clk-handling.patch [new file with mode: 0644]
queue-6.6/ila-serialize-calls-to-nf_register_net_hooks.patch [new file with mode: 0644]
queue-6.6/ip_tunnel-annotate-data-races-around-t-parms.link.patch [new file with mode: 0644]
queue-6.6/ipv4-ip_tunnel-unmask-upper-dscp-bits-in-ip_md_tunne.patch [new file with mode: 0644]
queue-6.6/ipv4-ip_tunnel-unmask-upper-dscp-bits-in-ip_tunnel_b.patch [new file with mode: 0644]
queue-6.6/ipv4-ip_tunnel-unmask-upper-dscp-bits-in-ip_tunnel_x.patch [new file with mode: 0644]
queue-6.6/net-dsa-microchip-fix-ksz9477-set_ageing_time-functi.patch [new file with mode: 0644]
queue-6.6/net-dsa-microchip-fix-lan937x-set_ageing_time-functi.patch [new file with mode: 0644]
queue-6.6/net-fix-memory-leak-in-tcp_conn_request.patch [new file with mode: 0644]
queue-6.6/net-fix-netns-for-ip_tunnel_init_flow.patch [new file with mode: 0644]
queue-6.6/net-llc-reset-skb-transport_header.patch [new file with mode: 0644]
queue-6.6/net-mlx5-dr-select-msix-vector-0-for-completion-queu.patch [new file with mode: 0644]
queue-6.6/net-mlx5e-macsec-maintain-tx-sa-from-encoding_sa.patch [new file with mode: 0644]
queue-6.6/net-mlx5e-skip-restore-tc-rules-for-vport-rep-withou.patch [new file with mode: 0644]
queue-6.6/net-mv643xx_eth-fix-an-of-node-reference-leak.patch [new file with mode: 0644]
queue-6.6/net-reenable-netif_f_ipv6_csum-offload-for-big-tcp-p.patch [new file with mode: 0644]
queue-6.6/net-restrict-so_reuseport-to-inet-sockets.patch [new file with mode: 0644]
queue-6.6/net-sfc-correct-key_len-for-efx_tc_ct_zone_ht_params.patch [new file with mode: 0644]
queue-6.6/net-stmmac-don-t-create-a-mdio-bus-if-unnecessary.patch [new file with mode: 0644]
queue-6.6/net-stmmac-restructure-the-error-path-of-stmmac_prob.patch [new file with mode: 0644]
queue-6.6/net-ti-icssg-prueth-fix-clearing-of-iep_cmp_cfg-regi.patch [new file with mode: 0644]
queue-6.6/net-wwan-iosm-properly-check-for-valid-exec-stage-in.patch [new file with mode: 0644]
queue-6.6/net-wwan-t7xx-fix-fsm-command-timeout-issue.patch [new file with mode: 0644]
queue-6.6/netfilter-nft_set_hash-unaligned-atomic-read-on-stru.patch [new file with mode: 0644]
queue-6.6/netrom-check-buffer-length-before-accessing-it.patch [new file with mode: 0644]
queue-6.6/nvme-pci-512-byte-aligned-dma-pool-segment-quirk.patch [new file with mode: 0644]
queue-6.6/rdma-bnxt_re-add-check-for-path-mtu-in-modify_qp.patch [new file with mode: 0644]
queue-6.6/rdma-bnxt_re-add-send-queue-size-check-for-variable-.patch [new file with mode: 0644]
queue-6.6/rdma-bnxt_re-add-support-for-variable-wqe-in-genp7-a.patch [new file with mode: 0644]
queue-6.6/rdma-bnxt_re-allow-msn-table-capability-check.patch [new file with mode: 0644]
queue-6.6/rdma-bnxt_re-avoid-initializing-the-software-queue-f.patch [new file with mode: 0644]
queue-6.6/rdma-bnxt_re-avoid-sending-the-modify-qp-workaround-.patch [new file with mode: 0644]
queue-6.6/rdma-bnxt_re-disable-use-of-reserved-wqes.patch [new file with mode: 0644]
queue-6.6/rdma-bnxt_re-fix-max_qp_wrs-reported.patch [new file with mode: 0644]
queue-6.6/rdma-bnxt_re-fix-msn-table-size-for-variable-wqe-mod.patch [new file with mode: 0644]
queue-6.6/rdma-bnxt_re-fix-reporting-hw_ver-in-query_device.patch [new file with mode: 0644]
queue-6.6/rdma-bnxt_re-fix-the-check-for-9060-condition.patch [new file with mode: 0644]
queue-6.6/rdma-bnxt_re-fix-the-locking-while-accessing-the-qp-.patch [new file with mode: 0644]
queue-6.6/rdma-bnxt_re-remove-always-true-dattr-validity-check.patch [new file with mode: 0644]
queue-6.6/rdma-hns-fix-mapping-error-of-zero-hop-wqe-buffer.patch [new file with mode: 0644]
queue-6.6/rdma-hns-fix-missing-flush-cqe-for-dwqe.patch [new file with mode: 0644]
queue-6.6/rdma-hns-fix-warning-storm-caused-by-invalid-input-i.patch [new file with mode: 0644]
queue-6.6/rdma-hns-refactor-mtr-find.patch [new file with mode: 0644]
queue-6.6/rdma-hns-remove-unused-parameters-and-variables.patch [new file with mode: 0644]
queue-6.6/rdma-mlx5-enforce-same-type-port-association-for-mul.patch [new file with mode: 0644]
queue-6.6/rdma-rtrs-ensure-ib_sge-list-is-accessible.patch [new file with mode: 0644]
queue-6.6/series

diff --git a/queue-6.6/af_packet-fix-vlan_get_protocol_dgram-vs-msg_peek.patch b/queue-6.6/af_packet-fix-vlan_get_protocol_dgram-vs-msg_peek.patch
new file mode 100644 (file)
index 0000000..761e135
--- /dev/null
@@ -0,0 +1,149 @@
+From eb2cda0a4e2d230511b3fe0c7416a6c8fd588732 Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Mon, 30 Dec 2024 16:10:04 +0000
+Subject: af_packet: fix vlan_get_protocol_dgram() vs MSG_PEEK
+
+From: Eric Dumazet <edumazet@google.com>
+
+[ Upstream commit f91a5b8089389eb408501af2762f168c3aaa7b79 ]
+
+Blamed commit forgot MSG_PEEK case, allowing a crash [1] as found
+by syzbot.
+
+Rework vlan_get_protocol_dgram() to not touch skb at all,
+so that it can be used from many cpus on the same skb.
+
+Add a const qualifier to skb argument.
+
+[1]
+skbuff: skb_under_panic: text:ffffffff8a8ccd05 len:29 put:14 head:ffff88807fc8e400 data:ffff88807fc8e3f4 tail:0x11 end:0x140 dev:<NULL>
+------------[ cut here ]------------
+ kernel BUG at net/core/skbuff.c:206 !
+Oops: invalid opcode: 0000 [#1] PREEMPT SMP KASAN PTI
+CPU: 1 UID: 0 PID: 5892 Comm: syz-executor883 Not tainted 6.13.0-rc4-syzkaller-00054-gd6ef8b40d075 #0
+Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 09/13/2024
+ RIP: 0010:skb_panic net/core/skbuff.c:206 [inline]
+ RIP: 0010:skb_under_panic+0x14b/0x150 net/core/skbuff.c:216
+Code: 0b 8d 48 c7 c6 86 d5 25 8e 48 8b 54 24 08 8b 0c 24 44 8b 44 24 04 4d 89 e9 50 41 54 41 57 41 56 e8 5a 69 79 f7 48 83 c4 20 90 <0f> 0b 0f 1f 00 90 90 90 90 90 90 90 90 90 90 90 90 90 90 90 90 f3
+RSP: 0018:ffffc900038d7638 EFLAGS: 00010282
+RAX: 0000000000000087 RBX: dffffc0000000000 RCX: 609ffd18ea660600
+RDX: 0000000000000000 RSI: 0000000080000000 RDI: 0000000000000000
+RBP: ffff88802483c8d0 R08: ffffffff817f0a8c R09: 1ffff9200071ae60
+R10: dffffc0000000000 R11: fffff5200071ae61 R12: 0000000000000140
+R13: ffff88807fc8e400 R14: ffff88807fc8e3f4 R15: 0000000000000011
+FS:  00007fbac5e006c0(0000) GS:ffff8880b8700000(0000) knlGS:0000000000000000
+CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
+CR2: 00007fbac5e00d58 CR3: 000000001238e000 CR4: 00000000003526f0
+DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
+DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400
+Call Trace:
+ <TASK>
+  skb_push+0xe5/0x100 net/core/skbuff.c:2636
+  vlan_get_protocol_dgram+0x165/0x290 net/packet/af_packet.c:585
+  packet_recvmsg+0x948/0x1ef0 net/packet/af_packet.c:3552
+  sock_recvmsg_nosec net/socket.c:1033 [inline]
+  sock_recvmsg+0x22f/0x280 net/socket.c:1055
+  ____sys_recvmsg+0x1c6/0x480 net/socket.c:2803
+  ___sys_recvmsg net/socket.c:2845 [inline]
+  do_recvmmsg+0x426/0xab0 net/socket.c:2940
+  __sys_recvmmsg net/socket.c:3014 [inline]
+  __do_sys_recvmmsg net/socket.c:3037 [inline]
+  __se_sys_recvmmsg net/socket.c:3030 [inline]
+  __x64_sys_recvmmsg+0x199/0x250 net/socket.c:3030
+  do_syscall_x64 arch/x86/entry/common.c:52 [inline]
+  do_syscall_64+0xf3/0x230 arch/x86/entry/common.c:83
+ entry_SYSCALL_64_after_hwframe+0x77/0x7f
+
+Fixes: 79eecf631c14 ("af_packet: Handle outgoing VLAN packets without hardware offloading")
+Reported-by: syzbot+74f70bb1cb968bf09e4f@syzkaller.appspotmail.com
+Closes: https://lore.kernel.org/netdev/6772c485.050a0220.2f3838.04c5.GAE@google.com/T/#u
+Signed-off-by: Eric Dumazet <edumazet@google.com>
+Cc: Chengen Du <chengen.du@canonical.com>
+Reviewed-by: Willem de Bruijn <willemb@google.com>
+Link: https://patch.msgid.link/20241230161004.2681892-2-edumazet@google.com
+Signed-off-by: Jakub Kicinski <kuba@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ include/linux/if_vlan.h | 16 +++++++++++++---
+ net/packet/af_packet.c  | 16 ++++------------
+ 2 files changed, 17 insertions(+), 15 deletions(-)
+
+diff --git a/include/linux/if_vlan.h b/include/linux/if_vlan.h
+index 3028af87716e..430749a0f362 100644
+--- a/include/linux/if_vlan.h
++++ b/include/linux/if_vlan.h
+@@ -585,13 +585,16 @@ static inline int vlan_get_tag(const struct sk_buff *skb, u16 *vlan_tci)
+  * vlan_get_protocol - get protocol EtherType.
+  * @skb: skbuff to query
+  * @type: first vlan protocol
++ * @mac_offset: MAC offset
+  * @depth: buffer to store length of eth and vlan tags in bytes
+  *
+  * Returns the EtherType of the packet, regardless of whether it is
+  * vlan encapsulated (normal or hardware accelerated) or not.
+  */
+-static inline __be16 __vlan_get_protocol(const struct sk_buff *skb, __be16 type,
+-                                       int *depth)
++static inline __be16 __vlan_get_protocol_offset(const struct sk_buff *skb,
++                                              __be16 type,
++                                              int mac_offset,
++                                              int *depth)
+ {
+       unsigned int vlan_depth = skb->mac_len, parse_depth = VLAN_MAX_DEPTH;
+@@ -610,7 +613,8 @@ static inline __be16 __vlan_get_protocol(const struct sk_buff *skb, __be16 type,
+               do {
+                       struct vlan_hdr vhdr, *vh;
+-                      vh = skb_header_pointer(skb, vlan_depth, sizeof(vhdr), &vhdr);
++                      vh = skb_header_pointer(skb, mac_offset + vlan_depth,
++                                              sizeof(vhdr), &vhdr);
+                       if (unlikely(!vh || !--parse_depth))
+                               return 0;
+@@ -625,6 +629,12 @@ static inline __be16 __vlan_get_protocol(const struct sk_buff *skb, __be16 type,
+       return type;
+ }
++static inline __be16 __vlan_get_protocol(const struct sk_buff *skb, __be16 type,
++                                       int *depth)
++{
++      return __vlan_get_protocol_offset(skb, type, 0, depth);
++}
++
+ /**
+  * vlan_get_protocol - get protocol EtherType.
+  * @skb: skbuff to query
+diff --git a/net/packet/af_packet.c b/net/packet/af_packet.c
+index 96eca4a290ad..4abf7e9ac4f2 100644
+--- a/net/packet/af_packet.c
++++ b/net/packet/af_packet.c
+@@ -568,21 +568,13 @@ static u16 vlan_get_tci(const struct sk_buff *skb, struct net_device *dev)
+       return ntohs(vh->h_vlan_TCI);
+ }
+-static __be16 vlan_get_protocol_dgram(struct sk_buff *skb)
++static __be16 vlan_get_protocol_dgram(const struct sk_buff *skb)
+ {
+       __be16 proto = skb->protocol;
+-      if (unlikely(eth_type_vlan(proto))) {
+-              u8 *skb_orig_data = skb->data;
+-              int skb_orig_len = skb->len;
+-
+-              skb_push(skb, skb->data - skb_mac_header(skb));
+-              proto = __vlan_get_protocol(skb, proto, NULL);
+-              if (skb_orig_data != skb->data) {
+-                      skb->data = skb_orig_data;
+-                      skb->len = skb_orig_len;
+-              }
+-      }
++      if (unlikely(eth_type_vlan(proto)))
++              proto = __vlan_get_protocol_offset(skb, proto,
++                                                 skb_mac_offset(skb), NULL);
+       return proto;
+ }
+-- 
+2.39.5
+
diff --git a/queue-6.6/af_packet-fix-vlan_get_tci-vs-msg_peek.patch b/queue-6.6/af_packet-fix-vlan_get_tci-vs-msg_peek.patch
new file mode 100644 (file)
index 0000000..9e7c6ed
--- /dev/null
@@ -0,0 +1,102 @@
+From b5f1bc7a43cd382ae22b1498db5e2e90972c349f Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Mon, 30 Dec 2024 16:10:03 +0000
+Subject: af_packet: fix vlan_get_tci() vs MSG_PEEK
+
+From: Eric Dumazet <edumazet@google.com>
+
+[ Upstream commit 77ee7a6d16b6ec07b5c3ae2b6b60a24c1afbed09 ]
+
+Blamed commit forgot MSG_PEEK case, allowing a crash [1] as found
+by syzbot.
+
+Rework vlan_get_tci() to not touch skb at all,
+so that it can be used from many cpus on the same skb.
+
+Add a const qualifier to skb argument.
+
+[1]
+skbuff: skb_under_panic: text:ffffffff8a8da482 len:32 put:14 head:ffff88807a1d5800 data:ffff88807a1d5810 tail:0x14 end:0x140 dev:<NULL>
+------------[ cut here ]------------
+ kernel BUG at net/core/skbuff.c:206 !
+Oops: invalid opcode: 0000 [#1] PREEMPT SMP KASAN PTI
+CPU: 0 UID: 0 PID: 5880 Comm: syz-executor172 Not tainted 6.13.0-rc3-syzkaller-00762-g9268abe611b0 #0
+Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 09/13/2024
+ RIP: 0010:skb_panic net/core/skbuff.c:206 [inline]
+ RIP: 0010:skb_under_panic+0x14b/0x150 net/core/skbuff.c:216
+Code: 0b 8d 48 c7 c6 9e 6c 26 8e 48 8b 54 24 08 8b 0c 24 44 8b 44 24 04 4d 89 e9 50 41 54 41 57 41 56 e8 3a 5a 79 f7 48 83 c4 20 90 <0f> 0b 0f 1f 00 90 90 90 90 90 90 90 90 90 90 90 90 90 90 90 90 f3
+RSP: 0018:ffffc90003baf5b8 EFLAGS: 00010286
+RAX: 0000000000000087 RBX: dffffc0000000000 RCX: 8565c1eec37aa000
+RDX: 0000000000000000 RSI: 0000000080000000 RDI: 0000000000000000
+RBP: ffff88802616fb50 R08: ffffffff817f0a4c R09: 1ffff92000775e50
+R10: dffffc0000000000 R11: fffff52000775e51 R12: 0000000000000140
+R13: ffff88807a1d5800 R14: ffff88807a1d5810 R15: 0000000000000014
+FS:  00007fa03261f6c0(0000) GS:ffff8880b8600000(0000) knlGS:0000000000000000
+CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
+CR2: 00007ffd65753000 CR3: 0000000031720000 CR4: 00000000003526f0
+DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
+DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400
+Call Trace:
+ <TASK>
+  skb_push+0xe5/0x100 net/core/skbuff.c:2636
+  vlan_get_tci+0x272/0x550 net/packet/af_packet.c:565
+  packet_recvmsg+0x13c9/0x1ef0 net/packet/af_packet.c:3616
+  sock_recvmsg_nosec net/socket.c:1044 [inline]
+  sock_recvmsg+0x22f/0x280 net/socket.c:1066
+  ____sys_recvmsg+0x1c6/0x480 net/socket.c:2814
+  ___sys_recvmsg net/socket.c:2856 [inline]
+  do_recvmmsg+0x426/0xab0 net/socket.c:2951
+  __sys_recvmmsg net/socket.c:3025 [inline]
+  __do_sys_recvmmsg net/socket.c:3048 [inline]
+  __se_sys_recvmmsg net/socket.c:3041 [inline]
+  __x64_sys_recvmmsg+0x199/0x250 net/socket.c:3041
+  do_syscall_x64 arch/x86/entry/common.c:52 [inline]
+  do_syscall_64+0xf3/0x230 arch/x86/entry/common.c:83
+
+Fixes: 79eecf631c14 ("af_packet: Handle outgoing VLAN packets without hardware offloading")
+Reported-by: syzbot+8400677f3fd43f37d3bc@syzkaller.appspotmail.com
+Closes: https://lore.kernel.org/netdev/6772c485.050a0220.2f3838.04c6.GAE@google.com/T/#u
+Signed-off-by: Eric Dumazet <edumazet@google.com>
+Cc: Chengen Du <chengen.du@canonical.com>
+Reviewed-by: Willem de Bruijn <willemb@google.com>
+Link: https://patch.msgid.link/20241230161004.2681892-1-edumazet@google.com
+Signed-off-by: Jakub Kicinski <kuba@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ net/packet/af_packet.c | 12 +++---------
+ 1 file changed, 3 insertions(+), 9 deletions(-)
+
+diff --git a/net/packet/af_packet.c b/net/packet/af_packet.c
+index 56e3ae3b6be9..96eca4a290ad 100644
+--- a/net/packet/af_packet.c
++++ b/net/packet/af_packet.c
+@@ -538,10 +538,8 @@ static void *packet_current_frame(struct packet_sock *po,
+       return packet_lookup_frame(po, rb, rb->head, status);
+ }
+-static u16 vlan_get_tci(struct sk_buff *skb, struct net_device *dev)
++static u16 vlan_get_tci(const struct sk_buff *skb, struct net_device *dev)
+ {
+-      u8 *skb_orig_data = skb->data;
+-      int skb_orig_len = skb->len;
+       struct vlan_hdr vhdr, *vh;
+       unsigned int header_len;
+@@ -562,12 +560,8 @@ static u16 vlan_get_tci(struct sk_buff *skb, struct net_device *dev)
+       else
+               return 0;
+-      skb_push(skb, skb->data - skb_mac_header(skb));
+-      vh = skb_header_pointer(skb, header_len, sizeof(vhdr), &vhdr);
+-      if (skb_orig_data != skb->data) {
+-              skb->data = skb_orig_data;
+-              skb->len = skb_orig_len;
+-      }
++      vh = skb_header_pointer(skb, skb_mac_offset(skb) + header_len,
++                              sizeof(vhdr), &vhdr);
+       if (unlikely(!vh))
+               return 0;
+-- 
+2.39.5
+
diff --git a/queue-6.6/alsa-usb-audio-us16x08-initialize-array-before-use.patch b/queue-6.6/alsa-usb-audio-us16x08-initialize-array-before-use.patch
new file mode 100644 (file)
index 0000000..7b7b14b
--- /dev/null
@@ -0,0 +1,43 @@
+From ed8d674047bd3fd357fc9bc6f1c8295c82328923 Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Sun, 29 Dec 2024 11:32:42 +0530
+Subject: ALSA: usb-audio: US16x08: Initialize array before use
+
+From: Tanya Agarwal <tanyaagarwal25699@gmail.com>
+
+[ Upstream commit b06a6187ef983f501e93faa56209169752d3bde3 ]
+
+Initialize meter_urb array before use in mixer_us16x08.c.
+
+CID 1410197: (#1 of 1): Uninitialized scalar variable (UNINIT)
+uninit_use_in_call: Using uninitialized value *meter_urb when
+calling get_meter_levels_from_urb.
+
+Coverity Link:
+https://scan7.scan.coverity.com/#/project-view/52849/11354?selectedIssue=1410197
+
+Fixes: d2bb390a2081 ("ALSA: usb-audio: Tascam US-16x08 DSP mixer quirk")
+Signed-off-by: Tanya Agarwal <tanyaagarwal25699@gmail.com>
+Link: https://patch.msgid.link/20241229060240.1642-1-tanyaagarwal25699@gmail.com
+Signed-off-by: Takashi Iwai <tiwai@suse.de>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ sound/usb/mixer_us16x08.c | 2 +-
+ 1 file changed, 1 insertion(+), 1 deletion(-)
+
+diff --git a/sound/usb/mixer_us16x08.c b/sound/usb/mixer_us16x08.c
+index 6eb7d93b358d..20ac32635f1f 100644
+--- a/sound/usb/mixer_us16x08.c
++++ b/sound/usb/mixer_us16x08.c
+@@ -687,7 +687,7 @@ static int snd_us16x08_meter_get(struct snd_kcontrol *kcontrol,
+       struct usb_mixer_elem_info *elem = kcontrol->private_data;
+       struct snd_usb_audio *chip = elem->head.mixer->chip;
+       struct snd_us16x08_meter_store *store = elem->private_data;
+-      u8 meter_urb[64];
++      u8 meter_urb[64] = {0};
+       switch (kcontrol->private_value) {
+       case 0: {
+-- 
+2.39.5
+
diff --git a/queue-6.6/drm-bridge-adv7511_audio-update-audio-infoframe-prop.patch b/queue-6.6/drm-bridge-adv7511_audio-update-audio-infoframe-prop.patch
new file mode 100644 (file)
index 0000000..6e132a2
--- /dev/null
@@ -0,0 +1,74 @@
+From 42c84bd646aeaf9547843a83911d391144f1c2cc Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Tue, 19 Nov 2024 08:40:29 +0100
+Subject: drm/bridge: adv7511_audio: Update Audio InfoFrame properly
+
+From: Stefan Ekenberg <stefan.ekenberg@axis.com>
+
+[ Upstream commit 902806baf3c1e8383c1fe3ff0b6042b8cb5c2707 ]
+
+AUDIO_UPDATE bit (Bit 5 of MAIN register 0x4A) needs to be set to 1
+while updating Audio InfoFrame information and then set to 0 when done.
+Otherwise partially updated Audio InfoFrames could be sent out. Two
+cases where this rule were not followed are fixed:
+ - In adv7511_hdmi_hw_params() make sure AUDIO_UPDATE bit is updated
+   before/after setting ADV7511_REG_AUDIO_INFOFRAME.
+ - In audio_startup() use the correct register for clearing
+   AUDIO_UPDATE bit.
+
+The problem with corrupted audio infoframes were discovered by letting
+a HDMI logic analyser check the output of ADV7535.
+
+Note that this patchs replaces writing REG_GC(1) with
+REG_INFOFRAME_UPDATE. Bit 5 of REG_GC(1) is positioned within field
+GC_PP[3:0] and that field doesn't control audio infoframe and is read-
+only. My conclusion therefore was that the author if this code meant to
+clear bit 5 of REG_INFOFRAME_UPDATE from the very beginning.
+
+Tested-by: Biju Das <biju.das.jz@bp.renesas.com>
+Fixes: 53c515befe28 ("drm/bridge: adv7511: Add Audio support")
+Signed-off-by: Stefan Ekenberg <stefan.ekenberg@axis.com>
+Reviewed-by: Dmitry Baryshkov <dmitry.baryshkov@linaro.org>
+Link: https://patchwork.freedesktop.org/patch/msgid/20241119-adv7511-audio-info-frame-v4-1-4ae68e76c89c@axis.com
+Signed-off-by: Dmitry Baryshkov <dmitry.baryshkov@linaro.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/gpu/drm/bridge/adv7511/adv7511_audio.c | 14 ++++++++++++--
+ 1 file changed, 12 insertions(+), 2 deletions(-)
+
+diff --git a/drivers/gpu/drm/bridge/adv7511/adv7511_audio.c b/drivers/gpu/drm/bridge/adv7511/adv7511_audio.c
+index 61f4a38e7d2b..8f786592143b 100644
+--- a/drivers/gpu/drm/bridge/adv7511/adv7511_audio.c
++++ b/drivers/gpu/drm/bridge/adv7511/adv7511_audio.c
+@@ -153,7 +153,16 @@ static int adv7511_hdmi_hw_params(struct device *dev, void *data,
+                          ADV7511_AUDIO_CFG3_LEN_MASK, len);
+       regmap_update_bits(adv7511->regmap, ADV7511_REG_I2C_FREQ_ID_CFG,
+                          ADV7511_I2C_FREQ_ID_CFG_RATE_MASK, rate << 4);
+-      regmap_write(adv7511->regmap, 0x73, 0x1);
++
++      /* send current Audio infoframe values while updating */
++      regmap_update_bits(adv7511->regmap, ADV7511_REG_INFOFRAME_UPDATE,
++                         BIT(5), BIT(5));
++
++      regmap_write(adv7511->regmap, ADV7511_REG_AUDIO_INFOFRAME(0), 0x1);
++
++      /* use Audio infoframe updated info */
++      regmap_update_bits(adv7511->regmap, ADV7511_REG_INFOFRAME_UPDATE,
++                         BIT(5), 0);
+       return 0;
+ }
+@@ -184,8 +193,9 @@ static int audio_startup(struct device *dev, void *data)
+       regmap_update_bits(adv7511->regmap, ADV7511_REG_GC(0),
+                               BIT(7) | BIT(6), BIT(7));
+       /* use Audio infoframe updated info */
+-      regmap_update_bits(adv7511->regmap, ADV7511_REG_GC(1),
++      regmap_update_bits(adv7511->regmap, ADV7511_REG_INFOFRAME_UPDATE,
+                               BIT(5), 0);
++
+       /* enable SPDIF receiver */
+       if (adv7511->audio_source == ADV7511_AUDIO_SOURCE_SPDIF)
+               regmap_update_bits(adv7511->regmap, ADV7511_REG_AUDIO_CONFIG,
+-- 
+2.39.5
+
diff --git a/queue-6.6/drm-i915-dg1-fix-power-gate-sequence.patch b/queue-6.6/drm-i915-dg1-fix-power-gate-sequence.patch
new file mode 100644 (file)
index 0000000..12b6e4f
--- /dev/null
@@ -0,0 +1,45 @@
+From c5180df8616181b2a8533d36c2e704c34c8e4842 Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Thu, 19 Dec 2024 16:00:19 -0500
+Subject: drm/i915/dg1: Fix power gate sequence.
+
+From: Rodrigo Vivi <rodrigo.vivi@intel.com>
+
+[ Upstream commit 20e7c5313ffbf11c34a46395345677adbe890bee ]
+
+sub-pipe PG is not present on DG1. Setting these bits can disable
+other power gates and cause GPU hangs on video playbacks.
+
+VLK: 16314, 4304
+
+Closes: https://gitlab.freedesktop.org/drm/i915/kernel/-/issues/13381
+Fixes: 85a12d7eb8fe ("drm/i915/tgl: Fix Media power gate sequence.")
+Cc: Vinay Belgaumkar <vinay.belgaumkar@intel.com>
+Cc: Himal Prasad Ghimiray <himal.prasad.ghimiray@intel.com>
+Reviewed-by: Vinay Belgaumkar <vinay.belgaumkar@intel.com>
+Reviewed-by: Himal Prasad Ghimiray <himal.prasad.ghimiray@intel.com>
+Link: https://patchwork.freedesktop.org/patch/msgid/20241219210019.70532-1-rodrigo.vivi@intel.com
+Signed-off-by: Rodrigo Vivi <rodrigo.vivi@intel.com>
+(cherry picked from commit de7061947b4ed4be857d452c60d5fb795831d79e)
+Signed-off-by: Tvrtko Ursulin <tursulin@ursulin.net>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/gpu/drm/i915/gt/intel_rc6.c | 2 +-
+ 1 file changed, 1 insertion(+), 1 deletion(-)
+
+diff --git a/drivers/gpu/drm/i915/gt/intel_rc6.c b/drivers/gpu/drm/i915/gt/intel_rc6.c
+index 9e113e947326..6e8c182b2559 100644
+--- a/drivers/gpu/drm/i915/gt/intel_rc6.c
++++ b/drivers/gpu/drm/i915/gt/intel_rc6.c
+@@ -133,7 +133,7 @@ static void gen11_rc6_enable(struct intel_rc6 *rc6)
+                       GEN9_MEDIA_PG_ENABLE |
+                       GEN11_MEDIA_SAMPLER_PG_ENABLE;
+-      if (GRAPHICS_VER(gt->i915) >= 12) {
++      if (GRAPHICS_VER(gt->i915) >= 12 && !IS_DG1(gt->i915)) {
+               for (i = 0; i < I915_MAX_VCS; i++)
+                       if (HAS_ENGINE(gt, _VCS(i)))
+                               pg_enable |= (VDN_HCP_POWERGATE_ENABLE(i) |
+-- 
+2.39.5
+
diff --git a/queue-6.6/eth-bcmsysport-fix-call-balance-of-priv-clk-handling.patch b/queue-6.6/eth-bcmsysport-fix-call-balance-of-priv-clk-handling.patch
new file mode 100644 (file)
index 0000000..939b12b
--- /dev/null
@@ -0,0 +1,82 @@
+From 851a0bd790818341e11bb33f5a07b8f4ff6e9d1d Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Fri, 27 Dec 2024 15:30:07 +0300
+Subject: eth: bcmsysport: fix call balance of priv->clk handling routines
+
+From: Vitalii Mordan <mordan@ispras.ru>
+
+[ Upstream commit b255ef45fcc2141c1bf98456796abb956d843a27 ]
+
+Check the return value of clk_prepare_enable to ensure that priv->clk has
+been successfully enabled.
+
+If priv->clk was not enabled during bcm_sysport_probe, bcm_sysport_resume,
+or bcm_sysport_open, it must not be disabled in any subsequent execution
+paths.
+
+Fixes: 31bc72d97656 ("net: systemport: fetch and use clock resources")
+Signed-off-by: Vitalii Mordan <mordan@ispras.ru>
+Reviewed-by: Florian Fainelli <florian.fainelli@broadcom.com>
+Link: https://patch.msgid.link/20241227123007.2333397-1-mordan@ispras.ru
+Signed-off-by: Jakub Kicinski <kuba@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/net/ethernet/broadcom/bcmsysport.c | 21 ++++++++++++++++++---
+ 1 file changed, 18 insertions(+), 3 deletions(-)
+
+diff --git a/drivers/net/ethernet/broadcom/bcmsysport.c b/drivers/net/ethernet/broadcom/bcmsysport.c
+index 49e890a7e04a..23cc2d85994e 100644
+--- a/drivers/net/ethernet/broadcom/bcmsysport.c
++++ b/drivers/net/ethernet/broadcom/bcmsysport.c
+@@ -1967,7 +1967,11 @@ static int bcm_sysport_open(struct net_device *dev)
+       unsigned int i;
+       int ret;
+-      clk_prepare_enable(priv->clk);
++      ret = clk_prepare_enable(priv->clk);
++      if (ret) {
++              netdev_err(dev, "could not enable priv clock\n");
++              return ret;
++      }
+       /* Reset UniMAC */
+       umac_reset(priv);
+@@ -2625,7 +2629,11 @@ static int bcm_sysport_probe(struct platform_device *pdev)
+               goto err_deregister_notifier;
+       }
+-      clk_prepare_enable(priv->clk);
++      ret = clk_prepare_enable(priv->clk);
++      if (ret) {
++              dev_err(&pdev->dev, "could not enable priv clock\n");
++              goto err_deregister_netdev;
++      }
+       priv->rev = topctrl_readl(priv, REV_CNTL) & REV_MASK;
+       dev_info(&pdev->dev,
+@@ -2639,6 +2647,8 @@ static int bcm_sysport_probe(struct platform_device *pdev)
+       return 0;
++err_deregister_netdev:
++      unregister_netdev(dev);
+ err_deregister_notifier:
+       unregister_netdevice_notifier(&priv->netdev_notifier);
+ err_deregister_fixed_link:
+@@ -2810,7 +2820,12 @@ static int __maybe_unused bcm_sysport_resume(struct device *d)
+       if (!netif_running(dev))
+               return 0;
+-      clk_prepare_enable(priv->clk);
++      ret = clk_prepare_enable(priv->clk);
++      if (ret) {
++              netdev_err(dev, "could not enable priv clock\n");
++              return ret;
++      }
++
+       if (priv->wolopts)
+               clk_disable_unprepare(priv->wol_clk);
+-- 
+2.39.5
+
diff --git a/queue-6.6/ila-serialize-calls-to-nf_register_net_hooks.patch b/queue-6.6/ila-serialize-calls-to-nf_register_net_hooks.patch
new file mode 100644 (file)
index 0000000..2eded38
--- /dev/null
@@ -0,0 +1,113 @@
+From cf7dbdddc2f63b3b9111d5fad7e067f99175e174 Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Mon, 30 Dec 2024 16:28:49 +0000
+Subject: ila: serialize calls to nf_register_net_hooks()
+
+From: Eric Dumazet <edumazet@google.com>
+
+[ Upstream commit 260466b576bca0081a7d4acecc8e93687aa22d0e ]
+
+syzbot found a race in ila_add_mapping() [1]
+
+commit 031ae72825ce ("ila: call nf_unregister_net_hooks() sooner")
+attempted to fix a similar issue.
+
+Looking at the syzbot repro, we have concurrent ILA_CMD_ADD commands.
+
+Add a mutex to make sure at most one thread is calling nf_register_net_hooks().
+
+[1]
+ BUG: KASAN: slab-use-after-free in rht_key_hashfn include/linux/rhashtable.h:159 [inline]
+ BUG: KASAN: slab-use-after-free in __rhashtable_lookup.constprop.0+0x426/0x550 include/linux/rhashtable.h:604
+Read of size 4 at addr ffff888028f40008 by task dhcpcd/5501
+
+CPU: 1 UID: 0 PID: 5501 Comm: dhcpcd Not tainted 6.13.0-rc4-syzkaller-00054-gd6ef8b40d075 #0
+Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 09/13/2024
+Call Trace:
+ <IRQ>
+  __dump_stack lib/dump_stack.c:94 [inline]
+  dump_stack_lvl+0x116/0x1f0 lib/dump_stack.c:120
+  print_address_description mm/kasan/report.c:378 [inline]
+  print_report+0xc3/0x620 mm/kasan/report.c:489
+  kasan_report+0xd9/0x110 mm/kasan/report.c:602
+  rht_key_hashfn include/linux/rhashtable.h:159 [inline]
+  __rhashtable_lookup.constprop.0+0x426/0x550 include/linux/rhashtable.h:604
+  rhashtable_lookup include/linux/rhashtable.h:646 [inline]
+  rhashtable_lookup_fast include/linux/rhashtable.h:672 [inline]
+  ila_lookup_wildcards net/ipv6/ila/ila_xlat.c:127 [inline]
+  ila_xlat_addr net/ipv6/ila/ila_xlat.c:652 [inline]
+  ila_nf_input+0x1ee/0x620 net/ipv6/ila/ila_xlat.c:185
+  nf_hook_entry_hookfn include/linux/netfilter.h:154 [inline]
+  nf_hook_slow+0xbb/0x200 net/netfilter/core.c:626
+  nf_hook.constprop.0+0x42e/0x750 include/linux/netfilter.h:269
+  NF_HOOK include/linux/netfilter.h:312 [inline]
+  ipv6_rcv+0xa4/0x680 net/ipv6/ip6_input.c:309
+  __netif_receive_skb_one_core+0x12e/0x1e0 net/core/dev.c:5672
+  __netif_receive_skb+0x1d/0x160 net/core/dev.c:5785
+  process_backlog+0x443/0x15f0 net/core/dev.c:6117
+  __napi_poll.constprop.0+0xb7/0x550 net/core/dev.c:6883
+  napi_poll net/core/dev.c:6952 [inline]
+  net_rx_action+0xa94/0x1010 net/core/dev.c:7074
+  handle_softirqs+0x213/0x8f0 kernel/softirq.c:561
+  __do_softirq kernel/softirq.c:595 [inline]
+  invoke_softirq kernel/softirq.c:435 [inline]
+  __irq_exit_rcu+0x109/0x170 kernel/softirq.c:662
+  irq_exit_rcu+0x9/0x30 kernel/softirq.c:678
+  instr_sysvec_apic_timer_interrupt arch/x86/kernel/apic/apic.c:1049 [inline]
+  sysvec_apic_timer_interrupt+0xa4/0xc0 arch/x86/kernel/apic/apic.c:1049
+
+Fixes: 7f00feaf1076 ("ila: Add generic ILA translation facility")
+Reported-by: syzbot+47e761d22ecf745f72b9@syzkaller.appspotmail.com
+Closes: https://lore.kernel.org/netdev/6772c9ae.050a0220.2f3838.04c7.GAE@google.com/T/#u
+Signed-off-by: Eric Dumazet <edumazet@google.com>
+Cc: Florian Westphal <fw@strlen.de>
+Cc: Tom Herbert <tom@herbertland.com>
+Link: https://patch.msgid.link/20241230162849.2795486-1-edumazet@google.com
+Signed-off-by: Jakub Kicinski <kuba@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ net/ipv6/ila/ila_xlat.c | 16 +++++++++++-----
+ 1 file changed, 11 insertions(+), 5 deletions(-)
+
+diff --git a/net/ipv6/ila/ila_xlat.c b/net/ipv6/ila/ila_xlat.c
+index 534a4498e280..fff09f5a796a 100644
+--- a/net/ipv6/ila/ila_xlat.c
++++ b/net/ipv6/ila/ila_xlat.c
+@@ -200,6 +200,8 @@ static const struct nf_hook_ops ila_nf_hook_ops[] = {
+       },
+ };
++static DEFINE_MUTEX(ila_mutex);
++
+ static int ila_add_mapping(struct net *net, struct ila_xlat_params *xp)
+ {
+       struct ila_net *ilan = net_generic(net, ila_net_id);
+@@ -207,16 +209,20 @@ static int ila_add_mapping(struct net *net, struct ila_xlat_params *xp)
+       spinlock_t *lock = ila_get_lock(ilan, xp->ip.locator_match);
+       int err = 0, order;
+-      if (!ilan->xlat.hooks_registered) {
++      if (!READ_ONCE(ilan->xlat.hooks_registered)) {
+               /* We defer registering net hooks in the namespace until the
+                * first mapping is added.
+                */
+-              err = nf_register_net_hooks(net, ila_nf_hook_ops,
+-                                          ARRAY_SIZE(ila_nf_hook_ops));
++              mutex_lock(&ila_mutex);
++              if (!ilan->xlat.hooks_registered) {
++                      err = nf_register_net_hooks(net, ila_nf_hook_ops,
++                                              ARRAY_SIZE(ila_nf_hook_ops));
++                      if (!err)
++                              WRITE_ONCE(ilan->xlat.hooks_registered, true);
++              }
++              mutex_unlock(&ila_mutex);
+               if (err)
+                       return err;
+-
+-              ilan->xlat.hooks_registered = true;
+       }
+       ila = kzalloc(sizeof(*ila), GFP_KERNEL);
+-- 
+2.39.5
+
diff --git a/queue-6.6/ip_tunnel-annotate-data-races-around-t-parms.link.patch b/queue-6.6/ip_tunnel-annotate-data-races-around-t-parms.link.patch
new file mode 100644 (file)
index 0000000..4b49ef5
--- /dev/null
@@ -0,0 +1,115 @@
+From 7ea3f45c52bb745cde58e68f9b13b8b11844ace0 Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Tue, 13 Feb 2024 06:32:34 +0000
+Subject: ip_tunnel: annotate data-races around t->parms.link
+
+From: Eric Dumazet <edumazet@google.com>
+
+[ Upstream commit f694eee9e1c00d6ca06c5e59c04e3b6ff7d64aa9 ]
+
+t->parms.link is read locklessly, annotate these reads
+and opposite writes accordingly.
+
+Signed-off-by: Eric Dumazet <edumazet@google.com>
+Signed-off-by: David S. Miller <davem@davemloft.net>
+Stable-dep-of: b5a7b661a073 ("net: Fix netns for ip_tunnel_init_flow()")
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ net/ipv4/ip_tunnel.c | 27 +++++++++++++--------------
+ 1 file changed, 13 insertions(+), 14 deletions(-)
+
+diff --git a/net/ipv4/ip_tunnel.c b/net/ipv4/ip_tunnel.c
+index 72b2d68ef4da..0f5cfe3caa2e 100644
+--- a/net/ipv4/ip_tunnel.c
++++ b/net/ipv4/ip_tunnel.c
+@@ -102,10 +102,9 @@ struct ip_tunnel *ip_tunnel_lookup(struct ip_tunnel_net *itn,
+               if (!ip_tunnel_key_match(&t->parms, flags, key))
+                       continue;
+-              if (t->parms.link == link)
++              if (READ_ONCE(t->parms.link) == link)
+                       return t;
+-              else
+-                      cand = t;
++              cand = t;
+       }
+       hlist_for_each_entry_rcu(t, head, hash_node) {
+@@ -117,9 +116,9 @@ struct ip_tunnel *ip_tunnel_lookup(struct ip_tunnel_net *itn,
+               if (!ip_tunnel_key_match(&t->parms, flags, key))
+                       continue;
+-              if (t->parms.link == link)
++              if (READ_ONCE(t->parms.link) == link)
+                       return t;
+-              else if (!cand)
++              if (!cand)
+                       cand = t;
+       }
+@@ -137,9 +136,9 @@ struct ip_tunnel *ip_tunnel_lookup(struct ip_tunnel_net *itn,
+               if (!ip_tunnel_key_match(&t->parms, flags, key))
+                       continue;
+-              if (t->parms.link == link)
++              if (READ_ONCE(t->parms.link) == link)
+                       return t;
+-              else if (!cand)
++              if (!cand)
+                       cand = t;
+       }
+@@ -150,9 +149,9 @@ struct ip_tunnel *ip_tunnel_lookup(struct ip_tunnel_net *itn,
+                   !(t->dev->flags & IFF_UP))
+                       continue;
+-              if (t->parms.link == link)
++              if (READ_ONCE(t->parms.link) == link)
+                       return t;
+-              else if (!cand)
++              if (!cand)
+                       cand = t;
+       }
+@@ -221,7 +220,7 @@ static struct ip_tunnel *ip_tunnel_find(struct ip_tunnel_net *itn,
+       hlist_for_each_entry_rcu(t, head, hash_node) {
+               if (local == t->parms.iph.saddr &&
+                   remote == t->parms.iph.daddr &&
+-                  link == t->parms.link &&
++                  link == READ_ONCE(t->parms.link) &&
+                   type == t->dev->type &&
+                   ip_tunnel_key_match(&t->parms, flags, key))
+                       break;
+@@ -774,7 +773,7 @@ void ip_tunnel_xmit(struct sk_buff *skb, struct net_device *dev,
+       ip_tunnel_init_flow(&fl4, protocol, dst, tnl_params->saddr,
+                           tunnel->parms.o_key, RT_TOS(tos),
+-                          dev_net(dev), tunnel->parms.link,
++                          dev_net(dev), READ_ONCE(tunnel->parms.link),
+                           tunnel->fwmark, skb_get_hash(skb), 0);
+       if (ip_tunnel_encap(skb, &tunnel->encap, &protocol, &fl4) < 0)
+@@ -894,7 +893,7 @@ static void ip_tunnel_update(struct ip_tunnel_net *itn,
+       if (t->parms.link != p->link || t->fwmark != fwmark) {
+               int mtu;
+-              t->parms.link = p->link;
++              WRITE_ONCE(t->parms.link, p->link);
+               t->fwmark = fwmark;
+               mtu = ip_tunnel_bind_dev(dev);
+               if (set_mtu)
+@@ -1084,9 +1083,9 @@ EXPORT_SYMBOL(ip_tunnel_get_link_net);
+ int ip_tunnel_get_iflink(const struct net_device *dev)
+ {
+-      struct ip_tunnel *tunnel = netdev_priv(dev);
++      const struct ip_tunnel *tunnel = netdev_priv(dev);
+-      return tunnel->parms.link;
++      return READ_ONCE(tunnel->parms.link);
+ }
+ EXPORT_SYMBOL(ip_tunnel_get_iflink);
+-- 
+2.39.5
+
diff --git a/queue-6.6/ipv4-ip_tunnel-unmask-upper-dscp-bits-in-ip_md_tunne.patch b/queue-6.6/ipv4-ip_tunnel-unmask-upper-dscp-bits-in-ip_md_tunne.patch
new file mode 100644 (file)
index 0000000..fe3ec49
--- /dev/null
@@ -0,0 +1,54 @@
+From f419f24263a4857557f2a5ac415f8bb53e54b528 Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Thu, 5 Sep 2024 19:51:34 +0300
+Subject: ipv4: ip_tunnel: Unmask upper DSCP bits in ip_md_tunnel_xmit()
+
+From: Ido Schimmel <idosch@nvidia.com>
+
+[ Upstream commit c34cfe72bb260fc49660d9e6a9ba95ba01669ae2 ]
+
+Unmask the upper DSCP bits when initializing an IPv4 flow key via
+ip_tunnel_init_flow() before passing it to ip_route_output_key() so that
+in the future we could perform the FIB lookup according to the full DSCP
+value.
+
+Note that the 'tos' variable includes the full DS field. Either the one
+specified via the tunnel key or the one inherited from the inner packet.
+
+Signed-off-by: Ido Schimmel <idosch@nvidia.com>
+Reviewed-by: Guillaume Nault <gnault@redhat.com>
+Signed-off-by: David S. Miller <davem@davemloft.net>
+Stable-dep-of: b5a7b661a073 ("net: Fix netns for ip_tunnel_init_flow()")
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ net/ipv4/ip_tunnel.c | 7 ++++---
+ 1 file changed, 4 insertions(+), 3 deletions(-)
+
+diff --git a/net/ipv4/ip_tunnel.c b/net/ipv4/ip_tunnel.c
+index 571cf7c2fa28..b5437755365c 100644
+--- a/net/ipv4/ip_tunnel.c
++++ b/net/ipv4/ip_tunnel.c
+@@ -43,6 +43,7 @@
+ #include <net/rtnetlink.h>
+ #include <net/udp.h>
+ #include <net/dst_metadata.h>
++#include <net/inet_dscp.h>
+ #if IS_ENABLED(CONFIG_IPV6)
+ #include <net/ipv6.h>
+@@ -609,9 +610,9 @@ void ip_md_tunnel_xmit(struct sk_buff *skb, struct net_device *dev,
+                       tos = ipv6_get_dsfield((const struct ipv6hdr *)inner_iph);
+       }
+       ip_tunnel_init_flow(&fl4, proto, key->u.ipv4.dst, key->u.ipv4.src,
+-                          tunnel_id_to_key32(key->tun_id), RT_TOS(tos),
+-                          dev_net(dev), 0, skb->mark, skb_get_hash(skb),
+-                          key->flow_flags);
++                          tunnel_id_to_key32(key->tun_id),
++                          tos & INET_DSCP_MASK, dev_net(dev), 0, skb->mark,
++                          skb_get_hash(skb), key->flow_flags);
+       if (!tunnel_hlen)
+               tunnel_hlen = ip_encap_hlen(&tun_info->encap);
+-- 
+2.39.5
+
diff --git a/queue-6.6/ipv4-ip_tunnel-unmask-upper-dscp-bits-in-ip_tunnel_b.patch b/queue-6.6/ipv4-ip_tunnel-unmask-upper-dscp-bits-in-ip_tunnel_b.patch
new file mode 100644 (file)
index 0000000..beb8f26
--- /dev/null
@@ -0,0 +1,39 @@
+From f47a26de9a493bf77ebbe807911ad2ddd8b6f059 Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Thu, 5 Sep 2024 19:51:33 +0300
+Subject: ipv4: ip_tunnel: Unmask upper DSCP bits in ip_tunnel_bind_dev()
+
+From: Ido Schimmel <idosch@nvidia.com>
+
+[ Upstream commit e7191e517a03d025405c7df730b400ad4118474e ]
+
+Unmask the upper DSCP bits when initializing an IPv4 flow key via
+ip_tunnel_init_flow() before passing it to ip_route_output_key() so that
+in the future we could perform the FIB lookup according to the full DSCP
+value.
+
+Signed-off-by: Ido Schimmel <idosch@nvidia.com>
+Reviewed-by: Guillaume Nault <gnault@redhat.com>
+Signed-off-by: David S. Miller <davem@davemloft.net>
+Stable-dep-of: b5a7b661a073 ("net: Fix netns for ip_tunnel_init_flow()")
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ net/ipv4/ip_tunnel.c | 2 +-
+ 1 file changed, 1 insertion(+), 1 deletion(-)
+
+diff --git a/net/ipv4/ip_tunnel.c b/net/ipv4/ip_tunnel.c
+index 0f5cfe3caa2e..571cf7c2fa28 100644
+--- a/net/ipv4/ip_tunnel.c
++++ b/net/ipv4/ip_tunnel.c
+@@ -293,7 +293,7 @@ static int ip_tunnel_bind_dev(struct net_device *dev)
+               ip_tunnel_init_flow(&fl4, iph->protocol, iph->daddr,
+                                   iph->saddr, tunnel->parms.o_key,
+-                                  RT_TOS(iph->tos), dev_net(dev),
++                                  iph->tos & INET_DSCP_MASK, dev_net(dev),
+                                   tunnel->parms.link, tunnel->fwmark, 0, 0);
+               rt = ip_route_output_key(tunnel->net, &fl4);
+-- 
+2.39.5
+
diff --git a/queue-6.6/ipv4-ip_tunnel-unmask-upper-dscp-bits-in-ip_tunnel_x.patch b/queue-6.6/ipv4-ip_tunnel-unmask-upper-dscp-bits-in-ip_tunnel_x.patch
new file mode 100644 (file)
index 0000000..fdf974c
--- /dev/null
@@ -0,0 +1,43 @@
+From 403599989afd031520cf277709611e59eaa92db6 Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Thu, 5 Sep 2024 19:51:35 +0300
+Subject: ipv4: ip_tunnel: Unmask upper DSCP bits in ip_tunnel_xmit()
+
+From: Ido Schimmel <idosch@nvidia.com>
+
+[ Upstream commit c2b639f9f3b7a058ca9c7349b096f355773f2cd8 ]
+
+Unmask the upper DSCP bits when initializing an IPv4 flow key via
+ip_tunnel_init_flow() before passing it to ip_route_output_key() so that
+in the future we could perform the FIB lookup according to the full DSCP
+value.
+
+Note that the 'tos' variable includes the full DS field. Either the one
+specified as part of the tunnel parameters or the one inherited from the
+inner packet.
+
+Signed-off-by: Ido Schimmel <idosch@nvidia.com>
+Reviewed-by: Guillaume Nault <gnault@redhat.com>
+Signed-off-by: David S. Miller <davem@davemloft.net>
+Stable-dep-of: b5a7b661a073 ("net: Fix netns for ip_tunnel_init_flow()")
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ net/ipv4/ip_tunnel.c | 2 +-
+ 1 file changed, 1 insertion(+), 1 deletion(-)
+
+diff --git a/net/ipv4/ip_tunnel.c b/net/ipv4/ip_tunnel.c
+index b5437755365c..fd8923561b18 100644
+--- a/net/ipv4/ip_tunnel.c
++++ b/net/ipv4/ip_tunnel.c
+@@ -773,7 +773,7 @@ void ip_tunnel_xmit(struct sk_buff *skb, struct net_device *dev,
+       }
+       ip_tunnel_init_flow(&fl4, protocol, dst, tnl_params->saddr,
+-                          tunnel->parms.o_key, RT_TOS(tos),
++                          tunnel->parms.o_key, tos & INET_DSCP_MASK,
+                           dev_net(dev), READ_ONCE(tunnel->parms.link),
+                           tunnel->fwmark, skb_get_hash(skb), 0);
+-- 
+2.39.5
+
diff --git a/queue-6.6/net-dsa-microchip-fix-ksz9477-set_ageing_time-functi.patch b/queue-6.6/net-dsa-microchip-fix-ksz9477-set_ageing_time-functi.patch
new file mode 100644 (file)
index 0000000..63df89f
--- /dev/null
@@ -0,0 +1,125 @@
+From 2f41924c9360698d52faddb92bc78724a2d69d98 Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Tue, 17 Dec 2024 18:02:23 -0800
+Subject: net: dsa: microchip: Fix KSZ9477 set_ageing_time function
+
+From: Tristram Ha <tristram.ha@microchip.com>
+
+[ Upstream commit 262bfba8ab820641c8cfbbf03b86d6c00242c078 ]
+
+The aging count is not a simple 11-bit value but comprises a 3-bit
+multiplier and an 8-bit second count.  The code tries to use the
+original multiplier which is 4 as the second count is still 300 seconds
+by default.
+
+Fixes: 2c119d9982b1 ("net: dsa: microchip: add the support for set_ageing_time")
+Signed-off-by: Tristram Ha <tristram.ha@microchip.com>
+Reviewed-by: Andrew Lunn <andrew@lunn.ch>
+Link: https://patch.msgid.link/20241218020224.70590-2-Tristram.Ha@microchip.com
+Signed-off-by: Jakub Kicinski <kuba@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/net/dsa/microchip/ksz9477.c     | 47 +++++++++++++++++++------
+ drivers/net/dsa/microchip/ksz9477_reg.h |  4 +--
+ 2 files changed, 37 insertions(+), 14 deletions(-)
+
+diff --git a/drivers/net/dsa/microchip/ksz9477.c b/drivers/net/dsa/microchip/ksz9477.c
+index a7e8fcdf2576..59134d117846 100644
+--- a/drivers/net/dsa/microchip/ksz9477.c
++++ b/drivers/net/dsa/microchip/ksz9477.c
+@@ -2,7 +2,7 @@
+ /*
+  * Microchip KSZ9477 switch driver main logic
+  *
+- * Copyright (C) 2017-2019 Microchip Technology Inc.
++ * Copyright (C) 2017-2024 Microchip Technology Inc.
+  */
+ #include <linux/kernel.h>
+@@ -916,26 +916,51 @@ void ksz9477_get_caps(struct ksz_device *dev, int port,
+ int ksz9477_set_ageing_time(struct ksz_device *dev, unsigned int msecs)
+ {
+       u32 secs = msecs / 1000;
+-      u8 value;
+-      u8 data;
++      u8 data, mult, value;
++      u32 max_val;
+       int ret;
+-      value = FIELD_GET(SW_AGE_PERIOD_7_0_M, secs);
++#define MAX_TIMER_VAL ((1 << 8) - 1)
+-      ret = ksz_write8(dev, REG_SW_LUE_CTRL_3, value);
+-      if (ret < 0)
+-              return ret;
++      /* The aging timer comprises a 3-bit multiplier and an 8-bit second
++       * value.  Either of them cannot be zero.  The maximum timer is then
++       * 7 * 255 = 1785 seconds.
++       */
++      if (!secs)
++              secs = 1;
+-      data = FIELD_GET(SW_AGE_PERIOD_10_8_M, secs);
++      /* Return error if too large. */
++      else if (secs > 7 * MAX_TIMER_VAL)
++              return -EINVAL;
+       ret = ksz_read8(dev, REG_SW_LUE_CTRL_0, &value);
+       if (ret < 0)
+               return ret;
+-      value &= ~SW_AGE_CNT_M;
+-      value |= FIELD_PREP(SW_AGE_CNT_M, data);
++      /* Check whether there is need to update the multiplier. */
++      mult = FIELD_GET(SW_AGE_CNT_M, value);
++      max_val = MAX_TIMER_VAL;
++      if (mult > 0) {
++              /* Try to use the same multiplier already in the register as
++               * the hardware default uses multiplier 4 and 75 seconds for
++               * 300 seconds.
++               */
++              max_val = DIV_ROUND_UP(secs, mult);
++              if (max_val > MAX_TIMER_VAL || max_val * mult != secs)
++                      max_val = MAX_TIMER_VAL;
++      }
++
++      data = DIV_ROUND_UP(secs, max_val);
++      if (mult != data) {
++              value &= ~SW_AGE_CNT_M;
++              value |= FIELD_PREP(SW_AGE_CNT_M, data);
++              ret = ksz_write8(dev, REG_SW_LUE_CTRL_0, value);
++              if (ret < 0)
++                      return ret;
++      }
+-      return ksz_write8(dev, REG_SW_LUE_CTRL_0, value);
++      value = DIV_ROUND_UP(secs, data);
++      return ksz_write8(dev, REG_SW_LUE_CTRL_3, value);
+ }
+ void ksz9477_port_queue_split(struct ksz_device *dev, int port)
+diff --git a/drivers/net/dsa/microchip/ksz9477_reg.h b/drivers/net/dsa/microchip/ksz9477_reg.h
+index a2ef4b18349c..d0886ed984c5 100644
+--- a/drivers/net/dsa/microchip/ksz9477_reg.h
++++ b/drivers/net/dsa/microchip/ksz9477_reg.h
+@@ -2,7 +2,7 @@
+ /*
+  * Microchip KSZ9477 register definitions
+  *
+- * Copyright (C) 2017-2018 Microchip Technology Inc.
++ * Copyright (C) 2017-2024 Microchip Technology Inc.
+  */
+ #ifndef __KSZ9477_REGS_H
+@@ -190,8 +190,6 @@
+ #define SW_VLAN_ENABLE                        BIT(7)
+ #define SW_DROP_INVALID_VID           BIT(6)
+ #define SW_AGE_CNT_M                  GENMASK(5, 3)
+-#define SW_AGE_CNT_S                  3
+-#define SW_AGE_PERIOD_10_8_M          GENMASK(10, 8)
+ #define SW_RESV_MCAST_ENABLE          BIT(2)
+ #define SW_HASH_OPTION_M              0x03
+ #define SW_HASH_OPTION_CRC            1
+-- 
+2.39.5
+
diff --git a/queue-6.6/net-dsa-microchip-fix-lan937x-set_ageing_time-functi.patch b/queue-6.6/net-dsa-microchip-fix-lan937x-set_ageing_time-functi.patch
new file mode 100644 (file)
index 0000000..92c81e4
--- /dev/null
@@ -0,0 +1,145 @@
+From efe8b8781f0ea811cba585159f9db5c12bc64882 Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Tue, 17 Dec 2024 18:02:24 -0800
+Subject: net: dsa: microchip: Fix LAN937X set_ageing_time function
+
+From: Tristram Ha <tristram.ha@microchip.com>
+
+[ Upstream commit bb9869043438af5b94230f94fb4c39206525d758 ]
+
+The aging count is not a simple 20-bit value but comprises a 3-bit
+multiplier and a 20-bit second time.  The code tries to use the
+original multiplier which is 4 as the second count is still 300 seconds
+by default.
+
+As the 20-bit number is now too large for practical use there is an option
+to interpret it as microseconds instead of seconds.
+
+Fixes: 2c119d9982b1 ("net: dsa: microchip: add the support for set_ageing_time")
+Signed-off-by: Tristram Ha <tristram.ha@microchip.com>
+Reviewed-by: Andrew Lunn <andrew@lunn.ch>
+Link: https://patch.msgid.link/20241218020224.70590-3-Tristram.Ha@microchip.com
+Signed-off-by: Jakub Kicinski <kuba@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/net/dsa/microchip/lan937x_main.c | 62 ++++++++++++++++++++++--
+ drivers/net/dsa/microchip/lan937x_reg.h  |  9 ++--
+ 2 files changed, 65 insertions(+), 6 deletions(-)
+
+diff --git a/drivers/net/dsa/microchip/lan937x_main.c b/drivers/net/dsa/microchip/lan937x_main.c
+index b479a628b1ae..dde37e61faa3 100644
+--- a/drivers/net/dsa/microchip/lan937x_main.c
++++ b/drivers/net/dsa/microchip/lan937x_main.c
+@@ -1,6 +1,6 @@
+ // SPDX-License-Identifier: GPL-2.0
+ /* Microchip LAN937X switch driver main logic
+- * Copyright (C) 2019-2022 Microchip Technology Inc.
++ * Copyright (C) 2019-2024 Microchip Technology Inc.
+  */
+ #include <linux/kernel.h>
+ #include <linux/module.h>
+@@ -257,10 +257,66 @@ int lan937x_change_mtu(struct ksz_device *dev, int port, int new_mtu)
+ int lan937x_set_ageing_time(struct ksz_device *dev, unsigned int msecs)
+ {
+-      u32 secs = msecs / 1000;
+-      u32 value;
++      u8 data, mult, value8;
++      bool in_msec = false;
++      u32 max_val, value;
++      u32 secs = msecs;
+       int ret;
++#define MAX_TIMER_VAL ((1 << 20) - 1)
++
++      /* The aging timer comprises a 3-bit multiplier and a 20-bit second
++       * value.  Either of them cannot be zero.  The maximum timer is then
++       * 7 * 1048575 = 7340025 seconds.  As this value is too large for
++       * practical use it can be interpreted as microseconds, making the
++       * maximum timer 7340 seconds with finer control.  This allows for
++       * maximum 122 minutes compared to 29 minutes in KSZ9477 switch.
++       */
++      if (msecs % 1000)
++              in_msec = true;
++      else
++              secs /= 1000;
++      if (!secs)
++              secs = 1;
++
++      /* Return error if too large. */
++      else if (secs > 7 * MAX_TIMER_VAL)
++              return -EINVAL;
++
++      /* Configure how to interpret the number value. */
++      ret = ksz_rmw8(dev, REG_SW_LUE_CTRL_2, SW_AGE_CNT_IN_MICROSEC,
++                     in_msec ? SW_AGE_CNT_IN_MICROSEC : 0);
++      if (ret < 0)
++              return ret;
++
++      ret = ksz_read8(dev, REG_SW_LUE_CTRL_0, &value8);
++      if (ret < 0)
++              return ret;
++
++      /* Check whether there is need to update the multiplier. */
++      mult = FIELD_GET(SW_AGE_CNT_M, value8);
++      max_val = MAX_TIMER_VAL;
++      if (mult > 0) {
++              /* Try to use the same multiplier already in the register as
++               * the hardware default uses multiplier 4 and 75 seconds for
++               * 300 seconds.
++               */
++              max_val = DIV_ROUND_UP(secs, mult);
++              if (max_val > MAX_TIMER_VAL || max_val * mult != secs)
++                      max_val = MAX_TIMER_VAL;
++      }
++
++      data = DIV_ROUND_UP(secs, max_val);
++      if (mult != data) {
++              value8 &= ~SW_AGE_CNT_M;
++              value8 |= FIELD_PREP(SW_AGE_CNT_M, data);
++              ret = ksz_write8(dev, REG_SW_LUE_CTRL_0, value8);
++              if (ret < 0)
++                      return ret;
++      }
++
++      secs = DIV_ROUND_UP(secs, data);
++
+       value = FIELD_GET(SW_AGE_PERIOD_7_0_M, secs);
+       ret = ksz_write8(dev, REG_SW_AGE_PERIOD__1, value);
+diff --git a/drivers/net/dsa/microchip/lan937x_reg.h b/drivers/net/dsa/microchip/lan937x_reg.h
+index 45b606b6429f..b3e536e7c686 100644
+--- a/drivers/net/dsa/microchip/lan937x_reg.h
++++ b/drivers/net/dsa/microchip/lan937x_reg.h
+@@ -1,6 +1,6 @@
+ /* SPDX-License-Identifier: GPL-2.0 */
+ /* Microchip LAN937X switch register definitions
+- * Copyright (C) 2019-2021 Microchip Technology Inc.
++ * Copyright (C) 2019-2024 Microchip Technology Inc.
+  */
+ #ifndef __LAN937X_REG_H
+ #define __LAN937X_REG_H
+@@ -48,8 +48,7 @@
+ #define SW_VLAN_ENABLE                        BIT(7)
+ #define SW_DROP_INVALID_VID           BIT(6)
+-#define SW_AGE_CNT_M                  0x7
+-#define SW_AGE_CNT_S                  3
++#define SW_AGE_CNT_M                  GENMASK(5, 3)
+ #define SW_RESV_MCAST_ENABLE          BIT(2)
+ #define REG_SW_LUE_CTRL_1             0x0311
+@@ -62,6 +61,10 @@
+ #define SW_FAST_AGING                 BIT(1)
+ #define SW_LINK_AUTO_AGING            BIT(0)
++#define REG_SW_LUE_CTRL_2             0x0312
++
++#define SW_AGE_CNT_IN_MICROSEC                BIT(7)
++
+ #define REG_SW_AGE_PERIOD__1          0x0313
+ #define SW_AGE_PERIOD_7_0_M           GENMASK(7, 0)
+-- 
+2.39.5
+
diff --git a/queue-6.6/net-fix-memory-leak-in-tcp_conn_request.patch b/queue-6.6/net-fix-memory-leak-in-tcp_conn_request.patch
new file mode 100644 (file)
index 0000000..2ca11e5
--- /dev/null
@@ -0,0 +1,64 @@
+From 8d2ef061580515be477e7c1c1643f368ded43f61 Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Thu, 19 Dec 2024 15:28:59 +0800
+Subject: net: fix memory leak in tcp_conn_request()
+
+From: Wang Liang <wangliang74@huawei.com>
+
+[ Upstream commit 4f4aa4aa28142d53f8b06585c478476cfe325cfc ]
+
+If inet_csk_reqsk_queue_hash_add() return false, tcp_conn_request() will
+return without free the dst memory, which allocated in af_ops->route_req.
+
+Here is the kmemleak stack:
+
+unreferenced object 0xffff8881198631c0 (size 240):
+  comm "softirq", pid 0, jiffies 4299266571 (age 1802.392s)
+  hex dump (first 32 bytes):
+    00 10 9b 03 81 88 ff ff 80 98 da bc ff ff ff ff  ................
+    81 55 18 bb ff ff ff ff 00 00 00 00 00 00 00 00  .U..............
+  backtrace:
+    [<ffffffffb93e8d4c>] kmem_cache_alloc+0x60c/0xa80
+    [<ffffffffba11b4c5>] dst_alloc+0x55/0x250
+    [<ffffffffba227bf6>] rt_dst_alloc+0x46/0x1d0
+    [<ffffffffba23050a>] __mkroute_output+0x29a/0xa50
+    [<ffffffffba23456b>] ip_route_output_key_hash+0x10b/0x240
+    [<ffffffffba2346bd>] ip_route_output_flow+0x1d/0x90
+    [<ffffffffba254855>] inet_csk_route_req+0x2c5/0x500
+    [<ffffffffba26b331>] tcp_conn_request+0x691/0x12c0
+    [<ffffffffba27bd08>] tcp_rcv_state_process+0x3c8/0x11b0
+    [<ffffffffba2965c6>] tcp_v4_do_rcv+0x156/0x3b0
+    [<ffffffffba299c98>] tcp_v4_rcv+0x1cf8/0x1d80
+    [<ffffffffba239656>] ip_protocol_deliver_rcu+0xf6/0x360
+    [<ffffffffba2399a6>] ip_local_deliver_finish+0xe6/0x1e0
+    [<ffffffffba239b8e>] ip_local_deliver+0xee/0x360
+    [<ffffffffba239ead>] ip_rcv+0xad/0x2f0
+    [<ffffffffba110943>] __netif_receive_skb_one_core+0x123/0x140
+
+Call dst_release() to free the dst memory when
+inet_csk_reqsk_queue_hash_add() return false in tcp_conn_request().
+
+Fixes: ff46e3b44219 ("Fix race for duplicate reqsk on identical SYN")
+Signed-off-by: Wang Liang <wangliang74@huawei.com>
+Link: https://patch.msgid.link/20241219072859.3783576-1-wangliang74@huawei.com
+Signed-off-by: Jakub Kicinski <kuba@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ net/ipv4/tcp_input.c | 1 +
+ 1 file changed, 1 insertion(+)
+
+diff --git a/net/ipv4/tcp_input.c b/net/ipv4/tcp_input.c
+index fb053942dba2..f6a213bae5cc 100644
+--- a/net/ipv4/tcp_input.c
++++ b/net/ipv4/tcp_input.c
+@@ -7192,6 +7192,7 @@ int tcp_conn_request(struct request_sock_ops *rsk_ops,
+                       if (unlikely(!inet_csk_reqsk_queue_hash_add(sk, req,
+                                                                   req->timeout))) {
+                               reqsk_free(req);
++                              dst_release(dst);
+                               return 0;
+                       }
+-- 
+2.39.5
+
diff --git a/queue-6.6/net-fix-netns-for-ip_tunnel_init_flow.patch b/queue-6.6/net-fix-netns-for-ip_tunnel_init_flow.patch
new file mode 100644 (file)
index 0000000..40424b3
--- /dev/null
@@ -0,0 +1,71 @@
+From 1f45e51ba6f108f6b75cc657b4ab8a4de8f79ded Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Thu, 19 Dec 2024 21:03:36 +0800
+Subject: net: Fix netns for ip_tunnel_init_flow()
+
+From: Xiao Liang <shaw.leon@gmail.com>
+
+[ Upstream commit b5a7b661a073727219fedc35f5619f62418ffe72 ]
+
+The device denoted by tunnel->parms.link resides in the underlay net
+namespace. Therefore pass tunnel->net to ip_tunnel_init_flow().
+
+Fixes: db53cd3d88dc ("net: Handle l3mdev in ip_tunnel_init_flow")
+Signed-off-by: Xiao Liang <shaw.leon@gmail.com>
+Reviewed-by: Ido Schimmel <idosch@nvidia.com>
+Link: https://patch.msgid.link/20241219130336.103839-1-shaw.leon@gmail.com
+Signed-off-by: Jakub Kicinski <kuba@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c | 3 +--
+ net/ipv4/ip_tunnel.c                                | 6 +++---
+ 2 files changed, 4 insertions(+), 5 deletions(-)
+
+diff --git a/drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c b/drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
+index dcd198104141..fa3fef2b74db 100644
+--- a/drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
++++ b/drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
+@@ -423,8 +423,7 @@ mlxsw_sp_span_gretap4_route(const struct net_device *to_dev,
+       parms = mlxsw_sp_ipip_netdev_parms4(to_dev);
+       ip_tunnel_init_flow(&fl4, parms.iph.protocol, *daddrp, *saddrp,
+-                          0, 0, dev_net(to_dev), parms.link, tun->fwmark, 0,
+-                          0);
++                          0, 0, tun->net, parms.link, tun->fwmark, 0, 0);
+       rt = ip_route_output_key(tun->net, &fl4);
+       if (IS_ERR(rt))
+diff --git a/net/ipv4/ip_tunnel.c b/net/ipv4/ip_tunnel.c
+index fd8923561b18..dd1803bf9c5c 100644
+--- a/net/ipv4/ip_tunnel.c
++++ b/net/ipv4/ip_tunnel.c
+@@ -294,7 +294,7 @@ static int ip_tunnel_bind_dev(struct net_device *dev)
+               ip_tunnel_init_flow(&fl4, iph->protocol, iph->daddr,
+                                   iph->saddr, tunnel->parms.o_key,
+-                                  iph->tos & INET_DSCP_MASK, dev_net(dev),
++                                  iph->tos & INET_DSCP_MASK, tunnel->net,
+                                   tunnel->parms.link, tunnel->fwmark, 0, 0);
+               rt = ip_route_output_key(tunnel->net, &fl4);
+@@ -611,7 +611,7 @@ void ip_md_tunnel_xmit(struct sk_buff *skb, struct net_device *dev,
+       }
+       ip_tunnel_init_flow(&fl4, proto, key->u.ipv4.dst, key->u.ipv4.src,
+                           tunnel_id_to_key32(key->tun_id),
+-                          tos & INET_DSCP_MASK, dev_net(dev), 0, skb->mark,
++                          tos & INET_DSCP_MASK, tunnel->net, 0, skb->mark,
+                           skb_get_hash(skb), key->flow_flags);
+       if (!tunnel_hlen)
+@@ -774,7 +774,7 @@ void ip_tunnel_xmit(struct sk_buff *skb, struct net_device *dev,
+       ip_tunnel_init_flow(&fl4, protocol, dst, tnl_params->saddr,
+                           tunnel->parms.o_key, tos & INET_DSCP_MASK,
+-                          dev_net(dev), READ_ONCE(tunnel->parms.link),
++                          tunnel->net, READ_ONCE(tunnel->parms.link),
+                           tunnel->fwmark, skb_get_hash(skb), 0);
+       if (ip_tunnel_encap(skb, &tunnel->encap, &protocol, &fl4) < 0)
+-- 
+2.39.5
+
diff --git a/queue-6.6/net-llc-reset-skb-transport_header.patch b/queue-6.6/net-llc-reset-skb-transport_header.patch
new file mode 100644 (file)
index 0000000..5d24424
--- /dev/null
@@ -0,0 +1,62 @@
+From e61922c0628de15bf158bbab1cbf7b18a4db72ab Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Tue, 24 Dec 2024 20:07:20 -0500
+Subject: net: llc: reset skb->transport_header
+
+From: Antonio Pastor <antonio.pastor@gmail.com>
+
+[ Upstream commit a024e377efed31ecfb39210bed562932321345b3 ]
+
+802.2+LLC+SNAP frames received by napi_complete_done with GRO and DSA
+have skb->transport_header set two bytes short, or pointing 2 bytes
+before network_header & skb->data. As snap_rcv expects transport_header
+to point to SNAP header (OID:PID) after LLC processing advances offset
+over LLC header (llc_rcv & llc_fixup_skb), code doesn't find a match
+and packet is dropped.
+
+Between napi_complete_done and snap_rcv, transport_header is not used
+until __netif_receive_skb_core, where originally it was being reset.
+Commit fda55eca5a33 ("net: introduce skb_transport_header_was_set()")
+only does so if not set, on the assumption the value was set correctly
+by GRO (and also on assumption that "network stacks usually reset the
+transport header anyway"). Afterwards it is moved forward by
+llc_fixup_skb.
+
+Locally generated traffic shows up at __netif_receive_skb_core with no
+transport_header set and is processed without issue. On a setup with
+GRO but no DSA, transport_header and network_header are both set to
+point to skb->data which is also correct.
+
+As issue is LLC specific, to avoid impacting non-LLC traffic, and to
+follow up on original assumption made on previous code change,
+llc_fixup_skb to reset the offset after skb pull. llc_fixup_skb
+assumes the LLC header is at skb->data, and by definition SNAP header
+immediately follows.
+
+Fixes: fda55eca5a33 ("net: introduce skb_transport_header_was_set()")
+Signed-off-by: Antonio Pastor <antonio.pastor@gmail.com>
+Reviewed-by: Eric Dumazet <edumazet@google.com>
+Link: https://patch.msgid.link/20241225010723.2830290-1-antonio.pastor@gmail.com
+Signed-off-by: Jakub Kicinski <kuba@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ net/llc/llc_input.c | 2 +-
+ 1 file changed, 1 insertion(+), 1 deletion(-)
+
+diff --git a/net/llc/llc_input.c b/net/llc/llc_input.c
+index 51bccfb00a9c..61b0159b2fbe 100644
+--- a/net/llc/llc_input.c
++++ b/net/llc/llc_input.c
+@@ -124,8 +124,8 @@ static inline int llc_fixup_skb(struct sk_buff *skb)
+       if (unlikely(!pskb_may_pull(skb, llc_len)))
+               return 0;
+-      skb->transport_header += llc_len;
+       skb_pull(skb, llc_len);
++      skb_reset_transport_header(skb);
+       if (skb->protocol == htons(ETH_P_802_2)) {
+               __be16 pdulen;
+               s32 data_size;
+-- 
+2.39.5
+
diff --git a/queue-6.6/net-mlx5-dr-select-msix-vector-0-for-completion-queu.patch b/queue-6.6/net-mlx5-dr-select-msix-vector-0-for-completion-queu.patch
new file mode 100644 (file)
index 0000000..781e081
--- /dev/null
@@ -0,0 +1,58 @@
+From 2c765d22d234ba151e97aaa045a01980a3390b17 Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Fri, 20 Dec 2024 10:15:02 +0200
+Subject: net/mlx5: DR, select MSIX vector 0 for completion queue creation
+
+From: Shahar Shitrit <shshitrit@nvidia.com>
+
+[ Upstream commit 050a4c011b0dfeb91664a5d7bd3647ff38db08ce ]
+
+When creating a software steering completion queue (CQ), an arbitrary
+MSIX vector n is selected. This results in the CQ sharing the same
+Ethernet traffic channel n associated with the chosen vector. However,
+the value of n is often unpredictable, which can introduce complications
+for interrupt monitoring and verification tools.
+
+Moreover, SW steering uses polling rather than event-driven interrupts.
+Therefore, there is no need to select any MSIX vector other than the
+existing vector 0 for CQ creation.
+
+In light of these factors, and to enhance predictability, we modify the
+code to consistently select MSIX vector 0 for CQ creation.
+
+Fixes: 297cccebdc5a ("net/mlx5: DR, Expose an internal API to issue RDMA operations")
+Signed-off-by: Shahar Shitrit <shshitrit@nvidia.com>
+Reviewed-by: Yevgeny Kliteynik <kliteyn@nvidia.com>
+Signed-off-by: Tariq Toukan <tariqt@nvidia.com>
+Link: https://patch.msgid.link/20241220081505.1286093-2-tariqt@nvidia.com
+Signed-off-by: Jakub Kicinski <kuba@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c | 4 +---
+ 1 file changed, 1 insertion(+), 3 deletions(-)
+
+diff --git a/drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c b/drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c
+index 6fa06ba2d346..f57c84e5128b 100644
+--- a/drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c
++++ b/drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c
+@@ -1067,7 +1067,6 @@ static struct mlx5dr_cq *dr_create_cq(struct mlx5_core_dev *mdev,
+       int inlen, err, eqn;
+       void *cqc, *in;
+       __be64 *pas;
+-      int vector;
+       u32 i;
+       cq = kzalloc(sizeof(*cq), GFP_KERNEL);
+@@ -1096,8 +1095,7 @@ static struct mlx5dr_cq *dr_create_cq(struct mlx5_core_dev *mdev,
+       if (!in)
+               goto err_cqwq;
+-      vector = raw_smp_processor_id() % mlx5_comp_vectors_max(mdev);
+-      err = mlx5_comp_eqn_get(mdev, vector, &eqn);
++      err = mlx5_comp_eqn_get(mdev, 0, &eqn);
+       if (err) {
+               kvfree(in);
+               goto err_cqwq;
+-- 
+2.39.5
+
diff --git a/queue-6.6/net-mlx5e-macsec-maintain-tx-sa-from-encoding_sa.patch b/queue-6.6/net-mlx5e-macsec-maintain-tx-sa-from-encoding_sa.patch
new file mode 100644 (file)
index 0000000..bc0e0c5
--- /dev/null
@@ -0,0 +1,63 @@
+From b58a9352b6e13932a6b87ddefc45aef3b1f0000e Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Fri, 20 Dec 2024 10:15:03 +0200
+Subject: net/mlx5e: macsec: Maintain TX SA from encoding_sa
+
+From: Dragos Tatulea <dtatulea@nvidia.com>
+
+[ Upstream commit 8c6254479b3d5bd788d2b5fefaa48fb194331ed0 ]
+
+In MACsec, it is possible to create multiple active TX SAs on a SC,
+but only one such SA can be used at a time for transmission. This SA
+is selected through the encoding_sa link parameter.
+
+When there are 2 or more active TX SAs configured (encoding_sa=0):
+  ip macsec add macsec0 tx sa 0 pn 1 on key 00 <KEY1>
+  ip macsec add macsec0 tx sa 1 pn 1 on key 00 <KEY2>
+
+... the traffic should be still sent via TX SA 0 as the encoding_sa was
+not changed. However, the driver ignores the encoding_sa and overrides
+it to SA 1 by installing the flow steering id of the newly created TX SA
+into the SCI -> flow steering id hash map. The future packet tx
+descriptors will point to the incorrect flow steering rule (SA 1).
+
+This patch fixes the issue by avoiding the creation of the flow steering
+rule for an active TX SA that is not the encoding_sa. The driver side
+tx_sa object and the FW side macsec object are still created. When the
+encoding_sa link parameter is changed to another active TX SA, only the
+new flow steering rule will be created in the mlx5e_macsec_upd_txsa()
+handler.
+
+Fixes: 8ff0ac5be144 ("net/mlx5: Add MACsec offload Tx command support")
+Signed-off-by: Dragos Tatulea <dtatulea@nvidia.com>
+Reviewed-by: Cosmin Ratiu <cratiu@nvidia.com>
+Reviewed-by: Lior Nahmanson <liorna@nvidia.com>
+Signed-off-by: Tariq Toukan <tariqt@nvidia.com>
+Link: https://patch.msgid.link/20241220081505.1286093-3-tariqt@nvidia.com
+Signed-off-by: Jakub Kicinski <kuba@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/net/ethernet/mellanox/mlx5/core/en_accel/macsec.c | 4 ++++
+ 1 file changed, 4 insertions(+)
+
+diff --git a/drivers/net/ethernet/mellanox/mlx5/core/en_accel/macsec.c b/drivers/net/ethernet/mellanox/mlx5/core/en_accel/macsec.c
+index cc9bcc420032..6ab02f3fc291 100644
+--- a/drivers/net/ethernet/mellanox/mlx5/core/en_accel/macsec.c
++++ b/drivers/net/ethernet/mellanox/mlx5/core/en_accel/macsec.c
+@@ -339,9 +339,13 @@ static int mlx5e_macsec_init_sa_fs(struct macsec_context *ctx,
+ {
+       struct mlx5e_priv *priv = macsec_netdev_priv(ctx->netdev);
+       struct mlx5_macsec_fs *macsec_fs = priv->mdev->macsec_fs;
++      const struct macsec_tx_sc *tx_sc = &ctx->secy->tx_sc;
+       struct mlx5_macsec_rule_attrs rule_attrs;
+       union mlx5_macsec_rule *macsec_rule;
++      if (is_tx && tx_sc->encoding_sa != sa->assoc_num)
++              return 0;
++
+       rule_attrs.macsec_obj_id = sa->macsec_obj_id;
+       rule_attrs.sci = sa->sci;
+       rule_attrs.assoc_num = sa->assoc_num;
+-- 
+2.39.5
+
diff --git a/queue-6.6/net-mlx5e-skip-restore-tc-rules-for-vport-rep-withou.patch b/queue-6.6/net-mlx5e-skip-restore-tc-rules-for-vport-rep-withou.patch
new file mode 100644 (file)
index 0000000..b126e8c
--- /dev/null
@@ -0,0 +1,78 @@
+From 1dd0d55a6ccd1f9b01d1a731fd75a9adffdb2e81 Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Fri, 20 Dec 2024 10:15:04 +0200
+Subject: net/mlx5e: Skip restore TC rules for vport rep without loaded flag
+
+From: Jianbo Liu <jianbol@nvidia.com>
+
+[ Upstream commit 5a03b368562a7ff5f5f1f63b5adf8309cbdbd5be ]
+
+During driver unload, unregister_netdev is called after unloading
+vport rep. So, the mlx5e_rep_priv is already freed while trying to get
+rpriv->netdev, or walk rpriv->tc_ht, which results in use-after-free.
+So add the checking to make sure access the data of vport rep which is
+still loaded.
+
+Fixes: d1569537a837 ("net/mlx5e: Modify and restore TC rules for IPSec TX rules")
+Signed-off-by: Jianbo Liu <jianbol@nvidia.com>
+Reviewed-by: Saeed Mahameed <saeedm@nvidia.com>
+Signed-off-by: Tariq Toukan <tariqt@nvidia.com>
+Link: https://patch.msgid.link/20241220081505.1286093-4-tariqt@nvidia.com
+Signed-off-by: Jakub Kicinski <kuba@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/net/ethernet/mellanox/mlx5/core/esw/ipsec_fs.c     | 6 +++---
+ drivers/net/ethernet/mellanox/mlx5/core/eswitch.h          | 3 +++
+ drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c | 3 ---
+ 3 files changed, 6 insertions(+), 6 deletions(-)
+
+diff --git a/drivers/net/ethernet/mellanox/mlx5/core/esw/ipsec_fs.c b/drivers/net/ethernet/mellanox/mlx5/core/esw/ipsec_fs.c
+index 13b5916b64e2..eed8fcde2613 100644
+--- a/drivers/net/ethernet/mellanox/mlx5/core/esw/ipsec_fs.c
++++ b/drivers/net/ethernet/mellanox/mlx5/core/esw/ipsec_fs.c
+@@ -150,11 +150,11 @@ void mlx5_esw_ipsec_restore_dest_uplink(struct mlx5_core_dev *mdev)
+       unsigned long i;
+       int err;
+-      xa_for_each(&esw->offloads.vport_reps, i, rep) {
+-              rpriv = rep->rep_data[REP_ETH].priv;
+-              if (!rpriv || !rpriv->netdev)
++      mlx5_esw_for_each_rep(esw, i, rep) {
++              if (atomic_read(&rep->rep_data[REP_ETH].state) != REP_LOADED)
+                       continue;
++              rpriv = rep->rep_data[REP_ETH].priv;
+               rhashtable_walk_enter(&rpriv->tc_ht, &iter);
+               rhashtable_walk_start(&iter);
+               while ((flow = rhashtable_walk_next(&iter)) != NULL) {
+diff --git a/drivers/net/ethernet/mellanox/mlx5/core/eswitch.h b/drivers/net/ethernet/mellanox/mlx5/core/eswitch.h
+index 9b771b572593..3e58e731b569 100644
+--- a/drivers/net/ethernet/mellanox/mlx5/core/eswitch.h
++++ b/drivers/net/ethernet/mellanox/mlx5/core/eswitch.h
+@@ -713,6 +713,9 @@ void mlx5e_tc_clean_fdb_peer_flows(struct mlx5_eswitch *esw);
+                         MLX5_CAP_GEN_2((esw->dev), ec_vf_vport_base) +\
+                         (last) - 1)
++#define mlx5_esw_for_each_rep(esw, i, rep) \
++      xa_for_each(&((esw)->offloads.vport_reps), i, rep)
++
+ struct mlx5_eswitch *__must_check
+ mlx5_devlink_eswitch_get(struct devlink *devlink);
+diff --git a/drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c b/drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c
+index 58529d1a98b3..7eba3a5bb97c 100644
+--- a/drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c
++++ b/drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c
+@@ -52,9 +52,6 @@
+ #include "lag/lag.h"
+ #include "en/tc/post_meter.h"
+-#define mlx5_esw_for_each_rep(esw, i, rep) \
+-      xa_for_each(&((esw)->offloads.vport_reps), i, rep)
+-
+ /* There are two match-all miss flows, one for unicast dst mac and
+  * one for multicast.
+  */
+-- 
+2.39.5
+
diff --git a/queue-6.6/net-mv643xx_eth-fix-an-of-node-reference-leak.patch b/queue-6.6/net-mv643xx_eth-fix-an-of-node-reference-leak.patch
new file mode 100644 (file)
index 0000000..6bff15b
--- /dev/null
@@ -0,0 +1,70 @@
+From 6799fa93f32455afee1a2d1de367c45aae14670d Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Sat, 21 Dec 2024 17:14:48 +0900
+Subject: net: mv643xx_eth: fix an OF node reference leak
+
+From: Joe Hattori <joe@pf.is.s.u-tokyo.ac.jp>
+
+[ Upstream commit ad5c318086e2e23b577eca33559c5ebf89bc7eb9 ]
+
+Current implementation of mv643xx_eth_shared_of_add_port() calls
+of_parse_phandle(), but does not release the refcount on error. Call
+of_node_put() in the error path and in mv643xx_eth_shared_of_remove().
+
+This bug was found by an experimental verification tool that I am
+developing.
+
+Fixes: 76723bca2802 ("net: mv643xx_eth: add DT parsing support")
+Signed-off-by: Joe Hattori <joe@pf.is.s.u-tokyo.ac.jp>
+Link: https://patch.msgid.link/20241221081448.3313163-1-joe@pf.is.s.u-tokyo.ac.jp
+Signed-off-by: Jakub Kicinski <kuba@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/net/ethernet/marvell/mv643xx_eth.c | 14 ++++++++++++--
+ 1 file changed, 12 insertions(+), 2 deletions(-)
+
+diff --git a/drivers/net/ethernet/marvell/mv643xx_eth.c b/drivers/net/ethernet/marvell/mv643xx_eth.c
+index 3b129a1c3381..07e5051171a4 100644
+--- a/drivers/net/ethernet/marvell/mv643xx_eth.c
++++ b/drivers/net/ethernet/marvell/mv643xx_eth.c
+@@ -2708,9 +2708,15 @@ static struct platform_device *port_platdev[3];
+ static void mv643xx_eth_shared_of_remove(void)
+ {
++      struct mv643xx_eth_platform_data *pd;
+       int n;
+       for (n = 0; n < 3; n++) {
++              if (!port_platdev[n])
++                      continue;
++              pd = dev_get_platdata(&port_platdev[n]->dev);
++              if (pd)
++                      of_node_put(pd->phy_node);
+               platform_device_del(port_platdev[n]);
+               port_platdev[n] = NULL;
+       }
+@@ -2773,8 +2779,10 @@ static int mv643xx_eth_shared_of_add_port(struct platform_device *pdev,
+       }
+       ppdev = platform_device_alloc(MV643XX_ETH_NAME, dev_num);
+-      if (!ppdev)
+-              return -ENOMEM;
++      if (!ppdev) {
++              ret = -ENOMEM;
++              goto put_err;
++      }
+       ppdev->dev.coherent_dma_mask = DMA_BIT_MASK(32);
+       ppdev->dev.of_node = pnp;
+@@ -2796,6 +2804,8 @@ static int mv643xx_eth_shared_of_add_port(struct platform_device *pdev,
+ port_err:
+       platform_device_put(ppdev);
++put_err:
++      of_node_put(ppd.phy_node);
+       return ret;
+ }
+-- 
+2.39.5
+
diff --git a/queue-6.6/net-reenable-netif_f_ipv6_csum-offload-for-big-tcp-p.patch b/queue-6.6/net-reenable-netif_f_ipv6_csum-offload-for-big-tcp-p.patch
new file mode 100644 (file)
index 0000000..ee54bd6
--- /dev/null
@@ -0,0 +1,83 @@
+From b4d1abf59cbd48f8e2830c22d30c2d661d281558 Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Wed, 1 Jan 2025 11:47:40 -0500
+Subject: net: reenable NETIF_F_IPV6_CSUM offload for BIG TCP packets
+
+From: Willem de Bruijn <willemb@google.com>
+
+[ Upstream commit 68e068cabd2c6c533ef934c2e5151609cf6ecc6d ]
+
+The blamed commit disabled hardware offoad of IPv6 packets with
+extension headers on devices that advertise NETIF_F_IPV6_CSUM,
+based on the definition of that feature in skbuff.h:
+
+ *   * - %NETIF_F_IPV6_CSUM
+ *     - Driver (device) is only able to checksum plain
+ *       TCP or UDP packets over IPv6. These are specifically
+ *       unencapsulated packets of the form IPv6|TCP or
+ *       IPv6|UDP where the Next Header field in the IPv6
+ *       header is either TCP or UDP. IPv6 extension headers
+ *       are not supported with this feature. This feature
+ *       cannot be set in features for a device with
+ *       NETIF_F_HW_CSUM also set. This feature is being
+ *       DEPRECATED (see below).
+
+The change causes skb_warn_bad_offload to fire for BIG TCP
+packets.
+
+[  496.310233] WARNING: CPU: 13 PID: 23472 at net/core/dev.c:3129 skb_warn_bad_offload+0xc4/0xe0
+
+[  496.310297]  ? skb_warn_bad_offload+0xc4/0xe0
+[  496.310300]  skb_checksum_help+0x129/0x1f0
+[  496.310303]  skb_csum_hwoffload_help+0x150/0x1b0
+[  496.310306]  validate_xmit_skb+0x159/0x270
+[  496.310309]  validate_xmit_skb_list+0x41/0x70
+[  496.310312]  sch_direct_xmit+0x5c/0x250
+[  496.310317]  __qdisc_run+0x388/0x620
+
+BIG TCP introduced an IPV6_TLV_JUMBO IPv6 extension header to
+communicate packet length, as this is an IPv6 jumbogram. But, the
+feature is only enabled on devices that support BIG TCP TSO. The
+header is only present for PF_PACKET taps like tcpdump, and not
+transmitted by physical devices.
+
+For this specific case of extension headers that are not
+transmitted, return to the situation before the blamed commit
+and support hardware offload.
+
+ipv6_has_hopopt_jumbo() tests not only whether this header is present,
+but also that it is the only extension header before a terminal (L4)
+header.
+
+Fixes: 04c20a9356f2 ("net: skip offload for NETIF_F_IPV6_CSUM if ipv6 header contains extension")
+Reported-by: syzbot <syzkaller@googlegroups.com>
+Reported-by: Eric Dumazet <edumazet@google.com>
+Closes: https://lore.kernel.org/netdev/CANn89iK1hdC3Nt8KPhOtTF8vCPc1AHDCtse_BTNki1pWxAByTQ@mail.gmail.com/
+Signed-off-by: Willem de Bruijn <willemb@google.com>
+Reviewed-by: Eric Dumazet <edumazet@google.com>
+Link: https://patch.msgid.link/20250101164909.1331680-1-willemdebruijn.kernel@gmail.com
+Signed-off-by: Jakub Kicinski <kuba@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ net/core/dev.c | 4 +++-
+ 1 file changed, 3 insertions(+), 1 deletion(-)
+
+diff --git a/net/core/dev.c b/net/core/dev.c
+index 4beb9acf2c18..69da7b009f8b 100644
+--- a/net/core/dev.c
++++ b/net/core/dev.c
+@@ -3628,8 +3628,10 @@ int skb_csum_hwoffload_help(struct sk_buff *skb,
+       if (features & (NETIF_F_IP_CSUM | NETIF_F_IPV6_CSUM)) {
+               if (vlan_get_protocol(skb) == htons(ETH_P_IPV6) &&
+-                  skb_network_header_len(skb) != sizeof(struct ipv6hdr))
++                  skb_network_header_len(skb) != sizeof(struct ipv6hdr) &&
++                  !ipv6_has_hopopt_jumbo(skb))
+                       goto sw_checksum;
++
+               switch (skb->csum_offset) {
+               case offsetof(struct tcphdr, check):
+               case offsetof(struct udphdr, check):
+-- 
+2.39.5
+
diff --git a/queue-6.6/net-restrict-so_reuseport-to-inet-sockets.patch b/queue-6.6/net-restrict-so_reuseport-to-inet-sockets.patch
new file mode 100644 (file)
index 0000000..118019a
--- /dev/null
@@ -0,0 +1,87 @@
+From ae090c47bc8c84e25ec16e76e49ac5df0024f6ca Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Tue, 31 Dec 2024 16:05:27 +0000
+Subject: net: restrict SO_REUSEPORT to inet sockets
+
+From: Eric Dumazet <edumazet@google.com>
+
+[ Upstream commit 5b0af621c3f6ef9261cf6067812f2fd9943acb4b ]
+
+After blamed commit, crypto sockets could accidentally be destroyed
+from RCU call back, as spotted by zyzbot [1].
+
+Trying to acquire a mutex in RCU callback is not allowed.
+
+Restrict SO_REUSEPORT socket option to inet sockets.
+
+v1 of this patch supported TCP, UDP and SCTP sockets,
+but fcnal-test.sh test needed RAW and ICMP support.
+
+[1]
+BUG: sleeping function called from invalid context at kernel/locking/mutex.c:562
+in_atomic(): 1, irqs_disabled(): 0, non_block: 0, pid: 24, name: ksoftirqd/1
+preempt_count: 100, expected: 0
+RCU nest depth: 0, expected: 0
+1 lock held by ksoftirqd/1/24:
+  #0: ffffffff8e937ba0 (rcu_callback){....}-{0:0}, at: rcu_lock_acquire include/linux/rcupdate.h:337 [inline]
+  #0: ffffffff8e937ba0 (rcu_callback){....}-{0:0}, at: rcu_do_batch kernel/rcu/tree.c:2561 [inline]
+  #0: ffffffff8e937ba0 (rcu_callback){....}-{0:0}, at: rcu_core+0xa37/0x17a0 kernel/rcu/tree.c:2823
+Preemption disabled at:
+ [<ffffffff8161c8c8>] softirq_handle_begin kernel/softirq.c:402 [inline]
+ [<ffffffff8161c8c8>] handle_softirqs+0x128/0x9b0 kernel/softirq.c:537
+CPU: 1 UID: 0 PID: 24 Comm: ksoftirqd/1 Not tainted 6.13.0-rc3-syzkaller-00174-ga024e377efed #0
+Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 09/13/2024
+Call Trace:
+ <TASK>
+  __dump_stack lib/dump_stack.c:94 [inline]
+  dump_stack_lvl+0x241/0x360 lib/dump_stack.c:120
+  __might_resched+0x5d4/0x780 kernel/sched/core.c:8758
+  __mutex_lock_common kernel/locking/mutex.c:562 [inline]
+  __mutex_lock+0x131/0xee0 kernel/locking/mutex.c:735
+  crypto_put_default_null_skcipher+0x18/0x70 crypto/crypto_null.c:179
+  aead_release+0x3d/0x50 crypto/algif_aead.c:489
+  alg_do_release crypto/af_alg.c:118 [inline]
+  alg_sock_destruct+0x86/0xc0 crypto/af_alg.c:502
+  __sk_destruct+0x58/0x5f0 net/core/sock.c:2260
+  rcu_do_batch kernel/rcu/tree.c:2567 [inline]
+  rcu_core+0xaaa/0x17a0 kernel/rcu/tree.c:2823
+  handle_softirqs+0x2d4/0x9b0 kernel/softirq.c:561
+  run_ksoftirqd+0xca/0x130 kernel/softirq.c:950
+  smpboot_thread_fn+0x544/0xa30 kernel/smpboot.c:164
+  kthread+0x2f0/0x390 kernel/kthread.c:389
+  ret_from_fork+0x4b/0x80 arch/x86/kernel/process.c:147
+  ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244
+ </TASK>
+
+Fixes: 8c7138b33e5c ("net: Unpublish sk from sk_reuseport_cb before call_rcu")
+Reported-by: syzbot+b3e02953598f447d4d2a@syzkaller.appspotmail.com
+Closes: https://lore.kernel.org/netdev/6772f2f4.050a0220.2f3838.04cb.GAE@google.com/T/#u
+Signed-off-by: Eric Dumazet <edumazet@google.com>
+Cc: Martin KaFai Lau <kafai@fb.com>
+Reviewed-by: Kuniyuki Iwashima <kuniyu@amazon.com>
+Link: https://patch.msgid.link/20241231160527.3994168-1-edumazet@google.com
+Signed-off-by: Jakub Kicinski <kuba@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ net/core/sock.c | 5 ++++-
+ 1 file changed, 4 insertions(+), 1 deletion(-)
+
+diff --git a/net/core/sock.c b/net/core/sock.c
+index bc2a4e38dcea..84ba3f67bca9 100644
+--- a/net/core/sock.c
++++ b/net/core/sock.c
+@@ -1133,7 +1133,10 @@ int sk_setsockopt(struct sock *sk, int level, int optname,
+               sk->sk_reuse = (valbool ? SK_CAN_REUSE : SK_NO_REUSE);
+               break;
+       case SO_REUSEPORT:
+-              sk->sk_reuseport = valbool;
++              if (valbool && !sk_is_inet(sk))
++                      ret = -EOPNOTSUPP;
++              else
++                      sk->sk_reuseport = valbool;
+               break;
+       case SO_TYPE:
+       case SO_PROTOCOL:
+-- 
+2.39.5
+
diff --git a/queue-6.6/net-sfc-correct-key_len-for-efx_tc_ct_zone_ht_params.patch b/queue-6.6/net-sfc-correct-key_len-for-efx_tc_ct_zone_ht_params.patch
new file mode 100644 (file)
index 0000000..2ffcf80
--- /dev/null
@@ -0,0 +1,45 @@
+From 2caf61fa7c544aebc024832eaf567f9ac100a1e9 Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Mon, 30 Dec 2024 17:37:09 +0800
+Subject: net: sfc: Correct key_len for efx_tc_ct_zone_ht_params
+
+From: Liang Jie <liangjie@lixiang.com>
+
+[ Upstream commit a8620de72e5676993ec3a3b975f7c10908f5f60f ]
+
+In efx_tc_ct_zone_ht_params, the key_len was previously set to
+offsetof(struct efx_tc_ct_zone, linkage). This calculation is incorrect
+because it includes any padding between the zone field and the linkage
+field due to structure alignment, which can vary between systems.
+
+This patch updates key_len to use sizeof_field(struct efx_tc_ct_zone, zone)
+, ensuring that the hash table correctly uses the zone as the key. This fix
+prevents potential hash lookup errors and improves connection tracking
+reliability.
+
+Fixes: c3bb5c6acd4e ("sfc: functions to register for conntrack zone offload")
+Signed-off-by: Liang Jie <liangjie@lixiang.com>
+Acked-by: Edward Cree <ecree.xilinx@gmail.com>
+Link: https://patch.msgid.link/20241230093709.3226854-1-buaajxlj@163.com
+Signed-off-by: Jakub Kicinski <kuba@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/net/ethernet/sfc/tc_conntrack.c | 2 +-
+ 1 file changed, 1 insertion(+), 1 deletion(-)
+
+diff --git a/drivers/net/ethernet/sfc/tc_conntrack.c b/drivers/net/ethernet/sfc/tc_conntrack.c
+index 44bb57670340..109d2aa34ae3 100644
+--- a/drivers/net/ethernet/sfc/tc_conntrack.c
++++ b/drivers/net/ethernet/sfc/tc_conntrack.c
+@@ -16,7 +16,7 @@ static int efx_tc_flow_block(enum tc_setup_type type, void *type_data,
+                            void *cb_priv);
+ static const struct rhashtable_params efx_tc_ct_zone_ht_params = {
+-      .key_len        = offsetof(struct efx_tc_ct_zone, linkage),
++      .key_len        = sizeof_field(struct efx_tc_ct_zone, zone),
+       .key_offset     = 0,
+       .head_offset    = offsetof(struct efx_tc_ct_zone, linkage),
+ };
+-- 
+2.39.5
+
diff --git a/queue-6.6/net-stmmac-don-t-create-a-mdio-bus-if-unnecessary.patch b/queue-6.6/net-stmmac-don-t-create-a-mdio-bus-if-unnecessary.patch
new file mode 100644 (file)
index 0000000..402c667
--- /dev/null
@@ -0,0 +1,248 @@
+From 136c39edf8eba4f5407c484e2c74531145ede3bb Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Tue, 12 Dec 2023 16:07:36 -0600
+Subject: net: stmmac: don't create a MDIO bus if unnecessary
+
+From: Andrew Halaney <ahalaney@redhat.com>
+
+[ Upstream commit f3c2caacee824ce4a331cdafb0b8dc8e987f105e ]
+
+Currently a MDIO bus is created if the devicetree description is either:
+
+    1. Not fixed-link
+    2. fixed-link but contains a MDIO bus as well
+
+The "1" case above isn't always accurate. If there's a phy-handle,
+it could be referencing a phy on another MDIO controller's bus[1]. In
+this case, where the MDIO bus is not described at all, currently
+stmmac will make a MDIO bus and scan its address space to discover
+phys (of which there are none). This process takes time scanning a bus
+that is known to be empty, delaying time to complete probe.
+
+There are also a lot of upstream devicetrees[2] that expect a MDIO bus
+to be created, scanned for phys, and the first one found connected
+to the MAC. This case can be inferred from the platform description by
+not having a phy-handle && not being fixed-link. This hits case "1" in
+the current driver's logic, and must be handled in any logic change here
+since it is a valid legacy dt-binding.
+
+Let's improve the logic to create a MDIO bus if either:
+
+    - Devicetree contains a MDIO bus
+    - !fixed-link && !phy-handle (legacy handling)
+
+This way the case where no MDIO bus should be made is handled, as well
+as retaining backwards compatibility with the valid cases.
+
+Below devicetree snippets can be found that explain some of
+the cases above more concretely.
+
+Here's[0] a devicetree example where the MAC is both fixed-link and
+driving a switch on MDIO (case "2" above). This needs a MDIO bus to
+be created:
+
+    &fec1 {
+            phy-mode = "rmii";
+
+            fixed-link {
+                    speed = <100>;
+                    full-duplex;
+            };
+
+            mdio1: mdio {
+                    switch0: switch0@0 {
+                            compatible = "marvell,mv88e6190";
+                            pinctrl-0 = <&pinctrl_gpio_switch0>;
+                    };
+            };
+    };
+
+Here's[1] an example where there is no MDIO bus or fixed-link for
+the ethernet1 MAC, so no MDIO bus should be created since ethernet0
+is the MDIO master for ethernet1's phy:
+
+    &ethernet0 {
+            phy-mode = "sgmii";
+            phy-handle = <&sgmii_phy0>;
+
+            mdio {
+                    compatible = "snps,dwmac-mdio";
+                    sgmii_phy0: phy@8 {
+                            compatible = "ethernet-phy-id0141.0dd4";
+                            reg = <0x8>;
+                            device_type = "ethernet-phy";
+                    };
+
+                    sgmii_phy1: phy@a {
+                            compatible = "ethernet-phy-id0141.0dd4";
+                            reg = <0xa>;
+                            device_type = "ethernet-phy";
+                    };
+            };
+    };
+
+    &ethernet1 {
+            phy-mode = "sgmii";
+            phy-handle = <&sgmii_phy1>;
+    };
+
+Finally there's descriptions like this[2] which don't describe the
+MDIO bus but expect it to be created and the whole address space
+scanned for a phy since there's no phy-handle or fixed-link described:
+
+    &gmac {
+            phy-supply = <&vcc_lan>;
+            phy-mode = "rmii";
+            snps,reset-gpio = <&gpio3 RK_PB4 GPIO_ACTIVE_HIGH>;
+            snps,reset-active-low;
+            snps,reset-delays-us = <0 10000 1000000>;
+    };
+
+[0] https://elixir.bootlin.com/linux/v6.5-rc5/source/arch/arm/boot/dts/nxp/vf/vf610-zii-ssmb-dtu.dts
+[1] https://elixir.bootlin.com/linux/v6.6-rc5/source/arch/arm64/boot/dts/qcom/sa8775p-ride.dts
+[2] https://elixir.bootlin.com/linux/v6.6-rc5/source/arch/arm64/boot/dts/rockchip/rk3368-r88.dts#L164
+
+Reviewed-by: Serge Semin <fancer.lancer@gmail.com>
+Co-developed-by: Bartosz Golaszewski <bartosz.golaszewski@linaro.org>
+Signed-off-by: Bartosz Golaszewski <bartosz.golaszewski@linaro.org>
+Signed-off-by: Andrew Halaney <ahalaney@redhat.com>
+Signed-off-by: David S. Miller <davem@davemloft.net>
+Stable-dep-of: 2b6ffcd7873b ("net: stmmac: restructure the error path of stmmac_probe_config_dt()")
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ .../ethernet/stmicro/stmmac/stmmac_platform.c | 91 +++++++++++--------
+ 1 file changed, 54 insertions(+), 37 deletions(-)
+
+diff --git a/drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c b/drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
+index b4fdd40be63c..d73b2c17cc6c 100644
+--- a/drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
++++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
+@@ -296,62 +296,80 @@ static int stmmac_mtl_setup(struct platform_device *pdev,
+ }
+ /**
+- * stmmac_dt_phy - parse device-tree driver parameters to allocate PHY resources
+- * @plat: driver data platform structure
+- * @np: device tree node
+- * @dev: device pointer
+- * Description:
+- * The mdio bus will be allocated in case of a phy transceiver is on board;
+- * it will be NULL if the fixed-link is configured.
+- * If there is the "snps,dwmac-mdio" sub-node the mdio will be allocated
+- * in any case (for DSA, mdio must be registered even if fixed-link).
+- * The table below sums the supported configurations:
+- *    -------------------------------
+- *    snps,phy-addr   |     Y
+- *    -------------------------------
+- *    phy-handle      |     Y
+- *    -------------------------------
+- *    fixed-link      |     N
+- *    -------------------------------
+- *    snps,dwmac-mdio |
+- *      even if       |     Y
+- *    fixed-link      |
+- *    -------------------------------
++ * stmmac_of_get_mdio() - Gets the MDIO bus from the devicetree.
++ * @np: devicetree node
+  *
+- * It returns 0 in case of success otherwise -ENODEV.
++ * The MDIO bus will be searched for in the following ways:
++ * 1. The compatible is "snps,dwc-qos-ethernet-4.10" && a "mdio" named
++ *    child node exists
++ * 2. A child node with the "snps,dwmac-mdio" compatible is present
++ *
++ * Return: The MDIO node if present otherwise NULL
+  */
+-static int stmmac_dt_phy(struct plat_stmmacenet_data *plat,
+-                       struct device_node *np, struct device *dev)
++static struct device_node *stmmac_of_get_mdio(struct device_node *np)
+ {
+-      bool mdio = !of_phy_is_fixed_link(np);
+       static const struct of_device_id need_mdio_ids[] = {
+               { .compatible = "snps,dwc-qos-ethernet-4.10" },
+               {},
+       };
++      struct device_node *mdio_node = NULL;
+       if (of_match_node(need_mdio_ids, np)) {
+-              plat->mdio_node = of_get_child_by_name(np, "mdio");
++              mdio_node = of_get_child_by_name(np, "mdio");
+       } else {
+               /**
+                * If snps,dwmac-mdio is passed from DT, always register
+                * the MDIO
+                */
+-              for_each_child_of_node(np, plat->mdio_node) {
+-                      if (of_device_is_compatible(plat->mdio_node,
++              for_each_child_of_node(np, mdio_node) {
++                      if (of_device_is_compatible(mdio_node,
+                                                   "snps,dwmac-mdio"))
+                               break;
+               }
+       }
+-      if (plat->mdio_node) {
++      return mdio_node;
++}
++
++/**
++ * stmmac_mdio_setup() - Populate platform related MDIO structures.
++ * @plat: driver data platform structure
++ * @np: devicetree node
++ * @dev: device pointer
++ *
++ * This searches for MDIO information from the devicetree.
++ * If an MDIO node is found, it's assigned to plat->mdio_node and
++ * plat->mdio_bus_data is allocated.
++ * If no connection can be determined, just plat->mdio_bus_data is allocated
++ * to indicate a bus should be created and scanned for a phy.
++ * If it's determined there's no MDIO bus needed, both are left NULL.
++ *
++ * This expects that plat->phy_node has already been searched for.
++ *
++ * Return: 0 on success, errno otherwise.
++ */
++static int stmmac_mdio_setup(struct plat_stmmacenet_data *plat,
++                           struct device_node *np, struct device *dev)
++{
++      bool legacy_mdio;
++
++      plat->mdio_node = stmmac_of_get_mdio(np);
++      if (plat->mdio_node)
+               dev_dbg(dev, "Found MDIO subnode\n");
+-              mdio = true;
+-      }
+-      if (mdio) {
+-              plat->mdio_bus_data =
+-                      devm_kzalloc(dev, sizeof(struct stmmac_mdio_bus_data),
+-                                   GFP_KERNEL);
++      /* Legacy devicetrees allowed for no MDIO bus description and expect
++       * the bus to be scanned for devices. If there's no phy or fixed-link
++       * described assume this is the case since there must be something
++       * connected to the MAC.
++       */
++      legacy_mdio = !of_phy_is_fixed_link(np) && !plat->phy_node;
++      if (legacy_mdio)
++              dev_info(dev, "Deprecated MDIO bus assumption used\n");
++
++      if (plat->mdio_node || legacy_mdio) {
++              plat->mdio_bus_data = devm_kzalloc(dev,
++                                                 sizeof(*plat->mdio_bus_data),
++                                                 GFP_KERNEL);
+               if (!plat->mdio_bus_data)
+                       return -ENOMEM;
+@@ -455,8 +473,7 @@ stmmac_probe_config_dt(struct platform_device *pdev, u8 *mac)
+       if (of_property_read_u32(np, "snps,phy-addr", &plat->phy_addr) == 0)
+               dev_warn(&pdev->dev, "snps,phy-addr property is deprecated\n");
+-      /* To Configure PHY by using all device-tree supported properties */
+-      rc = stmmac_dt_phy(plat, np, &pdev->dev);
++      rc = stmmac_mdio_setup(plat, np, &pdev->dev);
+       if (rc)
+               return ERR_PTR(rc);
+-- 
+2.39.5
+
diff --git a/queue-6.6/net-stmmac-restructure-the-error-path-of-stmmac_prob.patch b/queue-6.6/net-stmmac-restructure-the-error-path-of-stmmac_prob.patch
new file mode 100644 (file)
index 0000000..bc0aa32
--- /dev/null
@@ -0,0 +1,105 @@
+From 713950f9ee6210688a65af1740ed993aa853f23b Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Thu, 19 Dec 2024 11:41:19 +0900
+Subject: net: stmmac: restructure the error path of stmmac_probe_config_dt()
+
+From: Joe Hattori <joe@pf.is.s.u-tokyo.ac.jp>
+
+[ Upstream commit 2b6ffcd7873b7e8a62c3e15a6f305bfc747c466b ]
+
+Current implementation of stmmac_probe_config_dt() does not release the
+OF node reference obtained by of_parse_phandle() in some error paths.
+The problem is that some error paths call stmmac_remove_config_dt() to
+clean up but others use and unwind ladder.  These two types of error
+handling have not kept in sync and have been a recurring source of bugs.
+Re-write the error handling in stmmac_probe_config_dt() to use an unwind
+ladder. Consequently, stmmac_remove_config_dt() is not needed anymore,
+thus remove it.
+
+This bug was found by an experimental verification tool that I am
+developing.
+
+Fixes: 4838a5405028 ("net: stmmac: Fix wrapper drivers not detecting PHY")
+Signed-off-by: Joe Hattori <joe@pf.is.s.u-tokyo.ac.jp>
+Link: https://patch.msgid.link/20241219024119.2017012-1-joe@pf.is.s.u-tokyo.ac.jp
+Signed-off-by: Jakub Kicinski <kuba@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ .../ethernet/stmicro/stmmac/stmmac_platform.c | 27 ++++++++++++-------
+ 1 file changed, 17 insertions(+), 10 deletions(-)
+
+diff --git a/drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c b/drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
+index d73b2c17cc6c..4d570efd9d4b 100644
+--- a/drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
++++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
+@@ -474,8 +474,10 @@ stmmac_probe_config_dt(struct platform_device *pdev, u8 *mac)
+               dev_warn(&pdev->dev, "snps,phy-addr property is deprecated\n");
+       rc = stmmac_mdio_setup(plat, np, &pdev->dev);
+-      if (rc)
+-              return ERR_PTR(rc);
++      if (rc) {
++              ret = ERR_PTR(rc);
++              goto error_put_phy;
++      }
+       of_property_read_u32(np, "tx-fifo-depth", &plat->tx_fifo_size);
+@@ -564,8 +566,8 @@ stmmac_probe_config_dt(struct platform_device *pdev, u8 *mac)
+       dma_cfg = devm_kzalloc(&pdev->dev, sizeof(*dma_cfg),
+                              GFP_KERNEL);
+       if (!dma_cfg) {
+-              stmmac_remove_config_dt(pdev, plat);
+-              return ERR_PTR(-ENOMEM);
++              ret = ERR_PTR(-ENOMEM);
++              goto error_put_mdio;
+       }
+       plat->dma_cfg = dma_cfg;
+@@ -593,8 +595,8 @@ stmmac_probe_config_dt(struct platform_device *pdev, u8 *mac)
+       rc = stmmac_mtl_setup(pdev, plat);
+       if (rc) {
+-              stmmac_remove_config_dt(pdev, plat);
+-              return ERR_PTR(rc);
++              ret = ERR_PTR(rc);
++              goto error_put_mdio;
+       }
+       /* clock setup */
+@@ -646,6 +648,10 @@ stmmac_probe_config_dt(struct platform_device *pdev, u8 *mac)
+       clk_disable_unprepare(plat->pclk);
+ error_pclk_get:
+       clk_disable_unprepare(plat->stmmac_clk);
++error_put_mdio:
++      of_node_put(plat->mdio_node);
++error_put_phy:
++      of_node_put(plat->phy_node);
+       return ret;
+ }
+@@ -654,16 +660,17 @@ static void devm_stmmac_remove_config_dt(void *data)
+ {
+       struct plat_stmmacenet_data *plat = data;
+-      /* Platform data argument is unused */
+-      stmmac_remove_config_dt(NULL, plat);
++      clk_disable_unprepare(plat->stmmac_clk);
++      clk_disable_unprepare(plat->pclk);
++      of_node_put(plat->mdio_node);
++      of_node_put(plat->phy_node);
+ }
+ /**
+  * devm_stmmac_probe_config_dt
+  * @pdev: platform_device structure
+  * @mac: MAC address to use
+- * Description: Devres variant of stmmac_probe_config_dt(). Does not require
+- * the user to call stmmac_remove_config_dt() at driver detach.
++ * Description: Devres variant of stmmac_probe_config_dt().
+  */
+ struct plat_stmmacenet_data *
+ devm_stmmac_probe_config_dt(struct platform_device *pdev, u8 *mac)
+-- 
+2.39.5
+
diff --git a/queue-6.6/net-ti-icssg-prueth-fix-clearing-of-iep_cmp_cfg-regi.patch b/queue-6.6/net-ti-icssg-prueth-fix-clearing-of-iep_cmp_cfg-regi.patch
new file mode 100644 (file)
index 0000000..ada66e2
--- /dev/null
@@ -0,0 +1,61 @@
+From b503e64b9123d01b0dc559b8594c675833dd12de Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Mon, 23 Dec 2024 20:45:50 +0530
+Subject: net: ti: icssg-prueth: Fix clearing of IEP_CMP_CFG registers during
+ iep_init
+
+From: Meghana Malladi <m-malladi@ti.com>
+
+[ Upstream commit 9b115361248dc6cce182a2dc030c1c70b0a9639e ]
+
+When ICSSG interfaces are brought down and brought up again, the
+pru cores are shut down and booted again, flushing out all the memories
+and start again in a clean state. Hence it is expected that the
+IEP_CMP_CFG register needs to be flushed during iep_init() to ensure
+that the existing residual configuration doesn't cause any unusual
+behavior. If the register is not cleared, existing IEP_CMP_CFG set for
+CMP1 will result in SYNC0_OUT signal based on the SYNC_OUT register values.
+
+After bringing the interface up, calling PPS enable doesn't work as
+the driver believes PPS is already enabled, (iep->pps_enabled is not
+cleared during interface bring down) and driver will just return true
+even though there is no signal. Fix this by disabling pps and perout.
+
+Fixes: c1e0230eeaab ("net: ti: icss-iep: Add IEP driver")
+Signed-off-by: Meghana Malladi <m-malladi@ti.com>
+Reviewed-by: Roger Quadros <rogerq@kernel.org>
+Signed-off-by: David S. Miller <davem@davemloft.net>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/net/ethernet/ti/icssg/icss_iep.c | 8 ++++++++
+ 1 file changed, 8 insertions(+)
+
+diff --git a/drivers/net/ethernet/ti/icssg/icss_iep.c b/drivers/net/ethernet/ti/icssg/icss_iep.c
+index 3025e9c18970..f06cdec14ed7 100644
+--- a/drivers/net/ethernet/ti/icssg/icss_iep.c
++++ b/drivers/net/ethernet/ti/icssg/icss_iep.c
+@@ -290,6 +290,9 @@ static void icss_iep_enable_shadow_mode(struct icss_iep *iep)
+       for (cmp = IEP_MIN_CMP; cmp < IEP_MAX_CMP; cmp++) {
+               regmap_update_bits(iep->map, ICSS_IEP_CMP_STAT_REG,
+                                  IEP_CMP_STATUS(cmp), IEP_CMP_STATUS(cmp));
++
++              regmap_update_bits(iep->map, ICSS_IEP_CMP_CFG_REG,
++                                 IEP_CMP_CFG_CMP_EN(cmp), 0);
+       }
+       /* enable reset counter on CMP0 event */
+@@ -808,6 +811,11 @@ int icss_iep_exit(struct icss_iep *iep)
+       }
+       icss_iep_disable(iep);
++      if (iep->pps_enabled)
++              icss_iep_pps_enable(iep, false);
++      else if (iep->perout_enabled)
++              icss_iep_perout_enable(iep, NULL, false);
++
+       return 0;
+ }
+ EXPORT_SYMBOL_GPL(icss_iep_exit);
+-- 
+2.39.5
+
diff --git a/queue-6.6/net-wwan-iosm-properly-check-for-valid-exec-stage-in.patch b/queue-6.6/net-wwan-iosm-properly-check-for-valid-exec-stage-in.patch
new file mode 100644 (file)
index 0000000..1adf269
--- /dev/null
@@ -0,0 +1,47 @@
+From ef659dc2a5ccad66cf1260bcbe78757bf42a13ec Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Sun, 29 Dec 2024 17:46:58 +0100
+Subject: net: wwan: iosm: Properly check for valid exec stage in
+ ipc_mmio_init()
+
+From: Maciej S. Szmigiero <mail@maciej.szmigiero.name>
+
+[ Upstream commit a7af435df0e04cfb4a4004136d597c42639a2ae7 ]
+
+ipc_mmio_init() used the post-decrement operator in its loop continuing
+condition of "retries" counter being "> 0", which meant that when this
+condition caused loop exit "retries" counter reached -1.
+
+But the later valid exec stage failure check only tests for "retries"
+counter being exactly zero, so it didn't trigger in this case (but
+would wrongly trigger if the code reaches a valid exec stage in the
+very last loop iteration).
+
+Fix this by using the pre-decrement operator instead, so the loop counter
+is exactly zero on valid exec stage failure.
+
+Fixes: dc0514f5d828 ("net: iosm: mmio scratchpad")
+Signed-off-by: Maciej S. Szmigiero <mail@maciej.szmigiero.name>
+Link: https://patch.msgid.link/8b19125a825f9dcdd81c667c1e5c48ba28d505a6.1735490770.git.mail@maciej.szmigiero.name
+Signed-off-by: Jakub Kicinski <kuba@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/net/wwan/iosm/iosm_ipc_mmio.c | 2 +-
+ 1 file changed, 1 insertion(+), 1 deletion(-)
+
+diff --git a/drivers/net/wwan/iosm/iosm_ipc_mmio.c b/drivers/net/wwan/iosm/iosm_ipc_mmio.c
+index 63eb08c43c05..6764c13530b9 100644
+--- a/drivers/net/wwan/iosm/iosm_ipc_mmio.c
++++ b/drivers/net/wwan/iosm/iosm_ipc_mmio.c
+@@ -104,7 +104,7 @@ struct iosm_mmio *ipc_mmio_init(void __iomem *mmio, struct device *dev)
+                       break;
+               msleep(20);
+-      } while (retries-- > 0);
++      } while (--retries > 0);
+       if (!retries) {
+               dev_err(ipc_mmio->dev, "invalid exec stage %X", stage);
+-- 
+2.39.5
+
diff --git a/queue-6.6/net-wwan-t7xx-fix-fsm-command-timeout-issue.patch b/queue-6.6/net-wwan-t7xx-fix-fsm-command-timeout-issue.patch
new file mode 100644 (file)
index 0000000..6d1a031
--- /dev/null
@@ -0,0 +1,141 @@
+From 77b34c996589c3b54a61712abee1921731e531fa Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Tue, 24 Dec 2024 12:15:52 +0800
+Subject: net: wwan: t7xx: Fix FSM command timeout issue
+
+From: Jinjian Song <jinjian.song@fibocom.com>
+
+[ Upstream commit 4f619d518db9cd1a933c3a095a5f95d0c1584ae8 ]
+
+When driver processes the internal state change command, it use an
+asynchronous thread to process the command operation. If the main
+thread detects that the task has timed out, the asynchronous thread
+will panic when executing the completion notification because the
+main thread completion object has been released.
+
+BUG: unable to handle page fault for address: fffffffffffffff8
+PGD 1f283a067 P4D 1f283a067 PUD 1f283c067 PMD 0
+Oops: 0000 [#1] PREEMPT SMP NOPTI
+RIP: 0010:complete_all+0x3e/0xa0
+[...]
+Call Trace:
+ <TASK>
+ ? __die_body+0x68/0xb0
+ ? page_fault_oops+0x379/0x3e0
+ ? exc_page_fault+0x69/0xa0
+ ? asm_exc_page_fault+0x22/0x30
+ ? complete_all+0x3e/0xa0
+ fsm_main_thread+0xa3/0x9c0 [mtk_t7xx (HASH:1400 5)]
+ ? __pfx_autoremove_wake_function+0x10/0x10
+ kthread+0xd8/0x110
+ ? __pfx_fsm_main_thread+0x10/0x10 [mtk_t7xx (HASH:1400 5)]
+ ? __pfx_kthread+0x10/0x10
+ ret_from_fork+0x38/0x50
+ ? __pfx_kthread+0x10/0x10
+ ret_from_fork_asm+0x1b/0x30
+ </TASK>
+[...]
+CR2: fffffffffffffff8
+---[ end trace 0000000000000000 ]---
+
+Use the reference counter to ensure safe release as Sergey suggests:
+https://lore.kernel.org/all/da90f64c-260a-4329-87bf-1f9ff20a5951@gmail.com/
+
+Fixes: 13e920d93e37 ("net: wwan: t7xx: Add core components")
+Signed-off-by: Jinjian Song <jinjian.song@fibocom.com>
+Acked-by: Sergey Ryazanov <ryazanov.s.a@gmail.com>
+Link: https://patch.msgid.link/20241224041552.8711-1-jinjian.song@fibocom.com
+Signed-off-by: Jakub Kicinski <kuba@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/net/wwan/t7xx/t7xx_state_monitor.c | 26 ++++++++++++++--------
+ drivers/net/wwan/t7xx/t7xx_state_monitor.h |  5 +++--
+ 2 files changed, 20 insertions(+), 11 deletions(-)
+
+diff --git a/drivers/net/wwan/t7xx/t7xx_state_monitor.c b/drivers/net/wwan/t7xx/t7xx_state_monitor.c
+index 80edb8e75a6a..64868df3640d 100644
+--- a/drivers/net/wwan/t7xx/t7xx_state_monitor.c
++++ b/drivers/net/wwan/t7xx/t7xx_state_monitor.c
+@@ -97,14 +97,21 @@ void t7xx_fsm_broadcast_state(struct t7xx_fsm_ctl *ctl, enum md_state state)
+       fsm_state_notify(ctl->md, state);
+ }
++static void fsm_release_command(struct kref *ref)
++{
++      struct t7xx_fsm_command *cmd = container_of(ref, typeof(*cmd), refcnt);
++
++      kfree(cmd);
++}
++
+ static void fsm_finish_command(struct t7xx_fsm_ctl *ctl, struct t7xx_fsm_command *cmd, int result)
+ {
+       if (cmd->flag & FSM_CMD_FLAG_WAIT_FOR_COMPLETION) {
+-              *cmd->ret = result;
+-              complete_all(cmd->done);
++              cmd->result = result;
++              complete_all(&cmd->done);
+       }
+-      kfree(cmd);
++      kref_put(&cmd->refcnt, fsm_release_command);
+ }
+ static void fsm_del_kf_event(struct t7xx_fsm_event *event)
+@@ -396,7 +403,6 @@ static int fsm_main_thread(void *data)
+ int t7xx_fsm_append_cmd(struct t7xx_fsm_ctl *ctl, enum t7xx_fsm_cmd_state cmd_id, unsigned int flag)
+ {
+-      DECLARE_COMPLETION_ONSTACK(done);
+       struct t7xx_fsm_command *cmd;
+       unsigned long flags;
+       int ret;
+@@ -408,11 +414,13 @@ int t7xx_fsm_append_cmd(struct t7xx_fsm_ctl *ctl, enum t7xx_fsm_cmd_state cmd_id
+       INIT_LIST_HEAD(&cmd->entry);
+       cmd->cmd_id = cmd_id;
+       cmd->flag = flag;
++      kref_init(&cmd->refcnt);
+       if (flag & FSM_CMD_FLAG_WAIT_FOR_COMPLETION) {
+-              cmd->done = &done;
+-              cmd->ret = &ret;
++              init_completion(&cmd->done);
++              kref_get(&cmd->refcnt);
+       }
++      kref_get(&cmd->refcnt);
+       spin_lock_irqsave(&ctl->command_lock, flags);
+       list_add_tail(&cmd->entry, &ctl->command_queue);
+       spin_unlock_irqrestore(&ctl->command_lock, flags);
+@@ -422,11 +430,11 @@ int t7xx_fsm_append_cmd(struct t7xx_fsm_ctl *ctl, enum t7xx_fsm_cmd_state cmd_id
+       if (flag & FSM_CMD_FLAG_WAIT_FOR_COMPLETION) {
+               unsigned long wait_ret;
+-              wait_ret = wait_for_completion_timeout(&done,
++              wait_ret = wait_for_completion_timeout(&cmd->done,
+                                                      msecs_to_jiffies(FSM_CMD_TIMEOUT_MS));
+-              if (!wait_ret)
+-                      return -ETIMEDOUT;
++              ret = wait_ret ? cmd->result : -ETIMEDOUT;
++              kref_put(&cmd->refcnt, fsm_release_command);
+               return ret;
+       }
+diff --git a/drivers/net/wwan/t7xx/t7xx_state_monitor.h b/drivers/net/wwan/t7xx/t7xx_state_monitor.h
+index b6e76f3903c8..74f96fd2605e 100644
+--- a/drivers/net/wwan/t7xx/t7xx_state_monitor.h
++++ b/drivers/net/wwan/t7xx/t7xx_state_monitor.h
+@@ -109,8 +109,9 @@ struct t7xx_fsm_command {
+       struct list_head        entry;
+       enum t7xx_fsm_cmd_state cmd_id;
+       unsigned int            flag;
+-      struct completion       *done;
+-      int                     *ret;
++      struct completion       done;
++      int                     result;
++      struct kref             refcnt;
+ };
+ struct t7xx_fsm_notifier {
+-- 
+2.39.5
+
diff --git a/queue-6.6/netfilter-nft_set_hash-unaligned-atomic-read-on-stru.patch b/queue-6.6/netfilter-nft_set_hash-unaligned-atomic-read-on-stru.patch
new file mode 100644 (file)
index 0000000..f5432f4
--- /dev/null
@@ -0,0 +1,96 @@
+From 36d0396eac33ccf4873004acc1e34e7a214f6b7f Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Sat, 21 Dec 2024 00:29:20 +0100
+Subject: netfilter: nft_set_hash: unaligned atomic read on struct nft_set_ext
+
+From: Pablo Neira Ayuso <pablo@netfilter.org>
+
+[ Upstream commit 542ed8145e6f9392e3d0a86a0e9027d2ffd183e4 ]
+
+Access to genmask field in struct nft_set_ext results in unaligned
+atomic read:
+
+[   72.130109] Unable to handle kernel paging request at virtual address ffff0000c2bb708c
+[   72.131036] Mem abort info:
+[   72.131213]   ESR = 0x0000000096000021
+[   72.131446]   EC = 0x25: DABT (current EL), IL = 32 bits
+[   72.132209]   SET = 0, FnV = 0
+[   72.133216]   EA = 0, S1PTW = 0
+[   72.134080]   FSC = 0x21: alignment fault
+[   72.135593] Data abort info:
+[   72.137194]   ISV = 0, ISS = 0x00000021, ISS2 = 0x00000000
+[   72.142351]   CM = 0, WnR = 0, TnD = 0, TagAccess = 0
+[   72.145989]   GCS = 0, Overlay = 0, DirtyBit = 0, Xs = 0
+[   72.150115] swapper pgtable: 4k pages, 48-bit VAs, pgdp=0000000237d27000
+[   72.154893] [ffff0000c2bb708c] pgd=0000000000000000, p4d=180000023ffff403, pud=180000023f84b403, pmd=180000023f835403,
++pte=0068000102bb7707
+[   72.163021] Internal error: Oops: 0000000096000021 [#1] SMP
+[...]
+[   72.170041] CPU: 7 UID: 0 PID: 54 Comm: kworker/7:0 Tainted: G            E      6.13.0-rc3+ #2
+[   72.170509] Tainted: [E]=UNSIGNED_MODULE
+[   72.170720] Hardware name: QEMU QEMU Virtual Machine, BIOS edk2-stable202302-for-qemu 03/01/2023
+[   72.171192] Workqueue: events_power_efficient nft_rhash_gc [nf_tables]
+[   72.171552] pstate: 21400005 (nzCv daif +PAN -UAO -TCO +DIT -SSBS BTYPE=--)
+[   72.171915] pc : nft_rhash_gc+0x200/0x2d8 [nf_tables]
+[   72.172166] lr : nft_rhash_gc+0x128/0x2d8 [nf_tables]
+[   72.172546] sp : ffff800081f2bce0
+[   72.172724] x29: ffff800081f2bd40 x28: ffff0000c2bb708c x27: 0000000000000038
+[   72.173078] x26: ffff0000c6780ef0 x25: ffff0000c643df00 x24: ffff0000c6778f78
+[   72.173431] x23: 000000000000001a x22: ffff0000c4b1f000 x21: ffff0000c6780f78
+[   72.173782] x20: ffff0000c2bb70dc x19: ffff0000c2bb7080 x18: 0000000000000000
+[   72.174135] x17: ffff0000c0a4e1c0 x16: 0000000000003000 x15: 0000ac26d173b978
+[   72.174485] x14: ffffffffffffffff x13: 0000000000000030 x12: ffff0000c6780ef0
+[   72.174841] x11: 0000000000000000 x10: ffff800081f2bcf8 x9 : ffff0000c3000000
+[   72.175193] x8 : 00000000000004be x7 : 0000000000000000 x6 : 0000000000000000
+[   72.175544] x5 : 0000000000000040 x4 : ffff0000c3000010 x3 : 0000000000000000
+[   72.175871] x2 : 0000000000003a98 x1 : ffff0000c2bb708c x0 : 0000000000000004
+[   72.176207] Call trace:
+[   72.176316]  nft_rhash_gc+0x200/0x2d8 [nf_tables] (P)
+[   72.176653]  process_one_work+0x178/0x3d0
+[   72.176831]  worker_thread+0x200/0x3f0
+[   72.176995]  kthread+0xe8/0xf8
+[   72.177130]  ret_from_fork+0x10/0x20
+[   72.177289] Code: 54fff984 d503201f d2800080 91003261 (f820303f)
+[   72.177557] ---[ end trace 0000000000000000 ]---
+
+Align struct nft_set_ext to word size to address this and
+documentation it.
+
+pahole reports that this increases the size of elements for rhash and
+pipapo in 8 bytes on x86_64.
+
+Fixes: 7ffc7481153b ("netfilter: nft_set_hash: skip duplicated elements pending gc run")
+Signed-off-by: Pablo Neira Ayuso <pablo@netfilter.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ include/net/netfilter/nf_tables.h | 7 +++++--
+ 1 file changed, 5 insertions(+), 2 deletions(-)
+
+diff --git a/include/net/netfilter/nf_tables.h b/include/net/netfilter/nf_tables.h
+index b5f9ee5810a3..8321915dddb2 100644
+--- a/include/net/netfilter/nf_tables.h
++++ b/include/net/netfilter/nf_tables.h
+@@ -721,15 +721,18 @@ struct nft_set_ext_tmpl {
+ /**
+  *    struct nft_set_ext - set extensions
+  *
+- *    @genmask: generation mask
++ *    @genmask: generation mask, but also flags (see NFT_SET_ELEM_DEAD_BIT)
+  *    @offset: offsets of individual extension types
+  *    @data: beginning of extension data
++ *
++ *    This structure must be aligned to word size, otherwise atomic bitops
++ *    on genmask field can cause alignment failure on some archs.
+  */
+ struct nft_set_ext {
+       u8      genmask;
+       u8      offset[NFT_SET_EXT_NUM];
+       char    data[];
+-};
++} __aligned(BITS_PER_LONG / 8);
+ static inline void nft_set_ext_prepare(struct nft_set_ext_tmpl *tmpl)
+ {
+-- 
+2.39.5
+
diff --git a/queue-6.6/netrom-check-buffer-length-before-accessing-it.patch b/queue-6.6/netrom-check-buffer-length-before-accessing-it.patch
new file mode 100644 (file)
index 0000000..12b33c8
--- /dev/null
@@ -0,0 +1,105 @@
+From d58c985747fef13aceb57e0f1fa3d14f04b73609 Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Thu, 19 Dec 2024 08:23:07 +0000
+Subject: netrom: check buffer length before accessing it
+
+From: Ilya Shchipletsov <rabbelkin@mail.ru>
+
+[ Upstream commit a4fd163aed2edd967a244499754dec991d8b4c7d ]
+
+Syzkaller reports an uninit value read from ax25cmp when sending raw message
+through ieee802154 implementation.
+
+=====================================================
+BUG: KMSAN: uninit-value in ax25cmp+0x3a5/0x460 net/ax25/ax25_addr.c:119
+ ax25cmp+0x3a5/0x460 net/ax25/ax25_addr.c:119
+ nr_dev_get+0x20e/0x450 net/netrom/nr_route.c:601
+ nr_route_frame+0x1a2/0xfc0 net/netrom/nr_route.c:774
+ nr_xmit+0x5a/0x1c0 net/netrom/nr_dev.c:144
+ __netdev_start_xmit include/linux/netdevice.h:4940 [inline]
+ netdev_start_xmit include/linux/netdevice.h:4954 [inline]
+ xmit_one net/core/dev.c:3548 [inline]
+ dev_hard_start_xmit+0x247/0xa10 net/core/dev.c:3564
+ __dev_queue_xmit+0x33b8/0x5130 net/core/dev.c:4349
+ dev_queue_xmit include/linux/netdevice.h:3134 [inline]
+ raw_sendmsg+0x654/0xc10 net/ieee802154/socket.c:299
+ ieee802154_sock_sendmsg+0x91/0xc0 net/ieee802154/socket.c:96
+ sock_sendmsg_nosec net/socket.c:730 [inline]
+ __sock_sendmsg net/socket.c:745 [inline]
+ ____sys_sendmsg+0x9c2/0xd60 net/socket.c:2584
+ ___sys_sendmsg+0x28d/0x3c0 net/socket.c:2638
+ __sys_sendmsg net/socket.c:2667 [inline]
+ __do_sys_sendmsg net/socket.c:2676 [inline]
+ __se_sys_sendmsg net/socket.c:2674 [inline]
+ __x64_sys_sendmsg+0x307/0x490 net/socket.c:2674
+ do_syscall_x64 arch/x86/entry/common.c:52 [inline]
+ do_syscall_64+0x44/0x110 arch/x86/entry/common.c:83
+ entry_SYSCALL_64_after_hwframe+0x63/0x6b
+
+Uninit was created at:
+ slab_post_alloc_hook+0x129/0xa70 mm/slab.h:768
+ slab_alloc_node mm/slub.c:3478 [inline]
+ kmem_cache_alloc_node+0x5e9/0xb10 mm/slub.c:3523
+ kmalloc_reserve+0x13d/0x4a0 net/core/skbuff.c:560
+ __alloc_skb+0x318/0x740 net/core/skbuff.c:651
+ alloc_skb include/linux/skbuff.h:1286 [inline]
+ alloc_skb_with_frags+0xc8/0xbd0 net/core/skbuff.c:6334
+ sock_alloc_send_pskb+0xa80/0xbf0 net/core/sock.c:2780
+ sock_alloc_send_skb include/net/sock.h:1884 [inline]
+ raw_sendmsg+0x36d/0xc10 net/ieee802154/socket.c:282
+ ieee802154_sock_sendmsg+0x91/0xc0 net/ieee802154/socket.c:96
+ sock_sendmsg_nosec net/socket.c:730 [inline]
+ __sock_sendmsg net/socket.c:745 [inline]
+ ____sys_sendmsg+0x9c2/0xd60 net/socket.c:2584
+ ___sys_sendmsg+0x28d/0x3c0 net/socket.c:2638
+ __sys_sendmsg net/socket.c:2667 [inline]
+ __do_sys_sendmsg net/socket.c:2676 [inline]
+ __se_sys_sendmsg net/socket.c:2674 [inline]
+ __x64_sys_sendmsg+0x307/0x490 net/socket.c:2674
+ do_syscall_x64 arch/x86/entry/common.c:52 [inline]
+ do_syscall_64+0x44/0x110 arch/x86/entry/common.c:83
+ entry_SYSCALL_64_after_hwframe+0x63/0x6b
+
+CPU: 0 PID: 5037 Comm: syz-executor166 Not tainted 6.7.0-rc7-syzkaller-00003-gfbafc3e621c3 #0
+Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 11/17/2023
+=====================================================
+
+This issue occurs because the skb buffer is too small, and it's actual
+allocation is aligned. This hides an actual issue, which is that nr_route_frame
+does not validate the buffer size before using it.
+
+Fix this issue by checking skb->len before accessing any fields in skb->data.
+
+Found by Linux Verification Center (linuxtesting.org) with Syzkaller.
+
+Fixes: 1da177e4c3f4 ("Linux-2.6.12-rc2")
+Co-developed-by: Nikita Marushkin <hfggklm@gmail.com>
+Signed-off-by: Nikita Marushkin <hfggklm@gmail.com>
+Signed-off-by: Ilya Shchipletsov <rabbelkin@mail.ru>
+Link: https://patch.msgid.link/20241219082308.3942-1-rabbelkin@mail.ru
+Signed-off-by: Jakub Kicinski <kuba@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ net/netrom/nr_route.c | 6 ++++++
+ 1 file changed, 6 insertions(+)
+
+diff --git a/net/netrom/nr_route.c b/net/netrom/nr_route.c
+index bd2b17b219ae..0b270893ee14 100644
+--- a/net/netrom/nr_route.c
++++ b/net/netrom/nr_route.c
+@@ -754,6 +754,12 @@ int nr_route_frame(struct sk_buff *skb, ax25_cb *ax25)
+       int ret;
+       struct sk_buff *skbn;
++      /*
++       * Reject malformed packets early. Check that it contains at least 2
++       * addresses and 1 byte more for Time-To-Live
++       */
++      if (skb->len < 2 * sizeof(ax25_address) + 1)
++              return 0;
+       nr_src  = (ax25_address *)(skb->data + 0);
+       nr_dest = (ax25_address *)(skb->data + 7);
+-- 
+2.39.5
+
diff --git a/queue-6.6/nvme-pci-512-byte-aligned-dma-pool-segment-quirk.patch b/queue-6.6/nvme-pci-512-byte-aligned-dma-pool-segment-quirk.patch
new file mode 100644 (file)
index 0000000..4e323d9
--- /dev/null
@@ -0,0 +1,100 @@
+From 4bf8d6a45448632dcf55c336f5d19a6fcc50607f Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Tue, 12 Nov 2024 19:50:00 +0000
+Subject: nvme-pci: 512 byte aligned dma pool segment quirk
+MIME-Version: 1.0
+Content-Type: text/plain; charset=UTF-8
+Content-Transfer-Encoding: 8bit
+
+From: Robert Beckett <bob.beckett@collabora.com>
+
+[ Upstream commit ebefac5647968679f6ef5803e5d35a71997d20fa ]
+
+We initially introduced a quick fix limiting the queue depth to 1 as
+experimentation showed that it fixed data corruption on 64GB steamdecks.
+
+Further experimentation revealed corruption only happens when the last
+PRP data element aligns to the end of the page boundary. The device
+appears to treat this as a PRP chain to a new list instead of the data
+element that it actually is. This implementation is in violation of the
+spec. Encountering this errata with the Linux driver requires the host
+request a 128k transfer and coincidently be handed the last small pool
+dma buffer within a page.
+
+The QD1 quirk effectly works around this because the last data PRP
+always was at a 248 byte offset from the page start, so it never
+appeared at the end of the page, but comes at the expense of throttling
+IO and wasting the remainder of the PRP page beyond 256 bytes. Also to
+note, the MDTS on these devices is small enough that the "large" prp
+pool can hold enough PRP elements to never reach the end, so that pool
+is not a problem either.
+
+Introduce a new quirk to ensure the small pool is always aligned such
+that the last PRP element can't appear a the end of the page. This comes
+at the expense of wasting 256 bytes per small pool page allocated.
+
+Link: https://lore.kernel.org/linux-nvme/20241113043151.GA20077@lst.de/T/#u
+Fixes: 83bdfcbdbe5d ("nvme-pci: qdepth 1 quirk")
+Cc: PaweÅ‚ Anikiel <panikiel@google.com>
+Signed-off-by: Robert Beckett <bob.beckett@collabora.com>
+Signed-off-by: Keith Busch <kbusch@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/nvme/host/nvme.h | 5 +++++
+ drivers/nvme/host/pci.c  | 9 +++++++--
+ 2 files changed, 12 insertions(+), 2 deletions(-)
+
+diff --git a/drivers/nvme/host/nvme.h b/drivers/nvme/host/nvme.h
+index bddc068d58c7..e867ac859a87 100644
+--- a/drivers/nvme/host/nvme.h
++++ b/drivers/nvme/host/nvme.h
+@@ -172,6 +172,11 @@ enum nvme_quirks {
+        * MSI (but not MSI-X) interrupts are broken and never fire.
+        */
+       NVME_QUIRK_BROKEN_MSI                   = (1 << 21),
++
++      /*
++       * Align dma pool segment size to 512 bytes
++       */
++      NVME_QUIRK_DMAPOOL_ALIGN_512            = (1 << 22),
+ };
+ /*
+diff --git a/drivers/nvme/host/pci.c b/drivers/nvme/host/pci.c
+index d525fa1229d7..52c8fd3d5c47 100644
+--- a/drivers/nvme/host/pci.c
++++ b/drivers/nvme/host/pci.c
+@@ -2653,15 +2653,20 @@ static int nvme_disable_prepare_reset(struct nvme_dev *dev, bool shutdown)
+ static int nvme_setup_prp_pools(struct nvme_dev *dev)
+ {
++      size_t small_align = 256;
++
+       dev->prp_page_pool = dma_pool_create("prp list page", dev->dev,
+                                               NVME_CTRL_PAGE_SIZE,
+                                               NVME_CTRL_PAGE_SIZE, 0);
+       if (!dev->prp_page_pool)
+               return -ENOMEM;
++      if (dev->ctrl.quirks & NVME_QUIRK_DMAPOOL_ALIGN_512)
++              small_align = 512;
++
+       /* Optimisation for I/Os between 4k and 128k */
+       dev->prp_small_pool = dma_pool_create("prp list 256", dev->dev,
+-                                              256, 256, 0);
++                                              256, small_align, 0);
+       if (!dev->prp_small_pool) {
+               dma_pool_destroy(dev->prp_page_pool);
+               return -ENOMEM;
+@@ -3403,7 +3408,7 @@ static const struct pci_device_id nvme_id_table[] = {
+       { PCI_VDEVICE(REDHAT, 0x0010),  /* Qemu emulated controller */
+               .driver_data = NVME_QUIRK_BOGUS_NID, },
+       { PCI_DEVICE(0x1217, 0x8760), /* O2 Micro 64GB Steam Deck */
+-              .driver_data = NVME_QUIRK_QDEPTH_ONE },
++              .driver_data = NVME_QUIRK_DMAPOOL_ALIGN_512, },
+       { PCI_DEVICE(0x126f, 0x2262),   /* Silicon Motion generic */
+               .driver_data = NVME_QUIRK_NO_DEEPEST_PS |
+                               NVME_QUIRK_BOGUS_NID, },
+-- 
+2.39.5
+
diff --git a/queue-6.6/rdma-bnxt_re-add-check-for-path-mtu-in-modify_qp.patch b/queue-6.6/rdma-bnxt_re-add-check-for-path-mtu-in-modify_qp.patch
new file mode 100644 (file)
index 0000000..ddb2b56
--- /dev/null
@@ -0,0 +1,64 @@
+From e8184adbe11b0034172597ff9bb4d46fb39f2afa Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Wed, 11 Dec 2024 14:09:28 +0530
+Subject: RDMA/bnxt_re: Add check for path mtu in modify_qp
+
+From: Saravanan Vajravel <saravanan.vajravel@broadcom.com>
+
+[ Upstream commit 798653a0ee30d3cd495099282751c0f248614ae7 ]
+
+When RDMA app configures path MTU, add a check in modify_qp verb
+to make sure that it doesn't go beyond interface MTU. If this
+check fails, driver will fail the modify_qp verb.
+
+Fixes: 1ac5a4047975 ("RDMA/bnxt_re: Add bnxt_re RoCE driver")
+Reviewed-by: Kalesh AP <kalesh-anakkur.purayil@broadcom.com>
+Signed-off-by: Saravanan Vajravel <saravanan.vajravel@broadcom.com>
+Signed-off-by: Selvin Xavier <selvin.xavier@broadcom.com>
+Link: https://patch.msgid.link/20241211083931.968831-3-kalesh-anakkur.purayil@broadcom.com
+Signed-off-by: Leon Romanovsky <leon@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/infiniband/hw/bnxt_re/ib_verbs.c | 26 +++++++++++++-----------
+ 1 file changed, 14 insertions(+), 12 deletions(-)
+
+diff --git a/drivers/infiniband/hw/bnxt_re/ib_verbs.c b/drivers/infiniband/hw/bnxt_re/ib_verbs.c
+index fb6f15bb9d4f..9bf00fb666d7 100644
+--- a/drivers/infiniband/hw/bnxt_re/ib_verbs.c
++++ b/drivers/infiniband/hw/bnxt_re/ib_verbs.c
+@@ -2055,18 +2055,20 @@ int bnxt_re_modify_qp(struct ib_qp *ib_qp, struct ib_qp_attr *qp_attr,
+               }
+       }
+-      if (qp_attr_mask & IB_QP_PATH_MTU) {
+-              qp->qplib_qp.modify_flags |=
+-                              CMDQ_MODIFY_QP_MODIFY_MASK_PATH_MTU;
+-              qp->qplib_qp.path_mtu = __from_ib_mtu(qp_attr->path_mtu);
+-              qp->qplib_qp.mtu = ib_mtu_enum_to_int(qp_attr->path_mtu);
+-      } else if (qp_attr->qp_state == IB_QPS_RTR) {
+-              qp->qplib_qp.modify_flags |=
+-                      CMDQ_MODIFY_QP_MODIFY_MASK_PATH_MTU;
+-              qp->qplib_qp.path_mtu =
+-                      __from_ib_mtu(iboe_get_mtu(rdev->netdev->mtu));
+-              qp->qplib_qp.mtu =
+-                      ib_mtu_enum_to_int(iboe_get_mtu(rdev->netdev->mtu));
++      if (qp_attr->qp_state == IB_QPS_RTR) {
++              enum ib_mtu qpmtu;
++
++              qpmtu = iboe_get_mtu(rdev->netdev->mtu);
++              if (qp_attr_mask & IB_QP_PATH_MTU) {
++                      if (ib_mtu_enum_to_int(qp_attr->path_mtu) >
++                          ib_mtu_enum_to_int(qpmtu))
++                              return -EINVAL;
++                      qpmtu = qp_attr->path_mtu;
++              }
++
++              qp->qplib_qp.modify_flags |= CMDQ_MODIFY_QP_MODIFY_MASK_PATH_MTU;
++              qp->qplib_qp.path_mtu = __from_ib_mtu(qpmtu);
++              qp->qplib_qp.mtu = ib_mtu_enum_to_int(qpmtu);
+       }
+       if (qp_attr_mask & IB_QP_TIMEOUT) {
+-- 
+2.39.5
+
diff --git a/queue-6.6/rdma-bnxt_re-add-send-queue-size-check-for-variable-.patch b/queue-6.6/rdma-bnxt_re-add-send-queue-size-check-for-variable-.patch
new file mode 100644 (file)
index 0000000..545c2b2
--- /dev/null
@@ -0,0 +1,44 @@
+From 75191d4e0103587edce5dd2e2d2431cfa6b63a7d Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Tue, 17 Dec 2024 15:56:47 +0530
+Subject: RDMA/bnxt_re: Add send queue size check for variable wqe
+
+From: Damodharam Ammepalli <damodharam.ammepalli@broadcom.com>
+
+[ Upstream commit d13be54dc18baee7a3e44349b80755a8c8205d3f ]
+
+For the fixed WQE case, HW supports 0xFFFF WQEs.
+For variable Size WQEs, HW treats this number as
+the 16 bytes slots. The maximum supported WQEs
+needs to be adjusted based on the number of slots.
+Set a maximum WQE limit for variable WQE scenario.
+
+Fixes: de1d364c3815 ("RDMA/bnxt_re: Add support for Variable WQE in Genp7 adapters")
+Reviewed-by: Kalesh AP <kalesh-anakkur.purayil@broadcom.com>
+Signed-off-by: Damodharam Ammepalli <damodharam.ammepalli@broadcom.com>
+Signed-off-by: Selvin Xavier <selvin.xavier@broadcom.com>
+Link: https://patch.msgid.link/20241217102649.1377704-4-kalesh-anakkur.purayil@broadcom.com
+Signed-off-by: Leon Romanovsky <leon@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/infiniband/hw/bnxt_re/qplib_sp.c | 4 ++++
+ 1 file changed, 4 insertions(+)
+
+diff --git a/drivers/infiniband/hw/bnxt_re/qplib_sp.c b/drivers/infiniband/hw/bnxt_re/qplib_sp.c
+index 577a6eaca4ce..74c3f6b26c4d 100644
+--- a/drivers/infiniband/hw/bnxt_re/qplib_sp.c
++++ b/drivers/infiniband/hw/bnxt_re/qplib_sp.c
+@@ -138,6 +138,10 @@ int bnxt_qplib_get_dev_attr(struct bnxt_qplib_rcfw *rcfw,
+               attr->max_qp_wqes -= BNXT_QPLIB_RESERVED_QP_WRS + 1;
+       }
++      /* Adjust for max_qp_wqes for variable wqe */
++      if (cctx->modes.wqe_mode == BNXT_QPLIB_WQE_MODE_VARIABLE)
++              attr->max_qp_wqes = BNXT_VAR_MAX_WQE - 1;
++
+       attr->max_qp_sges = cctx->modes.wqe_mode == BNXT_QPLIB_WQE_MODE_VARIABLE ?
+                           min_t(u32, sb->max_sge_var_wqe, BNXT_VAR_MAX_SGE) : 6;
+       attr->max_cq = le32_to_cpu(sb->max_cq);
+-- 
+2.39.5
+
diff --git a/queue-6.6/rdma-bnxt_re-add-support-for-variable-wqe-in-genp7-a.patch b/queue-6.6/rdma-bnxt_re-add-support-for-variable-wqe-in-genp7-a.patch
new file mode 100644 (file)
index 0000000..d692e05
--- /dev/null
@@ -0,0 +1,341 @@
+From 9cf513f7e32d861646be7bb1b3d8588245fab8da Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Sun, 18 Aug 2024 21:47:23 -0700
+Subject: RDMA/bnxt_re: Add support for Variable WQE in Genp7 adapters
+
+From: Selvin Xavier <selvin.xavier@broadcom.com>
+
+[ Upstream commit de1d364c3815f9360a0945097ca2731950e914fa ]
+
+Variable size WQE means that each send Work Queue Entry to HW can use
+different WQE sizes as opposed to the static WQE size on the current
+devices. Set variable WQE mode for Gen P7 devices. Depth of the Queue will
+be a multiple of slot which is 16 bytes. The number of slots should be a
+multiple of 256 as per the HW requirement.
+
+Initialize the Software shadow queue to hold requests equal to the number
+of slots. Also, do not expose the variable size WQE capability until the
+last patch in the series.
+
+Link: https://patch.msgid.link/r/1724042847-1481-2-git-send-email-selvin.xavier@broadcom.com
+Signed-off-by: Hongguang Gao <hongguang.gao@broadcom.com>
+Signed-off-by: Selvin Xavier <selvin.xavier@broadcom.com>
+Signed-off-by: Jason Gunthorpe <jgg@nvidia.com>
+Stable-dep-of: d5a38bf2f359 ("RDMA/bnxt_re: Disable use of reserved wqes")
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/infiniband/hw/bnxt_re/ib_verbs.c |  8 +++++---
+ drivers/infiniband/hw/bnxt_re/main.c     | 21 +++++++++++----------
+ drivers/infiniband/hw/bnxt_re/qplib_fp.c | 18 +++++++++---------
+ drivers/infiniband/hw/bnxt_re/qplib_fp.h | 14 +++++++++++---
+ drivers/infiniband/hw/bnxt_re/qplib_sp.c |  7 +++++--
+ drivers/infiniband/hw/bnxt_re/qplib_sp.h |  6 ++++++
+ 6 files changed, 47 insertions(+), 27 deletions(-)
+
+diff --git a/drivers/infiniband/hw/bnxt_re/ib_verbs.c b/drivers/infiniband/hw/bnxt_re/ib_verbs.c
+index 9e8f86f48801..540998ddbb44 100644
+--- a/drivers/infiniband/hw/bnxt_re/ib_verbs.c
++++ b/drivers/infiniband/hw/bnxt_re/ib_verbs.c
+@@ -1154,6 +1154,7 @@ static struct bnxt_re_qp *bnxt_re_create_shadow_qp
+       /* Shadow QP SQ depth should be same as QP1 RQ depth */
+       qp->qplib_qp.sq.wqe_size = bnxt_re_get_wqe_size(0, 6);
+       qp->qplib_qp.sq.max_wqe = qp1_qp->rq.max_wqe;
++      qp->qplib_qp.sq.max_sw_wqe = qp1_qp->rq.max_wqe;
+       qp->qplib_qp.sq.max_sge = 2;
+       /* Q full delta can be 1 since it is internal QP */
+       qp->qplib_qp.sq.q_full_delta = 1;
+@@ -1165,6 +1166,7 @@ static struct bnxt_re_qp *bnxt_re_create_shadow_qp
+       qp->qplib_qp.rq.wqe_size = bnxt_re_get_rwqe_size(6);
+       qp->qplib_qp.rq.max_wqe = qp1_qp->rq.max_wqe;
++      qp->qplib_qp.rq.max_sw_wqe = qp1_qp->rq.max_wqe;
+       qp->qplib_qp.rq.max_sge = qp1_qp->rq.max_sge;
+       /* Q full delta can be 1 since it is internal QP */
+       qp->qplib_qp.rq.q_full_delta = 1;
+@@ -1226,6 +1228,7 @@ static int bnxt_re_init_rq_attr(struct bnxt_re_qp *qp,
+                */
+               entries = bnxt_re_init_depth(init_attr->cap.max_recv_wr + 1, uctx);
+               rq->max_wqe = min_t(u32, entries, dev_attr->max_qp_wqes + 1);
++              rq->max_sw_wqe = rq->max_wqe;
+               rq->q_full_delta = 0;
+               rq->sg_info.pgsize = PAGE_SIZE;
+               rq->sg_info.pgshft = PAGE_SHIFT;
+@@ -1285,6 +1288,7 @@ static int bnxt_re_init_sq_attr(struct bnxt_re_qp *qp,
+               0 : BNXT_QPLIB_RESERVED_QP_WRS;
+       entries = bnxt_re_init_depth(entries + diff + 1, uctx);
+       sq->max_wqe = min_t(u32, entries, dev_attr->max_qp_wqes + diff + 1);
++      sq->max_sw_wqe = bnxt_qplib_get_depth(sq, qplqp->wqe_mode, true);
+       sq->q_full_delta = diff + 1;
+       /*
+        * Reserving one slot for Phantom WQE. Application can
+@@ -2155,6 +2159,7 @@ int bnxt_re_modify_qp(struct ib_qp *ib_qp, struct ib_qp_attr *qp_attr,
+                       entries = bnxt_re_init_depth(qp_attr->cap.max_recv_wr, uctx);
+                       qp->qplib_qp.rq.max_wqe =
+                               min_t(u32, entries, dev_attr->max_qp_wqes + 1);
++                      qp->qplib_qp.rq.max_sw_wqe = qp->qplib_qp.rq.max_wqe;
+                       qp->qplib_qp.rq.q_full_delta = qp->qplib_qp.rq.max_wqe -
+                                                      qp_attr->cap.max_recv_wr;
+                       qp->qplib_qp.rq.max_sge = qp_attr->cap.max_recv_sge;
+@@ -4171,9 +4176,6 @@ int bnxt_re_alloc_ucontext(struct ib_ucontext *ctx, struct ib_udata *udata)
+       resp.cqe_sz = sizeof(struct cq_base);
+       resp.max_cqd = dev_attr->max_cq_wqes;
+-      resp.comp_mask |= BNXT_RE_UCNTX_CMASK_HAVE_MODE;
+-      resp.mode = rdev->chip_ctx->modes.wqe_mode;
+-
+       if (rdev->chip_ctx->modes.db_push)
+               resp.comp_mask |= BNXT_RE_UCNTX_CMASK_WC_DPI_ENABLED;
+diff --git a/drivers/infiniband/hw/bnxt_re/main.c b/drivers/infiniband/hw/bnxt_re/main.c
+index 0373d0e9db63..c7e51cc2ea26 100644
+--- a/drivers/infiniband/hw/bnxt_re/main.c
++++ b/drivers/infiniband/hw/bnxt_re/main.c
+@@ -128,13 +128,13 @@ static void bnxt_re_set_db_offset(struct bnxt_re_dev *rdev)
+       }
+ }
+-static void bnxt_re_set_drv_mode(struct bnxt_re_dev *rdev, u8 mode)
++static void bnxt_re_set_drv_mode(struct bnxt_re_dev *rdev)
+ {
+       struct bnxt_qplib_chip_ctx *cctx;
+       cctx = rdev->chip_ctx;
+-      cctx->modes.wqe_mode = bnxt_qplib_is_chip_gen_p5_p7(rdev->chip_ctx) ?
+-                             mode : BNXT_QPLIB_WQE_MODE_STATIC;
++      cctx->modes.wqe_mode = bnxt_qplib_is_chip_gen_p7(rdev->chip_ctx) ?
++                             BNXT_QPLIB_WQE_MODE_VARIABLE : BNXT_QPLIB_WQE_MODE_STATIC;
+       if (bnxt_re_hwrm_qcaps(rdev))
+               dev_err(rdev_to_dev(rdev),
+                       "Failed to query hwrm qcaps\n");
+@@ -155,7 +155,7 @@ static void bnxt_re_destroy_chip_ctx(struct bnxt_re_dev *rdev)
+       kfree(chip_ctx);
+ }
+-static int bnxt_re_setup_chip_ctx(struct bnxt_re_dev *rdev, u8 wqe_mode)
++static int bnxt_re_setup_chip_ctx(struct bnxt_re_dev *rdev)
+ {
+       struct bnxt_qplib_chip_ctx *chip_ctx;
+       struct bnxt_en_dev *en_dev;
+@@ -177,7 +177,7 @@ static int bnxt_re_setup_chip_ctx(struct bnxt_re_dev *rdev, u8 wqe_mode)
+       rdev->qplib_res.dattr = &rdev->dev_attr;
+       rdev->qplib_res.is_vf = BNXT_EN_VF(en_dev);
+-      bnxt_re_set_drv_mode(rdev, wqe_mode);
++      bnxt_re_set_drv_mode(rdev);
+       bnxt_re_set_db_offset(rdev);
+       rc = bnxt_qplib_map_db_bar(&rdev->qplib_res);
+@@ -1440,7 +1440,7 @@ static void bnxt_re_worker(struct work_struct *work)
+       schedule_delayed_work(&rdev->worker, msecs_to_jiffies(30000));
+ }
+-static int bnxt_re_dev_init(struct bnxt_re_dev *rdev, u8 wqe_mode)
++static int bnxt_re_dev_init(struct bnxt_re_dev *rdev)
+ {
+       struct bnxt_re_ring_attr rattr = {};
+       struct bnxt_qplib_creq_ctx *creq;
+@@ -1458,7 +1458,7 @@ static int bnxt_re_dev_init(struct bnxt_re_dev *rdev, u8 wqe_mode)
+       }
+       set_bit(BNXT_RE_FLAG_NETDEV_REGISTERED, &rdev->flags);
+-      rc = bnxt_re_setup_chip_ctx(rdev, wqe_mode);
++      rc = bnxt_re_setup_chip_ctx(rdev);
+       if (rc) {
+               bnxt_unregister_dev(rdev->en_dev);
+               clear_bit(BNXT_RE_FLAG_NETDEV_REGISTERED, &rdev->flags);
+@@ -1609,7 +1609,7 @@ static int bnxt_re_dev_init(struct bnxt_re_dev *rdev, u8 wqe_mode)
+       return rc;
+ }
+-static int bnxt_re_add_device(struct auxiliary_device *adev, u8 wqe_mode)
++static int bnxt_re_add_device(struct auxiliary_device *adev)
+ {
+       struct bnxt_aux_priv *aux_priv =
+               container_of(adev, struct bnxt_aux_priv, aux_dev);
+@@ -1626,7 +1626,7 @@ static int bnxt_re_add_device(struct auxiliary_device *adev, u8 wqe_mode)
+               goto exit;
+       }
+-      rc = bnxt_re_dev_init(rdev, wqe_mode);
++      rc = bnxt_re_dev_init(rdev);
+       if (rc)
+               goto re_dev_dealloc;
+@@ -1756,7 +1756,8 @@ static int bnxt_re_probe(struct auxiliary_device *adev,
+       int rc;
+       mutex_lock(&bnxt_re_mutex);
+-      rc = bnxt_re_add_device(adev, BNXT_QPLIB_WQE_MODE_STATIC);
++
++      rc = bnxt_re_add_device(adev);
+       if (rc) {
+               mutex_unlock(&bnxt_re_mutex);
+               return rc;
+diff --git a/drivers/infiniband/hw/bnxt_re/qplib_fp.c b/drivers/infiniband/hw/bnxt_re/qplib_fp.c
+index 8997f359b58b..2f85245d1285 100644
+--- a/drivers/infiniband/hw/bnxt_re/qplib_fp.c
++++ b/drivers/infiniband/hw/bnxt_re/qplib_fp.c
+@@ -807,13 +807,13 @@ static int bnxt_qplib_alloc_init_swq(struct bnxt_qplib_q *que)
+ {
+       int indx;
+-      que->swq = kcalloc(que->max_wqe, sizeof(*que->swq), GFP_KERNEL);
++      que->swq = kcalloc(que->max_sw_wqe, sizeof(*que->swq), GFP_KERNEL);
+       if (!que->swq)
+               return -ENOMEM;
+       que->swq_start = 0;
+-      que->swq_last = que->max_wqe - 1;
+-      for (indx = 0; indx < que->max_wqe; indx++)
++      que->swq_last = que->max_sw_wqe - 1;
++      for (indx = 0; indx < que->max_sw_wqe; indx++)
+               que->swq[indx].next_idx = indx + 1;
+       que->swq[que->swq_last].next_idx = 0; /* Make it circular */
+       que->swq_last = 0;
+@@ -849,7 +849,7 @@ int bnxt_qplib_create_qp1(struct bnxt_qplib_res *res, struct bnxt_qplib_qp *qp)
+       hwq_attr.res = res;
+       hwq_attr.sginfo = &sq->sg_info;
+       hwq_attr.stride = sizeof(struct sq_sge);
+-      hwq_attr.depth = bnxt_qplib_get_depth(sq);
++      hwq_attr.depth = bnxt_qplib_get_depth(sq, qp->wqe_mode, false);
+       hwq_attr.type = HWQ_TYPE_QUEUE;
+       rc = bnxt_qplib_alloc_init_hwq(&sq->hwq, &hwq_attr);
+       if (rc)
+@@ -877,7 +877,7 @@ int bnxt_qplib_create_qp1(struct bnxt_qplib_res *res, struct bnxt_qplib_qp *qp)
+               hwq_attr.res = res;
+               hwq_attr.sginfo = &rq->sg_info;
+               hwq_attr.stride = sizeof(struct sq_sge);
+-              hwq_attr.depth = bnxt_qplib_get_depth(rq);
++              hwq_attr.depth = bnxt_qplib_get_depth(rq, qp->wqe_mode, false);
+               hwq_attr.type = HWQ_TYPE_QUEUE;
+               rc = bnxt_qplib_alloc_init_hwq(&rq->hwq, &hwq_attr);
+               if (rc)
+@@ -1007,7 +1007,7 @@ int bnxt_qplib_create_qp(struct bnxt_qplib_res *res, struct bnxt_qplib_qp *qp)
+       hwq_attr.res = res;
+       hwq_attr.sginfo = &sq->sg_info;
+       hwq_attr.stride = sizeof(struct sq_sge);
+-      hwq_attr.depth = bnxt_qplib_get_depth(sq);
++      hwq_attr.depth = bnxt_qplib_get_depth(sq, qp->wqe_mode, true);
+       hwq_attr.aux_stride = psn_sz;
+       hwq_attr.aux_depth = psn_sz ? bnxt_qplib_set_sq_size(sq, qp->wqe_mode)
+                                   : 0;
+@@ -1049,7 +1049,7 @@ int bnxt_qplib_create_qp(struct bnxt_qplib_res *res, struct bnxt_qplib_qp *qp)
+               hwq_attr.res = res;
+               hwq_attr.sginfo = &rq->sg_info;
+               hwq_attr.stride = sizeof(struct sq_sge);
+-              hwq_attr.depth = bnxt_qplib_get_depth(rq);
++              hwq_attr.depth = bnxt_qplib_get_depth(rq, qp->wqe_mode, false);
+               hwq_attr.aux_stride = 0;
+               hwq_attr.aux_depth = 0;
+               hwq_attr.type = HWQ_TYPE_QUEUE;
+@@ -2493,7 +2493,7 @@ static int bnxt_qplib_cq_process_req(struct bnxt_qplib_cq *cq,
+       }
+       sq = &qp->sq;
+-      cqe_sq_cons = le16_to_cpu(hwcqe->sq_cons_idx) % sq->max_wqe;
++      cqe_sq_cons = le16_to_cpu(hwcqe->sq_cons_idx) % sq->max_sw_wqe;
+       if (qp->sq.flushed) {
+               dev_dbg(&cq->hwq.pdev->dev,
+                       "%s: QP in Flush QP = %p\n", __func__, qp);
+@@ -2885,7 +2885,7 @@ static int bnxt_qplib_cq_process_terminal(struct bnxt_qplib_cq *cq,
+       cqe_cons = le16_to_cpu(hwcqe->sq_cons_idx);
+       if (cqe_cons == 0xFFFF)
+               goto do_rq;
+-      cqe_cons %= sq->max_wqe;
++      cqe_cons %= sq->max_sw_wqe;
+       if (qp->sq.flushed) {
+               dev_dbg(&cq->hwq.pdev->dev,
+diff --git a/drivers/infiniband/hw/bnxt_re/qplib_fp.h b/drivers/infiniband/hw/bnxt_re/qplib_fp.h
+index 3a15ca7feb2b..b64746d484d6 100644
+--- a/drivers/infiniband/hw/bnxt_re/qplib_fp.h
++++ b/drivers/infiniband/hw/bnxt_re/qplib_fp.h
+@@ -251,6 +251,7 @@ struct bnxt_qplib_q {
+       struct bnxt_qplib_db_info       dbinfo;
+       struct bnxt_qplib_sg_info       sg_info;
+       u32                             max_wqe;
++      u32                             max_sw_wqe;
+       u16                             wqe_size;
+       u16                             q_full_delta;
+       u16                             max_sge;
+@@ -585,15 +586,22 @@ static inline void bnxt_qplib_swq_mod_start(struct bnxt_qplib_q *que, u32 idx)
+       que->swq_start = que->swq[idx].next_idx;
+ }
+-static inline u32 bnxt_qplib_get_depth(struct bnxt_qplib_q *que)
++static inline u32 bnxt_qplib_get_depth(struct bnxt_qplib_q *que, u8 wqe_mode, bool is_sq)
+ {
+-      return (que->wqe_size * que->max_wqe) / sizeof(struct sq_sge);
++      u32 slots;
++
++      /* Queue depth is the number of slots. */
++      slots = (que->wqe_size * que->max_wqe) / sizeof(struct sq_sge);
++      /* For variable WQE mode, need to align the slots to 256 */
++      if (wqe_mode == BNXT_QPLIB_WQE_MODE_VARIABLE && is_sq)
++              slots = ALIGN(slots, BNXT_VAR_MAX_SLOT_ALIGN);
++      return slots;
+ }
+ static inline u32 bnxt_qplib_set_sq_size(struct bnxt_qplib_q *que, u8 wqe_mode)
+ {
+       return (wqe_mode == BNXT_QPLIB_WQE_MODE_STATIC) ?
+-              que->max_wqe : bnxt_qplib_get_depth(que);
++              que->max_wqe : bnxt_qplib_get_depth(que, wqe_mode, true);
+ }
+ static inline u32 bnxt_qplib_set_sq_max_slot(u8 wqe_mode)
+diff --git a/drivers/infiniband/hw/bnxt_re/qplib_sp.c b/drivers/infiniband/hw/bnxt_re/qplib_sp.c
+index 0f6bae009af1..a46df2a5ab33 100644
+--- a/drivers/infiniband/hw/bnxt_re/qplib_sp.c
++++ b/drivers/infiniband/hw/bnxt_re/qplib_sp.c
+@@ -95,11 +95,13 @@ int bnxt_qplib_get_dev_attr(struct bnxt_qplib_rcfw *rcfw,
+       struct bnxt_qplib_cmdqmsg msg = {};
+       struct creq_query_func_resp_sb *sb;
+       struct bnxt_qplib_rcfw_sbuf sbuf;
++      struct bnxt_qplib_chip_ctx *cctx;
+       struct cmdq_query_func req = {};
+       u8 *tqm_alloc;
+       int i, rc;
+       u32 temp;
++      cctx = rcfw->res->cctx;
+       bnxt_qplib_rcfw_cmd_prep((struct cmdq_base *)&req,
+                                CMDQ_BASE_OPCODE_QUERY_FUNC,
+                                sizeof(req));
+@@ -133,8 +135,9 @@ int bnxt_qplib_get_dev_attr(struct bnxt_qplib_rcfw *rcfw,
+        * reporting the max number
+        */
+       attr->max_qp_wqes -= BNXT_QPLIB_RESERVED_QP_WRS + 1;
+-      attr->max_qp_sges = bnxt_qplib_is_chip_gen_p5_p7(rcfw->res->cctx) ?
+-                          6 : sb->max_sge;
++
++      attr->max_qp_sges = cctx->modes.wqe_mode == BNXT_QPLIB_WQE_MODE_VARIABLE ?
++                          min_t(u32, sb->max_sge_var_wqe, BNXT_VAR_MAX_SGE) : 6;
+       attr->max_cq = le32_to_cpu(sb->max_cq);
+       attr->max_cq_wqes = le32_to_cpu(sb->max_cqe);
+       if (!bnxt_qplib_is_chip_gen_p7(rcfw->res->cctx))
+diff --git a/drivers/infiniband/hw/bnxt_re/qplib_sp.h b/drivers/infiniband/hw/bnxt_re/qplib_sp.h
+index 2f16f3db093e..b91e6a85e75d 100644
+--- a/drivers/infiniband/hw/bnxt_re/qplib_sp.h
++++ b/drivers/infiniband/hw/bnxt_re/qplib_sp.h
+@@ -40,6 +40,7 @@
+ #ifndef __BNXT_QPLIB_SP_H__
+ #define __BNXT_QPLIB_SP_H__
++#include <rdma/bnxt_re-abi.h>
+ #define BNXT_QPLIB_RESERVED_QP_WRS    128
+ struct bnxt_qplib_dev_attr {
+@@ -352,4 +353,9 @@ int bnxt_qplib_qext_stat(struct bnxt_qplib_rcfw *rcfw, u32 fid,
+ int bnxt_qplib_modify_cc(struct bnxt_qplib_res *res,
+                        struct bnxt_qplib_cc_param *cc_param);
++#define BNXT_VAR_MAX_WQE       4352
++#define BNXT_VAR_MAX_SLOT_ALIGN 256
++#define BNXT_VAR_MAX_SGE        13
++#define BNXT_RE_MAX_RQ_WQES     65536
++
+ #endif /* __BNXT_QPLIB_SP_H__*/
+-- 
+2.39.5
+
diff --git a/queue-6.6/rdma-bnxt_re-allow-msn-table-capability-check.patch b/queue-6.6/rdma-bnxt_re-allow-msn-table-capability-check.patch
new file mode 100644 (file)
index 0000000..599740b
--- /dev/null
@@ -0,0 +1,188 @@
+From f21846f13df33c255e0bbfdce2cf89bb8260d412 Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Mon, 27 May 2024 23:11:36 -0700
+Subject: RDMA/bnxt_re: Allow MSN table capability check
+
+From: Selvin Xavier <selvin.xavier@broadcom.com>
+
+[ Upstream commit 8d310ba845827a38fcd463d86bfe3b730ce7ab8f ]
+
+FW reports the HW capability to use PSN table or MSN table and
+driver/library need to select it based on this capability.
+Use the new capability instead of the older capability check for HW
+retransmission while handling the MSN/PSN table. FW report
+zero (PSN table) for older adapters to maintain backward compatibility.
+
+Also, Updated the FW interface structures to handle the new fields.
+
+Signed-off-by: Selvin Xavier <selvin.xavier@broadcom.com>
+Link: https://lore.kernel.org/r/1716876697-25970-2-git-send-email-selvin.xavier@broadcom.com
+Signed-off-by: Leon Romanovsky <leon@kernel.org>
+Stable-dep-of: eb867d797d29 ("RDMA/bnxt_re: Remove always true dattr validity check")
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/infiniband/hw/bnxt_re/qplib_fp.c  | 12 ++++-----
+ drivers/infiniband/hw/bnxt_re/qplib_fp.h  |  2 +-
+ drivers/infiniband/hw/bnxt_re/qplib_res.h |  6 +++++
+ drivers/infiniband/hw/bnxt_re/qplib_sp.c  |  1 +
+ drivers/infiniband/hw/bnxt_re/qplib_sp.h  |  1 +
+ drivers/infiniband/hw/bnxt_re/roce_hsi.h  | 30 ++++++++++++++++++++++-
+ 6 files changed, 44 insertions(+), 8 deletions(-)
+
+diff --git a/drivers/infiniband/hw/bnxt_re/qplib_fp.c b/drivers/infiniband/hw/bnxt_re/qplib_fp.c
+index b624c255eee6..3e07500dcbcf 100644
+--- a/drivers/infiniband/hw/bnxt_re/qplib_fp.c
++++ b/drivers/infiniband/hw/bnxt_re/qplib_fp.c
+@@ -981,7 +981,7 @@ int bnxt_qplib_create_qp(struct bnxt_qplib_res *res, struct bnxt_qplib_qp *qp)
+       u16 nsge;
+       if (res->dattr)
+-              qp->dev_cap_flags = res->dattr->dev_cap_flags;
++              qp->is_host_msn_tbl = _is_host_msn_table(res->dattr->dev_cap_flags2);
+       sq->dbinfo.flags = 0;
+       bnxt_qplib_rcfw_cmd_prep((struct cmdq_base *)&req,
+@@ -999,7 +999,7 @@ int bnxt_qplib_create_qp(struct bnxt_qplib_res *res, struct bnxt_qplib_qp *qp)
+                        sizeof(struct sq_psn_search_ext) :
+                        sizeof(struct sq_psn_search);
+-              if (BNXT_RE_HW_RETX(qp->dev_cap_flags)) {
++              if (qp->is_host_msn_tbl) {
+                       psn_sz = sizeof(struct sq_msn_search);
+                       qp->msn = 0;
+               }
+@@ -1013,7 +1013,7 @@ int bnxt_qplib_create_qp(struct bnxt_qplib_res *res, struct bnxt_qplib_qp *qp)
+       hwq_attr.aux_depth = psn_sz ? bnxt_qplib_set_sq_size(sq, qp->wqe_mode)
+                                   : 0;
+       /* Update msn tbl size */
+-      if (BNXT_RE_HW_RETX(qp->dev_cap_flags) && psn_sz) {
++      if (qp->is_host_msn_tbl && psn_sz) {
+               hwq_attr.aux_depth = roundup_pow_of_two(bnxt_qplib_set_sq_size(sq, qp->wqe_mode));
+               qp->msn_tbl_sz = hwq_attr.aux_depth;
+               qp->msn = 0;
+@@ -1638,7 +1638,7 @@ static void bnxt_qplib_fill_psn_search(struct bnxt_qplib_qp *qp,
+       if (!swq->psn_search)
+               return;
+       /* Handle MSN differently on cap flags  */
+-      if (BNXT_RE_HW_RETX(qp->dev_cap_flags)) {
++      if (qp->is_host_msn_tbl) {
+               bnxt_qplib_fill_msn_search(qp, wqe, swq);
+               return;
+       }
+@@ -1820,7 +1820,7 @@ int bnxt_qplib_post_send(struct bnxt_qplib_qp *qp,
+       }
+       swq = bnxt_qplib_get_swqe(sq, &wqe_idx);
+-      bnxt_qplib_pull_psn_buff(qp, sq, swq, BNXT_RE_HW_RETX(qp->dev_cap_flags));
++      bnxt_qplib_pull_psn_buff(qp, sq, swq, qp->is_host_msn_tbl);
+       idx = 0;
+       swq->slot_idx = hwq->prod;
+@@ -2010,7 +2010,7 @@ int bnxt_qplib_post_send(struct bnxt_qplib_qp *qp,
+               rc = -EINVAL;
+               goto done;
+       }
+-      if (!BNXT_RE_HW_RETX(qp->dev_cap_flags) || msn_update) {
++      if (!qp->is_host_msn_tbl || msn_update) {
+               swq->next_psn = sq->psn & BTH_PSN_MASK;
+               bnxt_qplib_fill_psn_search(qp, wqe, swq);
+       }
+diff --git a/drivers/infiniband/hw/bnxt_re/qplib_fp.h b/drivers/infiniband/hw/bnxt_re/qplib_fp.h
+index 5d4c49089a20..3a15ca7feb2b 100644
+--- a/drivers/infiniband/hw/bnxt_re/qplib_fp.h
++++ b/drivers/infiniband/hw/bnxt_re/qplib_fp.h
+@@ -340,7 +340,7 @@ struct bnxt_qplib_qp {
+       struct list_head                rq_flush;
+       u32                             msn;
+       u32                             msn_tbl_sz;
+-      u16                             dev_cap_flags;
++      bool                            is_host_msn_tbl;
+ };
+ #define BNXT_QPLIB_MAX_CQE_ENTRY_SIZE sizeof(struct cq_base)
+diff --git a/drivers/infiniband/hw/bnxt_re/qplib_res.h b/drivers/infiniband/hw/bnxt_re/qplib_res.h
+index f9e7aa3757cf..c2152122a432 100644
+--- a/drivers/infiniband/hw/bnxt_re/qplib_res.h
++++ b/drivers/infiniband/hw/bnxt_re/qplib_res.h
+@@ -523,6 +523,12 @@ static inline bool _is_hw_retx_supported(u16 dev_cap_flags)
+ #define BNXT_RE_HW_RETX(a) _is_hw_retx_supported((a))
++static inline bool _is_host_msn_table(u16 dev_cap_ext_flags2)
++{
++      return (dev_cap_ext_flags2 & CREQ_QUERY_FUNC_RESP_SB_REQ_RETRANSMISSION_SUPPORT_MASK) ==
++              CREQ_QUERY_FUNC_RESP_SB_REQ_RETRANSMISSION_SUPPORT_HOST_MSN_TABLE;
++}
++
+ static inline u8 bnxt_qplib_dbr_pacing_en(struct bnxt_qplib_chip_ctx *cctx)
+ {
+       return cctx->modes.dbr_pacing;
+diff --git a/drivers/infiniband/hw/bnxt_re/qplib_sp.c b/drivers/infiniband/hw/bnxt_re/qplib_sp.c
+index 0b98577cd708..420f8613bcd5 100644
+--- a/drivers/infiniband/hw/bnxt_re/qplib_sp.c
++++ b/drivers/infiniband/hw/bnxt_re/qplib_sp.c
+@@ -165,6 +165,7 @@ int bnxt_qplib_get_dev_attr(struct bnxt_qplib_rcfw *rcfw,
+       attr->max_sgid = le32_to_cpu(sb->max_gid);
+       attr->max_sgid = min_t(u32, BNXT_QPLIB_NUM_GIDS_SUPPORTED, 2 * attr->max_sgid);
+       attr->dev_cap_flags = le16_to_cpu(sb->dev_cap_flags);
++      attr->dev_cap_flags2 = le16_to_cpu(sb->dev_cap_ext_flags_2);
+       bnxt_qplib_query_version(rcfw, attr->fw_ver);
+diff --git a/drivers/infiniband/hw/bnxt_re/qplib_sp.h b/drivers/infiniband/hw/bnxt_re/qplib_sp.h
+index 755765e68eaa..2f16f3db093e 100644
+--- a/drivers/infiniband/hw/bnxt_re/qplib_sp.h
++++ b/drivers/infiniband/hw/bnxt_re/qplib_sp.h
+@@ -73,6 +73,7 @@ struct bnxt_qplib_dev_attr {
+       u8                              tqm_alloc_reqs[MAX_TQM_ALLOC_REQ];
+       bool                            is_atomic;
+       u16                             dev_cap_flags;
++      u16                             dev_cap_flags2;
+       u32                             max_dpi;
+ };
+diff --git a/drivers/infiniband/hw/bnxt_re/roce_hsi.h b/drivers/infiniband/hw/bnxt_re/roce_hsi.h
+index 2909608f4b5d..cb4e7e19fbaf 100644
+--- a/drivers/infiniband/hw/bnxt_re/roce_hsi.h
++++ b/drivers/infiniband/hw/bnxt_re/roce_hsi.h
+@@ -2157,8 +2157,36 @@ struct creq_query_func_resp_sb {
+       __le32  tqm_alloc_reqs[12];
+       __le32  max_dpi;
+       u8      max_sge_var_wqe;
+-      u8      reserved_8;
++      u8      dev_cap_ext_flags;
++      #define CREQ_QUERY_FUNC_RESP_SB_ATOMIC_OPS_NOT_SUPPORTED         0x1UL
++      #define CREQ_QUERY_FUNC_RESP_SB_DRV_VERSION_RGTR_SUPPORTED       0x2UL
++      #define CREQ_QUERY_FUNC_RESP_SB_CREATE_QP_BATCH_SUPPORTED        0x4UL
++      #define CREQ_QUERY_FUNC_RESP_SB_DESTROY_QP_BATCH_SUPPORTED       0x8UL
++      #define CREQ_QUERY_FUNC_RESP_SB_ROCE_STATS_EXT_CTX_SUPPORTED     0x10UL
++      #define CREQ_QUERY_FUNC_RESP_SB_CREATE_SRQ_SGE_SUPPORTED         0x20UL
++      #define CREQ_QUERY_FUNC_RESP_SB_FIXED_SIZE_WQE_DISABLED          0x40UL
++      #define CREQ_QUERY_FUNC_RESP_SB_DCN_SUPPORTED                    0x80UL
+       __le16  max_inline_data_var_wqe;
++      __le32  start_qid;
++      u8      max_msn_table_size;
++      u8      reserved8_1;
++      __le16  dev_cap_ext_flags_2;
++      #define CREQ_QUERY_FUNC_RESP_SB_OPTIMIZE_MODIFY_QP_SUPPORTED             0x1UL
++      #define CREQ_QUERY_FUNC_RESP_SB_CHANGE_UDP_SRC_PORT_WQE_SUPPORTED        0x2UL
++      #define CREQ_QUERY_FUNC_RESP_SB_CQ_COALESCING_SUPPORTED                  0x4UL
++      #define CREQ_QUERY_FUNC_RESP_SB_MEMORY_REGION_RO_SUPPORTED               0x8UL
++      #define CREQ_QUERY_FUNC_RESP_SB_REQ_RETRANSMISSION_SUPPORT_MASK          0x30UL
++      #define CREQ_QUERY_FUNC_RESP_SB_REQ_RETRANSMISSION_SUPPORT_SFT           4
++      #define CREQ_QUERY_FUNC_RESP_SB_REQ_RETRANSMISSION_SUPPORT_HOST_PSN_TABLE  (0x0UL << 4)
++      #define CREQ_QUERY_FUNC_RESP_SB_REQ_RETRANSMISSION_SUPPORT_HOST_MSN_TABLE  (0x1UL << 4)
++      #define CREQ_QUERY_FUNC_RESP_SB_REQ_RETRANSMISSION_SUPPORT_IQM_MSN_TABLE   (0x2UL << 4)
++      #define CREQ_QUERY_FUNC_RESP_SB_REQ_RETRANSMISSION_SUPPORT_LAST \
++                      CREQ_QUERY_FUNC_RESP_SB_REQ_RETRANSMISSION_SUPPORT_IQM_MSN_TABLE
++      __le16  max_xp_qp_size;
++      __le16  create_qp_batch_size;
++      __le16  destroy_qp_batch_size;
++      __le16  reserved16;
++      __le64  reserved64;
+ };
+ /* cmdq_set_func_resources (size:448b/56B) */
+-- 
+2.39.5
+
diff --git a/queue-6.6/rdma-bnxt_re-avoid-initializing-the-software-queue-f.patch b/queue-6.6/rdma-bnxt_re-avoid-initializing-the-software-queue-f.patch
new file mode 100644 (file)
index 0000000..d7b17b4
--- /dev/null
@@ -0,0 +1,101 @@
+From 9b0c1439551947f766272603de2b9ba99ef22cb9 Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Wed, 4 Dec 2024 13:24:13 +0530
+Subject: RDMA/bnxt_re: Avoid initializing the software queue for user queues
+
+From: Selvin Xavier <selvin.xavier@broadcom.com>
+
+[ Upstream commit 5effcacc8a8f3eb2a9f069d7e81a9ac793598dfb ]
+
+Software Queues to hold the WRs needs to be created
+for only kernel queues. Avoid allocating the unnecessary
+memory for user Queues.
+
+Fixes: 1ac5a4047975 ("RDMA/bnxt_re: Add bnxt_re RoCE driver")
+Fixes: 159fb4ceacd7 ("RDMA/bnxt_re: introduce a function to allocate swq")
+Signed-off-by: Selvin Xavier <selvin.xavier@broadcom.com>
+Link: https://patch.msgid.link/20241204075416.478431-3-kalesh-anakkur.purayil@broadcom.com
+Signed-off-by: Leon Romanovsky <leon@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/infiniband/hw/bnxt_re/qplib_fp.c | 42 +++++++++++++-----------
+ 1 file changed, 23 insertions(+), 19 deletions(-)
+
+diff --git a/drivers/infiniband/hw/bnxt_re/qplib_fp.c b/drivers/infiniband/hw/bnxt_re/qplib_fp.c
+index 4098e01666d1..d38e7880cebb 100644
+--- a/drivers/infiniband/hw/bnxt_re/qplib_fp.c
++++ b/drivers/infiniband/hw/bnxt_re/qplib_fp.c
+@@ -639,13 +639,6 @@ int bnxt_qplib_create_srq(struct bnxt_qplib_res *res,
+       rc = bnxt_qplib_alloc_init_hwq(&srq->hwq, &hwq_attr);
+       if (rc)
+               return rc;
+-
+-      srq->swq = kcalloc(srq->hwq.max_elements, sizeof(*srq->swq),
+-                         GFP_KERNEL);
+-      if (!srq->swq) {
+-              rc = -ENOMEM;
+-              goto fail;
+-      }
+       srq->dbinfo.flags = 0;
+       bnxt_qplib_rcfw_cmd_prep((struct cmdq_base *)&req,
+                                CMDQ_BASE_OPCODE_CREATE_SRQ,
+@@ -674,9 +667,17 @@ int bnxt_qplib_create_srq(struct bnxt_qplib_res *res,
+       spin_lock_init(&srq->lock);
+       srq->start_idx = 0;
+       srq->last_idx = srq->hwq.max_elements - 1;
+-      for (idx = 0; idx < srq->hwq.max_elements; idx++)
+-              srq->swq[idx].next_idx = idx + 1;
+-      srq->swq[srq->last_idx].next_idx = -1;
++      if (!srq->hwq.is_user) {
++              srq->swq = kcalloc(srq->hwq.max_elements, sizeof(*srq->swq),
++                                 GFP_KERNEL);
++              if (!srq->swq) {
++                      rc = -ENOMEM;
++                      goto fail;
++              }
++              for (idx = 0; idx < srq->hwq.max_elements; idx++)
++                      srq->swq[idx].next_idx = idx + 1;
++              srq->swq[srq->last_idx].next_idx = -1;
++      }
+       srq->id = le32_to_cpu(resp.xid);
+       srq->dbinfo.hwq = &srq->hwq;
+@@ -1022,13 +1023,14 @@ int bnxt_qplib_create_qp(struct bnxt_qplib_res *res, struct bnxt_qplib_qp *qp)
+       if (rc)
+               return rc;
+-      rc = bnxt_qplib_alloc_init_swq(sq);
+-      if (rc)
+-              goto fail_sq;
+-
+-      if (psn_sz)
+-              bnxt_qplib_init_psn_ptr(qp, psn_sz);
++      if (!sq->hwq.is_user) {
++              rc = bnxt_qplib_alloc_init_swq(sq);
++              if (rc)
++                      goto fail_sq;
++              if (psn_sz)
++                      bnxt_qplib_init_psn_ptr(qp, psn_sz);
++      }
+       req.sq_size = cpu_to_le32(bnxt_qplib_set_sq_size(sq, qp->wqe_mode));
+       pbl = &sq->hwq.pbl[PBL_LVL_0];
+       req.sq_pbl = cpu_to_le64(pbl->pg_map_arr[0]);
+@@ -1054,9 +1056,11 @@ int bnxt_qplib_create_qp(struct bnxt_qplib_res *res, struct bnxt_qplib_qp *qp)
+               rc = bnxt_qplib_alloc_init_hwq(&rq->hwq, &hwq_attr);
+               if (rc)
+                       goto sq_swq;
+-              rc = bnxt_qplib_alloc_init_swq(rq);
+-              if (rc)
+-                      goto fail_rq;
++              if (!rq->hwq.is_user) {
++                      rc = bnxt_qplib_alloc_init_swq(rq);
++                      if (rc)
++                              goto fail_rq;
++              }
+               req.rq_size = cpu_to_le32(rq->max_wqe);
+               pbl = &rq->hwq.pbl[PBL_LVL_0];
+-- 
+2.39.5
+
diff --git a/queue-6.6/rdma-bnxt_re-avoid-sending-the-modify-qp-workaround-.patch b/queue-6.6/rdma-bnxt_re-avoid-sending-the-modify-qp-workaround-.patch
new file mode 100644 (file)
index 0000000..24818d5
--- /dev/null
@@ -0,0 +1,51 @@
+From 14ff715323d88c09e56b51d24d5af8a4d00c249f Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Wed, 4 Dec 2024 13:24:14 +0530
+Subject: RDMA/bnxt_re: Avoid sending the modify QP workaround for latest
+ adapters
+
+From: Kashyap Desai <kashyap.desai@broadcom.com>
+
+[ Upstream commit 064c22408a73b9e945139b64614c534cbbefb591 ]
+
+The workaround to modify the UD QP from RTS to RTS is required
+only for older adapters. Issuing this for latest adapters can caus
+some unexpected behavior. Fix it
+
+Fixes: 1801d87b3598 ("RDMA/bnxt_re: Support new 5760X P7 devices")
+Signed-off-by: Kashyap Desai <kashyap.desai@broadcom.com>
+Signed-off-by: Selvin Xavier <selvin.xavier@broadcom.com>
+Link: https://patch.msgid.link/20241204075416.478431-4-kalesh-anakkur.purayil@broadcom.com
+Signed-off-by: Leon Romanovsky <leon@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/infiniband/hw/bnxt_re/ib_verbs.c | 6 ++++--
+ 1 file changed, 4 insertions(+), 2 deletions(-)
+
+diff --git a/drivers/infiniband/hw/bnxt_re/ib_verbs.c b/drivers/infiniband/hw/bnxt_re/ib_verbs.c
+index df5897260601..fb6f15bb9d4f 100644
+--- a/drivers/infiniband/hw/bnxt_re/ib_verbs.c
++++ b/drivers/infiniband/hw/bnxt_re/ib_verbs.c
+@@ -2710,7 +2710,8 @@ static int bnxt_re_post_send_shadow_qp(struct bnxt_re_dev *rdev,
+               wr = wr->next;
+       }
+       bnxt_qplib_post_send_db(&qp->qplib_qp);
+-      bnxt_ud_qp_hw_stall_workaround(qp);
++      if (!bnxt_qplib_is_chip_gen_p5_p7(qp->rdev->chip_ctx))
++              bnxt_ud_qp_hw_stall_workaround(qp);
+       spin_unlock_irqrestore(&qp->sq_lock, flags);
+       return rc;
+ }
+@@ -2822,7 +2823,8 @@ int bnxt_re_post_send(struct ib_qp *ib_qp, const struct ib_send_wr *wr,
+               wr = wr->next;
+       }
+       bnxt_qplib_post_send_db(&qp->qplib_qp);
+-      bnxt_ud_qp_hw_stall_workaround(qp);
++      if (!bnxt_qplib_is_chip_gen_p5_p7(qp->rdev->chip_ctx))
++              bnxt_ud_qp_hw_stall_workaround(qp);
+       spin_unlock_irqrestore(&qp->sq_lock, flags);
+       return rc;
+-- 
+2.39.5
+
diff --git a/queue-6.6/rdma-bnxt_re-disable-use-of-reserved-wqes.patch b/queue-6.6/rdma-bnxt_re-disable-use-of-reserved-wqes.patch
new file mode 100644 (file)
index 0000000..faa0c01
--- /dev/null
@@ -0,0 +1,48 @@
+From 5cf48f928edb635d870f6efdd5d5f3e55ec53d0b Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Tue, 17 Dec 2024 15:56:46 +0530
+Subject: RDMA/bnxt_re: Disable use of reserved wqes
+
+From: Kalesh AP <kalesh-anakkur.purayil@broadcom.com>
+
+[ Upstream commit d5a38bf2f35979537c526acbc56bc435ed40685f ]
+
+Disabling the reserved wqes logic for Gen P5/P7 devices
+because this workaround is required only for legacy devices.
+
+Fixes: ecb53febfcad ("RDMA/bnxt_en: Enable RDMA driver support for 57500 chip")
+Signed-off-by: Kalesh AP <kalesh-anakkur.purayil@broadcom.com>
+Signed-off-by: Selvin Xavier <selvin.xavier@broadcom.com>
+Link: https://patch.msgid.link/20241217102649.1377704-3-kalesh-anakkur.purayil@broadcom.com
+Signed-off-by: Leon Romanovsky <leon@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/infiniband/hw/bnxt_re/qplib_sp.c | 12 +++++++-----
+ 1 file changed, 7 insertions(+), 5 deletions(-)
+
+diff --git a/drivers/infiniband/hw/bnxt_re/qplib_sp.c b/drivers/infiniband/hw/bnxt_re/qplib_sp.c
+index a46df2a5ab33..577a6eaca4ce 100644
+--- a/drivers/infiniband/hw/bnxt_re/qplib_sp.c
++++ b/drivers/infiniband/hw/bnxt_re/qplib_sp.c
+@@ -130,11 +130,13 @@ int bnxt_qplib_get_dev_attr(struct bnxt_qplib_rcfw *rcfw,
+               sb->max_qp_init_rd_atom > BNXT_QPLIB_MAX_OUT_RD_ATOM ?
+               BNXT_QPLIB_MAX_OUT_RD_ATOM : sb->max_qp_init_rd_atom;
+       attr->max_qp_wqes = le16_to_cpu(sb->max_qp_wr) - 1;
+-      /*
+-       * 128 WQEs needs to be reserved for the HW (8916). Prevent
+-       * reporting the max number
+-       */
+-      attr->max_qp_wqes -= BNXT_QPLIB_RESERVED_QP_WRS + 1;
++      if (!bnxt_qplib_is_chip_gen_p5_p7(rcfw->res->cctx)) {
++              /*
++               * 128 WQEs needs to be reserved for the HW (8916). Prevent
++               * reporting the max number on legacy devices
++               */
++              attr->max_qp_wqes -= BNXT_QPLIB_RESERVED_QP_WRS + 1;
++      }
+       attr->max_qp_sges = cctx->modes.wqe_mode == BNXT_QPLIB_WQE_MODE_VARIABLE ?
+                           min_t(u32, sb->max_sge_var_wqe, BNXT_VAR_MAX_SGE) : 6;
+-- 
+2.39.5
+
diff --git a/queue-6.6/rdma-bnxt_re-fix-max_qp_wrs-reported.patch b/queue-6.6/rdma-bnxt_re-fix-max_qp_wrs-reported.patch
new file mode 100644 (file)
index 0000000..dbb6907
--- /dev/null
@@ -0,0 +1,42 @@
+From a4ce37aee9e409f7768a587819bd5574c46d3f77 Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Tue, 17 Dec 2024 15:56:45 +0530
+Subject: RDMA/bnxt_re: Fix max_qp_wrs reported
+
+From: Selvin Xavier <selvin.xavier@broadcom.com>
+
+[ Upstream commit 40be32303ec829ea12f9883e499bfd3fe9e52baf ]
+
+While creating qps, driver adds one extra entry to the sq size
+passed by the ULPs in order to avoid queue full condition.
+When ULPs creates QPs with max_qp_wr reported, driver creates
+QP with 1 more than the max_wqes supported by HW. Create QP fails
+in this case. To avoid this error, reduce 1 entry in max_qp_wqes
+and report it to the stack.
+
+Fixes: 1ac5a4047975 ("RDMA/bnxt_re: Add bnxt_re RoCE driver")
+Reviewed-by: Kalesh AP <kalesh-anakkur.purayil@broadcom.com>
+Signed-off-by: Selvin Xavier <selvin.xavier@broadcom.com>
+Link: https://patch.msgid.link/20241217102649.1377704-2-kalesh-anakkur.purayil@broadcom.com
+Signed-off-by: Leon Romanovsky <leon@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/infiniband/hw/bnxt_re/qplib_sp.c | 2 +-
+ 1 file changed, 1 insertion(+), 1 deletion(-)
+
+diff --git a/drivers/infiniband/hw/bnxt_re/qplib_sp.c b/drivers/infiniband/hw/bnxt_re/qplib_sp.c
+index 420f8613bcd5..0f6bae009af1 100644
+--- a/drivers/infiniband/hw/bnxt_re/qplib_sp.c
++++ b/drivers/infiniband/hw/bnxt_re/qplib_sp.c
+@@ -127,7 +127,7 @@ int bnxt_qplib_get_dev_attr(struct bnxt_qplib_rcfw *rcfw,
+       attr->max_qp_init_rd_atom =
+               sb->max_qp_init_rd_atom > BNXT_QPLIB_MAX_OUT_RD_ATOM ?
+               BNXT_QPLIB_MAX_OUT_RD_ATOM : sb->max_qp_init_rd_atom;
+-      attr->max_qp_wqes = le16_to_cpu(sb->max_qp_wr);
++      attr->max_qp_wqes = le16_to_cpu(sb->max_qp_wr) - 1;
+       /*
+        * 128 WQEs needs to be reserved for the HW (8916). Prevent
+        * reporting the max number
+-- 
+2.39.5
+
diff --git a/queue-6.6/rdma-bnxt_re-fix-msn-table-size-for-variable-wqe-mod.patch b/queue-6.6/rdma-bnxt_re-fix-msn-table-size-for-variable-wqe-mod.patch
new file mode 100644 (file)
index 0000000..d3bfd3b
--- /dev/null
@@ -0,0 +1,46 @@
+From b4c5f4227398601d3c94393134595c9853ab2b4d Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Tue, 17 Dec 2024 15:56:48 +0530
+Subject: RDMA/bnxt_re: Fix MSN table size for variable wqe mode
+
+From: Damodharam Ammepalli <damodharam.ammepalli@broadcom.com>
+
+[ Upstream commit bb839f3ace0fee532a0487b692cc4d868fccb7cf ]
+
+For variable size wqe mode, the MSN table size should be
+half the size of the SQ depth. Fixing this to avoid wrap
+around problems in the retransmission path.
+
+Fixes: de1d364c3815 ("RDMA/bnxt_re: Add support for Variable WQE in Genp7 adapters")
+Reviewed-by: Kashyap Desai <kashyap.desai@broadcom.com>
+Reviewed-by: Kalesh AP <kalesh-anakkur.purayil@broadcom.com>
+Signed-off-by: Damodharam Ammepalli <damodharam.ammepalli@broadcom.com>
+Signed-off-by: Selvin Xavier <selvin.xavier@broadcom.com>
+Link: https://patch.msgid.link/20241217102649.1377704-5-kalesh-anakkur.purayil@broadcom.com
+Signed-off-by: Leon Romanovsky <leon@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/infiniband/hw/bnxt_re/qplib_fp.c | 7 ++++++-
+ 1 file changed, 6 insertions(+), 1 deletion(-)
+
+diff --git a/drivers/infiniband/hw/bnxt_re/qplib_fp.c b/drivers/infiniband/hw/bnxt_re/qplib_fp.c
+index 2f85245d1285..1355061d698d 100644
+--- a/drivers/infiniband/hw/bnxt_re/qplib_fp.c
++++ b/drivers/infiniband/hw/bnxt_re/qplib_fp.c
+@@ -1013,7 +1013,12 @@ int bnxt_qplib_create_qp(struct bnxt_qplib_res *res, struct bnxt_qplib_qp *qp)
+                                   : 0;
+       /* Update msn tbl size */
+       if (qp->is_host_msn_tbl && psn_sz) {
+-              hwq_attr.aux_depth = roundup_pow_of_two(bnxt_qplib_set_sq_size(sq, qp->wqe_mode));
++              if (qp->wqe_mode == BNXT_QPLIB_WQE_MODE_STATIC)
++                      hwq_attr.aux_depth =
++                              roundup_pow_of_two(bnxt_qplib_set_sq_size(sq, qp->wqe_mode));
++              else
++                      hwq_attr.aux_depth =
++                              roundup_pow_of_two(bnxt_qplib_set_sq_size(sq, qp->wqe_mode)) / 2;
+               qp->msn_tbl_sz = hwq_attr.aux_depth;
+               qp->msn = 0;
+       }
+-- 
+2.39.5
+
diff --git a/queue-6.6/rdma-bnxt_re-fix-reporting-hw_ver-in-query_device.patch b/queue-6.6/rdma-bnxt_re-fix-reporting-hw_ver-in-query_device.patch
new file mode 100644 (file)
index 0000000..905455c
--- /dev/null
@@ -0,0 +1,41 @@
+From 5e89bcfc8f48c134d5b24a780e6dc2f9adad334f Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Wed, 11 Dec 2024 14:09:31 +0530
+Subject: RDMA/bnxt_re: Fix reporting hw_ver in query_device
+
+From: Kalesh AP <kalesh-anakkur.purayil@broadcom.com>
+
+[ Upstream commit 7179fe0074a3c962e43a9e51169304c4911989ed ]
+
+Driver currently populates subsystem_device id in the
+"hw_ver" field of ib_attr structure in query_device.
+
+Updated to populate PCI revision ID.
+
+Fixes: 1ac5a4047975 ("RDMA/bnxt_re: Add bnxt_re RoCE driver")
+Reviewed-by: Preethi G <preethi.gurusiddalingeswaraswamy@broadcom.com>
+Signed-off-by: Kalesh AP <kalesh-anakkur.purayil@broadcom.com>
+Signed-off-by: Selvin Xavier <selvin.xavier@broadcom.com>
+Link: https://patch.msgid.link/20241211083931.968831-6-kalesh-anakkur.purayil@broadcom.com
+Signed-off-by: Leon Romanovsky <leon@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/infiniband/hw/bnxt_re/ib_verbs.c | 2 +-
+ 1 file changed, 1 insertion(+), 1 deletion(-)
+
+diff --git a/drivers/infiniband/hw/bnxt_re/ib_verbs.c b/drivers/infiniband/hw/bnxt_re/ib_verbs.c
+index 9bf00fb666d7..9e8f86f48801 100644
+--- a/drivers/infiniband/hw/bnxt_re/ib_verbs.c
++++ b/drivers/infiniband/hw/bnxt_re/ib_verbs.c
+@@ -147,7 +147,7 @@ int bnxt_re_query_device(struct ib_device *ibdev,
+       ib_attr->vendor_id = rdev->en_dev->pdev->vendor;
+       ib_attr->vendor_part_id = rdev->en_dev->pdev->device;
+-      ib_attr->hw_ver = rdev->en_dev->pdev->subsystem_device;
++      ib_attr->hw_ver = rdev->en_dev->pdev->revision;
+       ib_attr->max_qp = dev_attr->max_qp;
+       ib_attr->max_qp_wr = dev_attr->max_qp_wqes;
+       ib_attr->device_cap_flags =
+-- 
+2.39.5
+
diff --git a/queue-6.6/rdma-bnxt_re-fix-the-check-for-9060-condition.patch b/queue-6.6/rdma-bnxt_re-fix-the-check-for-9060-condition.patch
new file mode 100644 (file)
index 0000000..3af325c
--- /dev/null
@@ -0,0 +1,46 @@
+From 6e413f8ed1e82d20d4a86fc9aa5bede0c787a26f Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Wed, 11 Dec 2024 14:09:27 +0530
+Subject: RDMA/bnxt_re: Fix the check for 9060 condition
+
+From: Kalesh AP <kalesh-anakkur.purayil@broadcom.com>
+
+[ Upstream commit 38651476e46e088598354510502c383e932e2297 ]
+
+The check for 9060 condition should only be made for legacy chips.
+
+Fixes: 9152e0b722b2 ("RDMA/bnxt_re: HW workarounds for handling specific conditions")
+Reviewed-by: Kashyap Desai <kashyap.desai@broadcom.com>
+Signed-off-by: Kalesh AP <kalesh-anakkur.purayil@broadcom.com>
+Signed-off-by: Selvin Xavier <selvin.xavier@broadcom.com>
+Link: https://patch.msgid.link/20241211083931.968831-2-kalesh-anakkur.purayil@broadcom.com
+Signed-off-by: Leon Romanovsky <leon@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/infiniband/hw/bnxt_re/qplib_fp.c | 10 ++++++----
+ 1 file changed, 6 insertions(+), 4 deletions(-)
+
+diff --git a/drivers/infiniband/hw/bnxt_re/qplib_fp.c b/drivers/infiniband/hw/bnxt_re/qplib_fp.c
+index d38e7880cebb..8997f359b58b 100644
+--- a/drivers/infiniband/hw/bnxt_re/qplib_fp.c
++++ b/drivers/infiniband/hw/bnxt_re/qplib_fp.c
+@@ -2536,10 +2536,12 @@ static int bnxt_qplib_cq_process_req(struct bnxt_qplib_cq *cq,
+                       bnxt_qplib_add_flush_qp(qp);
+               } else {
+                       /* Before we complete, do WA 9060 */
+-                      if (do_wa9060(qp, cq, cq_cons, sq->swq_last,
+-                                    cqe_sq_cons)) {
+-                              *lib_qp = qp;
+-                              goto out;
++                      if (!bnxt_qplib_is_chip_gen_p5_p7(qp->cctx)) {
++                              if (do_wa9060(qp, cq, cq_cons, sq->swq_last,
++                                            cqe_sq_cons)) {
++                                      *lib_qp = qp;
++                                      goto out;
++                              }
+                       }
+                       if (swq->flags & SQ_SEND_FLAGS_SIGNAL_COMP) {
+                               cqe->status = CQ_REQ_STATUS_OK;
+-- 
+2.39.5
+
diff --git a/queue-6.6/rdma-bnxt_re-fix-the-locking-while-accessing-the-qp-.patch b/queue-6.6/rdma-bnxt_re-fix-the-locking-while-accessing-the-qp-.patch
new file mode 100644 (file)
index 0000000..aaf4897
--- /dev/null
@@ -0,0 +1,43 @@
+From 8f88a60a922a7fdce085cc5d357dbadd67b1811f Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Tue, 17 Dec 2024 15:56:49 +0530
+Subject: RDMA/bnxt_re: Fix the locking while accessing the QP table
+
+From: Selvin Xavier <selvin.xavier@broadcom.com>
+
+[ Upstream commit 9272cba0ded71b5a2084da3004ec7806b8cb7fd2 ]
+
+QP table handling is synchronized with destroy QP and Async
+event from the HW. The same needs to be synchronized
+during create_qp also. Use the same lock in create_qp also.
+
+Fixes: 76d3ddff7153 ("RDMA/bnxt_re: synchronize the qp-handle table array")
+Fixes: f218d67ef004 ("RDMA/bnxt_re: Allow posting when QPs are in error")
+Fixes: 84cf229f4001 ("RDMA/bnxt_re: Fix the qp table indexing")
+Signed-off-by: Selvin Xavier <selvin.xavier@broadcom.com>
+Link: https://patch.msgid.link/20241217102649.1377704-6-kalesh-anakkur.purayil@broadcom.com
+Signed-off-by: Leon Romanovsky <leon@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/infiniband/hw/bnxt_re/qplib_fp.c | 2 ++
+ 1 file changed, 2 insertions(+)
+
+diff --git a/drivers/infiniband/hw/bnxt_re/qplib_fp.c b/drivers/infiniband/hw/bnxt_re/qplib_fp.c
+index 1355061d698d..871a49315c88 100644
+--- a/drivers/infiniband/hw/bnxt_re/qplib_fp.c
++++ b/drivers/infiniband/hw/bnxt_re/qplib_fp.c
+@@ -1161,9 +1161,11 @@ int bnxt_qplib_create_qp(struct bnxt_qplib_res *res, struct bnxt_qplib_qp *qp)
+               rq->dbinfo.db = qp->dpi->dbr;
+               rq->dbinfo.max_slot = bnxt_qplib_set_rq_max_slot(rq->wqe_size);
+       }
++      spin_lock_bh(&rcfw->tbl_lock);
+       tbl_indx = map_qp_id_to_tbl_indx(qp->id, rcfw);
+       rcfw->qp_tbl[tbl_indx].qp_id = qp->id;
+       rcfw->qp_tbl[tbl_indx].qp_handle = (void *)qp;
++      spin_unlock_bh(&rcfw->tbl_lock);
+       return 0;
+ fail:
+-- 
+2.39.5
+
diff --git a/queue-6.6/rdma-bnxt_re-remove-always-true-dattr-validity-check.patch b/queue-6.6/rdma-bnxt_re-remove-always-true-dattr-validity-check.patch
new file mode 100644 (file)
index 0000000..8c9d41a
--- /dev/null
@@ -0,0 +1,46 @@
+From c207ca764b8962acc8ba8380c9e53e60e39e2b71 Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Tue, 26 Nov 2024 15:10:31 +0200
+Subject: RDMA/bnxt_re: Remove always true dattr validity check
+
+From: Leon Romanovsky <leonro@nvidia.com>
+
+[ Upstream commit eb867d797d294a00a092b5027d08439da68940b2 ]
+
+res->dattr is always valid at this point as it was initialized
+during device addition in bnxt_re_add_device().
+
+This change is fixing the following smatch error:
+drivers/infiniband/hw/bnxt_re/qplib_fp.c:1090 bnxt_qplib_create_qp()
+     error: we previously assumed 'res->dattr' could be null (see line 985)
+
+Fixes: 07f830ae4913 ("RDMA/bnxt_re: Adds MSN table capability for Gen P7 adapters")
+Reported-by: kernel test robot <lkp@intel.com>
+Reported-by: Dan Carpenter <dan.carpenter@linaro.org>
+Closes: https://lore.kernel.org/r/202411222329.YTrwonWi-lkp@intel.com/
+Link: https://patch.msgid.link/be0d8836b64cba3e479fbcbca717acad04aae02e.1732626579.git.leonro@nvidia.com
+Acked-by: Selvin Xavier <selvin.xavier@broadcom.com>
+Signed-off-by: Leon Romanovsky <leonro@nvidia.com>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/infiniband/hw/bnxt_re/qplib_fp.c | 4 +---
+ 1 file changed, 1 insertion(+), 3 deletions(-)
+
+diff --git a/drivers/infiniband/hw/bnxt_re/qplib_fp.c b/drivers/infiniband/hw/bnxt_re/qplib_fp.c
+index 3e07500dcbcf..4098e01666d1 100644
+--- a/drivers/infiniband/hw/bnxt_re/qplib_fp.c
++++ b/drivers/infiniband/hw/bnxt_re/qplib_fp.c
+@@ -980,9 +980,7 @@ int bnxt_qplib_create_qp(struct bnxt_qplib_res *res, struct bnxt_qplib_qp *qp)
+       u32 tbl_indx;
+       u16 nsge;
+-      if (res->dattr)
+-              qp->is_host_msn_tbl = _is_host_msn_table(res->dattr->dev_cap_flags2);
+-
++      qp->is_host_msn_tbl = _is_host_msn_table(res->dattr->dev_cap_flags2);
+       sq->dbinfo.flags = 0;
+       bnxt_qplib_rcfw_cmd_prep((struct cmdq_base *)&req,
+                                CMDQ_BASE_OPCODE_CREATE_QP,
+-- 
+2.39.5
+
diff --git a/queue-6.6/rdma-hns-fix-mapping-error-of-zero-hop-wqe-buffer.patch b/queue-6.6/rdma-hns-fix-mapping-error-of-zero-hop-wqe-buffer.patch
new file mode 100644 (file)
index 0000000..e0e4de7
--- /dev/null
@@ -0,0 +1,177 @@
+From 8cc9ac82061a71ae3b77e583a7c9c234b2ef8a6a Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Fri, 20 Dec 2024 13:52:46 +0800
+Subject: RDMA/hns: Fix mapping error of zero-hop WQE buffer
+
+From: wenglianfa <wenglianfa@huawei.com>
+
+[ Upstream commit 8673a6c2d9e483dfeeef83a1f06f59e05636f4d1 ]
+
+Due to HW limitation, the three region of WQE buffer must be mapped
+and set to HW in a fixed order: SQ buffer, SGE buffer, and RQ buffer.
+
+Currently when one region is zero-hop while the other two are not,
+the zero-hop region will not be mapped. This violate the limitation
+above and leads to address error.
+
+Fixes: 38389eaa4db1 ("RDMA/hns: Add mtr support for mixed multihop addressing")
+Signed-off-by: wenglianfa <wenglianfa@huawei.com>
+Signed-off-by: Junxian Huang <huangjunxian6@hisilicon.com>
+Link: https://patch.msgid.link/20241220055249.146943-2-huangjunxian6@hisilicon.com
+Signed-off-by: Leon Romanovsky <leon@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/infiniband/hw/hns/hns_roce_hem.c | 43 ++++++++++++++++--------
+ drivers/infiniband/hw/hns/hns_roce_mr.c  |  5 ---
+ 2 files changed, 29 insertions(+), 19 deletions(-)
+
+diff --git a/drivers/infiniband/hw/hns/hns_roce_hem.c b/drivers/infiniband/hw/hns/hns_roce_hem.c
+index d3fabf64e390..51ab6041ca91 100644
+--- a/drivers/infiniband/hw/hns/hns_roce_hem.c
++++ b/drivers/infiniband/hw/hns/hns_roce_hem.c
+@@ -986,6 +986,7 @@ struct hns_roce_hem_item {
+       size_t count; /* max ba numbers */
+       int start; /* start buf offset in this hem */
+       int end; /* end buf offset in this hem */
++      bool exist_bt;
+ };
+ /* All HEM items are linked in a tree structure */
+@@ -1014,6 +1015,7 @@ hem_list_alloc_item(struct hns_roce_dev *hr_dev, int start, int end, int count,
+               }
+       }
++      hem->exist_bt = exist_bt;
+       hem->count = count;
+       hem->start = start;
+       hem->end = end;
+@@ -1024,22 +1026,22 @@ hem_list_alloc_item(struct hns_roce_dev *hr_dev, int start, int end, int count,
+ }
+ static void hem_list_free_item(struct hns_roce_dev *hr_dev,
+-                             struct hns_roce_hem_item *hem, bool exist_bt)
++                             struct hns_roce_hem_item *hem)
+ {
+-      if (exist_bt)
++      if (hem->exist_bt)
+               dma_free_coherent(hr_dev->dev, hem->count * BA_BYTE_LEN,
+                                 hem->addr, hem->dma_addr);
+       kfree(hem);
+ }
+ static void hem_list_free_all(struct hns_roce_dev *hr_dev,
+-                            struct list_head *head, bool exist_bt)
++                            struct list_head *head)
+ {
+       struct hns_roce_hem_item *hem, *temp_hem;
+       list_for_each_entry_safe(hem, temp_hem, head, list) {
+               list_del(&hem->list);
+-              hem_list_free_item(hr_dev, hem, exist_bt);
++              hem_list_free_item(hr_dev, hem);
+       }
+ }
+@@ -1139,6 +1141,10 @@ int hns_roce_hem_list_calc_root_ba(const struct hns_roce_buf_region *regions,
+       for (i = 0; i < region_cnt; i++) {
+               r = (struct hns_roce_buf_region *)&regions[i];
++              /* when r->hopnum = 0, the region should not occupy root_ba. */
++              if (!r->hopnum)
++                      continue;
++
+               if (r->hopnum > 1) {
+                       step = hem_list_calc_ba_range(r->hopnum, 1, unit);
+                       if (step > 0)
+@@ -1232,7 +1238,7 @@ static int hem_list_alloc_mid_bt(struct hns_roce_dev *hr_dev,
+ err_exit:
+       for (level = 1; level < hopnum; level++)
+-              hem_list_free_all(hr_dev, &temp_list[level], true);
++              hem_list_free_all(hr_dev, &temp_list[level]);
+       return ret;
+ }
+@@ -1273,16 +1279,26 @@ static int alloc_fake_root_bt(struct hns_roce_dev *hr_dev, void *cpu_base,
+ {
+       struct hns_roce_hem_item *hem;
++      /* This is on the has_mtt branch, if r->hopnum
++       * is 0, there is no root_ba to reuse for the
++       * region's fake hem, so a dma_alloc request is
++       * necessary here.
++       */
+       hem = hem_list_alloc_item(hr_dev, r->offset, r->offset + r->count - 1,
+-                                r->count, false);
++                                r->count, !r->hopnum);
+       if (!hem)
+               return -ENOMEM;
+-      hem_list_assign_bt(hem, cpu_base, phy_base);
++      /* The root_ba can be reused only when r->hopnum > 0. */
++      if (r->hopnum)
++              hem_list_assign_bt(hem, cpu_base, phy_base);
+       list_add(&hem->list, branch_head);
+       list_add(&hem->sibling, leaf_head);
+-      return r->count;
++      /* If r->hopnum == 0, 0 is returned,
++       * so that the root_bt entry is not occupied.
++       */
++      return r->hopnum ? r->count : 0;
+ }
+ static int setup_middle_bt(struct hns_roce_dev *hr_dev, void *cpu_base,
+@@ -1326,7 +1342,7 @@ setup_root_hem(struct hns_roce_dev *hr_dev, struct hns_roce_hem_list *hem_list,
+               return -ENOMEM;
+       total = 0;
+-      for (i = 0; i < region_cnt && total < max_ba_num; i++) {
++      for (i = 0; i < region_cnt && total <= max_ba_num; i++) {
+               r = &regions[i];
+               if (!r->count)
+                       continue;
+@@ -1392,9 +1408,9 @@ static int hem_list_alloc_root_bt(struct hns_roce_dev *hr_dev,
+                            region_cnt);
+       if (ret) {
+               for (i = 0; i < region_cnt; i++)
+-                      hem_list_free_all(hr_dev, &head.branch[i], false);
++                      hem_list_free_all(hr_dev, &head.branch[i]);
+-              hem_list_free_all(hr_dev, &head.root, true);
++              hem_list_free_all(hr_dev, &head.root);
+       }
+       return ret;
+@@ -1457,10 +1473,9 @@ void hns_roce_hem_list_release(struct hns_roce_dev *hr_dev,
+       for (i = 0; i < HNS_ROCE_MAX_BT_REGION; i++)
+               for (j = 0; j < HNS_ROCE_MAX_BT_LEVEL; j++)
+-                      hem_list_free_all(hr_dev, &hem_list->mid_bt[i][j],
+-                                        j != 0);
++                      hem_list_free_all(hr_dev, &hem_list->mid_bt[i][j]);
+-      hem_list_free_all(hr_dev, &hem_list->root_bt, true);
++      hem_list_free_all(hr_dev, &hem_list->root_bt);
+       INIT_LIST_HEAD(&hem_list->btm_bt);
+       hem_list->root_ba = 0;
+ }
+diff --git a/drivers/infiniband/hw/hns/hns_roce_mr.c b/drivers/infiniband/hw/hns/hns_roce_mr.c
+index acdcd5c9f42f..408ef2a96149 100644
+--- a/drivers/infiniband/hw/hns/hns_roce_mr.c
++++ b/drivers/infiniband/hw/hns/hns_roce_mr.c
+@@ -767,11 +767,6 @@ int hns_roce_mtr_map(struct hns_roce_dev *hr_dev, struct hns_roce_mtr *mtr,
+       for (i = 0, mapped_cnt = 0; i < mtr->hem_cfg.region_count &&
+            mapped_cnt < page_cnt; i++) {
+               r = &mtr->hem_cfg.region[i];
+-              /* if hopnum is 0, no need to map pages in this region */
+-              if (!r->hopnum) {
+-                      mapped_cnt += r->count;
+-                      continue;
+-              }
+               if (r->offset + r->count > page_cnt) {
+                       ret = -EINVAL;
+-- 
+2.39.5
+
diff --git a/queue-6.6/rdma-hns-fix-missing-flush-cqe-for-dwqe.patch b/queue-6.6/rdma-hns-fix-missing-flush-cqe-for-dwqe.patch
new file mode 100644 (file)
index 0000000..251e42c
--- /dev/null
@@ -0,0 +1,58 @@
+From ae41e3d2944f30f19cc31376a93687f6559c1a3d Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Fri, 20 Dec 2024 13:52:49 +0800
+Subject: RDMA/hns: Fix missing flush CQE for DWQE
+
+From: Chengchang Tang <tangchengchang@huawei.com>
+
+[ Upstream commit e3debdd48423d3d75b9d366399228d7225d902cd ]
+
+Flush CQE handler has not been called if QP state gets into errored
+mode in DWQE path. So, the new added outstanding WQEs will never be
+flushed.
+
+It leads to a hung task timeout when using NFS over RDMA:
+    __switch_to+0x7c/0xd0
+    __schedule+0x350/0x750
+    schedule+0x50/0xf0
+    schedule_timeout+0x2c8/0x340
+    wait_for_common+0xf4/0x2b0
+    wait_for_completion+0x20/0x40
+    __ib_drain_sq+0x140/0x1d0 [ib_core]
+    ib_drain_sq+0x98/0xb0 [ib_core]
+    rpcrdma_xprt_disconnect+0x68/0x270 [rpcrdma]
+    xprt_rdma_close+0x20/0x60 [rpcrdma]
+    xprt_autoclose+0x64/0x1cc [sunrpc]
+    process_one_work+0x1d8/0x4e0
+    worker_thread+0x154/0x420
+    kthread+0x108/0x150
+    ret_from_fork+0x10/0x18
+
+Fixes: 01584a5edcc4 ("RDMA/hns: Add support of direct wqe")
+Signed-off-by: Chengchang Tang <tangchengchang@huawei.com>
+Signed-off-by: Junxian Huang <huangjunxian6@hisilicon.com>
+Link: https://patch.msgid.link/20241220055249.146943-5-huangjunxian6@hisilicon.com
+Signed-off-by: Leon Romanovsky <leon@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/infiniband/hw/hns/hns_roce_hw_v2.c | 4 ++++
+ 1 file changed, 4 insertions(+)
+
+diff --git a/drivers/infiniband/hw/hns/hns_roce_hw_v2.c b/drivers/infiniband/hw/hns/hns_roce_hw_v2.c
+index c9c9be122471..aded0a7f4283 100644
+--- a/drivers/infiniband/hw/hns/hns_roce_hw_v2.c
++++ b/drivers/infiniband/hw/hns/hns_roce_hw_v2.c
+@@ -673,6 +673,10 @@ static void write_dwqe(struct hns_roce_dev *hr_dev, struct hns_roce_qp *qp,
+ #define HNS_ROCE_SL_SHIFT 2
+       struct hns_roce_v2_rc_send_wqe *rc_sq_wqe = wqe;
++      if (unlikely(qp->state == IB_QPS_ERR)) {
++              flush_cqe(hr_dev, qp);
++              return;
++      }
+       /* All kinds of DirectWQE have the same header field layout */
+       hr_reg_enable(rc_sq_wqe, RC_SEND_WQE_FLAG);
+       hr_reg_write(rc_sq_wqe, RC_SEND_WQE_DB_SL_L, qp->sl);
+-- 
+2.39.5
+
diff --git a/queue-6.6/rdma-hns-fix-warning-storm-caused-by-invalid-input-i.patch b/queue-6.6/rdma-hns-fix-warning-storm-caused-by-invalid-input-i.patch
new file mode 100644 (file)
index 0000000..d344ea4
--- /dev/null
@@ -0,0 +1,47 @@
+From 447418ac54a63a2b5032d2c99feffc15ebb6f065 Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Fri, 20 Dec 2024 13:52:48 +0800
+Subject: RDMA/hns: Fix warning storm caused by invalid input in IO path
+
+From: Chengchang Tang <tangchengchang@huawei.com>
+
+[ Upstream commit fa5c4ba8cdbfd2c2d6422e001311c8213283ebbf ]
+
+WARN_ON() is called in the IO path. And it could lead to a warning
+storm. Use WARN_ON_ONCE() instead of WARN_ON().
+
+Fixes: 12542f1de179 ("RDMA/hns: Refactor process about opcode in post_send()")
+Signed-off-by: Chengchang Tang <tangchengchang@huawei.com>
+Signed-off-by: Junxian Huang <huangjunxian6@hisilicon.com>
+Link: https://patch.msgid.link/20241220055249.146943-4-huangjunxian6@hisilicon.com
+Signed-off-by: Leon Romanovsky <leon@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/infiniband/hw/hns/hns_roce_hw_v2.c | 4 ++--
+ 1 file changed, 2 insertions(+), 2 deletions(-)
+
+diff --git a/drivers/infiniband/hw/hns/hns_roce_hw_v2.c b/drivers/infiniband/hw/hns/hns_roce_hw_v2.c
+index aed9c403f3be..c9c9be122471 100644
+--- a/drivers/infiniband/hw/hns/hns_roce_hw_v2.c
++++ b/drivers/infiniband/hw/hns/hns_roce_hw_v2.c
+@@ -471,7 +471,7 @@ static inline int set_ud_wqe(struct hns_roce_qp *qp,
+       valid_num_sge = calc_wr_sge_num(wr, &msg_len);
+       ret = set_ud_opcode(ud_sq_wqe, wr);
+-      if (WARN_ON(ret))
++      if (WARN_ON_ONCE(ret))
+               return ret;
+       ud_sq_wqe->msg_len = cpu_to_le32(msg_len);
+@@ -575,7 +575,7 @@ static inline int set_rc_wqe(struct hns_roce_qp *qp,
+       rc_sq_wqe->msg_len = cpu_to_le32(msg_len);
+       ret = set_rc_opcode(hr_dev, rc_sq_wqe, wr);
+-      if (WARN_ON(ret))
++      if (WARN_ON_ONCE(ret))
+               return ret;
+       hr_reg_write(rc_sq_wqe, RC_SEND_WQE_SO,
+-- 
+2.39.5
+
diff --git a/queue-6.6/rdma-hns-refactor-mtr-find.patch b/queue-6.6/rdma-hns-refactor-mtr-find.patch
new file mode 100644 (file)
index 0000000..b968151
--- /dev/null
@@ -0,0 +1,406 @@
+From 46028eaaa134849ee067ff6b1c07ad5fbe84121e Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Sat, 13 Jan 2024 16:59:30 +0800
+Subject: RDMA/hns: Refactor mtr find
+
+From: Chengchang Tang <tangchengchang@huawei.com>
+
+[ Upstream commit a4ca341080758d847db155b97887bff6f84016a4 ]
+
+hns_roce_mtr_find() is a collection of multiple functions, and the
+return value is also difficult to understand, which is not conducive
+to modification and maintenance.
+
+Separate the function of obtaining MTR root BA from this function.
+And some adjustments has been made to improve readability.
+
+Signed-off-by: Chengchang Tang <tangchengchang@huawei.com>
+Signed-off-by: Junxian Huang <huangjunxian6@hisilicon.com>
+Link: https://lore.kernel.org/r/20240113085935.2838701-2-huangjunxian6@hisilicon.com
+Signed-off-by: Leon Romanovsky <leon@kernel.org>
+Stable-dep-of: 8673a6c2d9e4 ("RDMA/hns: Fix mapping error of zero-hop WQE buffer")
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/infiniband/hw/hns/hns_roce_cq.c     |  11 +--
+ drivers/infiniband/hw/hns/hns_roce_device.h |   7 +-
+ drivers/infiniband/hw/hns/hns_roce_hw_v2.c  | 102 ++++++++++----------
+ drivers/infiniband/hw/hns/hns_roce_mr.c     |  86 +++++++++++------
+ 4 files changed, 121 insertions(+), 85 deletions(-)
+
+diff --git a/drivers/infiniband/hw/hns/hns_roce_cq.c b/drivers/infiniband/hw/hns/hns_roce_cq.c
+index 9b91731a6207..5e0d78f4e545 100644
+--- a/drivers/infiniband/hw/hns/hns_roce_cq.c
++++ b/drivers/infiniband/hw/hns/hns_roce_cq.c
+@@ -133,14 +133,12 @@ static int alloc_cqc(struct hns_roce_dev *hr_dev, struct hns_roce_cq *hr_cq)
+       struct hns_roce_cq_table *cq_table = &hr_dev->cq_table;
+       struct ib_device *ibdev = &hr_dev->ib_dev;
+       u64 mtts[MTT_MIN_COUNT] = {};
+-      dma_addr_t dma_handle;
+       int ret;
+-      ret = hns_roce_mtr_find(hr_dev, &hr_cq->mtr, 0, mtts, ARRAY_SIZE(mtts),
+-                              &dma_handle);
+-      if (!ret) {
++      ret = hns_roce_mtr_find(hr_dev, &hr_cq->mtr, 0, mtts, ARRAY_SIZE(mtts));
++      if (ret) {
+               ibdev_err(ibdev, "failed to find CQ mtr, ret = %d.\n", ret);
+-              return -EINVAL;
++              return ret;
+       }
+       /* Get CQC memory HEM(Hardware Entry Memory) table */
+@@ -157,7 +155,8 @@ static int alloc_cqc(struct hns_roce_dev *hr_dev, struct hns_roce_cq *hr_cq)
+               goto err_put;
+       }
+-      ret = hns_roce_create_cqc(hr_dev, hr_cq, mtts, dma_handle);
++      ret = hns_roce_create_cqc(hr_dev, hr_cq, mtts,
++                                hns_roce_get_mtr_ba(&hr_cq->mtr));
+       if (ret)
+               goto err_xa;
+diff --git a/drivers/infiniband/hw/hns/hns_roce_device.h b/drivers/infiniband/hw/hns/hns_roce_device.h
+index 21ef00fdb656..a835368548e5 100644
+--- a/drivers/infiniband/hw/hns/hns_roce_device.h
++++ b/drivers/infiniband/hw/hns/hns_roce_device.h
+@@ -1129,8 +1129,13 @@ void hns_roce_cmd_use_polling(struct hns_roce_dev *hr_dev);
+ /* hns roce hw need current block and next block addr from mtt */
+ #define MTT_MIN_COUNT  2
++static inline dma_addr_t hns_roce_get_mtr_ba(struct hns_roce_mtr *mtr)
++{
++      return mtr->hem_cfg.root_ba;
++}
++
+ int hns_roce_mtr_find(struct hns_roce_dev *hr_dev, struct hns_roce_mtr *mtr,
+-                    u32 offset, u64 *mtt_buf, int mtt_max, u64 *base_addr);
++                    u32 offset, u64 *mtt_buf, int mtt_max);
+ int hns_roce_mtr_create(struct hns_roce_dev *hr_dev, struct hns_roce_mtr *mtr,
+                       struct hns_roce_buf_attr *buf_attr,
+                       unsigned int page_shift, struct ib_udata *udata,
+diff --git a/drivers/infiniband/hw/hns/hns_roce_hw_v2.c b/drivers/infiniband/hw/hns/hns_roce_hw_v2.c
+index 2824d390ec31..a396ba85cdce 100644
+--- a/drivers/infiniband/hw/hns/hns_roce_hw_v2.c
++++ b/drivers/infiniband/hw/hns/hns_roce_hw_v2.c
+@@ -3181,21 +3181,22 @@ static int set_mtpt_pbl(struct hns_roce_dev *hr_dev,
+       u64 pages[HNS_ROCE_V2_MAX_INNER_MTPT_NUM] = { 0 };
+       struct ib_device *ibdev = &hr_dev->ib_dev;
+       dma_addr_t pbl_ba;
+-      int i, count;
++      int ret;
++      int i;
+-      count = hns_roce_mtr_find(hr_dev, &mr->pbl_mtr, 0, pages,
+-                                min_t(int, ARRAY_SIZE(pages), mr->npages),
+-                                &pbl_ba);
+-      if (count < 1) {
+-              ibdev_err(ibdev, "failed to find PBL mtr, count = %d.\n",
+-                        count);
+-              return -ENOBUFS;
++      ret = hns_roce_mtr_find(hr_dev, &mr->pbl_mtr, 0, pages,
++                              min_t(int, ARRAY_SIZE(pages), mr->npages));
++      if (ret) {
++              ibdev_err(ibdev, "failed to find PBL mtr, ret = %d.\n", ret);
++              return ret;
+       }
+       /* Aligned to the hardware address access unit */
+-      for (i = 0; i < count; i++)
++      for (i = 0; i < ARRAY_SIZE(pages); i++)
+               pages[i] >>= 6;
++      pbl_ba = hns_roce_get_mtr_ba(&mr->pbl_mtr);
++
+       mpt_entry->pbl_size = cpu_to_le32(mr->npages);
+       mpt_entry->pbl_ba_l = cpu_to_le32(pbl_ba >> 3);
+       hr_reg_write(mpt_entry, MPT_PBL_BA_H, upper_32_bits(pbl_ba >> 3));
+@@ -3294,18 +3295,12 @@ static int hns_roce_v2_rereg_write_mtpt(struct hns_roce_dev *hr_dev,
+ static int hns_roce_v2_frmr_write_mtpt(struct hns_roce_dev *hr_dev,
+                                      void *mb_buf, struct hns_roce_mr *mr)
+ {
+-      struct ib_device *ibdev = &hr_dev->ib_dev;
++      dma_addr_t pbl_ba = hns_roce_get_mtr_ba(&mr->pbl_mtr);
+       struct hns_roce_v2_mpt_entry *mpt_entry;
+-      dma_addr_t pbl_ba = 0;
+       mpt_entry = mb_buf;
+       memset(mpt_entry, 0, sizeof(*mpt_entry));
+-      if (hns_roce_mtr_find(hr_dev, &mr->pbl_mtr, 0, NULL, 0, &pbl_ba) < 0) {
+-              ibdev_err(ibdev, "failed to find frmr mtr.\n");
+-              return -ENOBUFS;
+-      }
+-
+       hr_reg_write(mpt_entry, MPT_ST, V2_MPT_ST_FREE);
+       hr_reg_write(mpt_entry, MPT_PD, mr->pd);
+@@ -4333,17 +4328,20 @@ static int config_qp_rq_buf(struct hns_roce_dev *hr_dev,
+ {
+       u64 mtts[MTT_MIN_COUNT] = { 0 };
+       u64 wqe_sge_ba;
+-      int count;
++      int ret;
+       /* Search qp buf's mtts */
+-      count = hns_roce_mtr_find(hr_dev, &hr_qp->mtr, hr_qp->rq.offset, mtts,
+-                                MTT_MIN_COUNT, &wqe_sge_ba);
+-      if (hr_qp->rq.wqe_cnt && count < 1) {
++      ret = hns_roce_mtr_find(hr_dev, &hr_qp->mtr, hr_qp->rq.offset, mtts,
++                              MTT_MIN_COUNT);
++      if (hr_qp->rq.wqe_cnt && ret) {
+               ibdev_err(&hr_dev->ib_dev,
+-                        "failed to find RQ WQE, QPN = 0x%lx.\n", hr_qp->qpn);
+-              return -EINVAL;
++                        "failed to find QP(0x%lx) RQ WQE buf, ret = %d.\n",
++                        hr_qp->qpn, ret);
++              return ret;
+       }
++      wqe_sge_ba = hns_roce_get_mtr_ba(&hr_qp->mtr);
++
+       context->wqe_sge_ba = cpu_to_le32(wqe_sge_ba >> 3);
+       qpc_mask->wqe_sge_ba = 0;
+@@ -4407,23 +4405,23 @@ static int config_qp_sq_buf(struct hns_roce_dev *hr_dev,
+       struct ib_device *ibdev = &hr_dev->ib_dev;
+       u64 sge_cur_blk = 0;
+       u64 sq_cur_blk = 0;
+-      int count;
++      int ret;
+       /* search qp buf's mtts */
+-      count = hns_roce_mtr_find(hr_dev, &hr_qp->mtr, 0, &sq_cur_blk, 1, NULL);
+-      if (count < 1) {
+-              ibdev_err(ibdev, "failed to find QP(0x%lx) SQ buf.\n",
+-                        hr_qp->qpn);
+-              return -EINVAL;
++      ret = hns_roce_mtr_find(hr_dev, &hr_qp->mtr, hr_qp->sq.offset,
++                              &sq_cur_blk, 1);
++      if (ret) {
++              ibdev_err(ibdev, "failed to find QP(0x%lx) SQ WQE buf, ret = %d.\n",
++                        hr_qp->qpn, ret);
++              return ret;
+       }
+       if (hr_qp->sge.sge_cnt > 0) {
+-              count = hns_roce_mtr_find(hr_dev, &hr_qp->mtr,
+-                                        hr_qp->sge.offset,
+-                                        &sge_cur_blk, 1, NULL);
+-              if (count < 1) {
+-                      ibdev_err(ibdev, "failed to find QP(0x%lx) SGE buf.\n",
+-                                hr_qp->qpn);
+-                      return -EINVAL;
++              ret = hns_roce_mtr_find(hr_dev, &hr_qp->mtr,
++                                      hr_qp->sge.offset, &sge_cur_blk, 1);
++              if (ret) {
++                      ibdev_err(ibdev, "failed to find QP(0x%lx) SGE buf, ret = %d.\n",
++                                hr_qp->qpn, ret);
++                      return ret;
+               }
+       }
+@@ -5550,18 +5548,20 @@ static int hns_roce_v2_write_srqc_index_queue(struct hns_roce_srq *srq,
+       struct ib_device *ibdev = srq->ibsrq.device;
+       struct hns_roce_dev *hr_dev = to_hr_dev(ibdev);
+       u64 mtts_idx[MTT_MIN_COUNT] = {};
+-      dma_addr_t dma_handle_idx = 0;
++      dma_addr_t dma_handle_idx;
+       int ret;
+       /* Get physical address of idx que buf */
+       ret = hns_roce_mtr_find(hr_dev, &idx_que->mtr, 0, mtts_idx,
+-                              ARRAY_SIZE(mtts_idx), &dma_handle_idx);
+-      if (ret < 1) {
++                              ARRAY_SIZE(mtts_idx));
++      if (ret) {
+               ibdev_err(ibdev, "failed to find mtr for SRQ idx, ret = %d.\n",
+                         ret);
+-              return -ENOBUFS;
++              return ret;
+       }
++      dma_handle_idx = hns_roce_get_mtr_ba(&idx_que->mtr);
++
+       hr_reg_write(ctx, SRQC_IDX_HOP_NUM,
+                    to_hr_hem_hopnum(hr_dev->caps.idx_hop_num, srq->wqe_cnt));
+@@ -5593,20 +5593,22 @@ static int hns_roce_v2_write_srqc(struct hns_roce_srq *srq, void *mb_buf)
+       struct hns_roce_dev *hr_dev = to_hr_dev(ibdev);
+       struct hns_roce_srq_context *ctx = mb_buf;
+       u64 mtts_wqe[MTT_MIN_COUNT] = {};
+-      dma_addr_t dma_handle_wqe = 0;
++      dma_addr_t dma_handle_wqe;
+       int ret;
+       memset(ctx, 0, sizeof(*ctx));
+       /* Get the physical address of srq buf */
+       ret = hns_roce_mtr_find(hr_dev, &srq->buf_mtr, 0, mtts_wqe,
+-                              ARRAY_SIZE(mtts_wqe), &dma_handle_wqe);
+-      if (ret < 1) {
++                              ARRAY_SIZE(mtts_wqe));
++      if (ret) {
+               ibdev_err(ibdev, "failed to find mtr for SRQ WQE, ret = %d.\n",
+                         ret);
+-              return -ENOBUFS;
++              return ret;
+       }
++      dma_handle_wqe = hns_roce_get_mtr_ba(&srq->buf_mtr);
++
+       hr_reg_write(ctx, SRQC_SRQ_ST, 1);
+       hr_reg_write_bool(ctx, SRQC_SRQ_TYPE,
+                         srq->ibsrq.srq_type == IB_SRQT_XRC);
+@@ -6327,7 +6329,7 @@ static int config_eqc(struct hns_roce_dev *hr_dev, struct hns_roce_eq *eq,
+       u64 eqe_ba[MTT_MIN_COUNT] = { 0 };
+       struct hns_roce_eq_context *eqc;
+       u64 bt_ba = 0;
+-      int count;
++      int ret;
+       eqc = mb_buf;
+       memset(eqc, 0, sizeof(struct hns_roce_eq_context));
+@@ -6335,13 +6337,15 @@ static int config_eqc(struct hns_roce_dev *hr_dev, struct hns_roce_eq *eq,
+       init_eq_config(hr_dev, eq);
+       /* if not multi-hop, eqe buffer only use one trunk */
+-      count = hns_roce_mtr_find(hr_dev, &eq->mtr, 0, eqe_ba, MTT_MIN_COUNT,
+-                                &bt_ba);
+-      if (count < 1) {
+-              dev_err(hr_dev->dev, "failed to find EQE mtr\n");
+-              return -ENOBUFS;
++      ret = hns_roce_mtr_find(hr_dev, &eq->mtr, 0, eqe_ba,
++                              ARRAY_SIZE(eqe_ba));
++      if (ret) {
++              dev_err(hr_dev->dev, "failed to find EQE mtr, ret = %d\n", ret);
++              return ret;
+       }
++      bt_ba = hns_roce_get_mtr_ba(&eq->mtr);
++
+       hr_reg_write(eqc, EQC_EQ_ST, HNS_ROCE_V2_EQ_STATE_VALID);
+       hr_reg_write(eqc, EQC_EQE_HOP_NUM, eq->hop_num);
+       hr_reg_write(eqc, EQC_OVER_IGNORE, eq->over_ignore);
+diff --git a/drivers/infiniband/hw/hns/hns_roce_mr.c b/drivers/infiniband/hw/hns/hns_roce_mr.c
+index 7f29a55d378f..7b59b95f87c2 100644
+--- a/drivers/infiniband/hw/hns/hns_roce_mr.c
++++ b/drivers/infiniband/hw/hns/hns_roce_mr.c
+@@ -802,47 +802,53 @@ int hns_roce_mtr_map(struct hns_roce_dev *hr_dev, struct hns_roce_mtr *mtr,
+       return ret;
+ }
+-int hns_roce_mtr_find(struct hns_roce_dev *hr_dev, struct hns_roce_mtr *mtr,
+-                    u32 offset, u64 *mtt_buf, int mtt_max, u64 *base_addr)
++static int hns_roce_get_direct_addr_mtt(struct hns_roce_hem_cfg *cfg,
++                                      u32 start_index, u64 *mtt_buf,
++                                      int mtt_cnt)
+ {
+-      struct hns_roce_hem_cfg *cfg = &mtr->hem_cfg;
+-      int mtt_count, left;
+-      u32 start_index;
++      int mtt_count;
+       int total = 0;
+-      __le64 *mtts;
+       u32 npage;
+       u64 addr;
+-      if (!mtt_buf || mtt_max < 1)
+-              goto done;
+-
+-      /* no mtt memory in direct mode, so just return the buffer address */
+-      if (cfg->is_direct) {
+-              start_index = offset >> HNS_HW_PAGE_SHIFT;
+-              for (mtt_count = 0; mtt_count < cfg->region_count &&
+-                   total < mtt_max; mtt_count++) {
+-                      npage = cfg->region[mtt_count].offset;
+-                      if (npage < start_index)
+-                              continue;
++      if (mtt_cnt > cfg->region_count)
++              return -EINVAL;
+-                      addr = cfg->root_ba + (npage << HNS_HW_PAGE_SHIFT);
+-                      mtt_buf[total] = addr;
++      for (mtt_count = 0; mtt_count < cfg->region_count && total < mtt_cnt;
++           mtt_count++) {
++              npage = cfg->region[mtt_count].offset;
++              if (npage < start_index)
++                      continue;
+-                      total++;
+-              }
++              addr = cfg->root_ba + (npage << HNS_HW_PAGE_SHIFT);
++              mtt_buf[total] = addr;
+-              goto done;
++              total++;
+       }
+-      start_index = offset >> cfg->buf_pg_shift;
+-      left = mtt_max;
++      if (!total)
++              return -ENOENT;
++
++      return 0;
++}
++
++static int hns_roce_get_mhop_mtt(struct hns_roce_dev *hr_dev,
++                               struct hns_roce_mtr *mtr, u32 start_index,
++                               u64 *mtt_buf, int mtt_cnt)
++{
++      int left = mtt_cnt;
++      int total = 0;
++      int mtt_count;
++      __le64 *mtts;
++      u32 npage;
++
+       while (left > 0) {
+               mtt_count = 0;
+               mtts = hns_roce_hem_list_find_mtt(hr_dev, &mtr->hem_list,
+                                                 start_index + total,
+                                                 &mtt_count);
+               if (!mtts || !mtt_count)
+-                      goto done;
++                      break;
+               npage = min(mtt_count, left);
+               left -= npage;
+@@ -850,11 +856,33 @@ int hns_roce_mtr_find(struct hns_roce_dev *hr_dev, struct hns_roce_mtr *mtr,
+                       mtt_buf[total++] = le64_to_cpu(mtts[mtt_count]);
+       }
+-done:
+-      if (base_addr)
+-              *base_addr = cfg->root_ba;
++      if (!total)
++              return -ENOENT;
++
++      return 0;
++}
++
++int hns_roce_mtr_find(struct hns_roce_dev *hr_dev, struct hns_roce_mtr *mtr,
++                    u32 offset, u64 *mtt_buf, int mtt_max)
++{
++      struct hns_roce_hem_cfg *cfg = &mtr->hem_cfg;
++      u32 start_index;
++      int ret;
++
++      if (!mtt_buf || mtt_max < 1)
++              return -EINVAL;
+-      return total;
++      /* no mtt memory in direct mode, so just return the buffer address */
++      if (cfg->is_direct) {
++              start_index = offset >> HNS_HW_PAGE_SHIFT;
++              ret = hns_roce_get_direct_addr_mtt(cfg, start_index,
++                                                 mtt_buf, mtt_max);
++      } else {
++              start_index = offset >> cfg->buf_pg_shift;
++              ret = hns_roce_get_mhop_mtt(hr_dev, mtr, start_index,
++                                          mtt_buf, mtt_max);
++      }
++      return ret;
+ }
+ static int mtr_init_buf_cfg(struct hns_roce_dev *hr_dev,
+-- 
+2.39.5
+
diff --git a/queue-6.6/rdma-hns-remove-unused-parameters-and-variables.patch b/queue-6.6/rdma-hns-remove-unused-parameters-and-variables.patch
new file mode 100644 (file)
index 0000000..2e9a945
--- /dev/null
@@ -0,0 +1,268 @@
+From e16d6bebbc105a3eeaf61a21fbe289c91c7d37dc Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Fri, 12 Apr 2024 17:16:08 +0800
+Subject: RDMA/hns: Remove unused parameters and variables
+
+From: Chengchang Tang <tangchengchang@huawei.com>
+
+[ Upstream commit f4caa864af84f801a5821ea2ba6c1cc46f8252c1 ]
+
+Remove unused parameters and variables.
+
+Signed-off-by: Chengchang Tang <tangchengchang@huawei.com>
+Signed-off-by: Junxian Huang <huangjunxian6@hisilicon.com>
+Link: https://lore.kernel.org/r/20240412091616.370789-3-huangjunxian6@hisilicon.com
+Signed-off-by: Leon Romanovsky <leon@kernel.org>
+Stable-dep-of: 8673a6c2d9e4 ("RDMA/hns: Fix mapping error of zero-hop WQE buffer")
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/infiniband/hw/hns/hns_roce_alloc.c  |  3 +--
+ drivers/infiniband/hw/hns/hns_roce_device.h |  5 ++---
+ drivers/infiniband/hw/hns/hns_roce_hem.c    | 13 +++++--------
+ drivers/infiniband/hw/hns/hns_roce_hw_v2.c  | 20 +++++++-------------
+ drivers/infiniband/hw/hns/hns_roce_mr.c     |  4 ++--
+ drivers/infiniband/hw/hns/hns_roce_qp.c     |  4 +---
+ drivers/infiniband/hw/hns/hns_roce_srq.c    |  4 ++--
+ 7 files changed, 20 insertions(+), 33 deletions(-)
+
+diff --git a/drivers/infiniband/hw/hns/hns_roce_alloc.c b/drivers/infiniband/hw/hns/hns_roce_alloc.c
+index 11a78ceae568..950c133d4220 100644
+--- a/drivers/infiniband/hw/hns/hns_roce_alloc.c
++++ b/drivers/infiniband/hw/hns/hns_roce_alloc.c
+@@ -153,8 +153,7 @@ int hns_roce_get_kmem_bufs(struct hns_roce_dev *hr_dev, dma_addr_t *bufs,
+       return total;
+ }
+-int hns_roce_get_umem_bufs(struct hns_roce_dev *hr_dev, dma_addr_t *bufs,
+-                         int buf_cnt, struct ib_umem *umem,
++int hns_roce_get_umem_bufs(dma_addr_t *bufs, int buf_cnt, struct ib_umem *umem,
+                          unsigned int page_shift)
+ {
+       struct ib_block_iter biter;
+diff --git a/drivers/infiniband/hw/hns/hns_roce_device.h b/drivers/infiniband/hw/hns/hns_roce_device.h
+index a835368548e5..03b6546f63cd 100644
+--- a/drivers/infiniband/hw/hns/hns_roce_device.h
++++ b/drivers/infiniband/hw/hns/hns_roce_device.h
+@@ -892,8 +892,7 @@ struct hns_roce_hw {
+       int (*rereg_write_mtpt)(struct hns_roce_dev *hr_dev,
+                               struct hns_roce_mr *mr, int flags,
+                               void *mb_buf);
+-      int (*frmr_write_mtpt)(struct hns_roce_dev *hr_dev, void *mb_buf,
+-                             struct hns_roce_mr *mr);
++      int (*frmr_write_mtpt)(void *mb_buf, struct hns_roce_mr *mr);
+       int (*mw_write_mtpt)(void *mb_buf, struct hns_roce_mw *mw);
+       void (*write_cqc)(struct hns_roce_dev *hr_dev,
+                         struct hns_roce_cq *hr_cq, void *mb_buf, u64 *mtts,
+@@ -1193,7 +1192,7 @@ struct hns_roce_buf *hns_roce_buf_alloc(struct hns_roce_dev *hr_dev, u32 size,
+ int hns_roce_get_kmem_bufs(struct hns_roce_dev *hr_dev, dma_addr_t *bufs,
+                          int buf_cnt, struct hns_roce_buf *buf,
+                          unsigned int page_shift);
+-int hns_roce_get_umem_bufs(struct hns_roce_dev *hr_dev, dma_addr_t *bufs,
++int hns_roce_get_umem_bufs(dma_addr_t *bufs,
+                          int buf_cnt, struct ib_umem *umem,
+                          unsigned int page_shift);
+diff --git a/drivers/infiniband/hw/hns/hns_roce_hem.c b/drivers/infiniband/hw/hns/hns_roce_hem.c
+index 0ab514c49d5e..d3fabf64e390 100644
+--- a/drivers/infiniband/hw/hns/hns_roce_hem.c
++++ b/drivers/infiniband/hw/hns/hns_roce_hem.c
+@@ -1043,15 +1043,13 @@ static void hem_list_free_all(struct hns_roce_dev *hr_dev,
+       }
+ }
+-static void hem_list_link_bt(struct hns_roce_dev *hr_dev, void *base_addr,
+-                           u64 table_addr)
++static void hem_list_link_bt(void *base_addr, u64 table_addr)
+ {
+       *(u64 *)(base_addr) = table_addr;
+ }
+ /* assign L0 table address to hem from root bt */
+-static void hem_list_assign_bt(struct hns_roce_dev *hr_dev,
+-                             struct hns_roce_hem_item *hem, void *cpu_addr,
++static void hem_list_assign_bt(struct hns_roce_hem_item *hem, void *cpu_addr,
+                              u64 phy_addr)
+ {
+       hem->addr = cpu_addr;
+@@ -1222,8 +1220,7 @@ static int hem_list_alloc_mid_bt(struct hns_roce_dev *hr_dev,
+               if (level > 1) {
+                       pre = hem_ptrs[level - 1];
+                       step = (cur->start - pre->start) / step * BA_BYTE_LEN;
+-                      hem_list_link_bt(hr_dev, pre->addr + step,
+-                                       cur->dma_addr);
++                      hem_list_link_bt(pre->addr + step, cur->dma_addr);
+               }
+       }
+@@ -1281,7 +1278,7 @@ static int alloc_fake_root_bt(struct hns_roce_dev *hr_dev, void *cpu_base,
+       if (!hem)
+               return -ENOMEM;
+-      hem_list_assign_bt(hr_dev, hem, cpu_base, phy_base);
++      hem_list_assign_bt(hem, cpu_base, phy_base);
+       list_add(&hem->list, branch_head);
+       list_add(&hem->sibling, leaf_head);
+@@ -1304,7 +1301,7 @@ static int setup_middle_bt(struct hns_roce_dev *hr_dev, void *cpu_base,
+       /* if exist mid bt, link L1 to L0 */
+       list_for_each_entry_safe(hem, temp_hem, branch_head, list) {
+               offset = (hem->start - r->offset) / step * BA_BYTE_LEN;
+-              hem_list_link_bt(hr_dev, cpu_base + offset, hem->dma_addr);
++              hem_list_link_bt(cpu_base + offset, hem->dma_addr);
+               total++;
+       }
+diff --git a/drivers/infiniband/hw/hns/hns_roce_hw_v2.c b/drivers/infiniband/hw/hns/hns_roce_hw_v2.c
+index a396ba85cdce..aed9c403f3be 100644
+--- a/drivers/infiniband/hw/hns/hns_roce_hw_v2.c
++++ b/drivers/infiniband/hw/hns/hns_roce_hw_v2.c
+@@ -3292,8 +3292,7 @@ static int hns_roce_v2_rereg_write_mtpt(struct hns_roce_dev *hr_dev,
+       return ret;
+ }
+-static int hns_roce_v2_frmr_write_mtpt(struct hns_roce_dev *hr_dev,
+-                                     void *mb_buf, struct hns_roce_mr *mr)
++static int hns_roce_v2_frmr_write_mtpt(void *mb_buf, struct hns_roce_mr *mr)
+ {
+       dma_addr_t pbl_ba = hns_roce_get_mtr_ba(&mr->pbl_mtr);
+       struct hns_roce_v2_mpt_entry *mpt_entry;
+@@ -4208,8 +4207,7 @@ static void set_access_flags(struct hns_roce_qp *hr_qp,
+ }
+ static void set_qpc_wqe_cnt(struct hns_roce_qp *hr_qp,
+-                          struct hns_roce_v2_qp_context *context,
+-                          struct hns_roce_v2_qp_context *qpc_mask)
++                          struct hns_roce_v2_qp_context *context)
+ {
+       hr_reg_write(context, QPC_SGE_SHIFT,
+                    to_hr_hem_entries_shift(hr_qp->sge.sge_cnt,
+@@ -4231,7 +4229,6 @@ static inline int get_pdn(struct ib_pd *ib_pd)
+ }
+ static void modify_qp_reset_to_init(struct ib_qp *ibqp,
+-                                  const struct ib_qp_attr *attr,
+                                   struct hns_roce_v2_qp_context *context,
+                                   struct hns_roce_v2_qp_context *qpc_mask)
+ {
+@@ -4250,7 +4247,7 @@ static void modify_qp_reset_to_init(struct ib_qp *ibqp,
+       hr_reg_write(context, QPC_RQWS, ilog2(hr_qp->rq.max_gs));
+-      set_qpc_wqe_cnt(hr_qp, context, qpc_mask);
++      set_qpc_wqe_cnt(hr_qp, context);
+       /* No VLAN need to set 0xFFF */
+       hr_reg_write(context, QPC_VLAN_ID, 0xfff);
+@@ -4291,7 +4288,6 @@ static void modify_qp_reset_to_init(struct ib_qp *ibqp,
+ }
+ static void modify_qp_init_to_init(struct ib_qp *ibqp,
+-                                 const struct ib_qp_attr *attr,
+                                  struct hns_roce_v2_qp_context *context,
+                                  struct hns_roce_v2_qp_context *qpc_mask)
+ {
+@@ -4612,8 +4608,7 @@ static int modify_qp_init_to_rtr(struct ib_qp *ibqp,
+       return 0;
+ }
+-static int modify_qp_rtr_to_rts(struct ib_qp *ibqp,
+-                              const struct ib_qp_attr *attr, int attr_mask,
++static int modify_qp_rtr_to_rts(struct ib_qp *ibqp, int attr_mask,
+                               struct hns_roce_v2_qp_context *context,
+                               struct hns_roce_v2_qp_context *qpc_mask)
+ {
+@@ -4982,15 +4977,14 @@ static int hns_roce_v2_set_abs_fields(struct ib_qp *ibqp,
+       if (cur_state == IB_QPS_RESET && new_state == IB_QPS_INIT) {
+               memset(qpc_mask, 0, hr_dev->caps.qpc_sz);
+-              modify_qp_reset_to_init(ibqp, attr, context, qpc_mask);
++              modify_qp_reset_to_init(ibqp, context, qpc_mask);
+       } else if (cur_state == IB_QPS_INIT && new_state == IB_QPS_INIT) {
+-              modify_qp_init_to_init(ibqp, attr, context, qpc_mask);
++              modify_qp_init_to_init(ibqp, context, qpc_mask);
+       } else if (cur_state == IB_QPS_INIT && new_state == IB_QPS_RTR) {
+               ret = modify_qp_init_to_rtr(ibqp, attr, attr_mask, context,
+                                           qpc_mask, udata);
+       } else if (cur_state == IB_QPS_RTR && new_state == IB_QPS_RTS) {
+-              ret = modify_qp_rtr_to_rts(ibqp, attr, attr_mask, context,
+-                                         qpc_mask);
++              ret = modify_qp_rtr_to_rts(ibqp, attr_mask, context, qpc_mask);
+       }
+       return ret;
+diff --git a/drivers/infiniband/hw/hns/hns_roce_mr.c b/drivers/infiniband/hw/hns/hns_roce_mr.c
+index 7b59b95f87c2..acdcd5c9f42f 100644
+--- a/drivers/infiniband/hw/hns/hns_roce_mr.c
++++ b/drivers/infiniband/hw/hns/hns_roce_mr.c
+@@ -154,7 +154,7 @@ static int hns_roce_mr_enable(struct hns_roce_dev *hr_dev,
+       if (mr->type != MR_TYPE_FRMR)
+               ret = hr_dev->hw->write_mtpt(hr_dev, mailbox->buf, mr);
+       else
+-              ret = hr_dev->hw->frmr_write_mtpt(hr_dev, mailbox->buf, mr);
++              ret = hr_dev->hw->frmr_write_mtpt(mailbox->buf, mr);
+       if (ret) {
+               dev_err(dev, "failed to write mtpt, ret = %d.\n", ret);
+               goto err_page;
+@@ -714,7 +714,7 @@ static int mtr_map_bufs(struct hns_roce_dev *hr_dev, struct hns_roce_mtr *mtr,
+               return -ENOMEM;
+       if (mtr->umem)
+-              npage = hns_roce_get_umem_bufs(hr_dev, pages, page_count,
++              npage = hns_roce_get_umem_bufs(pages, page_count,
+                                              mtr->umem, page_shift);
+       else
+               npage = hns_roce_get_kmem_bufs(hr_dev, pages, page_count,
+diff --git a/drivers/infiniband/hw/hns/hns_roce_qp.c b/drivers/infiniband/hw/hns/hns_roce_qp.c
+index 88a4777d29f8..97d79c8d5cd0 100644
+--- a/drivers/infiniband/hw/hns/hns_roce_qp.c
++++ b/drivers/infiniband/hw/hns/hns_roce_qp.c
+@@ -1075,7 +1075,6 @@ static int set_qp_param(struct hns_roce_dev *hr_dev, struct hns_roce_qp *hr_qp,
+ }
+ static int hns_roce_create_qp_common(struct hns_roce_dev *hr_dev,
+-                                   struct ib_pd *ib_pd,
+                                    struct ib_qp_init_attr *init_attr,
+                                    struct ib_udata *udata,
+                                    struct hns_roce_qp *hr_qp)
+@@ -1229,7 +1228,6 @@ int hns_roce_create_qp(struct ib_qp *qp, struct ib_qp_init_attr *init_attr,
+       struct ib_device *ibdev = qp->device;
+       struct hns_roce_dev *hr_dev = to_hr_dev(ibdev);
+       struct hns_roce_qp *hr_qp = to_hr_qp(qp);
+-      struct ib_pd *pd = qp->pd;
+       int ret;
+       ret = check_qp_type(hr_dev, init_attr->qp_type, !!udata);
+@@ -1244,7 +1242,7 @@ int hns_roce_create_qp(struct ib_qp *qp, struct ib_qp_init_attr *init_attr,
+               hr_qp->phy_port = hr_dev->iboe.phy_port[hr_qp->port];
+       }
+-      ret = hns_roce_create_qp_common(hr_dev, pd, init_attr, udata, hr_qp);
++      ret = hns_roce_create_qp_common(hr_dev, init_attr, udata, hr_qp);
+       if (ret)
+               ibdev_err(ibdev, "create QP type 0x%x failed(%d)\n",
+                         init_attr->qp_type, ret);
+diff --git a/drivers/infiniband/hw/hns/hns_roce_srq.c b/drivers/infiniband/hw/hns/hns_roce_srq.c
+index 652508b660a0..80fcb1b0e8fd 100644
+--- a/drivers/infiniband/hw/hns/hns_roce_srq.c
++++ b/drivers/infiniband/hw/hns/hns_roce_srq.c
+@@ -249,7 +249,7 @@ static void free_srq_wqe_buf(struct hns_roce_dev *hr_dev,
+       hns_roce_mtr_destroy(hr_dev, &srq->buf_mtr);
+ }
+-static int alloc_srq_wrid(struct hns_roce_dev *hr_dev, struct hns_roce_srq *srq)
++static int alloc_srq_wrid(struct hns_roce_srq *srq)
+ {
+       srq->wrid = kvmalloc_array(srq->wqe_cnt, sizeof(u64), GFP_KERNEL);
+       if (!srq->wrid)
+@@ -365,7 +365,7 @@ static int alloc_srq_buf(struct hns_roce_dev *hr_dev, struct hns_roce_srq *srq,
+               goto err_idx;
+       if (!udata) {
+-              ret = alloc_srq_wrid(hr_dev, srq);
++              ret = alloc_srq_wrid(srq);
+               if (ret)
+                       goto err_wqe_buf;
+       }
+-- 
+2.39.5
+
diff --git a/queue-6.6/rdma-mlx5-enforce-same-type-port-association-for-mul.patch b/queue-6.6/rdma-mlx5-enforce-same-type-port-association-for-mul.patch
new file mode 100644 (file)
index 0000000..1806db3
--- /dev/null
@@ -0,0 +1,69 @@
+From 222dfa1b2907685ae58f74002f54f5a88ba9cc73 Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Tue, 3 Dec 2024 15:45:37 +0200
+Subject: RDMA/mlx5: Enforce same type port association for multiport RoCE
+
+From: Patrisious Haddad <phaddad@nvidia.com>
+
+[ Upstream commit e05feab22fd7dabcd6d272c4e2401ec1acdfdb9b ]
+
+Different core device types such as PFs and VFs shouldn't be affiliated
+together since they have different capabilities, fix that by enforcing
+type check before doing the affiliation.
+
+Fixes: 32f69e4be269 ("{net, IB}/mlx5: Manage port association for multiport RoCE")
+Reviewed-by: Mark Bloch <mbloch@nvidia.com>
+Signed-off-by: Patrisious Haddad <phaddad@nvidia.com>
+Link: https://patch.msgid.link/88699500f690dff1c1852c1ddb71f8a1cc8b956e.1733233480.git.leonro@nvidia.com
+Reviewed-by: Mateusz Polchlopek <mateusz.polchlopek@intel.com>
+Signed-off-by: Leon Romanovsky <leon@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/infiniband/hw/mlx5/main.c | 6 ++++--
+ include/linux/mlx5/driver.h       | 6 ++++++
+ 2 files changed, 10 insertions(+), 2 deletions(-)
+
+diff --git a/drivers/infiniband/hw/mlx5/main.c b/drivers/infiniband/hw/mlx5/main.c
+index c510484e024b..ada7dbf8eb1c 100644
+--- a/drivers/infiniband/hw/mlx5/main.c
++++ b/drivers/infiniband/hw/mlx5/main.c
+@@ -3372,7 +3372,8 @@ static int mlx5_ib_init_multiport_master(struct mlx5_ib_dev *dev)
+               list_for_each_entry(mpi, &mlx5_ib_unaffiliated_port_list,
+                                   list) {
+                       if (dev->sys_image_guid == mpi->sys_image_guid &&
+-                          (mlx5_core_native_port_num(mpi->mdev) - 1) == i) {
++                          (mlx5_core_native_port_num(mpi->mdev) - 1) == i &&
++                          mlx5_core_same_coredev_type(dev->mdev, mpi->mdev)) {
+                               bound = mlx5_ib_bind_slave_port(dev, mpi);
+                       }
+@@ -4406,7 +4407,8 @@ static int mlx5r_mp_probe(struct auxiliary_device *adev,
+       mutex_lock(&mlx5_ib_multiport_mutex);
+       list_for_each_entry(dev, &mlx5_ib_dev_list, ib_dev_list) {
+-              if (dev->sys_image_guid == mpi->sys_image_guid)
++              if (dev->sys_image_guid == mpi->sys_image_guid &&
++                  mlx5_core_same_coredev_type(dev->mdev, mpi->mdev))
+                       bound = mlx5_ib_bind_slave_port(dev, mpi);
+               if (bound) {
+diff --git a/include/linux/mlx5/driver.h b/include/linux/mlx5/driver.h
+index ffb98bc43b2d..38a8ff9c685c 100644
+--- a/include/linux/mlx5/driver.h
++++ b/include/linux/mlx5/driver.h
+@@ -1225,6 +1225,12 @@ static inline bool mlx5_core_is_vf(const struct mlx5_core_dev *dev)
+       return dev->coredev_type == MLX5_COREDEV_VF;
+ }
++static inline bool mlx5_core_same_coredev_type(const struct mlx5_core_dev *dev1,
++                                             const struct mlx5_core_dev *dev2)
++{
++      return dev1->coredev_type == dev2->coredev_type;
++}
++
+ static inline bool mlx5_core_is_ecpf(const struct mlx5_core_dev *dev)
+ {
+       return dev->caps.embedded_cpu;
+-- 
+2.39.5
+
diff --git a/queue-6.6/rdma-rtrs-ensure-ib_sge-list-is-accessible.patch b/queue-6.6/rdma-rtrs-ensure-ib_sge-list-is-accessible.patch
new file mode 100644 (file)
index 0000000..3ed2e6e
--- /dev/null
@@ -0,0 +1,71 @@
+From 84067fe288e1c296c21fb21a10b968c7957e8bb7 Mon Sep 17 00:00:00 2001
+From: Sasha Levin <sashal@kernel.org>
+Date: Tue, 31 Dec 2024 09:34:16 +0800
+Subject: RDMA/rtrs: Ensure 'ib_sge list' is accessible
+
+From: Li Zhijian <lizhijian@fujitsu.com>
+
+[ Upstream commit fb514b31395946022f13a08e06a435f53cf9e8b3 ]
+
+Move the declaration of the 'ib_sge list' variable outside the
+'always_invalidate' block to ensure it remains accessible for use
+throughout the function.
+
+Previously, 'ib_sge list' was declared within the 'always_invalidate'
+block, limiting its accessibility, then caused a
+'BUG: kernel NULL pointer dereference'[1].
+ ? __die_body.cold+0x19/0x27
+ ? page_fault_oops+0x15a/0x2d0
+ ? search_module_extables+0x19/0x60
+ ? search_bpf_extables+0x5f/0x80
+ ? exc_page_fault+0x7e/0x180
+ ? asm_exc_page_fault+0x26/0x30
+ ? memcpy_orig+0xd5/0x140
+ rxe_mr_copy+0x1c3/0x200 [rdma_rxe]
+ ? rxe_pool_get_index+0x4b/0x80 [rdma_rxe]
+ copy_data+0xa5/0x230 [rdma_rxe]
+ rxe_requester+0xd9b/0xf70 [rdma_rxe]
+ ? finish_task_switch.isra.0+0x99/0x2e0
+ rxe_sender+0x13/0x40 [rdma_rxe]
+ do_task+0x68/0x1e0 [rdma_rxe]
+ process_one_work+0x177/0x330
+ worker_thread+0x252/0x390
+ ? __pfx_worker_thread+0x10/0x10
+
+This change ensures the variable is available for subsequent operations
+that require it.
+
+[1] https://lore.kernel.org/linux-rdma/6a1f3e8f-deb0-49f9-bc69-a9b03ecfcda7@fujitsu.com/
+
+Fixes: 9cb837480424 ("RDMA/rtrs: server: main functionality")
+Signed-off-by: Li Zhijian <lizhijian@fujitsu.com>
+Link: https://patch.msgid.link/20241231013416.1290920-1-lizhijian@fujitsu.com
+Signed-off-by: Leon Romanovsky <leon@kernel.org>
+Signed-off-by: Sasha Levin <sashal@kernel.org>
+---
+ drivers/infiniband/ulp/rtrs/rtrs-srv.c | 2 +-
+ 1 file changed, 1 insertion(+), 1 deletion(-)
+
+diff --git a/drivers/infiniband/ulp/rtrs/rtrs-srv.c b/drivers/infiniband/ulp/rtrs/rtrs-srv.c
+index 758a3d9c2844..84d1654148d7 100644
+--- a/drivers/infiniband/ulp/rtrs/rtrs-srv.c
++++ b/drivers/infiniband/ulp/rtrs/rtrs-srv.c
+@@ -346,6 +346,7 @@ static int send_io_resp_imm(struct rtrs_srv_con *con, struct rtrs_srv_op *id,
+       struct rtrs_srv_mr *srv_mr;
+       bool need_inval = false;
+       enum ib_send_flags flags;
++      struct ib_sge list;
+       u32 imm;
+       int err;
+@@ -398,7 +399,6 @@ static int send_io_resp_imm(struct rtrs_srv_con *con, struct rtrs_srv_op *id,
+       imm = rtrs_to_io_rsp_imm(id->msg_id, errno, need_inval);
+       imm_wr.wr.next = NULL;
+       if (always_invalidate) {
+-              struct ib_sge list;
+               struct rtrs_msg_rkey_rsp *msg;
+               srv_mr = &srv_path->mrs[id->msg_id];
+-- 
+2.39.5
+
index 733ce81631edc83227921888985c05485bb84919..13c7c16abd84226cd62dbe44aa63b74a0f0d5b56 100644 (file)
@@ -126,3 +126,54 @@ seq_buf-introduce-declare_seq_buf-and-seq_buf_str.patch
 tracing-handle-old-buffer-mappings-for-event-strings.patch
 tracing-fix-trace_check_vprintf-when-tp_printk-is-us.patch
 tracing-check-s-dereference-via-the-field-and-not-th.patch
+rdma-bnxt_re-allow-msn-table-capability-check.patch
+rdma-bnxt_re-remove-always-true-dattr-validity-check.patch
+rdma-mlx5-enforce-same-type-port-association-for-mul.patch
+rdma-bnxt_re-avoid-initializing-the-software-queue-f.patch
+rdma-bnxt_re-avoid-sending-the-modify-qp-workaround-.patch
+nvme-pci-512-byte-aligned-dma-pool-segment-quirk.patch
+rdma-bnxt_re-fix-the-check-for-9060-condition.patch
+rdma-bnxt_re-add-check-for-path-mtu-in-modify_qp.patch
+rdma-bnxt_re-fix-reporting-hw_ver-in-query_device.patch
+rdma-bnxt_re-fix-max_qp_wrs-reported.patch
+rdma-bnxt_re-add-support-for-variable-wqe-in-genp7-a.patch
+rdma-bnxt_re-disable-use-of-reserved-wqes.patch
+rdma-bnxt_re-add-send-queue-size-check-for-variable-.patch
+rdma-bnxt_re-fix-msn-table-size-for-variable-wqe-mod.patch
+rdma-bnxt_re-fix-the-locking-while-accessing-the-qp-.patch
+drm-bridge-adv7511_audio-update-audio-infoframe-prop.patch
+net-dsa-microchip-fix-ksz9477-set_ageing_time-functi.patch
+net-dsa-microchip-fix-lan937x-set_ageing_time-functi.patch
+rdma-hns-refactor-mtr-find.patch
+rdma-hns-remove-unused-parameters-and-variables.patch
+rdma-hns-fix-mapping-error-of-zero-hop-wqe-buffer.patch
+rdma-hns-fix-warning-storm-caused-by-invalid-input-i.patch
+rdma-hns-fix-missing-flush-cqe-for-dwqe.patch
+net-stmmac-don-t-create-a-mdio-bus-if-unnecessary.patch
+net-stmmac-restructure-the-error-path-of-stmmac_prob.patch
+net-fix-memory-leak-in-tcp_conn_request.patch
+ip_tunnel-annotate-data-races-around-t-parms.link.patch
+ipv4-ip_tunnel-unmask-upper-dscp-bits-in-ip_tunnel_b.patch
+ipv4-ip_tunnel-unmask-upper-dscp-bits-in-ip_md_tunne.patch
+ipv4-ip_tunnel-unmask-upper-dscp-bits-in-ip_tunnel_x.patch
+net-fix-netns-for-ip_tunnel_init_flow.patch
+netrom-check-buffer-length-before-accessing-it.patch
+net-mlx5-dr-select-msix-vector-0-for-completion-queu.patch
+net-mlx5e-macsec-maintain-tx-sa-from-encoding_sa.patch
+net-mlx5e-skip-restore-tc-rules-for-vport-rep-withou.patch
+drm-i915-dg1-fix-power-gate-sequence.patch
+netfilter-nft_set_hash-unaligned-atomic-read-on-stru.patch
+net-llc-reset-skb-transport_header.patch
+alsa-usb-audio-us16x08-initialize-array-before-use.patch
+eth-bcmsysport-fix-call-balance-of-priv-clk-handling.patch
+net-mv643xx_eth-fix-an-of-node-reference-leak.patch
+net-wwan-t7xx-fix-fsm-command-timeout-issue.patch
+rdma-rtrs-ensure-ib_sge-list-is-accessible.patch
+net-sfc-correct-key_len-for-efx_tc_ct_zone_ht_params.patch
+net-reenable-netif_f_ipv6_csum-offload-for-big-tcp-p.patch
+net-restrict-so_reuseport-to-inet-sockets.patch
+net-wwan-iosm-properly-check-for-valid-exec-stage-in.patch
+af_packet-fix-vlan_get_tci-vs-msg_peek.patch
+af_packet-fix-vlan_get_protocol_dgram-vs-msg_peek.patch
+ila-serialize-calls-to-nf_register_net_hooks.patch
+net-ti-icssg-prueth-fix-clearing-of-iep_cmp_cfg-regi.patch