diff mbox series

[mptcp-net] mptcp: fix inconsistent state on fastopen race

Message ID c753d1520719374102ea5064d3f8eac46135a165.1700050135.git.pabeni@redhat.com (mailing list archive)
State Superseded, archived
Delegated to: Mat Martineau
Headers show
Series [mptcp-net] mptcp: fix inconsistent state on fastopen race | expand

Checks

Context Check Description
matttbe/build success Build and static analysis OK
matttbe/checkpatch warning total: 0 errors, 1 warnings, 0 checks, 90 lines checked
matttbe/KVM_Validation__normal__except_selftest_mptcp_join_ success Success! ✅
matttbe/KVM_Validation__debug__only_selftest_mptcp_join_ success Success! ✅
matttbe/KVM_Validation__normal__only_selftest_mptcp_join_ success Success! ✅
matttbe/KVM_Validation__debug__except_selftest_mptcp_join_ warning Unstable: 1 failed test(s): selftest_diag

Commit Message

Paolo Abeni Nov. 15, 2023, 12:09 p.m. UTC
The netlink PM can race with fastopen self-connect attempts, shutting
down the first subflow via:

MPTCP_PM_CMD_DEL_ADDR -> mptcp_nl_remove_id_zero_address ->
  mptcp_pm_nl_rm_subflow_received -> mptcp_close_ssk

and transitioning such subflow to FIN_WAIT1 status before the syn-ack
packet is processed. The MPTCP code does not react to such state change,
leaving the connection in not-fallback status and the subflow handshake
uncompleted, triggering the following splat:

WARNING: CPU: 0 PID: 10630 at net/mptcp/subflow.c:1405 subflow_data_ready+0x39f/0x690 net/mptcp/subflow.c:1405
Modules linked in:
CPU: 0 PID: 10630 Comm: kworker/u4:11 Not tainted 6.6.0-syzkaller-14500-g1c41041124bd #0
Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 10/09/2023
Workqueue: bat_events batadv_nc_worker
RIP: 0010:subflow_data_ready+0x39f/0x690 net/mptcp/subflow.c:1405
Code: 18 89 ee e8 e3 d2 21 f7 40 84 ed 75 1f e8 a9 d7 21 f7 44 89 fe bf 07 00 00 00 e8 0c d3 21 f7 41 83 ff 07 74 07 e8 91 d7 21 f7 <0f> 0b e8 8a d7 21 f7 48 89 df e8 d2 b2 ff ff 31 ff 89 c5 89 c6 e8
RSP: 0018:ffffc90000007448 EFLAGS: 00010246
RAX: 0000000000000000 RBX: ffff888031efc700 RCX: ffffffff8a65baf4
RDX: ffff888043222140 RSI: ffffffff8a65baff RDI: 0000000000000005
RBP: 0000000000000000 R08: 0000000000000005 R09: 0000000000000007
R10: 000000000000000b R11: 0000000000000000 R12: 1ffff92000000e89
R13: ffff88807a534d80 R14: ffff888021c11a00 R15: 000000000000000b
FS:  0000000000000000(0000) GS:ffff8880b9800000(0000) knlGS:0000000000000000
CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
CR2: 00007fa19a0ffc81 CR3: 000000007a2db000 CR4: 00000000003506f0
DR0: 000000000000d8dd DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Call Trace:
 <IRQ>
 tcp_data_ready+0x14c/0x5b0 net/ipv4/tcp_input.c:5128
 tcp_data_queue+0x19c3/0x5190 net/ipv4/tcp_input.c:5208
 tcp_rcv_state_process+0x11ef/0x4e10 net/ipv4/tcp_input.c:6844
 tcp_v4_do_rcv+0x369/0xa10 net/ipv4/tcp_ipv4.c:1929
 tcp_v4_rcv+0x3888/0x3b30 net/ipv4/tcp_ipv4.c:2329
 ip_protocol_deliver_rcu+0x9f/0x480 net/ipv4/ip_input.c:205
 ip_local_deliver_finish+0x2e4/0x510 net/ipv4/ip_input.c:233
 NF_HOOK include/linux/netfilter.h:314 [inline]
 NF_HOOK include/linux/netfilter.h:308 [inline]
 ip_local_deliver+0x1b6/0x550 net/ipv4/ip_input.c:254
 dst_input include/net/dst.h:461 [inline]
 ip_rcv_finish+0x1c4/0x2e0 net/ipv4/ip_input.c:449
 NF_HOOK include/linux/netfilter.h:314 [inline]
 NF_HOOK include/linux/netfilter.h:308 [inline]
 ip_rcv+0xce/0x440 net/ipv4/ip_input.c:569
 __netif_receive_skb_one_core+0x115/0x180 net/core/dev.c:5527
 __netif_receive_skb+0x1f/0x1b0 net/core/dev.c:5641
 process_backlog+0x101/0x6b0 net/core/dev.c:5969
 __napi_poll.constprop.0+0xb4/0x540 net/core/dev.c:6531
 napi_poll net/core/dev.c:6600 [inline]
 net_rx_action+0x956/0xe90 net/core/dev.c:6733
 __do_softirq+0x21a/0x968 kernel/softirq.c:553
 do_softirq kernel/softirq.c:454 [inline]
 do_softirq+0xaa/0xe0 kernel/softirq.c:441
 </IRQ>
 <TASK>
 __local_bh_enable_ip+0xf8/0x120 kernel/softirq.c:381
 spin_unlock_bh include/linux/spinlock.h:396 [inline]
 batadv_nc_purge_paths+0x1ce/0x3c0 net/batman-adv/network-coding.c:471
 batadv_nc_worker+0x9b1/0x10e0 net/batman-adv/network-coding.c:722
 process_one_work+0x884/0x15c0 kernel/workqueue.c:2630
 process_scheduled_works kernel/workqueue.c:2703 [inline]
 worker_thread+0x8b9/0x1290 kernel/workqueue.c:2784
 kthread+0x33c/0x440 kernel/kthread.c:388
 ret_from_fork+0x45/0x80 arch/x86/kernel/process.c:147
 ret_from_fork_asm+0x11/0x20 arch/x86/entry/entry_64.S:242
 </TASK>

To address the issue, catch the racing subflow state change and
use it to cause the MPTCP fallback. Such fallback is also used to
cause the first subflow state propagation to the msk socket via
mptcp_set_connected(). After this change, the first subflow can
additionally propagate the TCP_FIN_WAIT1 state, so rename the
helper accordingly.

Finally, if the state propagation is delayed to the msk release
callback, the first subflow can change to a different state in between.
Cache the relevant target state in a new msk-level field and use
such value to update the msk state at release time.

Fixes: 1e777f39b4d7 ("mptcp: add MSG_FASTOPEN sendmsg flag support")
Reported-by: syzbot+c53d4d3ddb327e80bc51@syzkaller.appspotmail.com
Closes: https://github.com/multipath-tcp/mptcp_net-next/issues/458
Signed-off-by: Paolo Abeni <pabeni@redhat.com>
---
Note: we can probably avoid the splat even preventing the PM from
removing any address/subflow before reaching the fully established
status, but we still miss catching a relevant state change.
---
 net/mptcp/protocol.c |  2 +-
 net/mptcp/protocol.h |  7 +++++--
 net/mptcp/subflow.c  | 23 ++++++++++++++---------
 3 files changed, 20 insertions(+), 12 deletions(-)

Comments

MPTCP CI Nov. 15, 2023, 1:38 p.m. UTC | #1
Hi Paolo,

Thank you for your modifications, that's great!

Our CI did some validations and here is its report:

- KVM Validation: normal (except selftest_mptcp_join):
  - Success! ✅:
  - Task: https://cirrus-ci.com/task/6397760975405056
  - Summary: https://api.cirrus-ci.com/v1/artifact/task/6397760975405056/summary/summary.txt

- KVM Validation: debug (only selftest_mptcp_join):
  - Success! ✅:
  - Task: https://cirrus-ci.com/task/6438726373474304
  - Summary: https://api.cirrus-ci.com/v1/artifact/task/6438726373474304/summary/summary.txt

- KVM Validation: normal (only selftest_mptcp_join):
  - Success! ✅:
  - Task: https://cirrus-ci.com/task/5875776420052992
  - Summary: https://api.cirrus-ci.com/v1/artifact/task/5875776420052992/summary/summary.txt

- KVM Validation: debug (except selftest_mptcp_join):
  - Unstable: 1 failed test(s): selftest_diag 
Mat Martineau Nov. 17, 2023, 8:10 p.m. UTC | #2
On Wed, 15 Nov 2023, Paolo Abeni wrote:

> The netlink PM can race with fastopen self-connect attempts, shutting
> down the first subflow via:
>
> MPTCP_PM_CMD_DEL_ADDR -> mptcp_nl_remove_id_zero_address ->
>  mptcp_pm_nl_rm_subflow_received -> mptcp_close_ssk
>
> and transitioning such subflow to FIN_WAIT1 status before the syn-ack
> packet is processed. The MPTCP code does not react to such state change,
> leaving the connection in not-fallback status and the subflow handshake
> uncompleted, triggering the following splat:
>
> WARNING: CPU: 0 PID: 10630 at net/mptcp/subflow.c:1405 subflow_data_ready+0x39f/0x690 net/mptcp/subflow.c:1405
> Modules linked in:
> CPU: 0 PID: 10630 Comm: kworker/u4:11 Not tainted 6.6.0-syzkaller-14500-g1c41041124bd #0
> Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 10/09/2023
> Workqueue: bat_events batadv_nc_worker
> RIP: 0010:subflow_data_ready+0x39f/0x690 net/mptcp/subflow.c:1405
> Code: 18 89 ee e8 e3 d2 21 f7 40 84 ed 75 1f e8 a9 d7 21 f7 44 89 fe bf 07 00 00 00 e8 0c d3 21 f7 41 83 ff 07 74 07 e8 91 d7 21 f7 <0f> 0b e8 8a d7 21 f7 48 89 df e8 d2 b2 ff ff 31 ff 89 c5 89 c6 e8
> RSP: 0018:ffffc90000007448 EFLAGS: 00010246
> RAX: 0000000000000000 RBX: ffff888031efc700 RCX: ffffffff8a65baf4
> RDX: ffff888043222140 RSI: ffffffff8a65baff RDI: 0000000000000005
> RBP: 0000000000000000 R08: 0000000000000005 R09: 0000000000000007
> R10: 000000000000000b R11: 0000000000000000 R12: 1ffff92000000e89
> R13: ffff88807a534d80 R14: ffff888021c11a00 R15: 000000000000000b
> FS:  0000000000000000(0000) GS:ffff8880b9800000(0000) knlGS:0000000000000000
> CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
> CR2: 00007fa19a0ffc81 CR3: 000000007a2db000 CR4: 00000000003506f0
> DR0: 000000000000d8dd DR1: 0000000000000000 DR2: 0000000000000000
> DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
> Call Trace:
> <IRQ>
> tcp_data_ready+0x14c/0x5b0 net/ipv4/tcp_input.c:5128
> tcp_data_queue+0x19c3/0x5190 net/ipv4/tcp_input.c:5208
> tcp_rcv_state_process+0x11ef/0x4e10 net/ipv4/tcp_input.c:6844
> tcp_v4_do_rcv+0x369/0xa10 net/ipv4/tcp_ipv4.c:1929
> tcp_v4_rcv+0x3888/0x3b30 net/ipv4/tcp_ipv4.c:2329
> ip_protocol_deliver_rcu+0x9f/0x480 net/ipv4/ip_input.c:205
> ip_local_deliver_finish+0x2e4/0x510 net/ipv4/ip_input.c:233
> NF_HOOK include/linux/netfilter.h:314 [inline]
> NF_HOOK include/linux/netfilter.h:308 [inline]
> ip_local_deliver+0x1b6/0x550 net/ipv4/ip_input.c:254
> dst_input include/net/dst.h:461 [inline]
> ip_rcv_finish+0x1c4/0x2e0 net/ipv4/ip_input.c:449
> NF_HOOK include/linux/netfilter.h:314 [inline]
> NF_HOOK include/linux/netfilter.h:308 [inline]
> ip_rcv+0xce/0x440 net/ipv4/ip_input.c:569
> __netif_receive_skb_one_core+0x115/0x180 net/core/dev.c:5527
> __netif_receive_skb+0x1f/0x1b0 net/core/dev.c:5641
> process_backlog+0x101/0x6b0 net/core/dev.c:5969
> __napi_poll.constprop.0+0xb4/0x540 net/core/dev.c:6531
> napi_poll net/core/dev.c:6600 [inline]
> net_rx_action+0x956/0xe90 net/core/dev.c:6733
> __do_softirq+0x21a/0x968 kernel/softirq.c:553
> do_softirq kernel/softirq.c:454 [inline]
> do_softirq+0xaa/0xe0 kernel/softirq.c:441
> </IRQ>
> <TASK>
> __local_bh_enable_ip+0xf8/0x120 kernel/softirq.c:381
> spin_unlock_bh include/linux/spinlock.h:396 [inline]
> batadv_nc_purge_paths+0x1ce/0x3c0 net/batman-adv/network-coding.c:471
> batadv_nc_worker+0x9b1/0x10e0 net/batman-adv/network-coding.c:722
> process_one_work+0x884/0x15c0 kernel/workqueue.c:2630
> process_scheduled_works kernel/workqueue.c:2703 [inline]
> worker_thread+0x8b9/0x1290 kernel/workqueue.c:2784
> kthread+0x33c/0x440 kernel/kthread.c:388
> ret_from_fork+0x45/0x80 arch/x86/kernel/process.c:147
> ret_from_fork_asm+0x11/0x20 arch/x86/entry/entry_64.S:242
> </TASK>
>
> To address the issue, catch the racing subflow state change and
> use it to cause the MPTCP fallback. Such fallback is also used to
> cause the first subflow state propagation to the msk socket via
> mptcp_set_connected(). After this change, the first subflow can
> additionally propagate the TCP_FIN_WAIT1 state, so rename the
> helper accordingly.
>
> Finally, if the state propagation is delayed to the msk release
> callback, the first subflow can change to a different state in between.
> Cache the relevant target state in a new msk-level field and use
> such value to update the msk state at release time.
>
> Fixes: 1e777f39b4d7 ("mptcp: add MSG_FASTOPEN sendmsg flag support")
> Reported-by: syzbot+c53d4d3ddb327e80bc51@syzkaller.appspotmail.com
> Closes: https://github.com/multipath-tcp/mptcp_net-next/issues/458
> Signed-off-by: Paolo Abeni <pabeni@redhat.com>
> ---
> Note: we can probably avoid the splat even preventing the PM from
> removing any address/subflow before reaching the fully established
> status, but we still miss catching a relevant state change.

Hi Paolo -

It's useful to be able to remove a subflow that is unable (or slow) to 
connect, so I think the approach in this patch is better.

> ---
> net/mptcp/protocol.c |  2 +-
> net/mptcp/protocol.h |  7 +++++--
> net/mptcp/subflow.c  | 23 ++++++++++++++---------
> 3 files changed, 20 insertions(+), 12 deletions(-)
>
> diff --git a/net/mptcp/protocol.c b/net/mptcp/protocol.c
> index 6956fde61c68..1eea04b11227 100644
> --- a/net/mptcp/protocol.c
> +++ b/net/mptcp/protocol.c
> @@ -3407,7 +3407,7 @@ static void mptcp_release_cb(struct sock *sk)
> 		 * On sk release avoid actions depending on the first subflow
> 		 */
> 		if (__test_and_clear_bit(MPTCP_CONNECTED, &msk->cb_flags) && msk->first)
> -			__mptcp_set_connected(sk);
> +			__mptcp_propagate_state(sk);

As part of my suggestion below (adding new state as a 2nd arg), it would 
be good to clear msk->pending_state here.

> 		if (__test_and_clear_bit(MPTCP_ERROR_REPORT, &msk->cb_flags))
> 			__mptcp_error_report(sk);
> 		if (__test_and_clear_bit(MPTCP_SYNC_SNDBUF, &msk->cb_flags))
> diff --git a/net/mptcp/protocol.h b/net/mptcp/protocol.h
> index 8529b3ae55e9..0148600e4f6d 100644
> --- a/net/mptcp/protocol.h
> +++ b/net/mptcp/protocol.h
> @@ -296,6 +296,9 @@ struct mptcp_sock {
> 	bool		use_64bit_ack; /* Set when we received a 64-bit DSN */
> 	bool		csum_enabled;
> 	bool		allow_infinite_fallback;
> +	u8		pending_state; /* A subflow asked to set this sk_state,
> +					* protected by the msk data lock
> +					*/
> 	u8		mpc_endpoint_id;
> 	u8		recvmsg_inq:1,
> 			cork:1,
> @@ -733,7 +736,7 @@ void mptcp_get_options(const struct sk_buff *skb,
> 		       struct mptcp_options_received *mp_opt);
>
> void mptcp_finish_connect(struct sock *sk);
> -void __mptcp_set_connected(struct sock *sk);
> +void __mptcp_propagate_state(struct sock *sk);
> void mptcp_reset_tout_timer(struct mptcp_sock *msk, unsigned long fail_tout);
>
> static inline void mptcp_stop_tout_timer(struct sock *sk)
> @@ -1129,7 +1132,7 @@ static inline bool subflow_simultaneous_connect(struct sock *sk)
> {
> 	struct mptcp_subflow_context *subflow = mptcp_subflow_ctx(sk);
>
> -	return sk->sk_state == TCP_ESTABLISHED &&
> +	return (1 << sk->sk_state) & (TCPF_ESTABLISHED | TCPF_FIN_WAIT1) &&
> 	       is_active_ssk(subflow) &&
> 	       !subflow->conn_finished;

is_active_ssk() will return true on joins as well - does it make sense to 
force a fallback in that case?

> }
> diff --git a/net/mptcp/subflow.c b/net/mptcp/subflow.c
> index a4f3c27f0309..43e3e54620d2 100644
> --- a/net/mptcp/subflow.c
> +++ b/net/mptcp/subflow.c
> @@ -419,22 +419,27 @@ static bool subflow_use_different_dport(struct mptcp_sock *msk, const struct soc
> 	return inet_sk(sk)->inet_dport != inet_sk((struct sock *)msk)->inet_dport;
> }
>
> -void __mptcp_set_connected(struct sock *sk)
> +void __mptcp_propagate_state(struct sock *sk)

Can you add the new state as a 2nd parameter, rather than using 
msk->pending_state directly? Direct calls then wouldn't have to set 
msk->pending_state, and it helps avoid future misuse with an invalid or 
unset pending_state.

> {
> -	__mptcp_propagate_sndbuf(sk, mptcp_sk(sk)->first);
> +	struct mptcp_sock *msk = mptcp_sk(sk);
> +
> +	__mptcp_propagate_sndbuf(sk, msk->first);
> 	if (sk->sk_state == TCP_SYN_SENT) {
> -		inet_sk_state_store(sk, TCP_ESTABLISHED);
> +		inet_sk_state_store(sk, msk->pending_state);
> 		sk->sk_state_change(sk);
> 	}
> }
>
> -static void mptcp_set_connected(struct sock *sk)
> +static void mptcp_propagate_state(struct sock *sk, struct sock *ssk)
> {
> +	struct mptcp_sock *msk = mptcp_sk(sk);
> +
> 	mptcp_data_lock(sk);
> +	msk->pending_state = ssk->sk_state;

Can move this to the else block below, so pending_state is only set when 
the transition is deferred.

> 	if (!sock_owned_by_user(sk))
> -		__mptcp_set_connected(sk);
> +		__mptcp_propagate_state(sk);
> 	else
> -		__set_bit(MPTCP_CONNECTED, &mptcp_sk(sk)->cb_flags);
> +		__set_bit(MPTCP_CONNECTED, &msk->cb_flags);

Suggest renaming the MPTCP_CONNECTED bit to be consistent with the new 
function name.

Thanks!

- Mat


> 	mptcp_data_unlock(sk);
> }
>
> @@ -496,7 +501,7 @@ static void subflow_finish_connect(struct sock *sk, const struct sk_buff *skb)
> 		subflow_set_remote_key(msk, subflow, &mp_opt);
> 		MPTCP_INC_STATS(sock_net(sk), MPTCP_MIB_MPCAPABLEACTIVEACK);
> 		mptcp_finish_connect(sk);
> -		mptcp_set_connected(parent);
> +		mptcp_propagate_state(parent, sk);
> 	} else if (subflow->request_join) {
> 		u8 hmac[SHA256_DIGEST_SIZE];
>
> @@ -540,7 +545,7 @@ static void subflow_finish_connect(struct sock *sk, const struct sk_buff *skb)
> 	} else if (mptcp_check_fallback(sk)) {
> fallback:
> 		mptcp_rcv_space_init(msk, sk);
> -		mptcp_set_connected(parent);
> +		mptcp_propagate_state(parent, sk);
> 	}
> 	return;
>
> @@ -1740,7 +1745,7 @@ static void subflow_state_change(struct sock *sk)
> 		mptcp_rcv_space_init(msk, sk);
> 		pr_fallback(msk);
> 		subflow->conn_finished = 1;
> -		mptcp_set_connected(parent);
> +		mptcp_propagate_state(parent, sk);
> 	}
>
> 	/* as recvmsg() does not acquire the subflow socket for ssk selection
> -- 
> 2.41.0
>
>
>
Paolo Abeni Nov. 22, 2023, 1:25 p.m. UTC | #3
On Fri, 2023-11-17 at 12:10 -0800, Mat Martineau wrote:
> On Wed, 15 Nov 2023, Paolo Abeni wrote:
> 
> > The netlink PM can race with fastopen self-connect attempts, shutting
> > down the first subflow via:
> > 
> > MPTCP_PM_CMD_DEL_ADDR -> mptcp_nl_remove_id_zero_address ->
> >  mptcp_pm_nl_rm_subflow_received -> mptcp_close_ssk
> > 
> > and transitioning such subflow to FIN_WAIT1 status before the syn-ack
> > packet is processed. The MPTCP code does not react to such state change,
> > leaving the connection in not-fallback status and the subflow handshake
> > uncompleted, triggering the following splat:
> > 
> > WARNING: CPU: 0 PID: 10630 at net/mptcp/subflow.c:1405 subflow_data_ready+0x39f/0x690 net/mptcp/subflow.c:1405
> > Modules linked in:
> > CPU: 0 PID: 10630 Comm: kworker/u4:11 Not tainted 6.6.0-syzkaller-14500-g1c41041124bd #0
> > Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 10/09/2023
> > Workqueue: bat_events batadv_nc_worker
> > RIP: 0010:subflow_data_ready+0x39f/0x690 net/mptcp/subflow.c:1405
> > Code: 18 89 ee e8 e3 d2 21 f7 40 84 ed 75 1f e8 a9 d7 21 f7 44 89 fe bf 07 00 00 00 e8 0c d3 21 f7 41 83 ff 07 74 07 e8 91 d7 21 f7 <0f> 0b e8 8a d7 21 f7 48 89 df e8 d2 b2 ff ff 31 ff 89 c5 89 c6 e8
> > RSP: 0018:ffffc90000007448 EFLAGS: 00010246
> > RAX: 0000000000000000 RBX: ffff888031efc700 RCX: ffffffff8a65baf4
> > RDX: ffff888043222140 RSI: ffffffff8a65baff RDI: 0000000000000005
> > RBP: 0000000000000000 R08: 0000000000000005 R09: 0000000000000007
> > R10: 000000000000000b R11: 0000000000000000 R12: 1ffff92000000e89
> > R13: ffff88807a534d80 R14: ffff888021c11a00 R15: 000000000000000b
> > FS:  0000000000000000(0000) GS:ffff8880b9800000(0000) knlGS:0000000000000000
> > CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
> > CR2: 00007fa19a0ffc81 CR3: 000000007a2db000 CR4: 00000000003506f0
> > DR0: 000000000000d8dd DR1: 0000000000000000 DR2: 0000000000000000
> > DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
> > Call Trace:
> > <IRQ>
> > tcp_data_ready+0x14c/0x5b0 net/ipv4/tcp_input.c:5128
> > tcp_data_queue+0x19c3/0x5190 net/ipv4/tcp_input.c:5208
> > tcp_rcv_state_process+0x11ef/0x4e10 net/ipv4/tcp_input.c:6844
> > tcp_v4_do_rcv+0x369/0xa10 net/ipv4/tcp_ipv4.c:1929
> > tcp_v4_rcv+0x3888/0x3b30 net/ipv4/tcp_ipv4.c:2329
> > ip_protocol_deliver_rcu+0x9f/0x480 net/ipv4/ip_input.c:205
> > ip_local_deliver_finish+0x2e4/0x510 net/ipv4/ip_input.c:233
> > NF_HOOK include/linux/netfilter.h:314 [inline]
> > NF_HOOK include/linux/netfilter.h:308 [inline]
> > ip_local_deliver+0x1b6/0x550 net/ipv4/ip_input.c:254
> > dst_input include/net/dst.h:461 [inline]
> > ip_rcv_finish+0x1c4/0x2e0 net/ipv4/ip_input.c:449
> > NF_HOOK include/linux/netfilter.h:314 [inline]
> > NF_HOOK include/linux/netfilter.h:308 [inline]
> > ip_rcv+0xce/0x440 net/ipv4/ip_input.c:569
> > __netif_receive_skb_one_core+0x115/0x180 net/core/dev.c:5527
> > __netif_receive_skb+0x1f/0x1b0 net/core/dev.c:5641
> > process_backlog+0x101/0x6b0 net/core/dev.c:5969
> > __napi_poll.constprop.0+0xb4/0x540 net/core/dev.c:6531
> > napi_poll net/core/dev.c:6600 [inline]
> > net_rx_action+0x956/0xe90 net/core/dev.c:6733
> > __do_softirq+0x21a/0x968 kernel/softirq.c:553
> > do_softirq kernel/softirq.c:454 [inline]
> > do_softirq+0xaa/0xe0 kernel/softirq.c:441
> > </IRQ>
> > <TASK>
> > __local_bh_enable_ip+0xf8/0x120 kernel/softirq.c:381
> > spin_unlock_bh include/linux/spinlock.h:396 [inline]
> > batadv_nc_purge_paths+0x1ce/0x3c0 net/batman-adv/network-coding.c:471
> > batadv_nc_worker+0x9b1/0x10e0 net/batman-adv/network-coding.c:722
> > process_one_work+0x884/0x15c0 kernel/workqueue.c:2630
> > process_scheduled_works kernel/workqueue.c:2703 [inline]
> > worker_thread+0x8b9/0x1290 kernel/workqueue.c:2784
> > kthread+0x33c/0x440 kernel/kthread.c:388
> > ret_from_fork+0x45/0x80 arch/x86/kernel/process.c:147
> > ret_from_fork_asm+0x11/0x20 arch/x86/entry/entry_64.S:242
> > </TASK>
> > 
> > To address the issue, catch the racing subflow state change and
> > use it to cause the MPTCP fallback. Such fallback is also used to
> > cause the first subflow state propagation to the msk socket via
> > mptcp_set_connected(). After this change, the first subflow can
> > additionally propagate the TCP_FIN_WAIT1 state, so rename the
> > helper accordingly.
> > 
> > Finally, if the state propagation is delayed to the msk release
> > callback, the first subflow can change to a different state in between.
> > Cache the relevant target state in a new msk-level field and use
> > such value to update the msk state at release time.
> > 
> > Fixes: 1e777f39b4d7 ("mptcp: add MSG_FASTOPEN sendmsg flag support")
> > Reported-by: syzbot+c53d4d3ddb327e80bc51@syzkaller.appspotmail.com
> > Closes: https://github.com/multipath-tcp/mptcp_net-next/issues/458
> > Signed-off-by: Paolo Abeni <pabeni@redhat.com>
> > ---
> > Note: we can probably avoid the splat even preventing the PM from
> > removing any address/subflow before reaching the fully established
> > status, but we still miss catching a relevant state change.
> 
> Hi Paolo -
> 
> It's useful to be able to remove a subflow that is unable (or slow) to 
> connect, so I think the approach in this patch is better.
> 
> > ---
> > net/mptcp/protocol.c |  2 +-
> > net/mptcp/protocol.h |  7 +++++--
> > net/mptcp/subflow.c  | 23 ++++++++++++++---------
> > 3 files changed, 20 insertions(+), 12 deletions(-)
> > 
> > diff --git a/net/mptcp/protocol.c b/net/mptcp/protocol.c
> > index 6956fde61c68..1eea04b11227 100644
> > --- a/net/mptcp/protocol.c
> > +++ b/net/mptcp/protocol.c
> > @@ -3407,7 +3407,7 @@ static void mptcp_release_cb(struct sock *sk)
> > 		 * On sk release avoid actions depending on the first subflow
> > 		 */
> > 		if (__test_and_clear_bit(MPTCP_CONNECTED, &msk->cb_flags) && msk->first)
> > -			__mptcp_set_connected(sk);
> > +			__mptcp_propagate_state(sk);
> 
> As part of my suggestion below (adding new state as a 2nd arg), it would 
> be good to clear msk->pending_state here.

I think that would not be needed, as we clear the relevant cb_flags bit
when taking action: the __test_and_clear_bit() above. Additionally
clearing pending_state would be IMHO rendundant.

> 
> > 		if (__test_and_clear_bit(MPTCP_ERROR_REPORT, &msk->cb_flags))
> > 			__mptcp_error_report(sk);
> > 		if (__test_and_clear_bit(MPTCP_SYNC_SNDBUF, &msk->cb_flags))
> > diff --git a/net/mptcp/protocol.h b/net/mptcp/protocol.h
> > index 8529b3ae55e9..0148600e4f6d 100644
> > --- a/net/mptcp/protocol.h
> > +++ b/net/mptcp/protocol.h
> > @@ -296,6 +296,9 @@ struct mptcp_sock {
> > 	bool		use_64bit_ack; /* Set when we received a 64-bit DSN */
> > 	bool		csum_enabled;
> > 	bool		allow_infinite_fallback;
> > +	u8		pending_state; /* A subflow asked to set this sk_state,
> > +					* protected by the msk data lock
> > +					*/
> > 	u8		mpc_endpoint_id;
> > 	u8		recvmsg_inq:1,
> > 			cork:1,
> > @@ -733,7 +736,7 @@ void mptcp_get_options(const struct sk_buff *skb,
> > 		       struct mptcp_options_received *mp_opt);
> > 
> > void mptcp_finish_connect(struct sock *sk);
> > -void __mptcp_set_connected(struct sock *sk);
> > +void __mptcp_propagate_state(struct sock *sk);
> > void mptcp_reset_tout_timer(struct mptcp_sock *msk, unsigned long fail_tout);
> > 
> > static inline void mptcp_stop_tout_timer(struct sock *sk)
> > @@ -1129,7 +1132,7 @@ static inline bool subflow_simultaneous_connect(struct sock *sk)
> > {
> > 	struct mptcp_subflow_context *subflow = mptcp_subflow_ctx(sk);
> > 
> > -	return sk->sk_state == TCP_ESTABLISHED &&
> > +	return (1 << sk->sk_state) & (TCPF_ESTABLISHED | TCPF_FIN_WAIT1) &&
> > 	       is_active_ssk(subflow) &&
> > 	       !subflow->conn_finished;
> 
> is_active_ssk() will return true on joins as well - does it make sense to 
> force a fallback in that case?

Right, not much! I think we should close the subflow. Note that
triggering a self connect on a subflow is very hard. AFAICS the only
way to get there is the remote sending an ADD_ADDR with port to a local
address, and the local kernel binding to such port at connect time by
random chance.

In any case this looks like a separate fix?

Ack to all the other suggestions!

Cheers,

Paolo
diff mbox series

Patch

diff --git a/net/mptcp/protocol.c b/net/mptcp/protocol.c
index 6956fde61c68..1eea04b11227 100644
--- a/net/mptcp/protocol.c
+++ b/net/mptcp/protocol.c
@@ -3407,7 +3407,7 @@  static void mptcp_release_cb(struct sock *sk)
 		 * On sk release avoid actions depending on the first subflow
 		 */
 		if (__test_and_clear_bit(MPTCP_CONNECTED, &msk->cb_flags) && msk->first)
-			__mptcp_set_connected(sk);
+			__mptcp_propagate_state(sk);
 		if (__test_and_clear_bit(MPTCP_ERROR_REPORT, &msk->cb_flags))
 			__mptcp_error_report(sk);
 		if (__test_and_clear_bit(MPTCP_SYNC_SNDBUF, &msk->cb_flags))
diff --git a/net/mptcp/protocol.h b/net/mptcp/protocol.h
index 8529b3ae55e9..0148600e4f6d 100644
--- a/net/mptcp/protocol.h
+++ b/net/mptcp/protocol.h
@@ -296,6 +296,9 @@  struct mptcp_sock {
 	bool		use_64bit_ack; /* Set when we received a 64-bit DSN */
 	bool		csum_enabled;
 	bool		allow_infinite_fallback;
+	u8		pending_state; /* A subflow asked to set this sk_state,
+					* protected by the msk data lock
+					*/
 	u8		mpc_endpoint_id;
 	u8		recvmsg_inq:1,
 			cork:1,
@@ -733,7 +736,7 @@  void mptcp_get_options(const struct sk_buff *skb,
 		       struct mptcp_options_received *mp_opt);
 
 void mptcp_finish_connect(struct sock *sk);
-void __mptcp_set_connected(struct sock *sk);
+void __mptcp_propagate_state(struct sock *sk);
 void mptcp_reset_tout_timer(struct mptcp_sock *msk, unsigned long fail_tout);
 
 static inline void mptcp_stop_tout_timer(struct sock *sk)
@@ -1129,7 +1132,7 @@  static inline bool subflow_simultaneous_connect(struct sock *sk)
 {
 	struct mptcp_subflow_context *subflow = mptcp_subflow_ctx(sk);
 
-	return sk->sk_state == TCP_ESTABLISHED &&
+	return (1 << sk->sk_state) & (TCPF_ESTABLISHED | TCPF_FIN_WAIT1) &&
 	       is_active_ssk(subflow) &&
 	       !subflow->conn_finished;
 }
diff --git a/net/mptcp/subflow.c b/net/mptcp/subflow.c
index a4f3c27f0309..43e3e54620d2 100644
--- a/net/mptcp/subflow.c
+++ b/net/mptcp/subflow.c
@@ -419,22 +419,27 @@  static bool subflow_use_different_dport(struct mptcp_sock *msk, const struct soc
 	return inet_sk(sk)->inet_dport != inet_sk((struct sock *)msk)->inet_dport;
 }
 
-void __mptcp_set_connected(struct sock *sk)
+void __mptcp_propagate_state(struct sock *sk)
 {
-	__mptcp_propagate_sndbuf(sk, mptcp_sk(sk)->first);
+	struct mptcp_sock *msk = mptcp_sk(sk);
+
+	__mptcp_propagate_sndbuf(sk, msk->first);
 	if (sk->sk_state == TCP_SYN_SENT) {
-		inet_sk_state_store(sk, TCP_ESTABLISHED);
+		inet_sk_state_store(sk, msk->pending_state);
 		sk->sk_state_change(sk);
 	}
 }
 
-static void mptcp_set_connected(struct sock *sk)
+static void mptcp_propagate_state(struct sock *sk, struct sock *ssk)
 {
+	struct mptcp_sock *msk = mptcp_sk(sk);
+
 	mptcp_data_lock(sk);
+	msk->pending_state = ssk->sk_state;
 	if (!sock_owned_by_user(sk))
-		__mptcp_set_connected(sk);
+		__mptcp_propagate_state(sk);
 	else
-		__set_bit(MPTCP_CONNECTED, &mptcp_sk(sk)->cb_flags);
+		__set_bit(MPTCP_CONNECTED, &msk->cb_flags);
 	mptcp_data_unlock(sk);
 }
 
@@ -496,7 +501,7 @@  static void subflow_finish_connect(struct sock *sk, const struct sk_buff *skb)
 		subflow_set_remote_key(msk, subflow, &mp_opt);
 		MPTCP_INC_STATS(sock_net(sk), MPTCP_MIB_MPCAPABLEACTIVEACK);
 		mptcp_finish_connect(sk);
-		mptcp_set_connected(parent);
+		mptcp_propagate_state(parent, sk);
 	} else if (subflow->request_join) {
 		u8 hmac[SHA256_DIGEST_SIZE];
 
@@ -540,7 +545,7 @@  static void subflow_finish_connect(struct sock *sk, const struct sk_buff *skb)
 	} else if (mptcp_check_fallback(sk)) {
 fallback:
 		mptcp_rcv_space_init(msk, sk);
-		mptcp_set_connected(parent);
+		mptcp_propagate_state(parent, sk);
 	}
 	return;
 
@@ -1740,7 +1745,7 @@  static void subflow_state_change(struct sock *sk)
 		mptcp_rcv_space_init(msk, sk);
 		pr_fallback(msk);
 		subflow->conn_finished = 1;
-		mptcp_set_connected(parent);
+		mptcp_propagate_state(parent, sk);
 	}
 
 	/* as recvmsg() does not acquire the subflow socket for ssk selection