You can subscribe to this list here.
2003 |
Jan
|
Feb
|
Mar
|
Apr
|
May
|
Jun
|
Jul
|
Aug
|
Sep
|
Oct
|
Nov
|
Dec
(6) |
---|---|---|---|---|---|---|---|---|---|---|---|---|
2004 |
Jan
(9) |
Feb
(11) |
Mar
(22) |
Apr
(73) |
May
(78) |
Jun
(146) |
Jul
(80) |
Aug
(27) |
Sep
(5) |
Oct
(14) |
Nov
(18) |
Dec
(27) |
2005 |
Jan
(20) |
Feb
(30) |
Mar
(19) |
Apr
(28) |
May
(50) |
Jun
(31) |
Jul
(32) |
Aug
(14) |
Sep
(36) |
Oct
(43) |
Nov
(74) |
Dec
(63) |
2006 |
Jan
(34) |
Feb
(32) |
Mar
(21) |
Apr
(76) |
May
(106) |
Jun
(72) |
Jul
(70) |
Aug
(175) |
Sep
(130) |
Oct
(39) |
Nov
(81) |
Dec
(43) |
2007 |
Jan
(81) |
Feb
(36) |
Mar
(20) |
Apr
(43) |
May
(54) |
Jun
(34) |
Jul
(44) |
Aug
(55) |
Sep
(44) |
Oct
(54) |
Nov
(43) |
Dec
(41) |
2008 |
Jan
(42) |
Feb
(84) |
Mar
(73) |
Apr
(30) |
May
(119) |
Jun
(54) |
Jul
(54) |
Aug
(93) |
Sep
(173) |
Oct
(130) |
Nov
(145) |
Dec
(153) |
2009 |
Jan
(59) |
Feb
(12) |
Mar
(28) |
Apr
(18) |
May
(56) |
Jun
(9) |
Jul
(28) |
Aug
(62) |
Sep
(16) |
Oct
(19) |
Nov
(15) |
Dec
(17) |
2010 |
Jan
(14) |
Feb
(36) |
Mar
(37) |
Apr
(30) |
May
(33) |
Jun
(53) |
Jul
(42) |
Aug
(50) |
Sep
(67) |
Oct
(66) |
Nov
(69) |
Dec
(36) |
2011 |
Jan
(52) |
Feb
(45) |
Mar
(49) |
Apr
(21) |
May
(34) |
Jun
(13) |
Jul
(19) |
Aug
(37) |
Sep
(43) |
Oct
(10) |
Nov
(23) |
Dec
(30) |
2012 |
Jan
(42) |
Feb
(36) |
Mar
(46) |
Apr
(25) |
May
(96) |
Jun
(146) |
Jul
(40) |
Aug
(28) |
Sep
(61) |
Oct
(45) |
Nov
(100) |
Dec
(53) |
2013 |
Jan
(79) |
Feb
(24) |
Mar
(134) |
Apr
(156) |
May
(118) |
Jun
(75) |
Jul
(278) |
Aug
(145) |
Sep
(136) |
Oct
(168) |
Nov
(137) |
Dec
(439) |
2014 |
Jan
(284) |
Feb
(158) |
Mar
(231) |
Apr
(275) |
May
(259) |
Jun
(91) |
Jul
(222) |
Aug
(215) |
Sep
(165) |
Oct
(166) |
Nov
(211) |
Dec
(150) |
2015 |
Jan
(164) |
Feb
(324) |
Mar
(299) |
Apr
(214) |
May
(111) |
Jun
(109) |
Jul
(105) |
Aug
(36) |
Sep
(58) |
Oct
(131) |
Nov
(68) |
Dec
(30) |
2016 |
Jan
(46) |
Feb
(87) |
Mar
(135) |
Apr
(174) |
May
(132) |
Jun
(135) |
Jul
(149) |
Aug
(125) |
Sep
(79) |
Oct
(49) |
Nov
(95) |
Dec
(102) |
2017 |
Jan
(104) |
Feb
(75) |
Mar
(72) |
Apr
(53) |
May
(18) |
Jun
(5) |
Jul
(14) |
Aug
(19) |
Sep
(2) |
Oct
(13) |
Nov
(21) |
Dec
(67) |
2018 |
Jan
(56) |
Feb
(50) |
Mar
(148) |
Apr
(41) |
May
(37) |
Jun
(34) |
Jul
(34) |
Aug
(11) |
Sep
(52) |
Oct
(48) |
Nov
(28) |
Dec
(46) |
2019 |
Jan
(29) |
Feb
(63) |
Mar
(95) |
Apr
(54) |
May
(14) |
Jun
(71) |
Jul
(60) |
Aug
(49) |
Sep
(3) |
Oct
(64) |
Nov
(115) |
Dec
(57) |
2020 |
Jan
(15) |
Feb
(9) |
Mar
(38) |
Apr
(27) |
May
(60) |
Jun
(53) |
Jul
(35) |
Aug
(46) |
Sep
(37) |
Oct
(64) |
Nov
(20) |
Dec
(25) |
2021 |
Jan
(20) |
Feb
(31) |
Mar
(27) |
Apr
(23) |
May
(21) |
Jun
(30) |
Jul
(30) |
Aug
(7) |
Sep
(18) |
Oct
|
Nov
(15) |
Dec
(4) |
2022 |
Jan
(3) |
Feb
(1) |
Mar
(10) |
Apr
|
May
(2) |
Jun
(26) |
Jul
(5) |
Aug
|
Sep
(1) |
Oct
(2) |
Nov
(9) |
Dec
(2) |
2023 |
Jan
(4) |
Feb
(4) |
Mar
(5) |
Apr
(10) |
May
(29) |
Jun
(17) |
Jul
|
Aug
|
Sep
(1) |
Oct
(1) |
Nov
(2) |
Dec
|
2024 |
Jan
|
Feb
(6) |
Mar
|
Apr
(1) |
May
(6) |
Jun
|
Jul
(5) |
Aug
|
Sep
(3) |
Oct
|
Nov
|
Dec
|
2025 |
Jan
|
Feb
(3) |
Mar
|
Apr
|
May
|
Jun
|
Jul
|
Aug
|
Sep
|
Oct
|
Nov
|
Dec
|
From: Hou, H. <HH...@rb...> - 2025-02-05 16:21:13
|
Hi Tung, Thanks for your reply. In my application, there is a need to dynamically add and remove the remote IP addresses. Can we add an option to the "tipc bearer" command to delete one of the remote IP? Since we have the "add", it makes sense to have a "delete" to pair with it. Is there any other way to remove a remote IP from a bearer without disabling the bearer. Thanks, Hao -----Original Message----- From: Tung Nguyen <tun...@en...> Sent: Tuesday, February 4, 2025 7:25 PM To: Hou, Hao <HH...@rb...>; tip...@li... Subject: [EXTERNAL] RE: Question on removing remote IP >TIPC supports adding multiple remote IP addresses to a UDP bearer with "tipc bearer add" command. >How do we remove a previously added remote IP address from a UDP bearer without impacting the TIPC communications with other existing remote IP addresses in the same UDP bearer? The "tipc bearer --help" doesn't show such option. No, there is no such option. The information in this email is confidential and may be legally privileged. It is intended solely for the addressee. Any opinions expressed are mine and do not necessarily represent the opinions of the Company. Emails are susceptible to interference. If you are not the intended recipient, any disclosure, copying, distribution or any action taken or omitted to be taken in reliance on it, is strictly prohibited and may be unlawful. If you have received this message in error, do not open any attachments but please notify the Endava Service Desk on (+44 (0)870 423 0187), and delete this message from your system. The sender accepts no responsibility for information, errors or omissions in this email, or for its use or misuse, or for any act committed or omitted in connection with this communication. If in doubt, please verify the authenticity of the contents with the sender. Please rely on your own virus checkers as no responsibility is taken by the sender for any damage rising out of any bug or virus infection. Endava plc is a company registered in England under company number 5722669 whose registered office is at 125 Old Broad Street, London, EC2N 1AR, United Kingdom. Endava plc is the Endava group holding company and does not provide any services to clients. Each of Endava plc and its subsidiaries is a separate legal entity and has no liability for another such entity's acts or omissions. Disclaimer This e-mail together with any attachments may contain information of Ribbon Communications Inc. and its Affiliates that is confidential and/or proprietary for the sole use of the intended recipient. Any review, disclosure, reliance or distribution by others or forwarding without express permission is strictly prohibited. If you are not the intended recipient, please notify the sender immediately and then delete all copies, including any attachments. |
From: Tung N. <tun...@en...> - 2025-02-05 03:59:34
|
>TIPC supports adding multiple remote IP addresses to a UDP bearer with "tipc bearer add" command. >How do we remove a previously added remote IP address from a UDP bearer without impacting the TIPC communications with other existing remote IP addresses in the same UDP bearer? The "tipc bearer --help" doesn't show such option. No, there is no such option. The information in this email is confidential and may be legally privileged. It is intended solely for the addressee. Any opinions expressed are mine and do not necessarily represent the opinions of the Company. Emails are susceptible to interference. If you are not the intended recipient, any disclosure, copying, distribution or any action taken or omitted to be taken in reliance on it, is strictly prohibited and may be unlawful. If you have received this message in error, do not open any attachments but please notify the Endava Service Desk on (+44 (0)870 423 0187), and delete this message from your system. The sender accepts no responsibility for information, errors or omissions in this email, or for its use or misuse, or for any act committed or omitted in connection with this communication. If in doubt, please verify the authenticity of the contents with the sender. Please rely on your own virus checkers as no responsibility is taken by the sender for any damage rising out of any bug or virus infection. Endava plc is a company registered in England under company number 5722669 whose registered office is at 125 Old Broad Street, London, EC2N 1AR, United Kingdom. Endava plc is the Endava group holding company and does not provide any services to clients. Each of Endava plc and its subsidiaries is a separate legal entity and has no liability for another such entity's acts or omissions. |
From: Hou, H. <HH...@rb...> - 2025-02-04 16:14:09
|
Greetings TIPC team, TIPC supports adding multiple remote IP addresses to a UDP bearer with "tipc bearer add" command. How do we remove a previously added remote IP address from a UDP bearer without impacting the TIPC communications with other existing remote IP addresses in the same UDP bearer? The "tipc bearer --help" doesn't show such option. Thanks, Hao Disclaimer This e-mail together with any attachments may contain information of Ribbon Communications Inc. and its Affiliates that is confidential and/or proprietary for the sole use of the intended recipient. Any review, disclosure, reliance or distribution by others or forwarding without express permission is strictly prohibited. If you are not the intended recipient, please notify the sender immediately and then delete all copies, including any attachments. |
From: prakash b. <ps1...@gm...> - 2024-09-18 17:09:27
|
Thank you Jon for your response. In some cases, we might be hitting a race-condition during the small window that you have mentioned. To give you more context about our use case which is little different than you have mentioned in your solution. We have one publisher and many subscribers. Typically the publisher is created first and the subscribers may come later at any time. The subscriber application needs to pull already published data from the publisher application. Currently the subscriber application is doing the pull just after the bind() system call through a different channel( not pub/sub). But if some data is published after the pull and before the subscriber socket becomes ready, then that data will never reach the subscriber application. Is there any way to know if the subscriber socket is fully ready inside the subscriber application ? Or If the subscriber application waits for a few milliseconds ( e.g.100ms ) after bind() system call() and before the pull, would it be a practical approach to avoid this race condition ? Kindly note that we are using tipc pub/sub sockets with scope as "TIPC_NODE_SCOPE" so the window we are talking about here should be relatively very small. Thanks and Regards, Prakash On Tue, Sep 17, 2024 at 1:01 AM Jon Maloy <jm...@re...> wrote: > > > On 2024-09-16 07:58, prakash bisht wrote: > > Hi Hi Tung,Xin,John, > > > > We are using tipc multicast sockets for publisher/subscriber type of > > application. > > Below are the socket details > > socket type : SOCK_RDM > > tipc addr type : TIPC_MCAST > > > > We would like to know if the successful bind() call on the subscriber > > socket guarantees that the subscriber is ready and no publication > > would be missed ? Or shall we wait for some time before assuming the > > subscriber socket is ready ? > There may be (very) a small lag between the return of the bind() call > until the bind is registered on all the other nodes. > The proper way to get around this is to let your "publishers" use the > TIPC subscription mechanism to keep track of your > "subscriber", i.e., they don´t start publishing until they have received > a notification that the subscriber is bound > to the service address/range. > ///jon > > > > > > Thanks and Regards, > > Prakash > > |
From: Jon M. <jm...@re...> - 2024-09-16 19:31:45
|
On 2024-09-16 07:58, prakash bisht wrote: > Hi Hi Tung,Xin,John, > > We are using tipc multicast sockets for publisher/subscriber type of > application. > Below are the socket details > socket type : SOCK_RDM > tipc addr type : TIPC_MCAST > > We would like to know if the successful bind() call on the subscriber > socket guarantees that the subscriber is ready and no publication > would be missed ? Or shall we wait for some time before assuming the > subscriber socket is ready ? There may be (very) a small lag between the return of the bind() call until the bind is registered on all the other nodes. The proper way to get around this is to let your "publishers" use the TIPC subscription mechanism to keep track of your "subscriber", i.e., they don´t start publishing until they have received a notification that the subscriber is bound to the service address/range. ///jon > > > Thanks and Regards, > Prakash |
From: prakash b. <ps1...@gm...> - 2024-09-16 11:59:12
|
Hi Hi Tung,Xin,John, We are using tipc multicast sockets for publisher/subscriber type of application. Below are the socket details socket type : SOCK_RDM tipc addr type : TIPC_MCAST We would like to know if the successful bind() call on the subscriber socket guarantees that the subscriber is ready and no publication would be missed ? Or shall we wait for some time before assuming the subscriber socket is ready ? Thanks and Regards, Prakash |
From: Tung N. <tun...@en...> - 2024-07-24 10:05:56
|
Hi Tung,Xin,John, Any thoughts about the previous issue ? We are seeing this quite often in our system now. Any help would be appreciated. Does this happen on a specific node or on some nodes of your cluster? We can only check if you provide syslog (/var/log/messages) of all nodes. The information in this email is confidential and may be legally privileged. It is intended solely for the addressee. Any opinions expressed are mine and do not necessarily represent the opinions of the Company. Emails are susceptible to interference. If you are not the intended recipient, any disclosure, copying, distribution or any action taken or omitted to be taken in reliance on it, is strictly prohibited and may be unlawful. If you have received this message in error, do not open any attachments but please notify the Endava Service Desk on (+44 (0)870 423 0187), and delete this message from your system. The sender accepts no responsibility for information, errors or omissions in this email, or for its use or misuse, or for any act committed or omitted in connection with this communication. If in doubt, please verify the authenticity of the contents with the sender. Please rely on your own virus checkers as no responsibility is taken by the sender for any damage rising out of any bug or virus infection. Endava plc is a company registered in England under company number 5722669 whose registered office is at 125 Old Broad Street, London, EC2N 1AR, United Kingdom. Endava plc is the Endava group holding company and does not provide any services to clients. Each of Endava plc and its subsidiaries is a separate legal entity and has no liability for another such entity's acts or omissions. |
From: prakash b. <ps1...@gm...> - 2024-07-24 09:43:13
|
Hi Tung,Xin,John, Any thoughts about the previous issue ? We are seeing this quite often in our system now. Any help would be appreciated. Thanks and Regards, Prakash On Wed, Jul 17, 2024 at 5:57 PM prakash bisht <ps1...@gm...> wrote: > Thanks Tung for your response. I verified the kernel logs in our setup. >> There is no memory error as in your case. >> >> I see the below pattern though. >> 1. There are multiple local publication related messages. But TIPC is not >> entering the infinite loop immediately. >> Jul 16 13:23:31 in-debbld39-pd kernel: [90527.681211] Failed to >> remove local publication {2,1359237139,1359237139}/0 >> 2. There are some errors of illegal FSM even after which the infinite >> loop is triggered. >> Jul 16 13:23:53 in-debbld39-pd kernel: [90550.396954] Illegal FSM >> event fa110ede in state d0000 on link >> 13500451:INTERCC_BEARER-135004a1:INTERCC_BEARER >> Not sure if the above error has some relation with hitting "Unable to >> remove publication from failed node" infinite loop issue. >> >> >> From kernel.log file - >> >> /**************************************************************************************************************************************************************************************************************************************/ >> Jul 16 13:23:31 in-debbld39-pd kernel: [90527.681211] Failed to remove >> local publication {2,1359237139,1359237139}/0 >> Jul 16 13:23:37 in-debbld39-pd kernel: [90533.911632] [UFW BLOCK] IN=eno1 >> OUT= MAC=5c:b9:01:fe:f6:d0:40:a8:f0:3b:ce:40:08:00 SRC=10.220.82.25 >> DST=10.220.82.39 LEN=60 TOS=0x00 PREC=0x00 TTL=64 ID=45831 DF PROTO=TCP >> SPT=46905 DPT=111 WINDOW=64240 RES=0x00 SYN URGP=0 >> Jul 16 13:23:37 in-debbld39-pd kernel: [90533.913988] [UFW BLOCK] IN=eno1 >> OUT= MAC=5c:b9:01:fe:f6:d0:d4:f5:ef:70:c9:10:08:00 SRC=10.220.82.110 >> DST=10.220.82.39 LEN=60 TOS=0x00 PREC=0x00 TTL=64 ID=64852 DF PROTO=TCP >> SPT=48891 DPT=111 WINDOW=64240 RES=0x00 SYN URGP=0 >> Jul 16 13:23:37 in-debbld39-pd kernel: [90533.915163] [UFW BLOCK] IN=eno1 >> OUT= MAC=5c:b9:01:fe:f6:d0:ec:b1:d7:83:83:90:08:00 SRC=10.220.82.28 >> DST=10.220.82.39 LEN=60 TOS=0x00 PREC=0x00 TTL=64 ID=27835 DF PROTO=TCP >> SPT=60415 DPT=111 WINDOW=64240 RES=0x00 SYN URGP=0 >> Jul 16 13:23:41 in-debbld39-pd kernel: [90537.913017] Failed to remove >> local publication {2,1359237139,1359237139}/0 >> Jul 16 13:23:42 in-debbld39-pd kernel: [90539.299673] ACPI Error: >> SMBus/IPMI/GenericSerialBus write requires Buffer of length 66, found >> length 32 (20180810/exfield-393) >> Jul 16 13:23:42 in-debbld39-pd kernel: [90539.299684] ACPI Error: Method >> parse/execution failed \_SB.PMI0._PMM, AE_AML_BUFFER_LIMIT >> (20180810/psparse-516) >> Jul 16 13:23:42 in-debbld39-pd kernel: [90539.299695] ACPI Error: >> AE_AML_BUFFER_LIMIT, Evaluating _PMM (20180810/power_meter-339) >> Jul 16 13:23:48 in-debbld39-pd kernel: [90544.825004] Failed to remove >> local publication {2,1359237139,1359237139}/0 >> Jul 16 13:23:51 in-debbld39-pd kernel: [90547.993176] Failed to remove >> local publication {2,1359237139,1359237139}/0 >> Jul 16 13:23:53 in-debbld39-pd kernel: [90549.816973] Failed to remove >> local publication {2,1359237139,1359237139}/0 >> Jul 16 13:23:53 in-debbld39-pd kernel: [90550.396954] Illegal FSM event >> fa110ede in state d0000 on link >> 13500451:INTERCC_BEARER-135004a1:INTERCC_BEARER >> Jul 16 13:23:53 in-debbld39-pd kernel: [90550.396971] Unable to remove >> publication from failed node >> Jul 16 13:23:53 in-debbld39-pd kernel: [90550.396971] (type=0, >> lower=2701414419, node=0xa1045013, port=0, key=2701414419) >> Jul 16 13:23:53 in-debbld39-pd kernel: [90550.396976] Unable to remove >> publication from failed node >> Jul 16 13:23:53 in-debbld39-pd kernel: [90550.396976] (type=19398666, >> lower=1, node=0xa1045013, port=3268044884, key=3268044885) >> Jul 16 13:23:53 in-debbld39-pd kernel: [90550.396978] Unable to remove >> publication from failed node >> >> /**************************************************************************************************************************************************************************************************************************************/ >> >> Thanks, >> Prakash >> >> On Mon, Jul 8, 2024 at 3:32 PM Tung Quang Nguyen < >> tun...@de...> wrote: >> >>> >*Jun 29 14:45:36 in-debbld-33 kernel: [1724399.196945] Unable to remove >>> publication from failed node Jun 29 14:45:36 in-debbld-33 >>> >kernel: >>> >[1724399.196945] (type=20185106, lower=1, node=0xd1045013, >>> port=2177385505, key=2177385506) Jun 29 14:45:36 in-debbld-33 >>> >kernel: >>> >[1724399.196948] Unable to remove publication from failed node Jun 29 >>> >14:45:36 in-debbld-33 kernel: [1724399.196948] (type=20185106, >>> lower=1, node=0xd1045013, port=2177385505, key=2177385506) >>> >Jun 29 14:45:36 >>> >in-debbld-33 kernel: [1724399.196954] Unable to remove publication from >>> failed node Jun 29 14:45:36 in-debbld-33 kernel: >>> >[1724399.196954] (type=20185106, lower=1, node=0xd1045013, >>> port=2177385505, key=2177385506)* >>> > >>> >>> >############################################################################################################### >>> >###### >>> > >>> > >>> > >>> >Any idea why this would be happening ? >>> Memory error could cause this. I observed the same thing in my system: >>> " >>> 2024-05-27T07:27:22.747+02:00 kernel:[425219.492047] {...}[Hardware >>> Error]: Hardware error from APEI Generic Hardware Error Source: 0 >>> ... >>> 2024-05-27T07:27:22.747+02:00 kernel:[425219.517622] {...}[Hardware >>> Error]: section_type: memory error >>> ... >>> 2024-05-27T07:27:22.973+02:00 kernel:[425221.704107] tipc: Unable to >>> remove publication from failed node >>> 2024-05-27T07:27:22.973+02:00 kernel:[425221.704107] (type=143322, >>> lower=9, node=0x1001009, port=1295954722, key=1295954723) >>> 2024-05-27T07:27:22.973+02:00 kernel:[425221.717912] tipc: Unable to >>> remove publication from failed node >>> 2024-05-27T07:27:22.973+02:00 kernel:[425221.717912] (type=143322, >>> lower=9, node=0x1001009, port=1295954722, key=1295954723) >>> " >>> >> |
From: prakash b. <ps1...@gm...> - 2024-07-17 12:28:04
|
> > Thanks Tung for your response. I verified the kernel logs in our setup. > There is no memory error as in your case. > > I see the below pattern though. > 1. There are multiple local publication related messages. But TIPC is not > entering the infinite loop immediately. > Jul 16 13:23:31 in-debbld39-pd kernel: [90527.681211] Failed to > remove local publication {2,1359237139,1359237139}/0 > 2. There are some errors of illegal FSM even after which the infinite loop > is triggered. > Jul 16 13:23:53 in-debbld39-pd kernel: [90550.396954] Illegal FSM > event fa110ede in state d0000 on link > 13500451:INTERCC_BEARER-135004a1:INTERCC_BEARER > Not sure if the above error has some relation with hitting "Unable to > remove publication from failed node" infinite loop issue. > > > From kernel.log file - > > /**************************************************************************************************************************************************************************************************************************************/ > Jul 16 13:23:31 in-debbld39-pd kernel: [90527.681211] Failed to remove > local publication {2,1359237139,1359237139}/0 > Jul 16 13:23:37 in-debbld39-pd kernel: [90533.911632] [UFW BLOCK] IN=eno1 > OUT= MAC=5c:b9:01:fe:f6:d0:40:a8:f0:3b:ce:40:08:00 SRC=10.220.82.25 > DST=10.220.82.39 LEN=60 TOS=0x00 PREC=0x00 TTL=64 ID=45831 DF PROTO=TCP > SPT=46905 DPT=111 WINDOW=64240 RES=0x00 SYN URGP=0 > Jul 16 13:23:37 in-debbld39-pd kernel: [90533.913988] [UFW BLOCK] IN=eno1 > OUT= MAC=5c:b9:01:fe:f6:d0:d4:f5:ef:70:c9:10:08:00 SRC=10.220.82.110 > DST=10.220.82.39 LEN=60 TOS=0x00 PREC=0x00 TTL=64 ID=64852 DF PROTO=TCP > SPT=48891 DPT=111 WINDOW=64240 RES=0x00 SYN URGP=0 > Jul 16 13:23:37 in-debbld39-pd kernel: [90533.915163] [UFW BLOCK] IN=eno1 > OUT= MAC=5c:b9:01:fe:f6:d0:ec:b1:d7:83:83:90:08:00 SRC=10.220.82.28 > DST=10.220.82.39 LEN=60 TOS=0x00 PREC=0x00 TTL=64 ID=27835 DF PROTO=TCP > SPT=60415 DPT=111 WINDOW=64240 RES=0x00 SYN URGP=0 > Jul 16 13:23:41 in-debbld39-pd kernel: [90537.913017] Failed to remove > local publication {2,1359237139,1359237139}/0 > Jul 16 13:23:42 in-debbld39-pd kernel: [90539.299673] ACPI Error: > SMBus/IPMI/GenericSerialBus write requires Buffer of length 66, found > length 32 (20180810/exfield-393) > Jul 16 13:23:42 in-debbld39-pd kernel: [90539.299684] ACPI Error: Method > parse/execution failed \_SB.PMI0._PMM, AE_AML_BUFFER_LIMIT > (20180810/psparse-516) > Jul 16 13:23:42 in-debbld39-pd kernel: [90539.299695] ACPI Error: > AE_AML_BUFFER_LIMIT, Evaluating _PMM (20180810/power_meter-339) > Jul 16 13:23:48 in-debbld39-pd kernel: [90544.825004] Failed to remove > local publication {2,1359237139,1359237139}/0 > Jul 16 13:23:51 in-debbld39-pd kernel: [90547.993176] Failed to remove > local publication {2,1359237139,1359237139}/0 > Jul 16 13:23:53 in-debbld39-pd kernel: [90549.816973] Failed to remove > local publication {2,1359237139,1359237139}/0 > Jul 16 13:23:53 in-debbld39-pd kernel: [90550.396954] Illegal FSM event > fa110ede in state d0000 on link > 13500451:INTERCC_BEARER-135004a1:INTERCC_BEARER > Jul 16 13:23:53 in-debbld39-pd kernel: [90550.396971] Unable to remove > publication from failed node > Jul 16 13:23:53 in-debbld39-pd kernel: [90550.396971] (type=0, > lower=2701414419, node=0xa1045013, port=0, key=2701414419) > Jul 16 13:23:53 in-debbld39-pd kernel: [90550.396976] Unable to remove > publication from failed node > Jul 16 13:23:53 in-debbld39-pd kernel: [90550.396976] (type=19398666, > lower=1, node=0xa1045013, port=3268044884, key=3268044885) > Jul 16 13:23:53 in-debbld39-pd kernel: [90550.396978] Unable to remove > publication from failed node > > /**************************************************************************************************************************************************************************************************************************************/ > > Thanks, > Prakash > > On Mon, Jul 8, 2024 at 3:32 PM Tung Quang Nguyen < > tun...@de...> wrote: > >> >*Jun 29 14:45:36 in-debbld-33 kernel: [1724399.196945] Unable to remove >> publication from failed node Jun 29 14:45:36 in-debbld-33 >> >kernel: >> >[1724399.196945] (type=20185106, lower=1, node=0xd1045013, >> port=2177385505, key=2177385506) Jun 29 14:45:36 in-debbld-33 >> >kernel: >> >[1724399.196948] Unable to remove publication from failed node Jun 29 >> >14:45:36 in-debbld-33 kernel: [1724399.196948] (type=20185106, lower=1, >> node=0xd1045013, port=2177385505, key=2177385506) >> >Jun 29 14:45:36 >> >in-debbld-33 kernel: [1724399.196954] Unable to remove publication from >> failed node Jun 29 14:45:36 in-debbld-33 kernel: >> >[1724399.196954] (type=20185106, lower=1, node=0xd1045013, >> port=2177385505, key=2177385506)* >> > >> >> >############################################################################################################### >> >###### >> > >> > >> > >> >Any idea why this would be happening ? >> Memory error could cause this. I observed the same thing in my system: >> " >> 2024-05-27T07:27:22.747+02:00 kernel:[425219.492047] {...}[Hardware >> Error]: Hardware error from APEI Generic Hardware Error Source: 0 >> ... >> 2024-05-27T07:27:22.747+02:00 kernel:[425219.517622] {...}[Hardware >> Error]: section_type: memory error >> ... >> 2024-05-27T07:27:22.973+02:00 kernel:[425221.704107] tipc: Unable to >> remove publication from failed node >> 2024-05-27T07:27:22.973+02:00 kernel:[425221.704107] (type=143322, >> lower=9, node=0x1001009, port=1295954722, key=1295954723) >> 2024-05-27T07:27:22.973+02:00 kernel:[425221.717912] tipc: Unable to >> remove publication from failed node >> 2024-05-27T07:27:22.973+02:00 kernel:[425221.717912] (type=143322, >> lower=9, node=0x1001009, port=1295954722, key=1295954723) >> " >> > |
From: Tung Q. N. <tun...@de...> - 2024-07-08 10:02:58
|
>*Jun 29 14:45:36 in-debbld-33 kernel: [1724399.196945] Unable to remove publication from failed node Jun 29 14:45:36 in-debbld-33 >kernel: >[1724399.196945] (type=20185106, lower=1, node=0xd1045013, port=2177385505, key=2177385506) Jun 29 14:45:36 in-debbld-33 >kernel: >[1724399.196948] Unable to remove publication from failed node Jun 29 >14:45:36 in-debbld-33 kernel: [1724399.196948] (type=20185106, lower=1, node=0xd1045013, port=2177385505, key=2177385506) >Jun 29 14:45:36 >in-debbld-33 kernel: [1724399.196954] Unable to remove publication from failed node Jun 29 14:45:36 in-debbld-33 kernel: >[1724399.196954] (type=20185106, lower=1, node=0xd1045013, port=2177385505, key=2177385506)* > >############################################################################################################### >###### > > > >Any idea why this would be happening ? Memory error could cause this. I observed the same thing in my system: " 2024-05-27T07:27:22.747+02:00 kernel:[425219.492047] {...}[Hardware Error]: Hardware error from APEI Generic Hardware Error Source: 0 ... 2024-05-27T07:27:22.747+02:00 kernel:[425219.517622] {...}[Hardware Error]: section_type: memory error ... 2024-05-27T07:27:22.973+02:00 kernel:[425221.704107] tipc: Unable to remove publication from failed node 2024-05-27T07:27:22.973+02:00 kernel:[425221.704107] (type=143322, lower=9, node=0x1001009, port=1295954722, key=1295954723) 2024-05-27T07:27:22.973+02:00 kernel:[425221.717912] tipc: Unable to remove publication from failed node 2024-05-27T07:27:22.973+02:00 kernel:[425221.717912] (type=143322, lower=9, node=0x1001009, port=1295954722, key=1295954723) " |
From: prakash b. <ps1...@gm...> - 2024-07-05 06:20:53
|
Hi All, We have been using TIPC sockets for a while now. Off late we are facing issues randomly where the CPU is frozen. The TIPC module seems to be entering in an infinite loop while trying to remove publication from a failed node. This is resulting in log flooding. Almost 50 GB worth of logs written in 1-2 hours. This also seems to be hogging the CPU resulting in the non-responsive system. We had to do a hard shutdown to recover from it. Below are the logs snippets from the kernel log file. ##################################################################################################################### *Jun 29 14:45:36 in-debbld-33 kernel: [1724399.196945] Unable to remove publication from failed node Jun 29 14:45:36 in-debbld-33 kernel: [1724399.196945] (type=20185106, lower=1, node=0xd1045013, port=2177385505, key=2177385506) Jun 29 14:45:36 in-debbld-33 kernel: [1724399.196948] Unable to remove publication from failed node Jun 29 14:45:36 in-debbld-33 kernel: [1724399.196948] (type=20185106, lower=1, node=0xd1045013, port=2177385505, key=2177385506) Jun 29 14:45:36 in-debbld-33 kernel: [1724399.196954] Unable to remove publication from failed node Jun 29 14:45:36 in-debbld-33 kernel: [1724399.196954] (type=20185106, lower=1, node=0xd1045013, port=2177385505, key=2177385506)* ##################################################################################################################### Any idea why this would be happening ? Would appreciate any help. *Kernel version : 4.19.0-26-amd64* Thanks, Prakash |
From: Duzan, G. D <Gar...@fi...> - 2024-05-21 19:27:28
|
FWIW, my code changes did manage to fix the problems I was having with idempotent datagram requests, where I could retransmit either way safely. Things got messier for a non-idempotent service with either dropped datagrams or dropped connections due to the link reset. So far, the only thing I have found to mitigate the problem is increasing the UDP media/bearer/link tolerance from 1500 to 5000. At least the tests I have run so far seem happier with it. Gary Duzan IT Architect Senior GT.M Core Team ________________________________ From: Duzan, Gary D via tipc-discussion <tip...@li...> Sent: Thursday, May 16, 2024 4:46 PM To: Tung Quang Nguyen <tun...@de...>; tip...@li... <tip...@li...> Subject: Re: [tipc-discussion] Tuning/Debugging of "Retransmission failure" Tung Quang Nguyen 5/15/2024 11:56 PM Not sure what you mean by "pushing TIPC a bit". If it means dropping TIPC messages, then "Retransmission failure" is expected. It just means that I increased the amount of traffic across TIPC in the cluster. I only noticed because it appeared from the application level that messages were being dropped. I do have TIPC_DEST_DROPPABLE and TIPC_SRC_DROPPABLE set to zero, but I just realized that I only have the TIPC_ERRINFO handling on one end. I should fix that. Tung Quang Nguyen 5/15/2024 11:56 PM If you did not intentionally drop TIPC messages, then issue could be due to packet drop at NIC (in your VM node or host or Switch/Router etc.). You need to do the tunning at NIC. Running some more tests, it does appear that my clusters of larger servers with eth bearers are not encountering this issue, but my clusters of smaller servers with udp bearers are. There is also some disparity in net.tipc.tipc_rmem settings, so I should address that. So it looks like I have some things to try. I'll follow up if that doesn't address the problem. Thanks. Gary Duzan IT Architect Senior GT.M Core Team ________________________________ From: Tung Quang Nguyen <tun...@de...> Sent: Wednesday, May 15, 2024 11:56 PM To: Duzan, Gary D <Gar...@fi...>; tip...@li... <tip...@li...> Subject: RE: Tuning/Debugging of "Retransmission failure" > I've started to notice messages like this when pushing TIPC a bit: Not sure what you mean by "pushing TIPC a bit". If it means dropping TIPC messages, then "Retransmission failure" is expected. >Is there any tuning I can do to avoid the problem, or other data to collect to better understand it? > If you did not intentionally drop TIPC messages, then issue could be due to packet drop at NIC (in your VM node or host or Switch/Router etc.). You need to do the tunning at NIC. The information contained in this message is proprietary and/or confidential. If you are not the intended recipient, please: (i) delete the message and all copies; (ii) do not disclose, distribute or use the message in any manner; and (iii) notify the sender immediately. In addition, please be aware that any message addressed to our domain is subject to archiving and review by persons other than the intended recipient. Thank you. Message Encrypted via TLS connection _______________________________________________ tipc-discussion mailing list tip...@li... https://eur02.safelinks.protection.outlook.com/?url=https%3A%2F%2Flists.sourceforge.net%2Flists%2Flistinfo%2Ftipc-discussion&data=05%7C02%7Cgary.duzan%40fisglobal.com%7Cd4e9b48e25094d3a586508dc75e962c0%7Ce3ff91d834c84b15a0b418910a6ac575%7C0%7C0%7C638514892444800272%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C0%7C%7C%7C&sdata=C%2F7M%2BcPUuQlPRZIQd7k7OicwrZFtRG6vYK9aFKyfP50%3D&reserved=0<https://lists.sourceforge.net/lists/listinfo/tipc-discussion> The information contained in this message is proprietary and/or confidential. If you are not the intended recipient, please: (i) delete the message and all copies; (ii) do not disclose, distribute or use the message in any manner; and (iii) notify the sender immediately. In addition, please be aware that any message addressed to our domain is subject to archiving and review by persons other than the intended recipient. Thank you. Message Encrypted via TLS connection |
From: Duzan, G. D <Gar...@fi...> - 2024-05-16 20:46:54
|
Tung Quang Nguyen 5/15/2024 11:56 PM Not sure what you mean by "pushing TIPC a bit". If it means dropping TIPC messages, then "Retransmission failure" is expected. It just means that I increased the amount of traffic across TIPC in the cluster. I only noticed because it appeared from the application level that messages were being dropped. I do have TIPC_DEST_DROPPABLE and TIPC_SRC_DROPPABLE set to zero, but I just realized that I only have the TIPC_ERRINFO handling on one end. I should fix that. Tung Quang Nguyen 5/15/2024 11:56 PM If you did not intentionally drop TIPC messages, then issue could be due to packet drop at NIC (in your VM node or host or Switch/Router etc.). You need to do the tunning at NIC. Running some more tests, it does appear that my clusters of larger servers with eth bearers are not encountering this issue, but my clusters of smaller servers with udp bearers are. There is also some disparity in net.tipc.tipc_rmem settings, so I should address that. So it looks like I have some things to try. I'll follow up if that doesn't address the problem. Thanks. Gary Duzan IT Architect Senior GT.M Core Team ________________________________ From: Tung Quang Nguyen <tun...@de...> Sent: Wednesday, May 15, 2024 11:56 PM To: Duzan, Gary D <Gar...@fi...>; tip...@li... <tip...@li...> Subject: RE: Tuning/Debugging of "Retransmission failure" > I've started to notice messages like this when pushing TIPC a bit: Not sure what you mean by "pushing TIPC a bit". If it means dropping TIPC messages, then "Retransmission failure" is expected. >Is there any tuning I can do to avoid the problem, or other data to collect to better understand it? > If you did not intentionally drop TIPC messages, then issue could be due to packet drop at NIC (in your VM node or host or Switch/Router etc.). You need to do the tunning at NIC. The information contained in this message is proprietary and/or confidential. If you are not the intended recipient, please: (i) delete the message and all copies; (ii) do not disclose, distribute or use the message in any manner; and (iii) notify the sender immediately. In addition, please be aware that any message addressed to our domain is subject to archiving and review by persons other than the intended recipient. Thank you. Message Encrypted via TLS connection |
From: Tung Q. N. <tun...@de...> - 2024-05-16 04:12:51
|
> I've started to notice messages like this when pushing TIPC a bit: Not sure what you mean by "pushing TIPC a bit". If it means dropping TIPC messages, then "Retransmission failure" is expected. >Is there any tuning I can do to avoid the problem, or other data to collect to better understand it? > If you did not intentionally drop TIPC messages, then issue could be due to packet drop at NIC (in your VM node or host or Switch/Router etc.). You need to do the tunning at NIC. |
From: Duzan, G. D <Gar...@fi...> - 2024-05-15 18:51:27
|
In case it matters, I'm seeing the same thing on older kernels, too. May 15 11:58:11 rhel04 kernel: tipc: Retransmission failure on link <qqqqqqqqqqqq:rhel04-wwwwwwwwwwww:ubu04> May 15 11:58:11 rhel04 kernel: tipc: State of link Link <qqqqqqqqqqqq:rhel04-wwwwwwwwwwww:ubu04> state e May 15 11:58:11 rhel04 kernel: tipc: XMTQ: 45 [40360-40425], BKLGQ: 0, SNDNX: 40426, RCVNX: 37551 May 15 11:58:11 rhel04 kernel: tipc: Failed msg: usr 12, typ 1, len 13928, err 0 May 15 11:58:11 rhel04 kernel: tipc: sqno 40360, prev: be5645f1, dest: 8356ce57 May 15 11:58:11 rhel04 kernel: tipc: retr_stamp -1694829503, retr_cnt 110 Linux rhel04 4.18.0-513.24.1.el8_9.x86_64 #1 SMP Thu Mar 14 14:20:09 EDT 2024 x86_64 x86_64 x86_64 GNU/Linux Linux ubu04 5.4.0-176-generic #196-Ubuntu SMP Fri Mar 22 16:46:39 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux Gary Duzan IT Architect Senior GT.M Core Team E: gar...@fi... FIS | Advancing the way the world pays, banks and invests™ ________________________________ From: Duzan, Gary D Sent: Wednesday, May 15, 2024 1:07 PM To: tip...@li... <tip...@li...> Subject: Tuning/Debugging of "Retransmission failure" I've started to notice messages like this when pushing TIPC a bit: [Wed May 15 11:27:32 2024] tipc: Retransmission failure on link <xxxxxxxxxxxx:deb03-yyyyyyyyyyyy:ubu06> [Wed May 15 11:27:32 2024] tipc: State of link Link <xxxxxxxxxxxx:deb03-yyyyyyyyyyyy:ubu06> state e [Wed May 15 11:27:32 2024] tipc: XMTQ: 15 [30965-31149], BKLGQ: 0, SNDNX: 31150, RCVNX: 27685 [Wed May 15 11:27:32 2024] tipc: Failed msg: usr 12, typ 0, len 13928, err 0 [Wed May 15 11:27:32 2024] tipc: sqno 30965, prev: ace568c, dest: ad7f568c [Wed May 15 11:27:32 2024] tipc: retr_stamp -1697900680, retr_cnt 73 The host reporting the problem is running: Linux vlltcgtmdeb03 6.7.9-amd64 #1 SMP PREEMPT_DYNAMIC Debian 6.7.9-2 (2024-03-13) x86_64 GNU/Linux and the other end of the link is: Linux vlltcgtmubu06 6.5.0-27-generic #28-Ubuntu SMP PREEMPT_DYNAMIC Thu Mar 7 18:21:00 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux Is there any tuning I can do to avoid the problem, or other data to collect to better understand it? Thanks. Gary Duzan IT Architect Senior GT.M Core Team E: gar...@fi... FIS | Advancing the way the world pays, banks and invests™ The information contained in this message is proprietary and/or confidential. If you are not the intended recipient, please: (i) delete the message and all copies; (ii) do not disclose, distribute or use the message in any manner; and (iii) notify the sender immediately. In addition, please be aware that any message addressed to our domain is subject to archiving and review by persons other than the intended recipient. Thank you. Message Encrypted via TLS connection |
From: Duzan, G. D <Gar...@fi...> - 2024-05-15 18:40:46
|
I've started to notice messages like this when pushing TIPC a bit: [Wed May 15 11:27:32 2024] tipc: Retransmission failure on link <xxxxxxxxxxxx:deb03-yyyyyyyyyyyy:ubu06> [Wed May 15 11:27:32 2024] tipc: State of link Link <xxxxxxxxxxxx:deb03-yyyyyyyyyyyy:ubu06> state e [Wed May 15 11:27:32 2024] tipc: XMTQ: 15 [30965-31149], BKLGQ: 0, SNDNX: 31150, RCVNX: 27685 [Wed May 15 11:27:32 2024] tipc: Failed msg: usr 12, typ 0, len 13928, err 0 [Wed May 15 11:27:32 2024] tipc: sqno 30965, prev: ace568c, dest: ad7f568c [Wed May 15 11:27:32 2024] tipc: retr_stamp -1697900680, retr_cnt 73 The host reporting the problem is running: Linux vlltcgtmdeb03 6.7.9-amd64 #1 SMP PREEMPT_DYNAMIC Debian 6.7.9-2 (2024-03-13) x86_64 GNU/Linux and the other end of the link is: Linux vlltcgtmubu06 6.5.0-27-generic #28-Ubuntu SMP PREEMPT_DYNAMIC Thu Mar 7 18:21:00 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux Is there any tuning I can do to avoid the problem, or other data to collect to better understand it? Thanks. Gary Duzan IT Architect Senior GT.M Core Team E: gar...@fi... FIS | Advancing the way the world pays, banks and invests™ The information contained in this message is proprietary and/or confidential. If you are not the intended recipient, please: (i) delete the message and all copies; (ii) do not disclose, distribute or use the message in any manner; and (iii) notify the sender immediately. In addition, please be aware that any message addressed to our domain is subject to archiving and review by persons other than the intended recipient. Thank you. Message Encrypted via TLS connection |
From: Tung Q. N. <tun...@de...> - 2024-05-02 00:58:32
|
>Subject: [PATCH net] tipc: fix a possible memleak in tipc_buf_append > >__skb_linearize() doesn't free the skb when it fails, so move '*buf = NULL' after __skb_linearize(), so that the skb can be freed on the >err path. > >Fixes: b7df21cf1b79 ("tipc: skb_linearize the head skb when reassembling msgs") >Reported-by: Paolo Abeni <pa...@re...> >Signed-off-by: Xin Long <luc...@gm...> >--- > net/tipc/msg.c | 2 +- > 1 file changed, 1 insertion(+), 1 deletion(-) > >diff --git a/net/tipc/msg.c b/net/tipc/msg.c index 5c9fd4791c4b..c52ab423082c 100644 >--- a/net/tipc/msg.c >+++ b/net/tipc/msg.c >@@ -142,9 +142,9 @@ int tipc_buf_append(struct sk_buff **headbuf, struct sk_buff **buf) > if (fragid == FIRST_FRAGMENT) { > if (unlikely(head)) > goto err; >- *buf = NULL; > if (skb_has_frag_list(frag) && __skb_linearize(frag)) > goto err; >+ *buf = NULL; > frag = skb_unshare(frag, GFP_ATOMIC); > if (unlikely(!frag)) > goto err; >-- >2.43.0 Reviewed-by: Tung Nguyen <tun...@de...> |
From: Xin L. <luc...@gm...> - 2024-04-30 14:03:46
|
__skb_linearize() doesn't free the skb when it fails, so move '*buf = NULL' after __skb_linearize(), so that the skb can be freed on the err path. Fixes: b7df21cf1b79 ("tipc: skb_linearize the head skb when reassembling msgs") Reported-by: Paolo Abeni <pa...@re...> Signed-off-by: Xin Long <luc...@gm...> --- net/tipc/msg.c | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/net/tipc/msg.c b/net/tipc/msg.c index 5c9fd4791c4b..c52ab423082c 100644 --- a/net/tipc/msg.c +++ b/net/tipc/msg.c @@ -142,9 +142,9 @@ int tipc_buf_append(struct sk_buff **headbuf, struct sk_buff **buf) if (fragid == FIRST_FRAGMENT) { if (unlikely(head)) goto err; - *buf = NULL; if (skb_has_frag_list(frag) && __skb_linearize(frag)) goto err; + *buf = NULL; frag = skb_unshare(frag, GFP_ATOMIC); if (unlikely(!frag)) goto err; -- 2.43.0 |
From: Tung Q. N. <tun...@de...> - 2024-02-20 07:19:36
|
>1. As TIPC creates a UDP socket for a bearer and multiple TIPC sockets are kind of multiplexed into one UDP socket. Can it slow the >performance when the incoming traffic is high? I do not see any performance down in my test when using UDP bearer. >2. Can we increase the buffer size of the TIPC UDP socket alone without changing the system level default receive buffer size ? > >I would like to increase the socket buffer size to reduce the packet drops and be efficient but don't want to change the default value at >the system level. There is no packet drop at TIPC socket level when using SOCK_SEQPACKET. You do not need to do anything. |
From: prakash b. <ps1...@gm...> - 2024-02-19 13:41:20
|
Hi John/Xin,Tung, I hope you guys are doing well. I need some suggestions about using a TIPC socket for the "SOCK_SEQPACKET" type while using a UDP bearer. There is a use case where the incoming traffic is very high for a short period intermittently. 1. As TIPC creates a UDP socket for a bearer and multiple TIPC sockets are kind of multiplexed into one UDP socket. Can it slow the performance when the incoming traffic is high? 2. Can we increase the buffer size of the TIPC UDP socket alone without changing the system level default receive buffer size ? I would like to increase the socket buffer size to reduce the packet drops and be efficient but don't want to change the default value at the system level. Please advise what is recommended in this case. Kernel version : 4.19.81 Socket type : SOCK_SEQPACKET Would appreciate any help. Thanks, Prakash |
From: Tung Q. N. <tun...@de...> - 2024-02-16 06:13:10
|
>As part of the shutdown, we are terminating all the processes using SIGKILL. We are expecting the tipc sockets to be closed >automatically by the kernel after some time. > >But sometimes ‘tipc socket list’ is still showing a few sockets as alive. > >Now when we restart the application, the system has two sockets with the same tipc address. > Did your check Gary Duzan's comment ? " If the program is forking off processes, perhaps the child processes aren't closing the socket file descriptor. Using fcntl() to set FD_CLOEXEC on the descriptor in the parent may help." Another thing is you need to make sure that your processes are actually killed (not hung or become zombie). > >Isn't the tipc sockets should have been closed automatically by the kernel once the application is killed ? > They are closed if application is actually killed. |
From: prakash b. <ps1...@gm...> - 2024-02-15 14:53:08
|
Hi John/Xin,Tung, We are facing an issue while closing the TIPC server socket. We are running multiple applications which are in turn creating TIPC sockets. As part of the shutdown, we are terminating all the processes using SIGKILL. We are expecting the tipc sockets to be closed automatically by the kernel after some time. But sometimes ‘tipc socket list’ is still showing a few sockets as alive. Now when we restart the application, the system has two sockets with the same tipc address. Isn't the tipc sockets should have been closed automatically by the kernel once the application is killed ? Kernel version : 4.19.81 Socket type : SOCK_SEQPACKET Would appreciate any help. Thanks, Prakash |
From: Tung Q. N. <tun...@de...> - 2024-02-04 15:34:30
|
> net/tipc/Makefile | 4 ++-- > 1 file changed, 2 insertions(+), 2 deletions(-) > >diff --git a/net/tipc/Makefile b/net/tipc/Makefile index ee49a9f1dd4f..18e1636aa036 100644 >--- a/net/tipc/Makefile >+++ b/net/tipc/Makefile >@@ -18,5 +18,5 @@ tipc-$(CONFIG_TIPC_MEDIA_IB) += ib_media.o > tipc-$(CONFIG_SYSCTL) += sysctl.o > tipc-$(CONFIG_TIPC_CRYPTO) += crypto.o > >- >-obj-$(CONFIG_TIPC_DIAG) += diag.o >+obj-$(CONFIG_TIPC_DIAG) += tipc_diag.o >+tipc_diag-y += diag.o >-- >2.39.1 > Reviewed-by: Tung Nguyen <tun...@de...> |
From: Xin L. <luc...@gm...> - 2024-02-02 20:11:17
|
It is not appropriate for TIPC to use "diag" as its diag module name while the other protocols are using "$(protoname)_diag" like tcp_diag, udp_diag and sctp_diag etc. So this patch is to rename diag.ko to tipc_diag.ko in tipc's Makefile. Signed-off-by: Xin Long <luc...@gm...> --- net/tipc/Makefile | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/net/tipc/Makefile b/net/tipc/Makefile index ee49a9f1dd4f..18e1636aa036 100644 --- a/net/tipc/Makefile +++ b/net/tipc/Makefile @@ -18,5 +18,5 @@ tipc-$(CONFIG_TIPC_MEDIA_IB) += ib_media.o tipc-$(CONFIG_SYSCTL) += sysctl.o tipc-$(CONFIG_TIPC_CRYPTO) += crypto.o - -obj-$(CONFIG_TIPC_DIAG) += diag.o +obj-$(CONFIG_TIPC_DIAG) += tipc_diag.o +tipc_diag-y += diag.o -- 2.39.1 |
From: Tung Q. N. <tun...@de...> - 2023-11-23 07:47:56
|
> net/tipc/node.c | 10 ++++++---- > 1 file changed, 6 insertions(+), 4 deletions(-) > >diff --git a/net/tipc/node.c b/net/tipc/node.c index 3105abe97bb9..2a036b8a7da3 100644 >--- a/net/tipc/node.c >+++ b/net/tipc/node.c >@@ -2154,14 +2154,15 @@ void tipc_rcv(struct net *net, struct sk_buff *skb, struct tipc_bearer *b) > /* Receive packet directly if conditions permit */ > tipc_node_read_lock(n); > if (likely((n->state == SELF_UP_PEER_UP) && (usr != TUNNEL_PROTOCOL))) { >+ tipc_node_read_unlock(n); > spin_lock_bh(&le->lock); > if (le->link) { > rc = tipc_link_rcv(le->link, skb, &xmitq); > skb = NULL; > } > spin_unlock_bh(&le->lock); >- } >- tipc_node_read_unlock(n); >+ } else >+ tipc_node_read_unlock(n); > > /* Check/update node state before receiving */ > if (unlikely(skb)) { >@@ -2169,12 +2170,13 @@ void tipc_rcv(struct net *net, struct sk_buff *skb, struct tipc_bearer *b) > goto out_node_put; > tipc_node_write_lock(n); > if (tipc_node_check_state(n, skb, bearer_id, &xmitq)) { >+ tipc_node_write_unlock(n); > if (le->link) { > rc = tipc_link_rcv(le->link, skb, &xmitq); > skb = NULL; > } >- } >- tipc_node_write_unlock(n); >+ } else >+ tipc_node_write_unlock(n); > } > > if (unlikely(rc & TIPC_LINK_UP_EVT)) >-- >2.15.2 > > This patch is wrong. le->link and link status must be protected by node lock. See what happens if tipc_node_timeout() is called, and the link goes down: tipc_node_timeout() tipc_node_link_down() { struct tipc_link *l = le->link; ... if (delete) { kfree(l); le->link = NULL; } ... } |