在程序里禁用了 Nagle 算法还会出现 200ms 延迟么 - V2EX
V2EX = way to explore
V2EX 是一个关于分享和探索的地方
现在注册
已注册用户请  登录
请不要在回答技术问题时复制粘贴 AI 生成的内容
flewover
V2EX    程序员

在程序里禁用了 Nagle 算法还会出现 200ms 延迟么

  •  2
     
  •   flewover 2015-05-20 17:59:03 +08:00 5891 次点击
    这是一个创建于 3801 天前的主题,其中的信息可能已经有所发展或是发生改变。

    如题

    简化的程序是在两个不同的 golang 程序建立一条 TCP 连接:发送端按照一定频率发送数据包(数据包最大 2048 字节,也存在少量较小的数据包);接收端接收数据伪代码如下

    for { rd := bufio.NewReaderSize(conn, bufsz) rlen, err := rd.Read(rbuf[:]) // handle rbuf } 

    建立 TCP 连接时发送端和接收端都默认禁用了 Nagle 算法(golang 默认 c.SetNoDelay(true)),但是测试中发现有时发送端的 Send-Q 堆积了很多数据,接收端接收数据很慢,大量 Read 出现 200ms 延迟,并且此时发送端输出端的 CPU 占用很低,远低于正常网络数据读写的情况。

    2015-05-20 17:34:44.418 [INFO] ac03 [processPushMsg] send 39 direct msg to connector conn0101 latency: 25270.388 2015-05-20 17:34:44.418 [DEBU] ac04 [handleTCPConn] before get new bufio reader 2015-05-20 17:34:44.418 [DEBU] ac05 [handleTCPConn] before read 2015-05-20 17:34:44.621 [DEBU] ac06 [handleTCPConn] end read 2015-05-20 17:34:44.621 [INFO] ac07 [processPushMsg] send 81 direct msg to connector conn0101 latency: 25474.038 2015-05-20 17:34:44.621 [INFO] ac08 [processPushMsg] send 38 direct msg to connector conn0101 latency: 25474.143 2015-05-20 17:34:44.622 [INFO] ac09 [processPushMsg] send 39 direct msg to connector conn0101 latency: 25374.197 2015-05-20 17:34:44.622 [DEBU] ac0a [handleTCPConn] before get new bufio reader 2015-05-20 17:34:44.622 [DEBU] ac0b [handleTCPConn] before read 2015-05-20 17:34:44.622 [DEBU] ac0c [handleTCPConn] end read 2015-05-20 17:34:44.622 [DEBU] ac0d [handleTCPConn] before get new bufio reader 2015-05-20 17:34:44.622 [DEBU] ac0e [handleTCPConn] before read 2015-05-20 17:34:44.825 [DEBU] ac0f [handleTCPConn] end read 2015-05-20 17:34:44.825 [INFO] ac10 [processPushMsg] send 81 direct msg to connector conn0101 latency: 25578.005 2015-05-20 17:34:44.825 [INFO] ac11 [processPushMsg] send 39 direct msg to connector conn0101 latency: 25578.103 

    我的理解是发送端数据已经发送到本地网络缓冲区,等待对端读取,但是不清楚接收端为什么延迟 200ms 读取。不知从何入手解决。

    第 1 条附言    2015-05-20 18:44:11 +08:00
    s/并且此时发送端输出端的 CPU 占用很低/并且此时发送端和接收端的 CPU 占用很低

    这种现象是随机出现的,有时就会正常的收发数据;并且一旦出现该现象,之后的数据收发,即使发送数据量很小,也会存在 200ms 延迟
    14 条回复    2015-05-21 21:35:05 +08:00
    zhicheng
        1
    zhicheng  
       2015-05-20 18:47:40 +08:00 via Android   1
    你知道拥塞算法吗?
    flewover
        2
    flewover  
    OP
       2015-05-20 18:56:53 +08:00
    @zhicheng 模糊的知道一些概念,但是不知道具体与哪些内核参数相关,如何调优,请问有哪些监控指标可以反映网络拥塞状况?还是需要应用层做流量控制?
    simomo2010
        3
    simomo2010  
       2015-05-20 20:02:17 +08:00   1
    @flewover
    你这跟tcp栈没关系吧~ 看log
    2015-05-20 17:34:44.418 [DEBU] ac05 [handleTCPConn] before read
    2015-05-20 17:34:44.621 [DEBU] ac06 [handleTCPConn] end read
    这两条记录中间应该是
    rlen, err := rd.Read(rbuf[:])
    吧?

    我不懂golang的这个接口, 猜想bufferReader在等数据填满它然后返回或者超时(200ms)返回。所以你发包小的时候,它总是需要超时才能返回~ 于是就有了200ms的延迟
    flewover
        4
    flewover  
    OP
       2015-05-20 20:29:06 +08:00
    @simomo2010

    那两条日志确实是在 rlen, err := rd.Read(rbuf[:]) 的前后,但是这个 read 接口只调用一次 reader 的 rea,产生延迟应该不是需要填满 buf 引发的。

    // Read reads data into p.
    // It returns the number of bytes read into p.
    // It calls Read at most once on the underlying Reader,
    // hence n may be less than len(p).
    // At EOF, the count will be zero and err will be io.EOF.
    func (b *Reader) Read(p []byte) (n int, err error) {
    ...
    }
    simomo2010
        5
    simomo2010  
       2015-05-20 21:24:40 +08:00
    @flewover
    看文档确实不会超时~~

    那或许需要wireshark之类的东西截个包看看~
    ryd994
        6
    ryd994  
       2015-05-20 23:15:46 +08:00 via Android   1
    不需要禁用啊,加nodelay就好
    https://access.redhat.com/documentation/en-US/Red_Hat_Enterprise_MRG/1.2/html/Realtime_Tuning_Guide/sect-Realtime_Tuning_Guide-Application_Tuning_and_Deployment-TCP_NODELAY_and_Small_Buffer_Writes.html
    wireshark截包没有用,因为无论用不用nagle,出来的都是合法的tcp包,大小不同而已。
    flewover
        7
    flewover  
    OP
       2015-05-20 23:51:37 +08:00
    @ryd994 说的禁用 Nagle 就是调用了 c.SetNoDelay(true)
    flewover
        8
    flewover  
    OP
       2015-05-20 23:58:21 +08:00
    用 strace 查看了系统调用,接收端 200ms 出现在 epoll_wait

    [<... epoll_wait resumed> {{EPOLLIN|EPOLLOUT, {u32=1626358040, u64=140674690203928}}}, 128, -1) = 1 <0.202147>
    ]

    接收端日志:
    2015-05-20 23:40:19.629 [INFO] 54f99 [processPushMsg] send 81 direct msg to connector conn0101 latency: 697548.524
    2015-05-20 23:40:19.629 [DEBU] 54f9a [handleTCPConn] before get new bufio reader
    2015-05-20 23:40:19.630 [DEBU] 54f9b [handleTCPConn] before read
    2015-05-20 23:40:19.834 [DEBU] 54f9c [handleTCPConn] end read rlen: 4068
    2015-05-20 23:40:19.834 [INFO] 54f9d [processPushMsg] send 39 direct msg to connector conn0101 latency: 697753.906


    strace 日志(sudo strace -f -v -x -tt -T -p 92324 -o strace_recv.log):

    94951 23:40:19.629828 write(3, "2015-05-20 23:40:19.629 [INFO] 5"..., 115) = 115 <0.000038>
    94951 23:40:19.629958 clock_gettime(CLOCK_REALTIME, {1432136419, 629994695}) = 0 <0.000031>
    94951 23:40:19.630103 clock_gettime(CLOCK_REALTIME, {1432136419, 630143536}) = 0 <0.000032>
    94951 23:40:19.630219 write(3, "2015-05-20 23:40:19.629 [DEBU] 5"..., 81) = 81 <0.000036>
    94951 23:40:19.630337 clock_gettime(CLOCK_REALTIME, {1432136419, 630373340}) = 0 <0.000030>
    94951 23:40:19.630479 clock_gettime(CLOCK_REALTIME, {1432136419, 630518521}) = 0 <0.000031>
    94951 23:40:19.630600 write(3, "2015-05-20 23:40:19.630 [DEBU] 5"..., 65) = 65 <0.000036>
    94951 23:40:19.630717 read(9, 0xc20805f01c, 4068) = -1 EAGAIN (Resource temporarily unavailable) <0.000033>
    94951 23:40:19.630833 write(8, "\x01\x03\xf1\x00\x01\x00\x0c\x55\x5c\xa8\x4d\x62\xbb\x97\x68\xde\x01\xd1\x12\x02\x00\x10\x31\x34\x33\x32\x31\x33\x35\x37\x32\x32"..., 1013) = 1013 <0.000103>
    94951 23:40:19.631027 write(8, "\x01\x01\xf9\x00\x03\x01\xd4\x55\x5c\x9c\x2a\x62\xbb\x97\x68\x8c\x00\x02\xa9\x55\x5c\x9c\x2a\x62\xbb\x97\x68\x8c\x00\x02\xfc\x55"..., 509) = 509 <0.000067>
    94951 23:40:19.631187 write(8, "\x01\x03\xf1\x00\x03\x03\xcc\x55\x5c\x9c\x29\x62\xbb\x97\x68\x8c\x00\x01\x3d\x55\x5c\x9c\x2a\x62\xbb\x97\x68\x8c\x00\x02\x93\x55"..., 1013) = 1013 <0.000058>
    94951 23:40:19.631332 epoll_wait(6, {}, 128, 0) = 0 <0.000035>
    94951 23:40:19.631440 epoll_wait(6, <unfinished ...>
    92326 23:40:19.631496 <... futex resumed> ) = 0 <0.205872>
    92326 23:40:19.631554 clock_gettime(CLOCK_MONOTONIC, {1298109, 942247030}) = 0 <0.000033>
    92326 23:40:19.631664 clock_gettime(CLOCK_REALTIME, {1432136419, 631698624}) = 0 <0.000031>
    92326 23:40:19.631767 select(0, NULL, NULL, NULL, {0, 20} <unfinished ...>
    92324 23:40:19.631845 <... futex resumed> ) = 0 <0.206158>
    92324 23:40:19.631911 epoll_wait(6, {}, 128, 0) = 0 <0.000064>
    92326 23:40:9.632048 <... select resumed> ) = 0 (Timeout) <0.000248>
    92326 23:40:19.632112 clock_gettime(CLOCK_MONOTONIC, {1298109, 942807377}) = 0 <0.000031>
    92326 23:40:19.632215 futex(0x8f4de0, FUTEX_WAIT, 0, {60, 0} <unfinished ...>
    92324 23:40:19.632279 futex(0x8f5838, FUTEX_WAIT, 0, NULL <unfinished ...>
    94951 23:40:19.833623 <... epoll_wait resumed> {{EPOLLIN|EPOLLOUT, {u32=1626358040, u64=140674690203928}}}, 128, -1) = 1 <0.202147>
    94951 23:40:19.833780 clock_gettime(CLOCK_MONOTONIC, {1298110, 144480859}) = 0 <0.000032>
    94951 23:40:19.833902 futex(0x8f5838, FUTEX_WAKE, 1) = 1 <0.000044>
    94951 23:40:19.834024 futex(0x8f4de0, FUTEX_WAKE, 1) = 1 <0.000042>
    94951 23:40:19.834131 read(9, "\x5c\x9c\x29\x62\xbb\x97\x68\x8c\x00\x01\x56\x63\x6f\x6e\x6e\x30\x31\x30\x31\x55\x5c\x9c\x2a\x62\xbb\x97\x68\x8c\x00\x03\x24\x63"..., 4068) = 4068 <0.000039>
    94951 23:40:19.834268 clock_gettime(CLOCK_REALTIME, {1432136419, 834305215}) = 0 <0.000030>
    94951 23:40:19.834476 clock_gettime(CLOCK_REALTIME, {1432136419, 834519537}) = 0 <0.000033>
    94951 23:40:19.834601 write(3, "2015-05-20 23:40:19.834 [DEBU] 5"..., 73) = 73 <0.000042>
    94951 23:40:19.834778 clock_gettime(CLOCK_REALTIME, {1432136419, 834819514}) = 0 <0.000032>
    94951 23:40:19.834896 clock_gettime(CLOCK_REALTIME, {1432136419, 834931088}) = 0 <0.000030>
    94951 23:40:19.835061 clock_gettime(CLOCK_REALTIME, {1432136419, 835102950}) = 0 <0.000033>
    94951 23:40:19.835180 write(3, "2015-05-20 23:40:19.834 [INFO] 5"..., 115) = 115 <0.000037>
    simomo2010
        9
    simomo2010  
       2015-05-21 00:12:56 +08:00
    @flewover
    感觉是发送端23:40:19.833才把包发出来~
    est
        10
    est  
       2015-05-21 00:21:12 +08:00   3
    http://weibo.com/p/1001603821691477346388

    如果是这个问题发个感谢呗。
    ryd994
        11
    ryd994  
       2015-05-21 01:57:57 +08:00 via Android
    @est nice!学习了
    @flewover 抱歉,看漏了
    simomo2010
        12
    simomo2010  
       2015-05-21 10:08:13 +08:00
    @est 这是计算RTT(超时重传时间)的方法,似乎也有可能,sender第一次发包,接受者没收到,然后超时重传了一遍,接受者就收到了,但是这个情况偶尔一次 两次还可能,可lz说了“并且一旦出现该现象,之后的数据收发,即使发送数据量很小,也会存在 200ms 延迟”,感觉不像啊~

    @flewover 还是截个包最清楚,debug 网络的issue,截包是最直观的方法~
    flewover
        13
    flewover  
    OP
       2015-05-21 10:11:15 +08:00
    @est 很有帮助 :)
    @simomo2010 我再跟踪一下看个究竟
    benjiam
       
    benjiam  
       2015-05-21 21:35:05 +08:00 via Android
    但是测试中发现有时发送端的 Send-Q 堆积了很多数据,接收端接收数据很慢,大量 Read 出现 200ms 延迟,
    说明不是客户端问题,server读取超时,更多地问题在于你read后处理问题吧,你简单的读不处理 还能重现吗?网络丢包引起的重传都是很快就能清理掉的,你的recv q有数据也证明不是网线问题。
    关于     帮助文档     自助推广系统     博客     API     FAQ     Solana     2600 人在线   最高记录 6679       Select Language
    创意工作者们的社区
    World is powered by solitude
    VERSION: 3.9.8.5 41ms UTC 15:33 PVG 23:33 LAX 08:33 JFK 11:33
    Do have faith in what you're doing.
    ubao msn snddm index pchome yahoo rakuten mypaper meadowduck bidyahoo youbao zxmzxm asda bnvcg cvbfg dfscv mmhjk xxddc yybgb zznbn ccubao uaitu acv GXCV ET GDG YH FG BCVB FJFH CBRE CBC GDG ET54 WRWR RWER WREW WRWER RWER SDG EW SF DSFSF fbbs ubao fhd dfg ewr dg df ewwr ewwr et ruyut utut dfg fgd gdfgt etg dfgt dfgd ert4 gd fgg wr 235 wer3 we vsdf sdf gdf ert xcv sdf rwer hfd dfg cvb rwf afb dfh jgh bmn lgh rty gfds cxv xcv xcs vdas fdf fgd cv sdf tert sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf sdf shasha9178 shasha9178 shasha9178 shasha9178 shasha9178 liflif2 liflif2 liflif2 liflif2 liflif2 liblib3 liblib3 liblib3 liblib3 liblib3 zhazha444 zhazha444 zhazha444 zhazha444 zhazha444 dende5 dende denden denden2 denden21 fenfen9 fenf619 fen619 fenfe9 fe619 sdf sdf sdf sdf sdf zhazh90 zhazh0 zhaa50 zha90 zh590 zho zhoz zhozh zhozho zhozho2 lislis lls95 lili95 lils5 liss9 sdf0ty987 sdft876 sdft9876 sdf09876 sd0t9876 sdf0ty98 sdf0976 sdf0ty986 sdf0ty96 sdf0t76 sdf0876 df0ty98 sf0t876 sd0ty76 sdy76 sdf76 sdf0t76 sdf0ty9 sdf0ty98 sdf0ty987 sdf0ty98 sdf6676 sdf876 sd876 sd876 sdf6 sdf6 sdf9876 sdf0t sdf06 sdf0ty9776 sdf0ty9776 sdf0ty76 sdf8876 sdf0t sd6 sdf06 s688876 sd688 sdf86