如何进行Linux下双网卡绑定bond的配置

2023-08-28 9847阅读

如何进行Linux下双网卡绑定bond的配置 第1张

如何进行Linux下双网卡绑定bond的配置,相信很多没有经验的人对此束手无策,为此本文总结了问题出现的原因和解决方法,通过这篇文章希望你能解决这个问题。


一、什么是bond?

网卡bond是通过多张网卡绑定为一个逻辑网卡,实现本地

网卡冗余,带宽扩容和负载均衡

,在生产场景中是一种常用的技术。kernels

2.4.12及以后的版本均供bondin开发云主机域名g模块,以前的版本可以通过patch实现。

二、bond的配置实例


1、测试环境

[root@wjq2

~]#

cat

/etc/redhat-release

Red

Hat Enterprise Linux Server release 6.4 (Santiago)[root@wjq2

~]#

uname -r

2.6.32-358.el6.x86_64

2、查看


linux


是否支持


bonding,


大部分发行版都支持

[root@wjq2

~]#

cat

/boot/config-2.6.32-358.el6.x86_64 | grep -i bonding

CONFIG_BONDING=m或者[root@wjq2

~]#

modinfo

bonding | more

filename:

/lib/modules/2.6.32-358.el6.x86_64/kernel/drivers/net/bonding/bonding.koauthor: Thomas Davis, tadavis@lbl.gov and many

othersdescription: Ethernet Channel Bonding Driver, v3.6.0version: 3.6.0license: GPLsrcversion: 990DDFEBDBA03F305236885depends: 8021q,ipv6vermagic: 2.6.32-358.el6.x86_64 SMP mod_unload

modversionsparm: max_bonds:Max number of bonded

devices (int)parm: tx_queues:Max number of transmit

queues (default = 16) (int)parm: num_grat_arp:Number of gratuitous

ARP packets to send on failover event (int)parm: num_unsol_na:Number of unsolicited

IPv6 Neighbor Advertisements packets to send on failover event (int)parm: miimon:Link check interval in

milliseconds (int)parm: updelay:Delay before considering

link up, in milliseconds (int)parm: downdelay:Delay before considering

link down, in milliseconds (int)parm: use_carrier:Use netif_carrier_ok (vs

MII ioctls) in miimon; 0 for off, 1 for on (default) (int)parm:

mode:Mode

of operation; 0 for balance-rr, 1 for active-backup, 2 for balance-xor, 3 for

broadcast, 4 fo


r 802.3ad, 5 for balance-tlb, 6 for

balance-alb

(charp)parm: primary:Primary network device to

use (charp)parm: primary_reselect:Reselect primary

slave once it comes up; 0 for always (default), 1 for only if speed of

primary is better, 2 for only on active slave failure (charp)parm: lacp_rate:LACPDU tx rate to request

from 802.3ad partner; 0 for slow, 1 for fast (charp)parm: ad_select:803.ad aggregation

selection logic; 0 for stable (default), 1 for bandwidth, 2 for count (charp)parm: xmit_hash_policy:balance-xor and

802.3ad hashing method; 0 for layer 2 (default), 1 for layer 3+4, 2 for

layer 2+3 (charp)parm: arp_interval:arp interval in

milliseconds (int)parm: arp_ip_target:arp targets in n.n.n.n

form (array of charp)parm: arp_validate:validate src/dst of ARP

probes; 0 for none (default), 1 for active, 2 for backup, 3 for all

(charp)

如输出以上信息,则说明支持bonding,如果没有,说明内核不支持bonding,需要重新编译内核

3、查看现有网卡的配置信息

[root@wjq2

~]#

ifconfig

eth0 Link encap:Ethernet HWaddr 00:50:56:B5:4F:F5 inet addr:10.1.1.58 Bcast:10.1.1.255 Mask:255.255.255.0 inet6 addr:

fe80::250:56ff:feb5:4ff5/64 Scope:Link UP BROADCAST RUNNING MULTICAST MTU:1500

Metric:1 RX packets:1302 errors:0 dropped:0

overruns:0 frame:0 TX packets:300 errors:0 dropped:0 overruns:0

carrier:0 collisions:0 txqueuelen:1000 RX bytes:105872 (103.3 KiB) TX bytes:38427 (37.5 KiB)eth2 Link encap:Ethernet HWaddr 00:50:56:B5:2D:0F inet addr:10.1.1.59 Bcast:10.1.1.255 Mask:255.255.255.0 inet6 addr: fe80::250:56ff:feb5:2d0f/64

Scope:Link UP BROADCAST RUNNING MULTICAST MTU:1500

Metric:1 RX packets:923 errors:0 dropped:0

overruns:0 frame:0 TX packets:18 errors:0 dropped:0

overruns:0 carrier:0 collisions:0 txqueuelen:1000 RX bytes:69365 (67.7 KiB) TX bytes:1240 (1.2 KiB)lo Link encap:Local Loopback inet addr:127.0.0.1 Mask:255.0.0.0 inet6 addr: ::1/128 Scope:Host UP LOOPBACK RUNNING MTU:16436

Metric:1 RX packets:8 errors:0 dropped:0

overruns:0 frame:0 TX packets:8 errors:0 dropped:0

overruns:0 carrier:0 collisions:0 txqueuelen:0 RX bytes:440 (440.0 b) TX bytes:440 (440.0 b)

4、修改物理网卡配置文件

两个物理网口分别是:eth0,eth2 绑定后的虚拟口是:bond0[root@wjq2

network-scripts]#

vim ifcfg-eth0

DEVICE=eth0TYPE=EthernetUUID=adc59d0e-8c6c-4d03-8f98-d77ef5131208ONBOOT=yesNM_CONTROLLED=noBOOTPROTO=none

MASTER=bond0


SLAVE=yes




#


可以没有此字段,就需要开机执行ifenslavebond0eth0eth2命令了

DEFROUTE=yesIPV4_FAILURE_FATAL=yesIPV6INIT=noNAME=”System

eth0″HWADDR=00:50:56:b5:4f:f5[root@wjq2

network-scripts]#

vim ifcfg-eth2

DEVICE=eth2TYPE=EthernetUUID=3fca3e29-7fd1-4355-86c2-2f76e43f4d9cONBOOT=yesNM_CONTROLLED=noBOOTPROTO=noneHWADDR=00:50:56:b5:2d:0f

MASTER=bond0


SLAVE=yes




#


可以没有此字段,就需要开机执行ifenslavebond0eth0eth2命令了

DEFROUTE=yesIPV4_FAILURE_FATAL=yesIPV6INIT=noNAME=”System

eth2″

5、配置逻辑网卡


bond0

ifcfg-bond0配置文件需要我们手工创建[root@wjq2

network-scripts]#

vim ifcfg-bond0

DEVICE=bond0NAME=’System

bond0’TYPE=EthernetNM_CONTROLLED=noUSERCTL=noONBOOT=yesBOOTPROTO=none

IPADDR=10.1.1.100


NETMASK=255.255.255.0

GATEWAY=10.1.1.249

6、加载模块,让系统支持


bonding

/etc/modprobe.d/modprobe.conf不存在的话,可以手动创建或者使用/etc/modprobe.d/dist.conf文件也可以[root@wjq2

~]#

vim

/etc/modprobe.d/modprobe.conf


alias

bond0 bonding


options

bond0 miimon=100 mode=0


说明

:配置bond0的链路检查时间为100ms,模式为0。(详细的bond七种模式在最后面详细讲解)

7、加载


bond module

[root@wjq2

~]#

modprobe

bonding


8、查看绑定结果

[root@wjq2

~]#

cat

/proc/net/bonding/bond0

Ethernet

Channel Bonding Driver: v3.6.0 (September 26, 2009)Bonding

Mode: load balancing (round-robin)MII

Status:

up

MII

Polling Interval (ms): 100Up

Delay (ms): 0Down

Delay (ms): 0Slave

Interface: eth0MII

Status:

up

Speed:

10000 MbpsDuplex:

fullLink

Failure Count: 0Permanent

HW addr: 00:50:56:b5:4f:f5Slave

queue ID: 0Slave

Interface: eth2MII

Status:

up

Speed:

10000 MbpsDuplex:

fullLink

Failure Count: 0Permanent

HW addr: 00:50:56:b5:2d:0fSlave

queue ID: 0查看物理网卡和逻辑网卡的信息[root@wjq2

~]#

ifconfig


bond0

Link

encap:Ethernet HWaddr

00:50:56:B5:4F:F5 inet addr:

10.1.1.100

Bcast:10.1.1.255 Mask:255.255.255.0 inet6 addr:

fe80::250:56ff:feb5:4ff5/64 Scope:Link UP BROADCAST RUNNING MASTER

MULTICAST MTU:1500 Metric:1 RX packets:903 errors:0 dropped:0

overruns:0 frame:0 TX packets:158 errors:0 dropped:0

overruns:0 carrier:0 collisions:0 txqueuelen:0 RX bytes:68972 (67.3 KiB) TX bytes:17037 (16.6 KiB)eth0 Link encap:Ethernet HWaddr 00:50:56:B5:4F:F5 UP BROADCAST RUNNING SLAVE

MULTICAST MTU:1500 Metric:1 RX packets:452 errors:0 dropped:0

overruns:0 frame:0 TX packets:79 errors:0 dropped:0

overruns:0 carrier:0 collisions:0 txqueuelen:1000 RX bytes:34507 (33.6 KiB) TX bytes:6935 (6.7 KiB)eth2 Link encap:Ethernet HWaddr 00:50:56:B5:4F:F5 UP BROADCAST RUNNING SLAVE

MULTICAST MTU:1500 Metric:1 RX packets:451 errors:0 dropped:0

overruns:0 frame:0 TX packets:79 errors:0 dropped:0

overruns:0 carrier:0 collisions:0 txqueuelen:1000 RX bytes:34465 (33.6 KiB) TX bytes:10102 (9.8 KiB)lo Link encap:Local Loopback inet addr:127.0.0.1 Mask:255.0.0.0 inet6 addr: ::1/128 Scope:Host UP LOOPBACK RUNNING MTU:16436

Metric:1 RX packets:16 errors:0 dropped:0

overruns:0 frame:0 TX packets:16 errors:0 dropped:0 overruns:0

carrier:0 collisions:0 txqueuelen:0 RX bytes:960 (960.0 b) TX bytes:960 (960.0 b)

9、测试


bond

由于使用的是mode=0,负载均衡的方式,这时我们ping114.114.114.114,然后手动断开一个网卡,此时ping不会中断[root@wjq2

~]#

ping

114.114.114.114

PING

114.114.114.开发云主机域名114 (114.114.114.114) 56(84) bytes of data.64

bytes from 114.114.114.114: icmp_seq=1 ttl=59 time=4.46 ms64

bytes from 114.114.114.114: icmp_seq=1 ttl=59 time=4.56 ms (DUP!)64

bytes from 114.114.114.114: icmp_seq=2 ttl=59 time=2.21 ms64

bytes from 114.114.114.114: icmp_seq=2 ttl=59 time=2.28 ms (DUP!)64

bytes from 114.114.114.114: icmp_seq=3 ttl=59 time=2.23 ms64

bytes from 114.114.114.114: icmp_seq=3 ttl=59 time=2.24 ms (DUP!)64

bytes from 114.114.114.114: icmp_seq=4 ttl=59 time=1.94 ms64

bytes from 114.114.114.114: icmp_seq=4 ttl=59 time=1.94 ms (DUP!)64

bytes from 114.114.114.114: icmp_seq=5 ttl=59 time=1.95 ms64

bytes from 114.114.114.114: icmp_seq=5 ttl=59 time=1.95 ms (DUP!)64

bytes from 114.114.114.114: icmp_seq=6 ttl=59 time=2.00 ms64

bytes from 114.114.114.114: icmp_seq=6 ttl=59 time=2.01 ms (DUP!)……用另一个终端手动关闭eth0网卡,ping并没有中断[root@wjq2

~]#

ifdown eth0

64

bytes from 114.114.114.114: icmp_seq=10 ttl=59 time=1.96 ms64

bytes from 114.114.114.114: icmp_seq=11 ttl=59 time=2.12 ms64

bytes from 114.114.114.114: icmp_seq=12 ttl=59 time=2.30 ms64

bytes from 114.114.114.114: icmp_seq=13 ttl=59 time=2.23 ms64

bytes from 114.114.114.114: icmp_seq=14 ttl=59 time=2.16 ms64

bytes from 114.114.114.114: icmp_seq=15 ttl=59 time=2.04 ms64

bytes from 114.114.114.114: icmp_seq=16 ttl=59 time=1.88 ms查看bond0状态,发现eth0没有了,但是bond正常(只显示处于UP状态的网卡信息)[root@wjq2

~]#

cat

/proc/net/bonding/bond0

Ethernet

Channel Bonding Driver: v3.6.0 (September 26, 2009)Bonding

Mode: load balancing (round-robin)MII

Status:

up

MII

Polling Interval (ms): 100Up

Delay (ms): 0Down

Delay (ms): 0Slave

Interface: eth2MII

Status:

up

Speed:

10000 MbpsDuplex:

fullLink

Failure Count: 0Permanent

HW addr: 00:50:56:b5:2d:0fSlave

queue ID: 0然后启动eth0网卡,发现有恢复up状态[root@wjq2

~]#

ifup eth0

[root@wjq2

~]#[root@wjq2

~]#

cat

/proc/net/bonding/bond0

Ethernet

Channel Bonding Driver: v3.6.0 (September 26, 2009)Bonding

Mode: load balancing (round-robin)MII

Status:

up

MII

Polling Interval (ms): 100Up

Delay (ms): 0Down

Delay (ms): 0Slave

Interface: eth0MII

Status:

up

Speed:

10000 MbpsDuplex:

fullLink

Failure Count: 0Permanent

HW addr: 00:50:56:b5:4f:f5Slave

queue ID: 0Slave

Interface: eth2MII

Status:

up

Speed:

10000 MbpsDuplex:

fullLink

Failure Count: 0Permanent

HW addr: 00:50:56:b5:2d:0fSlave

queue ID: 0

三、扩展(多网卡绑定多个


bond




上边是两个网卡(eth0、eth2)绑定成一个bond0,如果我们要设置多个bond口,比如物理网口eth0和eth2组成bond0,eth3和eth4组成bond1,eth5和eth6组成bond2,那么网口设置文件的设置方法和上面,是一样的,只是/etc/modprobe.d/dist.conf文件就不能叠加了。正确的设置方法有两种:


1


、第一种

aliasbond0bondingaliasbond1bondingaliasbond2bondingoptionsbonding

max_bonds=3

miimon=100mode=0这样所有的绑定只能使用一个mode了。



2


、第二种


alias bond0 bondingoptions bond0 miimon=100

mode=0 max_bonds=3

alias bond1 bondingoptions bond1 miimon=100

mode=1 max_bonds=3

alias bond2 bondingoptions bond2 miimon=100

mode=2 max_bonds=3

说明:这种方式不同的bond口可以设定为不同的mode,注意开机自动启动/etc/rc.d/rc.local文件的设置

特别注意:


如果在配置文件的options中没有max_bonds参数,就会出现下面的报错信息:Master ‘bond1’: Error: handshake with driver failed.因为在用modprobe加载bonding模块时,加上参数指定创建多少个bondingdevice,否则默认只会创建一个,就是bond0。



[root@wjq2

~]#

ifenslave

bond1 eth3 eth4



Master ‘bond1’: Error: handshake with driver failed. Aborting

多个bond的结果如下所示:

[root@wjq2 ~]#

ifconfig



bond0

Link encap:Ethernet HWaddr 00:50:56:B5:4F:F5

inet addr:

10.1.1.100

Bcast:10.1.1.255 Mask:255.255.255.0

inet6 addr:

fe80::250:56ff:feb5:4ff5/64 Scope:Link

UP BROADCAST RUNNING MASTER MULTICAST MTU:1500

Metric:1

RX packets:2012 errors:0

dropped:0 overruns:0 frame:0

TX packets:160 errors:0

dropped:0 overruns:0 carrier:0

collisions:0 txqueuelen:0

RX bytes:153964 (150.3

KiB) TX bytes:20786 (20.2 KiB)


bond1


Link encap:Ethernet HWaddr

00:50:56:B5:5A:25

inet addr:

10.1.1.101

Bcast:10.1.1.255 Mask:255.255.255.0

inet6 addr:

fe80::250:56ff:feb5:5a25/64 Scope:Link

UP BROADCAST RUNNING MASTER

MULTICAST MTU:1500 Metric:1

RX packets:1806 errors:0

dropped:0 overruns:0 frame:0

TX packets:8 errors:0 dropped:0

overruns:0 carrier:0

collisions:0 txqueuelen:0

RX bytes:131984 (128.8

KiB) TX bytes:536 (536.0 b)


bond2


Link encap:Ethernet HWaddr

00:50:56:B5:0B:EB

inet addr:

10.1.1.102

Bcast:10.1.1.255 Mask:255.255.255.0

inet6 addr:

fe80::250:56ff:feb5:beb/64 Scope:Link

UP BROADCAST RUNNING MASTER

MULTICAST MTU:1500 Metric:1

RX packets:1806 errors:0

dropped:0 overruns:0 frame:0

TX packets:9 errors:0 dropped:0

overruns:0 carrier:0

collisions:0 txqueuelen:0

RX bytes:131920 (128.8

KiB) TX bytes:606 (606.0 b)

eth0 Link encap:Ethernet HWaddr 00:50:56:B5:4F:F5

UP BROADCAST RUNNING SLAVE

MULTICAST MTU:1500 Metric:1

RX packets:1007 errors:0

dropped:0 overruns:0 frame:0

TX packets:81 errors:0

dropped:0 overruns:0 carrier:0

collisions:0 txqueuelen:1000

RX bytes:77033 (75.2 KiB) TX bytes:10939 (10.6 KiB)

eth2 Link encap:Ethernet HWaddr 00:50:56:B5:4F:F5

UP BROADCAST RUNNING SLAVE

MULTICAST MTU:1500 Metric:1

RX packets:1005 errors:0

dropped:0 overruns:0 frame:0

TX packets:80 errors:0

dropped:0 overruns:0 carrier:0

collisions:0 txqueuelen:1000

RX bytes:76931 (75.1 KiB) TX bytes:10929 (10.6 KiB)

eth3 Link encap:Ethernet HWaddr 00:50:56:B5:5A:25

UP BROADCAST RUNNING SLAVE

MULTICAST MTU:1500 Metric:1

RX packets:903 errors:0

dropped:0 overruns:0 frame:0

TX packets:4 errors:0 dropped:0

overruns:0 carrier:0

collisions:0 txqueuelen:1000

RX bytes:65983 (64.4 KiB) TX bytes:252 (252.0 b)

eth4 Link encap:Ethernet HWaddr 00:50:56:B5:5A:25

UP BROADCAST RUNNING SLAVE

MULTICAST MTU:1500 Metric:1

RX packets:903 errors:0

dropped:0 overruns:0 frame:0

TX packets:4 errors:0 dropped:0

overruns:0 carrier:0

collisions:0 txqueuelen:1000

RX bytes:66001 (64.4 KiB) TX bytes:284 (284.0 b)

eth5 Link encap:Ethernet HWaddr 00:50:56:B5:0B:EB

UP BROADCAST RUNNING SLAVE

MULTICAST MTU:1500 Metric:1

RX packets:903 errors:0

dropped:0 overruns:0 frame:0

TX packets:5 errors:0 dropped:0

overruns:0 carrier:0

collisio开发云主机域名ns:0 txqueuelen:1000

RX bytes:65951 (64.4 KiB) TX bytes:342 (342.0 b)

eth6 Link encap:Ethernet HWaddr 00:50:56:B5:0B:EB

UP BROADCAST RUNNING SLAVE

MULTICAST MTU:1500 Metric:1

RX packets:903 errors:0

dropped:0 overruns:0 frame:0

TX packets:4 errors:0 dropped:0 overruns:0 carrier:0

collisions:0 txqueuelen:1000

RX bytes:65969 (64.4 KiB) TX bytes:264 (264.0 b)

lo Link encap:Local Loopback

inet addr:127.0.0.1 Mask:255.0.0.0

inet6 addr: ::1/128 Scope:Host

UP LOOPBACK RUNNING MTU:16436

Metric:1

RX packets:4 errors:0 dropped:0

overruns:0 frame:0

TX packets:4 errors:0 dropped:0

overruns:0 carrier:0

collisions:0 txqueuelen:0

RX bytes:240 (240.0 b) TX bytes:240 (240.0 b)


四、


bond


的七种模式

像samba、nfs这种共享文件系统,网络的吞吐量非常大,就造成网卡的压力很大,网卡bond是通过把多个物理网卡绑定为一个逻辑网卡,实现本地网卡的冗余,带宽扩容和负载均衡,具体的功能取决于采用的哪种模式。


1


、mode=0(balance-rr)(平衡抡循环策略)

链路负载均衡,增加带宽,支持容错,一条链路故障会自动切换正常链路。交换机需要配置聚合口,思科叫port channel。


特点:

传输数据包顺序是依次传输(即:第1个包走eth0,下一个包就走eth2….一直循环下去,直到最后一个传输完毕),此模式提供负载平衡和容错能力;但是我们知道如果一个连接或者会话的数据包从不同的接口发出的话,中途再经过不同的链路,在客户端很有可能会出现数据包无序到达的问题,而无序到达的数据包需要重新要求被发送,这样网络的吞吐量就会下降


2


、mode=1(active-backup)(主-备份策略)

这个是主备模式,只有一块网卡是active,另一块是备用的standby,所有流量都在active链路上处理,交换机配置的是捆绑的话将不能工作,因为交换机往两块网卡发包,有一半包是丢弃的。


特点:

只有一个设备处于活动状态,当一个宕掉另一个马上由备份转换为主设备。mac地址是外部可见得,从外面看来,bond的MAC地址是唯一的,以避免switch(交换机)发生混乱。

此模式只提供了容错能力;由此可见此算法的优点是可以提供高网络连接的可用性,但是它的资源利用率较低,只有一个接口处于工作状态,在有 N 个网络接口的情况下,资源利用率为1/N


3


、mode=2(balance-xor)(平衡策略)

表示XOR Hash负载分担,和交换机的聚合强制不协商方式配合。(需要xmit_hash_policy,需要交换机配置port channel)


特点:

基于指定的传输HASH策略传输数据包。缺省的策略是:(源MAC地址 XOR 目标MAC地址) % slave数量。其他的传输策略可以通过xmit_hash_policy选项指定,此模式提供负载平衡和容错能力

本文从转载,原作者保留一切权利,若侵权请联系删除。

《如何进行Linux下双网卡绑定bond的配置》来自互联网同行内容,若有侵权,请联系我们删除!


    免责声明:我们致力于保护作者版权,注重分享,被刊用文章因无法核实真实出处,未能及时与作者取得联系,或有版权异议的,请联系管理员,我们会立即处理! 部分文章是来自自研大数据AI进行生成,内容摘自(百度百科,百度知道,头条百科,中国民法典,刑法,牛津词典,新华词典,汉语词典,国家院校,科普平台)等数据,内容仅供学习参考,不准确地方联系删除处理! 图片声明:本站部分配图来自人工智能系统AI生成,觅知网授权图片,PxHere摄影无版权图库和百度,360,搜狗等多加搜索引擎自动关键词搜索配图,如有侵权的图片,请第一时间联系我们,邮箱:ciyunidc@ciyunshuju.com。本站只作为美观性配图使用,无任何非法侵犯第三方意图,一切解释权归图片著作权方,本站不承担任何责任。如有恶意碰瓷者,必当奉陪到底严惩不贷!

    目录[+]