-
Notifications
You must be signed in to change notification settings - Fork 19
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Unable to create bridge on same subnet of VF #10
Comments
Here is a sample example, $ docker network create -d sriov --subnet=194.168.1.0/24 -o netdevice=ens1f0 mynet This will allow you to reach to other nodes in same subnet on other systems or VF to PF communication too. |
Thanks for the quick reply. I tried to create the network using SRIOV on same subnet but it didn't work.
Kernel version on host:
Kernel version on Container
So rdma stack is supported from kernel version 4.19 inside the container? I am asking this question because when I create a macvlan bridge on ROCE interface and assign to container , I am able to ping all other IPs but lspci output` 54:00.0 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex] 54:00.1 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex] 54:03.2 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex Virtual Function] 54:03.3 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex Virtual Function] 54:03.4 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex Virtual Function] 54:03.5 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex Virtual Function] 54:03.6 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex Virtual Function] 54:03.7 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex Virtual Function] 54:04.0 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex Virtual Function] 54:04.1 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex Virtual Function] 54:04.2 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex Virtual Function] 54:04.3 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex Virtual Function] 54:04.4 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex Virtual Function] 54:04.5 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex Virtual Function] 54:04.6 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex Virtual Function] 54:04.7 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex Virtual Function] 54:05.0 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex Virtual Function] 54:05.1 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex Virtual Function] 54:05.2 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex Virtual Function] 54:05.3 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex Virtual Function] 54:05.4 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex Virtual Function] 54:05.5 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex Virtual Function] 54:05.6 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex Virtual Function] 54:05.7 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex Virtual Function] 54:06.0 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex Virtual Function] 54:06.1 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex Virtual Function] 74:00.0 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex] 74:00.1 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex] 94:00.0 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex] 94:00.1 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex] b4:00.0 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex] b4:00.1 Ethernet controller: Mellanox Technologies MT28800 Family [ConnectX-5 Ex] ifconfig output` docker0: flags=4099<UP,BROADCAST,MULTICAST> mtu 1500 lo: flags=73<UP,LOOPBACK,RUNNING> mtu 65536 lo:1: flags=73<UP,LOOPBACK,RUNNING> mtu 65536 lo:2: flags=73<UP,LOOPBACK,RUNNING> mtu 65536 lo:3: flags=73<UP,LOOPBACK,RUNNING> mtu 65536 re0: flags=4163<UP,BROADCAST,RUNNING,MULTICAST> mtu 2300 re1: flags=4163<UP,BROADCAST,RUNNING,MULTICAST> mtu 2300 re2: flags=4163<UP,BROADCAST,RUNNING,MULTICAST> mtu 2300 re3: flags=4163<UP,BROADCAST,RUNNING,MULTICAST> mtu 2300 re4: flags=4163<UP,BROADCAST,RUNNING,MULTICAST> mtu 2300 re5: flags=4163<UP,BROADCAST,RUNNING,MULTICAST> mtu 2300 re6: flags=4163<UP,BROADCAST,RUNNING,MULTICAST> mtu 2300 re7: flags=4163<UP,BROADCAST,RUNNING,MULTICAST> mtu 2300 virbr0: flags=4099<UP,BROADCAST,MULTICAST> mtu 1500 |
@psaini79 @psaini2018 for duplicate subnet IP, please share the docker version and sriov plugin logs. |
Thanks, you are right. I am able to create network on same subnet and able to ping the target from container. ` eth0: flags=4163<UP,BROADCAST,RUNNING,MULTICAST> mtu 1500 ` PING 192.168.10.17 (192.168.10.17) 56(84) bytes of data. 64 bytes from 192.168.10.17: icmp_seq=1 ttl=64 time=0.120 ms 64 bytes from 192.168.10.17: icmp_seq=2 ttl=64 time=0.078 ms However, rping command is failing. I executed following command on server i.e. on 192.168.10.17:
cma event RDMA_CM_EVENT_ADDR_ERROR, error -19 waiting for addr/route resolution state 1
|
@psaini79 We also likely need to see kernel ftraces if it doesn't work. We haven't tried mofed user space and upstream kernel. With mofed kernel, mofed user space should be used. So post this, you might want to create rdma-core based container image. Please also share the docker run command that you run. |
Yes, I followed the steps given in https://community.mellanox.com/s/article/docker-rdma-sriov-networking-with-connectx4-connectx5. However, I executed steps from 5th step onward. Also, following commands exit the container without any error: ` client verbose created cm_id 0x97f1c0 cma_event type RDMA_CM_EVENT_ADDR_ERROR cma_id 0x97f1c0 (parent) cma event RDMA_CM_EVENT_ADDR_ERROR, error -19 waiting for addr/route resolution state 1 destroy cm_id 0x97f1c0 ` I have one more question, any ETA for rds_ping to work inside the container? |
@psaini2018 What is the command you ran to run this container? Please talk to Mellanox support for rds-ping. |
The above command failed inside the container: [root@9098b076cc9a mofed_installer]# ibdev2netdev uname -a |
@psaini2018 so as we discussed yesterday in this thread that you need kernel 4.20. Please upgrade to it. |
I need to upgrade the host kernel to 4.20? Just want to make sure to avoid any rework. Also, for rds_ping SR need to be opened through MLX support login? or is there any github repo for that to open the issue? |
@psaini2018, yes 4.20 or higher. 5.1 is even better. :-) rds-ping is owned by Oracle, you should first resolve at Oracle on supporting rds-ping before opening Mellanox support case. |
Ok and thanks a lot for your quick reply. As per the following link Ethernet card inside the container made available using IPoIB. I have a question, what is the difference between IPoIB device and VM IPoIB device? are they technically same? |
@psaini2018 |
I tried to create to the bridge using sriov plugin but it kept on failing on ROCE CX5 card. I tried to use the same subnet which is available on host devices i.e. 192.168.10.0/24 but I get following error:
docker network create -d sriov --subnet=192.168.10.0/24 -o netdevice=re6 -o mode=passthrough mynet1 Error response from daemon: Pool overlaps with other one on this address space
I am able to create the bridge if I use different subnet but for my usecase bridge must be on same subnet so that I can reach other nodes running on same subnet. Also, I want to know if rdma stack can work inside the container? will rds-ping will work inside the container?
The text was updated successfully, but these errors were encountered: