This is an automated archive made by the Lemmit Bot.
The original was posted on /r/homelab by /u/ConstructionSafe2814 on 2023-08-03 22:07:03.
OK so I have a c7000 enclosure in my home lab with a QDR BladeSystem Infiniband 40Gbps switch attached to it. The (BL460c gen8) blades have a 544M mezzanine QDR Mellanox ConnectX4 card in them. Then I used IPoIB to connect 4 hosts (mtu 64k)Yay so far so good
I put a 3.7GB ISO file on tmpfs (RAM) and exported it over NFS. Then I mounted the drive and cp’d it to the NFS client, also to RAM to avoid HDD bottlenecks.
I am “only” able to reach 1.2GB/s file transfer ratio. The bottleneck sits on the client side and is a single core saturated in kernel space. It’s always the last core of one of the 2 CPU’s installed in the blade.
I might try NFS over UDP? That might be friendlier to the CPU. But I noticed it’s NFSv4.2. Not sure if it can go over UDP.
IPoIB might not be ideal for performance. If I choose a higher clocked CPU, I will squeeze out a couple of 100MB/s or so. But how do I get 40Gbps? I’m roughly at one fourth/fifth now.
Would mounting NFS over RDMA be better? Or what other use cases transfer data from one host to another much faster?