安装 Mellanox 驱动程序后 modprobe 无法插入 beegfs
我有一个已经运行了几年的存储集群。它基于一个漂亮的 Centos 7.6 设置,使用 beegfs。
为了提高吞吐量,我决定对网络进行测试升级,从 10gig 到 40gig。然而,这个 40gig 卡所需的驱动程序在内核模块方面与 beegfs 冲突。现在我已经成功运行了 40gig 网络,beegfs-client 无法启动:
modprobe: ERROR: could not insert 'beegfs': Unknown symbol in module, or unknown parameter (see dmesg)
如何让这两者相处?
我安装的卡都是 ConnectX-3 FDR Infiniband(不过两个端口都配置为以太网)。我安装的驱动程序是 MLNX_OFED_LINUX-5.0-2.1.8.0-rhel7.6-x86_64
。卸载驱动程序并没有解决问题,但 40gig 网络仍然可以工作。仅需要将端口重新配置为以太网而不是 Infiniband。
更新:从表面上看,我需要向 beegfs-client-autobuild.conf
添加 infiniband 支持。不完全确定在哪里可以找到我需要参考的来源。
I have a storage cluster that has been churning along for a few years. It's based around a pretty stock Centos 7.6 setup, using beegfs.
In an effort to increase throughput I've decided to do a test-upgrade of the network, from 10gig to 40gig. However, it would appear that the necessary drivers for this 40gig card conflicts with beegfs in terms of kernel modules. Now that I have the 40gig network running successfully, beegfs-client fails to start:
modprobe: ERROR: could not insert 'beegfs': Unknown symbol in module, or unknown parameter (see dmesg)
How do I make these two get along?
The cards I've installed are all ConnectX-3 FDR Infiniband (both ports configured to Ethernet, though). The driver I installed is MLNX_OFED_LINUX-5.0-2.1.8.0-rhel7.6-x86_64
. Uninstalling the driver did not resolve the issue, but the 40gig network is still working. It was only needed for recorfiguring the ports to Ethernet instead of Infiniband.
Update: From the looks of it, I will need to add infiniband support to the beegfs-client-autobuild.conf
. Not entirely sure where to find the source that I need to reference.
如果你对这篇内容有疑问,欢迎到本站社区发帖提问 参与讨论,获取更多帮助,或者扫码二维码加入 Web 技术交流群。
绑定邮箱获取回复消息
由于您还没有绑定你的真实邮箱,如果其他用户或者作者回复了您的评论,将不能在第一时间通知您!
发布评论
评论(1)
事实证明答案比预期的更简单:升级到最新版本的 beegfs-client。新版本默认包含 infiniband 兼容性。无需重建。
升级并重新启动后,集群再次按预期运行,Mellanox 40Gb/s 卡运行。
Turns out the answer was simpler than anticipated: upgrade to the newest version of beegfs-client. The newer version includes infiniband compatibility by default. No rebuild needed.
After an upgrade and a reboot, the cluster behaved as intended again, with the Mellanox 40Gb/s cards operating.