mellanox

linux安装&卸载IB网卡(mellanox)驱动

浪尽此生 提交于 2020-03-24 03:49:41
由于工作上的需求,需要在redhat6.5服务器上面安装mellanox品牌的IB网卡,故整理了一个安装&卸载IB网卡驱动的教程: 一、安装 1) 下载相应操作系统驱动 http://www.mellanox.com/page/products_dyn?product_family=26&mtag=linux_sw_drivers 2 )解压安装驱动 tar zxf MLNX_OFED_LINUX-4.1-1.0.2.0-rhel6.5-x86_64.tar.gz cd MLNX_OFED_LINUX-4.1-1.0.2.0-rhel6.5-x86_64 ./mlnxofedinstall # 安装程序会检查系统库,如果缺少库会直接提示使用yum install ‘缺少的库’安装,根据提示安装即可,-q参数静默安装 /etc/init.d/openibd restart # 安装完成后重启主机 ,使用ifconfig -a或者ip a就可以看到ib0,ib1...等名字的IB网卡了 3 )重新编译添加针对当前内核版本的 IB 驱动 进行第2)步时如果下载的驱动不支持当前内核版本,则会提示编译安装,步骤如下: ◆在编译之前首先安装gcc编译环境和kernel开发包 yum install gcc gcc-c++ libstdc++-devel kernel-default-devel

What is the meaning of IB read, IB write, OB read and OB write. They came as output of Intel® PCM while monitoring PCIe bandwidth

早过忘川 提交于 2019-12-11 02:56:42
问题 I am trying to measure the PCIe bandwidth of NIC devices using Intel® Performance Counter Monitor (PCM) tools. But, I am not able to understand the output of it. To measure the PCIe bandwidth, I executed the binary pcm-iio. This binary helps to measure the monitor PCIe bandwidth per PCIe device. After executing the binary I got the following output. |IIO Stack 2 - PCIe1 |IB write|IB read|OB read|OB write|TLB Miss|VT-d L3 Miss|VT-d CTXT Miss|VT-d Lookup| |_____________________________|________

How to use GPUDirect RDMA with Infiniband

匿名 (未验证) 提交于 2019-12-03 01:23:02
可以将文章内容翻译成中文,广告屏蔽插件可能会导致该功能失效(如失效,请关闭广告屏蔽插件后再试): 问题: I have two machines. There are multiple Tesla cards on each machine. There is also an InfiniBand card on each machine. I want to communicate between GPU cards on different machines through InfiniBand. Just point to point unicast would be fine. I surely want to use GPUDirect RDMA so I could spare myself of extra copy operations. I am aware that there is a driver available now from Mellanox for its InfiniBand cards. But it doesn't offer a detailed development guide. Also I am aware that OpenMPI has support for the feature I am asking. But