Questions tagged [infiniband]

InfiniBand is a high-speed switched fabric communications link technology used in high-performance computing and enterprise data centers.

InfiniBand is a switched fabric communications link used in high-performance computing and enterprise data centers. Its features include scalability, high throughput, low latency, quality of service and failover. The InfiniBand architecture specification defines a connection between processor nodes and high performance I/O nodes such as storage devices. Infiniband host bus adapters and network switches are commonly manufactured by both Mellanox and Intel.

178 questions
4
votes
1 answer

How to Configure Chapel/GASNet for running multilocale codes on MXM Infiniband network with Partition Key?

I'm trying to run a multilocale Chapel code on a cluster that has an MXM Infiniband network(40 Gbps, model: Mellanox Technologies MT26428). I followed both Chapel and GASNet documentations, and I set export CHPL_COMM_SUBSTRATE=ibv export…
Tiago Carneiro
  • 199
  • 1
  • 9
4
votes
1 answer

How to use the RDMA on Azure

I purchased 1 H16r Virtual Machine on Azure, and I tried the setting method of RDMA on "About H-series and compute-intensive A-series VMs": azure config mode arm azure vm extension set RDMAUpdateForLinux…
ZhouYoung
  • 41
  • 2
4
votes
1 answer

Toy program Fails using OpenMPI 1.6 but works with Mvapich2

I am trying to figure out why my version of OpenMPI 1.6 does not work. I am using gcc-4.7.2 on CentOS 6.6. Given a toy program (i.e. hello.c) #include #include #include int main(int argc, char * argv[]) { int…
irritable_phd_syndrome
  • 4,631
  • 3
  • 32
  • 60
4
votes
1 answer

How to use GPUDirect RDMA with Infiniband

I have two machines. There are multiple Tesla cards on each machine. There is also an InfiniBand card on each machine. I want to communicate between GPU cards on different machines through InfiniBand. Just point to point unicast would be fine. I…
Hot.PxL
  • 1,902
  • 1
  • 17
  • 30
4
votes
4 answers

Infiniband in Java

As you all know, OFED's Socket Direct protocol is deprecated and OFED's 3.x releases do not come with SDP at all. Hence, Java's SDP also fails to work. I was wondering what is the proper method to program infiniband in Java? Is there any portable…
RoboAlex
  • 4,895
  • 6
  • 31
  • 37
4
votes
2 answers

In Infiniband, what mapping in PCIe-BAR, the internal buffer of Infiniband card or the remote computer's RAM?

As we know, Infiniband allows RDMA - direct access to the memory of the remote computer. It is also known, that the PCI-Express (endpoint) devices, including the PCIe-card Infiniband, are able to communicate in two ways: IO Ports (in / out)…
Alex
  • 12,578
  • 15
  • 99
  • 195
4
votes
3 answers

What is the maximum length of the cable can be for infiniband(RDMA)?

What is the maximum length of the cable can be for infiniband(RDMA)? For example. Here have said, that it can be up to 10 M for a four-channel copper cable. And using the Connects Cables using fiber, which fully electrically compatible with standard…
Alex
  • 12,578
  • 15
  • 99
  • 195
4
votes
1 answer

Can't connect to server through infiniband

I am trying to make a small server in c++ which would simply echo anything it receives through an infiniband connection. I am also using the Sockets Direct Protocol and POSIX sockets under Ubuntu. Unfortunately, the only advice i found on the…
Alin Huruba
  • 35,481
  • 2
  • 12
  • 9
3
votes
0 answers

Analogous to tcpkill for IP over InfiniBand

Does anybody know if there is an equivalent tool to tcpkill for killing TCP connections that travel over Infiniband? At the moment, when I use an infiniband interface, tcpkill complains: # tcpkill -i ib0 port 12345 tcpkill: unsupported datalink…
Luca Martini
  • 1,434
  • 1
  • 15
  • 35
3
votes
1 answer

RDMA between GPU and remote host

Is it possible to perform an RDMA operation between a GPU and a remote host? The documentation online on Nvidia's website only talks about doing RDMA between GPUs and does not mention the possibility of doing this between GPU-host. Note: I have…
JC1
  • 657
  • 6
  • 21
3
votes
1 answer

How to: Azure OpenMPI with Infiniband - Linux

I am new to using Microsoft Azure for scientific computing purposes and have encountered a few issues whilst setting up. I have a jump box set-up that acts as a license server for the software that I whish to use, is also has a common drive to store…
Owain1602
  • 31
  • 2
3
votes
1 answer

What's the doorbell in InfiniBand?

I am reading a document ablut InfiniBand, which is called 《InfiniBand™ Host Channel Adapter Verb Implementer’s Guide》--Intel, in Chapter 6.2.2, it mentions a concept called "Doorbell", I haven't found any information about this concept in the IB…
3
votes
1 answer

Segfaults when running OpenMPI job inside Slurm runscript

We are running a small cluster environment with Intel Xeon nodes connected via Infiniband. The login node is not attached to the infiniband interconnect. All nodes run Debian Jessie. We run Slurm 14.03.9 on the Login node. As the system OpenMPI is…
3
votes
1 answer

Does Distributed TensorFlow support InfiniBand interconnections "out of the box"

We have a cluster with InfiniBand interconnects. I know that Distributed TF uses gRPC to handle network calls, but I was wondering if there was anything special required to gain the speed of InfiniBand connections when using the gRPC.
Jacob Holloway
  • 887
  • 8
  • 24
3
votes
0 answers

What is the difference between SDP and VMA?

As known for Infiniband or RoCE-Ethernet we can use: VMA - Voltaire Messaging Accelerator (VMA) Mellanox's Messaging Accelerator (VMA) boosts performance for message-based and streaming applications... The result is a reduction in latency by as…
Alex
  • 12,578
  • 15
  • 99
  • 195
1
2
3
11 12