gchanan / nccl.torch

torch bindings for nccl

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

nccl.torch

Torch7 FFI bindings for NVidia NCCL library.

Installation

Collective operations supported

  • allReduce
  • reduce
  • broadcast
  • allGather

Example usage

Argument to the collective call should be a table of contiguous tensors located on the different devices. Example: perform in-place allReduce on the table of tensors:

require 'nccl'
nccl.allReduce(inputs)

where inputs is a table of contiguous tensors of the same size located on the different devices.

About

torch bindings for nccl


Languages

Language:Lua 97.5%Language:CMake 2.5%