mengke-mk / RING

RING is a ~5k lines pure C++11 runtime system for scaling irregular applications in PGAS programming praradigm.

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Intro

RING is a ~5k lines pure C++11 runtime system for scaling irregular applications in PGAS programming paradigm.

Dependences

You must have a 64-bit Linux system with the following installed to build RING.

  • Build
    • CMake >= 2.8.12
  • Compiler
    • GCC > 4.8
  • External:
    • Infiniband Verbs
    • MPI
    • pthread
    • libnuma
  • optional
    • gperftool

Quick Start

Like Grappa, Ring is primarly designed to be a "global view" model, whcih means that rather than coordinating where all parallel SPMD processes are at and how they devide up the data, the programmer is encourage to think of the system as a large single shared memory.

Section 1 : Hello world

You can use run() to capture the whole process you want to run. Use all_do() to spawn the same task on all cores just like what you do in SPMD. Use call() to perform a remote process call.

void run([]{ /*your codes*/ });
void all_do([]{ /*your works*/ });

void call(GlobalAddress<T> gt, [](T* t){} );
auto f = call(GlobalAddress<T> gt, [](T* t){} );
T result = f.get() // this will block explicitly just like C++11 future

void call(int which_core, [](){} );
auto f = call(int which_core, [](){} );
T result = f.get() // this will block explicitly just like C++11 future

#include <iostream>
#include <vector>
#include <cstdio>
#include "ring.hpp"

int x;

int main(int argc, char* argv[]){
  RING_Init(&argc, &argv);
  run([]{
    all_do([]{
      x=0;
      if(thread_rank() == 0){
        auto gx = make_global(0, 1, &x);
        call(gx, [](int *x){
           sync_printf("hello world", (*x));
        });//call
      }
    });//all_do
  });//run
  RING_Finalize();
  return 0;
}

Section 2 : Global memory

In Ring, all memory on all cores is addressable by any other core in spirit of PGAS programming paradigm.

You can allocate a global array by gmalloc(), free it by gfree().

auto A = gmalloc(size_t size);//only can be called inner run()
free(A);

run([]{
  auto x = gmalloc<int>(1<<20);
  auto y = gmalloc<int>(1<<20);
  auto x1 = x + 230;
  auto y1 = y + 333;
  call(x1, [](int * w){
    sync_printf("hello pgas x", id());
  });

  call(y1, [](int * w){
    sync_printf("hello pgas y", id());
  });
  /* free request will be executed after rpc is executed. */
  gfree(x);
  gfree(y);
});

Section 3 : parallel for

Instead of spawning tasks individually, it's almost always better to use a parallel loops of some sort. you can use pfor() to spwan loop iterations recursively untill hitting a threshold.

void pfor(Array<T> A, int s, int e, [](T* t){})
void pfor(Array<T> A, int s, int e, [](i, T* t){})

run([]{
  auto A = gmalloc<int>(1<<20);
  pfor(A, 0, 1024, [](int* t){
    auto a = id();
    call(1, [a]{
      sync_printf(id());
    });//call
  });//pfor
  gfree(std::move(A));
});//run

Benchmarks

  • GUPS (Ring vs. Grappa)
    • 1.23e8 vs. 8.04e7 (UPS), 120 core do 1<<20 updates
    • 5.30e7 vs. 5.12e7 (UPS), 12 core do 1<<20 updates
    • 2.36e7 vs. 2.10e7 (UPS), 1 core do 1<<20 updates
    • Baseline 2.62e8 (UPS), 1 core do 1<<20 updates.
  • Graph500 (Ring bs. Grappa)
    • 1.59e8 vs. 9.7e7 (TEPS), 120 core do scale=22 BFS. (10 nodes)
    • 1.55e7 vs 1.64e7 (TEPS), 12 core do scale=22 BFS. (1 node)

About

RING is a ~5k lines pure C++11 runtime system for scaling irregular applications in PGAS programming praradigm.


Languages

Language:C 61.5%Language:C++ 35.7%Language:Assembly 1.9%Language:CMake 0.6%Language:Shell 0.2%Language:Makefile 0.1%