mpl icon indicating copy to clipboard operation
mpl copied to clipboard

The MaPLe compiler for efficient and scalable parallel functional programming

MPL

MaPLe (MPL) is an extension of the MLton compiler for Standard ML which implements support for nested (fork-join) parallelism. MPL generates executables with excellent multicore performance, utilizing a novel approach to memory management based on the theory of disentanglement [1,2,3,4].

MPL is research software and is being actively developed.

If you are you interested in using MPL, consider checking out the tutorial. You might also be interested in exploring mpllib (a standard library for MPL) and the Parallel ML benchmark suite.

Docker

Try out MPL with Docker:

$ docker pull shwestrick/mpl
$ docker run -it shwestrick/mpl /bin/bash
...# examples/bin/primes @mpl procs 4 --

If you want to try out MPL by writing and compiling your own code, we recommend mounting a local directory inside the container. For example, here's how you can use MPL to compile and run your own main.mlb in the current directory. (To mount some other directory, replace $(pwd -P) with a different path.)

$ ls
main.mlb
$ docker run -it -v $(pwd -P):/root/mycode shwestrick/mpl /bin/bash
...# cd /root/mycode
...# mpl main.mlb
...# ./main @mpl procs 4 --

Build and Install (from source)

Requirements

MPL has only been tested on Linux with x86-64. The following software is required.

  • GCC
  • GMP (GNU Multiple Precision arithmetic library)
  • GNU Make, GNU Bash
  • binutils (ar, ranlib, strip, ...)
  • miscellaneous Unix utilities (diff, find, grep, gzip, patch, sed, tar, xargs, ...)
  • Standard ML compiler and tools:
    • Recommended: MLton (mlton, mllex, and mlyacc). Pre-built binary packages for MLton can be installed via an OS package manager or (for select platforms) obtained from http://mlton.org.
    • Supported but not recommended: SML/NJ (sml, ml-lex, ml-yacc).

Instructions

The following builds the compiler at build/bin/mpl.

$ make all

After building, MPL can then be installed to /usr/local:

$ make install

or to a custom directory with the PREFIX option:

$ make PREFIX=/opt/mpl install

Parallel and Concurrent Extensions

MPL extends SML with a number of primitives for parallelism and concurrency. Take a look at examples/ to see these primitives in action.

Note: Before writing any of your own code, make sure to read the section "Disentanglement" below.

The ForkJoin Structure

val par: (unit -> 'a) * (unit -> 'b) -> 'a * 'b
val parfor: int -> (int * int) -> (int -> unit) -> unit
val alloc: int -> 'a array

The par primitive takes two functions to execute in parallel and returns their results.

The parfor primitive is a "parallel for loop". It takes a grain-size g, a range (i, j), and a function f, and executes f(k) in parallel for each i <= k < j. The grain-size g is for manual granularity control: parfor splits the input range into approximately (j-i)/g subranges, each of size at most g, and each subrange is processed sequentially. The grain-size must be at least 1, in which case the loop is "fully parallel".

The alloc primitive takes a length and returns a fresh, uninitialized array of that size. Warning: To guarantee no errors, the programmer must be careful to initialize the array before reading from it. alloc is intended to be used as a low-level primitive in the efficient implementation of high-performance libraries. It is integrated with the scheduler and memory management system to perform allocation in parallel and be safe-for-GC.

The MLton.Parallel Structure

val compareAndSwap: 'a ref -> ('a * 'a) -> 'a
val arrayCompareAndSwap: ('a array * int) -> ('a * 'a) -> 'a

compareAndSwap r (x, y) performs an atomic CAS which attempts to atomically swap the contents of r from x to y, returning the original value stored in r before the CAS. Polymorphic equality is determined in the same way as MLton.eq, which is a standard equality check for simple types (char, int, word, etc.) and a pointer equality check for other types (array, string, tuples, datatypes, etc.). The semantics are a bit murky.

arrayCompareAndSwap (a, i) (x, y) behaves the same as compareAndSwap but on arrays instead of references. This performs a CAS at index i of array a, and does not read or write at any other locations of the array.

Using MPL

MPL uses .mlb files (ML Basis) to describe source files for compilation. A typical .mlb file for MPL is shown below. The first three lines of this file respectively load:

  • The SML Basis Library
  • The ForkJoin structure, as described above
  • The MLton structure, which includes the MPL extension MLton.Parallel as described above, as well as various MLton-specific features. Not all MLton features are supported (see "Unsupported MLton Features" below).
(* libraries *)
$(SML_LIB)/basis/basis.mlb
$(SML_LIB)/basis/fork-join.mlb
$(SML_LIB)/basis/mlton.mlb

(* your source files... *)
A.sml
B.sml

Compiling a Program

The command to compile a .mlb is as follows. By default, MPL produces an executable with the same base name as the source file, i.e. this would create an executable named foo:

$ mpl [compile-time options...] foo.mlb

MPL has a number of compile-time options derived from MLton, which are documented here. Note that MPL only supports C codegen and does not support profiling.

Some useful compile-time options are

  • -output <NAME> Give a specific name to the produced executable.
  • -default-type int64 -default-type word64 Use 64-bit integers and words by default.
  • -debug true -debug-runtime true -keep g For debugging, keeps the generated C files and uses the debug version of the runtime (with assertions enabled). The resulting executable is somewhat peruse-able with tools like gdb.
  • -detect-entanglement true enables the dynamic entanglement detector. See below for more information.

For example:

$ mpl -default-type -int64 -output foo sources.mlb

Running a Program

MPL executables can take options at the command line that control the run-time system. The syntax is

$ <program> [@mpl [run-time options...] --] [program args...]

The runtime arguments must begin with @mpl and end with --, and these are not visible to the program via CommandLine.arguments.

Some useful run-time options are

  • procs <N> Use N worker threads to run the program.
  • set-affinity Pin worker threads to processors. Can be used in combination with affinity-base <B> and affinity-stride <S> to pin thread i to processor number B + S*i.
  • block-size <X> Set the heap block size to X bytes. This can be written with suffixes K, M, and G, e.g. 64K is 64 kilobytes. The block-size must be a multiple of the system page size (typically 4K). By default it is set to one page.

For example, the following runs a program foo with a single command-line argument bar using 4 pinned processors.

$ foo @mpl procs 4 set-affinity -- bar

Disentanglement

Currently, MPL only supports programs that are disentangled, which (roughly speaking) is the property that concurrent threads remain oblivious to each other's allocations [3].

Here are a number of different ways to guarantee that your code is disentangled.

  • (Option 1) Use only purely functional data (no refs or arrays). This is the simplest but most restrictive approach.
  • (Option 2) If using mutable data, use only non-pointer data. MPL guarantees that simple types (int, word, char, real, etc.) are never indirected through a pointer, so for example it is safe to use int array. Other types such as int list array and int array array should be avoided. This approach is very easy to check and is surprisingly general. Data races are fine!
  • (Option 3) Make sure that your program is race-free. This can be tricky to check but allows you to use any type of data. Many of our example programs are race-free.

Entanglement Detection

Whenever a thread acquires a reference to an object allocated concurrently by some other thread, then we say that the two threads are entangled. This is a violation of disentanglement, which MPL currently does not allow.

To check if your program has entanglement, MPL has a built-in dynamic entanglement detector. You can enable the detector by using -detect-entanglement true at compile time. When entanglement detection is enabled, MPL will monitors individual reads and writes during execution; if entanglement is found, the program will terminate with an error message.

Entanglement detection is highly optimized, and often does not have a significant impact on performance. We recommend using entanglement detection liberally.

Note that the detector is execution-dependent: if your program is non-deterministic (e.g. racy), then entanglement may or may not occur depending on the outcome of a race condition. Similarly, entanglement could be input-dependent. Therefore, we recommend testing multiple inputs, as well as running on varying number of processors.

Bugs and Known Issues

Basis Library

In general, the basis library has not yet been thoroughly scrubbed, and many functions may not be safe for parallelism (#41). Some known issues:

  • Int.toString is racy when called in parallel.
  • Real.fromString may throw an error when called in parallel.

Garbage Collection

  • (#115) The GC is currently disabled at the "top level" (outside any calls to ForkJoin.par). For highly parallel programs, this has generally not been a problem so far, but it can cause a memory explosion for programs that are mostly (or entirely) sequential.

Unsupported MLton Features

Many MLton-specific features are unsupported, including (but not limited to):

  • share
  • shareAll
  • size
  • Finalizable
  • Profile
  • Signal
  • Thread (partially supported but not documented)
  • Cont (partially supported but not documented)
  • Weak
  • World

References

[1] Hierarchical Memory Management for Parallel Programs. Ram Raghunathan, Stefan K. Muller, Umut A. Acar, and Guy Blelloch. ICFP 2016.

[2] Hierarchical Memory Management for Mutable State. Adrien Guatto, Sam Westrick, Ram Raghunathan, Umut Acar, and Matthew Fluet. PPoPP 2018.

[3] Disentanglement in Nested-Parallel Programs. Sam Westrick, Rohan Yadav, Matthew Fluet, and Umut A. Acar. POPL 2020.

[4] Provably Space-Efficient Parallel Functional Programming. Jatin Arora, Sam Westrick, and Umut A. Acar. POPL 2021.