Neanderthal’s default option is to use use native libraries, so it is very important that you do not skip any part of this guide.
How to Get Started
- Walk through this guide, set up your development environment, and try the examples.
- Familiarize yourself with Neanderthal’s more detailed tutorials and API documentation.
This is what you’ll be able to do after the installation:
uncomplicate.neanderthal.native in your namespace, and you’ll be able to call appropriate functions from the Neanderthal library.
(ns example (:use [uncomplicate.neanderthal core native]))
Now you can create neanderthal vectors and matrices, and do computations with them.
Here is how we create two double-precision vectors and compute their dot product:
(def x (dv 1 2 3)) (def y (dv 10 20 30)) (dot x y)
This is one of the ways to multiply matrices:
(def a (dge 3 2 [1 2 3 4 5 6])) (def b (dge 2 3 [10 20 30 40 50 60])) (mm a b)
Overview and Features
Neanderthal is a Clojure library for fast matrix and linear algebra computations that supports pluggable engines:
- The native engine is based on a highly optimized native Intel’s MKL library of BLAS and LAPACK computation routines (MKL is not open-source, but it is free to use and redistribute since 2016).
- The CUDA GPU engine is based on cuBLAS and supports all modern Nvidia GPUs. It uses ClojureCUDA and JCuda libraries. Check out Uncomplicate ClojureCUDA.
- The OpenCL GPU engine is based on OpenCL BLAS routines from CLBlast library for even more computational power when needed. It uses ClojureCL and JOCL libraries. Check out Uncomplicate ClojureCL.
- Data structures: double and single precision vectors, dense matrices (GE), triangular matrices (TR), symmetric matrices (SY), banded, diagonal, etc.;
- BLAS Level 1, 2, and 3 routines;
- Lots of (but not all) LAPACK routines;
- Various Clojure vector and matrix functions (transpositions, submatrices etc.);
- Easy and efficient data mapping and transfer to and from GPUs;
- Fast map, reduce and fold implementations for the provided structures.
- OpenCL GPU support
- CUDA GPU support
On the TODO List
- “Tensors” (as defined in various deep learning libraries)
- Sparse matrices;
- Support for complex numbers;
- Add Neanderthal jars to your classpath (from the Clojars).
- To use the native engine: install Intel’s MKL on your system following Native Engine Requirements).
- To use the GPU engine: install the drivers and a CUDA and/or OpenCL platform software provided by the vendor of your graphic card (you probably already have that; see GPU Engine Requirements).
The most straightforward way to include Neanderthal in your project is with Leiningen. Add the following dependency to your
project.clj, just like in the Hello World project:
You need at least Java 8.
Neanderthal’s data structures are written in Clojure, so many functions work even without native engines. However, you probably need Neanderthal because of its fast BLAS native or GPU engines. Here is how to make sure they are available.
The native library used by Neanderthal’s native engine
- Works on Linux, OS X, and Windows!
Neanderthal uses the native Intel MKL library and expects that you make it available on your system, typically as shared xyz.so, xyz.dll, or xyz.dylib files. Intel MKL is highly optimized for various architectures; its installation comes with many optimized binaries for all supported architectures, that are then selected during runtime according to the hardware at hand. Neanderthal has been built and tested with Intel MKL 2018; please make sure that you use a compatible MKL version.
You do not need to compile or tune anything yourself.
There are two main steps to how to make MKL available on your system; either:
- (Optional) Install MKL using a GUI installer provided by Intel free of charge. In case you use this method, you may set environment variables as explained in this guide, but it is probably not required, since you do not need to compile anything.
- Put all required binary files (that you installed with MKL installer or copied from wherever you acquired them) in a directory that is available from your
PATH(windows). Those binary files are available from anyone who installed MKL and have the (free) license to redistribute it with a project.
Either way, Neanderthal does not care how MKL has been provided, as long as it is on the path of your OS. When it comes to distributing the software you build using Neanderthal, I guess the most convenient option is that you include the required MKL binary files in the uberjar or other package that you use to ship the product. Then, it would not require any additional action from your users.
This is the list of MKL files that Neanderthal requires:
- Mandatory MKL libraries (this is what I use on my Intel Core i7 4790k):
libmkl_avx2.so(if your CPU supports AVX2, which it probably does)
- Optionally, your processor might support additional set of instructions that may require an additional file from MKL. See the MKL guide and browse all available files once you install MKL to discover this. For example, I guess that you might also need
libmkl_avx512.soif you have a Xeon processor.
Please note that, if you use Windows or OS X, the binary file extensions are not
Note for OSX users: MKL installation on my OSX 10.11 placed
libiomp5.dylib in a different folder than the rest of the MKL libraries. In such case, copy that file where it is visible, or don’t rely on the MKL installation, but select the needed library files and put them somewhere on the
DYLD_LIBRARY_PATH. In newer versions of OSX, you’d have to configure the “system integrity protection (SIP)” settings for
DYLD_LIBRARY_PATH to be respected by the system see more here. If you want a quick & dirty solution without much fuss, copying the
dylib files and pasting them into
/usr/local/lib has been reported to work by multiple users.
Another important thing on Mac: CUDA GPU code that Neanderthal uses could not have been compiled for macOS since it is difficult to find a Mac with Nvidia hardware. That means that you have to explicitly exclude jcuda natives in your project. See how to do that in the Hello World project example. If you have a Mac with an Nvidia GPU and can help with the macOS build, please drop me a note. This will be permanently fixed in the following versions of Neanderthal.
Note for Windows users: MKL installation on my Windows 10 keeps all required
.dll files in the
<install dir>\redist folder. The usual folders that keep
dylib on Linux and OSX, keep
.lib files on Windows - you do not need those. Add the folder that contains the
dlls into the
PATH environment variable, and you’re ready to go. Some Windows users reported that
libiomp5.dll too is in another folder; see the note for OSX users and take the equivalent Windows action.
GPU drivers for the OpenCL GPU engine
Everything will magically work (no need to compile anything) on Nvidia, AMD, and Intel’s GPUs and CPUs as long as you have appropriate GPU drivers.
Works on Linux, Windows, and OS X!
Follow the ClojureCL getting started guide for the links for the GPU platform that you use and more detailed info.
If you use a pre-2.0 OpenCL platform (Nvidia and/or OS X), you’ll have to use
with-default-1 instead of
with-default that are used in the examples.
GPU drivers for the CUDA GPU engine
Everything will magically work (no need to compile anything) on Nvidia, provided that you installed the latest Nvidia’s CUDA Toolkit.
Follow the ClojureCUDA getting started guide for the links for the GPU platform that you use and more detailed info.
Where to Go Next
Hopefully this guide got you started and now you’d like to learn more. I expect to build a comprehensive base of articles and references for exploring this topic, so please check the All Guides page from time to time. Of course, you should also check the Neanderthal API for specific details, and feel free to take a glance at the source while you are there.
It is also a good idea to follow my blog at dragan.rocks since I’ll write about Neanderthal there.
Tell Us What You Think!
Please take some time to tell us about your experience with the library and this site. Let us know what we should be explaining or is not clear enough. If you are willing to contribute improvements, even better!