NNPACK is an acceleration package for neural network computations. NNPACK aims to provide high-performance implementations of convnet layers for multi-core CPUs.
NNPACK is not intended to be directly used by machine learning researchers; instead it provides low-level performance primitives leveraged in leading deep learning frameworks, such as PyTorch, Caffe2, MXNet, tiny-dnn, Caffe, Torch, and Darknet.
Environment | Architecture | CPU requirements |
---|---|---|
Linux | x86-64 | AVX2 and 3-level cache hierarchy |
Linux | ARM | NEON |
Linux | ARM64 | |
macOS | x86-64 | AVX2 and 3-level cache hierarchy |
Android | ARM | NEON |
Android | ARM64 | |
Android | x86 | |
Android | x86-64 | |
iOS | ARM | |
iOS | ARM64 | |
Emscripten | Asm.js | |
Emscripten | WebAssembly |
- Multiple algorithms for convolutional layers:
- Fast convolution based on Fourier transform (for kernels up to 16x16 without stride)
- Fast convolution based on Winograd transform (for 3x3 kernels without stride)
- Implicit matrix-matrix multiplication algorithm (no limitations)
- Direct convolution algorithm (for 1x1 kernels without stride)
- Multi-threaded SIMD-aware implementations of neural network layers
- Implemented in C99 and Python without external dependencies
- Extensive coverage with unit tests
- Convolutional layer
- Inference-optimized forward propagation (
nnp_convolution_inference
) - Training-optimized forward propagation (
nnp_convolution_output
) - Training-optimized backward input gradient update (
nnp_convolution_input_gradient
) - Training-optimized backward kernel gradient update (
nnp_convolution_kernel_gradient
)
- Inference-optimized forward propagation (
- Fully-connected layer
- Inference-optimized forward propagation (
nnp_fully_connected_inference
andnnp_fully_connected_inference_f16f32
version for FP16 weights) - Training-optimized forward propagation (
nnp_fully_connected_output
)
- Inference-optimized forward propagation (
- Max pooling layer
- Forward propagation, both for training and inference, (
nnp_max_pooling_output
)
- Forward propagation, both for training and inference, (
- ReLU layer (with parametrized negative slope)
- Forward propagation, both for training and inference, optionally in-place, (
nnp_relu_output
) - Backward input gradient update (
nnp_relu_input_gradient
)
- Forward propagation, both for training and inference, optionally in-place, (
- Softmax layer
- Forward propagation, both for training and inference, optionally in-place (
nnp_softmax_output
)
- Forward propagation, both for training and inference, optionally in-place (
For most users, the recommended way to build NNPACK is through CMake:
mkdir build
cd build
cmake -G Ninja ..
ninja
Note: if ninja
is not available on your system, configure without -G Ninja
, and use make
instead of ninja
.
To cross-compile for Android, add extra configuration options for cmake
: -DCMAKE_TOOLCHAIN_FILE=$ANDROID_NDK/build/cmake/android.toolchain.cmake
(where $ANDROID_NDK
is the path to Android NDK directorory, e.g. /opt/android-ndk-r15c
) AND arguments from the table below
ABI | Extra cmake args | Restrictions |
---|---|---|
armeabi | -DANDROID_ABI=armeabi -DANDROID_TOOLCHAIN=gcc |
Requires CPU with ARM NEON |
armeabi-v7a | -DANDROID_ABI=armeabi-v7a -DANDROID_TOOLCHAIN=gcc |
Requires CPU with ARM NEON |
arm64-v8a | -DANDROID_ABI=arm64-v8a -DANDROID_TOOLCHAIN=clang |
Requires clang toolchain |
x86 | -DANDROID_ABI=x86 |
|
x86_64 | -DANDROID_ABI=x86_64 |
Notes:
- On armeabi and armeabi-v7a
nnp_initialize
will fail withnnp_status_unsupported_hardware
if the mobile CPU does not support ARM NEON. Don't set-DANDROID_ARM_NEON=1
for NNPACK compilation as it can makennp_initialize
crash on CPUs without ARM NEON. - NNPACK builds for armeabi and armeabi-v7a are up to 2x slower if you use clang toolchain.
- mips and mips64 are not supported, and we have no plans to add it (pull request would be welcome, though)
- x86_64 build will use generic 128-bit (SSE2) micro-kernels rather than AVX2 micro-kernels in native build
- PyTorch supports NNPACK on mobile for inference in convolutional layers.
- TVM supports NNPACK for inference in convolutional layers. See these instructions to enable NNPACK in TVM.
- MXNet supports NNPACK for inference in convolutional layers, fully-connected, and max-pooling layers. See MXNet wiki for configuration instructions and performance benchmarks).
- Caffe2 supports NNPACK for inference in convolutional layers.
- darknet-nnpack - fork of Darknet framework with NNPACK support.
- tiny-dnn - header-only deep learning framework in C++11, which natively supports NNPACK.
- Maratyszcza/caffe - up-to-date integration of NNPACK (convolutional, fully-connected, max-pooling, and ReLU layers) into Caffe based on
nnpack-pr
branch in ajtulloch/caffe. - Maratyszcza/caffe-nnpack - older and unmaintained integration of NNPACK (convolutional layers only) into Caffe.
- szagoruyko/nnpack.torch - integration of NNPACK into Lua Torch via ffi
- See also discussion in Issue #1
- nnpack-windows - unofficial port for Windows
- node-nnpack - Node.js bindings
- peterhj/libnnpack - Rust bindings
The library is developed by Marat Dukhan of Georgia Tech with extensive advice from Nicolas Vasilache and Soumith Chintala of Facebook Artificial Intelligence Research. Andrew Tulloch of Facebook Artificial Intelligence Research contributed Caffe integration. We thank Andrew Lavin for fruitful discussions on Winograd transform-based implementations. NNPACK is a research project at Richard Vuduc's HPC Garage lab in the Georgia Institute of Technology, College of Computing, School of Computational Science and Engineering.
This material is based upon work supported by the U.S. National Science Foundation (NSF) Award Number 1339745. Any opinions, findings and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect those of NSF.