MShadow is a lightweight CPU/GPU Matrix/Tensor Template Library in C++/CUDA. The goal of mshadow is to support efficient, device invariant and simple tensor library for machine learning project that aims for maximum performance and control, while also emphasize simplicty.
MShadow also provides interface that allows writing Multi-GPU and distributed deep learning programs in an easy and unified way.
- Efficient: all the expression you write will be lazily evaluated and compiled into optimized code
- No temporal memory allocation will happen for expression you write
- mshadow will generate specific kernel for every expression you write in compile time.
- Device invariant: you can write one code and it will run on both CPU and GPU
- Simple: mshadow allows you to write machine learning code using expressions.
- Whitebox: put a float* into the Tensor struct and take the benefit of the package, no memory allocation is happened unless explicitly called
- Lightweight library: light amount of code to support frequently used functions in machine learning
- Extendable: user can write simple functions that plugs into mshadow and run on GPU/CPU, no experience in CUDA is required.
- MultiGPU and Distributed ML: mshadow-ps interface allows user to write efficient MultiGPU and distributed programs in an unified way.
- This version mshadow-2.x, there are a lot of changes in the interface and it is not backward compatible with mshadow-1.0
- If you use older version of cxxnet, you will need to use the legacy mshadow code
- For legacy code, refer to Here
- Change log in CHANGES.md
- CXXNET: large-scale deep learning backed by mshadow
- Parameter Server
- Parameter server project provides distributed back-end for mshadow-ps
- mshadow-ps extends original parameter server to support async updates for GPU Tensor