Distributed (Deep) Machine Learning Community's Projects
Caffe: a fast open framework for deep learning.
ccache – a fast compiler cache
move forward to https://github.com/dmlc/mxnet
An efficient video loader for deep learning with smart shuffling that's super easy to digest
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Python package built to ease deep learning on graph, on top of existing DL frameworks.
Distributed Factorization Machines
common in-memory tensor structure
A common bricks library for building scalable and portable distributed machine learning.
redirect mxnet.readthedocs.io to mxnet.io
Drat Repository for DMLC R packages
cache-friendly multithread matrix factorization
Gluon CV Toolkit
NLP made easy
Visualization tool for Graph Neural Networks
Symbolic Expression and Statement Module for new DSLs
Deep Learning library for Python. Convnets, recurrent neural networks, and more. Runs on MXNet, Theano or TensorFlow.
Minerva: a fast and flexible tool for deep learning on multi-GPU. It provides ndarray programming interface, just like Numpy. Python bindings and C++ bindings are both available. The resulting code can be run on CPU or GPU. Multi-GPU support is very easy.
NumPy interface with mixed backend execution
Matrix Shadow:Lightweight CPU/GPU Matrix and Tensor Template Library in C++/CUDA for (Deep) Machine Learning
Benchmark speed and other issues internally, before push to deep-mark
MXNet Example
MXNet Tutorial for NVidia GTC 2016.
Sublinear memory optimization for deep learning, reduce GPU memory cost to train deeper nets
Pre-trained Models of DMLC Project
Notebooks for MXNet