Skip to content
forked from dmlc/nnvm

Intermediate Computational Graph Representation for Deep Learning Systems

License

Notifications You must be signed in to change notification settings

zhang3wind/nnvm

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

53 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

NNVM: Build deep learning system by parts

Build Status GitHub license

NNVM is not a deep learning library. It is a modular, decentralized and lightweight part to help build deep learning libraries.

What is it

While most deep learning systems offer end to end solutions, it is interesting to assemble a deep learning system by parts. The goal is to enable user to customize optimizations, target platforms and set of operators they care about. We believe that the decentralized modular system is an interesting direction.

The hope is that effective parts can be assembled together just like you assemble your own desktops. So the customized deep learning solution can be minimax, minimum in terms of dependencies, while maxiziming the users' need.

NNVM offers one such part, it provides a generic way to do computation graph optimization such as memory reduction, device allocation and more while being agnostic to the operator interface defintion and how operators are executed. NNVM is inspired by LLVM, aiming to be a high level intermediate representation library for neural nets and computation graphs generation and optimizations.

See Overview for an introduction on what it provides.

Why build learning system by parts

This is essentially Unix philosophy applied to machine learning system.

  • Essential parts can be assembled in minimum way for embedding systems.
  • Developers can hack the parts they need and compose with other well defined parts.
  • Decentralized modules enable new extensions creators to own their project without creating a monothilic version.

Deep learning system itself is not necessary one part, for example here are some relative independent parts that can be isolated

  • Computation graph definition, manipulation.
  • Computation graph intermediate optimization.
  • Computation graph execution.
  • Operator kernel libraries.
  • Imperative task scheduling and parallel task coordination.

We hope that there will be more modular parts in the future, so system building can be fun and rewarding.

Links

MXNet is moving to NNVM as its intermediate representation layer for symbolic graphs.

About

Intermediate Computational Graph Representation for Deep Learning Systems

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Languages

  • C++ 84.1%
  • Python 14.7%
  • Other 1.2%