Latest API version: v4.20181215
Backend.AI is a hassle-free backend for AI programming and service. It runs arbitrary user codes safely in resource-constrained environments, using Docker and our own sandbox wrapper.
Backend.AI supports various programming languages and runtimes, such as Python 2/3, R, PHP, C/C++, Java, Javascript, Julia, Octave, Haskell, Lua and NodeJS, as well as AI-oriented libraries such as TensorFlow, Keras, Caffe, and MXNet.
Product | Role | Problem and Solution |
---|---|---|
Apache Zeppelin, Jupyter Notebook | Notebook-style document + code front-ends | Insecure host resource sharing |
Backend.AI | Pluggable back-end to any front-ends | Built for multi-tenancy: scalable and better isolation |
Product | Target | Value |
---|---|---|
Amazon ECS, Kubernetes | Long-running service daemons | Load balancing, fault tolerance, incremental deployment |
Backend.AI | Stateful compute sessions | Low-cost high-density computation |
Amazon Lambda | Stateless, light-weight functions | Serverless, zero-management |
Product | Role | Problem and Solution |
---|---|---|
TensorFlow, Apache Spark, Apache Hive | Computation runtime | Difficult to install, configure, and operate |
Amazon ML, Azure ML, GCP ML | Managed MLaaS | Still complicated for scientists, too restrictive for engineers |
Backend.AI | Host of computation runtimes | Pre-configured, versioned, reproducible, customizable (open-source) |
(All product names and trade-marks are the properties of their respective owners.)
gsg/overview gsg/clientlib
install/overview install/demo-setup install/development-setup
common-api/convention common-api/auth common-api/ratelimit common-api/objects
user-api/intro user-api/kernels user-api/exec-stream user-api/exec-query user-api/exec-batch user-api/vfolders
admin-api/intro admin-api/keypairs admin-api/sessions admin-api/vfolders admin-api/stats
install/development-setup dev/repl
genindex
modindex
search