Skip to content
develop
Switch branches/tags
Code

Latest commit

…35510)

* Create stateful OneDNNAXPYHandler object.

This makes it possible to call it multiple times without recreating the
oneDNN primitives every time.

* Prepare SGDOpKernel to reuse its implementation from OneDNN kernel.

* OneDNN SGD kernel.

* Update call to use new OneDNNAXPYHandler object api.

* Setup seed in proper place.

* Enable OneDNN kernel only for single case.

* For dense param and sparse grad.

* Small refactor.

* Enable oneDNN by op attr or by cmd line flag.

* Use int64_t type for number of elements.

* Support dense param and grad from OneDNN kernel.

* Enable SGD OneDNN kernel when use MP BF16 optimizer.

* Force non-copyable/movable OneDNNAXPYHandler.

* Reuse OneDNNAXPYHandler for spare tensors in SUM op.

* Fix SFINAE rules.

* Remove recording event inside AXPY.

* Get rid of internal primitive caching.

* Stop use PP cache mechanims to store mem and primitive obj.
* Handler obj store and reuse needed desc & prim

* Do not derive from MKLDNNHandlerT
799f386

Files

Permalink
Failed to load latest commit information.
Type
Name
Latest commit message
Commit time


English | 简体中文

Build Status Documentation Status Documentation Status Release License

Welcome to the PaddlePaddle GitHub.

PaddlePaddle, as the only independent R&D deep learning platform in China, has been officially open-sourced to professional communities since 2016. It is an industrial platform with advanced technologies and rich features that cover core deep learning frameworks, basic model libraries, end-to-end development kits, tools & components as well as service platforms. PaddlePaddle is originated from industrial practices with dedication and commitments to industrialization. It has been widely adopted by a wide range of sectors including manufacturing, agriculture, enterprise service, and so on while serving more than 2.3 million developers. With such advantages, PaddlePaddle has helped an increasing number of partners commercialize AI.

Installation

Latest PaddlePaddle Release: v2.1

Our vision is to enable deep learning for everyone via PaddlePaddle. Please refer to our release announcement to track the latest features of PaddlePaddle.

Install Latest Stable Release:

# CPU
pip install paddlepaddle
# GPU
pip install paddlepaddle-gpu

For more information about installation, please view Quick Install

Now our developers can acquire Tesla V100 online computing resources for free. If you create a program by AI Studio, you will obtain 8 hours to train models online per day. Click here to start.

FOUR LEADING TECHNOLOGIES

  • Agile Framework for Industrial Development of Deep Neural Networks

    The PaddlePaddle deep learning framework facilitates the development while lowering the technical burden, through leveraging a programmable scheme to architect the neural networks. It supports both declarative programming and imperative programming with both development flexibility and high runtime performance preserved. The neural architectures could be automatically designed by algorithms with better performance than the ones designed by human experts.

  • Support Ultra-Large-Scale Training of Deep Neural Networks

    PaddlePaddle has made breakthroughs in ultra-large-scale deep neural networks training. It launched the world's first large-scale open-source training platform that supports the training of deep networks with 100 billion features and trillions of parameters using data sources distributed over hundreds of nodes. PaddlePaddle overcomes the online deep learning challenges for ultra-large-scale deep learning models, and further achieved real-time model updating with more than 1 trillion parameters. Click here to learn more

  • High-Performance Inference Engines for Comprehensive Deployment Enviroments

    PaddlePaddle is not only compatible with models trained in 3rd party open-source frameworks , but also offers complete inference products for various production scenarios. Our inference product line includes Paddle Inference: Native inference library for high-performance server and cloud inference; Paddle Serving: A service-oriented framework suitable for distributed and pipeline productions; Paddle Lite: Ultra-Lightweight inference engine for mobile and IoT environments; Paddle.js: A frontend inference engine for browser and mini-apps. Furthermore, by great amounts of optimization with leading hardware in each scenario, Paddle inference engines outperform most of the other mainstream frameworks.

  • Industry-Oriented Models and Libraries with Open Source Repositories

    PaddlePaddle includes and maintains more than 100 mainstream models that have been practiced and polished for a long time in the industry. Some of these models have won major prizes from key international competitions. In the meanwhile, PaddlePaddle has further more than 200 pre-training models (some of them with source codes) to facilitate the rapid development of industrial applications. Click here to learn more

Documentation

We provide English and Chinese documentation.

  • Guides

    You might want to start from how to implement deep learning basics with PaddlePaddle.

  • Practice

    So far you have already been familiar with Fluid. And the next step should be building a more efficient model or inventing your original Operator.

  • API Reference

    Our new API enables much shorter programs.

  • How to Contribute

    We appreciate your contributions!

Communication

  • Github Issues: bug reports, feature requests, install issues, usage issues, etc.
  • QQ discussion group: 793866180 (PaddlePaddle).
  • Forums: discuss implementations, research, etc.

Courses

  • Server Deployments: Courses intorducing high performance server deployments via local and remote services.
  • Edge Deployments: Courses intorducing edge deployments from mobile, IoT to web and applets.

Copyright and License

PaddlePaddle is provided under the Apache-2.0 license.