Deep learning inference with Homomorphic Encryption (HE) through Intel nGraph
fboemer Fboemer/ngraph tf integration1 (#124)
* Update to latest ng-tf v0.9 / ng v 0.11.0

* Updates required compiler to GCC7. Clang not tested.

* Requires building tensorflow from source, slowing compilation times.

 - In future PR, we plan to enable using pre-built TF to speed up compilation. This would require editing `_D_GLIB_CXX11_ABI` flag, so we omit it here.

* Uses `ngraph-tf` to build `ngraph`

* Moves to `HE_SEAL_CKKS` instead of `HE:SEAL:CKKS` conventions

* Remove docker build for now. In future, will add and enable CI
Latest commit 805bcf5 Dec 15, 2018

HE Transformer for nGraph

The Intel® HE transformer for nGraph™ is a Homomorphic Encryption (HE) backend to the Intel® nGraph Compiler, Intel's graph compiler for Artificial Neural Networks.

Homomorphic encryption is a form of encryption that allows computation on encrypted data, and is an attractive remedy to increasing concerns about data privacy in the field of machine learning. For more information, see our paper.

This project is meant as a proof-of-concept to demonstrate the feasibility of HE on local machines. The goal is to measure performance of various HE schemes for deep learning. This is not intended to be a production-ready product, but rather a research tool.

Currently, we support two encryption schemes, implemented by the Simple Encrypted Arithmetic Library (SEAL) from Microsoft Research:

Additionally, we integrate with the Intel® nGraph™ Compiler and runtime engine for TensorFlow to allow users to run inference on trained neural networks through Tensorflow.


The examples directory contains a deep learning example which depends on the Intel® nGraph™ Compiler and runtime engine for TensorFlow.

Building HE Transformer


  • We currently only support Ubuntu 16.04
  • CMake >= 3.10, although different versions may work
  • GCC version 7, although different versions may work
  • OpenMP is strongly suggested, though not strictly necessary. You may experience slow runtimes without OpenMP
  • virtualenv
  • bazel v0.16.0

The following dependencies are built automatically

To install bazel

 chmod +x
 ./ --user

Make sure to add and source the bin path to your ~/.bashrc file in order to be able to call bazel from the user's installation we set up:

 export PATH=$PATH:~/bin
 source ~/.bashrc

1. Build HE-Transformer

Before building, make sure you deactivate any active virtual environments (i.e. run deactivate)

git clone
cd he-transformer
export HE_TRANSFORMER=$(pwd)
mkdir build
make -j install
source external/venv-tf-py3/bin/activate

2. Run C++ unit-tests

Ensure the virtual environment is active, i.e. run source $HE_TRANSFORMER /external/venv-tf-py3/bin/activate

# To run CKKS unit-test
./test/unit-test --gtest_filter="HE_SEAL_CKKS.*abc*"
# To run BFV unit-test
./test/unit-test --gtest_filter="HE_SEAL_BFV.*abc*
# To run all C++ unit-tests

3. Run Simple python example

Ensure the virtual environment is active, i.e. run source $HE_TRANSFORMER /external/venv-tf-py3/bin/activate

cd $HE_TRANSFORMER/examples
# Run with CPU
# To run CKKS unit-test
# To run BFV unit-test

For a deep learning example, see examples/cryptonets/.

Code formatting

Please run maint/ before submitting a pull request.