Deep learning gateway on Raspberry Pi
JavaScript Python Shell CSS
Latest commit 13a51ec Oct 9, 2017 Bofu Chen (bafu) Bofu Chen (bafu) Merge branch 'develop'
Permalink
Failed to load latest commit information.
config Make OpenCV to access RPi camera via V4L2. Sep 20, 2017
doc Update document for Nest camera setup. May 30, 2017
inference Fix inference result parsing issue. Oct 9, 2017
patch Make faster detection backend as systemd service. Sep 23, 2017
systemd Merge branch 'feature-line-notification' into develop Oct 8, 2017
test Init. Apr 27, 2017
utils Merge branch 'develop' Oct 9, 2017
.eslintrc.yaml Use ESLint as linter. May 8, 2017
.gitmodules Add darkflow as Git submodule. May 20, 2017
AUTHORS Add AUTHORS. Apr 27, 2017
CONTRIBUTING.md Add CONTRIBUTING.md (use DCO instead of CLA). Sep 4, 2017
LICENSE.txt Init. Apr 27, 2017
README.md Manually merged Paul's Nest streaming support: Sep 30, 2017
berrynet-manager Merge branch 'feature-line-notification' into develop Oct 8, 2017
broker.js Use ESLint as linter. May 8, 2017
camera.js Manually merged Paul's Nest streaming support: Sep 30, 2017
client.js Respect coding style. Sep 19, 2017
config.js Convert data collector's input text data format to JSON. Oct 9, 2017
configure Merge branch 'feature-line-notification' into develop Oct 8, 2017
data_collector.js Convert data collector's input text data format to JSON. Oct 9, 2017
journal.js Use the result image from object detection instead of camera May 30, 2017
line.js Merge branch 'feature-line-notification' into develop Oct 8, 2017
localimg.js Use ESLint as linter. May 8, 2017
mail.js Respect coding style. Sep 19, 2017
package.json Merge branch 'feature-line-notification' into develop Oct 8, 2017

README.md

BerryNet: Deep Learning Gateway on Raspberry Pi

This project turns Raspberry Pi 3 into an intelligent gateway with deep learning running on it. No internet connection is required, everything is done locally on the Raspberry Pi 3 itself. At DT42, we believe that bringing deep learning to edge devices is the trend towards the future. It not only saves costs of data transmission and storage but also makes devices able to respond according to the events shown in the images or videos without connecting to the cloud.

Figure 1

Figure 1 shows the software architecture of the project, we use Node.js, MQTT and an AI engine to analyze images or video frames with deep learning. So far, there are two supported AI engines, the classification engine and the object detection engine. Figure 2 shows the differences between classification and object detection.

Figure 2

One of the application of this intelligent gateway is to use the camera to monitor the place you care about. For example, Figure 3 shows the analyzed results from the camera hosted in the DT42 office. The frames were captured by the IP camera and they were submitted into the AI engine. The output from the AI engine will be shown in the dashboard. We are working on the Email and IM notification so you can get a notification when there is a dog coming into the meeting area with the next release.

Figure 3

AI Engines

The current supported AI Engines leverage work from the following projects:

The current supported classification model is Inception v3 [1] and the detection model is TinyYOLO [2]

Installation

$ git clone https://github.com/DT42/BerryNet.git
$ cd BerryNet
$ ./configure

Start and Stop BerryNet

BerryNet is managed by systemd. You can manage BerryNet via berrynet-manager:

$ berrynet-manager [start | stop | status | log]

Configuration

All the configurations are in config.js.

  • Choose AI Engine.

    • Two types of AI engines currently: object classifier and object detector.
  • Configure IP camera's snapshot access interface.

  • MQTT topics.

Dashboard

Open dashboard on RPi3 (with touch screen)

Open browser and enter the URL:

http://localhost:8080/index.html#source=dashboard.json

The default dashboard configuration file will be loaded.

Open dashboard on browser from any computer

Open browser and enter the URL:

http://<gateway-ip>:8080/index.html#source=dashboard.json

Click the data sources, and change MQTT broker's IP address to the gateway's IP.

For more details about dashboard configuration (e.g. how to add widgets), please refer to freeboard project.

Provide Image Input

To capture an image via configured IP camera

$ mosquitto_pub -h localhost -t berrynet/event/camera -m snapshot_ipcam

To capture an image via board-connected camera (RPi camera or USB webcam)

$ mosquitto_pub -h localhost -t berrynet/event/camera -m snapshot_boardcam

To provide a local image

$ mosquitto_pub -h localhost -t berrynet/event/localImage -m <image_path>

To start and stop streaming from board-connected camera

$ mosquitto_pub -h localhost -t berrynet/event/camera -m stream_boardcam_start
$ mosquitto_pub -h localhost -t berrynet/event/camera -m stream_boardcam_stop

To start and stop streaming from Nest IP camera

$ mosquitto_pub -h localhost -t berrynet/event/camera -m stream_nest_ipcam_start
$ mosquitto_pub -h localhost -t berrynet/event/camera -m stream_nest_ipcam_stop

Enable Data Collector

You might want to store the snapshot and inference results for data analysis.

To enable data collector, you can set the storage directory path in config.js:

config.storageDirPath = '<data-storage-dirpath>';

and restart BerryNet.

Use Your Data To Train

The original instruction of retraining YOLOv2 model see github repository of darknet

In the current of BerryNet, TinyYolo is used instead of YOLOv2. The major differences are:

  1. Create file yolo-obj.cfg with the same content as in tiny-yolo.cfg
  2. Download pre-trained weights of darknet reference model, darknet.weights.12, for the convolutional layers (6.1MB) https://drive.google.com/drive/folders/0B-oZJEwmkAObMzAtc2QzZDhyVGM?usp=sharing

The rest parts are the same as retraining YOLO.

If you use LabelMe to annotate data, utils/xmlTotxt.py can help convert the xml format to the text format that darknet uses.

Discussion

Please refer to the Google Group for questions, suggestions, or any idea discussion.