DRIVERLESS RC CAR
Because real cars are too expensive for self-driving car studies, we are building a self-driving RC car as a testbed. When you have
- an RC car with a steering servo,
- an Android device with WiFi, Camera and Bluetooth support (we use Samsung Galaxy S3 Mini),
- a selfie stick,
- a breadboard and many jumpers,
- an Arduino (we have an Arduino UNO),
- an HC-06 bluetooth module,
- a powerful GPU at your disposal (we use GTX 1070)
you are good to go. We also have a speed sensor, but it is only for monitoring purposes, you don't need it to train and drive your RC car.
While a human controlling the car through a transmitter, Arduino reads steering and speed commands, which are essentially servo signals, from the receiver. It also computes steering angles (just a linear mapping from steering commands to angles in degrees) and measures speed in m/s from the speed sensor. Arduino eventually sends them to the Android device installed on the car. We use bluetooth for the communication between Arduino and Android. Android device attaches a current image frame to the Arduino readings and forwards them all to a remote machine over TCP/IP. We capture 20 frames per second along with the Arduino readings. We had to create an access point on the remote machine to make the communication more reliable.
This is how our dataset looks like. The commands marked in red are the steering commands. We have speed commands (i.e. throttle) and image file names on the right and left of the red rectangle, respectively.
A row in
dataset.txt is as follows:
Teaching a Machine to Steer a Car
You might have heard about Udacity's open source self-driving car project. They organize various challenges to build the car step by step. Their steering control challenge led to brilliant deep learning models and we implemented the winner model by Team Komanda for our RC car. It has been a great starting point for our self-driving RC car journey. It should also be possible to experiment with other existing steering control models or develop and try out our own models using our RC car setup.
One difference between our training strategy and original Komanda model is that we augment our training set by vertically flipping images and mirroring their corresponding steering commands and angles with respect to the neutral position before training. We don't use steering angles for training, they are used for monitoring purposes just like speed values in m/s. Another difference is the size of input images. Our images have a size of 176x144, which was the minimum possible image size from our Android device. The model tries to predict both steering and speed commands given the input images.
You can see the results in the following video.
carinocontains Arduino-specific code.
caroidcontains Android-specific code.
stationcontains dataset acquisition code and our deep learning agent(s) on remote machine side. It also features various visualization tools.