This project aims to build an autonomous rc car using supervised learning of a neural network with a single hidden layer. We have not used any Machine Learning libraries since we wanted to implement the neural network from scratch to understand the concepts better. We have modified a remote controlled car to remove the dependency on the RF remote controller. A Raspberry Pi controls the DC motors via an L293D Motor Driver IC. You can find a post explaining this project in detail here. Here's a video of the car in action.
We will be referring the DC motor controlling the left/right direction as the front motor and the motor controlling the forward/reverse direction as the back motor. Connect the
BACK_MOTOR_DATA_TWO GPIO pins(
GPIO27) of the Raspberry Pi to the Input pins for Motor 1(
Input 2) and the
BACK_MOTOR_ENABLE_PIN GPIO pin(
GPIO22) to the Enable pin for Motor 1(
Enable 1,2) in the L293D Motor Driver IC. Connect the Output pins for Motor 1(
Output 2) of the IC to the back motor.
FRONT_MOTOR_DATA_TWO GPIO pins(
GPIO26) of the Raspberry Pi to the Input pins for Motor 2(
Input 4) in the IC. Connect the Output pins for Motor 2(
Output 4) of the IC to the front motor.
INITIAL_PWM_DUTY_CYCLE represent the initial frequency and duty cycle of the PWM output.
We have created five class labels namely
idle and assigned their expected values. All class labels would require a folder of the same name to be present in the current directory.
The input images resize to the dimension of the
IMAGE_DIMENSION tuple value during training. The
HIDDEN_LAYER_SIZE values represent the default lambda value and the number of nodes in the hidden layer while training the neural network.
All these values are configurable in
The images for training are captured using
interactive_control_train.py, the car is controlled using the direction arrows and all the images are recorded in the same folder along with the corresponding key press. At the command prompt, run the following command:
Data cleaning is done before segregating the images into their respective class folders based on the key press indicated in their filenames.
After segregating the images into their corresponding class folders, the neural network is trained using
train.py which takes two optional arguments -
hidden layer size; default values would be those specified in the configuration file. At the command prompt, run the following command:
python train.py 0.1 100
The images are loaded from the corresponding class folders and are assigned the class values indicated in the configuration file. The generated model is stored in the
optimized_thetas folder as a pickle file. Here's a sample dataset and trained model to get you started.
Once we have the trained model, the RC car is run autonomously using
autonomous.py which takes an optional argument for the trained model; default will use the latest model in the
optimized_thetas folder. At the command prompt, run the following command:
We are reducing the speed of the car when the model predicts a turn. The reduction is not being made by the neural network, though we hope to add a speed component in the future.
We intend to add the following capabilities in the future:
- Control speed of the car using neural networks
- Stop signal detection
- Obstacle detection
autonomous-rc-car is maintained and funded by Multunus Software Pvt. Ltd. The names and logos for Multunus are trademarks of Multunus Software Pvt. Ltd.
We built this to explore and learn Machine Learning concepts in our 20% investment time. We will be supporting this project during our investment time.