Skip to content

UT-ADL/e2e-rally-estonia

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Rally Estonia End-to-End Driving

This repository contains a framework to train End-to-End driving models using the Rally Estonia dataset. It includes scripts to train models as well dataset of over 750 km of driving in all seasons.

Articles and Theses

Following work has been produced using the framework:

Directory Structure

.
├── data_extract            # Code for extracting dataset from drive recordings
├── dataloading             # Dataloaders for training models with PyTorch
├── docs                    # Documentation, instructions for reproducing articles and theses
├── metrics                 # Code used to calculate open and closed loop metrics.
├── models                  # Models trained using the framework
└── notebooks               # Generic services
    ├── lidar-as-camera     # 'LiDAR-as-Camera for End-to-End Driving' article related notebooks
    ├── output-modalities   # 'Comparing Output Modalities in End-to-End Driving' master thesis related notebooks
├── tensorrt                # Inference with TensorRT library 
├── velocity_model          # Longitudinal model using predriven human trajectory to control vehicle's velocity.
├── viz                     # Scripts for interpret model's behavior using Visual BackProp, creating videos from drives with overlay etc.

Dataset

Dataset can be accessed in two forms: ROS bags containing all the raw information from the drives or the extracted dataset stored in the University of Tartu Rocket HPC. The latter option is currently only available for people with UT credentials. Without UT credentials, access to the dataset can be obtained by filling out this form. Once the ROS bags are downloaded, dataset can be extracted using scripts provided in data_extract package or custom extraction process can be developed. There is provided PyTorch data loaders in dataloading package for training models.

Following input modalities are supported:

  • Nvidia RGB cameras
  • Ouster LiDAR range, ambient, intensity image camera_crop

camera_crop

Following output modalities are supported:

  • Steering angle
  • Trajectory waypoints (longitudinal part fixed)

Current models only predict lateral control (steering), longitudinal control (throttle) is not predicted and must be controlled using other means.

Training

Environment setup

Environment can set up using conda by following commands:

# Set up Pytorch environment
conda create -n e2e pytorch torchvision cudatoolkit=11.1 jupyter pandas matplotlib tqdm scikit-learn scikit-image onnx seaborn -c pytorch -c nvidia
conda activate e2e

# Install TensorRT and pycuda
pip install nvidia-tensorrt==7.2.3.4
pip install 'pycuda<2021.1'

# Wandb
pip install wandb

# For visualising predictions
pip install opencv-contrib-python
# need to use specific version of moviepy as newer version did not work
pip install moviepy==1.0.0 

Run training

Model can be trained using following command:

python train.py --input-modality nvidia-camera --output-modality steering_angle --patience 10 --max-epochs 100 --model-name steering-angle --model-type pilotnet-conditional --wandb-project summer-models-6 --dataset-folder <path to extracted dataset>

Use --input-modality parameter to train using camera or lidar images.

Use --output-modality parameter to use different prediction targets like steering angle or trajectory waypoints.

Use --model-type parameter to use different model architectures like pilotnet-conditional and pilotnet-control.

Use --wandb-project parameter to use log using W&B. To use without W&B, just omit this parameter.

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages