Skip to content

🔘Official codes for "SocialCircle+: Learning the Angle-based Conditioned Interaction Representation for Pedestrian Trajectory Prediction"

License

Notifications You must be signed in to change notification settings

cocoon2wong/SocialCirclePlus

Repository files navigation

SocialCirclePlus

This is the official PyTorch codes for our paper "SocialCircle+: Learning the Angle-based Conditioned Interaction Representation for Pedestrian Trajectory Prediction".

SocialCircle+ is an extended journal-version of our previous work SocialCircle, and this paper will be made public soon. For our pre-trained model weights, please refer to this page.

Get Started

You can clone this repository by the following command:

git clone https://github.com/cocoon2wong/SocialCirclePlus.git

Then, run the following command to initialize all submodules:

git submodule update --init --recursive

Requirements

The codes are developed with Python 3.10. Additional packages used are included in the requirements.txt file.

{: .box-warning} Warning: We recommend installing all required Python packages in a virtual environment (like the conda environment). Otherwise, there COULD be other problems due to the package version conflicts.

Run the following command to install the required packages in your Python environment:

pip install -r requirements.txt

Dataset Prepare and Process

ETH-UCY, SDD, NBA, nuScenes

{: .box-warning} Warning: If you want to validate SocialCirclePlus models on these datasets, make sure you are getting this repository via git clone and that all gitsubmodules have been properly initialized via git submodule update --init --recursive.

You can run the following commands to prepare dataset files that have been validated in our paper:

  1. Run Python the script inner the dataset_original folder:

    cd dataset_original
    • For ETH-UCY and SDD, run

      python main_ethucysdd.py
    • For NBA or nuScenes, you can download their original dataset files, put them into the given path listed within dataset_original/main_nba.py or dataset_original/main_nuscenes.py, then run

      python main_nba.py
      python main_nuscenes.py
  2. Back to the repo folder and create soft links:

    cd ..
    ln -s dataset_original/dataset_processed ./
    ln -s dataset_original/dataset_configs ./

(NOTE: You can also download the processed dataset files manually from here, and put them into dataset_processed and dataset_configs folders.)

Click the following button to learn how we process these dataset files and the detailed dataset settings.

Prepare Your New Datasets

Before training SocialCirclePlus models on your own dataset, you should add your dataset information. See this document for details.

Pre-Trained Model Weights and Evaluation

We have provided our pre-trained model weights to help you quickly evaluate the SocialCirclePlus models' performance.

Click the following buttons to download our model weights. We recommend that you download the weights and place them in the weights/SocialCirclePlus folder.

You can start evaluating models by

python main.py --sc SOME_MODEL_WEIGHTS

Here, SOME_MODEL_WEIGHTS is the path of the weights folder, for example, weights/SocialCirclePlus/evspczara1_adaptive.

Training

You can start training a SocialCirclePlus model via the following command:

python main.py --model MODEL_IDENTIFIER --split DATASET_SPLIT

Here, MODEL_IDENTIFIER is the identifier of the model. These identifiers are supported in current codes:

  • The basic transformer model for trajectory prediction:
    • trans (named the Transformer in the paper);
    • transsc (SocialCircle variation Transformer-SC);
    • transscp (SocialCircle+ variation Transformer-SCP);
  • MSN (🔗homepage):
    • msna (original model);
    • msnsc (SocialCircle variation);
    • msnscp (SocialCircle+ variation).
  • V^2-Net (🔗homepage):
    • va (original model);
    • vsc (SocialCircle variation);
    • vscp (SocialCircle+ variation)
  • E-V^2-Net (🔗homepage):
    • eva (original model);
    • evsc (SocialCircle variation);
    • evscp (SocialCircle+ variation)

DATASET_SPLIT is the identifier (i.e., the name of dataset's split files in dataset_configs, for example eth is the identifier of the split list in dataset_configs/ETH-UCY/eth.plist) of the dataset or splits used for training. It accepts:

  • ETH-UCY: {eth, hotel, univ, zara1, zara2};
  • SDD: sdd;
  • NBA: nba50k.

For example, you can start training the E-V^2-Net-SC+ model by

python main.py --model evscp --split zara1

You can also specify other needed args, like the learning rate --lr, batch size --batch_size, etc. See detailed args in the Args Used Section.

In addition, the simplest way to reproduce our results is to copy all training args we used in the provided weights. For example, you can start a training of E-V^2-Net-SC+ on zara1 by:

python main.py --restore_args weights/SocialCirclePlus/evspczara1_adapative

Toy Example

You can run the following script to learn how the proposed SocialCirclePlus works in an interactive way:

python scripts/socialcircle_toy_example.py

In the toy example window, you can click Switch Mode to switch into three modes:

SC Mode PLT Mode PC Mode
Mode 1 Mode 2 Mode 3
  • Interactive(SC) You can directly click on the canvas or type in coordinates to set positions of the manual neighbor to see the model's outputs*;

  • PLT Mode You can type in coordinates to set positions of the manual neighbor to see the model's outputs without the scene image in matplotlib.pyplot mode;

  • Interactive(PC) You can directly click on the canvas or type in coordinates to set a pair of corners' positions in pixels to add a manual obstacle box to see the model's outputs*.

* These modes may need dataset videos. For copyright reasons, we do not provide them in our repo.

Args Used

Please specify your customized args when training or testing your model in the following way:

python main.py --ARG_KEY1 ARG_VALUE2 --ARG_KEY2 ARG_VALUE2 -SHORT_ARG_KEY3 ARG_VALUE3 ...

where ARG_KEY is the name of args, and ARG_VALUE is the corresponding value. All args and their usages are listed below.

About the argtype:

  • Args with argtype=static can not be changed once after training. When testing the model, the program will not parse these args to overwrite the saved values.
  • Args with argtype=dynamic can be changed anytime. The program will try to first parse inputs from the terminal and then try to load from the saved JSON file.
  • Args with argtype=temporary will not be saved into JSON files. The program will parse these args from the terminal at each time.

Basic Args

  • --K_train: type=int, argtype=static. The number of multiple generations when training. This arg only works for multiple-generation models. The default value is 10.
  • --K: type=int, argtype=dynamic. The number of multiple generations when testing. This arg only works for multiple-generation models. The default value is 20.
  • --anntype: type=str, argtype=static. Model's predicted annotation type. Can be 'coordinate' or 'boundingbox'. The default value is coordinate.
  • --auto_clear: type=int, argtype=temporary. Controls whether to clear all other saved weights except for the best one. It performs similarly to running python scripts/clear.py --logs logs. The default value is 1.
  • --batch_size (short for -bs): type=int, argtype=dynamic. Batch size when implementation. The default value is 5000.
  • --compute_loss: type=int, argtype=temporary. Controls whether to compute losses when testing. The default value is 0.
  • --dataset: type=str, argtype=static. Name of the video dataset to train or evaluate. For example, 'ETH-UCY' or 'SDD'. NOTE: DO NOT set this argument manually. The default value is Unavailable.
  • --draw_results (short for -dr): type=str, argtype=temporary. Controls whether to draw visualized results on video frames. Accept the name of one video clip. The codes will first try to load the video file according to the path saved in the plist file (saved in dataset_configs folder), and if it loads successfully it will draw the results on that video, otherwise it will draw results on a blank canvas. Note that test_mode will be set to 'one' and force_split will be set to draw_results if draw_results != 'null'. The default value is null.
  • --draw_videos: type=str, argtype=temporary. Controls whether to draw visualized results on video frames and save them as images. Accept the name of one video clip. The codes will first try to load the video according to the path saved in the plist file, and if successful it will draw the visualization on the video, otherwise it will draw on a blank canvas. Note that test_mode will be set to 'one' and force_split will be set to draw_videos if draw_videos != 'null'. The default value is null.
  • --epochs: type=int, argtype=static. Maximum training epochs. The default value is 500.
  • --experimental: type=bool, argtype=temporary. NOTE: It is only used for code tests. The default value is False.
  • --feature_dim: type=int, argtype=static. Feature dimensions that are used in most layers. The default value is 128.
  • --force_anntype: type=str, argtype=temporary. Assign the prediction type. It is now only used for silverballers models that are trained with annotation type coordinate but to be tested on datasets with annotation type boundingbox. The default value is null.
  • --force_clip: type=str, argtype=temporary. Force test video clip (ignore the train/test split). It only works when test_mode has been set to one. The default value is null.
  • --force_dataset: type=str, argtype=temporary. Force test dataset (ignore the train/test split). It only works when test_mode has been set to one. The default value is null.
  • --force_split: type=str, argtype=temporary. Force test dataset (ignore the train/test split). It only works when test_mode has been set to one. The default value is null.
  • --gpu: type=str, argtype=temporary. Speed up training or test if you have at least one NVidia GPU. If you have no GPUs or want to run the code on your CPU, please set it to -1. NOTE: It only supports training or testing on one GPU. The default value is 0.
  • --help (short for -h): type=str, argtype=temporary. Print help information on the screen. The default value is null.
  • --input_pred_steps: type=str, argtype=static. Indices of future time steps that are used as extra model inputs. It accepts a string that contains several integer numbers separated with '_'. For example, '3_6_9'. It will take the corresponding ground truth points as the input when training the model, and take the first output of the former network as this input when testing the model. Set it to 'null' to disable these extra model inputs. The default value is null.
  • --interval: type=float, argtype=static. Time interval of each sampled trajectory point. The default value is 0.4.
  • --load (short for -l): type=str, argtype=temporary. Folder to load model (to test). If set to null, the training manager will start training new models according to other given args. The default value is null.
  • --log_dir: type=str, argtype=static. Folder to save training logs and model weights. Logs will save at args.save_base_dir/current_model. DO NOT change this arg manually. (You can still change the path by passing the save_base_dir arg.) The default value is Unavailable.
  • --lr (short for -lr): type=float, argtype=static. Learning rate. The default value is 0.001.
  • --macos: type=int, argtype=temporary. (Experimental) Choose whether to enable the MPS (Metal Performance Shaders) on Apple platforms (instead of running on CPUs). The default value is 0.
  • --max_agents: type=int, argtype=static. Max number of agents to predict per frame. It only works when model_type == 'frame-based'. The default value is 50.
  • --model_name: type=str, argtype=static. Customized model name. The default value is model.
  • --model_type: type=str, argtype=static. Model type. It can be 'agent-based' or 'frame-based'. The default value is agent-based.
  • --model: type=str, argtype=static. The model type used to train or test. The default value is none.
  • --noise_depth: type=int, argtype=static. Depth of the random noise vector. The default value is 16.
  • --obs_frames (short for -obs): type=int, argtype=static. Observation frames for prediction. The default value is 8.
  • --output_pred_steps: type=str, argtype=static. Indices of future time steps to be predicted. It accepts a string that contains several integer numbers separated with '_'. For example, '3_6_9'. Set it to 'all' to predict points among all future steps. The default value is all.
  • --pmove: type=int, argtype=static. (Pre/post-process Arg) Index of the reference point when moving trajectories. The default value is -1.
  • --pred_frames (short for -pred): type=int, argtype=static. Prediction frames. The default value is 12.
  • --preprocess: type=str, argtype=static. Controls whether to run any pre-process before the model inference. It accepts a 3-bit-like string value (like '111'): - The first bit: MOVE trajectories to (0, 0); - The second bit: re-SCALE trajectories; - The third bit: ROTATE trajectories. The default value is 100.
  • --restore_args: type=str, argtype=temporary. Path to restore the reference args before training. It will not restore any args if args.restore_args == 'null'. The default value is null.
  • --restore: type=str, argtype=temporary. Path to restore the pre-trained weights before training. It will not restore any weights if args.restore == 'null'. The default value is null.
  • --save_base_dir: type=str, argtype=static. Base folder to save all running logs. The default value is ./logs.
  • --split (short for -s): type=str, argtype=static. The dataset split that used to train and evaluate. The default value is zara1.
  • --start_test_percent: type=float, argtype=temporary. Set when (at which epoch) to start validation during training. The range of this arg should be 0 <= x <= 1. Validation may start at epoch args.epochs * args.start_test_percent. The default value is 0.0.
  • --step: type=float, argtype=dynamic. Frame interval for sampling training data. The default value is 1.0.
  • --test_mode: type=str, argtype=temporary. Test settings. It can be 'one', 'all', or 'mix'. When setting it to one, it will test the model on the args.force_split only; When setting it to all, it will test on each of the test datasets in args.split; When setting it to mix, it will test on all test datasets in args.split together. The default value is mix.
  • --test_step: type=int, argtype=temporary. Epoch interval to run validation during training. The default value is 1.
  • --update_saved_args: type=int, argtype=temporary. Choose whether to update (overwrite) the saved arg files or not. The default value is 0.
  • --verbose (short for -v): type=int, argtype=temporary. Controls whether to print verbose logs and outputs to the terminal. The default value is 0.

V^2-Net Args

  • --Kc: type=int, argtype=static. The number of style channels in Agent model. The default value is 20.
  • --T (short for -T): type=str, argtype=static. Type of transformations used when encoding or decoding trajectories. It could be: - none: no transformations - fft: fast Fourier transform - fft2d: 2D fast Fourier transform - haar: haar wavelet transform - db2: DB2 wavelet transform The default value is fft.

SocialCircle Args

  • --Ts (short for -Ts): type=str, argtype=static. The transformation on SocialCircle. It could be: - none: no transformations - fft: fast Fourier transform - haar: haar wavelet transform - db2: DB2 wavelet transform The default value is none.
  • --partitions: type=int, argtype=static. Partitions in the SocialCircle. It should be manually set at each training run. The default value is -1.
  • --rel_speed: type=int, argtype=static. Choose whether to use the relative speed or the absolute speed as the speed factor in the SocialCircle. (Default to the absolute speed) The default value is 0.
  • --use_direction: type=int, argtype=static. Choose whether to use the direction factor in the SocialCircle. The default value is 1.
  • --use_distance: type=int, argtype=static. Choose whether to use the distance factor in the SocialCircle. The default value is 1.
  • --use_move_direction: type=int, argtype=static. Choose whether to use the move direction factor in the SocialCircle. The default value is 0.
  • --use_velocity: type=int, argtype=static. Choose whether to use the velocity factor in the SocialCircle. The default value is 1.

PhysicalCircle Args

  • --adaptive_fusion: type=int, argtype=static. Choose whether to use the adaptive fusion strategy to fuse SocialCircle and PhysicalCircle into the SocialCircle+. The default value is 0.
  • --seg_map_pool_size: type=int, argtype=temporary. Choose whether to max-pool the segmentation. It is used to speed up the model inference, which may cause a little bit performance drop. Set it to -1 to disable this function, and other integers will be treated as the kernel size of the pooling layer. The default value is -1.
  • --use_empty_seg_maps: type=int, argtype=temporary. Choose whether to use empty segmentation maps when computing the PhysicalCircle. The empty segmentation map means that EVERYWHERE in the scene is available for walking. This arg is only used when running ablation studies. The default value is 0.
  • --vision_radius: type=float, argtype=static. The radius of the target agent's vision field when constructing the PhysicalCircle. Radiuses are based on the length that the agent moves during the observation period. The default value is 2.0.

Visualization Args

  • --distribution_steps: type=str, argtype=temporary. Controls which time step(s) should be considered when visualizing the distribution of forecasted trajectories. It accepts one or more integer numbers (started with 0) split by '_'. For example, '4_8_11'. Set it to 'all' to show the distribution of all predictions. The default value is all.
  • --draw_distribution (short for -dd): type=int, argtype=temporary. Controls whether to draw distributions of predictions instead of points. If draw_distribution == 0, it will draw results as normal coordinates; If draw_distribution == 1, it will draw all results in the distribution way, and points from different time steps will be drawn with different colors. The default value is 0.
  • --draw_exclude_type: type=str, argtype=temporary. Draw visualized results of agents except for user-assigned types. If the assigned types are "Biker_Cart" and the draw_results or draw_videos is not "null", it will draw results of all types of agents except "Biker" and "Cart". It supports partial match, and it is case-sensitive. The default value is null.
  • --draw_extra_outputs: type=int, argtype=temporary. Choose whether to draw (put text) extra model outputs on the visualized images. The default value is 0.
  • --draw_full_neighbors: type=int, argtype=temporary. Choose whether to draw the full observed trajectories of all neighbor agents or only the last trajectory point at the current observation moment. The default value is 0.
  • --draw_index: type=str, argtype=temporary. Indexes of test agents to visualize. Numbers are split with _. For example, '123_456_789'. The default value is all.
  • --draw_lines: type=int, argtype=temporary. Choose whether to draw lines between each two 2D trajectory points. The default value is 0.
  • --draw_on_empty_canvas: type=int, argtype=temporary. Controls whether to draw visualized results on the empty canvas instead of the actual video. The default value is 0.

Toy Example Args

  • --draw_seg_map: type=int, argtype=temporary. Choose whether to draw segmentation maps on the canvas. The default value is 1.
  • --lite: type=int, argtype=temporary. Choose whether to show the lite version of tk window. The default value is 0.
  • --points: type=int, argtype=temporary. The number of points to simulate the trajectory of manual neighbor. It only accepts 2 or 3. The default value is 2.

Contact us

Conghao Wong (@cocoon2wong): conghaowong@icloud.com
Beihao Xia (@NorthOcean): xbh_hust@hust.edu.cn
Ziqian Zou (@LivepoolQ): ziqianzoulive@icloud.com

About

🔘Official codes for "SocialCircle+: Learning the Angle-based Conditioned Interaction Representation for Pedestrian Trajectory Prediction"

Resources

License

Stars

Watchers

Forks

Languages