Skip to content

Open source code for the paper Enlisting 3D Crop Models and GANs for More Data Efficient and Generalizable Fruit Detection

License

Notifications You must be signed in to change notification settings

plant-ai-biophysics-lab/CropGAN

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

14 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Enlisting 3D Crop Models and GANs for More Data Efficient and Generalizable Fruit Detection

We provide the pytorch implementation of a semantically constrained GAN to generate artificial realisitc fruit images for training to reduce the need of real image labeling in fruit detection.

Paper Link

1. Install Requirements

pip install -r requirements.txt

2. Using Trained CropGAN Model

Use generateGANImages.ipynb notebook to load the pre-trained Semantic Consistent GAN model and generate target domain images from source synthetic image.

3. Prepare Data

We provide our data in this CropGANData repo as an example, you can prepare your own dataset following this format.

Source domain (domain A) data

Crop images + Bounding box labels for each image
The dataset used in this research (you can use any domain data even its not synthetically generated, as long as you have labels):

  1. Synthetic Grape Data

Target domain (domain B) data

Images (with a few of them labeled, 1 at least), source and target image do not need to be paird. The dataset used in this research:

  1. Night Grape
  2. Day Grape

Data organization

Data used to train CropGAN

crop_gan_data
└── sytheticVis2bordenNight
    ├── labelA # (labels for domain A images)
    ├── trainA # (domain A images)
    └── trainB # (domain B images)

Data used to train object detection model

detection_datasets
└── TargetDomainData
    ├── class.txt
    ├── data_configs
    ├── test
    ├── train
    └── valid

3. Train the model

STEP 1: Finetuning using N train images and K validation images

  1. Run step1-finetuning.py script at yolov3 folder
"""
If not using synthetic pretrained
--model_def ./config/yolov3-tiny.cfg
--pretrained_weights /data2/zfei/data/cycleGAN/yolo/weights/yolov3-tiny.weights
--model_def ./config/yolov3.cfg
--pretrained_weights /data2/zfei/data/cycleGAN/yolo/weights/darknet53.conv.74
"""
# e.g ~/CropGANData/detection_datasets/BordenNight/
data_path="$your_folder/CropGANData/detection_datasets/BordenNight/"
save_dir="$your_folder/CropGANData/output/BordenNight"
pretrained_weights="$your_folder/CropGAN/data/models/yolo/synthetic_pretrained_yolov3.pth"
# You can change to other data config by change dataname="traina_valb"
dataname="train1_val1"

cd yolov3
python -u step1-finetuning.py --model_def ./config/yolov3-tiny.cfg \
                           --data_config $data_path/data_configs/$dataname/data.data \
                           --pretrained_weights $pretrained_weights \
                           --batch_size=8 \
                           --img_size=416 \
                           --save_dir $save_dir/$dataname \
                           --checkpoint_interval 10\
                           --epochs=100

STEP2 Train Semantically Constrained CycleGAN

  1. Run train_cropgan.py at src/
"""
yolo_b_weights can be selected from the weight you trained in STEP1
"""
dataroot="$your_folder/CropGANData/crop_gan_data/sytheticVis2bordenNight/"
checkpoints_dir="$your_folder/CropGANData/output/BordenNight/train1_val1/cropgan_checkpoints/"
yolo_a_weights="$your_folder/CropGAN/data/models/yolo/synthetic_pretrained_yolov3.pth"
yolo_b_weights="$your_folder/CropGANData/output/BordenNight/train1_val1/checkpoints/best_mAp_yolov3_ckpt.pth"
python -u train_cropgan.py --dataroot $dataroot \
             --num_threads 10\
             --name sythetic2bordenNight\
             --dataset_mode yolo_task_reverse\
             --checkpoints_dir  $checkpoints_dir \
             --no_flip\
             --preprocess aug\
             --model double_task_cycle_gan\
             --load_size 416\
             --crop_size 256\
             --lambda_yolo_b 0.1\
             --lambda_yolo_a 0.01\
             --batch_size 1\
             --cycle_gan_epoch 1\
             --yolo_eval_on_real_period 500\
             --yolo_epochs 0\
             --task_model_def ../yolov3/config/yolov3-tiny.cfg \
             --yolo_a_weights $yolo_a_weights \
             --yolo_b_weights $yolo_b_weights \
             --save_epoch_freq 1

You can view the training process in visdom http://localhost:8097/

3. Acknowledgement

  1. The basenet work CycleGAN code is from junyanz
    https://github.com/junyanz/pytorch-CycleGAN-and-pix2pix
    pytorch-CycleGAN-and-pix2pix

  2. The YOLOv3 implementation is from Erik Linder-Norén
    https://github.com/eriklindernoren/PyTorch-YOLOv3
    A minimal PyTorch implementation of YOLOv3, with support for training, inference and evaluation.

Funding

This project was partly funded by the National AI Institute for Food Systems (AIFS).

About

Open source code for the paper Enlisting 3D Crop Models and GANs for More Data Efficient and Generalizable Fruit Detection

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published