This is code implementation of vid2bmap, presented on Dec 20th, 2023 at Korea Software Congress, as the highly-improved version of previous paper, "Vision-based beatmap extraction in rhythm game toward platform-aware note generation", published at CoG 2021. Paper Link
Pre-trained model and its meta data are given in this repo, but training/test dataset are not.
- pretrained model:
detectionAI/checkpoints/best.ckpt
- meta data:
detectionAI/data/label.json
I. You need to prepare any Nostalgia
screen-recorded video, its metadata (and config file if necessary).
- video
- screen-recording required
- video sample link
- metadata
- json-formatted
- require 3 data:
- [starting_time, fps]. Its time unit is second.
- [ending_time, fps]. Its time unit is second.
- bounding box of cropped area [x_left, x_right, y_left, y_right]. Initial point (x,y)=(0,0) is located at leftmost-top. fps means the frame rate of your video.
- ex:
{"start": [10.499733333333367, 60.0], "end": [146.41826666666674, 60.0], "roi": [9, 121, 1687, 482]}
- (optional) config file
- check the
config.py
and usage of moduleyacs
, and then make configuration file as.yaml
, if needed.
- check the
Then configurate their paths at run_demo()
in run.py
.
II. Run the script:
python run.py