Skip to content

atecup/atec2025_software_algorithm

Repository files navigation

ATEC_competition_demo_code_2025

Note that the Unreal Environment Binaries and example dataset in JSONL format are not included in this repository. Please register on the ATEC official website to access and download the completed demo.

This project integrates Unreal Engine with OpenAI Gym for a rescue task based on UnrealCV.

🔥 News

  • 04/05/2025. We released two additional baseline methods: VLM-based and RL-based.
    # An example of implementing RL-based algorithm, user could try to improve it by modifying the reward function, integrating VFMs etc.  
    python train_ppo.py --num_envs 4 --num_epochs 100000 --device cuda --num_steps 60 --test_jsonl test_L1.jsonl --gamma 0.99 --ppo_epochs 20 --clip_param 0.2
    
    # An example of implementing VLM-based method for zero-shot reasoning and action execution.
    python Run_VLM_Agent.py
    

Introduction

In this competition, the agent will be given image and text cues at the initial stage to help it respond to distress signals from an injured person in need of rescue. The agent must navigate a complex 3D environment, locate the objects, and transport them to designated stretchers as quickly as possible.

1.Simulation Interaction Environment:
Different from traditional computer vision tasks, contestants need to conduct real-time dynamic interactions in the virtual environment provided by this competition. Contestants can use this virtual simulation platform for data collection and strategy training, and the final scoring of the competition will also be evaluated on the same platform.

Examples of the simulation environment used in the competition and the first-person perspective images of the intelligent agent are shown below: Description of the image

2.Interaction Interface:
Contestants can interact with the environment through a gym-like Python interface to obtain the initial clues (a color image and a text description), as well as the first-person perspective observation information (RGB image) of the intelligent agent. They can also control the real-time movement of the intelligent agent in the environment, execute actions, and receive reward signals.

Examples of the clues obtained by the intelligent agent at the initial stage.
Description of the image

Installation

Dependencies

UnrealCV, Gym, CV2, Numpy, Docker(Optional), Nvidia-Docker(Optional)

We recommend you use anaconda to install and manage your Python environment.

CV2 is used for image processing, like extracting object masks and bounding boxes. Matplotlib is used for visualization.

It is easy to install, just activate your python environment and install dependency package

pip install -r requirements.txt

Run the baseline code

./run_main.sh

Task Demonstration

We provide a task execution example to help participants better understand the task. Participants can:
- Control the agent via keyboard to complete the rescue mission --keyboard.
- Observe a randomly controlled agent navigating the environment.
By adding the parameters --render and --record_video, participants can visualize the agent’s first-person perspective and save the entire observation sequence as an MP4 file, gaining a clear understanding of the success and failure criteria.

Run a keyboard controlled agent, visualize and save to output.mp4

- Use `i`, `j`, `k`, `l` for agent movement  
- `1` for pick  
- `2` for drop  
- `e` for open the door  
- `space` for jump  
- `ctrl` for crouch  
python example/rescue_demo.py --render --record_video  --keyboard

Run a random agent, visualize it

python example/rescue_demo.py --render 

Acknowledgments

This repository is a specialized branch of the unrealzoo project, tailored to simulate rescue task scenarios. We acknowledge the following projects for their contributions:

About

Simulators and baselines for ATEC 2025 software algorithm track (online competition)

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 4

  •  
  •  
  •  
  •  

Languages