📹 Extremely low frame-rate (1 fps) video object tracking challenge
This project is developed by the Product R&D Department of the Digital Image Technology Division at ASUS Computer Inc. AI Solution Business Unit.
In recent years, surveillance camera systems have been widely used on roads due to the demands for home security and crime prevention. Since most surveillance systems are currently based on singlecamera recording, each camera operates independently, making it impossible to continue identifying moving objects once they leave the field of view. Additionally, in the event of accidents or criminal incidents, because each camera records independently and there is no mechanism for cooperative operation between cameras, law enforcement agencies must expend significant manpower resources to manually search through surveillance recordings to track the paths and trajectories of suspicious vehicles or pedestrians.
Example demo results 📷
- High movement speed
0902_130006_131041_camera_2.mp4
- Unstable connection (Heavy)
1001_130000_131559_camera_5.mp4
- Unstable connection (Slight)
0903_125957_131610_camera_1.mp4
- Flare issue
1001_180000_181558_camera_3.mp4
1015_180001_183846_camera_6.mp4
- Disconnect issue
1015_180001_183846_camera_8.mp4
Step 0: Environment Setting
-
Download the Repo
git clone https://github.com/FanChiMao/Competition-2024-PyTorch-Tracking.git git submodule update --init
-
Prepare the environment
❗ Noted: Please check your GPU and OS environment, and go to the PyTorch Website to install first.conda create --name AICUP_envs python=3.8 pip install -r requirements.txt
-
Prepare datasets
- Go to the official website to download the datasets.
- Place testing set (
32_33_AI_CUP_testdataset
folder) in ./datasets.
-
Prepare trained model weights
- Go to the download the pretrained weights in our release.
- Place all the model weights in ./weights.
- Or you can run the python script in ./weights/download_model_weights.py
Step 1: Set the yaml file correctly
-
Modify the inference setting (inference_testset.yaml) to prepare inference (following setting is our best submitted setting)
# [Default Setting] Default : RESULT_FOLDER: ./aicup_results FRAME_FOLDER: ./datasets/32_33_AI_CUP_testdataset/AI_CUP_testdata/images # write mot (AICUP submit) txt file WRITE_MOT_TXT: true # write final inference video SAVE_OUT_VIDEO: true SAVE_OUT_VIDEO_FPS: 3 # [Detector] Detector: ENSEMBLE: true # if set true, fill the detector_weight_list and corresponding score ENSEMBLE_MODEL_LIST: [ weights/yolov9-c_0902.pt, weights/yolov9-c_1016.pt, weights/yolov8-x_finetune.pt, weights/yolov8-x_worldv2_pretrained.pt ] ENSEMBLE_WEIGHT_LIST: [0.8, 0.76, 0.75, 0.7] DETECTOR_WEIGHT: weights/yolov9-c_0902.pt DETECTOR_CONFIDENCE: 0.05 # [Extractor] Extractor: EXTRACTOR_WEIGHT: weights/osnet_x1_0.pth.tar-50 EXTRACTOR_TYPE: osnet_x1_0 EXTRACTOR_THRESHOLD: 0.6 # [Tracker] Tracker: TRACKER_MOTION_PREDICT: lr # lr / kf (Linear / Kalman Filter) TRACKER_MAX_UNMATCH_FRAME : 3
Step 2: Run inference code
-
After setting the correct configuration of yaml file, simply run:
python inference_testset.py
-
⏱ We use single RTX 2070 (8 GB) to run the inference, here is the estimation inference time for single/ensemble model:
- 1 model (YOLOv9c model): about 45 minutes
- 2 model (YOLOv9c + YOLOv8x): about 1 hours
- 3 model (YOLOv9c + YOLOv8x + YOLOv8world): about 2 hours
- 4 model (YOLOv9c + YOLOv8x + YOLOv8world + YOLOv9c): about 2.5 hours
Step 0: Environment Setting
-
Download the Repo
git clone https://github.com/FanChiMao/Competition-2024-PyTorch-Tracking.git git submodule update --init
-
Prepare the environment
❗ Noted: Please check your GPU and OS environment, and go to the PyTorch Website to install first.conda create --name AICUP_envs python=3.8 pip install -r requirements.txt
-
Prepare datasets
- Go to the official website to download the datasets, and place them in the
./datasets
folder.
- Go to the official website to download the datasets, and place them in the
Step 1: Train Detector (YOLOv9-c model)
-
After downloading the dataset from official website, simply run
python .\Detector\yolov9\data_prepare.py --AICUP_dir ./datasets/32_33_train_v2/train --YOLOv9_dir ./datasets/detector_datasets --train_ratio 1
-
Set the correct data path
Correct thepath
argument in Detector\detector.yaml as the path after previous preprocessing
-
Start training by using following command
python .\Detector\yolov9\train_dual.py --weights .\Detector\yolov9-c.pt --cfg .\Detector\yolov9\models\detect\yolov9-c.yaml --data .\Detector\detector.yaml --device 0 --batch-size 4 --epochs 50 --hyp .\Detector\yolov9\data\hyps\hyp.scratch-high.yaml --name yolov9-c --close-mosaic 15 --cos-lr
💬 For more details about the Detector
of our method, you can check here.
Step 2: Train Detector with External Datasets (YOLOv8-x model)
-
BDD100K
DOWNLOAD LINK Download 3 of the following files 100k_images_train.zip, 100k_images_val.zip and bdd100k_det_20_labels_trainval.zip. Unzip 3 files and organize the directory:bdd100k - images - 100k - train - XXXX.jpg - ... - val - XXXX.jpg - ... - labels - det_20 - det_train.json - det_val.json
-
UA-DETRACE
DOWNLOAD LINK Download the dataset from Kaggle Unzip the file and keeps only DETRAC_Upload folderDETRAC_Upload - images - train - XXXX.jpg - ... - val - XXXX.jpg - ... - labels - train - XXX.txt - ... - val - XXX.txt - ...
This dataset is can only available on T-Brain Machine Learning Competition site (not v2 version)
Run the following command to merge 3 datasets into one
python prepare_dataset.py --aicup AICUP_DATASET_DIR --uadetrac UADETRAC_DATASET_DIR --bdd100k BDD100K_DATASET_DIR --output OUTPUT_DIR
- Note that in L44 of prepare_dataset.py, we only use images and labels in 10/16 for validation
Run the following command to train yolov8 detector
python train_detecctor.py --data_dir DATA_DIR
💬 For more details about the Detector
of our method, you can check here.
Step 3: Train Extractor (ReID model)
-
After downloading the dataset from official website, simply run
python .\Extractor\data_prepare.py --AICUP_dir ./datasets/32_33_train_v2/train --reid_dir ./datasets/extractor_datasets
-
Set the correct data path
Correct thepath
argument in Extractor\extractor.yaml as the path after previous preprocessing
-
Start training by using following command
python .\Extractor\train_reid_model.py
💬 For more details about the Extractor
of our method, you can check here.
- YOLOv9: https://github.com/WongKinYiu/yolov9
- YOLOv8: https://github.com/ultralytics/ultralytics
- YOLO-World: https://github.com/AILab-CVC/YOLO-World
- torchreid: https://github.com/KaiyangZhou/deep-person-reid
- Kelvin: [email protected]
- Jonathan: [email protected]
- Sam: [email protected]
- Henry: [email protected]
- Harry: [email protected]