✏️ Data for VidChain Excercise
VidChain: Chain-of-Tasks with Metric-based Direct Preference Optimization for Dense Video Captioning
Ji Soo Lee*, Jongha Kim*, Jeehye Na, Jinyoung Park, Hyunwoo J. Kim†.
AAAI 2025

🎯 Learning Objectives
By working through this exercise, you will:
- Reproduce baseline behavior of a video-language model (VTimeLLM, CVPR 2024 Highlight).
- Observe the limitations of existing approaches in temporal reasoning and coherence.
- Implement and experiment with VidChain's improvements using M-DPO.
- Run inference on videos to generate dense temporal captions (Dense Video Captioning).
- Evaluate how preference alignment improves performance over baselines.
- Discuss potential strategies for ensembling different reasoning paths of VidChain's CoTasks.
Citations 🌱
@inproceedings{lee2025vidchain,
title={VidChain: Chain-of-Tasks with Metric-based Direct Preference Optimization for Dense Video Captioning},
author={Lee, Ji Soo and Kim, Jongha and Na, Jeehye and Park, Jinyoung and Kim, Hyunwoo J},
booktitle={AAAI},
year={2025}
}