Dataset Viewer
Auto-converted to Parquet Duplicate
image
imagewidth (px)
70
2.05k
label
stringlengths
8
10
10:40 AM
04:58 PM
02:18 None
08:35 AM
02:00 PM
08:46 AM
12:18 PM
11:26 AM
01:25 PM
02:42 PM
06:18 AM
12:30 PM
02:25 PM
10:03 AM
01:43 None
09:14 AM
10:35 AM
12:17 PM
01:40 PM
06:15 AM
11:25 AM
02:10 PM
02:29 None
10:30 AM
12:06 PM
02:30 PM
09:57 PM
12:04 PM
12:43 PM
02:30 PM
04:53 PM
06:37 PM
11:46 AM
05:11 PM
03:00 PM
11:34 AM
10:50 AM
05:25 PM
12:00 PM
04:42 PM
08:55 AM
10:10 AM
06:20 AM
09:31 AM
02:35 PM
07:28 AM
04:05 PM
10:50 None
03:30 PM
04:45 PM
12:50 PM
09:35 PM
12:35 PM
09:05 PM
07:57 AM
10:03 PM
10:14 AM
08:06 AM
08:20 PM
08:05 PM
12:05 PM
04:55 PM
01:45 PM
05:15 PM
02:32 PM
02:36 PM
03:50 PM
05:05 PM
10:45 PM
10:08 AM
03:15 PM
10:10 AM
11:33 AM
02:35 PM
10:59 AM
02:53 PM
03:55 PM
05:16 AM
11:03 AM
02:40 PM
04:49 PM
01:03 PM
10:32 None
11:58 AM
05:30 PM
06:20 None
06:30 AM
03:20 PM
09:15 AM
02:38 PM
12:27 PM
04:15 PM
02:17 PM
04:45 PM
07:00 AM
04:05 PM
12:55 None
12:20 PM
10:55 None
04:03 PM
End of preview. Expand in Data Studio

πŸ• TickTockVQA

Analog clock reading dataset from It's Time to Get It Right: Improving Analog Clock Reading and Clock-Hand Spatial Reasoning in Vision-Language Models (CVPR 2026 Findings). Contains 12,483 images with time labels for training and evaluating VQA models on analog clock understanding.

arXiv GitHub Project Hugging Face

Dataset Notice: This dataset is collected from publicly available data corpora. The copyright and redistribution conditions of the dataset do not belong to the authors of this project. Please refer to the respective source data corpora and the license information in the annotations.json file for details on usage, attribution, and redistribution terms.


πŸ“– Overview

TickTockVQA is a benchmark dataset for reading analog clocks from images. Each sample includes an image and the ground-truth time displayed on the most prominent clock face.

Split Samples
Train 7,236
Test 5,247

Sources: OpenImages, COCO, ClockMovies, VisualGenome, CC12M, ImageNet, SBU


πŸ“Š Qualitative Results

Qualitative Results

Sample predictions and visualizations from our models on the TickTockVQA benchmark.

Effect of Swap-DPO

Hand-swap Error Correction

Qualitative examples of hand-swap error correction by Swap-DPO. SFT incorrectly swaps the hour and minute hands, whereas Swap-DPO successfully corrects this systematic error pattern.


πŸ“ Files

File Description
annotations.json Image annotations with image_name, image_path, time_string, hour, minute, license info
qualitative_results.png Qualitative results (sample predictions and visualizations)
handswap_qualitative_results.png Hand-swap error correction by Swap-DPO
dataset_statistics.json Dataset statistics (source distribution, splits, license distribution)

πŸš€ Quick Start

from datasets import load_dataset
dataset = load_dataset("jaeha-choi/TickTockVQA")

Or download annotations only:

huggingface-cli download jaeha-choi/TickTockVQA annotations.json --local-dir ./data

πŸ“š Citation

If you find our work useful, please cite:

@article{choi2026clockreasoning,
  title   = {It's Time to Get It Right: Improving Analog Clock Reading and Clock-Hand Spatial Reasoning in Vision-Language Models},
  author  = {Choi, Jaeha and Lee, Jin Won and You, Siwoo and Lee, Jangho},
  journal = {arXiv preprint arXiv:2603.08011},
  year    = {2026},
  url     = {https://arxiv.org/abs/2603.08011}
}

If you find our work useful, we would appreciate a Hugging Face πŸ€— like and a GitHub ⭐ Star. Thank you!


πŸ”— Links

Downloads last month
80

Paper for jaeha-choi/TickTockVQA