[go: up one dir, main page]

Follow
Zhaochong An
Title
Cited by
Cited by
Year
Indiscernible object counting in underwater scenes
G Sun, Z An, Y Liu, C Liu, C Sakaridis, DP Fan, L Van Gool
CVPR 2023, 2023
512023
Rethinking few-shot 3d point cloud semantic segmentation
Z An, G Sun, Y Liu, F Liu, Z Wu, D Wang, L Van Gool, S Belongie
CVPR 2024, 2024
432024
Object Segmentation by Mining Cross-Modal Semantics
Z Wu, J Wang, Z Zhou, Z An, Q Jiang, C Demonceaux, G Sun, R Timofte
ACM MM 2023, 2023
432023
Multimodality Helps Few-shot 3D Point Cloud Semantic Segmentation
Z An, G Sun, Y Liu, R Li, M Wu, MM Cheng, E Konukoglu, S Belongie
ICLR 2025 (Spotlight), 2024
232024
kNN-CLIP: Retrieval Enables Training-Free Segmentation on Continually Expanding Large Vocabularies
Z Gui, S Sun, R Li, J Yuan, Z An, K Roth, A Prabhu, P Torr
TMLR 2024, 2024
21*2024
Chatmotion: A multimodal multi-agent for human motion analysis
L Li, S Jia, J Wang, Z An, J Li, JN Hwang, S Belongie
arXiv preprint arXiv:2502.18180, 2025
152025
Generalized few-shot 3d point cloud segmentation with vision-language model
Z An, G Sun, Y Liu, R Li, J Han, E Konukoglu, S Belongie
CVPR 2025, 2025
92025
DreamBeast: Distilling 3D Fantastical Animals with Part-Aware Knowledge Transfer
R Li, J Han, L Melas-Kyriazi, C Sun, Z An, Z Gui, S Sun, P Torr, T Jakab
3DV 2025, 2024
8*2024
Temporal-aware Hierarchical Mask Classification for Video Semantic Segmentation
Z An, G Sun, Z Wu, H Tang, L Van Gool
BMVC 2023, Code: https://github.com/ZhaochongAn/THE-Mask, 2023
82023
What You Have is What You Track: Adaptive and Robust Multimodal Tracking
Y Tan, J Shao, E Zamfir, R Li, Z An, C Ma, D Paudel, L Van Gool, ...
ICCV 2025, 2025
42025
OneStory: Coherent Multi-Shot Video Generation with Adaptive Memory
Z An, M Jia, H Qiu, Z Zhou, X Huang, Z Liu, W Ren, K Kahatapitiya, D Liu, ...
arXiv preprint arXiv:2512.07802, 2025
2025
Scaling Zero-Shot Reference-to-Video Generation
Z Zhou, S Liu, H Liu, H Qiu, Z An, W Ren, Z Liu, X Huang, KW Ng, T Xie, ...
arXiv preprint arXiv:2512.06905, 2025
2025
TUNA: Taming Unified Visual Representations for Native Unified Multimodal Models
Z Liu, W Ren, H Liu, Z Zhou, S Chen, H Qiu, X Huang, Z An, F Yang, ...
arXiv preprint arXiv:2512.02014, 2025
2025
Cultural Evaluations of Vision-Language Models Have a Lot to Learn from Cultural Theory
S Yadav, L Tilton, M Antoniak, T Arnold, J Li, SM Pawar, A Karamolegkou, ...
arXiv preprint arXiv:2505.22793, 2025
2025
The system can't perform the operation now. Try again later.
Articles 1–14