paper_id
uint32
0
2.87k
title
stringlengths
15
149
authors
listlengths
1
69
cvf_url
stringlengths
94
199
pdf_url
stringlengths
95
200
supp_url
stringlengths
100
148
arxiv_id
stringlengths
10
10
bibtex
large_stringlengths
285
1.82k
abstract
large_stringlengths
547
2.44k
embedding
listlengths
768
768
2,800
SapiensID: Foundation for Human Recognition
[ "Minchul Kim", "Dingqiang Ye", "Yiyang Su", "Feng Liu", "Xiaoming Liu" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Kim_SapiensID_Foundation_for_Human_Recognition_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Kim_SapiensID_Foundation_for_Human_Recognition_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Kim_SapiensID_Foundation_for_CVPR_2025_supplemental.pdf
2504.04708
@InProceedings{Kim_2025_CVPR, author = {Kim, Minchul and Ye, Dingqiang and Su, Yiyang and Liu, Feng and Liu, Xiaoming}, title = {SapiensID: Foundation for Human Recognition}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}, month = {June}, year ...
Existing human recognition systems often rely on separate, specialized models for face and body analysis, limiting their effectiveness in real-world scenarios where pose, visibility, and context vary widely. This paper introduces SapiensID, a unified model that bridges this gap, achieving robust performance across dive...
[ -0.005507195368409157, -0.017942067235708237, 0.0026013441383838654, 0.03539692983031273, 0.03727925196290016, 0.015920283272862434, 0.05078093707561493, 0.024284357205033302, -0.01658656634390354, -0.05027376860380173, 0.012194529175758362, -0.006210691295564175, -0.10220412164926529, -0....
2,801
MIDI: Multi-Instance Diffusion for Single Image to 3D Scene Generation
[ "Zehuan Huang", "Yuan-Chen Guo", "Xingqiao An", "Yunhan Yang", "Yangguang Li", "Zi-Xin Zou", "Ding Liang", "Xihui Liu", "Yan-Pei Cao", "Lu Sheng" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Huang_MIDI_Multi-Instance_Diffusion_for_Single_Image_to_3D_Scene_Generation_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Huang_MIDI_Multi-Instance_Diffusion_for_Single_Image_to_3D_Scene_Generation_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Huang_MIDI_Multi-Instance_Diffusion_CVPR_2025_supplemental.pdf
2412.03558
@InProceedings{Huang_2025_CVPR, author = {Huang, Zehuan and Guo, Yuan-Chen and An, Xingqiao and Yang, Yunhan and Li, Yangguang and Zou, Zi-Xin and Liang, Ding and Liu, Xihui and Cao, Yan-Pei and Sheng, Lu}, title = {MIDI: Multi-Instance Diffusion for Single Image to 3D Scene Generation}, booktitle = ...
This paper introduces MIDI, a novel paradigm for compositional 3D scene generation from a single image. Unlike existing methods that rely on reconstruction or retrieval techniques or recent approaches that employ multi-stage object-by-object generation, MIDI extends pre-trained image-to-3D object generation models to m...
[ 0.002696372102946043, 0.005975804757326841, 0.019778776913881302, 0.05282891169190407, 0.03549489378929138, 0.03246266394853592, 0.012264224700629711, 0.017995070666074753, -0.031042464077472687, -0.06263817101716995, -0.03705582395195961, -0.0056355115957558155, -0.046271227300167084, 0.0...
2,802
S4-Driver: Scalable Self-Supervised Driving Multimodal Large Language Model with Spatio-Temporal Visual Representation
[ "Yichen Xie", "Runsheng Xu", "Tong He", "Jyh-Jing Hwang", "Katie Luo", "Jingwei Ji", "Hubert Lin", "Letian Chen", "Yiren Lu", "Zhaoqi Leng", "Dragomir Anguelov", "Mingxing Tan" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Xie_S4-Driver_Scalable_Self-Supervised_Driving_Multimodal_Large_Language_Model_with_Spatio-Temporal_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Xie_S4-Driver_Scalable_Self-Supervised_Driving_Multimodal_Large_Language_Model_with_Spatio-Temporal_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Xie_S4-Driver_Scalable_Self-Supervised_CVPR_2025_supplemental.pdf
null
@InProceedings{Xie_2025_CVPR, author = {Xie, Yichen and Xu, Runsheng and He, Tong and Hwang, Jyh-Jing and Luo, Katie and Ji, Jingwei and Lin, Hubert and Chen, Letian and Lu, Yiren and Leng, Zhaoqi and Anguelov, Dragomir and Tan, Mingxing}, title = {S4-Driver: Scalable Self-Supervised Driving Multimodal L...
The latest advancements in multi-modal large language models (MLLMs) have spurred a strong renewed interest in end-to-end motion planning approaches for autonomous driving. Many end-to-end approaches rely on human annotations to learn intermediate perception and prediction tasks, while purely self-supervised approaches...
[ 0.008104359731078148, -0.03546009212732315, 0.03391809016466141, 0.030481399968266487, 0.028538919985294342, 0.03802270069718361, 0.008866804651916027, 0.024421967566013336, -0.029046697542071342, -0.021733645349740982, -0.0008222378673963249, -0.0031492947600781918, -0.03732697293162346, ...
2,803
Extrapolating and Decoupling Image-to-Video Generation Models: Motion Modeling is Easier Than You Think
[ "Jie Tian", "Xiaoye Qu", "Zhenyi Lu", "Wei Wei", "Sichen Liu", "Yu Cheng" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Tian_Extrapolating_and_Decoupling_Image-to-Video_Generation_Models_Motion_Modeling_is_Easier_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Tian_Extrapolating_and_Decoupling_Image-to-Video_Generation_Models_Motion_Modeling_is_Easier_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Tian_Extrapolating_and_Decoupling_CVPR_2025_supplemental.pdf
2503.00948
@InProceedings{Tian_2025_CVPR, author = {Tian, Jie and Qu, Xiaoye and Lu, Zhenyi and Wei, Wei and Liu, Sichen and Cheng, Yu}, title = {Extrapolating and Decoupling Image-to-Video Generation Models: Motion Modeling is Easier Than You Think}, booktitle = {Proceedings of the Computer Vision and Pattern ...
Image-to-Video (I2V) generation aims to synthesize a video clip according to a given image and condition (e.g., text). The key challenge of this task lies in simultaneously generating natural motions while preserving the original appearance of the images. However, current I2V diffusion models (I2V-DMs) often produce vi...
[ 0.011334742419421673, 0.009406469762325287, 0.0066446359269320965, 0.06505485624074936, 0.054470568895339966, 0.011624102480709553, 0.0296525526791811, -0.01606125570833683, -0.025517761707305908, -0.05604008585214615, 0.018033809959888458, -0.03850791975855827, -0.04771270230412483, 0.022...
2,804
FreeCloth: Free-form Generation Enhances Challenging Clothed Human Modeling
[ "Hang Ye", "Xiaoxuan Ma", "Hai Ci", "Wentao Zhu", "Yizhou Wang" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Ye_FreeCloth_Free-form_Generation_Enhances_Challenging_Clothed_Human_Modeling_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Ye_FreeCloth_Free-form_Generation_Enhances_Challenging_Clothed_Human_Modeling_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Ye_FreeCloth_Free-form_Generation_CVPR_2025_supplemental.zip
2411.19942
@InProceedings{Ye_2025_CVPR, author = {Ye, Hang and Ma, Xiaoxuan and Ci, Hai and Zhu, Wentao and Wang, Yizhou}, title = {FreeCloth: Free-form Generation Enhances Challenging Clothed Human Modeling}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}, month ...
Achieving realistic animated human avatars requires accurate modeling of pose-dependent clothing deformations. Existing learning-based methods heavily rely on the Linear Blend Skinning (LBS) of minimally-clothed human models like SMPL to model deformation. However, they struggle to handle loose clothing, such as long d...
[ 0.01978619396686554, -0.05005129799246788, -0.022616956382989883, 0.011472110636532307, 0.04990708455443382, 0.029567856341600418, 0.01486177183687687, -0.0011560056591406465, -0.030665870755910873, -0.0496094711124897, -0.06101876124739647, -0.02757411077618599, -0.055600591003894806, 0.0...
2,805
ABC-Former: Auxiliary Bimodal Cross-domain Transformer with Interactive Channel Attention for White Balance
[ "Yu-Cheng Chiu", "Guan-Rong Chen", "Zihao Chen", "Yan-Tsung Peng" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Chiu_ABC-Former_Auxiliary_Bimodal_Cross-domain_Transformer_with_Interactive_Channel_Attention_for_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Chiu_ABC-Former_Auxiliary_Bimodal_Cross-domain_Transformer_with_Interactive_Channel_Attention_for_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Chiu_ABC-Former_Auxiliary_Bimodal_CVPR_2025_supplemental.pdf
null
@InProceedings{Chiu_2025_CVPR, author = {Chiu, Yu-Cheng and Chen, Guan-Rong and Chen, Zihao and Peng, Yan-Tsung}, title = {ABC-Former: Auxiliary Bimodal Cross-domain Transformer with Interactive Channel Attention for White Balance}, booktitle = {Proceedings of the Computer Vision and Pattern Recognit...
The primary goal of white balance (WB) for sRGB images is to correct inaccurate color temperatures, ensuring that images display natural, neutral colors. While existing WB methods yield reasonable results, their effectiveness is limited. They either focus solely on global color adjustments applied before the camera-spe...
[ 0.014240017160773277, -0.009746234863996506, 0.013423536904156208, 0.0052132392302155495, 0.016034161671996117, 0.03744306042790413, 0.000489598314743489, -0.001359766349196434, -0.02631286345422268, -0.04843702167272568, -0.026206716895103455, -0.011119834147393703, -0.07824518531560898, ...
2,806
Science-T2I: Addressing Scientific Illusions in Image Synthesis
[ "Jialuo Li", "Wenhao Chai", "Xingyu Fu", "Haiyang Xu", "Saining Xie" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Li_Science-T2I_Addressing_Scientific_Illusions_in_Image_Synthesis_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Li_Science-T2I_Addressing_Scientific_Illusions_in_Image_Synthesis_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Li_Science-T2I_Addressing_Scientific_CVPR_2025_supplemental.pdf
null
@InProceedings{Li_2025_CVPR, author = {Li, Jialuo and Chai, Wenhao and Fu, Xingyu and Xu, Haiyang and Xie, Saining}, title = {Science-T2I: Addressing Scientific Illusions in Image Synthesis}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}, month = {...
We present a novel approach to integrating scientific knowledge into generative models, enhancing their realism and consistency in image synthesis. First, we introduce Science-T2I, an expert-annotated adversarial dataset comprising adversarial 20k image pairs with 9k prompts, covering wide distinct scientific knowledge...
[ 0.02401631511747837, -0.03418651223182678, -0.0032243025489151478, 0.0737217366695404, 0.035312049090862274, 0.009864441119134426, 0.010609357617795467, -0.003132555168122053, -0.010446203872561455, -0.03843405842781067, -0.040561381727457047, 0.009908259846270084, -0.05581854283809662, 0....
2,807
Fingerprinting Denoising Diffusion Probabilistic Models
[ "Huan Teng", "Yuhui Quan", "Chengyu Wang", "Jun Huang", "Hui Ji" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Teng_Fingerprinting_Denoising_Diffusion_Probabilistic_Models_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Teng_Fingerprinting_Denoising_Diffusion_Probabilistic_Models_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Teng_Fingerprinting_Denoising_Diffusion_CVPR_2025_supplemental.pdf
null
@InProceedings{Teng_2025_CVPR, author = {Teng, Huan and Quan, Yuhui and Wang, Chengyu and Huang, Jun and Ji, Hui}, title = {Fingerprinting Denoising Diffusion Probabilistic Models}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}, month = {June}, ...
Diffusion models, especially denoising diffusion probabilistic models (DDPMs), are prevalent tools in generative AI, making their intellectual property (IP) protection increasingly important. Most existing IP protection methods for DDPMs are invasive, e.g., model watermarking, which alter model parameters and raise con...
[ -0.01227437611669302, 0.005213386844843626, -0.02063784934580326, 0.07623717933893204, 0.05680515244603157, 0.03084411658346653, 0.016508115455508232, -0.025145871564745903, -0.00873094517737627, -0.058476898819208145, 0.015718748793005943, -0.03339679539203644, -0.027431653812527657, -0.0...
2,808
MoST: Efficient Monarch Sparse Tuning for 3D Representation Learning
[ "Xu Han", "Yuan Tang", "Jinfeng Xu", "Xianzhi Li" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Han_MoST_Efficient_Monarch_Sparse_Tuning_for_3D_Representation_Learning_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Han_MoST_Efficient_Monarch_Sparse_Tuning_for_3D_Representation_Learning_CVPR_2025_paper.pdf
null
2503.18368
@InProceedings{Han_2025_CVPR, author = {Han, Xu and Tang, Yuan and Xu, Jinfeng and Li, Xianzhi}, title = {MoST: Efficient Monarch Sparse Tuning for 3D Representation Learning}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}, month = {June}, year...
We introduce Monarch Sparse Tuning (MoST), the first reparameterization-based parameter-efficient fine-tuning (PEFT) method tailored for 3D representation learning. Unlike existing adapter-based and prompt-tuning 3D PEFT methods, MoST introduces no additional inference overhead and is compatible with many 3D representa...
[ 0.013280694372951984, -0.027126502245664597, 0.0011936177033931017, 0.03193724900484085, 0.04096528887748718, 0.04682153835892677, 0.013737543486058712, -0.012978199869394302, -0.04995238035917282, -0.04363162815570831, 0.004143496509641409, -0.04325706884264946, -0.07684863358736038, -0.0...
2,809
Re-thinking Temporal Search for Long-Form Video Understanding
[ "Jinhui Ye", "Zihan Wang", "Haosen Sun", "Keshigeyan Chandrasegaran", "Zane Durante", "Cristobal Eyzaguirre", "Yonatan Bisk", "Juan Carlos Niebles", "Ehsan Adeli", "Li Fei-Fei", "Jiajun Wu", "Manling Li" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Ye_Re-thinking_Temporal_Search_for_Long-Form_Video_Understanding_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Ye_Re-thinking_Temporal_Search_for_Long-Form_Video_Understanding_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Ye_Re-thinking_Temporal_Search_CVPR_2025_supplemental.pdf
2504.02259
@InProceedings{Ye_2025_CVPR, author = {Ye, Jinhui and Wang, Zihan and Sun, Haosen and Chandrasegaran, Keshigeyan and Durante, Zane and Eyzaguirre, Cristobal and Bisk, Yonatan and Niebles, Juan Carlos and Adeli, Ehsan and Fei-Fei, Li and Wu, Jiajun and Li, Manling}, title = {Re-thinking Temporal Search fo...
Efficient understanding of long-form videos remains a significant challenge in computer vision. In this work, we revisit temporal search paradigms for long-form video understanding, studying a fundamental issue pertaining to all state-of-the-art (SOTA) long-context vision-language models (VLMs). In particular, our cont...
[ 0.01839444972574711, -0.024128928780555725, 0.005166979506611824, 0.04843585565686226, 0.019037505611777306, 0.03704748675227165, 0.009833276271820068, 0.026257263496518135, -0.025509420782327652, -0.027968764305114746, -0.028179405257105827, -0.0049339658580720425, -0.04418785870075226, 0...
2,810
InstanceGaussian: Appearance-Semantic Joint Gaussian Representation for 3D Instance-Level Perception
[ "Haijie Li", "Yanmin Wu", "Jiarui Meng", "Qiankun Gao", "Zhiyao Zhang", "Ronggang Wang", "Jian Zhang" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Li_InstanceGaussian_Appearance-Semantic_Joint_Gaussian_Representation_for_3D_Instance-Level_Perception_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Li_InstanceGaussian_Appearance-Semantic_Joint_Gaussian_Representation_for_3D_Instance-Level_Perception_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Li_InstanceGaussian_Appearance-Semantic_Joint_CVPR_2025_supplemental.pdf
2411.19235
@InProceedings{Li_2025_CVPR, author = {Li, Haijie and Wu, Yanmin and Meng, Jiarui and Gao, Qiankun and Zhang, Zhiyao and Wang, Ronggang and Zhang, Jian}, title = {InstanceGaussian: Appearance-Semantic Joint Gaussian Representation for 3D Instance-Level Perception}, booktitle = {Proceedings of the Com...
3D scene understanding is vital for applications in autonomous driving, robotics, and augmented reality. However, scene understanding based on 3D Gaussian Splatting faces three key challenges: (i) an imbalance between appearance and semantics, (ii) inconsistencies in object boundaries, and (iii) difficulties with top-d...
[ 0.00044243360753171146, 0.010472116991877556, 0.02209458500146866, 0.04085472598671913, -0.003000645898282528, 0.04840581864118576, 0.03120771422982216, 0.007955547422170639, -0.03500727564096451, -0.031318601220846176, -0.04866257682442665, 0.00013883874635212123, -0.07489282637834549, 0....
2,811
When Domain Generalization meets Generalized Category Discovery: An Adaptive Task-Arithmetic Driven Approach
[ "Vaibhav Rathore", "Shubhranil B", "Saikat Dutta", "Sarthak Mehrotra", "Zsolt Kira", "Biplab Banerjee" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Rathore_When_Domain_Generalization_meets_Generalized_Category_Discovery_An_Adaptive_Task-Arithmetic_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Rathore_When_Domain_Generalization_meets_Generalized_Category_Discovery_An_Adaptive_Task-Arithmetic_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Rathore_When_Domain_Generalization_CVPR_2025_supplemental.pdf
2503.14897
@InProceedings{Rathore_2025_CVPR, author = {Rathore, Vaibhav and B, Shubhranil and Dutta, Saikat and Mehrotra, Sarthak and Kira, Zsolt and Banerjee, Biplab}, title = {When Domain Generalization meets Generalized Category Discovery: An Adaptive Task-Arithmetic Driven Approach}, booktitle = {Proceeding...
Generalized Class Discovery (GCD) clusters base and novel classes in a target domain, using supervision from a source domain with only base classes. Current methods often falter with distribution shifts and typically require access to target data during training, which can sometimes be impractical. To address this issu...
[ -0.010133170522749424, -0.01025183591991663, 0.016385991126298904, 0.034066881984472275, 0.038030270487070084, 0.009081567637622356, 0.03223405405879021, -0.007034009788185358, -0.012521532364189625, -0.012275394052267075, -0.033196091651916504, 0.020674634724855423, -0.07153016328811646, ...
2,812
CSC-PA: Cross-image Semantic Correlation via Prototype Attentions for Single-network Semi-supervised Breast Tumor Segmentation
[ "Zhenhui Ding", "Guilian Chen", "Qin Zhang", "Huisi Wu", "Jing Qin" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Ding_CSC-PA_Cross-image_Semantic_Correlation_via_Prototype_Attentions_for_Single-network_Semi-supervised_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Ding_CSC-PA_Cross-image_Semantic_Correlation_via_Prototype_Attentions_for_Single-network_Semi-supervised_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Ding_CSC-PA_Cross-image_Semantic_CVPR_2025_supplemental.pdf
null
@InProceedings{Ding_2025_CVPR, author = {Ding, Zhenhui and Chen, Guilian and Zhang, Qin and Wu, Huisi and Qin, Jing}, title = {CSC-PA: Cross-image Semantic Correlation via Prototype Attentions for Single-network Semi-supervised Breast Tumor Segmentation}, booktitle = {Proceedings of the Computer Visi...
Accurate automatic breast ultrasound (BUS) image segmentation is essential for early breast cancer screening and diagnosis. However, it remains challenging owing to (1) breast lesions of various scale and shape, (2) ambiguous boundaries caused by speckle noise and artifacts, and (3) the scarcity of high-quality annotat...
[ 0.002151255728676915, -0.05091187730431557, -0.00712080392986536, -0.006074170116335154, 0.035217829048633575, 0.05178915336728096, 0.015673350542783737, -0.002105369931086898, -0.03898107260465622, -0.05254627391695976, -0.006974233780056238, -0.03876471519470215, -0.029437711462378502, 0...
2,813
BIP3D: Bridging 2D Images and 3D Perception for Embodied Intelligence
[ "Xuewu Lin", "Tianwei Lin", "Lichao Huang", "Hongyu Xie", "Zhizhong Su" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Lin_BIP3D_Bridging_2D_Images_and_3D_Perception_for_Embodied_Intelligence_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Lin_BIP3D_Bridging_2D_Images_and_3D_Perception_for_Embodied_Intelligence_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Lin_BIP3D_Bridging_2D_CVPR_2025_supplemental.pdf
2411.14869
@InProceedings{Lin_2025_CVPR, author = {Lin, Xuewu and Lin, Tianwei and Huang, Lichao and Xie, Hongyu and Su, Zhizhong}, title = {BIP3D: Bridging 2D Images and 3D Perception for Embodied Intelligence}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}, mon...
In embodied intelligence systems, a key component is 3D perception algorithm, which enables agents to understand their surrounding environments. Previous algorithms primarily rely on point cloud, which, despite offering precise geometric information, still constrain perception performance due to inherent sparsity, nois...
[ 0.02044943906366825, -0.02043375000357628, -0.00674846675246954, 0.013684920966625214, 0.017368579283356667, 0.04029405117034912, 0.0005889325984753668, 0.02234545536339283, -0.037063464522361755, -0.058463599532842636, -0.00952468067407608, -0.02363651618361473, -0.049755338579416275, 0.0...
2,814
Query Efficient Black-Box Visual Prompting with Subspace Learning
[ "Zhaogeng Liu", "Haozhen Zhang", "Hualin Zhang", "Xingchen Li", "Wanli Shi", "Bin Gu", "Yi Chang" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Liu_Query_Efficient_Black-Box_Visual_Prompting_with_Subspace_Learning_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Liu_Query_Efficient_Black-Box_Visual_Prompting_with_Subspace_Learning_CVPR_2025_paper.pdf
null
null
@InProceedings{Liu_2025_CVPR, author = {Liu, Zhaogeng and Zhang, Haozhen and Zhang, Hualin and Li, Xingchen and Shi, Wanli and Gu, Bin and Chang, Yi}, title = {Query Efficient Black-Box Visual Prompting with Subspace Learning}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition C...
Visual Prompt Learning (VPL) has emerged as a powerful strategy for harnessing the capabilities of large-scale pre-trained models (PTMs) to tackle specific downstream tasks. However, the opaque nature of PTMs in many real-world applications has led to a growing interest in gradient-free approaches within VPL. A signifi...
[ -0.017628898844122887, -0.03402460739016533, 0.01742001622915268, 0.031414277851581573, 0.003990026190876961, 0.01876058056950569, 0.02558494359254837, -0.0003219684585928917, -0.038699496537446976, -0.010010209865868092, -0.0552636943757534, -0.0022971711587160826, -0.07203590124845505, -...
2,815
VisionPAD: A Vision-Centric Pre-training Paradigm for Autonomous Driving
[ "Haiming Zhang", "Wending Zhou", "Yiyao Zhu", "Xu Yan", "Jiantao Gao", "Dongfeng Bai", "Yingjie Cai", "Bingbing Liu", "Shuguang Cui", "Zhen Li" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Zhang_VisionPAD_A_Vision-Centric_Pre-training_Paradigm_for_Autonomous_Driving_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Zhang_VisionPAD_A_Vision-Centric_Pre-training_Paradigm_for_Autonomous_Driving_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Zhang_VisionPAD_A_Vision-Centric_CVPR_2025_supplemental.pdf
2411.14716
@InProceedings{Zhang_2025_CVPR, author = {Zhang, Haiming and Zhou, Wending and Zhu, Yiyao and Yan, Xu and Gao, Jiantao and Bai, Dongfeng and Cai, Yingjie and Liu, Bingbing and Cui, Shuguang and Li, Zhen}, title = {VisionPAD: A Vision-Centric Pre-training Paradigm for Autonomous Driving}, booktitle = ...
This paper introduces VisionPAD, a novel self-supervised pre-training paradigm designed for vision-centric algorithms in autonomous driving. In contrast to previous approaches that employ neural rendering with explicit depth supervision, VisionPAD utilizes more efficient 3D Gaussian Splatting to reconstruct multi-view ...
[ 0.0021887917537242174, -0.024097485467791557, 0.008864613249897957, 0.04426548257470131, 0.013431800529360771, 0.04737604036927223, 0.024278556928038597, 0.007366017438471317, -0.013800220564007759, -0.06885917484760284, -0.02850051037967205, -0.0052197715267539024, -0.0332525372505188, -0...
2,816
Detecting Adversarial Data Using Perturbation Forgery
[ "Qian Wang", "Chen Li", "Yuchen Luo", "Hefei Ling", "Shijuan Huang", "Ruoxi Jia", "Ning Yu" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Wang_Detecting_Adversarial_Data_Using_Perturbation_Forgery_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Wang_Detecting_Adversarial_Data_Using_Perturbation_Forgery_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Wang_Detecting_Adversarial_Data_CVPR_2025_supplemental.pdf
2405.16226
@InProceedings{Wang_2025_CVPR, author = {Wang, Qian and Li, Chen and Luo, Yuchen and Ling, Hefei and Huang, Shijuan and Jia, Ruoxi and Yu, Ning}, title = {Detecting Adversarial Data Using Perturbation Forgery}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}...
As a defense strategy against adversarial attacks, adversarial detection aims to identify and filter out adversarial data from the data flow based on discrepancies in distribution and noise patterns between natural and adversarial data. Although previous detection methods achieve high performance in detecting gradient-...
[ 0.025693247094750404, -0.013363712467253208, 0.004469850566238165, 0.043239008635282516, 0.03844169154763222, 0.012473085895180702, 0.03549014776945114, -0.014903907664120197, -0.023038769140839577, -0.055273134261369705, -0.00997608806937933, 0.003417984815314412, -0.06900080293416977, -0...
2,817
CoA: Towards Real Image Dehazing via Compression-and-Adaptation
[ "Long Ma", "Yuxin Feng", "Yan Zhang", "Jinyuan Liu", "Weimin Wang", "Guang-Yong Chen", "Chengpei Xu", "Zhuo Su" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Ma_CoA_Towards_Real_Image_Dehazing_via_Compression-and-Adaptation_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Ma_CoA_Towards_Real_Image_Dehazing_via_Compression-and-Adaptation_CVPR_2025_paper.pdf
null
2504.05590
@InProceedings{Ma_2025_CVPR, author = {Ma, Long and Feng, Yuxin and Zhang, Yan and Liu, Jinyuan and Wang, Weimin and Chen, Guang-Yong and Xu, Chengpei and Su, Zhuo}, title = {CoA: Towards Real Image Dehazing via Compression-and-Adaptation}, booktitle = {Proceedings of the Computer Vision and Pattern ...
Learning-based image dehazing algorithms have shown remarkable success in synthetic domains. However, real image dehazing is still in suspense due to computational resource constraints and the diversity of real-world scenes. Therefore, there is an urgent need for an algorithm that excels in both efficiency and adaptabi...
[ 0.017899461090564728, -0.01862247660756111, 0.02397715300321579, 0.019717879593372345, 0.04405545815825462, 0.024240098893642426, -0.011001607403159142, 0.0064567201770842075, -0.012476403266191483, -0.048992693424224854, -0.026867778971791267, -0.005696069914847612, -0.060164209455251694, ...
2,818
NightAdapter: Learning a Frequency Adapter for Generalizable Night-time Scene Segmentation
[ "Qi Bi", "Jingjun Yi", "Huimin Huang", "Hao Zheng", "Haolan Zhan", "Yawen Huang", "Yuexiang Li", "Xian Wu", "Yefeng Zheng" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Bi_NightAdapter_Learning_a_Frequency_Adapter_for_Generalizable_Night-time_Scene_Segmentation_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Bi_NightAdapter_Learning_a_Frequency_Adapter_for_Generalizable_Night-time_Scene_Segmentation_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Bi_NightAdapter_Learning_a_CVPR_2025_supplemental.pdf
null
@InProceedings{Bi_2025_CVPR, author = {Bi, Qi and Yi, Jingjun and Huang, Huimin and Zheng, Hao and Zhan, Haolan and Huang, Yawen and Li, Yuexiang and Wu, Xian and Zheng, Yefeng}, title = {NightAdapter: Learning a Frequency Adapter for Generalizable Night-time Scene Segmentation}, booktitle = {Proceed...
Night-time scene segmentation is a critical yet challenging task in the real-world applications, primarily due to the complicated lighting conditions. However, existing methods lack sufficient generalization ability to unseen nigh-time scenes with varying illumination.In light of this issue, we focus on investigating g...
[ 0.0007492341683246195, -0.03112039901316166, 0.03873369097709656, 0.010847199708223343, 0.039414964616298676, -0.010015935637056828, 0.010310404002666473, -0.011140433140099049, -0.032352693378925323, -0.0407707579433918, -0.05068312585353851, 0.02741776406764984, -0.05704938620328903, 0.0...
2,819
UMFN: Unified Multi-Domain Face Normalization for Joint Cross-domain Prototype Learning and Heterogeneous Face Recognition
[ "Meng Pang", "Wenjun Zhang", "Nanrun Zhou", "Shengbo Chen", "Hong Rao" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Pang_UMFN_Unified_Multi-Domain_Face_Normalization_for_Joint_Cross-domain_Prototype_Learning_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Pang_UMFN_Unified_Multi-Domain_Face_Normalization_for_Joint_Cross-domain_Prototype_Learning_CVPR_2025_paper.pdf
null
null
@InProceedings{Pang_2025_CVPR, author = {Pang, Meng and Zhang, Wenjun and Zhou, Nanrun and Chen, Shengbo and Rao, Hong}, title = {UMFN: Unified Multi-Domain Face Normalization for Joint Cross-domain Prototype Learning and Heterogeneous Face Recognition}, booktitle = {Proceedings of the Computer Visio...
Face normalization aims to enhance the robustness and effectiveness of face recognition systems by mitigating intra-personal variations in expressions, poses, occlusions, illuminations, and domains. Existing methods face limitations in handling multiple variations and adapting to cross-domain scenarios. To address thes...
[ -0.03025917522609234, -0.02214129827916622, 0.024816570803523064, 0.023076767101883888, 0.04224889725446701, 0.012856923043727875, 0.03075266256928444, -0.0169320497661829, -0.006812617182731628, -0.04919581487774849, 0.002561810426414013, -0.003409244120121002, -0.0732649639248848, 0.0088...
2,820
TopV: Compatible Token Pruning with Inference Time Optimization for Fast and Low-Memory Multimodal Vision Language Model
[ "Cheng Yang", "Yang Sui", "Jinqi Xiao", "Lingyi Huang", "Yu Gong", "Chendi Li", "Jinghua Yan", "Yu Bai", "Ponnuswamy Sadayappan", "Xia Hu", "Bo Yuan" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Yang_TopV_Compatible_Token_Pruning_with_Inference_Time_Optimization_for_Fast_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Yang_TopV_Compatible_Token_Pruning_with_Inference_Time_Optimization_for_Fast_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Yang_TopV_Compatible_Token_CVPR_2025_supplemental.pdf
2503.18278
@InProceedings{Yang_2025_CVPR, author = {Yang, Cheng and Sui, Yang and Xiao, Jinqi and Huang, Lingyi and Gong, Yu and Li, Chendi and Yan, Jinghua and Bai, Yu and Sadayappan, Ponnuswamy and Hu, Xia and Yuan, Bo}, title = {TopV: Compatible Token Pruning with Inference Time Optimization for Fast and Low-Mem...
Vision-Language Models (VLMs) demand substantial computational resources during inference, largely due to the extensive visual input tokens for representing visual information. Previous studies have noted that visual tokens tend to receive less attention than text tokens, suggesting their lower importance during infere...
[ -0.00648347707465291, -0.004644227214157581, -0.002134795766323805, 0.03370136395096779, 0.005211507435888052, 0.02363959699869156, 0.029906094074249268, 0.030408645048737526, -0.041172094643116, -0.018807578831911087, -0.053474150598049164, 0.016249824315309525, -0.054068420082330704, 0.0...
2,821
Improving Autoregressive Visual Generation with Cluster-Oriented Token Prediction
[ "Teng Hu", "Jiangning Zhang", "Ran Yi", "Jieyu Weng", "Yabiao Wang", "Xianfang Zeng", "Zhucun Xue", "Lizhuang Ma" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Hu_Improving_Autoregressive_Visual_Generation_with_Cluster-Oriented_Token_Prediction_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Hu_Improving_Autoregressive_Visual_Generation_with_Cluster-Oriented_Token_Prediction_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Hu_Improving_Autoregressive_Visual_CVPR_2025_supplemental.pdf
2501.00880
@InProceedings{Hu_2025_CVPR, author = {Hu, Teng and Zhang, Jiangning and Yi, Ran and Weng, Jieyu and Wang, Yabiao and Zeng, Xianfang and Xue, Zhucun and Ma, Lizhuang}, title = {Improving Autoregressive Visual Generation with Cluster-Oriented Token Prediction}, booktitle = {Proceedings of the Computer...
Employing LLMs for visual generation has recently become a research focus. However, the existing methods primarily transfer the LLM architecture to visual generation but rarely investigate the fundamental differences between language and vision. This oversight may lead to suboptimal utilization of visual generation cap...
[ 0.029655108228325844, -0.028337471187114716, 0.021005043759942055, 0.010685059241950512, 0.046279024332761765, 0.04912764951586723, 0.00168604648206383, 0.010985196568071842, -0.0385037325322628, -0.028005454689264297, -0.02960428036749363, -0.01588182896375656, -0.058034781366586685, 0.00...
2,822
Learned Binocular-Encoding Optics for RGBD Imaging Using Joint Stereo and Focus Cues
[ "Yuhui Liu", "Liangxun Ou", "Qiang Fu", "Hadi Amata", "Wolfgang Heidrich", "Yifan Peng" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Liu_Learned_Binocular-Encoding_Optics_for_RGBD_Imaging_Using_Joint_Stereo_and_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Liu_Learned_Binocular-Encoding_Optics_for_RGBD_Imaging_Using_Joint_Stereo_and_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Liu_Learned_Binocular-Encoding_Optics_CVPR_2025_supplemental.pdf
null
@InProceedings{Liu_2025_CVPR, author = {Liu, Yuhui and Ou, Liangxun and Fu, Qiang and Amata, Hadi and Heidrich, Wolfgang and Peng, Yifan}, title = {Learned Binocular-Encoding Optics for RGBD Imaging Using Joint Stereo and Focus Cues}, booktitle = {Proceedings of the Computer Vision and Pattern Recogn...
Extracting high-fidelity RGBD information from two-dimensional (2D) images is essential for various visual computing applications. Stereo imaging, as a reliable passive imaging technique for obtaining three-dimensional (3D) scene information, has benefited greatly from deep learning advancements. However, existing ster...
[ -0.004758225753903389, 0.024636197835206985, -0.01853599213063717, 0.04138316586613655, 0.00903139729052782, 0.055654268711805344, 0.009969902224838734, 0.008906368166208267, -0.006326100789010525, -0.03849605843424797, -0.01348876953125, 0.006029585842043161, -0.02647622860968113, 0.01670...
2,823
Dual-view X-ray Detection: Can AI Detect Prohibited Items from Dual-view X-ray Images like Humans?
[ "Renshuai Tao", "Haoyu Wang", "Yuzhe Guo", "Hairong Chen", "Li Zhang", "Xianglong Liu", "Yunchao Wei", "Yao Zhao" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Tao_Dual-view_X-ray_Detection_Can_AI_Detect_Prohibited_Items_from_Dual-view_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Tao_Dual-view_X-ray_Detection_Can_AI_Detect_Prohibited_Items_from_Dual-view_CVPR_2025_paper.pdf
null
null
@InProceedings{Tao_2025_CVPR, author = {Tao, Renshuai and Wang, Haoyu and Guo, Yuzhe and Chen, Hairong and Zhang, Li and Liu, Xianglong and Wei, Yunchao and Zhao, Yao}, title = {Dual-view X-ray Detection: Can AI Detect Prohibited Items from Dual-view X-ray Images like Humans?}, booktitle = {Proceedin...
To detect prohibited items in challenging categories, human inspectors typically rely on images from two distinct views (vertical and side). Can AI detect prohibited items from dual-view X-ray images in the same way humans do? Existing X-ray datasets often suffer from limitations, such as single-view imaging or insuffi...
[ 0.011616267263889313, -0.00788940954953432, -0.031872235238552094, 0.00851253978908062, 0.05453498288989067, -0.0036183062475174665, 0.008433738723397255, -0.0051325359381735325, -0.009832500480115414, -0.06953763216733932, -0.04624132439494133, 0.024164637550711632, -0.02343815565109253, ...
2,824
LUCAS: Layered Universal Codec Avatars
[ "Di Liu", "Teng Deng", "Giljoo Nam", "Yu Rong", "Stanislav Pidhorskyi", "Junxuan Li", "Jason Saragih", "Dimitris N. Metaxas", "Chen Cao" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Liu_LUCAS_Layered_Universal_Codec_Avatars_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Liu_LUCAS_Layered_Universal_Codec_Avatars_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Liu_LUCAS_Layered_Universal_CVPR_2025_supplemental.zip
2502.19739
@InProceedings{Liu_2025_CVPR, author = {Liu, Di and Deng, Teng and Nam, Giljoo and Rong, Yu and Pidhorskyi, Stanislav and Li, Junxuan and Saragih, Jason and Metaxas, Dimitris N. and Cao, Chen}, title = {LUCAS: Layered Universal Codec Avatars}, booktitle = {Proceedings of the Computer Vision and Patte...
Photorealistic 3D head avatar reconstruction faces critical challenges in modeling dynamic face-hair interactions and achieving cross-identity generalization, particularly during expressions and head movements. We present LUCAS, a novel Universal Prior Model (UPM) for codec avatar modeling that disentangles face and ha...
[ 0.003153646131977439, 0.007218858227133751, 0.004543395712971687, 0.02885390818119049, 0.039354439824819565, 0.03268307074904442, 0.027671225368976593, 0.00474339397624135, -0.017638670280575752, -0.0485629066824913, -0.005560560151934624, -0.011606262065470219, -0.057950709015131, -0.0041...
2,825
MobilePortrait: Real-Time One-Shot Neural Head Avatars on Mobile Devices
[ "Jianwen Jiang", "Gaojie Lin", "Zhengkun Rong", "Chao Liang", "Yongming Zhu", "Jiaqi Yang", "Tianyun Zhong" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Jiang_MobilePortrait_Real-Time_One-Shot_Neural_Head_Avatars_on_Mobile_Devices_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Jiang_MobilePortrait_Real-Time_One-Shot_Neural_Head_Avatars_on_Mobile_Devices_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Jiang_MobilePortrait_Real-Time_One-Shot_CVPR_2025_supplemental.zip
2407.05712
@InProceedings{Jiang_2025_CVPR, author = {Jiang, Jianwen and Lin, Gaojie and Rong, Zhengkun and Liang, Chao and Zhu, Yongming and Yang, Jiaqi and Zhong, Tianyun}, title = {MobilePortrait: Real-Time One-Shot Neural Head Avatars on Mobile Devices}, booktitle = {Proceedings of the Computer Vision and Pa...
Existing neural head avatars methods have achieved significant progress in the image quality and motion range of portrait animation. However, these methods prioritize effectiveness over computational overhead. This paper presents MobilePortrait, a lightweight one-shot neural head avatars method that reduces learning co...
[ 0.011731795966625214, -0.00018728077702689916, 0.012019209563732147, 0.006577837280929089, 0.011573965661227703, 0.0366935059428215, 0.016488181427121162, 0.006247198209166527, -0.03231608867645264, -0.05186820775270462, -0.022353969514369965, -0.025471754372119904, -0.06524372100830078, -...
2,826
D^3: Scaling Up Deepfake Detection by Learning from Discrepancy
[ "Yongqi Yang", "Zhihao Qian", "Ye Zhu", "Olga Russakovsky", "Yu Wu" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Yang_D3_Scaling_Up_Deepfake_Detection_by_Learning_from_Discrepancy_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Yang_D3_Scaling_Up_Deepfake_Detection_by_Learning_from_Discrepancy_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Yang_D3_Scaling_Up_CVPR_2025_supplemental.pdf
null
@InProceedings{Yang_2025_CVPR, author = {Yang, Yongqi and Qian, Zhihao and Zhu, Ye and Russakovsky, Olga and Wu, Yu}, title = {D{\textasciicircum}3: Scaling Up Deepfake Detection by Learning from Discrepancy}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)},...
The boom of Generative AI brings opportunities entangled with risks and concerns. Existing literature emphasizes the generalization capability of deepfake detection on unseen generators, significantly promoting the detector's ability to identify more universal artifacts. This work seeks a step toward a universal deepfa...
[ 0.014811050146818161, -0.017834775149822235, 0.002041971543803811, 0.06694123148918152, 0.03818841278553009, 0.021892869845032692, 0.030128201469779015, -0.003241499187424779, -0.015649424865841866, -0.04993748292326927, 0.029729999601840973, -0.005840341094881296, -0.07765432447195053, -0...
2,827
Jailbreaking the Non-Transferable Barrier via Test-Time Data Disguising
[ "Yongli Xiang", "Ziming Hong", "Lina Yao", "Dadong Wang", "Tongliang Liu" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Xiang_Jailbreaking_the_Non-Transferable_Barrier_via_Test-Time_Data_Disguising_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Xiang_Jailbreaking_the_Non-Transferable_Barrier_via_Test-Time_Data_Disguising_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Xiang_Jailbreaking_the_Non-Transferable_CVPR_2025_supplemental.pdf
2503.17198
@InProceedings{Xiang_2025_CVPR, author = {Xiang, Yongli and Hong, Ziming and Yao, Lina and Wang, Dadong and Liu, Tongliang}, title = {Jailbreaking the Non-Transferable Barrier via Test-Time Data Disguising}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}, ...
Non-transferable learning (NTL) has been proposed to protect model intellectual property (IP) by creating a "non-transferable barrier" to restrict generalization from authorized to unauthorized domains. Recently, well-designed attack, which restores the unauthorized-domain performance by fine-tuning NTL models on few a...
[ -0.0027373419143259525, -0.031379397958517075, -0.03680575266480446, 0.0444517508149147, 0.051246579736471176, -0.005743732210248709, 0.05000339448451996, -0.026755044236779213, -0.016955845057964325, -0.010240375064313412, -0.0007109769503585994, -0.023875219747424126, -0.05259482562541962,...
2,828
Light3R-SfM: Towards Feed-forward Structure-from-Motion
[ "Sven Elflein", "Qunjie Zhou", "Laura Leal-Taixé" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Elflein_Light3R-SfM_Towards_Feed-forward_Structure-from-Motion_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Elflein_Light3R-SfM_Towards_Feed-forward_Structure-from-Motion_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Elflein_Light3R-SfM_Towards_Feed-forward_CVPR_2025_supplemental.pdf
null
@InProceedings{Elflein_2025_CVPR, author = {Elflein, Sven and Zhou, Qunjie and Leal-Taix\'e, Laura}, title = {Light3R-SfM: Towards Feed-forward Structure-from-Motion}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}, month = {June}, year = {...
We present Light3R-SfM, a feed-forward, end-to-end learnable framework for efficient large-scale Structure-from-Motion (SfM) from unconstrained image collections. Unlike existing SfM solutions that rely on costly matching and global optimization to achieve accurate 3D reconstructions, Light3R-SfM addresses this limitat...
[ 0.032609883695840836, 0.0013384554767981172, 0.01751243881881237, 0.01849684678018093, 0.018651753664016724, 0.02905472181737423, -0.012012605555355549, 0.01658523455262184, -0.045119814574718475, -0.03651951253414154, 0.008946344256401062, -0.018047621473670006, -0.06480301171541214, 0.00...
2,829
Robotic Visual Instruction
[ "Yanbang Li", "Ziyang Gong", "Haoyang Li", "Xiaoqi Huang", "Haolan Kang", "Guangping Bai", "Xianzheng Ma" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Li_Robotic_Visual_Instruction_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Li_Robotic_Visual_Instruction_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Li_Robotic_Visual_Instruction_CVPR_2025_supplemental.pdf
2505.00693
@InProceedings{Li_2025_CVPR, author = {Li, Yanbang and Gong, Ziyang and Li, Haoyang and Huang, Xiaoqi and Kang, Haolan and Bai, Guangping and Ma, Xianzheng}, title = {Robotic Visual Instruction}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}, month ...
Recently, natural language has been the primary medium for human-robot interaction. However, its inherent lack of spatial precision for robotic control introduces challenges such as ambiguity and verbosity. To address these limitations, we introduce the ***Robotic Visual Instruction (RoVI)***, a novel paradigm to guide...
[ 0.017503051087260246, 0.0029813903383910656, -0.0110533582046628, 0.03458656743168831, 0.027709882706403732, 0.023965712636709213, 0.020062420517206192, 0.019667470827698708, -0.02093866653740406, -0.02766268700361252, -0.06341243535280228, 0.022538889199495316, -0.09518919885158539, -0.01...
2,830
Solving Instance Detection from an Open-World Perspective
[ "Qianqian Shen", "Yunhan Zhao", "Nahyun Kwon", "Jeeeun Kim", "Yanan Li", "Shu Kong" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Shen_Solving_Instance_Detection_from_an_Open-World_Perspective_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Shen_Solving_Instance_Detection_from_an_Open-World_Perspective_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Shen_Solving_Instance_Detection_CVPR_2025_supplemental.pdf
2503.00359
@InProceedings{Shen_2025_CVPR, author = {Shen, Qianqian and Zhao, Yunhan and Kwon, Nahyun and Kim, Jeeeun and Li, Yanan and Kong, Shu}, title = {Solving Instance Detection from an Open-World Perspective}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}, ...
Instance detection (InsDet) aims to localize specific object instances within a novel scene imagery based on given visual references. Technically, it requires proposal detection to identify all possible object instances, followed by instance-level matching to pinpoint the ones of interest. Its open-world nature support...
[ 0.007009242661297321, 0.02413933165371418, 0.005049058701843023, 0.07296260446310043, 0.02621595375239849, 0.04460799694061279, 0.03096648119390011, 0.015639686957001686, -0.027752159163355827, -0.03282645717263222, -0.05894201621413231, 0.012999143451452255, -0.08625715970993042, -0.02445...
2,831
Percept, Memory, and Imagine: World Feature Simulating for Open-Domain Unknown Object Detection
[ "Aming Wu", "Cheng Deng" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Wu_Percept_Memory_and_Imagine_World_Feature_Simulating_for_Open-Domain_Unknown_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Wu_Percept_Memory_and_Imagine_World_Feature_Simulating_for_Open-Domain_Unknown_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Wu_Percept_Memory_and_CVPR_2025_supplemental.pdf
null
@InProceedings{Wu_2025_CVPR, author = {Wu, Aming and Deng, Cheng}, title = {Percept, Memory, and Imagine: World Feature Simulating for Open-Domain Unknown Object Detection}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}, month = {June}, year ...
To accelerate the safe deployment of object detectors, we focus on reducing the impact of both covariate and semantic shifts. And we consider a realistic yet challenging scenario, namely Open-Domain Unknown Object Detection (ODU-OD), which aims to detect unknown objects in unseen target domains without accessing any au...
[ -0.008078016340732574, 0.017321234568953514, 0.0041690366342663765, 0.05257410183548927, 0.04246114566922188, 0.018180735409259796, 0.013567603193223476, 0.006547384895384312, -0.035552896559238434, -0.0504951998591423, -0.03898179531097412, -0.015011349692940712, -0.0904456377029419, 0.00...
2,832
Efficient Depth Estimation for Unstable Stereo Camera Systems on AR Glasses
[ "Yongfan Liu", "Hyoukjun Kwon" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Liu_Efficient_Depth_Estimation_for_Unstable_Stereo_Camera_Systems_on_AR_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Liu_Efficient_Depth_Estimation_for_Unstable_Stereo_Camera_Systems_on_AR_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Liu_Efficient_Depth_Estimation_CVPR_2025_supplemental.pdf
2411.10013
@InProceedings{Liu_2025_CVPR, author = {Liu, Yongfan and Kwon, Hyoukjun}, title = {Efficient Depth Estimation for Unstable Stereo Camera Systems on AR Glasses}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}, month = {June}, year = {2025}, ...
Stereo depth estimation is a fundamental component in augmented reality (AR), which requires low latency for real-time processing. However, preprocessing such as rectification and non-ML computations such as cost volume require significant amount of latency exceeding that of an ML model itself, which hinders the real-t...
[ 0.03671673685312271, 0.05561302229762077, -0.0034146320540457964, 0.03597143664956093, 0.03509613871574402, 0.04794875159859657, 0.011399558745324612, 0.026936115697026253, -0.013108080253005028, -0.06497859954833984, -0.013122351840138435, -0.0027869236655533314, -0.06015048548579216, 0.0...
2,833
3D-GRAND: A Million-Scale Dataset for 3D-LLMs with Better Grounding and Less Hallucination
[ "Jianing Yang", "Xuweiyi Chen", "Nikhil Madaan", "Madhavan Iyengar", "Shengyi Qian", "David F. Fouhey", "Joyce Chai" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Yang_3D-GRAND_A_Million-Scale_Dataset_for_3D-LLMs_with_Better_Grounding_and_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Yang_3D-GRAND_A_Million-Scale_Dataset_for_3D-LLMs_with_Better_Grounding_and_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Yang_3D-GRAND_A_Million-Scale_CVPR_2025_supplemental.pdf
null
@InProceedings{Yang_2025_CVPR, author = {Yang, Jianing and Chen, Xuweiyi and Madaan, Nikhil and Iyengar, Madhavan and Qian, Shengyi and Fouhey, David F. and Chai, Joyce}, title = {3D-GRAND: A Million-Scale Dataset for 3D-LLMs with Better Grounding and Less Hallucination}, booktitle = {Proceedings of ...
The integration of language and 3D perception is crucial for embodied agents and robots that comprehend and interact with the physical world. While large language models (LLMs) have demonstrated impressive language understanding and generation capabilities, their adaptation to 3D environments (3D-LLMs) remains in its e...
[ -0.004285360220819712, 0.0024545970372855663, 0.035027727484703064, 0.011988606303930283, 0.03758537396788597, 0.03287498280405998, 0.020792830735445023, 0.0316123329102993, -0.020276721566915512, -0.023492325097322464, -0.022853044793009758, -0.002107601845636964, -0.0703388974070549, 0.0...
2,834
LiDAR-RT: Gaussian-based Ray Tracing for Dynamic LiDAR Re-simulation
[ "Chenxu Zhou", "Lvchang Fu", "Sida Peng", "Yunzhi Yan", "Zhanhua Zhang", "Yong Chen", "Jiazhi Xia", "Xiaowei Zhou" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Zhou_LiDAR-RT_Gaussian-based_Ray_Tracing_for_Dynamic_LiDAR_Re-simulation_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Zhou_LiDAR-RT_Gaussian-based_Ray_Tracing_for_Dynamic_LiDAR_Re-simulation_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Zhou_LiDAR-RT_Gaussian-based_Ray_CVPR_2025_supplemental.pdf
null
@InProceedings{Zhou_2025_CVPR, author = {Zhou, Chenxu and Fu, Lvchang and Peng, Sida and Yan, Yunzhi and Zhang, Zhanhua and Chen, Yong and Xia, Jiazhi and Zhou, Xiaowei}, title = {LiDAR-RT: Gaussian-based Ray Tracing for Dynamic LiDAR Re-simulation}, booktitle = {Proceedings of the Computer Vision an...
This paper targets the challenge of real-time LiDAR re-simulation in dynamic driving scenarios. Recent approaches utilize neural radiance fields combined with the physical modeling of LiDAR sensors to achieve high-fidelity re-simulation results. Unfortunately, these methods face limitations due to high computational de...
[ 0.014227678067982197, 0.006949173752218485, 0.029518969357013702, 0.037891071289777756, 0.03977517411112785, 0.018624743446707726, 0.0228542722761631, 0.030483465641736984, -0.029197722673416138, -0.059481535106897354, -0.027516746893525124, -0.012801236473023891, -0.04463379457592964, -0....
2,835
Generative Zero-Shot Composed Image Retrieval
[ "Lan Wang", "Wei Ao", "Vishnu Naresh Boddeti", "Ser-Nam Lim" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Wang_Generative_Zero-Shot_Composed_Image_Retrieval_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Wang_Generative_Zero-Shot_Composed_Image_Retrieval_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Wang_Generative_Zero-Shot_Composed_CVPR_2025_supplemental.pdf
null
@InProceedings{Wang_2025_CVPR, author = {Wang, Lan and Ao, Wei and Boddeti, Vishnu Naresh and Lim, Ser-Nam}, title = {Generative Zero-Shot Composed Image Retrieval}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}, month = {June}, year = {20...
Composed Image Retrieval (CIR) is a vision-language task utilizing queries comprising images and textual descriptions to achieve precise image retrieval. This task seeks to find images that are visually similar to a reference image while incorporating specific changes or features described textually (visual delta). CIR...
[ 0.02430691570043564, -0.02936943806707859, -0.023683730512857437, 0.07891496270895004, 0.05485543608665466, 0.014407831244170666, -0.008578473702073097, 0.039458632469177246, -0.025926977396011353, -0.02567289210855961, -0.05396077781915665, -0.002468432765454054, -0.04733596369624138, 0.0...
2,836
Large-Scale Text-to-Image Model with Inpainting is a Zero-Shot Subject-Driven Image Generator
[ "Chaehun Shin", "Jooyoung Choi", "Heeseung Kim", "Sungroh Yoon" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Shin_Large-Scale_Text-to-Image_Model_with_Inpainting_is_a_Zero-Shot_Subject-Driven_Image_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Shin_Large-Scale_Text-to-Image_Model_with_Inpainting_is_a_Zero-Shot_Subject-Driven_Image_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Shin_Large-Scale_Text-to-Image_Model_CVPR_2025_supplemental.pdf
2411.15466
@InProceedings{Shin_2025_CVPR, author = {Shin, Chaehun and Choi, Jooyoung and Kim, Heeseung and Yoon, Sungroh}, title = {Large-Scale Text-to-Image Model with Inpainting is a Zero-Shot Subject-Driven Image Generator}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (...
Subject-driven text-to-image generation aims to produce images of a new subject within a desired context by accurately capturing both the visual characteristics of the subject and the semantic content of a text prompt. Traditional methods rely on time- and resource-intensive fine-tuning for subject alignment, while rec...
[ 0.024058416485786438, -0.017595387995243073, -0.023422550410032272, 0.05066872015595436, 0.03496217727661133, 0.017022905871272087, 0.03499302268028259, 0.030508138239383698, -0.04773661866784096, -0.03256624937057495, -0.05100905895233154, 0.011219439096748829, -0.05747663974761963, 0.001...
2,837
MASt3R-SLAM: Real-Time Dense SLAM with 3D Reconstruction Priors
[ "Riku Murai", "Eric Dexheimer", "Andrew J. Davison" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Murai_MASt3R-SLAM_Real-Time_Dense_SLAM_with_3D_Reconstruction_Priors_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Murai_MASt3R-SLAM_Real-Time_Dense_SLAM_with_3D_Reconstruction_Priors_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Murai_MASt3R-SLAM_Real-Time_Dense_CVPR_2025_supplemental.pdf
null
@InProceedings{Murai_2025_CVPR, author = {Murai, Riku and Dexheimer, Eric and Davison, Andrew J.}, title = {MASt3R-SLAM: Real-Time Dense SLAM with 3D Reconstruction Priors}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}, month = {June}, year ...
We present a real-time monocular dense SLAM system designed bottom-up from MASt3R, a two-view 3D reconstruction and matching prior. Equipped with this strong prior, our system is robust on in-the-wild video sequences despite making no assumption on a fixed or parametric camera model beyond a unique camera centre. We ...
[ 0.013040868565440178, 0.006106006912887096, 0.006551829166710377, 0.02085103467106819, 0.016656193882226944, 0.04172820970416069, 0.026085857301950455, 0.03314802050590515, -0.05690362676978111, -0.04770669341087341, -0.019732030108571053, -0.027894435450434685, -0.05864588916301727, -0.00...
2,838
Flow-NeRF: Joint Learning of Geometry, Poses, and Dense Flow within Unified Neural Representations
[ "Xunzhi Zheng", "Dan Xu" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Zheng_Flow-NeRF_Joint_Learning_of_Geometry_Poses_and_Dense_Flow_within_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Zheng_Flow-NeRF_Joint_Learning_of_Geometry_Poses_and_Dense_Flow_within_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Zheng_Flow-NeRF_Joint_Learning_CVPR_2025_supplemental.pdf
null
@InProceedings{Zheng_2025_CVPR, author = {Zheng, Xunzhi and Xu, Dan}, title = {Flow-NeRF: Joint Learning of Geometry, Poses, and Dense Flow within Unified Neural Representations}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}, month = {June}, y...
Learning accurate scene reconstruction without pose priors in neural radiance fields is challenging due to inherent geometric ambiguity. Recent development either relies on correspondence priors for regularization or uses off-the-shelf flow estimators to derive analytical poses. However, the potential for jointly learn...
[ 0.013308018445968628, -0.012064259499311447, -0.0006497559952549636, 0.0239462461322546, 0.0026460199151188135, 0.05475090444087982, 0.0010651558404788375, 0.000981723889708519, -0.0325450636446476, -0.06120502948760986, -0.016874141991138458, -0.03322501853108406, -0.07260693609714508, -0...
2,839
Viewpoint Rosetta Stone: Unlocking Unpaired Ego-Exo Videos for View-invariant Representation Learning
[ "Mi Luo", "Zihui Xue", "Alex Dimakis", "Kristen Grauman" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Luo_Viewpoint_Rosetta_Stone_Unlocking_Unpaired_Ego-Exo_Videos_for_View-invariant_Representation_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Luo_Viewpoint_Rosetta_Stone_Unlocking_Unpaired_Ego-Exo_Videos_for_View-invariant_Representation_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Luo_Viewpoint_Rosetta_Stone_CVPR_2025_supplemental.pdf
null
@InProceedings{Luo_2025_CVPR, author = {Luo, Mi and Xue, Zihui and Dimakis, Alex and Grauman, Kristen}, title = {Viewpoint Rosetta Stone: Unlocking Unpaired Ego-Exo Videos for View-invariant Representation Learning}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (...
Egocentric and exocentric perspectives of human action differ significantly, yet overcoming this extreme viewpoint gap is critical for applications in augmented reality and robotics. We propose ViewpointRosetta, an approach that unlocks large-scale unpaired ego and exo video data to learn clip-level viewpoint-invariant...
[ 0.023456072434782982, -0.00235351943410933, -0.013844719156622887, 0.05441450700163841, 0.035631582140922546, 0.018769914284348488, 0.03786731883883476, -0.0031231038738042116, -0.01754126325249672, -0.01840144582092762, -0.02605668269097805, 0.000565664260648191, -0.0754697248339653, -0.0...
2,840
Cross-modal Information Flow in Multimodal Large Language Models
[ "Zhi Zhang", "Srishti Yadav", "Fengze Han", "Ekaterina Shutova" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Zhang_Cross-modal_Information_Flow_in_Multimodal_Large_Language_Models_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Zhang_Cross-modal_Information_Flow_in_Multimodal_Large_Language_Models_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Zhang_Cross-modal_Information_Flow_CVPR_2025_supplemental.pdf
2411.18620
@InProceedings{Zhang_2025_CVPR, author = {Zhang, Zhi and Yadav, Srishti and Han, Fengze and Shutova, Ekaterina}, title = {Cross-modal Information Flow in Multimodal Large Language Models}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}, month = {Jun...
The recent advancements in auto-regressive multimodal large language models (MLLMs) have demonstrated promising progress for vision-language tasks. While there exists a variety of studies investigating the processing of linguistic information within large language models, little is currently known about the inner worki...
[ 0.005568631459027529, -0.009237492457032204, 0.011435681954026222, 0.02208035998046398, 0.024713944643735886, 0.00027441472047939897, 0.03525738790631294, 0.05083790048956871, -0.03830321133136749, -0.0035728432703763247, -0.032571081072092056, 0.011887695640325546, -0.06413448601961136, -...
2,841
Consistent and Controllable Image Animation with Motion Diffusion Models
[ "Xin Ma", "Yaohui Wang", "Gengyun Jia", "Xinyuan Chen", "Tien-Tsin Wong", "Yuan-Fang Li", "Cunjian Chen" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Ma_Consistent_and_Controllable_Image_Animation_with_Motion_Diffusion_Models_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Ma_Consistent_and_Controllable_Image_Animation_with_Motion_Diffusion_Models_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Ma_Consistent_and_Controllable_CVPR_2025_supplemental.pdf
2407.15642
@InProceedings{Ma_2025_CVPR, author = {Ma, Xin and Wang, Yaohui and Jia, Gengyun and Chen, Xinyuan and Wong, Tien-Tsin and Li, Yuan-Fang and Chen, Cunjian}, title = {Consistent and Controllable Image Animation with Motion Diffusion Models}, booktitle = {Proceedings of the Computer Vision and Pattern ...
Diffusion models have achieved significant progress in the task of image animation due to their powerful generative capabilities. However, preserving appearance consistency to the static input image, and avoiding abrupt motion change in the generated animation, remains challenging. In this paper, we introduce Cinemo, a...
[ 0.006409382913261652, -0.013093093410134315, -0.011866465210914612, 0.053970593959093094, 0.05593743175268173, 0.015067663975059986, 0.00566480215638876, 0.00400832574814558, -0.034590646624565125, -0.07844587415456772, 0.001937385997734964, -0.03439966216683388, -0.029540542513132095, 0.0...
2,842
Towards Better Alignment: Training Diffusion Models with Reinforcement Learning Against Sparse Rewards
[ "Zijing Hu", "Fengda Zhang", "Long Chen", "Kun Kuang", "Jiahui Li", "Kaifeng Gao", "Jun Xiao", "Xin Wang", "Wenwu Zhu" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Hu_Towards_Better_Alignment_Training_Diffusion_Models_with_Reinforcement_Learning_Against_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Hu_Towards_Better_Alignment_Training_Diffusion_Models_with_Reinforcement_Learning_Against_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Hu_Towards_Better_Alignment_CVPR_2025_supplemental.pdf
2503.11240
@InProceedings{Hu_2025_CVPR, author = {Hu, Zijing and Zhang, Fengda and Chen, Long and Kuang, Kun and Li, Jiahui and Gao, Kaifeng and Xiao, Jun and Wang, Xin and Zhu, Wenwu}, title = {Towards Better Alignment: Training Diffusion Models with Reinforcement Learning Against Sparse Rewards}, booktitle = ...
Diffusion models have achieved remarkable success in text-to-image generation. However, their practical applications are hindered by the misalignment between generated images and corresponding text prompts. To tackle this issue, reinforcement learning (RL) has been considered for diffusion model fine-tuning. Yet, RL's ...
[ 0.00137622794136405, -0.0363784097135067, -0.02225121296942234, 0.08553862571716309, 0.040534816682338715, 0.029151173308491707, 0.019141707569360733, 0.009231467731297016, -0.01551155187189579, -0.056295089423656464, -0.01994958706200123, 0.008798605762422085, -0.07167192548513412, -0.010...
2,843
Spatial457: A Diagnostic Benchmark for 6D Spatial Reasoning of Large Mutimodal Models
[ "Xingrui Wang", "Wufei Ma", "Tiezheng Zhang", "Celso M de Melo", "Jieneng Chen", "Alan Yuille" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Wang_Spatial457_A_Diagnostic_Benchmark_for_6D_Spatial_Reasoning_of_Large_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Wang_Spatial457_A_Diagnostic_Benchmark_for_6D_Spatial_Reasoning_of_Large_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Wang_Spatial457_A_Diagnostic_CVPR_2025_supplemental.pdf
null
@InProceedings{Wang_2025_CVPR, author = {Wang, Xingrui and Ma, Wufei and Zhang, Tiezheng and de Melo, Celso M and Chen, Jieneng and Yuille, Alan}, title = {Spatial457: A Diagnostic Benchmark for 6D Spatial Reasoning of Large Mutimodal Models}, booktitle = {Proceedings of the Computer Vision and Patte...
Although large multimodal models (LMMs) have demonstrated remarkable capabilities in visual scene interpretation and reasoning, their capacity for complex and precise 3-dimensional spatial reasoning remains uncertain. Existing benchmarks focus predominantly on 2D spatial understanding and lack a framework to comprehens...
[ 0.012267369776964188, 0.007342570926994085, 0.004578731022775173, 0.03490391746163368, 0.035642869770526886, 0.02669256553053856, 0.017153127118945122, 0.024617725983262062, -0.03403163701295853, -0.02420402131974697, 0.003660303307697177, 0.030069418251514435, -0.05657785385847092, 0.0011...
2,844
Omnidirectional Multi-Object Tracking
[ "Kai Luo", "Hao Shi", "Sheng Wu", "Fei Teng", "Mengfei Duan", "Chang Huang", "Yuhang Wang", "Kaiwei Wang", "Kailun Yang" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Luo_Omnidirectional_Multi-Object_Tracking_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Luo_Omnidirectional_Multi-Object_Tracking_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Luo_Omnidirectional_Multi-Object_Tracking_CVPR_2025_supplemental.pdf
2503.04565
@InProceedings{Luo_2025_CVPR, author = {Luo, Kai and Shi, Hao and Wu, Sheng and Teng, Fei and Duan, Mengfei and Huang, Chang and Wang, Yuhang and Wang, Kaiwei and Yang, Kailun}, title = {Omnidirectional Multi-Object Tracking}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Co...
Panoramic imagery, with its 360deg field of view, offers comprehensive information to support Multi-Object Tracking (MOT) in capturing spatial and temporal relationships of surrounding objects. However, most MOT algorithms are tailored for pinhole images with limited views, impairing their effectiveness in panoramic se...
[ 0.022883949801325798, 0.002010146388784051, 0.019466927275061607, 0.028339240700006485, 0.026125719770789146, 0.026423342525959015, 0.01983395591378212, 0.029706459492444992, -0.06754394620656967, -0.06324783712625504, -0.03621811792254448, -0.008680465631186962, -0.05577903240919113, -0.0...
2,845
Potential Field Based Deep Metric Learning
[ "Shubhang Bhatnagar", "Narendra Ahuja" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Bhatnagar_Potential_Field_Based_Deep_Metric_Learning_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Bhatnagar_Potential_Field_Based_Deep_Metric_Learning_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Bhatnagar_Potential_Field_Based_CVPR_2025_supplemental.zip
2405.18560
@InProceedings{Bhatnagar_2025_CVPR, author = {Bhatnagar, Shubhang and Ahuja, Narendra}, title = {Potential Field Based Deep Metric Learning}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}, month = {June}, year = {2025}, pages = {25...
Deep metric learning (DML) involves training a network to learn a semantically meaningful representation space. Many current approaches mine n-tuples of examples and model interactions within each tuplets. We present a novel, compositional DML model that instead of in tuples, represents the influence of each example (e...
[ 0.0010093378368765116, -0.01678399369120598, -0.012281309813261032, 0.04561525210738182, 0.04318820685148239, 0.027247179299592972, -0.008830396458506584, -0.01569242961704731, -0.004639758262783289, -0.02928432822227478, -0.01996220275759697, 0.0025573621969670057, -0.07968149334192276, 0...
2,846
Enhancing Vision-Language Compositional Understanding with Multimodal Synthetic Data
[ "Haoxin Li", "Boyang Li" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Li_Enhancing_Vision-Language_Compositional_Understanding_with_Multimodal_Synthetic_Data_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Li_Enhancing_Vision-Language_Compositional_Understanding_with_Multimodal_Synthetic_Data_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Li_Enhancing_Vision-Language_Compositional_CVPR_2025_supplemental.pdf
2503.01167
@InProceedings{Li_2025_CVPR, author = {Li, Haoxin and Li, Boyang}, title = {Enhancing Vision-Language Compositional Understanding with Multimodal Synthetic Data}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}, month = {June}, year = {2025}...
Paired image-text data with subtle variations in-between (e.g., people holding surfboards vs. people holding shovels) hold the promise of producing Vision-Language Models with proper compositional understanding. Synthesizing such training data from generative models is a highly coveted prize due to the reduced cost of ...
[ 0.038487017154693604, -0.003992438782006502, -0.014269106090068817, 0.06595509499311447, 0.031113121658563614, 0.00940890982747078, 0.006036965176463127, 0.05191941559314728, -0.032208312302827835, -0.034680016338825226, -0.044100597500801086, 0.027595119550824165, -0.060013387352228165, -...
2,847
Directional Label Diffusion Model for Learning from Noisy Labels
[ "Senyu Hou", "Gaoxia Jiang", "Jia Zhang", "Shangrong Yang", "Husheng Guo", "Yaqing Guo", "Wenjian Wang" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Hou_Directional_Label_Diffusion_Model_for_Learning_from_Noisy_Labels_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Hou_Directional_Label_Diffusion_Model_for_Learning_from_Noisy_Labels_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Hou_Directional_Label_Diffusion_CVPR_2025_supplemental.pdf
null
@InProceedings{Hou_2025_CVPR, author = {Hou, Senyu and Jiang, Gaoxia and Zhang, Jia and Yang, Shangrong and Guo, Husheng and Guo, Yaqing and Wang, Wenjian}, title = {Directional Label Diffusion Model for Learning from Noisy Labels}, booktitle = {Proceedings of the Computer Vision and Pattern Recognit...
In image classification, the label quality of training data critically influences model generalization, especially for deep neural networks (DNNs). Traditionally, learning from noisy labels (LNL) can improve the generalization of DNNs through complex architectures or a series of robust techniques, but its performance i...
[ -0.0007907238323241472, -0.0006743291160091758, -0.021711351349949837, 0.04343636706471443, 0.029231635853648186, 0.033475153148174286, 0.01103205606341362, -0.018234387040138245, -0.01886117085814476, -0.05045889690518379, 0.001944458344951272, -0.034236203879117966, -0.034880977123975754, ...
2,848
AA-CLIP: Enhancing Zero-Shot Anomaly Detection via Anomaly-Aware CLIP
[ "Wenxin Ma", "Xu Zhang", "Qingsong Yao", "Fenghe Tang", "Chenxu Wu", "Yingtai Li", "Rui Yan", "Zihang Jiang", "S.Kevin Zhou" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Ma_AA-CLIP_Enhancing_Zero-Shot_Anomaly_Detection_via_Anomaly-Aware_CLIP_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Ma_AA-CLIP_Enhancing_Zero-Shot_Anomaly_Detection_via_Anomaly-Aware_CLIP_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Ma_AA-CLIP_Enhancing_Zero-Shot_CVPR_2025_supplemental.pdf
null
@InProceedings{Ma_2025_CVPR, author = {Ma, Wenxin and Zhang, Xu and Yao, Qingsong and Tang, Fenghe and Wu, Chenxu and Li, Yingtai and Yan, Rui and Jiang, Zihang and Zhou, S.Kevin}, title = {AA-CLIP: Enhancing Zero-Shot Anomaly Detection via Anomaly-Aware CLIP}, booktitle = {Proceedings of the Compute...
Anomaly detection (AD) identifies outliers for applications like defect and lesion detection. While CLIP shows promise for zero-shot AD tasks due to its strong generalization capabilities, its inherent Anomaly-Unawareness leads to limited discrimination between normal and abnormal features. To address this problem, we ...
[ 0.028867250308394432, -0.013538421131670475, 0.0017514403443783522, 0.03174412623047829, 0.066127248108387, 0.011927050538361073, 0.04891682043671608, -0.011943592689931393, -0.03883454576134682, -0.0528690330684185, -0.04373723268508911, 0.014519458636641502, -0.06677886843681335, -0.0141...
2,849
HybridGS: Decoupling Transients and Statics with 2D and 3D Gaussian Splatting
[ "Jingyu Lin", "Jiaqi Gu", "Lubin Fan", "Bojian Wu", "Yujing Lou", "Renjie Chen", "Ligang Liu", "Jieping Ye" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Lin_HybridGS_Decoupling_Transients_and_Statics_with_2D_and_3D_Gaussian_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Lin_HybridGS_Decoupling_Transients_and_Statics_with_2D_and_3D_Gaussian_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Lin_HybridGS_Decoupling_Transients_CVPR_2025_supplemental.zip
2412.03844
@InProceedings{Lin_2025_CVPR, author = {Lin, Jingyu and Gu, Jiaqi and Fan, Lubin and Wu, Bojian and Lou, Yujing and Chen, Renjie and Liu, Ligang and Ye, Jieping}, title = {HybridGS: Decoupling Transients and Statics with 2D and 3D Gaussian Splatting}, booktitle = {Proceedings of the Computer Vision a...
Generating high-quality novel view renderings of 3D Gaussian Splatting (3DGS) in scenes featuring transient objects is challenging. We propose a novel hybrid representation, termed as HybridGS, using 2D Gaussians for transient objects per image and maintaining traditional 3D Gaussians for the whole static scenes. 3DGS ...
[ 0.012778287753462791, 0.013557756319642067, 0.018034253269433975, 0.033151380717754364, 0.010727472603321075, 0.012039478868246078, 0.007821241393685341, 0.0021148615051060915, -0.04718371108174324, -0.060761675238609314, -0.024776460602879524, -0.011536783538758755, -0.06973139941692352, ...
2,850
Keyframe-Guided Creative Video Inpainting
[ "Yuwei Guo", "Ceyuan Yang", "Anyi Rao", "Chenlin Meng", "Omer Bar-Tal", "Shuangrui Ding", "Maneesh Agrawala", "Dahua Lin", "Bo Dai" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Guo_Keyframe-Guided_Creative_Video_Inpainting_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Guo_Keyframe-Guided_Creative_Video_Inpainting_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Guo_Keyframe-Guided_Creative_Video_CVPR_2025_supplemental.pdf
null
@InProceedings{Guo_2025_CVPR, author = {Guo, Yuwei and Yang, Ceyuan and Rao, Anyi and Meng, Chenlin and Bar-Tal, Omer and Ding, Shuangrui and Agrawala, Maneesh and Lin, Dahua and Dai, Bo}, title = {Keyframe-Guided Creative Video Inpainting}, booktitle = {Proceedings of the Computer Vision and Pattern...
Video inpainting, which aims to fill missing regions with visually coherent content, has emerged as a crucial technique for creative applications such as editing. While existing approaches achieve visual consistency or text-guided generation, they often struggle to balance coherence and creative diversity. In this work...
[ 0.06898590922355652, -0.022347914054989815, 0.008898869156837463, 0.066348597407341, 0.03459283336997032, 0.03467680886387825, 0.022528672590851784, -0.009162979200482368, -0.06859895586967468, -0.08354293555021286, -0.0469634011387825, -0.021868525072932243, -0.04771137982606888, 0.006137...
2,851
Channel Consistency Prior and Self-Reconstruction Strategy Based Unsupervised Image Deraining
[ "Guanglu Dong", "Tianheng Zheng", "Yuanzhouhan Cao", "Linbo Qing", "Chao Ren" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Dong_Channel_Consistency_Prior_and_Self-Reconstruction_Strategy_Based_Unsupervised_Image_Deraining_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Dong_Channel_Consistency_Prior_and_Self-Reconstruction_Strategy_Based_Unsupervised_Image_Deraining_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Dong_Channel_Consistency_Prior_CVPR_2025_supplemental.pdf
2503.18703
@InProceedings{Dong_2025_CVPR, author = {Dong, Guanglu and Zheng, Tianheng and Cao, Yuanzhouhan and Qing, Linbo and Ren, Chao}, title = {Channel Consistency Prior and Self-Reconstruction Strategy Based Unsupervised Image Deraining}, booktitle = {Proceedings of the Computer Vision and Pattern Recognit...
Recently, deep image deraining models based on paired datasets have made a series of remarkable progress. However, they cannot be well applied in real-world applications due to the difficulty of obtaining real paired datasets and the poor generalization performance. In this paper, we propose a novel Channel Consistency...
[ 0.03453259542584419, -0.03801308944821358, -0.00892771314829588, 0.06830441206693649, 0.04175076633691788, -0.0033336335327476263, 0.0475928820669651, 0.02127711847424507, -0.012559262104332447, -0.057356275618076324, -0.025954758748412132, -0.01330901775509119, -0.0350772887468338, 0.0176...
2,852
MobileMamba: Lightweight Multi-Receptive Visual Mamba Network
[ "Haoyang He", "Jiangning Zhang", "Yuxuan Cai", "Hongxu Chen", "Xiaobin Hu", "Zhenye Gan", "Yabiao Wang", "Chengjie Wang", "Yunsheng Wu", "Lei Xie" ]
https://openaccess.thecvf.com/content/CVPR2025/html/He_MobileMamba_Lightweight_Multi-Receptive_Visual_Mamba_Network_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/He_MobileMamba_Lightweight_Multi-Receptive_Visual_Mamba_Network_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/He_MobileMamba_Lightweight_Multi-Receptive_CVPR_2025_supplemental.zip
2411.15941
@InProceedings{He_2025_CVPR, author = {He, Haoyang and Zhang, Jiangning and Cai, Yuxuan and Chen, Hongxu and Hu, Xiaobin and Gan, Zhenye and Wang, Yabiao and Wang, Chengjie and Wu, Yunsheng and Xie, Lei}, title = {MobileMamba: Lightweight Multi-Receptive Visual Mamba Network}, booktitle = {Proceeding...
Previous research on lightweight models has primarily focused on CNNs and Transformer-based designs. CNNs, with their local receptive fields, struggle to capture long-range dependencies, while Transformers, despite their global modeling capabilities, are limited by quadratic computational complexity in high-resolution ...
[ 0.005360848270356655, -0.013485060073435307, 0.012852984480559826, 0.03145897388458252, 0.03163314238190651, 0.0494702085852623, 0.018922271206974983, -0.01047937385737896, -0.03696205094456673, -0.046778108924627304, 0.0011520680272951722, 0.015400772914290428, -0.07620080560445786, 0.004...
2,853
EdgeTAM: On-Device Track Anything Model
[ "Chong Zhou", "Chenchen Zhu", "Yunyang Xiong", "Saksham Suri", "Fanyi Xiao", "Lemeng Wu", "Raghuraman Krishnamoorthi", "Bo Dai", "Chen Change Loy", "Vikas Chandra", "Bilge Soran" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Zhou_EdgeTAM_On-Device_Track_Anything_Model_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Zhou_EdgeTAM_On-Device_Track_Anything_Model_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Zhou_EdgeTAM_On-Device_Track_CVPR_2025_supplemental.zip
2501.07256
@InProceedings{Zhou_2025_CVPR, author = {Zhou, Chong and Zhu, Chenchen and Xiong, Yunyang and Suri, Saksham and Xiao, Fanyi and Wu, Lemeng and Krishnamoorthi, Raghuraman and Dai, Bo and Loy, Chen Change and Chandra, Vikas and Soran, Bilge}, title = {EdgeTAM: On-Device Track Anything Model}, booktitle...
On top of Segment Anything Model (SAM), SAM 2 further extends its capability from image to video inputs through a memory bank mechanism and obtains a remarkable performance compared with previous methods, making it a foundation model for video segmentation task. In this paper, we aim at making SAM 2 much more efficient...
[ -0.008264019154012203, -0.018740268424153328, 0.013362222351133823, 0.0430944450199604, 0.03142671287059784, -0.011130339466035366, 0.035353418439626694, 0.013529296033084393, -0.037899382412433624, -0.060491275042295456, -0.018606849014759064, -0.04046059027314186, -0.04832275211811066, 0...
2,854
SimLTD: Simple Supervised and Semi-Supervised Long-Tailed Object Detection
[ "Phi Vu Tran" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Tran_SimLTD_Simple_Supervised_and_Semi-Supervised_Long-Tailed_Object_Detection_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Tran_SimLTD_Simple_Supervised_and_Semi-Supervised_Long-Tailed_Object_Detection_CVPR_2025_paper.pdf
null
2412.20047
@InProceedings{Tran_2025_CVPR, author = {Tran, Phi Vu}, title = {SimLTD: Simple Supervised and Semi-Supervised Long-Tailed Object Detection}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}, month = {June}, year = {2025}, pages = {46...
While modern visual recognition systems have made significant advancements, many continue to struggle with the open problem of learning from few exemplars. This paper focuses on the task of object detection in the setting where object classes follow a natural long-tailed distribution. Existing methods for long-tailed d...
[ 0.02772606536746025, -0.05521463230252266, 0.01241225004196167, 0.03602399677038193, 0.04485670477151871, 0.00169771583750844, -0.006763066630810499, -0.0013922702055424452, -0.03628925234079361, -0.03835068270564079, -0.046090077608823776, -0.0027376743964850903, -0.07338960468769073, -0....
2,855
EarthDial: Turning Multi-sensory Earth Observations to Interactive Dialogues
[ "Sagar Soni", "Akshay Dudhane", "Hiyam Debary", "Mustansar Fiaz", "Muhammad Akhtar Munir", "Muhammad Sohail Danish", "Paolo Fraccaro", "Campbell D Watson", "Levente J Klein", "Fahad Shahbaz Khan", "Salman Khan" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Soni_EarthDial_Turning_Multi-sensory_Earth_Observations_to_Interactive_Dialogues_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Soni_EarthDial_Turning_Multi-sensory_Earth_Observations_to_Interactive_Dialogues_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Soni_EarthDial_Turning_Multi-sensory_CVPR_2025_supplemental.pdf
2412.15190
@InProceedings{Soni_2025_CVPR, author = {Soni, Sagar and Dudhane, Akshay and Debary, Hiyam and Fiaz, Mustansar and Munir, Muhammad Akhtar and Danish, Muhammad Sohail and Fraccaro, Paolo and Watson, Campbell D and Klein, Levente J and Khan, Fahad Shahbaz and Khan, Salman}, title = {EarthDial: Turning Mult...
Automated analysis of vast Earth observation data via interactive Vision-Language Models (VLMs) can unlock new opportunities for environmental monitoring, disaster response, and resource management. Existing generic VLMs do not perform well on Remote Sensing data, while the recent Geo-spatial VLMs remain restricted to ...
[ 0.041802071034908295, -0.028157029300928116, 0.021842414513230324, 0.024789461866021156, 0.027414770796895027, 0.028055323287844658, -0.0010987783316522837, 0.022137565538287163, -0.017910931259393692, -0.033323150128126144, -0.07119715213775635, 0.05074884742498398, -0.08157467097043991, ...
2,856
Learning Endogenous Attention for Incremental Object Detection
[ "Xiang Song", "Yuhang He", "Jingyuan Li", "Qiang Wang", "Yihong Gong" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Song_Learning_Endogenous_Attention_for_Incremental_Object_Detection_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Song_Learning_Endogenous_Attention_for_Incremental_Object_Detection_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Song_Learning_Endogenous_Attention_CVPR_2025_supplemental.pdf
null
@InProceedings{Song_2025_CVPR, author = {Song, Xiang and He, Yuhang and Li, Jingyuan and Wang, Qiang and Gong, Yihong}, title = {Learning Endogenous Attention for Incremental Object Detection}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}, month =...
In this paper, we focus on a challenging Incremental Object Detection (IOD) problem. Existing IOD methods follow an image-to-annotation alignment paradigm, which attempts to complete the annotations for old categories and subsequently learns both new and old categories in new tasks. This paradigm inherently introduces ...
[ 0.000112354893644806, -0.0001597888913238421, -0.0037234437186270952, 0.012531968764960766, -0.0019188023870810866, 0.031766295433044434, 0.019419582560658455, 0.02924797311425209, -0.041933443397283554, -0.026006797328591347, -0.03604234382510185, 0.031746383756399155, -0.05666270852088928,...
2,857
StarGen: A Spatiotemporal Autoregression Framework with Video Diffusion Model for Scalable and Controllable Scene Generation
[ "Shangjin Zhai", "Zhichao Ye", "Jialin Liu", "Weijian Xie", "Jiaqi Hu", "Zhen Peng", "Hua Xue", "Danpeng Chen", "Xiaomeng Wang", "Lei Yang", "Nan Wang", "Haomin Liu", "Guofeng Zhang" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Zhai_StarGen_A_Spatiotemporal_Autoregression_Framework_with_Video_Diffusion_Model_for_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Zhai_StarGen_A_Spatiotemporal_Autoregression_Framework_with_Video_Diffusion_Model_for_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Zhai_StarGen_A_Spatiotemporal_CVPR_2025_supplemental.zip
2501.05763
@InProceedings{Zhai_2025_CVPR, author = {Zhai, Shangjin and Ye, Zhichao and Liu, Jialin and Xie, Weijian and Hu, Jiaqi and Peng, Zhen and Xue, Hua and Chen, Danpeng and Wang, Xiaomeng and Yang, Lei and Wang, Nan and Liu, Haomin and Zhang, Guofeng}, title = {StarGen: A Spatiotemporal Autoregression Framew...
Recent advances in large reconstruction and generative models have significantly improved scene reconstruction and novel view generation. However, due to compute limitations, each inference with these large models is confined to a small area, making long-range consistent scene generation challenging. To address this, w...
[ 0.016701508313417435, -0.03232324495911598, 0.041049450635910034, 0.03424915298819542, 0.062283117324113846, 0.03691811487078667, 0.02205866016447544, 0.007408980745822191, -0.03141564503312111, -0.05878279730677605, -0.0057992213405668736, -0.022433921694755554, -0.0426611565053463, 0.014...
2,858
HyperSeg: Hybrid Segmentation Assistant with Fine-grained Visual Perceiver
[ "Cong Wei", "Yujie Zhong", "Haoxian Tan", "Yong Liu", "Jie Hu", "Dengjie Li", "Zheng Zhao", "Yujiu Yang" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Wei_HyperSeg_Hybrid_Segmentation_Assistant_with_Fine-grained_Visual_Perceiver_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Wei_HyperSeg_Hybrid_Segmentation_Assistant_with_Fine-grained_Visual_Perceiver_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Wei_HyperSeg_Hybrid_Segmentation_CVPR_2025_supplemental.pdf
null
@InProceedings{Wei_2025_CVPR, author = {Wei, Cong and Zhong, Yujie and Tan, Haoxian and Liu, Yong and Hu, Jie and Li, Dengjie and Zhao, Zheng and Yang, Yujiu}, title = {HyperSeg: Hybrid Segmentation Assistant with Fine-grained Visual Perceiver}, booktitle = {Proceedings of the Computer Vision and Pat...
This paper aims to address universal segmentation for image and video perception with the strong reasoning ability empowered by Visual Large Language Models (VLLMs). Despite significant progress in current unified segmentation methods, limitations in adaptation to both image and video scenarios, as well as the complex ...
[ 0.006553725339472294, 0.005382797215133905, 0.031598787754774094, 0.05540941655635834, 0.020666474476456642, 0.02508641593158245, 0.03236908093094826, 0.0285036601126194, -0.053920160979032516, -0.020609933882951736, -0.03491921350359917, -0.007957360707223415, -0.04125959798693657, 0.0067...
2,859
Diffusion-based Event Generation for High-Quality Image Deblurring
[ "Xinan Xie", "Qing Zhang", "Wei-Shi Zheng" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Xie_Diffusion-based_Event_Generation_for_High-Quality_Image_Deblurring_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Xie_Diffusion-based_Event_Generation_for_High-Quality_Image_Deblurring_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Xie_Diffusion-based_Event_Generation_CVPR_2025_supplemental.pdf
null
@InProceedings{Xie_2025_CVPR, author = {Xie, Xinan and Zhang, Qing and Zheng, Wei-Shi}, title = {Diffusion-based Event Generation for High-Quality Image Deblurring}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}, month = {June}, year = {20...
While event-based deblurring have demonstrated impressive results, they are impractical for consumer photos captured by cell phones and digital cameras that are not equipped with the event sensor. To address this problem, we in this paper propose a novel deblurring framework called Event Generation Deblurring (EGDeblur...
[ 0.000678364303894341, -0.030284730717539787, 0.0023509685415774584, 0.06726235896348953, 0.05535203590989113, 0.0109882652759552, 0.006819094996899366, 0.006553097162395716, -0.047706324607133865, -0.06161009520292282, -0.02910730242729187, -0.024510664865374565, -0.028868475928902626, 0.0...
2,860
Video Summarization with Large Language Models
[ "Min Jung Lee", "Dayoung Gong", "Minsu Cho" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Lee_Video_Summarization_with_Large_Language_Models_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Lee_Video_Summarization_with_Large_Language_Models_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Lee_Video_Summarization_with_CVPR_2025_supplemental.pdf
2504.11199
@InProceedings{Lee_2025_CVPR, author = {Lee, Min Jung and Gong, Dayoung and Cho, Minsu}, title = {Video Summarization with Large Language Models}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}, month = {June}, year = {2025}, pages ...
The exponential increase in video content poses significant challenges in terms of efficient navigation, search, and retrieval, thus requiring advanced video summarization techniques. Existing video summarization methods, which heavily rely on visual features and temporal dynamics, often fail to capture the semantics o...
[ 0.003929843660444021, -0.02282130718231201, 0.024317586794495583, 0.04020427539944649, 0.03149588406085968, -0.00708374148234725, 0.014515145681798458, 0.029618823900818825, -0.0439913235604763, -0.013287066482007504, -0.05501970648765564, 0.02314506471157074, -0.05075893923640251, 0.02111...
2,861
Sketchtopia: A Dataset and Foundational Agents for Benchmarking Asynchronous Multimodal Communication with Iconic Feedback
[ "Mohd Hozaifa Khan", "Ravi Kiran Sarvadevabhatla" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Khan_Sketchtopia_A_Dataset_and_Foundational_Agents_for_Benchmarking_Asynchronous_Multimodal_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Khan_Sketchtopia_A_Dataset_and_Foundational_Agents_for_Benchmarking_Asynchronous_Multimodal_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Khan_Sketchtopia_A_Dataset_CVPR_2025_supplemental.pdf
null
@InProceedings{Khan_2025_CVPR, author = {Khan, Mohd Hozaifa and Sarvadevabhatla, Ravi Kiran}, title = {Sketchtopia: A Dataset and Foundational Agents for Benchmarking Asynchronous Multimodal Communication with Iconic Feedback}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition C...
We introduce Sketchtopia, a large-scale dataset and AI framework designed to explore goal-driven, multimodal communication through asynchronous interactions in a Pictionary-inspired setup. Sketchtopia captures natural human interactions, including freehand sketches, open-ended guesses, and iconic feedback gestures, sho...
[ -0.009786258451640606, -0.028870727866888046, -0.01029488630592823, 0.04316442832350731, 0.015750333666801453, 0.011437206529080868, 0.03229162469506264, 0.031512267887592316, -0.025688650086522102, -0.06879351288080215, -0.04458504170179367, 0.004850943107157946, -0.08806536346673965, -0....
2,862
Consistency-aware Self-Training for Iterative-based Stereo Matching
[ "Jingyi Zhou", "Peng Ye", "Haoyu Zhang", "Jiakang Yuan", "Rao Qiang", "Liu YangChenXu", "Wu Cailin", "Feng Xu", "Tao Chen" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Zhou_Consistency-aware_Self-Training_for_Iterative-based_Stereo_Matching_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Zhou_Consistency-aware_Self-Training_for_Iterative-based_Stereo_Matching_CVPR_2025_paper.pdf
null
2503.23747
@InProceedings{Zhou_2025_CVPR, author = {Zhou, Jingyi and Ye, Peng and Zhang, Haoyu and Yuan, Jiakang and Qiang, Rao and YangChenXu, Liu and Cailin, Wu and Xu, Feng and Chen, Tao}, title = {Consistency-aware Self-Training for Iterative-based Stereo Matching}, booktitle = {Proceedings of the Computer ...
Iterative-based methods have become mainstream in stereo matching due to their high performance. However, these methods heavily rely on labeled data and face challenges with unlabeled real-world data. To this end, we propose a consistency-aware self-training framework for iterative-based stereo matching for the first t...
[ 0.030088847503066063, -0.024158399552106857, -0.003897458780556917, 0.060476988554000854, 0.02167384885251522, 0.06994922459125519, 0.01614764705300331, 0.0019281980348750949, -0.0063834842294454575, -0.044743411242961884, -0.03120608814060688, 0.015279587358236313, -0.08006872981786728, 0...
2,863
MV-MATH: Evaluating Multimodal Math Reasoning in Multi-Visual Contexts
[ "Peijie Wang", "Zhong-Zhi Li", "Fei Yin", "Dekang Ran", "Cheng-Lin Liu" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Wang_MV-MATH_Evaluating_Multimodal_Math_Reasoning_in_Multi-Visual_Contexts_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Wang_MV-MATH_Evaluating_Multimodal_Math_Reasoning_in_Multi-Visual_Contexts_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Wang_MV-MATH_Evaluating_Multimodal_CVPR_2025_supplemental.pdf
null
@InProceedings{Wang_2025_CVPR, author = {Wang, Peijie and Li, Zhong-Zhi and Yin, Fei and Ran, Dekang and Liu, Cheng-Lin}, title = {MV-MATH: Evaluating Multimodal Math Reasoning in Multi-Visual Contexts}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}, m...
Multimodal Large Language Models (MLLMs) have shown promising capabilities in mathematical reasoning within visual contexts across various datasets. However, most existing multimodal math benchmarks are limited to single-visual contexts, which diverges from the multi-visual scenarios commonly encountered in real-world ...
[ 0.008100366219878197, 0.003138056257739663, 0.01618531532585621, 0.037215299904346466, 0.0404030866920948, 0.007046710699796677, 0.03250060975551605, 0.009049100801348686, -0.03739291429519653, -0.0022329173516482115, -0.011268848553299904, 0.038566410541534424, -0.057689350098371506, 0.00...
2,864
Balanced Rate-Distortion Optimization in Learned Image Compression
[ "Yichi Zhang", "Zhihao Duan", "Yuning Huang", "Fengqing Zhu" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Zhang_Balanced_Rate-Distortion_Optimization_in_Learned_Image_Compression_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Zhang_Balanced_Rate-Distortion_Optimization_in_Learned_Image_Compression_CVPR_2025_paper.pdf
null
2502.20161
@InProceedings{Zhang_2025_CVPR, author = {Zhang, Yichi and Duan, Zhihao and Huang, Yuning and Zhu, Fengqing}, title = {Balanced Rate-Distortion Optimization in Learned Image Compression}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}, month = {June...
Learned image compression (LIC) using deep learning architectures has seen significant advancements, yet standard rate-distortion (R-D) optimization often encounters imbalanced updates due to diverse gradients of the rate and distortion objectives. This imbalance can lead to suboptimal optimization, where one objective...
[ 0.02222595363855362, -0.020850731059908867, -0.02172544039785862, 0.028834840282797813, 0.037882354110479355, 0.07911751419305801, -0.0033734154421836138, -0.008570682257413864, -0.020168058574199677, -0.0513504222035408, 0.0049534449353814125, -0.00869401078671217, -0.05830303207039833, -...
2,865
Bridge the Gap: From Weak to Full Supervision for Temporal Action Localization with PseudoFormer
[ "Ziyi Liu", "Yangcen Liu" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Liu_Bridge_the_Gap_From_Weak_to_Full_Supervision_for_Temporal_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Liu_Bridge_the_Gap_From_Weak_to_Full_Supervision_for_Temporal_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Liu_Bridge_the_Gap_CVPR_2025_supplemental.pdf
2504.14860
@InProceedings{Liu_2025_CVPR, author = {Liu, Ziyi and Liu, Yangcen}, title = {Bridge the Gap: From Weak to Full Supervision for Temporal Action Localization with PseudoFormer}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}, month = {June}, year...
Weakly-supervised Temporal Action Localization (WTAL) has achieved notable success but still suffers from a lack of temporal annotations, leading to a performance and framework gap compared with fully-supervised methods. While recent approaches employ pseudo labels for training, three key challenges: generating high-qu...
[ 0.04384039342403412, -0.05111192166805267, -0.02620437927544117, 0.024743791669607162, 0.015241896733641624, 0.003155386308208108, 0.05571932718157768, -0.004386466462165117, -0.0226732287555933, -0.011901011690497398, 0.01483329851180315, 0.0026116662193089724, -0.03804323822259903, 0.005...
2,866
HomoGen: Enhanced Video Inpainting via Homography Propagation and Diffusion
[ "Ding Ding", "Yueming Pan", "Ruoyu Feng", "Qi Dai", "Kai Qiu", "Jianmin Bao", "Chong Luo", "Zhenzhong Chen" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Ding_HomoGen_Enhanced_Video_Inpainting_via_Homography_Propagation_and_Diffusion_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Ding_HomoGen_Enhanced_Video_Inpainting_via_Homography_Propagation_and_Diffusion_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Ding_HomoGen_Enhanced_Video_CVPR_2025_supplemental.pdf
null
@InProceedings{Ding_2025_CVPR, author = {Ding, Ding and Pan, Yueming and Feng, Ruoyu and Dai, Qi and Qiu, Kai and Bao, Jianmin and Luo, Chong and Chen, Zhenzhong}, title = {HomoGen: Enhanced Video Inpainting via Homography Propagation and Diffusion}, booktitle = {Proceedings of the Computer Vision an...
In this paper, we present HomoGen, an enhanced video inpainting method based on homography propagation and diffusion models. HomoGen leverages homography registration to propagate contextual pixels as priors for generating missing content in corrupted videos. Unlike previous flow-based propagation methods, which introd...
[ 0.03075614757835865, -0.0035110246390104294, 0.006296225357800722, 0.0680546909570694, 0.04555665701627731, 0.047692958265542984, 0.012848776765167713, -0.01804598607122898, -0.03901357203722, -0.0849742740392685, 0.003234220203012228, -0.026548227295279503, -0.025215165689587593, 0.001223...
2,867
Generalized Few-shot 3D Point Cloud Segmentation with Vision-Language Model
[ "Zhaochong An", "Guolei Sun", "Yun Liu", "Runjia Li", "Junlin Han", "Ender Konukoglu", "Serge Belongie" ]
https://openaccess.thecvf.com/content/CVPR2025/html/An_Generalized_Few-shot_3D_Point_Cloud_Segmentation_with_Vision-Language_Model_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/An_Generalized_Few-shot_3D_Point_Cloud_Segmentation_with_Vision-Language_Model_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/An_Generalized_Few-shot_3D_CVPR_2025_supplemental.pdf
2503.16282
@InProceedings{An_2025_CVPR, author = {An, Zhaochong and Sun, Guolei and Liu, Yun and Li, Runjia and Han, Junlin and Konukoglu, Ender and Belongie, Serge}, title = {Generalized Few-shot 3D Point Cloud Segmentation with Vision-Language Model}, booktitle = {Proceedings of the Computer Vision and Patter...
Generalized few-shot 3D point cloud segmentation (GFS-PCS) adapts models to new classes with few support samples while retaining base class segmentation. Existing GFS-PCS methods enhance prototypes via interacting with support or query features but remain limited by sparse knowledge from few-shot samples. Meanwhile, 3D...
[ 0.008647724986076355, -0.01819547638297081, 0.029394786804914474, 0.06367212533950806, 0.008202636614441872, 0.0608050711452961, 0.02297966554760933, 0.037685517221689224, -0.041709303855895996, -0.015522710047662258, -0.0523451492190361, -0.001537412405014038, -0.061863940209150314, 0.011...
2,868
Do ImageNet-trained Models Learn Shortcuts? The Impact of Frequency Shortcuts on Generalization
[ "Shunxin Wang", "Raymond Veldhuis", "Nicola Strisciuglio" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Wang_Do_ImageNet-trained_Models_Learn_Shortcuts_The_Impact_of_Frequency_Shortcuts_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Wang_Do_ImageNet-trained_Models_Learn_Shortcuts_The_Impact_of_Frequency_Shortcuts_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Wang_Do_ImageNet-trained_Models_CVPR_2025_supplemental.pdf
2503.03519
@InProceedings{Wang_2025_CVPR, author = {Wang, Shunxin and Veldhuis, Raymond and Strisciuglio, Nicola}, title = {Do ImageNet-trained Models Learn Shortcuts? The Impact of Frequency Shortcuts on Generalization}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}...
Frequency shortcuts refer to specific frequency patterns that models heavily rely on for correct classification. Previous studies have shown that models trained on small image datasets often exploit such shortcuts, potentially impairing their generalization performance. However, existing methods for identifying freque...
[ 0.0025149344000965357, -0.041230667382478714, -0.013982567004859447, 0.02458302490413189, 0.011914743110537529, 0.020902356132864952, 0.01714870147407055, 0.02566317841410637, -0.03857167810201645, -0.035750530660152435, -0.0411696694791317, 0.030015192925930023, -0.06630595028400421, -0.0...
2,869
HORP: Human-Object Relation Priors Guided HOI Detection
[ "Pei Geng", "Jian Yang", "Shanshan Zhang" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Geng_HORP_Human-Object_Relation_Priors_Guided_HOI_Detection_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Geng_HORP_Human-Object_Relation_Priors_Guided_HOI_Detection_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Geng_HORP_Human-Object_Relation_CVPR_2025_supplemental.pdf
null
@InProceedings{Geng_2025_CVPR, author = {Geng, Pei and Yang, Jian and Zhang, Shanshan}, title = {HORP: Human-Object Relation Priors Guided HOI Detection}, booktitle = {Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR)}, month = {June}, year = {2025}, pa...
Human-Object Interaction (HOI) detection aims to predict the <Human, Interaction, Object> triplets, where the core challenge lies in recognizing the interaction of each human-object pair. Despite recent progress thanks to more advanced model architectures, HOI performance remains unsatisfactory. In this work, we first ...
[ -0.014547202736139297, 0.05408894643187523, 0.0195439625531435, 0.021035846322774887, 0.03178224712610245, 0.01983560249209404, 0.022080063819885254, 0.016613835468888283, -0.03378768637776375, -0.022154897451400757, -0.0569782517850399, 0.039031825959682465, -0.09122291952371597, -0.03008...
2,870
Building a Mind Palace: Structuring Environment-Grounded Semantic Graphs for Effective Long Video Analysis with LLMs
[ "Zeyi Huang", "Yuyang Ji", "Xiaofang Wang", "Nikhil Mehta", "Tong Xiao", "Donghyun Lee", "Sigmund Vanvalkenburgh", "Shengxin Zha", "Bolin Lai", "Licheng Yu", "Ning Zhang", "Yong Jae Lee", "Miao Liu" ]
https://openaccess.thecvf.com/content/CVPR2025/html/Huang_Building_a_Mind_Palace_Structuring_Environment-Grounded_Semantic_Graphs_for_Effective_CVPR_2025_paper.html
https://openaccess.thecvf.com/content/CVPR2025/papers/Huang_Building_a_Mind_Palace_Structuring_Environment-Grounded_Semantic_Graphs_for_Effective_CVPR_2025_paper.pdf
https://openaccess.thecvf.com/content/CVPR2025/supplemental/Huang_Building_a_Mind_CVPR_2025_supplemental.pdf
2501.04336
@InProceedings{Huang_2025_CVPR, author = {Huang, Zeyi and Ji, Yuyang and Wang, Xiaofang and Mehta, Nikhil and Xiao, Tong and Lee, Donghyun and Vanvalkenburgh, Sigmund and Zha, Shengxin and Lai, Bolin and Yu, Licheng and Zhang, Ning and Lee, Yong Jae and Liu, Miao}, title = {Building a Mind Palace: Struct...
Long-form video understanding with Large Vision Language Models is challenged by the need to analyze temporally dispersed yet spatially concentrated key moments within limited context windows. In this work, we introduce VideoMindPalace, a new framework inspired by the "Mind Palace", which organizes critical video momen...
[ 0.024854285642504692, -0.00040079333120957017, 0.009719262830913067, 0.031212594360113144, 0.03469158336520195, 0.024453550577163696, 0.021321652457118034, 0.015731822699308395, -0.022004101425409317, -0.02182147279381752, -0.02879589982330799, -0.00833714660257101, -0.052126672118902206, ...