default search action
Yu Qiao 0001
Person information
- affiliation: Shanghai AI Laboratory, OpenGVLab, China
- affiliation: Chinese Academy of Sciences, Shenzhen Institutes of Advanced Technology, China
- affiliation (former): University of Tokyo, Graduate School of Information Science and Technology, Japan
- affiliation (PhD 2006): University of Electro-Communications, Tokyo, Japan
Other persons with the same name
- Yu Qiao — disambiguation page
- Yu Qiao 0002 — Biomedical Imaging Lab, Singapore
- Yu Qiao 0003 — Shanghai Jiao Tong University, Department of Automation, Institute of Image Processing and Pattern Recognition, China (and 1 more)
- Yu Qiao 0004 — Kyung Hee University, School of Computing, Department of Artificial Intelligence, Yongin, South Korea (and 1 more)
- Yu Qiao 0005 — RWTH Aachen University, Germany
- Yu Qiao 0006 — Nanjing University, National Key Laboratory for Novel Software Technology, Department of Computer Science and Technology, China
Refine list
refinements active!
zoomed in on ?? of ?? records
view refined list in
export refined list as
showing all ?? records
2020 – today
- 2025
- [j111]Yu Qiao, Xiaohui Yang, Jing Wang, Tongzhen Si, Qingbei Guo:
Driver Cognitive Distraction Detection based on eye movement behavior and integration of multi-view space-channel feature. Expert Syst. Appl. 266: 125975 (2025) - [j110]Boyu Chen, Siran Chen, Kunchang Li, Qinglin Xu, Yu Qiao, Yali Wang:
Percept, Chat, Adapt: Knowledge transfer of foundation models for open-world video recognition. Pattern Recognit. 160: 111189 (2025) - [j109]Qingsong Zhao, Yi Wang, Yinan He, Yu Qiao, Cairong Zhao:
Learning Discriminative Representations in Videos via Active Embedding Distance Correlation. IEEE Signal Process. Lett. 32: 56-60 (2025) - 2024
- [j108]Yihao Liu, Hengyuan Zhao, Kelvin C. K. Chan, Xintao Wang, Chen Change Loy, Yu Qiao, Chao Dong:
Temporally consistent video colorization with deep feature propagation and self-regularization learning. Comput. Vis. Media 10(2): 375-395 (2024) - [j107]Peng Gao, Shijie Geng, Renrui Zhang, Teli Ma, Rongyao Fang, Yongfeng Zhang, Hongsheng Li, Yu Qiao:
CLIP-Adapter: Better Vision-Language Models with Feature Adapters. Int. J. Comput. Vis. 132(2): 581-595 (2024) - [j106]Kaiyang Zhou, Yongxin Yang, Yu Qiao, Tao Xiang:
MixStyle Neural Networks for Domain Generalization and Adaptation. Int. J. Comput. Vis. 132(3): 822-836 (2024) - [j105]Peng Gao, Ziyi Lin, Renrui Zhang, Rongyao Fang, Hongyang Li, Hongsheng Li, Yu Qiao:
Mimic before Reconstruct: Enhancing Masked Autoencoders with Feature Mimicking. Int. J. Comput. Vis. 132(5): 1546-1556 (2024) - [j104]Haibin He, Xinyuan Chen, Chaoyue Wang, Juhua Liu, Bo Du, Dacheng Tao, Yu Qiao:
Diff-Font: Diffusion Model for Robust One-Shot Font Generation. Int. J. Comput. Vis. 132(11): 5372-5386 (2024) - [j103]Hao Zhang, Lumin Xu, Shenqi Lai, Wenqi Shao, Nanning Zheng, Ping Luo, Yu Qiao, Kaipeng Zhang:
Open-Vocabulary Animal Keypoint Detection with Semantic-Feature Matching. Int. J. Comput. Vis. 132(12): 5741-5758 (2024) - [j102]Yuhui Wang, Yahan Xie, Yu Qiao, Zhaohui Xia, Yanying Chen:
Chinese CSUQ: Cross-Cultural Adaptation and Evaluation of Measurement Properties. Int. J. Hum. Comput. Interact. 40(22): 7623-7641 (2024) - [j101]Yi Liu, Yu Qiao, Yali Wang:
F2S-Net: learning frame-to-segment prediction for online action detection. J. Real Time Image Process. 21(3): 73 (2024) - [j100]Hongyang Li, Chonghao Sima, Jifeng Dai, Wenhai Wang, Lewei Lu, Huijie Wang, Jia Zeng, Zhiqi Li, Jiazhi Yang, Hanming Deng, Hao Tian, Enze Xie, Jiangwei Xie, Li Chen, Tianyu Li, Yang Li, Yulu Gao, Xiaosong Jia, Si Liu, Jianping Shi, Dahua Lin, Yu Qiao:
Delving Into the Devils of Bird's-Eye-View Perception: A Review, Evaluation and Recipe. IEEE Trans. Pattern Anal. Mach. Intell. 46(4): 2151-2170 (2024) - [j99]Yuexin Ma, Tai Wang, Xuyang Bai, Huitong Yang, Yuenan Hou, Yaming Wang, Yu Qiao, Ruigang Yang, Xinge Zhu:
Vision-Centric BEV Perception: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 46(12): 10978-10997 (2024) - [j98]Yingqi Liu, Jingwen He, Yihao Liu, Xinqi Lin, Fanghua Yu, Jinfan Hu, Yu Qiao, Chao Dong:
AdaptBIR: Adaptive Blind Image Restoration with latent diffusion prior for higher fidelity. Pattern Recognit. 155: 110659 (2024) - [j97]Mingfei Han, Yali Wang, Mingjie Li, Xiaojun Chang, Yi Yang, Yu Qiao:
Progressive Frame-Proposal Mining for Weakly Supervised Video Object Detection. IEEE Trans. Image Process. 33: 1560-1573 (2024) - [j96]Siran Chen, Qinglin Xu, Yue Ma, Yu Qiao, Yali Wang:
Attentive Snippet Prompting for Video Retrieval. IEEE Trans. Multim. 26: 4348-4359 (2024) - [j95]Yuer Ma, Yi Liu, Limin Wang, Wenxiong Kang, Yu Qiao, Yali Wang:
Dual Masked Modeling for Weakly-Supervised Temporal Boundary Discovery. IEEE Trans. Multim. 26: 5694-5704 (2024) - [j94]Mingye Xu, Zhipeng Zhou, Hongbin Xu, Yu Qiao, Yali Wang:
CP-Net: Contour-Perturbed Reconstruction Network for Self-Supervised Point Cloud Learning. IEEE Trans. Multim. 26: 8799-8810 (2024) - [j93]Zhangwei Gao, Zhe Chen, Erfei Cui, Yiming Ren, Weiyun Wang, Jinguo Zhu, Hao Tian, Shenglong Ye, Junjun He, Xizhou Zhu, Lewei Lu, Tong Lu, Yu Qiao, Jifeng Dai, Wenhai Wang:
Mini-InternVL: a flexible-transfer pocket multi-modal model with 5% parameters and 90% performance. Vis. Intell. 2(1): 32 (2024) - [c358]Siran Chen, Yue Ma, Yu Qiao, Yali Wang:
M-BEV: Masked BEV Perception for Robust Autonomous Driving. AAAI 2024: 1183-1191 - [c357]Ziteng Cui, Lin Gu, Xiao Sun, Xianzheng Ma, Yu Qiao, Tatsuya Harada:
Aleth-NeRF: Illumination Adaptive NeRF with Concealing Field Assumption. AAAI 2024: 1435-1444 - [c356]Bo Peng, Xinyuan Chen, Yaohui Wang, Chaochao Lu, Yu Qiao:
ConditionVideo: Training-Free Condition-Guided Video Generation. AAAI 2024: 4459-4467 - [c355]Wenshuo Peng, Kaipeng Zhang, Yue Yang, Hao Zhang, Yu Qiao:
Data Adaptive Traceback for Vision-Language Foundation Models in Image Classification. AAAI 2024: 4506-4514 - [c354]Shilin Yan, Renrui Zhang, Ziyu Guo, Wenchao Chen, Wei Zhang, Hongyang Li, Yu Qiao, Hao Dong, Zhongjiang He, Peng Gao:
Referred by Multi-Modality: A Unified Temporal Transformer for Video Object Segmentation. AAAI 2024: 6449-6457 - [c353]Lingjun Zhang, Xinyuan Chen, Yaohui Wang, Yue Lu, Yu Qiao:
Brush Your Text: Synthesize Any Scene Text on Images via Diffusion Model. AAAI 2024: 7215-7223 - [c352]Yuanfu Wang, Chao Yang, Ying Wen, Yu Liu, Yu Qiao:
Critic-Guided Decision Transformer for Offline Reinforcement Learning. AAAI 2024: 15706-15714 - [c351]Yan Ma, Yu Qiao, Pengfei Liu:
MoPS: Modular Story Premise Synthesis for Open-Ended Automatic Story Generation. ACL (1) 2024: 2135-2169 - [c350]Lijun Li, Bowen Dong, Ruohui Wang, Xuhao Hu, Wangmeng Zuo, Dahua Lin, Yu Qiao, Jing Shao:
SALAD-Bench: A Hierarchical and Comprehensive Safety Benchmark for Large Language Models. ACL (Findings) 2024: 3923-3954 - [c349]Chen Qian, Jie Zhang, Wei Yao, Dongrui Liu, Zhenfei Yin, Yu Qiao, Yong Liu, Jing Shao:
Towards Tracing Trustworthiness Dynamics: Revisiting Pre-training Period of Large Language Models. ACL (Findings) 2024: 4864-4888 - [c348]Guoxin Chen, Kexin Tang, Chao Yang, Fuying Ye, Yu Qiao, Yiming Qian:
SEER: Facilitating Structured Reasoning and Explanation via Reinforcement Learning. ACL (1) 2024: 5901-5921 - [c347]Fanqing Meng, Wenqi Shao, Quanfeng Lu, Peng Gao, Kaipeng Zhang, Yu Qiao, Ping Luo:
ChartAssistant: A Universal Chart Multimodal Language Model via Chart-to-Table Pre-training and Multitask Instruction Tuning. ACL (Findings) 2024: 7775-7803 - [c346]Zhanhui Zhou, Jie Liu, Jing Shao, Xiangyu Yue, Chao Yang, Wanli Ouyang, Yu Qiao:
Beyond One-Preference-Fits-All Alignment: Multi-Objective Direct Preference Optimization. ACL (Findings) 2024: 10586-10613 - [c345]Zaibin Zhang, Yongting Zhang, Lijun Li, Jing Shao, Hongzhi Gao, Yu Qiao, Lijun Wang, Huchuan Lu, Feng Zhao:
PsySafe: A Comprehensive Framework for Psychological-based Attack, Defense, and Evaluation of Multi-agent System Safety. ACL (1) 2024: 15202-15231 - [c344]Zhanhui Zhou, Jie Liu, Zhichen Dong, Jiaheng Liu, Chao Yang, Wanli Ouyang, Yu Qiao:
Emulated Disalignment: Safety Alignment for Large Language Models May Backfire! ACL (1) 2024: 15810-15830 - [c343]Yuan Xu, Xiaoxuan Ma, Jiajun Su, Wentao Zhu, Yu Qiao, Yizhou Wang:
ScoreHypo: Probabilistic Human Mesh Estimation with Hypothesis Scoring. CVPR 2024: 979-989 - [c342]Xiaoliang Ju, Zhaoyang Huang, Yijiin Li, Guofeng Zhang, Yu Qiao, Hongsheng Li:
DiffInDScene: Diffusion-Based High-Quality 3D Indoor Scene Generation. CVPR 2024: 4526-4535 - [c341]Xiaoyang Wu, Li Jiang, Peng-Shuai Wang, Zhijian Liu, Xihui Liu, Yu Qiao, Wanli Ouyang, Tong He, Hengshuang Zhao:
Point Transformer V3: Simpler, Faster, Stronger. CVPR 2024: 4840-4851 - [c340]Yuwen Xiong, Zhiqi Li, Yuntao Chen, Feng Wang, Xizhou Zhu, Jiapeng Luo, Wenhai Wang, Tong Lu, Hongsheng Li, Yu Qiao, Lewei Lu, Jie Zhou, Jifeng Dai:
Efficient Deformable ConvNets: Rethinking Dynamic and Sparse Operator for Vision Applications. CVPR 2024: 5652-5661 - [c339]Ziyan Chen, Jingwen He, Xinqi Lin, Yu Qiao, Chao Dong:
Towards Real-world Video Face Restoration: A New Benchmark. CVPR Workshops 2024: 5929-5939 - [c338]Lirui Zhao, Yue Yang, Kaipeng Zhang, Wenqi Shao, Yuxin Zhang, Yu Qiao, Ping Luo, Rongrong Ji:
DiffAgent: Fast and Accurate Text-to-Image API Selection with Large Language Model. CVPR 2024: 6390-6399 - [c337]Yuming Jiang, Tianxing Wu, Shuai Yang, Chenyang Si, Dahua Lin, Yu Qiao, Chen Change Loy, Ziwei Liu:
VideoBooth: Diffusion-based Video Generation with Image Prompts. CVPR 2024: 6689-6700 - [c336]Shaobin Zhuang, Kunchang Li, Xinyuan Chen, Yaohui Wang, Ziwei Liu, Yu Qiao, Yali Wang:
Vlogger: Make Your Dream A Vlog. CVPR 2024: 8806-8817 - [c335]Zehuan Huang, Hao Wen, Junting Dong, Yaohui Wang, Yangguang Li, Xinyuan Chen, Yan-Pei Cao, Ding Liang, Yu Qiao, Bo Dai, Lu Sheng:
EpiDiff: Enhancing Multi-View Synthesis via Localized Epipolar-Constrained Diffusion. CVPR 2024: 9784-9794 - [c334]Bo Zou, Chao Yang, Yu Qiao, Chengbin Quan, Youjian Zhao:
LLaMA-Excitor: General Instruction Tuning via Indirect Feature Interaction. CVPR 2024: 14089-14099 - [c333]Jiazhi Yang, Shenyuan Gao, Yihang Qiu, Li Chen, Tianyu Li, Bo Dai, Kashyap Chitta, Penghao Wu, Jia Zeng, Ping Luo, Jun Zhang, Andreas Geiger, Yu Qiao, Hongyang Li:
Generalized Predictive Model for Autonomous Driving. CVPR 2024: 14662-14672 - [c332]Yiran Qin, Enshen Zhou, Qichang Liu, Zhenfei Yin, Lu Sheng, Ruimao Zhang, Yu Qiao, Jing Shao:
MP5: A Multi-modal Open-ended Embodied System in Minecraft via Active Perception. CVPR 2024: 16307-16316 - [c331]Hao Li, Xue Yang, Zhaokai Wang, Xizhou Zhu, Jie Zhou, Yu Qiao, Xiaogang Wang, Hongsheng Li, Lewei Lu, Jifeng Dai:
Auto MC-Reward: Automated Dense Reward Design with Large Language Models for Minecraft. CVPR 2024: 16426-16435 - [c330]Yi Yu, Xue Yang, Qingyun Li, Feipeng Da, Jifeng Dai, Yu Qiao, Junchi Yan:
Point2RBox: Combine Knowledge from Synthetic Visual Patterns for End-to-End Oriented Object Detection with Single Point Supervision. CVPR 2024: 16783-16793 - [c329]Zhiyu Zhao, Bingkun Huang, Sen Xing, Gangshan Wu, Yu Qiao, Limin Wang:
Asymmetric Masked Distillation for Pre-Training Small Foundation Models. CVPR 2024: 18516-18526 - [c328]Hao Wu, Huabin Liu, Yu Qiao, Xiao Sun:
DIBS: Enhancing Dense Video Captioning with Unlabeled Videos via Pseudo Boundary Enrichment and Online Refinement. CVPR 2024: 18699-18708 - [c327]Ziqi Huang, Yinan He, Jiashuo Yu, Fan Zhang, Chenyang Si, Yuming Jiang, Yuanhan Zhang, Tianxing Wu, Qingyang Jin, Nattapol Chanpaisit, Yaohui Wang, Xinyuan Chen, Limin Wang, Dahua Lin, Yu Qiao, Ziwei Liu:
VBench: Comprehensive Benchmark Suite for Video Generative Models. CVPR 2024: 21807-21818 - [c326]Yifei Huang, Guo Chen, Jilan Xu, Mingfang Zhang, Lijin Yang, Baoqi Pei, Hongjie Zhang, Lu Dong, Yali Wang, Limin Wang, Yu Qiao:
EgoExoLearn: A Dataset for Bridging Asynchronous Ego- and Exo-centric View of Procedural Activities in Real World. CVPR 2024: 22072-22086 - [c325]Yutao Hu, Tianbin Li, Quanfeng Lu, Wenqi Shao, Junjun He, Yu Qiao, Ping Luo:
OmniMedVQA: A New Large-Scale Comprehensive Evaluation Benchmark for Medical LVLM. CVPR 2024: 22170-22183 - [c324]Kunchang Li, Yali Wang, Yinan He, Yizhuo Li, Yi Wang, Yi Liu, Zun Wang, Jilan Xu, Guo Chen, Ping Lou, Limin Wang, Yu Qiao:
MVBench: A Comprehensive Multi-modal Video Understanding Benchmark. CVPR 2024: 22195-22206 - [c323]Zhe Chen, Jiannan Wu, Wenhai Wang, Weijie Su, Guo Chen, Sen Xing, Muyan Zhong, Qinglong Zhang, Xizhou Zhu, Lewei Lu, Bin Li, Ping Luo, Tong Lu, Yu Qiao, Jifeng Dai:
Intern VL: Scaling up Vision Foundation Models and Aligning for Generic Visual-Linguistic Tasks. CVPR 2024: 24185-24198 - [c322]Fanghua Yu, Jinjin Gu, Zheyuan Li, Jinfan Hu, Xiangtao Kong, Xintao Wang, Jingwen He, Yu Qiao, Chao Dong:
Scaling Up to Excellence: Practicing Model Scaling for Photo-Realistic Image Restoration In the Wild. CVPR 2024: 25669-25680 - [c321]Yufei Wang, Wenhan Yang, Xinyuan Chen, Yaohui Wang, Lanqing Guo, Lap-Pui Chau, Ziwei Liu, Yu Qiao, Alex C. Kot, Bihan Wen:
SinSR: Diffusion-Based Image Super-Resolution in a Single Step. CVPR 2024: 25796-25805 - [c320]Jiaming Han, Kaixiong Gong, Yiyuan Zhang, Jiaqi Wang, Kaipeng Zhang, Dahua Lin, Yu Qiao, Peng Gao, Xiangyu Yue:
OneLLM: One Framework to Align All Modalities with Language. CVPR 2024: 26574-26585 - [c319]Bo Zou, Chao Yang, Yu Qiao, Chengbin Quan, Youjian Zhao:
Language-aware Visual Semantic Distillation for Video Question Answering. CVPR 2024: 27103-27113 - [c318]Ziyi Lin, Dongyang Liu, Renrui Zhang, Peng Gao, Longtian Qiu, Han Xiao, Han Qiu, Wenqi Shao, Keqin Chen, Jiaming Han, Siyuan Huang, Yichi Zhang, Xuming He, Yu Qiao, Hongsheng Li:
SPHINX: A Mixer of Weights, Visual Embeddings and Image Scales for Multi-modal Large Language Models. ECCV (62) 2024: 36-55 - [c317]Yuchen Yang, Yu Qiao, Xiao Sun:
Mask as Supervision: Leveraging Unified Mask Information for Unsupervised 3D Pose Estimation. ECCV (44) 2024: 38-55 - [c316]Shuo Cao, Yihao Liu, Wenlong Zhang, Yu Qiao, Chao Dong:
GRIDS: Grouped Multiple-Degradation Restoration with Image Degradation Similarity. ECCV (70) 2024: 70-87 - [c315]Xiangyu Chen, Zheyuan Li, Yuandong Pu, Yihao Liu, Jiantao Zhou, Yu Qiao, Chao Dong:
A Comparative Study of Image Restoration Networks for General Backbone Network Design. ECCV (71) 2024: 74-91 - [c314]Zhaoyang Liu, Zeqiang Lai, Zhangwei Gao, Erfei Cui, Ziheng Li, Xizhou Zhu, Lewei Lu, Qifeng Chen, Yu Qiao, Jifeng Dai, Wenhai Wang:
ControlLLM: Augment Language Models with Tools by Searching on Graphs. ECCV (12) 2024: 89-105 - [c313]Yunsong Zhou, Linyan Huang, Qingwen Bu, Jia Zeng, Tianyu Li, Hang Qiu, Hongzi Zhu, Minyi Guo, Yu Qiao, Hongyang Li:
Embodied Understanding of Driving Scenarios. ECCV (62) 2024: 129-148 - [c312]Gang Li, Wenhai Wang, Xiang Li, Ziheng Li, Jian Yang, Jifeng Dai, Yu Qiao, Shanshan Zhang:
Distilling Knowledge from Large-Scale Image Models for Object Detection. ECCV (84) 2024: 142-160 - [c311]Renrui Zhang, Dongzhi Jiang, Yichi Zhang, Haokun Lin, Ziyu Guo, Pengshuo Qiu, Aojun Zhou, Pan Lu, Kai-Wei Chang, Yu Qiao, Peng Gao, Hongsheng Li:
MATHVERSE: Does Your Multi-modal LLM Truly See the Diagrams in Visual Math Problems? ECCV (8) 2024: 169-186 - [c310]Jiakang Yuan, Bo Zhang, Kaixiong Gong, Xiangyu Yue, Botian Shi, Yu Qiao, Tao Chen:
Reg-TTA3D: Better Regression Makes Better Test-Time Adaptive 3D Object Detection. ECCV (43) 2024: 197-213 - [c309]Kunchang Li, Xinhao Li, Yi Wang, Yinan He, Yali Wang, Limin Wang, Yu Qiao:
VideoMamba: State Space Model for Efficient Video Understanding. ECCV (26) 2024: 237-255 - [c308]Zhihang Zhong, Gurunandan Krishnan, Xiao Sun, Yu Qiao, Sizhuo Ma, Jian Wang:
Clearer Frames, Anytime: Resolving Velocity Ambiguity in Video Frame Interpolation. ECCV (33) 2024: 346-363 - [c307]Xin Liu, Yichen Zhu, Jindong Gu, Yunshi Lan, Chao Yang, Yu Qiao:
MM-SafetyBench: A Benchmark for Safety Evaluation of Multimodal Large Language Models. ECCV (56) 2024: 386-403 - [c306]Yi Wang, Kunchang Li, Xinhao Li, Jiashuo Yu, Yinan He, Guo Chen, Baoqi Pei, Rongkun Zheng, Zun Wang, Yansong Shi, Tianxiang Jiang, Songze Li, Jilan Xu, Hongjie Zhang, Yifei Huang, Yu Qiao, Yali Wang, Limin Wang:
InternVideo2: Scaling Foundation Models for Multimodal Video Understanding. ECCV (85) 2024: 396-416 - [c305]Xinqi Lin, Jingwen He, Ziyan Chen, Zhaoyang Lyu, Bo Dai, Fanghua Yu, Yu Qiao, Wanli Ouyang, Chao Dong:
DiffBIR: Toward Blind Image Restoration with Generative Diffusion Prior. ECCV (59) 2024: 430-448 - [c304]Weiyun Wang, Yiming Ren, Haowen Luo, Tiantong Li, Chenxiang Yan, Zhe Chen, Wenhai Wang, Qingyun Li, Lewei Lu, Xizhou Zhu, Yu Qiao, Jifeng Dai:
The All-Seeing Project V2: Towards General Relation Comprehension of the Open World. ECCV (33) 2024: 471-490 - [c303]Yutong Chen, Yifan Zhan, Zhihang Zhong, Wei Wang, Xiao Sun, Yu Qiao, Yinqiang Zheng:
Within the Dynamic Context: Inertia-Aware 3D Human Modeling with Pose Sequence. ECCV (49) 2024: 491-508 - [c302]Zhaoxun Ju, Chao Yang, Fuchun Sun, Hongbo Wang, Yu Qiao:
Rethinking Mutual Information for Language Conditioned Skill Discovery on Imitation Learning. ICAPS 2024: 301-309 - [c301]Yue Yang, Kaipeng Zhang, Yuying Ge, Wenqi Shao, Zeyue Xue, Yu Qiao, Ping Luo:
Align, Adapt and Inject: Audio-Guided Image Generation, Editing and Stylization. ICASSP 2024: 3475-3479 - [c300]Yuwei Guo, Ceyuan Yang, Anyi Rao, Zhengyang Liang, Yaohui Wang, Yu Qiao, Maneesh Agrawala, Dahua Lin, Bo Dai:
AnimateDiff: Animate Your Personalized Text-to-Image Diffusion Models without Specific Tuning. ICLR 2024 - [c299]Bo Zhang, Xinyu Cai, Jiakang Yuan, Donglin Yang, Jianfei Guo, Xiangchao Yan, Renqiu Xia, Botian Shi, Min Dou, Tao Chen, Si Liu, Junchi Yan, Yu Qiao:
ReSimAD: Zero-Shot 3D Domain Transfer for Autonomous Driving with Source Reconstruction and Target Simulation. ICLR 2024 - [c298]Xinyuan Chen, Yaohui Wang, Lingjun Zhang, Shaobin Zhuang, Xin Ma, Jiashuo Yu, Yali Wang, Dahua Lin, Yu Qiao, Ziwei Liu:
SEINE: Short-to-Long Video Diffusion Model for Generative Transition and Prediction. ICLR 2024 - [c297]Mengkang Hu, Yao Mu, Xinmiao Yu, Mingyu Ding, Shiguang Wu, Wenqi Shao, Qiguang Chen, Bin Wang, Yu Qiao, Ping Luo:
Tree-Planner: Efficient Close-loop Task Planning with Large Language Models. ICLR 2024 - [c296]Wenqi Shao, Mengzhao Chen, Zhaoyang Zhang, Peng Xu, Lirui Zhao, Zhiqian Li, Kaipeng Zhang, Peng Gao, Yu Qiao, Ping Luo:
OmniQuant: Omnidirectionally Calibrated Quantization for Large Language Models. ICLR 2024 - [c295]Weigao Sun, Zhen Qin, Weixuan Sun, Shidi Li, Dong Li, Xuyang Shen, Yu Qiao, Yiran Zhong:
CO2: Efficient Distributed Training with Full Communication-Computation Overlap. ICLR 2024 - [c294]Weiyun Wang, Min Shi, Qingyun Li, Wenhai Wang, Zhenhang Huang, Linjie Xing, Zhe Chen, Hao Li, Xizhou Zhu, Zhiguo Cao, Yushi Chen, Tong Lu, Jifeng Dai, Yu Qiao:
The All-Seeing Project: Towards Panoptic Visual Recognition and Understanding of the Open World. ICLR 2024 - [c293]Yi Wang, Yinan He, Yizhuo Li, Kunchang Li, Jiashuo Yu, Xin Ma, Xinhao Li, Guo Chen, Xinyuan Chen, Yaohui Wang, Ping Luo, Ziwei Liu, Yali Wang, Limin Wang, Yu Qiao:
InternVid: A Large-scale Video-Text Dataset for Multimodal Understanding and Generation. ICLR 2024 - [c292]Licheng Wen, Daocheng Fu, Xin Li, Xinyu Cai, Tao Ma, Pinlong Cai, Min Dou, Botian Shi, Liang He, Yu Qiao:
DiLu: A Knowledge-Driven Approach to Autonomous Driving with Large Language Models. ICLR 2024 - [c291]Peng Xu, Wenqi Shao, Mengzhao Chen, Shitao Tang, Kaipeng Zhang, Peng Gao, Fengwei An, Yu Qiao, Ping Luo:
BESA: Pruning Large Language Models with Blockwise Parameter-Efficient Sparsity Allocation. ICLR 2024 - [c290]Renrui Zhang, Zhengkai Jiang, Ziyu Guo, Shilin Yan, Junting Pan, Hao Dong, Yu Qiao, Peng Gao, Hongsheng Li:
Personalize Segment Anything Model with One Shot. ICLR 2024 - [c289]Renrui Zhang, Jiaming Han, Chris Liu, Aojun Zhou, Pan Lu, Yu Qiao, Hongsheng Li, Peng Gao:
LLaMA-Adapter: Efficient Fine-tuning of Large Language Models with Zero-initialized Attention. ICLR 2024 - [c288]Wenlong Zhang, Xiaohui Li, Xiangyu Chen, Xiaoyun Zhang, Yu Qiao, Xiao-Ming Wu, Chao Dong:
SEAL: A Framework for Systematic Evaluation of Real-World Super-Resolution. ICLR 2024 - [c287]Mingzhou Liu, Xinwei Sun, Yu Qiao, Yizhou Wang:
Causal Discovery via Conditional Independence Testing with Proxy Variables. ICML 2024 - [c286]Yihao Liu, Xiangyu Chen, Xianzheng Ma, Xintao Wang, Jiantao Zhou, Yu Qiao, Chao Dong:
Unifying Image Processing as Visual Prompting Question Answering. ICML 2024 - [c285]Yao Mu, Junting Chen, Qinglong Zhang, Shoufa Chen, Qiaojun Yu, Chongjian Ge, Runjian Chen, Zhixuan Liang, Mengkang Hu, Chaofan Tao, Peize Sun, Haibao Yu, Chao Yang, Wenqi Shao, Wenhai Wang, Jifeng Dai, Yu Qiao, Mingyu Ding, Ping Luo:
RoboCodeX: Multimodal Code Generation for Robotic Behavior Synthesis. ICML 2024 - [c284]Dongyang Liu, Renrui Zhang, Longtian Qiu, Siyuan Huang, Weifeng Lin, Shitian Zhao, Shijie Geng, Ziyi Lin, Peng Jin, Kaipeng Zhang, Wenqi Shao, Chao Xu, Conghui He, Junjun He, Hao Shao, Pan Lu, Yu Qiao, Hongsheng Li, Peng Gao:
SPHINX-X: Scaling Data and Parameters for a Family of Multi-modal Large Language Models. ICML 2024 - [c283]Yue Yang, Yuqi Lin, Hong Liu, Wenqi Shao, Runjian Chen, Hailong Shang, Yu Wang, Yu Qiao, Kaipeng Zhang, Ping Luo:
Position: Towards Implicit Prompt For Text-To-Image Models. ICML 2024 - [c282]Kaining Ying, Fanqing Meng, Jin Wang, Zhiqian Li, Han Lin, Yue Yang, Hao Zhang, Wenbo Zhang, Yuqi Lin, Shuo Liu, Jiayi Lei, Quanfeng Lu, Runjian Chen, Peng Xu, Renrui Zhang, Haozhe Zhang, Peng Gao, Yali Wang, Yu Qiao, Ping Luo, Kaipeng Zhang, Wenqi Shao:
MMT-Bench: A Comprehensive Multimodal Benchmark for Evaluating Large Vision-Language Models Towards Multitask AGI. ICML 2024 - [c281]Xin Liu, Yichen Zhu, Yunshi Lan, Chao Yang, Yu Qiao:
Safety of Multimodal Large Language Models on Images and Text. IJCAI 2024: 8151-8159 - [c280]Daocheng Fu, Wenjie Lei, Licheng Wen, Pinlong Cai, Song Mao, Min Dou, Botian Shi, Yu Qiao:
LimSim++: A Closed-Loop Platform for Deploying Multimodal LLMs in Autonomous Driving. IV 2024: 1084-1090 - [c279]Xiangyu Chen, Yihao Liu, Yuandong Pu, Wenlong Zhang, Jiantao Zhou, Yu Qiao, Chao Dong:
Learning A Low-Level Vision Generalist via Visual Task Prompt. ACM Multimedia 2024: 2671-2680 - [c278]Yixu Wang, Yan Teng, Kexin Huang, Chengqi Lyu, Songyang Zhang, Wenwei Zhang, Xingjun Ma, Yu-Gang Jiang, Yu Qiao, Yingchun Wang:
Fake Alignment: Are LLMs Really Aligned Well? NAACL-HLT 2024: 4696-4712 - [c277]Daocheng Fu, Xin Li, Licheng Wen, Min Dou, Pinlong Cai, Botian Shi, Yu Qiao:
Drive Like a Human: Rethinking Autonomous Driving with Large Language Models. WACV (Workshops) 2024: 910-919 - [c276]Zeyu Lu, Chengyue Wu, Xinyuan Chen, Yaohui Wang, Lei Bai, Yu Qiao, Xihui Liu:
Hierarchical Diffusion Autoencoders and Disentangled Image Manipulation. WACV 2024: 5362-5371 - [i421]