Liunian Harold Li
Cited by
Cited by
Visualbert: A simple and performant baseline for vision and language
LH Li, M Yatskar, D Yin, CJ Hsieh, KW Chang
arXiv preprint arXiv:1908.03557, 2019
Grounded language-image pre-training
LH Li, P Zhang, H Zhang, J Yang, C Li, Y Zhong, L Wang, L Yuan, ...
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern …, 2022
How much can clip benefit vision-and-language tasks?
S Shen, LH Li, H Tan, M Bansal, A Rohrbach, KW Chang, Z Yao, ...
arXiv preprint arXiv:2107.06383, 2021
Regionclip: Region-based language-image pretraining
Y Zhong, J Yang, P Zhang, C Li, N Codella, LH Li, L Zhou, X Dai, L Yuan, ...
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern …, 2022
Glipv2: Unifying localization and vision-language understanding
H Zhang, P Zhang, X Hu, YC Chen, L Li, X Dai, L Wang, L Yuan, ...
Advances in Neural Information Processing Systems 35, 36067-36080, 2022
What Does BERT with Vision Look At?
LH Li, M Yatskar, D Yin, CJ Hsieh, KW Chang
Elevater: A benchmark and toolkit for evaluating language-augmented visual models
C Li, H Liu, L Li, P Zhang, J Aneja, J Yang, P Jin, H Hu, Z Liu, YJ Lee, ...
Advances in Neural Information Processing Systems 35, 9287-9301, 2022
Unsupervised vision-and-language pre-training without parallel images and captions
LH Li, H You, Z Wang, A Zareian, SF Chang, KW Chang
arXiv preprint arXiv:2010.12831, 2020
On the paradox of learning to reason from data
H Zhang, LH Li, T Meng, KW Chang, GV Broeck
arXiv preprint arXiv:2205.11502, 2022
Symbolic Chain-of-Thought Distillation: Small Models Can Also" Think" Step-by-Step
LH Li, J Hessel, Y Yu, X Ren, KW Chang, Y Choi
arXiv preprint arXiv:2306.14050, 2023
Broaden the vision: Geo-diverse visual commonsense reasoning
D Yin, LH Li, Z Hu, N Peng, KW Chang
arXiv preprint arXiv:2109.06860, 2021
Point precisely: Towards ensuring the precision of data in generated texts using delayed copy mechanism
L Li, X Wan
Proceedings of the 27th International Conference on Computational …, 2018
Berthop: An effective vision-and-language model for chest x-ray disease diagnosis
M Monajatipoor, M Rouhsedaghat, LH Li, CC Jay Kuo, A Chien, ...
International Conference on Medical Image Computing and Computer-Assisted …, 2022
Geomlama: Geo-diverse commonsense probing on multilingual pre-trained language models
D Yin, H Bansal, M Monajatipoor, LH Li, KW Chang
arXiv preprint arXiv:2205.12247, 2022
SGEITL: Scene graph enhanced image-text learning for visual commonsense reasoning
Z Wang, H You, LH Li, A Zareian, S Park, Y Liang, KW Chang, SF Chang
Proceedings of the AAAI Conference on Artificial Intelligence 36 (5), 5914-5922, 2022
Efficient contextual representation learning with continuous outputs
LH Li, PH Chen, CJ Hsieh, KW Chang
Transactions of the Association for Computational Linguistics 7, 611-624, 2019
MetaVL: Transferring In-Context Learning Ability From Language Models to Vision-Language Models
M Monajatipoor, LH Li, M Rouhsedaghat, LF Yang, KW Chang
arXiv preprint arXiv:2306.01311, 2023
Desco: Learning object recognition with rich language descriptions
L Li, ZY Dou, N Peng, KW Chang
Advances in Neural Information Processing Systems 36, 2024
Disinfomeme: A multimodal dataset for detecting meme intentionally spreading out disinformation
J Qu, LH Li, J Zhao, S Dev, KW Chang
arXiv preprint arXiv:2205.12617, 2022
Proceedings of the Workshop on Multilingual Multimodal Learning
E Bugliarello, KW Cheng, D Elliott, S Gella, A Kamath, LH Li, F Liu, ...
Proceedings of the Workshop on Multilingual Multimodal Learning, 2022
The system can't perform the operation now. Try again later.
Articles 1–20