Folgen
Linjie (Lindsey) Li
Linjie (Lindsey) Li
Senior Researcher, Microsoft
Bestätigte E-Mail-Adresse bei microsoft.com
Titel
Zitiert von
Zitiert von
Jahr
UNITER: Learning UNiversal Image-TExt Representations
YC Chen, L Li, L Yu, AE Kholy, F Ahmed, Z Gan, Y Cheng, J Liu
ECCV 2020, 2020
2131*2020
Less is More: ClipBERT for Video-and-Language Learning via Sparse Sampling
J Lei, L Li, L Zhou, Z Gan, TL Berg, M Bansal, J Liu
CVPR 2021, 2021
5272021
Large-Scale Adversarial Training for Vision-and-Language Representation Learning
Z Gan, YC Chen, L Li, C Zhu, Y Cheng, J Liu
NeurIPS 2020, 2020
4362020
HERO: Hierarchical Encoder for Video+ Language Omni-representation Pre-training
L Li, YC Chen, Y Cheng, Z Gan, L Yu, J Liu
EMNLP 2020, 2020
4322020
Relation-aware graph attention network for visual question answering
L Li, Z Gan, Y Cheng, J Liu
ICCV 2019, 2019
3642019
GIT: A Generative Image-to-text Transformer for Vision and Language
J Wang, Z Yang, X Hu, L Li, K Lin, Z Gan, Z Liu, C Liu, L Wang
TMLR, 2022
2832022
MM-REACT: Prompting ChatGPT for Multimodal Reasoning and Action
Z Yang, L Li, J Wang, K Lin, E Azarnasab, F Ahmed, Z Liu, C Liu, M Zeng, ...
arXiv preprint arXiv:2303.11381, 2023
1582023
VIOLET: End-to-End Video-Language Transformers with Masked Visual-token Modeling
TJ Fu, L Li, Z Gan, K Lin, WY Wang, L Wang, Z Liu
arXiv preprint arXiv:2111.12681, 2021
1582021
Segment Everything Everywhere All at Once
X Zou, J Yang, H Zhang, F Li, L Li, J Gao, YJ Lee
NeurIPS 2023, 2023
1522023
SwinBERT: End-to-End Transformers with Sparse Attention for Video Captioning
K Lin, L Li, CC Lin, F Ahmed, Z Gan, Z Liu, Y Lu, L Wang
CVPR 2022, 2021
1522021
The dawn of lmms: Preliminary explorations with gpt-4v (ision)
Z Yang, L Li, K Lin, J Wang, CC Lin, Z Liu, L Wang
arXiv preprint arXiv:2309.17421 9, 1, 2023
1442023
Graph Optimal Transport for Cross-Domain Alignment
L Chen, Z Gan, Y Cheng, L Li, L Carin, J Liu
ICML 2020, 2020
1382020
Multi-step reasoning via recurrent dual attention for visual dialog
Z Gan, Y Cheng, AEI Kholy, L Li, J Liu, J Gao
ACL 2019, 2019
1072019
Generalized Decoding for Pixel, Image, and Language
X Zou, ZY Dou, J Yang, Z Gan, L Li, C Li, X Dai, H Behl, J Wang, L Yuan, ...
CVPR 2023, 2022
1062022
Vision-Language Pre-training: Basics, Recent Advances, and Future Trends
Z Gan, L Li, C Li, L Wang, Z Liu, J Gao
Foundations and Trends® in Computer Graphics and Vision 14 (3–4), 163-352, 2022
1062022
VALUE: A Multi-Task Benchmark for Video-and-Language Understanding Evaluation
L Li, J Lei, Z Gan, L Yu, YC Chen, R Pillai, Y Cheng, L Zhou, XE Wang, ...
NeurIPS 2021 Data and Benchmark Track, 2021
902021
Mitigating hallucination in large multi-modal models via robust instruction tuning
F Liu, K Lin, L Li, J Wang, Y Yacoob, L Wang
ICLR 2024, 2023
89*2023
LightningDOT: Pre-training Visual-Semantic Embeddings for Real-Time Image-Text Retrieval
S Sun, YC Chen, L Li, S Wang, Y Fang, J Liu
NAACL 2021, 2021
772021
Coarse-to-Fine Vision-Language Pre-training with Fusion in the Backbone
ZY Dou, A Kamath, Z Gan, P Zhang, J Wang, L Li, Z Liu, C Liu, Y LeCun, ...
NeurIPS 2022, 2022
732022
UC2: Universal Cross-lingual Cross-modal Vision-and-Language Pre-training
M Zhou, L Zhou, S Wang, Y Cheng, L Li, Z Yu, J Liu
CVPR 2021, 2021
702021
Das System kann den Vorgang jetzt nicht ausführen. Versuchen Sie es später erneut.
Artikel 1–20