Follow
Tianlu Wang
Tianlu Wang
Facebook AI Research
Verified email at fb.com - Homepage
Title
Cited by
Cited by
Year
Opt: Open pre-trained transformer language models
S Zhang, S Roller, N Goyal, M Artetxe, M Chen, S Chen, C Dewan, ...
arXiv preprint arXiv:2205.01068, 2022
19012022
Men also like shopping: Reducing gender bias amplification using corpus-level constraints
J Zhao, T Wang, M Yatskar, V Ordonez, KW Chang
Proceedings of the 2017 Conference on Empirical Methods in Natural Language …, 2017
10972017
Gender bias in coreference resolution: Evaluation and debiasing methods
J Zhao, T Wang, M Yatskar, V Ordonez, KW Chang
Gender Bias in Coreference Resolution: Evaluation and Debiasing Methods, 15–20, 2018
8832018
Balanced datasets are not enough: Estimating and mitigating gender bias in deep image representations
T Wang, J Zhao, M Yatskar, KW Chang, V Ordonez
Proceedings of the IEEE/CVF international conference on computer vision …, 2019
4592019
Gender bias in contextualized word embeddings
J Zhao, T Wang, M Yatskar, R Cotterell, V Ordonez, KW Chang
Proceedings of the 2019 Conference of the North American Chapter of the …, 2019
4122019
General multi-label image classification with transformers
J Lanchantin, T Wang, V Ordonez, Y Qi
Proceedings of the IEEE/CVF conference on computer vision and pattern …, 2021
2992021
Selective annotation makes language models better few-shot learners
H Su, J Kasai, CH Wu, W Shi, T Wang, J Xin, R Zhang, M Ostendorf, ...
The Eleventh International Conference on Learning Representations, 2023
1272023
Visual news: Benchmark and challenges in news image captioning
F Liu, Y Wang, T Wang, V Ordonez
Proceedings of the 2021 Conference on Empirical Methods in Natural Language …, 2021
912021
Cat-gen: Improving robustness in nlp models via controlled adversarial text generation
T Wang, X Wang, Y Qin, B Packer, K Li, J Chen, A Beutel, E Chi
Proceedings of the 2020 Conference on Empirical Methods in Natural Language …, 2020
892020
Opt: Open pre-trained transformer language models, 2022
S Zhang, S Roller, N Goyal, M Artetxe, M Chen, S Chen, C Dewan, ...
URL https://arxiv. org/abs/2205.01068 3, 19-0, 2023
872023
Few-shot learning with multilingual generative language models
XV Lin, T Mihaylov, M Artetxe, T Wang, S Chen, D Simig, M Ott, N Goyal, ...
Proceedings of the 2022 Conference on Empirical Methods in Natural Language …, 2022
86*2022
Jingfei Du, et al. 2021. Few-shot learning with multilingual language models
XV Lin, T Mihaylov, M Artetxe, T Wang, S Chen, D Simig, M Ott, N Goyal, ...
arXiv preprint arXiv:2112.10668, 35-40, 2021
812021
Opt-iml: Scaling language model instruction meta learning through the lens of generalization
S Iyer, XV Lin, R Pasunuru, T Mihaylov, D Simig, P Yu, K Shuster, T Wang, ...
arXiv preprint arXiv:2212.12017, 2022
752022
Scaling autoregressive multi-modal models: Pretraining and instruction tuning
L Yu, B Shi, R Pasunuru, B Muller, O Golovneva, T Wang, A Babu, B Tang, ...
arXiv preprint arXiv:2309.02591 2 (3), 2023
742023
Identifying and mitigating spurious correlations for improving robustness in nlp models
T Wang, R Sridhar, D Yang, X Wang
Findings of the Association for Computational Linguistics: NAACL 2022, 2022
582022
Double-Hard Debias: Tailoring Word Embeddings for Gender Bias Mitigation
T Wang, XV Lin, NF Rajani, B McCann, V Ordonez, C Xiong
Proceedings of the 58th Annual Meeting of the Association for Computational …, 2020
52*2020
Name tagging for low-resource incident languages based on expectation-driven learning
B Zhang, X Pan, T Wang, A Vaswani, H Ji, K Knight, D Marcu
Proceedings of the 2016 conference of the North American chapter of the …, 2016
442016
Shepherd: A critic for language model generation
T Wang, P Yu, XE Tan, S O'Brien, R Pasunuru, J Dwivedi-Yu, ...
arXiv preprint arXiv:2308.04592, 2023
332023
Understanding in-context learning via supportive pretraining data
X Han, D Simig, T Mihaylov, Y Tsvetkov, A Celikyilmaz, T Wang
arXiv preprint arXiv:2306.15091, 2023
242023
Tao Yu
H Su, W Shi, J Kasai, Y Wang, Y Hu, M Ostendorf, W Yih, NA Smith, ...
Selective annotation makes language models better few-shot learners, 2022
192022
The system can't perform the operation now. Try again later.
Articles 1–20