OpenAlex Citation Counts

OpenAlex Citations Logo

OpenAlex is a bibliographic catalogue of scientific papers, authors and institutions accessible in open access mode, named after the Library of Alexandria. It's citation coverage is excellent and I hope you will find utility in this listing of citing articles!

If you click the article title, you'll navigate to the article, as listed in CrossRef. If you click the Open Access links, you'll navigate to the "best Open Access location". Clicking the citation count will open this listing for that article. Lastly at the bottom of the page, you'll find basic pagination options.

Requested Article:

BLIVA: A Simple Multimodal LLM for Better Handling of Text-Rich Visual Questions
Wenbo Hu, Yifan Xu, Yi Li, et al.
Proceedings of the AAAI Conference on Artificial Intelligence (2024) Vol. 38, Iss. 3, pp. 2256-2264
Open Access | Times Cited: 21

Showing 21 citing articles:

A Survey on Multimodal Large Language Models
Shukang Yin, Chaoyou Fu, Sirui Zhao, et al.
National Science Review (2024) Vol. 11, Iss. 12
Open Access | Times Cited: 61

Monkey: Image Resolution and Text Label are Important Things for Large Multi-Modal Models
Zhang Li, Biao Yang, Qiang Liu, et al.
2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2024) Vol. 35, pp. 26753-26763
Closed Access | Times Cited: 22

Multimodal Large Language Models in Healthcare: Applications, Challenges, and Future Outlook (Preprint)
Rawan AlSaad, Alaa Abd‐Alrazaq, Sabri Boughorbel, et al.
Journal of Medical Internet Research (2024) Vol. 26, pp. e59505-e59505
Open Access | Times Cited: 16

LVLM-EHub: A Comprehensive Evaluation Benchmark for Large Vision-Language Models
Peng Xu, Wenqi Shao, Kaipeng Zhang, et al.
IEEE Transactions on Pattern Analysis and Machine Intelligence (2024) Vol. 47, Iss. 3, pp. 1877-1893
Open Access | Times Cited: 12

From Large Language Models to Large Multimodal Models: A Literature Review
Dawei Huang, C.-W. Yan, Qing Li, et al.
Applied Sciences (2024) Vol. 14, Iss. 12, pp. 5068-5068
Open Access | Times Cited: 8

Large language model to multimodal large language model: A journey to shape the biological macromolecules to biological sciences and medicine
Manojit Bhattacharya, Soumen Pal, Srijan Chatterjee, et al.
Molecular Therapy — Nucleic Acids (2024) Vol. 35, Iss. 3, pp. 102255-102255
Open Access | Times Cited: 6

VLAAD: Vision and Language Assistant for Autonomous Driving
SungYeon Park, Min Jae Lee, JiHyuk Kang, et al.
(2024), pp. 980-987
Closed Access | Times Cited: 5

LION : Empowering Multimodal Large Language Model with Dual-Level Visual Knowledge
Gongwei Chen, Leyang Shen, Rui Shao, et al.
2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2024) Vol. 35, pp. 26530-26540
Closed Access | Times Cited: 5

Temporal Closeness for Enhanced Cross-Modal Retrieval of Sensor and Image Data
Shuhei Yamamoto, Noriko Kando
Lecture notes in computer science (2025), pp. 170-183
Closed Access

Stimulating conversation-style emergencies of multi-modal LMs
順時 湯浅, Bingquan Liu, Chengjie Sun, et al.
Information Fusion (2025), pp. 103047-103047
Closed Access

A Multimodal Framework Embedding Retrieval-Augmented Generation with MLLMs for Eurobarometer Data
George Papageorgiou, Vangelis Sarlis, Manolis Μaragoudakis, et al.
AI (2025) Vol. 6, Iss. 3, pp. 50-50
Open Access

Question Aware Vision Transformer for Multimodal Reasoning
Roy Ganz, Yair Kittenplon, Aviad Aberdam, et al.
2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2024) Vol. 35, pp. 13861-13871
Closed Access | Times Cited: 3

Equation-based and data-driven modeling: Open-source software current state and future directions
LaGrande Gunnell, Bethany L. Nicholson, John D. Hedengren
Computers & Chemical Engineering (2023) Vol. 181, pp. 108521-108521
Open Access | Times Cited: 9

iSMELL: Assembling LLMs with Expert Toolsets for Code Smell Detection and Refactoring
Di Wu, Fangwen Mu, Lin Shi, et al.
(2024), pp. 1345-1357
Closed Access | Times Cited: 1

Instruction Makes a Difference
Tosin Adewumi, Nudrat Habib, Lama Alkhaled, et al.
Lecture notes in computer science (2024), pp. 71-88
Closed Access

Exploring the Capabilities of Large Multimodal Models on Dense Text
Shuo Zhang, Biao Yang, Zhang Li, et al.
Lecture notes in computer science (2024), pp. 281-298
Closed Access

Multimodal Large Language Models in Health Care: Applications, Challenges, and Future Outlook (Preprint)
Rawan AlSaad, Alaa Abd‐Alrazaq, Sabri Boughorbel, et al.
(2024)
Closed Access

Collaborative Training of Tiny-Large Vision Language Models
Shichen Lu, Longteng Guo, Wenxuan Wang, et al.
(2024), pp. 4928-4937
Closed Access

Explore the Textual Perception Ability on the Images for Multimodal Large Language Models
Jiayi Kuang, Jiarui Ouyang, Ying Shen
Lecture notes in computer science (2024), pp. 300-311
Closed Access

M3DBench: Towards Omni 3D Assistant with Interleaved Multi-modal Instructions
Mingsheng Li, Xin Chen, Chi Zhang, et al.
Lecture notes in computer science (2024), pp. 41-59
Closed Access

A review of deep learning-based approaches to sign language processing
S.H. Jessica Tan, Nadeem A. Khan, Zhaoyi An, et al.
Advanced Robotics (2024), pp. 1-19
Open Access

Page 1

Scroll to top