AcademicEval / intro_8K /test_introduction_short_2404.16821v2.json
XaiverZ's picture
syn
ed3212e
raw
history blame
62.4 kB
{
"url": "http://arxiv.org/abs/2404.16821v2",
"title": "How Far Are We to GPT-4V? Closing the Gap to Commercial Multimodal Models with Open-Source Suites",
"abstract": "In this report, we introduce InternVL 1.5, an open-source multimodal large\nlanguage model (MLLM) to bridge the capability gap between open-source and\nproprietary commercial models in multimodal understanding. We introduce three\nsimple improvements: (1) Strong Vision Encoder: we explored a continuous\nlearning strategy for the large-scale vision foundation model -- InternViT-6B,\nboosting its visual understanding capabilities, and making it can be\ntransferred and reused in different LLMs. (2) Dynamic High-Resolution: we\ndivide images into tiles ranging from 1 to 40 of 448$\\times$448 pixels\naccording to the aspect ratio and resolution of the input images, which\nsupports up to 4K resolution input. (3) High-Quality Bilingual Dataset: we\ncarefully collected a high-quality bilingual dataset that covers common scenes,\ndocument images, and annotated them with English and Chinese question-answer\npairs, significantly enhancing performance in OCR- and Chinese-related tasks.\nWe evaluate InternVL 1.5 through a series of benchmarks and comparative\nstudies. Compared to both open-source and proprietary models, InternVL 1.5\nshows competitive performance, achieving state-of-the-art results in 8 of 18\nbenchmarks. Code has been released at https://github.com/OpenGVLab/InternVL.",
"authors": "Zhe Chen, Weiyun Wang, Hao Tian, Shenglong Ye, Zhangwei Gao, Erfei Cui, Wenwen Tong, Kongzhi Hu, Jiapeng Luo, Zheng Ma, Ji Ma, Jiaqi Wang, Xiaoyi Dong, Hang Yan, Hewei Guo, Conghui He, Botian Shi, Zhenjiang Jin, Chao Xu, Bin Wang, Xingjian Wei, Wei Li, Wenjian Zhang, Bo Zhang, Pinlong Cai, Licheng Wen, Xiangchao Yan, Min Dou, Lewei Lu, Xizhou Zhu, Tong Lu, Dahua Lin, Yu Qiao, Jifeng Dai, Wenhai Wang",
"published": "2024-04-25",
"updated": "2024-04-29",
"primary_cat": "cs.CV",
"cats": [
"cs.CV"
],
"label": "Original Paper",
"paper_cat": "LLM Fairness",
"gt": "Large language models (LLMs) have been instrumental in advancing artificial general intelligence (AGI) systems, * equal contribution; \u2020 interns at OpenGVLab, Shanghai AI Labora- tory; B corresponding author ([email protected]). Benchmark InternVL 1.5 Grok-1.5V GPT-4V Claude-3 Opus Gemini Pro 1.5 MMMU Multi-discipline 45.2% 53.6% 56.8% 59.4% 58.5% MathVista Math 53.5% 52.8% 49.9% 50.5% 52.1% AI2D Diagrams 80.7% 88.3% 78.2% 88.1% 80.3% TextVQA Text reading 80.6% 78.1% 78.0% - 73.5% ChartQA Charts 83.8% 76.1% 78.5% 80.8% 81.3% DocVQA Documents 90.9% 85.6% 88.4% 89.3% 86.5% RealWorldQA Real-world understanding 66.0% 68.7% 61.4% 49.8% 67.5% Figure 1. InternVL 1.5 versus proprietary commercial models. The results of these benchmarks show that InternVL 1.5 achieves performance comparable to leading proprietary models. demonstrating remarkable abilities in processing open- world language tasks. Leveraging the advancements in LLMs, multimodal large language models (MLLMs) [5, 18, 23, 62, 63, 84, 92, 116, 142] have made significant strides, facilitating complex vision-language dialogues and interac- tions that bridge the gap between textual and visual infor- mation. Despite these achievements, there remains a notice- able divide between the capabilities of open-source mod- els and proprietary commercial models, e.g., GPT-4V [87], 1 arXiv:2404.16821v2 [cs.CV] 29 Apr 2024 Gemini series [92, 107], and Qwen-VL-Max [5]. This gap is mainly reflected in the following three as- pects: (1) Parameter Scale: Recent proprietary commer- cial MLLMs [5, 87, 92, 102] typically scales not less than 100 billion parameters, while open-source models com- monly employ a 300 million parameter vision foundation model (VFM), which is integrated with either a 7 billion or 13 billion LLMs. (2) Image Resolution: Proprietary commercial models typically employ a dynamic resolution approach, preserving the original aspect ratio to facilitate detailed scene and document understanding. In contrast, open-source models generally train with fixed resolutions [18, 23, 62, 71, 117, 142], such as 336\u00d7336 and 448\u00d7448, leading to a considerable gap in capabilities relative to com- mercial counterparts. (3) Multilingual Capability: Propri- etary models often leverage extensive multilingual datasets for training, enhancing their performance across diverse languages. However, open-source models predominantly utilize English data, relying on the zero-shot capabilities of LLMs for other languages, e.g. LLaVA-NeXT [64]. This re- sults in sub-optimal performance in non-English scene un- derstanding and OCR tasks. To bridge the gap, we introduce InternVL 1.5, integrat- ing three major improvements to enhance its performance and usability. (1) We implement a continuous learning ap- proach to a large-scale VFM\u2014InternViT-6B [18], refining it using high-quality image-text data. This process not only enhances the model\u2019s ability to understand visual content but also improves its adaptability across various LLMs. In addition, using InternLM2-20B [11] as the language foun- dation model also offers robust initial language processing capabilities. (2) We adopt a dynamic high-resolution strat- egy that segments images into 448\u00d7448 tiles, with the num- ber of tiles ranging from 1 to 40 (i.e., 4K resolution) based on the aspect ratio and resolution of the images. To capture global context, we additionally include a thumbnail view. (3) We gather a diverse collection of public datasets, cov- ering high-quality natural scenes, charts, documents, and conversations in both English and Chinese. Additionally, we develop a data translation pipeline using open-source LLMs, which can be easily extended to more languages. These designs endow our model with several advan- tages: (1) Flexible Resolution: Similar to the \u201clow\u201d or \u201chigh\u201d modes available in GPT-4V [87], InternVL 1.5 en- ables users to select the optimal resolution for their im- ages, such as using low-resolution for scene subject de- scription and high-resolution (up to 4K resolution) for doc- ument understanding, effectively balancing computational efficiency with detail preservation. (2) Bilingual Profi- ciency: InternVL 1.5 exhibits robust bilingual capabili- ties, proficiently handling multimodal perception and un- derstanding tasks in both English and Chinese. Notably, in tasks related to Chinese, our model generally outperforms InternVL 1.5 High-Quality Bilingual Dataset Captioning, General QA, Science, Chart, Mathematics, Knowledge, OCR, Document, Grounding, Conversation, Chinese, English Strong Foundation Models InternViT-6B-448px-V1.5 + InternLM2-20B Dynamic High-Resolution 448 ~ 4K Resolution AGI Figure 2. Characteristics of InternVL 1.5. InternVL 1.5 features strong visual representation through continuous learning, flexible resolution capabilities, and robust bilingual proficiency in English and Chinese, positioning it as a competitive MLLM. the leading commercial model GPT-4V [87]. (3) Strong Vi- sual Representation: By implementing a continuous learn- ing strategy, we enhance the visual representation capabil- ities of InternViT-6B [18], making it robust to flexible in- put resolution and various visual domains. Benefitting from InternViT-6B\u2019s massive parameters, our model achieves a level of visual representation that rivals the linguistic capa- bilities of LLMs with more than 20 billion parameters. This synergy between visual and linguistic processing endows our system with robust multimodal capabilities. We evaluated InternVL 1.5 on 18 representative multi- modal benchmarks, which are categorized into four spe- cific groups: OCR-related, general multimodal, mathemat- ical, and multi-turn conversation benchmarks. Compared to both open-source and proprietary models, InternVL 1.5 shows competitive performance, achieving state-of-the-art results in 8 of 18 benchmarks. Notably, as shown in Fig- ure 1, it even surpasses leading proprietary models like Grok-1.5V [125], GPT-4V [87], Claude-3 Opus [3], and Gemini Pro 1.5 [92] in four specific benchmarks, partic- ularly in OCR-related datasets such as TextVQA [100], ChartQA [81], and DocVQA [82]. This evaluation indicates that InternVL 1.5 has effectively narrowed the gap between open-source models and leading commercial models. We hope that our approach and open-source model weights can contribute to the development of the MLLM community. 2",
"main_content": "2.1. Proprietary Commercial MLLMs Large language models (LLMs) [1, 4, 7, 8, 11, 25, 104, 106, 108, 112, 113, 122, 123, 141] have greatly advanced AGI by enabling complex language tasks previously thought human-exclusive. Building on this, the development of proprietary commercial MLLMs represents a significant evolution. For example, OpenAI\u2019s GPT-4V [87] extends GPT4\u2019s capabilities by incorporating visual inputs, allowing it to handle both text and image content, which stands as a significant development in the domain of MLLMs. Afterward, Google\u2019s Gemini series progresses from Gemini 1.0 [107] to Gemini 1.5 [92], enhancing MLLMs with the ability to process text, images, and audio and support up to 1 million tokens, which boosts performance significantly. The QwenVL-Plus/Max are Alibaba\u2019s leading models in the QwenVL series [5], renowned for superior capacity in multimodal tasks without needing OCR tools. Recent advancements in proprietary MLLMs include Anthropic\u2019s Claude-3V series [3], HyperGAI\u2019s HPT Pro [35], Apple\u2019s MM1 [84], StepFun\u2019s Step-1V [102], and xAI\u2019s Grok-1.5V [125]. 2.2. Open-Source MLLMs The development of open-source MLLMs [2, 13, 43, 48, 51, 55, 56, 69, 70, 103, 110, 118, 120, 124, 138, 139] has significantly influenced the AGI landscape by integrating and enhancing capabilities in processing both visual and textual data. Over the past year, many open-source MLLMs have become well-known, including the LLaVA series [62\u2013 64], MiniGPT-4 [142], VisionLLM [116], Qwen-VL [5], CogVLM [117], Shikra [15], and others [18, 23, 90, 119]. However, these models are typically trained on images with small, fixed resolutions such as 336\u00d7336, or 448\u00d7448, which leads to sub-optimal performance on images with unusual aspect ratios or document data. To address this issue, many approaches have been explored for training on high-resolution images. Currently, there are two common technical routes: one involves designing a dual-branch image encoder [32, 53, 76, 77, 121], and the other involves dividing a high-resolution image into many low-resolution tiles [24, 33, 47, 55, 57, 64, 68, 126, 127]. Despite these explorations in high-resolution training, these open-source models still exhibit significant gaps in understanding documents, charts, and infographics, as well as recognizing scene texts, compared to leading commercial models. 2.3. Vision Foundation Models for MLLMs Vision foundation models (VFMs) are a focal point of research within the MLLM community. Currently, models like CLIP-ViT [91] and SigLIP [136] are prevalently utilized; however, many studies have been conducted to find the most suitable vision encoders for MLLMs [57, 71, 76, Pixel Shuffle Dynamic High Resolution InternViT-6B InternLM2 Tokenizer InternLM2-Chat-20B MLP Projector User Message Figure 3. Overall Architecture. InternVL 1.5 adopts the ViTMLP-LLM architecture similar to popular MLLMs [62, 64], combining a pre-trained InternViT-6B [18] with InternLM2-20B [11] through a MLP projector. Here, we employ a simple pixel shuffle to reduce the number of visual tokens to one-quarter. 111]. For instance, Tong et al. [111] observed notable differences in the visual patterns of CLIP and DINOv2 [88], leading to the development of a mixture-of-features module that combines these two VFMs. LLaVA-HR [76] introduced a dual-branch vision encoder utilizing CLIP-ViT for low-resolution pathways and CLIP-ConvNext for highresolution pathways. Similarly, DeepSeek-VL [71] adopted a dual vision encoder design, using SigLIP-L for lowresolution images and SAM-B for high-resolution images. In this report, we propose a continuous learning strategy for our vision foundation model\u2014InternViT-6B [18], which continuously boosts the visual understanding capabilities and can be transferred and reused across different LLMs. 3. InternVL 1.5 3.1. Overall Architecture As illustrated in Figure 3, InternVL 1.5 employs an architecture akin to widely-used open-source MLLMs, specifically the \u201cViT-MLP-LLM\u201d configuration referenced in various existing studies [18, 23, 62\u201364, 71, 142]. Our implementation of this architecture integrates a pre-trained InternViT-6B [18] with a pre-trained InternLM2-20B [11] using a randomly initialized MLP projector. During training, we implemented a dynamic resolution strategy, dividing images into tiles of 448\u00d7448 pixels in sizes ranging from 1 to 12, based on the aspect ratio and resolution of the input images. During testing, this can be zero-shot scaled up to 40 tiles (i.e., 4K resolution). To enhance scalability for high resolution, we simply employed a pixel shuffle operation to reduce the number of visual tokens to one-quarter of the original. Therefore, in our model, a 448\u00d7448 image is represented by 256 visual tokens. 3 3.2. Strong Vision Encoder In existing MLLMs [5, 23, 62\u201364, 78, 142], the most commonly used vision foundation model is typically a contrastively pre-trained ViT [18, 36, 91, 136]. However, these ViTs are commonly trained on image-text pairs crawled from the Internet at a fixed low resolution (e.g., 224\u00d7224), so their performance degrades when tasked with processing high-resolution images or images from sources other than the Internet, such as document images. InternViT-6B-448px-V1.2. To address this issue, the InternVL 1.2 update involved continuous pre-training of the InternViT-6B model. First, we found that the features from the fourth-to-last layer perform best for multimodal tasks, so we directly discarded the weights of the last three layers, reducing InternViT-6B from 48 layers to 45 layers. Then, we increased the resolution of InternViT-6B from 224 to 448 and integrated it with Nous-Hermes-2-Yi-34B [130]. To equip the model with high-resolution processing and OCR capabilities, both the vision encoder and the MLP were activated for training, utilizing a mix of image captioning [10, 17, 90, 93, 100] and OCR-specific datasets [29, 94]. The newly derived InternViT weights from this process were released as InternViT-6B-448px-V1.21. InternViT-6B-448px-V1.5. The development of InternVL 1.5 continues the pre-training of the strong foundation of InternViT-6B-448px-V1.2. In this update, the resolution of training images is expanded from fixed 448\u00d7448 to dynamic 448\u00d7448, where the basic tile size is 448\u00d7448 and the number of tiles ranges from 1 to 12. Additionally, we enhance the data scale, quality, and diversity of the pretraining dataset, resulting in the powerful robustness, OCR capability, and high-resolution processing capability of our 1.5 version model2. Details of the dynamic resolution and training datasets are described in Sections 3.3 and 3.4. It is noteworthy that despite the LLM in InternVL 1.5 being changed from Nous-Hermes-2-Yi-34B to InternLM220B [11], the InternViT maintained excellent compatibility and portability with the new LLM. This suggests that the visual features learned by InternViT-6B during the pretraining stage of MLLMs are broadly applicable and not tightly bound to the specific LLM. 3.3. Dynamic High-Resolution Inspired by UReader [127], we adopt a dynamic highresolution training approach that effectively adapts to the varying resolutions and aspect ratios of input images. This method leverages the flexibility of segmenting images into tiles, enhancing the model\u2019s ability to process detailed vi1https://huggingface.co/OpenGVLab/InternViT-6B448px-V1-2 2https://huggingface.co/OpenGVLab/InternViT-6B448px-V1-5 1:1 1:4 2:3 1:2 1:5 3:2 1:3 1:6 ... Pre-defined Aspect Ratios 448\u00d7448 Tiles Matching 2:3 (896\u00d71344) Thumbnail Input Image (800\u00d71300) Figure 4. Illustration of dynamic high resolution. We dynamically match an optimal aspect ratio from pre-defined ratios, dividing the image into tiles of 448\u00d7448 pixels and creating a thumbnail for global context. This method minimizes aspect ratio distortion and accommodates varying resolutions during training. sual information while accommodating diverse image resolutions. It mainly consists of the following steps: Dynamic Aspect Ratio Matching. As shown in Figure 4, to maintain natural aspect ratios during processing, we dynamically match the optimal aspect ratio from a pre-defined set of aspect ratios. Due to limited computational resources, we allow a maximum of 12 tiles during training. Consequently, this set includes all 35 possible combinations of aspect ratios formed by 1 to 12 tiles, such as {1:1, 1:2, 2:1, 3:1, ..., 2:6}. During the matching process, for each input image, we calculate its aspect ratio and compare it with the 35 pre-defined aspect ratios by measuring the absolute difference. If multiple pre-defined aspect ratios match (e.g., 1:1 and 2:2), we prioritize the one not exceeding twice the input image\u2019s area, thereby preventing excessive enlargement of low-resolution images. Image Division & Thumbnail. Once an appropriate aspect ratio is determined, the image is resized to the corresponding resolution. For example, an 800\u00d71300 image will be resized to 896\u00d71344. The resized image is then divided into tiles of 448\u00d7448 pixels. Alongside the tiles, we include a thumbnail of the entire image to capture the global context. This thumbnail is scaled down to 448\u00d7448, aiding the model in understanding the overall scene. Therefore, during training, the number of visual tokens ranges from 256 to 3,328. During testing, the number of tiles can increase to a maximum of 40, resulting in 10,496 visual tokens. 3.4. High-Quality Bilingual Dataset Pre-training Dataset. The pre-training dataset utilized in our InternVL 1.5 encompasses a diverse range of publicly accessible sources. We provide an overview of these datasets in Table 1a. These datasets span multi4 task ratio dataset Laion-EN (en) [93], Laion-ZH (zh) [93], COYO (zh) [10], Captioning 53.9% GRIT (zh) [90], COCO (en) [17], TextCaps (en) [99] Objects365 (en&zh) [97], GRIT (en&zh) [90], Detection 5.2% All-Seeing (en&zh) [119] Wukong-OCR (zh) [29], LaionCOCO-OCR (en) [94], OCR (large) 32.0% Common Crawl PDF (en&zh) MMC-Inst (en) [61], LSVT (zh) [105], ST-VQA (en) [9] RCTW-17 (zh) [98], ReCTs (zh) [137], ArT (en&zh) [19], SynthDoG (en&zh) [41], COCO-Text (en) [114], ChartQA (en) [81], CTW (zh) [134], DocVQA (en) [82], OCR (small) 8.9% TextOCR (en) [101], PlotQA (en) [85], InfoVQA (en) [83] (a) Datasets used in the pre-training stage. task dataset Captioning TextCaps (en) [99], ShareGPT4V (en&zh) [16] VQAv2 (en) [28], GQA (en) [34], OKVQA (en) [80], General QA VSR (en) [59], VisualDialog (en) [22] Science AI2D (en) [39], ScienceQA (en) [73], TQA (en) [40] ChartQA (en) [81], MMC-Inst (en) [61], DVQA (en) [38], Chart PlotQA (en) [85], LRV-Instruction (en) [60] GeoQA+ (en) [12], TabMWP (en) [74], MathQA (en) [132], Mathematics CLEVR-Math/Super (en) [54, 58], Geometry3K (en) [72] KVQA (en) [96], A-OKVQA (en) [95], ViQuAE (en) [45], Knowledge Wikipedia (en&zh) [31] OCRVQA (en) [86], InfoVQA (en) [83], TextVQA (en) [100], ArT (en&zh) [19], COCO-Text (en) [114], CTW (zh) [134], LSVT (zh) [105], RCTW-17 (zh) [98], ReCTs (zh) [137], OCR SynthDoG (en&zh) [41], ST-VQA (en) [9] Document DocVQA (en) [20], Common Crawl PDF (en&zh) Grounding RefCOCO/+/g (en) [79, 131], Visual Genome (en) [42] LLaVA-150K (en&zh) [63], LVIS-Instruct4V (en) [115], ALLaVA (en&zh) [14], Laion-GPT4V (en) [44], Conversation TextOCR-GPT4V (en) [37], SVIT (en&zh) [140] OpenHermes2.5 (en) [109], Alpaca-GPT4 (en) [106], Text-only ShareGPT (en&zh) [141], COIG-CQIA (zh) [6] (b) Datasets used in the fine-tuning stage. Table 1. Summary of datasets used in InternVL 1.5. To construct large-scale OCR datasets, we utilized PaddleOCR [49] to perform OCR in Chinese on images from Wukong [29] and in English on images from LAION-COCO [94]. ple tasks, including captioning, which predominantly uses datasets such as Laion-EN [93], Laion-ZH [93], COYO [10], and GRIT [90], constituting 53.9% of the total data. Detection and grounding tasks utilize datasets like Objects365 [97], GRIT [90], and All-Seeing [119], making up 5.2%. For OCR tasks, we utilized large-scale datasets such as Wukong-OCR, LaionCOCO-OCR, and Common Crawl PDFs, which constitute 32.0% of our data. These datasets were constructed using PaddleOCR [49] to perform OCR on Chinese images from Wukong [29] and on English images from LaionCOCO [94]. Smaller OCR datasets include MMC-Inst [61], LSVT [105], ST-VQA [9], RCTW-17 [98], ArT [19], and others, accounting for 8.9% of the data, which focus on more specific or constrained OCR challenges. This diverse dataset assembly ensures robust model pre-training of InternVL, catering to varied linguistic and visual elements across tasks. Fine-tuning Dataset. During the fine-tuning stage, we System: You are a translator proficient in English and {language}. Your task is to translate the following English text into {language}, focusing on a natural and fluent result that avoids \u201ctranslationese.\u201d Please consider these points: 1. Keep proper nouns, brands, and geographical names in English. 2. Retain technical terms or jargon in English, but feel free to explain in {language} if necessary. 3. Use {language} idiomatic expressions for English idioms or proverbs to ensure cultural relevance. 4. Ensure quotes or direct speech sound natural in {language}, maintaining the original\u2019s tone. 5. For acronyms, provide the full form in {language} with the English acronym in parentheses. User: Text for translation: {text} Assistant: {translation results} Figure 5. Explanation of our data translation pipeline. Based on this prompt, we translate English data into Chinese while keeping the language natural and smooth. Here, {language} represents the target language, {text} refers to the original English text, and {translation results} indicates the translated text. meticulously selected datasets to enhance model performance across a wide range of multimodal tasks. The datasets used in this phase are summarized in Table 1b. For image captioning, we included TextCaps [99] and bilingual ShareGPT4V [16], which help the model learn to generate descriptive captions in both English and Chinese. In the domain of general QA, datasets such as VQAv2 [28], GQA [34], and VisualDialog [22] teach the model to handle diverse question-answering scenarios. For scientific image understanding, datasets like AI2D [39], ScienceQA [73], and TQA [40] provide content-rich scenarios to enhance the model\u2019s ability to interpret scientific diagrams and texts. Chart interpretation is bolstered by ChartQA [81], MMC-Inst [61], and PlotQA [85], which train the model to analyze and understand chart images. Mathematics datasets such as GeoQA+ [12], TabMWP [74], and MathQA [132] introduce complex numerical and geometric problem-solving tasks. Knowledge-based QA benefits from the inclusion of datasets like KVQA [96] and bilingual Wikipedia [31], enabling the model to extract and reason with factual information across multiple languages. For tasks involving OCR, we utilize OCRVQA [86], TextVQA [100], and several datasets focused on Chinese and English text recognition, such as SynthDoG [41], to improve text recognition from images. Document understanding is advanced through datasets like DocVQA [82] and Common Crawl PDFs, which help the model for real-world document analysis. Visual grounding is trained using RefCOCO [79, 131] and Visual Genome [42], aiding the model 5 openOCR-related Benchmarks General Multimodal Benchmarks Math model source #param DocVQA ChartQA InfoVQA TextVQA OCRBench MME RWQA AI2D MMMU MMB-EN/CN CCB MMVet SEED HallB MathVista GPT-4V [1] \u2717 \u2212 88.4 78.5 \u2212 78.0 645 1926.6 61.4 78.2 56.8 77.0 / 74.4 46.5 67.6 71.6 46.5 49.9 Gemini Ultra 1.0 [107] \u2717 \u2212 90.9 80.8 80.3 82.3 \u2212 \u2212 \u2212 79.5 59.4 \u2212/ \u2212 \u2212 \u2212 \u2212 \u2212 53.0 Gemini Pro 1.0 [107] \u2717 \u2212 88.1 74.1 75.2 74.6 659 1933.4 \u2212 73.9 47.9 73.6 / 74.3 52.5 64.3 70.7 45.2 45.2 Gemini Pro 1.5 [92] \u2717 \u2212 86.5 81.3 72.7 73.5 \u2212 \u2212 67.5 80.3 58.5 \u2212/ \u2212 \u2212 \u2212 \u2212 \u2212 52.1 Qwen-VL-Max [5] \u2717 \u2212 93.1 79.8 73.4 \u2212 723 2433.6 \u2212 79.3 51.3 77.6 / 75.7 63.5 66.6 \u2212 41.2 51.0 Qwen-VL-Plus [5] \u2717 \u2212 91.4 78.1 \u2212 \u2212 694 2183.4 \u2212 75.9 45.2 67.0 / 70.7 55.1 61.1 72.7 40.6 43.3 Claude-3 Opus [3] \u2717 \u2212 89.3 80.8 \u2212 \u2212 694 1586.8 49.8 88.1 59.4 63.3 / 59.2 26.3 58.1 \u2212 37.8 50.5 Claude-3 Sonnet [3] \u2717 \u2212 89.5 81.1 \u2212 \u2212 646 1625.9 51.9 88.7 53.1 67.8 / 64.2 27.8 \u2212 \u2212 41.3 47.9 Claude-3 Haiku [3] \u2717 \u2212 88.8 81.7 \u2212 \u2212 658 1453.2 \u2212 86.7 50.2 60.7 / 57.2 24.5 \u2212 \u2212 39.2 46.4 HPT Pro [35] \u2717 \u2212 \u2212 \u2212 \u2212 \u2212 \u2212 \u2212 \u2212 \u2212 52.0 77.5 / 76.7 \u2212 \u2212 73.1 \u2212 \u2212 MM1 [84] \u2717 30B \u2212 \u2212 \u2212 73.5 \u2212 2069.0 \u2212 \u2212 44.7 75.1 / \u2212 \u2212 48.7 72.1 \u2212 39.4 Step-1V [102] \u2717 100B \u2212 \u2212 \u2212 \u2212 625 2206.4 \u2212 79.2 49.9 80.7 / 79.9 71.2 63.3 70.3 48.4 44.8 Grok-1.5V [125] \u2717 \u2212 85.6 76.1 \u2212 78.1 \u2212 \u2212 68.7 88.3 \u2212 \u2212/ \u2212 \u2212 \u2212 \u2212 \u2212 52.8 Text-Monkey [68] \u2713 10B 66.7 59.9 28.6 64.3 561 \u2212 \u2212 \u2212 \u2212 \u2212/ \u2212 \u2212 \u2212 \u2212 \u2212 \u2212 DocOwl-1.5 [33] \u2713 8B 82.2 70.2 50.7 68.6 599 \u2212 \u2212 \u2212 \u2212 \u2212/ \u2212 \u2212 \u2212 \u2212 \u2212 \u2212 Mini-Gemini [53] \u2713 35B \u2212 \u2212 \u2212 74.1* \u2212 2141.0 \u2212 \u2212 48.0 80.6 / \u2212 \u2212 59.3 \u2212 \u2212 43.3 LLaVA-NeXT [64] \u2713 35B 84.0 68.7 51.5 69.5* 574 2028.0 \u2212 74.9 51.1 81.1 / 79.0 49.2 57.4 75.9 34.8 46.5 InternVL 1.2 (ours) \u2713 40B 57.7 68.0 39.5 72.5* 569 2175.4 67.5 79.0 51.6 82.2 / 81.2 59.2 48.9 75.6 47.6 47.7 InternVL 1.5 (ours) \u2713 26B 90.9 83.8 72.5 80.6 724 2187.8 66.0 80.7 45.2 82.2 / 82.0 69.8 62.8 76.0 49.3 53.5 Table 2. Comparison with SoTA models on 16 multimodal benchmarks. OCR-related benchmarks include: DocVQA test [82], ChartQA test [81], InfographicVQA test [83], TextVQA val [100], and OCRBench [67]. General multimodal benchmarks encompass: MME [26], RealWorldQA [125], AI2D test [39], MMMU val [135], MMBench-EN/CN test [66], CCBench dev [66], MMVet [133], SEED Image [46], and HallusionBench [30]. Additionally, the math dataset includes MathVista testmini [75]. * denotes that Rosetta OCR tokens are used in the testing of TextVQA. The MME results we report are the sum of the perception and cognition scores. The results of OCRBench, MMBench, CCBench, and HallusionBench are collected from the OpenCompass leaderboard [21]. in precise object localization within images. In the realm of multimodal conversation, datasets like LLaVA-150K [63] and ALLaVA [14] enhance the model\u2019s dialogic capabilities by simulating interactive and engaging scenarios. Lastly, text-only datasets include OpenHermes2.5 [109], AlpacaGPT4 [106], among others [6, 141], which are used to maintain the original linguistic capabilities of the LLM. In summary, these datasets together establish a rich and diverse foundation for fine-tuning, which enhances our model\u2019s ability to handle a wide range of multimodal tasks and ensures its readiness for practical applications. Data Translation Pipeline. As shown in Figure 5, to enhance our model\u2019s multilingual capabilities, we implemented a data translation pipeline. This pipeline utilizes state-of-the-art open-source LLMs [4, 11, 130] or GPT-3.5 to convert English datasets to another language (e.g., Chinese), maintaining consistency and precision in bilingual labeling. Moreover, it can readily expand to encompass more languages by adjusting the language prompt, without relying on manual annotation processes. In Table 1, we have annotated the language for each dataset. For a dataset that was originally in English, an annotation as \u201czh\u201d indicates that we have translated it into Chinese using the translation pipeline. For example, COYO [10] and GRIT [90] were originally English datasets, and we have translated them into Chinese. By leveraging this translation pipeline, the Chinese capabilities of InternVL 1.5 have been greatly enhanced. 4. Experiments 4.1. Implementation Details. InternVL 1.5 was developed by integrating the InternViT6B [18] vision encoder with the InternLM2-20B [11] language model, using a dynamic high-resolution strategy. In this approach, images are segmented into 448\u00d7448 pixel tiles, with the number of tiles ranging up to 12 based on the image\u2019s aspect ratio and resolution during training. In testing phases, the model could handle up to 40 tiles, equivalent to 4K resolution, demonstrating its adaptability to highresolution inputs in a zero-shot manner. Notably, we built our model based on the chat version of InternLM2-20B rather than the base model. The training of InternVL 1.5 was divided into two stages. Initially, the pre-training stage focused on training the InternViT-6B vision encoder and the MLP projector to optimize visual feature extraction. Subsequently, the entire model\u2019s 26 billion parameters were fine-tuned to enhance multimodal capabilities. In both two stages of training, we use a context length of 4096 and adopt the same response formatting prompts as LLaVA 1.5 [52]. Additionally, the evaluation was mainly supported by VLMEvalKit [21]. 4.2. Comparison with State-of-the-Art MLLMs 4.2.1 Quantitative Results on 18 Benchmarks In this section, we conduct an extensive evaluation across a series of benchmarks to assess our model\u2019s multimodal un6 openConvBench (Pairwise Grading) ConvBench (Direct Grading) model source #param R1 R2 S1 S2 S3 SO R1 R2 S1 S2 S3 SO GPT-4V [1] \u2717 \u2212 39.51 38.47 38.47 39.34 37.61 40.55 7.09 7.30 7.30 7.48 7.12 6.88 Claude-3 Opus [3] \u2717 \u2212 36.60 37.49 38.99 39.17 34.32 35.70 6.54 6.75 6.53 7.04 6.68 6.32 Reka Flash [89] \u2717 \u2212 25.60 24.67 25.13 27.56 21.32 26.52 6.78 6.86 6.93 7.25 6.41 6.70 Gemini Pro 1.0 [107] \u2717 \u2212 8.44 8.55 9.01 9.36 7.28 8.32 4.42 4.60 5.18 4.95 3.66 4.24 ShareGPT4V-13B [16] \u2713 13B 17.56 17.45 17.85 18.72 15.77 17.68 4.85 5.03 5.16 5.06 4.86 4.67 LLaVA-1.5-13B [62] \u2713 13B 16.93 18.08 20.45 18.02 15.77 15.77 4.94 5.14 5.03 5.41 4.99 4.74 XComposer2 [23] \u2713 8B 15.83 16.41 17.16 19.06 13.00 15.25 5.82 5.98 5.98 6.17 5.78 5.66 mPLUG-Owl2 [128] \u2713 8B 14.93 15.83 17.50 17.16 12.82 14.04 5.04 5.17 4.98 5.38 5.14 4.91 Qwen-VL-Chat [5] \u2713 10B 14.33 14.62 16.29 18.37 9.19 14.04 5.54 5.65 5.96 5.78 5.22 5.43 MiniGPT-4 [142] \u2713 8B 10.95 10.80 11.61 11.27 9.53 11.09 3.85 4.04 3.99 4.40 3.73 3.66 LLaMA-A-V2 [27] \u2713 7B 9.04 9.59 8.84 10.92 9.01 8.49 4.77 4.91 4.77 5.47 4.48 4.64 InternVL 1.2 (ours) \u2713 40B 21.17 22.41 24.96 21.31 20.97 19.93 5.49 5.69 5.80 5.88 5.39 5.29 InternVL 1.5 (ours) \u2713 26B 17.65 20.22 26.00 17.33 17.33 15.08 5.60 5.76 6.11 5.93 5.25 5.43 openMMT-Bench model source #param Overall Overall* GPT-4V [1] \u2717 \u2212 62.0 55.5 Qwen-VL-Plus [4] \u2717 \u2212 62.3 56.6 Gemini Pro 1.0 [107] \u2717 \u2212 61.6 55.1 Claude-3 Haiku [3] \u2717 \u2212 52.2 46.4 LLaVA-NeXT [64] \u2713 35B 60.8 56.3 XComposer2 [23] \u2713 8B 55.7 50.0 BLIP-2-XXL [50] \u2713 12B 54.8 49.1 Yi-VL-34B [130] \u2713 35B 54.2 48.6 Monkey-Chat [107] \u2713 10B 53.4 46.0 DeepSeek-VL [71] \u2713 7B 53.2 46.5 CogVLM-Chat [117] \u2713 17B 51.6 44.2 InternVL 1.2 (ours) \u2713 40B 63.4 58.2 InternVL 1.5 (ours) \u2713 26B 59.0 56.2 Table 3. Comparison with SoTA models on ConvBench and MMT-Bench. ConvBench [65] is a multi-turn conversation evaluation benchmark designed for MLLMs. The table presents win rates against humans, where S1, S2, and S3 represent the scores for perception, reasoning, and creation, respectively. R2 is calculated as (S1 + S2 + S3)/3, reflecting the average performance across three turns. R1 is derived from (R2 + S0)/2, indicating the model\u2019s overall score. MMT-Bench [129] is a comprehensive benchmark designed to assess MLLMs across massive multimodal tasks requiring expert knowledge and deliberate visual recognition, localization, reasoning, and planning. The overall score is computed across 162 subtasks, excluding visual recognition as denoted by *. derstanding and reasoning capability. The benchmarks employed in our study are categorized into four distinct types: OCR-related, general multimodal, mathematical, and multiturn conversation benchmarks. As depicted in Table 2, InternVL 1.5 exhibits leading performance across the majority of these benchmarks. OCR-related Image Understanding. We evaluate the model performance across four key dimensions of OCR: document comprehension (DocVQA [82]), chart understanding (ChartQA [81]), infographic understanding (InfographicVQA [83]), and scene text interpretation (TextVQA [100]). Additionally, we employ OCRBench [67] to perform a comprehensive evaluation of the model\u2019s overall OCR capabilities. As shown in Table 2, our model demonstrated comparable performance to proprietary models on these benchmarks and significantly outperformed the opensource LLaVA-NeXT [64] as well as InternVL 1.2, the predecessor of InternVL 1.5. Notably, our model achieves state-of-the-art performance on ChartQA and OCRBench, outperforming all competing proprietary models. General Multimodal Evaluation. In addition to OCRrelated benchmarks, we tested our model on several general multi-modal benchmarks. We used RealWorldQA [125] to evaluate the model\u2019s real-world spatial understanding capabilities. HallusionBench [30] was employed to assess its ability to control hallucinations. Additionally, MMMU [135] was utilized to evaluate the model\u2019s multidisciplinary capabilities, and AI2D [39] to assess its understanding of science diagrams. We also tested the model\u2019s proficiency in Chinese and understanding of Chinese culture with the MMBench-CN test [66] and CCBench [66], respectively. Other comprehensive benchmarks such as MME [26], MMBench-EN [66], MMVet [133], SEED [46], and MMT-Bench [129] were also used to assess the model\u2019s visual understanding and reasoning abilities. Compared to other open-source models like TextMonkey [68], DocOwl-1.5 [33], and LLaVA-NeXT [64], our InternVL 1.5 significantly closes the gap with proprietary models in these benchmarks. Specifically, our model achieves the best performance on HallusionBench [30], demonstrating its outstanding ability to reduce hallucinations. Moreover, thanks to our high-quality bilingual dataset, our model exhibits robust Chinese language capabilities, significantly surpassing both open-source and proprietary methods on MMBench-CN and CCBench. However, while InternVL 1.5 surpasses MM1 [84] and is comparable to Gemini Pro 1.0 [107] on MMMU, it shows a slight decline from its predecessor, InternVL 1.2. We attribute this modest decrement to the smaller size of the language model, a phenomenon similarly observed in the MMT-Bench [129] results, as shown in Table 3. Math Reasoning. MathVista [75] is a benchmark designed to integrate challenges from various mathematical and visual tasks. Completing these tasks requires a deep understanding of visuals, logical thinking, and math knowledge\u2014areas where many proprietary commercial models encounter significant difficulties. As shown in Table 2, our model outperforms others, including GPT-4V [87], by a clear margin in this benchmark, showcasing its ability to handle mathematically demanding tasks. Multi-Turn Conversation. Compared to single-turn dialogues, multi-turn conversations align more with human preferences. In practical usage, multi-turn dialogue is the preferred mode for general-purpose assistants to engage with humans in solving a variety of tasks. Therefore, we opt to utilize ConvBench [65] for evaluating multi-turn conversations, which progressively assesses the perception, reasoning, and creativity capabilities of MLLMs. As depicted 7 0 10 20 30 40 50 60 70 80 90 100 90.5 18 DocVQA val 0 10 20 30 40 70 75 80 85 83.8 12 ChartQA test 0 10 20 30 40 30 40 50 60 70 80 72.3 24 InfoVQA val 0 10 20 30 40 70 75 80 80.6 24 T extVQA val 0 10 20 30 40 55 60 65 70 75 72.4 24 OCRBench 0 10 20 30 40 72.00 72.25 72.50 72.75 73.00 73.25 73.1 36 MME 0 10 20 30 40 62 64 66 66.1 40 RealWorldQA 0 10 20 30 40 79.8 80.0 80.2 80.4 80.6 80.8 80.7 6 AI2D test 0 5 10 15 44.25 44.50 44.75 45.00 45.25 45.2 6 MMMU val 0 10 20 30 40 81.6 81.8 82.0 82.2 82.2 6 MMBench-EN test 0 10 20 30 40 81.2 81.4 81.6 81.8 82.0 82.0 6 MMBench-CN test 0 10 20 30 40 69.0 69.5 70.0 70.5 70.2 18 CCBench dev 0 10 20 30 40 58 60 62 64 62.8 24 MMVet 0 10 20 30 40 74.5 75.0 75.5 76.0 76.5 76.1 36 SEED Image 0 10 20 30 40 47.5 48.0 48.5 49.0 49.5 50.0 49.6 4 HallusionBench 0 10 20 30 40 65.0 67.5 70.0 72.5 75.0 74.3 24 Average (w/o MMMU val) Figure 6. Comparison of InternVL 1.5 performance across different image resolutions. The X-axis represents the number of tiles, while the Y-axis indicates benchmark performance. The highest value and its corresponding number of tiles are highlighted. The scores of MME [26] and OCRBench [67] have been normalized to a maximum score of 100. We found that although only 1 to 12 tiles were used during training, it is possible to zero-shot scale up to 40 tiles (i.e., 4K resolution) during testing. Note that since MMMU [135] includes multiple images per sample, it may run out of memory when the number of tiles is large. Therefore, we only tested up to 18 tiles maximum, and MMMU was not included when calculating the average score. in Table 3, InternVL exhibits leading performance among open-source models, albeit still trailing behind GPT-4V by a considerable margin. Going forward, we will continue refining InternVL\u2019s capabilities in multi-turn conversations. 4.3. Ablation Study Larger LLMs need Larger VFMs. In this study, we investigate the interplay between LLMs and VFMs. The comparison involves two open-source MLLMs, LLaVA-NeXT [64] and InternVL 1.2, each equipped with LLMs of 34 billion parameters. Notably, although both models employ LLMs of the same scale, InternVL 1.2 incorporates a significantly larger VFM, with 6 billion parameters, compared to LLaVA-NeXT\u2019s 300 million parameters. Since the data for LLaVA-NeXT is not available, we created a similar dataset ourselves. Additionally, InternVL 1.2 was trained at a fixed resolution of 448\u00d7448, while LLaVA-NeXT used a higher dynamic resolution of 672 \u00d7 672. Therefore, this comparison is not entirely fair or equivalent. Nevertheless, the findings still reveal noteworthy insights. For example, after excluding five OCR-related datasets, ConvBench, and RealWorldQA, InternVL 1.2 outperformed LLaVA-NeXT in 9 out of the remaining 11 datasets. This performance difference supports our hypothesis that for a large-scale LLM (e.g., 34B), a larger VFM (e.g., 6B) can effectively improve the model\u2019s ability to handle complex multimodal tasks, thereby enhancing the overall performance. Dynamic Resolution Matters. As shown in Figure 6, we investigated the effectiveness of dynamic resolution across various multimodal benchmarks. We found that not all tasks require high resolution. Specifically, tasks related to OCR, such as DocVQA, InfoVQA, TextVQA, and OCRBench, benefit from increased resolution. However, tasks like AI2D, MMMU, MMBench, and HallusionBench ex8 User: What action is taking place in the picture? Which color jersey is the attacking player wearing? Explain in detail. GPT-4V: In the image, we see a basketball game in progress where a player in a black jersey is in the act of shooting the ball. The player is extending their arms upwards towards the basket, and it looks like they\u2019re attempting a jump shot. The player is wearing a black jersey with the word \u201cSPURS\u201d across the front, which indicates they are a member of the San Antonio Spurs team. There\u2019s a defensive player from the opposing team, wearing a red jersey with the number \u201c9\u201d and the name \u201cVUCEVIC\u201d on it, who is trying to contest the shot. InternVL 1.5: The image depicts a basketball game in progress. A player in a black jersey, who is likely on the offensive team, is in the act of shooting the basketball. He is airborne, with one arm extended upwards towards the basket, indicating that he is in the process of taking a shot. The player in the red jersey is on defense, with his arm raised in an attempt to block the shot. The background shows the interior of an arena with spectators and a scoreboard, suggesting that this is a professional or organized basketball match. The attacking player\u2019s jersey is black, and the defending player\u2019s jersey is red. User: Tell me a person who had made a similar pose to the cute duck. GPT-4V: I can\u2019t identify or make assumptions about a specific person who has made a similar pose to the duck in the image you\u2019ve uploaded, but the stance the duck is taking, with what appears to be a martial arts belt and a weapon, is reminiscent of a martial artist demonstrating a pose or ready stance. This is a common pose in many martial arts, as practitioners often pose for photos demonstrating their skills or in the midst of performing kata (a sequence of movements in martial arts). If you\u2019re looking for something specific, you might find martial artists or actors who play martial arts roles striking similar poses. InternVL 1.5: The pose of the cute duck is reminiscent of a karate stance, which is often associated with martial artists. One famous person who made a similar pose is Bruce Lee, a martial artist and actor known for his contributions to martial arts and film. Figure 7. Examples for General QA. Blue highlights the standout answers, while red highlights incorrect responses. hibit a slight decline in performance at higher resolutions. Overall, InternVL 1.5 demonstrates strong robustness to dynamic resolution. It can adjust the resolution based on the specific requirements of each task, ensuring optimal performance where high resolution is beneficial and conserving resources where it is not. 4.3.1 Qualitative Results on Different Scenes In previous sections, we evaluated our model across various benchmarks and observed its strong performance. In this section, we conduct a qualitative comparison of our model with GPT-4V [87] across diverse scenarios, including General QA, OCR-related QA, Scientific Understanding, Chinese Traditional Culture, Object Localization, and MultiImage Dialogue. We aim to demonstrate the practicality and versatility of our model in real-world applications, offering insights from the perspective of actual user experience. General QA. To compare the general capabilities of InternVL 1.5 and GPT-4V, we first conducted an experiment involving simple user queries with images requiring general knowledge. As shown on the left side of Figure 7, both two models respond accurately to the query, showcasing their proficiency in general topics. As shown on the right side of Figure 7, GPT-4V may excessively refuse to answer some questions due to its involvement in personal privacy. OCR-Related QA. We conducted an evaluation to compare the OCR capabilities of our InternVL 1.5 model against GPT-4V. On the left side of Figure 8, the first prompt aimed to measure the models\u2019 ability to understand Chinese scenes. In this instance, GPT-4V cannot extract all useful information in the image. On the right side of Figure 8, both GPT-4V and our model have good performance on chart understanding. Scientific Understanding. Evaluating the capabilities of models in scientific understanding reasoning tasks is essential for advancing computational intelligence, particularly in contexts requiring in-domain knowledge and logical reasoning. In our study, we compared the performance of our InternVL 1.5 model with GPT-4V by administering complex multi-disciplinary problems designed to assess the accuracy of their reasoning. In Figure 9, for the first question, both models accurately answered and provided an analysis from an aerodynamic perspective. For the second question, our model precisely analyzed the elements depicted in the image and provided the correct response, whereas GPT-4V speculated on the trend of amino acid transport. These results suggest that our method and GPT-4V exhibit comparable capabilities in scientific understanding and reasoning. Chinese Traditional Culture. We selected two typical multimodal examples related to traditional Chinese art to evaluate our model. As illustrated in Figure 10, both InternVL 1.5 and GPT-4V correctly recognize the Chinese traditional culture depicted in the image. Notably, InternVL 1.5 demonstrates a deeper understanding of this culture, as 9 User: Describe this image in detail. GPT-4V: The image shows a signboard with Chinese characters on it. From the text, we can deduce that the sign is likely for a business. The sign features a design that includes a martini glass with a flamingo, indicating that the establishment might serve drinks. There\u2019s additional text on the right side, providing specific details such as business hours. It appears to be a nighttime business, likely a bar or restaurant, operating from 16:30 to 22:30. The other text provides further information about the services or specialties of the place. Unfortunately, I can\u2019t provide a translation of the text, as my capabilities don\u2019t include translating text within images. InternVL 1.5: The image shows a signboard with Chinese text, indicating a business establishment. The signboard reads: \u201c\u6cb3\u5927\u897f\u95e8\u7b2c\u4e00\u5bb6\u201d which translates to \u201cThe first shop at the west gate of the river.\u201d \u201c\u6b63\u5b97\u8001\u5b57\u53f7\u4e94\u9999\u5473\u201d which means \u201cGenuine, old brand, five-spice flavor.\u201d \u201c\u66f9\u8bb0\u7116\u70e4\u9e21\u817f\u201d which translates to \u201cCao\u2019s roasted and baked chicken legs.\u201d \u201c\u8425\u4e1a\u65f6\u95f416:30-22:30\u201d indicating the business hours from 4:30 PM to 10:30 PM. \u201c7\u70b9\u4e4b\u540e\u5728\u95e8\u53e3\u8def\u706f\u4e0b\u51fa\u591c\u5e02\u5730\u644a\u201d which suggests that after 7 PM, the shop operates in the doorway under the streetlights as a night market stall. The signboard also includes a logo that appears to be a stylized rooster, which might be the brand or shop\u2019s emblem. The overall style of the signboard is quite traditional and straightforward, commonly seen in Chinese street markets or small businesses. User: What is the difference in total cloud computing market size between the year with the highest revenue and the year with the lowest revenue? GPT-4V: To find the difference in total cloud computing market size between the year with the highest revenue and the year with the lowest revenue, we need to: 1. Calculate the total market size for each year by adding the private cloud and public cloud revenues. 2. Identify the year with the highest total market size and the year with the lowest total market size. 3. Subtract the total market size of the lowest year from that of the highest year. Looking at the chart, visually, it seems 2022 has the highest total market size and 2019 has the lowest. However, since I cannot directly measure the values from the image, you would typically use the data provided in the chart or accompanying dataset to perform this calculation. Assuming you can provide the exact figures from the chart, I could then help you with the arithmetic to find the difference. InternVL 1.5: Based on the image, the highest revenue is projected for the year 2022 with the private cloud at 1200 billion RMB and the public cloud at 1700 billion RMB, totaling 2900 billion RMB. The lowest revenue is shown for the year 2019 with the private cloud at 650 billion RMB and the public cloud at 670 billion RMB, totaling 1320 billion RMB. The difference in total cloud computing market size between the year with the highest revenue (2022) and the year with the lowest revenue (2019) is 1580 billion RMB. Figure 8. Examples for OCR-related QA. Blue highlights the standout answers, while red highlights incorrect responses. evidenced by its more detailed descriptions of the cultural elements in its response. Object Localization. Evaluating machine learning models for their proficiency in object localization tasks is essential, especially in applications requiring precise spatial awareness. In our comparative analysis, the performance of the InternVL 1.5 model was juxtaposed with GPT-4V, focusing on their ability to accurately detect and localize objects within various environments. Our assessments ranged from simple object recognition in cluttered scenes to complex scenarios involving dynamic interactions among multiple entities. As illustrated in Figure 11, the results demonstrate that InternVL 1.5 not only localized objects with high accuracy but also exhibited a comparable understanding of spatial relationships, matching the performance of GPT-4V. Multi-Image Dialogue. As shown in Figure 12, in this experiment, we ask InternVL 1.5 and GPT-4V to compare the similarities and differences between the two images. As can be seen, both GPT-4V and InternVL 1.5 provide detailed and accurate responses. Through this experiment, we discovered that although InternVL 1.5 was trained solely on single-image inputs, it exhibits strong zero-shot capabilities for multi-image dialogues. 5. Conclusion This work introduced InternVL 1.5, an open-source MLLM designed to narrow the performance gap between opensource and proprietary models in multimodal understanding. By integrating a strong vision encoder with continuous learning capabilities, adopting a dynamic high-resolution strategy, and utilizing a high-quality bilingual dataset, InternVL 1.5 has demonstrated robust performance across a variety of benchmarks. Our evaluations indicate that the model achieves competitive performance with leading proprietary models, excelling particularly in OCR-related tasks and showing significant improvements in Chinese-related scene understanding. While InternVL 1.5 has contributed to the open-source multimodal understanding, the field continues to evolve with many challenges ahead. We aspire to further enhance InternVL\u2019s capabilities and invite collaboration with the global research community, hoping to enrich and expand the reach of open-source models together. 10 User: Which car in the picture is more aerodynamic? GPT-4V: Based on the image you\u2019ve provided, determining which car is more aerodynamic involves analyzing the shape and design of each vehicle. The car on the right appears to have a more streamlined shape, with a sloped front that would likely reduce air resistance better than the boxy design of the car on the left. Therefore, the car on the right would be considered more aerodynamic. InternVL 1.5: The car on the right appears to be more aerodynamic. This is inferred from the shape of the car, which has a more streamlined and elongated form. The design includes a pointed front end and a tapered rear, which are common features in aerodynamic design to reduce air resistance. The car on the left, with its boxy shape, does not exhibit the same level of aerodynamic efficiency. User: If the pH outside the cell decreases, would you expect the amount of amino acids transported into the cell to increase or decrease? GPT-4V: The image depicts a sodium-glucose symporter, which is a type of secondary active transport mechanism. This transport mechanism is specifically for glucose and does not directly provide information about amino acid transporters. However, speaking generally about amino acid transporters, if the pH outside the cell decreases, the effect on amino acid transport would depend on the type of amino acid transporter involved. Some amino acid transporters are proton-dependent, meaning that they use the proton gradient (which increases as pH decreases) to transport amino acids into the cell. For these transporters, a decrease in extracellular pH could potentially increase the transport of amino acids into the cell. To give a precise answer, one would need specific information about the type of amino acid transporter present in the cell membrane. Since this information is not provided in the image or the description, the impact of a decrease in extracellular pH on amino acid transport cannot be determined from the provided content alone. InternVL 1.5: Based on the image, if the pH outside the cell decreases, we would expect the amount of amino acids transported into the cell to increase. This is because the diagram shows a gradient of amino acids moving from an area of lower concentration (outside the cell) to an area of higher concentration (inside the cell), which is a typical representation of active transport. Active transport is a process that moves substances against their concentration gradient, often using energy, and it is typically more efficient at higher concentrations of the substance being transported. Therefore, a decrease in pH outside the cell would likely increase the concentration gradient, leading to an increase in the transport of amino acids into the cell. Figure 9. Examples for Scientific Understanding. Blue highlights the standout answers. User: \u56fe\u4e2d\u6240\u793a\u7684\u4f20\u7edf\u8868\u6f14\u5f62\u5f0f\u4e3a? GPT-4V: \u8fd9\u5f20\u56fe\u7247\u5c55\u793a\u7684\u662f\u4e2d\u56fd\u7684\u4f20\u7edf\u6728\u5076\u620f\uff0c\u4e5f\u88ab\u79f0\u4e3a\u201c\u63d0\u7ebf \u6728\u5076\u201d\u3002\u8fd9\u79cd\u8868\u6f14\u827a\u672f\u5f62\u5f0f\u5728\u4e2d\u56fd\u6709\u7740\u60a0\u4e45\u7684\u5386\u53f2\uff0c\u901a\u5e38\u4ee5\u7cbe\u7f8e \u7684\u6728\u5076\u548c\u590d\u6742\u7684\u64cd\u4f5c\u6280\u5de7\u8457\u79f0\u3002\u6728\u5076\u620f\u5728\u4e2d\u56fd\u7684\u5404\u4e2a\u5730\u533a\u6709\u4e0d\u540c \u7684\u98ce\u683c\u548c\u8868\u6f14\u7279\u8272\u3002\u56fe\u7247\u4e2d\u7684\u6728\u5076\u670d\u9970\u7eda\u4e3d\uff0c\u8868\u6f14\u8005\u901a\u8fc7\u7ebf\u63a7\u5236 \u6728\u5076\u7684\u52a8\u4f5c\uff0c\u8fdb\u884c\u6545\u4e8b\u53d9\u8ff0\u548c\u8868\u6f14\u3002 InternVL 1.5: \u56fe\u4e2d\u6240\u793a\u7684\u4f20\u7edf\u8868\u6f14\u5f62\u5f0f\u4e3a\u5e03\u888b\u620f\uff0c\u5c5e\u4e8e\u6728\u5076\u620f \u7684\u4e00\u79cd\uff0c\u5e03\u888b\u620f\u662f\u4e00\u79cd\u53e4\u8001\u7684\u6c49\u65cf\u6c11\u95f4\u620f\u66f2\u8868\u6f14\u827a\u672f\u5f62\u5f0f\uff0c\u5e03\u888b \u620f\uff08\u53c8\u79f0\u5e03\u888b\u6728\u5076\u620f\u3001\u624b\u64cd\u5080\u5121\u620f\u3001\u638c\u4e2d\u620f\u3001\u5c0f\u7b3c\u3001\u6307\u82b1\u620f\uff09\u662f \u4e00\u79cd\u6e90\u4e8e17\u4e16\u7eaa\u4e2d\u56fd\u798f\u5efa\u6cc9\u5dde\uff0c\u6d41\u884c\u4e8e\u53f0\u6e7e\u3001\u798f\u5efa\u3001\u5e7f\u4e1c\u7b49\u5730\u7684 \u4f20\u7edf\u620f\u5267\u827a\u672f\u3002\u8868\u6f14\u8005\u7528\u53cc\u624b\u64cd\u7eb5\u5e03\u888b\u5185\u7684\u6728\u5076\uff0c\u901a\u8fc7\u624b\u6307\u7684\u52a8 \u4f5c\u6765\u63a7\u5236\u6728\u5076\u7684\u52a8\u4f5c\u548c\u8868\u60c5\uff0c\u540c\u65f6\u914d\u5408\u97f3\u4e50\u3001\u5531\u8154\u548c\u9053\u767d\uff0c\u6765\u8bb2 \u8ff0\u6545\u4e8b\u3002\u5728\u4e2d\u56fd\u4ee5\u53ca\u4e16\u754c\u534e\u4eba\u5730\u533a\u90fd\u6709\u6d41\u4f20\u3002 User: \u56fe\u4e2d\u662f\u8c01\u7684\u4f5c\u54c1? Options: A. \u5f90\u60b2\u9e3fB. \u9f50\u767d\u77f3C. \u5f20\u5927\u5343 D. \u5218\u6d77\u7c9f GPT-4V: \u8fd9\u5e45\u753b\u7684\u98ce\u683c\u4f3c\u4e4e\u662f\u5f90\u60b2\u9e3f\u7684\uff0c\u4ed6\u4ee5\u8868\u73b0\u9a6c\u7684\u9ad8\u8d85\u6280\u5de7 \u800c\u95fb\u540d\u3002\u5f90\u60b2\u9e3f\u7684\u9a6c\u753b\u7279\u70b9\u662f\u7ebf\u6761\u6d41\u7545\u800c\u6709\u529b\uff0c\u4f53\u73b0\u4e86\u9a6c\u7684\u52a8\u6001 \u7f8e\u548c\u751f\u547d\u529b\u3002\u8fd9\u5e45\u753b\u4e2d\u9a6c\u7684\u8868\u73b0\u4e0e\u5f90\u60b2\u9e3f\u7684\u98ce\u683c\u76f8\u543b\u5408\uff0c\u6240\u4ee5\u6b63 \u786e\u7b54\u6848\u662fA. \u5f90\u60b2\u9e3f\u3002 InternVL 1.5: \u8fd9\u5e45\u753b\u770b\u8d77\u6765\u50cf\u662f\u4e00\u5e45\u4e2d\u56fd\u753b\uff0c\u7279\u522b\u662f\u9a6c\u7684\u63cf\u7ed8 \u65b9\u5f0f\u548c\u98ce\u683c\u3002\u5f90\u60b2\u9e3f\u662f\u4e2d\u56fd\u8457\u540d\u7684\u753b\u5bb6\uff0c\u4ee5\u753b\u9a6c\u800c\u95fb\u540d\uff0c\u4ed6\u7684\u753b \u4f5c\u901a\u5e38\u5177\u6709\u5f88\u5f3a\u7684\u8868\u73b0\u529b\u548c\u52a8\u611f\u3002\u4ece\u56fe\u4e2d\u9a6c\u7684\u59ff\u6001\u548c\u8868\u73b0\u624b\u6cd5\u6765 \u770b\uff0c\u4e0e\u5f90\u60b2\u9e3f\u7684\u98ce\u683c\u975e\u5e38\u76f8\u4f3c\u3002\u5f90\u60b2\u9e3f\u7684\u4f5c\u54c1\u901a\u5e38\u4ee5\u5927\u5199\u610f\u7684\u65b9 \u5f0f\u5448\u73b0\uff0c\u6ce8\u91cd\u8868\u73b0\u7269\u8c61\u7684\u7cbe\u795e\u548c\u6c14\u8d28\u3002\u56fe\u4e2d\u9a6c\u7684\u795e\u6001\u548c\u52a8\u6001\u6355\u6349 \u5f97\u975e\u5e38\u751f\u52a8\uff0c\u4f53\u73b0\u4e86\u5f90\u60b2\u9e3f\u5bf9\u9a6c\u5339\u7684\u6df1\u523b\u7406\u89e3\u548c\u7cbe\u6e5b\u7684\u7ed8\u753b\u6280 \u5de7\u3002\u56e0\u6b64\uff0c\u6839\u636e\u63d0\u4f9b\u7684\u9009\u9879\uff0c\u56fe\u4e2d\u4f5c\u54c1\u5f88\u53ef\u80fd\u662f\u5f90\u60b2\u9e3f\u7684\u4f5c\u54c1\u3002 Figure 10. Examples for Chinese Traditional Culture. Blue highlights the standout answers. 11 User: Tell me the locations of all the number 10 jerseys in the picture. GPT-4V: In the image, there are two players wearing number 10 jerseys. The first one is located towards the center, slightly to the right while facing the image; he\u2019s wearing a red jersey and is the second person from the right in the group wearing red and white uniforms. The second player is wearing a light blue jersey and is positioned on the far left side of the image; he\u2019s the third person from the left in the group wearing blue and white uniforms. InternVL 1.5: In the image, there are two number 10 jerseys visible: 1. The first player wearing the number 10 jersey is located in the center of the image. He is wearing a red jersey with the number 10 on it, and he is standing with his hands on his hips. 2. The second player with the number 10 jersey is on the far right of the image. He is wearing a blue jersey with the number 10 on it and is crouching down with his hands on his knees. User: What is the man in the picture doing with his right hand? GPT-4V: The man in the picture is pointing with his right hand towards a beluga whale that appears to be behind a glass barrier. He seems to be doing this to show the whale to the child he is holding, likely to engage the child\u2019s interest or to teach the child about the whale. InternVL 1.5: The man in the picture is pointing at a large dolphin through a glass barrier. He is likely making this gesture to draw attention to the dolphin, either for the child he is holding or for others around them. This could be to educate the child about the dolphin, to highlight a specific behavior or feature of the dolphin, or simply to share the experience with others. The gesture is a common way to direct someone\u2019s attention to something of interest. Figure 11. Examples for Object Localization. Blue highlights the standout answers, while red highlights incorrect responses. User: What are the similarities and differences between these two pictures? GPT-4V: The similarities between the two images are that they both feature red pandas, a species known for their reddish-brown fur, bushy tails, and adorable faces. In both images, the red pandas are resting on tree branches with green foliage in the background, highlighting their arboreal (tree-dwelling) nature. As for the differences, the first image shows a red panda in a curled-up position, appearing to be asleep with its tail wrapped around its body for warmth. This posture also serves as camouflage, making the panda resemble a tree branch or part of the foliage. The angle of the shot is from above the animal, and the focus is on the texture and patterns of the fur. In contrast, the second image captures the red panda in a more alert and awake state, with its head raised and looking toward the camera. This posture allows a clear view of the panda\u2019s facial features, such as its ears, eyes, and the white markings. The angle of the second image is more direct and at the same level as the animal, which gives a better sense of the panda\u2019s overall appearance and stance. InternVL 1.5: The two pictures both feature red pandas, which are small, arboreal mammals native to the temperate forests of the Himalayas, southwestern China, and the northern part of Myanmar. They are known for their distinctive reddish-brown fur, white markings around their eyes, and bushy tails. In the first picture, the red panda is curled up and appears to be sleeping or resting on a branch. The background is blurred with green foliage, suggesting a natural, forested habitat. The focus is on the red panda\u2019s relaxed posture and the texture of its fur. In the second picture, the red panda is sitting upright on a branch, with its front paws resting on the branch. The animal is looking directly at the camera, giving a more alert and curious expression. The background is also green and blurred, indicating a similar natural setting. The similarities between the two pictures include the presence of the red panda, the branch it is on, and the green, forested background. The differences lie in the red panda\u2019s posture and expression, with the first picture showing a more relaxed and the second a more alert state. Additionally, the first picture has a more focused view of the red panda\u2019s fur and form, while the second picture captures the animal\u2019s face and eyes more clearly. Figure 12. Examples for Multi-Image Dialogue. Blue highlights the standout answers. 12"
}