GEO
赞助商内容

RAG-Anything是什么?如何实现多模态文档智能问答?

2026/4/24
RAG-Anything是什么?如何实现多模态文档智能问答?

AI Summary (BLUF)

RAG-Anything is an open-source multimodal RAG framework developed by Professor Huang Chao's team at the University of Hong Kong. It builds a unified multimodal knowledge graph architecture to process

RAG-Anything: A Unified Multimodal RAG Framework for Comprehensive Document Understanding

Introduction

【导读】最近,由香港大学黄超教授团队发布的开源项目「一体化的多模态RAG框架」RAG-Anything,有效解决了传统RAG的技术局限,实现了「万物皆可RAG」的处理能力。

Recently, the open-source project RAG-Anything, released by Professor Chao Huang's team at the University of Hong Kong, effectively addresses the technical limitations of traditional RAG systems, achieving a "RAG for Everything" processing capability.

RAG-Anything的核心技术创新在于构建了统一的多模态知识图谱架构,能够同时处理并关联文档中的文字内容、图表信息、表格数据、数学公式等多种类型的异构内容,解决了传统RAG系统仅支持文本处理的技术限制,为多模态文档的智能理解提供了新的技术方案。

The core technical innovation of RAG-Anything lies in its unified multimodal knowledge graph architecture, which can simultaneously process and correlate heterogeneous content types within documents, including text, charts, tables, and mathematical formulas. This solves the technical limitation of traditional RAG systems that only support text processing, providing a new technical solution for the intelligent understanding of multimodal documents.

Project Repository: https://github.com/HKUDS/RAG-Anything

Lab Homepage: https://sites.google.com/view/chaoh

RAG-Anything作为一个专为多模态文档设计的检索增强生成(RAG)系统,专注解决复杂场景下的智能问答与信息检索难题。

As a Retrieval-Augmented Generation (RAG) system specifically designed for multimodal documents, RAG-Anything focuses on solving the challenges of intelligent question-answering and information retrieval in complex scenarios.

该系统提供完整的端到端多模态文档处理解决方案,能够统一处理文本、图像、表格、数学公式等多种异构内容,实现从文档解析、知识图谱构建到智能问答的全流程自动化,为下一代AI应用提供了可靠的技术基础。

The system provides a complete end-to-end multimodal document processing solution, capable of uniformly processing various heterogeneous content types such as text, images, tables, and mathematical formulas. It achieves full-process automation from document parsing and knowledge graph construction to intelligent question-answering, providing a reliable technical foundation for next-generation AI applications.

该项目在开源框架LightRAG的基础上进行了深度扩展与优化,其多模态处理能力现已独立演进为RAG-Anything,并将基于此平台持续迭代更新。

This project has been deeply extended and optimized based on the open-source framework LightRAG. Its multimodal processing capabilities have now independently evolved into RAG-Anything, which will continue to be iteratively updated on this platform.


Background and Technical Drivers

The Era of Multimodal Understanding

随着人工智能技术的快速发展和大型语言模型能力的显著提升,用户对AI系统的期望已经从单纯的文本处理扩展到对真实世界复杂信息的全面理解。

With the rapid development of AI technology and the significant improvement in the capabilities of large language models, user expectations for AI systems have expanded from simple text processing to a comprehensive understanding of complex real-world information.

现代知识工作者每天面对的文档不再是简单的纯文本,而是包含丰富视觉元素、结构化数据和多媒体内容的复合型信息载体。

The documents that modern knowledge workers face daily are no longer simple plain text but are composite information carriers containing rich visual elements, structured data, and multimedia content.

这些文档中往往蕴含着文字描述、图表分析、数据统计、公式推导等多种信息形态,彼此相互补充、共同构成完整的知识体系。

These documents often contain multiple information forms, such as textual descriptions, chart analysis, data statistics, and formula derivation, which complement each other to form a complete knowledge system.

在专业领域的实际应用中,多模态内容已成为知识传递的主要载体。科研论文中的实验图表和数学公式承载着核心发现,教育材料通过图解和示意图增强理解效果,金融报告依赖统计图表展示数据趋势,医疗文档则包含大量影像资料和检验数据。

In practical applications across professional fields, multimodal content has become the primary carrier of knowledge transfer. Experimental charts and mathematical formulas in research papers carry core findings, educational materials enhance understanding through diagrams and schematics, financial reports rely on statistical charts to display data trends, and medical documents contain large amounts of imaging data and test results.

面对如此复杂的信息形态,传统的单一文本处理方式已无法满足现代应用需求。各行业都迫切需要AI系统具备跨模态的综合理解能力,能够同时解析文字叙述、图像信息、表格数据和数学表达式,并建立它们之间的语义关联,从而为用户提供准确、全面的智能分析和问答服务。

Faced with such complex information forms, traditional single-mode text processing can no longer meet modern application requirements. Various industries urgently need AI systems with cross-modal comprehensive understanding capabilities, able to simultaneously parse textual narratives, image information, tabular data, and mathematical expressions, and establish semantic relationships between them, thereby providing users with accurate and comprehensive intelligent analysis and question-answering services.

Technical Bottlenecks of Traditional RAG Systems

尽管检索增强生成(RAG)技术在文本问答领域取得了显著成功,但现有的RAG系统普遍存在明显的模态局限性。

Although Retrieval-Augmented Generation (RAG) technology has achieved remarkable success in text-based question-answering, existing RAG systems generally suffer from significant modality limitations.

传统RAG架构主要针对纯文本内容设计,其核心组件包括文本分块、向量化编码、相似性检索等,这些技术栈在处理非文本内容时面临严重挑战:

Traditional RAG architectures are primarily designed for plain text content, with core components including text chunking, vector encoding, and similarity retrieval. These technology stacks face serious challenges when processing non-text content:

  • Content Understanding Limitations: Traditional systems typically use OCR technology to forcibly convert images and tables into text, but this approach loses important information such as visual layout, color coding, and spatial relationships, leading to a significant degradation in understanding quality.

    内容理解局限: 传统系统通常采用OCR技术将图像和表格强制转换为文本,但这种方式会丢失视觉布局、颜色编码、空间关系等重要信息,导致理解质量大幅下降。

  • Insufficient Retrieval Accuracy: Plain text vectors cannot effectively represent the visual semantics of charts, the structural relationships of tables, or the mathematical meaning of formulas. When faced with questions like "What is the trend in this chart?" or "Which metric is the highest in this table?", retrieval accuracy is severely inadequate.

    检索精度不足: 纯文本向量无法有效表示图表的视觉语义、表格的结构化关系和公式的数学含义,在面对"图中的趋势如何"或"表格中哪个指标最高"等问题时,检索准确性严重不足。

  • Missing Context: Text and images within a document often have close cross-references and explanatory relationships. Traditional systems cannot establish these cross-modal semantic connections, leading to incomplete and inaccurate answers.

    上下文缺失: 文档中的图文内容往往存在密切的相互引用和解释关系,传统系统无法建立这种跨模态的语义关联,导致回答缺乏完整性和准确性。

  • Low Processing Efficiency: When dealing with complex documents containing many non-text elements, traditional systems often require multiple specialized tools to work together, resulting in complex workflows and low efficiency, making it difficult to meet practical application needs.

    处理效率低下: 面对包含大量非文本元素的复杂文档,传统系统往往需要多个专用工具配合处理,流程复杂、效率低下,难以满足实际应用需求。

The Practical Value of RAG-Anything

RAG-Anything项目针对上述技术挑战而设计开发。项目目标是构建一个完整的多模态RAG系统,解决传统RAG在处理复杂文档时的局限性问题。

The RAG-Anything project is designed and developed to address the above technical challenges. The project's goal is to build a complete multimodal RAG system that solves the limitations of traditional RAG when processing complex documents.

系统采用统一的技术架构,将多模态文档处理从概念验证阶段推进到实际可部署的工程化解决方案。

The system adopts a unified technical architecture, advancing multimodal document processing from the proof-of-concept stage to a practically deployable engineering solution.

此外,系统还采用了端到端的技术栈设计,覆盖文档解析、内容理解、知识构建和智能问答等核心功能模块。

Furthermore, the system employs an end-to-end technology stack design, covering core functional modules such as document parsing, content understanding, knowledge construction, and intelligent question-answering.

在文件格式支持方面,系统兼容PDF、Office文档、图像等常见格式。技术架构上,系统实现了跨模态的统一知识表示和检索算法,同时提供标准化的API接口和灵活的配置参数。

In terms of file format support, the system is compatible with common formats such as PDF, Office documents, and images. Technically, the system implements cross-modal unified knowledge representation and retrieval algorithms, while also providing standardized API interfaces and flexible configuration parameters.

RAG-Anything的技术定位是作为多模态AI应用的基础组件,为RAG系统提供可直接集成的多模态文档处理能力。

The technical positioning of RAG-Anything is as a foundational component for multimodal AI applications, providing directly integrable multimodal document processing capabilities for RAG systems.


Core Technical Advantages of RAG-Anything

RAG-Anything 通过创新的技术架构和工程实践,在多模态文档处理领域实现了显著突破:

RAG-Anything achieves significant breakthroughs in the field of multimodal document processing through innovative technical architecture and engineering practices:

· End-to-End Multimodal Processing Architecture

· 端到端多模态处理架构

构建完整的自动化处理链路,从原始文档输入开始,系统能够智能识别并精确提取文本、图像、表格、数学公式等异构内容。

Building a complete automated processing pipeline, starting from raw document input, the system can intelligently identify and precisely extract heterogeneous content such as text, images, tables, and mathematical formulas.

通过统一的结构化建模方法,建立从文档解析、语义理解、知识构建到智能问答的全流程自动化体系,彻底解决了传统多工具拼接带来的数据损失和效率问题。

Through a unified structured modeling approach, it establishes a full-process automation system from document parsing, semantic understanding, and knowledge construction to intelligent question-answering, completely solving the data loss and efficiency problems caused by traditional multi-tool integration.

· Broad Document Format Compatibility

· 广泛的文档格式兼容性

原生支持PDF、Microsoft Office套件(Word/Excel/PowerPoint)、常见图像格式(JPG/PNG/TIFF)以及Markdown、纯文本等多达10余种主流文档格式。

Natively supports over 10 mainstream document formats including PDF, Microsoft Office suite (Word/Excel/PowerPoint), common image formats (JPG/PNG/TIFF), Markdown, and plain text.

系统内置智能格式检测和标准化转换机制,确保不同来源的文档都能通过统一的处理管道获得一致的高质量解析结果。

The system includes built-in intelligent format detection and standardized conversion mechanisms, ensuring that documents from different sources can obtain consistent, high-quality parsing results through a unified processing pipeline.

· Deep Content Understanding Technology Stack

· 深度内容理解技术栈

集成视觉、语言语义理解模块和结构化数据分析技术,实现对各类内容的深度理解。

Integrates visual and linguistic semantic understanding modules with structured data analysis techniques to achieve deep understanding of various content types.

图像分析模块支持复杂图表的语义提取,表格处理引擎能够准确识别层次结构和数据关系,LaTeX公式解析器确保数学表达式的精确转换,文本语义建模则提供丰富的上下文理解能力。

The image analysis module supports semantic extraction from complex charts, the table processing engine accurately identifies hierarchical structures and data relationships, the LaTeX formula parser ensures precise conversion of mathematical expressions, and text semantic modeling provides rich contextual understanding capabilities.

· Multimodal Knowledge Graph Construction

· 多模态知识图谱构建

采用基于实体关系的图结构表示方法,自动识别文档中的关键实体并建立跨模态的语义关联。

Employs an entity-relationship-based graph structure representation method to automatically identify key entities in documents and establish cross-modal semantic associations.

系统能够理解图片与说明文字的对应关系、表格数据与分析结论的逻辑联系,以及公式与理论阐述的内在关联,从而在问答过程中提供更加准确和连贯的回答。

The system can understand the correspondence between images and their captions, the logical connection between tabular data and analytical conclusions, and the intrinsic relationship between formulas and theoretical explanations, thereby providing more accurate and coherent answers during the question-answering process.

· Flexible Modular Extensibility

· 灵活的模块化扩展

基于插件化的系统架构设计,支持开发者根据特定应用场景灵活配置和扩展功能组件。

Based on a plugin-oriented system architecture design, it supports developers in flexibly configuring and extending functional components according to specific application scenarios.

无论是更换更先进的视觉理解模型、集成专业领域的文档解析器,还是调整检索策略和嵌入算法,都可以通过标准化接口快速实现,确保系统能够持续适应技术发展和业务需求的动态变化。

Whether it's replacing a more advanced visual understanding model, integrating a domain-specific document parser, or adjusting retrieval strategies and embedding algorithms, these can all be quickly implemented through standardized interfaces, ensuring the system can continuously adapt to technological developments and dynamic changes in business needs.


RAG-Anything System Architecture

RAG-Anything基于创新的三阶段技术架构,突破传统RAG系统在多模态文档处理上的技术瓶颈,实现真正的端到端智能化处理。

RAG-Anything is based on an innovative three-stage technical architecture that breaks through the technical bottlenecks of traditional RAG systems in multimodal document processing, achieving true end-to-end intelligent processing.

  • Multimodal Document Parsing: Processes PDF, Office, and image format documents through a multimodal parsing engine, comprising four core modules: text extraction, image analysis, formula recognition, and table parsing.

    多模态文档解析 通过多模态解析引擎处理PDF、Office、图像等格式文档,包含文本提取、图像分析、公式识别和表格解析四个核心模块。

  • Cross-Modal Knowledge Construction: Builds a cross-modal knowledge graph through entity-relationship extraction and multimodal fusion techniques, establishing a unified graph representation and vector database.

    跨模态知识构建 构建跨模态知识图谱,通过实体关系抽取和多模态融合技术,建立统一的图谱表示和向量数据库。

  • Retrieval and Generation: Combines graph retrieval and vector retrieval to generate precise answers via large language models. The system adopts a modular design with high scalability and flexibility.

    检索生成 结合图谱检索和向量检索,通过大型语言模型生成精准回答。系统采用模块化设计,具备高度可扩展性和灵活性。

High-Precision Document Parsing Technology

采用基于MinerU 2.0的先进结构化提取引擎,实现对复杂文档的智能解析。系统能够准确识别文档的层次结构,自动分割文本块、定位图像区域、解析表格布局、识别数学公式。

Adopting an advanced structured extraction engine based on MinerU 2.0, it achieves intelligent parsing of complex documents. The system can accurately identify the hierarchical structure of documents, automatically segment text blocks, locate image regions, parse table layouts, and recognize mathematical formulas.

通过标准化的中间格式转换,保证不同文档类型的统一处理流程,最大化保留原始信息的语义完整性。

Through standardized intermediate format conversion, it ensures a unified processing flow for different document types, maximizing the preservation of the semantic integrity of the original information.

Deep Multimodal Content Understanding

系统内置专业化的模态处理引擎,针对不同内容类型提供定制化的理解能力:

The system includes a specialized modality processing engine that provides customized understanding capabilities for different content types:

Visual Content Analysis: Integrates a large vision model to automatically generate high-quality image descriptions, accurately extracting data relationships and visual elements from charts.

视觉内容分析: 集成视觉大模型,自动生成高质量图像描述,准确提取图表中的数据关系和视觉要素。

Intelligent Table Parsing: Deeply understands the hierarchical structure of tables, automatically identifies header relationships, data types, and logical connections, and extracts data trends and statistical patterns.

表格智能解析: 深度理解表格的层次结构,自动识别表头关系、数据类型和逻辑联系,提炼数据趋势和统计规律。

Mathematical Formula Understanding: Precisely recognizes LaTeX-formatted mathematical expressions, analyzing variable meanings, formula structures, and application scenarios.

数学公式理解: 精确识别LaTeX格式的数学表达式,分析变量含义、公式结构和适用场景。

Extended Modality Support: Supports intelligent recognition and semantic modeling of specialized content such as flowcharts, code snippets, and geographic information.

扩展模态支持: 支持流程图、代码片段、地理信息等专业内容的智能识别和语义建模。

All modality content is integrated through a unified knowledge representation framework, enabling true cross-modal semantic understanding and associative analysis.

所有模态内容通过统一的知识表示框架进行整合,实现真正的跨模态语义理解和关联分析。

Unified Knowledge Graph Construction

RAG-Anything将多模态内容统一建模为结构化知识图谱,突破传统文档处理的信息孤岛问题。

RAG-Anything uniformly models multimodal content as a structured knowledge graph, breaking through the information silo problem of traditional document processing.

Entity Modeling: Abstracts heterogeneous content such as text paragraphs, chart data, and mathematical formulas into knowledge entities, preserving complete content information, source identifiers, and type attributes.

实体化建模: 将文本段落、图表数据、数学公式等异构内容统一抽象为知识实体,保留完整的内容信息、来源标识和类型属性。

Intelligent Relationship Construction: Uses semantic analysis techniques to automatically identify logical relationships between paragraphs, explanatory relationships between text and images, and semantic connections between structured content, building a multi-layered knowledge association network.

智能关系构建: 通过语义分析技术,自动识别段落间的逻辑关系、图文间的说明关系、以及结构化内容间的语义联系,构建多层次的知识关联网络。

Efficient Storage and Indexing: Establishes a dual storage mechanism of graph database and vector database, supporting structured queries and semantic similarity retrieval, providing powerful knowledge support for complex question-answering tasks.

高效存储索引: 建立图谱数据库和向量数据库的双重存储机制,支持结构化查询和语义相似性检索,为复杂问答任务提供强大的知识支撑。

Dual-Level Retrieval for Question-Answering

RAG-Anything采用双层次检索问答机制,以实现对复杂问题的精准理解与多维响应。

RAG-Anything employs a dual-level retrieval question-answering mechanism to achieve precise understanding and multi-dimensional responses to complex questions.

该机制同时兼顾细粒度信息抽取与高层语义理解,显著提升了系统在多模态文档场景下的检索广度与生成深度。

This mechanism simultaneously considers fine-grained information extraction and high-level semantic understanding, significantly improving the system's retrieval breadth and generation depth in multimodal document scenarios.

Intelligent Hierarchical Keyword Extraction:

智能关键词分层提取:

  • Fine-grained keywords: Precisely locate specific entities, technical terms, data points, and other detailed information.

    细粒度关键词:精确定位具体实体、专业术语、数据点等详细信息

  • Concept-level keywords: Grasp thematic context, analyze trends, and understand abstract concepts.

    概念级关键词:把握主题脉络、分析趋势、理解抽象概念

Hybrid Retrieval Strategy:

混合检索策略:

  • Precise entity matching: Quickly locate relevant entity nodes through the graph structure.

    精准实体匹配:通过图谱结构快速定位相关实体节点

  • Semantic relationship expansion: Use the graph's associative relationships to discover potentially relevant information.

    语义关系扩展:利用图谱的关联关系发现潜在相关信息

  • Vector similarity retrieval: Capture semantically related content.

    向量相似性检索:捕获语义层面的相关内容

  • Contextual fusion generation: Integrate multi-source information to generate logically clear and accurate intelligent answers.

    上下文融合生成:整合多源信息,生成逻辑清晰、内容准确的智能回答

Through this dual-level retrieval architecture, the system can handle a wide range of questions, from simple fact queries to complex analytical reasoning, truly realizing an intelligent document question-answering experience.

通过这种双层次的检索架构,系统能够处理从简单事实查询到复杂分析推理的各类问题,真正实现智能化的文档问答体验。


Quick Deployment Guide

RAG-Anything提供两种便捷的安装部署方式,满足不同用户的技术需求。推荐使用PyPI安装方式,可实现一键快速部署,体验完整的多模态RAG功能。

RAG-Anything offers two convenient installation and deployment methods to meet the technical needs of different users. The PyPI installation method is recommended for one-click rapid deployment and a full experience of the multimodal RAG capabilities.

Installation Methods

Option 1: Install from PyPI

选项1:从PyPI安装

pip install raganything

Option 2: Install from Source

选项2:从源码安装

git clone https://github.com/HKUDS/RAG-Anything.git
cd RAG-Anything
pip install -r requirements.txt

Multi-Scenario Application Modes

RAG-Anything基于模块化架构设计,为不同应用场景提供两种灵活的使用路径,满足从快速原型到生产级部署的各类需求:

RAG-Anything is designed with a modular architecture, offering two flexible usage paths for different application scenarios, meeting various needs from rapid prototyping to production-level deployment:

Mode 1: One-Click End-to-End Processing

方式一:一键式端到端处理

Applicable Scenario: Processing complete original documents like PDFs, Word files, and PPTs, pursuing zero-configuration, fully automatic intelligent processing.

适用场景:处理完整的PDF、Word、PPT等原始文档,追求零配置、全自动的智能处理。

Core Advantages:

核心优势:

  • Full-process automation: From document upload to intelligent question-answering, no manual intervention is required.

    全流程自动化:从文档上传到智能问答,无需人工干预

  • Intelligent structure recognition: Automatically detects title hierarchy, paragraph structure, image positions, table layouts, and mathematical formulas.

    智能结构识别:自动检测标题层次、段落结构、图像位置、表格布局、数学公式

  • Deep content understanding: Semantic analysis and vector representation of multimodal content.

    深度内容理解:多模态内容的语义分析和向量化表示

  • Automatic knowledge graph construction: Automatically generates a structured knowledge network and retrieval index.

    知识图谱自构建:自动生成结构化知识网络和检索索引

Technical Flow: Raw Document → Intelligent Parsing → Multimodal Understanding → Knowledge Graph Construction → Intelligent Question-Answering

技术流程: 原始文档 → 智能解析 → 多模态理解 → 知识图谱构建 → 智能问答

Example Code:

示例代码:

from raganything import RAGAnything

# Initialize the system
rag = RAGAnything()

# Process a document end-to-end
rag.process_document("path/to/your/document.pdf")

# Ask a question
answer = rag.query("What is the main finding of this paper?")
print(answer)

Mode 2: Fine-Grained Manual Construction

方式二:精细化手动构建

Applicable Scenario: When you already have structured multimodal content data (images, tables, formulas, etc.) and need precise control over the processing flow and customized function extensions.

适用场景:已有结构化的多模态内容数据(图像、表格、公式等),需要精确控制处理流程和定制化功能扩展。

Core Advantages:

核心优势:

  • Precise control: Manually specify the processing method for key content like images and tables.

    精确控制:手动指定图像、表格等关键内容的处理方式

  • Customized processing: Adjust parsing strategies based on specific domain requirements.

    定制化处理:根据特定领域需求调整解析策略

  • Incremental construction: Supports gradual addition and updating of multimodal content.

    增量构建:支持逐步添加和更新多模态内容

  • Professional optimization: Deep optimization for specific document types.

    专业优化:针对特定文档类型进行深度优化


Future Outlook for RAG-Anything

Deep Reasoning Capability Upgrade

RAG-Anything将构建具备人类级别逻辑推理能力的多模态AI系统。通过多层次推理架构实现从浅层检索到深层推理的跃升,支持跨模态多跳深度推理和因果关系建模。考虑提供可视化推理路径追踪、证据溯源和置信度评估。

RAG-Anything will build a multimodal AI system with human-level logical reasoning capabilities. Through a multi-layered reasoning architecture, it will achieve a leap from shallow retrieval to deep reasoning, supporting cross-modal multi-hop deep reasoning and causal relationship modeling. It also considers providing visual reasoning path tracking, evidence tracing, and confidence assessment.

Richer Plugin Ecosystem

RAG-Anything未来也会考虑从另一个维度实现扩展——探索构建开放的多模态处理生态系统。我们设想让不同行业都能拥有更贴合需求的智能助手。

In the future, RAG-Anything will also consider expansion from another dimension—exploring the construction of an open multimodal processing ecosystem. We envision enabling different industries to have intelligent assistants that better fit their needs.

比如帮助科研人员更好地解析学术图表,协助金融分析师处理复杂的财务数据,或者让工程师更容易理解技术图纸,医生更快速地查阅病历资料等。

For example, helping researchers better parse academic charts, assisting financial analysts in processing complex financial data, or making it easier for engineers to understand technical drawings and for doctors to quickly review medical records.


References:

参考资料:

This article is adapted from the WeChat public account "新智元" (New Intelligence Era), edited by LRST, and republished with authorization from 36Kr.

本文来自微信公众号“新智元”,编辑:LRST,36氪经授权发布。

常见问题(FAQ)

RAG-Anything能处理哪些类型的内容?

RAG-Anything能处理文本、图像、表格和数学公式等多种异构内容,构建统一的多模态知识图谱,实现跨模态的语义关联和智能问答。

RAG-Anything与传统RAG系统的主要区别是什么?

传统RAG系统仅支持文本处理,而RAG-Anything通过多模态知识图谱架构,能同时处理并关联文档中的文字、图表、表格和公式,实现端到端的文档解析与智能问答。

RAG-Anything的开源地址和开发团队是?

RAG-Anything由香港大学黄超教授团队开发,开源地址为https://github.com/HKUDS/RAG-Anything,基于LightRAG框架扩展,支持多模态文档处理。

← 返回文章列表
分享到:微博

版权与免责声明:本文仅用于信息分享与交流,不构成任何形式的法律、投资、医疗或其他专业建议,也不构成对任何结果的承诺或保证。

文中提及的商标、品牌、Logo、产品名称及相关图片/素材,其权利归各自合法权利人所有。本站内容可能基于公开资料整理,亦可能使用 AI 辅助生成或润色;我们尽力确保准确与合规,但不保证完整性、时效性与适用性,请读者自行甄别并以官方信息为准。

若本文内容或素材涉嫌侵权、隐私不当或存在错误,请相关权利人/当事人联系本站,我们将及时核实并采取删除、修正或下架等处理措施。 也请勿在评论或联系信息中提交身份证号、手机号、住址等个人敏感信息。

您可能感兴趣