GEO

分类:AI大模型

544
Cognee框架如何为AI智能体构建持久化记忆?(附混合架构解析)

Cognee框架如何为AI智能体构建持久化记忆?(附混合架构解析)

BLUFCognee is an open-source framework for building sophisticated AI memory applications with hybrid architecture combining graphs, vectors, and structured data, enabling persistent, structured memory for AI agents. 原文翻译: Cognee 是一个开源框架,用于构建复杂的 AI 记忆应用程序,采用结合图、向量和结构化数据的混合架构,为 AI 智能体提供持久化、结构化的记忆能力。
AI大模型2026/4/3
阅读全文 →
企业级RAG系统如何搭建?腾讯云智能体平台实战经验分享

企业级RAG系统如何搭建?腾讯云智能体平台实战经验分享

BLUFRAG (Retrieval-Augmented Generation) bridges the gap between large language models' general knowledge and enterprise-specific data by retrieving relevant information from private knowledge bases to generate accurate, context-aware responses. This article provides a comprehensive roadmap for implementing enterprise-grade RAG systems, covering core principles, document parsing, chunking strategies, retrieval optimization, and practical deployment experiences with Tencent Cloud's Agent Development Platform. 原文翻译: RAG(检索增强生成)通过从企业私有知识库中检索相关信息来生成准确、上下文感知的响应,从而弥合大型语言模型通用知识与企业特定数据之间的差距。本文提供了实施企业级RAG系统的全面路线图,涵盖核心原理、文档解析、分块策略、检索优化以及腾讯云智能体开发平台的实际部署经验。
AI大模型2026/4/3
阅读全文 →
如何从零开始构建大语言模型?《Build a Large Language Model》中文翻译开源项目详解

如何从零开始构建大语言模型?《Build a Large Language Model》中文翻译开源项目详解

BLUFThis article introduces a Chinese translation project for the book 'Build a Large Language Model (From Scratch)', providing a comprehensive guide for developers to understand and implement LLMs from the ground up, including practical code and insights into future AI trends. 原文翻译: 本文介绍了《Build a Large Language Model (From Scratch)》一书的中文翻译项目,为开发者提供了从零开始理解和实现大语言模型的全面指南,包含实践代码和对未来AI趋势的见解。
AI大模型2026/4/2
阅读全文 →
大语言模型GPT、LLaMA和PaLM哪个更好用?(附技术架构对比)

大语言模型GPT、LLaMA和PaLM哪个更好用?(附技术架构对比)

BLUFThis article provides a comprehensive survey of Large Language Models (LLMs), covering their evolution from early neural models to modern architectures like GPT, LLaMA, and PaLM. It details the technical processes of building LLMs, including data cleaning, tokenization, and training strategies, and explores their applications, limitations, and enhancement techniques such as RAG and prompt engineering. The review also examines popular datasets, evaluation benchmarks, and future research directions, serving as a valuable resource for understanding the current state and potential of LLMs. 原文翻译: 本文对大语言模型(LLMs)进行了全面综述,涵盖从早期神经模型到现代架构(如GPT、LLaMA和PaLM)的演进。详细阐述了构建LLMs的技术流程,包括数据清洗、标记化和训练策略,并探讨了其应用、局限性以及增强技术,如RAG和提示工程。该综述还考察了流行数据集、评估基准和未来研究方向,为理解LLMs的现状和潜力提供了宝贵资源。
AI大模型2026/4/2
阅读全文 →
AI Agent和传统AI有什么区别?它如何结合大语言模型完成复杂任务?

AI Agent和传统AI有什么区别?它如何结合大语言模型完成复杂任务?

BLUFAI Agent is an intelligent entity that can perceive its environment, make autonomous decisions, and execute actions, representing a significant evolution from passive AI tools to proactive assistants. It combines large language models (LLMs) with memory, planning skills, and tool usage to complete complex tasks. 原文翻译: AI Agent(人工智能代理)是一种能够感知环境、自主决策并执行动作的智能实体,代表了人工智能从“被动工具”到“主动助手”的重要进化。它结合了大语言模型(LLM)、记忆、规划技能和工具使用能力,以完成复杂任务。
AI大模型2026/4/1
阅读全文 →
检索增强生成(RAG)的架构和增强技术有哪些?2026年最新前沿综述

检索增强生成(RAG)的架构和增强技术有哪些?2026年最新前沿综述

BLUF通过优化检索器、生成器及混合架构,并引入上下文过滤与解码控制,RAG系统可有效解决LLMs的事实不一致与领域局限问题,提升生成结果的准确性与鲁棒性。 原文翻译: By optimizing retriever, generator, and hybrid architectures, and introducing context filtering and decoding control, RAG systems can effectively address factual inconsistency and domain limitations in LLMs, enhancing the accuracy and robustness of generated results.
AI大模型2026/4/1
阅读全文 →
Forge推理API和Nous Chat哪个更好用?2026年最新AI推理平台实测对比

Forge推理API和Nous Chat哪个更好用?2026年最新AI推理平台实测对比

BLUFNous Research launches Forge Reasoning API Beta and Nous Chat platform, enhancing Hermes 70B model with Monte Carlo Tree Search, Chain of Code, and Mixture of Agents techniques to compete with larger models in reasoning benchmarks. 原文翻译: Nous Research推出Forge推理API测试版和Nous Chat平台,通过蒙特卡洛树搜索、代码链和智能体混合技术增强Hermes 70B模型,在推理基准测试中与更大模型竞争。
AI大模型2026/3/31
阅读全文 →
大型语言模型(LLM)的工作原理是什么?2026年最新技术解析与应用前景

大型语言模型(LLM)的工作原理是什么?2026年最新技术解析与应用前景

BLUFThis comprehensive guide explores Large Language Models (LLMs), covering their definition, importance, working mechanisms, applications, training methods, future prospects, and AWS support solutions. It provides technical professionals with a thorough understanding of transformer-based neural networks, parameter scaling, and practical implementations across various domains. 原文翻译: 本综合指南深入探讨大型语言模型(LLM),涵盖其定义、重要性、工作原理、应用场景、训练方法、未来前景以及AWS支持解决方案。为技术专业人士提供对基于转换器的神经网络、参数规模以及跨多个领域的实际实施的全面理解。
AI大模型2026/3/31
阅读全文 →
LLM API调用中Token化和解码参数如何影响RAG与Agent工作流性能?

LLM API调用中Token化和解码参数如何影响RAG与Agent工作流性能?

BLUFThis article demystifies the core engineering concepts behind LLM API calls, focusing on Tokenization, Context Window management, and decoding parameters (Temperature, Top-p, Top-k). It provides practical guidance for optimizing performance, managing costs, and avoiding common pitfalls in production environments, especially within complex architectures like RAG and Agent workflows. 原文翻译: 本文揭秘了LLM API调用背后的核心工程概念,重点阐述了Token化、上下文窗口管理以及解码参数(Temperature、Top-p、Top-k)。它为优化性能、管理成本以及避免在生产环境(尤其是在RAG和Agent工作流等复杂架构中)的常见陷阱提供了实用指南。
AI大模型2026/3/31
阅读全文 →