Multi-Agent Framework
AutoGPT — The 2023 Autonomous Agent That Started the Movement logo

AutoGPT — 2023 引发 AI Agent 浪潮的自主 Agent 项目

AutoGPT 2023 年爆红,是首个公开的"自主 Agent"——LLM 自己规划、调用工具、循环执行无需人工介入。如今演进为 AutoGPT Platform 无代码 Agent 构建器。

为什么选它

The original AutoGPT (2023) was a Python script that fed GPT-4 a goal, let it plan steps, call tools (web, file, shell), execute them, and iterate. It went viral on GitHub (150K+ stars) and kicked off the public conversation about "autonomous agents". Its practical usefulness was limited — loops got stuck, hallucinated, and burned tokens — but as a concept demo it shaped the entire field.

In 2024-2025 the project pivoted to AutoGPT Platform: a visual, no-code agent builder with a node-based interface, block library, marketplace of pre-built agents, and cloud execution. The core autonomy-by-looping idea remains, but wrapped in a product that non-engineers can use.

For developers in 2026, AutoGPT is worth studying historically and potentially useful as a no-code frontend for simple agent workflows. For production agents you’re building in code, use CrewAI, LangGraph, or AutoGen — AutoGPT’s code path isn’t where the innovation is happening anymore.

Quick Start — AutoGPT Platform (Docker)

The 2023 Python package (autogpt / pyautogpt) is deprecated. New users should start with the Platform. If you want the original "autonomous loop" pattern in code, look at tools like BabyAGI, Agno, or roll a minimal loop with any agent framework — AutoGPT Platform is optimized for no-code users now.

# The modern AutoGPT is the Platform — a local web app, not a Python script.
# https://docs.agpt.co/platform/getting-started/

git clone https://github.com/Significant-Gravitas/AutoGPT.git
cd AutoGPT/autogpt_platform

# Create .env and set OPENAI_API_KEY (or any supported provider)
cp .env.example .env
nano .env      # fill OPENAI_API_KEY=sk-...

# Boot the full stack (backend, frontend, Postgres, Redis, S3-compat) via Docker
docker compose up -d --build

# Open http://localhost:3000 and sign in
# Build agents in the visual editor using blocks:
#   - Triggers (manual, cron, webhook)
#   - LLM blocks (OpenAI, Anthropic, Ollama, ...)
#   - Logic blocks (branch, loop, store)
#   - Integrations (Gmail, Slack, web search, GitHub, ...)
# Publish runs on the local scheduler, or push to AutoGPT cloud.

核心能力

Visual node editor

Drag-and-drop blocks on a canvas. Connect outputs to inputs. Run or schedule the resulting graph. Lowest barrier-to-entry for building agents without writing code.

Block library

100+ built-in blocks: LLM calls (OpenAI, Claude, Gemini, Ollama), logic, storage, integrations (Gmail, Discord, Slack, Notion, GitHub, web search, scraping). Extendable with custom Python blocks.

Agent marketplace

Community-shared agents ready to fork. Helpful for discovery — though quality varies widely. Treat as starter templates, not production-ready.

Cron + webhook triggers

Schedule runs or trigger via HTTP webhook. Enables "autonomous" agents that fire on events or on a schedule without human intervention.

Self-host + cloud

Full stack runs via docker-compose. AutoGPT Cloud (paid) for managed hosting and higher-tier runs.

Historical codebase preserved

The classic Python loop is archived in the repo for study. Useful as a teaching example for "minimum autonomous agent loop" even if you use a modern framework for real work.

对比

 AudienceProduction PathCode RequiredMaturity
AutoGPT PlatformthisNo-code + devSelf-host or cloudMinimal (custom blocks)Stable, evolving rapidly
n8n + LLM nodesNo-code automationSelf-host or cloudMinimalMature in automation space
CrewAI / LangGraphDevelopersCode-firstYesProduction-grade
BabyAGIEducationalForks onlyYesPattern reference

实际用例

01. No-code automation with LLMs

Business users and PMs who want agents but can’t or won’t write Python. AutoGPT Platform + a block marketplace covers many "schedule this and email me" workflows.

02. Quick prototyping of multi-step agents

Even developers sometimes prefer to sketch an agent in a visual canvas to validate the flow before translating to code. AutoGPT is fine for that.

03. Teaching the autonomous agent pattern

The archived classic codebase is still the cleanest minimal example of plan → act → reflect → iterate.

价格与许可

AutoGPT Platform: MIT open source. Free to self-host via Docker.

AutoGPT Cloud: managed hosting with tiered pricing — free hobby tier, paid tiers for higher run limits. See agpt.co.

Model cost: whatever LLMs your blocks use. Cheap models (gpt-4o-mini, haiku) are usually enough for block-level reasoning; reserve big models for blocks that materially benefit.

相关 TokRepo 资产

常见问题

Should I use the old AutoGPT Python package?+

No — it is deprecated. The new direction is AutoGPT Platform (Docker, visual editor). For code-first agent development, CrewAI / LangGraph / AutoGen are the modern defaults.

AutoGPT Platform vs n8n?+

n8n is a mature general-purpose automation platform with LLM nodes. AutoGPT Platform is LLM-agent-first. If your workflows are "traditional automations with AI sprinkled in", n8n wins on stability and integrations. If they are "agent-led with optional automation", AutoGPT is more natural.

Can I self-host the Platform fully?+

Yes — the docker-compose stack runs everything locally (frontend, backend, Postgres, Redis, blob storage). You can swap the blob store for S3 and scale out components individually.

Is the autonomy real or a marketing term?+

Agents in the Platform run through a graph you define — they are not free-floating loops like the 2023 script. "Autonomy" means "runs without a human present at each step", not "AGI that decides its own goals". Good that this is clearer now than in 2023.

Historical question — why did AutoGPT get stuck so much in 2023?+

Short context windows, no structured planning, weak tool-use reliability, and a strong tendency to loop on verification failures. The field has since added planning frameworks, structured tool schemas, longer context, and grounding tools. Modern agents are dramatically more reliable.

同类推荐