Forge Agent / v1.0 / MIT

The AI agent runtime that does not lock you in.

Open-source. Type-safe tool use. Plug-and-play for any LLM. Ship a production agent in under 50 lines of code without Python circus tricks or 14-layer abstractions. Forge Agent is the runtime we wish we had when we started building voice receptionists, trading bots, and autonomous research agents.

Star on GitHub → Read the Docs
MIT licensed
TypeScript + Python
Multi-model (Anthropic, OpenAI, Google)
Self-hostable
Quickstart

Ship a working agent in under a minute.

No hidden chain-of-chain-of-agents. One class, typed tools, your LLM of choice. Run it locally, deploy it anywhere a Node or Python process runs.

// npm install @forge/agent @forge/tools
import { Agent } from "@forge/agent";
import { webSearch, sendEmail } from "@forge/tools";

const researcher = new Agent({
  model: "claude-opus-4-7",
  tools: [webSearch, sendEmail],
  systemPrompt: "You are a research analyst. Investigate leads, then email the summary.",
});

await researcher.run({
  input: "Research Forge Dev Studio and email findings to corey@forgedev.studio",
  maxSteps: 10,
});
Core Features

Built for the 90% case, not the 0.01%.

Every feature earns its place. No graph-of-graphs. No config file gymnastics. Just the primitives you need to build an agent that runs in production without babysitting.

Type-safe tools

Declare a tool's input and output schemas once in TypeScript or Zod. The runtime enforces them, catches hallucinated arguments, and returns typed results your code can trust.

Multi-model

Swap between Anthropic Claude, OpenAI, Google Gemini, and local models (Ollama, vLLM) with one line. Prompt caching, tool use, and streaming work the same across providers.

Plugin architecture

Drop in first-party plugins for MCP servers, vector stores, CRM integrations, voice, browser automation. Or write your own in ten lines and publish to npm.

Streaming + interruption

Server-sent events out of the box. Users can see tokens land in real time and interrupt the agent mid-stream without losing state.

Structured output

Force JSON schemas, enum choices, or typed object trees. The runtime retries on shape mismatch and surfaces typed objects to your caller.

Built-in observability

Every tool call, every reasoning step, every retry is logged with OpenTelemetry-compatible traces. Plug into Grafana, Datadog, or Forge Vault and watch agents work in real time.

Use Cases

What people are shipping with it.

01 / Voice

Forge Voice itself

The production bilingual voice agent that handles real calls at (234) 235-2314 is built on Forge Agent. One instance per call, typed CRM and calendar tools, ElevenLabs streaming voice plugin.

02 / Trading

Autonomous quant agents

Market-regime classifiers feed signals to execution agents that submit orders to Kraken and Alpaca. Runs 24/7 with budget tracking and human override.

03 / Intake

Lead triage at scale

Inbound web forms feed a triage agent that enriches with firmographic data, scores urgency, and routes to the right partner or calendar. Handles 2,000+ leads per day for our staffing partner.

04 / Research

Deep-research agents

Multi-step web research with source citations. Plugin chain: webSearch → fetchUrl → summarize → synthesize. Replaces a junior analyst for early-stage diligence memos.

Integrations

Works with the stack you already use.

First-party plugins covered. Everything else is a ten-line plugin.

Anthropic
Claude Opus/Sonnet/Haiku
OpenAI
GPT-4o, o1
Google
Gemini 2.5/3.1
Ollama
Local models
MCP
Server + client
Pinecone
Vector memory
Postgres
pgvector, pooled
Firestore
Session + state
Stripe
Billing tools
Twilio
SMS + voice
SendGrid
Transactional email
Playwright
Browser automation
Questions

Before you npm install.

Is Forge Agent production-ready?

Yes. It runs our production bilingual voice agent on a real phone number, our quant trading bots, and our lead triage system. Every major release is smoke-tested against those workloads before shipping.

How is this different from LangChain or CrewAI?

Two things. One: no fake abstractions. We do not wrap every LLM call in a Chain of a Chain of a Graph of a Runnable. The runtime is the model, your tools, and a loop. Two: typed tool use is first-class. Tool arguments and results are fully typed, schema-validated, and retried on shape mismatch without config.

Do I need to use a cloud service?

No. Forge Agent runs on any Node or Python host. Bring your own LLM API key and you are done. Forge Vault (separate product) is our cloud option if you want hosted multi-tenant agent management, but the runtime itself never phones home.

What license?

MIT. Use it commercially, modify it, redistribute it. No strings, no telemetry.

What about MCP servers?

First-class. Forge Agent can consume MCP servers as plugins or expose its own tools as an MCP server for Claude Desktop, Cursor, and any other MCP-aware client.

How do I learn it?

Read the quickstart on GitHub, then pick one of the example agents in /examples (voice receptionist, lead-scoring agent, research agent). If you can read 200 lines of TypeScript you can ship your first agent today.

Build an agent that ships this weekend.

Star the repo. Read the quickstart. By Monday you have a production-ready agent with typed tools and your LLM of choice. No abstractions hunt, no plumbing tax.

Star Forge Agent → Talk to Corey
The Full Stack

Every Forge product, in one platform.

Forge Vault

Multi-tenant platform

Enterprise agent management. RBAC, audit, credentials.

Forge Voice

AI receptionist

Bilingual voice agent that answers every call.

Forge Desk

Virtual desktops

Managed dev environments in the cloud.

Forge Studio

AI-native IDE

Pair program, ship, deploy. All in one place.