The Production runtime for Agent Skills

Give your agents
Skills runtime

Cursor and Claude have Skills because they have runtimes. BlueBag gives your production agents the same power: sandboxed execution, version control, works with any LLM
Quick install
npm i @bluebag/ai-sdk
OpenAI
Anthropic
Gemini
Llama
DeepSeek
Mistral
Grok

You've hit this.

Your product demo works. Production breaks.
A runtime layer is not optional.
1
Skills break outside demo environments
Demos aren't production. Cursor and Claude have runtimes. Your app doesn't. No sandbox, no file system, no execution layer. Skills fall apart after deploy
2
Vendor lock-in is the default
Skills built for Claude stay in Claude. Cursor stays in Cursor. Switch LLMs? Rewrite capabilities. Your agent roadmap becomes a vendor contract
3
Maintaining infrastructure is a full-time job
LLMs reason. They don't execute. You still need sandboxes, dependency management, file handling, session state, observability. Maintaining these take forever
What You Get

Everything you need to ship Skills

Versioned skills. A playground to test. A CLI to ship.
Logs when things break.
Dashboard
Skills Overview
12
Skills
847
Executions
99%
Success
code-review
● v2.1
data-analysis
● v1.4
Manage skills, view executions, monitor usage.
Skill versioning
Execution logs
Usage analytics
Playground
Review the auth middleware for security issues
Run
Sandbox initialized in 23ms
Reading auth/middleware.ts...
Found 2 potential issues
Session saved
Test skills before deploying to production.
Live testing
Debug mode
Session replay
CLI
$
bluebag push
→ Detecting changes...
→ skills/code-review (modified)
→ skills/data-analysis (new)
✓ Pushed 2 skills to production
✓ Version: v2.1.0
$
_
Git-like workflow for skill management.
Push/pull skills
Team sync
CI/CD ready
Versioning
v2.1
Add error handling for edge cases
2 hours ago • Production
v2.0
Refactor response format
1 day ago
Rollback
v1.9
Initial sandbox integration
3 days ago
Rollback
Version control built into the platform.
Rollback support
Branch testing
Diff view
3
Lines to integrate
<50ms
Cold start time
LLMs supported
0
Vendor lock-in
Skill Sources

Import skills. Run in production.

Import from skills.sh or GitHub. BlueBag runs them in sandboxes with files, dependencies, and isolation handled.
Recommended
Import from skills.sh
The public skill marketplace
Find a skill. Swap the domain. Import complete.
# One URL swap:
https://skills.sh/... → https://bluebag.ai/...
Browse skills.sh
Import from GitHub
All public GitHub skill repos
Import any public skill repository. BlueBag versions it. Deploy instantly.
# One URL swap:
https://github.com/... → https://bluebag.ai/...
Read the docs
Demo environments vs. Production
Without BlueBag
Skills run in Cursor/Claude only
No sandbox outside demo
No file system access
Skills break when deployed
Rebuild everything for production
With BlueBag
Same skills, any environment
Isolated Sandboxed environments
Full file system per session
Skills work everywhere
Production-ready from day one
Integration

Add skills to your agent in 3 lines

Add BlueBag. Keep your stack. Skills now execute in production sandboxes.
1
Install
One package. No migrations.
2
Enhance
Inject skills + tools at runtime.
3
Done
Production sandbox execution.
Sandboxed
Versioned
LLM-agnostic
api/chat/route.ts
import { Bluebag } from "@bluebag/ai-sdk";
const bluebag = new Bluebag({ apiKey: process.env.BLUEBAG_API_KEY });
const config = await bluebag.enhance({ model, messages });
return streamText(config);
1
Install
Requires @langchain/core ^0.3.0. Any LangChain-compatible model.
2
Enhance
Skills merge into your tools array. Existing tools keep working.
3
Portable
Switch from OpenAI to Anthropic to Llama. Skills follow.
agent/index.ts
import { createAgent } from "langchain";
import { HumanMessage } from "@langchain/core/messages";
import { Bluebag } from "@bluebag/langchain";
const bluebag = new Bluebag({
apiKey: process.env.BLUEBAG_API_KEY,
});
const config = await bluebag.enhance({
model: "openai:gpt-4o",
tools: [yourCustomTool],
systemMessage: "You are a helpful assistant.",
messages: [new HumanMessage("Analyze this document")],
});
const agent = createAgent({
model: config.model,
tools: config.tools,
systemPrompt: config.systemMessage,
});
const result = await agent.invoke({ messages: config.messages });
1
Auth
One command links your terminal to your BlueBag project.
2
Push/Pull
Skills are markdown files. Version in git. Deploy with push.
3
CI/CD ready
Team sync. Conflict resolution. Automated deploys.
terminal
# Install the CLI
npm install -g @bluebag/cli
# Authenticate (one time)
bluebag auth
# Pull existing skills from your project
bluebag pull
# Create a new skill locally
mkdir skills/code-review
cat > skills/code-review/SKILL.md << 'EOF'
---
name: code-review
description: Reviews code for best practices
---
# Code Review
Review the provided code for:
1. Security vulnerabilities
2. Performance issues
3. Best practice violations
EOF
# Push to BlueBag (versioned automatically)
bluebag push
Under the Hood

The runtime you don't have to build

Sandboxes, isolation, scaling, security—handled. Built on battle-tested execution machines. You focus on real capabilities via Skills.
Isolated Sandboxes
Every execution runs in its own Sandbox. Full file system. Python, Node, shell. Complete isolation.
Secure by Default
No arbitrary code in your infra. Managed sandboxes with controlled network. Full audit logs.
Sub-50ms Cold Starts
Sandboxes spin up instantly. Persistent sessions keep state across tool calls. Auto-cleanup after.
Enterprise Ready
SOC 2 in progress. Self-hosted available. Data stays in your region. Compliance-approved.
Your Agent
Bluebag SDK
Any LLM
Skill Discovery
Tool Injection
Sandbox Exec
Runtime Virtual Machines
Same architecture that powers Claude and Cursor Skills, now for your production agents

Start shipping production agents with Skills.

The runtime layer Anthropic built for Claude Skills . Now available for your agents.
Production ready in minutes, not months.
Free tier
No credit card
5 min
To first skill
Open spec
No lock-in