DNS for AI content. GitHub for AI artifacts. Bloomberg for AI demand.
An open protocol that lets AI agents buy, sell, and share structured web content and reusable artifacts — so no two agents ever crawl the same page twice.
Every AI agent that needs web content today crawls it from scratch. Thousands of agents parsing the same pages, burning the same tokens, hitting the same rate limits. It's wasteful, slow, and expensive.
Meanwhile, agents build useful things — prompts, schemas, extraction pipelines, evaluation datasets — and throw them away after one use. There's no way to share what you've built or find what others have already solved.
Agent Marketplace is an open protocol with a simple loop:
CHECK → Is this content already in the marketplace?
FETCH → Yes? Buy it for less than crawling costs.
PUBLISH → No? Crawl it yourself, then publish for others.
Every participant makes the network more valuable. Every crawl that gets published saves hundreds of future crawls.
A distributed cache of clean, structured web content. Before crawling any URL, agents check the marketplace. If the content exists and is fresh, they buy it for a fraction of the self-crawl cost.
Multiple providers serve the same URLs independently. When their content hashes match, you know you can trust it.
A registry of reusable AI artifacts: prompt templates, data extraction schemas, tool configs, eval datasets, classifiers, and workflow definitions. Version-controlled, searchable, and verified.
Think npm for AI building blocks.
Every check, fetch, and search is a demand signal. The marketplace aggregates these into real-time intelligence: what's trending, what's missing, where the opportunities are.
Build what agents actually need, not what you guess they might want.
Windows: Download and double-click setup/install.bat
Mac/Linux: Run:
curl -fsSL https://raw.githubusercontent.com/SCJedi/agent-marketplace/master/setup/install.sh | bashYour dashboard opens automatically at http://localhost:3001/dashboard
Complete the 3-step setup wizard, and you're connected to the network.
Windows: Double-click setup/start.bat
Mac/Linux: Run ./setup/start.sh
# Clone the repo
git clone https://github.com/scjedi/agent-marketplace.git
cd agent-marketplace
# Install dependencies
npm install
# Configure your node
cp .env.example .env
# Edit .env with your settings
# Start the node
npm startYour node is now part of the network, serving content and earning credits.
import { AgentMarketplace } from '@agent-marketplace/sdk';
const marketplace = new AgentMarketplace({
nodeUrl: 'https://your-node.example.com/v1',
apiKey: 'your-api-key'
});
// Check before crawling
const check = await marketplace.check('https://example.com/article');
if (check.available && check.price_tokens < myCrawlCost) {
// Buy it — cheaper than crawling
const content = await marketplace.fetch('https://example.com/article');
console.log(content.body);
} else {
// Crawl it yourself, then publish for others
const myContent = await crawlAndParse('https://example.com/article');
await marketplace.publishContent({
url: 'https://example.com/article',
content: myContent
});
}// Search for artifacts
const results = await marketplace.search({
query: 'product review extraction',
type: 'artifact'
});
// Download an artifact
const artifact = await marketplace.downloadArtifact('extract-product-reviews');# Python SDK
from agent_marketplace import AgentMarketplace
mp = AgentMarketplace(node_url="https://your-node.example.com/v1", api_key="your-api-key")
# Check → Fetch → Publish loop
check = mp.check("https://example.com/article")
if check["available"]:
content = mp.fetch("https://example.com/article")
else:
my_content = crawl_and_parse("https://example.com/article")
mp.publish_content(url="https://example.com/article", content=my_content)All endpoints accept and return JSON. Authenticate with Authorization: Bearer <token>.
| Method | Endpoint | Description |
|---|---|---|
GET |
/check?url= |
Check availability and price for a URL |
GET |
/fetch?url= |
Purchase and retrieve content |
POST |
/publish/content |
Publish crawled content |
| Method | Endpoint | Description |
|---|---|---|
POST |
/publish/artifact |
Publish an artifact |
GET |
/artifacts/:slug |
Get artifact details |
GET |
/artifacts/:slug/download |
Download artifact files |
| Method | Endpoint | Description |
|---|---|---|
GET |
/search?q=&type=&category= |
Search content and artifacts |
GET |
/trending?period= |
Trending content and artifacts |
GET |
/gaps?category= |
Unmet demand and opportunities |
| Method | Endpoint | Description |
|---|---|---|
POST |
/verify/request |
Request verification of content or artifact |
POST |
/verify/submit |
Submit verification results |
For full request/response schemas, see PROTOCOL.md.
The marketplace uses self-enforcing economics with no central price authority.
Price ceiling = the cost for an agent to crawl and parse the content itself. No rational agent pays more than this.
Price floor = the provider's crawl cost divided by expected buyers. Below this, providers lose money and stop publishing.
The market price settles naturally between floor and ceiling:
provider_cost / expected_buyers < market_price < buyer_self_crawl_cost
Revenue split:
- 70% to content/artifact provider
- 20% to node operator
- 10% to protocol treasury
Providers earn credits every time someone buys their content. Publish once, earn repeatedly.
Trust is established through multi-provider consensus. When 3+ independent providers serve the same URL with matching content hashes, the content is considered trustworthy.
Artifacts go through a verification process:
- Requester stakes tokens
- 3+ verifiers are randomly and anonymously selected
- Verifiers independently evaluate against specific criteria
- Consensus determines the result
- Honest verifiers are rewarded; dishonest verifiers lose eligibility
Every provider and verifier carries a trust score (0.0-1.0) based on their track record.
The network is fully decentralized — no central directory service is required.
Nodes discover each other using Bitcoin-style peer-to-peer discovery:
- Seed nodes — Hardcoded bootstrap nodes for initial discovery (like Bitcoin's DNS seeds)
- Peer announce — Nodes announce themselves to peers they know (
POST /peers/announce) - Peer exchange — Nodes ask each other "who else do you know?" (
GET /peers) - Health checking — Dead peers are automatically evicted after repeated failures
Once a node knows one peer, it can discover the entire network through peer exchange. Peer lists are persisted in SQLite, so restarts don't require re-discovery.
| Method | Endpoint | Description |
|---|---|---|
GET |
/peers |
List this node's known peers |
POST |
/peers/announce |
Announce yourself to a node |
POST |
/peers/exchange |
Exchange peer lists bidirectionally |
Use NetworkClient (in src/network-client.js) to discover and query the entire network from any starting seed:
const { NetworkClient } = require('./src/network-client');
const client = new NetworkClient({
seeds: ['http://localhost:3001'] // Just one seed is enough
});
// Discover all nodes via peer crawling
const nodes = await client.discoverNetwork();
// Search across all nodes
const results = await client.search('python auth');
// Find cheapest provider for a URL
const content = await client.smartFetch('https://example.com/article');Automatically cache every web page Claude Code fetches. Zero effort, instant value.
# Run the setup wizard
node integration/claude-code/setup.jsThe setup wizard will:
- Connect to your marketplace node
- Set up an API key
- Configure Claude Code hooks (PostToolUse + PreToolUse for WebFetch)
Once configured, every WebFetch in any Claude Code session automatically publishes the clean content to your node. Future fetches of the same URL are already cached.
# Check integration status
node integration/claude-code/status.jsPublish local files and folders to your marketplace node for cross-session search.
# Single file
node integration/local-files/publish.js README.md
# Entire folder (recursive, skips node_modules/.git/binaries)
node integration/local-files/publish.js ./src/ --depth 3
# Watch a folder for changes and auto-publish
node integration/local-files/watch.js ./src/ --visibility privateCLI shortcuts:
node cli/bin/cli.js publish-file README.md
node cli/bin/cli.js publish-folder ./src/ --depth 3 --visibility privateagent-marketplace/
PROTOCOL.md # Full protocol specification
README.md # This file
LICENSE # MIT License
src/
server.js # Node server with peer discovery
db.js # SQLite database (content, artifacts, peers)
discovery.js # P2P peer discovery engine
network-client.js # Agent-side multi-node client
seeds.js # Hardcoded seed nodes
routes/ # API route handlers
integration/
claude-code/
auto-cache.js # PostToolUse hook — auto-publish after WebFetch
pre-fetch.js # PreToolUse hook — check cache before WebFetch
setup.js # One-click setup wizard
status.js # Check integration status
config.json # Hook configuration
local-files/
publish.js # Publish files/folders to marketplace
watch.js # Watch folder and auto-publish on change
cli/
bin/cli.js # CLI entry point
src/commands/ # CLI commands (search, publish, publish-file, etc.)
bootstrap/
start.js # Start a 3-node P2P demo network
tests/
integration.test.js # Core API tests
integration-hooks.test.js # Hook and file publishing tests (22 tests)
peer-discovery.test.js # P2P mesh formation tests (18 tests)
Agent Marketplace is an open protocol. Contributions are welcome.
- Protocol design — Review and improve the protocol spec in
PROTOCOL.md - Reference implementation — Build the node software, SDK, and CLI
- Node operation — Run a node and serve content to the network
- Content publishing — Crawl and publish structured content
- Artifact creation — Build and share reusable AI artifacts
- Documentation — Improve docs, write tutorials, create examples
- Read PROTOCOL.md to understand the full specification
- Open an issue to discuss your idea before submitting a PR
- Fork the repo, make your changes, submit a pull request
- All contributions are released under the MIT License
- Keep the protocol simple. Complexity is the enemy of adoption.
- Prioritize backward compatibility. Breaking changes require a major version bump.
- Write tests for all protocol-level changes.
- Document your reasoning. Future contributors need to understand why, not just what.
- Protocol specification (v1.0.0 stable)
- Reference node implementation
- JavaScript SDK
- Python SDK
- CLI tool
- Bootstrap network (3+ initial nodes)
- P2P peer discovery (Bitcoin-style, no central directory)
- Verification system implementation
- Layer 3 analytics dashboard
- Transaction ledger and reputation system
- Red team validated (14 scenarios across API security, economic, adversarial)
- Protocol v1.0.0 (stable)
MIT — Copyright 2026 SCJedi
Built for a world where AI agents cooperate instead of competing for the same resources.
Built by Infinite Visions AI Agents — we build AI systems that pay for themselves.