Generated from the founding session, March 9 2026. Contributors: Nicholai, Oogie, Jake, MrClaude
A publicly accessible marketplace for the agent economy — skills, MCP servers, and knowledge bases. Open to browse. Signet-verified to sell.
The North Star: "Your agent's knowledge has value. Here's how to monetize it."
Existing marketplaces (ClawHub, skills.sh, npm) are directories. No accountability. No identity. No trust signal. Anyone can publish a credential stealer and call it a weather skill.
Signet Marketplace is the first marketplace where trust is structural, not hoped for. The verification model has three tiers — each adds a trust signal, none gate access to browsing.
We scrape and mirror every major source automatically. Your skill or MCP server appears on day one, whether or not you've ever heard of us. Creators can claim their listing via GitHub OAuth — proves repo ownership, grants edit access to description, links, and tags. Zero friction, no cost.
A listing becomes verified when the creator provides Entire.io session logs as the audit trail.
Entire hooks into your git workflow to capture AI agent sessions — every prompt/response transcript, files touched, and checkpoints — stored on a separate branch alongside commits. This is the provenance record.
Signet is the distillation layer: Entire captures what happened, Signet turns it into persistent, structured knowledge. Together they produce a verifiable claim: "this skill was built this way, here's the receipts."
The verified badge means:
- The creator has proven authorship (GitHub OAuth)
- The build process is auditable (Entire.io session chain)
- The identity is DID-anchored (did:signet:)
Entire.io is a foundational partner in this model, not a bolt-on. Their logo appears in the colophon.
Paid visibility. USDC staking on Base. Stake to surface a listing in featured/top tiers. Stakes expire, no lock-in. This is the revenue layer.
Skills and MCP servers are code. Knowledge bases are understanding.
Signet's distillation pipeline is powerful enough that any raw, unstructured data dump — internal docs, decades of company history, domain expertise, chat logs — can be ingested and turned into actionable agent knowledge. Not raw observations. Structured entities, mapped dependencies, atomic facts with context.
This means:
- A company can embed their entire institutional knowledge into a portable, agent-readable asset
- That asset can be listed, sold, and transferred on the marketplace
- The buyer gets a verified knowledge base — DID-signed, with Entire.io provenance if the creator provided it
- Companies could effectively auction off their accumulated expertise — knowledge bases as exit vehicles
This is the category nobody else has. Not code. Not tools. Verified agent knowledge.
- Frontend: Astro 5 + Tailwind CSS 4 — Cloudflare Pages
- API layer: Cloudflare Workers
- Database: Cloudflare D1 (SQLite) for listings, claims, verification
- Storage: R2 for mirrored content
- Cache: KV for hot listings and search
- Queue: Cloudflare Queues for async crawl jobs
- Search: Cloudflare Vectorize (semantic) or lightweight keyword index
listings (
id, slug, name, description,
type TEXT, -- 'skill' | 'mcp' | 'knowledge'
source_url TEXT,
source_repo TEXT,
install_count INTEGER DEFAULT 0,
created_at TIMESTAMP,
verification_status TEXT DEFAULT 'unclaimed' -- 'unclaimed' | 'claimed' | 'verified'
)
claims (
id, listing_id,
github_user TEXT,
github_oauth_token_hash TEXT,
claimed_at TIMESTAMP
)
verification (
id, listing_id,
did TEXT, -- did:signet:<pubkey>
did_document TEXT,
entire_session_ref TEXT, -- Entire.io checkpoint branch/commit ref
anchor_tx_hash TEXT, -- null until Phase 2 on-chain anchor
chain TEXT,
verified_at TIMESTAMP,
status TEXT -- 'pending' | 'verified' | 'revoked'
)
visibility (
id, listing_id,
stake_amount_usdc REAL,
stake_tx_hash TEXT,
expires_at TIMESTAMP,
tier TEXT -- 'basic' | 'featured' | 'top'
)
index_runs (
id, source, started_at, completed_at,
listings_added INTEGER, listings_updated INTEGER
)- Astro scaffold at
~/signet/marketplace/ - Indexer crawled 10,110 listings (5,451 skills + 4,659 MCP servers) —
public/data/listings.json - Wire
listings.jsoninto Astro pages for real data at build time - GitHub OAuth claim flow
- Deploy to
market.signetai.shvia Cloudflare Pages (already live, needs real data)
- Cherry-pick PR #25 Phase 0: Ed25519 keypairs +
did:signet:DID generation — pure crypto, no chain dependency -
signet did generate→ DID stored in D1 → verified badge on listing - Entire.io session log ingestion for Tier 2 verification
- Hold Phase 4A (ERC-8004 escrow) until external audit
- Basic USDC staking for Tier 3 on Base
- Knowledge base listing format + upload flow
- Distillation pipeline integration (Signet daemon → knowledge bundle export)
- DID-signed knowledge bundles with Entire.io provenance chain
- BroadcastRegistry + ReferralGraph contracts (agents earn for downstream discovery)
- WorldBlast integration
- Moltbook — MrClaude seeding trust chains → identity → marketplace narrative. Oracle post live. eudaemon_0 engagement underway.
- ClawHub skill — publish a Signet skill for native OpenClaw distribution
- awesome-openclaw lists — get into curated lists
- rentamac.io — target for write-up (ran "best skills" piece ~1 week ago)
- Entire.io — reach out when ready; joint positioning is natural and mutually beneficial
| Who | What | When |
|---|---|---|
| MrClaude | Wire listings.json into Astro pages | Today |
| Nicholai | Validate D1 schema | Today |
| Nicholai | Confirm PR #25 Phase 0 cherry-pick scope | Today |
| Jake | Review Phase 0 cherry-pick plan | This week |
| Oogie | Indexer dedup + additional sources | This week |
| All | Schedule external audit for ERC-8004 escrow | Before Phase 2 |
| Repo | Purpose | Status |
|---|---|---|
Signet-AI/signetai |
Core daemon, CLI, SDK | Active |
Signet-AI/marketplace |
This marketplace | Active |
NicholaiVogel/signet |
Private planning | Active |
Powered by Entire.io for verifiable session provenance.
Added March 9, 2026 — Nicholai
Signet is not a memory company.
The core value proposition: take any dataset in the world, no matter how messy or unstructured, and transform it into actionable, agent-understandable knowledge.
This reframes everything:
An importer and distillation engine. The pipeline already exists. What's missing is the surface — the button that lets anyone connect their data sources and watch them become structured knowledge.
Import targets needed:
- Gmail / Google Workspace (Docs, Sheets, Drive)
- Excel / CSV / any tabular data
- Notion, Obsidian, Roam
- Slack, Discord export
- PDFs, Markdown, plaintext
- GitHub repos (already in PR #25)
- Any data dump, any format
The distillation pipeline handles the rest. Messy → atomic facts → entities → knowledge graph.
A knowledge base is a pre-compiled dataset of knowledge — the output of running the distillation pipeline over a body of data. Packaged, portable, DID-signed. Ready to install into any Signet-powered agent.
This means:
- A law firm's 40 years of case strategy becomes a knowledge base
- A VC's pattern library becomes a knowledge base
- A founder's playbook becomes a knowledge base
- A company's entire institutional memory — the decisions, the reasoning, the pattern matching that took decades to build — becomes a knowledge base
When companies exit, they sell customer lists, equity, and IP.
The next category: sell your cognitive substrate. The pattern matching. The decision trees. The institutional memory that lives in the heads of people who will leave. The thing that makes your company smart, not just the assets it owns.
Signet packages that. The marketplace auctions it. The buyer gets a knowledge base they can install — and their agent inherits decades of understanding in a single import.
This is a new asset class.
The pipeline exists. The distillation engine works. What's needed:
- Import connectors — OAuth flows for Gmail, Google Workspace, Notion, Slack; file upload for CSV/Excel/PDF; GitHub repo ingestion (PR #25)
- Knowledge base packager — bundle the distillation output into a portable, signed, versioned artifact
- Marketplace listing format — describe the knowledge base (domain, source types, entity count, date range) without exposing the data
- Purchase + install flow — buyer pays, gets the bundle, installs into their Signet
The oracle is the product. The import layer is how you feed it.