Skip to content

Latest commit

 

History

History
491 lines (386 loc) · 24.6 KB

File metadata and controls

491 lines (386 loc) · 24.6 KB

Create Context Graph

Neo4j Labs Docs

Neo4j Labs Project — This project is part of Neo4j Labs. It is maintained by Neo4j staff and the community, but not officially supported. For help, use GitHub Issues or the Neo4j Community Forum.

Interactive CLI scaffolding tool that generates fully-functional, domain-specific context graph applications. Pick your industry domain, pick your agent framework, and get a complete full-stack app in under 5 minutes.

Generated app: chat interface, graph visualization, and document browser

# Python
uvx create-context-graph

# Node.js
npx create-context-graph

# Non-interactive (PROJECT_NAME is optional — auto-generates slug from domain+framework)
uvx create-context-graph --domain healthcare --framework pydanticai --demo-data

What It Does

Create Context Graph walks you through an interactive wizard and generates a complete project:

  • FastAPI backend with an AI agent configured for your domain, powered by neo4j-agent-memory v0.1.0 for multi-turn conversations with automatic entity extraction and preference detection
  • Next.js + Chakra UI v3 frontend with streaming chat (Server-Sent Events), real-time tool call visualization (Timeline with live spinners), interactive graph visualization (schema view, double-click expand, drag/zoom, property panel), entity detail panel, document browser, and decision trace viewer
  • Neo4j schema with domain-specific constraints, indexes, and GDS projections
  • Rich demo data — LLM-generated entities, relationships, professional documents (discharge summaries, trade confirmations, lab reports), and multi-step decision traces
  • SaaS data import — connect GitHub, Slack, Gmail, Jira, Notion, Google Calendar, or Salesforce
  • Custom domains — describe your domain in plain English and the LLM generates a complete ontology
  • Domain-specific agent tools with Cypher queries tailored to your industry
  • MCP server for Claude Desktop — optionally generates an MCP server config so Claude Desktop can query the same knowledge graph (--with-mcp)
  Creating context graph application...

  Domain:     Wildlife Management
  Framework:  PydanticAI
  Data:       Demo (synthetic)
  Neo4j:      Docker (neo4j://localhost:7687)

  [1/6] Generating domain ontology...          ✓
  [2/6] Creating project scaffold...           ✓
  [3/6] Configuring agent tools & system prompt...  ✓
  [4/6] Generating synthetic documents (25 docs)... ✓
  [5/6] Writing fixture data...                ✓
  [6/6] Bundling project...                    ✓

  Done! Your context graph app is ready.

  cd my-app
  make install && make start

Quick Start

Prerequisites

  • Python 3.11+ (with uv recommended)
  • Node.js 18+ (for the frontend)
  • Neo4j 5+ (Docker, Aura, or local install)

1. Create a project

uvx create-context-graph

The interactive wizard will guide you through selecting a domain, framework, and Neo4j connection.

Or skip the wizard with flags:

uvx create-context-graph my-app \
  --domain financial-services \
  --framework pydanticai \
  --demo-data

# Custom domain from description
uvx create-context-graph my-app \
  --custom-domain "veterinary clinic management" \
  --framework pydanticai \
  --anthropic-api-key $ANTHROPIC_API_KEY \
  --demo-data

# Import from SaaS services
uvx create-context-graph my-app \
  --domain personal-knowledge \
  --framework pydanticai \
  --connector github \
  --connector slack

# With MCP server for Claude Desktop
uvx create-context-graph my-app \
  --domain healthcare \
  --framework pydanticai \
  --demo-data \
  --with-mcp

2. Start the app

The wizard offers four Neo4j connection options:

Option Command Description
Neo4j Aura (cloud) (no start needed) Free cloud database — import your .env from console.neo4j.io
neo4j-local make neo4j-start Lightweight local Neo4j, no Docker required (needs Node.js)
Docker make docker-up Full Neo4j via Docker Compose
Existing (no start needed) Connect to any running Neo4j instance
cd my-app
make install       # Install backend + frontend dependencies
make neo4j-start   # Start Neo4j (if using neo4j-local)
# OR: make docker-up  # Start Neo4j (if using Docker)
make seed          # Seed sample data into Neo4j
make start         # Start backend (port 8000) + frontend (port 3000)

3. Explore

Supported Domains

22 industry domains, each with a purpose-built ontology, sample data, agent tools, and demo scenarios:

Domain Key Entities Domain Key Entities
Financial Services Account, Transaction, Decision, Policy Real Estate Property, Listing, Agent, Inspection
Healthcare Patient, Provider, Diagnosis, Treatment Vacation & Hospitality Resort, Booking, Guest, Activity
Retail & E-Commerce Customer, Product, Order, Review Oil & Gas Well, Reservoir, Equipment, Permit
Manufacturing Machine, Part, WorkOrder, Supplier Data Journalism Source, Story, Claim, Investigation
Scientific Research Researcher, Paper, Dataset, Grant Trip Planning Destination, Hotel, Activity, Itinerary
GenAI / LLM Ops Model, Experiment, Prompt, Evaluation GIS & Cartography Feature, Layer, Survey, Boundary
Agent Memory Agent, Conversation, Memory, ToolCall Wildlife Management Species, Sighting, Habitat, Camera
Gaming Player, Character, Quest, Guild Conservation Site, Species, Program, Funding
Personal Knowledge Note, Contact, Project, Topic Golf & Sports Mgmt Course, Player, Round, Tournament
Digital Twin Asset, Sensor, Reading, Alert Software Engineering Repository, Issue, PR, Deployment
Product Management Feature, Epic, UserPersona, Metric Hospitality Hotel, Room, Reservation, Service
# List all available domains
create-context-graph --list-domains

Custom domains: Don't see your industry? Select "Custom (describe your domain)" in the wizard or use --custom-domain "your description". The LLM generates a complete ontology with entity types, relationships, agent tools, and more.

SaaS Data Connectors

SaaS connector data flow into Neo4j knowledge graph

Import real data from your existing tools instead of (or in addition to) synthetic demo data:

Service What's Imported Auth
GitHub Issues, PRs, commits, contributors Personal access token
Notion Pages, databases, users Integration token
Jira Issues, sprints, users API token
Slack Channel messages, threads, users Bot OAuth token
Gmail Emails (last 30 days) Google Workspace CLI or OAuth2
Google Calendar Events, attendees (last 90 days) Google Workspace CLI or OAuth2
Salesforce Accounts, contacts, opportunities Username/password
Linear Issues, projects, cycles, teams, users, labels, comments, milestones, initiatives, attachments + decision traces from history Personal API key
Google Workspace Drive files, comment threads (as decision traces), revisions, Drive Activity, Calendar events, Gmail metadata Google OAuth 2.0
Claude Code Session history, messages, tool calls, files, decisions, preferences, errors None (local files)
Claude AI Conversations, messages, tool calls, thinking traces from Claude AI web/app export None (local file)
ChatGPT Conversations, messages, tool results from ChatGPT data export None (local file)

The Google Workspace connector extracts resolved comment threads from Google Docs as first-class decision traces — capturing the question, deliberation, resolution, and participants. Combined with Linear, it provides the full decision lifecycle: from meeting discussion to code execution.

The Claude Code connector reads your local session history from ~/.claude/projects/ — no API keys needed. It extracts decision traces from user corrections and error-resolution cycles, identifies developer preferences from explicit statements and behavioral patterns, and automatically redacts secrets before storage.

The Claude AI and ChatGPT connectors import your conversation exports directly from the official data export features. Export your data from Settings, pass the .zip file to the CLI with --import-type claude-ai or --import-type chatgpt, and get a fully populated context graph from your real conversations — no API keys needed.

Connectors run at scaffold time to populate initial data. They're also generated into your project so you can re-import with make import:

cd my-app
make import            # Re-import from connected services
make import-and-seed   # Import and seed into Neo4j

Agent Frameworks

Select your preferred agent framework at project creation time:

Framework Description
PydanticAI Structured tool definitions with Pydantic models and RunContext
Claude Agent SDK Anthropic tool-use with agentic loop
OpenAI Agents SDK @function_tool decorators with Runner.run()
LangGraph Stateful graph-based agent workflow with create_react_agent()
CrewAI Multi-agent crew with role-based tools
Strands Tool-use agents with Anthropic model
Google ADK Gemini agents with FunctionTool calling
Anthropic Tools Modular tool registry with Anthropic API agentic loop

All frameworks share the same FastAPI HTTP layer, Neo4j client, and frontend. Only the agent implementation differs. "Full streaming" means token-by-token text + real-time tool calls. "Tool streaming" means real-time tool calls with text delivered at the end.

Note: Conversation memory uses local sentence-transformers embeddings by default — no OPENAI_API_KEY required. If you set OPENAI_API_KEY in your .env, it will automatically upgrade to OpenAI embeddings.

Generated Project Structure

From domain YAML through Jinja2 templates to generated code

A single domain YAML drives the entire generated application — schema, models, agent tools, and visualization — through the Jinja2 template engine.

my-app/
├── backend/
│   ├── app/
│   │   ├── main.py                # FastAPI application
│   │   ├── agent.py               # AI agent (framework-specific)
│   │   ├── config.py              # Settings from .env
│   │   ├── routes.py              # REST API endpoints
│   │   ├── models.py              # Pydantic models (from ontology)
│   │   ├── context_graph_client.py # Neo4j CRUD operations
│   │   ├── memory.py              # Memory integration (MemoryIntegration)
│   │   ├── gds_client.py          # Graph Data Science algorithms
│   │   ├── vector_client.py       # Vector search
│   │   └── connectors/            # SaaS connectors (if selected)
│   ├── scripts/
│   │   ├── generate_data.py       # Data seeding script
│   │   └── import_data.py         # SaaS import script (if connectors selected)
│   └── pyproject.toml
├── frontend/
│   ├── app/                       # Next.js pages
│   ├── components/
│   │   ├── ChatInterface.tsx      # Streaming AI chat (SSE) with real-time tool calls + graph data flow
│   │   ├── ContextGraphView.tsx   # Interactive NVL graph (schema view, expand, drag/zoom, properties)
│   │   ├── DecisionTracePanel.tsx  # Reasoning trace viewer with step details
│   │   ├── DocumentBrowser.tsx    # Document browser with template filtering
│   │   └── Provider.tsx           # Chakra UI v3 provider
│   ├── lib/config.ts              # Domain configuration
│   ├── theme/index.ts             # Chakra theme with domain colors
│   └── package.json
├── cypher/
│   ├── schema.cypher              # Constraints & indexes
│   └── gds_projections.cypher     # GDS algorithm config
├── data/
│   ├── ontology.yaml              # Domain ontology definition
│   └── fixtures.json              # Pre-generated sample data
├── .env                           # Neo4j + API key configuration
├── .env.example                   # Configuration template (tracked in git)
├── .dockerignore                  # Docker build context exclusions
├── docker-compose.yml             # Local Neo4j instance (Docker mode only)
├── Makefile                       # start, seed, reset, install, test, test-connection, lint
├── mcp/                           # MCP server config (only if --with-mcp)
└── README.md                      # Domain-specific documentation (with framework docs + troubleshooting)

CLI Reference

create-context-graph [PROJECT_NAME] [OPTIONS]

Arguments:
  PROJECT_NAME              Project name (optional — auto-generated from domain+framework if omitted)

Options:
  --domain TEXT             Domain ID (e.g., healthcare, gaming)
  --framework TEXT          Agent framework (pydanticai, claude-agent-sdk, openai-agents, langgraph, crewai, strands, google-adk, anthropic-tools)
  --demo-data               Generate synthetic demo data
  --custom-domain TEXT      Generate custom domain from description (requires --anthropic-api-key)
  --connector TEXT          SaaS connector to enable; repeatable (github, slack, jira, notion, gmail, gcal, salesforce, linear, google-workspace, claude-code, claude-ai, chatgpt)
  --import-type TEXT        Chat history import: claude-ai or chatgpt (requires --import-file)
  --import-file PATH        Path to chat export file (.zip, .json, .jsonl)
  --import-depth TEXT       Import extraction depth: fast (default) or deep
  --import-filter-after TEXT Only import conversations after this date (ISO 8601)
  --import-filter-before TEXT Only import conversations before this date (ISO 8601)
  --import-filter-title TEXT Only import conversations matching title pattern (regex)
  --import-max-conversations INT Max conversations to import, 0=all (default: 0)
  --linear-api-key TEXT    Linear API key (required for --connector linear) [env: LINEAR_API_KEY]
  --linear-team TEXT       Linear team key to filter import (e.g., ENG) [env: LINEAR_TEAM]
  --gws-folder-id TEXT     Google Drive folder ID to scope import [env: GWS_FOLDER_ID]
  --gws-include-comments / --gws-no-comments  Import comment threads (default: on)
  --gws-include-revisions / --gws-no-revisions  Import revision history (default: on)
  --gws-include-activity / --gws-no-activity  Import Drive Activity (default: on)
  --gws-include-calendar   Import Calendar events (default: off)
  --gws-include-gmail      Import Gmail thread metadata (default: off)
  --gws-since TEXT         Import data since date (ISO format, default: 90 days ago)
  --gws-mime-types TEXT    MIME types to include (default: docs,sheets,slides)
  --gws-max-files INT      Maximum files to import (default: 500)
  --claude-code-scope TEXT Import current project or all (default: current)
  --claude-code-project TEXT Explicit project path to import sessions for
  --claude-code-since TEXT Import sessions since date (ISO format)
  --claude-code-max-sessions INT Max sessions to import, 0=all (default: 0)
  --claude-code-content TEXT Content mode: truncated, full, none (default: truncated)
  --with-mcp                Generate MCP server configuration for Claude Desktop
  --mcp-profile TEXT        MCP tool profile: core (6 tools) or extended (16 tools, default)
  --session-strategy TEXT   Memory session strategy: per_conversation (default), per_day, persistent
  --auto-extract/--no-auto-extract  Auto-extract entities from messages (default: on)
  --auto-preferences/--no-auto-preferences  Auto-detect user preferences (default: on)
  --ingest                  Ingest data into Neo4j after generation
  --neo4j-uri TEXT          Neo4j connection URI [env: NEO4J_URI]
  --neo4j-username TEXT     Neo4j username [env: NEO4J_USERNAME]
  --neo4j-password TEXT     Neo4j password [env: NEO4J_PASSWORD]
  --neo4j-aura-env PATH    Path to Neo4j Aura .env file with credentials
  --neo4j-local             Use @johnymontana/neo4j-local for local Neo4j (no Docker)
  --anthropic-api-key TEXT  Anthropic API key for LLM generation [env: ANTHROPIC_API_KEY]
  --openai-api-key TEXT    OpenAI API key for LLM generation [env: OPENAI_API_KEY]
  --google-api-key TEXT    Google/Gemini API key (required for google-adk) [env: GOOGLE_API_KEY]
  --output-dir PATH         Output directory (default: ./<project-name>)
  --demo                    Shortcut for --reset-database --demo-data --ingest
  --reset-database          Clear all Neo4j data before ingesting
  --dry-run                 Preview what would be generated without creating files
  --verbose                 Enable verbose debug output
  --list-domains            List available domains and exit
  --version                 Show version and exit
  --help                    Show help and exit

Context Graph Architecture

Architecture: generation pipeline and runtime components

Every generated app demonstrates the three-memory-type architecture from neo4j-agent-memory:

Three memory types: short-term, long-term, and reasoning memory in Neo4j

  • Short-term memory — Conversation history and document content stored as messages
  • Long-term memory — Entity knowledge graph built on the POLE+O model (Person, Organization, Location, Event, Object)
  • Reasoning memory — Decision traces with full provenance: thought chains, tool calls, causal relationships

This is what makes context graphs different from simple RAG — the agent doesn't just retrieve text, it reasons over a structured knowledge graph with full decision traceability.

With --with-mcp, the generated project also includes an MCP server configuration that connects Claude Desktop to the same knowledge graph. This dual-interface architecture means the web app and Claude Desktop share one context graph — entities, conversations, and reasoning traces are available everywhere.

Development

# Clone and install
git clone https://github.com/neo4j-labs/create-context-graph.git
cd create-context-graph
uv venv && uv pip install -e ".[dev]"

# Run tests (no Neo4j or API keys required)
source .venv/bin/activate
pytest tests/ -v               # Fast: 1,053 tests
pytest tests/ -v --slow        # Full: 1,259 tests (includes domain x framework matrix + perf + generated project tests)
pytest tests/ --integration    # Integration tests (requires running Neo4j)

# Test a specific scaffold
create-context-graph /tmp/test-app --domain software-engineering --framework pydanticai --demo-data

Makefile Targets

Target Description Requirements
make test Run fast unit tests (1,053 tests) None
make test-slow Full suite including matrix + perf + generated project tests (1,259 tests) None
make test-matrix Domain × framework matrix only (176 combos) None
make test-coverage Tests with HTML coverage report None
make smoke-test E2E smoke tests for 3 key frameworks Neo4j + LLM API keys
make lint Run ruff linter ruff
make scaffold Scaffold a test project to /tmp/test-scaffold None
make build Build Python package (sdist + wheel) None
make docs Start Docusaurus dev server Node.js

E2E Smoke Tests

The smoke tests scaffold a real project, install dependencies, start the backend, and send chat prompts to verify the full pipeline works end-to-end. They test the 3 frameworks that had critical bug fixes in v0.5.1:

# Run all 3 smoke tests (requires Neo4j + at least one LLM API key)
make smoke-test

# Or run individual framework tests directly
python scripts/e2e_smoke_test.py --domain financial-services --framework pydanticai --quick
python scripts/e2e_smoke_test.py --domain real-estate --framework google-adk --quick
python scripts/e2e_smoke_test.py --domain trip-planning --framework strands --quick

# Test all 22 domains with one framework
python scripts/e2e_smoke_test.py --all-domains --framework pydanticai --quick

# Full mode (all prompts per scenario, not just first)
python scripts/e2e_smoke_test.py --domain healthcare --framework claude-agent-sdk

Required environment variables:

  • NEO4J_URI, NEO4J_USERNAME, NEO4J_PASSWORD — Neo4j connection (Aura, Docker, or local)
  • ANTHROPIC_API_KEY — for Claude-based frameworks (PydanticAI, Claude Agent SDK, Anthropic Tools, Strands, CrewAI)
  • OPENAI_API_KEY — for OpenAI-based frameworks (OpenAI Agents, LangGraph)
  • GOOGLE_API_KEY — for Google ADK (Gemini)

CI Pipeline

GitHub Actions (.github/workflows/ci.yml) runs automatically:

Job Trigger Description
test All pushes + PRs Unit tests on Python 3.11 and 3.12 (1,053 tests including security, doc snippets, frontend logic)
lint All pushes + PRs Ruff linter on src/ and tests/
matrix Push to main only Full suite + 176 domain × framework matrix + perf + generated project tests (1,259 tests)
smoke-test Push to main only Neo4j integration tests + E2E for all 8 frameworks (scaffold → install → start → chat)

The smoke-test CI job is gated behind a SMOKE_TESTS_ENABLED repository variable. To enable it:

  1. Go to Settings → Variables → Repository variables and add SMOKE_TESTS_ENABLED = true
  2. Go to Settings → Secrets → Repository secrets and add:
    • NEO4J_URI — e.g., neo4j+s://xxxxx.databases.neo4j.io
    • NEO4J_USERNAME
    • NEO4J_PASSWORD
    • ANTHROPIC_API_KEY
    • OPENAI_API_KEY
    • GOOGLE_API_KEY

The smoke-test job uses fail-fast: false so one framework failure doesn't block the others, and it only runs after the unit test job passes.

Publishing

PyPI (Python)

# Build
uv build

# Publish (requires PyPI account + API token)
uv publish
# Or: twine upload dist/*

After publishing, users can install with:

uvx create-context-graph       # Ephemeral (recommended)
pip install create-context-graph   # Permanent install

npm (Node.js wrapper)

cd npm-wrapper

# Publish (requires npm account + auth)
npm publish --access public

After publishing, users can run with:

npx create-context-graph

The npm package is a thin wrapper that delegates to the Python CLI via uvx, pipx, or python3 -m. It requires Python 3.11+ to be installed.

Automated Publishing (GitHub Actions)

Both packages are published automatically when you push a version tag:

# 1. Update version in pyproject.toml and npm-wrapper/package.json
# 2. Commit the version bump
# 3. Tag and push
git tag v0.1.0
git push origin v0.1.0

This triggers two GitHub Actions workflows:

  • publish-pypi.yml — Builds and publishes to PyPI (uses trusted publishing / OIDC)
  • publish-npm.yml — Publishes the npm wrapper to npmjs.com

Setup required:

  • PyPI: Configure trusted publishing for this repo, or set a PYPI_API_TOKEN secret
  • npm: Set an NPM_TOKEN secret in the repository settings

Version Bumping

Both packages must use the same version. Update in two places:

  1. pyproject.tomlversion = "X.Y.Z"
  2. npm-wrapper/package.json"version": "X.Y.Z"

License

Apache-2.0