Problem/Opportunity
When developers evaluate an AI agent framework, they often look at stars and recent activity. But community composition tells a deeper story about long-term sustainability:
- Company concentration risk: Is 80% of contributions coming from one company? What happens if they pivot?
- Geographic distribution: Is the community concentrated in one timezone, or globally distributed with 24/7 coverage?
- New contributor retention: Does the community successfully onboard and retain new contributors, or is it a closed circle of veterans?
- Contribution velocity distribution: Are contributions evenly distributed, or does the project rely on 2-3 heroes?
Currently, there's no systematic way to quantify these signals across the AI agent ecosystem.
Implementation Plan
Phase 1: Data Collection
- For each tracked AI agent framework repo, collect:
- All contributors (last 12 months)
- Commit counts per contributor
- PR counts per contributor
- Issue comments per contributor
- Contributor company affiliation (from GitHub profile/company field)
- Contributor location (from GitHub profile/location field, normalized to country/region)
- First contribution date (to calculate tenure)
Phase 2: Metric Calculation
Calculate the following scores:
-
Company Concentration Index (0-100)
- Herfindahl-Hirschman Index (HHI) of contributions by company
- 100 = perfectly distributed, 0 = single-company monopoly
-
Geographic Diversity Score (0-100)
- Number of countries represented weighted by contribution share
- Bonus for multi-continent coverage
-
New Contributor Retention Rate (%)
- % of first-time contributors who made a 2nd contribution within 90 days
- Benchmark against top 10% of projects
-
Contribution Distribution Gini (0-1)
- Gini coefficient of contribution counts
- Lower = more egalitarian, Higher = concentrated among few
-
Composite Community Health Score (0-100)
- Weighted average of above metrics
- Display as badge: 🟢 Healthy / 🟡 Moderate / 🔴 At-Risk
Phase 3: Visualization and Integration
- Add a Community Health tab to each AI framework analysis page
- Create a leaderboard: Most Sustainable AI Agent Communities
- Add filters to collections page: Show only frameworks with healthy community health
- Optional: Weekly digest highlighting Rising Stars (new contributors gaining traction)
Phase 4: API Exposure
- Expose metrics via OSSInsight API for programmatic access
- Enable MCP server queries like "Which agent frameworks have the healthiest communities?"
Why AI Builders Would Care
For Founders
- Benchmark their community health against competitors
- Identify red flags before adopting a framework (e.g., single-company risk)
- Use as a recruiting signal: Join our diverse, healthy community
For Enterprise Teams
- De-risk framework selection by avoiding projects with concentration risk
- Justify technology choices to leadership with data-driven community analysis
- Assess long-term viability beyond current star count
For Contributors
- Discover projects where their contributions will be valued (high new-contributor retention)
- Avoid projects dominated by corporate agendas (high company concentration)
Estimated Impact
Traffic
- 15-20% increase in time-on-page for framework analysis pages (new tab + deeper engagement)
- 10% increase in returning visitors (community health tracking becomes a habit)
Engagement
- High shareability: Our framework scored 87/100 on Community Health (Twitter/LinkedIn)
- Framework maintainers will actively promote their scores, driving referral traffic
Retention
- Stickiness: Teams will bookmark and monitor their favorite frameworks community health over time
- Differentiation: Positions OSSInsight as the go-to for sustainable open source intelligence, not just growth metrics
Strategic
- First-mover advantage: No other platform systematically tracks community composition for AI projects
- Aligns with enterprise procurement needs (vendor risk assessment)
- Complements existing metrics (stars, growth) with a sustainability dimension
Data Sources
- GitHub GraphQL API (contributor data, commit history)
- GitHub REST API (user profiles for company/location)
- Optional: LinkedIn/company domain enrichment for more accurate company mapping
Technical Considerations
- Cache contributor data (update weekly, not real-time)
- Handle missing company/location gracefully (show Unknown bucket)
- Privacy: Only use public GitHub profile data
- Performance: Pre-compute scores in nightly batch job, serve from cache
Problem/Opportunity
When developers evaluate an AI agent framework, they often look at stars and recent activity. But community composition tells a deeper story about long-term sustainability:
Currently, there's no systematic way to quantify these signals across the AI agent ecosystem.
Implementation Plan
Phase 1: Data Collection
Phase 2: Metric Calculation
Calculate the following scores:
Company Concentration Index (0-100)
Geographic Diversity Score (0-100)
New Contributor Retention Rate (%)
Contribution Distribution Gini (0-1)
Composite Community Health Score (0-100)
Phase 3: Visualization and Integration
Phase 4: API Exposure
Why AI Builders Would Care
For Founders
For Enterprise Teams
For Contributors
Estimated Impact
Traffic
Engagement
Retention
Strategic
Data Sources
Technical Considerations