Skip to content

viplee110/econ-theorist-ai

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

31 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Econ Theorist AI

An AI system for economic theorists, from idea discovery to theory paper development.

This is an English-first research workflow system with multilingual command understanding. You can talk to the assistant in Chinese or English, but research artifacts, referee reports, theorem notes, and manuscripts are English by default.

Originally developed with Econometrica-level theory standards in mind, this unofficial local workflow system helps researchers move from rough ideas to primitive hunting, model tournaments, theorem candidates, absorption tests, target journal calibration, human gates, simulated review, and controlled revision paths.

This project is not affiliated with Econometrica. The phrase Econometrica-level is used only as shorthand for high standards of theoretical clarity, novelty, and rigor.

The workflow is target-calibrated rather than Econometrica-only. A confirmed target_journal_profile.md can calibrate the reader path, referee mix, exposition style, and fit standard for RAND, JET, Theoretical Economics, GEB, ReStud, AER, or another venue. Target journal changes calibration, not quality.

If this workflow helps your research, please consider giving the repository a star so more researchers can find and improve their own research process.

Workflow Map

flowchart LR
    A["Topic or rough idea"] --> B["Literature probe<br/>field_profile.md"]
    B --> T["Target profile<br/>target_journal_profile.md"]
    T --> C["Primitive Hunter<br/>Exploration Quota"]
    C --> D["Model tournament"]
    D --> E["First-pass derivation"]
    E --> F["Absorption test<br/>main theorem gate"]
    F --> G["5-8 page model note"]
    G --> H["Manuscript development"]
    H --> S["Style calibration<br/>style_calibration.md"]
    S --> I["Target-calibrated<br/>simulated review"]
    S -->|structural weakness| C
    I --> J["Revision tree"]
    J --> K{"Human decision"}
    K -->|preserve or revise| H
    K -->|simplify or pivot| C
    K -->|submit or retarget| L["Final readiness"]
Loading

Control Layer

flowchart TD
    AC["active_context.md<br/>compact dashboard"] --> OR["ECONOMETRICA_ORCHESTRATOR.md<br/>router"]
    HD["human_decisions.md<br/>append-only gates"] --> OR
    FP["field_profile.md<br/>confirmed field and referee roles"] --> OR
    TJ["target_journal_profile.md<br/>target ladder and quality floor"] --> OR
    GL["generality_ledger.md<br/>special-case and assumption drift"] --> OR
    OR --> WG["Workflow modules"]
    WG --> HP{"Human gate"}
    HP --> HD
Loading

active_context.md is only a dashboard, not a source of truth. Important claims must still be checked against the underlying workflow artifacts.

What This System Does

  • Searches for research directions without immediately forcing mainstream taste.
  • Preserves 1-2 non-mainstream but internally coherent directions during discovery.
  • Runs Primitive Hunter / Theorem Generator panels when the primitive is unclear.
  • Compares model variants before manuscript writing.
  • Requires absorption tests against closest literature and known theory families.
  • Uses field_profile.md to assign field-sensitive simulated referees.
  • Uses target_journal_profile.md to recommend and confirm a primary, stretch, and fallback target without lowering the quality floor.
  • Records human gate decisions in human_decisions.md instead of relying on chat memory.
  • Tracks assumption and generality drift in generality_ledger.md.
  • Uses Scientific Judge / Nugget Test safeguards against defensive complexity.
  • Runs target-calibrated simulated review with dynamic referee roles.
  • Routes local-optimum traps back to discovery before manuscript polishing.
  • Calibrates exposition style with elegance without rhetoric after the contribution is locked.
  • Supports Python, Mathematica, Lean, LaTeX, and git-based verification workflows.

Start In 5 Minutes

  1. Download the repository as a ZIP file or clone it.
  2. Copy these files into the root directory of your paper project:
AGENTS.md
ECONOMETRICA_ORCHESTRATOR.md
ECONOMETRICA_PANEL_PROTOCOL.md
ECONOMETRICA_DISCOVERY_WORKFLOW.md
ECONOMETRICA_VERIFICATION_WORKFLOW.md
ECONOMETRICA_AI_HUMAN_WORKFLOW.md
ECONOMETRICA_VERSION_CONTROL.md
FIRST_RUN.md
TOOLCHAIN_README.md
verify_toolchain.ps1
verification_templates/
  1. Open the paper folder in Codex Desktop or another agent IDE. AGENTS.md should be read automatically, and ECONOMETRICA_ORCHESTRATOR.md acts as the router.

  2. Run a first-run setup check:

Use the system: first-run setup check

Small Chinese command example:

按系统处理:初始化检测
  1. Initialize the paper project:
Use the system: initialize this paper project

Common Commands

Use the system: first-run setup check.
Use the system: initialize this paper project.
Use the system: I want to explore a new research topic.
Use the system: run a model tournament and absorption test before writing.
Use the system: run Primitive Hunter and identify the deepest primitive.
Use the system: rigorously verify Proposition 1 with Python, Mathematica, or Lean if useful.
Use the system: recommend a target journal ladder for this project.
Use the system: run a full target-calibrated simulated review.
Use the system: revise with agentic tree search instead of a defensive patch.
按系统继续。
按系统处理:这个项目是不是陷入局部最优?

Files

File Purpose
AGENTS.md Project-level instructions for Codex and compatible agents.
ECONOMETRICA_ORCHESTRATOR.md Natural-language router for workflow modules and stages.
ECONOMETRICA_DISCOVERY_WORKFLOW.md Topic discovery, primitive hunting, model tournaments, and theorem gates.
ECONOMETRICA_PANEL_PROTOCOL.md Independent panels, dynamic referee assignment, AE synthesis, and Co-Editor decisions.
ECONOMETRICA_AI_HUMAN_WORKFLOW.md Manuscript development, simulated review, revision trees, and human gates.
ECONOMETRICA_VERIFICATION_WORKFLOW.md Mathematical derivation, counterexample search, symbolic checks, and formal verification.
FIRST_RUN.md First-run setup guide for non-technical users.
ECONOMETRICA_VERSION_CONTROL.md Git checkpoints, branches, diffs, rollback safety, and version logs.
TOOLCHAIN_README.md Local Python, Lean, Mathematica, and verification setup.
verify_toolchain.ps1 Quick local toolchain self-test and status writer.
verification_templates/ Starter templates for counterexample search and Lean lemmas.

Runtime Artifacts

These files are created inside paper projects, not maintained as fixed files in this workflow repository:

Artifact Purpose
active_context.md 80-120 line continuation dashboard.
human_decisions.md Append-only human gate decisions, reversals, and reasons.
field_profile.md Confirmed or provisional field profile for literature and referee routing.
target_journal_profile.md Confirmed or provisional target ladder, fit standard, quality floor, and reader calibration.
generality_ledger.md Record of special-case moves, assumptions, and theorem-sentence drift.
style_calibration.md Confirmed or provisional guide for elegant, field-calibrated exposition without rhetoric.
toolchain_status.md Computer-level diagnostic status, usually stored globally outside the paper project.
model_tournament.md Comparison of model variants and documented winners/losers.
absorption_tests.md Tests for whether the result is absorbed by existing theory.
referee_reports/round_N/ Simulated referee, AE, Co-Editor, and summary reports.

Toolchain

Keep Python, Lean/elan, Mathlib, and large package caches outside paper folders. On Windows the recommended shared tool root is:

C:\Tools\CodexVerification

You can choose another location by setting CODEX_VERIFICATION_HOME or by passing -ToolRoot to verify_toolchain.ps1.

Minimum setup check:

.\verify_toolchain.ps1 -WriteStatus

If the script cannot find Python, Lean, or Mathematica, the workflow still works as prompts and checklists, but mathematical verification is weaker until the tool root is configured. See TOOLCHAIN_README.md for details.

If You Do Not Have Python, Lean, Or Mathematica

You can still use the research workflow. Missing tools only weaken mathematical verification, local compilation, or formal proof support. They do not block idea discovery, field profiling, target journal profiling, model tournaments, simulated review, style calibration, or revision planning.

Language Policy

The workflow files, project artifacts, referee reports, theorem notes, revision logs, and manuscripts are English by default.

Chinese commands are supported for convenience, but generated research files remain English unless the user explicitly requests a separate Chinese explanatory note outside the manuscript workflow.

Design Principles

  • Scientific taste is a filter, not the sole objective.
  • Target journal changes calibration, not quality.
  • Token economy must never override research quality.
  • Main theorem discovery, proof verification, closest-literature checks, and simulated review require enough context and tools.
  • Human gate decisions must be written to persistent artifacts.
  • Simulated acceptance is a diagnostic benchmark, not a publication guarantee.
  • If the paper is stuck in local polishing, return to primitive hunting and model discovery.
  • The preferred path is theorem note first, manuscript second.

Feedback and Issues

This project currently uses an issue-only feedback model.

Please open a GitHub Issue if you find:

  • unclear workflow routing
  • internal contradictions
  • missing safeguards
  • confusing documentation
  • toolchain setup problems
  • examples where the system behaves poorly
  • suggestions that may help researchers use the system better

Pull requests are not the preferred contribution path at this stage. The maintainer reviews issues, decides which suggestions to adopt, and integrates accepted changes directly to preserve consistency across the workflow system.

See CONTRIBUTING.md for the current contribution policy.

License

See LICENSE.