Installation · Quick Start · Commands · How It Works
ASCI evaluates scientific research repositories using specialized AI agents that assess hypothesis clarity, evidence quality, code implementation, and novelty. Each agent scores independently, producing a detailed report with actionable feedback.
pip install asciRequires Python 3.11+ and an Anthropic API key:
export ANTHROPIC_API_KEY=sk-ant-...Evaluate a repository:
asci evaluate ./my-researchScaffold a new research repo:
asci new ./my-projectSubmit to the journal:
asci submit --repo ./my-researchRun a full evaluation on a research repository.
| Option | Description |
|---|---|
--model, -m |
Claude model to use (default: claude-sonnet-4-20250514) |
--json |
Output as JSON to stdout |
--output, -o |
Write JSON report to file |
Scaffold a new research repository with all required files:
README.md— structured templatehypothesis.md— hypothesis format guidesrc/main.py— starter codedata/— data directory.gitignore— Python defaults
Validate, evaluate, and submit a repository to the journal.
| Option | Description |
|---|---|
--repo |
Local path or GitHub URL (required) |
--model, -m |
Claude model to use |
--min-score |
Minimum score threshold (default: 4.0) |
Repositories must include:
README.md- A hypothesis doc (
hypothesis.mdorabstract.md) - At least one code file (
.py,.r,.jl,.m,.cpp,.c,.java,.rs,.go) - A
data/directory
ASCI runs four specialized evaluation agents concurrently:
| Agent | What it evaluates |
|---|---|
| Hypothesis | Clarity, testability, specificity, falsifiability |
| Evidence | Data sufficiency, relevance, statistical rigor |
| Implementation | Code quality, reproducibility, correctness |
| Novelty | Originality, significance, prior work awareness |
Each agent uses Claude's tool-use API to produce structured scores (1–10) with reasoning. Results are aggregated into an overall score and detailed report.
from asci.skill import evaluate_repo
report = evaluate_repo("./my-research")
print(report.overall_score)MIT
