Research moves faster when evidence is easy to trust.
Arxie is an AI research assistant that reads real papers (Semantic Scholar + arXiv), reasons across sources, and writes citation-grounded outputs you can verify.
- Current:
v0.1.0(released) - Next:
v0.2.0(in planning)
- Live paper retrieval and citation-grounded Q&A
- Deep search (multi-hop paper analysis)
- Full-text PDF parsing for methods/results-level reasoning
- Literature review generation (
ra lit-review) - Citation influence tracing (
ra trace) - Confidence annotations (supporting vs contradicting evidence)
- Conversational mode (
ra chat) - FastAPI + Docker support
- Dashboard-based proposal workspace (not terminal-first)
- Iterative research proposal co-creation workflow
- Visual artifacts (mindmap, evidence map, logical tree, method pipeline, outcome matrix)
- Cross-artifact sync when users revise hypotheses/assumptions
(See docs/PRE-PRD-v0.2.md for discussion draft.)
Most assistants stop at summaries. Arxie is designed for researchers who need a defensible reasoning trail:
- read full papers, not just abstracts
- compare methods and contradictions across papers
- keep citations tied to claims
- show confidence based on evidence landscape
git clone https://github.com/mmTheBest/arxie.git cd arxie python -m venv .venv source .venv/bin/activate pip install -e . export OPENAI_API_KEY="sk-..."# Ask a question ra query "What are recent approaches to long-context LLMs?" # Deeper multi-hop analysis ra query --deep "Compare LoRA vs QLoRA methodologies" # Literature review draft ra lit-review "attention mechanisms in computer vision" # Citation timeline ra trace "Attention Is All You Need" # Interactive session ra chatuvicorn ra.api.app:app --host 0.0.0.0 --port 8000curl -X POST http://localhost:8000/api/query \ -H "Content-Type: application/json" \ -d '{"query":"What are retrieval-augmented generation trade-offs?"}'docker build -t arxie . docker run -e OPENAI_API_KEY="sk-..." arxie ra query "Your question here"Using Arxie’s internal 100-question benchmark with GPT-4o-mini:
| Metric | Result |
|---|---|
| Citation precision | 86% |
| Claim support ratio | 100% |
| Tool success rate | 99.8% |
These are reported benchmark results, not a user quick-start workflow.
src/ra/ ├── agents/ # research, lit-review, chat behaviors ├── api/ # FastAPI app + request models ├── citation/ # citation formatting + confidence scoring ├── parsing/ # PDF parsing ├── retrieval/ # Semantic Scholar + arXiv + cache ├── tools/ # tool interfaces for the agent loop └── utils/ # config, logging, rate limiting MIT