Give your AI agent long-term memory and decision-making ability for feature planning.
Your AI agent is great at writing code. But ask it "what should we build next?" and it gives you generic suggestions β it doesn't remember last week's architecture decisions, doesn't know which features you already rejected, and can't connect the dots between user feedback and technical possibilities.
Feature Radar fixes this. It creates an external brain (.feature-radar/) that persists across sessions β so your agent accumulates knowledge instead of starting from zero every time.
Before Feature Radar:
You: "What should we build next?"
Agent: "You could add dark mode, improve performance, or add user authentication..." (generic suggestions that ignore your project's context)
With Feature Radar:
You: "feature radar"
Agent: "Based on
base.md, your project is a CLI tool focused on local-first workflow. I found 3 features you've already shipped that aren't documented, 2 open opportunities with high user demand, and a pattern fromarchive/suggesting your streaming approach needs rethinking. I recommend building the plugin system next β it has the highest impact-to-effort ratio and aligns with your architecture."
The difference? Your agent now has context that compounds over time.
The fastest way to inject your domain knowledge:
1. Say "feature radar" β agent generates opportunities 2. Open any file, add: > NOTE: this is too complex, split into two phases 3. Say "address my notes" 4. Agent reads your notes, applies corrections, removes markers 5. Repeat until perfect Your agent learns from your corrections. Architecture constraints, naming conventions, strategic priorities β all captured in files that persist forever.
With skillshare
skillshare install runkids/feature-radar --into feature-radarWith Vercel Skills CLI
npx skills add runkids/feature-radar# Claude Code cp -r skills/* ~/.claude/skills/ # Codex cp -r skills/* ~/.codex/skills/Then say "feature radar" in your next session. That's it.
You can also say "feature radar quick" for a fast scan, "feature radar evaluate" to jump to prioritization, or "feature radar #2" to deep-dive a specific opportunity.
Your agent analyzes your project β language, architecture, key features β and builds a structured knowledge base:
.feature-radar/ βββ base.md β Project dashboard: what you've built, where you're going βββ archive/ β Shipped, rejected, or covered features (with extracted learnings) βββ opportunities/ β Open features ranked by impact and effort βββ specs/ β Reusable patterns and architectural decisions βββ references/ β External inspiration, ecosystem trends, research The data format is defined in SPEC.md β a language-agnostic specification that any AI tool can implement.
Every feature goes through a lifecycle β discovered, evaluated, built, archived. And every archived feature feeds back into the system:
flowchart TD subgraph Discovery S1["scan"] --> OPP[opportunities/] S2["ref"] --> REF[references/] end subgraph Evaluation FR["feature-radar"] --> |Phase 1-3| CLASSIFY{Classify} CLASSIFY --> |Open| OPP CLASSIFY --> |Done/Rejected| ARC[archive/] CLASSIFY --> |Pattern| SPEC[specs/] CLASSIFY --> |External| REF FR --> |Phase 5-6| RANK[Rank & Propose] RANK --> BUILD["Enter plan mode"] end subgraph Completion DONE["archive"] --> ARC DONE --> |extract learnings| SPEC DONE --> |derive opportunities| OPP DONE --> |update references| REF S3["learn"] --> SPEC end OPP --> FR BUILD --> DONE Archiving is not the end β it's a checkpoint. Every shipped feature produces learnings, reveals new gaps, and opens new directions. Knowledge compounds instead of evaporating.
Just say the trigger phrase β the right workflow kicks in automatically.
| Skill | Say this | What happens |
|---|---|---|
| feature-radar | "feature radar", "what should we build next" | Full 6-phase analysis: scan β archive β organize β gap analysis β evaluate β propose |
| scan | "scan opportunities", "brainstorm ideas" | Discover new ideas from multiple sources β opportunities/ |
| archive | "archive feature", "this feature is done" | Move to archive/ + mandatory learning extraction |
| learn | "extract learnings", "save this decision" | Capture patterns & decisions β specs/ |
| ref | "add reference", "interesting approach" | Record external observations β references/ |
| validate | "validate", "check format", "lint skills" | Validate SKILL.md frontmatter + .feature-radar/ files against format rules |
Skill details
The full 6-phase workflow. Analyzes your project, creates .feature-radar/ with base.md (project dashboard), then runs: scan, archive, organize, gap analysis, evaluate, propose. Checkpoints after Phase 1, 3, and 5 let you steer mid-flow.
Modes β pass an argument to run a subset of phases:
| Argument | What it does |
|---|---|
(none) or full | All 6 phases (default) |
quick | Phases 1-3 only β fast scan + archive + organize |
evaluate | Phases 5-6 only β prioritize existing opportunities |
#N (e.g. #2) | Deep-dive a single opportunity β evaluate + propose |
Discover new ideas β from creative brainstorming, user pain points, ecosystem evolution, technical possibilities, or cross-project research. Evaluates each candidate on 6 criteria including value uplift and innovation potential. Deduplicates against existing tracking.
Archive a shipped, rejected, or covered feature. Then runs the mandatory extraction checklist: extract learnings β specs, derive new opportunities, update references, update trends. This is where knowledge compounds β the checklist ensures nothing gets lost.
Capture reusable patterns, architectural decisions, and pitfalls from completed work. Names files by the pattern (e.g., yaml-config-merge.md), not the feature that produced it.
Record external observations β ecosystem trends, creative approaches, research findings, user feedback. Cites sources and dates, assesses implications, suggests new opportunities when relevant.
Checks SKILL.md frontmatter (description β€ 1024 chars, kebab-case name) and .feature-radar/ files against SPEC.md format rules. Reports errors and warnings, then auto-fixes what it can. Runs proactively after any edit to skills or .feature-radar/ files.
Every skill follows the same model β deep understanding before action, structured checkpoints, verified completion:
flowchart TD A[Trigger phrase received] --> B[Deep Read] B --> B1[Read base.md thoroughly] B1 --> B2[Scan existing files] B2 --> B3[State understanding] B3 --> C{Understanding<br/>confirmed?} C -->|No| B C -->|Yes| D[Behavioral Directives<br/>loaded] D --> E[Execute Workflow Steps] E --> F{Important<br/>output?} F -->|Yes| G[Write file +<br/>annotation review] F -->|No| H[Conversational<br/>confirm] G --> I{User annotated?} I -->|Yes| J[Address notes] J --> K{Approved?} I -->|No / approved| L[Continue] K -->|No| J K -->|Yes| L H --> L L --> M{More steps?} M -->|Yes| E M -->|No| N[Completion Summary] - Compound knowledge β Every completed feature feeds back into the system. Learnings accumulate, patterns emerge, decisions are remembered.
- Value-driven β Chase user value and innovation, not feature checklists. Ask "what problem does this solve?" before "what does this do?"
- Honest evaluation β Evaluate fit with YOUR architecture and users, not someone else's roadmap.
- Signal over noise β 1 issue with no comments = weak signal. Multiple independent asks = strong signal.
- Evidence over assumptions β Rank by real demand and creative potential, not hypothetical value.
Any AI agent that supports SKILL.md β Claude Code, Codex, and others.
Skills live in the skills/ directory. Fork, branch, add your skill under skills/{skill-name}/SKILL.md, and submit a PR.
MIT License β see LICENSE file for details.