Roadmap
Roadmap
Where Prompt2PR is headed. This roadmap reflects the original product vision and community feedback. Items are grouped by phase — not by timeline.
Current state: Prompt2PR v1 is a single-shot action. One LLM call, one response, one PR. No iteration, no tool use. Internet access depends on the model and provider.
Completed (v1.0 - v1.1)
Everything in the MVP and early Growth phases is shipped:
- Core pipeline: prompt + file context -> LLM -> PR
- 4 LLM providers: Mistral, OpenAI, Anthropic, GitHub Models
- Safety guardrails:
max_files,max_changes, path scoping,.github/block - Dry-run mode
- AI-generated PR summaries and descriptive titles
- Configurable labels, branch prefix, base URL override
- 13 example workflows across 4 categories
- GitHub Pages documentation site
- 97%+ test coverage (274 tests)
Phase 2 — Growth
Near-term improvements that build on the current single-shot architecture.
PR Deduplication
Skip PR creation if an identical open PR already exists. Prevents duplicate PRs when a scheduled workflow runs multiple times before the previous PR is merged.
Auto-Assign Reviewers
Automatically assign reviewers or teams to generated PRs. Configuration via a
new reviewers input:
with:
reviewers: 'alice,bob'
team_reviewers: 'platform-team'
LiteLLM Proxy Mode
Support any LLM provider via a LiteLLM proxy. Users who run a LiteLLM gateway can point Prompt2PR at it:
with:
provider: litellm
model: mistral/mistral-large-latest
base_url: http://my-litellm-proxy:4000
Bootstrap CLI
A command-line tool or companion action that scaffolds Prompt2PR workflows:
npx prompt2pr init
# Interactive prompt: What do you want to automate?
# Generates .github/workflows/prompt2pr.yml
More Providers
- Google Gemini
- AWS Bedrock
- Azure OpenAI
- Ollama (local models)
Phase 3 — Agentic Mode
The biggest architectural evolution. Instead of a single LLM call, Prompt2PR would run an agent loop that can use tools between LLM calls.
How It Would Work
- The action calls the LLM with the prompt + file context + available tools
- The LLM responds with a tool call (e.g.,
read_file,run_tests,search_codebase) - The action executes the tool and sends the result back to the LLM
- Loop until the LLM returns a final answer (the file changes)
- Submit the PR as usual
Why This Matters
The single-shot model has a fundamental limitation: the LLM can only reason about files it was given upfront. It cannot:
- Explore the codebase to find related files
- Run tests to verify its changes work
- Iterate on its output based on feedback
An agentic mode would unlock tasks that are currently impossible:
| Task | Single-shot | Agentic |
|---|---|---|
| Add JSDoc to visible files | Works well | Works well |
| Fix a bug described in an issue | Limited | Can explore codebase |
| Refactor across many files | Limited by context | Can navigate freely |
| Fix failing tests | Cannot run tests | Can run and iterate |
What the API Could Look Like
- uses: davd-gzl/Prompt2PR@v2
with:
prompt: 'Fix all failing tests in src/'
provider: openai
mode: 'agentic' # vs 'single-shot' (default)
max_iterations: 10 # safety cap on agent loops
tools: 'read_file,run_command' # which tools to expose
Provider Support
All major providers already support tool use / function calling:
| Provider | API | Mechanism |
|---|---|---|
| OpenAI | Responses API | tool_calls in response, loop on requires_action |
| Anthropic | Messages API | tool_use content blocks, return tool_result |
| Mistral | Chat API | tool_calls field, same loop pattern |
| GitHub Models | Inherited | Depends on upstream model provider |
Trade-offs
| Single-shot (current) | Agentic | |
|---|---|---|
| Predictability | High | Lower — could loop 2x or 20x |
| Cost | Fixed, proportional to context | Variable, needs caps |
| Capability | Reasons about given files only | Can explore, test, verify |
| Latency | Seconds | Minutes |
| Complexity | Simple | Significantly more complex |
| Safety | Easy to guardrail | Harder — agent takes actions |
Single-shot would remain the default. Agentic mode would be opt-in.
Phase 4 — Structured Prompts & DSL
Evolve from plain-English prompts to a richer declarative language for precision and reusability.
Structured Prompts
Break prompts into semantic parts: task, scope, and rules:
with:
prompt:
task: 'Update copyright year to 2026'
scope: 'All files containing copyright notices'
rules:
- 'Use range format: 2024-2026'
- 'Do not modify files in vendor/'
- 'Preserve original author names'
Conditional Execution
Control when prompts actually run, beyond cron schedules:
with:
prompt: 'Update copyright year'
only_if: 'january' # Only run in January
Chained Prompts
Multi-step workflows where one prompt’s output feeds the next:
jobs:
step1:
uses: davd-gzl/Prompt2PR@v2
with:
prompt: 'Identify all deprecated API calls'
dry_run: true
step2:
needs: step1
uses: davd-gzl/Prompt2PR@v2
with:
prompt:
'Replace the deprecated calls found: $'
Phase 5 — Community & Scale
Community Prompt Templates
Reference shared, curated prompts by name instead of writing your own:
with:
prompt: 'community/dead-link-fixer'
A public registry of prompts that the community can browse, share, and rate.
Prompt Marketplace
A GitHub Pages-hosted gallery where users can:
- Browse prompts by category (code quality, docs, security, maintenance)
- See real-world results (example PRs generated by each prompt)
- Rate and review prompts
- Submit their own
Cross-Repo Dashboard
A companion GitHub App or dashboard that shows:
- All Prompt2PR workflows across an organization
- PR creation rate, merge rate, rejection rate per prompt
- Cost tracking (estimated API usage per workflow)
Self-Improving Prompts
Learn from rejected PRs to improve future runs:
- Track which PRs get merged vs. closed
- Analyze rejection patterns (what kinds of changes get rejected?)
- Suggest prompt refinements based on historical outcomes
- Optionally auto-adjust prompts over time
Organization-Wide Prompt Policies
For teams managing Prompt2PR across many repositories:
- Central prompt library with org-level governance
- Mandatory guardrail policies (e.g., “all repos must use max_files <= 5”)
- Audit log of all LLM-generated changes across the org
- Role-based access to prompt creation and modification
Contributing to the Roadmap
Have an idea? Open an issue or start a discussion. Community input shapes what gets built next.