Skip to content

tinycrops/PlanExe

 
 

Repository files navigation

PlanExe

PlanExe - Turn your idea into a comprehensive plan in minutes, not months.

Turn your idea into a comprehensive plan in minutes, not months.


Example plans generated with PlanExe

What is PlanExe?

PlanExe is an open-source tool, that turns a single plain-english goal statement into a 40-page, strategic plan in ~15 minutes using a local or cloud models. It's an accelerator for outlines, but no silver bullet for polished plans.

Typical output contains:

  • Executive summary
  • Gantt chart
  • Governance structure
  • Role descriptions
  • Stakeholder maps
  • Risk registers
  • SWOT analyses

The technical quality of structure, formatting, and coherence is consistently excellent—often superior to human junior/mid-tier consulting drafts. However, budgets remain headline-only, timelines contain errors, metrics are usually vague, and legal/operational realism is weak on high-stakes topics. A usable, client-ready version still requires weeks to months of skilled human refinement.

PlanExe removes 70–90 % of the labor for the planning scaffold on any topic, but the final 10–30 % that separates a polished document from a credible, defensible plan remains human-only work.


New to PlanExe? Follow the Getting Started guide.

Try it out now (Click to expand)

You can generate 1 plan for free.

Try it here →


Run locally with Docker (Click to expand)

Prerequisite: Docker with Docker Compose installed; you only need basic Docker knowledge. No local Python setup is required because everything runs in containers.

Quickstart: single-user UI + worker (frontend_single_user + worker_plan)

  1. Clone the repo and enter it:
git clone https://github.com/neoneye/PlanExe.git
cd PlanExe
  1. Provide an LLM provider. Copy .env.example to .env and fill in OPENROUTER_API_KEY with your key from OpenRouter. The containers mount .env and llm_config.json; pick a model profile there. For host-side Ollama, use the docker-ollama-llama3.1 entry and ensure Ollama is listening on http://host.docker.internal:11434.

  2. Start the stack (first run builds the images):

docker compose up worker_plan frontend_single_user

The worker listens on http://localhost:8000 and the UI comes up on http://localhost:7860 after the worker healthcheck passes.

  1. Open http://localhost:7860 in your browser. Optional: set PLANEXE_PASSWORD in .env to require a password. Enter your idea, click the generate button, and watch progress with:
docker compose logs -f worker_plan

Outputs are written to run/ on the host (mounted into both containers).

  1. Stop with Ctrl+C (or docker compose down). Rebuild after code/dependency changes:
docker compose build --no-cache worker_plan frontend_single_user

For compose tips, alternate ports, or troubleshooting, see extra/docker.md or docker-compose.md.

Configuration

Config A: Run a model in the cloud using a paid provider. Follow the instructions in OpenRouter.

Config B: Run models locally on a high-end computer. Follow the instructions for either Ollama or LM Studio. When using host-side tools with Docker, point the model URL at the host (for example http://host.docker.internal:11434 for Ollama).

Recommendation: I recommend Config A as it offers the most straightforward path to getting PlanExe working reliably.


Screenshots (Click to expand)

You input a vague description of what you want and PlanExe outputs a plan.

YouTube video: Using PlanExe to plan a lunar base

Screenshot of PlanExe


Help (Click to expand)

For help or feedback.

Join the PlanExe Discord.

About

AI planner similar to OpenAI's deep research

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 92.9%
  • HTML 6.9%
  • Dockerfile 0.2%