White Collar Agent is your Personal AI Assistant that lives inside your machine and works 24/7 for you.
It can autonomously interpret tasks, plan actions, and execute actions to achieve complex goals.
This is an open-source project and is still in development, so we welcome any suggestions, contributions, and feedback! You are free to use, host, and monetize this project (with credit given in case of distribution and monetization).
- 🧠 Single Base Agent Architecture — Simple, extendable core that handles reasoning, planning, and execution.
- ⚙️ CLI/GUI mode — Agent can switch between CLI and GUI mode according to the complexity of the task. GUI mode is still in experimental phase 🧪.
- 🧩 Subclass & Extend — Build your own agents by inheriting from the base class.
- 🔍 Task Document Interface — Define structured tasks for the agent to perform in-context learning.
- 🧰 Actions Library — Reusable tools (web search, code execution, I/O, etc.).
- ⚡ Lightweight & Cross-Platform — Works seamlessly across Linux and Windows.
- 💻 Support multiple LLM providers — Bring your own API keys (Anthropic, OpenAI, Gemini, BytePlus, or even running your own Ollama endpoint).
Important
Note for GUI mode: The GUI mode is still in experimental phase. This means you will encounter a lot of issues when the agent decides to switch to GUI mode. We are still working on it.
- Memory Module — Coming next!
- External Tool integration — Pending
- MCP Layer — Pending
- Proactive Behaviour — Pending
- Python 3.9+
git,conda, andpip- An API key for your chosen LLM provider (e.g., OpenAI or Gemini)
git clone https://github.com/zfoong/WhiteCollarAgent.git
cd WhiteCollarAgent
conda env create -f environment.ymlExport your API key:
export OPENAI_API_KEY=<YOUR_KEY_HERE>
or
export GOOGLE_API_KEY=<YOUR_KEY_HERE>Run:
python start.pyThis executes the built-in White Collar Agent, that you can communicate to:
- Talk to the agent
- Ask it to perform complex series of tasks
- Run command /help to seek help
- Get along with the AI agent
- Do advanced computer-use tasks with a dedicated but lightweight WebRTC Linux VM
| Argument | Description |
|---|---|
--only-cpu |
Run the agent on CPU mode |
--fast |
Skip unecessary update checks and launch agent faster. NOTE: You have to run without --fast the first time you launch |
--no-omniparser |
Disable the use of microsoft omniparser to analyse UI - will greatly reduce GUI action accuracy. It is highly encouraged to use omniparser |
--no-conda |
Installs all packages globally instead of inside a conda environment |
EXAMPLE
python start.py --only-cpu --fastThe repository root included a Docker configuration with Python 3.10, key system packages (including Tesseract for OCR), and all Python dependencies defined in environment.yml/requirements.txt so the agent can run consistently in isolated environments.
Below are the setup instruction of running our agent with container.
From the repository root:
docker build -t white-collar-agent .The image is configured to launch the agent with python -m core.main by default. To run it interactively:
docker run --rm -it white-collar-agentIf you need to supply environment variables, pass an env file (for example, based on .env.example):
docker run --rm -it --env-file .env white-collar-agentMount any directories that should persist outside the container (such as data or cache folders) using -v, and adjust ports or additional flags as needed for your deployment. The container ships with system dependencies for OCR (tesseract), screen automation (pyautogui, mss, X11 utilities, and a virtual framebuffer), and common HTTP clients so the agent can work with files, network APIs, and GUI automation inside the container.
GUI actions (mouse/keyboard events, screenshots) require an X11 server. You can either attach to your host display or run headless with xvfb:
-
Use the host display (requires Linux with X11):
docker run --rm -it -e DISPLAY=$DISPLAY \ -v /tmp/.X11-unix:/tmp/.X11-unix \ -v $(pwd)/data:/app/core/data \ white-collar-agent
Add extra
-vmounts for any folders the agent should read/write. -
Run headlessly with a virtual display:
docker run --rm -it --env-file .env white-collar-agent bash -lc "Xvfb :99 -screen 0 1920x1080x24 & export DISPLAY=:99 && exec python -m core.main"
By default the image uses Python 3.10 and bundles the Python dependencies from environment.yml/requirements.txt, so python -m core.main works out of the box.
You can easily create your own specialized agent by extending the base agent:
import asyncio
from core.agent_base import AgentBase
class MyCustomAgent(AgentBase):
def __init__(
self,
*,
data_dir: str = "core/data",
chroma_path: str = "./chroma_db",
):
super().__init__(
data_dir=data_dir,
chroma_path=chroma_path,
)
# Your implementation
def _generate_role_info_prompt(self) -> str:
"""
Defines this agent's role, behaviour, and purpose.
"""
return (
"You are MyCustomAgent — an intelligent research assistant. "
"Your role is to find, summarize, and synthesize information from multiple sources. "
"You respond concisely, prioritize factual accuracy, and cite sources when relevant. "
"If you cannot find something, you explain why and suggest alternatives."
)
agent = MyCustomAgent(
data_dir=os.getenv("DATA_DIR", "core/data"),
chroma_path=os.getenv("CHROMA_PATH", "./chroma_db"),
)
asyncio.run(agent.run())Here, you’re reusing all the core planning, reasoning, and execution logic —
just plugging in your own personality, actions, and task documents.
| Component | Description |
|---|---|
| BaseAgent | The core reasoning and execution engine — can be subclassed or used directly. |
| Action / Tool | Reusable atomic functions (e.g., web search, API calls, file ops). |
| Task Document | Describes what the agent must achieve and how. |
| Planner / Executor | Handles goal decomposition, script generation, and execution. |
| LLM Wrapper | Unified layer for model interactions (OpenAI, Gemini, etc.). |
Contributions and suggestions are welcome! You can contact @zfoong @ thamyikfoong(at)craftos.net. We currently don't have checks set up, so we can't allow direct contributions but we appreciate any suggestions and feedback.
This project is licensed under the MIT License. You are free to use, host, and monetize this project (you must credit this project in case of distribution and monetization).
Developed and maintained by CraftOS and contributors @zfoong and @ahmad-ajmal.
If you find White Collar Agent useful, please ⭐ the repository and share it with others!
