Wanwu AI Agent Platform is an enterprise-grade one-stop commercially friendly AI agent development platform designed for business scenarios. It is committed to providing enterprises with a safe, efficient, and compliant one-stop AI solution. With the core philosophy of "technology openness and ecological co-construction", we integrate cutting-edge technologies such as large language models and business process automation to build an AI engineering platform with a complete functional system covering model full life-cycle management, MCP, web search, AI agent rapid development, enterprise knowledge base construction, and complex workflow orchestration. The platform adopts a modular architecture design, supports flexible functional expansion and secondary development, and greatly reduces the application threshold of AI technology while ensuring the security and privacy protection of enterprise data. Whether it is for small and medium-sized enterprises to quickly build intelligent applications or for large enterprises to achieve intelligent transformation of complex business scenarios, the Wanwu AI Agent Platform can provide strong technical support to help enterprises accelerate the process of digital transformation, achieve cost reduction and efficiency improvement, and business innovation.
🔥 Adopt a permissive and friendly Apache 2.0 License, supporting developers to freely expand and develop secondary
✔ Enterprise-level engineering: Provides a complete toolchain from model management to application landing, solving the "last mile" problem of LLM technology landing
✔ Open-source ecological: Adopt a permissive and friendly Apache 2.0 License, supporting developers to freely expand and develop
✔ Full-stack technology support: Equipped with a professional team to provide architecture consulting, performance optimization and full-cycle empowerment for ecological partners
✔ Multi-tenant architecture: Provides a multi-tenant account system to meet the core needs of users in cost control, data security isolation, business elasticity expansion, industry customization, rapid online and ecological collaboration
✔ XinChuang adaptation: Already adapted to domestic XinChuang databases TiDB and OceanBase
▸ Supports the unified access and lifecycle management of hundreds of proprietary/open-source large models (including GPT, Claude, Llama, etc.)
▸ Deeply adapts to OpenAI API standards and Unicom Yuanjing ecological models, realizing seamless switching of heterogeneous models
▸ Provides multi-inference backend support (vLLM, TGI, etc.) and self-hosted solutions to meet the computing power needs of enterprises of different scales
▸ Standardized interfaces: Enable AI models to seamlessly connect to various external tools (such as GitHub, Slack, databases, etc.) without the need to develop adapters for each data source separately
▸ Built-in rich and selected recommendations: Integrates 100+ industry MCP interfaces, making it easy for users to call up quickly and easily
▸ Real-time information acquisition: Possesses powerful web search capabilities, capable of obtaining the latest information from the Internet in real-time. In question and answer scenarios, when a user's question requires the latest news, data, and other information, the platform can quickly search and return accurate results, enhancing the timeliness and accuracy of the answers
▸ Multi-source data integration: Integrates various Internet data sources, including news websites, academic databases, industry reports, etc. Through the integration and analysis of multi-source data, it provides users with more comprehensive and in-depth information. For example, in market research scenarios, relevant data can be obtained from multiple data sources at the same time for comprehensive analysis and evaluation
▸ Intelligent search strategy: Adopt intelligent search algorithms, automatically optimize search strategies based on user questions to improve search efficiency and accuracy. Support keyword search, semantic search and other search methods to meet the needs of different users. At the same time, intelligently sort and filter search results, prioritize the display of the most relevant and valuable information
▸ Quickly build complex AI business processes through low-code drag-and-drop canvas
▸ Built-in conditional branching, API, large model, knowledge base, code, MCP and other nodes, support end-to-end process debugging and performance analysis
▸ Provides the whole process knowledge management capabilities of knowledge base creation → document parsing → vectorization → retrieval → fine sorting, supports multiple formats such as pdf/docx/txt/xlsx/csv/pptx documents, and also supports the capture and access of web resources
▸ Integrates multi-modal retrieval, cascading segmentation and adaptive segmentation, significantly improves the accuracy of Q&A
▸ Can be based on the function call (Function Calling) agent construction paradigm, supports tool expansion, private knowledge base association and multi-round dialogue
▸ Support online debugging
▸ Provides RESTful API, supports deep integration with existing enterprise systems (OA/CRM/ERP, etc.)
▸ Provides fine-grained permission control to ensure stable operation in production environments
Function | Wanwu | Dify.AI | Fastgpt | Ragflow | Coze open source version |
---|---|---|---|---|---|
Model import | ✅ | ✅ | ❌(Built-in models) | ✅ | ❌(Built-in models) |
RAG engine | ✅ | ✅ | ✅ | ✅ | ✅ |
MCP | ✅ | ✅ | ✅ | ✅(Need to install tools to use) | ❌ |
Direct OCR import | ✅ | ❌ | ❌ | ❌ | ❌ |
Search enhancement | ✅ | ✅(Need to install tools to use) | ✅ | ✅(Need to install tools to use) | ✅ |
Agent | ✅ | ✅ | ✅ | ✅ | ✅ |
Workflow | ✅ | ✅ | ✅ | ✅ | ✅ |
Local deployment | ✅ | ✅ | ✅ | ✅ | ✅ |
license friendly | ✅ | ❌(Commercially restricted) | ❌(Commercially restricted) | Not fully open source | ✅ |
Multi-tenant | ✅ | ❌(Commercially restricted) | ❌(Commercially restricted) | ✅ | ✅(Users are not interconnected) |
As of August 1, 2025.
- Intelligent Customer Service: Realize high-accuracy business consultation and ticket processing based on RAG + Agent
- Knowledge Management: Build an exclusive enterprise knowledge base, support semantic search and intelligent summary generation
- Process Automation: Realize AI-assisted decision-making for business processes such as contract review and reimbursement approval through the workflow engine
The platform has been successfully applied in multiple industries such as finance, industry, and government, helping enterprises transform the theoretical value of LLM technology into actual business benefits. We sincerely invite developers to join the open source community and jointly promote the democratization of AI technology.
-
The workflow module of the Wanwu AI Agent Platform uses the following project, you can go to its warehouse to view the details.
- v0.1.8 and earlier: wanwu-agentscope project
- v0.2.0 and later: wanwu-workflow project
-
Docker Installation (Recommended)
-
Before the first run
1.1 Copy the environment variable file
cp .env.bak .env
1.2 Modify the
WANWU_ARCH
andWANWU_EXTERNAL_IP
variables in the .env file according to the system# amd64 / arm64 WANWU_ARCH=amd64 # external ip port (Note: if the browser accesses Wanwu deployed on a non-localhost server, you need to change localhost to the external IP, for example, 192.168.xx.xx) WANWU_EXTERNAL_IP=localhost
1.3 Configure the
WANWU_BFF_JWT_SIGNING_KEY
variable in the .env file, a custom complex random string used for generating JWT tokens# bff WANWU_BFF_JWT_SIGNING_KEY=
1.4 Create a Docker running network
docker network create wanwu-net
-
Start the service (the image will be automatically pulled from Docker Hub during the first run)
# For amd64 system: docker compose --env-file .env --env-file .env.image.amd64 up -d # For arm64 system: docker compose --env-file .env --env-file .env.image.arm64 up -d
-
Log in to the system: http://localhost:8081
Default user: admin Default password: Wanwu123456
-
Stop the service
# For amd64 system: docker compose --env-file .env --env-file .env.image.amd64 down # For arm64 system: docker compose --env-file .env --env-file .env.image.arm64 down
- Source Code Start (Development)
-
Based on the above Docker installation steps, start the system service completely
-
Take the backend bff-service service as an example
2.1 Stop bff-service
make -f Makefile.develop stop-bff
2.2 Compile the bff-service executable file
# For amd64 system: make build-bff-amd64 # For arm64 system: make build-bff-arm64
2.3 Start bff-service
make -f Makefile.develop run-bff
-
Based on the above Docker installation steps, completely stop the system service
-
Update to the latest version of the code
2.1 In the wanwu repository directory, update the code
# Switch to the main branch git checkout main # Pull the latest code git pull
2.2 Recopy the environment variable file (if there are changes to the environment variables, please modify them again)
# Backup the current .env file cp .env .env.old # Copy the .env file cp .env.bak .env
-
Based on the above Docker installation steps, completely start the system service
To help you quickly get started with this project, we strongly recommend that you first check out the Documentation Operation Manual. We provide users with interactive and structured operation guides, where you can directly view operation instructions, interface documents, etc., greatly reducing the threshold for learning and use. The detailed function list is as follows:
Feature | Detailed Description |
---|---|
Model Management | Supports users to import LLM, Embedding, and Rerank models from various model providers, including Unicom Yuanjing, OpenAI-API-compatible, Ollama, Tongyi Qianwen, and Volcano Engine. Model Import Methods - Detailed Version |
Knowledge Base | In terms of document parsing capabilities: supports uploading of 12 file types and URL parsing; document parsing methods include OCR and high-precision model parsing (titles/tables/formulas); document segmentation settings support both general segmentation and parent-child segmentation. In terms of optimization capabilities: supports metadata management and metadata filtering queries, supports adding, deleting, and modifying segmented content, supports setting keyword tags for segments to improve recall performance, supports segment enable/disable operations, and supports hit testing. In terms of retrieval capabilities: supports multiple retrieval modes including vector search, full-text search, and hybrid search. In terms of Q&A capabilities: supports automatic citation of sources and generating answers with both text and images.<br |
Resource Library | Supports importing your own MCP services or custom tools for use in workflows and agents. |
Safety Guardrails | Users can create sensitive word lists to control the safety of the model's output. |
Text Q&A | A dedicated knowledge advisor based on a private knowledge base. It supports features like knowledge base management, Q&A, knowledge summarization, personalized parameter configuration, safety guardrails, and retrieval configuration to improve the efficiency of knowledge management and learning. Supports publishing text Q&A applications publicly or privately, and can be published as an API. |
Workflow | Extends the capabilities of agents. Composed of nodes, it provides a visual workflow editor. Users can orchestrate multiple different workflow nodes to implement complex and stable business processes. Supports publishing workflow applications publicly or privately, can be published as an API, and supports import/export. |
Agent | Create agents based on user scenarios and business requirements. Supports model selection, prompt setting, web search, knowledge base selection, MCP, workflows, and custom tools. Supports publishing agent applications publicly or privately, and can be published as an API and a Web URL. |
App Marketplace | Allows users to experience published applications, including Text Q&A, Workflows, and Agents. |
MCP Hub | Features 100+ pre-selected industry-specific MCP servers, ready for immediate use. |
Settings | The platform supports multi-tenancy, allowing users to manage organizations, roles, users, and perform basic platform configuration. |
- Multi-modal model access
- Support custom MCP Server, which means that workflows, agents, or APIs that conform to the OpenAPI specification can be added to the MCP Server for release
- Knowledge base sharing
- Agent and model evaluation
- Agent monitoring statistics
- Model experience
- Prompt engineering
-
[Q] Error when starting Elastic (elastic-wanwu) on Linux system: Memory limited without swap. [A] Stop the service, run
sudo sysctl -w vm.max_map_count=262144
, and then restart the service. -
[Q] After the system services start normally, the mysql-wanwu-setup and elastic-wanwu-setup containers exit with status code Exited (0). [A] This is normal. These two containers are used to complete some initialization tasks and will automatically exit after execution.
-
[Q] Regarding model import [A] Taking the import of Unicom Yuanjing LLM as an example (the process is similar for importing OpenAI-API-compatible models, Embedding, or Rerank types):
1. The Open API interface for Unicom Yuanjing MaaS Cloud LLM is, for example: https://maas.ai-yuanjing.com/openapi/compatible-mode/v1/chat/completions 2. The API Key applied for by the user on Unicom Yuanjing MaaS Cloud looks like: sk-abc********************xyz 3. Confirm that the API and Key can correctly request the LLM. Taking a request to yuanjing-70b-chat as an example: curl --location 'https://maas.ai-yuanjing.com/openapi/compatible-mode/v1/chat/completions' \ --header 'Content-Type: application/json' \ --header 'Accept: application/json' \ --header 'Authorization: Bearer sk-abc********************xyz' \ --data '{ "model": "yuanjing-70b-chat", "messages": [{ "role": "user", "content": "你好" }] }' 4. Import the model: 4.1 [Model Name] must be the model that can be correctly requested in the curl command above; for example, yuanjing-70b-chat. 4.2 [API Key] must be the key that can be correctly requested in the curl command above; for example, sk-abc********************xyz (note: do not include the 'Bearer' prefix). 4.3 [Inference URL] must be the URL that can be correctly requested in the curl command above; for example, https://maas.ai-yuanjing.com/openapi/compatible-mode/v1 (note: do not include the /chat/completions suffix). 5. Importing an Embedding model is the same as importing an LLM as described above. Note that the inference URL should not include the /embeddings suffix. 6. Importing a Rerank model is the same as importing an LLM as described above. Note that the inference URL should not include the /rerank suffix.
The Yuanjing Wanwu AI Agent Platform is released under the Apache License 2.0.
QQ Group1(Full):490071123 | QQ Group2:1026898615 |
---|---|
![]() |
![]() |