Skip to content

Commit d01f1e3

Browse files
mudlergithub-actions[bot]
authored andcommitted
chore(model gallery): 🤖 add new models via gallery agent
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
1 parent 88cb379 commit d01f1e3

File tree

1 file changed

+50
-0
lines changed

1 file changed

+50
-0
lines changed

gallery/index.yaml

Lines changed: 50 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -23181,3 +23181,53 @@
2318123181
- filename: Qwen3-Grand-Horror-Light-1.7B.Q4_K_M.gguf
2318223182
sha256: cbbb0c5f6874130a8ae253377fdc7ad25fa2c1e9bb45f1aaad88db853ef985dc
2318323183
uri: huggingface://mradermacher/Qwen3-Grand-Horror-Light-1.7B-GGUF/Qwen3-Grand-Horror-Light-1.7B.Q4_K_M.gguf
23184+
- !!merge <<: *qwen3
23185+
name: "qwen3-vl-235b-a22b-instruct-mxfp4_moe"
23186+
urls:
23187+
- https://huggingface.co/noctrex/Qwen3-VL-235B-A22B-Instruct-MXFP4_MOE-GGUF
23188+
description: |
23189+
**Model Name:** Qwen3-VL-235B-A22B-Instruct
23190+
**Model Type:** Vision-Language Model (VLM)
23191+
**Architecture:** MoE (Mixture of Experts) with 235B parameters
23192+
**Base Model:** Qwen3-VL-235B-A22B-Instruct (original by Alibaba)
23193+
**Quantization:** MXFP4_MOE (quantized version by noctrex, not original)
23194+
**License:** Apache 2.0
23195+
23196+
---
23197+
23198+
### 🌟 Description:
23199+
23200+
Qwen3-VL-235B-A22B-Instruct is a state-of-the-art **vision-language model** developed by Alibaba, designed to understand and generate rich, multimodal content. It combines powerful visual perception with advanced language capabilities, enabling seamless interaction between images, videos, and text.
23201+
23202+
This model supports **long-context reasoning (up to 1M tokens)**, making it ideal for processing books, lengthy documents, and extended video content. It excels in **spatial reasoning, visual coding, OCR across 32 languages**, and **agent-based GUI interaction**, allowing it to perform complex tasks like navigating interfaces or generating code from diagrams.
23203+
23204+
Equipped with **interleaved-MRoPE**, **DeepStack**, and **text-timestamp alignment**, it delivers superior performance in video understanding and fine-grained visual analysis.
23205+
23206+
The **Instruct** variant is optimized for dialogue and task completion, making it suitable for chatbots, intelligent assistants, and multimodal agents.
23207+
23208+
> ⚠️ **Note:** The model hosted at `noctrex/Qwen3-VL-235B-A22B-Instruct-MXFP4_MOE-GGUF` is a **quantized version** (MXFP4_MOE) of the original. The true base model is available at: [Qwen/Qwen3-VL-235B-A22B-Instruct](https://huggingface.co/Qwen/Qwen3-VL-235B-A22B-Instruct).
23209+
23210+
---
23211+
23212+
### ✅ Key Features:
23213+
- 235B-parameter MoE architecture
23214+
- 256K native context, expandable to 1M tokens
23215+
- Advanced spatial & video understanding
23216+
- 32-language OCR with high accuracy
23217+
- Visual agent capabilities (GUI interaction)
23218+
- Supports image, video, and text inputs
23219+
- Optimized for reasoning, coding, and multimodal tasks
23220+
23221+
---
23222+
23223+
### 🔗 Resources:
23224+
- **Original Model:** [Qwen/Qwen3-VL-235B-A22B-Instruct](https://huggingface.co/Qwen/Qwen3-VL-235B-A22B-Instruct)
23225+
- **Technical Report:** [Qwen3 Technical Report (arXiv)](https://arxiv.org/abs/2505.09388)
23226+
- **Chat Demo:** [Qwen Chat](https://chat.qwenlm.ai/)
23227+
23228+
---
23229+
23230+
📌 *Perfect for researchers, developers, and enterprises building intelligent, multimodal AI systems.*
23231+
overrides:
23232+
parameters:
23233+
model: noctrex/Qwen3-VL-235B-A22B-Instruct-MXFP4_MOE-GGUF

0 commit comments

Comments
 (0)