|
23181 | 23181 | - filename: Qwen3-Grand-Horror-Light-1.7B.Q4_K_M.gguf |
23182 | 23182 | sha256: cbbb0c5f6874130a8ae253377fdc7ad25fa2c1e9bb45f1aaad88db853ef985dc |
23183 | 23183 | uri: huggingface://mradermacher/Qwen3-Grand-Horror-Light-1.7B-GGUF/Qwen3-Grand-Horror-Light-1.7B.Q4_K_M.gguf |
| 23184 | +- !!merge <<: *qwen3 |
| 23185 | + name: "qwen3-vl-235b-a22b-instruct-mxfp4_moe" |
| 23186 | + urls: |
| 23187 | + - https://huggingface.co/noctrex/Qwen3-VL-235B-A22B-Instruct-MXFP4_MOE-GGUF |
| 23188 | + description: | |
| 23189 | + **Model Name:** Qwen3-VL-235B-A22B-Instruct |
| 23190 | + **Model Type:** Vision-Language Model (VLM) |
| 23191 | + **Architecture:** MoE (Mixture of Experts) with 235B parameters |
| 23192 | + **Base Model:** Qwen3-VL-235B-A22B-Instruct (original by Alibaba) |
| 23193 | + **Quantization:** MXFP4_MOE (quantized version by noctrex, not original) |
| 23194 | + **License:** Apache 2.0 |
| 23195 | + |
| 23196 | + --- |
| 23197 | + |
| 23198 | + ### 🌟 Description: |
| 23199 | + |
| 23200 | + Qwen3-VL-235B-A22B-Instruct is a state-of-the-art **vision-language model** developed by Alibaba, designed to understand and generate rich, multimodal content. It combines powerful visual perception with advanced language capabilities, enabling seamless interaction between images, videos, and text. |
| 23201 | + |
| 23202 | + This model supports **long-context reasoning (up to 1M tokens)**, making it ideal for processing books, lengthy documents, and extended video content. It excels in **spatial reasoning, visual coding, OCR across 32 languages**, and **agent-based GUI interaction**, allowing it to perform complex tasks like navigating interfaces or generating code from diagrams. |
| 23203 | + |
| 23204 | + Equipped with **interleaved-MRoPE**, **DeepStack**, and **text-timestamp alignment**, it delivers superior performance in video understanding and fine-grained visual analysis. |
| 23205 | + |
| 23206 | + The **Instruct** variant is optimized for dialogue and task completion, making it suitable for chatbots, intelligent assistants, and multimodal agents. |
| 23207 | + |
| 23208 | + > ⚠️ **Note:** The model hosted at `noctrex/Qwen3-VL-235B-A22B-Instruct-MXFP4_MOE-GGUF` is a **quantized version** (MXFP4_MOE) of the original. The true base model is available at: [Qwen/Qwen3-VL-235B-A22B-Instruct](https://huggingface.co/Qwen/Qwen3-VL-235B-A22B-Instruct). |
| 23209 | + |
| 23210 | + --- |
| 23211 | + |
| 23212 | + ### ✅ Key Features: |
| 23213 | + - 235B-parameter MoE architecture |
| 23214 | + - 256K native context, expandable to 1M tokens |
| 23215 | + - Advanced spatial & video understanding |
| 23216 | + - 32-language OCR with high accuracy |
| 23217 | + - Visual agent capabilities (GUI interaction) |
| 23218 | + - Supports image, video, and text inputs |
| 23219 | + - Optimized for reasoning, coding, and multimodal tasks |
| 23220 | + |
| 23221 | + --- |
| 23222 | + |
| 23223 | + ### 🔗 Resources: |
| 23224 | + - **Original Model:** [Qwen/Qwen3-VL-235B-A22B-Instruct](https://huggingface.co/Qwen/Qwen3-VL-235B-A22B-Instruct) |
| 23225 | + - **Technical Report:** [Qwen3 Technical Report (arXiv)](https://arxiv.org/abs/2505.09388) |
| 23226 | + - **Chat Demo:** [Qwen Chat](https://chat.qwenlm.ai/) |
| 23227 | + |
| 23228 | + --- |
| 23229 | + |
| 23230 | + 📌 *Perfect for researchers, developers, and enterprises building intelligent, multimodal AI systems.* |
| 23231 | + overrides: |
| 23232 | + parameters: |
| 23233 | + model: noctrex/Qwen3-VL-235B-A22B-Instruct-MXFP4_MOE-GGUF |
0 commit comments