Qwen3.5-9b-Sushi-Coder-RL-GGUF

Qwen3.5-9b-Sushi-Coder-RL-GGUF

Lineage

Training

The upstream SFT model was trained with Unsloth on:

The RL stage was then run for coding with NousResearch/hermes-agent using NousResearch/atropos.

During that run, vLLM was patched with vllm-project/vllm PR #36395, fix(lora): add bounds checking for TP configurations, to address the LoRA tensor-parallel bounds issue.

Files

  • Qwen3.5-9b-Sushi-Coder-RL.Q4_K_M.gguf
  • Qwen3.5-9b-Sushi-Coder-RL.Q8_0.gguf
  • Qwen3.5-9b-Sushi-Coder-RL.BF16-mmproj.gguf

Usage Note

This is a multimodal Qwen 3.5 export. Use the text GGUF together with the BF16-mmproj file.

Quick Start

Example download commands with the Hugging Face CLI:

hf download bigatuna/Qwen3.5-9b-Sushi-Coder-RL-GGUF \
  Qwen3.5-9b-Sushi-Coder-RL.Q4_K_M.gguf \
  Qwen3.5-9b-Sushi-Coder-RL.BF16-mmproj.gguf

Alternative quant:

hf download bigatuna/Qwen3.5-9b-Sushi-Coder-RL-GGUF \
  Qwen3.5-9b-Sushi-Coder-RL.Q8_0.gguf \
  Qwen3.5-9b-Sushi-Coder-RL.BF16-mmproj.gguf

Metadata

  • License: Apache-2.0
  • Architecture: Qwen 3.5
  • Format: GGUF
  • Tags: llama.cpp, qwen3_5, multimodal, code, rl, conversational
Downloads last month
3,608
GGUF
Model size
9B params
Architecture
qwen35
Hardware compatibility
Log In to add your hardware

4-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for bigatuna/Qwen3.5-9b-Sushi-Coder-RL-GGUF

Quantized
(2)
this model

Datasets used to train bigatuna/Qwen3.5-9b-Sushi-Coder-RL-GGUF