Qwen3.5-9b-Sushi-Coder-RL-GGUF
Lineage
- Base model lineage:
bigatuna/Qwen3.5-9b-Sushi-Coder - RL model:
bigatuna/Qwen3.5-9b-Sushi-Coder-RL - RL pipeline: NousResearch/atropos
Training
The upstream SFT model was trained with Unsloth on:
The RL stage was then run for coding with NousResearch/hermes-agent using NousResearch/atropos.
During that run, vLLM was patched with vllm-project/vllm PR #36395, fix(lora): add bounds checking for TP configurations, to address the LoRA tensor-parallel bounds issue.
Files
Qwen3.5-9b-Sushi-Coder-RL.Q4_K_M.ggufQwen3.5-9b-Sushi-Coder-RL.Q8_0.ggufQwen3.5-9b-Sushi-Coder-RL.BF16-mmproj.gguf
Usage Note
This is a multimodal Qwen 3.5 export. Use the text GGUF together with the BF16-mmproj file.
Quick Start
Example download commands with the Hugging Face CLI:
hf download bigatuna/Qwen3.5-9b-Sushi-Coder-RL-GGUF \
Qwen3.5-9b-Sushi-Coder-RL.Q4_K_M.gguf \
Qwen3.5-9b-Sushi-Coder-RL.BF16-mmproj.gguf
Alternative quant:
hf download bigatuna/Qwen3.5-9b-Sushi-Coder-RL-GGUF \
Qwen3.5-9b-Sushi-Coder-RL.Q8_0.gguf \
Qwen3.5-9b-Sushi-Coder-RL.BF16-mmproj.gguf
Metadata
- License: Apache-2.0
- Architecture: Qwen 3.5
- Format: GGUF
- Tags:
llama.cpp,qwen3_5,multimodal,code,rl,conversational
- Downloads last month
- 3,608
Hardware compatibility
Log In to add your hardware
4-bit
8-bit
Model tree for bigatuna/Qwen3.5-9b-Sushi-Coder-RL-GGUF
Base model
bigatuna/Qwen3.5-9b-Sushi-Coder